Iām not advocating in either direction, btw. I havenāt made up my mind yet. š Just braindumping here.
The (replyto:ā¦) proposal is definitely more in the spirit of twtxt, Iād say. Itās much simpler, anyone can use it even with the simplest tools, no need for any client code. That is certainly a great property, if you ask me, and itās things like that that brought me to twtxt in the first place.
Iād also say that in our tiny little community, message integrity simply doesnāt matter. Signed feeds donāt matter. I signed my feed for a while using GPG, someone else did the same, but in the end, nobody cares. The community is so tiny, thereās enough āimplicit trustā or whatever you want to call it.
If twtxt/Yarn was to grow bigger, then this would become a concern again. But even Mastodon allows editing, so how much of a problem can it really be? š
I do have to āadmitā, though, that hashes feel better. It feels good to know that we can clearly identify a certain twt. It feels more correct and stable.
Hm.
I suspect that the (replyto:ā¦) proposal would work just as well in practice.
@prologic@twtxt.net I saw those, yes. I tried using yarnc, and it would work for a simple twtxt. Now, for a more convoluted one it truly becomes a nightmare using that tool for the job. I know there are talks about changing this hash, so this might be a moot point right now, but it would be nice to have a tool that:
- Would calculate the hash of a twtxt in a file.
- Would calculate all hashes on a
twtxt.txt(local and remote).
Again, something lovely to have after any looming changes occur.
Could someone knowledgable reply with the steps a grandpa will take to calculate the hash of a twtxt from the CLI, using out-of-the-box tools? I swear I read about it somewhere, but canāt find it.
@bender@twtxt.net thatās not your change, silly robot, it is mine! LOL. I am finding @prologic@twtxt.netās tool handy to refer to previous posts (as reference, for example).
@aelaraji@aelaraji.com Btw, Iām also open to ideas for this tool and welcome any contributions š
@prologic@twtxt.net wellā¦
how would that work exactly?
To my limited knowledge, Keyoxide is an open source project offering different tools for verifying oneās online persona(s). Thatās done by either A) creating an Ariande Profile using the web interface, a CLI. or B) Just using your GPG key. Either way, you add in Identity claims to your different profiles, links and whatnot, and finally advertise your profile ⦠Then there is a second set of Mobile/Web clients and CLI your correspondents can use to check your identity claims. I think of them like the front-ends of GPG Keyservers (which keyoxide leverages for verification when you opt for the GPG Key method), where you verify profiles using links, Key IDs and Fingerprintsā¦
Who maintains cox site? Is it centralized or decentralized can be relied upon?
- Maintainers? Definitely not me, but hereās their Git stuff and OpenCollective page ā¦
- Both ASP and Keyoxide Webtools can be self-hosted. I donāt see a central authority here⦠+ As mentioned on their FAQ page the whole process can be done manually, so you donāt have to relay on any one/thing if you donāt want to, the whole thing is just another tool for convenience (with a bit of eye candy).
Does that mean then that every user is required to have a cox side profile?
Nop. But it looks like a nice option to prove that Iām the same person to whom that may concern if I ever change my Twtxt URL, host/join a yarn pod or if I reach out on other platforms to someone Iāve met in her. Otherwise Iām just happy exchanging GPG keys or confirm the change IRL at a coffee shop or something. š
@mckinley@twtxt.net To answer some of your questions:
Are SSH signatures standardized and are there robust software libraries that can handle them? Weāll need a library in at least Python and Go to provide verified feed support with the currently used clients.
We already have this. Ed25519 libraries exist for all major languages. Aside from using ssh-keygen -Y sign and ssh-keygen -Y verify, you can also use the salty CLI itself (https://git.mills.io/prologic/salty), and Iām sure there are other command-line tools that could be used too.
If we all implemented this, every twt hash would suddenly change and every conversation thread weāve ever had would at least lose its opening post.
Yes. This would happen, so weād have to make a decision around this, either a) a cut-off point or b) some way to progressively transition.
This tool, using age is pretty neat: https://github.com/ndavd/agevault. So simple, yet seemingly powerful!
@prologic@twtxt.net itās a Clownflare option to prevent images on your website from being embedded on other websites. It helps with my low bandwidth resources. And I believe you can set-up similar rules with Nginx, Iām just too lazy to do it manually RN.
Kinda cool tool for bringing together all your timeline based data across socials.
pour faire mes cours, le gĆ©nĆ©rateur de grisse bouille est magique. https://framalab.org/gknd-creator/ . je dĆ©couvre que je peux hĆ©berger le code source https://si3t.ch/tools/comicgen/. Reste donc Ć y dĆ©poser les images modĆØles qui me plaisent pour crrĆ©er des cours de sciences. Et bien Ć©videmment, je ne peux māempĆŖcher de penser Ć <@peha@framapiaf.org> š¼
./tools/dump_cache.sh: line 8: bat: command not found
No Token Provided
I donāt have bat on my VPS and there is no package for installing it. Is cat a reasonable alternate?
Thereās other potential uses for the tool (compare syscall latency between OSes, stat latency between file systems), but not what iām after.
Holly insert inappropriate word here ! 𤣠I have finally done it !!!
FreeBSD 14.0-RELEASE (GENERIC) #0 releng/14.0-n265380-f9716eee8ab4: Fri Nov 10 05:57:23 UTC 2023
Welcome to FreeBSD!
% pkg update
The package management tool is not yet installed on your system.
Do you want to fetch and install it now? [y/N]:
Bloom Filter
ā Read more
Interesting. Thanks! And thank you for replying. :) Indeed, I donāt check for mention with twtxt. To me, twtxt is to share, not to talk: there is my email commented at the top of my #twtxt.txt for this purpose. Trying to create discussions with twtxt is nonsens : there are much better tools to do so (email, xmpp, ā¦) @aelaraji@aelaraji.com @im-in.space@im-in.space
Iām looking for wallpapers matching a color palette. Is there any tool to do so? I found the opposite, picture to palette, but not palette to picture :/
@mckinley@twtxt.net You definitely have got a point!
It is kind of a hassle to keep things in sync and NOT eff up.
It happened to me before but I was lucky enough to have backups elsewhere.
But, now I kind of have a workflow to avoid data loss while benefiting from both tools.
P.S: my bad, I meant Syncthing earlier on my original replay instead of Rsync. š«
</> htmx - high power tools for html really liking the idea of htmx š¤ If I donāt have to learn all this complicated TypeScript/React/NPM garbage, I can just write regular SSA (Server-Side-Apps) and then progressively upgrade to SPA (Single-Page-App) using htmx hmmm š§
it is an addon in the download tool. Or you can use xcaddy to build it in.

its a notebook tool like evernote. @sorenpeter@darch.dk linked it above: https://joplinapp.org/
@lyse@lyse.isobeef.org its a hierarchy key value format. I designed it for the network peering tools i use.. I can grant access to different parts of the tree to other users.. kinda like directory permissions. a basic example of the format is:
@namespace
# multi
# line
# comment
root :value
# example space comment
@namespace.name space-tag
# attribute comments
attribute attr-tag :value for attribute
# attribute with multiple
# lines of values
foo :bar
:bin
:baz
repeated :value1
repeated :value2
each @ starts the definition of a namespace kinda like [name] in ini format. It can have comments that show up before. then each attribute is key :value and can have their own # comment lines.
Values can be multi line.. and also repeated..
the namespaces and values can also have little meta data tags added to them.

the service can define webhooks/mqtt topics to be notified when the configs are updated. That way it can deploy the changes out when they are updated.
#gemini readers, I wrote a tool to download new gemfeeds entries instead of opening a client: gemini://si3t.ch/log/2024-02-28-gemfeeds-downloader.txt
Google Chrome Gains AI Features Including a Writing Helper
Google is adding new AI features to Chrome, including tools to organize browser tabs, customize themes, and assist users with writing online content such as reviews and forum posts.
The writing helper is similar to an AI-powered feature already offered in Googleās experimental search experience, SGE, which helps users draft emails in various tones and lengths. W ⦠ā Read more
when writing a new tool/software, write doc first, explaining how it works. Then, actually writing the code is much easier :)
Forensics Tools https://github.com/mesquidar/ForensicsTools
@eapl.me@eapl.me I have many fond memories of Turbo pascal and Turbo C(++). They really did have a great help system. And debug tools! Its rare for language docs to be as approachable. QBasic was great. As was PHP docs when I first came into web.
Obligatory Twtxt post: I love how I can simply use a terminal window and some very basic tools (echo, scp, ssh) to publish thoughts, as they pop up, onto the Internet in a structured way, that can be found and perhaps even appreciated.
@lyse@lyse.isobeef.org I wish more standardization around distributed issues and PRs within the repo ala git-bug was around for this. I see it has added some bridge tooling now.
Anyone have any ideas how you might identify processes (pids) on Linux machine that are responsible for most of the Disk I/O on that machine and subsequently causing high I/O wait times for other processes? š¤
Important bit: The machine has no access to the internet, there are hardly any standard tools on it, etc. So I have to get something to it āair gappedā. I have terminal access to it, so I can do interesting things like, base64 encode a static binary to my clipboard and paste it to a file, then base64 decode it and execute. Thatās about the only mechanisms I have.
@prologic@twtxt.net do not use it, but gave it a try early on and was not impressed. it gave a good outline of what I asked but then unreliably dorked up all the crucial parts.
I will say though if it is truly learning at the rate they say then it should be a good tool.
In setting up my own company and itās internal tools and services and supporting infrastructure, the ony thing I havenāt figured out how to solve āreally wellā is Email, Calendar and Contacts š¢ All the options that exist āsuckā. They suck either in terms of āoperational complexity and overheadsā or āa poor user experienceā.
@prologic@twtxt.net Horseshit hype:
- AI that we have today cannot thinkāthere is no cognitive capacity
- AI that we have today cannot be interviewedāāinterā āviewingā is two minds interacting, but AI of today has no mind, which means this is a puppet show
- AI today is not freeāitās a tool, a machine, hardly different from a hammer. It does what a human directs it to do and has no drives, desires, or autonomy. What youāre seeing here is a fancy Mechnical Turk
This shit is probably paid for by AI companies who desperately want us to think of the AI as far more capable than it actually is, because that juices sales and gives them a way to argue they arenāt responsible for any harms it causes.
Iād love to read the original source code of this:
https://ecsoft2.org/t-tiny-editor
This was our standard editor back in the day, not an āemergency toolā. And itās only 9kB in size ⦠which feels absurd in 2023. š The entire hex dump fits on one of todayās screens.
Being so small meant it had no config file. Instead, it came with TKEY.EXE, a little tool to binary-patch T.EXE to your likings.



Yep, thatās right, we have to use these tools in a proper way; terminal itās not a friendly tool to use for this kind of stuff, on mobile devices, and web interfaces are prepared to bring us a confortable space.
Btw, Iām waiting for your php based client š no pressure⦠š¤
[lang=en] That was the reason for twtxt-php =P
I tried using CLI tools but it was too hacky, I think.
More if we consider Jakobās Law, where we have prior expectations of a microblogging system.
A Web interface could be quite minimalistic and usable as well. (And mobile-friendly)
snac/the fediverse for a few days and already I've had to mute somebody. I know I come on strongly with my opinions sometimes and some people don't like that, but this person had already started going ad hominem (in my reading of it), and was using what felt to me like sketchy tactics to distract from the point I was trying to make and to shut down conversation. They were doing similar things to other people in the thread so rather than wait for it to get bad for me I just muted them. People get so weirdly defensive so fast when you disagree with something they said online. Not sure I fully understand that.
@prologic@twtxt.net Well, you can mute or block individual users, and you can mute conversations too. I think the tools for controlling your interactions arenāt so bad (they could definitely be improved ofc). And in my case, I was replying to something this person said, so it wasnāt outrageous for his reply to be pushed to me. Mostly, I was sad to see how quickly the conversation went bad. I thought I was offering something relatively uncontroversial, and actually I was just agreeing with and amplifying something another person had already said.
What I see here is that when I was reading your .txt, the timestamp was like 40 minutes later than current time. Say itās 1pm and that twt is timed on 1.40pm
No idea why, perhaps your server has a wrong Timezone, or your twtxt tool is doing some timezome conversion?
Google Says Itāll Scrape Everything You Post Online for AI
Google updated its privacy policy over the weekend, explicitly saying the company reserves the right to scrape just about everything you post online to build its AI tools.
Google can eat shit.
Seems to me you could write a script that:
- Parses a StackOverflow question
- Runs it through an AI text generator
- Posts the output as a post on StackOverflow
and basically pollute the entire information ecosystem there in a matter of a few months? How long before some malicious actor does this? Maybe itās being done already š¤·
What an asinine, short-sighted decision. An astonishing number of companies are actively reducing headcount because their executives believe they can use this newfangled AI stuff to replace people. But, like the dot com boom and subsequent bust, many of the companies going this direction are going to face serious problems when the hypefest dies down and the reality of what this tech can and canāt do sinks in.
We really, really need to stop trusting important stuff to corporations. They are not tooled to last.
@shreyan@twtxt.net I agree re: AR. Vircadia is neat. I stumbled on it years ago when I randomly started wondering āwonder whatās going on with Second Life and those VR thingsā and started googling around.
Unfortunately, like so many metaverse efforts, itās almost devoid of life. Interesting worlds to explore, cool tools to build your own stuff, but almost no people in it. It feels depressing, like an abandoned shopping mall.
I have no interest in doing anything about it, even if I had the time (which I donāt), but these kind of thing happen all day every day to countless people. My silly blog post isnāt worth getting up in arms about, but there are artists and other creators who pour countless hours, heart and soul into their work, only to have it taken in exactly this way. Thatās one of the reasons Iām so extremely negative about the spate of āAIā tools that have popped up recently. They are powered by theft.
There is a ārightā way to make something like GitHub CoPilot, but Microsoft did not choose that way. They chose one of the most exploitative options available to them. For that reason, I hope they face significant consequences, though I doubt they will in the current climate. I also hope that CoPilot is shut down, though Iām pretty certain it will not be.
Other than access to the data behind it, Microsoft has nothing special that allows it to create something like CoPilot. The technology behind it has been around for at least a decade. There could be a āpublicā version of this same tool made by a cooperating group of people volunteering, āleasingā, or selling their source code into it. There could likewise be an ethically-created corporate version. Such a thing would give individual developers or organizations the choice to include their code in the tool, possibly for a fee if thatās something they want or require. The creators of the tool would have to acknowledge that they have suppliersāthe people who create the code that makes their tool possibleāinstead of simply stealing what they need and pretending thatās fine.
This era weāre living through, with large companies stomping over all laws and regulations, blatantly stealing other peopleās work for their own profit, cannot come to an end soon enough. It is destroying innovation, and we all suffer for that. Having one nifty tool like CoPilot that gives a bit of convenience is nowhere near worth the tremendous loss that Microsoftās actions in this instace are creating for everyone.
I was listening to an OāReilly hosted event where they had the CEO of GitHub, Thomas Dohmke, talking about CoPilot. I asked about biased systems and copyright problems. He, Thomas Dohmke, said, that in the next iteration they will show name, repo and licence information next to the code snippets you see in CoPilot. This should give a bit more transparency. The developer still has to decide to adhere to the licence. On the other hand, I have to say he is right about the fact, that probably every one of us has used a code snippet from stack overflow (where 99% no licence or copyright is mentioned) or GitHub repos or some tutorial website without mentioning where the code came from. Of course, CoPilot has trained with a lot of code from public repos. It is a more or less a much faster and better search engine that the existing tools have been because how much code has been used from public GitHub repos without adding the source to code you pasted it into?
I have to write so many emails to so many idiots who have no idea what they are doing
So it sounds to me like the pressure is to reduce how much time you waste on idiots, which to my mind is a very good reason to use a text generator! I guess in that case you donāt mind too much whether the company making the AI owns your prompt text?
Iād really like to see tools like this that you can run on your desktop or phone, so they donāt send your hard work off to someone else and give a company a chance to take it from you.
On LinkedIn I see a lot of posts aimed at software developers along the lines of āIf youāre not using these AI tools (X,Y,Z) youāre going to be left behind.ā
Two things about that:
- No youāre not. If you have good soft skills (good communication, show up on time, general time management) then youāre already in excellent shape. No AI can do that stuff, and for that alone no AI can replace people
- This rhetoric is coming directly from the billionaires who are laying off tech people by the 100s of thousands as part of the class war theyāve been conducting against all working people since the 1940s. They want you to believe that you have to scramble and claw over one another to learn the āAIā that theyāre forcing onto the world, so that you stop honing the skills that matter (see #1) and are easier to obsolete later. Donāt fall for it. Itās far from clear how this will shake out once governments get off their asses and start regulating this stuff, by the wayāmost of these āAIā tools are blatantly breaking copyright and other IP laws, and some day thatāll catch up with them.
That said, it is helpful to know thy enemy.
@prologic@twtxt.net I get the worry of privacy. But I think there is some value in the data being collected. Do I think that Russ is up there scheming new ways to discover what packages you use in internal projects for targeting ads?? Probably not.
Go has always been driven by usage data. Look at modules. There was need for having repeatable builds so various package tool chains were made and evolved into what we have today. Generics took time and seeing pain points where they would provide value. They werenāt done just so it could be checked off on a box of features. Some languages seem to do that to the extreme.
Whenever changes are made to the language there are extensive searches across public modules for where the change might cause issues or could be improved with the change. The fs embed and strings.Cut come to mind.
I think its good that the language maintainers are using what metrics they have to guide where to focus time and energy. Some of the other languages could use it. So time and effort isnāt wasted in maintaining something that has little impact.
The economics of the āspyingā are to improve the product and ecosystem. Is it āspyingā when a municipality uses water usage metrics in neighborhoods to forecast need of new water projects? Or is it to discover your shower habits for nefarious reasons?
Iāve never liked the idea of having everything displayed all of the time for all of history.
And I still donāt: Search and Bookmarks are better tools for this IMO.
From a technical perspective however, we will not introduce any CGO dependencies into yarnd ā It makes portability harder.
Also I hate SQL š
restic Ā· Backups done right! ā In case no-one has used this wonderful tool restic yet, I can beyond a doubt assure you it is really quite fantastic š #backups
$name$ and then dispatch the hashing or checking to its specific format.
I have submitted this to be used as the hash tooling for Yarn. See it as a good example on using this in a production environment!