Searching yarn

Twts matching #training
Sort by: Newest, Oldest, Most Relevant

We went on a 14 kilometers long hike in the heat, only a few spots were in the shade, most of our trip was in the open fields with the sun beating down on us. We reapplied the sun blocker after about two hours or so. All in all it took us about three and a half hours before we reached our destination Besigheim.

Last time I was there it was rainy, now we had the exact opposite. After some yummy Chinese lunch we visited the old town. There’s some gorgeous timer framing to see. When kept in decent shape, it just looks so dang cool.

Since it was too hot, we rode back by train. Despite the heat and some sections near the roaring Autobahn, this was a nice hike. Would do it again. Only in colder weather, though. I certainly don’t wanna trade my comperatively larger (still nothing to other more rural areas), covering forests with the wide open fields and vineyards in summer. That’s for sure.

Image

https://lyse.isobeef.org/wanderung-von-asperg-nach-besigheim-2025-05-01/

​ Read More

yeah! I’ve passed the critical point at factorio
 I managed to automate blue science, trains and oil production
 now it’s a great fun again


​ Read More

Short summary of Project2025 and Trump’s plans for the US:

  • Abolish the Federal Reserve
    Why? To end what is seen as an unelected, centralized body that exerts too much influence over the economy and monetary policy, replacing it with a more transparent, market-driven approach.

  • Implement a national consumption tax
    Why? To replace the current federal income tax system, simplify taxation, and increase government revenue through a broader base that includes all consumers.

  • Lower corporate tax rates
    Why? To promote business growth, increase investment, and stimulate job creation by reducing the financial burden on companies.

  • Deregulate environmental policies
    Why? To reduce government intervention in the economy, particularly in energy and natural resources sectors, and to foster a more business-friendly environment.

  • Restrict abortion access
    Why? To align with conservative pro-life values and overturn or limit abortion rights, seeking to restrict the practice at a federal level.

  • Dismantle LGBTQ+ protections
    Why? To roll back protections viewed as promoting LGBTQ+ rights in areas like employment and education, in line with traditional family values.

  • Eliminate diversity, equity, and inclusion (DEI) programs
    Why? To end policies that are seen as divisive and to promote a merit-based system that prioritizes individual achievements over group identity.

  • Enforce stricter immigration policies, including mass deportations and detentions
    Why? To prioritize border security, reduce illegal immigration, and enforce existing laws more aggressively, as part of a broader strategy to safeguard U.S. sovereignty.

  • Eliminate the Department of Education
    Why? To reduce federal control over education and shift responsibilities back to local governments and private sectors, arguing that education decisions should be made closer to the community level.

  • Restructure the Department of Justice
    Why? To ensure the department aligns more closely with the administration’s priorities, potentially reducing its scope or focus on areas like civil rights in favor of law-and-order policies.

  • Appoint political loyalists to key federal positions
    Why? To ensure that government agencies are headed by individuals who are committed to advancing the administration’s policies, and to reduce the influence of career bureaucrats.

  • Develop training programs for appointees to execute reforms effectively
    Why? To ensure that political appointees are equipped with the knowledge and skills necessary to implement the proposed changes quickly and effectively.

  • Provide a 180-day transition plan with immediate executive orders
    Why? To ensure that the incoming administration can swiftly implement its agenda and make major changes early in its term without delay.

Do y’all agree with any/all/some of these poliices? Hmmm đŸ€”

#Project2025 #US #Trump

​ Read More
In-reply-to » I got promoted today to try using Passkeys on Github.com. Fine 😅 I did that, but I discovered that when you use your Passkey to login, Chrome prompts you for your device's password (i.e: The password you use to login to your macOS Desktop). Is that intentional? Kind of defeats the point no? I mean sure, now there's no Password being transmitted, stored or presented to Github.com but still, all an attacker has to do is somehow be on my device and know my login password to my device right? Is that better or worse? đŸ€”

@prologic@twtxt.net That boycott didn’t last very long, eh!?

Yeah, sounds like another hype train arriving at the station.

​ Read More

OpenAI doesn’t like it when you use “their” generated slop without permission
OpenAI says it has found evidence that Chinese artificial intelligence start-up DeepSeek used the US company’s proprietary models to train its own open-source competitor, as concerns grow over a potential breach of intellectual property. ↫ Cristina Criddle and Eleanor Olcott for the FT This is more ironic than writing a song called Ironic that lists situations that aren’t actually 
 ⌘ Read more

​ Read More

OpenAI Says It Has Evidence DeepSeek Used Its Model To Train Competitor
OpenAI says it has evidence suggesting Chinese AI startup DeepSeek used its proprietary models to train a competing open-source system through “distillation,” a technique where smaller models learn from larger ones’ outputs.

The San Francisco-based company, along with partner Microsoft, blocked suspected DeepSeek accounts from accessing 
 ⌘ Read more

​ Read More

Bluesky’s Open API Means Anyone Can Scrape Your Data for AI Training. It’s All Public
Bluesky says it will never train generative AI on its users’ data. But despite that, “one million public Bluesky posts — complete with identifying user information — were crawled and then uploaded to AI company Hugging Face,” reports Mashable (citing an article by 404 Media).

“Shortly after the article’s p 
 ⌘ Read more

​ Read More

OpenAI, Google, Anthropic admit they can’t scale up their chatbots any further

Once you’ve trained your large language model on the entire written output of humanity, where do you go?

https://pivot-to-ai.com/2024/11/14/openai-google-anthropic-admit-they-cant-scale-up-their-chatbots-any-further/

So we’re going to destroy the environment for AI slop that isn’t fit for purpose now and, if you believe the above post, never will be.

​ Read More

Portion of the modified Twitter TOS that goes into effect today (it’s on right now), as summarised (ironically) by Google’s Gemini:

“In simpler terms, this means that when you share your content (like text, images, or videos) on the service, you’re giving the company permission to use it in various ways. They can copy, modify, distribute, and even use it to train their AI models. This includes sharing your content with others and using it on other platforms. You won’t be paid for this, but using the service itself is considered enough compensation.”

​ Read More
In-reply-to » Generative AI Doesn't Have a Coherent Understanding of the World, MIT Researchers Find Long-time Slashdot reader Geoffrey.landis writes: Despite its impressive output, a recent study from MIT suggests generative AI doesn't have a coherent understanding of the world. While the best-performing large language models have surprising capabilities that make it seem like the models are implicitly learn ... ⌘ Read more

I like this comment on Slashdot in the above link:

LLMs don’t have an understanding of anything. They can only regurgitate derivations of what they’ve been trained on and can’t apply that to something new in the same ways that humans or even other animals can. The models are just so large that the illusion is impressive.

So true.

​ Read More

hop, entraĂźnement terminĂ©, j’ai fait le plein d’énergie avant d’aller donner un sang de qualittĂ© ^^ #EFS #dondusang. Le niveau 7 de la mĂ©thode #lafay est par contre trop longue, je n’ai pas assez de temps pour faire ça bien. Tant pis dans ce cas, retour Ă  la n°6 et j’y ajouterai 1 exercice jusqu’à Ă©puisement tirĂ© au sort. Ou alors je ressort le #TRX, il faut qu eje trouve oĂč l’accrocher. #sport #training

​ Read More

Insecure Robot Vacuums From Chinese Company Deebot Collect Photos and Audio to Train Their AI
Long-time Slashdot reader schwit1 shared this report from Australia’s public broadcaster ABC:

Ecovacs robot vacuums, which have been found to suffer from critical cybersecurity flaws, are collecting photos, videos and voice recordings — taken inside customers’ houses — to train the company’ 
 ⌘ Read more

​ Read More
In-reply-to » @movq The success of large neural nets. People love to criticize today's LLMs and image models, but if you compare them to what we had before, the progress is astonishing.

@prologic@twtxt.net I don’t know what you mean when you call them stochastic parrots, or how you define understanding. It’s certainly true that current language models show an obvious lack of understanding in many situations, but I find the trend impressive. I would love to see someone achieve similar results with much less power or training data.

​ Read More

Regarding complexity budget, slow software, all that:

Very few people do take pride in building simple, elegant, high-quality systems, do they? Why is that? Why are huge shiny things with tons of features more attractive? đŸ€”

I never explicitly thought about this, to be honest. It was only at the back of my head. And I never tried to teach our younger “students” at work: “Hey, it’s a great achievement to build something simple and elegant. That’s something to be proud of!”

Worse, simple software is often described as “boring”. Yes, in a way, it is boring, because your brain doesn’t have to get into overdrive to understand it. But that’s exactly the point. And it’s hard to achieve that! Simple software isn’t just “fewer lines of code”, you have to be pretty clever to solve a problem in a simple and elegant way. So it’s something to be proud of.

Could this be an intuitive, emotional way to get more people on board the “simple software”-train? đŸ€”

​ Read More

Pinellas County Running: 3.16 miles, 00:08:30 average pace, 00:26:52 duration
aiming for whatever felt easy. the humidity really was heavy with a light fog. woke up with no real pain and it was not until the end of the run where i felt a slightly sharp pain around my left glute and a bit in the left hip as well. thinking i need to reduce mileage a bit and try to train around it until i feel good enough to get back in to a routine again.
#running

​ Read More
In-reply-to » Not making THREADING the default view of e-mail clients and thus teaching users that e-mail is “chaotic” (if you get a lot of mail, it becomes unusable without threading) and “needs” full quoting all the time was one of the worst mistakes ever.

My email is such a cluster of noise. The only time i actually use it is to find out I have to do my security training or something. All communication is slack now days.

​ Read More
In-reply-to » @prologic High five, I’m “generation Java” as well! 😂 There were some leftovers of C++, we used that in the computer graphics courses in Uni a lot. But pretty much anything else that involved programming was Java.

@movq@www.uninformativ.de Haha! yeah sounds about like my HS CS program. A math teacher taught visual basic and pascal. and over on the other end of the school we had “electronics” which was a room next to the auto body class where they had a bunch of random computer parts scavenged from the district decommissioned surplus storage.

The advanced class would piece together training kits for the basic class to put together.

​ Read More

St Petersburg Distance Classic Marathon: 26.41 miles, 00:11:23 average pace, 05:00:39 duration
first marathon down. everything that could go wrong did. honestly i am just proud i did not quit. now i have to look at the run and figure out what i can tweak or add to my training. had a cramp start in my right quad at around mile 15. then around mile 18 both of my calves started to feel odd as if someone was lightly strumming my tendons. then they seized! this continued for the remainder of the marathon where i would walk then try to run and then stop when i had to. then during the entirety of the pace my nose would not stop dripping making it difficult to breathe. ha! also my shorts almost came down twice and i had to re-tie them while carrying my handheld water in my teeth. seriously, so many things i did not expect and had not happened in any previous runs.

really happy to be able to eat spicy food and have some alcoholic beverages again though!
#running #race

​ Read More

Pinellas County - Easy: 5.05 miles, 00:09:08 average pace, 00:46:08 duration
everything clicked today. kept a steady but mildly progressive pace whilst keeping the heart rate in zone 2 for the most part. nothing felt strained and breathing was easy. this one was a great boost in confidence seeing the progress made in the training and very happy with it.
#running

​ Read More

Having fun with React - yet again. A large part of my job entails (re)learning technologies - luckily I have access to some good resources in the form of training- and tutorial sites, all provided by my employer.

​ Read More

Pyramids: 5.39 miles, 00:10:53 average pace, 00:58:37 duration
called it early due to spicy food and opted for treadmill. was trying to get to zone 5 and was sure the pace would get me there at the peak, but barely reached threshold with conditions. curious if i was outside how different it would have been. new training block. 10:55, 9:41, 6:59
#running #treadmill

​ Read More

SPRF Half Marathon: 13.20 miles, 00:09:29 average pace, 02:05:12 duration
still host and humid (no surprise) but more cloud cover today. no kids but beth came and was able to cheer me on in a couple of places which was fun. the last bit she yelled “five to go!” which kind of got in my head a bit, albeit i think the heat started to get to me as well. had to take a couple of brief walks just to recollect and focus again. pretty good training run. keeping it in the green for the most part and around marathon pace. can’t wait to see how cooler weather and more training will pay off!
#running #race

​ Read More
In-reply-to » Google Says It'll Scrape Everything You Post Online for AI

@marado@twtxt.net It can’t possibly be defensible, which to me always signals an attempt at a power grab. They never explicitly said “we will use anything we scrape from the web to train our AI” before–that’s new. There is growing pushback against that practice, with numerous legal cases winding through the legal system right now. Some day those cases will be heard and decided on by judges. So they’re trying to get out ahead of that, in my opinion, and cement their claims to this data before there’s a precedent set.

​ Read More
In-reply-to » Home | Tabby This is actually pretty cool and useful. Just tried this on my Mac locally of course and it seems to have quite good utility. What would be interesting for me would be to train it on my code and many projects 😅

Most of the can run locally have such a small training set they arnt worth it. Are more like the Markov chains from the subreddit simulator days.

There is one called orca that seems promising that will be released as OSS soon. Its running at comparable numbers to OpenAI 3.5.

https://youtube.com/watch?v=Dt_UNg7Mchg&feature=share9

​ Read More
In-reply-to » I played with nlpodyssey/verbaflow: Neural Language Model for Go today a little bit today.... First I had to download a ~2GB file (the model), then convert that to a format the program verbaflow understands which came out to roughly ~5GB. Then I tried some of the samples in the README. My god, this this is so goddamn awfully slow its like watching paint dry đŸ˜± All just to predict the next few tokens?! 😳 I had a look at the resource utilisation as well as it was trying to do this "work", using 100% of 1.5 Cores and ~10GB of Memory 😳 Who da fuq actually thinks any of this large language model (LLM) and neural network crap is actually any good or useful? đŸ€” Its just garbage đŸ€Ł

@prologic@twtxt.net You more or less need a data center to run one of these adequately (well, train
you can run a trained one with a little less hardware). I think that’s the idea–no one can run them locally, they have to rent them (and we know how much SaaS companies and VCs love the rental model of computing).

There’s a lot of promising research-grade work being done right now to produce models that can be run on a human-scale (not data-center-scale) computing setup. I suspect those will become more commonly deployed in the next few years.

​ Read More
In-reply-to » @prologic hmm, dunno about the recency of that line of thought. I suspect though that given his (recent or not) history, if someone directly asked him "do you support rape" he would not say "no", he'd go on one of these rambling answers about property crime like he did in the video. Maybe I'm mind poisoned by being around academics my whole career, but that way of talking is how an academic gives you an answer they know will be unpopular. PhD = Piled Higher And Deeper, after all right? In other words, if he doesn't say "no" right away, he's saying "yes", except with so many words there's some uncertainty about whether he actually meant yes. And he damn well knows that, and that's why I give him no slack.

@prologic@twtxt.net It’s a fun challenge to see how many words you can say without expressing any ideas at all. Maybe this GPT stuff should be trained to do that!

​ Read More

I was listening to an O’Reilly hosted event where they had the CEO of GitHub, Thomas Dohmke, talking about CoPilot. I asked about biased systems and copyright problems. He, Thomas Dohmke, said, that in the next iteration they will show name, repo and licence information next to the code snippets you see in CoPilot. This should give a bit more transparency. The developer still has to decide to adhere to the licence. On the other hand, I have to say he is right about the fact, that probably every one of us has used a code snippet from stack overflow (where 99% no licence or copyright is mentioned) or GitHub repos or some tutorial website without mentioning where the code came from. Of course, CoPilot has trained with a lot of code from public repos. It is a more or less a much faster and better search engine that the existing tools have been because how much code has been used from public GitHub repos without adding the source to code you pasted it into?

​ Read More