Searching yarn

Twts matching #training
Sort by: Newest, Oldest, Most Relevant
In-reply-to » @movq The success of large neural nets. People love to criticize today's LLMs and image models, but if you compare them to what we had before, the progress is astonishing.

@prologic@twtxt.net I don’t know what you mean when you call them stochastic parrots, or how you define understanding. It’s certainly true that current language models show an obvious lack of understanding in many situations, but I find the trend impressive. I would love to see someone achieve similar results with much less power or training data.

⤋ Read More

Regarding complexity budget, slow software, all that:

Very few people do take pride in building simple, elegant, high-quality systems, do they? Why is that? Why are huge shiny things with tons of features more attractive? šŸ¤”

I never explicitly thought about this, to be honest. It was only at the back of my head. And I never tried to teach our younger ā€œstudentsā€ at work: ā€œHey, it’s a great achievement to build something simple and elegant. That’s something to be proud of!ā€

Worse, simple software is often described as ā€œboringā€. Yes, in a way, it is boring, because your brain doesn’t have to get into overdrive to understand it. But that’s exactly the point. And it’s hard to achieve that! Simple software isn’t just ā€œfewer lines of codeā€, you have to be pretty clever to solve a problem in a simple and elegant way. So it’s something to be proud of.

Could this be an intuitive, emotional way to get more people on board the ā€œsimple softwareā€-train? šŸ¤”

⤋ Read More

Pinellas County Running: 3.16 miles, 00:08:30 average pace, 00:26:52 duration
aiming for whatever felt easy. the humidity really was heavy with a light fog. woke up with no real pain and it was not until the end of the run where i felt a slightly sharp pain around my left glute and a bit in the left hip as well. thinking i need to reduce mileage a bit and try to train around it until i feel good enough to get back in to a routine again.
#running

⤋ Read More
In-reply-to » Not making THREADING the default view of e-mail clients and thus teaching users that e-mail is ā€œchaoticā€ (if you get a lot of mail, it becomes unusable without threading) and ā€œneedsā€ full quoting all the time was one of the worst mistakes ever.

My email is such a cluster of noise. The only time i actually use it is to find out I have to do my security training or something. All communication is slack now days.

⤋ Read More
In-reply-to » @prologic High five, I’m ā€œgeneration Javaā€ as well! šŸ˜‚ There were some leftovers of C++, we used that in the computer graphics courses in Uni a lot. But pretty much anything else that involved programming was Java.

@movq@www.uninformativ.de Haha! yeah sounds about like my HS CS program. A math teacher taught visual basic and pascal. and over on the other end of the school we had ā€œelectronicsā€ which was a room next to the auto body class where they had a bunch of random computer parts scavenged from the district decommissioned surplus storage.

The advanced class would piece together training kits for the basic class to put together.

⤋ Read More

St Petersburg Distance Classic Marathon: 26.41 miles, 00:11:23 average pace, 05:00:39 duration
first marathon down. everything that could go wrong did. honestly i am just proud i did not quit. now i have to look at the run and figure out what i can tweak or add to my training. had a cramp start in my right quad at around mile 15. then around mile 18 both of my calves started to feel odd as if someone was lightly strumming my tendons. then they seized! this continued for the remainder of the marathon where i would walk then try to run and then stop when i had to. then during the entirety of the pace my nose would not stop dripping making it difficult to breathe. ha! also my shorts almost came down twice and i had to re-tie them while carrying my handheld water in my teeth. seriously, so many things i did not expect and had not happened in any previous runs.

really happy to be able to eat spicy food and have some alcoholic beverages again though!
#running #race

⤋ Read More

Pinellas County - Easy: 5.05 miles, 00:09:08 average pace, 00:46:08 duration
everything clicked today. kept a steady but mildly progressive pace whilst keeping the heart rate in zone 2 for the most part. nothing felt strained and breathing was easy. this one was a great boost in confidence seeing the progress made in the training and very happy with it.
#running

⤋ Read More

Having fun with React - yet again. A large part of my job entails (re)learning technologies - luckily I have access to some good resources in the form of training- and tutorial sites, all provided by my employer.

⤋ Read More

Pyramids: 5.39 miles, 00:10:53 average pace, 00:58:37 duration
called it early due to spicy food and opted for treadmill. was trying to get to zone 5 and was sure the pace would get me there at the peak, but barely reached threshold with conditions. curious if i was outside how different it would have been. new training block. 10:55, 9:41, 6:59
#running #treadmill

⤋ Read More

SPRF Half Marathon: 13.20 miles, 00:09:29 average pace, 02:05:12 duration
still host and humid (no surprise) but more cloud cover today. no kids but beth came and was able to cheer me on in a couple of places which was fun. the last bit she yelled ā€œfive to go!ā€ which kind of got in my head a bit, albeit i think the heat started to get to me as well. had to take a couple of brief walks just to recollect and focus again. pretty good training run. keeping it in the green for the most part and around marathon pace. can’t wait to see how cooler weather and more training will pay off!
#running #race

⤋ Read More
In-reply-to » Google Says It'll Scrape Everything You Post Online for AI

@marado@twtxt.net It can’t possibly be defensible, which to me always signals an attempt at a power grab. They never explicitly said ā€œwe will use anything we scrape from the web to train our AIā€ before–that’s new. There is growing pushback against that practice, with numerous legal cases winding through the legal system right now. Some day those cases will be heard and decided on by judges. So they’re trying to get out ahead of that, in my opinion, and cement their claims to this data before there’s a precedent set.

⤋ Read More
In-reply-to » Home | Tabby This is actually pretty cool and useful. Just tried this on my Mac locally of course and it seems to have quite good utility. What would be interesting for me would be to train it on my code and many projects šŸ˜…

Most of the can run locally have such a small training set they arnt worth it. Are more like the Markov chains from the subreddit simulator days.

There is one called orca that seems promising that will be released as OSS soon. Its running at comparable numbers to OpenAI 3.5.

https://youtube.com/watch?v=Dt_UNg7Mchg&feature=share9

⤋ Read More
In-reply-to » I played with nlpodyssey/verbaflow: Neural Language Model for Go today a little bit today.... First I had to download a ~2GB file (the model), then convert that to a format the program verbaflow understands which came out to roughly ~5GB. Then I tried some of the samples in the README. My god, this this is so goddamn awfully slow its like watching paint dry 😱 All just to predict the next few tokens?! 😳 I had a look at the resource utilisation as well as it was trying to do this "work", using 100% of 1.5 Cores and ~10GB of Memory 😳 Who da fuq actually thinks any of this large language model (LLM) and neural network crap is actually any good or useful? šŸ¤” Its just garbage 🤣

@prologic@twtxt.net You more or less need a data center to run one of these adequately (well, train…you can run a trained one with a little less hardware). I think that’s the idea–no one can run them locally, they have to rent them (and we know how much SaaS companies and VCs love the rental model of computing).

There’s a lot of promising research-grade work being done right now to produce models that can be run on a human-scale (not data-center-scale) computing setup. I suspect those will become more commonly deployed in the next few years.

⤋ Read More
In-reply-to » @prologic hmm, dunno about the recency of that line of thought. I suspect though that given his (recent or not) history, if someone directly asked him "do you support rape" he would not say "no", he'd go on one of these rambling answers about property crime like he did in the video. Maybe I'm mind poisoned by being around academics my whole career, but that way of talking is how an academic gives you an answer they know will be unpopular. PhD = Piled Higher And Deeper, after all right? In other words, if he doesn't say "no" right away, he's saying "yes", except with so many words there's some uncertainty about whether he actually meant yes. And he damn well knows that, and that's why I give him no slack.

@prologic@twtxt.net It’s a fun challenge to see how many words you can say without expressing any ideas at all. Maybe this GPT stuff should be trained to do that!

⤋ Read More

I was listening to an O’Reilly hosted event where they had the CEO of GitHub, Thomas Dohmke, talking about CoPilot. I asked about biased systems and copyright problems. He, Thomas Dohmke, said, that in the next iteration they will show name, repo and licence information next to the code snippets you see in CoPilot. This should give a bit more transparency. The developer still has to decide to adhere to the licence. On the other hand, I have to say he is right about the fact, that probably every one of us has used a code snippet from stack overflow (where 99% no licence or copyright is mentioned) or GitHub repos or some tutorial website without mentioning where the code came from. Of course, CoPilot has trained with a lot of code from public repos. It is a more or less a much faster and better search engine that the existing tools have been because how much code has been used from public GitHub repos without adding the source to code you pasted it into?

⤋ Read More