Now this is a properly balanced piece which looks beyond the hype. Usable yes, if hallucinations don't have a high impact. Can the hallucinations be solved? To be seen, I personally have my doubts with the current architecture... at least banking it all on human feedback is being very naive about the scale of the task.
This is an excellent piece. Very nice portrait of Emily M. Bender a really gifted computational linguist and really bad ass if you ask me. She's out there asking all the difficult questions about the current moment regarding large language models and so far the answers are (I find) disappointing. We collectively seem to be way too fascinated by the shiny new toy and the business opportunities to pay really attention to the impact on the social fabric of all of this.
Interesting strategy, shows a fascinating blind spot in the typical AIs used for Go nowadays. It kind of hints to the fact that the neural networks abstract knowledge much less than advertised.
Interesting work, trying to get back to the source material used by a generative model. This is definitely necessary as well.
A few interesting points in there. Too much hype and important points are glanced over, we'd all benefit from them being more actively explored.
Very nice summary of the architecture in the latest trend of transformer models. Long but comprehensive, a good way to start diving in the topic.
Such generative models are getting more and more accessible. You can play with them using a few lines of python now.
The human labor behind AI training is still on going. This is clearly gruesome and sent over to other countries... ignoring the price for a minute this is also a good way to hide its consequences I guess.
Interesting reverse engineering job of Copilot's client side to have a better idea at which information it actually feeds to the model. A couple of funny tricks to prepare the prompt are involved. Obviously some telemetry involved as well, again with interesting heuristics to try to figure out if the user kept the suggestion or not.
At least a good balanced post about Generative AI and programming. It's not overestimating abilities of the latest trend in large language models and moves away from the "I'll loose my job, developers will be replaced with AI" stance.
A few months old but a good piece to put things in perspective after the recent craze around large language models in general and GPT in particular. Noteworthy is the "wishful mnemonics" phrase mentioned and how it impacts the debate. Let's have less talks about AIs and more about SALAMIs please?
Nice article, gives a few clues to get a grasp on how GPT-3 works.
Words of caution regarding the use of language models for producing code. This can derail fairly quickly and earlier than you'd expect... without noticing it.
Hard not to have at least some ChatGPT related news this week. Plenty of impressive experiments out there... I think this one is hands down the most impressive one. I'm biased though, I like linguistics inspired works.
Interesting food for thought. Not necessarily easy to see it used in as many fields as the article claims. Maybe a bit too much on the techno solutionist side at times. Still, that sounds like an interesting guideline and path to explore.
Very early days for research on this topic and the sample is rather small still. That said the results are interesting, there seems to have a few biases inherent to the use of such an assistant, there's also clearly a link between the AI agency and the quality of what gets produced. We'll see if those result holds to larger studies.
Not much new in this article regarding Stable Diffusion. That being said the section about ethics is spot on. This is the toughest part to grapple with regarding the latest batch of generative AIs. The legal traceability of the training set might end up being required (even though I doubt it will happen).
It was only a matter of time I guess... this is sad.
I'd lie if I said I'm not slightly fascinated by what you can do with Stable Diffusion...
Interesting first exploration of a tiny part of the data set. If you read closely, this shows some of the potential biases in there.