84 private links
Clearly aims to demonstrate the superiority of their specialized hardware for training. That said it's nice to have proper open models available (architecture, training data, weights... it's all in the open).
Now, this starts to become interesting. This is a first example of trying to plug symbolic and sub-symbolic approaches together in the wild. This highlights some limitations of this particular (quite a bit rough) approach, we'll see how far that can go before another finer approach is needed.
Training sets are obviously already contaminated... now it'll be a race of hiding such mistake under the carpet with human interventions. That'll be a boon for misinformation. That's what we get for a useless large models arm race.
Now this is a properly balanced piece which looks beyond the hype. Usable yes, if hallucinations don't have a high impact. Can the hallucinations be solved? To be seen, I personally have my doubts with the current architecture... at least banking it all on human feedback is being very naive about the scale of the task.
The climate constraints are currently not compatible with the ongoing arm race on large neural networks models. The training seems kinda OK, but the inferences... and it's currently just rolled out as shiny gadgets. This really need to be rethought.
Now, this is interesting research. With all that complexity, emergence is bound to happen. There's a chance to explain how and why. The links with the training data quality and the prompts themselves are interesting. It also explains a lot of the uncertainty.
The lack of transparency is staggering... this is purely about hype and at that point they're not making any effort to push science forward anymore.
Well, people asking relevant questions slow you down obviously... since the goal about the latest set of generative models is to "move them into customers hands at a very high speed" this creates tension. Instead of slowing down they seem hell bent at throwing ethics out of the window.
This is an excellent piece. Very nice portrait of Emily M. Bender a really gifted computational linguist and really bad ass if you ask me. She's out there asking all the difficult questions about the current moment regarding large language models and so far the answers are (I find) disappointing. We collectively seem to be way too fascinated by the shiny new toy and the business opportunities to pay really attention to the impact on the social fabric of all of this.
When they changed their statutes it was the first sign... now it's clear all ethics went through the window. It's about fueling the hype to drive money home.
Or why they are definitely not a magic tool for programming. Far from it. This might help developers a tiny bit, at the expense of killing the learning of students falling for it and the creation of a massive amount of low quality content.
Excellent piece as usual from Cory Doctorow. It quite clearly point out why Google is anxious and running off the chatbot cliff
Inaccuracies, contradicting itself, conflating events, misquoting sources... all of that mixed with some correct facts, it's a perfect misinformation spreading machine. This just can't be trusted at this point. Those experiments should be stopped in my opinion, better do proper homeworks first, then relaunch when this can be better trusted.
Are we surprised? Not really no... you don't own any of the data you're feeding it. Keep it away from your secrets.
There's really something rotten in this AI "arms race"... they're clearly making mistakes to go fast for PR purposes and using tools the wrong way. This can only lead to large scale disinformation if they don't correct course quickly. This has more political impacts than it looks at first sight.
So transformer models produce things that look plausible... and that's it. What would it look like if we started to make hybrid models in which a transformer model is also tied to proper computation tools with general knowledge? This piece is a good illustration of what it could provide.
For all the conversations about how chat GPT might displace jobs, there's a big untold: how much of copyright is violated in the process? It's also very concerning about how much data it collects when interacted with.
Or why you can't trust large language model for any fact or knowledge related tasks...
It's limits and biases are well documented. But, what about the ideologies of the people behind those models? What can it tell us about their aims behind those models? Questions worth exploring in my opinion.
A few interesting points in there. Too much hype and important points are glanced over, we'd all benefit from them being more actively explored.