71 private links
This also carries privacy concerns indeed even for local models. It all depends how it's inserted in the system.
This is an important position paper in my opinion. The whole obsession towards the ill defined "AGI" is shadowing important problems in the research field which have to be overcome.
This move is really unsurprising... It's bound to become another channel for advertisements to try to cover the costs of running it all.
Of course it helps also against DDoS attacks... tells something about the state of AI scrapers I guess.
Interesting point of view... what makes a tool really?
Clearly to really benefit from LLMs there's quite some thinking around UX design to be had. This can't be only chat bots all the way.
I willadmit it... I laughed. And that's just one business risk among many.
This is a question which I have been pondering for a while... what will be left when the generative AI bursts. And indeed it won't be the models as they won't age well. The conclusion of this article got a chill running down my spine. It's indeed likely that the conclusion will be infrastructure for a bigger surveillance apparatus.
Sourcehut pulled the trigger on their crawler deterrent. Good move, good explanations of the reasons too.
This matches what I see. For some tasks these can be helpful tools, but it definitely need a strong hand to steer them in the right direction and to know when to not use them. If you're a junior you'd better invest in the craft rather than such tools. If you got experience, use with care and keep the ethical conundrum in mind.
Don't confuse scenarios for predictions... Big climate improvements due to AI tomorrow after accepting lots of emissions today is just a belief. There's nothing to back up it would really happen.
I hope people using Grok enjoy their queries... Because they come with direct environmental and health consequences.
This is very interesting research. This confirms that LLMs can't be trusted on any output they make about their own inference. The example about simple maths is particularly striking, the real inference and what it outputs if you ask about its inference process are completely different.
Now for the topic dearest to my heart: It looks like there's some form of concept graph hiding in there which is reapplied across languages. Now we don't know if a particular language influences that graph. I don't expect the current research to explore this question yet, but looking forward to someone tackling it.
Unsurprisingly, hiring scams are becoming more elaborate. Keep it in mind for your upcoming interviews.
The "asleep at the wheel" effect is real with such tools. The consequences can be dire in quite a few fields. Here is a good illustration with OSINT.
Even if you use LLMs, make sure you don't depend on them in your workflows. Friction can indeed have value. Also if you're a junior you should probably seldom use them, build your skill and knowledge first... otherwise you'll forever be a beginner and that will bite you hard.
We just can't leave the topic of how the big model makers are building their training corpus unaddressed. This is both an ethics and economics problem. The creators of the content used to train such large models should be compensated in a way.
Between this, the crawlers they use and the ecological footprint of the data centers, there are so many negative externalities to those systems that law makers should have cease the topic a while ago. The paradox is that if nothing is done about it, the reckless behavior of the model makers will end up hurting them as well.
Unsurprisingly, Wikimedia is also badly impacted by the LLM crawlers... That puts access to curated knowledge at risk if the trend continues.
I somehow recognise myself in this piece. Not completely though, I disagree with some of the points... but we share some baggage so I recognize another fellow.
Sure, a filter which turns pictures into something with the Ghibli style looks cute. But make no mistake, it has utter political motives. They need a distraction from their problems and it's yet another way to breach a boundary. Unfortunately I expect people will comply and use the feature with enthusiasm...