Inspired by a recent talk from Richard Stallman.
From Slashdot:
Speaking about AI, Stallman warned that “nowadays, people often use the term artificial intelligence for things that aren’t intelligent at all…” He makes a point of calling large language models “generators” because “They generate text and they don’t understand really what that text means.” (And they also make mistakes “without batting a virtual eyelash. So you can’t trust anything that they generate.”) Stallman says “Every time you call them AI, you are endorsing the claim that they are intelligent and they’re not. So let’s let’s refuse to do that.”
Sometimes I think that even though we are in a “FuckAI” community, we’re still helping the “AI” companies by tacitly agreeing that their LLMs and image generators are in fact “AI” when they’re not. It’s similar to how the people saying “AI will destroy humanity” give an outsized aura to LLMs that they don’t deserve.
Personally I like the term “generators” and will make an effort to use it, but I’m curious to hear everyone else’s thoughts.


“Language evolves and changes” describes evolving and changing language, not keeping it the same. Language evolves into more specific definitions, not less specific.
For example: you might say “LLMs are one form of intelligence”, I don’t agree with that, but it makes logical sense. But claiming “LLMs are the same thing as intelligence” changes the definition of “intelligence” to a much broader umbrella. If you want to change that definition then you also need to invent a new word that means “non-LLM intelligence”.