We’ve been productively using AI for decades now – just not the AI you think of when you hear the term. Fuzzy logic, expert systems, basic automatic translation… Those are all things that were researched as artificial intelligence. We’ve been using neural nets (aka the current hotness) to recognize hand-written zip codes since the 90s.
Of course that’s an expert definition of artificial intelligence. You might expect something different. But saying that AI isn’t AI unless it’s sentient is like saying that space travel doesn’t count if it doesn’t go faster than light. It’d be cool if we had that but the steps we’re actually taking are significant.
Even if the current wave of AI is massively overhyped, as usual.
The issue is AI is a buzz word to move product. The ones working on it call it an LLM, the one seeking buy-ins call it AI.
Wile labels change, its not great to dilute meaning because a corpo wants to sell some thing but wants a free ride on the collective zeitgeist. Hover boards went from a gravity defying skate board to a rebranded Segway without the handle that would burst into flames. But Segway 2.0 didn’t focus test with the kids well and here we are.
The people working on LLMs also call it AI. Just that LLMs are a small subset in the AI research area. That is every LLM is AI but not every AI is an LLM.
Just look at the conference names the research is published in.
Maybe, still doesn’t mean that the label AI was ever warranted, nor that the ones who chose it had a product to sell. The point still stands. These systems do not display intelligence any more than a Rube Goldberg machine is a thinking agent.
These systems do not display intelligence any more than a Rube Goldberg machine is a thinking agent.
Well now you need to define “intelligence” and that’s wandering into some thick philosophical weeds. The fact is that the term “artificial intelligence” is as old as computing itself. Go read up on Alan Turing’s work.
That’s just kicking the can down the road, because now you have to define agency. Do you have agency? If you didn’t, would you even know? Can you prove it either way? In any case, this is no longer a scientific discussion, but a philosophical one, because whether or not an entity has “intelligence” or “agency” are not testable questions.
We’ve been using neural nets (aka the current hotness) to recognize hand-written zip codes since the 90s.
Not to go way offtop here but this reminds me: Palm’s “Graffiti” handwriting recognition was a REALLY good input method back when I used it. I bet it did something similar.
We’ve been productively using AI for decades now – just not the AI you think of when you hear the term. Fuzzy logic, expert systems, basic automatic translation… Those are all things that were researched as artificial intelligence. We’ve been using neural nets (aka the current hotness) to recognize hand-written zip codes since the 90s.
Of course that’s an expert definition of artificial intelligence. You might expect something different. But saying that AI isn’t AI unless it’s sentient is like saying that space travel doesn’t count if it doesn’t go faster than light. It’d be cool if we had that but the steps we’re actually taking are significant.
Even if the current wave of AI is massively overhyped, as usual.
The issue is AI is a buzz word to move product. The ones working on it call it an LLM, the one seeking buy-ins call it AI.
Wile labels change, its not great to dilute meaning because a corpo wants to sell some thing but wants a free ride on the collective zeitgeist. Hover boards went from a gravity defying skate board to a rebranded Segway without the handle that would burst into flames. But Segway 2.0 didn’t focus test with the kids well and here we are.
The people working on LLMs also call it AI. Just that LLMs are a small subset in the AI research area. That is every LLM is AI but not every AI is an LLM.
Just look at the conference names the research is published in.
Maybe, still doesn’t mean that the label AI was ever warranted, nor that the ones who chose it had a product to sell. The point still stands. These systems do not display intelligence any more than a Rube Goldberg machine is a thinking agent.
Well now you need to define “intelligence” and that’s wandering into some thick philosophical weeds. The fact is that the term “artificial intelligence” is as old as computing itself. Go read up on Alan Turing’s work.
Does “AI” have agency?
It’s still an unsettled question if we even do
That’s just kicking the can down the road, because now you have to define agency. Do you have agency? If you didn’t, would you even know? Can you prove it either way? In any case, this is no longer a scientific discussion, but a philosophical one, because whether or not an entity has “intelligence” or “agency” are not testable questions.
Not to go way offtop here but this reminds me: Palm’s “Graffiti” handwriting recognition was a REALLY good input method back when I used it. I bet it did something similar.