But you can't claim that something "wasn't possible 5 years" ago, if 7 years ago said feature was included in inexpensive consumer product (LG TV).
I'm not acting cynical, but it's tiresome for me to see people who claim that 20-30 years ago we all were living in a caves and catching bugs with wooden sticks, and now boom, ML!
Regarding "something truly different", well, my personal computing / mobile experience not changed that much from 2015. Honestly speaking, progress from 1995 to 2000 felt much more impressive and 'truly different'.
I mean, think of it, during this timeframe we went from DOOM via V.34 modems to amazon.com and ordering pizza online.
It matters if it actually works and if it works broadly or just for limited use cases. As I said in the original comment, I think that you will see many applications in the near future. The last 5 years have been focused on research (comparable to 1990-1995), only now are we getting ready for commercial applications.
Yeah - entire classes of problems went from unsolvable to solved. Some of that is in the consumer space and some of it is not.
I feel like an AGI could accidentally wipe out half of humanity and there would still be people commenting on HN about how the exact same technology already existed in a roomba seven years ago.
Honest question, no snark --- which consumer space problems were solved, if I don't play Go and don't have FB account to recognize me on a group photos (both of these two statements are true)?
Which shows some generality, the best way to accurately predict an arithmetic answer is to deduce how the mathematical rules work. That paper shows some evidence of that and that’s just from a relatively dumb predict what comes next model.
It’s hard to predict timelines for this kind of thing, and people are notoriously bad at it. Nobody would have predicted the results we’re seeing today in 2010. What would you expect to see in the years leading up to AGI? Does what we’re seeing look like failure?
Tesla autopilot is notoriously unsafe, and still extremely bad even at simple problems. Not sure how it could be called a "solved problem" - especially when you have the CEO of Waymo saying publically that he doesn't believe we'll "ever" reach level 5 autonomous driving.
There have been massive improvements in automated driving, but if you want to talk about solved problems, parking assisst is as far as you can get.
Translation is much better, and is often understandable, but it is far from a solved problem.
Colorizing/repairing old photos also often introduces strange artifacts in places where they are unnecessary. Again, workable technology, not a solved problem.
Voice transcription is also decent, but far from a solved problem. You need only look at YouTube auto-generated captions to see both how far it has come and how many trivial errors it still has.
And regarding "generalizable intelligence" and arithmetic in GPT-3, the paper can't even definitively confirm that the examples that they showed are not part of the corpus (they note that they made some attempts to find them that didn't turn out anything, but they can't go so far as to say they are certain that the particular calculations were not found within the corpus). They also make no attempts to check the model itself to find if any sub-structure may have simply encoded an addition table for 2-digit numbers.
Also, AGI will certainly require at least some attempts to get models to learn about the rules of the real world, the myriad bits of knowledge that we are born with that are not normally captured in any kinds of text you might train your AI on (the idea of objects and object permanence, the intelligent agent model of the world, the mechanical interaction model of the world etc.).
Autopilot is distinct from full self driving - I was talking about driver assist.
'Solved' is doing a lot of work here and is an unnecessary threshold I'm willing to concede. I think we would be more likely to agree that things went from unusably bad or impossible to usably good, but imperfect in a lot of consumer categories in the last five years due to ML and deep learning approaches.
The more clearly 'solved' cases of previously open problems (Go, protein folding, facial recognition, etc.) are mostly not in the consumer space.
As far as the gpt-3 bit I encourage others to read that excerpt, they explicitly state they excluded the problems they asked from the test data so it's not memorization. The types of failures it makes are failures like failing to carry the one, it certainly seems like it's deducing the rules. It'll be interesting to see what happens in gpt-4 as they continue to scale it up.
I keep hearing voice transcription is "solved", but both where I see it in consumer products (e.g. YT auto subtitles) and in dedicated transcription services I've tried it's far from solved.
Even the good translation services will randomly produce garbage, struggle with content that's not in nicely fully formed sentences, and are severely limited in language support.
First, nobody is claiming that people were living in caves before ML. I understand you're exaggerating for effect -- but that's the same thing the parent comment is doing when they say something "wasn't possible" 5 years ago. They don't mean that it was literally impossible, they mean that it was sufficiently bad that a typical consumer would be unlikely to use it back then -- whereas now the quality has improved to the point where these things are ubiquitous.
Similarly, both Amazon [1] and online pizza ordering [2] existed before 1995. They were just not commonly used.
>they mean that it was uncommon for a typical consumer to experience it back then.
Siri from Apple was launched in 2011, as some other commenter noted below. Also, "On June 14, 2011, Google announced at its Inside Google Search event that it would start to roll out Voice Search on Google.com during the coming days".
If it does not count as 'typical consumer to experience it', well, I do not know what counts then.
9 years ago, I mind you, not 5. And I think that 5 years ago voice recognition was more-or-less good already. In 4 years both Apple and Google acquired large enough datasets to learn from, afer initial launch of their products in 2011.
What we are still struggling with is proccesing of fuzzy queries, something among the lines of 'Siri tell me which restaurant in my area serves the most delicious sushi according to yelp reviews and also allows takeout', but this is not a voice recognition problem (though typical consumer can think it is).
> ‘Siri tell me which restaurant in my area serves the most delicious sushi according to yelp reviews and also allows takeout’
Siri stumbles at way less complex queries than that. Every year or so I retry using it, and give up due to the error rate.
An accuracy of 99% and 10x slower is apparently preferable for me.
My experience has been very different. I use an Alexa purely to control lights and set alarms, and have enough misses at just those tasks that I don't consider it particularly good at them.
I'd take a literal clapper that hooked into smartbulbs over it at this point.
But you can't claim that something "wasn't possible 5 years" ago, if 7 years ago said feature was included in inexpensive consumer product (LG TV).
I'm not acting cynical, but it's tiresome for me to see people who claim that 20-30 years ago we all were living in a caves and catching bugs with wooden sticks, and now boom, ML!
Regarding "something truly different", well, my personal computing / mobile experience not changed that much from 2015. Honestly speaking, progress from 1995 to 2000 felt much more impressive and 'truly different'. I mean, think of it, during this timeframe we went from DOOM via V.34 modems to amazon.com and ordering pizza online.