Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's a proof-of-concept. What they've proved is that the same kind of intelligence required to play Go can be implemented with computer hardware. Before now, software couldn't beat a ranked human player at Go no matter how much computing power we threw at it. Now we can. Give it ten years and, between algorithmic optimizations and advances in processing, you'll have an unbeatable Go app on your phone.


Indeed. The first time a computer defeated a human in Chess it was this[1] size (1997). In 2009 it became possible to fit a grandmaster into this[2].

> Pocket Fritz 4 won the Copa Mercosur tournament in Buenos Aires, Argentina with 9 wins and 1 draw on August 4–14, 2009. Pocket Fritz 4 searches fewer than 20,000 positions per second. This is in contrast to supercomputers such as Deep Blue that searched 200 million positions per second. Pocket Fritz 4 achieves a higher performance level than Deep Blue.[3]

The first steps are always the most inefficient. Make it work, make it right, make it fast.

[1]: https://en.wikipedia.org/wiki/Deep_Blue_%28chess_computer%29... [2]: http://cdn.slashgear.com/wp-content/uploads/2008/10/htc_touc... [3]: https://en.wikipedia.org/wiki/Human%E2%80%93computer_chess_m...


GM Michael Stean lost to Cyber 176 (a mainframe 'supercomputer') in 1977 (at blitz). AFAIK this was the first time a computer defeated a GM; they began defeating IMs and experts some ten years before that. Kasparov himself lost to Fritz 2 at blitz as early as 1992.


"Under tournament conditions" is the condition everyone forgets. Go AIs were competing with ranked players given handicaps of varying degrees of absurdity.


> Give it ten years and, between algorithmic optimizations and advances in processing, you'll have an unbeatable Go app on your phone.

I find this overly optimistic because of the huge amount of power required to run the Go application. Remember, we're getting closer and closer to the theoretical lower limit in the size of silicon chips, which is around 4nm (that's about a dozen silicon atoms). That's a 3-4x improvement over the current state of the art.

The computer to run AlphaGo requires thousands of watts of power. A smartphone can do about one watt. A 3-4x increase in perf per watt isn't going to cut it.

If there will be a smartphone capable of beating the best human Go players, my guess is that it won't be based on general purpose silicon chips running on lithium ion batteries.

On the other hand, a desktop computer with a ~1000 watt power supply (ie. a gaming pc) might be able to do this in a matter of years or a few decades.


As solid as your argument may be, everyone saw arguments like this over and over. Every single time they were solid. For a time, it was the high frequency noise that would not be manageable (80s), then heat dissipation (90s), then limits on pipeline optimization (00s) and now size constraints on transistors. They were all hard barriers, deemed impossible and all were overcome.

I already know that your answer will be: "but this time it is a fundamental physics limit". Whatever. I'm jaded by previous doomsday predictions. We'll go clockless, or 3D, or tri-state or quantum. It'll be something that is fringe, treated as idiotic by current standards and an obvious choice in hindsight.


This looks like a good example of the Normalcy bias logical fallacy: https://en.wikipedia.org/wiki/Normalcy_bias

That previous constraints have been beaten in no way supports the argument that we will beat the laws of physics this time.


Our brains use roughly ~20 watts though, so we know that the power constraints can be overcome, if not in silicon then it may be biological machines we use in the future.


The previous problems were solved because people were willing to spend hundreds of billions of dollars to solve them. And they are still spending that kinds of money.

If the normalcy bias was in effect, they wouldn't be spending that money.


Actually Normalcy Bias may in fact feed that kind of money spending until such time as reality hits. Assuming that people will automatically act more logically when large amounts of money is in play flies in the face of recent history. Just look at the recent housing loan crisis. Normalcy Bias played a part there.

It's certainly possible that we'll break more barriers with clever engineering and new scientific breakthroughs. But that doesn't mean the Normalcy Bias isn't in play here.


Normalcy bias may have people spending lots of money on fabs assuming that the problems would be solved by the time the fabs are built.

However, I'm talking hundreds of billions spent on R&D to specifically to solve problems associated with chip manufacture. It took on the order of 25 years to solve each of the problems listed in the grandparent's post. Nobody would spend that kind of money or time on something that they think somebody else would solve.


People have probably spent billions of dollars to find a cure for cancer, but there isn't one that works for all cancers and most are still very bad news.

Say you spent a hundred billion dollars to extinguish the sun- that wouldn't work. How much money you spend is irrelevant when you're up against what people call "hard physical limits".


Isn't our inability to cure all cancers a limitation of our knowledge more than a hard physical limit?

I've read several articles saying that different cancers are not exactly the same disease, but more like different diseases with the same symptom (uncontrolled tumor growth) and different etiology, even sometimes different from person to person, not just from tissue to tissue. This was said to be a reason that a general cancer cure is so elusive. But is it really thought of as impossible, not just elusive?

Maybe our inability to extinguish the sun is also a limitation of knowledge more than a hard physical limit!

Even if I'm right about this, your description of the situation would still be accurate in that there would be no way to simply throw more money at the problems and guarantee a solution; there would need to be qualitative breakthroughs which aren't guaranteed to happen at any particular level of expenditure. If people had spent multiples of the entire world GDP on a space program in the 1500s, they would still not have been able to get people to the moon, though not because it's physically impossible to do so in an absolute sense.


>> there would need to be qualitative breakthroughs which aren't guaranteed to happen at any particular level of expenditure

Yep, that's my point, thanks. Sorry, I'm not in my most eloquent today :)


And the cost of building a fab is increasing exponentially; eventually that trend has to come to an end.


It also looks like a fully general argument against anything new ever being accomplished.


There is a lot of room for improvement with the implementation. The way we are using deep neural networks at the moment is exellent for prototyping, but far from optimal. For instance, this paper http://arxiv.org/abs/1511.00363 shows that you can replace floating point operations by simple bitwise operations without losing too much precision in DNNs for image recognition. Together with a better (that is, compiled, instead of interpreted) representation of the inference step I would expect an order of magnitude improvement at a small loss in precision. More software tuning, especially the kind of low-level optimizations that most chess programs do, should yield another big improvement.

Finally, the hardware we are using to run these programs is insane. Sure the silicon is approaching some hard physical limits, but your processor spends most of that power trying to make old programs run fast...

My prediction is that with enough ressources it is possible to write a Go AI which runs on general purpose hardware that's manufactured on current process nodes and fits in your pocket.


I don't think you appreciate how much of this is good algorithms, and how little you need sheer computing power to get good results.

If you look at http://googleresearch.blogspot.com/2016/01/alphago-mastering... you'll find that Google's estimate of the strength difference between the full distributed system and their trained system on a single PC is around 4 professional dan. Let's suppose that squeezing it from a PC to a phone takes about the same off. Now a pocket phone is about 8 professional dan weaker than the full distributed system.

If their full trained system is now 9 dan, that means that they can likely squeeze it into a phone and get a 1 dan professional. So the computing power on a phone already allows us to play at the professional level!

You can get to an unbeatable device on a phone in 10 years, if self-training over a decade can create about as much improvement they have done in the last 6 months, AND phones in 10 years are about as capable as a PC is today. Those two trade off, so a bigger algorithmic improvement gets you there with a weaker device.

You consider this result "overly optimistic". I consider this estimate very conservative. If Google continues to train it, I wouldn't be surprised if there is a PC program in a year that can beat any Go player in the world.


You're right, it won't be a general purpose computing device the way we conceive of it with the von Neumon architecture.

It'll likely be hardware that can be generalized to run any kind of deep net. The iPhone 5S is already capable of running some deep nets.

As a friend mentioned, it isn't the running of the net, it's the training that takes a lot more computational power (leaving aside data normalization). A handheld device that is not only capable of running a deep net, but also training one -- yeah, that will be the day.

There are non von Neuman architectures that are capable of this. Someone had figured out how to build general-purpose CPUs on silicon made for memory. You can shrink down a full rack of computers down into a single mother board, and use less wattage while you are at it.

This really isn't about having a phone be able to beat a Go player. Go is a transformative game that, when learned, it teaches the player how to think strategically. There is value for a human to learn Go, but this is no longer about being able to be the best player in the absolute sense. Go will undergo the same transformation that martial arts in China and Japan has gone through with the proliferation and use of guns in warfare.

Rather, what we're really talking about is a shot at having AIs do things that we never thought they could do -- handle ambiguity. What I think we will see is -- not the replacement of blue collar workers by robots -- but the replacement of white collar workers by deep nets. Coupled with the problems in the US educational system (optimizing towards passing tests rather than critical thinking, handling ambiguity, and making decisions in face of uncertainty), we're on a verge of some very interesting times.


Your making the same assumption people made about computing in the 50s, then 70s, then 90s, etc.


Please do elaborate. I try to base my assumptions (which I accept may turn out to be completely wrong) on physics and experience in working in semiconductors.

I just don't see a 1000x+ decrease in the power required happening in a decade or two without some revolutionary technology I can't even imagine. Is this what you meant? I'm sure most people couldn't imagine modern silicon chips in the 1950s vacuum tube era. But now we're getting close to the theoretical, well-understood minimums in silicon chips, so another revolutionary step is required if another giant leap like that is to be achieved.


    > physics and experience in working in semiconductors

    > without some revolutionary technology I can't even
    > imagine
I suspect (in the nicest possible way) that in a lineup of your imagination (on current assumptions) vs the combined ingenuiety of the human race driven by the hidden hand, the latter wins.


> > Give it ten years and […]

> I find this overly optimistic

exDM69 never said it's not gonna happen, he just said that it's not going to happen in ten years, and I agree with him. Revolutions never occurs that quickly. To achieve that we don't just need an improvement of the current state of the art, we need a massive change and we don't even know what it's going to look like yet ! This kind of revolution may occur one day but not in ten year.

And it could even never happen, remember that we don't have flying cars yet ;)


The thing is though we could already be 10+ years along the path to that next revolution, it wont start being talked about until its basically here


It seems to me that the people who say "it won't happen" do tend to have a much better reason to say it won't happen (or rather that it _probably_ won't happen) than the people who insist the next big revolution is just round the corner just because the last big revolution did happen.

The optimistic position is a bit like saying: "I 've lived 113 years, I'm not going to die now!". It's entirely possible for a trend to reverse itself. If machine learning has taught us something is that background knowledge (in this case, of processor technology) gives you much better results than just guessing based on what happened in the past.


Here's some possibilities:

Stacked 3D chips (HBM, etc), Heterogenous computing (OpenCL, Vulkan), Optical computing, Memristors, Graphene-based microchips, Superconductors, Spintronics, Quantum computers, Genetic computers (self-reconfigurable)


Heterogenous computing is already used in AlphaGo (and your smartphone). 3d chips will come to mainstream devices in a few years, but will give "only" a modest performance boost, say 2x or so.

The rest of the technologies you mention have great potential but will they be available in a smartphone in one decade? I don't think so.


You might ultimately only need some specialized "neural processing instruction set" for either the GPU cores or for the CPU cores. Or at least, I don't see any obvious obstacles to that.


I feel the same way about the chips reaching their physical limits. But I keep waiting for a new way we use them. We used to just churn out MHz and that was the metric. Then we got hyper-threading, multi-cores, GPU and other specific processors and new ways of programming to go with it all. I imagine we'll see the same. Just like the brain has different areas of processing, I'm hoping we'll see the same in silicon chips. Just like how we offload work from the general purpose cpu to the more efficient purpose build gpu or sound card etcs. Not saying every computer is going to have a GO chip in it, but maybe someday we'll have machine learning processors or who knows what. But yeah the advancements will be new designs and new ways of processing instead of more power.


Sure. But so far, we've found that revolutionary step every time we've hit these sorts of walls, and if I was a betting man I'd wager we'll do the same again.


Right, but just as a contrast: Technological progress speed has been at an all-time high since the begin of the industrial revolution.

It might as well slow down again and we have to remember that most humans in history saw little to no advances in technology over their lifetime.

I'm excited for the possibilities modern science opens up but I also think we might reach a point where fundamental progress stalls for a century or two.


I guess free worldwide information transfer (aka Internet) just opened this era and we are not close to see any kind of stalling (IMHO).


Amongst other things, you're assuming hardware is where the speed will come from. But it's as likely to come from better software.


How many watts does Lee Sedol's brain require?


About 25.

(2000 kilocalories / day -> ~100W; the brain uses about a quarter of your calories.)


A Go app likely wouldn't rely on the native processing power of the smartphone. An AlphaGo app could be created today for a smartphone. The bottleneck isn't the phone it's the cost of the cloud computing resources behind it. Perhaps a combination of Moore's law and economy of scale would make it affordable sooner than we think. The Xbox One, for example, already subcontracts difficult problems out to Azure.


The unbeatable GO app on your phone doesn't have to do the processing locally.


Yes, but that's just a silly argument and definitely not what GP meant. You can go and play a Go bot on KGS network with your smartphone today.


No, they haven't shown that the same kind of intelligence required to play go can be implemented in computer software. The methods AlphaGo uses are not the same as the intelligence a human uses at all. What they have done is prove an implementation of computer go in software is capable of beating a human player, not that they have implemented the same kind of intelligence as the human player.


"What they've proved is that the same kind of intelligence required to play Go can be implemented with computer hardware"

Not necessarily the same kind, and, if I had to make the call, I would say they aren't of the same kind.


> What they've proved is that the same kind of intelligence required to play Go can be implemented with computer hardware. Before now, software couldn't beat a ranked human player at Go no matter how much computing power we threw at it.

I don't think that's quite true as a description of what we knew about computer Go previously, though it depends on what precisely you mean. Recent systems (meaning the past 10 years, post the resurgence of MCTS) appear to scale to essentially arbitrarily good play as you throw more computing power at them. Play strength scales roughly with the log of computing power, at least as far as anyone tested them (maybe it plateaus at some point, but if so, that hasn't been demonstrated).

So we've had systems that can in principle play to any arbitrary strength, if you can throw enough computing power at them. Though you might legitimately argue: by "in principle" do you mean some truly absurd amount, like more computing power than could conceivably fit in the universe? The answer to that is also no; scaling trends have been such that people expected computer Go to beat humans anywhere from, well, around now [1], to 5 to 10 years from now [2].

The two achievements of the team here, at least as I see them, are: 1) they managed to actually throw orders of magnitude more computing power at it than other recent systems have used, in part by making use of GPUs, which the other strong computer-Go systems don't use (the AlphaGo cluster as reported in the Nature paper uses 1202 CPUs and 176 GPUs), and 2) improved the scaling curve by algorithmic improvements over vanilla MCTS (the main subject of their Nature paper). Those are important achievements, but I think not philosophical ones, in the sense of figuring out how to solve something that we previously didn't know how to solve even given arbitrary computing power.

While I don't agree with everything in it, I also found this recent blog post / paper on the subject interesting: http://www.milesbrundage.com/blog-posts/alphago-and-ai-progr...

[1] A 2007 survey article suggested that mastering Go within 10 years was probably feasible; not certain, but something that the author wouldn't bet against. I think that was at least a somewhat widely held view as of 2007. http://spectrum.ieee.org/computing/software/cracking-go

[2] A 2012 interview though that mastering Go would need a mixture of inevitable scaling improvements plus probably one significant new algorithmic idea, also a reasonably widely held view as of 2012. https://gogameguru.com/computer-go-demystified-interview-mar...


"Recent systems (meaning the past 10 years, post the resurgence of MCTS) appear to scale to essentially arbitrarily good play as you throw more computing power at them. Play strength scales roughly with the log of computing power, at least as far as anyone tested them (maybe it plateaus at some point, but if so, that hasn't been demonstrated)."

This is exactly the opposite of my sense based on following the computer go mailing list (which featured almost all the top program designers prior to Google/Facebook entering the race). They said that scaling was quite bad past a certain point. The programs had serious blindspots when dealing with capturing races and kos[1] that you couldn't overcome with more power.

Also, DNNs were novel for Go--Google wasn't the first one to use them, but no one was talking about them until sometime in 2014-2015.

[0] Not the kind of weaknesses that can be mechanically exploited by a weak player, but the kind of weaknesses that prevented them from reaching professional level.


> Play strength scales roughly with the log of computing power

That means that the problem is exponentially hard. EXPTIME, actually. You couldn't possibly scale it much.


> Play strength scales roughly with the log of computing power

To be fair, a lot of the progress in recent years has been due to taking a different approach to solving the problem, and not just due to pure computing power. Due to the way go works, you can't do what we do with chess and try all combinations, no matter how powerful of a computer you have. Using deep learning, we have recently helped computers develop what you might call intuition -- they're now much better at figuring out when they should stop going deeper into the tree (of all possible combinations).


There've definitely been algorithmic improvements, but from what I've read so far, the change in search algorithms, from traditional minimax search to MCTS, has been the biggest improvement, more than deep learning.


   Play strength scales roughly with the log 
   of computing power
The rumor I have heard is that the new Deep Mind learning algorithm really improves on this and scales linearly with computing power.


The game itself, however, scales exponentially, and there's nothing to do about that, so if you enlargen the board, no computer... And no human may be able to play it well.

The achivement was a leap towards the human level of play (and quite possibly over it). There might be additional leaps, which will take AIs WAY beyond humans, but none of those will scale linearily in the end. (And yeah, I guess you didn't want to say that either)


Branch and bound my friend, branch and bound. If you can build an awesome bounding function, even exponentially large spaces can be manageable.


Then you can say that, in 10 years, if we indeed have reached that point. Otherwise it'd just an empty prediction, and his perfectly valid point stands.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: