AlphaZero v Stockfish (2018): Lecture by GM Ben Finegold

  Рет қаралды 38,081

GMBenjaminFinegold

GMBenjaminFinegold

5 ай бұрын

This lecture was recorded on December 20, 2023 in Roswell, Georgia. Thank you to the Zatloukal Family for sponsoring!
Games/Positions:
09:07 AlphaZero vs Stockfish
24:56 Stockfish vs AlphaZero
37:01 Stockfish vs AlphaZero
Check out Ben's Chessable courses here! www.chessable.com/author/BenF...
If you're interested in sponsoring a lecture of your choice, email Karen at karen@atlchessclub.com
Signup or gift a chess.com Premium membership to get access to their lessons, remove ads, improve your chess, AND help Ben at the same time!! Upgrade your chess.com membership or join! - go.chess.com/finegold.
Buy Merch today! ccscatlmerch.com/
Watch live at / itsbenandkaren
Donations are appreciated at: streamlabs.com/itsBenandKaren...
Follow me on Twitter: / ben_finegold
#benfinegold #chess #AlphaZero #Stockfish

Пікірлер: 229
@AdamCHowell
@AdamCHowell 5 ай бұрын
Beating stockfish is easy. It’s cheating to use an engine. Stockfish is an engine. I automatically win as a result. Checkmate stockfish.
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
But can you prove that Stonkfish ls cheating?
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
Also: Cheating basically means getting help from a third party? Like when I as a 500 elo fish get hints from a master while I play some ofishel chessgame like a real nerd and stuff.
@buuythbuuyth1412
@buuythbuuyth1412 5 ай бұрын
Stockfish doesn't cheat. Its chess speaks for itself.
@vibovitold
@vibovitold 14 күн бұрын
It's also illegal to get human assistance during a game, so wouldn't you get disqualified too? : ) And if the assistance provided by your own brain doesn't count... you see where it is going then; does Stockfish use itself?
@zvonimirtosic6171
@zvonimirtosic6171 5 ай бұрын
V. Kramnik: 'Alpha 0 beats Stockfish 8? With only 9 hours of self-learning? And without openings book? How interesting ...'
@Corteum
@Corteum 5 ай бұрын
It has a chip in its butt lol
@askthepizzaguy
@askthepizzaguy 5 ай бұрын
Congrats on using the word "suffice" twice in a row in a grammatically correct sentence. Achievement unlocked.
@ernietollar407
@ernietollar407 5 ай бұрын
GM Finegold at his funniest while talking about computer chess.. Way better than the stockfish channels cheesy scripts
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
That channel is just loads of AI generated crap. The games they show are probably just Stockfish vs Stockfish itself.
@f.d.3289
@f.d.3289 5 ай бұрын
come on, millionaires, sponsor GM Finegold analyzing all 1000 games in full
@jayyy5270
@jayyy5270 5 ай бұрын
best comment 2023
@-zelda-
@-zelda- 5 ай бұрын
I don't even think they made all 1000 games public, I think just around 100 or so are available
@xianglong2871
@xianglong2871 5 ай бұрын
Has anybody consulted Vladimir Kramnik and his mathematicians about these games?
@Corteum
@Corteum 5 ай бұрын
Kramnik is several hundred points lower rated. His opinions dont apply here. lol
@JM-tj5qm
@JM-tj5qm 5 ай бұрын
AlphaZero was great for it's time. But I don't think most people realize Stockfish is way stronger than AlphaZero ever was now.
@keithoma7265
@keithoma7265 5 ай бұрын
I think StockFish would roflstomp AlphaZero even from the starting position. We live in crazy times.
@josephmathes
@josephmathes 5 ай бұрын
For real. It incorporated techniques from AlphaZero, just didn't replace itself with them entirely. Current stockfish is like old stockfish and AlphaZero combined.
@kingsolo6241
@kingsolo6241 5 ай бұрын
It just takes time for Alpha Zero. It has a much stronger computer that Stockfish will never see. It a matter of minutes Alpha Zero will always surpass Stockfish.
@MadMetalMacho
@MadMetalMacho 5 ай бұрын
Wasn't Alphazero trained for just 5 hours? It's just that they stopped after it became the best in the world and moved on to other things, but had they continued, with 5 years of development/training it's hard to see how Alphazero wouldn't be on top today...
@JM-tj5qm
@JM-tj5qm 5 ай бұрын
@@kingsolo6241 Alphazero is not a computer and neither is Stockfish. Both are computer programs. Alphazero advantage at the time was the fact that it was a NNUE (Efficiently updatable neural network) Once Stockfish updated it's evaluation function too it surpassed AlphaZero. And Since then Stockfish has gotten way stronger still. AlphaZero doesn't have any advantage now, while Stockfish has the advantage of being open source and is constantly tested by it's community.
@sage5296
@sage5296 5 ай бұрын
The main reason A0 is more willing to sac material is that since it's using AI/NN to evaluate the position, it can more or less see the compensation, where as SF can't due to the horizon effect. Very similar things happen in some of the SF v Leela0 games. One of the main things at play I think is long term piece mobility.
@raskalnekov
@raskalnekov 5 ай бұрын
Not only did I stay awake long enough to see the Stockfish win, I also stood awake long enough to write this comme
@xwngdrvr
@xwngdrvr 5 ай бұрын
Show ALL the games! Five THOUSAND minutes with Ben!
@art58wong
@art58wong 12 күн бұрын
Great lecture. I remember the AZ headlines and even saw a video of your first relentless example. Your explanation filled my swiss cheese understanding of why AZ isn't mopping up the chess world and leaving the game in tatters. Again my thanks for giving much needed perspective. Your series is quite refreshing.
@wreckim
@wreckim 2 ай бұрын
Thanks GM!
@-_Nuke_-
@-_Nuke_- 5 ай бұрын
Oh yes! I was waiting for these!
@miroljubbelic2007
@miroljubbelic2007 5 ай бұрын
Thanks , very instructive . Your comments are so funny. Keep going
@kmarasin
@kmarasin 5 ай бұрын
Those endgames by AlphaZero remind me of Capablanca, but on steroids. Capa loved to push his pawns and gain space.
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
Great video. Love the Berlin Defense. They are never as drawish as top players make it look. Especially this endgame is very tricky. I actually like much less the lines that don't head to this early queen trade. The old mainline. Then the structure is just symmetrical. The dull exchange french basically.
@ek5688
@ek5688 5 ай бұрын
Go Ben!
@johnolah8853
@johnolah8853 5 ай бұрын
But stay right there!
@talhakonjic3197
@talhakonjic3197 5 ай бұрын
hi Ben, this game is awesome, its almost as awesome as the games you play are!
@grahamsmith2753
@grahamsmith2753 3 күн бұрын
As I understand it, AlphaZero was playing on massive hardware, and Google was still so worried, it had the Stockfish opening book disabled. Lc0 has just lost the TCEC season 26 Superfinal to Stockfish by 57 points to 43. The Superfinal consisted of 100 games, played at a time control of 120' + 12". The headlines that Stockfish was crushed by AlphaZero were conveniently always taken out of context...but that's Google for you.
@jdubrule
@jdubrule 5 ай бұрын
IIRC, the resignation rule was that both engines had to agree that one side was up by at least 10 pawn-equivalents. So, one engine could be forced to play on, even though the other engine was the only one that had hope for its position
@archsys307
@archsys307 5 ай бұрын
Just like human chess 😂😂😂
@curiousnerd3444
@curiousnerd3444 5 ай бұрын
There is a very good reason for not using opening book. chess engines at that time played at high depths and evaluated positions based on materials (Queen 9 points, rook 5, pawn 1) or simple rules which were coded into the engine by humans. This made chess engines great at tactics but chess grandmasters would still outplay them because the engines didn’t understand strategy. This is the reason stockfish uses opening book to compensate for lack of understanding chess People at that time used to believe that learning chess was therefore a sign of human intelligence. The point AlphaZero was trying to make was an AI can learn to play chess at a grandmasters level with zero human training just by getting the rules of the game. Pitting Stockfish against AplhaZero without opening book is a test of “Artificial Intelligence” learning strategies that beat calculations
@ClarkPotter
@ClarkPotter 5 ай бұрын
Even in 2018 the eval functions were more sophisticated than what you outline. I would guess they used a genetic algorithm with parameters as genes to converge upon optimum values, and to determine which to jettison to improve speed.
@mohammadhaider8946
@mohammadhaider8946 5 ай бұрын
Your claim that engines were great at tactics but they would still be outplayed by grandmasters due to a lack of strategical understanding is just plainly wrong. Even years before this match, the best engines had reached a point far beyond human ability to match. There might have been deficiencies, but they were not going to be outplayed by humans.
@curiousnerd3444
@curiousnerd3444 5 ай бұрын
I guess I could frame it better, the main point is after finishing AlphaGo, they made AplhaZero which didn’t use human games and trained entirely on self play aka it learned long term strategies on its own. They could quickly train it to play other games than go (including chess) They were not trying to compete with stockfish and that’s why they quickly moved to other things instead of improving AplhaZero unlike stockfish
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
Evaluating based only on material will give you a significantly inferior engine than Stockfish 8. HCE was the result of three decades of development and this is just disrespectful
@vibovitold
@vibovitold 15 күн бұрын
You've oversimplified it to the point where it's blatantly untrue. The rules were anything but trivial. If you ever looked into (old at least) Stockfish source code, it contained a huge pile of "manual" corrections - surprisingly specific, like "in an endgame with just Queens and minor pieces value the mobility of minor pieces 16% less" etc. Not literally ofc, but this sort of stuff. Of course it's still heuristics (albeit very complex and multi-layered), and it doesn't quite convey the notion of long-term strategical sense (this part is correct). But modern engines were very far from evaluating positions only based on the material value of pieces. This is really unfair to their creators (who btw often were strong players themselves, or strong players assisted with the development). If you built an engine based on such simple criteria, it would never be stronger than an amateur. It's true that engines relied on opening books because they weren't very good at grasping the strategical implications of opening choices (the "fog" was too dense from the starting position). But still, even if you disabled an opening book, an engine would still play a pretty strong opening based on general principles (mobility, control of the center etc.). The main reason for including an opening book is just optimization. Similarly, (serious) engines use endgame tables, even though endgame is "pure" calculation, and relatively easy at that (as you can calculate much farther when there are only a few pieces left on the board). But the reason is optimization. In an engine vs. engine match time matters: just like in human chess, they both have only got so much time allocated for the game. So why waste time on computations when you can replace them with the equivalent of "home preparation".
@sachatostevin6435
@sachatostevin6435 5 ай бұрын
Thanks, Ben! Great video as always. You might be interested to know that I've just downloaded a free version of LeelaZero, and my first test was to setup board position: 1. d4 d5 2. Nc3 Nf6 3. Bg4... Onwards from here I let LeelaZero play both sides. Strangley enough it accepted a 3-fold-repetition at move 29. Maybe I'm supposed to tweak some settings, I don't know.
@sage5296
@sage5296 5 ай бұрын
the setting to look for is "Contempt", which basically just tells the engine how much it should avoid draws, or in other terms, how much worse of a position it's willing to accept to avoid a draw (on the assumption that it can outplay the opponent in the long term and win anyways)
@sachatostevin6435
@sachatostevin6435 5 ай бұрын
@@sage5296 oh yeah. I think i remember ben using that word now. Thanx!
@baoboumusic
@baoboumusic 5 ай бұрын
Stockfish 16 really likes what AlphaZero did and predicts pretty much all moves. I expect it's about twice as good as Stockfish 8.
@addisonmigash8227
@addisonmigash8227 5 ай бұрын
8 * 2 = 16
@nilsp9426
@nilsp9426 5 ай бұрын
If you can now tell me what "twice as good means" :'D
@baoboumusic
@baoboumusic 5 ай бұрын
@@nilsp9426 the answer is fries.
@DarthVader-wk9sd
@DarthVader-wk9sd 5 ай бұрын
Way more than twice as good but ok
@archsys307
@archsys307 5 ай бұрын
@@DarthVader-wk9sdHow do you quantify twice as good Obviously elo only went up some hundreds of points. Not to 6000 What does it mean to be twice as good? A 2 to 1 win probability? Then you could use the elo formula to compute the exact elo gap, should be about +100 points.
@pinofshame6253
@pinofshame6253 5 ай бұрын
46:44 my guess AZ saw the idea of promoting the H pawn and didn't want to open the a1-h8 diagonal for the dark square bishop, so that he can put his rook to h8. Notice how the Bishop on f2 can't get to f6 or g7 any way
@-zelda-
@-zelda- 5 ай бұрын
8:31 Some versions of Stockfish will gladly play the Berlin draw most of the time with black so you can get lots of easy draws that way
@jessealtum2627
@jessealtum2627 5 ай бұрын
I wonder why at 30:30 stock fish chose not to capture the dark squared bishop with the knight eliminating the bishop pair. Is the light squared bishop that much stronger in this position, maybe because of the A2 target?
@ennerz-hq8pq
@ennerz-hq8pq 5 ай бұрын
It wouldnt have used human reasoning like you and I. The answer lies in huge branches of calculations
@Uerdue
@Uerdue 5 ай бұрын
Does anyone know a game of this match (or the one from 2017) where AlphaZero sacrificed and Stockfish won by defending and holding on to the extra material? Like, did it ever misevaluate / overestimate its attacking chances?
@timharris559
@timharris559 3 ай бұрын
Precisely this, it lost most of its games by attacking in a drawn position, but A0 believed it had an overwhelming attack but stockfish had a strong defence to counter.
@rhiehn
@rhiehn 5 ай бұрын
Most versions of stockfish will accept the berlin draw, so magnus(or anyone else) can draw it with white, but I don't know that i'd really count that, but if a GM said that I'd assume that's what they meant, since it's obviously absurd otherwise.
@CTJ2619
@CTJ2619 4 ай бұрын
i like BF discussion about compensation and judgement
@unclvinny
@unclvinny 5 ай бұрын
My nasty comment is that Janet Jackson had a song years ago called "Nasty Boys".
@Corteum
@Corteum 5 ай бұрын
Lol. I dont think anyone even knows who that is now!
@dannygjk
@dannygjk 5 ай бұрын
Incorrect off the bat. Alphazero is/has an engine. It's just that at the time AZ used neural net tech whereas SF didn't use NN tech, (SF started using NN tech starting with SF 12). Technically any machine that plays chess has an engine built into it.
@nilsp9426
@nilsp9426 5 ай бұрын
From my look at AlphaZero vs. Stockfish it seems A0 is increadibly good at punishing any setup with b7, Bc8 and e6. Black not getting their bishop out is such a common theme in these games.
@JohSmith
@JohSmith 5 ай бұрын
Interesting
@ocudagledam
@ocudagledam 4 ай бұрын
So, I was curious about the last game and it turns out that Ben was right about a couple of things. I let SF 14 take a look at the position where, at 47:18, Ben says that it looks to him like it should still be a draw and the newer SF engine agrees, and when Ben says that, after white pushes h5, he would go g:h5, well, he's absolutely right. The way to draw was to exchange pawns on h5 and then go Kc4, with the intention of bringing it to d3. Instead, Kb3 that the A0 played instead of g:h5 straight up loses, the eval jumps from something like 0.03 to around 3.5.
@jefftaylor1186
@jefftaylor1186 5 ай бұрын
If I remember correctly, when they didn’t have the opening book, both “engines” decided that the Evans gambit was optimal. Which is awesome because I only play 1.e4 with the hopes of being able to go 4.b4
@ethangilworth7891
@ethangilworth7891 5 ай бұрын
Extremely incorrect. Maybe at very low depth, but certainly not at the massive depths they are typically run.
@jefftaylor1186
@jefftaylor1186 5 ай бұрын
@@ethangilworth7891 You’re right. It was just alphazero that did it. Not stockfish.
@ethangilworth7891
@ethangilworth7891 5 ай бұрын
@@jefftaylor1186 I’m an Evan’s player myself, so I would love to see your source for this.
@askthepizzaguy
@askthepizzaguy 5 ай бұрын
At a certain point black's best idea is to give away all his pieces every move in order to stop white's attack. That's why being single minded about gaining a material advantage was the flaw in the armor of Stockfish 8. At a certain point, it was willing to trade all that material back to stop getting its ass beat, and considered giving away material like that to be its best move. Alphazero was definitely on the right track with its judgment call that the attack was strong enough even at a 4 pawn deficit to be worthwhile.
@RoxxorzYourBoxxorz
@RoxxorzYourBoxxorz 5 ай бұрын
AlphaZero actually flipped the board before Stockfish was able to play Be7
@ThunderChickenBucket
@ThunderChickenBucket 5 ай бұрын
nice
@mrpocock
@mrpocock 5 ай бұрын
One thing i love about computer chess is when the engines radically disagree about the evaluation.
@alfiecollins5617
@alfiecollins5617 5 ай бұрын
37:03
@askthepizzaguy
@askthepizzaguy 5 ай бұрын
Didn't stockfish "resign" if its own evaluation of its position was like, -5 whole points or something like that? I may be getting the number wrong, but I thought they called the game lost if either program felt it was losing by a sufficient amount. I could be confusing this with a different computer chess matchup event, but I know I didn't invent this out of whole cloth, I definitely read or heard about it somewhere. edit- literally one second after I wrote this comment you mentioned white resigned. Great timing on my part. /sarcasm
@Corteum
@Corteum 5 ай бұрын
_"Didn't stockfish "resign" if its own evaluation of its position was like, -5 whole points or something"_ They can set the resign score threshold for the match. Once it reaches that threshold (e.g. -8.00), and that score stays or worsens over three more subsequent moves, then it resigns. You can download a good chess gui like Arena 3.5 (it's free) and run your own matches with Stockfish 8 and other newer NNUE engines.
@McWerp
@McWerp 5 ай бұрын
I always wonder at what point, what move, the engine realizes it was wrong and its losing. What moves did one engine miss that the other saw.
@ethangilworth7891
@ethangilworth7891 5 ай бұрын
With Stockfish, it uses Alpha Beta pruning for its tree search, so it saw every move. The question is more “what move did it evaluate incorrectly.”
@sage5296
@sage5296 5 ай бұрын
@@ethangilworth7891 not exactly, it can easily be the case that it prunes away the key line too early. Ig that would be an incorrect evaluation to some degree but non-NN engines rely more on seeing far ahead than evaluating accurately in any given position
@vibovitold
@vibovitold 14 күн бұрын
@@ethangilworth7891 that's pretty much the whole definition (and the purpose) of pruning - to NOT see every move.
@grahamdugan
@grahamdugan 10 күн бұрын
I would have been more interested in the opening sequences
@ra1u
@ra1u 5 ай бұрын
In final position of last game, after Bf8 Rxf8 . h8Q Rxh8 Rxh8 is still hard to break the fortress as there are no pawns available for attack.
@MrHeroicDemon
@MrHeroicDemon 5 ай бұрын
I will allow the ads to run only for some FineGold.
@Corteum
@Corteum 5 ай бұрын
You dont use ad blockers?
@user-hl7mc8mm6x
@user-hl7mc8mm6x 5 ай бұрын
Remember, stockfish, at that time, was calculating around 4 million positions a second to alpha zeros 80 thousand... If AO were to have the same calculating power as SF then no stockfish would be able beat Alpha Zero😮
@ocudagledam
@ocudagledam 4 ай бұрын
They probably had the same computational power on hand. The thing is that A0's evaluation function is much more complex and much more computationally demanding than that of a traditional engine, so if A0 is to match SF8 in term of positions per second, A0 has to have many, many times more computational power at its disposal
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
They do have the same calculating power
@askthepizzaguy
@askthepizzaguy 5 ай бұрын
22:00 black decides that the white bishop is worth more than the black rook, and wants to trade those two pieces. White also decides that the white bishop is worth more than the black rook, and decides to take the black rook with his own rook instead of with his bishop. Both computer programs agree that white's bishop is worth more than a rook. It's not three points versus five points at all, each piece has a fluid value depending on where it is placed in the current position. Sometimes it's basically worthless, like when it is trapped, or when its only value is in stopping checkmate by blocking an attack and putting itself in an absolute pin to stop the attack. Sometimes, like this bishop, the piece is so strongly placed that it is worth more than a supposedly much stronger piece that is having to play defense to stop the assault of the bishops. Being passive and forced into the role of being a meat shield for the king truly lowers the value of a piece that normally gets its value from the range of spaces it can move to and the speed at which it can attack an opponent's position. Lowering that mobility and putting it on passive guard-only duty really wrecks the value of the piece, to the point you'd gladly trade it away for the minor piece that is utterly dominating your king.
@juliandiazromero4101
@juliandiazromero4101 5 ай бұрын
OG EBN!
@movatrova3539
@movatrova3539 5 ай бұрын
Nasty comments! Rawr!
@vibovitold
@vibovitold 14 күн бұрын
I'd really like to show one of these games to someone like Capablanca or Alekhine (without telling them where it came from), i wonder what they'd think of it.: ) would they "get it"? would they think it was played by genius players? or dismiss it as nonsense
@user-sf8cs3sz2i
@user-sf8cs3sz2i 5 ай бұрын
AlphaZero "cheated". Stockfish 2018 would have drawn almost all the lost games had it been given a proper computer to run on. The deep mind team put Stockfish on the crappiest computer they could find, the worst time controls Stockfish played at, and only published the results where AlphaZero won. Oh and they removed Stockfish's opening book and endgame database. AlphaZero on the other ran on a custom gpu-driven machine. The match was a complete farce.
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
Idk it’s pretty hard to come up with fair hardware for GPU/CPU engine matches. Even modern engine tournaments get criticized over them.
@user-sf8cs3sz2i
@user-sf8cs3sz2i 4 ай бұрын
​@@user-ws7kp1yh9l I agree it is hard to compare apples to apples CPU vs GPU, but look at what Stockfish was given vs. Google's machine. It was totally lopsided. If you Google "Hardware used in AlphaZero vs Stockfish match", and read this line from the StackExchange discussion: "[AlphaZero] used 4 TPUs for the games, so a processing power of 180 TFLOPS. Note TFLOPS = 1000 billion floating point operations per second." The GPU network AlphaZero ran on was roughly 10,000 times more powerful than the hardware SF was allowed, and AlphaZero was *designed* to work without an opening book, whereas Stockfish was designed to work with an opening book an endgame database, and at classical time controls. The game structure was obviously designed handicap SF to the maximum.
@ReyBasilisko
@ReyBasilisko 5 ай бұрын
You are not cool Ben. There you go! 1:58
@TymexComputing
@TymexComputing 5 ай бұрын
Can A0 play F3 - only as white?
@Corteum
@Corteum 5 ай бұрын
It can if it's forced to. But it will never choose to play f3 if left to its own calaculations.
@user-rh4qv4gd8t
@user-rh4qv4gd8t 5 ай бұрын
My favorite game in this match is when Alpha zero plays as white in the Queen's indian
@MrBonified66
@MrBonified66 5 ай бұрын
Kasparov's claim was that GM's on the Deep Blue team were inserting specific recommendations for certain positions, and the key position in Game 6 was one of them. GK knew his move wasn't sound but he knew DB wouldn't respond correctly - which it wouldn't have without the specific human input. In any case, my favourite bit is that a few days after Kasparov bumps into Charles Bronson, they have a chat, GK says something like "I think I can beat it in a rematch". Bronson says "they aren't going to let you have a rematch". And he was right.
@archsys307
@archsys307 5 ай бұрын
Y they didn’t let him rematch
@zackarysemancik5491
@zackarysemancik5491 2 ай бұрын
​@@archsys307because they won
@vibovitold
@vibovitold 15 күн бұрын
They had never been obliged to keep on granting him rematches until he's satisfied with the outcome. And why would they grant him a rematch anyway after he publicly accused them of cheating without providing a shred of evidence? "I just know that computers can't play that way" is an allegation, nothing else.
@bobi5202
@bobi5202 5 ай бұрын
Rawrrrr
@Writerscabin
@Writerscabin 5 ай бұрын
Leela is also free right ?
@Corteum
@Corteum 5 ай бұрын
Indeed. But you need a decent GPU to run it.
@ernietollar407
@ernietollar407 5 ай бұрын
has Alpha Zero run away hiding from new Stockfishes
@Corteum
@Corteum 5 ай бұрын
The Stockfish derivative engines (engines based on Stokcfish but with modifications), are stronger than standard SF, for sure. Have you played around with any of them?
@victorfinberg8595
@victorfinberg8595 5 ай бұрын
never play f3, never start a land war in asia, and never trust a headline (and 7 times never kill man)
@ahrrydepp493
@ahrrydepp493 5 ай бұрын
Thats the old stockfish, 12 mybe, but now you have stockfish 15 who can easly defeat alpha
@DekarNL
@DekarNL 5 ай бұрын
A human brain works on roughly 20 Watts of power. I wonder how strong an engine is when limiting it's computing power to a cell phone's vs a GM. @Ben: thoughts?
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
Stockfish on your phone is already stronger than all GMs out there
@DekarNL
@DekarNL 4 ай бұрын
Noted, but what if the brain power required to play chess is between 1% and 10% the computing power of a phone. Limiting computing power to that: Who would win? It'd be a cool experiment 😇
@perakojot6524
@perakojot6524 5 ай бұрын
That black pushing pawn to e4 (move 28.) is an instant loss. It's like a ??? type of a blunder for a strong engine. There is really no reason to analyse the game any more since after that move it's practically a forced loss (even though engines of that era - 5 years ago didn't really have a clue).
@Rubrickety
@Rubrickety 5 ай бұрын
"It seems like AlphaZero chooses the move that's most likely to win" is an almost spot-on description of how these deep-learning game AIs actually work. They don't really "evaluate" any position; they just simulate the entire game to the end and see what happened. My understanding is that getting AIs to give a traditional "-2.3" type of evaluation pretty much required a hack; what it's actually "thinking" is "from this position, Black wins 74% of games" or whatever.
@sage5296
@sage5296 5 ай бұрын
my understanding is that it's basically using the standard alpha-beta pruning, but being much more selective/aggressive with the pruning. The AI part is the algorithm that evaluates a position, which is what basically assigns that %win or +- score. The tree is then pruned accordingly in the same fashion as SF would, except it checks less positions overall (often 100x or higher less positions), generally to a comparable depth tho In other words, it plays moves towards a position that's most likely to win, but it still evaluates many moves deep
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
@@sage5296No, Alphazero does not use alpha-beta pruning. Evaluation is too computationally expensive for that
@vibovitold
@vibovitold 15 күн бұрын
No, chess AIs (or engines based on neural networks) don't work that way. They don't simulate the entire game to the end (that's known as a Monte Carlo simulation, this technique predates the AI as we know it, and it can only be executed with a rather low accuracy / randomness, due to computational complexity). Quite the contrary - they've got a highly developed (via deep learning) evaluation function, which is essentially a black-box algorithm that "wrote itself" (by trial and error, in the training phase). Noone really knows how it works, and what patterns trigger the AI into thinking a position is "promising" or "seeds of defeat". It's the same as with eg. image recognition. Noone knows what exactly it was that made the AI recognize a dog in the picture. AZ, in a way, played more like a human (human brains are biological neural networks, after all). It uses highly sophisticated pattern recognition insteadf of a huge tree search with a slimmer (manually encoded) evaluation. That's why it analyzes much fewer nodes per second than the likes of Stockfish. But its analysis is of higher quality, and it compensates for that.
@Rubrickety
@Rubrickety 15 күн бұрын
@@vibovitold In fact, AlphaZero (and all similarly-architected game-playing AIs) use Monte Carlo Tree Search (MCTS) as a fundamental part of their training (and, it seems, during actual play, but see below). MCTS is exactly how AZ learns "by trial and error, in the training phase", and is indeed based on playing a large number of games to the end to determine win/loss efficacy for a given move, which it uses to update the network. It's surprisingly difficult to find clear info on whether current AIs use MCTS during play to choose candidate moves. DeepMind's paper on AlphaGo Zero (the Go-playing AI which preceded generalized AlphaZero) states specifically that it does not, but the AlphaZero description is unclear on the matter, and Leela Zero definitely does use runtime MCTS (i.e., "playouts" or "rollouts"). Apparently without them its Elo was only ~2500. In any case, though, the "evaluation function" of deep learning AIs does literally boil down to "Will this move lead to a win, on average?"
@vibovitold
@vibovitold 14 күн бұрын
@@Rubrickety as part of the training, sure, i meant the "over the board" phase. According to LZ website "Many people call this MCTS (Monte-Carlo Tree Search), because it is very similar to the search algorithm the Go programs started using in 2006. But the PUCT [Predictor + Upper Confidence Bound tree search] used in AGZ and Lc0 replaces rollouts (sampling playouts to a terminal game state) with a neural network that estimates what a rollout would do" so if it doesn't "sample playouts to a terminal game state", only "estimates what a rollout would do", it directly contradicts the claim that "they just simulate the entire game to the end and see what happened". no? i do admit that while i'm a programmer with some interest in chess (and conversely chess engines), i am not working on chess engines, let alone an expert in the field
@atwarwithdust
@atwarwithdust 5 ай бұрын
And AlphaZero says no kibitzing.
@goodlaugh1094
@goodlaugh1094 5 ай бұрын
Stockfish 13 was the last version of stockfish to be weaker than Alphazero. Stockfish 14 was already considered stronger. Stockfish 13 and Alphazero do not maximize their advantages from the opening, while stockfish 14, 14.1 have near to complete drawable foresight to the game. Stockfish 15 does better sacrifices slightly. 15.1 and 16 reached a new super level by world record sacrifices and depth to the game. Leelazero, Dragon, and even Torch are a bit behind since they are still only around the level of brilliancy as Stockfish 15.1.
@Corteum
@Corteum 5 ай бұрын
I'd agree with your assessment. When I tested SF13 by having it analyze some of AZ's wins, it couldnt find some of the moves that AZ played in critical positions. But SF15 and 16 are most definitely stronger than AZ. I wish someone would hack DeepMInd, steal the code for AZ, and make a a new "Frankenstein" version of AZ. 😂 I reckon it'll be even stronger!
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
@@CorteumGoogle Leela Chess Zero
@ocudagledam
@ocudagledam 4 ай бұрын
Ginger GM was pushing Harry up the board before A0!
@EliasMheart
@EliasMheart 4 ай бұрын
So is this "Great Engines of the Past"? Or ... "Great Systems of the Past"... No, I don't think we have a good word that specifically includes AI(ML) and also engines/expert systems, but is still narrow enough... Ah well, I like the first idea, anyway
@vibovitold
@vibovitold 14 күн бұрын
Chess engine is basically any program that plays chess. Pretty much all strong chess engines (including Stockfish) incorporate NNUE these days. So whether it's based on neural networks or manually encoded heuristics, it's still a chess engine.
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
Ya engines don't get tired and won't mind playing a 24 hours long game because they ain't got no life. Just like me, except for the tired part.
@noobmaster006
@noobmaster006 5 ай бұрын
21:14 hahaha 🤣
@user-bh5zl9nx3m
@user-bh5zl9nx3m 5 ай бұрын
1000 games? Tyler1: Hold by beer...
@victorfinberg8595
@victorfinberg8595 5 ай бұрын
actually, 155-6, with even a million draws DOES qualify as a "crush"
@-zelda-
@-zelda- 5 ай бұрын
+155 -6 =1000000 is just a 0.05 elo difference
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
1000 Games is just too small of a sample size.
@mrcleanisin
@mrcleanisin 5 ай бұрын
The stockfish16(GM) on my phone says 2850. So, there's another version that is 3350? Very difficult to even get a draw with this one.
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
What app are you using
@mrcleanisin
@mrcleanisin 4 ай бұрын
Chessify
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
@@mrcleanisin Chessify’s Stockfish seems to be handicapped to 2800 elo
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
If you want to play the non handicapped version use either Droidfish or SmallFish (depending on whether you use iOS or Android)
@mrcleanisin
@mrcleanisin 4 ай бұрын
I'm not sure I understand you. I have chessify on my android phone that says it uses stockfish rated 2850. It beats me really easy, but I have gotten a draw when I take back some weak moves. I would imagine this chessify on my phone would beat Magnus Carlson. Why don't you try it?
@exitar1
@exitar1 5 ай бұрын
ChessGenius for the win...
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
I believe what is cool about AZ is that it doesn't know the concept of material in a way that humans do and therefore Stonkfish.
@f.d.3289
@f.d.3289 5 ай бұрын
what is cool is that AZ doesn't know ANYTHING at all. it only knows how to learn.
@MrBonified66
@MrBonified66 5 ай бұрын
It didn't know the concept of *anything* outside of the rules. That's the amazing part. It's just here are the rules, play a billion games against yourself and see what works
@BobChess
@BobChess 5 ай бұрын
Smartest fish in the world is Stockfish
@donovanwarren5012
@donovanwarren5012 5 ай бұрын
kf1!!
@mikeweaver8495
@mikeweaver8495 5 ай бұрын
Everyone is ignoring the elephant in the room. Alpha zero surpassed stockfish in a MUCH shorter development period. Stockfish had years of development at the time alpha zero was better, but alpha zero had mere months of development. So yes BECAUSE alpha zero was shut down in 2018 and stockfish continued developing for years after, stockfish is better now than alpha zero was then. This is like comparing Magnus Carlson to Paul Morphy or even Bobby Fischer. Yes, Magnus Carlson of today could beat them then, but they were so much farther ahead of their competitors than Magnus is from his competitors. Now, if we look at leela chess zero, an AI that has been in development for a few years instead of months, we see that it CAN beat the best modern stockfish. Now think about how much more advanced alpha zero was than leela and apply the same or faster advancement for alpha zero. I don’t understand how anyone would think that if alpha zero had continued development, it wouldn’t be the best chess entity in the world. It just doesn’t make sense.
@vibovitold
@vibovitold 15 күн бұрын
Carlsen, not Carlson. And the fact that every strong language (including Stockfish) includes a neural network now (NNUE) pretty much answers the question of whether the approach prototyped by AZ is superior or not. It is even more obvious where you look at Go. Due to (much) greater computational complexity, there have NEVER been any "traditional" Go engines capable of beating the best humans (without handicap). No "Go-Stockfish" existed. Which is why AlphaGo was tested against the best human players (beating them). This has only become possible once deep learning / neural networks were used, as opposed to traditional algorithms.
@kmarasin
@kmarasin 5 ай бұрын
In the last game, a GM playing black would've taken the draw by repetition. Probably all those 800+ draws went a lot like this. Which begs the question: why this game? I think AlphaZero evaluated the positron as winning every time. By the barest margin. Somewhere deep in its behavior model, there's gotta be a place where it decided "when I'm winning by any margin, I refuse to repeat." And that's even if the move that breaks repetition is worse, as long as the overall evaluation is equal or in its favor. So it allowed its position to become worse from every refusal, until its usual model didn't work. Truly a 0.6% error.
@andsviat
@andsviat 5 ай бұрын
Leaving a nasty comment even before you started explaining anything and everything.
@-_Nuke_-
@-_Nuke_- 5 ай бұрын
StockFish doesn't know how to play chess. It uses a negamax function that mini maxes thousands of positions per second with alpha beta pruning and iterative deepening, by comparing static evaluations that come from a static evaluation function that doesn't necesserily understand how chess is being played, its just good enough to beat all Humans and most other engines... But if a Human GM had the ability to remember a perfect oppening book, had the ability to have a perfect ending book, and the ability to calculate equal amount of positions per second and acess them perfectly at any second... Then the Human GM would always win against StockFish and he would in fact prove that the engine doesn't know what it is doing... But a neural network does! A neural network indeed thinks like a Human, and sort of understands that chess more deeply... Of course only a saintient being could understand what chess is, but neural networks show sparks of AGI - artificial general inteligence, and that is much closer to playing stronger chess overall. Maybe StockFish 16 or whatever we have now, is already a bruteforce - neural network AI hybrid and that is the only way to reach such high elo. And by StockFish 16 I don't mean the one you have on your phones...
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
Stockfish 16 uses NN and Stockfish 16 runs on your phone. It also has a branching factor of under 1.7, which is way less “brute force” than even human GMs
@Wurto
@Wurto 5 ай бұрын
Hey Ben, Leela is also an engine just like Stockfish and they are both "AI". The difference is mainly how they evaluate positions. And even though they are similar strength, I find Leela much better for analysis since it can give you stats (such as estimated win loss draw probabilities) that Stockfish cant (Stockfish "thinks" in centipawns). So I think the chess world/coverage could benefit from adapting Leela as the main engine. Cheers, thanks for the content
@vibovitold
@vibovitold 16 күн бұрын
Stockfish's algorithm was written manually. (It's got a NNUE component now, but it didn't back then). What we refer to as "AI" in this context are implementations based on neural networks, meaning they're black-box algorithms which have kind of "written themselves" by trial and error..
@jaroslavtavgen3939
@jaroslavtavgen3939 5 ай бұрын
AlphaZero never existed (at least as a fully functional and strong chess engine). Nobody has seen it, the chess players were never allowed to "play with it" (although many of them wanted to), the source code was never released (why not? the engine is abandoned), and the notation of 862 games (out of 1072) which were allegedly played against Stockfish 8 were never published.
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
Classic google move
@vibovitold
@vibovitold 14 күн бұрын
According to my info (my friend works at Google), "AlphaZero" was Hans Niemann disguised as a neural network.
@Corteum
@Corteum 5 ай бұрын
Having looked at the games, I can say that AlphaZero was better than SF8, SF9, SF10, and I'd say even SF11. But SF12 would be pretty close to AlphaZero's level. AZ was at least 3600-3650 FIDE classical level. SF8 was about 3400-3450 classical on the rating lists.
@zvonimirtosic6171
@zvonimirtosic6171 5 ай бұрын
SF8 I think was, in reality, 3200-3250. Many of its moves were predictable. And still, Stockfish had years and years of development and learning and openings repertoire. All of that sunk at the bottom of the sea when torpedoed by mere 9 hours of self-learning by A0. And A0 moves were utterly mind-blowing and unpredictable. We got an all new insight into the game.
@Corteum
@Corteum 5 ай бұрын
@@zvonimirtosic6171 I think if you had SF8 competing in disguise in human elite events, it would wipe the floor clean. Former WCCC Rybka 3 was already over 3000. Rybka 4 was not quite as strong as SF8 and had a rating of around 3350. But i remember clearly SF8 having been rated at over 3400 on multiple ratings lists. A good way to compare AZ with today's engines is to have SF12 or higher play a match against SF8 using the exact same openings as it played vs AZ. You'll learn a lot from those games and have a much better idea of how strong SF12 (or higher) is compared to AZ.
@Corteum
@Corteum 5 ай бұрын
@@zvonimirtosic6171 I've gone through a lot of A'Z's wins vs SF8, and even SF12 struggled to find some of AZ's best moves. In some cases, it never found the moves at all. Only later versions of SF did find AZ's moves.
@zvonimirtosic6171
@zvonimirtosic6171 5 ай бұрын
What I wanted to say, perhaps, is that chess engines do not perform at presumed ELO rating across all possible aspects of playing. On some aspects they are good and maybe better than their alleged ELO rating, but on some they are not as good as their alleged rating. Maybe this is "human way" of assessing engine's performance (which does not suffer from human issues, but it does have its own weaknesses). I do remember using engines before A0, and they played "predictably boring" hi-level chess, without blunders. But with A0, though, for the first time I saw something I never saw before, and which Magnus Carlsen called "something we can mistake for creativity". A0 approached the game in a wholly different way and that chess was exciting to watch.
@Corteum
@Corteum 5 ай бұрын
@@zvonimirtosic6171 A0 is definitely something else. I remember testing Stockfish 12 NNUE on some of AO's brilliant moves, and SF12 struggled to find the correct solution. It wasn't until SF14 and 15 that it started to find A0's moves. I agree that the top engines have their strengths and a few (very few) weaknesses. But in general, their performance on different aspects of chess is superior to elite GMs. In saying that, there are still some positions they struggle with that i've come across. But those are very few. It would be interesting if they ran an average engine (e.g. Houdini 1.5a or Rybka 3 - both former WCCC's), using just 1-cpu and a small hash, in a top level human tournament, without the human players knowing about it. And just see what the approximate human elo level is for these engines. Then do the same with a more modern NNUE neuralnetwork engine. My guess is that a few GMs may get some draws against Rybka or Houdini, but probably no wins. Meanwhile, the NNUE engine will not lose, or even draw, a single game vs elite GMs. Maybe more advanced AI's in the near future will be able to accurately model the performance and style of different elite GMs from past or present, and then test to see how they might have performed against chess computers from past or present using different hardware configurations. Would love to see it. 👍
@kamilhorvat8290
@kamilhorvat8290 4 ай бұрын
Stockfish 2018 ? That's to outdates, not worth watching.
@TheMasterboi1
@TheMasterboi1 5 ай бұрын
No Talking
@askthepizzaguy
@askthepizzaguy 5 ай бұрын
I have only watched 12 seconds of this video but I feel strangely compelled to leave a nasty comment for no reason. Odd since I usually damn this channel with praise coming from the likes of someone such as me. My approval brings you shame, etc.
@loophazard
@loophazard 5 ай бұрын
obligatory nasty comment: *shakes fist*
@p1god2
@p1god2 5 ай бұрын
nasty comment #1
@aanon4019
@aanon4019 5 ай бұрын
Rawr!!
@sugarcravings1797
@sugarcravings1797 5 ай бұрын
They didn't use Stockfish 9 because they couldn't afford it, and so with Stockfish 10, and stockfish 11, and stockfish 12, and stockfish 13, and stockfish 14.
@Corteum
@Corteum 5 ай бұрын
Only SF13 ror later wouldve had a chance against AZ.
@sugarcravings1797
@sugarcravings1797 5 ай бұрын
@@Corteum I'm pretty sure Stockfish 16 is good enough to defeat her consistently.
@Corteum
@Corteum 5 ай бұрын
@@sugarcravings1797 Most definitely. SF16 is super strong. But there are some SF derivatives that are even stronger.
@user-ws7kp1yh9l
@user-ws7kp1yh9l 4 ай бұрын
@@CorteumDerivatives are just scam
@xtripx4273
@xtripx4273 5 ай бұрын
So many idiots in comment. Nice video Ben! Keep it up 😊
@oriondx72
@oriondx72 5 ай бұрын
Stockfish is just overrated!
@user-ws7kp1yh9l
@user-ws7kp1yh9l 5 ай бұрын
Stockfish’s tournament wins speaks for itself
@JannisSicker
@JannisSicker 5 ай бұрын
leaving a nasty comment for good measure
@bugzbunny109
@bugzbunny109 5 ай бұрын
Hi guys, I'm here to leave a nasty comment.
@JojenReed
@JojenReed 5 ай бұрын
Very nasty comment. Nice glasses though Ben
@iAm7amdoh
@iAm7amdoh 5 ай бұрын
A very nasty comment
@King.Mark.
@King.Mark. Ай бұрын
zero had over 100 cpu and sf 1 cpu sf 8 elo 3400 sf 16 elo 3650
@vibovitold
@vibovitold 15 күн бұрын
You seem to be confusing two things here. AZ was TRAINED on a strong machne, but when it played, it didn't run on one anymore. Training a neural network is pretty much the equivalent of home preparation.
@King.Mark.
@King.Mark. 14 күн бұрын
@@vibovitold thanks for the info was something l read on an other post and did not look into it any more than that
@iAm7amdoh
@iAm7amdoh 5 ай бұрын
Very nasty comment
@jaydubbs3902
@jaydubbs3902 5 ай бұрын
A bit much chatter and not enough analysis. Just get into it! Typically enjoy your work but not this.
@knightmare4874
@knightmare4874 5 ай бұрын
booo!!! ben sucks!!!1111 🥰😍
@paulgoogol2652
@paulgoogol2652 5 ай бұрын
But in the most enjoyable nevermind.
Diagonal Attacks on the Uncastled King with GM Ben Finegold
54:48
GMBenjaminFinegold
Рет қаралды 10 М.
Jobava London: Lecture by GM Ben Finegold
50:15
GMBenjaminFinegold
Рет қаралды 72 М.
Make me the happiest man on earth... 🎁🥹
00:34
A4
Рет қаралды 7 МЛН
когда достали одноклассники!
00:49
БРУНО
Рет қаралды 4 МЛН
Sprinting with More and More Money
00:29
MrBeast
Рет қаралды 147 МЛН
Alpha Zero's "Immortal Zugzwang Game" against Stockfish
15:37
agadmator's Chess Channel
Рет қаралды 1,6 МЛН
Great Players of the Past: Duncan Suttles
47:49
GMBenjaminFinegold
Рет қаралды 20 М.
Stockfish Has Ruined Chess
8:20
Chess Vibes
Рет қаралды 463 М.
This Insane Move Crushed Stockfish
20:55
GothamChess
Рет қаралды 629 М.
Converting Winning Positions with GM Ben Finegold
54:05
GMBenjaminFinegold
Рет қаралды 111 М.
The TRUE values of all chess pieces (according to maths)
11:55
Famous GM Blunders
56:59
GMBenjaminFinegold
Рет қаралды 78 М.
5 Checkmates Chess ENGINES CAN'T Find!
13:22
Chess.com
Рет қаралды 112 М.
😨 Никогда Не смотри на Эту Луну в Майнкрафт
37:02