DeepMind’s New AI Saw 15,000,000,000 Chess Boards!

  Переглядів 96,931

Two Minute Papers

Two Minute Papers

29 днів тому

❤️ Check out Microsoft Azure AI and try it out for free:
azure.microsoft.com/en-us/sol...
📝 The paper "Grandmaster-Level Chess Without Search" is available here:
arxiv.org/abs/2402.04494
+1:
www.anthropic.com/news/decomp...
transformer-circuits.pub/2023...
📝 My paper on simulations that look almost like reality is available for free here:
rdcu.be/cWPfD
Or this is the orig. Nature Physics link with clickable citations:
www.nature.com/articles/s4156...
🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
Alex Balfanz, Alex Haro, B Shang, Benji Rabhan, Gaston Ingaramo, Gordon Child, John Le, Kyle Davis, Lukas Biewald, Martin, Michael Albrecht, Michael Tedder, Owen Skarpness, Richard Sundvall, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Tybie Fitzhugh, Ueli Gallizzi.
If you wish to appear here or pick up other perks, click here: / twominutepapers
Thumbnail background design: Felícia Zsolnai-Fehér - felicia.hu
Károly Zsolnai-Fehér's research works: cg.tuwien.ac.at/~zsolnai/
Twitter: / twominutepapers

КОМЕНТАРІ: 311
@burger6178
@burger6178 27 днів тому
So this can deduct algorithms from data? Wow!
@Deathington.
@Deathington. 27 днів тому
No more black boxes!
@herp_derpingson
@herp_derpingson 27 днів тому
Literally all machine learning.
@thomasamara1453
@thomasamara1453 27 днів тому
What do you mean? Does it mean the previous one is not from data or what?
@YoussefMohamed-er6zy
@YoussefMohamed-er6zy 27 днів тому
WTF?!, it pattern matches, and it doesn't chose optimum moves
@xaxfixho
@xaxfixho 27 днів тому
Next a cure for cancer 😂
@antarctic214
@antarctic214 27 днів тому
1:42 stockfish isn't handcrafted anymore. It also uses a neural network for eval, just a very small one that is optimized for incremental evaluation on a cpu (NNUE)
@fireninja8250
@fireninja8250 26 днів тому
Does that mean it could be smarter/faster on a GPU?
@szymonmilosz99
@szymonmilosz99 26 днів тому
I think you still need do consider it handcrafted, as those NNUEs are trained on those handcrafted evaluations themselves, just on some depth search.
@user-ws7kp1yh9l
@user-ws7kp1yh9l 26 днів тому
search is handcrafted no?
@charis7854
@charis7854 25 днів тому
Search seems to be handcrafted but its parameters tuned with some blackbox optimiser. NNUE's dataset also contains games from the Leela engine 🤔
@ThePositiev3x
@ThePositiev3x 23 дні тому
@@szymonmilosz99 actually SF authors used many Leela evals in the training process which were not handcrafted. So it's a far-fetched claim to say SF is handcrafted
@cosmosmythos
@cosmosmythos 27 днів тому
Now we need an algorithm to analyze why the chicken really crossed the road
@Axiomatic75
@Axiomatic75 27 днів тому
I wish we lived in a world where chickens didn't have their motives questioned 😊
@itzhexen0
@itzhexen0 27 днів тому
Because there was a road in the way to where it was going otherwise the chicken wouldn't have cross a road.
@wisdomking8305
@wisdomking8305 27 днів тому
Can somebody explain this joke
@2019inuyasha
@2019inuyasha 27 днів тому
Chicken doesn't notice road just walks and runs around as desired
@Axiomatic75
@Axiomatic75 27 днів тому
@@wisdomking8305 The joke "Why did the chicken cross the road?" is a classic with origins dating back to the mid-19th century. Its humor lies in the unexpected simplicity of the punchline, "To get to the other side," which subverts the expectation of a clever or humorous answer. The joke's enduring popularity stems from its versatility and ability to be interpreted in various ways. While it's traditionally seen as a play on words, some may also interpret it metaphorically, suggesting deeper meanings related to life and death.
@wacky3022
@wacky3022 27 днів тому
What a time to be alive!
@jimmykrochmalska3501
@jimmykrochmalska3501 26 днів тому
saying a chess ai can outperform gpt 4 in chess is like saying a sumo master can knock over a child
@tobirivera-garcia1692
@tobirivera-garcia1692 12 днів тому
more like a sumo master beating a child in a sumo match, and the child doesnt even know the rules
@kpoiii7795
@kpoiii7795 27 днів тому
Does it mean that this is the start of making full algorithms out of neural networks instead of black boxes which they are currently?
@calebbarton4293
@calebbarton4293 27 днів тому
This is a fantastic question, so I am engaging with it in hopes more people see it!
@dustinandrews89019
@dustinandrews89019 27 днів тому
This is the real implication. Can we next show it snippets of code and have it decompile the algorithm from the weights and biases in the original model? That sounds like a typical "Two papers down the road" type leap. Inspection is lacking in current models, and that's a big problem.
@jorgwei8590
@jorgwei8590 27 днів тому
I wonder that as well. But: Maybe the algorithm is just another output of the model, not necessarily the logic it used to chose the move. A bit like when you ask GPT why it did something and it gives you a plausible explanation that makes sense but isn't really what is going on ... just with maths.
@KaworuSylph
@KaworuSylph 27 днів тому
The footage accompanying that part of the video looked like frequency analysis (of node weights?) to pattern match a known chess move (algorithm). I don't know how you'd be able to isolate new algorithms the AI comes up with on its own - maybe something like a fourier analysis to see if it's composed of the freqencies of multiple algorithms combined?
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
I don't know where you got that from... because what happened here it is the exact opposite. It takes an algorithm or more specifically, two algorithms: position evaluation and search and turns it into a black box that can somehow figure out the right move (or at least a very good move) instantly.
@Thomas-ot5ei
@Thomas-ot5ei 27 днів тому
If you thought you couldn't be more surprised… wow!
@Chris.Davies
@Chris.Davies 26 днів тому
This is a clear derivation of the word prediction of the large language models. All these pseudo-smart systems are deeply impressive tools for humans, and I can't wait to see what we'll do with them!
@Jackson_Zheng
@Jackson_Zheng 27 днів тому
Sounds like the AI just learned to play blitz really well. It's just pattern recognition, no calculations are needed and humans can do it too.
@ayoCC
@ayoCC 27 днів тому
seems like the purpose of thise research was to discover how to make an AI output repeatable algorythms that then can maybe run more cheaply.
@PlayOfLifeOfficial
@PlayOfLifeOfficial 27 днів тому
Raw intuition
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
Yes, it runs on pure intuition like human blitz players, but it can presumably not only beat the best blitz players in the world, but does so with only 270 million parameters. If you want to correlate parameters with synapses, that's smaller than the "brain" of a bee.
@Jackson_Zheng
@Jackson_Zheng 24 дні тому
@@miriamkapeller6754 yes, and humans haven't seen billions of games and do not play chess for 24/7 without rest. The network is highly optimised for this one specific narrow task and does nothing else. Not really a fair comparison is it? ...Especially as bees and other insects have to pack sensory neurons, an flight controller, accleration and gyroscopic processors, and homeostasis systems, and much much more, into a similar amount of neurons.
@miriamkapeller6754
@miriamkapeller6754 24 дні тому
@@Jackson_Zheng I'm not saying a bee can learn chess, I'm just saying "a human can do it too" is not much of an argument when a human has about 1 quadrillion synapses compared to this tiny neural network.
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
I tried this approach a while ago. The network I trained could at least beat me (~1000 elo), which I was satisfied with. I could have trained it longer, but this was 6 years ago and my graphics cards was really bad back then. It was also a smaller network and not a transformer, but an AlphaZero style convolutional network. One fun thing you can do with this is to condition on the rating of the player/chess engine to have the network mimic a certain skill level.
@shoobidyboop8634
@shoobidyboop8634 27 днів тому
The neural net embeds the lookahead library.
@cannot-handle-handles
@cannot-handle-handles 27 днів тому
Agreed, the video's take that no search was used was weird. The search used by Stockfish was translated into weights and biases. Still a cool result.
@chritical_ep
@chritical_ep 27 днів тому
​@@cannot-handle-handlesfrom my understanding, it's "no-search" because it doesn't search more than one move ahead when playing a game. How it was trained doesn't matter
@cannot-handle-handles
@cannot-handle-handles 27 днів тому
@@chritical_ep Yes, that's also how I'd understand "no search", but the framing in the video was weird / a bit too sensationalist. So it's technically no search, but the neural net contains equivalent information.
@jswew12
@jswew12 27 днів тому
​@@cannot-handle-handles If I am understanding the paper from my skim, the data it is trained on uses binned positional assessments generated by StockFish. Are you saying that because of this, calling it "no search" is sensational? I feel like the underlying data doesn't really matter much as long as the system itself isn't using the approach, kind of like we wouldn't call a sentence-generating transformer "conscious" just because the data underneath was generated by conscious beings
@Xamy-
@Xamy- 27 днів тому
Guys stop being bloody stupid. It’s novel because it means it can run in your browser, where as you cannot run the original could not
@johnmarmalade4345
@johnmarmalade4345 26 днів тому
Leela Chess Zero (or just Leela) tried a similar test to this one after the team saw the DeepMind paper. They compared how the most recent Leela weights performed against DeepMind's AI. The cool thing is that Leela performed a little better than the DeepMind one. The results are published on Leela's Blog.
@pinkserenade
@pinkserenade 16 днів тому
This AI proves that it's possible to reach 2800 using pure intuition (pattern recognition) without calculating (searching)
@benjaminlynch9958
@benjaminlynch9958 27 днів тому
I’m really curious about the real world benefit of this model beyond the fact that it’s small and fast and can be run efficiently locally on low cost and low power devices. I know this supposedly isn’t about chess (???), but they essentially managed to build a worse version of Alpha Zero 7 years later with a completely different model architecture. And yes, the new model can be applied to many domains beyond chess or even board games in general, so there’s some usefulness in that, but… this just seems like a repackaged form of supervised learning that has been around for ages, and the result / accuracy doesn’t seem that different than what older methods of supervised learning would yield. What am I missing here?
@geli95us
@geli95us 27 днів тому
The fact that it didn't use search is very significant, as search is one of the most important factors in creating a strong chess engine, even a simple hand-crafted evaluation function can be extremely strong when paired with a strong computer and a good search algorithm. There are 2 aspects I consider relevant here: 1) This AI has not been given any "hints" by humans, it had to figure everything itself, unlike alphazero which was given the search algorithm and only had to develop the evaluation function. 2) This AI is forced to spend a constant amount of compute for each position. It's not very useful at all in terms of chess, stockfish is faster and stronger (it's the model it was trained on), but it can teach us a bunch of stuff about how transformers learn (the architecture LLMs are based on), and if we could decipher the algorithm this is using it would be incredibly interesting I imagine, so it could be helpful in developing stronger chess engines in the future
@Chris-b-2
@Chris-b-2 27 днів тому
It strikes me as a thrust towards a two tier modeling system. The primary model learns about the system it is trying to infer on. The secondary models try to predict what the primary model would do. The second is _much_ easier to train and can be smaller for deployment onto low power devices.
@mtdfs5147
@mtdfs5147 27 днів тому
​@@geli95usstill seems like a pretty small advancement to make a whole video on. Even given your explanation it sounds like a pretty meh change. With how rapidly AI is advancing, this shouldn't even be a footnote. Let alone a whole video... Maybe I'm missing something too 😅
@percywhitehead9228
@percywhitehead9228 27 днів тому
@@geli95us wasn't deepmind first go at alpha zero using human games as data then played lots of more games using that data to improve? I honestly don't understand what this video is showing us, why is it better than alpha zero in anyway? isn't inference just using an algorithm in AI. Also it's abit weird to say this doesn't use search, because stockfish uses search... so it's tying to imitate stockfish which has had over a decade worth of work put in it and even then it's no where near stockfish and stockfish has stupid amount of search. such a bad video , can't be that hard to explain the advantages over traditional methods. seems like it was rushed
@Aerxis
@Aerxis 27 днів тому
You can train chess bots that imitate humans a lot better, for lower level ELOs.
@vladthemagnificent9052
@vladthemagnificent9052 27 днів тому
Dear Karoly, thank you for extensive coverage of the most exciting reakthrough in computer science and more. But this one, you did not convinse me this paper has any significant result. They deleted the search part and it preformed worse than tree-search algorithms... duh... In fact it couldn't even make mate in 3 on its own. That whole speach about the algorithms in the end seems to be on the opposite side of what was done in the paper. I don't get it whats new or exciting about this particular paper. And I do not like the general sentiment that now the transformer is a soluion for everything, just throw more compute into it. No, I believe, it is still very important to develop different algorithms and different architecture to do really impressive stuff, and other videos on this channel illustrate this idea perfectly.
@EstrangedEstranged
@EstrangedEstranged 26 днів тому
The transformer is not a solution for everything but it solves way more than our ego would like to admit. The paper proves that even training on isolated examples can create inside the model structures similar to the things in the real world that produced the examples. The examples are enough to create (through training which is an evolutionary process) the behaviour that produced the examples. It's a chess proof against the "stochastic parrot" argument.
@joshuascholar3220
@joshuascholar3220 26 днів тому
Didn't it say that it's using a one move look ahead? I should read the paper, but I bet if you understood the details enough, you'd find that all kinds of look aheads are actually happening. In large language models, transformers make many passes with many specialized "key and query" networks.
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
@@joshuascholar3220 No. The "one move" it looks ahead is just the move it's going to make. There is no search. Now internally the transformer is obviously going to do some analysis, it will likely generate maps of where the opposite pieces could move, create some higher level danger maps and so on. But it's still only a single inference pass.
@vladthemagnificent9052
@vladthemagnificent9052 26 днів тому
​ @EstrangedEstranged I don't have my ego hurt in any way, if they put 1000 times more compute and memory into it they would have had a program that plays chess better than everything seen before (although it still wouldn't know if it is allowed to do castling or en passant, without search and game history, lol). The problem I see here is that a bunch of people put a lot of resources to build a useless model to prove a point that had no need to be proven. ok deepmind can waste their money and time all as mush as they want, but I am confused as to why this is reported about as an exciting piece of research.
@1qmqmqm1
@1qmqmqm1 27 днів тому
I've been waiting for this since several YEARS! I'm doing my own research and it's always interesting to see where others are at. And I feel the urge to solve chess too lol
@peterkonrad4364
@peterkonrad4364 25 днів тому
i tried something similar about 15 years ago for connect 4. i had a very good traditional connect 4 program (like stockfish for chess here), and i let a neural net learn boards and what my player would do. but i never got it working as intended. i always had a feeling that i simply didnt give it enough training examples. i wanted to go to tic tac toe, so i could give it maybe enough examples with the computing power i had available back then, but then i lost interest. nice to see, that it works! even on chess!
@STONJAUS_FILMS
@STONJAUS_FILMS 10 днів тому
even tho i dont completely understand the how, this sounds like the biggest leap I've heard in the last year. sounds like there is room for crazy optimization of way bigger models. ... wow
@shinyless
@shinyless 25 днів тому
What about the fact that the input/output parameter count is drastically reduced by iteration ? Better to compute "just a move" rather than a whole board ? Or am I misled ? :)
@jasonruff1270
@jasonruff1270 27 днів тому
what kind of computer is used to handle all these operations and parameters, do they just used multiple gpus for deep learning like this?
@AviweZathu
@AviweZathu 27 днів тому
😌 I need this AI for checkers as AR Sun Glassea, like those Meta + Rey-ben
@wealthycow5625
@wealthycow5625 27 днів тому
If you adjusted the amount of moves it could see, this could be a great engine for chess players to see actually human playable lines
@Blattealkiller
@Blattealkiller 27 днів тому
Can they use this as a new evaluation function that they would use inside a search algorithm ?
@user-dh8oi2mk4f
@user-dh8oi2mk4f 23 дні тому
No, you need a separate network for evaluating positions
@galmacky
@galmacky 21 день тому
So this is just a distilled student model from Stockfish?
@quellepls2568
@quellepls2568 27 днів тому
Were can i Download that ai?
@user-ws7kp1yh9l
@user-ws7kp1yh9l 26 днів тому
You can’t, but this neural network just a marketing stunt. It is weaker than Leela Chess Zero network even at 1 node.
@ananthakrishnank3208
@ananthakrishnank3208 27 днів тому
2:10 "Lichess elo" of 2895 sounds more like IM (International Master) strength, than that of a GM's strength. Alphazero (AI chess engine of Deepmind) could easily be 3300+. Leela Chess Zero (Lc0) built using the same architecture is 3600 today. 3:24 ChatGPT is not even good. In the sense it was not made for chess. Gothamchess channel has covered a video with GPT playing chess. At times it played moves that are illegal. Regardless, incredible results 👍🏻
@BooBaddyBig
@BooBaddyBig 27 днів тому
No. GMs start at 2500. IMs are 2400-2500. Above 2700 are sometimes called 'super grandmasters'.
@ananthakrishnank3208
@ananthakrishnank3208 27 днів тому
@@BooBaddyBig That is in "real elo" right? I always thought that "Lichess elo" is 300-400 more from real elo.
@ruinenlust_
@ruinenlust_ 27 днів тому
@@BooBaddyBig Lichess consistently rates everyone 300+ ELOpoints higher than their actual ELO
@andrewboldi47
@andrewboldi47 27 днів тому
the main breakthrough is not that it's the best chess engine. It's that it can be run on very minimal hardware and none of its essential algorithms are specific for chess
@Tymon0000
@Tymon0000 27 днів тому
and lichess uses glicko-2 not elo
@ansklfnaskidfhn-hi6zg
@ansklfnaskidfhn-hi6zg 16 днів тому
Leela Chess Zero surpassed it last year. Hopefully Google gives their vast resources to the Leela team so they can make it even better.
@aliefrat
@aliefrat 26 днів тому
Your enthusiasm is always inspiring!
@MrSongib
@MrSongib 27 днів тому
Next, we do Shogi.
@MoritzvonSchweinitz
@MoritzvonSchweinitz 27 днів тому
Huh. Way back in the day, AI researchers asked human experts tons of questions in order to understand their internal "algorithms" - with the problem that often, humans didn't even know how the came to certain conclusions. This algorithm-from-data thing could be groundbreaking for this - and, if I understood correctly - should reduce the 'model' size by a lot, because you could replace millions of neurons with algoithms.
@Ghost-pb4ts
@Ghost-pb4ts 26 днів тому
"If we could use the Sharingan, this is how it would have felt."
@pierrecurie
@pierrecurie 27 днів тому
Did the title change just a few hrs after release? Was this channel always so reliant on clickbait?
@alanESV2
@alanESV2 27 днів тому
“New” from 2017. AI goes by quick
@DeepThinker193
@DeepThinker193 23 дні тому
lol "I love the smell of an amazing dusty old paper". xD
@caveman-cp9tq
@caveman-cp9tq 18 днів тому
Now give it various openings and positions to see where its performance are the maximum and minimum. Then we would have an idea of which openings would lead to play that’s easier and more intuitive
@TreeYogaSchool
@TreeYogaSchool 27 днів тому
Looks like AI has us in checkmate, but I want a rematch! Great video. Thank you for this information.
@vik24oct1991
@vik24oct1991 20 днів тому
this is akin to rote learning with some pattern recognition, you analyse so many board positions from top engine that you literally know the best move in each of those position, for positions which are unseen you do pattern recognition (this part is shady because there is no real way to know if logic was used to make the move as search is not used).
@DeepThinker193
@DeepThinker193 23 дні тому
Imma call it now. This is a true breakthrough to the actual start of AGI...provided the paper is true. The ability to derive pattern learning algorithms from the model and duplicate them rapidly and place them in other models would potentially lead to the exponential growth we were actually expecting when AI was announced.
@exsurgemechprints2671
@exsurgemechprints2671 27 днів тому
this paper proved that you can have capable neural networks just with datasets that are large enough.
@feynstein1004
@feynstein1004 26 днів тому
Aha! So size does matter. I knew it
@henrytep8884
@henrytep8884 26 днів тому
@@feynstein1004uhh no, because size is too costly. What really matters is efficiency, so getting the job done really really really fast
@Adhil_parammel
@Adhil_parammel 27 днів тому
From this papper we can infer that inorder to get human level general language model we we need super human level data.
@research417
@research417 27 днів тому
Well, yes and no. Human language isn't a zero-sum game, you aren't trying to beat the other person (unless you're in a debate). For a general language model like Gemini and GPT-4 it's more about the breadth and size of the amount of data, you want it to be able to imitate very high level academic language, yes, but you also want it to be able to imitate your average casual conversation. I think the next step for a human level general language model is inventing a transformer that's more novel, being able to learn from itself or on its own without humans, so it can infer data on its own.
@jnevercast
@jnevercast 27 днів тому
I agree with 417, the ability to reach better than human artificial intelligence really hinges on being able to before deductive reasoning faster than humans on more data they could see in their life time. Subjectively, It doesn't seem unlikely that we could achieve that with in context learning alone, which could be a year away.
@dustinandrews89019
@dustinandrews89019 27 днів тому
I think you have a point. GPT type models write like an average, if overly polite, internet user. It takes a lot of prompt engineering to get it to be brief, stop apologizing, and stop explaining like I am five. So that data's not going to cut it.
@jnevercast
@jnevercast 27 днів тому
@@dustinandrews89019 They write however they're tuned to, but I'd say the RLHF had a big effect on OpenAI GPTs feeling mid. I'm not sure how they accomplished the difference, be it fine tuning or just better data cleaning, but Claude 3 feels a lot better. Though there is a tradeoff, Claude doesn't fear jargon, and sometimes that makes communication slightly more challenging. Though of course it can explain what it meant, so we all get smarter.
@vik24oct1991
@vik24oct1991 20 днів тому
humans use search so no.
@lasagnadipalude8939
@lasagnadipalude8939 27 днів тому
Give it eyes and it could tell you the laws of physics we still don't understand or know
@propotkunin445
@propotkunin445 26 днів тому
does the comparison to gpt-4 make sense? i know little about ai but is it really such a surprise that this algorithm is better at chess than a large LANGUAGE model?
@menjolno
@menjolno 27 днів тому
6:52 very important disclaimer: if you try to create csam using microsoft AI, they will catch you. Please make sure to disclose that you don't really own the service.
@punpck
@punpck 27 днів тому
Chess without search 🤯🤯🤯
@P-G-77
@P-G-77 26 днів тому
Great method. The game. Basic, elementary, mathematical... but superb for the search for superiority, elasticity, stubbornness I would say almost perfection. Something that leads us to be proud to see how much we are doing for our future in the end.
@JoshKings-tr2vc
@JoshKings-tr2vc 27 днів тому
If only they could learn from fewer datapoints. That would be an absolutely crazy time.
@ILLUSIONgroup
@ILLUSIONgroup 27 днів тому
What happens if it plays tic-tac-toe with itself like in the movie War Games? 😄
@MrVbarroso
@MrVbarroso 27 днів тому
Damn, I just read this paper yesterday! That's a coincidence and a half.
@usualatoms4868
@usualatoms4868 27 днів тому
Let us know if someone leaks a military tactics AI. I'd be utterly surprised if there weren't several already.
@downey6666
@downey6666 27 днів тому
This is a gaint leap forward.
@spacejamzyt6461
@spacejamzyt6461 27 днів тому
I wonder how far we are from AI being able to examine its own code and make improvements so profound that no human could match them. It’s akin to how no human chess player can defeat today’s best chess engines. There was a time when people doubted that computers could outperform human chess players, but now the situation has completely reversed, and no human can beat the computer. Imagine the possibilities when computers achieve this level of mastery across all aspects of life. Eventually, we may find ourselves with nothing left to teach, as AI begins to enlighten us with its own discoveries. It's scary as hell but also incredibly fascinating, it's like getting a glimpse into the future.
@Julzaa
@Julzaa 27 днів тому
Love Anthropic
@DanFrederiksen
@DanFrederiksen 25 днів тому
top human players don't calculate either for the most part (non classical). It's shocking how shallow their compute is. they have learned a function to make the call without branching. the game with magnus and polgar in the park showed me that
@paroxysm6437
@paroxysm6437 24 дні тому
this is false lol humans especially in classical do very deep thinks/calculations - even Magnus. although in other time formats like blitz, it's more intuition
@DanFrederiksen
@DanFrederiksen 24 дні тому
@@paroxysm6437 true in classical they do calculate but in faster time control it's just a glance. Hikaru plays a lot on camera and you can hear his decision process. He has an immediate impression of the board and he'll say something like I guess I'll go here? and there is no calculation going on at all. And it's typically out of theory. And I have no doubt he's still playing in the 90s of accuracy. So much so that Nepo has accused him of cheating.
@PlayOfLifeOfficial
@PlayOfLifeOfficial 27 днів тому
There are a lot of unanswered questions here
@dreadowen616
@dreadowen616 12 днів тому
Can you imagine specs of machine to run this?
@Chorm
@Chorm 27 днів тому
That would be crazy if P=NP turned out to be true.
@yanushkowalsky1402
@yanushkowalsky1402 21 день тому
I thought chess ai is already on the top level that can lead to a draw anytime it can't win
@Maltebyte2
@Maltebyte2 27 днів тому
Its pre AGI! its here soon!
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
This is as far from AGI as it could possibly get. This is a classical and highly specialized neural network that can do only a single thing and nothing else.
@Maltebyte2
@Maltebyte2 26 днів тому
@@miriamkapeller6754 That is true but this capability can be expanded bejond chess. And OpenAI probably already has AGI but its not available to the public.
@caveman-cp9tq
@caveman-cp9tq 18 днів тому
@@miriamkapeller6754This translates to other fields. Give AI a bunch of data about how a human acts in certain situations and the AI would be able to mimic humans
@panzerofthelake4460
@panzerofthelake4460 26 днів тому
I'm all in for those raytracing algorithms 🤤 if this makes better raytracing algos, I'll be a happy scholar
@manavkumar348
@manavkumar348 27 днів тому
We should have chess tournaments where different types AIs play against each other
@Benjamin-yq8yl
@Benjamin-yq8yl 27 днів тому
They already exist 😁
@Adhil_parammel
@Adhil_parammel 27 днів тому
Ccrl
@c4fusion1
@c4fusion1 26 днів тому
For anyone confused, the real breakthrough is that this further proves transformers can reason better at the zero shot level with more scale. By making the neural network 10x bigger than AlphaZero, they were able to get way better zero shot performance.
@user-dh8oi2mk4f
@user-dh8oi2mk4f 23 дні тому
Leela Chess Zero had already been experimenting transformer networks for more than a year before this paper, this isn’t a breakthrough
@marcio.oliveira
@marcio.oliveira 24 дні тому
What a time to be alive!!!! 😄
@andreaspetrov5951
@andreaspetrov5951 27 днів тому
The cadence of speech in this video differs from earlier ones so much that I can only assume it's AI-generated. It's borderline robotic in it's precise timing.
@jorgwei8590
@jorgwei8590 27 днів тому
Wait...wouldn't that be huge for interpretability? Or isn't the algorithm it gives out necessarily the logic it used to make the move?
@VeeZzz123
@VeeZzz123 27 днів тому
It develops an algorithm in order to choose its action, unlike other AIs who rely on memory, not raw logic. Now they are trying to "extract" that algorithm. And chess was just an example of use. If they are successful, chess will become a "solved game" thanks to the algorithm this AI developed. After that they will want to do this for other things other than chess, which is huge for any industry.
@ekkehard8
@ekkehard8 27 днів тому
So it performs at 1000 less elo than stockfish...
@gergopool
@gergopool 26 днів тому
I think I'm missing the point. They did a supervised training on pre-generated, powerful chess moves and made a paper out of it? That sounds like something one would do as a simple baseline for further experiments.
@paroxysm6437
@paroxysm6437 24 дні тому
Essentially, an AI learned by watching games instead of playing them. Traditionally, you train AI through playing itself and searching/calculating the most "optimal" move. Tens of millions of games are played/analyzed then stored into massive datasets. This AI didn't play any games and purely just "watched" another strong AI play. This is big not just for chess but a bunch of applications as you could theoretically train an AI for a fraction of the computing power and get relatively similar performance.
@gergopool
@gergopool 24 дні тому
​@@paroxysm6437 They trained AlphaZero with no human training data because that was the goal they wanted to achieve. They were aware they could have done it with supervised data similarly to AlphaGo, but AlphaZero stood as a proof that RL can do that task alone, tabula rasa. But just because something can be solved by RL, it's often not optimal. So now, years later they present a transformer can be trained supervised to learn moves. Of course it can, their AlphaZero policy network did the same thing with a much harder learning curve. So I am missing the main contribution here. For me it sounds like a university project, but I'm very much open for arguments.
@Dark_Brandon_2024
@Dark_Brandon_2024 27 днів тому
algorithm that creates an algorithm that creates an algorithm that creates an algorithm - god
@banggiangle8258
@banggiangle8258 25 днів тому
To me, the breakthrough of this paper is kind of limited. Firstly, it seems that this is only supervised learning on a massive dataset of chess games. Secondly, the ground truth of the action is taken from ST, which is already a strong chess engine. ST 16 elo rating is around 3k6, while the learned model achieves a 2k8 rating, which makes sense but is not surprising to me at all. Finally, the whole point of searching and self-play is that we don't know what the right moves are, so the machines have to come up with new moves on their own. In this way, the machines can make up things that can even surprise experts. The hard part of finding good moves in the first place is already done by ST, and given that the labels are provided by ST, the end performance is always upper-bounded by the performance of ST.
@DanLinder
@DanLinder 25 днів тому
Thanks!
@blackmartini7684
@blackmartini7684 26 днів тому
But can it play 5D chess with multiversal time travel
@adamweishaupt3733
@adamweishaupt3733 27 днів тому
If this is ChatGPT's next brain compartment (after GPT and DALI) we might be getting awfully close to an AI that can tell truth from fiction, which sounds pretty general to me...
@milseq
@milseq 27 днів тому
AI has been beating grandmasters since the 90s...
@dustinandrews89019
@dustinandrews89019 27 днів тому
Potentially running on a phone? Granted the phones are a lot better now...
@oyashi7
@oyashi7 26 днів тому
Isnt this basically the ReBeL algorithm for poker…
@Nonkel_Jef
@Nonkel_Jef 21 день тому
No selfplay? Why would they do that?
@EobardUchihaThawne
@EobardUchihaThawne 24 дні тому
lol, i made similiar project for fun with only 10K dataset😂
@EobardUchihaThawne
@EobardUchihaThawne 24 дні тому
if i only had 14.99999B more😢
@stevemeisternomic
@stevemeisternomic 27 днів тому
We need a ai that can learn concepts and strategies, so that it can generalize from much less information.
@tubesteaknyouri
@tubesteaknyouri 27 днів тому
I was thinking something similar. In light of its training set of 15 billion examples, it seems somewhat trivial to boast that it doesn't use explicit search.
@caveman-cp9tq
@caveman-cp9tq 18 днів тому
But they do have concepts and strategies. They understand chess the same way we do, but better
@mahdoosh1907
@mahdoosh1907 27 днів тому
i saw you changed the title
@AndersHaalandverby
@AndersHaalandverby 27 днів тому
Very interesting, as usual. One problem with engines like stockfish is that it plays.. well.. like an engine. It would be interesting to train this not on stockfish, but on human/grandmaster games, maybe it would play more "human-like" moves? I think that would be a really powerful learning tool for chess players.
@janAkaliKilo
@janAkaliKilo 26 днів тому
Is this voice AI-generated? Intonations are all kinds of wrong, even more so than usually.
@craigsurbrook5702
@craigsurbrook5702 24 дні тому
Why did the robot cross the road? Because it was carbon bonded to the chicken.
@atakama2380
@atakama2380 27 днів тому
It's not about winning, it's about sending a message :)
@mathsciencefundamentals3168
@mathsciencefundamentals3168 27 днів тому
Just to add, stockfish is not one of the best, but is the best😅
@littlegravitas9898
@littlegravitas9898 27 днів тому
Given it creates specific algorithms rather than data piles, does this help with explainability?
@gilboewer4107
@gilboewer4107 27 днів тому
I'm always thinking could Dr. TMP be an AI himself from the way he talks
@TwoMinutePapers
@TwoMinutePapers 27 днів тому
I can assure you that every episode is me flipping out behind the microphone while looking at these papers! 😀
@shadowskullG
@shadowskullG 27 днів тому
​@@TwoMinutePaperswith the quality delivered , you get the credit 😅
@davechaffey3493
@davechaffey3493 27 днів тому
Brilliant! If AI was ever going to take over the world, this would be the perfect plan!
@juhotuho10
@juhotuho10 27 днів тому
Imitation learning on stockfish? That is probably the most boring approach they could come up with
@lvutodeath
@lvutodeath 27 днів тому
You'd be surprised on how much discoveries are made from "boring experiments".
@Daniel-xh9ot
@Daniel-xh9ot 27 днів тому
Why didn't they use alpha go zero instead of stockfish I wonder
@jnevercast
@jnevercast 27 днів тому
Sure, but there could be good reasons for that. Stockfish isn't a neural network for one, another is that the paper is about demonstrating learning from expert data which is hard to do well.
@miriamkapeller6754
@miriamkapeller6754 26 днів тому
@@Daniel-xh9ot Why? Stockfish is a lot stronger.
@Daniel-xh9ot
@Daniel-xh9ot 26 днів тому
@@miriamkapeller6754 I meant an implementation of alpha zero for chess instead of stockfish, wouldnt that output better moves than stockfish?
@diamondjazz2000
@diamondjazz2000 27 днів тому
But if this is trained on board positions generated by algorithms (or people) looking ahead, that information would be encoded in the board position. This is just an artifact of a game like chess that doesn’t depend on psychology or history. This same approach wouldn’t work well in poker. This feels overhyped. Of course it’s better than a large language predicting model.
@Culpride
@Culpride 27 днів тому
I see a notable step forward in the alignment problem in this approach. Machines not learning from machines with human set goals, but machines learning approaches and goals from humans. Now we only need a lot of data of desirable human behaviour. Oh ... i see a new problem arise ....
@OpreanMircea
@OpreanMircea 27 днів тому
I'm so excited for the future
@MindBlowingXR
@MindBlowingXR 26 днів тому
Interesting video!
@Leto2ndAtreides
@Leto2ndAtreides 27 днів тому
This sounds like pattern recognition - across the various states of the game.
@phen-themoogle7651
@phen-themoogle7651 27 днів тому
Humans aren't very good at chess (compared to top engines), a lot make guesses and aren't completely sure about the position in many positions. It doesn't surprise me that an AI learning from billions of stockfish moves in various positions transfer over to very high playing, even if the positions are a bit more random, it can still compare how similar something is to another position and sometimes extra pawns or something doesn't matter when positionally you have faster tactics or combinations. Maybe it learns combinations in context from so much data which makes it strong at tactics giving it that kind of rating. Heck, it can even learn openings and every part of the game from that many moves...that's a lot. Imagine a human remembering 15 billion random moves (if I remembered the number correctly), even top players only have a few thousand positions/games memorized (except Magnus which has 10k+) if there's on average 50 moves per game, maybe Magnus has 50,000 random positions/moves down and so yeah...actually compared to how little humans train on, that AI thing kinda sucks for having to look at so much data and can only be around a human GM level. Quite inefficient. When you make an AI that you didn't teach to play chess, and it somehow plays at Stockfish level or better, than I'll be extremely impressed. GPT6 or GPT7? Right now it's just doing copy-cat with some RNG at a high level.
@the_master_of_cramp
@the_master_of_cramp 5 днів тому
So basically, first we solve a problem using a neural net, then we apply this method to learn the underlying algorithm that best resembles the decisions made by the neural net? Yea sounds good. That's probably how humans also do it. First we experience the world, and get some intuition about it, learn to reflexively act. Then, with some conscious thought, we infer rules from these intuitions.
@ChristopherMcIntosh1
@ChristopherMcIntosh1 27 днів тому
SHALL WE PLAY A GAME? ... WOULDN'T YOU PREFER A NICE GAME OF CHESS?
@joannot6706
@joannot6706 27 днів тому
A step towards a Neural Network based AI that creates a symbolic or neuro-symbolic AI that runs blazingly fast?
@invisiblevfx
@invisiblevfx 24 дні тому
Consider upgrading your voice robot.
@thehealthofthematter1034
@thehealthofthematter1034 27 днів тому
Great...but...can we play with it?
@khaliffoster3777
@khaliffoster3777 26 днів тому
Ok, so explain the process of algorithms from data?? So, using less parameter, but can do better than more than hundren time of higher GTC.., so it is base on calculcate anything, so that means can can calculcate how everything work so by divide the whole process of whole into fraction of whole, so what that fraction represent, for an examble the whole pie is 1, and you slice 10, so it is part pie, so the fraction is 1/10, so the gap is bigger than other gap, so 1/10 is represent of gap of pie that is one out of ten, so I did algorithms the data that is pie into fraction of pie that I can explain what it means, from whole to part. Algorithms is the part as fraction into whole. Human can be very logical since humans can create infinity than AI since AI has higher need than human, so AI is inferior. So, ya are saying so can calculcate better so will become infinity as adapt to any situation so always right, so never wrong compare to us.
@vnagaravi
@vnagaravi 27 днів тому
How many did you see? AI: 15 billion How many did i win? AI: 0 so there isn't a chance until it freeze/crashed
@timrose4310
@timrose4310 27 днів тому
How is that different than any old neural network? Nothing new here??
@paroxysm6437
@paroxysm6437 24 дні тому
One learned via watching another person/AI play. Another learned via playing tens of millions of games/attempting to figure out the best move.
@davidetondini94
@davidetondini94 27 днів тому
What's about with his voice?
DeepMind AlphaFold 3 - This Will Change Everything!
9:47
Two Minute Papers
Переглядів 118 тис.
Повістки у Києві: «Яке право вони мають забирати всіх мужиків?» #війна #мобілізація #військові
00:41
Слідство.Інфо | Розслідування, репортажі, викриття
Переглядів 1,7 млн
Зомби Апокалипсис  часть 1 🤯#shorts
00:29
Дурнєв дивиться сторіс ZОМБІ #47
53:48
Aleksey Durnev
Переглядів 668 тис.
NVIDIA’s AI Puts You In a Video Game 75,000x Faster!
7:07
Two Minute Papers
Переглядів 145 тис.
Why the Future of AI & Computers Will Be Analog
17:36
Undecided with Matt Ferrell
Переглядів 407 тис.
The Hydra Game - Numberphile
21:54
Numberphile
Переглядів 306 тис.
Something Strange Happens When You Follow Einstein's Math
37:03
Veritasium
Переглядів 7 млн
Has Generative AI Already Peaked? - Computerphile
12:48
Computerphile
Переглядів 155 тис.
Coding Adventure: Chess
29:22
Sebastian Lague
Переглядів 3,7 млн
2024 Perovskite Breakthroughs are the Future of Solar
14:27
Undecided with Matt Ferrell
Переглядів 1 млн
I Asked Photoshop AI to Zoom Out Infinitely. Here’s What Happened.
22:37
DeepMind’s New Robots: An AI Revolution!
8:39
Two Minute Papers
Переглядів 196 тис.
AI Deception: How Tech Companies Are Fooling Us
18:59
ColdFusion
Переглядів 1,3 млн
СЛОМАЛСЯ ПК ЗА 2000$🤬
0:59
Корнеич
Переглядів 1,9 млн
Эволюция телефонов!
0:30
ТРЕНДИ ШОРТС
Переглядів 4,9 млн