Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful youāll near-instantly regret.
Any awful.systems sub may be subsneered in this subthread, techtakes or no.
If your sneer seems higher quality than you thought, feel free to cutānāpaste it into its own post ā thereās no quota for posting and the bar really isnāt that high.
The post Xitter web has spawned soo many āesotericā right wing freaks, but thereās no appropriate sneer-space for them. Iām talking redscare-ish, reality challenged āculture criticsā who write about everything but understand nothing. Iām talking about reply-guys who make the same 6 tweets about the same 3 subjects. Theyāre inescapable at this point, yet I donāt see them mocked (as much as they should be)
Like, there was one dude a while back who insisted that women couldnāt be surgeons because they didnāt believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I canāt escape them, I would love to sneer at them.
(Semi-obligatory thanks to @dgerard for starting this)
Dude discovers that one LLM model is not entirely shit at chess, spends time and tokens proving that other models are actually also not shit at chess.
The irony? Heās comparing it against Stockfish, a computer chess engine. Computers playing chess at a superhuman level is a solved problem. LLMs have now slightly approached that level.
Writeup https://dynomight.net/more-chess/
HN discussion https://news.ycombinator.com/item?id=42206817
I remember when several months (a year ago?) when the news got out that gpt-3.5-turbo-papillion-grumpalumpgus could play chess around ~1600 elo. I was skeptical the apparent skill wasnāt just a hacked-on patch to stop folks from clowning on their models on xitter. Like if an LLM had just read the instructions of chess and started playing like a competent player, that would be genuinely impressive. But if what happened is they generated 10^12 synthetic games of chess played by stonk fish and used that to train the model- that aināt an emergent ability, thatās just brute forcing chess. The fact that larger, open-source models that perform better on other benchmarks, still flail at chess is just a glaring red flag that something funky was going on w/ gpt-3.5-turbo-instruct to drive home the āeMeRgEnCeā narrative. Iād bet decent odds if you played with modified rules, (knights move a one space longer L shape, you cannot move a pawn 2 moves after it last moved, etc), gpt-3.5 would fuckin suck.
Edit: the author asks āwhy skill go down thoā on later models. Like isnāt it obvious? At that moment of time, chess skills werenāt a priority so the trillions of synthetic games werenāt included in the training? Like this isnāt that big of a mysteryā¦? Itās not like other NN havenāt been trained to play chessā¦
Particularly hilarious at how thoroughly theyāre missing the point. The fact that it suggests illegal moves at all means that no matter how good itās openings are the scaling laws and emergent behaviors havenāt magicked up an internal model of the game of Chess or even the state of the chess board itās working with. I feel like playing games is a particularly powerful example of this because the game rules provide a very clear structure to model and itās very obvious when that model doesnāt exist.
uhh
Battlechess both could choose legal moves and also had cool animations. Battlechess wins again!
Iām not a Chess person or familiar with Stockfish so take this with a grain of salt, but I found a few interesting things perusing the code / docs which I think makes useful context.
Skill Level
I assume ālevelā refers to Stockfishās Skill Level option.
If I mathed right, Stockfish roughly estimates Skill Level 1 to be around 1445 ELO (source). However it says āThis Elo rating has been calibrated at a time control of 60s+0.6sā so it may be significantly lower here.
Skill Level affects the search depth (appears to use depth of 1 at Skill Level 1). It also enables MultiPV 4 to compute the four best principle variations and randomly pick from them (more randomly at lower skill levels).
Move Time & Hardware
This is all independent of move time. This author used a move time of 10 milliseconds (for stockfish, no mention on how much time the LLMs got). ā¦ or at least they did if they accounted for the āMove Overheadā option defaulting to 10 milliseconds. If they left that at itās default then 10ms - 10ms = 0ms so š¤·āāļø.
There is also no information about the hardware or number of threads they ran this one, which I feel is important information.
Evaluation Function
Stockfishās FAQ mentions that they have gone beyond centipawns for evaluating positions, because itās strong enough that material advantage is much less relevant than it used to be. I assume it doesnāt really matter at level 1 with ~0 seconds to produce moves though.
Still since the author has Stockfish handy anyway, itād be interesting to use it in itās not handicapped form to evaluate who won.
@gerikson @BlueMonday1984 the only analysis of computer chess anybody needs https://youtu.be/DpXy041BIlA