Well, letās not let Baldur be a complete dumbass. There is something bad here, and weāve discussed it before (1, 2), but itās not āUS authoritiesā gaining ācontrolā over ābigotry and biasesā. The actual harm here is appointing AI-safety dorks to positions in NIST. For those outside the USA, NIST is our metrologist organization, and thereās no good reason for AI safety to show up there.
Hacker News is truly a study in masculinity. This brave poster is willing to stand up and ask whether Bluey harms men by making them feel emotions. Credit to the top replies for simply linking him to WPās article on catharsis.
Hereās some food for thought; ha ha, only serious. What if none of this is new?
If this is a dealbreaker today, then it should have been a dealbreaker over a decade ago, when Google first rolled out Knowledge panels, which were also often inaccurate and unhelpful.
If this isnāt acceptable from Google, then it shouldnāt be acceptable from DuckDuckGo, which has the same page-one results including an AI summary and panels, nor any other search engines. If summaries are unacceptable from Gemini, which has handily topped the leaderboards for weeks, then itās not acceptable using models from any other vendor, including Alibaba, High-Flyer, Meta, Microsoft, or Twitter.
If fake, hallucinated, confabulated, or synthetic search results are ruining the Web today, then they were ruining the Web over two decades ago and have not lessened since. The economic incentives and actors have shifted slightly, but the overall goal of fraudulent clicks still underlies the presentation.
If machine learning isnāt acceptable in collating search results today, then search engines would not exist. The issue is sheer data; ever since about 1991, before the Web existed, there has been too much data available on the Internet to search exhaustively and quickly. The problem is recursive: when a user queries a popular search engine, their results are populated by multiple different searchers using different techniques to learn what is relevant, because no one search strategy works at scale for most users asking most things.
Iām not saying this to defend Google but to steer yāall away from uncanny-valley reactionism. The search-engine business model was always odious, but we were willing to tolerate it because it was very inaccurate and easy to game, like a silly automaton which obeys simple rules. Now we are approaching the ability to conduct automated reference interviews and suddenly we have an āoops, all AI!ā moment as if it werenāt always generative AI from the beginning.
For posterity: English Wikipedia is deletionist, so your burden of proof is entirely backwards. I know this because I quit English WP over it; the sibling replies are from current editors who have fully internalized it. English WPās notability bar is very high and not moved by quantity of sources; it also has suffered from many cranks over the years, and we should not legitimize cranks merely because they publish on ArXiv.
We can read between the lines for ourselves. From OpenAIās announcement of Stargate in January, the only equity-holder who has built datacenters is Oracle, and the only other technology partner who has built datacenters is Microsoft. They claim that OpenAI will be operationally responsible, but OpenAI doesnāt have a team dedicated to building out and staffing datacenters. In related reporting, Microsoft relaxed its exclusive rights to OpenAIās infrastructure specifically for Oracle and Stargate. As for the motives, Iāll highlight Edās reporting:
The Oracle/Stargate situation was a direct result ā according to reporting from The Information ā of OpenAI becoming frustrated with Microsoft for not providing it with servers fast enough, including an allotment of 300,000 of NVIDIAās GB200 chips by the end of 2025.
Why is Microsoft canceling a Gigawatt of data center capacity while telling everybody that it didnāt have enough data centers to handle demand for its AI products? I suppose thereās one way of looking at it: that Microsoft may currently have a capacity issue, but soon wonāt, meaning that further expansion is unnecessary.
This is precisely it. Internally, Microsoftās SREs perform multiple levels of capacity planning, so that a product might individually be growing and requiring more resources over the next few months, but a department might be overall shrinking and using less capacity over the next few years. A datacenter requires at least 4yrs of construction before its capacity is available (usually more like 5yrs) which is too long of a horizon for any individual productā¦unless, of course, your product is ChatGPT and it requires a datacenterās worth of resources. Even if OpenAI were siloed from Microsoft or Azure, they would still know that OpenAI is among their neediest customers and include them in planning.
Source: Scuttlebutt from other SREs, mostly. An analogous situation happened with Googleās App Engine product: App Engineās biggest users impacted App Engineās internal capacity planning at the product level, which impacted datacenter planning because App Engine was mostly built from one big footprint in one little Oklahoma datacenter.
Conclusion: Microsoftās going to drop OpenAI as a customer. Oracleās going to pick up the responsibility. Microsoft knows that thereās no money to be made here, and is eager to see how expensive that lesson will be for Oracle; Oracle is fairly new to the business of running a public cloud and likely thinks they can offer a better platform than Azure, especially when fueled by delicious Arabian oil-fund money. Folks may want to close OpenAI accounts if they donāt want Oracle billing them someday.
Reading through the docket, he is entitled to a hearing for relief and has a modicum of standing due to the threat of deportation from the USA to China; itās not unreasonable to go to federal court. The judge was fairly courteous in referring him to the Pro Se Project a week ago. Iām a little jealous of how detached he is from reality; from 36(a) of the Amended Complaint:
The Plaintiff asserts that completing a Ph.D. in Health Services Research significantly increases earning potential. The average salary for individuals with such a Ph.D. is $120,000 annually, compared to $30,000 annually in China, where Plaintiffās visa cancellation forces him to seek employment. Over an estimated 30-year working career, this represents a lifetime income loss of $2,700,000.
He really went up to the judge and said, āyour honor, my future career is dependent on how well I prompt ChatGPT, but statistically I should be paid more if I have a second doctorate,ā and the judge patted him on his head and gave him a lollipop for being so precocious.
Well, how do you feel about robotics?
On one hand, I fully agree with you. AI is a rebranding of cybernetics, and both fields are fundamentally inseparable from robotics. The goal of robotics is to create artificial slaves who will labor without wages or solidarity. Weāre all ethically obliged to question the way that robots affect our lives.
On the other hand, machine learning (ML) isnāt going anywhere. In my oversimplification of history, ML was originally developed by Markov and Shannon to make chatbots and predict the weather; we still want to predict the weather, so even a complete death of the chatbot industry wonāt kill ML. Similarly, some robotics and cybernetics research is still useful even when not applied to replacing humans; robotics is where we learned to apply kinematics, and cybernetics gave us the concept of a massive system that we only partially see and interact with, leading to systems theory.
Hereās the kicker: at the end of the day, most people will straight-up refuse to grok that robotics is about slavery. Theyāll usually refuse to even examine the etymology, let alone the history of dozens of sci-fi authors exploring how robots are slaves or the reality today of robots serving humans in a variety of scenarios. They fundamentally donāt see that humans are aggressively chauvinist and exceptionalist in their conception of work and labor. Itās a painful and slow conversation just to get them to see the word robota
.
Starting the week with yet another excellent sneer about Dan Gackle on HN. The original post is in reply to a common complaint: politics shouldnāt be flagged so quickly. First, the scene is set:
The story goes, at least a few people donāt like hearing about Musk so often, and so we need to let all news about the rapid strip-mining of our government and economy be flagged without question.
The capital class are set to receive trillions in tax breaks off the gutting of things like Medicaid and foreign aid to the poorest and most vulnerable people in the world. The CEO of YC and Paul Graham are cheer-leading the provably racist and inexperienced DOGE team. That dozens of stories about their incredibly damaging antics are being flagged on HN is purely for the good of us tech peasants, and nothing to do with the massive tax breaks for billionaires.
But this sneer goes above and beyond, accusing Gackle of steering the communityās politics through abuse of the opaque flagging mechanism and lack of moderator logs:
Remember, dang wants us all to know that these flags are for the good of the community, and by our own hand. All the flaggers of these stories that heās seen are ālegitā. No you canāt look at the logs.
And no, you canāt make a thread to discuss this without it getting flagged; how dare you even ask that. Now let Musk reverse Robin Hood those trillions in peace, and stop trying to rile up the tech-peasantry.
Iām not really surprised to see folks accusing the bartender of the Nazi Bar of being a member of the Nazi Party; itās a reasonable conclusion given the shitty moderation over there. Edit: Restored original formatting in quote.
The sibling comment gives a wider perspective. Iām going to only respond narrowly on that final paragraphās original point.
String theories arise naturally from thinking about objects vibrating in spacetime. As such, theyāve generally been included in tests of particle physics whenever feasible. The LHC tested and (statistically) falsified some string theories. String theorists also have a sort of self-regulating ratchet which excludes unphysical theories, most recently excluding swampland theories. Most money in particle physics is going towards nuclear power, colliders like LHC or Fermilabās loops, or specialized detectors like SK (a giant tank of water) or LIGO (artfully-arranged laser beams) which mostly have to sit still and not be disturbed; in all cases, that money is going towards verification and operationalization of the Standard Model, and any non-standard theories are only coincidentally funded.
So just by double-checking the history, we see that some string theories have been falsified and that the Standard Model, not any string theory, is where most funding goes. Hossenfelder and Woit both know better, but knowing better doesnāt sell books. Gutmann doesnāt realize, I think.
Itās been frustrating to watch Gutmann slowly slide. He hasnāt slid that far yet, I suppose. Donāt discount his voice, but donāt let him be the only resource for you to learn about quantum computing; fundamentally, post-quantum concerns are a sort of hard read in one direction, and Gutmann has decided to try a hard read in the opposite direction.
Page 19, complaining about lattice-based algorithms, is hypocritical; lattice-based approaches are roughly as well-studied as classical cryptography (Feistel networks, RSA) and elliptic curves. Yes, we havenāt proven that lattice-based algorithms have the properties that we want, but we havenāt proven them for classical circuits or over elliptic curves, either, and we nonetheless use those today for TLS and SSH.
Pages 28 and 29 are outright science denial and anti-intellectualism. By quoting Woit and Hossenfelder ā who are sneerable in their own right for writing multiple anti-science books each ā he is choosing anti-maths allies, which is not going to work for a subfield of maths like computer science or cryptography. In particular, p28 lies to the reader with a doubly-bogus analogy, claiming that both string theory and quantum computing are non-falsifiable and draw money away from other research. This sort of closing argument makes me doubt the entire premise.
Look, I get your perspective, but zooming out there is a context that nobodyās mentioning, and the thread deteriorated into name-calling instead of looking for insight.
In theory, a training pass needs one readthrough of the input data, and we know of existing systems that achieve that, from well-trodden n-gram models to the wholly-hypothetical large Lempel-Ziv models. Viewed that way, most modern training methods are extremely wasteful: Transformers, Mamba, RWKV, etc. are trading time for space to try to make relatively small models, and itās an expensive tradeoff.
From that perspective, we should expect somebody to eventually demonstrate that the Transformers paradigm sucks. Mamba and RWKV are good examples of modifying old ideas about RNNs to take advantage of GPUs, but are still stuck in the idea that having a GPU perform lots of gradient descent is good. If you want to critique something, critique the gradient worship!
I swear, itās like whenever Chinese folks do anything the rest of the blogosphere goes into panic. Iām not going to insult anybody directly but Iām so fucking tired of mathlessness.
Also, point of order: Meta open-sourced Llama so that their employees would stop using Bittorrent to leak it! Not to ākeep the rabble quietā but to appease their own developers.
West Coast of USA, late 2000s to early 2010s, yes, the thick squared dark eyeglass frames were popular. Every time I see photos of these folks, Iām reminded of a couple people I know IRL as well as folks I know professionally who still prefer the thicker frames. Personally, Iāve always needed a very heavy prescription, and so Iāve always looked for the thinnest frames, but it really was a trend a decade ago.
Somebody pointed out that HNās management is partially to blame for the situation in general, on HN. Copying their comment here because itās the sort of thing Dan might blank:
but I donāt want to get hellbanned by dang.
Who gives a fuck about HN. Consider the notion that dang is, in fact, partially to blame for this entire fiasco. He runs an easy-to-propagandize platform due how much control of information is exerted by upvotes/downvotes and unchecked flagging. Itās caused a very noticeable shift over the past decade among tech/SV/hacker voices ā the dogmatic following of anything that Musk or Thiel shit out or say, this community laps it up without hesitation. Users on HN learn what sentiment on a given topic is rewarded and repeat it in exchange for upvotes.
I look forward to all of it burning down so we can, collectively, learn our lessons and realize that building platforms where discourse itself is gamified (hn, twitter, facebook, and reddit) is exactly what led us down this path today.
Elon is an Expert Beginner: he has become proficient in executing the basics of the craft by sheer repetition, but failed to develop meaningful generalizations.
The original Expert Beginner concept was defined here in terms of the Dreyfus model, but I think itās compatible with Leeās model as well. In your wording of Leeās model, one becomes an Expert Beginner when their intuition is specialized for seeing the thing; they have seen so many punches that now everything looks like a punch and must be treated like a punch, but donāt worry, Iām a punch expert, Iāve seen so many punches, I definitely know what to do when punches are involved.
Thereās a good insight from this armchair psychoanalysis. The typical narcissist is technically capable of performing the whole pretend-to-care-for-game-theoretic-reasons behavior, provided that there is an incentive for them. However, if Elon genuinely believes himself to be Christ or Buddha or Roy, then his abilities donāt matter, because he will never have the incentive to deflate his beliefs and face his own limitations and mortality. In short, Elonās attitude canāt be adjusted and his mental health will never improve.
You may have heard that Catturd doesnāt have any fiber in his diet and was hospitalized for bowel blockage. (Best sneer Iāve seen so far: ācanāt turd.ā) Along similar lines, Srid isnāt taking his statins for high cholesterol caused by a carnivore diet.
Meta: Iām kind of pissed that Catturd is WP notable but laughing my ass off at the page for carnivore diets. Life takes and gives.
Yeah, as somebody in the USA, I think that both you and @gerikson@awful.systems are pearl-clutching over laboratory conditions while ignoring the other, more serious safety problems being addressed; the presentation was not exaggerating when they were talking about the lifesaving impact of gender-affirming therapy. Last thread, you sheepishly admitted that part of the synthesis is complicated by criminalization and over-regulation; this thread, Iād like a sheepish admission that about a third of the USA (by population) suffers from restrictions on their reproductive rights.
Like, yes, you shouldnāt brew your own high-proof alcohol at home, because you can go blind from methanol poisoning. But also, there was a time in the USA when high-proof alcohol was over-regulated, and it incentivized a lot of people to homebrew.
Todayās āLuigi isnāt sexyā poster is Thomas Ptacek. The funniest example is probably this reply on the orange site:
Thatās an extrapolation from a poll, not literally 50 million peopleā¦
A cryptographer not believing in statistical analysis! I canāt stop giggling, sorry.
In lesser corruption news, California Governor Gavin Newsom has been caught distributing burner phones to California-based CEOs. These are people that likely already have Newsomās personal and business numbers, so itās not hard to imagine that these phones are likely to facilitate extralegal conversations beyond the existing
briberylegitimate business lobbying before the Legislature. With this play, Newsomās putting a lot of faith into his sexting game.