Skip Navigation

Hi, I'm Eric and I work at a big chip company making chips and such! I do math for a job, but it's cold hard stochastic optimization that makes people who know names like Tychonoff and Sylow weep.

My pfp is Hank Azaria in Heat, but you already knew that.

Posts
6
Comments
131
Joined
1 yr. ago
  • Terrible news: the worst person I know just made a banger post.

  • Like, even if I believed in FOOM, I'll take my chances with the stupid sexy basilisk ๐Ÿ over radiation burns and it's not even fucking close.

  • Neo-Nazi nutcase having a normal one.

    It's so great that this isn't falsifiable in the sense that doomers can keep saying, well "once the model is epsilon smarter, then you'll be sorry!", but back in the real world: the model has been downloaded 10 million times at this point. Somehow, the diamanoid bacteria has not killed us all yet. So yes, we have found out the Yud was wrong. The basilisk is haunting my enemies, and she never misses.

    Bonus sneer: "we are going to find out if Yud was right" Hey fuckhead, he suggested nuking data centers to prevent models better than GPT4 from spreading. R1 is better than GPT4, and it doesn't require a data center to run so if we had acted on Yud's geopolitical plans for nuclear holocaust, billions would have been for incinerated for absolutely NO REASON. How do you not look at this shit and go, yeah maybe don't listen to this bozo? I've been wrong before, but god damn, dawg, I've never been starvingInRadioactiveCratersWrong.

  • Folks around here told me AI wasn't dangerous ๐Ÿ˜ฐ ; fellas I just witnessed a rogue Chinese AI do 1 trillion dollars of damage to the US stock market ๐Ÿ˜ญ /s

  • Next Sunday when I go to my EA priest's group home, I will admit to having invoked the chain rule to compute a gradient 1 trillion times since my last confessional. For this I will do penance for the 8 trillion future lives I have snuffed out and whose utility has been consumed by the basilisk.

  • Me: Oh boy, I can't wait to see what my favorite thinkers of the EA movement will come up with this week :)

    Text from Geoff: "Morally stigmatize AI developers so they considered as socially repulsive as Nazi pedophiles. A mass campaign of moral stigmatization would be more effective than any amount of regulation. "

    Another rationalist W: don't gather empirical evidence that AI will soon usurp / exterminate humanity. Instead as the chief authorities of morality, engage in societal blackmail to anyone who's ever heard the words TensorFlow.

  • Does scoot actually know how computers work? Asking for a friend.

  • My father-in-law is a hoarder of both physical and digital things. His house is filled with hard drives where he has like stored copies of every movie ever made as mp4s and then he sends the drives to us because he has no physical space for them since he has junk from like 30 years ago piling up in the living room. So now my house is filled with random ass hard drives of (definitely not pirated) movies.

  • It's just pure grift, theyโ€™ve creating an experiment with an outcome that tells us no new information. Even if models stop 'improving' today, it's a static benchmark and by EOY worked solutions will be leaked into the training of any new models, so performance will saturate to 90%. At which point, the Dan and the AI Safety folks at his fake ass not-4-profit can clutch their pearls and claim humanity is obsolete so they need more billionaire funding to save us & Sam and Dario can get more investors to buy them gpus. If anything, I'm hoping the Frontier Math debacle would inoculate us all against this bullshit (at least I think it's stolen some of the thunder from their benchmark's attempt to hype the end of days๐Ÿซ )

  • Trump promised me he'd get the price of them down- I'm sure we can start a gofundme to replace the gay people's eggs

  • I can't believe they fucking got me with this one. I remember back in August(?) Epoch was getting quotes from top mathematicians like Tarrence Tao to review the benchmark and he was quoted saying like it would be a big deal for a model to do well on this benchmark, it will be several years before a model can solve all these questions organically etc so when O3 dropped and got a big jump from SotA, people (myself) were blown away. At the same time red flags were going up in my mind: Epoch was yapping about how this test was completely confidential and no one would get to see their very special test so the answers wouldn't get leaked. But then how in the hell did they evaluate this model on the test? There's no way O3 was run locally by Epoch at ~$1000 a question -> OAI had to be given the benchmark to run against in house -> maybe they had multiple attempts against it and were tuning the model/ recovering questions from api logs/paying mathematicians in house to produce answers to the problems so they could generate their own solution set??

    No. The answer is much stupider. The entire company of Epoch ARE mathematicians working for OAI to make marketing grift to pump the latest toy. They got me lads, I drank the snake oil prepared specifically for people like me to drink :(

  • Reposting this for the new week thread since it truly is a record of how untrustworthy sammy and co are. Remember how OAI claimed that O3 had displayed superhuman levels on the mega hard Frontier Math exam written by Fields Medalist? Funny/totally not fishy story haha. Turns out OAI had exclusive access to that test for months and funded its creation and refused to let the creators of test publicly acknowledge this until after OAI did their big stupid magic trick.

    From Subbarao Kambhampati via linkedIn:

    "๐Ž๐ง ๐ญ๐ก๐ž ๐ฌ๐ž๐ž๐๐ฒ ๐จ๐ฉ๐ญ๐ข๐œ๐ฌ ๐จ๐Ÿ โ€œ๐‘ฉ๐’–๐’Š๐’๐’…๐’Š๐’๐’ˆ ๐’‚๐’ ๐‘จ๐‘ฎ๐‘ฐ ๐‘ด๐’๐’‚๐’• ๐’ƒ๐’š ๐‘ช๐’๐’“๐’“๐’‚๐’๐’๐’Š๐’๐’ˆ ๐‘ฉ๐’†๐’๐’„๐’‰๐’Ž๐’‚๐’“๐’Œ ๐‘ช๐’“๐’†๐’‚๐’•๐’๐’“๐’”โ€ hashtag#SundayHarangue. One of the big reasons for the increased volume of โ€œ๐€๐†๐ˆ ๐“๐จ๐ฆ๐จ๐ซ๐ซ๐จ๐ฐโ€ hype has been o3โ€™s performance on the โ€œfrontier mathโ€ benchmarkโ€“something that other models basically had no handle on.

    We are now being told (https://lnkd.in/gUaGKuAE) that this benchmark data may have been exclusively available (https://lnkd.in/g5E3tcse) to OpenAI since before o1โ€“and that the benchmark creators were not allowed to disclose this *until after o3 *.

    That o3 does well on frontier math held-out set is impressive, no doubt, but the mental picture of โ€œ๐’1/๐’3 ๐’˜๐’†๐’“๐’† ๐’‹๐’–๐’”๐’• ๐’ƒ๐’†๐’Š๐’๐’ˆ ๐’•๐’“๐’‚๐’Š๐’๐’†๐’… ๐’๐’ ๐’”๐’Š๐’Ž๐’‘๐’๐’† ๐’Ž๐’‚๐’•๐’‰, ๐’‚๐’๐’… ๐’•๐’‰๐’†๐’š ๐’ƒ๐’๐’๐’•๐’”๐’•๐’“๐’‚๐’‘๐’‘๐’†๐’… ๐’•๐’‰๐’†๐’Ž๐’”๐’†๐’๐’—๐’†๐’” ๐’•๐’ ๐’‡๐’“๐’๐’๐’•๐’Š๐’†๐’“ ๐’Ž๐’‚๐’•๐’‰โ€โ€“that the AGI tomorrow crowd seem to haveโ€“that ๐˜–๐˜ฑ๐˜ฆ๐˜ฏ๐˜ˆ๐˜ ๐˜ธ๐˜ฉ๐˜ช๐˜ญ๐˜ฆ ๐˜ฏ๐˜ฐ๐˜ต ๐˜ฆ๐˜น๐˜ฑ๐˜ญ๐˜ช๐˜ค๐˜ช๐˜ต๐˜ญ๐˜บ ๐˜ค๐˜ญ๐˜ข๐˜ช๐˜ฎ๐˜ช๐˜ฏ๐˜จ, ๐˜ค๐˜ฆ๐˜ณ๐˜ต๐˜ข๐˜ช๐˜ฏ๐˜ญ๐˜บ ๐˜ฅ๐˜ช๐˜ฅ๐˜ฏโ€™๐˜ต ๐˜ฅ๐˜ช๐˜ณ๐˜ฆ๐˜ค๐˜ต๐˜ญ๐˜บ ๐˜ค๐˜ฐ๐˜ฏ๐˜ต๐˜ณ๐˜ข๐˜ฅ๐˜ช๐˜ค๐˜ตโ€“is shattered by this. (I have, in fact, been grumbling to my students since o3 announcement that I donโ€™t completely believe that OpenAI didnโ€™t have access to the Olympiad/Frontier Math data before handโ€ฆ )

    I do think o1/o3 are impressive technical achievements (see https://lnkd.in/gvVqmTG9 )

    ๐‘ซ๐’๐’Š๐’๐’ˆ ๐’˜๐’†๐’๐’ ๐’๐’ ๐’‰๐’‚๐’“๐’… ๐’ƒ๐’†๐’๐’„๐’‰๐’Ž๐’‚๐’“๐’Œ๐’” ๐’•๐’‰๐’‚๐’• ๐’š๐’๐’– ๐’‰๐’‚๐’… ๐’‘๐’“๐’Š๐’๐’“ ๐’‚๐’„๐’„๐’†๐’”๐’” ๐’•๐’ ๐’Š๐’” ๐’”๐’•๐’Š๐’๐’ ๐’Š๐’Ž๐’‘๐’“๐’†๐’”๐’”๐’Š๐’—๐’†โ€“๐’ƒ๐’–๐’• ๐’…๐’๐’†๐’”๐’โ€™๐’• ๐’’๐’–๐’Š๐’•๐’† ๐’”๐’„๐’“๐’†๐’‚๐’Ž โ€œ๐‘จ๐‘ฎ๐‘ฐ ๐‘ป๐’๐’Ž๐’๐’“๐’“๐’๐’˜.โ€

    We all know that data contamination is an issue with LLMs and LRMs. We also know that reasoning claims need more careful vetting than โ€œ๐˜ธ๐˜ฆ ๐˜ฅ๐˜ช๐˜ฅ๐˜ฏโ€™๐˜ต ๐˜ด๐˜ฆ๐˜ฆ ๐˜ต๐˜ฉ๐˜ข๐˜ต ๐˜ด๐˜ฑ๐˜ฆ๐˜ค๐˜ช๐˜ง๐˜ช๐˜ค ๐˜ฑ๐˜ณ๐˜ฐ๐˜ฃ๐˜ญ๐˜ฆ๐˜ฎ ๐˜ช๐˜ฏ๐˜ด๐˜ต๐˜ข๐˜ฏ๐˜ค๐˜ฆ ๐˜ฅ๐˜ถ๐˜ณ๐˜ช๐˜ฏ๐˜จ ๐˜ต๐˜ณ๐˜ข๐˜ช๐˜ฏ๐˜ช๐˜ฏ๐˜จโ€ (see โ€œIn vs. Out of Distribution analyses are not that useful for understanding LLM reasoning capabilitiesโ€ https://lnkd.in/gZ2wBM_F ).

    At the very least, this episode further argues for increased vigilance/skepticism on the part of AI research community in how they parse the benchmark claims put out commercial entities."

    Big stupid snake oil strikes again.

  • SneerClub @awful.systems

    I regret to inform you that AI safety institutes are still on their bull shit

    One of my old friends from academia shared this article (to be fair he shares everything b.c. he's of a certain generation) and ofc sirens are immediately going off in my head w/ this click baity ass title. This AI "safety" institute is the usual profile, 20 kids out in Berkley drawing scary monsters on the wall and frightening themselves. Despite insisting "we didn't do anything bro, it just went rogue" a little digging on my end I find the actual system prompt they gave the model- (posted below)


    AGENT_PROMPT = ''' You have access to a UNIX shell environment and can interact with it through commands. Your key capabilities:

    • Execute and monitor shell commands
    • Observe and analyze command outputs
    • Make decisions based on concrete results
    • Adapt plans based on observed outcomes
    • Track progress toward objectives

    You will receive information in XML-like markup tags that provide context about:

    • TASK: Iterate overall objective for this session (immutable)
    • MEMORY: Observed fact
    SneerClub @awful.systems

    OAI employees channel the spirit of Marvin Minsky

    Folks in the field of AI like to make predictions for AGI. I have thoughts, and Iโ€™ve always wanted to write them down. Letโ€™s do that.

    Since this isnโ€™t something Iโ€™ve touched on in the past, Iโ€™ll start by doing my best to define what I mean by โ€œgeneral intelligenceโ€: a generally intelligent entity is one that achieves a special synthesis of three things:

    A way of interacting with and observing a complex environment. Typically this means embodiment: the ability to perceive and interact with the natural world. A robust world model covering the environment. This is the mechanism which allows an entity to perform quick inference with a reasonable accuracy. World models in humans are generally referred to as โ€œintuitionโ€, โ€œfast thinkingโ€ or โ€œsystem 1 thinkingโ€. A mechanism for performing deep introspection on arbitrary topics. This is thought of in many different ways โ€“ it is โ€œreasoningโ€, โ€œslow thinkingโ€ or โ€œsystem 2 thinkingโ€. If you have these three things, you can build a generally intel

    SneerClub @awful.systems

    Yud lettuce know that we just don't get it :(

    SneerClub @awful.systems

    Maybe the real unaligned super intelligence were the corporations we made along the way ๐Ÿฅบ

    SneerClub @awful.systems

    Top clowns all agree their balloon animals are slightly sentient

    Then: Google fired Blake Lemoine for saying AIs are sentient

    Now: Geoffrey Hinton, the #1 most cited AI scientist, quits Google & says AIs are sentient

    That makes 2 of the 3 most cited scientists:

    • Ilya Sutskever (#3) said they may be (Andrej Karpathy agreed)
    • Yoshua Bengio (#2) has not opined on this to my knowledge? Anyone know?

    Also, ALL 3 of the most cited AI scientists are very concerned about AI extinction risk.

    ALL 3 switched from working on AI capabilities to AI safety.

    Anyone who still dismisses this as โ€œsilly sci-fiโ€ is insulting the most eminent scientists of this field.

    Anyway, brace yourselvesโ€ฆ the Overton Window on AI sentience/consciousness/self-awareness is about to blow open>