this post was submitted on 20 Jan 2025
26 points (100.0% liked)

TechTakes

1638 readers
54 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
 

Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.

Any awful.systems sub may be subsneered in this subthread, techtakes or no.

If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.

The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)

Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.

(Semi-obligatory thanks to @dgerard for starting this.)

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 19 points 1 month ago* (last edited 1 month ago) (1 children)

PSA: the domain shown in a google search result is not necessarily the domain it will actually link you to.

Maybe this is common knowledge, but I had no idea before. What an absolutely horrible decision from google to allow this. What are they thinking?? This is great for phishing and malware, but I don't know what else. (Yeah ok, the reason has probably something to do with "line must go up".)

[–] [email protected] 11 points 1 month ago

I recall seeing something of this sort happening on goog for about 12~18mo - every so often a researcher post does the rounds where someone finds Yet Another way goog is fucking it up

the advertising dept has completely captured all mindshare and it is (demonstrably) the only part that goog-the-business cares about

[–] [email protected] 15 points 4 weeks ago (3 children)

From the "flipping through LessWrong for entertainment" department:

What effect does LLM use have on the quality of people's thinking / knowledge?

  • I'd expect a large positive effect from just making people more informed / enabling them to interpret things correctly / pointing out fallacies etc.
[–] [email protected] 13 points 3 weeks ago (1 children)

You'd think the AI safety chuds would have more reservations about using GPT, which they believe has sapience, to learn things. They have the concept of an AI being a good convincer, which, hey, idiots, how have none of you thought the great convincing has started? Also, how have none of you realised that maybe you should be a little harder to convince in general???

load more comments (1 replies)
load more comments (2 replies)
[–] [email protected] 14 points 3 weeks ago (19 children)
load more comments (19 replies)
[–] [email protected] 14 points 1 month ago (1 children)

Reposting this for the new week thread since it truly is a record of how untrustworthy sammy and co are. Remember how OAI claimed that O3 had displayed superhuman levels on the mega hard Frontier Math exam written by Fields Medalist? Funny/totally not fishy story haha. Turns out OAI had exclusive access to that test for months and funded its creation and refused to let the creators of test publicly acknowledge this until after OAI did their big stupid magic trick.

From Subbarao Kambhampati via linkedIn:

"𝐎𝐧 𝐭𝐡𝐞 𝐬𝐞𝐞𝐝𝐲 𝐨𝐩𝐭𝐢𝐜𝐬 𝐨𝐟 “𝑩𝒖𝒊𝒍𝒅𝒊𝒏𝒈 𝒂𝒏 𝑨𝑮𝑰 𝑴𝒐𝒂𝒕 𝒃𝒚 𝑪𝒐𝒓𝒓𝒂𝒍𝒍𝒊𝒏𝒈 𝑩𝒆𝒏𝒄𝒉𝒎𝒂𝒓𝒌 𝑪𝒓𝒆𝒂𝒕𝒐𝒓𝒔” hashtag#SundayHarangue. One of the big reasons for the increased volume of “𝐀𝐆𝐈 𝐓𝐨𝐦𝐨𝐫𝐫𝐨𝐰” hype has been o3’s performance on the “frontier math” benchmark–something that other models basically had no handle on.

We are now being told (https://lnkd.in/gUaGKuAE) that this benchmark data may have been exclusively available (https://lnkd.in/g5E3tcse) to OpenAI since before o1–and that the benchmark creators were not allowed to disclose this *until after o3 *.

That o3 does well on frontier math held-out set is impressive, no doubt, but the mental picture of “𝒐1/𝒐3 𝒘𝒆𝒓𝒆 𝒋𝒖𝒔𝒕 𝒃𝒆𝒊𝒏𝒈 𝒕𝒓𝒂𝒊𝒏𝒆𝒅 𝒐𝒏 𝒔𝒊𝒎𝒑𝒍𝒆 𝒎𝒂𝒕𝒉, 𝒂𝒏𝒅 𝒕𝒉𝒆𝒚 𝒃𝒐𝒐𝒕𝒔𝒕𝒓𝒂𝒑𝒑𝒆𝒅 𝒕𝒉𝒆𝒎𝒔𝒆𝒍𝒗𝒆𝒔 𝒕𝒐 𝒇𝒓𝒐𝒏𝒕𝒊𝒆𝒓 𝒎𝒂𝒕𝒉”–that the AGI tomorrow crowd seem to have–that 𝘖𝘱𝘦𝘯𝘈𝘐 𝘸𝘩𝘪𝘭𝘦 𝘯𝘰𝘵 𝘦𝘹𝘱𝘭𝘪𝘤𝘪𝘵𝘭𝘺 𝘤𝘭𝘢𝘪𝘮𝘪𝘯𝘨, 𝘤𝘦𝘳𝘵𝘢𝘪𝘯𝘭𝘺 𝘥𝘪𝘥𝘯’𝘵 𝘥𝘪𝘳𝘦𝘤𝘵𝘭𝘺 𝘤𝘰𝘯𝘵𝘳𝘢𝘥𝘪𝘤𝘵–is shattered by this. (I have, in fact, been grumbling to my students since o3 announcement that I don’t completely believe that OpenAI didn’t have access to the Olympiad/Frontier Math data before hand… )

I do think o1/o3 are impressive technical achievements (see https://lnkd.in/gvVqmTG9 )

𝑫𝒐𝒊𝒏𝒈 𝒘𝒆𝒍𝒍 𝒐𝒏 𝒉𝒂𝒓𝒅 𝒃𝒆𝒏𝒄𝒉𝒎𝒂𝒓𝒌𝒔 𝒕𝒉𝒂𝒕 𝒚𝒐𝒖 𝒉𝒂𝒅 𝒑𝒓𝒊𝒐𝒓 𝒂𝒄𝒄𝒆𝒔𝒔 𝒕𝒐 𝒊𝒔 𝒔𝒕𝒊𝒍𝒍 𝒊𝒎𝒑𝒓𝒆𝒔𝒔𝒊𝒗𝒆–𝒃𝒖𝒕 𝒅𝒐𝒆𝒔𝒏’𝒕 𝒒𝒖𝒊𝒕𝒆 𝒔𝒄𝒓𝒆𝒂𝒎 “𝑨𝑮𝑰 𝑻𝒐𝒎𝒐𝒓𝒓𝒐𝒘.”

We all know that data contamination is an issue with LLMs and LRMs. We also know that reasoning claims need more careful vetting than “𝘸𝘦 𝘥𝘪𝘥𝘯’𝘵 𝘴𝘦𝘦 𝘵𝘩𝘢𝘵 𝘴𝘱𝘦𝘤𝘪𝘧𝘪𝘤 𝘱𝘳𝘰𝘣𝘭𝘦𝘮 𝘪𝘯𝘴𝘵𝘢𝘯𝘤𝘦 𝘥𝘶𝘳𝘪𝘯𝘨 𝘵𝘳𝘢𝘪𝘯𝘪𝘯𝘨” (see “In vs. Out of Distribution analyses are not that useful for understanding LLM reasoning capabilities” https://lnkd.in/gZ2wBM_F ).

At the very least, this episode further argues for increased vigilance/skepticism on the part of AI research community in how they parse the benchmark claims put out commercial entities."

Big stupid snake oil strikes again.

[–] [email protected] 15 points 1 month ago (4 children)

Every time they go 'this wasnt in the data' it turns out it was. A while back they did the same with translating rareish languages. Turns out it was trained on it. Fucked up. But also, wtf how are they expecting this to stay secret and there being no backlash? This world needs a better class of criminals.

[–] [email protected] 13 points 1 month ago (1 children)

The conspiracy theorist who lives in my brain wants to say its intentional to make us more open to blatant cheating as something that's just a "cost of doing business." (I swear I saw this phrase a half dozen times in the orange site thread about this)

The earnest part of me tells me no, these guys are just clowns, but I dunno, they can't all be this dumb right?

[–] [email protected] 12 points 1 month ago (7 children)

holy shit, that’s the excuse they’re going for? they cheated on a benchmark so hard the results are totally meaningless, sold their most expensive new models yet on the back of that cheated benchmark, further eroded the scientific process both with their cheating and by selling those models as better for scientific research… and these weird fucks want that to be fine and normal? fuck them

load more comments (7 replies)
load more comments (3 replies)
[–] [email protected] 14 points 4 weeks ago (17 children)

This is a thought I've been entertaining for some time, but this week's discussion about Ars Technica's article on Anthropic, as well as the NIH funding freeze, finally prodded me to put it out there.

A core strategic vulnerability that Musk, his hangers-on, and geek culture more broadly haven't cottoned onto yet: Space is 20th-century propaganda. Certainly, there is still worthwhile and inspirational science to be done with space probes and landers; and the terrestrial satellite network won't dwindle in importance. I went to high school with a guy who went on to do his PhD and get into research through working with the first round of micro-satellites. Resources will still be committed to space. But as a core narrative of technical progress to bind a nation together? It's gassed. The idea that "it might be ME up there one day!" persisted through the space shuttle era, but it seems more and more remote. Going back to the moon would be a remake of an old television show, that went off the air because people ended up getting bored with it the first time. Boots on Mars (at least healthy boots with a solid chance to return home) are decades away, even if we start throwing Apollo money at it immediately. The more outlandish ideas like orbital data centers and asteroid mining don't have the same inspirational power, because they are meant to be private enterprises operated by thoroughly unlikeable men who have shackled themselves to a broadly destructive political program.

For better or worse, biotechnology and nanotechnology are the most important technical programs of the 21st century, and by backgrounding this and allowing Trump to threaten funding, the tech oligarchs kowtowing to him right now are undermining themselves. Biotech should be obvious, although regulatory capture and the impulse for rent-seeking will continue to hold it back in the US. I expect even more money to be thrown at nanotechnology manufacturing going into the 2030s, to try to overcome the fact that semiconductor scaling is hitting a wall, although most of what I've seen so far is still pursuing the Drexlerian vision of MEMS emulating larger mechanical systems... which, if it's not explicitly biocompatible, is likely going down a cul-de-sac.

Everybody's looking for a positive vision of the future to sell, to compete with and overcome the fraudulent tech-fascists who lead the industry right now. A program of accessible technology at the juncture of those two fields would not develop overnight, but could be a pathway there. Am I off base here?

[–] [email protected] 12 points 3 weeks ago (1 children)

This seems like yet another disconnect between however the fuck science communication has been failing the general public and myself.

Like when you say space I think, fuck yeah, space! Those crisp pictures of Pluto! Pictures of black holes! The amazing JWST data! Gravitational waves detection! Recreating the conditions of the early universe in particle accelerators to unlock the secrets of spacetime! Just most amazing geek shit that makes me as excited as I was when I was 12 looking at the night sky through my cheap-ass telescope.

Who gives a single fuck about sending people up there when we have probes and rovers, true marvels of engineering, feeding us data back here? Did you know Voyager 1, Voyager Fucking ONE, almost 50 years old probe, over 150 AU away from Earth, is STILL SENDING US DATA? We engineered the fuck of that bolt bucket so that even the people that designed it are surprised by how long it lasted. You think a human would last 50 years in the interstellar medium? I don't fucking think so.

We're unlocking the secrets of the universe and confirming theories from decades ago, has there been a more exciting time to be a scientist? Wouldn't you want to run a particle accelerator? Do science on the ISS? Be the engineer behind the next legendary probe that will benefit mankind even after you're gone? If you can't spin this into a narrative of technical progrees and humans being amazing then that's a skill issue, you lack fucking whimsy.

And I don't think there's a person in the world less whimsical than Elon fucking Musk.

load more comments (1 replies)
[–] [email protected] 11 points 4 weeks ago

Agree with space travel being retro-futurist fluff. It's very rich men badly remembering mediocre science fiction.

The US could lead the world in innovation in green technology but that's now tainted by wokeness.

load more comments (15 replies)
[–] [email protected] 14 points 1 month ago (8 children)

Good news, everyone: critihype is canceled until the next tweet.

https://xcancel.com/sama/status/1881258443669172470#m

[–] [email protected] 11 points 1 month ago

Hmm, surely there is no downside to doing all of one's marketing, both personal* and professional, through the false certainty and low signal of short-form social media. The leopard has only licked Sam's face, it will never bite and begin chewing!

*You and I may find the concept of a "personal brand" to be horrifying, but these guys clearly want to become brands more fervently than Bruce Wayne wanted to become a bat

load more comments (7 replies)
[–] [email protected] 13 points 1 month ago* (last edited 1 month ago) (4 children)

til that there's not one millionaire with family business in south african mining in current american oligarchy, but at least two. (thiel's father was an exec at mine in what is today Namibia). (they mined uranium). (it went towards RSA nuclear program). (that's easily most ghoulish thing i've learned today, but i'm up only for 2h)

load more comments (4 replies)
[–] [email protected] 13 points 1 month ago (1 children)
load more comments (1 replies)
[–] [email protected] 12 points 4 weeks ago* (last edited 4 weeks ago) (5 children)

I hope everyone is ready for the constant overlap between politics and AI / Silicon Valley; because I'm not.

Trump Admin Accused of Using AI to Draft Executive Orders (Source Bluesky Thread).

I'm not 100% sure I buy that the EOs were written by AI rather than people who simply don't care about or don't know the details; but it certainly looks possible. Especially that example about the Gulf of Mexico. Either way I am heartened that this is the conclusion people jump to.

Aside: I also like how much media is starting to cite bluesky (and activitypub to a lesser extent). I assume a bunch of journalists moved off of twitter or went multi-platform.

load more comments (5 replies)
[–] [email protected] 11 points 1 month ago (2 children)

Banner start to the next US presidency, with Wiener Von Wrong tossing a Nazi salute and the ADL papering that one over as an "awkward gesture". 2025 is going to be great for my country.

Incidentally is "Wiener Von Wrong" or "Wernher Von Brownnose" better?

[–] [email protected] 9 points 1 month ago (1 children)

Perhaps "Wanker von Clown"?

load more comments (1 replies)
load more comments (1 replies)
[–] [email protected] 11 points 3 weeks ago* (last edited 3 weeks ago) (6 children)

so I ran into this fucking garbage earlier, which goes so hard on the constituent parts of "the spam is the point", an ouroborosian self-reinforcing loop of Just More Media Bro Just One More Video Bro You'll See Bro It'll Be The Best Listicle Bro Just Watch Bro, and the insufferably cancerous "the medium is the message" videos-made-for-youtube-because-youtube that if it were a voltron it'd probably have its own unique Special Moment sequence instead of being one of the canned assembly shots

[–] [email protected] 11 points 3 weeks ago (5 children)

various topics (e.g., AI news, crypto, fitness, personal finance)

That sure is a specifc selection of topics.

load more comments (5 replies)
[–] [email protected] 10 points 3 weeks ago (2 children)

I wish YouTube would ban this shit wholesale, but it’s Google and of course they won’t.

Aside: I’ve been hammering “Don’t recommend this channel” on every video that remotely smells like AI slop for a while and so far that seems to keep the feed fairly clean.

load more comments (2 replies)
load more comments (4 replies)
[–] [email protected] 10 points 4 weeks ago* (last edited 4 weeks ago) (5 children)

Buckle up humans; because humanity's last exam just dropped: https://lastexam.ai/ (Hacker News discussion). May the odds be ever in your favor.

Edit: Per NyTimes, whom I hate, they were apparently trying to avoid an over-dramatic name. Amazing:

The test’s original name, “Humanity’s Last Stand,” was discarded for being overly dramatic.

[–] [email protected] 13 points 4 weeks ago* (last edited 4 weeks ago) (1 children)

i only want to notice that the example chemistry question has two steps out of three that are very similar to last image in wikipedia article https://en.wikipedia.org/wiki/Electrocyclic_reaction (question explicitly mentions that it is electrocyclic reaction and mentions the same class of natural product)

e: the exact reaction scheme that is answer to that question is in article linked just above that image. taking last image from wiki article and one of schemes from cited article gives the exact same compound as in question, and provides answer. considering how these spicy autocomplete rainforest incinerators work, this sounds like some serious ratfucking, right? you don't even have to know how this all works to get that and it's an isolated and a bit obscure subsubfield

[–] [email protected] 11 points 4 weeks ago (2 children)

You think people would secretly submit easy questions just for the reward money, and that since the question database is so big and inscrutable no one bothered to verify one way or another? No, that could never happen.

[–] [email protected] 10 points 4 weeks ago* (last edited 4 weeks ago) (2 children)

well, it's not the most obvious thing but not because it's easy, it's because it's almost a trivia, a sort of thing you can see once in textbook and then never use it ever for anything and that doesn't really connects readily to anything else, most of the time. i haven't done electrocyclic reaction once in my entire phd programme, and last time i've seen them was in second year ochem course. these kinds of reactions are not very controllable or clean, synthesis of precursors looks like a major PITA, precursors would probably have to be kept in freezer under argon for maybe days before they decompose, and introduction of any modifications requires you to redo multistep synthesis, and then it might fail to work. i also suspect that this exact example might be in some undergrad textbook verbatim, and it will be in scihub pdfs at any rate. it's also kinda old stuff with research starting in 60s

load more comments (2 replies)
load more comments (1 replies)
[–] [email protected] 11 points 4 weeks ago* (last edited 4 weeks ago)

oh cool, the logo’s just a barely modified sparkle emoji so you know it’s horseshit, and it’s directly funded by Scale AI and a Rationalist thinktank so the chances the models weren’t directly trained on the problem set are vanishingly thin. this is just the FrontierMath grift with new, more dramatic, paint.

e: also, slightly different targeting — FrontierMath was looking to grift institutional dollars, I feel. this one’s designed to look good in a breathless thinkpiece about how, I dunno…

When A.I. Passes This Test, Look Out

yeah, whatever the fuck they think this means. this one’s designed to be talked about, to be brought up behind closed doors as a reason why your pay’s being cut. this is vile shit.

load more comments (3 replies)
[–] [email protected] 10 points 1 month ago

Polish commentary on Hitlergruß: https://bsky.app/profile/smutnehistorie.bsky.social/post/3lgaoyezhgc2c

Translation:

  • it’s just a Hindu symbol of prosperity
  • a normal Roman salute
  • regular rail car
  • wait a second
[–] [email protected] 9 points 1 month ago* (last edited 1 month ago) (2 children)

hackernews: We're going to build utopia on Mars, reinvent money, and construct god.

also hackernews: moving off facebook is too hard :( :( :(

https://news.ycombinator.com/item?id=42779776

load more comments (2 replies)
[–] [email protected] 9 points 4 weeks ago (3 children)

d'ya think this post on awful.systems, the lemmy instance (which is known as awful.systems), is the location of this awful.systems thread? let me hear your thoughts, awful.systems

[–] [email protected] 10 points 4 weeks ago (1 children)

you get even more awful.systems on remote instances

[–] [email protected] 12 points 4 weeks ago (3 children)
load more comments (3 replies)
load more comments (2 replies)
load more comments
view more: next ›