Stubsack: weekly thread for sneers not worth an entire post, week ending 3rd November 2024
Need to let loose a primal scream without collecting footnotes first? Have a sneer percolating in your system but not enough time/energy to make a whole post about it? Go forth and be mid: Welcome to the Stubsack, your first port of call for learning fresh Awful you’ll near-instantly regret.
Any awful.systems sub may be subsneered in this subthread, techtakes or no.
If your sneer seems higher quality than you thought, feel free to cut’n’paste it into its own post — there’s no quota for posting and the bar really isn’t that high.
The post Xitter web has spawned soo many “esoteric” right wing freaks, but there’s no appropriate sneer-space for them. I’m talking redscare-ish, reality challenged “culture critics” who write about everything but understand nothing. I’m talking about reply-guys who make the same 6 tweets about the same 3 subjects. They’re inescapable at this point, yet I don’t see them mocked (as much as they should be)
Like, there was one dude a while back who insisted that women couldn’t be surgeons because they didn’t believe in the moon or in stars? I think each and every one of these guys is uniquely fucked up and if I can’t escape them, I would love to sneer at them.
We're also using AI internally to improve our coding processes, which is boosting productivity and efficiency. Today, more than a quarter of all new code at Google is generated by AI, then reviewed and accepted by engineers. This helps our engineers do more and move faster.
Firstly, if this is literally true they're completely fucking cooked.
Google has a gigantic code generation culture, because the engineers there strongly prefer complexity to drudgery.
If you asked them to write fizzbuzz and left them in a room for twelve hours they would deliver a new programming language that generalized repetitive string printing, with an extension language for potential non-string-printing actions.
I left in ‘22 but feel fairly confident that “25% of code generated by AI” is going to be more of the same.
God knows I like a good DSL, but "complexity over drudgery" just sounds miserable. I also wonder what kind of stuff they're coding that's supposedly trivial enough to be generated by AI.
Best case scenario they are using a loose definition of AI to mean any code generated by other code in order to signal to investors that google isn’t the hulking, sluggish monolith that it is and is agile enough to use AI.
Worst case scenario: “hey chatgpt pls write me new search algorithm to print money, thanks, sundar”
At least for me in the US, performance was very good. I was able to 100% Sekiro, for example.
The reason I think it was a freebie is:
Everyone was stuck in-doors about six months after launch
Everybody wanted to play videogames, but no one could get GPUs and the console situation was not great
Cyberpunk 2022 2077 came out and tons of people wanted to play it. It ran terribly on consoles and on PCs, but surprisingly well on Stadia at launch
It may have still failed altogether anyway, but the fact that they didn't seize this opportunity, and instead stuck by their absolutely confusing-as-fuck "like Netflix but not really; first let me explain how this works" subscription model, always gets me.
but the fact that they didn’t seize this opportunity
honestly, I think they did try, and ran into the unfortunate reality of physics
to make that product work, you need reliable high throughput (this is helped by codecs), sufficient juggle-able GPU space (this is helped by being a gear-hogging first-in-line monopolist), and lastly the casual little requirement of actually being close enough to your customer base
iirc US cost to coast latency is around 65~70ms (so 2x that is the upper timebound for player interactivity, obvs there it'd be less because more local DCs though). just from me to europe is 165msec+, with a far less predictable path throughput. the scale economics to launch a DC for this in ZA (even to serve subsaharan africa all the way up to kenya) just plain doesn't work, and there are many more places in the world where it doesn't
it'll be interesting to see if a retrospective as to why it failed leaks out of that biz someday
If the purpose of a metric is to show adoption, the metric can be defined in a way to show adoption. Could be just an effect of promo driven culture, AI push and good'ol Goodhart's law.
Like, how do you even measure when code is ai authored and when not. If you insert 25% of a variable name and the autocompleter guesses the rest of the name correctly, are the remaining 75% AI generated?