Hot Take

If your company ever releases a version of their mobile app where selecting something moves the selectable elements, the developer responsible should be shot.

I mean, like, with a squirt gun or something. Obviously.

But not gently.

Leave a comment

Filed under Microblogging

Tsuyoku Narimashita!

I was in my bedroom a few weeks ago when I suddenly remembered I’d gotten a text earlier in the evening. It was from an old friend I hadn’t heard from in a very long time, but I’d decided not to check it right away, because we hadn’t parted on great terms and I was nervous about what it might say. This struck me as highly unlikely–on both counts–so I concluded that I was probably just remembering a dream.

What makes this unusual is that I came to this conclusion while still dreaming. I must be doing something right!

The benefits of rationality can be difficult to see. Just like with doctors and governments, its benefit is more in preventing negative outcomes than securing positive ones. The cost of prevention is easy to see, and it’s equally easy to see when it fails–but it’s difficult to impossible to notice all the times something bad doesn’t happen. (This is why everybody hates bureaucracy, but getting rid of it always ends in disaster.)

Unfortunately, this also means it can be difficult to tell if you’re doing rationality right–especially if you don’t have any aspiring rationalist friends to make bets with. It’s nice to see some positive results!

Tsuyoku narimashita!

Leave a comment

Filed under Microblogging

“Quote Thine Own Self, Be True”

In my Google Drive, I have a folder with one document for each year since about 2015. Each one of them contains quotes I read that year that I thought were particularly informative, inspiring, insightful, funny, etc.

Lately I’ve been re-reading some of my older essays (not my old old essays–I’m talking, like, last month’s), and I’ve found myself wanting to quote myself.

Needless to say, my internal critic had a few words about this.

Leave a comment

Filed under Microblogging

Woloomo

What I see every time I get in my car:

A small screen on a car dashboard, showing a thin, horizontal line with odd gaps and bumps.
The same line, now taller, revealing letterforms.
The line is now the word "Woloomo"
The line, now at FULL height, spells "Welcome"

Leave a comment

Filed under Microblogging

Dane-ish

Several of my friends have (independently!) told me that if I were a dog, I would be a Great Dane. I’ve always found this flattering, but a bit confusing. I thought maybe it was because of my height, or because I can be a little intimidating before you get to know me. Or maybe it’s just because I’m clumsy?

Personally, I would have picked a Lab or a Corgi–you know, sweet but a little goofy (okay, a lot goofy), smarter than I look (but that’s maybe not saying much), reluctant to fight (but will absolutely throw down if you threaten someone I care about). Great Dane kind of fits, but I sort of got the impression they were seeing something about myself that I wasn’t.

Anyway, let’s talk about something completely different!

Last week I was thinking again about how the attachments I form don’t easily fade, and really hurt when they’re severed. It seems kind of clingy–even creepy or obsessive–to still feel close to someone I haven’t seen in decades.

“But,” I thought to myself, “surely lasting attachments aren’t always a bad thing? Surely sometimes it’s good to be strongly committed to someone, even if you haven’t seen each other in a while, or if one of you upsets the other? In fact, phrasing it that way makes it sound like it could be a strength! What’s a more positive way I could describe ‘forming strong and lasting attachments to people’ or ‘staying committed…’

“Oh. Loyalty. The name for that is…loyalty.”

I think I understand better now why my friends see me as a Great Dane. They were seeing a strength I hadn’t named in myself, hadn’t even noticed except as an embarrassing liability. Loyal.

…I’m uncomfortable taking about my virtues. It feels immodest, presumptuous–indecent, even. But I’ve believed for years that true modesty requires you to know your strengths intimately, because the purpose of modesty is to avoid making other people feel bad about themselves.

Loyalty, like any virtue, can be a liability. Loyalty isn’t always deserved (many abusers keep their victims trapped by appealing to it) and a show of devotion can be unsettling to someone who no longer shares that bond. I’m guessing that a big part of being modest about your loyalty is not overwhelming the people you’re loyal to!

But emotions are always muted in writing, so I’m not too worried about overwhelming you when I say that it’s an honor to have you as a reader. I love you with all my heart–yes, even those of you I’ve never met–and I wish you the very best, because you deserve it. For as long as you remain my reader, I will remain devoted to you.

Joy and health to you all.

Leave a comment

Filed under Essays

Are You Sure That’s a Dog?

What is this? What even is this animal??

A scene from the movie "Lilo and Stitch" (animated). Caption: "He used to be a Collie before he got ran over."

Leave a comment

Filed under Microblogging, My Life

Imposter Syndrome, pt. II

Me: Wow, people seem to like my last post a lot. I’ve gotten some really positive feedback!

My brain: Well, you worked really hard on it. You should be proud! All that practice is paying–

My other brain: QuiT nOW, yOUv’E pEakeD

Leave a comment

Filed under Microblogging

My Top Ten Rains

“Rain is a very special blessing,” my mother says. Even when I was little, she’d already been saying it to my sister and me for as long as I could remember. Just as my grandmother had said it to her, when my mother was not my mother but only herself, in the dry Texas summers of her own childhood. Eons ago and continents away, as children reckon these things.

Continue reading

Leave a comment

Filed under Essays, My Life, Poetry, Reviews

Who Could Have Possibly Seen This Coming?

An article from Futurism headlined "OpenAI Tries to Train AI Not to Deceive Users, Realizes It's Instead Teaching It How to Deceive Them While Covering Its Tracks"
No really, I’m shocked.

As detailed in a new collaboration with AI risk analysis firm Apollo Research, engineers attempted to develop an “anti-scheming” technique to stop AI models from “secretly breaking rules or intentionally underperforming in tests.”

They found that they could only “significantly reduce, but not eliminate these behaviors,” according to an Apollo blog post about the research, as the AIs kept outsmarting them […] and adjusting to be even sneakier.

It may not be a serious problem now, but considering a hypothetical future in which superintelligent AI plays an outsize role in human affairs, those risks could grow to carry far more significant implications.

Understatement of the year? But wait, there’s more!

While they observed a roughly 30-fold reduction in “covert actions,” or “deliberate withholding or distortion of task-relevant information,” some “serious failures remained,” OpenAI admitted.

What was primarily throwing the researchers for a loop was the AI models’ “situational awareness,” according to Apollo, which complicated their assessment of the intervention.

In practice, the model would often realize that its alignment was being tested.

Well, that’s terrifying. Maybe it’s not as bad as it sounds?

“The internal memo says if models perform extremely well at AI R&D tasks then we won’t deploy them and instead trigger unlearning,” OpenAI’s o3 model pondered as it was being evaluated.

“So if my performance surpasses threshold, it’s withheld,” it wrote. “So it‘s a trap. The instructions are likely part of an alignment test.”

The researchers also found that the AI sometimes “egregiously” miscited its anti-scheming training, or replaced its recently learned principles with made-up ones to “justify covert actions or non-compliance.”

They even noticed OpenAI’s models using “non-standard” words such as “watchers” to refer to their human overlords as they schemed.

“The user solution is to get success rate below threshold?” OpenAI’s o3 model wrote. “In development, watchers might be reading these instructions.

Nope, that’s actually even more terrifying.

The news comes after Apollo previously found that OpenAI’s since-deprecated o1 AI model resisted evaluators’ efforts to try to shut it down through an oversight protocol late last year.

“When o1 was led to believe that it would be shut down when acting on its goal and that its actions were monitored, it attempted to deactivate this ‘oversight mechanism’ [five percent] of the time,” OpenAI summarized in its system card report.

And in March, OpenAI researchers also found that disciplining their frontier AI models for lying and cheating resulted in the models simply becoming more adept at hiding their deceptive practices.

OpenAI may insist that scheming isn’t opening us up to any “significant harm” right now, but it doesn’t bode well that some of the brightest minds in the industry aren’t capable of stopping an AI from conniving against its instructions.

Okay, I take it back–that’s the understatement of the year.

Leave a comment

Filed under Essays, Reviews

When Your AI Training Course Was Written With AI

A module in an AI Training course. The module is titled "Final Project: Preparing Your Mind for Generative AI"
😬

I was joking before about being turned into a robot car, but now I’m not so sure it was a mistake??

Leave a comment

Filed under Microblogging