Blood in the Machine

Blood in the Machine

Share this post

Blood in the Machine
Blood in the Machine
An 'always on' OpenAI device is a massive backlash waiting to happen
Copy link
Facebook
Email
Notes
More

An 'always on' OpenAI device is a massive backlash waiting to happen

If the thing ever gets made, it will be like Google Glass all over again, but worse

Brian Merchant's avatar
Brian Merchant
May 29, 2025
∙ Paid
74

Share this post

Blood in the Machine
Blood in the Machine
An 'always on' OpenAI device is a massive backlash waiting to happen
Copy link
Facebook
Email
Notes
More
7
17
Share

Last week, the big tech news was that OpenAI had acquired Jony Ive’s design studio in yet another deal worth so many billions of dollars, to build some vaguely defined “AI devices.” While the announcement itself was void of meaningful information about the apparently forthcoming products, details of OpenAI’s plans and what the they’re supposed to look like soon started leaking out.

As those details leaked, all I could think about was how much people are going to hate these things. If an OpenAI hardware product ever does launch, there’s going to be plenty of, let’s say, rather justified hostility towards them. An OpenAI-branded device might make the public vitriol spewed at the Google Glass—which was banished by backlash for a decade—tame by comparison.

<BLOOD IN THE MACHINE is 100% reader-supported and made possible by my excellent paying subscribers. I’m able to keep the vast majority of my work free to read and open to all thanks to that support. If you can, for the cost of a beer a month, or, uh, a Nintendo Switch game a year, consider helping me keep this thing running. On top of that, you’ll get access to the Critical AI reports, a short installment of which you’ll find at the bottom of this article. Today, we’ll cover artists taking the fight to AI companies in Sweden, the self-driving trucking boom in Texas, and a study that reveals that AI makes for a bad, even dangerous, therapist. Thanks everyone, for all you do. Onwards./>

According to the Wall Street Journal, Altman told OpenAI staff that the incoming device won’t be a phone or glasses, but that it will be portable and always on, to gather data about its users’ environment, and, of course, that it will be a phenomenal success: Altman said it will add $1 trillion to OpenAI’s value. He also said that the products, which he referred to as “companions,” will ship by the end of next year, and that they will sell them “faster than any company has ever shipped 100 million of something new before.”

The supply chain analyst Ming-Chi Kuo, meanwhile, reported that according to his research, mass production is expected to begin in 2027, and that

3. The current prototype is slightly larger than the AI Pin, with a form factor as compact and elegant as an iPod Shuffle…
4. One of the intended use cases is wearing the device around the neck.
5. It will have cameras and microphones for environmental detection, with no display functionality.
6. It is expected to connect to smartphones and PCs, utilizing their computing and display capabilities.

The self-described AR/VR/AI Content Creator Ben Geskin made an AI-generated design mockup using the above details, which looked like this:

Obviously, this is not a rendering of the final product, or even a prototype of the product, it’s one guy on social media sharing an AI-generated image (which, humorously, he watermarked). But it’s effective enough at conveying what an always-listening, always data-collecting OpenAI device would be, so it’s been passed around social media a good deal since. Surprise, people hate it.

For good reason:

Now, this is still mostly speculation; some reports are saying it will be worn like a necklace, others that Ive doesn’t want it to be a “wearable,” others just that it will be a “third device” in addition to your phone and computer etc etc etc.

But given what we do know—that it will always be on to receive data—and will necessarily contain instruments to collect that data—microphone, camera, sensors— it is safe to assume that it will be a surveillance machine of considerable proportions. So that’s one strike against it right out the gate, and a valid reason for people to reject its very premise; ChatGPT collects user input to train its models already, now just imagine that everything you hear, see, or say becomes training data owned by OpenAI.

If you’re old enough to remember the debut of Google Glass in 2013, then you’re old enough to remember one of the most pronounced bouts of public backlash against a tech product in the 21st Century. Early users of Google Glass—the search giant’s augmented reality glasses—were dubbed Glassholes and endlessly mocked on social and regular media. In particular, people recoiled against the idea of normalizing a device that allowed them to be photographed or documented against their will. The device flopped among developers and early users, too, and that public outcry, and the sense of utter rejection of Glass it stamped into popular culture, basically banished the very concept of such a device for nearly a decade.

A lot has changed since then, of course. There’s a case to be made that courtesy of other “always on” devices like Alexa, Ring, and a decade-plus of total smartphone saturation, public tolerance for mass surveillance and tech company-orchestrated privacy violations has been chipped away. Then again, sometimes we just need a blunt reminder of how bad things have gotten to get angry all over again. Notably, Google had been involved in a number of privacy-violating scandals in the run-up to Glass’s launch, which the glasses then crystalized; the device, in other words, also became something of a beacon for the discomfort, distrust, and ambient anger people had at Google in general.

It’s very easy to see the same happening to OpenAI.

And remember, Google wasn’t particularly unpopular in 2013. People had privacy concerns, but Google still carried its don’t be evil aura, and Silicon Valley was widely seen as a beneficent engine of progress. It’s a different landscape today, and a very different story with OpenAI and AI products more generally.

Poll after poll finds that the public is more concerned than excited about AI. There’s a widespread distrust of Sam Altman and OpenAI. And, perhaps most importantly, there’s a sustained anger at the very foundations that the modern AI industry is built on. There’s anger among artists, writers, creatives, and other workers at the way AI systems were built on their work without consent or compensation, at the way AI output is being used to degrade their livelihoods, at the way AI companies have benefitted at their expense. There’s anger over the growing energy, water, and resource costs of powering the AI systems. There’s anger among the AI safety crowd at the way the major AI companies are developing consumer products without guardrails or accountability. There’s widespread anger over the specter potential job loss. There’s a lot of anger.

Throw a bunch of OpenAI-branded pendants—or orbs, or AI iPod shuffles, or whatever it ends up being—around early adopters’ necks, and my guess is you’ll see another serious round of public shaming, of what you might call AI-generated Glassholing. The stakes are higher; we’re not just talking about privacy, which is of course very important, but often not as viscerally felt as threats to people’s livelihoods. And there are a lot of people who would no doubt see OpenAI device users as gleefully wearing their contempt for others around their necks. This is a recipe for mass backlash, in other words.

Look, the thing could be total vaporware—we might hear a couple updates on io progress from Altman here and there before it fades away, the announcement alone having done its part to keep the AI narrative rolling. It could also launch and be laughably awkward and bad, with the most obnoxious people you can imagine repeating questions too loud in the restaurant to the hunk of plastic strapped to their necks, and thus go the way of the reviled Humane Pin—which Altman also invested in, recall—all on its own. I doubt Altman and Ive launch something that sucks so obviously overtly, they’d probably rather not launch at all. So if the thing comes to fruition, then we really might see users boasting OpenAI pendants in the open, embracing their a roles as a consumer avatars of technological inequality, as proud real-world AI influencer blue X checkmarks.

In that case, well, in a world where Google’s Waymos are coned and torched, where Tesla dealerships are firebombed for what they stand for—well, things might get interesting.

In this context, it’s also important to remember that backlashes are often not only well-deserved, but hold the potential for political and social progress, too. The successful pushback against Glass, as writers like Rose Eveleth have argued, helped establish boundaries for what rights we were willing to cede to tech companies. “Google Glass is a story about human beings setting boundaries and pushing back against surveillance,” Eveleth wrote for Wired, “a tale of how a giant company’s crappy product allowed us to envision a better future.” Looks like we’re going to need to tell a story like that again pretty soon.

The CRITICAL AI REPORT: Artists take a hard line against AI, the brewing self-driving truck crisis, and why AI should never replace your therapist

It’s not all grim news out there—artists and workers are banding together to fight for their livelihoods, and finding legally binding ways to push back on big AI. One example:

Keep reading with a 7-day free trial

Subscribe to Blood in the Machine to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Brian Merchant
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More