AI and Ambient Computing, Zuckerberg and The Verge, Amazon and Anthropic

Good morning,

On last Friday’s Sharp Tech we discussed last Thursday’s Article AI, Hardware, and Virtual Reality. More on this below.

On to the Update:

AI and Ambient Computing

In case you missed it, AI, Hardware, and Virtual Reality came out quite late on Thursday; I rewrote the Article a couple of times, and for that reason alone it’s probably not one of my best. I feel like AI is one of those topics where the edges of what is going to happen next are only faintly visible, and it’s sometimes hard to articulate that. To that end, I wanted to add another framing that I was considering.

In 2020’s The End of the Beginning I posited that we had reached a natural endpoint in terms of the personal computer and cloud revolutions:

There is an implication in the “generational change is inevitable” argument that paradigm shifts are sui generis. The personal computer was a discrete event, the Internet another, and mobile a third. Now we are simply waiting to see what is next — perhaps augmented reality, or voice assistants.

In fact, I would argue the opposite: the critical paradigm shifts in technology, which drove the generational changes that Evans wrote about, are part of a larger pattern.

Start with the mainframe: the primary interaction model was punched cards; to execute a program you had to insert your cards into a card reader and wait for the computer to read the program into memory, execute it, and give you the results. Computing was done in batches, because the I/O layer was directly linked to the application and data layer.

This explains why personal computers were so revolutionary: instead of one large shared computer for which you had to wait your turn, a user could access their own computer on their own desk whenever they wanted. Still, the personal computer, particularly in a corporate environment, lived alongside not just mainframes but increasingly servers on an intranet. The I/O layer and application and data layers were being pulled apart, but both were destinations: you had to go to your desk and be on the network to compute.

This last point gets at why the cloud and mobile, which are often thought of as two distinct paradigm shifts, are very much connected: the cloud meant applications and data could be accessed from anywhere; mobile made the I/O layer available anywhere. The combination of the two make computing continuous.

A drawing of The Evolution of Computing

What is notable is that the current environment appears to be the logical endpoint of all of these changes: from batch-processing to continuous computing, from a terminal in a different room to a phone in your pocket, from a tape drive to data centers all over the globe. In this view the personal computer/on-premises server era was simply a stepping stone between two ends of a clearly defined range.

I think that this framework is correct, but I described the endpoint incorrectly: even with smartphones computing is not yet continuous. Yes, a smartphone is always available, but it is still an intentional act to take your phone out of your pocket and interact with it. To put it another way, a PC was a destination, while smartphones were always available, but the next step is truly ambient computing.

The key to ambient computing, though — and this was the point I was trying to tease out with the opening about TMT and virtualizing space, time, and interactivity — is that there needs to be something to interact with continuously. No human can fill this role: yes, you can communicate with anyone anywhere, but not on-demand 24/7. Media can’t fill this role either: you can listen to or watch something at any time or any place, but it’s not interactive. Software is always available, but there is friction in terms of figuring out what app to use, actually launching it, interacting with it, etc. This is where generative AI is so compelling: it is the essential technological underpinning of continuous ambient computing.

That then leads to my second point: there is a necessity for new hardware to enable this sort of frictionless ambient computing. You can get an idea of what that ambient computing is like with chatbots — this is why the ChatGPT voice application was so compelling to me — but the app paradigm is not the right one. There is a hardware breakthrough waiting to happen just like the Internet created the conditions for the smartphone breakthrough to happen. In that context, Meta’s hardware investments are suddenly drastically more compelling than they might have seemed previously, and it makes sense that OpenAI is exploring the space.

The other point that I made more forcefully in the above Sharp Tech episode is that this framework — where seamless integration into your environment is the primary state of this ambient computing, which I am calling virtual reality — means that immersive environments like that provided by a headset are downstream from lower bandwidth and lower fidelity augmented experiences. That’s why I suddenly find Meta’s Smart Glasses much more compelling in the short-term than full-on headsets.

Zuckerberg and The Verge

It’s pretty clear that Meta CEO Mark Zuckerberg agrees. I highly recommend his interview with Alex Heath of The Verge that touches on several of these points. First, with regard to Meta’s Smart Glasses:

When I was thinking about what would be the key features for smart glasses, I kind of thought that we were going to get holograms in the world, and that was one. That’s kind of like augmented reality. But then there was always some vague notion that you’d have an assistant that could do something.

I thought that things like Siri or Alexa were very limited. So I was just like, “Okay, well, over the time period of building AR glasses, hopefully the AI will advance.” And now it definitely has. So now I think we’re at this point where it may actually be the case that for smart glasses, the AI is compelling before the holograms and the displays are, which is where we got to with the new version of the Ray-Bans that we’re shipping this year, right? When we started working on the product, all this generative AI stuff hadn’t happened yet…

Again, this is all really novel stuff. So I’m not pretending to know exactly what the key use cases are or how people are going to use that. But smart glasses are very powerful for AI because, unlike having it on your phone, glasses, as a form factor, can see what you see and hear what you hear from your perspective.

So if you want to build an AI assistant that really has access to all of the inputs that you have as a person, glasses are probably the way that you want to build that. It’s this whole new angle on smart glasses that I thought might materialize over a five- to 10-year period but, in this odd twist of the tech industry, I think actually is going to show up maybe before even super high-quality holograms do.

Second, consider how VR fits in with the idea of personalized assistants, including the celebrity-based characters Meta launched last week:

I think that people will probably want the AIs that they interact with, I think it’ll be more exciting and interesting if they do, too. So part of what I’m interested in is this isn’t just chat, right? Chat will be where most of the interaction happens. But these AIs are going to have profiles on Instagram and Facebook, and they’ll be able to post content, and they’ll be able to interact with people and interact with each other, right?

There’s this whole interesting set of flywheels around how that interaction can happen and how they can sort of evolve over time. I think that’s going to be very compelling and interesting, and obviously, we’re kind of starting slowly on that. So we wanted to build it so that it kind of worked across the whole Meta universe of products, including having them be able to, in the near future, be embodied as avatars in the metaverse, right?

So you go into VR and you have an avatar version of the AI, and you can talk to them there. I think that’s gonna be really compelling, right? It’s, at a minimum, creating much better NPCs and experiences when there isn’t another actual person who you want to play a game with. You can just have AIs that are much more realistic and compelling to interact with.

But I think having this crossover where you have an assistant or you have someone who tells you jokes and cracks you up and entertains you, and then they can show up in some of your metaverse worlds and be able to be there as an avatar, but you can still interact with them in the same way — I think it’s pretty cool.

Note the progression here: this isn’t about building out a Metaverse and then giving a poor facsimile on your phone; rather, you connect with and develop a relationship with an AI through the low-fidelity means available today, and those capabilities increase over time and culminate in an immersive experience, sort of like how you might travel to see a friend that you mostly talk to via chat or phone calls. It’s a subtle distinction, but I think an important one.

I must emphasize, these are very early days; for the record I found the celebrity AI’s incredibly banal. That’s why things are fuzzy, though: the future is starting to take shape, but details are not yet in focus.

Amazon and Anthropic

From the Wall Street Journal:

Amazon.com said it has agreed to invest up to $4 billion in artificial-intelligence company Anthropic, the latest big startup investment by tech giants jockeying for an edge in the AI arms race. Amazon said that, as part of the deal, Anthropic would be using its custom chips to build and deploy its AI software. Amazon also agreed to incorporate Anthropic’s technology into products across its business.

People familiar with the deal said Amazon has committed to an initial $1.25 billion investment in two-year-old Anthropic, a number that could grow to $4 billion over time depending on certain conditions. As part of the agreement, Anthropic has agreed to spend a certain amount of the capital on Amazon’s cloud infrastructure business, Amazon Web Services, one of the people said. The specifics of that arrangement couldn’t be learned.

Dylan Patel has more details at SemiAnalysis:

The match of Anthropic and Amazon is sort of perfect on the surface. Amazon needs frontier model capabilities, and this is how the empire strikes back. Anthropic, with Claude 2, has the 2nd best publicly accessible model after OpenAI’s GPT-4. Amazon gets direct access to Claude 2 for serving customers and can also offer fine tuning services. Future models will also be available to Amazon. Amazon also stated they will be leveraging these models for drug discovery and many other aspects of healthcare, further strengthening their Amazon HealthOmics platform.

On Anthropic’s side of the fence, while on the surface, it seems they get to stick to their core beliefs of AI safety without signing control away to Amazon, in reality, the deal represents Anthropic effectively betting it all. We hear there are pretty meaty IP transfers with regards to giving away certain current and upcoming models to Amazon. Allegedly Amazon can build pretty much anything they’d like to leveraging Anthropic’s technology…

As Patel astutely notes, although this deal appears to be broadly similar to Microsoft’s arrangement with OpenAI, Anthropic’s relative lack of mindshare, both in terms of consumers and developers, means that Amazon is much more likely to be the API of choice for these models, whereas OpenAI’s API is used more than Azure’s, particularly for startups.

At the same time, I agree with Patel that this was a deal both companies needed to make: Anthropic needs access to compute, and Amazon needs a foundational model to anchor its Bedrock managed AI service and compete with Azure/OpenAI and Google (what is not clear is what happened to Anthropic’s relationship with Google, who invested in the AI startup just six months ago). Anthropic also gives a core model around which to focus AWS’s chip design efforts, which have urgency not just for long-term cost reasons, but also because AWS is severely limited in just how many Nvidia GPUs it can buy.

One other note on Bedrock, which officially launched last week, via the press release:

Amazon Web Services, Inc. (AWS), an Amazon.com, Inc. company, today announced five generative artificial intelligence (AI) innovations, so organizations of all sizes can build new generative AI applications, enhance employee productivity, and transform their businesses. Today’s announcement includes the general availability of Amazon Bedrock, a fully managed service that makes foundation models (FMs) from leading AI companies available through a single application programming interface (API). To give customers an even greater choice of FMs, AWS also announced that Amazon Titan Embeddings model is generally available and that Llama 2 will be available as a new model on Amazon Bedrock – making it the first fully managed service to offer Meta’s Llama 2 via an API.

I’ve been wondering for a couple of months when/if Amazon would accept it needed to cut a deal with Meta to provide Llama as a managed service: it looks like it didn’t take long.

This suggests two things: first, it reaffirms that Meta’s open source approach with Llama is absolutely driving meaningful adoption such that AWS had to be concerned about being left behind, and second, it suggests that AWS is indeed worried about being left behind. Indeed, perhaps this is the other bit my illustration above got wrong: mobile drove a new kind of cloud computing, and it doesn’t seem unreasonable to think that AI will not only require new hardware but also create the opportunity for new clouds as well. Incumbents beware!


This Update will be available as a podcast later today. To receive it in your podcast player, visit Stratechery.

The Stratechery Update is intended for a single recipient, but occasional forwarding is totally fine! If you would like to order multiple subscriptions for your team with a group discount (minimum 5), please contact me directly.

Thanks for being a subscriber, and have a great day!