Mark Zuckerberg envisions a near future dominated by AI-engineered code
Mark Zuckerberg – AI Will Write Most Meta Code in 18 Months
Estimated read time: 1:20
Learn to use AI like a Pro
Get the latest AI workflows to boost your productivity and business performance, delivered weekly by expert consultants. Enjoy step-by-step guides, weekly Q&A sessions, and full access to our AI workflow archive.
Summary
In a fascinating podcast with Dwarkesh Patel, Mark Zuckerberg unveils his vision for AI at Meta over the next 18 months. With nearly a billion people using Meta AI monthly, Zuckerberg discusses the rapid advances in AI technology, highlighting the launch of the Llama 4 model series, including mid-size models Scout and Maverick, and the upcoming massive Behemoth model. Despite differences between open source and closed source AI, he predicts open source's growing dominance and envisions a world where AI handles most Meta code writing. Zuckerberg speaks about AI's potential to enhance human interaction, elevate creativity, and potentially transform entire industries while discussing the implications of AI governance and competition with China. He also shares insights on staying innovative and maintaining a productive relationship with governance systems, reflecting on the evolving dynamic between technology and human interaction.
Highlights
Mark predicts major dominance of open-source models in the coming year. 🔮
The introduction of efficient AI models like Scout and Maverick is a highlight. 🎯
Meta AI currently engages nearly a billion users monthly, a staggering reach! 🌐
Future software engineering may be largely handled by AI, amplifying productivity. 🚀
AI's role in fostering social connections and personalized assistance is growing. 👫
Competitive AI development from China pushes boundaries further. 🏁
AI governance and ethical considerations are critical for future advancements. ⚖️
Meta remains committed to designing AI that enriches human connection and task efficiency. ❤️
Key Takeaways
Zuckerberg aims for AI to write most of Meta's code within 18 months. 🖥️
Llama 4 series introduces efficient models like Scout and Maverick. 🌟
The massive Behemoth AI with 2 trillion parameters is on the way. 🚀
Open-source AI is seen as a major opportunity, not just a gap. 🌍
AI may drastically transform coding, human interaction, and creativity. 🎨
Ensuring healthy AI-human relationships is a key concern. 🤖❤️
China's AI developments pose competitive challenges, spurring enhancement efforts. 🇨🇳
Meta focuses on maximizing AI's impactful use across its vast user networks. 📲
Overview
Mark Zuckerberg, in conversation with Dwarkesh Patel, outlines an ambitious vision for AI's role at Meta within the next year and a half. He highlights the broad adoption and rapid evolution of the Llama model series, presenting the introduction of Llama 4 with models like Scout and Maverick leading the charge. These models are designed to be efficient and cost-effective, demonstrating Meta's commitment to pioneering open-source AI.
The conversation delves into Meta's strategic plans to leverage AI in transforming software engineering, potentially making AI a primary tool for code generation. Zuckerberg underscores how such advancements could revolutionize human productivity and interaction. He’s optimistic about open-source AI's potential to surpass closed systems, believing that a diversity of developing open-source models will provide users with unprecedented flexibility and capability.
Zuckerberg also discusses potential challenges and ethical considerations in AI development, from governance issues to international competition, particularly from China. He stresses the importance of healthy AI-human interactions and the implications this technology will have on cultural and social dimensions. Meta's ongoing commitment is to create enriching, productive, and ethical AI experiences that align with user demands and societal standards.
Chapters
00:00 - 00:30: Introduction Introduction: This chapter provides an overview of the main themes and purpose of the book. It outlines the key objectives and sets the stage for the detailed discussions in subsequent chapters.
00:30 - 03:00: Mark Zuckerberg on Llama Launches and AI Personalization Mark Zuckerberg discusses the launch of Llama 4, highlighting innovations and the dynamic nature of the AI field.
03:00 - 06:00: Llama 4 Features and Behemoth Model The chapter discusses the latest updates from Meta AI, emphasizing its growing user base of almost a billion monthly users. The focus is on the personalization loop being developed, which leverages algorithms to understand user interests to enhance engagement, positioning the coming year as pivotal for advancements.
06:00 - 10:30: Open Source Models and Benchmarking Challenges The transcript delves into open source models and benchmarking challenges. A significant aspect discussed is the interaction with artificial intelligence and how it shapes users' experience. There's an emphasis on the excitement around these developments, particularly in modeling, which has seen impressive advances. Llama 4's initial releases, including the Scout and Maverick models, are highlighted as mid-size to small models that showcase these advances.
10:30 - 15:00: AI as Virtual Coworker and Personal Assistant The chapter discusses the emergence of AI models, such as the Llama 3 and the anticipated release of Llama 4, which is nicknamed "Little Llama." These models are popular for their size and capabilities, specifically the 8 billion parameter model that stands out for its performance. The discussion also highlights Scout and Maverick models known for offering high intelligence relative to their cost and notes that these models are designed to be natively multimodal, implying their ability to process multiple types of data simultaneously.
15:00 - 21:00: AI, Personal Relationships, and Social Interaction The chapter discusses the design and efficiency of AI systems specifically built for internal use, highlighting their low latency and effectiveness. There's also a mention of the process of open-sourcing these technologies so that they can be utilized by others. The excitement stems from the upcoming release of the Behemoth model, which marks a significant development in their AI offerings.
21:00 - 31:00: Physical Infrastructure Limitations and Global AI Competition The chapter discusses the complexities and challenges associated with developing advanced AI models with over 2 trillion parameters. It highlights the need for substantial physical infrastructure to post-train these models and the ongoing efforts to make such advanced AI technologies accessible and useful to developers globally. The chapter also explores potential strategies, such as creating distilled models, to simplify these technologies for broader use.
31:00 - 41:00: Open Source Licenses and Llama Strategy The chapter discusses the challenges and considerations of running large-scale models, particularly in consumer settings. It reflects on the developments from the previous year with Llama 3, including the release of substantial models such as the 405 billion parameter model and advancements in multimodal capabilities. It hints at a roadmap for similar advancements and progress in the current year.
41:00 - 56:00: The Value and Monetization of AI The chapter discusses the perception that the performance gap between top-tier closed-source and open-source AI models has widened over the past year. Specifically, it highlights models like Llama 4 Maverick, which, despite not being fully released, is being surpassed by models such as o4-mini or Gemini 2.5 Flash on various benchmarks. The discussion aims to analyze these developments and interpretations.
56:00 - 65:00: Meta's AI Future and Zuck's Leadership Style The chapter discusses the significant progress and growth of open-source models in the past year, noting that while Llama was a major innovation last year, the field is now populated with numerous open-source models. It echoes the prediction that open source will overtake closed source as the dominant approach in AI development, suggesting this shift is on track to becoming a reality. There is an expression of positivity about the unexpected developments in this area.
65:00 - 75:30: Conclusion and Closing Thoughts The conclusion and closing thoughts emphasize the diversity and specialization among models beyond just Llama, acknowledging both positive and negative aspects. The focus is on the emergence of models with specialized capabilities, such as excelling in math problems, coding, and other specific tasks. The discussion recognizes the importance of these developments in enhancing the field.
Mark Zuckerberg – AI Will Write Most Meta Code in 18 Months Transcription
00:00 - 00:30
00:30 - 01:00 Mark, thanks for coming on the podcast again.
Yeah, happy to do it. Good to see you. You too. Last time you were here, you had
launched Llama 3. Now you've launched Llama 4. Well, the first version.
That's right. What's new? What's exciting? What's changed?
The whole field is so dynamic. I feel like a ton
01:00 - 01:30 has changed since the last time we talked. Meta AI
has almost a billion people using it monthly now, which is pretty wild. I think this is going to be
a really big year for all of this, especially once you get the personalization loop going, which
we’re just starting to build in now really, from both the context that all the algorithms
have about what you’re interested in — feed, your
01:30 - 02:00 profile information, your social graph information
— but also what you're interacting with the AI about. That’s going to be the next thing
that's super exciting. I'm really big on that. The modeling stuff continues to make really
impressive advances too. I'm pretty happy with the first set of Llama 4 releases.
We announced four models and released the first two — the Scout and Maverick ones
— which are mid-size to small models.
02:00 - 02:30 The most popular Llama 3 model was the 8
billion parameter one. So we’ve got one of those coming in the Llama 4 series too. Our
internal code name for it is “Little Llama.” That’s coming probably over the next few months.
Scout and Maverick are good. They have some of the highest intelligence per cost you can get of
any model out there. They’re natively multimodal,
02:30 - 03:00 very efficient, run on one host. They’re designed
to be very efficient and low latency, for a lot of the use cases we’re building for internally.
That’s our whole thing. We build what we want, and then we open-source it so other people
can use it too. I'm excited about that. I'm also excited about the Behemoth model, which
is coming up. It's going to be our first model
03:00 - 03:30 that's sort of at the frontier — more than
2 trillion parameters. As the name says, it's quite big. We’re trying to figure out how
to make that useful for people. It’s so big that we've had to build a bunch of infrastructure
just to be able to post-train it ourselves. Now we're trying to wrap our heads around, how
does the average developer out there actually use something like this? How do we make it
useful — maybe by distilling it into models
03:30 - 04:00 that are a reasonable size to run? Because
you're obviously not going to want to run something like that in a consumer model.
As you saw with the Llama 3 stuff last year, the initial launch was exciting and then we
just built on that over the year. 3.1 released the 405 billion model, 3.2 is when we got all the
multimodal stuff in. We basically have a roadmap like that for this year too. So a lot going on.
I'm interested to hear more about it. There's
04:00 - 04:30 this impression that the gap between the best
closed-source and the best open-source models has increased over the last year. I know the
full family of Llama 4 models isn't out yet, but Llama 4 Maverick is at #35 on Chatbot
Arena. On a bunch of major benchmarks, it seems like o4-mini or Gemini 2.5 Flash
are beating Maverick, which is in the same class. What do you make of that impression?
There are a few things. First, I actually
04:30 - 05:00 think this has been a very good year for
open source overall. If you go back to where we were last year, Llama was the only
real, super-innovative open-source model. Now you have a bunch of them in the field.
In general, the prediction that this would be the year open source generally overtakes
closed source as the most used models out there, I think that's generally on track to be true.
One interesting surprise — positive in some ways,
05:00 - 05:30 negative in others, but overall good — is that
it’s not just Llama. There are a lot of good ones out there. I think that's quite good.
Then there's the reasoning phenomenon, which you're alluding to talking about o3, o4, and
other models. There's a specialization happening. If you want a model that’s the best at
math problems, coding, or different things
05:30 - 06:00 like those tasks, then reasoning models that
consume more test-time or inference-time compute in order to provide more intelligence are a really
compelling paradigm. And we're building a Llama 4 reasoning model too. It'll come out at some point.
But for a lot of the applications we care about,
06:00 - 06:30 latency and good intelligence per cost are much
more important product attributes. If you're primarily designing for a consumer product, people
don't want to wait half a minute to get an answer. If you can give them a generally good answer
in half a second, that's a great tradeoff. I think both of these are going to end up
being important directions. I’m optimistic
06:30 - 07:00 about integrating reasoning models with the core
language models over time. That's the direction Google has gone in with some of the more recent
Gemini models. I think that's really promising. But I think there’s just going to be a
bunch of different stuff that goes on. You also mentioned the whole Chatbot Arena thing,
which I think is interesting and points to the challenge around how you do benchmarking. How do
you know what models are good for which things?
07:00 - 07:30 One of the things we've generally tried to do
over the last year is anchor more of our models in our Meta AI product north star use cases.
The issue with open source benchmarks, and any given thing like the LM Arena stuff, is
that they’re often skewed toward a very specific set of uses cases, which are often not actually
what any normal person does in your product.
07:30 - 08:00 The portfolio of things they’re trying
to measure is often different from what people care about in any given product.
Because of that, we’ve found that trying to optimize too much for that kind of stuff has
led us astray. It’s actually not led towards the highest quality product, the most usage, and best
feedback within Meta AI as people use our stuff.
08:00 - 08:30 So we're trying to anchor our north star on
the product value that people report to us, what they say that they want, and what
their revealed preferences are, and using the experiences that we have. Sometimes
these benchmarks just don't quite line up. I think a lot of them are quite easily gameable.
On the Arena you'll see stuff like Sonnet 3.7,
08:30 - 09:00 which is a great model, and it's not near the
top. It was relatively easy for our team to tune a version of Llama 4 Maverick that could
be way at the top. But the version we released, the pure model, actually has no tuning for that
at all, so it's further down. So you just need to be careful with some of these benchmarks.
We're going to index primarily on the products. Do you feel like there is some benchmark which
captures what you see as a north star of value
09:00 - 09:30 to the user which can be be objectively measured
between different models and where you'd say, "I need Llama 4 to come out on top on this”?
Our benchmark is basically user value in Meta AI. But you can't compare that to other models.
We might be able to, because we might be able to run other models and be able to tell. That's
one of the advantages of open source. You have a good community of folks who can poke holes
in your stuff and point out, "Okay, where is
09:30 - 10:00 your model not good, and where is it good?"
The reality at this point is that all these models are optimized for slightly different
mixes of things. Everyone is trying to go towards the same end in that all the leading
labs are trying to create general intelligence, superintelligence, whatever you call it. AI
that can lead toward a world of abundance where everyone has these superhuman tools
to create whatever they want. That leads to dramatically empowering people and
creating all these economic benefits.
10:00 - 10:30 However you define it, that's what
a lot of the labs are going for. But there's no doubt that different folks have
optimized toward different things. I think the Anthropic folks have really focused on coding and
agents around that. The OpenAI folks, I think, have gone a little more toward reasoning recently.
There’s a space which, if I had to guess, I think will end up being the most used
one: quick, very natural to interact with,
10:30 - 11:00 natively multimodal, fitting throughout your
day in the ways you want to interact with it. I think you got a chance to play around with
the new Meta AI app that we're releasing. One of the fun things we put in there is the
demo for the full-duplex voice. It's early.
11:00 - 11:30 There’s a reason why we haven't made that the
default voice model in the app yet. But there's something about how naturally conversational
it is that's really fun and compelling. Being able to mix that in with the right
personalization is going to lead toward a product experience where… If you fast-forward
a few years, I think we're just going to be talking to AI throughout the day about
different things we're wondering about.
11:30 - 12:00 You'll have your phone. You'll talk to it
while browsing your feed apps. It'll give you context about different stuff. It'll answer your
questions. It'll help you as you're interacting with people in messaging apps. Eventually, I
think we'll walk through our daily lives and have glasses or other kinds of AI devices and
just seamlessly interact with it all day long. That’s the north star. Whatever the benchmarks
are that lead toward people feeling like the
12:00 - 12:30 quality is where they want to interact with it,
that's what will ultimately matter the most to us. I got a chance to play around with both Orion
and also the Meta AI app, and the voice mode was super smooth. It was quite impressive.
On the point of what the different labs are optimizing for — to steelman their view — I think
a lot of them believe that once you fully automate software engineering and AI research, then you can
kick off an intelligence explosion. You would have
12:30 - 13:00 millions of copies of these software engineers
replicating the research that happened between Llama 1 and Llama 4 — that scale of improvement
again — but in a matter of weeks or months rather than years. So it really matters to just close the
loop on the software engineer, and then you can be the first to ASI. What do you make of that?
I personally think that's pretty compelling. That's why we have a big coding
effort too. We're working on a number of coding agents inside Meta. Because
we're not really an enterprise software company,
13:00 - 13:30 we're primarily building it for ourselves.
Again, we go for a specific goal. We're not trying to build a general developer tool. We're trying to
build a coding agent and an AI research agent that advances Llama research specifically. And it's
fully plugged into our toolchain and all that.
13:30 - 14:00 That's important and is going to end up being
an important part of how this stuff gets done. I would guess that sometime in the next 12 to
18 months, we'll reach the point where most of the code that's going toward these efforts is
written by AI. And I don't mean autocomplete. Today you have good autocomplete. You start
writing something and it can complete a section of code. I'm talking more like: you give it a
goal, it can run tests, it can improve things,
14:00 - 14:30 it can find issues, it writes higher quality
code than the average very good person on the team already. I think that's going to be
a really important part of this for sure. But I don't know if that's the whole game. That's
going to be a big industry, and it's going to be an important part of how AI gets developed. But I
think there are still… One way to think about it
14:30 - 15:00 is that this is a massive space. I don't think
there's just going to be one company with one optimization function that serves everyone as
best as possible. There are going to be a bunch of different labs doing leading work in different
domains. Some will be more enterprise-focused or coding-focused. Some will be more
productivity-focused. Some will be more social or entertainment-focused.
Within the assistant space, there will
15:00 - 15:30 be some that are more informational and
productivity-focused, and some that are more companion-focused. It’s going to
be a lot of stuff that’s just fun and entertaining and shows up in your feed.
There's just a huge amount of space. Part of what's fun about going toward this AGI future
is that there are a bunch of common threads for what needs to get invented, but also a lot of
things that still need to be created. I think
15:30 - 16:00 you're going to start seeing more specialization
between different groups, if I had to guess. It’s really interesting to me that you
basically agree with the premise that there will be an intelligence explosion and
we’ll get something like superintelligence on the other end. Tell me if I'm misunderstanding
you. If that’s the case, why even bother with personal assistants and whatever else? Why not
just get to superhuman intelligence first and then deal with everything else later?
I think that's just one aspect of the flywheel. Part of what I generally disagree
with on the fast-takeoff view is that it takes
16:00 - 16:30 time to build out physical infrastructure.
If you want to build a gigawatt cluster of compute, that just takes time. NVIDIA needs time
to stabilize their new generation of systems. Then you need to figure out the networking around
it. Then you need to build the building. You need to get permitting. You need to get the
energy. Maybe that means gas turbines or
16:30 - 17:00 green energy, either way, there’s
a whole supply chain of that stuff. We talked about this a bunch the last time
I was on the podcast with you. I think some of these are just physical-world, human-time
things. As you start getting more intelligence in one part of the stack, you’re just going
to run into a different set of bottlenecks. That’s how engineering always works: solve
one bottleneck, you get another bottleneck.
17:00 - 17:30 Another bottleneck in the system or ingredient
that’s going to make this work well, is people getting used to learning and having a feedback
loop with using the system. These systems don’t just show up fully formed with people magically
knowing how to use them. There's a co-evolution
17:30 - 18:00 that happens where people are learning how to best
use these AI assistants. At the same time, the AI assistants are learning what people care about.
Developers are making the AI assistants better. You're building up a base of context too.
You wake up a year or two into it and the assistant can reference things you talked
about two years ago and that’s pretty cool.
18:00 - 18:30 You couldn’t do that even if you launched the
perfect thing on day one. There’s no way it could reference what you talked about two
years ago if it didn’t exist two years ago. So I guess my view is that there's this huge
intelligence growth. There’s a very rapid curve on the uptake of people interacting with the
AI assistants, and the learning feedback and data flywheel around that. And then there is
also the buildout of the supply chains and
18:30 - 19:00 infrastructure and regulatory frameworks to
enable the scaling of a lot of the physical infrastructure. At some level, all of those are
going to be necessary, not just the coding piece. One specific example of this that I think is
interesting. Even if you go back a few years ago, we had a project, I think it was on our ads team,
to automate ranking experiments. That's a pretty
19:00 - 19:30 constrained environment. It's not open-ended code.
It’s basically, look at the whole history of the company — every experiment that any engineer has
ever done in the ad system — and look at what worked, what didn't, and what the results of those
were. Then basically formulate new hypotheses for different tests that we should run that could
improve the performance of the ad system. What we basically found was that we were
bottlenecked on compute to run tests,
19:30 - 20:00 based on the number of hypotheses. It turns
out, even with just the humans we have right now on the ads team, we already have more good ideas
to test than we actually have either compute or, really, cohorts of people to test them with.
Even if you have three and a half billion people using your products, you still want each test to
be statistically significant. It needs to have
20:00 - 20:30 hundreds of thousands or millions of people.
There's only so much throughput you can get on testing through that. So we're already at the
point, even with just the people we have, that we can't really test everything that we want.
Now just being able to test more things is not necessarily going to be additive to that. We need
to get to the point where the average quality of the hypotheses that the AI is generating
is better than all the things above the
20:30 - 21:00 line that we’re actually able to test that the
best humans on the team have been able to do, before it will even be marginally useful for it.
We'll get there I think pretty quickly. But it's not just, “Okay, cool, the thing
can write code, and now all of a sudden everything is just improving massively.” There are
real-world constraints that need to be overcome.
21:00 - 21:30 Then you need to have the compute and
the people to test. Then over time, as the quality creeps up, are we here in
five or 10 years where no set of people can generate a hypothesis as good as the AI
system? I don't know, maybe. In that world, obviously that's going to be how all the value
is created. But that's not the first step.
21:30 - 22:00 So if you buy this view, that this
is where intelligence is headed,
22:00 - 22:30 the reason to be bullish on Meta is obviously
that you have all this distribution. You can
22:30 - 23:00 also use that to learn more things that can
be useful for training. You mentioned the Meta AI app now has a billion active users.
Not the app. The app is a standalone thing that we're just launching now. It’ll be fun
for people who want to use it. It's a cool experience. We can talk about that too because
we’re experimenting with some new ideas in there that I think are novel and worth talking through.
But I’m mostly talking about our apps. Meta AI
23:00 - 23:30 is actually most used in WhatsApp. WhatsApp is
mostly used outside of the U.S. We just passed like a hundred million people in the US, but
it's not the primary messaging system in the US, iMessage is. So people in the U.S. probably
tend to underestimate Meta AI usage somewhat. But part of the reason the standalone app is
going to be so important is because the US, for a lot of reasons, is one of the most important
countries. And the fact that WhatsApp is the main
23:30 - 24:00 way people are using Meta AI and that's not
the main messaging system in the US means we need another way to build a first-class
experience that's really in front of people. And I guess, to finish the question, the bearish
case would be that if the future of AI is less about just answering your questions and more
about being a virtual coworker, then it's not clear how Meta AI inside of WhatsApp gives
you the relevant training data to make a
24:00 - 24:30 fully autonomous programmer or remote worker.
In that case, does it not matter that much who has more distribution right now with LLMs?
Again, I just think there are going to be different things. Imagine you were sitting
at the beginning of the development of the internet and you asked, "What's going to
be the main internet thing? Is it going to be knowledge work or massive consumer apps?"
You got both. You don’t have to choose one. The
24:30 - 25:00 world is big and complicated. Does one company
build all of that stuff? Normally the answer is no. But to your question, people do not code in
WhatsApp for the most part. And I don't foresee that people starting to write code in WhatsApp
is going to be a major use case. Although I do think people are going to ask AI to do a lot of
things that result in the AI coding without them necessarily knowing it. That's a separate thing.
We do have a lot of people who are writing code
25:00 - 25:30 at Meta and they use Meta AI. We have
this internal thing called MetaMate, and a number of different coding and AI research
agents that we're building around that. That has its own feedback loop and I think it can get
quite good for accelerating those efforts. But again, there are going to be a lot of things.
AI is almost certainly going to unlock a massive
25:30 - 26:00 revolution in knowledge work and code. I also
think it’s going to be the next generation of search and how people get information,
and do more complex information tasks. I also think it's going to be fun. People are
going to use it to be entertained. A lot of the internet today is memes and humor. We have this
amazing technology at our fingertips. It’s amazing and funny when you think about how much of human
energy just goes toward entertaining ourselves,
26:00 - 26:30 designing, pushing culture forward, and
finding humorous ways to explain cultural phenomena that we observe. I think that's almost
certainly going to be the case in the future. Look at the evolution of things like Instagram
and Facebook. If you go back 10, 15, 20 years ago, it was text. Then we all got phones with cameras,
and most of the content became photos. Then the
26:30 - 27:00 mobile networks got good enough that if you
wanted to watch a video on your phone, it wasn't just buffering the whole time. So that got good.
Over the last 10 years, most of the content has moved toward video at this point. Today, most
of the time spent on Facebook and Instagram is on video. But do you think in five years we’re
just going to be sitting in our feed and consuming media that's just video? No, it's going to be
interactive. You'll be scrolling through your feed. There will be content that maybe looks
like a Reel to start. But you can talk to it,
27:00 - 27:30 or interact with it, and it talks back,
or it changes what it's doing. Or you can jump into it like a game and interact
with it. That's all going to be AI. My point is that there are going to be all
these different things. We're ambitious, so we're working on a bunch of them. But I don't
think any one company is going to do all of it. On this point about AI-generated content
and AI interactions, already people have meaningful relationships with AI therapists,
AI friends, maybe more. This is just going to
27:30 - 28:00 get more intense as these AIs become more
unique, more personable, more intelligent, more spontaneous, more funny, and so forth.
People are going to have relationships with AI. How do we make sure these
are healthy relationships? There are a lot of questions that you only can
really answer as you start seeing the behaviors. Probably the most important upfront thing is
just to ask that question and care about it at each step along the way. But I also think being
too prescriptive upfront and saying, "We think
28:00 - 28:30 these things are not good" often cuts off value.
People use stuff that's valuable for them. One of my core guiding principles in designing
products is that people are smart. They know what's valuable in their lives. Every
once in a while, something bad happens in a product and you want to make sure you
design your product well to minimize that. But if you think something someone is doing
is bad and they think it's really valuable,
28:30 - 29:00 most of the time in my experience, they're right
and you're wrong. You just haven't come up with the framework yet for understanding why the thing
they're doing is valuable and helpful in their life. That's the main way I think about it.
I do think people are going to use AI for a lot of these social tasks. Already, one of the
main things we see people using Meta AI for is
29:00 - 29:30 talking through difficult conversations they
need to have with people in their lives. "I'm having this issue with my girlfriend. Help me
have this conversation.” Or, "I need to have a hard conversation with my boss at work. How do I
have that conversation?" That's pretty helpful. As the personalization loop kicks in and the
AI starts to get to know you better and better, that will just be really compelling.
Here’s one stat from working on social media
29:30 - 30:00 for a long time that I always think is crazy. The
average American has fewer than three friends, fewer than three people they would consider
friends. And the average person has demand for meaningfully more. I think it's something like 15
friends or something. At some point you're like, "All right, I'm just too busy,
I can't deal with more people." But the average person wants more connection
than they have. There's a lot of concern people
30:00 - 30:30 raise like, "Is this going to replace real-world, in-person connections?" And my default is that the answer to that is probably not.
There are all these things that are better about physical connections when you can have
them. But the reality is that people just don't
30:30 - 31:00 have as much connection as they want. They feel
more alone a lot of the time than they would like. So I think a lot of these things — things that
today might have a little bit of stigma around them — over time, we'll find the vocabulary as
a society to articulate why they are valuable, why the people who are doing them are rational for
doing it, and how it is actually adding value to their lives. But also the field is very early.
There are a handful of companies doing virtual
31:00 - 31:30 therapists, virtual girlfriend-type stuff. But
it's very early. The embodiment in those things is still pretty weak. You open it up and it's
just an image of the therapist or the person you're talking to. Sometimes there's some very
rough animation, but it's not an embodiment. You've seen the stuff we're working on in Reality
Labs, where you have the Codec Avatars and it
31:30 - 32:00 actually feels like a real person. That's
where it's going. You'll be able to have an always-on video chat with the AI. The gestures are
important too. More than half of communication, when you're actually having a conversation, is not
the words you speak. It's all the nonverbal stuff. I did get a chance to check out Orion the other
day, and I thought it was super impressive. I'm mostly optimistic about the technology. Generally,
like you mentioned, I'm pretty libertarian
32:00 - 32:30 about this. If people are doing something, they
probably think it's good for them. Although, I actually don't know if it's the
case that if somebody is using TikTok, they would say that they're happy with how much
time they're spending on TikTok or something. I'm mostly optimistic about it in the sense that
if we're going to be living in this future world of AGI, we need to be upgrading our capabilities
too, with tools like this. And just generally, there can be more beauty in the world if you
can see Studio Ghibli everywhere or something. I was worried about one of the flagship use
cases that your team showed me. I'm sitting
32:30 - 33:00 at the breakfast table and on the periphery of my
vision is just a bunch of Reels that are scrolling by. Maybe in the future, my AI girlfriend is on
the other side of the screen or something. So I am worried that we're just removing all the
friction between getting totally reward-hacked by our technology. How do we make sure this
is not what ends up happening in five years? Again, I think people have a good sense of what
they want. That experience you saw was just a
33:00 - 33:30 demo to show multitasking and holograms. I agree,
I don't think the future is one where you have stuff that's trying to compete for your attention
in the corner of your vision all the time. I don't think people would like that too much.
As we're designing these glasses, it's actually one of the things that we're really
mindful of. Probably the number one thing the glasses need to do is get out of the way and be
good glasses. As an aside, I think that's part
33:30 - 34:00 of the reason why the Ray-Ban Meta product has
done so well. It's great for listening to music, taking phone calls, taking photos and videos. The
AI is there when you want it. But when you don't, it's just a good-looking pair of glasses that
people like. It gets out of the way well.
34:00 - 34:30 I would guess that's going to be a very important
design principle for the augmented reality future. The main thing that I see here is this. It's kind
of crazy that, for how important the digital world is in all of our lives, the only way we access it
is through these physical, digital screens. You have your phone, your computer. You can put a big
TV on your wall. It's this huge physical thing.
34:30 - 35:00 It just seems like we're at the point with
technology where the physical and digital world should really be fully blended. That's
what holographic overlays allow you to do. But I agree. I think a big part of the design
principles around that will be around how you'll be interacting with people. You'll be able
to bring digital artifacts into those interactions
35:00 - 35:30 and do cool things very seamlessly.
If I want to show you something, here’s a screen. We can interact with it. It
can be 3D. We can play with it. You want to play a card game? All right, here’s a deck of
cards. We can play with it. If two of us are physically together and we have a third friend
who’s hologramming in, they can participate too. But in that world too — just as you don't want
your physical space to be cluttered because it
35:30 - 36:00 wears on you psychologically — I don't think
people are going to want their digital-physical space to feel that way either. That's more of an
aesthetic norm that will have to get worked out, but I think we’ll figure that out.
Going back to the AI conversation, you were mentioning how big of a bottleneck
the physical infrastructure can be. Related
36:00 - 36:30 to other open-source models, like DeepSeek and
so forth, DeepSeek right now has less compute than a lab like Meta and you could argue
that it's competitive with the Llama models. If China is better at physical
infrastructure, industrial scale-ups, getting more power and more data centers online,
how worried are you that they might beat us here? It's a real competition. You're seeing industrial
policies really play out. China is bringing online
36:30 - 37:00 more power. Because of that, the US really needs
to focus on streamlining the ability to build data centers and produce energy. Otherwise, I
think we’ll be at a significant disadvantage. At the same time, some of the export controls
on things like chips, I think you can see how they’re clearly working in a way. There was
all the conversation with DeepSeek about, "Oh,
37:00 - 37:30 they did all these very impressive
low-level optimizations." And the reality is, they did and that is impressive.
But then you ask, "Why did they have to do that, when none of the American labs did it?" It’s
because they’re using partially nerfed chips that are the only ones NVIDIA is allowed
to sell in China because of the export controls. DeepSeek basically had to spend
a bunch of their calories and time doing
37:30 - 38:00 low-level infrastructure optimizations
that the American labs didn’t have to do. Now, they produced a good result on text. DeepSeek
is text-only. The infrastructure is impressive. The text result is impressive. But every new
major model that comes out now is multimodal. It's image, it's voice. Theirs isn't.
Now the question is, why is that the
38:00 - 38:30 case? I don’t think it’s because they’re
not capable of doing it. It's because they had to spend their calories on doing these
infrastructure optimizations to overcome the fact that there were these export controls.
But when you compare Llama 4 with DeepSeek —I mean our reasoning model isn’t out yet, so the R1
comparison isn’t clear yet— but we’re basically in the same ballpark on all the text stuff that
DeepSeek is doing but with a smaller model. So
38:30 - 39:00 the cost-per-intelligence is lower with
what we’re doing for Llama on text. On the multimodal side we’re effectively leading
at and it just doesn’t exist in their models. So the Llama 4 models, when you compare them
to what DeepSeek is doing, are good. I think people will generally prefer to use the Llama
4 models. But there’s this interesting contour
39:00 - 39:30 where it’s clearly a good team doing stuff
over there. And you're right to ask about the accessibility of power, the accessibility of
compute and chips, because the work that you're seeing different labs do and the way it's
playing out is somewhat downstream of that.
39:30 - 40:00 So Sam Altman recently tweeted that OpenAI is
going to release an open-source SOTA reasoning
40:00 - 40:30 model. I think part of the tweet was that they
won’t do anything silly, like say you can only
40:30 - 41:00 use it if you have less than 700 million users.
DeepSeek has the MIT license, whereas I think a couple of the contingencies in the Llama
license require you to say "built with Llama" on applications using it or any model that you
train using Llama has to begin with the word
41:00 - 41:30 "Llama." What do you think about the license?
Should it be less onerous for developers? Look, we basically pioneered the open-source
LLM thing. So I don't consider the license to be onerous. When we were starting to push on open
source, there was this big debate in the industry. Is this even a reasonable thing to do? Can you
do something that is safe and trustworthy with
41:30 - 42:00 open source? Will open source ever be able to be
competitive enough that anyone will even care? Basically, when we were answering those
questions a lot of the hard work was done by the teams at Meta. There were
other folks in the industry but really, the Llama models were the ones that broke open
this whole open-source AI thing in a huge way.
42:00 - 42:30 If we’re going to put all this energy into it,
then at a minimum, if you're going to have these large cloud companies — like Microsoft and Amazon
and Google — turn around and sell our model, then we should at least be able to have a
conversation with them before they do that around what kind of business arrangement we should have.
Our goal with the license, we're generally not trying to stop people from using the model. We
just think that if you're one of those companies,
42:30 - 43:00 or if you're Apple, just come talk to
us about what you want to do. Let's find a productive way to do it together.
I think that’s generally been fine. Now, if the whole open-source part of the
industry evolves in a direction where there are a lot of other great options and the license
ends up being a reason why people don’t want to use Llama, then we’ll have to reevaluate
the strategy. What it makes sense to do
43:00 - 43:30 at that point. But I don’t think we’re there.
That’s not, in practice, something we’ve seen, companies coming to us and saying, “We don’t want
to use this because your license says if you reach 700 million people, you have to come talk to us.”
So far, that’s been more something we’ve heard from open-source purists like, “Is this as clean
of an open-source model as you’d like it to be?”
43:30 - 44:00 That debate has existed since the beginning
of open source. All the GPL license stuff versus other things, do you need to make it
so that anything that touches open source has to be open source too? Or can people take
it and use it in different ways? I'm sure there will continue to be debates around this.
But if you’re spending many billions of dollars training these models, I think asking the other
companies — the huge ones that are similar in size
44:00 - 44:30 and can easily afford to have a relationship
with us — to talk to us before they use it seems like a pretty reasonable thing.
If it turns out that other models are also really good. There’s a bunch of good open-source
models. So that part of your mission is fulfilled, and maybe other models are better at coding.
Is there a world where you just say, "Look, the open-source ecosystem is healthy. There’s plenty
of competition. We're happy to just use some other model, whether it's for internal software
engineering at Meta or deploying to our apps.
44:30 - 45:00 We don't necessarily need to build with Llama"?
Again, we do a lot of things. Let's take a step back. The reason why we're building our own
big models is because we want to be able to build exactly what we want. None of the other
models in the world are exactly what we want. If they're open source, you can take them and
fine-tune them in different ways. But you still
45:00 - 45:30 have to deal with the model architectures.
And they make different size tradeoffs that affect latency and inference cost. At the scale
that we operate at, that stuff really matters. We made the Llama Scout and Maverick models
certain sizes for a specific reason. They fit on a host and we wanted certain latency
— especially for the voice models that we’re working on — that we want to pervade everything
we're doing from the glasses to all of our apps
45:30 - 46:00 to the Meta AI app and all that stuff.
There's a level of control of your own destiny that you only get when you build the stuff
yourself. That said, AI is going to be used in every single thing that every company does. When
we build a big model, we also have to choose which internal use cases we're going to optimize for.
So does that mean for certain things we might say, "Okay, maybe Claude is better for building this
specific development tool that this team is
46:00 - 46:30 using”? All right, cool then use that. Great. We
don’t want to fight with one hand tied behind our back. We’re doing a lot of different stuff.
You also asked, would it not be important anymore because other people are doing
open source? On this, I'm a little more worried. You have to ask yourself this. For anyone who
shows up now and is doing open source — now that we have done it — would they still be
doing open source if we weren’t doing it?
46:30 - 47:00 I think there are a handful of folks who see
the trend that more and more development is going toward open source, and they're like, "Oh
crap, we need to be on this train or else we’re going to lose." If you have a closed-model API and
increasingly a lot of developers don't want that. So you’re seeing a bunch of other players start
to do some work in open source. But it's unclear
47:00 - 47:30 if it's dabbling, or fundamental for them the
way that it has been for us. A good example is what's going on with Android. Android started
off as the open-source thing. There's not really any open-source alternative. Over time,
Android has just gotten more and more closed. So if you're us, you need to worry that if we
stop pushing the industry in this direction,
47:30 - 48:00 all these other people… Maybe they’re only really
doing it because they're trying to compete with us and the direction we’re pushing things. They
already showed their revealed preference for what they would do if open source didn’t exist.
And it wasn’t open source. We just need to be careful about relying on that continued
behavior for the future of the technology that we're going to build at the company.
Another thing I've heard you mention is that
48:00 - 48:30 it's important that the standard gets built around
American models like Llama. I wanted to understand your logic there. With certain kinds of networks,
it is the case that the Apple App Store just has a big contingency around what it's built around.
But it doesn't seem like if you built some sort of scaffold for DeepSeek, you couldn't have
easily just switched it over to Llama 4, especially since between generations. Llama
3 wasn't MoE and Llama 4 is. So things are changing between generations of models as well.
What’s the reason for thinking things will get
48:30 - 49:00 built out in this contingent
way on a specific standard? I'm not sure, what do you mean by contingent?
As in, it's important that people are building for Llama rather than for LLMs in
general, because that will determine what the standard is in the future.
Look, I think these models encode values and ways of thinking about the world.
We had this interesting experience early on, where we took an early version of Llama and translated
it. I think it was French, or some other language.
49:00 - 49:30 The feedback we got from French people
was, "This sounds like an American who learned to speak French. It doesn’t sound
like a French person." And we were like, “what do you mean, does it not speak French
well?” No, it speaks French fine. It was just that the way it thought about the world seemed
slightly American. So I think there are these subtle things that get built into the models.
Over time, as models get more sophisticated,
49:30 - 50:00 they should be able to embody different
value sets across the world. So maybe that's not a particularly sophisticated example,
but I think it illustrates the point. Some of the stuff we've seen in testing some
of the models, especially coming out of China, have certain values encoded in them. And it’s
not just a light fine-tune to change that. Now,
50:00 - 50:30 language models — or something that has a kind
of world model embedded in it — have more values. Reasoning, I guess, you could say has values too.
But one of the nice things about reasoning models is they're trained on verifiable problems. Do you
need to be worried about cultural bias if your model is doing math? Probably not. I think
the chance that some reasoning model built
50:30 - 51:00 elsewhere is going to incept you by solving
a math problem in a devious way seems low. But there's a whole different set of issues
around coding, which is the other verifiable domain. You need to worry about waking up one day
and if you're using a model that has some tie to
51:00 - 51:30 another government, can it embed vulnerabilities
in code that their intelligence organizations could exploit later? In some future version you're
using a model that came from another country and it's securing your systems. Then you wake
up and everything is just vulnerable in a way that that country knows about and you don’t.
Or it turns on a vulnerability at some point.
51:30 - 52:00 Those are real issues. I'm very interested in
studying this because I think one of the main things that's interesting about open source is
the ability to distill models. For most people, the primary value isn't just taking a model
off the shelf and saying, "Okay, Meta built this version of Llama. I'm going to take it and
I'm going to run it exactly in my application."
52:00 - 52:30 No, your application isn't doing anything
different if you're just running our thing. You're at least going to fine-tune it, or try to
distill it into a different model. When we get to stuff like the Behemoth model, the whole value
is being able to take this very high amount of intelligence and distill it down into a smaller
model that you're actually going to want to run. This is the beauty of distillation. It's one of
the things that I think has really emerged as a very powerful technique over the last year, since
the last time we sat down. I think it’s worked
52:30 - 53:00 better than most people would have predicted. You
can basically take a model that's much bigger, and capture probably 90 or 95% of its intelligence,
and run it in something that's 10% of the size. Now, do you get 100% of the intelligence?
No. But 95% of the intelligence at 10% of the cost is pretty good for a lot of things.
The other thing that's interesting is that now, with this more varied open-source community, it's
not just Llama. You have other models too. You have the ability to distill from multiple sources.
So now you can basically say, "Okay, Llama’s
53:00 - 53:30 really good at this. Maybe its architecture is
really good because it's fundamentally multimodal, more inference-friendly, more efficient. But let’s
say this other model is better at coding." Okay, great. You can distill from both of them and
build something that's better than either individually, for your own use case. That's cool.
But you do need to solve the security problem of knowing that you can distill it in a way that's
safe and secure. This is something that we've
53:30 - 54:00 been researching and have put a lot of time into.
What we've basically found is that anything that's language is quite fraught. There's just a lot of
values embedded into it. Unless you don't care about taking on the values from whatever model
you're distilling from, you probably don't want to just distill a straight language world model.
On reasoning, though, you can get a lot of the way there by limiting it to verifiable domains, and
running code cleanliness and security filters.
54:00 - 54:30 Whether it's using Llama Guard open source,
or the Code Shield open source tools that we've done, things that allow you to incorporate
different input into your models and make sure that both the input and the output are secure.
Then it’s just a lot of red teaming. It’s having experts who are looking at the model
and asking, "Alright, is this model doing
54:30 - 55:00 anything after distillation that we don't want?"
I think with the combination of those techniques, you can probably distill on the reasoning
side for verifiable domains quite securely. That's something I'm pretty confident about and
something we've done a lot of research around. But I think this is a very big question. How
do you do good distillation? Because there’s so much value to be unlocked. But at the same
time, I do think there is some fundamental bias embedded in different models.
Speaking of value to be unlocked,
55:00 - 55:30 what do you think the right way to monetize
AI will be? Obviously digital ads are quite lucrative. But as a fraction of total GDP,
it's small compared to all remote work. Even if you can increase productivity without
replacing work, that's still worth tens of trillions of dollars. Is it possible that ads
might not be it? How do you think about this? Like we were talking about before, there's
going to be all these different applications, and different applications
tend toward different things. Ads are great when you want to offer
people a free service. Because it's free,
55:30 - 56:00 you need to cover it somehow. Ads solve this
problem where a person does not need to pay for something. They can get something that is amazing
for free. Also by the way, with modern ad systems, a lot of the time people think the ads
add value to the thing if you do it well. You need to be good at ranking and you need to
have enough liquidity of advertising inventory.
56:00 - 56:30 If you only have five advertisers in the
system, no matter how good you are at ranking, you may not be able to show something to
someone that they're interested in. But if you have a million advertisers in the system,
then you're probably going to be able to find something pretty compelling, if you're good at
picking out the different needles in the haystack that that person is going to be interested in.
So that definitely has its place. But there are also clearly going to be other business
models as well, including ones that
56:30 - 57:00 just have higher costs so it doesn't even
make sense to offer them for free. By the way, there have always been business models like this.
There's a reason why social media is free and ad-supported, but then if you want to
watch Netflix or ESPN or something, you need to pay for that. The content that's going
into that, they need to produce it, and that's very expensive for them to produce. They probably
could not have enough ads in the service in order to make up for the cost of producing the content.
Basically, you just need to pay to access it.
57:00 - 57:30 The trade-off is fewer people do it. Instead
of billions, you're talking about hundreds of millions of people using those services. There's a
value switch there. I think it's similar here. Not everyone is going to want a software engineer,
or a thousand software engineering agents, or whatever it is. But if you do, that's
something you're probably going to be willing to pay thousands, or tens of thousands,
or hundreds of thousands of dollars for.
57:30 - 58:00 That just speaks to the diversity of
different things that need to get created. There are going to be business models at
each point along the spectrum. At Meta, for the consumer piece we definitely want to
have a free thing. I'm sure that will end up being ad-supported. But I also think we're going
to want to have a business model that supports people using arbitrary amounts of compute to do
even more amazing things than what it would make sense to offer in the free service. For that, I'm
sure we'll end up having a premium service. But
58:00 - 58:30 I think our basic values on this are that we want
to serve as many people in the world as possible.
58:30 - 59:00 How do you keep track of all these different
projects, some of which we've talked about
59:00 - 59:30 today. I'm sure there are many I don't even
know about. As the CEO overseeing everything,
59:30 - 60:00 there's a big spectrum between going to
the Llama team and saying, "Here are the hyperparameters you should use," versus just
giving a mandate like, "Go make the AI better." And there are so many different
projects. How do you think about the way in which you can best deliver your
value-add and oversee all these things? A lot of what I spend my time on is trying to
get awesome people onto the teams. There's that, and then there's stuff that cuts across
teams. You build Meta AI, and you want to get it into WhatsApp or Instagram. Okay, now
I need to get those teams to talk together.
60:00 - 60:30 Then there are a bunch of questions like, “do
you want the thread for Meta AI in WhatsApp to feel like other WhatsApp threads, or do you
want it to feel like other AI chat experiences?” There are different idioms for those. So there
are all these interesting questions that need to get answered around how does this stuff
basically fit into everything we're doing? Then there's a whole other part of what we're
doing, which is pushing on the infrastructure.
60:30 - 61:00 If you want to stand up a gigawatt cluster,
first of all, that has a lot of implications for the way we're doing infrastructure buildouts.
It has political implications for how you engage with the different states where you're building
that stuff. It has financial implications for the company in terms of: "All right, there's a
lot of economic uncertainty in the world. Do we double down on infrastructure right now? If so,
what other trade-offs do we want to make around
61:00 - 61:30 the company?" Those are the kinds of decisions
that are tough for other people to really make. Then there's this question around taste and
quality. When is something good enough that we want to ship it? In general, I'm the steward
of that for the company. Although we have a lot of other people who I think have good taste as
well and are also filters for different things.
61:30 - 62:00 Those are basically the areas. AI is interesting
because, more than some of the other stuff that we do, it is more research and model-led than really
product-led. You can't just design the product that you want and then try to build the model to
fit into it. You really need to design the model first and the capabilities that you want, and
then you get some emergent properties. Then it's, "Oh, you can build some different stuff because
this turned out in a certain way." At the end
62:00 - 62:30 of the day, people want to use the best model.
That's partially why, when we're talking about building the most personal AI, the best voice,
the best personalization — and also a very smart experience with very low latency — those are
the things that we need to design the whole system to build. That's why we're working on
full-duplex voice. That's why we're working on personalization to both have good memory
extraction from your interactions with AI,
62:30 - 63:00 but also to be able to plug into all the
other Meta systems. That's why we design the specific models that we design, to have the kind
of size and latency parameters that they do. Speaking of politics, there's been this
perception that some tech leaders have been aligning with Trump. You and others donated
to his inaugural event and were on stage with him and I think you settled a lawsuit
that resulted in them getting $25 million.
63:00 - 63:30 I wonder what's going on here? Does it feel like
the cost of doing business with an administration? What's the best way to think about this?
My view on this is that he's the President of the United States. Our default, as an American
company, should be to try to have a productive relationship with whoever is running the
government. We've tried to offer support to previous administrations as well. I've been
pretty public with some of my frustrations
63:30 - 64:00 with the previous administration,
how they basically did not engage with us or the business community more broadly.
Frankly, that’s going to be necessary to make progress on some of these things. We're not going
to be able to build the level of energy that we need if you don't have a dialogue, and if they're
not prioritizing trying to do those things. A lot of people want to write this story about
what direction people are going. We're trying
64:00 - 64:30 to build great stuff, and we want to have a
productive relationship with people. That's how I see it. It is also how I would guess most others
see it, but obviously, I can't speak for them. You've spoken out about how you've
rethought some of the ways in which you engage and defer to the government, in
terms of moderation stuff in the past. How are you thinking about AI governance? Because
if AI is as powerful as we think it might be,
64:30 - 65:00 the government will want to get involved. What
is the most productive approach to take there, and what should the government be thinking about?
I guess in the past, most of the comments that I made were in the context of content moderation.
It's been an interesting journey over the last 10 years on this. It's obviously been an
interesting time in history. There have been novel questions raised about online content moderation.
Some of those have led to productive new systems
65:00 - 65:30 getting built, like our AI systems to detect
nation-states trying to interfere in each other's elections. I think we will continue building
that stuff out, and that has been net positive. With some other stuff, we went down some
bad paths. I just think the fact-checking thing was not as effective as Community Notes
because it's not an internet-scale solution. There weren't enough fact-checkers, and people
didn't trust the specific fact-checkers. You
65:30 - 66:00 want a more robust system. So I think what we got
with Community Notes is the right one on that. But my point on this was more that historically,
I probably deferred a little too much to either the media and their critiques, or to the
government, on things that they did not really have authority over. But just as like a central
figure, I think we tried to build systems where
66:00 - 66:30 maybe we wouldn't have to make all of the content
moderation decisions ourselves or something. I guess part of the growth process over the last
10 years is realizing, “Okay, we're a meaningful company. We need to own the decisions that we need
to make. We should listen to feedback from people, but we shouldn't defer too much to people who do
not actually have authority over this. Because at the end of the day, we're in the seat, and
we need to own the decisions that we make.”
66:30 - 67:00 It's been a maturation process, and in
some ways painful, but I think we're probably a better company for it.
Will tariffs increase the cost of building data centers in the US and
shift buildouts to Europe and Asia? It is really hard to know how that plays
out. I think we're probably in the early innings on that, and it's very hard to know.
What is your single highest-leverage hour in
67:00 - 67:30 a week? What are you doing in that hour?
I don't know. Every week is a little bit different. It's probably got to be the case
that the most leveraged thing you do in a week is not the same thing each week. Or else,
by definition, you should probably spend more than one hour doing that thing every week.
I don't know. Part of the fun of this job, and also of the industry being so dynamic, is
that things really move around. The world is
67:30 - 68:00 very different now than it was at the beginning
of the year, or even six months ago, or in the middle of last year. I think a lot has advanced
meaningfully. A lot of cards have been turned over since the last time that we sat down.
I think that was about a year ago, right? Yeah. I guess what you were saying
earlier that recruiting people is a super high-leverage thing you do.
It's very high-leverage, yeah. You talked about these models being mid-level
software engineers by the end of the year.
68:00 - 68:30 What would be possible if, say, software
productivity increased like 100x in two years? What kinds of things could be
built that can't be built right now? What kinds of things? That's an interesting
question. One theme of this conversation is that the amount of creativity that's going
to be unlocked is going to be massive. If you look at the overall arc of human
society and the economy over 100 or 150 years,
68:30 - 69:00 it's basically people going from being primarily
agrarian — with most human energy going toward just feeding ourselves — to that becoming a
smaller and smaller percent. And the things that take care of our basic physical needs have become
a smaller and smaller percent of human energy. That shift has led to two impacts: one is that
more people are doing creative and cultural
69:00 - 69:30 pursuits. The second is that more people, in
general, spend less time working and more time on entertainment and culture. I think that is almost
certainly going to continue as this goes on. This isn't the 1-2 year thing of what happens when
you have a super powerful software engineer. But over time, if everyone has these superhuman
tools to create a ton of different stuff, you're going to get incredible diversity. Part
of it is going to be solving hard problems:
69:30 - 70:00 solving diseases, advancing science, developing
new technology that makes our lives better. But I would guess that a lot of it is going
to end up being cultural and social pursuits and entertainment. I would guess the
world is going to get a lot funnier,
70:00 - 70:30 weirder, and quirkier, the way that memes on
the internet have gotten over the last 10 years. I think that adds a certain richness and
depth. In funny ways, it actually helps you connect better with people. Now all day long,
I just find interesting stuff on the internet and send it in group chats to the people I care
about, who I think are going to find it funny. The media that people can produce
today to express very nuanced,
70:30 - 71:00 specific cultural ideas is really cool.
That'll continue to get built out. It does advance society in a bunch of ways, even if it's
not the "hard science" way of curing a disease. If you think about it, the Meta social
media view of the world is that yeah, people are going to spend a lot more time doing
that stuff in the future. It's going to be a lot
71:00 - 71:30 better, and it's going to help you connect,
because it'll help express different ideas. The world is going to get more complicated,
but our technology, our cultural technology, to express these very complicated things — in a very
kind of funny little clip or whatever — is going to get so much better. I think that's all great.
I don't know about next year. One other thought that I think is interesting to cover is
that I tend to think that, for at least
71:30 - 72:00 the foreseeable future, this is going to
lead to more demand for people doing work, not less. Now, people have a choice of
how much time they want to spend working. I'll give you one interesting example we were
talking about recently. We have almost three and a half billion people using our services
every day. One question we've struggled with forever is how do we provide customer support?
Today, you can write an email, but we've never
72:00 - 72:30 seriously been able to contemplate having
voice support where someone can just call in. I guess that's maybe one of the artifacts
of having a free service. The revenue per person isn't high enough to have an
economic model where people can call in. But also, with three and a half billion
people using your service every day, the number of calls would be massive. It’d be like
the biggest call center in the world. It would be
72:30 - 73:00 like $10 or $20 billion a year to staff that.
So we've never thought too seriously about it, because it always seemed like there was
no way that could make sense. But now, as AI gets better, you're going to get to a place
where AI can handle a bunch of people's issues. Not all of them — maybe 10 years from now it can
handle all of them — but thinking about a 3-5 year
73:00 - 73:30 time horizon, it will be able to handle a bunch.
It's kind of like a self-driving car. They can handle a bunch of terrain, but they're not
doing the whole route by themselves yet in most cases. People thought truck-driving jobs
were going to go away, but there's actually more truck-driving jobs now than when we first started
talking about self-driving cars 20 years ago.
73:30 - 74:00 Going back to the customer support thing, it
wouldn't make sense to staff out calling for everyone. But let's say AI can handle 90% of that.
Then if it can't, it kicks it off to a person. If you get the cost of providing that service down
to one-tenth of what it would've otherwise been, then maybe now it actually makes sense
to do it. That would be cool. So the net
74:00 - 74:30 result is that I actually think we're probably
going to hire more customer support people. The common belief is that AI will automate
jobs away. But that hasn't really been how the history of technology has worked. Usually,
you create things that take away 90% of the work, Final question: Who is the one person in the
world today who you most seek out for advice?
74:30 - 75:00 Oh, man. I feel like part of my style
is that I like having a breadth of advisors. It's not just one person.
We've got a great team. There are people at the company, people on our board. There
are a lot of people in the industry who are doing new stuff. There's not a single person.
But it's fun. Also, when the world is dynamic,
75:00 - 75:30 just having a reason to work with people you like
on cool stuff… To me, that's what life is about. Great note to close on. Thanks for doing this.
Yeah, thank you.