Dear friends,
Recently Meta made headlines with unprecedented, massive compensation packages for AI model builders exceeding $100M (sometimes spread over multiple years). With the company planning to spend $66B-72B this year on capital expenses such as data centers, a meaningful fraction of which will be devoted to AI, from a purely financial point of view, it’s not irrational to spend a few extra billion dollars on salaries to make sure this hardware is used well.
A typical software-application startup that’s not involved in training foundation models might spend 70-80% of its dollars on salaries, 5-10% on rent, and 10-25% on other operating expenses (cloud hosting, software licenses, marketing, legal/accounting, etc.). But scaling up models is so capital-intensive, salaries are a small fraction of the overall expense. This makes it feasible for businesses in this area to pay their relatively few employees exceptionally well. If you’re spending tens of billions of dollars on GPU hardware, why not spend just a tenth of that on salaries? Even before Meta’s recent offers, salaries of AI model trainers have been high, with many being paid $5-10M/year, although Meta has raised these numbers to new heights.
Meta carries out many activities, including run Facebook, Instagram, WhatsApp, and Oculus. But the Llama/AI-training part of its operations is particularly capital-intensive. Many of Meta’s properties rely on user-generated content (UGC) to attract attention, which is then monetized through advertising. AI is a huge threat and opportunity to such businesses: If AI-generated content (AIGC) substitutes for UGC to capture people's attention to sell ads against, this will transform the social-media landscape.
This is why Meta — like TikTok, YouTube, and other social-media properties — is paying close attention to AIGC, and why making significant investments in AI is rational. Further, when Meta hires a key employee, not only does it gain the future work output of that person, but it also potentially gets insight into a competitor’s technology, which also makes its willingness to pay high salaries a rational business move (so long as it does not adversely affect the company’s culture).
The pattern of capital-intensive businesses compensating employees extraordinarily well is not new. For example, Netflix expects to spend a huge $18B this year on content. This makes the salary expense of paying its 14,000 employees a small fraction of the total expense, which allows the company to routinely pay above-market salaries. Its ability to spend this way also shapes a distinctive culture that includes elements of “we’re a sports team, not a family” (which seems to work for Netflix but isn’t right for everyone). In contrast, a labor-intensive manufacturing business like Foxconn, which employs over 1 million people globally, has to be much more price-sensitive in what it pays people.
Keep building! Andrew
A MESSAGE FROM DEEPLEARNING.AIUse the agentic coding assistant Claude Code to build applications faster.
News
The Re-Opening of OpenAI
The “open” is back in play at OpenAI.
What’s new: OpenAI released its first open-weights model since 2019’s GPT-2. The gpt-oss family comprises two mixture-of-experts (MoE) models, gpt-oss-120b and gpt-oss-20b, that are designed for agentic applications and free to use and modify.
How it works: The team pretrained the gpt-oss models on trillions of tokens of text including general knowledge, coding, math, and science. Fine-tuning focused on reasoning and tool use.
Results: Set to high reasoning effort, the models generally performed midway between o3-mini, o3, and o4-mini in OpenAI’s tests. Unless otherwise noted, OpenAI results come from OpenAI’s reporting, and DeepSeek R1’s results come from its report on its latest update of the model.
Behind the news: Founded in 2015 as a nonprofit corporation, OpenAI initially was devoted to open source development on the theory that AI would produce greater benefits and advance more safely if members of the community at large could inspect, use, and improve upon each others’ work. However, in 2019, the high cost of building cutting-edge AI models led the organization to form a for-profit subsidiary, and it stopped releasing large language model weights (although it continued to publish weights for models such as Clip, which produces similar embeddings for related images and text, and Whisper, a speech-to-text engine).
We’re thinking: A vibrant open source community is vital to AI’s ongoing progress! Every open model holds valuable knowledge and functionality.
Reasoning Boosts Carbon Emissions
In the era of reasoning models, delivering better answers to questions has an environmental cost. A new study quantifies the impact.
What’s new: Researchers estimated the emissions of carbon dioxide and other heat-trapping gases associated with using 14 open-weights large language models. (The information needed to study closed models is not publicly available.) Reasoning, total tokens generated, and accuracy on question-answering benchmarks were associated with higher greenhouse-gas emissions, according to findings by Maximilian Dauner at Munich Center for Digital Sciences and AI and Gudrun Socher at HM Hochschule München University of Applied Sciences.
How it works: The authors tested models of various sizes, with and without reasoning capabilities, using questions that required short and long answers.
Results: The authors found a clear trade-off between reasoning (and the higher resulting numbers of tokens generated and output accuracy) and greenhouse-gas emissions.
Yes, but: The authors’ estimates of carbon emissions likely are overestimates. Older GPUs such as the A100 are less energy-efficient than newer ones; and much cloud computing takes place in data centers powered by renewable energy sources that emit less carbon than global average energy consumption. For example, Google and Amazon match their electricity consumption with renewable energy, and Meta has powered its data centers solely by renewable energy since 2020.
Why it matters: The International Energy Agency projects that AI will consume increasing amounts of energy, and thus produce more greenhouse-gas emissions, as companies focus on training and serving ever larger models. Current AI poses a double-barreled challenge: The more accurate a model’s output, (i) the more emissions it will produce and (ii) the more people will query it. Much of the thinking about how to manage this issue has pointed to leaner parameter counts: Smaller models consume less energy. But the authors’ findings instead point to strategic deployment: The right model for the right task. AI providers can reduce emissions by routing inputs to models that can process them both accurately and efficiently, and by limiting outputs to appropriate lengths. These strategies don’t require building new infrastructure or models.
We’re thinking: We must continue to work toward improving AI’s energy efficiency and reducing its carbon emissions. That said, in many tasks, using AI produces fewer emissions than other approaches, such as using human labor.
Learn More About AI With Data Points!
AI is moving faster than ever. Data Points helps you make sense of it just as fast. Data Points arrives in your inbox twice a week with six brief news stories. This week, we covered Google’s release of AlphaEarth Foundations, a new AI model that builds better maps from satellite data to help track deforestation, urban growth, and more. Subscribe today!
GLM-4.5, an Open, Agentic Contender
The race is on to develop large language models that can drive agentic interactions. Following the one-two punch of Moonshot’s Kimi K2 and Alibaba’s Qwen3-235B-A22B update, China’s Z.ai aims to one-up the competition.
What’s new: GLM-4.5 is a family of open-weights models trained to excel at tool use and coding. The family includes GLM-4.5 and the smaller GLM-4.5-Air, both of which offer reasoning that can be switched on or off.
How it works: GLM-4.5 models include several architectural features that differ from other recent MoE models. Instead of adding more experts or making the experts use more parameters per layer (which would make the models wider), the team increased the number of layers per expert (which makes them deeper). The pretraining/fine-tuning process distilled three models into one.
Results: The team compared GLM-4.5 and GLM-4.5-Air to top open and closed models across 12 benchmarks that assess reasoning, coding, and tool use.
Behind the news: A rapid run of releases by teams in China — Kimi K2, Qwen3’s updates, and now GLM-4.5 — has established momentum in open-weights, large language models that are tuned for agentic behavior.
Why it matters: It’s not uncommon to distill larger models into smaller ones, sometimes to shrink the parameter count, sometimes to improve an existing small model’s performance. Z.ai’s approach distilled not a larger model but three specialized variations on the base model.
We’re thinking: The “best” open model for agentic applications is shifting weekly, creating both exciting opportunities and daunting challenges for developers.
Robot Surgeon Cuts and Clips
An autonomous robot performed intricate surgical operations without human intervention.
Results: Tested on 8 pig tissues, SRT-H successfully performed each operation, correcting its own mistakes along the way.
Yes, but: The authors tested SRT-H on tissues that had been removed from an animal’s body. Real-world surgery involves the body as a whole, and surgeons must manage bleeding, tissue motion from respiration, and visual occlusions that might challenge SRT-H.
Work With Andrew Ng
Join the teams that are bringing AI to the world! Check out job openings at DeepLearning.AI, AI Fund, and Landing AI.
Subscribe and view previous issues here.
Thoughts, suggestions, feedback? Please send to thebatch@deeplearning.ai. Avoid our newsletter ending up in your spam folder by adding our email address to your contacts list.
|