Fundamental Thinking featured image

DeepSeek Rattles Markets But Not the Outlook for AI

A one-year-old artificial-intelligence (AI) startup born out of a Chinese hedge fund released a powerful AI model on January 20 that is challenging investors’ assumptions about the economics of building such systems.

R1, as the model is called, is an open-source, advanced-reasoning model—the kind that is designed to mimic the way humans think through problems. It was developed by DeepSeek, whose founder, Liang Wenfeng, reportedly accumulated 10,000 of NVIDIA’s graphics processing units (GPUs) while at his quantitative hedge fund, which relied on machine-learning investment strategies. The kicker: DeepSeek says it spent less than US$6 million to train the model that was used as a base for R1—a fraction of the billions of dollars that Western companies such as OpenAI have spent on their foundational models. This detail stunned the market and walloped the share prices of large tech companies and other parts of the burgeoning AI industry on January 27.

The knee-jerk reactions are a reminder that we are still in the early stages of a potential AI revolution, and that no matter the conviction some insiders and onlookers may seem to have, no one knows with certainty where the path will lead, let alone how many twists and turns the industry will encounter along the way. As more details become available, companies and investors will be able to better assess the broader implications of DeepSeek’s achievement, which may reveal that the initial market reaction was overdone in some cases. However, should DeepSeek’s claims that its methods lead to dramatic improvements in cost efficiency be substantiated, it may actually bode well for the adoption of AI tools over time.

Although the techniques employed by DeepSeek are not entirely new, the company’s execution is considered extraordinary for how it significantly reduced training and inference costs. The company reportedly not only used fewer and less-powerful chips but also trained its model in a matter of weeks. There is considerable debate regarding the accuracy of the roughly US$6 million training cost figure that has been reported as well as speculation that DeepSeek may have utilized rival models such as Meta Platforms’ Llama, Alibaba’s Qwen, and OpenAI’s GPT-4 in a process known as distillation to create a student model. Therefore, it may be that the true training cost of DeepSeek’s model is understated.

There is the risk that more efficient AI models may reduce the growth rate of demand for computing power and, in turn, the pace of investment in AI infrastructure. But while investors seem to have equated improved efficiency with a diminished need for AI infrastructure, the history of technology shows that better performance and lower costs typically lead to wider adoption and faster growth over time. AI could follow the same trajectory.

Among the likely beneficiaries of cheaper, more efficient AI are the software providers building applications on top of these models. As more users discover the advantages of AI, increased demand for user-facing software also helps the cloud-services companies that provide the computing power and data storage. However, the commoditization of large language models may negatively affect creators of these models, particularly those that have eschewed open-source technology.

Also, greater computing efficiency may lead to higher volumes of cloud data and thus the need for more data centers and electrical power over time. This would benefit infrastructure providers on the condition that higher demand more than offsets price declines.

The field of AI is still evolving, and further advances in the technology and the methods for developing it are likely. Innovation is always disruptive, but on the whole, DeepSeek’s breakthrough would seem to be a good sign for the industry.

OOOM featured image

TSMC Leads Emerging-Markets Returns due to AI Enthusiasm

Momentum strategies were a powerful force for global equities in 2024, primarily ones that followed artificial intelligence (AI) and its extended value chain. The trend was particularly acute in emerging markets (EMs).

For the calendar year, virtually all of the returns in the MSCI EM Index could be traced to just five stocks, led by TSMC, which on its own contributed nearly half of the index’s return. The dominance of these stocks was most evident in the fourth quarter. Every sector except Information Technology fell, and the only regions that rose were the Middle East, which eked out a small gain, and Taiwan, home to a number of prominent tech stocks including TSMC (the “T” stands for Taiwan.)

TSMC rocketed on surging demand for AI-related chips and a seemingly unassailable leadership position in the industry following failed attempts by competitors to take market share in advanced process technologies. The company has more than 80% market share in leading-edge semiconductors globally and fabricates almost all of the chips designed by NVIDIA. TSMC expects AI-related revenue to grow at a cumulative rate of 50% over the next five years.

Hon Hai Precision, the giant Taiwanese electronics contract manufacturer, also saw its shares soar as growth prospects have been boosted by the demand for AI servers. The other top-five contributors, however, came from China and outside the AI-momentum trade: Tencent, Meituan, and Xiaomi.

OOOM featured image

Salesforce, ServiceNow Benefit from Next Phase of the AI Wave

By now, it is clear to most investors that the companies benefiting from the move to artificial intelligence (AI) include more than just a couple of chipmakers. One analogy for the increasing size and breadth of the AI industry is a tsunami: As the wave of corporate investment in AI builds, and the underlying hardware, foundational machine-learning models, and early-stage software applications all continue to improve, a broader set of tech providers has been able to benefit from the associated demand.

For example, Alphabet and Meta are among large cloud-services companies—also known as “hyperscalers”—that are building both the physical infrastructure and large-language models that are needed for AI technologies to be adopted by corporations more broadly. While the servers in their data centers continue to require powerful graphics processing units designed by NVIDIA and manufactured by TSMC, these hyperscalers are also designing their own custom chips, called application-specific integrated circuits (ASICs), which they are developing in partnership with chipmaker Broadcom, boosting that company’s growth outlook as well.

Advances in AI technology are also benefiting enterprise-software providers, as the rise of agentic AI in the fourth quarter brought the long-term promise of computers with human-like problem-solving capabilities into sharper focus. Agentic AI is capable of sophisticated reasoning and can automatically come up with ways to solve complex, multi-step problems—a significant step forward by models such as OpenAI’s o3 and Google’s Gemini 2 as compared to the more limited tasks performed by the previous generation of AI technology. Therefore, agentic AI is likely to be more broadly useful in business software, with companies such as Salesforce and ServiceNow well positioned to offer powerful tools based on this technology. At a recent event in San Francisco, Salesforce Chief Executive Officer Mark Benioff said the company is focused on its agentic-AI platform, Agentforce, adding, “The only thing we’re going to do at Salesforce is Agentforce.”

OOOM featured image

Chipmaking Is Getting More Complex. Daifuku’s Smart Monorails Keep Fabs Running Smoothly

In semiconductor manufacturing, a single speck of dust poses a threat to production. It’s why cleanrooms, the sterile labs where silicon wafers get etched and cut into pieces, and then packaged as finished chips—with thousands of steps in between—contain few humans. To reduce the risk of contamination and defects, materials are largely transported by automated monorail systems that travel along the ceiling.

Source: Daifuku.
While advances in generative artificial intelligence (AI) have put a spotlight on the companies that design and manufacture chips, as well as their data-center customers, providers of cleanroom technology play an increasingly critical role in a world of high-performance computing. Not only is the industry for cleanroom automation characterized by an attractive competitive structure, but new trends and challenges in chipmaking are also improving the growth outlook for this specialized material-handling technology. One player in particular may stand to benefit, and that is Daifuku.