Nvidia: Small LLMs, Bigger Problems
Summary:
- Nvidia Corporation has experienced impressive growth in revenue and profits, but the rise of smaller, more cost-effective AI models and advancements by competitors like Intel may hinder future growth.
- Intel’s 5th Gen Intel Xeon Scalable processors offer efficient AI workloads for smaller local models, potentially reducing the demand for Nvidia’s high-end GPUs.
- Open-source models like Mistral 7B and the increase in viable alternatives from competitors are making AI technology more accessible and affordable, posing a challenge to Nvidia’s EPS growth.
Investment Thesis
Nvidia Corporation (NASDAQ:NVDA), one of the dominant forces in the artificial intelligence (“AI”) and data center chip markets, has reached a crucial tipping point in its scale & growth. In 2023, the company has had impressive performance, with revenue growing 206% alone in the for the third quarter of FY 2024 to 18.12 billion. However, the rapid rise of an emerging trend in the AI sector may start to hamper their growth.
More efficient Large Language Models (LLMs) capable of running locally and the significant advancements in AI chip technology by competitors like Intel (INTC) propose a possible roadblock for Nvidia. These smaller LLMs require much less data and computing power for training. Not only that, but they are much more cost-effective than the larger models that require Nvidia’s high-end GPUs.
What makes Intel’s 5th Gen Intel Xeon Scalable processors so special is their ability to operate smaller AI workloads efficiently (like Nvidia’s GPU solutions) but for smaller local models. This is due to their built-in AI acceleration. Nvidia’s current growth predictions rely heavily on the demand for their top line GPU solutions, which is an issue because of these new processors diminishing the demand for Nvidia’s GPU solutions. Therefore, I believe that these predictions are an overestimation of future growth for Nvidia. As developments in open-source models like Mistral 7B and viable alternatives from competitors continue, AI technology is becoming more accessible and affordable. Given this, Nvidia’s stock growth could face a reality check, so I think it’s a hold.
Background
In the recent fast-paced rise of AI technologies, Nvidia has emerged as one of the key players. 2023 has been huge for Nvidia, marked by a 206% year-on-year growth in the most recent quarter, with third-quarter FY 2024 revenue hitting $18.12 billion. This performance is based on Nvidia’s current dominance in the AI chip market, where they possess approximately 70% market share, and their pioneering role in AI technology due to the success of its CUDA platform.
Nvidia has been proactive when it comes to advancing their AI technology (yet another way they’ve strengthened their current leading position in the AI world). For example in 2021, Nvidia met face-to-face with one of their competitors, Intel, with the release of Grace CPU by Nvidia. This software directly challenged Intel’s dominance in AI technology within the data center CPU market, allowing Nvidia to excel in this field. Nvidia later released the BlueField DPU (data processing unit), which is designed for AI and accelerated computing. Along with that they also unveiled the DGX SuperPOD, a turnkey AI supercomputer. Their innovation has been strong, this supports their 70% market share. How much more market share can they get?
While Nvidia’s growth has been strong, the creation of smaller, more efficient LLMs and advancements by competitors like Intel could add headwinds. The recent trend of localized and cost-effective AI solutions is causing a shift in the demand for LLMs. This could (in turn) shift demand for GPUs.
Changes in the Last Month: New, Smaller LLMs Pose a Threat
My hold on Nvidia stems from the industry’s hard shift towards smaller, more efficient LLMs. These models can be run more locally, which reduces the demand for extensive and costly data center infrastructure.
The infrastructures used for the development and training of large models like GPT-3 can be expensive, costing millions of dollars. On the other hand these new, smaller models, require less data and computing power. This decrease in power requirement will drastically lower operational costs for developers. For example, GPT-3, one of the more prevalent LLMs in this year’s AI sector, cost OpenAI over $4.6 million to train. This cost could now be easily avoided using smaller models which can be developed and deployed at a fraction of this cost. Because of the drop in development cost, these smaller LLMs will be much more accessible to businesses. The opportunity for AI development (as costs drop) will expand to small and medium-sized enterprises (SMEs).
Not only do smaller LLMs vastly reduce cost, but they are also domain-specific to a company’s needs. By domain-specific, I mean that these LLMs are able to adapt specifically to a company’s needs. This can be done by prompt and adapter tuning, which then allows those companies to adjust these models. This increased adaptability allows businesses to achieve more accurate and relevant AI-powered solutions without the overhead associated with larger models.
As smaller LLMs gain in popularity large-scale models such as those produced by Nvidia are faced with a serious competitor. The growing trend of developing and applying smaller, more efficient models are reducing the need for high-end GPUs. Intel’s new chips are an example of where these models could go.
Mistral’s Advancements are Lowering the API Costs of Models. These Margins may be too Low for many Firms to Afford New NVDA Chips
The creation of open-source LLMs like Mistral 7B, a 7-billion-parameter model, has significantly impacted the AI model market, especially when it comes to API costs and hardware requirements. This model is presented as a favorable open-source alternative to other AI solutions, therefore in some cases providing superior performances. The release of this model is consistent with the trend of more affordable, accessible, and cost efficient AI models.
As my research of Mistral 7B has advanced, I’ve become more and more impressed. Mistral 7B is already making its own name in the world of AI, outperforming the Llama 2 13B model in all tests and competing closely with the Llama 1 34B in many metrics. Although this model has similar capabilities to Llama 2, the computational overhead is much lower, therefore presented as a more cost effective solution for businesses. Along with that, Mistral AI’s approach to open-source, community-driven development is predicted to provide a more adaptable and ethical oversight in AI applications.
With the development of Mistrak 7B, there has been a shift in the charging model. Instead of their cost being calculated per token of GPU usage, it is now being calculated per minute. The change can lead to companies saving vast amounts of money, especially if they are using Mistrak 7B for applications that don’t need a continuous high GPU usage.
With changes means dropping prices. Major AI model providers like Anthropic and OpenAI have cut their API costs meaningfully over the last two months. With cheaper models comes margin compression and an obvious emphasis to cut costs, especially at smaller AI startups.
Nvidia’s new H100 chips cost twice as much to train on according to data from CoreWeave (an Nvidia-backed startup). While these H100s can train models faster, many new smaller models do not need such intensive training. This is where Intel’s new cost-conscious (yet powerful) chips come in.
What’s Changed in the Last Week: Intel’s New Chips Make Cheaper Hosting Possible
On December 14th, 2023, Intel revealed their newest technology, the 5th Gen Intel Xeon Scalable processor. The release of this processor resulted in a pivotal moment in AI advancement. It presented a new technology that was capable of a large increase in overall performance for a variety of critical workloads, including artificial intelligence, analytics, networking, security, storage, and high-performance computing. What makes these processors so important is they provide an increased performance per watt, thereby lowering the operating cost. By producing a more cost-effective solution to AI, businesses could start shifting towards the use of Intel’s new technologies rather than Nvidia.
As I mentioned above, one of the features of the 5th Gen Intel Xeon processors is that performance is being increased per watt. This can be affiliated with the built-in AI acceleration in every core. This allows processors to handle demanding AI workloads, including deep learning inference and fine-tuning on models up to 20 billion parameters, while still producing effective results. This positions the 5th Gen Xeon processors as a viable option for running smaller, more localized AI models, such as LLMs.
There are significant improvements in performance in these new processors. For example, they provide a 21% average performance gain for general compute performance and result in a 36% higher average performance per watt. The cost of ownership compared if other processors can decrease by up to 77%. These are incredibly cost conscious.
Intel’s processors have already shown promising results in various applications. For example, IBM’s watsonx.data platform achieved up to 2.7x better query throughput with these processors compared to previous-generation Xeon processors. Google Cloud plans to implement the 5th Gen Xeon next year and is predicting increased performance in AI-based applications. These processors are being used across all types of companies, for example an indie game studio reported a 6.5x improvement in inference performance using Xeon processors over a GPU-based cloud instance.
Growth Will Slow and Year-Over-Year Hardware Comps are Challenging
The growth outlook for Nvidia in the coming years presents a complex picture. The 3rd quarter FY 2024 year-on-year revenue growth of 206% reflects Nvidia’s current dominance in the AI chip market, where it holds approximately 70% of the market share. Growth like this will be hard to sustain.
However, while Nvidia’s current performance is strong, the future might look a little different. The increasing development and adoption of smaller, more efficient LLMs and advancements by competitors in the AI chip market could reduce the demand for Nvidia’s high-end GPUs. Because of these new LLMs advancements, Nvidia’s growth might not maintain its current performance.
In FY 2024 (current year) Nvidia is projected to hit revenue of $58.77 billion. Next fiscal year revenue is supposed to grow an additional 54% to $90.66 billion. Keep in mind that most of Nvidia’s revenue comes from Chip sales (which is not a recurring revenue model like a Software as a Service, or SAAS, company). While this is an obvious statement, this part is key: they have to sell the same number of chips they will sell this fiscal year next year, but then plus another 54%. This assumes the pricing mix on these chips does not fall.
As models get smaller and costs to run the models are compressed as pricing compresses, AI companies will look to make the most of the hardware they have. This will mean longer cycle times going forward for hardware upgrades, and making the most of the hardware they currently have (in many cases this hardware is sufficient for smaller models).
Many of these companies that contributed to the sales of chips this year may not be marginal buyers next year as models get smaller and chip needs are not as strong. This will likely not cause revenue to drop, but growth to slow below the market expectations.
Valuation
With Nvidia facing tougher year-over-year comps for growth (and expecting to grow another 54% above that), I think some of their key valuation metrics are high and warrant a ‘wait and see’ approach, which is why I rated this stock a hold.
For example, Nvidia’s forward price to sales ratio is 20.22, almost 7 times higher than the industry average of 2.94. While the company’s strong growth has helped push this price to sales ratio up, part of the key here is the firm’s extremely high margins on their chips, which means while the high forward P/E ratio of 39.11 is notable, it’s not as challenging to understand.
However, these margins (rumored to be over 90% on some chips or a 1000% markup), are insane margins for hardware. These margins could compress if Nvidia’s demand growth starts to slow. This could decrease effective EPS growth making the P/E ratio seem high and justify pressure on the price to sales ratio as well.
Takeaway
As Nvidia navigates the evolving AI world, its future is somewhat murky with growth. While 2023 was a great year for them, I am hesitant to predict the same for the future of this company. The growing trend of smaller and most cost effective LLM technology may reduce the growth demand for Nvidia’s high-end GPUs. Nvidia’s competitors have significantly contributed to this shift in the AI market by producing new technologies such as Mistral 7B and Intel’s 5th Gen Intel Xeon Scalable processors.
Nvidia has a solid foundation due to their observed strong market position, however in order to stay consistent with their current stock performance, they will need to be very strategic in how they respond to their competitors and the ever-changing world of AI. In essence, Nvidia Corporation will have to keep these incredible growth rates up in order to justify the valuation they have (this is despite increased competition and different needs from LLM AI companies). I am not saying it’s impossible, I am just saying it is difficult. In essence, until we have more clarity, Nvidia Corporation stock is a hold.
Analyst’s Disclosure: I/we have no stock, option or similar derivative position in any of the companies mentioned, but may initiate a beneficial Long position through a purchase of the stock, or the purchase of call options or similar derivatives in INTC over the next 72 hours. I wrote this article myself, and it expresses my own opinions. I am not receiving compensation for it (other than from Seeking Alpha). I have no business relationship with any company whose stock is mentioned in this article.
Noah Cox (account author) is the Co-Managing partner of Noahs' Arc Capital Management. His views in this article are not necessarily reflective of the firms. Nothing contained in this note is intended as investment advice. It is solely for informational purposes. Invest at your own risk.
Seeking Alpha’s Disclosure: Past performance is no guarantee of future results. No recommendation or advice is being given as to whether any investment is suitable for a particular investor. Any views or opinions expressed above may not reflect those of Seeking Alpha as a whole. Seeking Alpha is not a licensed securities dealer, broker or US investment adviser or investment bank. Our analysts are third party authors that include both professional investors and individual investors who may not be licensed or certified by any institute or regulatory body.