Nvidia: Upside Potential Is Greater Than You May Think
Summary:
- Many analysts are eager to declare the peak of Nvidia Corporation stock, with the share price having multiplied by 4x since spring 2023.
- In reality, the upside potential is much greater once we holistically understand how Nvidia has cleverly strategized and planned out the company’s positioning in the AI revolution beyond chips.
- Nvidia stock remains a “buy,” and we discuss what price the stock should be trading at.
Nvidia Corporation (NASDAQ:NVDA) stock has multiplied by 4x since spring 2023, when this incredible AI-driven rally really began. The rally has also attracted many skeptics, doubting the stock’s ability to continue rallying from here. The truth is that Nvidia is much more than a semiconductor company selling popular AI chips, rather it has built an extensive AI platform. Once investors understand the extent to which Nvidia has gone to optimally capitalize on this AI revolution, the greater upside potential becomes more visible.
In our previous article, we discussed how Nvidia will capitalize on both the training and inferencing phases of this AI revolution, and cited that Nvidia should be worth at least $190 (post-split), implying at least 50% upside from here.
The market seems to be too fixated on the future growth potential of AI chips sales. However, Nvidia’s growth story extends well beyond hardware sales, as there is a massive opportunity in the new AI software industry for Nvidia to capitalize on, which we will cover in this article. Nvidia stock remains a “buy.”
Nvidia’s massive opportunity in the software industry
Now, to appreciate the greater upside potential from here, it is crucial to holistically understand how Nvidia has cleverly strategized and planned out the company’s positioning in the AI revolution that they anticipated before most of the world did. To fully understand how Nvidia will capitalize on this new era, we would have to go back to the comprehensive explanation CEO Jensen Huang had actually provided on the Q3 2024 Nvidia Earnings Call back in November 2023. Now, these excerpts are long, so we will break it down into two parts. The terminology may seem complex, but we will be breaking down what he is conveying in a simplified manner for all investors to understand.
There is a glaring opportunity in the world for AI foundry, and it makes so much sense. First, every company has its core intelligence. It makes up our company. Our data, our domain expertise, in the case of many companies, we create tools, and most of the software companies in the world are tool platforms, and those tools are used by people today. And in the future, it’s going to be used by people augmented with a whole bunch of AIs that we hire.
And these platforms just got to go across the world and you’ll see and we’ve only announced a few; SAP, ServiceNow, Dropbox, Getty, many others are coming. And the reason for that is because they have their own proprietary AI. They want their own proprietary AI. They can’t afford to outsource their intelligence and handout their data, and handout their flywheel for other companies to build the AI for them. And so, they come to us. We have several things that are really essential in a foundry. Just as TSMC as a foundry, you have to have AI technology. And as you know, we have just an incredible depth of AI capability — AI technology capability.
And then second, you have to have the best practice known practice, the skills of processing data through the invention of AI models to create AIs that are guardrails, fine-tuned, so on and so forth, that are safe, so on and so forth. And the third thing is you need factories. And that’s what DGX Cloud is. Our AI models are called AI Foundations. Our process, if you will, our CAD system for creating AIs are called NeMo and they run on NVIDIA’s factories we call DGX Cloud.
Our monetization model is that with each one of our partners they rent a sandbox on DGX Cloud, where we work together, they bring their data, they bring their domain expertise, we bring our researchers and engineers, we help them build their custom AI. We help them make that custom AI incredible.
That’s a lot to process, so let’s break it down. So back in November 2023, Nvidia introduced its first AI foundry in partnership with Microsoft Azure. The AI foundry combines Nvidia’s “generative AI model technologies, LLM training expertise and giant-scale AI factory [DGX Cloud]”. Essentially, Nvidia is offering its own generative AI capabilities to enterprises through the platforms of Cloud Service Providers (CSPs), to help companies build their own AI models.
In a previous article from June 2023, we delved deeply into Nvidia’s savvy strategy to offer its own DGX cloud services in partnership with the leading cloud providers, whereby around 10% of the chips sold to these CSPs is leased back by Nvidia for running its own cloud service.
CEO Jensen Huang had shared on the FYQ1 2024 Nvidia earnings call that:
It allows us…to create new applications together and develop new markets together. And we go-to-market as one team and we benefit by getting our customers on our computing platform and they benefit by having us in their cloud.
Indeed, Nvidia’s technology has become highly sought-after in the corporate world amid the AI revolution, making cloud providers willing to offer Nvidia’s unique expertise and tools in addition to their GPUs, in a bid to attract more enterprises to their cloud platforms. And Nvidia has been leveraging this prominence to vertically integrate upwards to reach enterprises directly, as outlined in the previous article:
while CSPs are striving to design chips optimized for their own cloud services, in order to offer enhanced performance more cost-effectively, Nvidia is offering DGX Cloud services that are optimized for Nvidia’s GPUs, thereby ensuring the availability of cloud services that leverage the full potential of Nvidia’s technology, offering optimum performance and cost-efficiency, and countering the threat of CSPs’ services powered by their own chips gaining prevalence.
So Nvidia is striving hard to sustain its presence in the cloud platforms. Now in that same article it was also suggested that (emphasis added):
Nvidia’s cloud endeavors are only likely to keep growing from here. Whether it continues to focus on the partnerships route, or increasingly serve enterprises directly instead over time.
And this seems to be playing out now, whereby Nvidia is not only partnering with CSPs to offer DGX Cloud, but is now partnering with leading Server OEMs like Dell to build what they call “AI factories”.
Interestingly, on each of the leading cloud providers’ earnings calls, the CEOs like to proclaim just how large the total addressable market remains for their cloud businesses, by highlighting how many businesses’ workloads still remain “on-premises.”
For example, on Amazon’s Q1 2024 earnings call, CEO Andy Jassy mentioned that:
85% or more of the global IT spend remains on-premises. And this is before you even calculate gen AI, most of which will be created over the next 10 to 20 years from scratch and on the cloud. There is a very large opportunity in front of us.
So the cloud providers are essentially striving to capitalize on more and more enterprises having to migrate their workloads to the cloud to be able to fully leverage the power of generative AI.
But the reality is, many businesses still remain averse to shifting to the cloud, primarily due to data security/ privacy concerns. And even the ones that do shift certain workloads to the cloud, oftentimes they are there to experiment with generative AI services, and repatriate their workloads back to on-premises.
In a recent podcast interview, Head of Groq Cloud, Sunny Madra shared that:
I’ve heard this you know in conversation with hyperscalers where customers will not trust and the hyperscaler will legally sign that they will not train using their data, and they still will not trust them. They just believe all these stories around the data making these models better that everyone just wants a way to get access to that data to make the models better
…
what happens you know with us which is basically maybe a pattern people come and try something in the cloud make sure that it works and then immediately want to get on the phone with you and say hey can I get this on Prem
This insight was further corroborated by a Barclays survey earlier this year revealing that 83% of CIOs plan to repatriate at least some workloads back to either private cloud or on-premises in the first half of 2024.
And this is the trend that Nvidia is now seeking to capitalize on.
Basically, they are helping enterprises build AI applications through DGX Cloud/ AI factories, taking advantage of the fact that these companies don’t want to give their data to other companies (public cloud providers).
Through AI factories/ DGX cloud, Nvidia is also taking advantage of the fact that AI talent is scarce, given the nascence of this generative AI revolution. In fact, research from venture capital giant Andreessen Horowitz on how enterprises are deploying AI revealed that:
Implementing and scaling generative AI requires the right technical talent, which currently isn’t in-house for many enterprises
Hence, companies are having to rely on expertise offered by leading tech companies, including Nvidia, to set up companies’ AI infrastructure and assist them in building AI models. This creates a beautiful opportunity for Nvidia to lock enterprises into the Nvidia ecosystem, as each company’s AI models/ software applications will be optimized to only run on Nvidia’s GPUs and hardware.
And that is just the training/ production phase. Once enterprises’ AI models/ applications have been built, they then run on Nvidia’s GPUs and use NVIDIA AI Enterprise software, which is $4,500 per GPU per year, as Jensen Huang went on to explain on the Q3 2024 earnings call:
Then that custom AI becomes theirs. And they deploy it on the runtime that is enterprise grade, enterprise optimized or outperformance optimized, runs across everything NVIDIA. We have a giant installed base in the cloud, on-prem, anywhere.
And it’s secure, securely patched, constantly patched and optimized and supported. And we call that NVIDIA AI Enterprise. NVIDIA AI Enterprise is $4,500 per GPU per year, that’s our business model. Our business model is basically a license. Our customers then with that basic license can build their monetization model on top of.
In a lot of ways we’re wholesale, they become retail. They could have a per — they could have subscription license base, they could per instance or they could do per usage, there is a lot of different ways that they could take a — create their own business model, but ours is basically like a software license, like an operating system. And so our business model is help you create your custom models, you run those custom models on NVIDIA AI Enterprise. And it’s off to a great start. NVIDIA AI Enterprise is going to be a very large business for us.
Now, this is a very key concept to understand to truly grasp the long-term bull case for Nvidia.
CEO Jensen Huang essentially wants the world’s software applications to run on Nvidia’s hardware and operating system. That is what NVIDIA AI Enterprise is, it is an operating system for running AI workloads.
Take ServiceNow (NOW) as an example, which has been one of the early movers in partnering with Nvidia and Dell (DELL) to build an AI factory for creating and running their generative AI-powered software services.
The enterprise software giant has already been leveraging NVIDIA AI Enterprise to build new generative AI capabilities “that can boost productivity for IT operations, customer service teams, and employees.” The way ServiceNow is monetizing these capabilities is by selling them as “add-ons” on top of customers’ existing subscription plans, essentially collecting additional recurring subscription revenues for access to the Gen AI features, similar to how Microsoft charges additional recurring subscription fees for “Copilot” on top of existing Microsoft 365 subscription prices.
So while ServiceNow charges additional subscription fees for generative AI-powered capabilities on the front-end, it will be paying Nvidia $4,500 per GPU every year for processing these new service requests at the back-end.
And this is just one example. Nvidia is striving to become the core operating system for all software applications. So Jensen Huang is essentially creating a business where the world’s software companies are paying him rent on a perpetual basis. These are computing dollars that have traditionally flowed to public/ private cloud service providers, or operating system providers for servers on-premises.
Keep in mind that NVIDIA AI Enterprise is not only accessible through AI factories on-premises, but also available on the marketplaces of all major cloud service providers, making it truly accessible everywhere, and creating an alluring selling point for Nvidia’s AI platform, as it can support enterprises’ hybrid-cloud strategies too (half workloads in public cloud, the other half on-premises).
Now that being said, while the public cloud providers have been keen to offer Nvidia’s technology through their platforms, they will certainly strive
to keep software companies on their public cloud services, and increasingly encourage them to use their own in-house hardware/ software solutions for running their software applications. For instance, on Amazon’s (AMZN) Q4 2023 earnings call, CEO Andy Jassy boasted about the number of software companies using AWS’s chips:
we built custom AI training chips, named Trainium, and inference chips, name Inferentia…We already have several customers using our AI chips including Anthropic, Airbnb, Hugging Face, Qualtrics, Ricoh and Snap.
Nevertheless, Nvidia’s GPUs are not likely to fall out of favor anytime soon given the extensive partner ecosystem built around its products helping sustain its moat, including 5.1 million developers.
Additionally, Nvidia will certainly strive to take advantage of individual enterprises’ needs for optimum data security and privacy, offering them on-premises solutions by building AI factories for them. In fact, on the FYQ1 2025 earnings call, CFO Colette Kress proclaimed the number of AI factories they are building for all sorts of companies beyond enterprise software companies like ServiceNow (emphasis added):
Large clusters like the ones built by Meta and Tesla are examples of the essential infrastructure for AI production, what we refer to as AI factories.
These next-generation data centers host advanced full-stack accelerated computing platforms where the data comes in and intelligence comes out. In Q1, we worked with over 100 customers building AI factories ranging in size from hundreds to tens of thousands of GPUs, with some reaching 100,000 GPUs.
Considering the fact that each AI factory can consist of tens of thousands of GPUs already, and Jensen Huang anticipating that “the days of millions of GPU data centers are coming” at Computex recently, $4,500 per chip on an annual basis should generate massive streams of recurring revenue streams through NVIDIA AI Enterprise. So this would be companies running their customized AI models/ software applications on Nvidia’s platform to provide all sorts of services to their own end-customers, whether that is in the form of a chatbot providing an answer to a user’s question, or a software application generating a video based on a creator’s prompts.
Last year, Nvidia revealed that they expect the annual market opportunity for NVIDIA AI Enterprise to be worth $150 billion, as part of the grander $1 trillion market opportunity for the entire company.
Now, how much of this total addressable market, or TAM, that Nvidia will be able to capture remains to be seen. But the more GPUs Nvidia sells now, the more AI factories Nvidia builds for enterprises across the world, the greater its share of the AI software market will be too in the future.
Will Nvidia stock have a major pullback?
Now because we are still very early in the phase of enterprises building out AI data centers, Nvidia’s sales revenue and earnings are likely to remain high. CEO Jensen Huang highlighted on the earnings call where he thinks we currently stand at in terms of the build-out phase:
If you’re 5% into the build-out versus if you’re 95% into the build-out, you’re going to feel very differently. And because you’re only 5% into the build-out anyhow, you build as fast as you can.
Now once the cloud service providers, consumer internet companies and early enterprise customers are almost done building out their AI infrastructure, their purchases of Nvidia’s chips will inevitably slow from current levels. And it is at this stage, where the massive stock price pullback could occur.
After all, what caused the Dot-com bubble to burst over two decades ago was tech companies’ sales and earnings growth rates slowing down and not living up to market expectations. A recent news piece from Seeking Alpha revealed insights from a Goldman Sachs note on the Dot-com burst:
According to the analysis, the “failure to meet lofty growth expectations ended the Dot-com valuation boom.”
The sales CAGR as of January 2000, during 1999-2001, for Cisco (CSCO) was 28% but realized 11%; for Microsoft Corp. (MSFT) was 16% but realized 10%; and for Intel (INTC) was 17% but realized -5%.
Now today, Wall Street analysts are estimating that this AI infrastructure build-out phase will last for 5 years, with BofA analyst Vivek Arya estimating that:
We think that the market for generative AI is in the second year of what could be a three to five year deployment cycle. The demand for hardware could triple to 300 billion over the next few years.
Hence, we are still in the early stages of this build out, and Nvidia’s sales and earnings growth rates are likely to remain high over the next 5 years, with the EPS FWD Long-Term Growth (3-5Y CAGR) estimated to be 32.70%.
But does that mean you should sell your NVDA shares once we are approaching the end of this estimated 5-year deployment cycle?
Truth is, no one really knows how long this AI infrastructure build-out phase is going to last.
As generative AI increasingly revolutionizes more aspects of the digital economy, and new types of services arise that weren’t possible before, more and more enterprises could decide to set up AI factories of their own over the next decade, keeping demand for Nvidia’s hardware high without a major slowdown in sales revenue/ profit growth.
Furthermore, even if we do witness a notable slowdown in hardware sales in a few years’ time, Nvidia has a burgeoning software market opportunity ahead of it to capitalize on, as discussed earlier. Hence, continuous growth in NVIDIA AI Enterprise revenue could cushion any drop off in hardware sales revenue.
In fact, locking customers into the NVIDIA AI Enterprise software ecosystem should also benefit the company’s hardware business in the form of high-volume replacement cycles (i.e., data centers replacing the oldest generation chips with latest generation GPUs).
Nvidia has impressively moved to a one-year rhythm cycle, launching new generation GPUs every year. How frequently customers replace old chips with new GPUs remains to be seen.
If customers can reap the kind of Return-on-Investment figures that Jensen shared on the last earnings call, then customers could indeed be inclined to continue buying new chips to stay competitive:
Training and inferencing AI on NVIDIA CUDA is driving meaningful acceleration in cloud rental revenue growth, delivering an immediate and strong return on cloud provider’s investment. For every $1 spent on NVIDIA AI infrastructure, cloud providers have an opportunity to earn $5 in GPU instant hosting revenue over four years.
…
using Llama 3 with 700 billion parameters, a single NVIDIA HGX H200 server can deliver 24,000 tokens per second, supporting more than 2,400 users at the same time. That means for every $1 spent on NVIDIA HGX H200 servers at current prices per token, an API provider serving Llama 3 tokens can generate $7 in revenue over four years.
As per Nvidia’s own long-term annual market opportunity projections, the tech giant expects the “Chips & Systems” total market size to be worth $300 billion per annum. Note that this does not just include AI GPUs, where Nvidia is estimated to possess “between 70% and 95% of the market,” but also adjacent hardware like networking equipment, including InfiniBand, Ethernet, switches and NICs. In a previous article, we discussed Nvidia’s savvy strategy of increasingly selling its GPUs as part of pre-built HGX supercomputers, whereby clusters of 4 or 8 AI chips are combined using InfiniBand and NVLink technology, and also includes the NVIDIA AI Enterprise software layer. This approach allows Nvidia to optimally capture as much market share as possible of the $300 billion “Chips & Systems” market ahead (as well as the AI software market), fending off competition from networking equipment manufacturers like Broadcom and Arista Networks, and of course chipmakers like AMD and Intel.
But the point is, if Nvidia generated over $47 billion in data center revenue last fiscal year (including some software revenue), and the company expects the total market opportunity for “Chips & Systems” to grow to $300 billion on an annual basis (of which it is likely to hold a dominant market share), the tech giant indeed still has multiplicative revenue growth potential ahead of it. Nvidia’s executives are clearly expecting a lot more data centers to be built out, and the chips/hardware upgrade cycle is also likely to entail high-volume orders on an annual basis if the total market opportunity is expected to be $300 billion per year.
Risks to bull case for Nvidia
Nvidia’s largest customers are designing their own chips: A favorite argument among the bears is that the tech giant’s biggest customers are also designing their own chips, which includes the major cloud providers (Microsoft Azure, Amazon AWS and Google Cloud), as well as consumer internet giant Meta Platforms.
In fact, an enterprise survey conducted by UBS recently revealed that:
Nearly 70% of respondents indicated they were using some sort of Nvidia hardware as an LLM training platform. Another 21% said they were using AWS Trainium and 4% were using Intel’s (INTC) Gaudi.
Nvidia still holds the dominant lead in the AI chips space, but the fact that Amazon’s AWS has been able to successfully encourage a growing number of its cloud customers to use its own in-house chips for training/ inferencing is certainly noteworthy.
However, this bearish argument for not buying NVDA stock is becoming weaker for several reasons.
Firstly, as CFO Colette Kress highlighted on the last earnings call (emphasis added):
Strong sequential data center growth was driven by all customer types, led by enterprise and consumer internet companies. Large cloud providers continue to drive strong growth as they deploy and ramp NVIDIA AI infrastructure at scale and represented the mid-40s as a percentage of our Data Center revenue.
While CSPs still make up almost half of Nvidia’s Data Center sales, the concentration risk in this segment is subduing as demand for Nvidia’s GPUs broadens to other customer segments.
Although that being said, the CFO also went on to highlight that the fact that the main source of “enterprise” and “consumer Internet companies” revenue came from Tesla (TSLA) and Meta Platforms (META), respectively:
Enterprises drove strong sequential growth in Data Center this quarter. We supported Tesla’s expansion of their training AI cluster to 35,000 H100 GPUs.
…
Consumer Internet companies are also a strong growth vertical. A big highlight this quarter was Meta’s announcement of Llama 3, their latest large language model, which was trained on a cluster of 24,000 H100 GPUs.
Nonetheless, the customer base for Nvidia’s GPUs should continue to diversify and broaden. As we discussed earlier in-depth, enterprises’ needs for data privacy/ security could impede their shift to cloud platforms and encourage them to stay on-premises instead. This should create immense sales growth opportunities for Nvidia in the enterprise space, as companies decide to build AI factories instead.
So while the bears argue that Nvidia’s future sales growth could be in danger of a major slowdown going forward due to CSPs designing their own chips, the opposite may actually be the case, whereby cloud providers’ future sales growth could be undermined by Nvidia building on-premises AI factories for an increasing number of enterprises, reducing their need to migrate to the cloud to take advantage of generative AI.
Speculative behavior could exacerbate stock volatility: a growing number of market participants, particularly on the retail side, are seeking leveraged exposure to NVDA through leveraged ETFs. This includes the Direxion Daily Semiconductor Bull 3X Shares ETF (SOXL), and the GraniteShares 2x Long NVDA Daily ETF (NVDL), whose Assets Under Management have grown to $12 billion and $5 billion, respectively. While such investment vehicles can deliver lucrative returns on the upside, any short-term pullbacks in NVDA stock would also multiply the losses to the downside.
As a result, weak hands that are simply seeking to ride the AI hype for quick returns are also highly likely to be the first ones to sell out of the leveraged ETFs and NVDA stock on the slightest hint of the bull run potentially coming to an end. This could exacerbate pullbacks in the stock, but also create new buying opportunities for long-term investors that comprehensively understand the growth opportunities ahead.
Nvidia’s financial performance and valuation
Nvidia has been delivering outstanding growth rates for its data center business on a consistent basis over the past year, growing by another 427% last quarter. And there are no signs of this slowing down as demand is broadening out beyond CSPs.
Furthermore, over the past several quarters, skeptics have been arguing that competitors like AMD and Intel could catch up to Nvidia and undermine the tech giant’s growth rates and pricing power. However, no competitor has been able to introduce technology that matches the performance of Nvidia’s GPUs/ AI platform, enabling CEO Jensen Huang to maintain immensely strong pricing power for its next-generation chips.
Consequently, the high top-line revenue growth rates continue to flow down to the bottom line, with diluted EPS growing by 803% over the past year, while boasting a net profit margin of 57%.
And Nvidia remains in a strong position to sustain strong pricing power with its next-generation Blackwell platform that will start shipping this year, with analysts at Jefferies confirming that “industry checks suggest further upside, with beats accelerating in FY26 as Blackwell ramps.”
With Nvidia estimated to hike the price of its new edition GPUs by around 40% over the previous generation, complemented by high-margin software revenue growth through NVIDIA AI Enterprise (as discussed earlier), the company’s top and bottom-line growth rates are unlikely to slow down any time soon.
This is clearly going to be a long-term growth story, as Nvidia is well-positioned to dominate the AI era beyond just GPUs. Although surprisingly, Nvidia stock’s valuation continues to be relatively reasonable at less than 47x forward earnings, which is roughly in-line with its 5-year average.
Market participants like to compare Nvidia’s performance to that of Cisco Systems during the dot-com era. Nonetheless, with the valuation for NVDA still reasonable at 47x forward earnings, current buyers of the stock have little to hesitate about. For context, Cisco’s valuation at its peak before the crash amid the dot-com bubble was 196x forward earnings.
Moreover, when we adjust the Forward P/E ratio by the expected future EPS growth rate to determine the stock’s Forward Price-Earnings-Growth [PEG] multiple, the stock appears attractively cheap at 1.42, which is well below its 5-year average Forward PEG ratio of 2.13.
As discussed earlier, a growing proportion of Nvidia’s revenue will compose of recurring software revenue going forward, which should also benefit hardware upgrade cycles. Hence, as the market perceives Nvidia evolving from a cyclical stock to a secular growth stock, NVDA should certainly be able to sustain a Forward PEG ratio of above 2 over the long term.
Nvidia should be trading closer to its 5-year average Forward PEG multiple of 2.13 given the company’s growth potential. This would imply 50% upside potential from here. Hence, the stock should be worth at least $190 at present.
Though keep in mind that while Nvidia’s valuation is reasonable today, if the stock climbs to an unsustainable high valuation multiple due to extremely speculative behavior (e.g., through the leveraged ETFs mentioned earlier in the Risks section), then the stock may not remain a good investment despite the anticipated AI growth story playing out over the next decade.
Nonetheless, for now, NVDA remains attractively valued relative to its growth story, making it a stock worth buying.
Analyst’s Disclosure: I/we have a beneficial long position in the shares of NVDA either through stock ownership, options, or other derivatives. I wrote this article myself, and it expresses my own opinions. I am not receiving compensation for it (other than from Seeking Alpha). I have no business relationship with any company whose stock is mentioned in this article.
Seeking Alpha’s Disclosure: Past performance is no guarantee of future results. No recommendation or advice is being given as to whether any investment is suitable for a particular investor. Any views or opinions expressed above may not reflect those of Seeking Alpha as a whole. Seeking Alpha is not a licensed securities dealer, broker or US investment adviser or investment bank. Our analysts are third party authors that include both professional investors and individual investors who may not be licensed or certified by any institute or regulatory body.