SI
SI
discoversearch

We've detected that you're using an ad content blocking browser plug-in or feature. Ads provide a critical source of revenue to the continued operation of Silicon Investor.  We ask that you disable ad blocking while on Silicon Investor in the best interests of our community.  If you are not using an ad blocker but are still receiving this message, make sure your browser's tracking protection is set to the 'standard' level.
Technology Stocks : Qualcomm Moderated Thread - please read rules before posting
QCOM 178.63-1.3%Oct 29 3:59 PM EDT

 Public ReplyPrvt ReplyMark as Last ReadFilePrevious 10Next 10PreviousNext  
From: METMAN10/27/2025 1:15:13 PM
5 Recommendations

Recommended By
Art Bechhoefer
Jon Koplik
manning18
ryhack
VinnieBagOfDonuts

  Read Replies (2) of 196311
 
Here's a recent paper/study (10/22/2025) published just a week ago on the efficiency (token/watt) of Q's new AI chip offerings vs NVDA:

(added this comment after I posted: I also just realized that the chip news today refers to the AI200 and AI250, not the AI100, so this study is somewhat dated in terms of technology, perhaps - making today's announcement very significant. Still, an interesting read)

Serving LLMs in HPC Clusters: A Comparative Study of Qualcomm Cloud AI 100 Ultra and NVIDIA Data Center GPUs

arxiv.org

Can't speak much on the topic myself, but QHTrain likely can help make sense of it. Seems like Q's solutions are quite competitive right now, based on the application with some of the older NVDA chips. Someone on the Q Yahoo board (Skuldur) posted this (sort of) rebuttal to today's news:

"You can't compete in Nvidia's arena by going cheap. AMD could give their chips away for free and it would still make more sense to go with Nvidia. It's about tokens per watt.
I don't know enough about Qualcomm to even guess whether their new chips are any good, but with release dates in 2026 they're going up against Rubin, and in 2027 it's Rubin Ultra. Good luck."


That prompted me to search for any comparisons on token/watt. Don't know anything about NDVA's roadmap, but at first glance, this study is interesting on what QCOM can provide out of the gate.

Upon further review - NVDA is already 2 generations beyond the A100's used to compare in this study, per google AI overview:

While the A100 GPU is available through many cloud providers and data center vendors, it is not NVIDIA's latest offering; the company's newer and more performant GPUs include the H100 and H200. The A100 is based on the NVIDIA Ampere architecture and remains widely available for AI, data analytics, and high-performance computing workloads. However, NVIDIA announced an End-of-Life (EOL) for the A100 in January 2024.

Hopefully the power savings makes Q's option as, or more attractive for many more basic applications. Baby steps can lead to sprints eventually! It might take some time to gain a lot of traction vs NVDA, but optimistic.

Not trying to throw water on the fire here - just wishing to understand the (extent of the) significance of today's announcement.

regards,

metman
Report TOU ViolationShare This Post
 Public ReplyPrvt ReplyMark as Last ReadFilePrevious 10Next 10PreviousNext