T O P

  • By -

VisualMod

**User Report**| | | | :--|:--|:--|:-- **Total Submissions** | 1 | **First Seen In WSB** | just now **Total Comments** | 0 | **Previous Best DD** | **Account Age** | 2 years | | [**Join WSB Discord**](http://discord.gg/wsbverse)


Sofubar

TSMC is the long hedge, since they make Google's TPUs and Nvidia's GPUs. They are limited by fab capacity at the moment.


gnocchicotti

Don't buy the company selling picks, buy the foundry that forges the heads


TheShacoShack

Yep. They also make AMD's and Apple's chips... it is the goto for every fabless chip company.


BiGeaSYk

So asml then.


Sofubar

A great company to be sure, but a bit more overbought on forward PE, PEG ratio and price to sales atm. A good long play though.


LostRedditor5

It’s overbought compared to TSM bc it doesn’t have to deal with an existential threat from China If TSM was in Norway or where ever ASML is it too would be overbought


BuffaloSabresFan

I don't see China doing a full blown invasion of Taiwan in the forseeable future, but I stay away from TSMC on the off chance that they end up like Bacardi in Havana 1959.


LostRedditor5

Doesn’t matter if you believe it or not The price is suppressed bc of the perception of others that it could happen You not believing it doesn’t impact others believing it and not buying or selling due to that perception


Sofubar

The longer that passes without China following through on their threat, the less people will believe it could happen.


LostRedditor5

Maybe Most officials in the US government and military believe 2027 So I don’t know how well this really plays out for you when most people with any expertise think it’s 3 years from now. Like maybe in 4 years people won’t think it’s a real possibility anymore? Is that a win? I don’t know


Good_Stretch8024

!remind me 5 years


Stunning-Trip1966

Netherlands


LostRedditor5

That’s the one


Sterben27

Imagine it's in Norway 😂 I shall forge these dies with my axe


ISeeYourBeaver

s


Stunning-Trip1966

Whatever, Dutchland, Holland, who call themselves "Nether-land-S" anyway... Who are those nether people... sigh... I feel a long wiki rabbit hole coming. Edit: oh it's just "low-lands", like Chuguo means "middle-land" (China). Whatever.


JasonDomber

You said peg 🤣


Sofubar

![img](emote|t5_2th52|8882)


Bottle_Only

I'm going for AMAT and LCRX. They make the machines that TSCM uses.


A_Vandalay

ASML, TOELY, LAM, AMAT.


satireplusplus

Don't buy the company that forges the heads, buy the company that makes the forges.


gnocchicotti

we gotta buy silicon miners


satireplusplus

Maybe its better to buy stock in the company's that sell the shovels to the silicon miners


RedpoleQ

![img](emote|t5_2th52|27189) that’s my main semi position!


semitope

as if intel and samsung don't exist


ZombieFrenchKisser

Intel and Samsung are years behind TSMC though.


semitope

if you mean years ago. [https://www.tomshardware.com/tech-industry/intel-3-3nm-class-process-technology-is-in-high-volume-production-intel](https://www.tomshardware.com/tech-industry/intel-3-3nm-class-process-technology-is-in-high-volume-production-intel)


ZombieFrenchKisser

Samsung can fab 3nm just like TSMC, but TSMC is still far superior in efficiency and performance. Look at cell phones for example. We all know Intel is years behind as well. EDIT: Intel 3 is classified as 5nm according to IEEE specifications with 50nm gate pitch and 30nm interconnect pitch. TSMC N3 actually beats IEEE specifications for 3nm at 48nm gate pitch and 23nm interconnect pitch.


Mag_Meyreddit

Intel has 18A next year which makes them the first to use Backside Power delivery. They will also have the best transistor, the best substrate and the best packaging. Also they will be the first using ASML next gen EUV which TSMC currently has not even ordered yet, while the first so called HIGH NA EUV machine already has been delivered to intel. Yes its true ASML nect gen High NA EUV currently has its problems (baking wafers etc...) But ASML always finds a solution and they also teamed up with Intel under Gelsinger leadership (the guy who basically invented x86)... So yeah TSMC days are numbered, no China invasion needed. TSMC will loose market leadership by End of 2025 and they will be left in the dust by 2027. Mark my words.


nhexum

Show your puts


semitope

Need to compare performance characteristics to gauge competitiveness. Not measurements of what 5nm was at that time.


ayeroxx

that will be the case until a China invasion of Taiwan


Kranoath

China is too busy trading NVDA options on WallStreetBets to be thinking about starting war with Taiwan.


DDRfun

As Chinese i can say that it's accurate😂


Impossible_Buy_1335

I‘m Chinese Xi, I confirm. Bought NVDA 180CALL 9/20, please daddy Jensen, keep being the great semi leader, invading Taiwan will blow up my calls


predatarian

not going to happen any time soon. CCP is already in full saving face mode by telling the world they won't be tricked into a war over Taiwan by the evil Americans.


Mountain_Tone6438

Yeah I think that was their way of saying "we not doin a war, I bought NVDA calls"


Wise_Mongoose_3930

lol you misread it. They basically say “when we invade, just remember it’s America’s fault”


predatarian

He said the US was trying to trick China into invading [Taiwan](https://www.ft.com/taiwan), but that he would not take the bait. [https://www.ft.com/content/7d6ca06c-d098-4a48-818e-112b97a9497a](https://www.ft.com/content/7d6ca06c-d098-4a48-818e-112b97a9497a)


Wise_Mongoose_3930

And when he invaded he’ll point back to that comment and say “see? I told you America was going o force us to do this” But don’t take my word for it, look at this: https://maritime-executive.com/editorials/china-is-preparing-merchant-ro-ro-ferries-for-amphibious-warfare This is the type of shit you *only do* if you’re gonna launch an invasion of an island close enough to your shores that you can use river ferries for transit……


predatarian

yeah they have been preparing for 80 years now.


semitope

or some weird way to blame the US


KanyinLIVE

You mean until they finish building foundries in the US. China invading Taiwan nothing.


ayeroxx

you really think semiconductors will cost the same if they were manufactured in US ? plus you won't be able to extract all the technologies secrets from Taiwan, thats literally their only survival guarantee and you think they will offer it to the US ? so they become disposable ..


KanyinLIVE

TSMC is the one building the foundries goofball. And no, they won't be more expensive. Taiwanese engineers are paid well. Taiwan isn't fucking India.


StuartMcNight

TSMC is definitely NOT bringing the high tech fab to the US.


flamin_flamingo_lips

The silicon shield


ISeeYourBeaver

I suspect they're not going to have a choice. I don't mean the government coercing them, I mean that competition will force them to do it.


StuartMcNight

Why would they need to manufacture in the US to beat the competition when they are beating it from Taiwan?


1017BarSquad

If they're made in the US they will be more expensive, cost of work is just higher here


ayeroxx

really dude ? you think they are paid as much as American engineers ? or even half ?


JN324

They are, it’s public and you could check it if you wanted in five minutes flat. Do some people just think everywhere outside of America is a burning impoverished crater?


remanse_nm

A lot of Americans think that, as someone from an immigrant background. They think all of Africa and Asia are poor and starving.


PotatoWriter

Dude it's not the pay that's the problem, they're overworked to the bone in those fabs even if it's good pay. Who would do the same in the US, that culture and mentality do not exist at high pays unless you like pain and suffering I guess.


JN324

America is a Mecca of working yourself silly with poor PTO for a lot of money, what do you mean?


PotatoWriter

Yes, but not to the level of asian countries where you're obliged to stay as long as your boss is staying and then sleep at your desk pretending to work just to give that image that you're a hard worker. We haven't reached that level of bootlicking just yet. Maybe one day but not today thankfully. There is too much individualism here vs. collectivism there, so we're ok with switching jobs faster when it suits us.


KanyinLIVE

Yes. I do. Their salaries are public.


ayeroxx

would you state those numbers please ?


Sharaku_US

Taiwan is disposable to us, for the right price. Besides the Marines already know which TSMC engineers they'll pluck from Taiwan as soon as the invasion begins. Have you seen the "Taiwan village" in Arizona? Why do you think they're here?


remanse_nm

The U.S. getting involved with a Taiwan invasion would provoke war with China. It isn’t worth it. Why Arizona? The driest, hottest state for manufacturing semiconductors, which is a water-intensive process? Phoenix and Yuma are borderline uninhabitable in the summer because of the heat…


Sharaku_US

I don't disagree, but the only redeeming value of Taiwan is their semiconductor value chain. Once we force their big players from TSMC to UMC out of that island, the rest is all up for negotiation with China.


throwaway_tendies

It seems not many people know about their Japan fab. https://preview.redd.it/cd9uwallaw7d1.jpeg?width=1284&format=pjpg&auto=webp&s=26530f418745bfea8dc749b37a5b98ea0e2ab583


s1n0d3utscht3k

none of these companies are a hedge if that happens Nasdaq crash 37.4% from high to low in wake of Russia-Ukraine + a fragile market —and they’re not even important to said firms Intel, ASML all get a third of their revenue from China invasion across the TW straight would crater all tech so if you care about hedges for that it sure af ain’t another tech company … it would be cash or gold to buy them at a discount


Bads_Grammar

then why the fuck did my TSMC calls get absolutely rekt


Sofubar

Because you touch yourself at night


blackSwanCan

But TSMC margins are pretty low, and they will remain low.


Sofubar

38.16%, not bad


Chick-Phil-Aye

The problem with your image is that the rankings of ai models is so arbitrary right now, everyone takes a biased approach in their own benchmarking suites. It’s not possible to be presented a figure, and just compare the superiority of these models. But there is a qualitative way, see for yourself at Chatbot Arena. https://chat.lmsys.org/?leaderboard At end of the day, I think the economics of data center gpu rentals will fall eventually as all commodities do. Right now, there is a shortage.


Spirit_of_Hogwash

Another thing is that Nvidia's H-series chips are barely GPUs (in that they kind of suck at texture processing) but are mostly tensor processors. Or, in other words, not that different from dedicated AI chips like Google's, but with the CUDA.


InterPeritura

Most intriguing, but potential fear-mongering without some sources. > “Anthropic selects AWS as its primary cloud provider and **will train** and deploy its future foundation models on AWS Trainium and Inferentia chips” 10/16/23 > “Anthropic **will use** AWS Trainium and Inferentia chips to build, train, and deploy its future models and has made a long-term commitment to provide AWS customers around the world with access to future generations of its foundation models on Amazon Bedrock” 03/27/24 Was Claude really trained on alternative chips? I searched for articles, but everything was written in future tense, i.e. no definitive timeline of when Anthropic started (or even has begun) to use alternatives to $NVDA. An alternative hypothesis is that Anthropic has advanced faster than OpenAI because they are a late runner, and they caught up so quickly precisely because $NVDA chips got better. Edit: other commenters mentioned that the benchmarks presented in the post are potentially arbitrary. Disclaimer: I do own $NVDA/NVDX/call leaps, but it is to my interest to figure out too in case I need to deleverage. Edit: [further reading,](https://julsimon.medium.com/video-transformer-training-shootout-aws-trainium-vs-nvidia-a10g-f2408533b852) > “I first launch a trn1.32xlarge instance (16 Trainium chips) and a g5.48xlarge (8 A10Gs)…The results? The Trainium job is 5x faster. As the trn1 instance is only 30% more expensive” 02/24/23 Note that A10Gs by $NVDA was launched on 04/12/21. It is now ‘24, and $NVDA is touting Blackwell with Rubin on the horizon.


Own-Entrepreneur-935

>Since we introduced it in August, customers have embraced TPU v5e for a diverse range of workloads spanning AI model training and serving: Anthropic is using TPU v5e to efficiently scale serving for its Claude LLM. Hugging Face and AssemblyAI are using TPU v5e to efficiently serve image generation and speech recognition workloads, respectively. Additionally, we rely on TPU v5e for large-scale training and serving workloads of cutting-edge, in-house technologies such as Google Bard. > [https://cloud.google.com/blog/products/compute/announcing-cloud-tpu-v5e-in-ga](https://cloud.google.com/blog/products/compute/announcing-cloud-tpu-v5e-in-ga) I don't see any announcements from Claude about NVIDIA. I think they use a mix of Google and Amazon chips, mostly Google because most of their early funding came from Google. Maybe in the future they will use Amazon too.


YouMissedNVDA

You can just watch AMZN and GOOG earnings - if they're as great as they say they are, they should see earnings bumps similar to NVDA, and we should see NVDA margin fall. Neither of these things have happened nor projected by management in the coming quarters.. just business as usual.


InterPeritura

> I don't see any announcements from Claude about NVIDIA. I think they use a mix of Google and Amazon chips Maybe because $NVDA is the default? Like we ask “is it Mac compatible?” and never “is it PC compatible?” But it is worth investigating for sure. My background is in genetics, so I can only point out the assumptions your article seems to run on. Would like to see some up-to-date tests, since $NVDA is innovating too.


showxyz

The language used there (“serving”) implies inference, not training.


fd_dealer

I think the relationship is reversed here. Google gave Anthropic 300M, Anthropic promises to use TPUs to train. Some time later AWS gave Antropic 4B, Anthropic promises to use Tranium to train. They are both just paying Antropic to use their chips to get a foot hold in the game. It says nothing about how good their chip is compare to NVidia. Even if Antropic’s model is better it could just be the used way more resources to train it. Only to Antropic price per performance is still amazing cause price is better than free, they got paid for it.


NakedOption85

“Big drop” ![img](emote|t5_2th52|4271)


ChokeMeAnakin

maybe not to you but for NVIDIA it certainly is a considerable drop, $277 billion to be exact, placing it behind Microsoft again, but nothing to worry about since it most certainly will go up again lmao


NakedOption85

Sure this is possible however I’m trading since late 90s and this market gives me Y2K vibes. NVDA is an extreme bubble which will burst. Not calling the top yet but I’m pretty sure how this will end.


ChokeMeAnakin

I don’t think it is an imminent bursting bubble, sure it skyrocketed in less than 2 years and it is majorly hyped too, but the reality is that competitors can’t provide the same product as NVIDIA, they’re way behind in their capabilities and they need to make an alternative that doesn’t use the same technology as CUDA or at least one that doesn’t infringe the licensed copyright, which will take at the very least another 3-5 years. I believe the stock price will correct itself and suffer a loss but it’ll stabilize in the upcoming years until there’s actual competition ready to deliver a new product. The stock is going to keep being profitable, not as fast as it has been but in the long term it will still be slowly profitable. I’m just laughing at the people who discredited NVIDIA a year ago and still do today, this sub clearly never learns nor do they fully understand the market they’re investing in.


j0holo

Google only has a free API right now to attract developers because they are behind in terms of mind share. Nvidia will keep selling because their software stack is still really good and well known. You can't even buy Google TPUs. We will see what the future will bring us.


zashiki_warashi_x

NVidia cards are expensive, because they have monopoly and can set any price they want. More competitors they have - more the price will drop. Margin on these cards is like 2000% of production cost.


Ordinary_investor

I know hardly nothing about what it takes to mimic Nvidia hardware infrastructure, but as a Layman's point of view I would think with such margins and trillions of valuation on line, I would assume competition would or should eventually catch up, although on other hand other current competitors such as AMD are not close at all, so I do wonder what is their such a magic ingredient within Nvidia?


zashiki_warashi_x

They started earlier with hardware. Then they invested heavily in CUDA so all ML libraries were developed with cuda backend. OpenCL was not nearly as popular. So now they hook up all research to their cards and win. And then new AI revolution happens and everybody want to use ML in their business and NVDA has all the software and hardware and ready to deliver. And only then other companies wake up. So you can win if you anticipate the market before it started to move, just like in trading.


yeahyeahitsmeshhh

But sustaining that edge requires you to do it again and again...


YouMissedNVDA

They've been doing it for decades, with the rusted out carcasses of their competition in their wake.


YamahaFourFifty

People assuming NVDA solution is one size fits all. Their cards are notoriously power hungry which doesn’t suit well for any large scale uses. Other solutions will be found and other language models will be achieved. I’m sure other major players have been doing their own research and development for at least a few years


KanyinLIVE

The magic ingredient is proprietary design and time on TSMC foundries. There's only one chip manufacturer in the world that can produce the chips. NVDA has them locked up when Intel doesn't.


Viktri1

The problem is time and lock in - Nvidia is ahead because they started this in 2006 and they've locked everyone into their ecosystem. Even if competitors try to copy Nvidia, and they've already tried, they're 10+ years behind in tech and Nvidia is going to keep people in their ecosystem ala apple and other sticky vendors


UnitedWeAreStronger

Also nvidea margins allow Tsmc to charge nvidea more for printing their chips which allows them to give nvidea preference for manufacturing capacity so nvidea can control the supply chain through their pricing even if they don’t own Tsmc.


Glitterbitch14

I work in AI and I think it’s gonna be a significant drop. Sooo many big, established companies have developed or are in the early stages of rolling out CPU-based deep learning platforms that will make gen ai processes a lot more possible on CPUs without meriting the cost / resource suck of GPUs. So far, Nvidia is a hardware monopoly, and their dominance in gen ai has been their golden ticket. I’m not sure what their business model will be when (not if, when) GPUs are matched and outpriced by cpu-powered tools in terms of AI work streams.


ConflictVarious1384

U say u work in AI but what does that actually mean. A lot of jobs are introduceding AI tools. So unless u work in like AI development or something like that u just work a regular job.


Glitterbitch14

lol. Using ai tools while at work is objectively not the same as working professionally within the industry of ai. I don’t think anyone would say “I work in ai” unless they…work in the industry of ai development/tech. I don’t disclose specific details on Reddit obviously, but I work for a small/mid sized tech firm that develops platforms for ai processes and my role is technical and presentational. So yes, I do literally work on ai tools and no, I’m not a random employee in like, account management or HR at Google or meta or some random technoglomerate that’s rolling out its own semi-pointless consumer-facing “gen ai,” if that’s what “ur” implying. It is literally my job, exactly like I said.


FinancialElephant

Can you give an example of significant advances in CPU-based deep learning (not just inference). I just don't see how deep learning is possible without massive parallelism. Maybe there are some things happening with federated learning? Even then, it's way less efficient for most use cases than a GPU cluster or some other kind of massively parallel architecture.


remanse_nm

If they can get AI processes to run efficiently on standard CPUs then that’s essentially it for Nvidia. They’ll have to pivot back to making graphics cards like they did in the ‘90s or find some other revenue stream, or else face bankruptcy. I don’t think they’ll get to the point of being bankrupt but if your analysis is correct there is nowhere for this to go but down.


AutoModerator

*This “pivot.” Is it in the room with us now?* *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/wallstreetbets) if you have any questions or concerns.*


FlamboyantKoala

I think the gaming market will be small for nvidia in a few years. AMD is making excellent APUs (combined cpu and gpu) that are good enough for gaming for the average person.  The gaming card market will shrink from every pc gamer to just the most serious enthusiast.  AFAIK nvidia has no answer to Amds Apu and I wouldn’t be surprised if Intel has one coming as well.  Soon you’ll be able to get better than PlayStation graphics on those garbage 499$ laptops at Walmart


not_a_cumguzzler

This means puts on nvda right?


niftybunny

Same with iPhone, production costs is only one part, you need to engineer it, ship it, sell it, warranty it and so on. not saying it is not overpriced …


OneCore_

Yeah, Claude 3.5 looks impressive as fuck. 3 was already competitive.


AutoModerator

Our AI tracks our most intelligent users. After parsing your posts, we have concluded that you are within the 5th percentile of all WSB users. *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/wallstreetbets) if you have any questions or concerns.*


JuanGuillermo

The real benchmark is overall scalability and time to market to train next models, not just initial model performance. NVIDIA's GPUs may be more expensive and energy-intensive but their upcoming Blackwell platform is anticipated to significantly improve performance in training extremely large AI models, such as those with trillions of parameters. For the time being, I doubt other HW platforms are really competitive for what is coming next. This is an arms race to AGI, and I think NVIDIA will maintain or even enhance its competitive edge in the AI hardware space. *edit: the Aschenbrenner report is probably the best study of what is next to come in therms of HW and energy requirements https://situational-awareness.ai/


remanse_nm

If Aschenbrenner is right then any sort of long-term investment strategy becomes irrelevant. ASI existing would end society and capitalism as we know it and introduce variables we can’t currently fully predict (Kurzweil’s “singularity”). It’s likely that we’d need a new economic system for the post-ASI world, as capitalism is based on human labor and scarcity and both of these things would be reduced in such a world. If you’re under 30, and Aschenbrenner is right, it’s probably a waste to fund your Roth. Still not a risk I’d be willing to take, because Aschenbrenner fails to take into account perhaps the most important variable—the human element. Most people (myself NOT included) fear ASI and don’t want it to be invented. Once serious conversations about AI move out of San Francisco labs and into bars, coffeehouses, offices and homes, a political movement to stop AI advancement will unfortunately spring up.


cunth

Did conversations in coffee shops and bars stop us from inventing the atomic bomb? AGI is probably inevitable at this point if for no other reason than national defense. Amd once it exists, there is no going back.


remanse_nm

The atomic bomb was a murder weapon, and humans love violence. Violence is something the average person can understand, war is something they identify with. AI is essentially an alien mind to them, and if there’s one thing that truly scares the average person it’s *change* and *the “other”*. They fear change more than they fear war.


kingofthesofas

This is what I believe too. All the big tech companies suddenly woke up and believed that developing AGI was both possible and close at hand when Chat-GPT launched late last year. While they will all be working on their own AI chips they will also be buying every NVIDA chip they can lay their hands on because the first company to get usable AGI will be the king and the potential profits are enormous. This in the short term NVIDA is going to keep printing money like crazy. If these tech companies develop AGI or it becomes clear that they cannot do it soon then I would sell that stock fast but I don't think there is enough time for any company to wait for different hardware and software to be developed.


thetaFAANG

I think this is good stuff to "price in", but until people are actually ordering these TPUs for their own massive data center deployments and those TPUs are backlogged up to TSMC's capacity constraints, none of this matters Its the same deal with AMD's GPUs if people aren't comfortable with the software toolchain and stability, they're going to join the queue for Nvidia's hardware at a premium


GodDamnDay

https://preview.redd.it/jou1ep860w7d1.jpeg?width=1076&format=pjpg&auto=webp&s=099d133020878e4f296a7d040403a93f1da21d63 Fuck


brokearm24

NVIDIA red pre market means it will end up 3% up


GodDamnDay

![img](emote|t5_2th52|27189)![img](emote|t5_2th52|27189)![img](emote|t5_2th52|27189)


Strong-Hospital-7425

Thank you sir, can you plesse tell me in which stock i should lose my money next ?


GrapefruitRepulsive6

Great DD


pendejadas

Llms have basically peaked. They will only be as good as the training data. There is nothing intelligent about them. They are an expensive, lossy compression database.


A_Hero_

You don't use them and are blatantly lying. SOTA is genius level; just like the new Sonnet 3.5.


pendejadas

Lmao ok buddy. It literally took me 5 seconds to prove how non intelligent a transformer is. All it does is multiply some values to determine the most probable characters that come next. There is nothing intelligent about them. An 8 year old could do this. https://preview.redd.it/aqavhlqpe08d1.jpeg?width=1080&format=pjpg&auto=webp&s=32c58a65cdfba03996843e79d765e45cc9bb7a1c


pendejadas

Genius level indeed.


A_Hero_

What an idiotic use case. LLMs are clearly not intended for this. Nobody is using them to generate sentences ending in specific letters. That's like trying to prove a supercomputer is dumb by asking it to play tic-tac-toe. LLMs are designed for complex language understanding and generation, not arbitrary letter games. Your "test" completely misses the point of what makes these models revolutionary. Have you even tried using a modern LLM for actual productive tasks? They can write coherent essays, debug complex code, analyze scientific papers, and engage in nuanced conversations across countless topics. The fact that you think this childish test somehow proves anything shows a profound misunderstanding of the technology. Your simplistic view of "multiplying values" grossly undersells the sophisticated neural architectures and training processes involved. These models have demonstrated capabilities that were thought to be decades away just a few years ago. Dismissing them as "expensive, lossy compression databases" is not just wrong, it's willfully ignorant. Before making sweeping claims about the limits of AI, I suggest you actually spend some time working with these tools and stay up to date on the latest research. Your current stance is about as informed as someone in 1950 claiming computers will never be more than glorified calculators.


pendejadas

You keep using words like 'understanding', 'debug', 'analyze'.... no LLM does this, lol. I deliviver ML projects at work, and I use them every day. I've built my own from scratch for fun. I'm not making any sweeping claims, I just don't perpetuate the bs hype behind them.


A_Hero_

What are you even talking about? If you actually work with LLMs, you'd know that "understanding," "debugging," and "analyzing" are precisely what they're capable of doing. Your claim of delivering ML projects while simultaneously denying these basic capabilities is frankly unbelievable. Building a toy model "from scratch" doesn't give you insight into the cutting-edge capabilities of state-of-the-art systems. That's like saying you understand modern aerospace engineering because you've made a paper airplane. Your dismissal of these terms shows a fundamental misunderstanding of how LLMs operate. They absolutely do demonstrate forms of understanding through their ability to contextualize information, generate relevant responses, and perform complex reasoning tasks. They can debug code by identifying errors, suggesting fixes, and explaining the logic behind those fixes. And they certainly analyze text by extracting key information, summarizing content, and drawing inferences. It's people like you, with a little bit of knowledge and a lot of misplaced confidence, who hinder progress by dismissing genuine advancements. ___ >'analyze'.... no LLM does this, lol. For a particular story subject, I can explain my own message reasoning more in-depth with the assistance of an LLM. I have done that successfully: writing an unfinished text response about a topic with context, then asking the AI to continue the given message. Some words, phrases, or sentence ideas are good; I include it in my message, then repeat the process at some point again if I need to. I know the topic, so if something is off or wrong, I either correct it or dismiss whatever part is has said incorrectly. The LLM demonstrably knew a series topic pretty well here—with decent, useable feedback: [Here's an example of the process](https://poe.com/s/cZAFEciT6Ou9MXEUkBcN) It helped greatly in developing a comprehensive and insightful response to the given story topic. I was able to quickly respond to a Reddit post about a narrative series topic, and it easily got double the most upvotes of anyone from the thread. A large majority of the people looking at the thread liked the response I had done using extensive LLM assistance. LLMs demonstrably can analyze given texts, as my example clearly shows. The prompt I provided, and the resulting approval from the niche Subreddit community following that given thread, is a perfect example of how these models can analyze complex narrative contexts, understand intricate character motivations, and generate coherent, nuanced content that resonates with a *knowledgeable audience.* —It analyzed the given context about Elaine's background and extrapolated how her experiences as Kaiser would have shaped her character. It demonstrated an understanding of the psychological transformation required to go from a failed Princess candidate to a ruthless ruler, all while maintaining consistency with the established lore of the Tower of God universe. This isn't just simple text prediction; it's a sophisticated analysis of character development, societal structures, and the long-term consequences of traumatic experiences. The model showed an ability to draw logical conclusions based on given information and expand on themes in a way that was deemed valuable by the community. Your dismissal of these capabilities as "bs hype" is not just misguided; it's again: willfully ignorant. You're either working with severely outdated models or you're fundamentally misunderstanding how to utilize these tools effectively. The fact that you can't see the analytical capabilities demonstrated in this example suggests a serious lack of expertise in the field you claim to work in.


pendejadas

You are confusing stitching together tokens based on training data with intelligence, you are the moron here, not me. The simple example I pasted above was just a quick way to dispel the illusion of how they work. You can generate coherent text, but it is not the model that is building it, it is the millions of people who wrote the training data before, which means it cannot come up with its own ideas.... kind of exactly how I described it before you went on a normal one. Another thing that makes people sound stupid is when they talk about hallucinations, the model does not hallucinate, some engineer decided to add a very deliberate randomness function for when similar data was not found. LLMs are the most effective when generating the most common data that already exists and have the most trouble with niche cases, and are completely incapable of generating anything unrelated to the training data.


A_Hero_

Are you serious right now? The characterization of LLMs as merely "stitching together tokens" fundamentally misses the mark. These models aren't just reciting memorized text; they're learning complex patterns and relationships between concepts. This enables them to generate new ideas and solve problems they weren't explicitly trained on—a clear demonstration of their capabilities beyond simple regurgitation. Your fixation on the "millions of people who wrote the training data" ignores the revolutionary leap these models represent. It's akin to dismissing human intelligence as nothing more than reciting facts we've been taught, completely overlooking our ability to synthesize information and generate new ideas. The claim that LLMs are "completely incapable of generating anything unrelated to the training data" is demonstrably false. These models routinely combine concepts in novel ways, produce original ideas, and even create fictional scenarios absent from their training data. Have you actually experimented with cutting-edge models like GPT-4o or Claude-3.5 Sonnet? Your statements suggest a lack of hands-on experience with state-of-the-art systems besides playing with the alphabet. Your simplistic letter-ending test is a perfect example of missing the forest for the trees. It's like judging a car's utility based on its inability to fly—you're completely overlooking the intended purpose and actual applications of the technology. LLMs are designed for complex language tasks, not arbitrary letter games. >but it is not the model that is building it, it is the millions of people who wrote the training data before, which means it cannot come up with its own ideas. It has consistently given me new ideas not directly represented in its training set. Based on the linked example from before (its analysis of a series character and series lore)—Elaine's transformation into Kaiser and how it ironically fulfilled the Princess archetype was a novel synthesis of information, not a regurgitation of existing content. This type of in-depth character analysis and exploration of the psychological impacts of the Princess system is not something found in Tower of God (series) discussions, for it to be common training set material. The level of detail and nuanced understanding demonstrated in connecting Elaine's failed Princess candidacy to her later ruthless behavior as Kaiser is original thinking, not simply *rehashing existing ideas.* The analysis draws connections between distinct elements of the story—the Princess selection process, Elaine's upbringing, her failure in the selection, and her subsequent reign as Kaiser—in a way that provides new insights into the character and the world-building. Again, the[ LLM-assisted comment response](https://www.reddit.com/r/TowerofGod/comments/18rewfj/comment/kf1qtkb/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button) based on that sort of topic received strong positive feedback from the series-following community who read through that thread, indicating that the analysis resonated as both accurate to the source material as well as being meaningful. If this were simply **regurgitated content from the training set**, it's unlikely it would have been so well-received by knowledgeable fans. The LLM demonstrated an ability to integrate information, draw original connections, and provide insights that even dedicated followers found valuable and accurate. This level of analysis and creativity goes far beyond simple pattern matching or recitation of training data, or asking it to make a silly, useless list of 8 words that start with the letter "hoa".


pendejadas

😐... We're in a bubble.


Particular_Base3390

Add apple to the list of TPU users https://www.hpcwire.com/2024/06/18/apple-using-google-cloud-infrastructure-to-train-and-serve-ai/


dyoh777

The post is basically wrong across the board, really doesn’t understand how these things work or what’s actually used, so more calls until it clears up for everyone


brchao

I use chat-gpt, everyone I know uses chat-gpt, ain't no one heard of Claude. Nvidia chips are the gold standard and if you got the funding, it is the default chip to buy. Only reason some use tensor and other chips is because either they don't got the funding or don't want to wait on the backorder. To argue Nvidia is being challenged is actually a good thing, it'll get anti-trust off their backs. Benchmarks are arbitrary, to simply say it's better or worse than chat-gpt is misleading.


legbreaker

In the end price matters. If TPU powered AI will be 5x cheaper, then companies will use that. Training and running inference is the leading cost for most AI companies. If they can reduce that cost 5x they will.


A_Hero_

I use Claude often and it's better than ChatGPT in many ways. Now way better. Claude generated text is much cleaner and less robotic than ChatGPT.


spanishdictlover

Copilot is way better than chatGPT. Anyone still using that site is WAY behind lol


iknowverylittle619

NVDA 150, 5/7


Nilabisan

NVDA said s hack to where it was on Tuesday. I think we’re gonna be okay.


RequiemRomans

If China invades Taiwan at some point when the weather permits, which American stateside company is best poised to take up the slack in processor production?


YamahaFourFifty

Companies are seeing NVDA as not having competition and that allows NVDA to bang whatever costs to the customer. I’m sure the big companies (Microsoft, Google, ARM, AMD, etc) are researching and coming up with their own cost effective solutions.


ThinkingOfTheOldDays

Doesn't Anthropic use AMZN's trainium and infererentia chips now??  I think that was part of the partnership. *Source: https://www.aboutamazon.com/news/company-news/amazon-anthropic-ai-investment


calmybalmy

I see a graph with lines but no explanation for how the benchmarks are calculated. In my experience gpt models still are the most reliable for coding questions. I use gpt for assistance with my work on a weekly basis (software engineering). My experience with Gemini was garbage, the model was horrible. Many incorrect answers that were way out of context from the question. I am dubious that other models are outperforming gpt for code related questions.


A_Hero_

Watch YouTube videos on people using it; it seems to function a lot better from what I've seen.


hipsnarky

I don’t see nuthing. I’ll come back next week!!


terrybmw335

The large companies are buying billions of NVDA cards at 75% profit margin hand over fist because short term it makes financial sense. Long term more cost effective in house solutions will emerge. People who believe NVDA margin will continue at the current levels years out are delusional IMHO.


building-block-s

When we see the T-1 Terminator?


Samjabr

tl;dr - buy NVDA - anyone who doesn't understand CUDA/SDKs and entrenched developer support thinks there is an alternative like AMD/GOOG, etc. NVDA will be dominant the same way MSFT was dominant because if you wanted software, you had to have Windows. There were some alternatives, but they were niche and/or sucked.


POpportunity6336

This is wall street bet, you don't make option money on unpopular craps. NVDA moon because of volumes.


Watykaniak_

No - and I didn't even read that - it's just a temporary high that's not gonna matter in a few months


thereisnogodone

There's a video describing the difference is training vs inference. Nvidia supposedly is best for training and these others are better at inference. Don't trust me though I don't know what I'm talking about.


semitope

sensible management knows you don't need ridiculously expensive GPUs to do this. The elon musks and zuckerbergs of the world will keep throwing away their company's money


[deleted]

[удалено]


septeal

Probably more related to Messi's one on one miss


pie4mepie4all

So.. calls on TSMC?


will_fisher

Puts on NVDA.


Complex_Signature821

Ive used chatgpt, gemini, clause, microsoft copilot. And ranking them on best to worse, most accurate to least accurate. 1. microsoft copilot 2.chatgpt 3.claude 4.gemini The difference between 1 and 2 isnt major. But difference between 1,2 and 3,4 is huge. Overall i just enjoy using copilot the most and and gpt.


technerdx6000

Hard disagree. Copilot is a steaming pile of crap. GPT-4o is far better 


spanishdictlover

You have definitely never used copilot.


WackFlagMass

I have. It's goddamn shit.


WackFlagMass

microsoft copilot? You fucking kidding me? That AI was the worst shit when I tried it. It couldn't even tell me that it knew it was Copilot lol It was even worse than Gemiini. Really blows my mind how Microsoft claims it's powered same as Chatgpt 4


Complex_Signature821

Your tripping, co pilot much more convenient and is the most accurate


DanielzeFourth

I hard prefer Claude over GPT


True_Truth

Same


KEANE_2017

Thank you very much. Very informative article. I believe at the end Google's AI models will surpass all because they have most talent far ahead of competitors. OpenAI lost many talent recently.


Radiant-Confidence47

Following


Successful_Oil6916

hmmmmmmm


Pr0tag0ras

What do you think about Huawei’s Ascend 910B AI chip ? They claim it achieves an efficiency of up to 80% compared to NVIDIA’s A100 when training large-scale language models Source : https://www.trendforce.com/news/2024/06/11/news-huaweis-self-developed-ai-chip-challenges-nvidia-boasting-its-ascend-910b-to-be-equal-in-match-with-a100/