3 Comments
User's avatar
Jared Rand's avatar

Nice to hear an update from you Ryan. I remember back in 2017 when you were so bullish on Nvidia and convinced me to buy. I held it until last year (sold too soon) because they’re wildly overpriced at this point. Now I own puts on Nvidias. The reasons are exactly as you stated: competition from ASICs for LLMs and small fine tuned LLMs driving more business value than mega models like GPT4.

Expand full comment
Ryan's avatar

Jared! So good to hear from you. Hope you're doing well. Glad you held on to it ;)

Yeah, something has to give here and these new generation of semiconductors are easier to manufacturing, cheaper to build, better performance, etc.

The thing that will take time is the programming language. CUDA gives them such a great moat but I believe that's why each of these neruomorphic companies are going fully vertical (software/hardware). I suppose we'll see. Perhaps we get a black swan event where we're forced to make the leap in architecture (eg. China invades Taiwan; rapid supply chain shift).

Great to hear from you dude! We should catch up soon, do a collaborative post/podcast. Would love to hear your perspective on all the craziness happening.

Expand full comment
Jared Rand's avatar

Agreed, CUDA is a huge moat and a huge reason other hardware isn't used widely. Just today my teammate reported on how he is blocked on trying to convert our model to run on AWS' custom hardware called Inferentia. We've been overpaying by 4x for years using GPUs instead of Inferentia but are stuck due to the software hurdles of "compiling" our model for custom hardware. But we're a small team operating at a relatively small scale without the expertise to do this - bigger companies with bigger scale will have the motivation and know how to use custom hardware.

Expand full comment