I first got into deep learning in 2012, when AlexNet came out. I was CTO of Jetpac, a startup that aimed to provide information about bars, hotels, and restaurants by analyzing public photos, for e…
He’s also making some false equivalencies. For instance, his startup (Moonshine) seems to be a plain STT dev… STT is dirt cheap and dime-a-dozen these days. Not to speak of more advanced speech ingestion coming out daily, like https://huggingface.co/stepfun-ai/Step-Audio-R1
He’s also assuming the same level ‘unpicked’ optimization exists as we had in 2012, with Fermi GPUs and barebones software support. It does not. Look at frontier stuff, like bitnet models with sparsity and CPU LUT kernels, and there is just not much theoretical room for dramatic optimization there.
I don’t mean to refute his general point; his sentiment on what investors are (stupidly) doing is spot on. But also, it’s not 2012.
+1
He’s also making some false equivalencies. For instance, his startup (Moonshine) seems to be a plain STT dev… STT is dirt cheap and dime-a-dozen these days. Not to speak of more advanced speech ingestion coming out daily, like https://huggingface.co/stepfun-ai/Step-Audio-R1
He’s also assuming the same level ‘unpicked’ optimization exists as we had in 2012, with Fermi GPUs and barebones software support. It does not. Look at frontier stuff, like bitnet models with sparsity and CPU LUT kernels, and there is just not much theoretical room for dramatic optimization there.
I don’t mean to refute his general point; his sentiment on what investors are (stupidly) doing is spot on. But also, it’s not 2012.