Pro@programming.dev to Programming@programming.devEnglish · 1 month agoSurprisingly Fast AI-Generated Kernels We Didn’t Mean to Publish (Yet)crfm.stanford.eduexternal-linkmessage-square5fedilinkarrow-up123arrow-down16cross-posted to: hackernews
arrow-up117arrow-down1external-linkSurprisingly Fast AI-Generated Kernels We Didn’t Mean to Publish (Yet)crfm.stanford.eduPro@programming.dev to Programming@programming.devEnglish · 1 month agomessage-square5fedilinkcross-posted to: hackernews
minus-squareSpicyToaster420@sopuli.xyzlinkfedilinkarrow-up4·1 month agoAwesome use of LLMs. I wonder they didn’t use FP8 quantization though, especially since their target hardware was an L40s.
Awesome use of LLMs. I wonder they didn’t use FP8 quantization though, especially since their target hardware was an L40s.