WalnutLum@lemmy.mltoTechnology@lemmy.world•OpenAI, Google, Anthropic admit they can’t scale up their chatbots any furtherEnglish
3·
2 days agoSeeing as how the full unquantized FP16 for Llama 3.1 405B requires around a terabyte of VRAM (16 bits per parameter + context), I’d say way more than several.
Pages that don’t work well with darkreader aren’t many, but when they do fail they fail somewhat entertainingly: