site banner

Small-Scale Question Sunday for March 2, 2025

Do you have a dumb question that you're kind of embarrassed to ask in the main thread? Is there something you're just not sure about?

This is your opportunity to ask questions. No question too simple or too silly.

Culture war topics are accepted, and proposals for a better intro post are appreciated.

1
Jump in the discussion.

No email address required.

I can only hope that they're simply strapped for GPUs, especially for inference, and are using the bulk of their compute on the 4.0 models they're cooking.

This is somewhat unlikely. The GPUs that you need for training cost a fortune (or rather, NVidia can charge a fortune for them since they have almost zero large scale competition) while much cheaper ones can be good enough for inference.