Jump to content

NVidia's Breakthrough AI Chip Defies Physics

6 hours ago, Quackers101 said:

We actually have models using parameters with less than 2 bits (look into Microsoft's 1.5-something bit LLM), so FP4 is going to be really useful.

If you take an existing model, then you can quantize it to a smaller size with a minor perf loss, or you can train a model from scratch using such small parameters.

The docs you linked are old and are meant for automatic use of the tensor cores when they first came out. We're on cuda 12 nowadays and most ml frameworks allow you to do much more than that聽馃槢

FX6300 @ 4.2GHz | Gigabyte GA-78LMT-USB3 R2 | Hyper 212x | 3x 8GB + 1x 4GB @ 1600MHz | Gigabyte 2060 Super | Corsair CX650M | LG 43UK6520PSA
ASUS X550LN | i5 4210u | 12GB
Lenovo N23 Yoga

Link to comment
Share on other sites

Link to post
Share on other sites

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now