AI Assistant
Blog
Pricing
Log In
Sign Up
Fast and Efficient 2-bit LLM Inference on GPU: 2/4/16-bit in a Weight Matrix with Asynchronous Dequantization
Details
Cite
Export
Add to List
The content you want is available to Zendy users.
Already have an account? Click
here.
to sign in.