-
Notifications
You must be signed in to change notification settings - Fork 800
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Tinygrad quantization support #213
base: main
Are you sure you want to change the base?
Conversation
Great work @varshith15 would really like to see this through together with #200 - I want to see how viable it is for interoperability between different inference engines. This would open the door to a lot of really useful workflows like running tinygrad on Qualcomm chips (where it's by far the fastest) and MLX on Apple M chips (assuming it's faster, tinygrad might also be fastest here). |
Is this still in draft? @varshith15 |
@AlexCheema yeah, the tinygrad quantized_mat_mul i wrote is quite slow, debugging that, will need a couple of days |
#148
#200