Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tinygrad quantization support #213

Draft
wants to merge 10 commits into
base: main
Choose a base branch
from

Conversation

varshith15
Copy link
Contributor

@varshith15 varshith15 commented Sep 10, 2024

@AlexCheema
Copy link
Contributor

Great work @varshith15 would really like to see this through together with #200 - I want to see how viable it is for interoperability between different inference engines. This would open the door to a lot of really useful workflows like running tinygrad on Qualcomm chips (where it's by far the fastest) and MLX on Apple M chips (assuming it's faster, tinygrad might also be fastest here).

@AlexCheema
Copy link
Contributor

Is this still in draft? @varshith15

@varshith15
Copy link
Contributor Author

@AlexCheema yeah, the tinygrad quantized_mat_mul i wrote is quite slow, debugging that, will need a couple of days

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants