How to convert Llama2 and Mistral models to TFlite ? #1428
Unanswered
RageshAntonyHM
asked this question in
Q&A
Replies: 1 comment 2 replies
-
I am interested in the same thing. Did you get any luck with Llama2 with tflite? Not sure whether it will be possible though with a 6 GB ram device. Maybe the solution is to go with llama.cpp? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I saw that it is possible to run GPT2 in Android using KerasNLP
https://www.tensorflow.org/lite/examples/auto_complete/overview
I am eager to run Llama 2 or Mistral in the same way
I am currently using MLC-LLM and also interested other ways.
I saw this Colab to convert but it is only in GPT2
https://colab.research.google.com/github/tensorflow/codelabs/blob/main/KerasNLP/io2023_workshop.ipynb#scrollTo=2fux7O64KHQ6 (remove @googleio2023 branch in pip install)
But the KerasNLP also has Mistral and LLama as per the repo https://github.com/keras-team/keras-nlp/tree/master/keras_nlp/models/
But I unable to find how to convert these models to Tflite.
I am asking what is the equivalent for this block in Mistral and Llama.
Please help me
Beta Was this translation helpful? Give feedback.
All reactions