Replies: 27 comments
-
might be something to due with how web browser handle RAM management |
Beta Was this translation helpful? Give feedback.
-
It could be for WASM, cause I tried some llama-cpp built (current one), but my maximum was around 300 MB. Would be nice if you could help me with building it or maybe some WebGPI solution |
Beta Was this translation helpful? Give feedback.
-
hmmm might be a limitation of wasm itself... I could help |
Beta Was this translation helpful? Give feedback.
-
quick thought: |
Beta Was this translation helpful? Give feedback.
-
Are you referring to the llama-cpp-wasm? What do you mean by split it up?
Yeah, I'm already working on it (check the
How? All the build files included (and simplified) in Yuna are from this repo: https://github.com/tangledgroup/llama-cpp-wasm. It's not using WebGPU, so I have no idea. (There's also a
|
Beta Was this translation helpful? Give feedback.
-
I'm using Firefox since chromium doesn't work |
Beta Was this translation helpful? Give feedback.
-
and I just realized the base model is 4.5gb 🤦♂️ |
Beta Was this translation helpful? Give feedback.
-
I could make it to the ~3.2 GB to fit the WASM limit of 4 GB! |
Beta Was this translation helpful? Give feedback.
-
it might work then... I'll try it |
Beta Was this translation helpful? Give feedback.
-
Sure! Let me know if you'll figure it out. I'll try to modify stuff on the HF today, and you can probably get a more quantized model before tomorrow morning 👍🏻 |
Beta Was this translation helpful? Give feedback.
-
sure... I'll try getting one |
Beta Was this translation helpful? Give feedback.
-
Updated model link: https://huggingface.co/yukiarimo/yuna-ai-v1 |
Beta Was this translation helpful? Give feedback.
-
I'll get back to you once I get home... I'm at school rn and they blocked huggingface |
Beta Was this translation helpful? Give feedback.
-
Sure thing! Lol, is ChatGPT or Perplexity also blocked in your school? |
Beta Was this translation helpful? Give feedback.
-
yep... classified as ai |
Beta Was this translation helpful? Give feedback.
-
I've been home for awhile so let me get the model... I'll test it |
Beta Was this translation helpful? Give feedback.
-
Sure, you can grab any (better q5) from the HF repo above. By the way, I'm also starting training V2 in a few days, so keep updated (150k+ tokens)! |
Beta Was this translation helpful? Give feedback.
-
im using the light model... but i do want to see if the heavy version works too |
Beta Was this translation helpful? Give feedback.
-
Are you doing a light model in WASM? Where? Which model? |
Beta Was this translation helpful? Give feedback.
-
the model is "yuna-ai-v1-q3_k_m.gguf" |
Beta Was this translation helpful? Give feedback.
-
Is it working in WASM? How exactly did you try? |
Beta Was this translation helpful? Give feedback.
-
I tried enabling the setting... and then nothing happened (probably since I was using the pi to try and chat) |
Beta Was this translation helpful? Give feedback.
-
I'll try my phone today |
Beta Was this translation helpful? Give feedback.
-
Sure. And don’t forget to check the console for logs! (because I was lazy to implement popup-errors) |
Beta Was this translation helpful? Give feedback.
-
Everyone! I think, I'll convert this issue into a discussion! |
Beta Was this translation helpful? Give feedback.
-
took a small break: I'm going try my phone again (my phone was dead and I forgot) |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
All reactions