Replies: 58 comments 156 replies
-
i have tried to run memGPT using the Chat Completion API in LM Studio using a few different models which have been finetuned on function calling datasets. the model rizerphe/CodeLlama-function-calling-6320-7b-Instruct-GGUF, seems like it might work but LM studio throws an error regarding the function calling syntax. ive heard somewhere that falcon-180b can correctly do function calls. i believe this is because falcon-180b-chat is finetuned on the airoboros dataset which features examples of function calling... so perhaps airoboros models will work as well. |
Beta Was this translation helpful? Give feedback.
-
I'd love to see this as well |
Beta Was this translation helpful? Give feedback.
-
In a utopian world someone explains to me how to resolve this issue of function calling within LM Studio by selecting a LLM that actually runs on my machine. Then turn it into an absolute savage set of agents with AutoGen, which is also on the roadmap. Patience is a virtue it seems. |
Beta Was this translation helpful? Give feedback.
-
attempting to run memgpt with the oobabooga openai endpoint gives a step() failed with openai.InvalidRequestError, but didn't recognize the error message: 'functions is not supported.' it would seem that the api endpoint emulator thingie needs to be updated to support openai style function calling https://github.com/oobabooga/text-generation-webui/blob/main/extensions/openai/script.py or perhaps memgpt could be changed to do its function calling in a different format? |
Beta Was this translation helpful? Give feedback.
-
it seems like it should be possible to implement memgpt by replacing the function calling api with open-interpreter and the codelllama 34b. |
Beta Was this translation helpful? Give feedback.
-
What about using LocalAI? If memgpt supports OPENAI_API_HOST/BASE env vars, it should "just work" for the most part, as LocalAI provides a drop-in replacement for OpenAI chat/completion/functions and provides the Go-bindings for Llama.cpp: https://github.com/go-skynet/LocalAI, so can run on just about anything. |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
So it is planed to run MemGPT with own fine tuned GPT 3.5 Models? |
Beta Was this translation helpful? Give feedback.
-
https://huggingface.co/THUDM/agentlm-7b (13b,70b) This project looks promising for local llm instruction tuned LLMs, I wonder if it could mesh into your testing? How goes fine tuning a Mistral 7b, Im very excited at the prospect of a local model being able to perform tasks that are dependent on AutoGPT API Access... Local data handling by internal AI llms is very appealing to some. :) |
Beta Was this translation helpful? Give feedback.
-
⭐ We've added support for running MemGPT with local LLMs!You can find instructions on how to set this up on this README. The tl;dr is:
We include an example for running MemGPT with Airoboros - we include example function call parsers so you can just set up Airoboros behind WebUI and it should "work" (we did some very limited testing, and it can make basic calls to Things we're still working on (welcoming contributions!):
This is just a start and we're excited to work with the community to make MemGPT with open models a viable alternative to GPT-4 🚀 |
Beta Was this translation helpful? Give feedback.
-
Maybe integrate AutoGPT with MemGPT? Or making a fork of it and refactor to it to work only as a function caller/parser and extension functionality. They have a pretty decent features out there, along with api integrations and hooks/callbacks (i.e. module integration) |
Beta Was this translation helpful? Give feedback.
-
Also they are using |
Beta Was this translation helpful? Give feedback.
-
Please consider prioritizing the wrapper for LiteLLM https://github.com/BerriAI/litellm |
Beta Was this translation helpful? Give feedback.
-
i am interested in fine tuning a model specifically for memGPT though i am not exactly sure of the best way to do this. it seems that we would need to gather raw communication logs from a variety of memGPT use cases, probably using the gpt4-32k model? |
Beta Was this translation helpful? Give feedback.
-
Currently testing Zephyr 7B beta, The problem i am getting is sometimes the model fails to realize that the inner thought part are supposed not present in the conversation: ? for example like this: As you can see here, after i asked whether i could give them another name, the agent should have give me a yes/no answer, which it did in the inner thought session, but failed to say it out loud, but instead they just continue their inner thought. |
Beta Was this translation helpful? Give feedback.
-
❯ python ../tests/test_cli.py |
Beta Was this translation helpful? Give feedback.
-
Since MemGPT has been severely upgraded, could you upgrade also this guide to show what are the correct steps for the env vars, for memgpt configure and for memgpt run ? Many thanks |
Beta Was this translation helpful? Give feedback.
-
I'm giving up. Maybe I can look at this again in 6 months? |
Beta Was this translation helpful? Give feedback.
-
At this point, what is the most feature rich and reliable back-end to use? LM Studio is not an option for me since it's closed source. |
Beta Was this translation helpful? Give feedback.
-
Having unexpected troubles with Backend: Ollama 1.8 What happened:
Checking ollama server log I've out that ollama apparently run out of memory More than that: it looks like MemGpt (or at least this current instance) is
Ollama server log, same-ish every time I run memgpt
|
Beta Was this translation helpful? Give feedback.
-
the agentlm 70b 4_k_m gguf model works great with memgpt also this dataset might be useful for training the memgpt model https://github.com/THUDM/AgentTuning |
Beta Was this translation helpful? Give feedback.
-
This just got released yesterday, I'm going to see how it holds up. It's supposed to be specific for function calling which I could see helping a lot. |
Beta Was this translation helpful? Give feedback.
-
A quick question: Does anyone have a colab notebook on how to run this with a local llm without autogen? I'm using ollama and litellm, but any other setup that calls a local endpoint would be much appreciated. Thank you. |
Beta Was this translation helpful? Give feedback.
-
Is there a way to get this to work on vllm? I attempted this but was unsuccessful. |
Beta Was this translation helpful? Give feedback.
-
NexusRaven-V2-13B (https://huggingface.co/Nexusflow/NexusRaven-V2-13B) new model just dropped today. idk how i missed the first one but anyway version two works decently with current version of MemGPT with default settings. what they did is very impressive and i think memGPT could benefit alot from this. they have a slightly different prompting style and syntax and i think it would perform much better with some modifications to the memgpt code. on a sorta related note; ive always thought that writing prompts with python style indentation would somehow help since most models are familiar with that. |
Beta Was this translation helpful? Give feedback.
-
I think |
Beta Was this translation helpful? Give feedback.
-
Yeah, i have also headed back here to check whether MemGPT is already working fine with mixtral-7x8 and others fine-tuned versions of it. I think it's ready for MemGPT function calling requirements. |
Beta Was this translation helpful? Give feedback.
-
Can we get support for TabbyAPI in the future? This is a local backend made by the author of exl2 https://github.com/theroyallab/tabbyAPI |
Beta Was this translation helpful? Give feedback.
-
If anyone is looking for the documentation then the correct link is https://memgpt.readme.io/docs/local_llm |
Beta Was this translation helpful? Give feedback.
-
⭐ We've added support for running MemGPT with open/local LLMs!
Instructions on how to connect MemGPT to open/local LLMs can be found on our docs page.
🙋 Need help with local LLMs? Check Discord!
If you need help visit our Discord server and post in the #support channel.
You can also use this GitHub discussions page, but the Discord server is the official support channel and is monitored more actively.
To help us (and the entire MemGPT community) help you, please provide the following information when asking a new question about debugging a local model:
dolphin-2.1-mistral-7b.Q6_K.gguf
(not justdolphin-2.1-mistral
)Managing memory in MemGPT requires a lot of instruction following (the LLM needs to follow instructions in the system prompt on how to use the memory functions). GPT-4 can do this well, but even the best open LLMs may struggle to do this correctly, so you will likely observe MemGPT + open LLMs not working very well. This problem gets worse as the LLM gets worse, eg if you're trying a small quantized llama2 model, expect MemGPT to perform very poorly.
If your model outputs bad function calls / bad JSON, things will fail. Even if the model outputs good JSON, if you don't parse it correctly, it will also fail.
Beta Was this translation helpful? Give feedback.
All reactions