Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The context length exceeds the max token of api #16

Open
Jarvis-K opened this issue Jan 6, 2024 · 2 comments · Fixed by #8
Open

The context length exceeds the max token of api #16

Jarvis-K opened this issue Jan 6, 2024 · 2 comments · Fixed by #8
Assignees

Comments

@Jarvis-K
Copy link
Collaborator

Jarvis-K commented Jan 6, 2024

No description provided.

@Jarvis-K Jarvis-K self-assigned this Jan 6, 2024
@Jarvis-K Jarvis-K linked a pull request Jan 22, 2024 that will close this issue
@Jarvis-K
Copy link
Collaborator Author

solved by auto truncating

@ewanlee
Copy link
Member

ewanlee commented Feb 16, 2024

A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts

https://papers.cool/arxiv/2402.09727

Authors: Kuang-Huei Lee ; Xinyun Chen ; Hiroki Furuta ; John Canny ; Ian Fischer

Summary: Current Large Language Models (LLMs) are not only limited to some maximum context length, but also are not able to robustly consume long inputs. To address these limitations, we propose ReadAgent, an LLM agent system that increases effective context length up to 20x in our experiments. Inspired by how humans interactively read long documents, we implement ReadAgent as a simple prompting system that uses the advanced language capabilities of LLMs to (1) decide what content to store together in a memory episode, (2) compress those memory episodes into short episodic memories called gist memories, and (3) take actions to look up passages in the original text if ReadAgent needs to remind itself of relevant details to complete a task. We evaluate ReadAgent against baselines using retrieval methods, using the original long contexts, and using the gist memories. These evaluations are performed on three long-document reading comprehension tasks: QuALITY, NarrativeQA, and QMSum. ReadAgent outperforms the baselines on all three tasks while extending the effective context window by 3-20x.

@ewanlee ewanlee reopened this Feb 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants