Skip to content

Commit

Permalink
feat: migrate to new session API
Browse files Browse the repository at this point in the history
  • Loading branch information
jhen0409 committed Apr 26, 2024
1 parent f2a4c53 commit 725117c
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 4 deletions.
4 changes: 2 additions & 2 deletions android/src/main/jni.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -231,7 +231,7 @@ Java_com_rnllama_LlamaContext_loadSession(
auto result = createWriteableMap(env);
size_t n_token_count_out = 0;
llama->embd.resize(llama->params.n_ctx);
if (!llama_load_session_file(llama->ctx, path_chars, llama->embd.data(), llama->embd.capacity(), &n_token_count_out)) {
if (!llama_state_load_file(llama->ctx, path_chars, llama->embd.data(), llama->embd.capacity(), &n_token_count_out)) {
env->ReleaseStringUTFChars(path, path_chars);

putString(env, result, "error", "Failed to load session");
Expand Down Expand Up @@ -262,7 +262,7 @@ Java_com_rnllama_LlamaContext_saveSession(
std::vector<llama_token> session_tokens = llama->embd;
int default_size = session_tokens.size();
int save_size = size > 0 && size <= default_size ? size : default_size;
if (!llama_save_session_file(llama->ctx, path_chars, session_tokens.data(), save_size)) {
if (!llama_state_save_file(llama->ctx, path_chars, session_tokens.data(), save_size)) {
env->ReleaseStringUTFChars(path, path_chars);
return -1;
}
Expand Down
4 changes: 2 additions & 2 deletions ios/RNLlamaContext.mm
Original file line number Diff line number Diff line change
Expand Up @@ -390,7 +390,7 @@ - (NSDictionary *)loadSession:(NSString *)path {

size_t n_token_count_out = 0;
llama->embd.resize(llama->params.n_ctx);
if (!llama_load_session_file(llama->ctx, [path UTF8String], llama->embd.data(), llama->embd.capacity(), &n_token_count_out)) {
if (!llama_state_load_file(llama->ctx, [path UTF8String], llama->embd.data(), llama->embd.capacity(), &n_token_count_out)) {
@throw [NSException exceptionWithName:@"LlamaException" reason:@"Failed to load session" userInfo:nil];
}
llama->embd.resize(n_token_count_out);
Expand All @@ -408,7 +408,7 @@ - (int)saveSession:(NSString *)path size:(int)size {
std::vector<llama_token> session_tokens = llama->embd;
int default_size = session_tokens.size();
int save_size = size > 0 && size <= default_size ? size : default_size;
if (!llama_save_session_file(llama->ctx, [path UTF8String], session_tokens.data(), save_size)) {
if (!llama_state_save_file(llama->ctx, [path UTF8String], session_tokens.data(), save_size)) {
@throw [NSException exceptionWithName:@"LlamaException" reason:@"Failed to save session" userInfo:nil];
}
return session_tokens.size();
Expand Down

0 comments on commit 725117c

Please sign in to comment.