File tree Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Expand file tree Collapse file tree 2 files changed +4
-4
lines changed Original file line number Diff line number Diff line change 11
2- LLAMA_VERSION? =4abef75f2cf2eee75eb5083b30a94cf981587394
2+ LLAMA_VERSION? =d82b7a7c1d73c0674698d9601b1bbb0200933f29
33LLAMA_REPO? =https://github.com/ggerganov/llama.cpp
44
55CMAKE_ARGS? =
Original file line number Diff line number Diff line change @@ -1199,7 +1199,7 @@ class BackendServiceImpl final : public backend::Backend::Service {
11991199 task.id_slot = json_value (data, " id_slot" , -1 );
12001200
12011201 // OAI-compat
1202- task.params .oaicompat = OAICOMPAT_TYPE_NONE ;
1202+ task.params .res_type = TASK_RESPONSE_TYPE_NONE ;
12031203 task.params .oaicompat_cmpl_id = completion_id;
12041204 // oaicompat_model is already populated by params_from_json_cmpl
12051205
@@ -1930,7 +1930,7 @@ class BackendServiceImpl final : public backend::Backend::Service {
19301930 task.id_slot = json_value (data, " id_slot" , -1 );
19311931
19321932 // OAI-compat
1933- task.params .oaicompat = OAICOMPAT_TYPE_NONE ;
1933+ task.params .res_type = TASK_RESPONSE_TYPE_NONE ;
19341934 task.params .oaicompat_cmpl_id = completion_id;
19351935 // oaicompat_model is already populated by params_from_json_cmpl
19361936
@@ -2056,7 +2056,7 @@ class BackendServiceImpl final : public backend::Backend::Service {
20562056 task.index = i;
20572057 task.tokens = std::move (tokenized_prompts[i]);
20582058
2059- task.params .oaicompat = OAICOMPAT_TYPE_NONE ;
2059+ task.params .res_type = TASK_RESPONSE_TYPE_NONE ;
20602060 task.params .embd_normalize = embd_normalize;
20612061 tasks.push_back (std::move (task));
20622062 }
You can’t perform that action at this time.
0 commit comments