File tree Expand file tree Collapse file tree 1 file changed +3
-1
lines changed Expand file tree Collapse file tree 1 file changed +3
-1
lines changed Original file line number Diff line number Diff line change @@ -2082,10 +2082,13 @@ struct server_context {
2082
2082
2083
2083
// keep only the common part
2084
2084
int p0 = slot.n_past ;
2085
+
2085
2086
if (!llama_kv_cache_seq_rm (ctx, slot.id + 1 , p0, -1 )) {
2086
2087
// could not partially delete (likely using a non-Transformer model)
2087
2088
llama_kv_cache_seq_rm (ctx, slot.id + 1 , -1 , -1 );
2088
2089
2090
+ p0 = 0 ;
2091
+
2089
2092
// there is no common part left
2090
2093
slot.n_past = 0 ;
2091
2094
slot.n_past_se = 0 ;
@@ -2773,7 +2776,6 @@ int main(int argc, char ** argv) {
2773
2776
2774
2777
const auto handle_props = [&ctx_server, &res_ok](const httplib::Request &, httplib::Response & res) {
2775
2778
json data = {
2776
- { " system_prompt" , " [unavailable]" },
2777
2779
{ " default_generation_settings" , ctx_server.default_generation_settings_for_props },
2778
2780
{ " total_slots" , ctx_server.params .n_parallel },
2779
2781
{ " chat_template" , llama_get_chat_template (ctx_server.model ) },
You can’t perform that action at this time.
0 commit comments