mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-09-25 16:16:01 +00:00
Refa: ollama keep alive issue. (#8216)
### What problem does this PR solve? #8122 ### Type of change - [x] Refactoring
This commit is contained in:
parent
e7c85e569b
commit
d5236b71f4
@ -871,7 +871,7 @@ class OllamaChat(Base):
|
||||
ctx_size = self._calculate_dynamic_ctx(history)
|
||||
|
||||
gen_conf["num_ctx"] = ctx_size
|
||||
response = self.client.chat(model=self.model_name, messages=history, options=gen_conf)
|
||||
response = self.client.chat(model=self.model_name, messages=history, options=gen_conf, keep_alive=-1)
|
||||
ans = response["message"]["content"].strip()
|
||||
token_count = response.get("eval_count", 0) + response.get("prompt_eval_count", 0)
|
||||
return ans, token_count
|
||||
@ -898,7 +898,7 @@ class OllamaChat(Base):
|
||||
|
||||
ans = ""
|
||||
try:
|
||||
response = self.client.chat(model=self.model_name, messages=history, stream=True, options=options)
|
||||
response = self.client.chat(model=self.model_name, messages=history, stream=True, options=options, keep_alive=-1)
|
||||
for resp in response:
|
||||
if resp["done"]:
|
||||
token_count = resp.get("prompt_eval_count", 0) + resp.get("eval_count", 0)
|
||||
|
@ -506,7 +506,8 @@ class OllamaCV(Base):
|
||||
response = self.client.chat(
|
||||
model=self.model_name,
|
||||
messages=history,
|
||||
options=options
|
||||
options=options,
|
||||
keep_alive=-1
|
||||
)
|
||||
|
||||
ans = response["message"]["content"].strip()
|
||||
@ -536,7 +537,8 @@ class OllamaCV(Base):
|
||||
model=self.model_name,
|
||||
messages=history,
|
||||
stream=True,
|
||||
options=options
|
||||
options=options,
|
||||
keep_alive=-1
|
||||
)
|
||||
for resp in response:
|
||||
if resp["done"]:
|
||||
|
Loading…
x
Reference in New Issue
Block a user