mirror of
https://github.com/infiniflow/ragflow.git
synced 2025-09-26 16:51:23 +00:00
Refa: ollama keep alive issue. (#8216)
### What problem does this PR solve? #8122 ### Type of change - [x] Refactoring
This commit is contained in:
parent
e7c85e569b
commit
d5236b71f4
@ -871,7 +871,7 @@ class OllamaChat(Base):
|
|||||||
ctx_size = self._calculate_dynamic_ctx(history)
|
ctx_size = self._calculate_dynamic_ctx(history)
|
||||||
|
|
||||||
gen_conf["num_ctx"] = ctx_size
|
gen_conf["num_ctx"] = ctx_size
|
||||||
response = self.client.chat(model=self.model_name, messages=history, options=gen_conf)
|
response = self.client.chat(model=self.model_name, messages=history, options=gen_conf, keep_alive=-1)
|
||||||
ans = response["message"]["content"].strip()
|
ans = response["message"]["content"].strip()
|
||||||
token_count = response.get("eval_count", 0) + response.get("prompt_eval_count", 0)
|
token_count = response.get("eval_count", 0) + response.get("prompt_eval_count", 0)
|
||||||
return ans, token_count
|
return ans, token_count
|
||||||
@ -898,7 +898,7 @@ class OllamaChat(Base):
|
|||||||
|
|
||||||
ans = ""
|
ans = ""
|
||||||
try:
|
try:
|
||||||
response = self.client.chat(model=self.model_name, messages=history, stream=True, options=options)
|
response = self.client.chat(model=self.model_name, messages=history, stream=True, options=options, keep_alive=-1)
|
||||||
for resp in response:
|
for resp in response:
|
||||||
if resp["done"]:
|
if resp["done"]:
|
||||||
token_count = resp.get("prompt_eval_count", 0) + resp.get("eval_count", 0)
|
token_count = resp.get("prompt_eval_count", 0) + resp.get("eval_count", 0)
|
||||||
|
@ -506,7 +506,8 @@ class OllamaCV(Base):
|
|||||||
response = self.client.chat(
|
response = self.client.chat(
|
||||||
model=self.model_name,
|
model=self.model_name,
|
||||||
messages=history,
|
messages=history,
|
||||||
options=options
|
options=options,
|
||||||
|
keep_alive=-1
|
||||||
)
|
)
|
||||||
|
|
||||||
ans = response["message"]["content"].strip()
|
ans = response["message"]["content"].strip()
|
||||||
@ -536,7 +537,8 @@ class OllamaCV(Base):
|
|||||||
model=self.model_name,
|
model=self.model_name,
|
||||||
messages=history,
|
messages=history,
|
||||||
stream=True,
|
stream=True,
|
||||||
options=options
|
options=options,
|
||||||
|
keep_alive=-1
|
||||||
)
|
)
|
||||||
for resp in response:
|
for resp in response:
|
||||||
if resp["done"]:
|
if resp["done"]:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user