mirror of
https://github.com/deepset-ai/haystack.git
synced 2025-09-14 18:52:11 +00:00
fix broken serialization of HFAPI components (#7661)
This commit is contained in:
parent
94467149c1
commit
7c9532b200
@ -175,7 +175,7 @@ class HuggingFaceAPIDocumentEmbedder:
|
||||
"""
|
||||
return default_to_dict(
|
||||
self,
|
||||
api_type=self.api_type,
|
||||
api_type=str(self.api_type),
|
||||
api_params=self.api_params,
|
||||
prefix=self.prefix,
|
||||
suffix=self.suffix,
|
||||
|
@ -142,7 +142,7 @@ class HuggingFaceAPITextEmbedder:
|
||||
"""
|
||||
return default_to_dict(
|
||||
self,
|
||||
api_type=self.api_type,
|
||||
api_type=str(self.api_type),
|
||||
api_params=self.api_params,
|
||||
prefix=self.prefix,
|
||||
suffix=self.suffix,
|
||||
|
@ -158,7 +158,7 @@ class HuggingFaceAPIChatGenerator:
|
||||
callback_name = serialize_callable(self.streaming_callback) if self.streaming_callback else None
|
||||
return default_to_dict(
|
||||
self,
|
||||
api_type=self.api_type,
|
||||
api_type=str(self.api_type),
|
||||
api_params=self.api_params,
|
||||
token=self.token.to_dict() if self.token else None,
|
||||
generation_kwargs=self.generation_kwargs,
|
||||
|
@ -142,7 +142,7 @@ class HuggingFaceAPIGenerator:
|
||||
callback_name = serialize_callable(self.streaming_callback) if self.streaming_callback else None
|
||||
return default_to_dict(
|
||||
self,
|
||||
api_type=self.api_type,
|
||||
api_type=str(self.api_type),
|
||||
api_params=self.api_params,
|
||||
token=self.token.to_dict() if self.token else None,
|
||||
generation_kwargs=self.generation_kwargs,
|
||||
|
@ -0,0 +1,5 @@
|
||||
---
|
||||
fixes:
|
||||
- |
|
||||
Fix the broken serialization of HuggingFaceAPITextEmbedder, HuggingFaceAPIDocumentEmbedder,
|
||||
HuggingFaceAPIGenerator, and HuggingFaceAPIChatGenerator.
|
@ -109,7 +109,7 @@ class TestHuggingFaceAPIDocumentEmbedder:
|
||||
assert data == {
|
||||
"type": "haystack.components.embedders.hugging_face_api_document_embedder.HuggingFaceAPIDocumentEmbedder",
|
||||
"init_parameters": {
|
||||
"api_type": HFEmbeddingAPIType.SERVERLESS_INFERENCE_API,
|
||||
"api_type": "serverless_inference_api",
|
||||
"api_params": {"model": "BAAI/bge-small-en-v1.5"},
|
||||
"token": {"env_vars": ["HF_API_TOKEN"], "strict": False, "type": "env_var"},
|
||||
"prefix": "prefix",
|
||||
|
@ -95,7 +95,7 @@ class TestHuggingFaceAPITextEmbedder:
|
||||
assert data == {
|
||||
"type": "haystack.components.embedders.hugging_face_api_text_embedder.HuggingFaceAPITextEmbedder",
|
||||
"init_parameters": {
|
||||
"api_type": HFEmbeddingAPIType.SERVERLESS_INFERENCE_API,
|
||||
"api_type": "serverless_inference_api",
|
||||
"api_params": {"model": "BAAI/bge-small-en-v1.5"},
|
||||
"token": {"env_vars": ["HF_API_TOKEN"], "strict": False, "type": "env_var"},
|
||||
"prefix": "prefix",
|
||||
|
@ -138,7 +138,7 @@ class TestHuggingFaceAPIGenerator:
|
||||
result = generator.to_dict()
|
||||
init_params = result["init_parameters"]
|
||||
|
||||
assert init_params["api_type"] == HFGenerationAPIType.SERVERLESS_INFERENCE_API
|
||||
assert init_params["api_type"] == "serverless_inference_api"
|
||||
assert init_params["api_params"] == {"model": "HuggingFaceH4/zephyr-7b-beta"}
|
||||
assert init_params["token"] == {"env_vars": ["ENV_VAR"], "strict": False, "type": "env_var"}
|
||||
assert init_params["generation_kwargs"] == {"temperature": 0.6, "stop": ["stop", "words"], "max_tokens": 512}
|
||||
|
@ -131,7 +131,7 @@ class TestHuggingFaceAPIGenerator:
|
||||
result = generator.to_dict()
|
||||
init_params = result["init_parameters"]
|
||||
|
||||
assert init_params["api_type"] == HFGenerationAPIType.SERVERLESS_INFERENCE_API
|
||||
assert init_params["api_type"] == "serverless_inference_api"
|
||||
assert init_params["api_params"] == {"model": "HuggingFaceH4/zephyr-7b-beta"}
|
||||
assert init_params["token"] == {"env_vars": ["ENV_VAR"], "strict": False, "type": "env_var"}
|
||||
assert init_params["generation_kwargs"] == {
|
||||
|
Loading…
x
Reference in New Issue
Block a user