| 
									
										
										
										
											2024-01-02 23:42:00 +08:00
										 |  |  | import os | 
					
						
							| 
									
										
										
										
											2024-04-18 20:24:05 +08:00
										 |  |  | from collections.abc import Generator | 
					
						
							| 
									
										
										
										
											2024-01-02 23:42:00 +08:00
										 |  |  | from time import sleep | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-01-12 12:34:01 +08:00
										 |  |  | import pytest | 
					
						
							| 
									
										
										
										
											2024-04-18 20:24:05 +08:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-01-12 12:34:01 +08:00
										 |  |  | from core.model_runtime.entities.llm_entities import LLMResult, LLMResultChunk, LLMResultChunkDelta | 
					
						
							| 
									
										
										
										
											2024-01-02 23:42:00 +08:00
										 |  |  | from core.model_runtime.entities.message_entities import AssistantPromptMessage, UserPromptMessage | 
					
						
							|  |  |  | from core.model_runtime.entities.model_entities import AIModelEntity | 
					
						
							|  |  |  | from core.model_runtime.errors.validate import CredentialsValidateFailedError | 
					
						
							|  |  |  | from core.model_runtime.model_providers.minimax.llm.llm import MinimaxLargeLanguageModel | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-01-12 12:34:01 +08:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2024-01-02 23:42:00 +08:00
										 |  |  | def test_predefined_models(): | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  |     model_schemas = model.predefined_models() | 
					
						
							|  |  |  |     assert len(model_schemas) >= 1 | 
					
						
							|  |  |  |     assert isinstance(model_schemas[0], AIModelEntity) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def test_validate_credentials_for_chat_model(): | 
					
						
							|  |  |  |     sleep(3) | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     with pytest.raises(CredentialsValidateFailedError): | 
					
						
							|  |  |  |         model.validate_credentials( | 
					
						
							|  |  |  |             model='abab5.5-chat', | 
					
						
							|  |  |  |             credentials={ | 
					
						
							|  |  |  |                 'minimax_api_key': 'invalid_key', | 
					
						
							|  |  |  |                 'minimax_group_id': 'invalid_key' | 
					
						
							|  |  |  |             } | 
					
						
							|  |  |  |         ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     model.validate_credentials( | 
					
						
							|  |  |  |         model='abab5.5-chat', | 
					
						
							|  |  |  |         credentials={ | 
					
						
							|  |  |  |             'minimax_api_key': os.environ.get('MINIMAX_API_KEY'), | 
					
						
							|  |  |  |             'minimax_group_id': os.environ.get('MINIMAX_GROUP_ID') | 
					
						
							|  |  |  |         } | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def test_invoke_model(): | 
					
						
							|  |  |  |     sleep(3) | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     response = model.invoke( | 
					
						
							|  |  |  |         model='abab5-chat', | 
					
						
							|  |  |  |         credentials={ | 
					
						
							|  |  |  |             'minimax_api_key': os.environ.get('MINIMAX_API_KEY'), | 
					
						
							|  |  |  |             'minimax_group_id': os.environ.get('MINIMAX_GROUP_ID') | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         prompt_messages=[ | 
					
						
							|  |  |  |             UserPromptMessage( | 
					
						
							|  |  |  |                 content='Hello World!' | 
					
						
							|  |  |  |             ) | 
					
						
							|  |  |  |         ], | 
					
						
							|  |  |  |         model_parameters={ | 
					
						
							|  |  |  |             'temperature': 0.7, | 
					
						
							|  |  |  |             'top_p': 1.0, | 
					
						
							|  |  |  |             'top_k': 1, | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         stop=['you'], | 
					
						
							|  |  |  |         user="abc-123", | 
					
						
							|  |  |  |         stream=False | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert isinstance(response, LLMResult) | 
					
						
							|  |  |  |     assert len(response.message.content) > 0 | 
					
						
							|  |  |  |     assert response.usage.total_tokens > 0 | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def test_invoke_stream_model(): | 
					
						
							|  |  |  |     sleep(3) | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     response = model.invoke( | 
					
						
							|  |  |  |         model='abab5.5-chat', | 
					
						
							|  |  |  |         credentials={ | 
					
						
							|  |  |  |             'minimax_api_key': os.environ.get('MINIMAX_API_KEY'), | 
					
						
							|  |  |  |             'minimax_group_id': os.environ.get('MINIMAX_GROUP_ID') | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         prompt_messages=[ | 
					
						
							|  |  |  |             UserPromptMessage( | 
					
						
							|  |  |  |                 content='Hello World!' | 
					
						
							|  |  |  |             ) | 
					
						
							|  |  |  |         ], | 
					
						
							|  |  |  |         model_parameters={ | 
					
						
							|  |  |  |             'temperature': 0.7, | 
					
						
							|  |  |  |             'top_p': 1.0, | 
					
						
							|  |  |  |             'top_k': 1, | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         stop=['you'], | 
					
						
							|  |  |  |         stream=True, | 
					
						
							|  |  |  |         user="abc-123" | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert isinstance(response, Generator) | 
					
						
							|  |  |  |     for chunk in response: | 
					
						
							|  |  |  |         assert isinstance(chunk, LLMResultChunk) | 
					
						
							|  |  |  |         assert isinstance(chunk.delta, LLMResultChunkDelta) | 
					
						
							|  |  |  |         assert isinstance(chunk.delta.message, AssistantPromptMessage) | 
					
						
							|  |  |  |         assert len(chunk.delta.message.content) > 0 if chunk.delta.finish_reason is None else True | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def test_invoke_with_search(): | 
					
						
							|  |  |  |     sleep(3) | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     response = model.invoke( | 
					
						
							|  |  |  |         model='abab5.5-chat', | 
					
						
							|  |  |  |         credentials={ | 
					
						
							|  |  |  |             'minimax_api_key': os.environ.get('MINIMAX_API_KEY'), | 
					
						
							|  |  |  |             'minimax_group_id': os.environ.get('MINIMAX_GROUP_ID') | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         prompt_messages=[ | 
					
						
							|  |  |  |             UserPromptMessage( | 
					
						
							|  |  |  |                 content='北京今天的天气怎么样' | 
					
						
							|  |  |  |             ) | 
					
						
							|  |  |  |         ], | 
					
						
							|  |  |  |         model_parameters={ | 
					
						
							|  |  |  |             'temperature': 0.7, | 
					
						
							|  |  |  |             'top_p': 1.0, | 
					
						
							|  |  |  |             'top_k': 1, | 
					
						
							|  |  |  |             'plugin_web_search': True, | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         stop=['you'], | 
					
						
							|  |  |  |         stream=True, | 
					
						
							|  |  |  |         user="abc-123" | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert isinstance(response, Generator) | 
					
						
							|  |  |  |     total_message = '' | 
					
						
							|  |  |  |     for chunk in response: | 
					
						
							|  |  |  |         assert isinstance(chunk, LLMResultChunk) | 
					
						
							|  |  |  |         assert isinstance(chunk.delta, LLMResultChunkDelta) | 
					
						
							|  |  |  |         assert isinstance(chunk.delta.message, AssistantPromptMessage) | 
					
						
							|  |  |  |         total_message += chunk.delta.message.content | 
					
						
							|  |  |  |         assert len(chunk.delta.message.content) > 0 if not chunk.delta.finish_reason else True | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert '参考资料' in total_message | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | def test_get_num_tokens(): | 
					
						
							|  |  |  |     sleep(3) | 
					
						
							|  |  |  |     model = MinimaxLargeLanguageModel() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     response = model.get_num_tokens( | 
					
						
							|  |  |  |         model='abab5.5-chat', | 
					
						
							|  |  |  |         credentials={ | 
					
						
							|  |  |  |             'minimax_api_key': os.environ.get('MINIMAX_API_KEY'), | 
					
						
							|  |  |  |             'minimax_group_id': os.environ.get('MINIMAX_GROUP_ID') | 
					
						
							|  |  |  |         }, | 
					
						
							|  |  |  |         prompt_messages=[ | 
					
						
							|  |  |  |             UserPromptMessage( | 
					
						
							|  |  |  |                 content='Hello World!' | 
					
						
							|  |  |  |             ) | 
					
						
							|  |  |  |         ], | 
					
						
							|  |  |  |         tools=[] | 
					
						
							|  |  |  |     ) | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     assert isinstance(response, int) | 
					
						
							|  |  |  |     assert response == 30 |