165 Commits

Author SHA1 Message Date
zrguo
f868c1da78
Merge branch 'main' into code-cleaning 2025-02-16 19:41:05 +08:00
Yannick Stephan
eaf1d553d2 improved typing 2025-02-15 22:37:12 +01:00
Yannick Stephan
621540a54e cleaned code 2025-02-15 00:23:14 +01:00
Yannick Stephan
e6520ad6a2 added typing 2025-02-14 23:49:39 +01:00
Yannick Stephan
66f555677a cleaning the message and project no needed 2025-02-14 23:31:27 +01:00
Pankaj Kaushal
cd81312659 Enhance Neo4j graph storage with error handling and label validation
- Add label existence check and validation methods in Neo4j implementation
- Improve error handling in get_node, get_edge, and upsert methods
- Add default values and logging for missing edge properties
- Ensure consistent label processing across graph storage methods
2025-02-14 16:07:41 +01:00
yangdx
ed73ea4076 Fix linting 2025-02-13 04:12:00 +08:00
yangdx
11c7af7fd8 refactor: use vdb instance's cosine threshold instead of global constant
• Remove global COSINE_THRESHOLD
• Use instance-level threshold config
• Update logging statements
• Reference vdb threshold directly
2025-02-13 03:34:31 +08:00
yangdx
3308ecfa69 Refactor logging for vector similarity search with configurable threshold 2025-02-13 02:14:32 +08:00
zrguo
0c3b754108 Fix bugs 2025-02-11 11:42:46 +08:00
Yannick Stephan
acbe3e2ff2 cleaned code 2025-02-09 13:18:47 +01:00
Yannick Stephan
1f8fc44591 cleaned type 2025-02-09 11:46:01 +01:00
Yannick Stephan
f4287804ce added typing 2025-02-09 10:39:48 +01:00
Yannick Stephan
020fdecc73 cleaned typed not useful 2025-02-08 23:33:11 +01:00
yangdx
5d14ab03eb Fix linting 2025-02-02 01:56:32 +08:00
yangdx
b45ae1567c Refactor LLM cache handling and entity extraction
- Removed custom LLM function in entity extraction
- Simplified cache handling logic
- Added `force_llm_cache` parameter
- Updated cache handling conditions
2025-02-02 01:28:46 +08:00
yangdx
c9481c81b9 Add cache type "extract" for entity extraction 2025-02-01 23:05:02 +08:00
yangdx
2d387fa6de Save keywords to cache only when it's no empty 2025-02-01 22:54:23 +08:00
yangdx
b22a8b216c Fix linting 2025-02-01 10:36:25 +08:00
yangdx
2a010c985e Add logging for chunk truncation in mix_kg_vector_query 2025-02-01 10:36:25 +08:00
yangdx
b0d87b2e29 Fix linting 2025-01-31 15:33:50 +08:00
yangdx
cc50ade14e Fix concurrent problem on extract_entities function.
- Abandon the approach of temporarily replacing the global llm_model_func configuration
- Introduce custom_llm function with new_config for handle_cache while extracting entities
- Update handle_cache to accept custom_llm
2025-01-30 02:45:33 +08:00
yangdx
90c765c724 Fix linting 2025-01-29 22:14:18 +08:00
yangdx
4302c65a04 Avoid unnecessary LLM response generation 2025-01-29 21:01:35 +08:00
yangdx
c423f37f62 Lower log level for keyword outputs 2025-01-29 21:00:42 +08:00
MdNazishArmanShorthillsAI
f0b2024667 Query with your custom prompts 2025-01-27 10:32:22 +05:30
yangdx
86282ba434 Fix history_context handling error in kg_query 2025-01-25 22:54:12 +08:00
yangdx
06673b6fb2 Add logging for high and low-level keywords 2025-01-25 20:11:52 +08:00
yangdx
cf2f83377c fix linting 2025-01-25 16:57:47 +08:00
yangdx
d1969515d3 Add conversation history to naive_query prompt 2025-01-25 16:29:18 +08:00
zrguo
13cc0e662f
Merge pull request #642 from dimatill/main
asyncio optimizations
2025-01-25 01:52:24 +08:00
Dmytro Til
f7b66d2c22 asyncio optimizations 2025-01-24 16:06:04 +01:00
Magic_yuan
8c121f50e9 支持多轮对话 2025-01-24 19:21:03 +08:00
Magic_yuan
5719aa8882 支持多轮对话 2025-01-24 19:03:36 +08:00
zrguo
b84aab5cd0
Merge pull request #590 from jin38324/main
Enhance Robustness of insert Method with Pipeline Processing and Caching Mechanisms
2025-01-16 14:20:08 +08:00
jin
6ae8647285 support pipeline mode 2025-01-16 12:58:15 +08:00
jin
d5ae6669ea support pipeline mode 2025-01-16 12:52:37 +08:00
jin
17a2ec2bc4
Merge branch 'HKUDS:main' into main 2025-01-16 09:59:27 +08:00
Gurjot Singh
bc79f6650e Fix linting errors 2025-01-14 22:23:14 +05:30
Gurjot Singh
ef61ffe444 Add custom function with separate keyword extraction for user's query and a separate prompt 2025-01-14 22:10:47 +05:30
jin
85331e3fa2 update Oracle support
add cache support, fix bug
2025-01-10 11:36:28 +08:00
adikalra
acde4ed173 Add custom chunking function. 2025-01-09 17:20:24 +05:30
zrguo
b93203804c
Merge branch 'main' into main 2025-01-09 15:28:57 +08:00
童石渊
dd213c95be 增加仅字符分割参数,如果开启,仅采用字符分割,不开启,在分割完以后如果chunk过大,会继续根据token size分割,更新测试文件 2025-01-09 11:55:49 +08:00
zrguo
6c78c96854 fix linting errors 2025-01-07 22:02:34 +08:00
zrguo
fe7f7086b1
Merge pull request #547 from n3A87/main
Fix:Optimized logic for automatic switching modes when keywords do not exist
2025-01-07 21:51:51 +08:00
童石渊
6b19401dc6 chunk split retry 2025-01-07 16:26:12 +08:00
童石渊
536d6f2283 添加字符分割功能,在“insert”函数中如果增加参数split_by_character,则会按照split_by_character进行字符分割,此时如果每个分割后的chunk的tokens大于max_token_size,则会继续按token_size分割(todo:考虑字符分割后过短的chunk处理) 2025-01-07 00:28:15 +08:00
xYLiuuuuuu
79646fced8
Fix:Optimized logic for automatic switching modes when keywords do not exist 2025-01-06 16:54:53 +08:00
Samuel Chan
6ae27d8f06 Some enhancements:
- Enable the llm_cache storage to support get_by_mode_and_id, to improve the performance for using real KV server
- Provide an option for the developers to cache the LLM response when extracting entities for a document. Solving the paint point that sometimes the process failed, the processed chunks we need to call LLM again, money and time wasted. With the new option (by default not enabled) enabling, we can cache that result, can significantly save the time and money for beginners.
2025-01-06 12:50:05 +08:00