zrguo
f868c1da78
Merge branch 'main' into code-cleaning
2025-02-16 19:41:05 +08:00
Yannick Stephan
eaf1d553d2
improved typing
2025-02-15 22:37:12 +01:00
Yannick Stephan
621540a54e
cleaned code
2025-02-15 00:23:14 +01:00
Yannick Stephan
e6520ad6a2
added typing
2025-02-14 23:49:39 +01:00
Yannick Stephan
66f555677a
cleaning the message and project no needed
2025-02-14 23:31:27 +01:00
Pankaj Kaushal
cd81312659
Enhance Neo4j graph storage with error handling and label validation
...
- Add label existence check and validation methods in Neo4j implementation
- Improve error handling in get_node, get_edge, and upsert methods
- Add default values and logging for missing edge properties
- Ensure consistent label processing across graph storage methods
2025-02-14 16:07:41 +01:00
yangdx
ed73ea4076
Fix linting
2025-02-13 04:12:00 +08:00
yangdx
11c7af7fd8
refactor: use vdb instance's cosine threshold instead of global constant
...
• Remove global COSINE_THRESHOLD
• Use instance-level threshold config
• Update logging statements
• Reference vdb threshold directly
2025-02-13 03:34:31 +08:00
yangdx
3308ecfa69
Refactor logging for vector similarity search with configurable threshold
2025-02-13 02:14:32 +08:00
zrguo
0c3b754108
Fix bugs
2025-02-11 11:42:46 +08:00
Yannick Stephan
acbe3e2ff2
cleaned code
2025-02-09 13:18:47 +01:00
Yannick Stephan
1f8fc44591
cleaned type
2025-02-09 11:46:01 +01:00
Yannick Stephan
f4287804ce
added typing
2025-02-09 10:39:48 +01:00
Yannick Stephan
020fdecc73
cleaned typed not useful
2025-02-08 23:33:11 +01:00
yangdx
5d14ab03eb
Fix linting
2025-02-02 01:56:32 +08:00
yangdx
b45ae1567c
Refactor LLM cache handling and entity extraction
...
- Removed custom LLM function in entity extraction
- Simplified cache handling logic
- Added `force_llm_cache` parameter
- Updated cache handling conditions
2025-02-02 01:28:46 +08:00
yangdx
c9481c81b9
Add cache type "extract" for entity extraction
2025-02-01 23:05:02 +08:00
yangdx
2d387fa6de
Save keywords to cache only when it's no empty
2025-02-01 22:54:23 +08:00
yangdx
b22a8b216c
Fix linting
2025-02-01 10:36:25 +08:00
yangdx
2a010c985e
Add logging for chunk truncation in mix_kg_vector_query
2025-02-01 10:36:25 +08:00
yangdx
b0d87b2e29
Fix linting
2025-01-31 15:33:50 +08:00
yangdx
cc50ade14e
Fix concurrent problem on extract_entities function.
...
- Abandon the approach of temporarily replacing the global llm_model_func configuration
- Introduce custom_llm function with new_config for handle_cache while extracting entities
- Update handle_cache to accept custom_llm
2025-01-30 02:45:33 +08:00
yangdx
90c765c724
Fix linting
2025-01-29 22:14:18 +08:00
yangdx
4302c65a04
Avoid unnecessary LLM response generation
2025-01-29 21:01:35 +08:00
yangdx
c423f37f62
Lower log level for keyword outputs
2025-01-29 21:00:42 +08:00
MdNazishArmanShorthillsAI
f0b2024667
Query with your custom prompts
2025-01-27 10:32:22 +05:30
yangdx
86282ba434
Fix history_context handling error in kg_query
2025-01-25 22:54:12 +08:00
yangdx
06673b6fb2
Add logging for high and low-level keywords
2025-01-25 20:11:52 +08:00
yangdx
cf2f83377c
fix linting
2025-01-25 16:57:47 +08:00
yangdx
d1969515d3
Add conversation history to naive_query prompt
2025-01-25 16:29:18 +08:00
zrguo
13cc0e662f
Merge pull request #642 from dimatill/main
...
asyncio optimizations
2025-01-25 01:52:24 +08:00
Dmytro Til
f7b66d2c22
asyncio optimizations
2025-01-24 16:06:04 +01:00
Magic_yuan
8c121f50e9
支持多轮对话
2025-01-24 19:21:03 +08:00
Magic_yuan
5719aa8882
支持多轮对话
2025-01-24 19:03:36 +08:00
zrguo
b84aab5cd0
Merge pull request #590 from jin38324/main
...
Enhance Robustness of insert Method with Pipeline Processing and Caching Mechanisms
2025-01-16 14:20:08 +08:00
jin
6ae8647285
support pipeline mode
2025-01-16 12:58:15 +08:00
jin
d5ae6669ea
support pipeline mode
2025-01-16 12:52:37 +08:00
jin
17a2ec2bc4
Merge branch 'HKUDS:main' into main
2025-01-16 09:59:27 +08:00
Gurjot Singh
bc79f6650e
Fix linting errors
2025-01-14 22:23:14 +05:30
Gurjot Singh
ef61ffe444
Add custom function with separate keyword extraction for user's query and a separate prompt
2025-01-14 22:10:47 +05:30
jin
85331e3fa2
update Oracle support
...
add cache support, fix bug
2025-01-10 11:36:28 +08:00
adikalra
acde4ed173
Add custom chunking function.
2025-01-09 17:20:24 +05:30
zrguo
b93203804c
Merge branch 'main' into main
2025-01-09 15:28:57 +08:00
童石渊
dd213c95be
增加仅字符分割参数,如果开启,仅采用字符分割,不开启,在分割完以后如果chunk过大,会继续根据token size分割,更新测试文件
2025-01-09 11:55:49 +08:00
zrguo
6c78c96854
fix linting errors
2025-01-07 22:02:34 +08:00
zrguo
fe7f7086b1
Merge pull request #547 from n3A87/main
...
Fix:Optimized logic for automatic switching modes when keywords do not exist
2025-01-07 21:51:51 +08:00
童石渊
6b19401dc6
chunk split retry
2025-01-07 16:26:12 +08:00
童石渊
536d6f2283
添加字符分割功能,在“insert”函数中如果增加参数split_by_character,则会按照split_by_character进行字符分割,此时如果每个分割后的chunk的tokens大于max_token_size,则会继续按token_size分割(todo:考虑字符分割后过短的chunk处理)
2025-01-07 00:28:15 +08:00
xYLiuuuuuu
79646fced8
Fix:Optimized logic for automatic switching modes when keywords do not exist
2025-01-06 16:54:53 +08:00
Samuel Chan
6ae27d8f06
Some enhancements:
...
- Enable the llm_cache storage to support get_by_mode_and_id, to improve the performance for using real KV server
- Provide an option for the developers to cache the LLM response when extracting entities for a document. Solving the paint point that sometimes the process failed, the processed chunks we need to call LLM again, money and time wasted. With the new option (by default not enabled) enabling, we can cache that result, can significantly save the time and money for beginners.
2025-01-06 12:50:05 +08:00