yuanxiaobin
|
ad4b0d1ba9
|
Merge remote-tracking branch 'origin/main'
# Conflicts:
# lightrag/llm.py
# lightrag/operate.py
|
2024-12-06 15:06:00 +08:00 |
|
yuanxiaobin
|
584258078f
|
重构缓存处理逻辑
- 提取通用缓存处理逻辑到新函数 handle_cache 和 save_to_cache
- 使用 CacheData 类统一缓存数据结构
- 优化嵌入式缓存和常规缓存的处理流程
- 添加模式参数以支持不同查询模式的缓存策略
- 重构 get_best_cached_response 函数,提高缓存查询效率
|
2024-12-06 14:29:16 +08:00 |
|
zrguo
|
ad991f904d
|
Merge branch 'main' into main
|
2024-12-06 11:38:27 +08:00 |
|
yuanxiaobin
|
7c4bbe2474
|
修复 args_hash在使用常规缓存时候才计算导致embedding缓存时没有计算的bug
|
2024-12-06 10:40:48 +08:00 |
|
yuanxiaobin
|
8a69604966
|
修复 args_hash在使用常规缓存时候才计算导致embedding缓存时没有计算的bug
|
2024-12-06 10:28:35 +08:00 |
|
yuanxiaobin
|
f2a1897b61
|
修复 args_hash在使用常规缓存时候才计算导致embedding缓存时没有计算的bug
|
2024-12-06 10:21:53 +08:00 |
|
partoneplay
|
335179196a
|
Add support for Ollama streaming output and integrate Open-WebUI as the chat UI demo
|
2024-12-06 10:13:16 +08:00 |
|
yuanxiaobin
|
525c971a23
|
feat(lightrag): 添加 查询时使用embedding缓存功能
- 在 LightRAG 类中添加 embedding_cache_config配置项
- 实现基于 embedding 相似度的缓存查询和存储
- 添加量化和反量化函数,用于压缩 embedding 数据
- 新增示例演示 embedding 缓存的使用
|
2024-12-06 08:17:20 +08:00 |
|
Larfii
|
da73ba9b6b
|
fix: unexpected keyword argument error
|
2024-12-05 11:47:56 +08:00 |
|
LarFii
|
db9b9f69f8
|
update insert custom kg
|
2024-12-04 19:44:04 +08:00 |
|
zrguo
|
f3ae4fccfa
|
Merge pull request #383 from MRX760/main
added nvidia text-embedding API and example of using nvidia API llm a…
|
2024-12-04 11:21:46 +08:00 |
|
MRX760
|
0b87e4649f
|
added nvidia text-embedding API and example of using nvidia API llm and text-embedding
|
2024-12-03 17:15:10 +07:00 |
|
partoneplay
|
18e86a1825
|
embedding deprecated in favor of embed
|
2024-12-03 08:42:36 +08:00 |
|
zrguo
|
1420669891
|
Merge pull request #360 from ahmadhatahet/azure_openai_embedding
Azure OpenAI Embedding
|
2024-12-02 16:02:13 +08:00 |
|
Ahmad Hatahet
|
137315ec18
|
add api_version to azure_openai_complete_if_cache
|
2024-11-30 17:47:33 +01:00 |
|
Ahmad Hatahet
|
9b92d425f6
|
update max_token_size according to openai doc
|
2024-11-30 17:16:07 +01:00 |
|
Ahmad Hatahet
|
38c8a6e97c
|
add api_version to args
|
2024-11-30 17:11:38 +01:00 |
|
b10902118
|
2e95c633cf
|
fix error and tested
|
2024-11-30 00:00:51 +08:00 |
|
b10902118
|
b0dd600429
|
support JSON output for ollama and openai
|
2024-11-29 21:41:37 +08:00 |
|
jin
|
5bde05ed53
|
add LightRAG init parameters in readme
also fix some error
|
2024-11-26 10:19:28 +08:00 |
|
jin
|
21f161390a
|
Logic Optimization
|
2024-11-25 13:40:38 +08:00 |
|
jin
|
bf5815be8f
|
Merge branch 'main' of https://github.com/jin38324/LightRAG
|
2024-11-25 13:32:33 +08:00 |
|
jin
|
af3aef5d88
|
Optimization logic
|
2024-11-25 13:29:55 +08:00 |
|
david
|
38e1956395
|
fix hf embedding to support loading to different device
|
2024-11-13 14:20:36 +08:00 |
|
Ken Wiltshire
|
ac0b7fd775
|
fix event loop conflict
|
2024-11-06 11:18:14 -05:00 |
|
Ken Wiltshire
|
f19af82db1
|
using neo4j async
|
2024-11-02 18:35:07 -04:00 |
|
Andrii Lazarchuk
|
478998bc27
|
Merge branch 'main' into before-sync-28-10-2024
|
2024-10-28 15:54:43 +02:00 |
|
tackhwa
|
81d5b904fb
|
update do_preprocess
|
2024-10-26 16:24:35 +08:00 |
|
tackhwa
|
88f4e34528
|
support lmdeploy backend
|
2024-10-26 16:11:15 +08:00 |
|
zrguo
|
fd68757c4f
|
Merge pull request #138 from tackhwa/main
[FIX] fix hf output bug (current output contain user prompt which cause logical error in entity extraction phase)
|
2024-10-26 14:10:14 +08:00 |
|
tackhwa
|
5bfd107f5e
|
Update token length
|
2024-10-26 02:42:40 +08:00 |
|
tackhwa
|
d9054c6e4f
|
fix hf output bug
|
2024-10-26 02:20:23 +08:00 |
|
Sanketh Kumar
|
5e3ab98d83
|
Manually reformatted files
|
2024-10-25 13:32:25 +05:30 |
|
tackhwa
|
fd30ae4e45
|
move_code
|
2024-10-23 15:25:46 +08:00 |
|
tackhwa
|
63c0283514
|
fix hf bug
|
2024-10-23 15:02:28 +08:00 |
|
zrguo
|
5e388f3f9f
|
Merge pull request #87 from Soumil32/main
Added a class to use multiple models
|
2024-10-23 11:12:44 +08:00 |
|
zrguo
|
cd41fa1b25
|
Merge branch 'main' into main
|
2024-10-23 11:08:40 +08:00 |
|
tpoisonooo
|
274d0fcc92
|
feat(examples): support siliconcloud free API
|
2024-10-22 15:16:57 +08:00 |
|
Abyl Ikhsanov
|
c69a3606c6
|
Update llm.py
|
2024-10-21 20:40:49 +02:00 |
|
Soumil
|
d517ef9c20
|
added a class to use multiple models
|
2024-10-21 18:34:43 +01:00 |
|
Andrii Lazarchuk
|
216813c300
|
Add ability to passadditional parameters to ollama library like host and timeout
|
2024-10-21 11:55:46 +00:00 |
|
Sanketh Kumar
|
744dad339d
|
chore: added pre-commit-hooks and ruff formatting for commit-hooks
|
2024-10-19 09:43:17 +05:30 |
|
João Galego
|
75a91d9300
|
Fixed retry strategy, message history and inference params; Cleaned up Bedrock example
|
2024-10-18 16:50:02 +01:00 |
|
João Galego
|
7050875295
|
Added support for Amazon Bedrock models
|
2024-10-18 14:17:14 +01:00 |
|
LarFii
|
10d1ac4885
|
ollama test
|
2024-10-16 15:15:10 +08:00 |
|
Sung Kim
|
b0ad8775f4
|
Added OpenAI compatible options and examples
|
2024-10-15 12:55:05 -07:00 |
|
LarFii
|
756133512e
|
fix bug
|
2024-10-15 21:21:57 +08:00 |
|
LarFii
|
a92f7bfd61
|
update llm.py
|
2024-10-15 20:06:59 +08:00 |
|
LarFii
|
ea126a7108
|
Add huggingface model support
|
2024-10-15 19:40:08 +08:00 |
|
TianyuFan0504
|
d47dd465ce
|
update hf_model_complete
|
2024-10-14 20:33:46 +08:00 |
|