Update since 'File manager' is renamed to 'File' (#11698)

### What problem does this PR solve?

Update some docs and comments, since 'File manager' is rename to 'File'

### Type of change

- [x] Documentation Update
- [x] Refactoring

---------

Signed-off-by: Jin Hai <haijin.chn@gmail.com>
Co-authored-by: writinwaters <93570324+writinwaters@users.noreply.github.com>
This commit is contained in:
Jin Hai 2025-12-03 18:32:15 +08:00 committed by GitHub
parent 648342b62f
commit a7d40e9132
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
5 changed files with 17 additions and 17 deletions

View File

@ -39,7 +39,7 @@ async def upload(tenant_id):
Upload a file to the system.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -155,7 +155,7 @@ async def create(tenant_id):
Create a new file or folder.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -233,7 +233,7 @@ async def list_files(tenant_id):
List files under a specific folder.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -325,7 +325,7 @@ async def get_root_folder(tenant_id):
Get user's root folder.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
responses:
@ -361,7 +361,7 @@ async def get_parent_folder():
Get parent folder info of a file.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -406,7 +406,7 @@ async def get_all_parent_folders(tenant_id):
Get all parent folders of a file.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -454,7 +454,7 @@ async def rm(tenant_id):
Delete one or multiple files/folders.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -528,7 +528,7 @@ async def rename(tenant_id):
Rename a file.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:
@ -589,7 +589,7 @@ async def get(tenant_id, file_id):
Download a file.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
produces:
@ -637,7 +637,7 @@ async def move(tenant_id):
Move one or multiple files to another folder.
---
tags:
- File Management
- File
security:
- ApiKeyAuth: []
parameters:

View File

@ -170,7 +170,7 @@ TZ=Asia/Shanghai
# Uncomment the following line if your operating system is MacOS:
# MACOS=1
# The maximum file size limit (in bytes) for each upload to your knowledge base or File Management.
# The maximum file size limit (in bytes) for each upload to your dataset or RAGFlow's File system.
# To change the 1GB file size limit, uncomment the line below and update as needed.
# MAX_CONTENT_LENGTH=1073741824
# After updating, ensure `client_max_body_size` in nginx/nginx.conf is updated accordingly.

View File

@ -76,5 +76,5 @@ No. Files uploaded to an agent as input are not stored in a dataset and hence wi
There is no _specific_ file size limit for a file uploaded to an agent. However, note that model providers typically have a default or explicit maximum token setting, which can range from 8196 to 128k: The plain text part of the uploaded file will be passed in as the key value, but if the file's token count exceeds this limit, the string will be truncated and incomplete.
:::tip NOTE
The variables `MAX_CONTENT_LENGTH` in `/docker/.env` and `client_max_body_size` in `/docker/nginx/nginx.conf` set the file size limit for each upload to a dataset or **File Management**. These settings DO NOT apply in this scenario.
The variables `MAX_CONTENT_LENGTH` in `/docker/.env` and `client_max_body_size` in `/docker/nginx/nginx.conf` set the file size limit for each upload to a dataset or RAGFlow's File system. These settings DO NOT apply in this scenario.
:::

View File

@ -9,7 +9,7 @@ Initiate an AI-powered chat with a configured chat assistant.
---
Knowledge base, hallucination-free chat, and file management are the three pillars of RAGFlow. Chats in RAGFlow are based on a particular dataset or multiple datasets. Once you have created your dataset, finished file parsing, and [run a retrieval test](../dataset/run_retrieval_test.md), you can go ahead and start an AI conversation.
Chats in RAGFlow are based on a particular dataset or multiple datasets. Once you have created your dataset, finished file parsing, and [run a retrieval test](../dataset/run_retrieval_test.md), you can go ahead and start an AI conversation.
## Start an AI chat

View File

@ -5,7 +5,7 @@ slug: /configure_knowledge_base
# Configure dataset
Most of RAGFlow's chat assistants and Agents are based on datasets. Each of RAGFlow's datasets serves as a knowledge source, *parsing* files uploaded from your local machine and file references generated in **File Management** into the real 'knowledge' for future AI chats. This guide demonstrates some basic usages of the dataset feature, covering the following topics:
Most of RAGFlow's chat assistants and Agents are based on datasets. Each of RAGFlow's datasets serves as a knowledge source, *parsing* files uploaded from your local machine and file references generated in RAGFlow's File system into the real 'knowledge' for future AI chats. This guide demonstrates some basic usages of the dataset feature, covering the following topics:
- Create a dataset
- Configure a dataset
@ -82,10 +82,10 @@ Some embedding models are optimized for specific languages, so performance may b
### Upload file
- RAGFlow's **File Management** allows you to link a file to multiple datasets, in which case each target dataset holds a reference to the file.
- RAGFlow's File system allows you to link a file to multiple datasets, in which case each target dataset holds a reference to the file.
- In **Knowledge Base**, you are also given the option of uploading a single file or a folder of files (bulk upload) from your local machine to a dataset, in which case the dataset holds file copies.
While uploading files directly to a dataset seems more convenient, we *highly* recommend uploading files to **File Management** and then linking them to the target datasets. This way, you can avoid permanently deleting files uploaded to the dataset.
While uploading files directly to a dataset seems more convenient, we *highly* recommend uploading files to RAGFlow's File system and then linking them to the target datasets. This way, you can avoid permanently deleting files uploaded to the dataset.
### Parse file
@ -142,6 +142,6 @@ As of RAGFlow v0.22.1, the search feature is still in a rudimentary form, suppor
You are allowed to delete a dataset. Hover your mouse over the three dot of the intended dataset card and the **Delete** option appears. Once you delete a dataset, the associated folder under **root/.knowledge** directory is AUTOMATICALLY REMOVED. The consequence is:
- The files uploaded directly to the dataset are gone;
- The file references, which you created from within **File Management**, are gone, but the associated files still exist in **File Management**.
- The file references, which you created from within RAGFlow's File system, are gone, but the associated files still exist.
![delete dataset](https://raw.githubusercontent.com/infiniflow/ragflow-docs/main/images/delete_datasets.jpg)