| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  | import logging | 
					
						
							|  |  |  | import time | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | import click | 
					
						
							|  |  |  | from celery import shared_task | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | from langchain.schema import Document | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | from core.index.index import IndexBuilder | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  | from extensions.ext_database import db | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | from models.dataset import DocumentSegment, Dataset | 
					
						
							|  |  |  | from models.dataset import Document as DatasetDocument | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-07-31 13:13:08 +08:00
										 |  |  | @shared_task(queue='dataset') | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  | def deal_dataset_vector_index_task(dataset_id: str, action: str): | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     Async deal dataset from index | 
					
						
							|  |  |  |     :param dataset_id: dataset_id | 
					
						
							|  |  |  |     :param action: action | 
					
						
							|  |  |  |     Usage: deal_dataset_vector_index_task.delay(dataset_id, action) | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     logging.info(click.style('Start deal dataset vector index: {}'.format(dataset_id), fg='green')) | 
					
						
							|  |  |  |     start_at = time.perf_counter() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     try: | 
					
						
							|  |  |  |         dataset = Dataset.query.filter_by( | 
					
						
							|  |  |  |             id=dataset_id | 
					
						
							|  |  |  |         ).first() | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  |         if not dataset: | 
					
						
							|  |  |  |             raise Exception('Dataset not found') | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  |         if action == "remove": | 
					
						
							| 
									
										
										
										
											2023-09-22 14:21:26 +08:00
										 |  |  |             index = IndexBuilder.get_index(dataset, 'high_quality', ignore_high_quality_check=True) | 
					
						
							|  |  |  |             index.delete_by_group_id(dataset.id) | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |         elif action == "add": | 
					
						
							|  |  |  |             dataset_documents = db.session.query(DatasetDocument).filter( | 
					
						
							|  |  |  |                 DatasetDocument.dataset_id == dataset_id, | 
					
						
							|  |  |  |                 DatasetDocument.indexing_status == 'completed', | 
					
						
							|  |  |  |                 DatasetDocument.enabled == True, | 
					
						
							|  |  |  |                 DatasetDocument.archived == False, | 
					
						
							|  |  |  |             ).all() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |             if dataset_documents: | 
					
						
							|  |  |  |                 # save vector index | 
					
						
							| 
									
										
										
										
											2023-08-29 03:37:45 +08:00
										 |  |  |                 index = IndexBuilder.get_index(dataset, 'high_quality', ignore_high_quality_check=False) | 
					
						
							| 
									
										
										
										
											2023-07-18 15:02:33 +08:00
										 |  |  |                 documents = [] | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |                 for dataset_document in dataset_documents: | 
					
						
							|  |  |  |                     # delete from vector index | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  |                     segments = db.session.query(DocumentSegment).filter( | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |                         DocumentSegment.document_id == dataset_document.id, | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  |                         DocumentSegment.enabled == True | 
					
						
							|  |  |  |                     ) .order_by(DocumentSegment.position.asc()).all() | 
					
						
							|  |  |  |                     for segment in segments: | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |                         document = Document( | 
					
						
							|  |  |  |                             page_content=segment.content, | 
					
						
							|  |  |  |                             metadata={ | 
					
						
							|  |  |  |                                 "doc_id": segment.index_node_id, | 
					
						
							|  |  |  |                                 "doc_hash": segment.index_node_hash, | 
					
						
							|  |  |  |                                 "document_id": segment.document_id, | 
					
						
							|  |  |  |                                 "dataset_id": segment.dataset_id, | 
					
						
							|  |  |  |                             } | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  |                         ) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |                         documents.append(document) | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-07-18 15:02:33 +08:00
										 |  |  |                 # save vector index | 
					
						
							| 
									
										
										
										
											2023-08-25 18:29:55 +08:00
										 |  |  |                 index.create(documents) | 
					
						
							| 
									
										
										
										
											2023-05-23 22:54:59 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  |         end_at = time.perf_counter() | 
					
						
							|  |  |  |         logging.info( | 
					
						
							|  |  |  |             click.style('Deal dataset vector index: {} latency: {}'.format(dataset_id, end_at - start_at), fg='green')) | 
					
						
							|  |  |  |     except Exception: | 
					
						
							|  |  |  |         logging.exception("Deal dataset vector index failed") |