| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | import datetime | 
					
						
							|  |  |  | import logging | 
					
						
							|  |  |  | import time | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  | import click | 
					
						
							|  |  |  | from celery import shared_task | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | from core.index.index import IndexBuilder | 
					
						
							| 
									
										
										
										
											2024-01-12 12:34:01 +08:00
										 |  |  | from core.indexing_runner import DocumentIsPausedException, IndexingRunner | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | from extensions.ext_database import db | 
					
						
							| 
									
										
										
										
											2024-01-12 12:34:01 +08:00
										 |  |  | from models.dataset import Dataset, Document, DocumentSegment | 
					
						
							|  |  |  | from werkzeug.exceptions import NotFound | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-07-31 13:13:08 +08:00
										 |  |  | @shared_task(queue='dataset') | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | def document_indexing_update_task(dataset_id: str, document_id: str): | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     Async update document | 
					
						
							|  |  |  |     :param dataset_id: | 
					
						
							|  |  |  |     :param document_id: | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     Usage: document_indexing_update_task.delay(dataset_id, document_id) | 
					
						
							|  |  |  |     """
 | 
					
						
							|  |  |  |     logging.info(click.style('Start update document: {}'.format(document_id), fg='green')) | 
					
						
							|  |  |  |     start_at = time.perf_counter() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     document = db.session.query(Document).filter( | 
					
						
							|  |  |  |         Document.id == document_id, | 
					
						
							|  |  |  |         Document.dataset_id == dataset_id | 
					
						
							|  |  |  |     ).first() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     if not document: | 
					
						
							|  |  |  |         raise NotFound('Document not found') | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     document.indexing_status = 'parsing' | 
					
						
							|  |  |  |     document.processing_started_at = datetime.datetime.utcnow() | 
					
						
							|  |  |  |     db.session.commit() | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |     # delete all document segment and index | 
					
						
							|  |  |  |     try: | 
					
						
							|  |  |  |         dataset = db.session.query(Dataset).filter(Dataset.id == dataset_id).first() | 
					
						
							|  |  |  |         if not dataset: | 
					
						
							|  |  |  |             raise Exception('Dataset not found') | 
					
						
							|  |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |         vector_index = IndexBuilder.get_index(dataset, 'high_quality') | 
					
						
							|  |  |  |         kw_index = IndexBuilder.get_index(dataset, 'economy') | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  |         segments = db.session.query(DocumentSegment).filter(DocumentSegment.document_id == document_id).all() | 
					
						
							|  |  |  |         index_node_ids = [segment.index_node_id for segment in segments] | 
					
						
							|  |  |  | 
 | 
					
						
							|  |  |  |         # delete from vector index | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |         if vector_index: | 
					
						
							|  |  |  |             vector_index.delete_by_ids(index_node_ids) | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  |         # delete from keyword index | 
					
						
							|  |  |  |         if index_node_ids: | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |             kw_index.delete_by_ids(index_node_ids) | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  | 
 | 
					
						
							|  |  |  |         for segment in segments: | 
					
						
							|  |  |  |             db.session.delete(segment) | 
					
						
							| 
									
										
										
										
											2023-06-06 23:16:51 +08:00
										 |  |  |         db.session.commit() | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  |         end_at = time.perf_counter() | 
					
						
							|  |  |  |         logging.info( | 
					
						
							|  |  |  |             click.style('Cleaned document when document update data source or process rule: {} latency: {}'.format(document_id, end_at - start_at), fg='green')) | 
					
						
							|  |  |  |     except Exception: | 
					
						
							|  |  |  |         logging.exception("Cleaned document when document update data source or process rule failed") | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  | 
 | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  |     try: | 
					
						
							|  |  |  |         indexing_runner = IndexingRunner() | 
					
						
							| 
									
										
										
										
											2023-06-16 21:47:51 +08:00
										 |  |  |         indexing_runner.run([document]) | 
					
						
							| 
									
										
										
										
											2023-06-01 23:19:36 +08:00
										 |  |  |         end_at = time.perf_counter() | 
					
						
							|  |  |  |         logging.info(click.style('update document: {} latency: {}'.format(document.id, end_at - start_at), fg='green')) | 
					
						
							| 
									
										
										
										
											2023-06-25 16:49:14 +08:00
										 |  |  |     except DocumentIsPausedException as ex: | 
					
						
							|  |  |  |         logging.info(click.style(str(ex), fg='yellow')) | 
					
						
							|  |  |  |     except Exception: | 
					
						
							|  |  |  |         pass |