import datetime import logging import time import click from celery import shared_task from llama_index.data_structs import Node from llama_index.data_structs.node_v2 import DocumentRelationship from werkzeug.exceptions import NotFound from core.index.keyword_table_index import KeywordTableIndex from core.index.vector_index import VectorIndex from extensions.ext_database import db from extensions.ext_redis import redis_client from models.dataset import DocumentSegment @shared_task def add_segment_to_index_task(segment_id: str): """ Async Add segment to index :param segment_id: Usage: add_segment_to_index.delay(segment_id) """ logging.info(click.style('Start add segment to index: {}'.format(segment_id), fg='green')) start_at = time.perf_counter() segment = db.session.query(DocumentSegment).filter(DocumentSegment.id == segment_id).first() if not segment: raise NotFound('Segment not found') if segment.status != 'completed': return indexing_cache_key = 'segment_{}_indexing'.format(segment.id) try: relationships = { DocumentRelationship.SOURCE: segment.document_id, } previous_segment = segment.previous_segment if previous_segment: relationships[DocumentRelationship.PREVIOUS] = previous_segment.index_node_id next_segment = segment.next_segment if next_segment: relationships[DocumentRelationship.NEXT] = next_segment.index_node_id node = Node( doc_id=segment.index_node_id, doc_hash=segment.index_node_hash, text=segment.content, extra_info=None, node_info=None, relationships=relationships ) dataset = segment.dataset if not dataset: raise Exception('Segment has no dataset') vector_index = VectorIndex(dataset=dataset) keyword_table_index = KeywordTableIndex(dataset=dataset) # save vector index if dataset.indexing_technique == "high_quality": vector_index.add_nodes( nodes=[node], duplicate_check=True ) # save keyword index keyword_table_index.add_nodes([node]) end_at = time.perf_counter() logging.info(click.style('Segment added to index: {} latency: {}'.format(segment.id, end_at - start_at), fg='green')) except Exception as e: logging.exception("add segment to index failed") segment.enabled = False segment.disabled_at = datetime.datetime.utcnow() segment.status = 'error' segment.error = str(e) db.session.commit() finally: redis_client.delete(indexing_cache_key)