mirror of
https://github.com/langgenius/dify.git
synced 2024-11-16 19:59:50 +08:00
89 lines
2.7 KiB
Python
89 lines
2.7 KiB
Python
import datetime
|
|
import logging
|
|
import time
|
|
|
|
import click
|
|
from celery import shared_task
|
|
from llama_index.data_structs import Node
|
|
from llama_index.data_structs.node_v2 import DocumentRelationship
|
|
from werkzeug.exceptions import NotFound
|
|
|
|
from core.index.keyword_table_index import KeywordTableIndex
|
|
from core.index.vector_index import VectorIndex
|
|
from extensions.ext_database import db
|
|
from extensions.ext_redis import redis_client
|
|
from models.dataset import DocumentSegment
|
|
|
|
|
|
@shared_task
|
|
def add_segment_to_index_task(segment_id: str):
|
|
"""
|
|
Async Add segment to index
|
|
:param segment_id:
|
|
|
|
Usage: add_segment_to_index.delay(segment_id)
|
|
"""
|
|
logging.info(click.style('Start add segment to index: {}'.format(segment_id), fg='green'))
|
|
start_at = time.perf_counter()
|
|
|
|
segment = db.session.query(DocumentSegment).filter(DocumentSegment.id == segment_id).first()
|
|
if not segment:
|
|
raise NotFound('Segment not found')
|
|
|
|
if segment.status != 'completed':
|
|
return
|
|
|
|
indexing_cache_key = 'segment_{}_indexing'.format(segment.id)
|
|
|
|
try:
|
|
relationships = {
|
|
DocumentRelationship.SOURCE: segment.document_id,
|
|
}
|
|
|
|
previous_segment = segment.previous_segment
|
|
if previous_segment:
|
|
relationships[DocumentRelationship.PREVIOUS] = previous_segment.index_node_id
|
|
|
|
next_segment = segment.next_segment
|
|
if next_segment:
|
|
relationships[DocumentRelationship.NEXT] = next_segment.index_node_id
|
|
|
|
node = Node(
|
|
doc_id=segment.index_node_id,
|
|
doc_hash=segment.index_node_hash,
|
|
text=segment.content,
|
|
extra_info=None,
|
|
node_info=None,
|
|
relationships=relationships
|
|
)
|
|
|
|
dataset = segment.dataset
|
|
|
|
if not dataset:
|
|
raise Exception('Segment has no dataset')
|
|
|
|
vector_index = VectorIndex(dataset=dataset)
|
|
keyword_table_index = KeywordTableIndex(dataset=dataset)
|
|
|
|
# save vector index
|
|
if dataset.indexing_technique == "high_quality":
|
|
vector_index.add_nodes(
|
|
nodes=[node],
|
|
duplicate_check=True
|
|
)
|
|
|
|
# save keyword index
|
|
keyword_table_index.add_nodes([node])
|
|
|
|
end_at = time.perf_counter()
|
|
logging.info(click.style('Segment added to index: {} latency: {}'.format(segment.id, end_at - start_at), fg='green'))
|
|
except Exception as e:
|
|
logging.exception("add segment to index failed")
|
|
segment.enabled = False
|
|
segment.disabled_at = datetime.datetime.utcnow()
|
|
segment.status = 'error'
|
|
segment.error = str(e)
|
|
db.session.commit()
|
|
finally:
|
|
redis_client.delete(indexing_cache_key)
|