mirror of
https://github.com/langgenius/dify.git
synced 2024-11-16 11:42:29 +08:00
94 lines
3.1 KiB
Python
94 lines
3.1 KiB
Python
|
import json
|
||
|
import logging
|
||
|
import time
|
||
|
|
||
|
import click
|
||
|
from celery import shared_task
|
||
|
|
||
|
from core.indexing_runner import DocumentIsPausedException
|
||
|
from extensions.ext_database import db
|
||
|
from extensions.ext_storage import storage
|
||
|
from models.dataset import Dataset, ExternalKnowledgeApis
|
||
|
from models.model import UploadFile
|
||
|
from services.external_knowledge_service import ExternalDatasetService
|
||
|
|
||
|
|
||
|
@shared_task(queue="dataset")
|
||
|
def external_document_indexing_task(
|
||
|
dataset_id: str, external_knowledge_api_id: str, data_source: dict, process_parameter: dict
|
||
|
):
|
||
|
"""
|
||
|
Async process document
|
||
|
:param dataset_id:
|
||
|
:param external_knowledge_api_id:
|
||
|
:param data_source:
|
||
|
:param process_parameter:
|
||
|
Usage: external_document_indexing_task.delay(dataset_id, document_id)
|
||
|
"""
|
||
|
start_at = time.perf_counter()
|
||
|
|
||
|
dataset = db.session.query(Dataset).filter(Dataset.id == dataset_id).first()
|
||
|
if not dataset:
|
||
|
logging.info(
|
||
|
click.style("Processed external dataset: {} failed, dataset not exit.".format(dataset_id), fg="red")
|
||
|
)
|
||
|
return
|
||
|
|
||
|
# get external api template
|
||
|
external_knowledge_api = (
|
||
|
db.session.query(ExternalKnowledgeApis)
|
||
|
.filter(
|
||
|
ExternalKnowledgeApis.id == external_knowledge_api_id, ExternalKnowledgeApis.tenant_id == dataset.tenant_id
|
||
|
)
|
||
|
.first()
|
||
|
)
|
||
|
|
||
|
if not external_knowledge_api:
|
||
|
logging.info(
|
||
|
click.style(
|
||
|
"Processed external dataset: {} failed, api template: {} not exit.".format(
|
||
|
dataset_id, external_knowledge_api_id
|
||
|
),
|
||
|
fg="red",
|
||
|
)
|
||
|
)
|
||
|
return
|
||
|
files = {}
|
||
|
if data_source["type"] == "upload_file":
|
||
|
upload_file_list = data_source["info_list"]["file_info_list"]["file_ids"]
|
||
|
for file_id in upload_file_list:
|
||
|
file = (
|
||
|
db.session.query(UploadFile)
|
||
|
.filter(UploadFile.tenant_id == dataset.tenant_id, UploadFile.id == file_id)
|
||
|
.first()
|
||
|
)
|
||
|
if file:
|
||
|
files[file.id] = (file.name, storage.load_once(file.key), file.mime_type)
|
||
|
try:
|
||
|
settings = ExternalDatasetService.get_external_knowledge_api_settings(
|
||
|
json.loads(external_knowledge_api.settings)
|
||
|
)
|
||
|
# assemble headers
|
||
|
headers = ExternalDatasetService.assembling_headers(settings.authorization, settings.headers)
|
||
|
|
||
|
# do http request
|
||
|
response = ExternalDatasetService.process_external_api(settings, headers, process_parameter, files)
|
||
|
job_id = response.json().get("job_id")
|
||
|
if job_id:
|
||
|
# save job_id to dataset
|
||
|
dataset.job_id = job_id
|
||
|
db.session.commit()
|
||
|
|
||
|
end_at = time.perf_counter()
|
||
|
logging.info(
|
||
|
click.style(
|
||
|
"Processed external dataset: {} successful, latency: {}".format(dataset.id, end_at - start_at),
|
||
|
fg="green",
|
||
|
)
|
||
|
)
|
||
|
except DocumentIsPausedException as ex:
|
||
|
logging.info(click.style(str(ex), fg="yellow"))
|
||
|
|
||
|
except Exception:
|
||
|
pass
|