dify/api
小羽 56b43f62d1
Some checks are pending
Build and Push API & Web / build (api, DIFY_API_IMAGE_NAME, linux/amd64, build-api-amd64) (push) Waiting to run
Build and Push API & Web / build (api, DIFY_API_IMAGE_NAME, linux/arm64, build-api-arm64) (push) Waiting to run
Build and Push API & Web / build (web, DIFY_WEB_IMAGE_NAME, linux/amd64, build-web-amd64) (push) Waiting to run
Build and Push API & Web / build (web, DIFY_WEB_IMAGE_NAME, linux/arm64, build-web-arm64) (push) Waiting to run
Build and Push API & Web / create-manifest (api, DIFY_API_IMAGE_NAME, merge-api-images) (push) Blocked by required conditions
Build and Push API & Web / create-manifest (web, DIFY_WEB_IMAGE_NAME, merge-web-images) (push) Blocked by required conditions
feat: nvidia add llama3.1 model (#6844)
2024-07-31 21:24:02 +08:00
..
configs fix(api/core/moderation/output_moderation.py): Fix config call. (#6769) 2024-07-29 18:30:29 +08:00
constants
contexts
controllers fix notion internal setting (#6836) 2024-07-31 17:17:46 +08:00
core feat: nvidia add llama3.1 model (#6844) 2024-07-31 21:24:02 +08:00
docker feat: support Celery auto-scale (#6249) 2024-07-31 14:34:44 +08:00
events
extensions
fields
libs refactor(api): Switch to dify_config (#6750) 2024-07-30 11:15:26 +08:00
migrations Fix/6615 40 varchar limit on DatasetCollectionBinding and Embedding model name (#6723) 2024-07-28 09:42:58 +08:00
models Fix/6615 40 varchar limit on DatasetCollectionBinding and Embedding model name (#6723) 2024-07-28 09:42:58 +08:00
schedule
services
tasks
templates
tests add deepseek-coder-v2 in siliconflow (#6149) 2024-07-29 18:45:19 +08:00
.dockerignore
.env.example
app.py chore: Add processId field for metrics of threads/db-pool-stat/health (#6797) 2024-07-31 00:21:16 +08:00
commands.py refactor(api): Switch to dify_config (#6750) 2024-07-30 11:15:26 +08:00
Dockerfile
poetry.lock chore: remove redundant version pinning for indirect dependencies (#6772) 2024-07-30 08:45:57 +08:00
poetry.toml
pyproject.toml chore: remove redundant version pinning for indirect dependencies (#6772) 2024-07-30 08:45:57 +08:00
README.md

Dify Backend API

Usage

Important

In the v0.6.12 release, we deprecated pip as the package management tool for Dify API Backend service and replaced it with poetry.

  1. Start the docker-compose stack

    The backend require some middleware, including PostgreSQL, Redis, and Weaviate, which can be started together using docker-compose.

    cd ../docker
    cp middleware.env.example middleware.env
    # change the profile to other vector database if you are not using weaviate
    docker compose -f docker-compose.middleware.yaml --profile weaviate -p dify up -d
    cd ../api
    
  2. Copy .env.example to .env

  3. Generate a SECRET_KEY in the .env file.

    sed -i "/^SECRET_KEY=/c\SECRET_KEY=$(openssl rand -base64 42)" .env
    
    secret_key=$(openssl rand -base64 42)
    sed -i '' "/^SECRET_KEY=/c\\
    SECRET_KEY=${secret_key}" .env
    
  4. Create environment.

    Dify API service uses Poetry to manage dependencies. You can execute poetry shell to activate the environment.

  5. Install dependencies

    poetry env use 3.10
    poetry install
    

    In case of contributors missing to update dependencies for pyproject.toml, you can perform the following shell instead.

    poetry shell                                               # activate current environment
    poetry add $(cat requirements.txt)           # install dependencies of production and update pyproject.toml
    poetry add $(cat requirements-dev.txt) --group dev    # install dependencies of development and update pyproject.toml
    
  6. Run migrate

    Before the first launch, migrate the database to the latest version.

    poetry run python -m flask db upgrade
    
  7. Start backend

    poetry run python -m flask run --host 0.0.0.0 --port=5001 --debug
    
  8. Start Dify web service.

  9. Setup your application by visiting http://localhost:3000...

  10. If you need to debug local async processing, please start the worker service.

poetry run python -m celery -A app.celery worker -P gevent -c 1 --loglevel INFO -Q dataset,generation,mail,ops_trace,app_deletion

The started celery app handles the async tasks, e.g. dataset importing and documents indexing.

Testing

  1. Install dependencies for both the backend and the test environment

    poetry install --with dev
    
  2. Run the tests locally with mocked system environment variables in tool.pytest_env section in pyproject.toml

    cd ../
    poetry run -C api bash dev/pytest/pytest_all_tests.sh