# ============================================================================= # DictIA 16 — Docker Compose # GPU : RTX 5070 Ti (16 Go VRAM) # ============================================================================= # # Services : # - dictia : Application principale DictIA # - whisperx-asr : Service de transcription WhisperX Large-v3 # - ollama : LLM local Mistral 7B (résumés, chat, Q&A) # # Démarrage : # 1. cp config/env.dictia16.example .env # 2. docker compose -f config/docker-compose.dictia16.yml up -d # 3. Télécharger Mistral : docker exec ollama ollama pull mistral # # Note : Aucune clé API nécessaire — tout tourne en local (100% privé). # ============================================================================= services: # --------------------------------------------------------------------------- # Application DictIA # --------------------------------------------------------------------------- dictia: image: dictia:latest container_name: dictia restart: unless-stopped ports: - "8899:8899" env_file: - ../.env environment: - LOG_LEVEL=ERROR volumes: - ../uploads:/data/uploads - ../instance:/data/instance # Décommenter pour l'export automatique : # - ../exports:/data/exports # Décommenter pour le traitement automatique : # - ../auto-process:/data/auto-process depends_on: - whisperx-asr - ollama networks: - dictia-net # --------------------------------------------------------------------------- # WhisperX ASR — Transcription locale (WhisperX Large-v3) # RTX 5070 Ti : BATCH_SIZE=32, COMPUTE_TYPE=float16 # --------------------------------------------------------------------------- whisperx-asr: image: murtazanasir/whisperx-asr-service:latest container_name: whisperx-asr restart: unless-stopped environment: - HF_TOKEN=${HF_TOKEN} - DEVICE=cuda - COMPUTE_TYPE=float16 - BATCH_SIZE=32 - DEFAULT_MODEL=large-v3 volumes: - whisperx-models:/root/.cache deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] networks: - dictia-net # --------------------------------------------------------------------------- # Ollama — LLM local Mistral 7B # Résumés, points d'action, Q&A — 100% local, aucune donnée externe # --------------------------------------------------------------------------- ollama: image: ollama/ollama:latest container_name: ollama restart: unless-stopped volumes: - ollama-models:/root/.ollama deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu] networks: - dictia-net networks: dictia-net: driver: bridge volumes: whisperx-models: driver: local ollama-models: driver: local