Merged
Conversation
…a.cpp - Add YAML anchors for common configurations (x-common-config, x-huggingface-cache, x-auth-config, etc.) - Reduce code duplication by ~200+ lines across services - Make llama.cpp image configurable via LLAMACPP_IMAGE environment variable - Resolve ARM64/AMD64 platform compatibility issues - Improve maintainability through centralized configuration patterns
- Add --timeout 300 and --retries 3 flags to pip install - Resolve intermittent build failures when downloading large packages (onnxruntime) - Improve build reliability for CI/CD and slower network connections
- Document configurable llama.cpp Docker image option - Provide examples for different architectures (ARM64, AMD64, CUDA) - Keep .env.example in sync with docker-compose.yml capabilities
🤖 Augment PR SummarySummary: Improves Docker-based dev/deploy ergonomics by making the llama.cpp decoder image configurable and increasing build robustness. 🤖 Was this summary useful? React with 👍 or 👎 |
| # Llama.cpp decoder service configuration | ||
| # Default: ghcr.io/ggml-org/llama.cpp:server (multi-arch) | ||
| # ARM64 specific: ghcr.io/ggml-org/llama.cpp:server-cuda (if needed) | ||
| # Alternative: local builds or custom images |
Collaborator
Author
There was a problem hiding this comment.
Updated comment.
- Fix misleading comment about server-cuda being ARM64-specific - CUDA images are for NVIDIA GPU support, not ARM64 architecture - Clarify that server-cuda is for NVIDIA GPUs (typically x86_64)
…rameter - Add missing on_disk_payload parameter to FakeClient mock in test_ingest_schema_mode.py - Resolves TypeError: FakeClient.create_collection() got an unexpected keyword argument 'on_disk_payload' - Ensures test mocks match the real Qdrant client interface which includes this parameter
m1rl0k
approved these changes
Jan 26, 2026
m1rl0k
added a commit
that referenced
this pull request
Mar 1, 2026
…-optimizations Feature/docker compose optimizations
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
🐳 Docker Compose Optimization: YAML Anchors & Configurable Services
Summary
Significantly improves Docker Compose maintainability by introducing YAML anchors and making llama.cpp image configurable.
Changes
x-common-config,x-huggingface-cache, etc.) reducing ~200+ lines of duplicationLLAMACPP_IMAGEenvironment variable for ARM64/AMD64 compatibilityBenefits
Testing
Migration
No breaking changes. Optionally set
LLAMACPP_IMAGEin your.envfor custom images.