Skip to content

mcaimi/llamachat

Repository files navigation

LlamaChat

Docker Repository on Quay

LlamaChat is a web application that allows users to chat with Large Language Models (LLMs). It provides flexibility to interact with LLMs either locally (on-prem) using technologies like vLLM or Ollama or through remote OpenAI-compatible endpoints.

With LlamaChat, users can easily chat with powerful LLMs both in private environments (for enhanced privacy and control) and using cloud-based APIs for scalability.

Features

  • Unified Chat: Interact with LLMs hosted on-premises or remotely using llama-stack
  • User-Friendly Interface: Simple web interface to initiate and maintain conversations with LLMs.
  • API Key Support: Securely manage and switch between different API keys (for remote services).
  • Multiple Models: Support for multiple LLM models such as GPT-based models, Ollama models, and others.
  • RAG Support: Chat with your documents, using a vector database as embeddings backend
  • Safety Shields: Add safety guardrails to user prompts
  • Agents Support: Use tools to enhance the LLM capabilities

Screenshot

LlamaChat Screenshot

Installation

Prerequisites

  • Python 3.12+
  • Streamlit
  • An Ollama/vLLM instance or a public OpenAI-Compatible API endpoint.
  • Llama-Stack instance running on-prem or in the cloud.

Issues

On MacOS 26, there could be problems during importing torchcodec into the application due to missing DYLD_LIBRARY_PATH settings. Project needs ffmpeg@7, so a workaround could be:

$ brew install ffmpeg@7
$ DYLD_LIBRARY_PATH="/opt/homebrew/opt/ffmpeg@7/lib:$DYLD_LIBRARY_PATH" uv run streamlit run main.py

About

This repository has been moved to https://codeberg.org/mcaimi/llamachat. Please visit the new location for the latest updates.

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published