Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
-
Updated
Nov 6, 2023 - Python
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
This Repositry is an experiment with an agent that searches documents and asks questions repeatedly in response to the main question. It automatically determines the optimal answer from the current documents or recognizes when there is no answer.
🐋 DeepSeek-R1: Retrieval-Augmented Generation for Document Q&A 📄
⚡ Production-ready .NET Standard 2.1 RAG library with 🤖 multi-AI provider support, 🏢 enterprise vector storage, 📄 intelligent document processing, and 🗄️ multi-database query coordination. 🌍 Cross-platform compatible.
ContextAgent is a production-ready AI assistant backend with RAG, LangChain, and FastAPI. It ingests documents, uses OpenAI embeddings, and stores vectors in ChromaDB 🐙
An advanced, fully local, and GPU-accelerated RAG pipeline. Features a sophisticated LLM-based preprocessing engine, state-of-the-art Parent Document Retriever with RAG Fusion, and a modular, Hydra-configurable architecture. Built with LangChain, Ollama, and ChromaDB for 100% private, high-performance document Q&A.
An LLM-powered Slack bot built with Langchain.
Sistema RAG (Retrieval Augmented Generation) para asistencia de documentación técnica en español utilizando LangChain, OpenAI Y Streamlit para la interfaz visual
RAG chatbot designed for domain-specific queries using Ollama, Langchain, phi-3 and Faiss
⚡️ Local RAG API using FastAPI + LangChain + Ollama | Upload PDFs, DOCX, CSVs, XLSX and ask questions using your own documents — fully offline!
AI assistant backend for document-based question answering using RAG (LangChain, OpenAI, FastAPI, ChromaDB). Features modular architecture, multi-tool agents, conversational memory, semantic search, PDF/Docx/Markdown processing, and production-ready deployment with Docker.
AI agent API (Python/FastAPI) to upload documents (PDF/TXT) and answer questions using RAG with Azure OpenAI and LangChain.
A Document QA chatbot using LangChain, Pinecone for vector storage, and Amazon Bedrock (mistral.mixtral-8x7b-instruct for LLM and titan-embed-text for embeddings). Built with a Streamlit frontend for document uploads and contextual question answering.
Sub-second RAG-based chatbot for medical Q&A over 10+ textbooks with source-cited responses.
AI-powered document Q&A system with vector search
A full-stack RAG application that enables intelligent document Q&A. Upload PDFs, DOCX, or TXT files and ask questions powered by LangChain, ChromaDB, and Claude/GPT. Features smart chunking, semantic search, conversation memory, and source citations. Built with FastAPI & React + TypeScript.
Create your own Retrieval-Augmented Generation (RAG) chatbot for PDFs. This project uses LangChain, Flask, and an LLM (IBM WatsonX/Hugging Face) to build a conversational AI that understands your documents.
An intelligent document assistant powered by Open-Source Large Language Models
Add a description, image, and links to the document-qa topic page so that developers can more easily learn about it.
To associate your repository with the document-qa topic, visit your repo's landing page and select "manage topics."