This project is a simple full-stack prototype for generating cinematic scene concepts, storyboards, shot lists, and images.
- Backend API server (FastAPI) that returns director analysis, storyboard, and shot list.
- Frontend (React + Vite) that calls the backend and displays results.
- Image generation falls back to a placeholder image when Stable Diffusion is not available.
- Open a terminal at
visionforge-ai/backend - (Optional) Create a virtual environment:
python -m venv .venv .\.venv\Scripts\Activate.ps1 - Install dependencies:
pip install -r requirements.txt - Start the backend server:
python main.py
The backend will be available at: http://127.0.0.1:8000
- Open a new terminal at
visionforge-ai/frontend - Install dependencies:
npm install
- Start the dev server:
npm run dev
The frontend will open in your browser at http://localhost:5173 and proxy requests to the backend.
- Add a
GROQ_API_KEYtobackend/.env. - Optionally change
GROQ_MODELinbackend/.env.
- Set
ENABLE_LOCAL_SD=trueinbackend/.env. - Install the required packages (can be large):
pip install diffusers[torch] torch
- Add
HUGGINGFACE_API_TOKEN=<your_token>tobackend/.env(go to https://huggingface.co/settings/tokens to create one). - Optionally set
HF_MODELto a model you have access to (defaults tostabilityai/stable-diffusion-xl-base-1.0).
- Add
OPENAI_API_KEY=<your_key>tobackend/.env. - Optionally set
OPENAI_IMAGE_MODEL=gpt-image-1inbackend/.env. - The app will use the OpenAI Images generations endpoint to generate a landscape image.
If no image backend is available, the app will serve a placeholder image and continue working.