From f4369daf1a9a843de192563bf41fadae05988456 Mon Sep 17 00:00:00 2001 From: Colin Urban Date: Thu, 17 Jul 2025 19:22:39 -0400 Subject: [PATCH 1/3] Add contributing guidelines --- README.md | 20 +++++--------------- 1 file changed, 5 insertions(+), 15 deletions(-) diff --git a/README.md b/README.md index 1bdaf0fe..125b2b46 100644 --- a/README.md +++ b/README.md @@ -37,23 +37,13 @@ For information on installation, configuration, and usage, please visit our [doc Please see [this guide](https://docs.framepackstudio.com/docs/get_started/) on our documentation site to get FP-Studio installed. -## LoRAs +## Contributing -Add LoRAs to the /loras/ folder at the root of the installation. Select the LoRAs you wish to load and set the weights for each generation. Most Hunyuan LoRAs were originally trained for T2V, it's often helpful to run a T2V generation to ensure they're working before using input images. +We would love your help building FramePack Studio! To make collaboration effective, please adhere to the following: +- Keep Pull Requests Focused: Each Pull Request should address a single issue or add one specific feature. Please do not mix bug fixes, new features, and code refactoring in the same PR. +- Target the develop Branch: All Pull Requests must be opened against the develop branch. PRs opened against the main branch will be closed. +- Discuss Big Changes First: If you plan to work on a large feature or a significant refactor, please announce it first in the #contributors channel on our [Discord server](https://discord.com/invite/MtuM7gFJ3V). This helps us coordinate efforts and prevent duplicate work. -NOTE: Slow lora loading is a known issue - -## Working with Timestamped Prompts - -You can create videos with changing prompts over time using the following syntax: - -``` -[0s: A serene forest with sunlight filtering through the trees ] -[5s: A deer appears in the clearing ] -[10s: The deer drinks from a small stream ] -``` - -Each timestamp defines when that prompt should start influencing the generation. The system will (hopefully) smoothly transition between prompts for a cohesive video. ## Credits From 34be04ef4e8c3c603d6811ab92ba6282bc5dc391 Mon Sep 17 00:00:00 2001 From: ai_anchorite Date: Sat, 15 Nov 2025 02:46:38 +1300 Subject: [PATCH 2/3] Update package versions in requirements.txt * pin peft to avoid 0.18.0 transformers incompatibility * add pinned huggingface_hub to avoid 1.0.0 update incompatibilities. --- requirements.txt | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/requirements.txt b/requirements.txt index 4a74de83..133c6a4d 100644 --- a/requirements.txt +++ b/requirements.txt @@ -5,12 +5,13 @@ diffusers==0.33.1 einops ffmpeg-python==0.2.0 gradio==5.25.2 +huggingface_hub<0.35.1 imageio-ffmpeg==0.4.8 imageio==2.31.1 jinja2>=3.1.2 numpy==1.26.2 opencv-contrib-python -peft +peft<0.18.0 pillow==11.1.0 requests==2.31.0 safetensors @@ -28,4 +29,4 @@ devicetorch facexlib>=0.2.5 gfpgan>=1.3.5 psutil -realesrgan \ No newline at end of file +realesrgan From 39324e7ea6c382584255427b9f2a4e7dd4070cdd Mon Sep 17 00:00:00 2001 From: hadxp Date: Wed, 28 Jan 2026 17:33:12 +0100 Subject: [PATCH 3/3] fix_lora_load_in_high_vram_mode --- modules/pipelines/worker.py | 16 +++++++++------- 1 file changed, 9 insertions(+), 7 deletions(-) diff --git a/modules/pipelines/worker.py b/modules/pipelines/worker.py index 67148273..3871488d 100644 --- a/modules/pipelines/worker.py +++ b/modules/pipelines/worker.py @@ -1302,13 +1302,15 @@ def fmt_eta(sec): ), ) ) - move_model_to_device_with_memory_preservation( - studio_module.current_generator.transformer, - target_device=gpu, - preserved_memory_gb=settings.get("gpu_memory_preservation"), - ) - if selected_loras: - studio_module.current_generator.move_lora_adapters_to_device(gpu) + + move_model_to_device_with_memory_preservation( + studio_module.current_generator.transformer, + target_device=gpu, + preserved_memory_gb=settings.get("gpu_memory_preservation"), + ) + + if selected_loras: + studio_module.current_generator.move_lora_adapters_to_device(gpu) from diffusers_helper.pipelines.k_diffusion_hunyuan import sample_hunyuan