diff --git a/README.md b/README.md
index 777beec5..25946f74 100644
--- a/README.md
+++ b/README.md
@@ -1,4 +1,115 @@
-# LTX-2
+# LTX-2 Optimized (8GB of VRAM Edition) + Web UI
+
+This repository contains a **modified and optimized version of the LTX-2 Video Generation Model**, designed specifically to run on consumer hardware with as little as **8GB of VRAM**.
+
+It includes a fully-featured **Gradio Web Interface** to make generating videos, managing presets, and applying LoRAs easy without needing to remember complex command-line arguments.
+## Web UI v2
+
+
+## Web UI v4
+
+
+## CinemaMaker UI
+
+
+* https://youtu.be/eGOq0hUiri4
+* https://youtu.be/HAQqzPdDIj0
+
+## Music to Video UI
+
+
+* https://youtu.be/HzK1nW-OVtQ
+
+
+
+## 🚀 Features
+
+* **8GB VRAM Optimization:** Runs locally on cards like the RTX 3070/4060Ti using FP8 quantization and memory management tweaks.
+* **Windows 11 support!!!** You can even run it on Windows (not supported in the original model).
+* **User-Friendly Web UI:** Control everything from your browser.
+* **Smart "Safe Mode":** The UI automatically limits the frame count based on selected resolution to prevent Out-Of-Memory (OOM) errors. (If you do not have 8GB of free VRAM, try decreasing the frame count.)
+* **Real-time Logging:** View the generation progress and console output directly in the web interface.
+* **Advanced Features:**
+ * **Image Conditioning:** Upload reference images.
+ * **LoRA Support:** Checkbox selection for Camera Control.
+ * **Seed Control:** Reproducible generations.
+
+## 📥 Model Download & Setup
+
+To run this, you need to download the specific FP8 distilled checkpoints and the Text Encoder.
+
+**1. Create a `models` directory in the root folder:**
+```bash
+mkdir models
+mkdir models/loras
+```
+
+**2. Download the models:**
+* [`ltx-2-19b-distilled-fp8.safetensors`](https://huggingface.co/Lightricks/LTX-2/blob/main/ltx-2-19b-distilled-fp8.safetensors) - [Download](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-19b-distilled-fp8.safetensors)
+* [`ltx-2-spatial-upscaler-x2-1.0.safetensors`](https://huggingface.co/Lightricks/LTX-2/blob/main/ltx-2-spatial-upscaler-x2-1.0.safetensors) - [Download](https://huggingface.co/Lightricks/LTX-2/resolve/main/ltx-2-spatial-upscaler-x2-1.0.safetensors)
+* [`Gemma 3`](https://huggingface.co/google/gemma-3-12b-it-qat-q4_0-unquantized/tree/main)
+```
+./models/
+ ltx-2-19b-distilled-fp8.safetensors
+ ltx-2-spatial-upscaler-x2-1.0.safetensors
+
+./models/gemma3/
+ gemma-3 files
+
+./models/loras/
+ LoRA files here
+```
+**3. Install all required modules:**
+```
+required modules
+pip install -e packages/ltx-pipelines
+pip install -e packages/ltx-core
+
+Python 3.12.8
+accelerate==1.10.1
+torch==2.8.0+cu128
+torchaudio==2.8.0+cu128
+torchvision==0.23.0+cu128
+xformers==0.0.32.post2
+...
+```
+**🖥️ Usage**
+Run the web interface with a single command:
+```Bash
+python web_ui_v2.py
+
+or
+
+python web_ui_v4.py
+```
+
+**📊 Performance & Presets (8GB of VRAM)**
+* The Web UI includes an "8GB VRAM Safe Mode" checkbox. When enabled, it enforces the following limits to ensure you don't crash your GPU. Est. inference time on RTX 3070 Ti laptop GPU ~300sec for all presets.
+```
+| Resolution | Max Frames i2v| t2v | Est. Time (3070ti laptop 8gb vram) |
+| :---------- | :------------ |:---- |:---------------------------------- |
+| 1280 x 704 | 177 | 257 | ~300..400 sec |
+| 1536 x 1024 | 121 | 185 | ~300..400 sec |
+| 1920 x 1088 | 81 | 121 | ~300..400 sec |
+| 2560 x 1408 | 49 | 65 | ~300..400 sec |
+| 3840 x 2176 | 17 | 25 | ~300..400 sec |
+* +60 sec for prompt (if not empty/not cached)
+* time to stage 1 preview 80..150 sec
+```
+* UPD: optimized transformer code, increased max frames by 40% for text to video, generation speed 300..315 -> 385..415 sec, (1280x704 11sec 24fps, 1920x1088 5sec 24fps)
+* UPD2: added web ui v4, stage 1 video preview, task queue, prompt constructor, disable audio option (faster inference 10-30%)
+
+
+**Credits**
+* Original Model: Lightricks (LTX-2)
+* Optimization: nalexand
+* Web UI: Created for the community to make this powerful model accessible.
+
+Original Model:
+* (you can find links to all model files and loras below)
+
+
+## LTX-2
[](https://ltx.io)
[](https://huggingface.co/Lightricks/LTX-2)
diff --git a/film_maker_ui_v4.py b/film_maker_ui_v4.py
new file mode 100644
index 00000000..856548c5
--- /dev/null
+++ b/film_maker_ui_v4.py
@@ -0,0 +1,574 @@
+import gradio as gr
+import subprocess
+import os
+import datetime
+import threading
+import json
+import sys
+import google.generativeai as genai
+from collections import deque
+import cv2 # For frame extraction
+
+# --- Configuration & Defaults ---
+DEFAULT_CHECKPOINT = "./models/ltx-2-19b-distilled-fp8.safetensors"
+DEFAULT_GEMMA = "./models/gemma3"
+DEFAULT_UPSAMPLER = "./models/ltx-2-spatial-upscaler-x2-1.0.safetensors"
+
+# --- Master Prompt ---
+SYSTEM_INSTRUCTION = """
+You are a Creative Assistant. Given a user's raw input prompt describing a scene or concept, expand it into a detailed video generation script split into 5-8 short scenes (5 seconds each).
+Each scene must guide a text-to-video model with specific visuals and integrated audio.
+
+#### Crucial Generation Context
+- We generate scenes in CHRONOLOGICAL ORDER (starting from the first scene and moving towards the last).
+- The FIRST SCENE must be the MOST DETAILED, describing the environment, primary characters, and lighting with high precision to set the standard for the entire chain.
+- Subsequent scenes should maintain this description while focusing on their specific action and ensuring continuity from the previous scene.
+
+#### Continuity & Scene Construction
+- All scenes are connected by shared end/start frames.
+- Environment changes MUST OCCUR INSIDE a scene, not between scenes.
+- Each scene must be a direct continuation of the previous one.
+- Describe explicit CAMERA MOVEMENTS (e.g., "slow dolly in," "pan left," "handheld shake") within each scene.
+- Transitions or scene changes must be described as part of the visual action within the 5-second block.
+
+#### Guidelines
+- Strictly follow all aspects of the user's raw input.
+- If the input is vague, invent concrete details: lighting, textures, materials, scene settings, etc.
+- For characters: describe gender, clothing, hair, expressions. DO NOT invent unrequested characters.
+- NO SPEECH: Characters do not speak (this model produces video and background audio only). Describe reactions, expressions, and physical movements instead.
+- Use active language: present-progressive verbs ("is walking," "is grasping").
+- Maintain chronological flow within scenes: use temporal connectors ("as," "then," "while").
+- Audio layer: Describe complete soundscape integrated chronologically. Be specific (e.g., "distant thunder," "rustling leaves," "mechanical hum").
+- Style: Include visual style at the beginning: "Style: