floyo logo
Powered by
ThinkDiffusion
floyo logo
Powered by
ThinkDiffusion

LTX 2.0 – Prompting & Dynamic Camera Movement

21

LTX-2 Video & Audio Generator

What this workflow does: This workflow creates AI-generated videos with synchronized audio using the powerful LTX-2 model. Simply describe what you want to see and hear, and the AI will generate a short video clip for you.


📝 How to Use:

1. Write Your Prompt (CR Prompt Text boxes)

  • Describe your video scene in detail

  • Include what's happening, how things look, sounds, and camera movements

  • Example: "A cheerful puppet singing in the rain, holding a red umbrella"

2. Set Video Dimensions

  • Width: 1280 pixels (default)

  • Height: 720 pixels (default)

  • Length: 401 frames (about 16 seconds at 24fps)

3. Optional AI Prompt Enhancement (LLM_floyo - currently disabled)

  • This can automatically improve your prompt for better results

  • Transforms simple ideas into detailed, cinematic descriptions

4. Generate Your Video

  • The workflow will create both video and audio

  • Uses advanced AI models to ensure high quality

  • Final output is saved as an MP4 file


🎯 Key Components Explained:

Prompt Inputs: Where you describe what you want Model Loaders: Load the AI brain that creates the video Video Settings: Control size, length, and frame rate Samplers: How the AI creates the video (like brush strokes for a painter) Upscaler: Makes the video sharper and more detailed Output: Your finished video with sound!


💡 Tips for Better Results:

Be specific - "A golden retriever puppy playing in a sunny park" works better than just "dog" ✅ Include sounds - Mention what you want to hear: "birds chirping, children laughing" ✅ Describe camera work - "slow zoom in," "handheld camera," "drone shot from above" ✅ Add atmosphere - Mention lighting, weather, and mood

Avoid these - Don't use quality tags like "4K" or "best quality" - they don't help with this model


⚙️ Technical Settings (Advanced Users):

  • Steps: 20 (how many times the AI refines the image)

  • CFG Scale: 4 (how closely it follows your prompt)

  • Frame Rate: 24 fps (standard cinema speed)

  • Sampler: euler_ancestral (the rendering method)

Read more

N
Generates in about -- secs

Nodes & Models

JWInteger
LTXVAudioVAELoader
ltx-2-19b-dev-fp8.safetensors
PrimitiveInt
PrimitiveFloat
RandomNoise
KSamplerSelect
LatentUpscaleModelLoader
ltx-2-spatial-upscaler-x2-1.0.safetensors
ManualSigmas
LTXAVTextEncoderLoader
gemma_3_12B_it.safetensors
ltx-2-19b-dev-fp8.safetensors
CheckpointLoaderSimple
ltx-2-19b-dev-fp8.safetensors
MarkdownNote
WorkflowGraphics
CLIPTextEncode
LoraLoaderModelOnly
ltx-2-19b-distilled-lora-384.safetensors
ltx-2-19b-lora-camera-control-dolly-left-ltx-2-0-prompting-dy-1TFIV6LG.safetensors
EmptyImage
LTXVEmptyLatentAudio
LTXVConditioning
ImageScaleBy
CFGGuider
GetImageSize
EmptyLTXVLatentVideo
LTXVConcatAVLatent
LTXVScheduler
SamplerCustomAdvanced
LTXVSeparateAVLatent
LTXVCropGuides
LTXVLatentUpsampler
VAEDecode
VAEDecodeTiled
LTXVAudioVAEDecode
CreateVideo
SaveVideo

LTX-2 Video & Audio Generator

What this workflow does: This workflow creates AI-generated videos with synchronized audio using the powerful LTX-2 model. Simply describe what you want to see and hear, and the AI will generate a short video clip for you.


📝 How to Use:

1. Write Your Prompt (CR Prompt Text boxes)

  • Describe your video scene in detail

  • Include what's happening, how things look, sounds, and camera movements

  • Example: "A cheerful puppet singing in the rain, holding a red umbrella"

2. Set Video Dimensions

  • Width: 1280 pixels (default)

  • Height: 720 pixels (default)

  • Length: 401 frames (about 16 seconds at 24fps)

3. Optional AI Prompt Enhancement (LLM_floyo - currently disabled)

  • This can automatically improve your prompt for better results

  • Transforms simple ideas into detailed, cinematic descriptions

4. Generate Your Video

  • The workflow will create both video and audio

  • Uses advanced AI models to ensure high quality

  • Final output is saved as an MP4 file


🎯 Key Components Explained:

Prompt Inputs: Where you describe what you want Model Loaders: Load the AI brain that creates the video Video Settings: Control size, length, and frame rate Samplers: How the AI creates the video (like brush strokes for a painter) Upscaler: Makes the video sharper and more detailed Output: Your finished video with sound!


💡 Tips for Better Results:

Be specific - "A golden retriever puppy playing in a sunny park" works better than just "dog" ✅ Include sounds - Mention what you want to hear: "birds chirping, children laughing" ✅ Describe camera work - "slow zoom in," "handheld camera," "drone shot from above" ✅ Add atmosphere - Mention lighting, weather, and mood

Avoid these - Don't use quality tags like "4K" or "best quality" - they don't help with this model


⚙️ Technical Settings (Advanced Users):

  • Steps: 20 (how many times the AI refines the image)

  • CFG Scale: 4 (how closely it follows your prompt)

  • Frame Rate: 24 fps (standard cinema speed)

  • Sampler: euler_ancestral (the rendering method)

Read more

N