floyo logo
Powered by
ThinkDiffusion
floyo logo
Powered by
ThinkDiffusion

Vertical to Horizontal Video Reframe

33

Generates in about 1 sec

Nodes & Models

VHS_LoadVideoFFmpeg
VHS_VideoInfoSource
VHS_LoadVideoFFmpeg
VHS_VideoInfoSource
UNETLoader
wan2.1_vace_14B_fp16.safetensors
CLIPLoader
umt5_xxl_fp16.safetensors
PrimitiveInt
VAELoader
wan_2.1_vae.safetensors
MarkdownNote
WorkflowGraphics
LoraLoader
#community_models/loras/Wan2.1_I2V_14B_FusionX_LoRA-vertical-to-horizontal-v-M6cxEed7.safetensors
ModelSamplingSD3
CLIPTextEncode
ImagePadForOutpaint
MaskToImage
RepeatImageBatch
ImageToMask
WanVaceToVideo
KSampler
VAEDecode
CreateVideo
SaveVideo
ImageResize+
SimpleMath+
ImageResize+
DisplayAny
SimpleMath+
DisplayAny
easy compare

Vertical to Horizontal Video Outpainting Workflow (Wan 2.1)

This workflow converts vertical videos into horizontal cinematic videos by intelligently expanding the frame using Wan 2.1 VACE video generation and AI outpainting. It preserves the original content while generating new visual areas on the sides to match a wider aspect ratio.

The process begins by loading the source video using the VHS Load Video FFmpeg node, which extracts frames and retrieves video metadata such as frame count, FPS, width, and height. These values are analyzed to determine the original aspect ratio and ensure proper resizing and processing of the input frames.

The workflow then resizes the frames while maintaining proportions, targeting a horizontal format (typically 1280×720). Mathematical nodes calculate the padding required to center the original vertical content within the new horizontal canvas.

To generate the missing side regions, the workflow performs outpainting. A mask is created for the empty padded areas, and these masked regions are filled using the Wan 2.1 diffusion model. The system uses:

  • Wan 2.1 VACE 14B model for high-quality generation

  • FusionX LoRA to improve image-to-video consistency

  • UMT5 text encoder for prompt conditioning

  • Wan 2.1 VAE for decoding latent frames

The workflow optionally accepts prompts to guide how the expanded regions should look, while negative prompts help suppress artifacts such as blur or noise.

A KSampler performs the diffusion process to generate new visual content for the masked areas across the full frame sequence. The generated latent frames are then decoded into images using the VAE.

Finally, the processed frames are assembled back into a video using the CreateVideo node while preserving the original FPS, producing a smooth horizontal video with expanded scenery.

Key Features

  • Converts vertical video to horizontal format

  • Uses AI outpainting instead of cropping

  • Maintains original frame timing and motion

  • Supports prompt-based scene extension

  • Optimized with Wan 2.1 + LoRA acceleration

Output

The result is a fully reconstructed horizontal video where the original vertical content remains centered and the newly generated areas blend naturally with the scene.

Read more

N