Generate spectrograms and feature-panel visualizations from audio with the songsee CLI.
74 AI agent skills for Video Generation. Part of the ๐ค AI & Agents category.
Generate spectrograms and feature-panel visualizations from audio with the songsee CLI.
Best practices for Remotion - Video creation in React
Best practices for Remotion - Video creation in React
Long-form AI video production: the frontier of multi-agent coordination. CellCog orchestrates 6-7 foundation models to produce up to 4-minute videos from a single prompt โ scripted, filmed, voiced, lipsync'd, scored, and edited automatically. Create marketing videos, product demos, explainer videos, educational content, spokesperson videos, training materials, UGC content, news reports.
HeyGen AI video creation API. Use when: (1) Using Video Agent for one-shot prompt-to-video generation, (2) Generating AI avatar videos with /v2/video/generat...
Complete toolkit for programmatic video creation with Remotion + React. Covers animations, timing, rendering (CLI/Node.js/Lambda/Cloud Run), captions, 3D, charts, text effects, transitions, and media handling. Use when writing Remotion code, building video generation pipelines, or creating data-driven video templates.
Generate video using Google Veo (Veo 3.1 / Veo 3.0).
Process video and audio using FFmpeg CLI for transcoding, cutting, merging, audio extraction, thumbnails, GIFs, speed, filters, subtitles, and watermarks.
Full video production from a single prompt. Script, shoot, stitch, score โ automatically. 30s to 4-minute Instagram Reels, TikToks, Stories, and carousels with consistent characters and agentic editing. The most advanced AI video suite for social media content, powered by #1 on DeepResearch Bench (Feb 2026).
Generate FFmpeg commands from natural language video editing requests - cut, trim, convert, compress, change aspect ratio, extract audio, and more.
Generate AI-powered notes from videos (document, outline, or graphic-text formats)
End-to-end AI video generation - create videos from text prompts using image generation, video synthesis, voice-over, and editing. Supports OpenAI DALL-E, Replicate models, LumaAI, Runway, and FFmpeg editing.
Get transcripts from any YouTube video โ for summarization, research, translation, quoting, or content analysis. Use when the user shares a video link or asks "what did they say", "get the transcript", "transcribe this video", "summarize this video", or wants to analyze spoken content.
Launch a smart teleprompter with mobile remote control for video recording. Use when the user wants to read scripts while recording video, use a teleprompter...
Process video and audio with correct codec selection, filtering, and encoding settings.
Use when performing video/audio processing tasks including transcoding, filtering, streaming, metadata manipulation, or complex filtergraph operations with FFmpeg.
Generate videos from text prompts or reference images using OpenAI Sora. โ USE WHEN: - Need AI-generated video from text description - Want image-to-video (animate a still image) - Creating cinematic/artistic video content - Need motion/animation without lip-sync โ DON'T USE WHEN: - Need lip-sync (person speaking) โ use veed-ugc or ugc-manual - Just need image generation โ use nano-banana-pro or morpheus - Editing existing videos โ use Remotion - Need UGC-style talking head โ use veed-ugc INPUT: Text prompt + optional reference image OUTPUT: MP4 video (various resolutions/durations)
YouTube video summarizer with speaker detection, formatted documents, and audio output. Works out of the box with macOS built-in TTS. Optional recommended tools (pandoc, ffmpeg, mlx-audio) enhance quality. Requires internet for YouTube access. No paid APIs or subscriptions. Use when user sends a YouTube URL or asks to summarize/transcribe a YouTube video.
Recreate low-budget AI video ad workflows using Nano Banana image generation plus Kling 3.0 video synthesis with dialogue, including prompt design, scene planning, cost control, and export handoff. Use when a user wants to produce a cinematic ad quickly (often in a few hours) with a small credit budget, or asks for a Deon-style Nano Banana + Kling pipeline.
If you can imagine it, CellCog can film it. Grand widescreen cinematics with consistent characters โ what previously required million-dollar production budgets, now generated from a single prompt. Short films, music videos, brand films, cinematic productions โ epic compositions, cinematic lighting, visual storytelling at scale. Grand cinema, accessible to everyone.
Generate videos using SiliconFlow API with Wan2.2 model. Supports both Text-to-Video and Image-to-Video.
Create films with AI video generation by managing scripts, prompts, consistency, and production workflows from concept to final cut.
Generate new videos from text prompts, images, or reference inputs using EachLabs AI models. Supports text-to-video, image-to-video, transitions, motion control, talking head, and avatar generation. Use when the user wants to create new video content. For editing existing videos, see eachlabs-video-edit.
Complete A/B video pipeline โ storyboard, Veo 3 batch generation, browser preview with feedback loop, and ffmpeg assembly into final videos. Use when creatin...