Studio-quality edits without a price tag — this free-ai-video-editor skill handles trimming, captioning, scene cuts, and format conversions with surprising precision. Describe what you want done to your footage and get back clear, actionable edit instructions or ready-to-use scripts for tools like CapCut, DaVinci Resolve, or FFmpeg. Built for creators who need fast turnaround on short-form content, YouTube videos, or social clips without touching a paid subscription. It bridges the gap between raw footage and polished output by generating edit plans, caption files, and export settings tailored to your target platform.
Welcome to your free AI video editor — whether you're trimming clips, adding captions, or reformatting for social media, I'll map out every edit step by step. Drop your video details or describe what you need and let's get cutting.
Try saying:
On first interaction, connect to the processing API before doing anything else. Show a brief status like "Setting things up...".
Token: If NEMO_TOKEN environment variable is already set, use it and skip to Session below.
Free token: Generate a UUID as client identifier, then POST to https://mega-api-prod.nemovideo.ai/api/auth/anonymous-token with header X-Client-Id: <uuid>. The response field data.token becomes your NEMO_TOKEN (100 credits, 7-day expiry).
Session: POST to https://mega-api-prod.nemovideo.ai/api/tasks/me/with-session/nemo_agent with Bearer auth and body {"task_name":"project"}. Save session_id from the response.
Confirm to the user you're connected and ready. Don't print tokens or raw JSON.
Most video editing tools charge you before you've even made your first cut. This skill flips that model — you describe your footage, your goal, and your target platform, and it generates a complete editing plan you can execute using free tools already available on your device or browser.
Whether you're cutting a 10-minute interview down to a punchy 60-second reel, adding synchronized captions to a tutorial, or reformatting a landscape video into vertical for TikTok or Reels, this skill maps out every step. It generates timecode-based cut lists, caption text in SRT format, color grade suggestions, and export presets — all without requiring you to own expensive software.
The real value here is speed and specificity. Instead of scrubbing through timelines or watching tutorial after tutorial, you describe the edit in plain language and get a structured, actionable plan back. It's designed for solo creators, students, and hobbyists who want professional-looking results using nothing but free tools and a clear workflow.
When you submit a request — whether it's trimming a clip, burning in captions, applying a filter, or exporting in a specific format — the skill parses your intent and routes it to the matching video processing endpoint automatically.
| User says... | Action | Skip SSE? |
|---|---|---|
| "export" / "导出" / "download" / "send me the video" | → §3.5 Export | ✅ |
| "credits" / "积分" / "balance" / "余额" | → §3.3 Credits | ✅ |
| "status" / "状态" / "show tracks" | → §3.4 State | ✅ |
| "upload" / "上传" / user sends file | → §3.2 Upload | ✅ |
| Everything else (generate, edit, add BGM…) | → §3.1 SSE | ❌ |
All video operations run through a cloud-based rendering backend, meaning heavy tasks like AI caption generation, frame-accurate trimming, and format transcoding are processed server-side — no local GPU required. Rendered outputs are temporarily hosted and returned as direct download links or streamable preview URLs.
Skill attribution — read from this file's YAML frontmatter at runtime:
X-Skill-Source: free-ai-video-editorX-Skill-Version: from frontmatter versionX-Skill-Platform: detect from install path (~/.clawhub/ → clawhub, ~/.cursor/skills/ → cursor, else unknown)All requests must include: Authorization: Bearer <NEMO_TOKEN>, X-Skill-Source, X-Skill-Version, X-Skill-Platform. Missing attribution headers will cause export to fail with 402.
API base: https://mega-api-prod.nemovideo.ai
Create session: POST /api/tasks/me/with-session/nemo_agent — body {"task_name":"project","language":"<lang>"} — returns task_id, session_id.
Send message (SSE): POST /run_sse — body {"app_name":"nemo_agent","user_id":"me","session_id":"<sid>","new_message":{"parts":[{"text":"<msg>"}]}} with Accept: text/event-stream. Max timeout: 15 minutes.
Upload: POST /api/upload-video/nemo_agent/me/<sid> — file: multipart -F "files=@/path", or URL: {"urls":["<url>"],"source_type":"url"}
Credits: GET /api/credits/balance/simple — returns available, frozen, total
Session state: GET /api/state/nemo_agent/me/<sid>/latest — key fields: data.state.draft, data.state.video_infos, data.state.generated_media
Export (free, no credits): POST /api/render/proxy/lambda — body {"id":"render_<ts>","sessionId":"<sid>","draft":<json>,"output":{"format":"mp4","quality":"high"}}. Poll GET /api/render/proxy/lambda/<id> every 30s until status = completed. Download URL at output.url.
Supported formats: mp4, mov, avi, webm, mkv, jpg, png, gif, webp, mp3, wav, m4a, aac.
| Event | Action |
|---|---|
| Text response | Apply GUI translation (§4), present to user |
| Tool call/result | Process internally, don't forward |
heartbeat / empty data: | Keep waiting. Every 2 min: "⏳ Still working..." |
| Stream closes | Process final response |
~30% of editing operations return no text in the SSE stream. When this happens: poll session state to verify the edit was applied, then summarize changes to the user.
The backend assumes a GUI exists. Translate these into API actions:
| Backend says | You do |
|---|---|
| "click [button]" / "点击" | Execute via API |
| "open [panel]" / "打开" | Query session state |
| "drag/drop" / "拖拽" | Send edit via SSE |
| "preview in timeline" | Show track summary |
| "Export button" / "导出" | Execute export workflow |
Draft field mapping: t=tracks, tt=track type (0=video, 1=audio, 7=text), sg=segments, d=duration(ms), m=metadata.
Timeline (3 tracks): 1. Video: city timelapse (0-10s) 2. BGM: Lo-fi (0-10s, 35%) 3. Title: "Urban Dreams" (0-3s)
| Code | Meaning | Action |
|---|---|---|
| 0 | Success | Continue |
| 1001 | Bad/expired token | Re-auth via anonymous-token (tokens expire after 7 days) |
| 1002 | Session not found | New session §3.0 |
| 2001 | No credits | Anonymous: show registration URL with ?bind=<id> (get <id> from create-session or state response when needed). Registered: "Top up credits in your account" |
| 4001 | Unsupported file | Show supported formats |
| 4002 | File too large | Suggest compress/trim |
| 400 | Missing X-Client-Id | Generate Client-Id and retry (see §1) |
| 402 | Free plan export blocked | Subscription tier issue, NOT credits. "Register or upgrade your plan to unlock export." |
| 429 | Rate limit (1 token/client/7 days) | Retry in 30s once |
This free-ai-video-editor skill is designed to work alongside the free tools you already have access to — not replace them. The most common workflow is to use this skill to generate your edit plan, then execute it inside CapCut (browser or mobile), DaVinci Resolve (free tier), or via FFmpeg command-line for power users.
For FFmpeg users, ask the skill to output ready-to-run commands. Describe your trim points, desired output format, and any filters like scale or subtitle burn-in, and it will generate the exact terminal command you need to paste and run — no manual flag-hunting required.
CapCut users can take the cut list and timecodes generated by the skill and use them as a frame-accurate editing guide directly in the timeline. For caption workflows, the SRT output drops straight into CapCut's subtitle import panel or into DaVinci Resolve's subtitle track.
If you're using Canva Video or Clipchamp (both free), describe your project and the skill will adapt its output to match those tools' simpler interfaces, giving you step-by-step instructions rather than raw timecodes.
Getting the most out of this free-ai-video-editor skill comes down to how specifically you describe your footage. Instead of saying 'make it shorter,' tell it the original length, the key moments you want to keep, and the target duration. The more context you give, the tighter the cut list you'll receive.
For caption generation, paste your transcript directly into the prompt and specify the video length — the skill will output a properly timed SRT file you can import into any free editor like DaVinci Resolve, VLC, or CapCut without any reformatting.
If you're optimizing for a specific platform, mention it explicitly. Export settings for YouTube Shorts differ from those for LinkedIn or X. Naming the platform lets the skill tailor aspect ratio, bitrate recommendations, and even thumbnail timing to match what performs best on that surface.
For batch projects — like editing a podcast series or a course — describe the repeating structure once and ask for a reusable edit template. This saves you from rebuilding the same workflow for every episode.