AI-powered media editor that includes advanced generation, editing, & post-processing features, with a focus on efficiency & user control. Key idea: allow intermediate file export & local (CPU/GPU) processing to reduce server load.
Core Features Summary
- Text & Drawing Tools
Text input box with font, color (hex/name), size, and position (top, center, bottom)
Painter tools: brush , pen
, pencil, with shading and adjustable width for brush/eraser
Geometry input (circle, square, polygon) with shape size and color control
- Zoom & Precision Editing
Zoom to select exact regions for brush/eraser or regeneration (instead of regenerating entire image/video)
Ability to select part of image/video for AI regeneration (useful for touch-ups)
- Scene & Style Separation
Separate prompt boxes for: scene (or upload scene), characters (or upload custom character), objects (cropped uploads), and style
Style choices: Ghibli, cinematic, oil painting, hyper-realistic, etc.
Camera shot options: close-up, zoom-in, zoom-out
- Audio & Emotion Control
Separate audio input sections:
Dialogue box (character-wise with function-like labels)
Music box (create music with Sonu AI or select existing)
Sound effects box (e.g., climate, background sounds)
Audio settings: pitch, emotional tone, and emotion intensity (scale 1 to 10)
- Visual & Special Effects
Filters: black & white, saturation, hue, brightness, exposure (non-AI real-time)
Effects: snow , blur specific parts of image/video
Lip sync with dialogue
Option to edit or enhance AI-generated videos/images after generation
- Diagram Support
Science-specific diagrams (e.g., chemistry: benzene)
Export diagrams to TikZ for LaTeX
Intermediate File & Local Processing
After AI generation, the system should export an intermediate file (raw editable format)
Users can edit locally in the browser or app (CPU/GPU processing)
After edits, the system can export to MP4/image without hitting the AI servers again
Why This Matters
Reduces server load by offloading post-processing to client devices
Increases user control and allows fine editing without repeated AI calls
Enables modular workflows for creators and educators using AI media