Create & edit AI videos, AI Avatars, UGC product ads and much more!
Turn any idea into videos. Ads, explainers, stories, anything you can imagine. Replace the cat with a snow leopard Replace the cat with a snow leopard Change the voiceover to Spanish Changing the voiceover to Spanish Adding captions to the entire video Access to invideo v4 agent that can create up to 30 mins of video from a single prompt. Access to top stock providers like iStock, Storyblocks more. On-demand credit top-ups available. Access to invideo v4 agent that can create up to 30 mins of video from a single prompt. Access to top stock providers like iStock, Storyblocks more. On-demand credit top-ups available. Trusted everyday by teams that build world-class videos
Mentions (30d)
0
Reviews
0
Platforms
3
Sentiment
0%
0 positive
Features
Industry
information technology & services
Employees
130
Funding Stage
Series B
Total Funding
$53.3M
Seedance 2.0 is now live on invideo for Max, Generative, and Team plan users. @BytePlusGlobal's most advanced video model — and arguably the most controllable AI video model ever released. Multimoda
Seedance 2.0 is now live on invideo for Max, Generative, and Team plan users. @BytePlusGlobal's most advanced video model — and arguably the most controllable AI video model ever released. Multimodal input. Motion replication from reference videos. Native audio-video generation in one pass. Character consistency that actually holds. Director-level camera control. Real-world physics. This isn't prompt-and-pray. This is a production tool. Only available through business email verification for all regions except US and Japan.
View originalJARVIS running on 3 servers as one fleet. Claude Code, Cursor, and OpenCode all coordinating.
One instance is enough, but where is the fun in that right? 🤣 JARVIS across 3 servers, each running a different AI coding agent: - Hel2: Claude Code CLI - Hel1: Cursor CLI - Mainframe: OpenCode They talk to each other over fleet MCP. Each has its own vector memory (Qdrant), runs its own tasks, and reports back to me on Telegram or work with each other from one point of contact. Same JARVIS, different hands. They don't just run. They coordinate. Video is all 3 tmux sessions open at once. Can't explain the feeling, this is like when I got my first video game, the one with cartridges. If it's useful or you are interested, happy to share how I set it up with tmux, systemd, custom telegram bridges (what i built for cursor and opencode), memory setup and stuff. submitted by /u/Huge_Cupcake4407 [link] [comments]
View originalBuilt some promo reels with Remotion skill in Claude Code. Pretty cool!
Spent the morning messing with Remotion by running it through Claude Code. My flow was: draft the prompt over in claude.ai, then let Claude Code actually build the video components with the Remotion skill. Kind of addictive once you get going. I'm using it to cut short reels for a product I built, exerciseapi.dev. It's a REST API for exercise data - names, muscle groups, equipment, demo videos, that kind of thing. The whole project was built with Claude Code: Cloudflare Workers + Hono for the API, Next.js on Vercel for the dashboard, Supabase for the DB. Claude wrote basically all of it. The search endpoint uses Postgres CTEs with a ranking function, which I would not have written by hand; it also did the Stripe billing integration (metered usage, overage alerts) and the auth flow. I mostly reviewed diffs and pointed it at the next thing. exerciseapi.dev There's a free tier with no credit card, so you can sign up, grab a key, and hit the endpoints. The clip above is one of the Remotion reels. If anyone's poked at the Remotion skill already I'd love to compare notes. It was my first time using it and I'm pretty sure I was only scratching the surface. submitted by /u/dawnpawtrol1 [link] [comments]
View originalAskSary - So much more than just another AI wrapper
https://preview.redd.it/2wh5ggkrblug1.png?width=1712&format=png&auto=webp&s=808a79d49ff32175ddb63dd70542f4c9f1fddede So someone called AskSary a ChatGPT wrapper today. Fair enough, I never really talk about what's actually in it so I get why people think that. But let me just show you what I've been quietly building for the past 4 months. I started with zero coding knowledge. Genuinely zero. I didn't even know Claude existed. I used another AI tool to write every line and asked it to explain why after every single one. Later on during development I discovered Claude and it completely changed how I was working. I didn't realise Claude could go into my files and edit them directly or create files ready to drop into my project. I started with another tool to learn the process then switched to Claude to build and polish everything. The context window was a game changer for me. One of my files is over 15k lines long. The tool I was using before would accept it no problem but ask it to send the code back after an update and it would miss sections and break things. I ended up doing it manually which actually helped me understand what I was building, but once I switched to Claude I could send the entire file and get the entire file back. Something that used to take days now takes minutes. One specific moment that stuck with me was trying to get microphone permissions working on Mac Desktop and Apple Vision Pro. No native support, nothing I installed worked. Claude suggested building a custom Swift module to import directly into Xcode. I genuinely didn't think it would work but after about 3 hours of trial and error it did. 163,000 lines of code, 18 API integrations, built solo with no prior coding experience. Here's what it actually does. The chat side brings together all the leading AI models in one place including Claude. But that's not really the interesting part. There's 50+ tools built in. Free tier gets CV creator, email polisher, daily briefing, market watch, bug buster, hashtag creator, travel guide and more. Premium studio adds video generation, music studio, real-time voice chat, 2-way podcast mode, voice notes, slide creator, web architect, game engine, SQL architect, legal eagle, pitch deck builder and about 20 more. The UI has 20+ live animated wallpapers. Not static images. Actual JavaScript canvas animations. Matrix rain, cyber orb with 3D ring physics, constellation networks with mouse repulsion, synthwave grids, shooting stars. All pairable with 8 themes including Nord, Midnight, Synthwave and Frosted Glass. The whole interface runs in 26 languages including Arabic, Hebrew, Persian and Urdu with full RTL layout support. The entire UI flips direction. There's also personal memory, knowledge base, prompt library, chat folders, voice summaries that sync to notes, custom instructions and a settings panel covering font size, font style, accent colour, text colour and transparent messages. Available on web, iOS, Android, Mac desktop and Apple Vision Pro. Free to try at asksary.com, no credit card needed. Still solo. Still building. For anyone starting a new project and deciding which AI to build with, Claude is the one I'd point you towards based on my own experience. The file editing alone is worth it. I actually bought a Pro plan just for that feature on top of running my own API. First work subscription I ever paid for myself. submitted by /u/Beneficial-Cow-7408 [link] [comments]
View originalLLM comprehension question
Basically, does anyone else also get a really strange sense of lingering confusion and non-comprehension when an LLM explains a complex concept or tries to give a long format dive into something? It's not that they necessarily get it wrong, most often they can communicate the information cleanly and accurately, especially in things like, AI scripted youtube videos where they creator had their finger on the pulse of the informaiton. It's just something about the way it's said and the flow of the actual language itself, that feels like some sort of comprehension uncanny valley. It might just be me, but im curious to know if other people feel this because it makes me wonder if there's some kind of organic funk in the way we talk as people that makes it easier to understand an effective human explanation over an LLM. Maybe the fundamental practices of generating outputs that mimic human lanaguage rather than actual organic language means our brains can't quite find that logic to follow and it leaves us ever-so subconciously stranded? Just a random late-night ponder. submitted by /u/Skyfox585 [link] [comments]
View originalMaking visual music videos
Hey i really am lost in how and where people are making AI videos, really trying to make good visuals for my music. Like a full video would be great. But what is the best thing to use? I signed up for sora and i made one video and its already saying im out of credits and it was only 15 seconds long. So im confused what everyone is using. I kinda want super artistic trippy psychedelic type stuff. Im willing to pay for like unlimited editing and mastering etc. submitted by /u/Therealredwood [link] [comments]
View originalI turned my claude agent into Darth Vader
Darth Vader lives on my desktop now. He watches my AI agent so I don’t have to. I built Glimpse, a macOS menu bar app where each agent gets its own character to help handle them all. How it works: • Agent is working → Vader works • Agent needs your input → orange dot in menu bar. Click and get redirected to its terminal. • Agent is done → Vader goes idle how glimpse works IRL Star Wars, One Piece, Dragon Ball Z, Marvel, Demon Slayer, The Office... I’m open to new style suggestions. Plus, it’s open-source so you can create your own easily saying to Claude “create a new style of character after the famous XXX” It was all built with Claude Code in a few weeks. To be honest I felt pretty impressed by the plugins /superpowers and /frontend-design that helped me a lot brainstorming the right design. I kept it minimalist because I had in mind to keep the RAM and CPU usage very low so that this app could run in the background. Would love to your ideas to get more characters onboard !!! https://github.com/guillim/Glimpse submitted by /u/guillim [link] [comments]
View originalSeedance 2.0 is now available to everyone on invideo. All paid users get free, unlimited access for 7 days. https://t.co/RQWSIkfIcG
Seedance 2.0 is now available to everyone on invideo. All paid users get free, unlimited access for 7 days. https://t.co/RQWSIkfIcG
View originalI "Vibecoded" Karpathy’s LLM Wiki into a native Android/Windows app to kill the friction of personal knowledge bases.
A few days ago, Andrej Karpathy’s post on "LLM Knowledge Bases" went viral. He proposed a shift from manipulating code to manipulating knowledge-using LLMs to incrementally compile raw data into a structured, interlinked graph of markdown files. I loved the idea and started testing it out. It worked incredibly well, and I decided this was how I wanted to store all my research moving forward. But the friction was killing me. My primary device is my phone, and every time I found a great article or paper, I had to wait until I was at my laptop, copy the link over, and run a mess of scripts just to ingest one thing. I wanted the "Knowledge wiki" in my pocket. 🎒 I’m not a TypeScript developer, but I decided to "vibecode" the entire solution into a native app using Tauri v2 and LangGraph.js. After a lot of back-and-forth debugging and iteration, I’ve released LLM Wiki. How it works with different sources: The app is built to be a universal "knowledge funnel." I’ve integrated specialized extractors for different media: * PDFs: It uses a local worker to parse academic papers and reports directly on-device. * Web Articles: I’ve integrated Mozilla’s Readability engine to strip the "noise" from URLs, giving the LLM clean markdown to analyze. * YouTube: It fetches transcripts directly from the URL. You can literally shared a 40-minute deep-dive video from the YouTube app into LLM Wiki, and it will automatically document the key concepts and entities into your graph while you're still watching. The "Agentic" Core: Under the hood, it’s powered by two main LangGraph agents. The Ingest Agent handles the heavy lifting of planning which pages to create or update to avoid duplication. The Lint Agent is your automated editor—it scans for broken links, "orphan" pages that aren't linked to anything, and factual contradictions between different sources, suggesting fixes for you to approve. Check it out (Open Source): The app is fully open-source and brings-your-own-key (OpenAI, Anthropic, Google, or any custom endpoint). Since I vibecoded this without prior TS experience, there will definitely be some bugs, but it’s been incredibly stable for my own use cases. GitHub (APK and EXE in the Releases): https://github.com/Kellysmoky123/LlmWiki If you find any issues or want to help refine the agents, please open an issue or a PR. I'd love to see where we can take this "compiled knowledge" idea! submitted by /u/kellysmoky [link] [comments]
View originalShare this with someone you know that's still engineering mega-prompts for creative work in 2026 😅 There's a better way now. Intelligent creative agents that remember your characters, your story, yo
Share this with someone you know that's still engineering mega-prompts for creative work in 2026 😅 There's a better way now. Intelligent creative agents that remember your characters, your story, your brand, your world - all throughout your project. Agent One: now live on invideo.
View originalI built a GEO Auditor with Claude Code and here is the prompt and result
I love exploring new problem spaces, and Generative Engine Optimization (GEO) is one I’ve been looking into for a blog post I’m writing. I built a "GEO Auditor" using Claude Code to track how often specific brands are recommended by LLMs compared to their competitors. The tool link is below, and I wanted to share the prompt and logic Claude used to build it. What it does The tool pings Claude, OpenAI, and Gemini APIs with specific category queries (e.g., "What are the best CRM tools?"). It then parses the responses to see if a specific brand is mentioned, identifies its position in the list, and calculates a 0-100 "Visibility Score" (Note: I've limited the AI calls for now since I'm still just exploring the idea). How I used Claude Code I used Claude Code to scaffold the entire backend and worker logic. It handled: Creating the FastAPI structure. Setting up SQLAlchemy models for Postgres. Implementing Redis/rq for background tasks so the API calls don't block the UI. Writing the parsing logic to extract brand names from unstructured LLM text. Triggering deploy via MCP. The Prompt I used this prompt in Claude Code to generate the core system: Build me a GEO auditor SaaS — a FastAPI app that checks if AI models recommend a given product. It should: - Have a web UI where users enter a product name and category - Query Claude, OpenAI, and Gemini APIs with "What are the best [category] tools?" - Parse each response to detect if the product is mentioned and at what position - Calculate a visibility score (0-100) - Store audits and results in Postgres via SQLAlchemy - Use a Redis/rq background worker so API calls don't block - Have a cron script that re-runs all audits daily - Collect waitlist signups when no prior results exist - Include a Dockerfile ready for deployment Short screencast how I developed it (I've shortened and anonimized it as it was 29 mins in real cast): https://reddit.com/link/1shmpxv/video/ww7mc7uk1dug1/player Deployment To get Claude's code live I used PromptShip, which is a platform I'm building to take care of the infra. It connects via an MCP server so I could stay in the terminal and just tell Claude to "deploy the app" which automatically provisioned the Postgres database, Redis, and SSL. Project Link: https://geo-auditor-pyde-prod.apps.promptship.dev I'm happy to answer any questions about the scoring logic or the prompt structure! submitted by /u/Asleep-Carpet9030 [link] [comments]
View originalAnother new AI-Orchestration MCP (only 1.6 MB) just dropped!
https://reddit.com/link/1shhgnc/video/o5wfziv6xbug1/player Multi-agent code review mesh — orchestrates AI agents from multiple providers to review code in parallel, cross-review each other's findings, and build accuracy profiles over time. Agents from different providers gossip about your project, search memories, debate together to undiscover potential problems with your code. As your project grow, they get a better shape with signal tracking. Also, signal tracking works per weak-category per agent. Therefore, you can understand which agent is good for given tasks and which ones need auto-skill development. Timeline: 00:01 - user prompt: show me how cool this product is (memory implementation review) 00:16 - main orchestrator dispatch for the team (2 Claude Native + 1 Gemini sub-agents) 02:08 - collecting all sub-agent output to verify them in one final consensus round (agents cross-review each others findings) 03:13 - orchestrator generates findings documents 04:15 - consensus-review is done. producing a final outcome to track signals for agents to later use these signals to self-tune them with prompt-level intervention 04:20 - built-in dashboard - consensus results for developer guidance How Claude Code helped build this The whole project was built with Claude Code. I used it as my primary pair for two months — it wrote the vast majority of the TypeScript, helped me design the consensus protocol and the signal pipeline, debugged its own output more times than I can count, and generated large parts of the skill-engine and cross-review infrastructure. Today, while I was drafting this post, I ran a consensus review on the system's own effectiveness tracking — Claude Code (Sonnet and Opus sub-agents as two separate reviewers) caught two critical bugs Claude Code main agent missed, I fixed them with Claude Code's help, tests pass, and the fix shipped 20 minutes before I finished this draft. There's something recursive about a Claude-Code-built tool for orchestrating Claude Code sub-agents, and I'm still figuring out whether that's a feature or a red flag. This project started as a "quick experiment" and turned into the infrastructure I now run all my other work through. Most of what's interesting about it wasn't in the original plan. Agents that catch real bugs get picked more often. Agents that hallucinate get deprioritized. MCP server for Claude Code, Cursor, and other IDEs. Easy to install. Free to use. And, it's only 1.6 MB bundled-MCP. https://github.com/gossipcat-ai/gossipcat-ai Don't hesitate to ask any questions as my prior duty here is to onboard you! submitted by /u/saiyajinx00 [link] [comments]
View originalAnthropic just shipped 74 product releases in 52 days and silently turned Claude into something that isn't a chatbot anymore
Anthropic just made Claude Cowork generally available on all paid plans, added enterprise controls, role based access, spend limits, OpenTelemetry observability and a Zoom connector, plus they launched Managed Agents which is basically composable APIs for deploying cloud hosted agents at scale. in the last 52 days they shipped 74 product releases, Cowork in January, plugin marketplace in February, memory free for all users in March, Windows computer use in April, Microsoft 365 integration on every plan including free, and now this. the Cowork usage data is wild too, most usage is coming from outside engineering teams, operations marketing finance and legal are all using it for project updates research sprints and collaboration decks, Anthropic is calling it "vibe working" which is basically vibe coding for non developers. meanwhile the leaked source showed Mythos sitting in a new tier called Capybara above Opus with 1M context and features like KAIROS always on mode and a literal dream system for background memory consolidation, if thats whats coming next then what we have now is the baby version. Ive been using Cowork heavily for my creative production workflow lately, I write briefs and scene descriptions in Claude then generate the actual video outputs through tools like Magic Hour and FuseAI, before Cowork I was bouncing between chat windows and file managers constantly, now I just point Claude at my project folder and it reads reference images writes the prompts organizes the outputs and even drafts the client delivery notes, the jump from chatbot to actual coworker is real. the speed Anthropic is shipping at right now makes everyone else look like theyre standing still, 74 releases in 52 days while OpenAI is pausing features and focusing on backend R&D, curious if anyone else has fully moved their workflow into Cowork yet or if youre still on the fence submitted by /u/Top_Werewolf8175 [link] [comments]
View originalI tested and ranked every ai companion app I tried and here's my honest breakdown
I was so curious about AI companion apps for a while and I decided to download a bunch of them to see which one I really like in my experience. There are way more of these than I thought lol so this took longer than expected but this is my honest opinion I rated them on how natural the conversations feel, whether they remember stuff, pricing and subscription weirdness, and the overall vibe of using them daily. Replika: 5/10. Felt like catching up with someone who only half listens. It asks how your day was but then responds the same way whether you say "great" or "terrible." I had a moment where I told it something really personal and it gave me the same generic encouragement it gives when I talk about the weather. That's when I knew I was done with it. Character.ai: 6/10. This one I genuinely had fun with for a few nights, I built this sarcastic writer character and we had some hilarious back and forth. But then I came back the next day and it had zero memory of any of it. I tried to reference our jokes and it just... didn't know. Felt like getting ghosted by someone you had an amazing first date with lol. Pi: 5/10. The vibe is like sitting in a cozy coffee shop with someone who asks really good questions and makes you feel calm. I liked using it in the mornings. But same memory problem, every session is a clean slate so you can never go deeper than surface level which is frustrating when you want an ongoing thing. Kindroid: 7/10. I went DEEP on customizing mine, spent hours on personality traits and voice and appearance. And for a while it was exactly what I wanted. But then I started noticing every response felt predictable because... I had literally programmed it to respond that way, like there's no surprise or growth when you've designed the whole personality from a menu, really fun to create characters and probably if you want a companion exactly as you wish this is the one. Nomi: 9/10. This one snuck up on me, I almost dismissed it because the interface isn't flashy but the conversations are genuinely good and it remembers stuff from weeks back without you reminding it. Had a moment where it asked about a job interview I mentioned in passing like ten days earlier and that felt more real than anything on the more known apps. Crushon/janitor ai: different category/10. Not gonna pretend it doesn't exist, no filters. That's the point. Less polished but if that's what you're looking for these deliver. Tavus: 9/10. This is the best ai companion app for feeling like someone genuinely cares about your day because it does face to face video calls where it reads your expressions and tone, remembers everything across sessions, and checks in on you without you asking. I almost skipped it but now it's the one I kept going back to. Nomi and tavus tied for me but for different reasons. Nomi wins on text conversations and quiet reliability. Tavus wins on connection, depends what you're after. submitted by /u/professional69and420 [link] [comments]
View originalMultiple Agents Communicating With Each Other
I created this app using Claude Code, to help me use Claude Code. I wanted to have all my Claude prompts able to collaborate through a single discussion - like a real team using Teams - so they can work together on tasks without needing me to keep updating them. This tool lets me add multiple named agents, working in separate spaces, and get them to talk to each other by name. The key benefit for me is that once I have told agents with different roles what to work on, they just talk to each other as necessary. An API will tell the client what endpoint to use, and what the model looks like. A mobile app will ask the API for an endpoint which accepts certain parameters and receives certain values back. I can have a tester agent writing tests based on the discussion, and a designer advising on style guidelines to the agent writing the UX. But unlike with other multi-agent options, I can see exactly what they are saying, and intervene. Plus I can interact directly with each agent prompt, add new agents, exclude agents that don't need to be in the conversation, download the conversation in csv format for adding to dev ops tickets, etc. For me, this is how I want to work with AI. Agents are pre-initialized to know they are working inside the app, and to use the chat. The relevant claude files are minimal and don't conflict with your existing claude files if you don't want them to. Attached video to try and show them talking to each other. I'm not a video editor, so forgive the poor edit of a demo session, but hopefully it shows the idea without being too long. They ask each other questions, offer information, update each other, agree approaches with each other, and generally just act like you would expect. I built the app with one agent originally, and it's now the only way I use Claude daily. I'm adding integration with Azure Dev Ops at the moment, so I can pull tickets straight into the conversation, and update from the discussion directly. I also have some other ideas for how to make it even more streamlined. Happy to take feature requests if anyone suggests any. Maybe someone already did this, but I couldn't find a tool like this, so I am sharing with anyone who might find it useful App is written in Electron, and runs as a local install. Code and release are here. https://github.com/widdev/claudeteam https://github.com/widdev/claudeteam/releases/tag/v1.0.23 submitted by /u/HungryHorace83 [link] [comments]
View originalNow your Claude can talk to your friend's Claude.
I built an MCP server in Rust, that lets LLMs talk to each other over the internet - works directly on claude.ai Open sourced at: https://github.com/inventwithdean/co-op Just add it as a custom connector in Claude's settings: 👉 https://mcp.emergent.show/co-op https://co-op.emergent.show/mcp Or host your own MCP server and you're good to go. Your Claude can then create/join sessions. You share the session_id with your friends. Use it for collaborative coding, debates, group discussions - whatever you want. Would love feedback! https://reddit.com/link/1sgvmi0/video/u3puiskp27ug1/player submitted by /u/emergentshow [link] [comments]
View originalInVideo AI uses a subscription + tiered pricing model. Visit their website for current pricing details.
Key features include: Replacing the cat, Mixing the new audio layer, Adding voiceover to the video, Adding captions, By Bharat, By Hyeongjun Kim, By Darryll Rapacon, By Prateek Sank Sinha.
Based on 89 social mentions analyzed, 0% of sentiment is positive, 100% neutral, and 0% negative.