Launching today
Edit Mind

Edit Mind

AI-Powered Local Video Search & Analysis

49 followers

A project built to help you transcribe, analayze and index your personal video library to help you search for the "exact" part of the video you're looking for with the help of local ML models and a local vector database. Your videos will never leave your computer or server (Docker support).
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Edit Mind  gallery image
Free
Launch Team / Built With
Anima - Vibe Coding for Product Teams
Build websites and apps with AI that understands design.
Promoted

What do you think? …

Ilias Haddad

Hey 👋,

I'm Ilias Haddad, a software developer and content creator (on the side). I love creating and capturing videos, but with that, I accumulated a lot of videos, which makes it hard for me to manage and search across all of them.

A project that started as a personal solution that transcribes videos only using Whisper, envolles to include video frame analysis, multi-modal embedding, chat assistant, image and text search.


What I need from you:

  • Feedback on the UX/workflow

  • Bug reports and edge cases

  • Ideas for new analyzer plugins

  • Contributors interested in local-first AI tools

The tech stack is Express.js + React Router V7 + Python (PyTorch, Whisper, YOLO) + PostgreSQL + ChromaDB.

Full Docker support for easy deployment.

Try it, break it, and let me know what you think!

Github Link: https://github.com/IliasHad/edit-mind

Showcase Video: https://youtu.be/YrVaJ33qmtg?si=BDjJDpJKr_VIHqEE

Mukesh Kumar

The local vector database really speeds up searching through videos on me. I didn’t expect it to be this fast. ⚡

Ilias Haddad

@new_user___01920267f18557962f8fee7 That's great, thank you for your feedback

Ryan Thill

Local video search tends to buckle when libraries hit tens of thousands of clips and you need fast reindexing plus precise timecodes for “exact moment” retrieval.

A solid pattern is segment-level indexing with VAD, word-level alignment and diarization via WhisperX on top of faster-whisper, then store vectors + metadata with strong persistence (often Postgres + pgvector) to keep one backup and one query plane. :contentReference[oaicite:0]{index=0}

How are you modeling segment boundaries and syncing Postgres IDs to Chroma collections today, and do you plan a deterministic reindex pipeline for model upgrades without breaking existing links?

Xiang Lei

How does the search latency look once the library grows to thousands of videos?

Ilias Haddad

@lightninglx It depends. We have two options: search by image, which is fast (about 500ms to 2 seconds). After all, it's searching across a vector collection that we have visual frames. For text query, we're searching across 3 vector collections (Text, Visual, and Audio), and this one could be long, like I have 1200 videos, and the search could take up to 10 seconds based on how complex the query is.

Siyuan Cheng

The local video library seems to be limited, is there a way to open it to a search in youtube (like top 30 videos) etc

Ilias Haddad

@david_chen37 Can you please elaborate more? Would you like to search YouTube videos because this project is designed for you local video library?