Skip to content

AI that explains what’s happening inside a video in plain English — step-by-step.

Notifications You must be signed in to change notification settings

akhileshmanitiwari06/VisionExplainr

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VisionExplainr

VisionExplainr is a starter project that explains what is happening in short videos using MediaPipe (pose + hands) and simple heuristics. It produces a timeline of events and human-friendly explanations, and can optionally use OpenAI to polish text and gTTS for audio narration.

Quick start

  1. Create & activate venv:
    python -m venv venv
    source venv/bin/activate   # Windows: venv\Scripts\activate
  2. Install requirements:
    pip install -r requirements.txt
  3. Add a short test video at example_inputs/sample_video.mp4 (<= 30s recommended).
  4. Run app:
    streamlit run app.py
  5. (Optional) To enable OpenAI polishing, set OPENAI_API_KEY in your environment.

Notes

  • gTTS requires internet to synthesize audio.
  • MediaPipe works better with clear, well-lit videos.

About

AI that explains what’s happening inside a video in plain English — step-by-step.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages