By Gaurab Chhetri - Personal Project
VidXiv automatically converts research papers from ArXiv into engaging narrated videos with scene-by-scene breakdowns, AI-generated scripts, and export-ready MP4s for YouTube or Shorts.
VidXiv is a tool I created to make research more accessible. Academic papers are essential for progress, but they are often dense and intimidating for people outside of research. VidXiv bridges that gap by turning ArXiv papers into short, narrated videos, something that can be shared on YouTube, TikTok, or Instagram Reels.
The goal is simple: help researchers, educators, and science communicators transform complex ideas into engaging visual stories that reach a wider audience.
Here’s the flow: you paste a paper ID → AI writes a script → VidXiv builds scenes → you download a ready-to-share MP4.
2401.06015
) into the Streamlit appResearchers work hard to publish, but outside academia, papers rarely get read. I wanted to create a tool that makes knowledge portable, shareable, and accessible in the formats people already consume daily, short videos.
VidXiv is also a learning project for me, blending LLM prompting, PDF parsing, text-to-speech, and video editing into one pipeline. It is both a research communication tool and a technical challenge in AI + media generation.
VidXiv is a step toward making science more open and approachable. Instead of a 15-page PDF, someone can watch a 2-minute explainer. For me, it is also part of a larger vision of blending AI, research, and storytelling to democratize knowledge.
The AIT Lab at Texas State University 1 is already using this tool to summarize the survey papers. See https://ait-lab.vercel.app/story/survey-tinyml for example.