- NotebookLM now turns notes and documents into "cinematic" videos with AI-generated narrative and visuals.
- The feature combines models such as Gemini 3, Nano Banana Pro and Veo 3 for scripting, narration and animations.
- The videos are based solely on the user's files, reducing hallucinations and being useful for study and presentations.
- For now, it is limited to English, to those over 18 years of age and already subscribers of Google AI Ultra, with a maximum of 20 videos per day.

The race to create AI-generated video really useful It's not just in the entertainment sector: it has also fully entered the world of study and productivity. Google has just taken a significant leap with NotebookLM, its AI-powered research platform, by allowing users to transform their notes and documents into "cinematic" videos designed to better explain dense and complex content.
This new function of cinematic video summaries It fits within Google's broader commitment to AI-powered video, alongside its Veo model, the Flow tool, and experimental projects like Project Genie and the AI features in ChromeOSThe idea is clear: you don't have to rack your brains manually designing presentations or videos, and AI takes care of the narrative, pacing, and visual resources so you only have to review the final result.
What are NotebookLM cinematic videos and how are they different?
NotebookLM is an AI-powered research platform that lets you Upload notes, PDFs, articles, or complex documents and work on them: summarizing, asking questions, organizing ideas, or generating derivative content. Until now, one of its most striking features was video summaries, which basically created something similar to a slideshow with AI-generated voiceover.
With the arrival of the new Cinematic Video SummariesGoogle wants to go a step further. Instead of simply showing static slides while a voice reads the summary, NotebookLM generates much more elaborate audiovisual pieces, with animations, movement, and a visual style closer to a presentation. explanatory mini-documentary than a PowerPoint with audio.
According to Google, these new videos not only condense the content, but also They structure the information with a clear narrative.The video consists of an introduction, development, and conclusion, supported by visual resources that reinforce key ideas. The tool aims to give the video rhythm, coherence, and a certain "production" feel to help maintain the viewer's attention, especially when dealing with complex topics.
While the previous video mode focused on offering a Simple presentation with automatic narrationThis cinematic approach aims to create something more akin to a short documentary: with linked scenes, transitions, synthetic camera movement, and visuals that accompany each part of the explanation. The goal is for you to consume complex information in a more engaging way, almost as if you were watching an educational video on YouTube, but created directly from your own notes.
However, one of the big questions among those who follow these developments is to what extent these videos are actually “cinematic” or simply more eye-catching than slides with voiceover. There are still reasonable doubts as to whether the leap in quality compared to traditional presentations will be as spectacular as the name promises, or whether, in the end, it will be more like a souped-up version of the video summaries that already existed.
How AI-powered video generation works in NotebookLM
The magic of this function is not based on a single model, but on the coordinated combination of several advanced AIs from Google. To create these cinematic videos, NotebookLM relies on three technological pillars: Gemini 3, Nano Banana Pro and the house video model, identified in some sources as Video 3 or Veo 3, depending on the context.
In this system, Gemini 3 acts as “creative director” of the projectBased on the documents you upload to NotebookLM—notes, reports, academic papers, manuals, long PDFs, etc.—Gemini analyzes the content, identifies the main concepts, and decides how to turn all that material into a coherent audiovisual story.
This “director” defines the narrative skeleton of the videoThis includes determining the introduction, the main body of the piece, the examples or quotes to highlight, and how to conclude the piece so that the viewer clearly grasps the key points. It also defines the overall visual style, tone, approximate duration, and how to link the different sections together without creating a simple collage of scenes.
For its part, the video model (Veo 3 / Video 3) is responsible for generate the visual resources and animations that you'll see on screen. This is where Google is trying to distance itself from old slide-based presentations: the goal is to have real movement, more complex compositions, camera effects, and a more "cinematic" staging within the current limitations of generative AI.
Nano Banana Pro comes into play as a support component, with tasks that can range from optimize content flow They even help ensure consistency between the narration and the visuals in each video segment. Together, these models help align the narrative, audio, and visuals, preventing the script and images from appearing to be in disarray.
Step by step: from dense document to mini-documentary
The workflow within NotebookLM is designed so that anyone with access to the tool can Convert your study material into video without editing knowledgeThe general process, broadly speaking, follows these steps:
First you have to upload your notes or reference files to NotebookLM. These can be long documents, multi-page PDFs, compilations of class notes, work reports, or even a mix of various file types. It's important that the content is reasonably organized, because the quality of the final video will depend largely on how the original information is structured.
Once your documents are uploaded, the platform itself allows you to organize and group the contentIt's worth taking a few minutes here to clarify which blocks are most important, which chapters or sections you want to prioritize, and which parts might not need to be included. While AI can do a good job extracting the essentials, clearly defining the scope makes its task easier and usually yields better results.
Once you have the base ready, it's time to activate the Video Overview function within NotebookLM. That's where the option to select the mode appears. Cinematic Video SummariesThis is the new feature we're discussing. By choosing this mode, you're telling the system that you don't want a simple presentation, but a more elaborate audiovisual piece.
From that point, NotebookLM puts its models into operation: analyze your notes, generate the script, create the automatic narration (the voice-over) and produces the animations and visuals with AI. The result is a video with a narrative structure, transitions between blocks, scenes with movement, and a more "polished" presentation of the concepts that were scattered throughout your initial material.
Finally, a key phase arrives: review and adjust the resulting videoAlthough the idea is for AI to do most of the work, it's still important that you watch the video, check if the ideas are well represented, if anything important is missing, or if there are any parts that should be cut or clarified. With minor adjustments, you can get the video ready for studying, sharing with your team, or presenting in class or at a meeting.
Practical uses and advantages for students, professionals, and researchers
One of the great advantages of this feature is that the videos are generated exclusively from the notes and files that you upload to NotebookLM. This means that, instead of inventing content or supplementing it with external information, the AI sticks to your own material. This reduces one of the typical problems with some generative tools: hallucinations, irrelevant data, or ideas that don't appear anywhere in your documents.
In practice, this makes NotebookLM especially useful for working with complex investigationsIf you've spent hours compiling papers, technical reports, or extensive documentation, being able to turn all of that into an explanatory video that summarizes the main ideas and presents them with a clear narrative can save you a lot of time when sharing results with others.
For students, this feature opens the door to a different way of studying extensive topics: instead of reviewing notes of dozens of pages over and over again, you can generate personalized educational videos with your own notes. Watching and listening to the content in audiovisual format helps many people retain information and makes it easier to tackle theoretical blocks that, on paper, are a real struggle.
In the professional environment, the cinematic video summaries They can be used to prepare internal presentations, explain the results of a complex project to a non-technical team, or condense a lot of corporate documentation into a visual resource that can be viewed in just a few minutes. It's a way to democratize access to information within a company without each department having to produce its own videos from scratch.
It also opens up an interesting field for new forms of visual learningThis applies to both formal education and continuing professional development. Teachers, trainers, or educational content creators can use NotebookLM to generate visual materials from lengthy texts, and then refine those videos or integrate them into larger courses. The key is that the raw material is your own documents, not generic content generated without clear sources.
Limitations, available language and access model
As is often the case with many of Google's new features in artificial intelligence, this function comes with certain access restrictions at launchAt the moment, the Cinematic Video Summaries are only available in English, which significantly limits their usefulness for those who work primarily in other languages, especially in educational contexts where the language of the materials is crucial.
Furthermore, this feature is reserved for users with a subscription. Google AI UltraIn other words, simply having NotebookLM isn't enough: you need the premium layer of Google's AI offering, designed for those who require the most advanced capabilities. It also requires being over 18 years old, which puts it out of reach for younger students who might be the ones who would benefit most from this type of resource.
Another important point is that Google has established a daily creation limitEach user can generate a maximum of 20 cinematic video summaries per day. This is more than enough for most people, but it's worth considering if you plan to use the tool intensively, for example, to convert a large amount of documentation into video at once.
All of this clearly places the feature in the segment of those who can invest in subscriptions and work, at least for the time being, in environments where English is the dominant languageIt is expected that, if the feature becomes established and proves stable, Google will expand the number of languages supported and make access more flexible, but for now its use is limited to a relatively specific niche.
This strategy fits with the rest of Google's AI-powered video roadmap: many of the most cutting-edge features have first appeared in English and in paid plans, as was the case with the improvement of its Veo video model, the expansion of access to Flow, or the demos of Project Genie, that experiment capable of creating small interactive worlds like a video game based on simple descriptions.
The videos are “cinematic”… to what extent?
The term “cinematic” sounds very appealing in marketing, but it also generates a certain expectation difficult to meetMany users are wondering if these new summaries are truly comparable to a high-production video, or if, in practice, they are more like the usual video summaries with a few additional transitions and effects.
The big difference from the previous mode is that now the system can generate scenes with real movementNot just static slides strung together in sequence. That, on paper, is a significant leap: it allows you to play with shots, synthetic zooms, changes in composition, and animations that make the content seem more like a fluid video than a looped presentation with voice-over.
However, it should be noted that the main objective of this function is not to compete with a professional filming with real camerasIt's not about millimeter-precise lighting and post-production equipment, but about offering a productivity tool that transforms documents into clear and attractive audiovisual pieces with the least possible effort from the user.
In that sense, the adjective “cinematic” probably refers more to the intention of giving the video a more elaborate structure and aesthetics that doesn't promise a level of finish worthy of a blockbuster. Even so, the fact that Google claims to use its most advanced video models leaves the door open to reasonably high visual quality for the intended use: education, outreach, presentations, and quick information consumption.
The opinions of those already testing NotebookLM Ultra will be key to gauging the extent to which these videos represent a radical change from voice-based presentations, or if they fall short. What does seem clear is that, although there are still doubts about the exact level of "cinematic" quality in these summaries, Google's move reinforces a trend: AI no longer just writes texts or summarizes PDFs, it is also starting to convert that knowledge into complex audiovisual formats without using traditional video editing tools.
With all this in mind, it's clear that NotebookLM is trying to become a kind of "express documentary producer" using your notes. For those who work daily with tons of written information, having the option to press a button and get a narrated, animated, and structured video that condenses all that material can make a considerable difference in how we understand, share, and study complex content.
Overall, the arrival of the cinematic video summaries in NotebookLM This adds to Google's series of moves in the field of AI-powered video—improvements to Veo, expansion of Flow, experiments like Project Genie—and points to a future in which much of the heavy lifting of producing explanatory materials will fall to artificial intelligence models, while people focus more on choosing the right sources, reviewing the result, and deciding what is worth sharing.