top of page

Concept Videos

Week 1: February 20, 2026
Duration: 16 seconds

Week 2: March 9, 2026
Duration: 1 minute, 7 seconds

Concept Video Overview

Role: UX Designer/ Prompt Engineer/ Producer/ Composer
Tools: Google Vids, Logic Pro

Concept Summary

This concept video was created to showcase my multidisciplinary skill set across generative AI, UX design, motion design, and audio production. I began by feeding a generative image I created years ago into Google Vids, using a prompt to generate motion with the Veo feature. I repeated this process up to ten times (the platform’s daily limit) to produce multiple visual variations.

 

From there, I evaluated the generated 8-second clips and selected the strongest sequences, merging and trimming them to construct a cohesive narrative.

 

Although the platform generated audio alongside the visuals, I removed it to maintain creative control and emotional clarity. I then imported the merged video into Logic Pro, where I composed an original score designed to enhance the pacing, visual rhythm, and emotional tone of the piece.

Impressions of Google Vids: First Look & Second Glance

Week 1: February 20, 2026

While using Google Vids, I was genuinely impressed by how accurately the AI interpreted my prompt:\

“An African-American boy standing in his backyard, in awe and enamored by the night sky and celestial bodies. The video should be black and white, in the style of a coloring book, with much white space.”

​

The initial output closely matched the intended aesthetic and tone. Although the first render was not saved to the cloud, rerunning the prompt produced a comparable result. I particularly appreciated how the platform serves as a multimodal hub, with complementary tools and features accessible through the right-hand toolbar—supporting a streamlined creative workflow.

Week 2: March 9, 2026

General: I discovered the Version History feature, which allows users to recover previously generated clips that were inserted into the timeline. This proved valuable when experimenting with multiple generations and wanting to revisit earlier outputs.

​

Veo Model: 

Prompts: Through experimentation, I found there is a “sweet spot” when writing prompts. Overly detailed prompts often attempted to describe actions that were unrealistic for an 8-second generation window, while prompts that were too minimal lacked sufficient context for the model to produce strong results. To mitigate this, I kept a Google Doc open to structure prompts into smaller segments and generate clips iteratively. I also created a second Google Vids project file to store clips that were useful but not immediately needed in the timeline, preventing potentially valuable assets from being lost.

 

Uploading Images: Uploading up to three reference images generally worked well. However, it was unclear whether the model consistently recognized the intended order of the images (e.g., first, second, third) when referencing specific details. Because of this uncertainty—and the need to regenerate clips multiple times, sometimes waiting until the next day for the generation limit to reset—I experimented with preparing images beforehand using the Nano Banana model. This allowed me to add details and generate stronger point-of-view compositions before feeding them into Veo.

​

Audio: The auto-generated audio sounded strong when evaluated individually. However, when multiple clips were combined, the audio transitions became disorienting and disrupted the overall emotional continuity. To maintain narrative cohesion, I removed the generated audio and instead composed an original score in Logic Pro, which allowed me to control pacing, emotional tone, and rhythm across the full sequence.

​

Nano Banana Model: The Nano Banana model proved particularly effective for refining images before using them as references in Veo. It was useful for enhancing visual detail and establishing clearer contextual cues for the video generation process.

​

Voiceover Model: I expected a broader range of voice models similar to those available in AI Studio with ElevenLabs models. Ultimately, I chose not to use voiceover in this concept piece because it did not align with the visual storytelling approach or the tone of the narrative.

​

Upload Feature: I appreciated the integration with Google Drive and Google Photos, which made accessing assets convenient. However, a direct drag-and-drop upload option within the interface would improve workflow efficiency.

​

Captions: The preloaded caption styles were useful and visually polished. A valuable addition would be the ability to create and save custom caption styles as reusable templates.

​

​

Opportunities for Enhancement

Week 1: February 20, 2026

Through hands-on experimentation, I identified several opportunities that could further strengthen the platform:
   •  Frame-level isolation for more precise editing and iteration
   •  The ability to insert generated clips directly into the timeline
   •  Tagging and organizational tools for managing multiple generated assets
   •  Native audio stripping or replacement options within the platform

To maintain visual continuity between clips, I isolated the final frame of one video and fed it back into Vids as the starting reference for the next generation. This helped ensure character and environmental consistency across sequences.

Week 2: March 9, 2026

Through hands-on experimentation with the platform, I identified several opportunities that could further strengthen the creative workflow and user control:

 

  • Expanded Prompt Workspace – A prompt window that expands when selected would make it easier to structure longer prompts, review context, and iterate without feeling constrained by a small input field.

  • Regional Regeneration Controls – The ability to isolate and regenerate specific regions or moments within a generated clip would help creators refine outputs without needing to regenerate an entire sequence.

  • Voiceover Tone Controls – More granular controls for adjusting voiceover tone, style, and emotional delivery would allow users to better align narration with the intended narrative mood.

  • Continuity Generation Between Clips – A feature that allows users to select the ending frame of one clip and the beginning frame of another, prompting the Veo model to generate a short bridging sequence, could help create smoother visual transitions and stronger narrative continuity.

Reflection

Overall, Google Vids represents a significant step forward in accessible generative media creation. The platform lowers technical barriers while offering powerful creative potential. With continued refinement in editing controls and asset management, it has the capacity to become an industry-defining tool for AI-assisted storytelling.

bottom of page