In a recent demonstration video titled “Hands-on with Gemini: Interacting with multimodal AI,” Google showcased its GPT-4 competitor, Gemini, with high expectations. However, a Bloomberg opinion piece highlights concerns about the video’s accuracy and transparency.
According to Bloomberg, Google admitted that parts of the video were staged, with edits made to accelerate the outputs, as disclosed in the video description. The implied voice interaction between a human user and the AI, touted in the demonstration, was revealed to be non-existent. Instead, the actual demo involved the creation of interactions by “using still image frames from the footage and prompting via text,” rather than responding to real-time drawing or object changes on the table.
The lack of a disclaimer about the actual input method raises questions about the readiness of Gemini, portraying a less impressive capability than the video implies. While Google denies any wrongdoing, referencing a post by Gemini’s co-lead, Oriol Vinyals, stating that “all the user prompts and outputs in the video are real,” critics argue that the tech giant should exercise more sensitivity in its presentations, especially given the increased scrutiny from both the industry and regulatory authorities on AI practices.