TL;DR: Google Bard Gemini model
- Google’s Bard AI chatbot receives a significant upgrade with the Gemini model, impressing users and rivalling OpenAI’s GPT-4.
- Bard’s Pro model secures second place in Chatbot Arena benchmark, marking a crucial win against GPT-4 models.
- Google’s AI supremacy ambitions heat up as Gemini Ultra release looms, promising even greater performance.
Introduction
The AI landscape in Asia is undergoing a seismic shift as Google’s Bard, the ever-evolving chatbot, receives a substantial upgrade with the new Gemini model. Early users are raving about its performance, with many drawing comparisons to OpenAI’s highly praised GPT-4.
Gemini Model: A Powerful Upgrade
Jeff Dean, Google’s AI chief, announced the new model, which belongs to the “Gemini Pro” family with the intriguing suffix “scale.” According to Dean, this upgrade makes Bard “much better” and unlocks “many more capabilities” compared to its March launch.
Speculation suggests that the “scale” suffix indicates a bigger, more powerful version of the previous Pro model. Although benchmarks have not yet placed it ahead of the free version of GPT-3.5 (ChatGPT), the Gemini Pro “scale” model has achieved an impressive feat in the neutral Chatbot Arena benchmark.
A Crucial Win in the Chatbot Arena
Despite being Google’s second-tier offering, the Pro model with its “scale” suffix secured a remarkable second place in the Chatbot Arena benchmark. This victory places it ahead of two existing GPT-4 models, marking a significant milestone for Bard and breaking into the GPT-4 stronghold for the first time.
The Chatbot Arena evaluation system relies on user-based assessments, with models competing against each other in anonymous duels. Users decide the winner based on their interaction and preference, ensuring an objective evaluation.
Potential Limitations and Future Prospects
While user-based evaluations offer valuable insights, it is essential to consider potential limitations. Bard’s current rating is based on roughly 3,000 interactions, significantly less than the 30,000 received by some GPT-4 models. This disparity could influence the outcome, and small gaps in the leaderboard should not be overemphasised.
Oriol Vinyals, Google’s deep learning head and co-lead of Gemini, acknowledges the challenge of AI evaluation, describing it as “hard and nuanced.” He expresses his delight with Bard’s performance in the Chatbot Arena and hints at even greater capabilities with the upcoming release of Gemini Ultra, Google’s top-tier model.
As Google’s Bard continues to evolve and challenge existing AI leaders, we ask you: Which AI chatbot do you prefer using, and what features stand out to you? Share your thoughts in the comments below.
You may also like: