OpenAI vs. Google: Ready, Go? Comparing GPT-4 versus Gemini: Benchmarking for Language Understanding and Code Generation
OpenAI launched ChatGPT a year and a week ago, and the company and product immediately became the biggest things in AI. It is clear that the company that created much of the technology behind the current trend in artificial intelligence was caught off guard by how well openai was doing.
So, let’s just get to the important question, shall we? OpenAI’s GPT-4 versus Google’s Gemini: ready, go. This was something that had been on their mind for a while. “We’ve done a very thorough analysis of the systems side by side, and the benchmarking,” Hassabis says. From the multi- task Language Understanding benchmark to the only one that compared the two models, there were 32 well-established tests that the Internet giant ran. Hassabis said he believed that we were substantially ahead on 30 out of 32 benchmarks. “Some of them are very narrow. Some of them are larger.”
Right now, Gemini’s most basic models are text in and text out, but more powerful models like Gemini Ultra can work with images, video, and audio. And “it’s going to get even more general than that,” Hassabis says. “There’s still things like action, and touch — more like robotics-type things.” He says that over the course of time, Gemini will get more senses, become aware and be more grounded in the process. These models are better at understanding the world around them, but they still have biases and other problems. But the more they know, Hassabis says, the better they’ll get.
The true test of Gemini will come from everyday users who want to use the tool to look up information, write code and much more. Google seems to see coding in particular as a killer app for Gemini; it uses a new code-generating system called AlphaCode 2 that it says performs better than 85 percent of coding competition participants, up from 50 percent for the original AlphaCode. The models will see an improvement in just about everything, according to Pichai.
Demis Hassabis has never been shy about proclaiming big leaps in artificial intelligence. Most notably, he became famous in 2016 after a bot called AlphaGo taught itself to play the complex and subtle board game Go with superhuman skill and ingenuity.
Multimodality in Bard: What Do We Need to Know Before You Train Your Chatbot? A Commentary on Hassabis’s Disclaimer
Hassabis says that until now, most models have approximated multimodality by training separate modules and stitching them together. “That’s OK for some tasks, but you can’t have this sort of deep complex reasoning in multimodal space.”
OpenAI launched an upgrade to ChatGPT in September that gave the chatbot the ability to take images and audio as input in addition to text. OpenAI has not disclosed technical details about how GPT-4 does this or the technical basis of its multimodal capabilities.
Only a sliver of Gemini is currently available. Future releases are expected to include multimodal capabilities, where a chatbot processes multiple forms of input and produces outputs in different ways, just the text-based version has been added to Bard.
You might see a software glitch in your responses if you are involved in an experiment like this. Bard’s current strengths are its integrations with other services, which actually work. For example, to have a chatbot summarize your daily messages or to explore topics with videos, you should tag@Gmail in your prompt. Our previous tests of the Bard chatbot showed potential for these integrations, but there are still plenty of kinks to be worked out.
Gemini Ultra vs. Pixel 8 Pro: What Can Users Expect to Learn from Live Demos of Artificial Intelligence Models?
So how is the anticipated Gemini Ultra different from the currently available Gemini Pro model? According to Google, Ultra is its “most capable mode” and is designed to handle complex tasks across text, images, audio, video, and code. The smaller model of theArtificial Intelligence model, which is fitted to work as part of theSmartphone features, is called the “Gemini” and can be found in the Pixel 8 Pro.
“All the user prompts and outputs in the video are real, shortened for brevity,” Vinyals says. “The video illustrates what the multimode user experiences built with Gemini could look like. We made it to inspire developers.”
Many companies edit demo videos to make them easier to watch, as they want to avoid any technical difficulties that live demos bring. It’s common to tweak things a little. There is a history of questionable video demos by the search engine company, led by founder Sergey Brin. People wondered if Google’s Duplex demo (remember Duplex, the AI voice assistant that called hair salons and restaurants to book reservations?) was real because there was a distinct lack of ambient noise and too-helpful employees. People are more suspicious when there’s pre recorded videos of artificial intelligence models. Remember when Baidu launched its Ernie Bot with edited videos and its shares tanked?