OpenAI vs. Google: Ready, Go? A Benchmark Analysis of the Open AI-Gemini Models and its Relationship to Gemini Ultra
The small is the start of the search engine. The Smart Reply feature only works in Messenger, and it is being developed for other apps next year. And Gemini as a whole is only rolling out in English right now, which means many users won’t be able to use it at all. It might take time for you to notice that your Pixel 8 Pro won’t be a massive upgrade, though if it is possible, it will be. Next year, when there is a Bard to Assistant on thePixel telephones, you will get even more of the Gemini experience.
After declaring a “code red” after its launch, and being seen to be playing catch-up ever since, google seems to be holding on to its “bold and responsible” slogan. Hassabis and Pichai both say they’re not willing to move too fast just to keep up, especially as we get closer to the ultimate AI dream: artificial general intelligence, the term for an AI that is self-improving, smarter than humans, and poised to change the world. “As we approach AGI, things are going to be different,” Hassabis says. “It’s kind of an active technology, so I think we have to approach that cautiously. Cautiously, but optimistically.”
Let us get to the important question, shall we? OpenAI’s GPT-4 versus Google’s Gemini: ready, go. This topic has been on the mind of the search giant for a while. “We’ve done a very thorough analysis of the systems side by side, and the benchmarking,” Hassabis says. A lot ofBenchmarks were run comparing the two models from broad tests to one that looks at how much Python code can be generated. Hassabis says he thinks we are considerably ahead on 30 of the benchmarks. “Some of them are very narrow. Some of them are bigger.
Right now, Gemini’s most basic models are text in and text out, but more powerful models like Gemini Ultra can work with images, video, and audio. Hassabis says that it will get even more general. “There’s still things like action, and touch — more like robotics-type things.” He says that over time, he says, it will get more senses, become more aware, and become more grounded in the process. “These models just sort of understand better about the world around them.” These models still hallucinate, of course, and they still have biases and other problems. Hassabis says the better they will get, the more they know.
The real test of Gemini’s capability will come from everyday users, who want to use it to find and make new friends, look up information, and much more. Google seems to see coding in particular as a killer app for Gemini; it uses a new code-generating system called AlphaCode 2 that it says performs better than 85 percent of coding competition participants, up from 50 percent for the original AlphaCode. Users will notice an improvement when they look at the model.