Introducing Artificial Intelligence to Silicon Valley: How a Generative AI Company Grows to Compute Content and Capture Languages
A buzzy salon at a trendy bar in San Francisco last week drew an animated crowd of engineers, entrepreneurs and financiers. The thing on all of their minds: the blossoming creative capabilities of artificial intelligence.
Social media has been overrun with stunning and strange images thanks to advances by Hugging Face and others. Machine learning technology is able to create reams of coherent text on a given subject. A few of what are now styled as generative AI companies have collectively raised hundreds of millions of dollars, spurring a hunt for a new generation of AI unicorns.
The party was held in San Francisco by StabilityAI, which offers tools for generating images with few restrictions. The company is valued at $1 billion, after announcing $101 million in new funding. The gathering attracted tech celebrities including Google cofounder Sergey Brin.
Song works with Everyprompt, a startup that makes it easier for companies to use text generation. He says testing the generativeAI tools that make images, text, or code left him with a sense of wonder. “It’s been a long time since I used a website or technology that felt immensely helpful or magical,” he says. “Using generative AI makes me feel like I’m using magic.”
The world’s “1,000 most spoken languages” will be supported by a new project to develop a single artificial intelligence language model. As a first step towards this goal, the company is unveiling an AI model trained on over 400 languages, which it describes as “the largest language coverage seen in a speech model today.”
Google has already begun integrating these language models into products like Google Search, while fending off criticism about the systems’ functionality. Language models have a number of flaws, including a tendency to regurgitate harmful societal biases like racism and xenophobia, and an inability to parse language with human sensitivity. The researchers were fired by Google after their papers were published.
Speaking to The Verge, Zoubin Ghahramani, vice president of research at Google AI, said the company believes that creating a model of this size will make it easier to bring various AI functionalities to languages that are poorly represented in online spaces and AI training datasets (also known as “low-resource languages”).
“By having a single model that is exposed to and trained on many different languages, we get much better performance on our low resource languages,” says Ghahramani. “The way we get to 1,000 languages is not by building 1,000 different models. Languages have some similarities and they have evolved from one another. And we can find some pretty spectacular advances in what we call zero-shot learning when we incorporate data from a new language into our 1,000 language model and get the ability to translate [what it’s learned] from a high-resource language to a low-resource language.”
When training across so many languages, it is not easy to obtain data, and that is the reason why Google will be funding the collection of data for low-resource languages.
The company has no plans to apply the features of this model to any specific product, only that it expects it will have a variety of uses across some of the company’s products.
“One of the really interesting things about large language models and language research in general is that they can do lots and lots of different tasks,” says Ghahramani. One model can be used to create code for a robot, solve problems in mathematics, and perform translation. The really interesting thing about language models is that they are becoming a place to keep a lot of knowledge and by probing them you can get to different bits of useful function.
The artificial intelligence at the core of chatsgpt is not new. It is a version of an AI model called GPT-3 that generates text based on patterns it digested from huge quantities of text gathered from the web. That model, which is available as a commercial API for programmers, has already shown that it can answer questions and generate text very well some of the time. But getting the service to respond in a particular way required crafting the right prompt to feed into the software.
But Google’s competitors don’t seem to have “slow” in their vocabularies. While LaMDA is available in a protected Test Kitchen app, other companies have created their own products with their own image generators. Only a few weeks after the Google event came the most consequential release yet: OpenAI’s latest version of its own powerful text generation technology, ChatGPT, a lightning-fast, logorrheic gadfly that spits out coherent essays, poems, plays, songs, and even obituaries at the merest hint of a prompt. Taking advantage of the chatbot’s wide availability, millions of people have tinkered with it and shared its amazing responses, to the point where it’s become an international obsession, as well as a source of wonder and fear. Will ChatGPT kill the college essay? Do you want to destroy traditional internet search? Put millions of copywriters, journalists, artists, songwriters, and legal assistants out of a job?
The company shared some details in a post, but it did not reveal full details about how it gives its software a naturalistic new interface. It says the team fed human-written answers to GPT-3.5 as training data, and then used a form of simulated reward and punishment known as reinforcement learning to push the model to provide better answers to example questions.
Jacob Andreas, an assistant professor who works on AI and language at MIT, says the system seems likely to widen the pool of people able to tap into AI language tools. “Here’s a thing being presented to you in a familiar interface that causes you to apply a mental model that you are used to applying to other agents—humans—that you interact with,” he says.
The launch of chatgtp has opened up new conversations about the potential of machine learning to replace traditional search engines but the question of whether or not to use it has been under discussion for a long time. The same ethical and technical issues that Pichai and Dean are now talking about are the ones that brought down Timnit Gebru and Margaret Mitchell. In May last year, a group of researchers explored the issue of artificial intelligence in search. As the researchers noted in their paper, one of the biggest issues is that LLMs “do not have a true understanding of the world, they are prone to hallucinating, and crucially they are incapable of justifying their utterances by referring to supporting documents in the corpus they were trained over.”
The creators of such models admit that they face a challenge of acknowledging inappropriate responses that do not accurately reflect the contents of authoritative external sources. A scientific paper on benefits of eating crushed glass and a text about how porcelain can be added to breast milk have been created by Galactica. Stack Overflow had to temporarily cease the use of the generated answers due to the fact that it was obvious that the LLM generated the wrong answers to the coding questions.
There are ongoing differences of blame and praise in the wake of this work. Model builders and tech Evangelists alike say that a mythically autonomously-generated model is a technological marvel. There is no decision-making involved in the model development, and model feats are not related to the design and implementation choices of engineers. But without naming and recognizing the engineering choices that contribute to the outcomes of these models, it becomes almost impossible to acknowledge the related responsibilities. Functional failures are framed as devoid of engineering choices due to society at large or supposed “naturally occurring” datasets, factors that those developing models will claim they have little control over. The models we are seeing now are not inevitable because they have control. It would have been entirely feasible for different choices to have been made, resulting in an entirely different model being developed and released.
A pre-event message from the CEO of the company stated that they were starting with machine learning in search to make it easier to see the big picture. Despite recent layoffs, the company remains an assertive force in Silicon Valley. The success of other generative models put pressure on the company to speed up their research for the public.
It seems to be trying to damp down expectations. Sam Altman has said that there isn’t enough of CHATGPP to create a misleading impression of greatness. It is not a good idea to rely on it for anything important right now. it’s a preview of progress; we have lots of work to do on robustness and truthfulness.”
At the New York pier, where the toys were displayed, most of them showed the fruits of generative models like LaMDA. It is able to answer questions and help writers make stories. Other projects can produce 3D images from text prompts or even help to produce videos by cranking out storyboard-like suggestions on a scene-by-scene basis. But a big piece of the program dealt with some of the ethical issues and potential dangers of unleashing robot content generators on the world. The company made a point of emphasizing how cautious it was in its use of powerful creations. The statement that made the most sense came from Douglas Eck. “Generative AI models are powerful—there’s no doubt about that,” he said. We need to acknowledge that the technology may pose real risks if we don’t take care of it, which is why we have been slow to release them. And I’m proud we’ve been slow to release them.”
On February 8th, at 9:00 am eastern, Google will announce Artificial Intelligence integrations for its search engine. You can watch it on the internet for free.
One of the questions that remains is Is generative artificial intelligence ready to help you surf the web? These models are costly to power and hard to keep updated, and they love to make shit up. As more people test out the new tools, public engagement with the technology is rapidly changing but generative AI’s positive impact is still largely unknown.
Some early programmers will have access to a more powerful version of Bing, which will roll out today, according to Microsoft executives. The company is going to launch a wider-ranging product in the coming weeks.
The response also included a disclaimer: “However, this is not a definitive answer and you should always measure the actual items before attempting to transport them.” Microsoft will use a feedback box at the top of responses to help train its calculations. A demonstration yesterday was made of how text generation could be used to improve search results.