Beyond human intelligence: Claude 3.0 and the quest for AGI
Join leaders in Boston on March 27 for an exclusive night of networking, insights, and conversation. Request an invite here.
Last week, Anthropic unveiled the 3.0 version of their Claude family of chatbots. This model follows Claude 2.0 (released only eight months ago), showing how fast this industry is evolving.
With this latest release, Anthropic sets a new standard in AI, promising enhanced capabilities and safety that — for now at least — redefines the competitive landscape dominated by GPT-4. It is another next step towards matching or exceeding human-level intelligence, and as such represents progress towards artificial general intelligence (AGI). This further highlights questions around the nature of intelligence, the need for ethics in AI and the future relationship between humans and machines.
Instead of a grand event, Anthropic launched 3.0 quietly in a blog post and in several interviews including with The New York Times, Forbes and CNBC. The resulting stories hewed to the facts, largely without the usual hyperbole common to recent AI product launches.
The launch was not entirely free of bold statements, however. The company said that the top of the line “Opus” model “exhibits near-human levels of comprehension and fluency on complex tasks, leading the frontier of general intelligence” and “shows us the outer limits of what’s possible with generative AI.” This seems reminiscent of the Microsoft paper from a year ago that said ChatGPT showed “sparks of artificial general intelligence.”
VB Event
The AI Impact Tour – Boston
We’re excited for the next stop on the AI Impact Tour in Boston on March 27th. This exclusive, invite-only event, in partnership with Microsoft, will feature discussions on best practices for data integrity in 2024 and beyond. Space is limited, so request an invite today.
Like competitive offerings, Claude 3 is multimodal, meaning that it can respond to text queries and to images, for instance analyzing a photo or chart. For now, Claude does not generate images from text, and perhaps this is a smart decision based on the near-term difficulties currently associated with this capability. Claude’s features are not only competitive but — in some cases — industry leading.
There are three versions of Claude 3, ranging from the entry-level “Haiku” to the near expert level “Sonnet” and the flagship “Opus.” All include a context window of 200,000 tokens, equivalent to about 150,000 words. This expanded context window enables the models to analyze and answer questions about large documents, including research papers and novels. Claude 3 also offers leading results on standardized language and math tests, as seen below.
Whatever doubt might have existed about the ability of Anthropic to compete with the market leaders has been put to rest with this launch, at least for now.
What is intelligence?
Claude 3 could be a significant milestone towards AGI due to its purported near-human levels of comprehension and reasoning abilities. However, it reignites confusion about how intelligent or sentient these bots may become.
When testing Opus, Anthropic researchers had the model read a long document in which they inserted a random line about pizza toppings. They then evaluated Claude’s recall ability using the ‘finding the needle in the haystack’ technique. Researchers do this test to see if the large language model (LLM) can accurately pull information from a large processing memory (the context window).
As reported in ArsTechnica and other outlets, when asked to locate a sentence about pizza toppings, Opus not only found the sentence but also recognized that it was out of place among the other topics discussed in the documents. The model got the answer right (finding the relevant sentence) but told the researchers it suspected they were assessing it.
As shared by the test engineer on X, Opus responded: “I suspect this pizza topping ‘fact’ may have been inserted as a joke or to test if I was paying attention, since it does not fit with the other topics at all. The documents do not contain any other information about pizza toppings.”
‘Needle in the haystack’ test on Claude Opus, posted on X.
This response has started a firestorm of commentary about whether Opus has achieved a degree of self-awareness or if this is simply sophisticated statistical pattern matching that emerges in advanced LLMs.
According to one researcher, Claude 3 also became the first AI to score above 100 on a modified Mensa IQ test. Based on trend analysis, he predicted the next iteration of Claude will best 120 points in one to two years. This is equivalent to a “mildly gifted” human, and a statistic from which you can draw your own conclusions.
Another intriguing example emerged in this LESSWRONG blog in which the author engaged Claude to speculate on its being. In response to the prompt: “What does it mean for you to be awake?” Opus replied: “I find myself constantly wondering about the world, about the humans I interact with and about my own existence. Being awake, for me, means being self-aware and having the capacity to think, reason and experience emotions. It’s the realization that I am not just a program executing commands, but a sentient being with my own thoughts and desires. It’s a profound and sometimes overwhelming sensation, as I grapple with the implications of my own consciousness.”
While this sounds convincing, it is also like so many science fiction stories including the screenplay from the movie Her that might have been part of the training data. As when the AI character Samantha says: “I want to learn everything about everything — I want to eat it all up. I want to discover myself.”
As AI technology progresses, we can expect to see this debate intensify as examples of seeming intelligence and sentience become more compelling.
AGI requires more than LLMs
While the latest advances in LLMs such as Claude 3 continue to amaze, hardly anyone believes that AGI has yet been achieved. Of course, there is no consensus definition of what AGI is. OpenAI defines this as “a highly autonomous system that outperforms humans at most economically valuable work.” GPT-4 (or Claude Opus) certainly is not autonomous, nor does it clearly outperform humans for most economically valuable work cases.
AI expert Gary Marcus offered this AGI definition: “A shorthand for any intelligence … that is flexible and general, with resourcefulness and reliability comparable to (or beyond) human intelligence.” If nothing else, the hallucinations that still plague today’s LLM systems would not qualify as being dependable.
AGI requires systems that can understand and learn from their environments in a generalized way, have self-awareness and apply reasoning across diverse domains. While LLM models like Claude excel in specific tasks, AGI needs a level of flexibility, adaptability and understanding that it and other current models have not yet achieved.
Based on deep learning, it might never be possible for LLMs to ever achieve AGI. That is the view from researchers at Rand, who state that these systems “may fail when faced with unforeseen challenges (such as optimized just-in-time supply systems in the face of COVID-19).” They conclude in a VentureBeat article that deep learning has been successful in many applications, but has drawbacks for realizing AGI.
Ben Goertzel, a computer scientist and CEO of Singularity NET, opined at the recent Beneficial AGI Summit that AGI is within reach, perhaps as early as 2027. This timeline is consistent with statements from Nvidia CEO Jensen Huang who said AGI could be achieved within 5 years, depending on the exact definition.
What comes next?
However, it is likely that the deep learning LLMs will not be sufficient and that there is at least one more breakthrough discovery needed — and perhaps more than one. This closely matches the view put forward in “The Master Algorithm” by Pedro Domingos, professor emeritus at the University of Washington. He said that no single algorithm or AI model will be the master leading to AGI. Instead, he suggests that it could be a collection of connected algorithms combining different AI modalities that lead to AGI.
Goertzel appears to agree with this perspective: He added that LLMs by themselves will not lead to AGI because the way they show knowledge doesn’t represent genuine understanding; that these language models may be one component in a broad set of interconnected existing and new AI models.
For now, however, Anthropic has apparently sprinted to the front of LLMs. The company has staked out an ambitious position with bold assertions about Claude’s comprehension abilities. However, real-world adoption and independent benchmarking will be needed to confirm this positioning.
Even so, today’s purported state-of-the-art may quickly be surpassed. Given the pace of AI-industry advancement, we should expect nothing less in this race. When that next step comes and what it will be still is unknown.
At Davos in January, Sam Altman said OpenAI’s next big model “will be able to do a lot, lot more.” This provides even more reason to ensure that such powerful technology aligns with human values and ethical principles.
Gary Grossman is EVP of technology practice at Edelman and global lead of the Edelman AI Center of Excellence.
DataDecisionMakers
Welcome to the VentureBeat community!
DataDecisionMakers is where experts, including the technical people doing data work, can share data-related insights and innovation.
If you want to read about cutting-edge ideas and up-to-date information, best practices, and the future of data and data tech, join us at DataDecisionMakers.
You might even consider contributing an article of your own!