Associate Professor Sandra Peter
Imagine a group of young men gathered at a picturesque college campus in New England, in the United States, during the northern summer of 1956.
Itās a small casual gathering. But the men are not here for campfires and nature hikes in the surrounding mountains and woods. Instead, these pioneers are about to embark on an experimental journey that will spark countless debates for decades to come and change not just the course of technology ā but the course of humanity.
Welcome to the Dartmouth Conference ā the birthplace of artificial intelligence (AI) as we know it today.
What transpired here would ultimately lead to ChatGPT and the many other kinds of AI which now help us diagnose disease, detect fraud, put together playlists and write articles (well, not this one). But it would also create some of the many problems the field is still trying to overcome. Perhaps by looking back, we can find a better way forward.
In the mid 1950s, rockānāroll was taking the world by storm. Elvisās Heartbreak Hotel was topping the charts, and teenagers started embracing James Deanās rebellious legacy.
But in 1956, in a quiet corner of New Hampshire, a different kind of revolution was happening.
°Õ³ó±šĢż, often remembered as the Dartmouth Conference, kicked off on June 18 and lasted for about eight weeks. It was the brainchild of four American computer scientists ā John McCarthy, Marvin Minsky, Nathaniel Rochester and Claude Shannon ā and brought together some of the brightest minds in computer science, mathematics and cognitive psychology at the time.
These scientists, along with some of the 47 people they invited, set out to tackle an ambitious goal: to make intelligent machines.
“”²õĢż, they aimed to find out āhow to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humansā.
The Dartmouth Conference didnāt just coin the term āartificial intelligenceā; it coalesced an entire field of study. Itās like a mythical Big Bang of AI ā everything we know about machine learning, neural networks and deep learning now traces its origins back to that summer in New Hampshire.
But the legacy of that summer is complicated.
At the 1956 Dartmouth AI workshop: (back row, from left) Oliver Selfridge, Nathaniel Rochester, Marvin Minsky and John McCarthy, and (front row, from left) Ray Solomonoff, Pete Milner and Claude Shannon. Picture: The Minsky Family
Artificial intelligence won out as a name over others proposed or in use at the time. Shannon preferred the term āautomata studiesā, while two other conference participants (and the soon-to-be creators of the first AI program), Allen Newell and Herbert Simon, continued to use ācomplex information processingā for a few years still.
But hereās the thing: having settled on AI, no matter how much we try, today we canāt seem to get away from comparing AI to human intelligence.
This comparison is both a blessing and a curse.
On the one hand, it drives us to create AI systems that can match or exceed human performance in specific tasks. We celebrate when AI outperforms humans in games such as chess or Go, or when it can detect cancer in medical images with greater accuracy than human doctors.
On the other hand, this constant comparison leads to misconceptions.
When aĀ , it is easy to jump to the conclusion that machines are now smarter than us in all aspects ā or that we are at least well on our way to creating such intelligence. But AlphaGo is no closer to writing poetry than a calculator.
And when a large language model sounds human,Ā .
But ChatGPT is no more alive than a talking ventriloquistās dummy.
The scientists at the Dartmouth Conference were incredibly optimistic about the future of AI. They were convinced they could solve the problem of machine intelligence in a single summer.
This overconfidence has been a recurring theme in AI development, and it has led to several cycles of hype and disappointment.
Ā that āmachines will be capable, within 20 years, of doing any work a man can doā.Ā Ā that āwithin a generation [ā¦] the problem of creating āartificial intelligenceā will substantially be solvedā.
Popular futuristĀ Ā itās only five years away: āweāre not quite there, but we will be there, and by 2029 it will match any personā.
So, how can AI researchers, AI users, governments, employers and the broader public move forward in a more balanced way?
A key step is embracing the difference and utility of machine systems. Instead of focusing on the race to āartificial general intelligenceā, we can focus onĀ Ā ā for example, the enormous creative capacity of image models.
Shifting the conversation from automation to augmentation is also important. Rather than pitting humans against machines, letās focus onĀ .
Letās also emphasise ethical considerations. The Dartmouth participants didnāt spend much time discussing the ethical implications of AI. Today, we know better, and must do better.
We must also refocus research directions. Letās emphasise research into AI interpretability and robustness, interdisciplinary AI research and explore new paradigms of intelligence that arenāt modelled on human cognition.
Finally, we must manage our expectations about AI. Sure, we can be excited about its potential. But we must also have realistic expectations, so that we can avoid the disappointment cycles of the past.
As we look back at that summer camp 68 years ago, we can celebrate the vision and ambition of the Dartmouth Conference participants. Their work laid the foundation for the AI revolution weāre experiencing today.
By reframing our approach to AI ā emphasising utility, augmentation, ethics and realistic expectations ā we can honour the legacy of Dartmouth while charting a more balanced and beneficial course for the future of AI.
After all, the real intelligence lies not just in creating smart machines, but in how wisely we choose to use and develop them.
This article originally appeared in . Associate Professor Sandra Peter is the Director of Sydney Executive Plus at the University of Sydney.