A Year Ago ChatGPT Ushered in an AI Era: How Has It Impacted Society?
ChatGPT was launched on 30 November 2022, ushering in what many have called artificial intelligence’s breakout year. Within days of its release, ChatGPT went viral
Screenshots of conversations snowballed across social media, and the use of ChatGPT skyrocketed to an extent that seems to have surprised even its maker, OpenAI.
By January, ChatGPT was seeing 13 million unique visitors each day, setting a record for the fastest-growing user base of a consumer application.
Throughout this breakout year, ChatGPT has revealed the power of a good interface and the perils of hype, and it has sown the seeds of a new set of human behaviors.
As a researcher who studies technology and human information behavior, I find that ChatGPT’s influence in society comes as much from how people view and use it as the technology itself.
Generative AI systems like ChatGPT are becoming pervasive. Since ChatGPT’s release, some mention of AI has seemed obligatory in presentations, conversations and articles.
Today, OpenAI claims 100 million people use ChatGPT every week.
Besides people interacting with ChatGPT at home, employees at all levels up to the C-suite in businesses are using the AI chatbot. In tech, generative AI is being called the biggest platform since the iPhone, which debuted in 2007.
All the major players are making AI bets, and venture funding in AI startups is booming.
Along the way, ChatGPT has raised numerous concerns, such as its implications for disinformation, fraud, intellectual property issues and discrimination. In my world of higher education, much of the discussion has surrounded cheating, which has become a focus of my own research this year.
Lessons from ChatGPT’s first year
The success of ChatGPT speaks foremost to the power of a good interface.
AI has already been part of countless everyday products for well over a decade, from Spotify and Netflix to Facebook and Google Maps. The first version of GPT, the AI model that powers ChatGPT, dates back to 2018. And even OpenAI’s other products, such as DALL-E, did not make the waves that ChatGPT did immediately upon its release.
It was the chat-based interface that set off AI’s breakout year.
There is something uniquely beguiling about chat. Humans are endowed with language, and conversation is a primary way people interact with each other and infer intelligence. A chat-based interface is a natural mode for interaction and a way for people to experience the “intelligence” of an AI system.
The phenomenal success of ChatGPT shows again that user interfaces drive widespread adoption of technology, from the Macintosh to web browsers and the iPhone. Design makes the difference.
At the same time, one of the technology’s principal strengths – generating convincing language – makes it well suited for producing false or misleading information. ChatGPT and other generative AI systems make it easier for criminals and propagandists to prey on human vulnerabilities.
The potential of the technology to boost fraud and misinformation is one of the key rationales for regulating AI.
Amid the real promises and perils of generative AI, the technology has also provided another case study in the power of hype. This year has brought no shortage of articles on how AI is going to transform every aspect of society and how the proliferation of the technology is inevitable.
ChatGPT is not the first technology to be hyped as “the next big thing,” but it is perhaps unique in simultaneously being hyped as an existential risk. Numerous tech titans and even some AI researchers have warned about the risk of superintelligent AI systems emerging and wiping out humanity, though I believe that these fears are far-fetched.
The media environment favors hype, and the current venture funding climate further fuels AI hype in particular. Playing to people’s hopes and fears is a recipe for anxiety with none of the ingredients for wise decision making.
What the future may hold
The AI floodgates opened in 2023, but the next year may bring a slowdown. AI development is likely to meet technical limitations and encounter infrastructural hurdles such as chip manufacturing and server capacity. Simultaneously, AI regulation is likely to be on the way.
This slowdown should give space for norms in human behavior to form, both in terms of etiquette, as in when and where using ChatGPT is socially acceptable, and effectiveness, like when and where ChatGPT is most useful.
ChatGPT and other generative AI systems will settle into people’s workflows, allowing workers to accomplish some tasks faster and with fewer errors. In the same way that people learned “to google” for information, humans will need to learn new practices for working with generative AI tools.
But the outlook for 2024 isn’t completely rosy. It is shaping up to be a historic year for elections around the world, and AI-generated content will almost certainly be used to influence public opinion and stoke division.
Meta may have banned the use of generative AI in political advertising, but this isn’t likely to stop ChatGPT and similar tools from being used to create and spread false or misleading content.
Political misinformation spread across social media in 2016 as well as in 2020, and it is virtually certain that generative AI will be used to continue those efforts in 2024.
Even outside social media, conversations with ChatGPT and similar products can be sources of misinformation on their own.
As a result, another lesson that everyone – users of ChatGPT or not – will have to learn in the blockbuster technology’s second year is to be vigilant when it comes to digital media of all kinds.
See more here sciencealert.com
Header imager: Klippa
Please Donate Below To Support Our Ongoing Work To Defend The Scientific Method
PRINCIPIA SCIENTIFIC INTERNATIONAL, legally registered in the UK as a company incorporated for charitable purposes. Head Office: 27 Old Gloucester Street, London WC1N 3AX.
Trackback from your site.
Tom
| #
Everything A/I has to be programmed by humans. Thus, it is highly susceptible to derangement, corruption and indoctrination agendas. Even if you have A/I systems writing other A/I systems, the basis is still human. No A/I system can ever think entirely for itself as a human brain might do.
The only purpose of A/I is to further the agenda of death-to-us-all. This is just another means to that end disguised as something meaningful and useful. What a freaking joke as we are being played for fools once again by the anti-humans. Haven’t the last four years of BS been enough? Nada, nope and no way.
Reply
Howdy
| #
“AI designed to exploit vulnerabilities”
https://professionalsecurity.co.uk/news/interviews/ai-designed-to-exploit-vulnerabilities/
“Google warns of surge in generative AI-enhanced attacks, zero-day exploit use in 2024”
https://www.cshub.com/attacks/news/google-warns-of-surge-in-generative-ai-enhanced-attacks-zero-day-exploit-use-in-2024
You knew this is where it would lead though, didn’t you?
Reply