Sunday, April 28, 2024

The year we played with artificial intelligence — and weren’t sure what to do about it



Artificial intelligence went mainstream in 2023 — it was once a very long time coming but has far to opt for the generation to fit other people’s science fiction fantasies of human-like machines.

Catalyzing a year of AI fanfare was ChatGPT. The chatbot gave the arena a glimpse of latest advances in laptop science even supposing now not everybody discovered somewhat how it works or what to do with it.

- Advertisement -

“I would call this an inflection moment,” pioneering AI scientist Fei-Fei Li mentioned. “2023 is, in history, hopefully going to be remembered for the profound changes of the technology as well as the public awakening. It also shows how messy this technology is.”

It was once a year for other people to work out “what this is, how to use it, what’s the impact — all the good, the bad and the ugly,” she said.

PANIC OVER AI

- Advertisement -

The first AI panic of 2023 set in soon after New Year’s Day when classrooms reopened and schools from Seattle to Paris started blocking ChatGPT. Teenagers were already asking the chatbot — released in late 2022 — to compose essays and answer take-home tests.

AI large language models behind technology such as ChatGPT work by repeatedly guessing the next word in a sentence after having “learned” the patterns of an enormous trove of human-written works. They regularly get info fallacious. But the outputs seemed so herbal that it sparked interest about the following AI advances and its attainable use for trickery and deception.

Worries escalated as this new cohort of generative AI tools — spitting out not just words but novel images, music and synthetic voices — threatened the livelihoods of anyone who writes, draws, strums or codes for a living. It fueled moves via Hollywood writers and actors and criminal demanding situations from visual artists and bestselling authors.

- Advertisement -

Some of the AI box’s maximum esteemed scientists warned that the generation’s unchecked growth was once marching towards outsmarting people and possibly threatening their existence, whilst different scientists referred to as their issues overblown or introduced consideration to extra rapid dangers.

By spring, AI-generated deepfakes — some more convincing than others — had leaped into U.S. election campaigns, where one falsely showed Donald Trump embracing the nation’s former top infectious disease expert. The technology made it increasingly difficult to distinguish between real and fabricated war footage in Ukraine and Gaza.

By the end of the year, the AI crises had shifted to ChatGPT’s own maker, the San Francisco startup OpenAI, nearly destroyed by corporate turmoil over its charismatic CEO, and to a government meeting room in Belgium, where exhausted political leaders from across the European Union emerged after days of intense talks with a deal for the world’s first major AI legal safeguards.

The new AI law won’t take effect until 2025, and other lawmaking bodies — including the U.S. Congress — are still a long way from enacting their own.

TOO MUCH HYPE?

There’s no question that commercial AI products unveiled in 2023 incorporated technological achievements not possible in earlier stages of AI research, which trace back to the mid-20th century.

But the latest generative AI trend is at peak hype, according to the market research firm Gartner, which has tracked what it calls the “hype cycle” of emerging technology since the 1990s. Picture a wooden rollercoaster ticking up to its highest hill, about to careen down into what Gartner describes as a “trough of disillusionment” before coasting back to reality.

“Generative AI is right in the peak of inflated expectations,” Gartner analyst Dave Micko said. “There’s massive claims by vendors and producers of generative AI around its capabilities, its ability to deliver those capabilities.”

Google drew grievance this month for modifying a video demonstration of its maximum succesful AI fashion, called Gemini, in some way that made it seem extra spectacular — and human-like.

Micko said leading AI developers are pushing certain ways of applying the latest technology, most of which correspond to their current line of products — be they search engines or workplace productivity software. That doesn’t mean that’s how the world will use it.

“As much as Google and Microsoft and Amazon and Apple would love us to adopt the way that they think about their technology and that they deliver that technology, I think adoption actually comes from the bottom up,” he said.

IS IT DIFFERENT THIS TIME?

It’s easy to forget that this isn’t the first wave of AI commercialization. Computer vision techniques developed by Li and other scientists helped sort through a huge database of photos to recognize objects and individual faces and help guide self-driving cars. Speech recognition advances made voice assistants like Siri and Alexa a fixture in many people’s lives.

“When we launched Siri in 2011, it was at that point the fastest-growing consumer app and the only major mainstream application of AI that people had ever experienced,” said Tom Gruber, co-founder of Siri Inc., which Apple bought and made an integral iPhone feature.

But Gruber believes what’s happening now is the “biggest wave ever” in AI, unleashing new possibilities as well as dangers.

“We’re surprised that we could accidentally encounter this astonishing ability with language, by training a machine to play solitaire on all of the internet,” Gruber said. “It’s kind of amazing.”

The dangers could come fast in 2024, as major national elections in the U.S., India and elsewhere could get flooded with AI-generated deepfakes.

In the longer term, AI technology’s rapidly improving language, visual perception and step-by-step planning capabilities could supercharge the vision of a digital assistant — but only if granted access to the “inner loop of our digital life stream,” Gruber said.

“They can manage your attention as in, ‘You should watch this video. You should read this book. You should respond to this person’s communication,’” Gruber mentioned. “That is what a real executive assistant does. And we could have that, but with a really big risk of personal information and privacy.”

More articles

- Advertisement -
- Advertisement -

Latest article