Tesla CEO Elon Musk urges pause on AI systems, citing risks to society

Elon Musk has been vocal about his concerns about AI, a technology used by his EV maker Tesla for autopilot system in its electric cars.
(FILES) In this file photo taken on February 10, 2022 Elon Musk speaks during a press conference at SpaceX's Starbase facility near Boca Chica Village in South Texas. - Musk has put the current value of Twitter at $20 billion, less than half the $44 billion he paid for the social media platform just five months ago, according to an internal email seen by American news media. (Photo by JIM WATSON / AFP) (AFP)
(FILES) In this file photo taken on February 10, 2022 Elon Musk speaks during a press conference at SpaceX's Starbase facility near Boca Chica Village in South Texas. - Musk has put the current value of Twitter at $20 billion, less than half the $44 billion he paid for the social media platform just five months ago, according to an internal email seen by American news media. (Photo by JIM WATSON / AFP)

Elon Musk and a group of artificial intelligence experts and industry executives are calling for a six-month pause in developing systems more powerful than OpenAI's newly launched GPT-4, in an open letter citing potential risks to society and humanity. The letter, issued by the non-profit Future of Life Institute and signed by more than 1,000 people including Musk, called for a pause on advanced AI development until shared safety protocols for such designs were developed, implemented and audited by independent experts.

"Powerful AI systems should be developed only once we are confident that their effects will be positive and their risks will be manageable," the letter said.

The letter detailed potential risks to society and civilization by human-competitive AI systems in the form of economic and political disruptions, and called on developers to work with policymakers on governance and regulatory authorities.

Also check these Cars

Find more Cars
Tesla Model S (HT Auto photo)
UPCOMING
BatteryCapacity Icon75 kWh Range Icon396 km
₹70 Lakhs - 1 Cr
View Details
Lexus Es (HT Auto photo)
Engine Icon2487.0 cc FuelType IconMultiple
₹ 56.55 - 62.19 Lakhs
Compare
View Offers
Bmw I4 (HT Auto photo)
BatteryCapacity Icon83.9 kWh Range Icon493 Km
₹ 69.90 Lakh
Compare
Volkswagen Id.7 (HT Auto photo)
UPCOMING
BatteryCapacity Icon77 kWh Range Icon621 Km
₹ 70 Lakhs
View Details
Tesla Model 3 (HT Auto photo)
UPCOMING
BatteryCapacity Icon82kWh Range Icon 555 km
₹ 70 - 90 Lakhs
View Details
Hyundai Creta (HT Auto photo)
Engine Icon1497 cc FuelType IconMultiple
₹ 11 - 20.15 Lakhs
Compare
View Offers

Co-signatories included Stability AI CEO Emad Mostaque, researchers at Alphabet-owned DeepMind, as well as AI heavyweights Yoshua Bengio and Stuart Russell.

According to the European Union's transparency register, the Future of Life Institute is primarily funded by the Musk Foundation, as well as London-based effective altruism group Founders Pledge, and Silicon Valley Community Foundation.

The concerns come as EU police force Europol on Monday joined a chorus of ethical and legal concerns over advanced AI like ChatGPT, warning about the potential misuse of the system in phishing attempts, disinformation and cybercrime.

Meanwhile, the UK government unveiled proposals for an "adaptable" regulatory framework around AI.

Also read: ChatGPT in cars? General Motors exploring tie up with Microsoft

The government's approach, outlined in a policy paper published on Wednesday, would split responsibility for governing artificial intelligence (AI) between its regulators for human rights, health and safety, and competition, rather than create a new body dedicated to the technology.

Musk, whose carmaker Tesla is using AI for an autopilot system, has been vocal about his concerns about AI.

Since its release last year, Microsoft-backed OpenAI's ChatGPT has prompted rivals to accelerate developing similar large language models, and companies to integrate generative AI models into their products.

Sam Altman, chief executive at OpenAI, hasn't signed the letter, a spokesperson at Future of Life told Reuters. OpenAI didn't immediately respond to request for comment.

"The letter isn’t perfect, but the spirit is right: we need to slow down until we better understand the ramifications," said Gary Marcus, a professor at New York University who signed the letter. “They can cause serious harm... the big players are becoming increasingly secretive about what they are doing, which makes it hard for society to defend against whatever harms may materialize."

First Published Date: 29 Mar 2023, 16:11 PM IST
NEXT ARTICLE BEGINS

Please provide your details to get Personalized Offers on

Choose city
+91 | Choose city
Choose city
Choose city

Want to get the best price for your existing car?

Powered by: Spinny Logo
By clicking "View Offers" you Agree to our Terms and Privacy Policy
Dear Name

Please verify your mobile number.

+91 | Choose city
Couldn't verify the OTP.
It's either expired or it's incorrect.