Skip to content ↓

Topic

Artificial intelligence

Download RSS feed: News Articles / In the Media / Audio

Displaying 181 - 195 of 1222 news clips related to this topic.
Show:

Fast Company

In an excerpt from her new book, “The Mind’s Mirror: Risk and Reward in the Age of AI," Prof. Daniela Rus, director of CSAIL, addresses the fear surrounding new AI technologies, while also exploring AI’s vast potential. “New technologies undoubtedly disrupt existing jobs, but they also create entirely new industries, and the new roles needed to support them,” writes Rus.  

NPR

Prof. Daron Acemoglu speaks with NPR Planet Money hosts Greg Rosalsky and Darian Woods about the anticipated economic impacts of generative AI. Acemoglu notes he believes AI is overrated because humans are underrated. "A lot of people in the industry don't recognize how versatile, talented, multifaceted human skills and capabilities are," Acemoglu says. "And once you do that, you tend to overrate machines ahead of humans and underrate the humans."

Fortune

Writing for Fortune, Prof. Daron Acemoglu explores the estimated scale of AI’s impact on the labor market and productivity. “The problem with the AI bubble isn’t that it is bursting and bringing the market down,” writes Acemoglu. “It’s that the hype will likely go on for a while and do much more damage in the process than experts are anticipating." 

Forbes

MIT researchers have found that “when nudged to review LLM-generated outputs, humans are more likely to discover and fix errors,” reports Carter Busse for Forbes. The findings suggest that, “when given the chance to evaluate results from AI systems, users can greatly improve the quality of the outputs,” explains Busse. “The more information provided about the origins and accuracy of the results, the better the users are at detecting problems.” 

GBH

Prof. Jonathan Gruber joins GBH’s All Things Considered to discuss stock market jitters, AI hype and interest rates, urging calm and a long-term view. “No one who’s in the market should be overreacting to one day’s movement,” Gruber says. “These short-run reactions are really overreactions to individual bits of news.”

New York Times

Prof. Simon Johnson and Prof. David Autor speak with New York Times reporter Emma Goldberg about the anticipated impact of AI on the job market. “We should be concerned about eliminating them,” says Prof. Simon Johnson, of the risks posed by automating jobs. “This is the hollowing out of the middle class.”

Forbes

Chanyeol Choi MS '19, PhD '21 and Subeen Pang MS '21, PhD '24 cofounded Linq, an AI startup that “helps hedge funds speed up their research into thousands of listed companies worldwide,” reports John Kang for Forbes. The company’s software “automates time-consuming equity research tasks, such as scanning for company announcements and news, building financial models and summarizing earnings reports and call transcripts,” explains Kang. 

Tech Briefs

Research Scientist Mathieu Huot speaks with Tech Briefs reporter Andrew Corselli about his work with GenSQL, a generative AI system for databases that “could help users make predictions, detect anomalies, guess missing values, fix errors, or generate synthetic data with just a few keystrokes.” 

TechCrunch

Intelmatix, an AI startup founded by by Almaha Almalki MS '18, Anas Alfaris MS '09, PhD '09 and Ahmad Alabdulkareem PhD '18, aims to provide businesses in the Middle East and North Africa with access to AI for decision-making, reports Annie Njanja for TechCrunch. . “The idea of democratizing access to AI has always been something that we’ve been very passionate about,” says Alfaris. 

 

Scientific American

Prof. Sherry Turkle shares the benefits of being polite when interacting with AI technologies, reports Webb Wright for Scientific American, underscoring the risks of becoming habituated to using crass, disrespectful and dictatorial language. “We have to protect ourselves,” says Turkle. “Because we’re the only ones that have to form relationships with real people.”

Forbes

Penny Abeywardena of Forbes spotlights Joy Buolamwini PhD '22 and her work focused on the impacts of AI technologies on privacy. “We have to think about what we do when something goes wrong for the people I call the excoded, or in other words, those harmed by AI systems,” says Buolamwini. “We need to have pathways for redress and design with redress in mind.”

Forbes

Forbes reporter Rodger Dean Duncan spotlights “The Skill Code: How to Save Human Ability in an Age of Intelligent Machines,” a new book by Research Affiliate Matt Bean SM '14, PhD '17. Duncan “explains Beane’s take on AI tools, collaboration and remote work, who suggests traditional mentoring is at risk in the workplace. Beane says today’s successful people have ‘discovered new tactics that others can use to get skills without throwing out the benefits of hybrid working arrangements.’”

The New York Times

Researchers from the Data Provenance Initiative, a research group led by MIT engineers, have found that “important web sources used for training AI models have restricted the use of their data,” reports Kevin Roose for The New York Times. “We’re seeing a rapid decline in consent to use data across the web that will have ramifications not just for A.I. companies, but for researchers, academics and noncommercial entities,” explains graduate student Shayne Longpre.

Kathimerini

Dean Anantha Chandrakasan, MIT’s Chief Innovation and Strategy Officer, and Pavlos-Petros Sotiriadis PhD '02 sit down with Tassoula Eptakili of Kathimerini to discuss MIT’s unique approach to entrepreneurship, the future of AI and the importance of mentorship. MIT’s approach to education “focuses both on foundational knowledge and its practical application. This prepares [students] to tackle even the most challenging problems,” says Chandrakasan. “It is no coincidence that MIT’s impact on the world has been immense over time: In recent decades, behind almost every significant discovery or new lifesaving technology, you will find our graduates leading or contributing decisively to these efforts.”

The Wall Street Journal

Prof. Armando Solar-Lezama speaks with The Wall Street Journal reporter Isabelle Bousquette about large language models (LLMs) in academia. Instead of building LLMs from scratch, Solar-Lezama suggests “students and researchers are focused on developing applications and even creating synthetic data that could be used to train LLMs,” writes Bousquette.