How AI and ChatGPT are full of promise and peril, explained by experts – Vox.com
At this point, you have tried ChatGPT. Even Joe Biden has tried ChatGPT, and this week, his administration made a big show of inviting AI leaders like Microsoft CEO Satya Nadella and OpenAI CEO Sam Altman to the White House to discuss ways they could make responsible AI.
But maybe, just maybe, you are still fuzzy on some very basics about AI like, how does this stuff work, is it magic, and will it kill us all? but dont want to admit to that.
No worries. We have you covered: Weve spent much of the spring talking to people working in AI, investing in AI, trying to build businesses in AI as well as people who think the current AI boom is overblown or maybe dangerously misguided. We made a podcast series about the whole thing, which you can listen to over at Recode Media.
But weve also pulled out a sampling of insightful and oftentimes conflicting answers we got to some of these very basic questions. Theyre questions that the White House and everyone else needs to figure out soon, since AI isnt going away.
Read on and dont worry, we wont tell anyone that youre confused. Were all confused.
Kevin Scott, chief technology officer, Microsoft: I was a 12-year-old when the PC revolution was happening. I was in grad school when the internet revolution happened. I was running a mobile startup right at the very beginning of the mobile revolution, which coincided with this massive shift to cloud computing. This feels to me very much like those three things.
Dror Berman, co-founder, Innovation Endeavors: Mobile was an interesting time because it provided a new form factor that allowed you to carry a computer with you. I think we are now standing in a completely different time: Weve now been introduced to a foundational intelligence block that has become available to us, one that basically can lean on all the publicly available knowledge that humanity has extracted and documented. It allows us to retrieve all this information in a way that wasnt possible in the past.
Gary Marcus, entrepreneur; emeritus professor of psychology and neural science at NYU: I mean, its absolutely interesting. I would not want to argue against that for a moment. I think of it as a dress rehearsal for artificial general intelligence, which we will get to someday.
But right now we have a trade-off. There are some positives about these systems. You can use them to write things for you. And there are some negatives. This technology can be used, for example, to spread misinformation, and to do that at a scale that weve never seen before which may be dangerous, might undermine democracy.
And I would say that these systems arent very controllable. Theyre powerful, theyre reckless, but they dont necessarily do what we want. Ultimately, theres going to be a question, Okay, we can build a demo here. Can we build a product that we can actually use? And what is that product?
I think in some places people will adopt this stuff. And theyll be perfectly happy with the output. In other places, theres a real problem.
James Manyika, SVP of technology and society, Google: Youre trying to make sure the outputs are not toxic. In our case, we do a lot of generative adversarial testing of these systems. In fact, when you use Bard, for example, the output that you get when you type in a prompt is not necessarily the first thing that Bard came up with.
Were running 15, 16 different types of the same prompt to look at those outputs and pre-assess them for safety, for things like toxicity. And now we dont always get every single one of them, but were getting a lot of it already.
One of the bigger questions that we are going to have to face, by the way and this is a question about us, not about the technology, its about us as a society is how do we think about what we value? How do we think about what counts as toxicity? So thats why we try to involve and engage with communities to understand those. We try to involve ethicists and social scientists to research those questions and understand those, but those are really questions for us as society.
Emily M. Bender, professor of linguistics, University of Washington: People talk about democratizing AI, and I always find that really frustrating because what theyre referring to is putting this technology in the hands of many, many people which is not the same thing as giving everybody a say in how its developed.
I think the best way forward is cooperation, basically. You have sensible regulation coming from the outside so that the companies are held accountable. And then youve got the tech ethics workers on the inside helping the companies actually meet the regulation and meet the spirit of the regulation.
And to make all that happen, we need broad literacy in the population so that people can ask for whats needed from their elected representatives. So that the elected representatives are hopefully literate in all of this.
Scott: Weve spent from 2017 until today rigorously building a responsible AI practice. You just cant release an AI to the public without a rigorous set of rules that define sensitive uses, and where you have a harms framework. You have to be transparent with the public about what your approach to responsible AI is.
Marcus: Dirigibles were really popular in the 1920s and 1930s. Until we had the Hindenburg. Everybody thought that all these people doing heavier-than-air flight were wasting their time. They were like, Look at our dirigibles. They scale a lot faster. We built a small one. Now we built a bigger one. Now we built a much bigger one. Its all working great.
So, you know, sometimes you scale the wrong thing. In my view, were scaling the wrong thing right now. Were scaling a technology that is inherently unstable.
Its unreliable and untruthful. Were making it faster and have more coverage, but its still unreliable, still not truthful. And for many applications thats a problem. There are some for which its not right.
ChatGPTs sweet spot has always been making surrealist prose. It is now better at making surrealist prose than it was before. If thats your use case, its fine, I have no problem with it. But if your use case is something where theres a cost of error, where you do need to be truthful and trustworthy, then that is a problem.
Scott: It is absolutely useful to be thinking about these scenarios. Its more useful to think about them grounded in where the technology actually is, and what the next step is, and the step beyond that.
I think were still many steps away from the things that people worry about. There are people who disagree with me on that assertion. They think theres gonna be some uncontrollable, emergent behavior that happens.
And were careful enough about that, where we have research teams thinking about the possibility of these emergent scenarios. But the thing that you would really have to have in order for some of the weird things to happen that people are concerned about is real autonomy a system that could participate in its own development and have that feedback loop where you could get to some superhumanly fast rate of improvement. And thats not the way the systems work right now. Not the ones that we are building.
Bender: We already have WebMD. We already have databases where you can go from symptoms to possible diagnoses, so you know what to look for.
There are plenty of people who need medical advice, medical treatment, who cant afford it, and that is a societal failure. And similarly, there are plenty of people who need legal advice and legal services who cant afford it. Those are real problems, but throwing synthetic text into those situations is not a solution to those problems.
If anything, its gonna exacerbate the inequalities that we see in our society. And to say, people who can pay get the real thing; people who cant pay, well, here, good luck. You know: Shake the magic eight ball that will tell you something that seems relevant and give it a try.
Manyika: Yes, it does have a place. If Im trying to explore as a research question, how do I come to understand those diseases? If Im trying to get medical help for myself, I wouldnt go to these generative systems. I go to a doctor or I go to something where I know theres reliable factual information.
Scott: I think it just depends on the actual delivery mechanism. You absolutely dont want a world where all you have is some substandard piece of software and no access to a real doctor. But I have a concierge doctor, for instance. I interact with my concierge doctor mostly by email. And thats actually a great user experience. Its phenomenal. It saves me so much time, and Im able to get access to a whole bunch of things that my busy schedule wouldnt let me have access to otherwise.
So for years Ive thought, wouldnt it be fantastic for everyone to have the same thing? An expert medical guru that you can go to that can help you navigate a very complicated system of insurance companies and medical providers and whatnot. Having something that can help you deal with the complexity, I think, is a good thing.
Marcus: If its medical misinformation, you might actually kill someone. Thats actually the domain where Im most worried about erroneous information from search engines
Now people do search for medical stuff all the time, and these systems are not going to understand drug interactions. Theyre probably not going to understand particular peoples circumstances, and I suspect that there will actually be some pretty bad advice.
We understand from a technical perspective why these systems hallucinate. And I can tell you that they will hallucinate in the medical domain. Then the question is: What becomes of that? Whats the cost of error? How widespread is that? How do users respond? We dont know all those answers yet.
Berman: I think society will need to adapt. A lot of those systems are very, very powerful and allow us to do things that we never thought would be possible. By the way, we dont yet understand what is fully possible. We dont also fully understand how some of those systems work.
I think some people will lose jobs. Some people will adjust and get new jobs. We have a company called Canvas that is developing a new type of robot for the construction industry and actually working with the union to train the workforce to use this kind of robot.
And a lot of those jobs that a lot of technologies replace are not necessarily the jobs that a lot of people want to do anyway. So I think that we are going to see a lot of new capabilities that will allow us to train people to do much more exciting jobs as well.
Manyika: If you look at most of the research on AIs impact on work, if I were to summarize it in a phrase, Id say its jobs gained, jobs lost, and jobs changed.
All three things will happen because there are some occupations where a number of the tasks involved in those occupations will probably decline. But there are also new occupations that will grow. So theres going to be a whole set of jobs gained and created as a result of this incredible set of innovations. But I think the bigger effect, quite frankly what most people will feel is the jobs changed aspect of this.
Read the original here:
How AI and ChatGPT are full of promise and peril, explained by experts - Vox.com
- What Apples AI deal with Google means for the two tech giants, and for $500 billion upstart OpenAI - Fortune - January 14th, 2026 [January 14th, 2026]
- Whats Expensive in AI? The Answer is Changing Fast. - SaaStr - January 14th, 2026 [January 14th, 2026]
- Four Ways I Use AI as a Principal (and One Way I Never Will) (Opinion) - Education Week - January 14th, 2026 [January 14th, 2026]
- Pentagon rolls out major reforms of R&D, AI - Breaking Defense - January 14th, 2026 [January 14th, 2026]
- Pentagon task force to deploy AI-powered UAS systems to capture drones - Defense News - January 14th, 2026 [January 14th, 2026]
- Buy These 3 AI ETFs Now: They Could Be Worth $15 Million in 30 Years - The Motley Fool - January 14th, 2026 [January 14th, 2026]
- ServiceNow Patches Critical AI Platform Flaw Allowing Unauthenticated User Impersonation - The Hacker News - January 14th, 2026 [January 14th, 2026]
- Partnering with Sandstone: An AI-Native Platform for In-House Legal Teams - Sequoia Capital - January 14th, 2026 [January 14th, 2026]
- Bandcamps Mission and Our Approach to Generative AI - Bandcamp - January 14th, 2026 [January 14th, 2026]
- Mom of one of Elon Musk's kids says AI chatbot Grok generated sexual deepfake images of her: "Make it stop" - CBS News - January 14th, 2026 [January 14th, 2026]
- Bill Gates Says 'AI Will Change Society the Most'Job Disruption Has Already Begun, 'Less Labor' Will Be Needed, And 5-Day Work Week May Disappear -... - January 14th, 2026 [January 14th, 2026]
- Prediction: This Artificial Intelligence (AI) Chip Stock Will Outperform Nvidia in 2026 (Hint: It's Not AMD) - The Motley Fool - January 14th, 2026 [January 14th, 2026]
- Microsoft responds to AI data center revolt, vowing to cover full power costs and reject local tax breaks - GeekWire - January 14th, 2026 [January 14th, 2026]
- War Department 'SWAT Team' Removes Barriers to Efficient AI Development - U.S. Department of War (.gov) - January 14th, 2026 [January 14th, 2026]
- South Koreas Revised AI Basic Act to Take Effect January 22 With New Oversight, Watermarking Rules - BABL AI - January 14th, 2026 [January 14th, 2026]
- Musks AI tool Grok will be integrated into Pentagon networks, Hegseth says - The Guardian - January 14th, 2026 [January 14th, 2026]
- You cant afford not to use it: Inderpal Bhandari speaks about the future of AI in sports - The Daily Northwestern - January 14th, 2026 [January 14th, 2026]
- How AI image tools can be tricked into making political propaganda - Help Net Security - January 14th, 2026 [January 14th, 2026]
- Mesa County to test AI software for housing development reviews - KKCO 11 News - January 14th, 2026 [January 14th, 2026]
- 'Most Severe AI Vulnerability to Date' Hits ServiceNow - Dark Reading | Security - January 14th, 2026 [January 14th, 2026]
- Self-learning AI generates NFL picks, score predictions for every 2026 divisional round matchup - CBS Sports - January 14th, 2026 [January 14th, 2026]
- Gen AI Is Threatening the Platforms That Dominate Online Travel - Harvard Business Review - January 14th, 2026 [January 14th, 2026]
- NVIDIA and Lilly Announce Co-Innovation AI Lab to Reinvent Drug Discovery In the Age of AI - Eli Lilly - January 14th, 2026 [January 14th, 2026]
- AI Fraud Has Exploded. This Background-Check Startup Is Cashing In. - Forbes - January 14th, 2026 [January 14th, 2026]
- Caterpillar Briefly Tops $300 Billion Valuation on AI Rally - Yahoo Finance - January 14th, 2026 [January 14th, 2026]
- Google has the best AI for enterprise right now, Ray Wang - Fox Business - January 14th, 2026 [January 14th, 2026]
- What AI is actually good for, according to developers - The GitHub Blog - January 14th, 2026 [January 14th, 2026]
- Apple and Google are teaming up on AI. What it means for both stocks - CNBC - January 14th, 2026 [January 14th, 2026]
- A Look At Cisco Systems (CSCO) Valuation As AI And Cybersecurity Expansion Gain Traction - simplywall.st - January 14th, 2026 [January 14th, 2026]
- US allows Nvidia to send advanced AI chips to China with restrictions - Yahoo Finance - January 14th, 2026 [January 14th, 2026]
- AI industry insiders launch site to poison the data that feeds them - theregister.com - January 11th, 2026 [January 11th, 2026]
- The agentic commerce platform: Shopify connects any merchant to every AI conversation - Shopify - January 11th, 2026 [January 11th, 2026]
- Google teams up with Walmart and other retailers to enable shopping within Gemini AI chatbot - AP News - January 11th, 2026 [January 11th, 2026]
- This CEO laid off nearly 80% of his staff because they refused to adopt AI fast enough. 2 years later, he says hed do it again - Yahoo Finance - January 11th, 2026 [January 11th, 2026]
- Bridging the gap between AI and medicine: Claude in Microsoft Foundry advances capabilities for healthcare and life sciences customers - Microsoft - January 11th, 2026 [January 11th, 2026]
- Artificial Intelligence (AI) Is Driving a New Wave of Infrastructure Spending. This Stock Is Key. - Yahoo Finance - January 11th, 2026 [January 11th, 2026]
- Job Seekers Find a New Source of Income: Training AI to Do Their Old Roles - The Wall Street Journal - January 11th, 2026 [January 11th, 2026]
- The AI platform shift and the opportunity ahead for retail - blog.google - January 11th, 2026 [January 11th, 2026]
- Applied Digital Just Solved AI's Biggest Bottleneck with Technology From the 1800s - The Motley Fool - January 11th, 2026 [January 11th, 2026]
- Can Agentic AI reduce the burden of compliance? - Security Boulevard - January 11th, 2026 [January 11th, 2026]
- Americas AI Boom Is Running Into An Unplanned Water Problem - Forbes - January 11th, 2026 [January 11th, 2026]
- AI, edge, and security: Shaping the need for modern infrastructure management - Network World - January 11th, 2026 [January 11th, 2026]
- Your next primary care doctor could be online only, accessed through an AI tool : Shots - Health News - NPR - January 11th, 2026 [January 11th, 2026]
- Brad Gerstner breaks from the crowd on one AI stock - thestreet.com - January 11th, 2026 [January 11th, 2026]
- Are we in an AI bubble? What 40 tech leaders and analysts are saying, in one chart - CNBC - January 11th, 2026 [January 11th, 2026]
- AI is intensifying a 'collapse' of trust online, experts say - NBC News - January 11th, 2026 [January 11th, 2026]
- Anthropic follows OpenAI in rolling out healthcare AI tools - Investing.com - January 11th, 2026 [January 11th, 2026]
- Behind Anthropic's stunning growth is a sibling team that may hold the key to generative AI - CNBC - January 11th, 2026 [January 11th, 2026]
- Fears of an AI bubble were nowhere to be found at the worlds biggest tech show - CNN - January 11th, 2026 [January 11th, 2026]
- 'No one verified the evidence': Woman says AI-generated deepfake text sent her to jail | Action News Investigation - 6abc Philadelphia - January 11th, 2026 [January 11th, 2026]
- Global AI adoption rose in 2025 but regional gaps widened | ETIH EdTech News - EdTech Innovation Hub - January 11th, 2026 [January 11th, 2026]
- AI isn't making us smarter it's training us to think backward, an innovation theorist says - Business Insider - January 11th, 2026 [January 11th, 2026]
- The "Safest" Trillion-Dollar Artificial Intelligence (AI) Stock to Invest $50,000 In Right Now - The Motley Fool - January 11th, 2026 [January 11th, 2026]
- Prediction: These 2 AI Stocks Will Be Worth More Than Palantir by the End of 2026 - The Motley Fool - January 11th, 2026 [January 11th, 2026]
- A new AI tool could dramatically speed up the discovery of life-saving medicines - Phys.org - January 11th, 2026 [January 11th, 2026]
- After 7 years at McKinsey, I left to build an AI healthtech startup. I had to unlearn the pursuit of perfection. - Business Insider - January 11th, 2026 [January 11th, 2026]
- I never expected an AI app to replace my keyboard, but I was wrong - Android Authority - January 11th, 2026 [January 11th, 2026]
- YouHodler CEO Ilya Volkovs 2026 Crypto Predictions for Stablecoins, AI, Lightning, Bitcoin and More - Crowdfund Insider - January 11th, 2026 [January 11th, 2026]
- I asked AI to beat the S&P 500 it gave me this strategy - MSN - January 11th, 2026 [January 11th, 2026]
- Ant International Partners with Googles Universal Commerce Protocol to Expand AI Capabilities - Business Wire - January 11th, 2026 [January 11th, 2026]
- CES 2026: Follow live for the best, weirdest, most interesting tech as this robot and AI-heavy event wraps up - TechCrunch - January 9th, 2026 [January 9th, 2026]
- Physical AI dominates CES but humanity will still have to wait a while for humanoid servants - Reuters - January 9th, 2026 [January 9th, 2026]
- OpenAI and SoftBank announce $1 billion investment in SB Energy as part of massive AI buildout - CNBC - January 9th, 2026 [January 9th, 2026]
- DeepSeek To Release Next Flagship AI Model With Strong Coding Ability - The Information - January 9th, 2026 [January 9th, 2026]
- AI on Campus: Rethinking the Core Goals of Higher Education - GovTech - January 9th, 2026 [January 9th, 2026]
- 3 Brilliant AI Stocks That Could Double in 2026 - The Motley Fool - January 9th, 2026 [January 9th, 2026]
- DeepSeek to launch new AI model focused on coding in February, The Information reports - Reuters - January 9th, 2026 [January 9th, 2026]
- Marsha Blackburn: My convictions on AI have been clear, not all over - Chattanooga Times Free Press - January 9th, 2026 [January 9th, 2026]
- OpenAI Is Asking Contractors to Upload Work From Past Jobs to Evaluate the Performance of AI Agents - WIRED - January 9th, 2026 [January 9th, 2026]
- AI Hallucinations, Sanctions, and Context: What a Florida Disciplinary Case Really Teaches - JD Supra - January 9th, 2026 [January 9th, 2026]
- Californias budget projections rely heavily on a continued AI boom - capradio.org - January 9th, 2026 [January 9th, 2026]
- No 10 condemns insulting move by X to restrict Grok AI image tool - The Guardian - January 9th, 2026 [January 9th, 2026]
- How to regulate AI by learning from the United States - omnesmag.com - January 9th, 2026 [January 9th, 2026]
- The Change You Need Is the Change You Make. SaaStr AI Annual 2026 Will Help You Get There. - SaaStr - January 9th, 2026 [January 9th, 2026]
- Reimagining Merchandising in the Era of Agentic AI - Bain & Company - January 9th, 2026 [January 9th, 2026]
- Bill Gates says AI could be used as a bioterrorism weapon akin to the COVID pandemic if it falls into the wrong hands - Fortune - January 9th, 2026 [January 9th, 2026]
- So are we in an AI bubble? Here are clues to look for. - NCPR: North Country Public Radio - January 9th, 2026 [January 9th, 2026]
- #679: Why AI Taking Your Job Isnt the Real Problem, with Fmr. OpenAI Exec Zack Kass - Afford Anything - January 9th, 2026 [January 9th, 2026]
- Idaho Statesman AI falsely said a brewery closed, hurting business. Union calls for guardrails on the tech. - BoiseDev - January 9th, 2026 [January 9th, 2026]
- What effect will AI have on the radiologist workforce? - AuntMinnie - January 9th, 2026 [January 9th, 2026]