• Home
  • Investing
  • Global
  • Business
  • Economy
  • Tech/AI
  • Lifestyle
  • About Us
  • Contact
  • Home
  • Investing
  • Global
  • Business
  • Economy
  • Tech/AI
  • Lifestyle
  • About Us
  • Contact
LOGIN
Saturday, May 16, 2026
Top Posts
Costco: Compounding Power of Trust and Discipline
Uber: The Rulebreaker’s Playbook
Google: Search Box to Empires
Y Combinator: Accelerator or University
Investing Guidance – Oct 24, 2025
Investing Guidance – Oct 17, 2025
Investing Guidance – Nov 19, 2025
Intel: The Traitorous Eight
Investing Guidance – Nov 12, 2025
Investing Guidance – Nov 7, 2025
SUBSCRIBE NEWSLETTERS
  • Home
  • Investing
  • Global
  • Business
  • Economy
  • Tech/AI
  • Lifestyle
  • About Us
  • Contact
Copyright 2021 - All Right Reserved
Why Taiwan became the defining issue in the Trump-Xi talks
Economy

Why Taiwan became the defining issue in the Trump-Xi talks

by admin May 16, 2026
written by admin

Aly Song | Reuters

BEIJING — U.S. President Donald Trump has kept up an uneasy silence about Taiwan following his meeting with Chinese leader Xi Jinping this week, despite the U.S.’ announcement in December of a record $11 billion in arms sales to the island against Beijing’s wishes.

Trump had said the Taiwan arms sales would be on the agenda for his talks with Chinese President Xi Jinping which ended on Friday.

But after the two leaders’ first day of meetings on Thursday, Secretary of State Marco Rubio told NBC News the topic “did not feature primarily in today’s discussion.”

The initial White House readout also did not mention Taiwan – home to manufacturers of some of the world’s most advanced semiconductors – although Treasury Secretary Scott Bessent told CNBC he expected Trump would say more on Taiwan in coming days.

The silence persisted — more than 24 hours after China published its official readout with a stark warning from Xi that mishandling Taiwan would put the U.S.-China relationship in “great jeopardy.”

“This is a pretty direct and strong comment by President Xi,” Wendy Cutler, former acting deputy U.S. trade representative, said Friday on CNBC’s “The China Connection.”

“The way I interpret it too is that he really tied economic stability to developments with respect to Taiwan,” she said.

Beijing’s readout of the closing Trump-Xi meeting Friday morning emphasized the benefits of cooperation and did not mention Taiwan.

‘Cool it’

Trump said that China and Taiwan “ought to both cool it“.

In an interview with Fox News that aired Friday afternoon, Trump insisted that long-standing U.S. policy on Taiwan remains unchanged after his two days of meetings with Xi.

The people of Taiwan should feel “neutral” about his visit, Trump said.

But he also appeared to express some opposition to the prospect of the U.S. leaping to Taiwan’s defense if it is attacked, while framing Taipei’s decision to pursue independence from China as the deciding factor.

“I will say this: I’m not looking to have somebody go independent, and you know, we’re supposed to travel 9,500 miles to fight a war,” Trump said. “I’m not looking for that. I want them to cool down, I want China to cool down.”

He added that he has yet to approve another potential large sale of weapons to Taiwan: “I may do it, I may not do it.”

“We’re not looking to have somebody say ‘Let’s go independent because the United States is backing us,'” Trump said.

“Taiwan would be very smart to cool it a little bit. China would be very smart to cool it a little bit. They ought to both cool it,” he said.

Earlier, Trump said he refused to directly answer Xi when asked if the U.S. would defend Taiwan against a Chinese attack.

Trump also said Taiwan was not part of the discussion when he met with Xi in South Korea last fall.

Trump’s decision not to answer is in line with the U.S.′ long-standing “One China” policy, which leaves the status of Taiwan, an island that Beijing claims as its own, undefined.

The approach of “strategic ambiguity” leaves open whether Washington would come to Taipei’s aid in the event of a Chinese attack.

A statue of a soldier with its gun pointed towards Xiamen on the Chinese mainland across the Taiwan Strait on Lieyu Island in Kinmen, Taiwan.
Bloomberg | Bloomberg | Getty Images

As for arms sales, the 1979 Taiwan Relations Act adds that the U.S. “will make available to Taiwan such defense articles and defense services” as may be necessary to “enable Taiwan to maintain sufficient self-defense capabilities.”

Maintaining the status quo

Taiwan, meanwhile, said comments by Trump and Rubio signal that U.S. policy toward the island remains unchanged.

“It is a clear fact that [Taiwanese] President Lai Ching-te has consistently advocated for continuing to contribute to regional peace and stability and remaining committed to maintaining the status quo across the Taiwan Strait,” Taiwan’s presidential spokesperson Karen Kuo said in a statement on Saturday.

“China’s escalating military threat is the sole destabilizing factor within the Indo-Pacific region, including the Taiwan Strait,” Kuo added.

“If you look at the readouts of all Trump-Xi meetings before this [week], just the last several that have occurred since maybe April of last year, you see the U.S. readouts have a much smaller portion focused on Taiwan,” Rush Doshi, director of the China strategy initiative, Council on Foreign Relations, said Friday on CNBC’s “Squawk Box Asia.”

“There’s really no sign that there’s been a significant change in [the U.S.] Taiwan policy, at least not yet from the summit,” Doshi said.

Taiwan is a democratically self-ruled island that Beijing claims is part of its territory. Since 1979, the U.S. has recognized Beijing and not Taipei, and acknowledges the Chinese position that there is one China and Taiwan is part of China. The U.S. maintains an unofficial relationship with the island.

– CNBC’s Eunice Yoon, Dan Mangan, Kevin Breuninger and Azhar Sukri contributed to this story.

Choose CNBC as your preferred source on Google and never miss a moment from the most trusted name in business news.

May 16, 2026 0 comments
0 FacebookTwitterPinterestEmail
Musk v. Altman week 3: Musk and Altman traded blows over each other’s credibility. Now the jury will pick a side.
Tech/AI

Musk v. Altman week 3: Musk and Altman traded blows over each other’s credibility. Now the jury will pick a side.

by admin May 15, 2026
written by admin

In the final week of the Musk v. Altman trial, lawyers traded blows over Elon Musk’s and OpenAI CEO Sam Altman’s credibility. Altman was grilled on his alleged history of lying and self-dealing involving companies that do business with OpenAI. But he fired back, painting Musk as a power-seeker who wanted to control the development of artificial general intelligence (AGI)—powerful AI that can compete with humans on most cognitive tasks. 

As evidence of their commitment to AI safety, OpenAI brought out a golden trophy of a donkey’s ass that was gifted to an employee after he was called a “jackass” for standing up to Musk’s plans to race toward AGI. 

Lawyers for both sides also presented their closing arguments, floating unflattering mugshot-style photos of Musk and Altman next to each other on a giant screen. Musk’s lawyer Steven Molo argued that Altman and OpenAI president Greg Brockman broke their promise to use money Musk donated to maintain OpenAI as a nonprofit that develops AI for the benefit of humanity. Instead, they created a for-profit subsidiary that made them extraordinarily wealthy.

OpenAI’s lawyer Sarah Eddy argued that Altman and Brockman never promised to keep OpenAI a nonprofit. She added that even though it’s been restructured, OpenAI remains a nonprofit dedicated to developing AI safely.

She claimed that Musk sued too late—and that his real motive is to sabotage a competitor to his own AI company, xAI, which he launched in 2023. 

Musk is asking the court to unwind the 2025 restructuring that converted OpenAI’s for-profit subsidiary into a public benefit corporation and to remove Altman and Brockman from their roles. He is also seeking as much as $134 billion in damages from OpenAI and Microsoft, to be awarded to OpenAI’s nonprofit. 

The jury will begin deliberating on Monday and deliver an advisory verdict as soon as next week. The jury verdict is not binding on the judge, who will decide the case.

If the judge rules in Musk’s favor, it could upend OpenAI’s race toward an IPO at a valuation approaching $1 trillion. Meanwhile, xAI is expected to go public as a part of Musk’s rocket company SpaceX as early as June, at a target valuation of $1.75 trillion.

Musk the power-seeker, Altman the liar.

In the first week of the trial, Musk said he was suing to save OpenAI’s mission to build AI safely for the benefit of humanity. This week, Altman denied Musk was a paladin of AI safety and painted him as a power-seeker who wanted to control OpenAI. 

Altman told the jury that in 2017, when Musk and other cofounders were discussing creating a for-profit arm, they asked Musk what would happen to his control over such an entity if he died. “Maybe the control of OpenAI should pass to my children,” Musk said, according to Altman.

Musk’s lawyer shot back, grilling Altman on his alleged history of lying. He pointed out that OpenAI’s former executives Ilya Sutskever and Mira Murati, and former board members Helen Toner and Tasha McCauley, all testified that Altman had lied to them. In 2023, Altman was briefly fired as CEO over the alleged behavior.

Molo also pressed Altman about his personal investments in startups that do business with OpenAI. Altman testified that he tried to steer OpenAI to buying power from the nuclear energy company Helion Energy, a third of which he owns.

(Last Friday, the US House oversight committee launched an investigation into Altman’s potential conflicts of interest. Attorneys general from more than a half-dozen states called for the Securities and Exchange Commission to review them.)

During his closing statement, Molo put Altman’s credibility on the stand again. “Imagine that you’re on a hike, and you come upon one of those wooden bridges that you see on a trail, and it’s over a gorge,” he said. “A woman standing by the entry to the bridge says, ‘Don’t worry—the bridge is built on Sam Altman’s version of the truth.’ Would you walk across that bridge?”

Altman, who sat behind his lawyers, looked up uneasily every time his name was mentioned. 

During her closing argument, Eddy fired back. Musk “never cared about the nonprofit structure,” she said. “What he cared about was winning.” 

Musk, though, was absent. Despite the judge’s order that he remain available, he flew to China with President Trump.

Did Altman promise to keep OpenAI a nonprofit?

During her closing argument, Eddy argued that no testimony or evidence showed any conditions on Musk’s donations, or any promises made by Altman and Brockman to keep the company a nonprofit. “No commitments or promises were made. No restrictions were placed on Mr. Musk’s donations,” she said.

Eddy added that it was evident Musk wasn’t truly committed to keeping OpenAI a nonprofit. She noted that in 2017, he tried to create a for-profit subsidiary and fought a bitter battle with Altman and Brockman to have control over it.

“I was not opposed to there being a small for-profit that provides funding to the nonprofit,” Musk told the jury earlier in the trial, “as long as the tail didn’t wag the dog.” 

Eddy then argued that Musk sued too late, filing in 2024 after the statutes of limitations on his claims ran out. In 2019, OpenAI created a for-profit subsidiary, under which employees and investors received a capped return on their investment. 

But Musk testified that he discovered OpenAI had abandoned its nonprofit mission only in 2022, when Microsoft was preparing to invest $10 billion in OpenAI—a deal that closed in 2023. “I was disturbed to see OpenAI with a $20B valuation,” he texted Altman after reading the news. “This is a bait and switch.”

Musk told the jury that the $20 billion valuation made him realize “the for-profit is the tail wagging the dog.” 

“The 2023 deal was different,” Molo hammered home during his closing argument.

Is OpenAI still a nonprofit committed to its mission?

A central question raised in the last week of trial was whether OpenAI remains a nonprofit committed to developing AGI safely for the benefit of humanity. Eddy, the OpenAI lawyer, argued that the nonprofit still controls the for-profit and seeks to “help AGI turn out well for humanity.” “The OpenAI nonprofit is the best-resourced nonprofit in the world,” thanks to the for-profit, she added.

Molo countered that while the OpenAI’s nonprofit nominally controls the company, it does not do so in practice. OpenAI’s nonprofit and for-profit are controlled by the same people—seven of the nonprofit’s eight board members are on the for-profit’s board. The nonprofit hired employees only a month before the trial started and does work only in grant-making rather than AI research. 

Molo played a video interview of Altman saying that the nonprofit board’s failure to fire him in 2023 was “its own kind of governance failure.”

“We’re left with this nonprofit that doesn’t have any voice,” Jill Horwitz, a law professor at Northwestern University who studies nonprofits, told MIT Technology Review. “It doesn’t have much money, and OpenAI doesn’t think it has any obligation to fund it. It barely has a staff,” she says. “It’s unclear how on earth the nonprofit is supposed to exercise its duties and control the entire company.” 

Civil society groups and policymakers have spoken out against OpenAI’s restructuring over the years. So has Musk, although his own stake in the AI race makes him a dubious champion for the public interest. 

“The public interest in the nonprofit loses, no matter who wins or loses this trial,” says Horwitz.

Jackass for AI safety

Despite US District Judge Yvonne Gonzalez Rogers’s warning during the first week that this trial was not about AI safety, the issue stole the show again. Throughout the trial, the lawyers from both sides traded barbs over the safety track records of ChatGPT (which has allegedly caused teen suicides) and Grok (which has flooded X with porn).  

On the last day of testimony, OpenAI’s lawyer Bradley Wilson handed the judge a small golden trophy of a donkey’s ass, inscribed: “Never stop being a jackass for safety.” 

The trophy belonged to Joshua Achiam, OpenAI’s chief futurist. He testified that he’d warned, when Musk announced in 2018 that he was leaving OpenAI to race toward building AGI, that speed could compromise safety. Musk snapped and called him a “jackass,” said Achiam. His colleagues, including Dario Amodei, now CEO of Anthropic, gave him the trophy to enshrine the diss.

“I don’t want it,” said the judge.
The shenanigans spilled out into the street too. In front of the Oakland courthouse, a protester paraded around wearing a costume of Musk holding a bag of ketamine and driving a Cybertruck. Another held a photo of Sam Altman and a poster reading, “Stop AGI or we’re all gonna die.”

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
Xbox is now XBOX
Tech/AI

Xbox is now XBOX

by admin May 15, 2026
written by admin

Believe it or not, Microsoft appears to be rebranding Xbox.

Believe it or not, Microsoft appears to be rebranding Xbox.

May 15, 2026, 11:10 PM UTC
Vector illustration the Xbox logo.
Vector illustration the Xbox logo.
Tom Warren
Tom Warren is a senior correspondent and author of Notepad, who has been covering all things Microsoft, PC, and tech for over 20 years.

Xbox just allcapsmaxxed: Meet XBOX. This isn’t a joke; Microsoft appears to be actually rebranding Xbox to XBOX. Asha Sharma, Xbox CEO, ran a poll on X earlier this week, asking fans whether Microsoft should use Xbox or XBOX. The results were in favor of XBOX, and the company has now renamed its X account.

Curiously, the Threads and Bluesky accounts for Xbox haven’t been renamed yet, but if Microsoft is going ahead with a rebranding then I expect those will change soon. I asked Microsoft to comment on this potential Xbox rebranding and the company simply referred me to Sharma’s post.

The use of all caps for Xbox is a return to original form, though. Microsoft’s first Xbox logo for its console was all caps, and the company has favored using similar capped versions for the Xbox 360, Xbox One, and Xbox Series X / S console logos.

The apparent rebranding comes just a few weeks after Sharma scrapped Microsoft Gaming and renamed Microsoft’s gaming division back to Xbox. It’s part of Sharma’s continued promise of a “return of Xbox,” which has involved fan-focused console updates, a new Xbox logo, Game Pass pricing changes, and lots more in recent weeks.

Last week Sharma also outlined her first major organizational changes at Xbox and revealed a new Xbox boot-up animation. The reorganization of the Xbox platform team is designed to “build a platform that is affordable, personal, and open by staying close to the work and the people we serve,” according to Sharma.

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.

  • Tom Warren

Most Popular

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
Russia pressures university students to become wartime drone pilots
Tech/AI

Russia pressures university students to become wartime drone pilots

by admin May 15, 2026
written by admin

Nonetheless, Russia’s effort to recruit student drone pilots goes toward its goal of having 168,000 drone operators by the end of 2026, according to the Kyiv Independent. In that sense, Russia is copying the success of Ukraine’s Unmanned Systems Force that became the world’s first standalone military branch focused on drones in June 2024.

The Russian recruitment efforts have typically promised that university students can serve as drone pilots without risking their lives in bloody infantry assaults on Ukrainian trenches and fortifications. But safety is a relative term as constant surveillance and the threat of drone strikes or artillery fire has created a “kill zone” stretching as far as 25 kilometers on both sides of the frontlines, according to the commander of Ukraine’s Unmanned Systems Force in an interview with Ukrainksa Pravda.

The Russian-language news service of BBC News identified 23-year-old Valery Averin as the first known death among the new wave of Russian university students who trained and deployed as drone operators. Averin’s adoptive mother, Oksana Afanasyeva, was informed of her son’s death in a mortar attack on April 6 near the Russian-occupied city of Luhansk in eastern Ukraine.

“The child had been training on a drone for three months, and now we’re throwing him into an assault, into the meat grinder, someone who had never served in the army,” Afanasyeva told BBC News.

Russia has lost an estimated 1.3 million soldiers as battlefield casualties since the start of its full-scale invasion of Ukraine, according to a NATO official cited by news reporting in February 2026. By comparison, Ukrainian casualties were estimated as being between 500,000 and 600,000 over approximately the same period, including killed, wounded, and missing soldiers.

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
Trump and Xi conclude 'very successful' talks but few deals confirmed
Global

Trump and Xi conclude ‘very successful’ talks but few deals confirmed

by admin May 15, 2026
written by admin

According to US trade representative Jamieson Greer, deals on Chinese purchases of US agricultural products have been firmed up. But China’s foreign ministry did not confirm any such new deals, saying only that both sides had agreed to maintain stable trade ties and expand co-operation based on “equality, mutual respect and mutual benefit”.

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
We're Talking About Tipping at Restaurants Again
Lifestyle

We’re Talking About Tipping at Restaurants Again

by admin May 15, 2026
written by admin

Welcome to Open Tab, a weekly roundup of news, gossip, and stories that have stayed open in my tabs all week. Last week we covered the upcoming Anthony Bourdain biopic.

This week is a big one for New York bagel shop owners (and bagel lovers). Bromated flour (which includes the ingredient potassium bromate) is set to be banned in New York state. The ingredient works as extra insurance that your dough will be springy and pliable, but it has shown to be carcinogenic. This has massive implications for bagel and pizza shops, an estimated 80%–90% of whom use brominated flour, as they’ll likely have to find new suppliers. Which may mean a higher price for your next schmear or slice.

In other news, expect more changes for Crumbl, the once viral cookie chain that’s seen slowed growth in recent years. This week the chain’s two cofounders announced they’re stepping down. “This is not goodbye,” reads a statement shared on X. Does this mean America is preparing to enter the next era of desserts? Should we all get into ice cream again?

Also this week: New tipping data from payment platform Square holds good news, there’s a looming protein powder crisis, tomatoes are really expensive, and we dive into the wonders of eating on Route 66.

New data from payment platform Square reveals that tipping averages at full-service restaurants have started trending upward for the first time in four years. Back in 2022, tipping averages were around 15.16%, but over the years they steadily lowered. For the last three quarters, they’ve begun to rise, though hovering at 14.82%.

What actually is the correct percentage to tip? (We here at BA talk about tipping, like, a lot. Seriously.) Personally, I tip at least 20% on every sit-down meal I experience, and I assumed that we all agreed that was the standard. Speaking as a service industry vet, if you’re tipping 14.82% on dinner, you should be made to work a Mother’s Day brunch.

If you were hoping this would be the year that you finally get yoked, I’ve got some bad news: A USDA report reveals a looming whey protein shortage. Suppliers reportedly can’t keep up with the very high demand. According to the report, “Some market contacts indicated that certain suppliers are sold out for the remainder of the year.” I guess we’re going to have to go plant-based this summer.

If you’ve started to find yourself wondering if tomatoes are getting pricier, you aren’t losing your mind. Prices have jumped nearly 40% since this time last year due to issues with weather, tariffs, and transportation costs. Also more expensive? Basically everything else you’re buying at the supermarket, if you hadn’t noticed.

Overall prices were up nearly 3% in April compared to that time last year. Looks like we’re going to have to start adjusting our grocery spending, gals—though I will not be relinquishing my expensive soft cheeses. Gotta make life worth living somehow, you know?

Route 66 occupies a special place in American culture—it’s a celebration of freedom, mobility, and the romance of the open road. But did you know it’s also home to some of the most diverse cuisine in the country? Cornish pasties, Lebanese steakhouses, and, of course, fried onion smashburgers. Read on to discover the wonders of the Mother Road in time to celebrate its recent centennial. Road trip, anyone?

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
AI research papers are getting better, and it’s a big problem for scientists
Tech/AI

AI research papers are getting better, and it’s a big problem for scientists

by admin May 15, 2026
written by admin
268512_PHOTO-_Research_papers_are_overrun_with_AI_slop_CVirginia
268512_PHOTO-_Research_papers_are_overrun_with_AI_slop_CVirginia

Journal editors and peer reviewers are being flooded with AI-generated papers that are almost impossible to detect.

May 15, 2026, 11:00 AM UTC

Last summer, Peter Degen’s postdoctoral supervisor came to him with an unusual problem: One of his papers was being cited too much. Citations are the currency of academia, but there was something unusual about these. Published in 2017, the paper had assessed the accuracy of a particular type of statistical analysis on epidemiological data and had received a respectable few dozen citations in other research papers over the years, but now it was being referenced every few days, hundreds of times, placing it among the most cited papers of his career. Another professor might be thrilled. Degen’s adviser asked him to investigate.

Degen, a postdoctoral researcher at the University of Zurich Center for Reproducible Science and Research Synthesis, found that the citing papers all followed a similar pattern. Like the original, they were analyzing the Global Burden of Disease study, a publicly available dataset compiled by the Institute for Health Metrics and Evaluation at the University of Washington. But they were using the dataset to churn out a seemingly endless supply of predictions: about the future likelihood of stroke among adults over 20 years old, of testicular cancer among young adults, of falls among elderly people in China, of colorectal cancer among people who eat minimal whole grains, of disease X among population Y, and so on.

Searching on GitHub for code that would be used to do this sort of analysis, Degen followed some links and wound up on the Chinese social media site Bilibili, where he discovered a Guangzhou-based company touting tutorials on how to produce publishable research in under two hours using its software tools and AI writing assistance. These studies were not very good. Researchers who analyzed a subset of studies about headaches found they were rife with errors and misrepresentations. But they were also not as flagrantly wrong as AI-generated papers of the recent past, making them more difficult to filter out.

“It’s a huge burden on the peer-review system, which is already at the limit,” Degen said. “There’s just too many papers being published and there’s not enough peer reviewers, and if the LLMs make it so much easier to mass produce papers, then this will reach a breaking point.”

Optimists about generative AI have high hopes for its ability to produce future scientific breakthroughs — accelerating discovery, eliminating most types of cancer — but the technology is currently undermining one of the pillars of scientific research, inundating editors and reviewers with an endless stream of papers. Paradoxically, the better the technology gets at producing competent papers, the worse the crisis becomes.

For the past decade, academic publishing has been contending with so-called “paper mills,” black-market companies that mass-produce papers and sell authorship slots to academics, doctors, or others who hope to gain a competitive edge by having published research on their resumes. It has been a game of cat and mouse, with publishers — often pressed by so-called science sleuths, researchers who specialize in ferreting out fraudulent research — closing one vulnerability only to have the mills find a new one. Generative AI was a boon to the mills, helping them to skirt plagiarism detectors by creating wholly new images and text. Still, the technology’s telltale hallucinations meant that publishers could at least theoretically screen out much of their work. In practice, papers still got through, only to get retracted when sleuths encountered a diagram of a rat with inexplicably gargantuan genitals labeled “testtomcels” or prose sprinkled with “as an AI assistant”s that someone forgot to delete.

But now AI has improved to the point where it can produce convincing papers almost wholesale, allowing desperate academics in need of a publication to mill papers of their own. The result is a deluge of scientific slop that threatens to swamp publishing, peer review, grant making, and the research system as it exists today.

Matt Spick, a lecturer in health and biomedical data analytics at the University of Surrey and an associate editor at Scientific Reports, first noticed the phenomenon when he received three strikingly similar papers analyzing the US National Health and Nutrition Examination Survey (NHANES), another public dataset. He checked Google Scholar and realized that it wasn’t a coincidence: There had been a sudden explosion in papers citing NHANES that all followed a similar formula, each purporting to discover an association between, for example, eating walnuts and cognitive function or drinking skim milk and depression.

“If you’ve got enough computing power, you go through and you measure every single pairwise association, and eventually you find some that haven’t been written on before and you just publish: There is a correlation between this and that,” Spick said. These correlations are often misleading simplifications of phenomena with multiple causes or random statistical flukes. “One was that how many years you spend in education will cause postoperative hernia complications. That is just a random correlation. What am I supposed to do with that? Leave school early so that I won’t get a postoperative hernia complication later?”

Over the years, sleuths have developed a variety of methods for detecting inauthentic papers. Some search for “tortured phrases,” instances where someone was trying to skirt plagiarism detectors by feeding an existing paper through a synonym generator, which often has the effect of turning technical terms like “reinforcement learning” into nonsense like “reinforcement getting to know,” to cite one recent example. Other sleuths track duplicated images, perform network analysis of authors, or check citations for hallucinated publications, a classic sign of LLM use. Spick searches for masses of papers following the same template as they analyze public datasets.

These papers may not necessarily be wrong, though they are often misleading. Nor are they strictly speaking fraudulent. They’re just useless, and suddenly very easy to make. Last year, several journals began restricting submissions of papers analyzing public datasets, citing a flood of redundant research.

Spick fears these measures may be fighting the last battle. In recent months, AI companies have released a range of “agentic” science assistants capable of analyzing data, generating hypotheses, and writing research papers with a high degree of autonomy. While a possible step toward the goal of AI-accelerated science, these systems also come with novel risks. When Carnegie Mellon researchers tested several agentic tools, they found that they sometimes invented data or used misleading techniques, but that these errors were only apparent upon close analysis of the full workflow; the final papers looked polished.

Announcing an AI paper writing assistant earlier this year, OpenAI’s then-vice president for science, Kevin Weil, predicted, “I think 2026 will be for AI and science what 2025 was for AI and software engineering.” Spick and some colleagues, curious what it could do, gave the tool, called Prism, some data from an already published paper documenting ripening times of eggplants and peppers. Prism analyzed the data, proposed a new statistical method that could be applied to it, and wrote an entire paper complete with charts and correct citations.

“We were all looking at each other like, ‘What the [expletive], this is actually a decent piece of work!’” Spick recalled. Unlike the generated papers he’d encountered previously, this one didn’t follow a template, nor was it using a single well-known database. It took 25 minutes and 50 seconds to produce.

“I’m genuinely not sure at what point we will suddenly realize that more are getting through than we realize because we can’t easily tell the difference anymore,” Spick said.

This raises some philosophical questions, Spick said, like: Does it matter who or what writes the paper if the information is accurate? And should science be in the business of publishing every possible fact?

“Part of science is supposed to be the filter. We’re supposed to publish the stuff that we think is interesting, not publish literally everything that we can possibly find,” Spick said. “Because if we do that, science is just spamming the world with all the data, irrespective of whether it constitutes actual new knowledge or not, and in any kind of medium-term time frame, it’s almost impossible to work out what’s meaningful and what isn’t.”

This is the immediate practical challenge posed by AI agents. They threaten to overwhelm the human systems that create and organize knowledge. Research funders are contending with onslaughts of proposals perfectly tailored to their particular grant, unable to parse which projects represent the next step in years of work and which were generated in minutes. Conference organizers, journal editors, and peer reviewers are all struggling to sort through a flood of material that all seems good enough at first glance to warrant a close read. There is an enormous and growing asymmetry between the time it takes to produce new work and the time it takes a subject-matter expert to vet it.

For Marit Moe-Pryce, the managing editor of the international relations journal Security Dialogue, submissions are up 100 percent over where they were a year before. Just as problematic: All the submissions have become pretty good. Gone are the blatant hallucinations and leftover prompts; everything has suddenly become coherent, well structured, and stylistically similar, difficult to say whether it is a wholly generated paper, an experienced academic, or a young scholar using AI as an editor.

“The main problem that we see currently from the desk is that the fraudulent side and the academic side are conflating, which ends up with a big gray mass of articles that we as editors need to sit and try to figure out, ‘What is this? Is this something that we need to engage with? Is it not?’” Moe-Pryce said.

One paper made it past at least 10 editors and two rounds of peer review before she noticed a fake citation — a very plausible one, involving several former editors of the journal on a topic they could have written about but never did. She then found several more. She doesn’t know at what stage of revision the hallucinations were introduced, but the close call underscored the level of care required to ensure nothing false gets published. Now that models increasingly cite real papers, she has to read for whether the works cited are the ones an expert would actually use, AI not yet having mastered the difference between canonical literature and more peripheral work.

“It’s incredibly detailed, and this is a normal part of the editorial work. The difference is that now you have to do that for all the rubbish that comes through the door,” Moe-Pryce said. “That’s why our workload becomes so unmanageable.”

Academic papers go through a multi-stage review process before publication. First, manuscripts are triaged for obvious problems, then sent to a journal’s editor, who decides whether it might be worth publishing. The editor then sends it to an associate editor with experience in the field, who again vets it before recruiting two or three subject-matter specialists — the “peers” in peer review — to read the paper and write responses. The editors and reviewers are typically working for free, volunteering their time in addition to their primary academic job.

The review system was already struggling under increasing volumes of submissions, and now AI is increasing those volumes while also making the bad ones more difficult to filter out. Moe-Pryce now spends more time sorting papers before deciding what to send out for review, and prospective reviewers, swamped themselves, are less and less likely to respond. Where she previously could send four queries out and get three replies, it now takes her a dozen tries to get two people. Increasingly, she reaches out to 20 reviewers and hears nothing.

“It’s fatigue. Academic journals have mushroomed, and then you have AI helping everyone fraudulent or not generate more, faster, so you have a massive increase in volume,” she said. “AI currently holds the potential to bring down the publishing system as we know it.”

The journal Accountability in Research has seen a 60 percent surge in submissions this year, according to David Resnik, an associate editor at the journal. Ironically, he has been besieged by likely AI-generated papers about fraudulent academic papers that have mined public data compiled by the organization Retraction Watch.

He, too, is struggling to find reviewers. At times, he’s had to send out 20 requests just to get two responses — and he’s suspected that some of the responses he’s received are AI-generated themselves. He has reason to be suspicious. A survey conducted by the publishing company Frontiers last year found that more than half of researchers have used AI assistance in their peer review.

“I’m very worried about this straining, breaking the back of the peer-review system,” said Resnik.

AI agents arrive at a time when the quality filters of academia are already struggling to cope with a superabundance of papers. The number of scientific papers published has grown exponentially in recent years, according to an analysis of data published in Quantitative Science Studies, while the number of PhDs who might review them has not. Unfortunately, the authors attribute this explosion in productivity not to rapid progress in science but to the fact that commercial and professional incentives align to publish the maximum quantity of papers.

Many journals have shifted to an “open access” model where they earn revenue by charging authors processing fees to have their papers published, as opposed to charging for subscriptions. In earnings calls, publishing companies tout the recent 20 percent or more increase in submissions as a positive growth story. Universities and funding agencies, meanwhile, look at researchers’ publication metrics when deciding whom to fund or promote, which means researchers are under pressure to “publish or perish.” Nor is it only traditional academics who are under this pressure to publish. Overseas medical students can improve their chance at a US residency program by having a few peer-reviewed papers on their resume. In China, medical doctors have strong incentives to publish despite neither having the time nor resources to conduct research, making quick paper generation an attractive option.

If you introduce an infinite paper-writing machine to a system that defines productivity by the number of papers written, people will use it to write a lot of papers. A study published in Nature this year found that scientists who adopted AI published three times more papers and received nearly five times more citations than those who didn’t. They also became research project leaders 1.37 years earlier than those who did not use AI. While individually beneficial, the embrace of AI to mass-produce papers may be detrimental to science as a collective endeavor, beyond exhausting journal editors and peer reviewers. The same study found a collective narrowing of focus as these newly productive scientists gravitated toward well-studied fields with abundant existing data for AI to synthesize.

There are no easy solutions to this problem. In 2022, the scientific organization STM launched an initiative called Integrity Hub to contend with paper mills. Since then, it has been engaged in an “arms race” with AI, according to Joris van Rossum, the project’s program director — assembling automated tools that check for plagiarism, then tortured phrases, then fake citations — but the group must now consider more sweeping remedies.

“We anticipate a future where it’s going to be more realistic to enable submitters to demonstrate authenticity rather than trying to detect fabrication,” he said. That is, once fraudulent manuscripts are impossible to detect, publishers will have to find a way for researchers to prove their work is real — perhaps by working with instrument manufacturers to develop ways of watermarking their images, he said, or having researchers submit more of the data behind their work so it can be analyzed for suspicious signals.

This would entail changing the way research is done on a massive scale, and while it might stem outright fraud, it would do little to reduce the volume problem. Using AI to assist with peer review, as some have proposed — and some reviewers are already doing, permitted or not — raises a nest of other possible risks. Studies have found that models often continue to cite retracted studies as valid and write superficially good reviews while overlooking methodological problems. AI reviewers also appear to prefer AI-generated writing.

“It’s not really a tractable problem,” said Reese Richardson, a postdoctoral fellow at Northwestern University who studies mass-produced papers. “I think that the only way out of this situation is to actually change the way that the scientific enterprise awards prestige and awards resources. As long as we have this hyper-competitive, hyper-unequal rat race where people’s productivity and their worth as scientists is being measured by how many publications they put out and how many times they get cited, it’s just going to incentivize this behavior.”

Vincent Larivière, the editor-in-chief of Quantitative Science Studies, had a similar diagnosis. His journal has seen a 40 percent increase in submissions this year.

“We need a reform of what matters in science,” Larivière said. The conflation of scientific productivity with publication counts has had a distorting effect on science, causing research to gravitate toward small, tractable problems that are guaranteed to result in something publishable. AI could do great things, he said — help cure cancer, develop fusion energy — but right now it is being used to generate papers to “pad CVs.”

“Of course we need more science,” he said, “but do we need more papers?”

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.

  • Joshua Dzieza
May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
Trump-Xi summit: the 3 big takeaways from historic meeting in Beijing
Economy

Trump-Xi summit: the 3 big takeaways from historic meeting in Beijing

by admin May 15, 2026
written by admin

The national flags of the United States and China hang in front of the portrait of late communist leader Mao Zedong at Tiananmen Gate in Beijing on May 15, 2026.
Brendan Smialowski | Afp | Getty Images

BEIJING — U.S. President Donald Trump’s closely watched visit to China this week has gone a long way toward strengthening a fragile trade truce with Beijing and stabilizing the bilateral relationship.

While the visit was delayed by more than a month due to the Iran war, Trump’s two-day summit with Chinese President Xi Jinping wrapped up Friday with plans for another meeting this fall.

Here’s what’s changed since the leaders met:

U.S.-China geopolitical alignment

Xi’s warning to Trump that mishandling Taiwan would put the U.S.-China relationship into “great jeopardy,” according to official English-language state media, dominated headlines at the start of talks.

Oil prices also rose after Trump told Fox News in a pre-recorded interview that China has agreed to buy U.S. oil and would help with Iran negotiations. He did not reveal when purchases would begin or at what volume.

China has yet to confirm plans to buy U.S. oil, while Washington has yet to say anything on Taiwan.

“I do think each side has delivered. There was no substantive discussion on Taiwan, though, which is not surprising,” said Yue Su, principal economist, China, at the Economist Intelligence Unit. “More discussion on Iran highlighted that they do have common ground. The fact that both sides want to describe the meeting as a win shows goodwill, at least.”

“There are limits to what China can realistically do, as the Iranian regime is operating in survival mode and will prioritize its own interests and agenda above all else,” she said.

Trade truce holds

The U.S. and Chinese sides have not yet released details on specific agreements. But Trump’s invitation to Xi to visit the U.S. on Sept. 24 means the two leaders can talk in person again before the expiration of the one-year trade truce set in October 2025.

The agreement lowered tariffs and rolled back rare earths restrictions after an escalation in tensions between the two countries earlier in 2025.

Xi said the U.S. and China agreed to constructive “strategic stability” as a framework for the next three years, according to state media. 

“Strategically, Beijing appears to be trying to turn Trump’s transactional willingness to stabilize ties into a longer-term operating framework for U.S.-China relations,” said Jack Lee, analyst at China Macro Group, noting the framework could become a baseline on dealing with Beijing for the next U.S. president.

Wins for business

Trump told Fox News that China will order 200 Boeing jets, which he said was more than the 150 units the company had expected. But that was less than half the 500 planes that many initially expected.

Nvidia also reportedly got the green light from the U.S. to sell its H200 chips to major Chinese companies, sending tech stocks higher.

Both Boeing CEO Kelly Ortberg and Nvidia CEO Jensen Huang accompanied Trump to Beijing. The executives and more than a dozen U.S. business leaders — including Apple CEO Tim Cook and Tesla’s Elon Musk — participated in a meeting Thursday with Chinese Premier Li Qiang.

Opening remarks and readouts offered no details beyond China’s pledge to open up its market further to foreign business, which has occurred gradually over recent decades.

The U.S. business delegation was far smaller than the more than 30 leaders that joined Trump on his trip to Saudi Arabia last year.

“I don’t think the purpose was to have every CEO sign a deal,” said Gary Dvorchak, Blueshirt Group managing director. “I think the purpose was just to kind of flex America’s muscles and just show from an economic standpoint what a powerhouse we are.”

“It also shows a high level of unity amongst the American government and private sectors,” he said.

Choose CNBC as your preferred source on Google and never miss a moment from the most trusted name in business news.

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
The world is on track to miss its health targets
Tech/AI

The world is on track to miss its health targets

by admin May 15, 2026
written by admin

Every year the World Health Organization publishes a global health statistics report. It features the numbers behind world health trends and, importantly, assesses whether we’re on track to reach ambitious goals set in 2015. It’s a bit like a health grade.

The 2026 report was published on Wednesday. And the results aren’t looking brilliant. While we are seeing some improvements, they are uneven, and they’re far too slow.

The targets themselves are part of the United Nations’ Sustainable Development Goals, a sprawling and ambitious plan focused on improving life around the world. The 17 goals were set to tackle poverty and climate change and to boost education, gender equality, health, and well-being, among many other quality of life issues. Those targets were meant to be met by 2030.

Perhaps they were a little too ambitious. Here are the numbers and statistics that stood out to me on this year’s world health report card.

1.3 million new cases of HIV in 2024

Before the SDGs, there were the Millennium Development Goals. One MDG target was to halt and reverse the spread of HIV—and that target was exceeded by 2015. Back then, we were considered on track to “end the AIDS epidemic by 2030.”

How depressing, then, to see that in 2024 there were an estimated 1.3 million new cases of HIV. That’s 40% lower than the figure from 2010. But it’s still 1.3 million additional people with HIV. The SDG target is to reduce HIV incidence by 90% by 2030—we’re not likely to meet it.

10.7 million new cases of TB

The picture is even bleaker for tuberculosis, which ranks 10th on the WHO’s list of top global causes of death. The goal was to reduce cases by 80% between 2015 and 2030. So far, cases have only fallen by a measly 12%. And when you break the change down by region, the Americas saw an increase of 13%

An 8.5% rise in malaria cases

And then there’s malaria, the mosquito-borne disease with a 7% fatality rate. The European region has been free of malaria since 2015, but the disease is a significant concern in many countries in the Global South, particularly in Africa. The goal was to lower rates by 90% between 2015 and 2030. In 2024, there were an estimated 282 million cases of malaria globally—representing an 8.5% increase in incidence rates.

Antimalarial drug resistance is a major challenge here—forms of the malaria virus that are resistant to drugs have been confirmed or suspected in eight countries in Africa, according to a separate WHO report. Mosquitoes that are resistant to commonly used insecticides are present in nine African countries. And climate change, which can alter mosquito habitats, may be making things worse.

42.8 million children are wasting

We’re not meeting child health targets, either. Take malnutrition, for example. As of 2024, the global prevalence of wasting in children was 6.6%—that’s a staggering 42.8 million children who are literally wasting away because of a lack of adequate food. On the other end of the spectrum, 5.5% of children are now considered overweight. Both figures were meant to be below 5% by 2030, which now seems unlikely.

Vaccination rates are dropping in the Americas

Progress in improving childhood vaccination coverage has stalled. Globally, an estimated 76% of children are getting their second dose of a measles vaccine—a figure far below the the approximately 95% needed to prevent outbreaks. The Americas currently has lower rates of vaccine coverage for three of the four “core” vaccines than it did in 2015.

22.1 million pandemic-related deaths

And of course the pandemic affected progress toward health goals in more direct ways: 7 million people died of covid-19. The WHO report estimates that, for each of these, there were an additional two “excess” deaths related to the pandemic, due to disruptions in health care, for example. That puts the total figure at 22.1 million pandemic-related deaths.

A woman dies every two minutes from “maternal causes”

Maternal mortality rates fell by about 40% between 2020 and 2023. But today’s rate equates to 712 maternal deaths every single day. That’s one every two minutes. The WHO report notes that we’d have to reduce the mortality rate by almost 15% per year in order to meet the 2030 target. This seems incredibly unlikely, particularly given the recent decimation of US funding for global aid programs, which is expected to result in thousands of additional maternal deaths.

Progress has also slowed in reducing the risk of death from noninfectious diseases like cancer, diabetes and cardiovascular disease. “Overall, neither the world nor any WHO region is currently on track to meet the 2030 SDG target,” the report states.

2.1 billion people struggle to afford health care

Despite plans to make health care more affordable, a significant chunk of the population is being pushed into poverty by health-care costs. In 2022, 2.1 billion people faced financial hardship due to health spending—and 1.6 billion of them were living in or had been pushed into poverty.

Across the board, there have been some important improvements in global health. But the achievements have not gone far enough. “The good news is that there is progress,” says Danaei. “But as always, the glass is half empty.”

This article first appeared in The Checkup, MIT Technology Review’s weekly biotech newsletter. To receive it in your inbox every Thursday, and read articles like this first, sign up here.

May 15, 2026 0 comments
0 FacebookTwitterPinterestEmail
CIA chief visits Cuba as energy crisis worsens
Global

CIA chief visits Cuba as energy crisis worsens

by admin May 14, 2026
written by admin

“During the meeting, Director Ratcliffe and Cuban officials discussed intelligence cooperation, economic stability, and security issues, all against the backdrop that Cuba can no longer be a safe haven for adversaries in the Western Hemisphere,” the official added.

May 14, 2026 0 comments
0 FacebookTwitterPinterestEmail
Newer Posts
Older Posts

Recent Comments

No comments to show.

Follow Us

Recent Posts

  • Why Taiwan became the defining issue in the Trump-Xi talks

    May 16, 2026
  • Musk v. Altman week 3: Musk and Altman traded blows over each other’s credibility. Now the jury will pick a side.

    May 15, 2026
  • Xbox is now XBOX

    May 15, 2026
  • Russia pressures university students to become wartime drone pilots

    May 15, 2026
  • Trump and Xi conclude ‘very successful’ talks but few deals confirmed

    May 15, 2026

Newsletter

Join the BusinessStory newsletter for fresh insights, market analysis, and new stories!

Categories

  • Business (18)
  • Economy (430)
  • Global (449)
  • Investing (8)
  • Lifestyle (117)
  • Tech/AI (1,201)
  • Uncategorized (10)

Our Company

We’re dedicated to telling true stories from all around the world.

  • Ilulissat 3952, Greenland
  • Phone: (686) 587 6876
  • Email: [email protected]
  • Support: [email protected]

About Links

  • About Us
  • Contact
  • Advertise With Us
  • Media Relations
  • Corporate Information
  • Compliance
  • Apps & Products

Useful Links

  • Privacy Policy
  • Terms of Use
  • Closed Captioning Policy
  • Accessibility Statement
  • Personal Information
  • Data Tracking
  • Register New Account

Newsletter

Join the BusinessStory newsletter for fresh insights, market analysis, and new stories!

Latest Posts

Trump and Xi conclude ‘very successful’ talks but few deals confirmed
We’re Talking About Tipping at Restaurants Again
AI research papers are getting better, and it’s a big problem for scientists
Trump-Xi summit: the 3 big takeaways from historic meeting in Beijing

@2025 – All Right Reserved. Designed and Developed by BusinessStory.org

Facebook Twitter Instagram Linkedin Youtube Email
  • Home
  • Investing
  • Global
  • Business
  • Economy
  • Tech/AI
  • Lifestyle
  • About Us
  • Contact