Accurate, Focused Research on Law, Technology and Knowledge Discovery Since 2002

The risks of AI for scientific research

Ars Technica – “Producing more but understanding less. A psychologist and an anthropologist ponder the epistemic risks AI could pose for science: Last month, we witnessed the viral sensation of several egregiously bad AI-generated figures published in a peer-reviewed article in Frontiers, a reputable scientific journal. Scientists on social media expressed equal parts shock and ridicule at the images, one of which featured a rat with grotesquely large and bizarre genitals. As Ars Senior Health Reporter Beth Mole reported, looking closer only revealed more flaws, including the labels “dissilced,” “Stemm cells,” “iollotte sserotgomar,” and “dck.” Figure 2 was less graphic but equally mangled, rife with nonsense text and baffling images. Ditto for Figure 3, a collage of small circular images densely annotated with gibberish. The paper has since been retracted, but that eye-popping rat organ image will remain indelibly imprinted on our collective consciousness. The incident reinforces a growing concern that the increasing use of AI will make published scientific research less trustworthy, even as it increases productivity. While the proliferation of errors is a valid concern, especially in the early days of AI tools like ChatGPT, two researchers argue in a new perspective published in the journal Nature that AI also poses potential long-term epistemic risks to the practice of science. Molly Crockett is a psychologist at Princeton University who routinely collaborates with researchers from other disciplines in her research into how people learn and make decisions in social situations. Her co-author, Lisa Messeri, is an anthropologist at Yale University whose research focuses on science and technology studies (STS), analyzing the norms and consequences of scientific and technological communities as they forge new fields of knowledge and invention—like AI…”

Sorry, comments are closed for this post.