The Byte – “A new study highlights a glaring hole in AI models’ ability to learn new information: turns out, they can’t! According to the study, conducted by a team of scientists at Canada’s University of Alberta and published this week in the journal Nature, AI algorithms trained via deep learning — in short, AI models like large language models built by finding patterns in heaps of data — fail to work in “continual learning settings,” or when new concepts are introduced to a model’s existing training. In other words, if you want to teach an existing deep learning model something new, you’ll likely have to retrain it from the ground up — otherwise, according to the research, the artificial neurons in their proverbial minds will sink to a value of zero. This results in a loss of “plasticity,” or their ability to learn at all. “If you think of it like your brain, then it’ll be like 90 percent of the neurons are dead,” University of Alberta computer scientist and lead study author Shibhansh Dohare told New Scientist. “There’s just not enough left for you to learn.” And training advanced AI models, as the researchers point out, is a cumbersome and wildly expensive process — making this a major financial obstacle for AI companies, which burn through a ton of cash as it is…”
Sorry, comments are closed for this post.