The internal metrics are one thing, but what do real scientists think? Google had human biomedical researchers evaluate the robot's proposals, and they reportedly rated the AI co-scientist higher than other, less specialized agentic AI systems. The experts also agreed the AI co-scientist's outputs showed greater potential for impact and novelty compared to standard AI models.
Too bad you can't trust its output on contextual datasets, especially as the data sets get larger than the token limits.That doesn't mean the AI-co-scientist won't be useful, though. Google's new AI could help humans interpret and contextualize expansive data sets and bodies of research, even if it can't understand or offer true insights.
Indeed! That's one of the main tools in the field.The entire field of computational biology has been about eeking out slightly better hypotheses from millions or billions of options.
They mistook the Copilot key for their CTRL.That's a very oddly specific measurement. Wouldn't it be better to rate it vs someone formulating their own hypotheses? It's like saying "scientists rated toddlers higher than babies at research proposals" -- it doesn't really mean anything because it's comparing things that were never built to do such a thing.
Too bad you can't trust its output on contextual datasets, especially as the data sets get larger than the token limits.
This is literally a lower bar for experimental quality than what we did for bio 201 lab in college.I find it a bit amusing that they have a system that is trying to find the best experimental design to test a given hypothesis and the best they can do to test their own system is to compare it to older systems in a bar graph without any statistical tests to assess if the differences are significant.
this probably will save someone's life.For example, the AI suggested repurposing certain drugs for treating acute myeloid leukemia, and laboratory testing suggested it was a viable idea.
Suggesting a new avenue of treatment for cancer is not considered to be a "true insight" and it still "can't understand". A bit more optimism is warranted.Google's new AI could help humans interpret and contextualize expansive data sets and bodies of research, even if it can't understand or offer true insights.
I can't remember the last time I read a paper that didn't include a "additional study in [various areas] is warranted" statement in the conclusion. If a model was fed those papers, it has a handy list from which to start.[EXAMPLE]this probably will save someone's life.
It's followed by [QUALIFIER]
Suggesting a new avenue of treatment for cancer is not considered to be a "true insight" and it still "can't understand". A bit more optimism is warranted.
I think this is more a divergence between the actual reported news and the community cynicism.i like how we keep oscillating between "AI isn't really doing anything new" and "but hey it's actually getting real results in labs"
If there’s anything that might change people’s views, that might actually work. Or not.this probably will save someone's life
People have invested a lot of money in AI, they need a return on their investment.Ah, a stochastic system in search for an application.
Refreshing use of time and money.
I mean, some of us have been using LLMs for the last year or more to do exactly this? Helping with brainstorming is one the strong-suits of LLMs already.Meanwhile, this technique could also be useful to find the gaps in existing research that might make good starting points for new research.
Have you taken a look at the US president lately? There are very clearly zero limits here.I’d like to think eventually truth is undeniable but the human capacity for self deception doesn’t seem to have any actual limits.
For example, the AI suggested repurposing certain drugs for treating acute myeloid leukemia, and laboratory testing suggested it was a viable idea.
If I had a dollar for every drug repurposing experiment that had initial promising lab results and went nowhere, I could probably fund my lab for a year.this probably will save someone's life.
Well. That’s the cult leader deceiving others. Most don’t actual buy what they are selling.Have you taken a look at the US president lately? There are very clearly zero limits here.
i would also not call "showing greater potential for impact and novelty" as a metric appropriate to make comparisons. it seems discretionary.That's a very oddly specific measurement. Wouldn't it be better to rate it vs someone formulating their own hypotheses? It's like saying "scientists rated toddlers higher than babies at research proposals" -- it doesn't really mean anything because it's comparing things that were never built to do such a thing.
I think it's more like most of the low-hanging fruit has already been picked.If I had a dollar for every drug repurposing experiment that had initial promising lab results and went nowhere, I could probably fund my lab for a year.
We all wish this wasn't so, but drug repurposing isn't proving to be the cheap panacea many hoped it would. (It does sometimes work; but it's pretty rare.)
The truth is that not all things are usefully quantifiable.i would also not call "showing greater potential for impact and novelty" as a metric appropriate to make comparisons. it seems discretionary.
and yes, saying that spoon is better at digging than a fork is useless without information how good a fork is at digging to have a frame of reference.
It's Google man so everybody's gonna hate on top of the AI hate in general.i like how we keep oscillating between "AI isn't really doing anything new" and "but hey it's actually getting real results in labs"
the framing of "this is just a chatbot" feels disconnected when we're seeing actual wet lab validation of its hypotheses. like, yes, it's "just" pattern matching, but so is most of early stage research? connecting dots others haven't connected is literally how science progresses
this probably will save someone's life.
Just look at the fights we have now over vaccines, and hormones, and pharma, and Waymo, and so on. "Lives saved" is just wood for the fire.If there’s anything that might change people’s views, that might actually work. Or not.
Did the AI have access to his computer and unpublished work? No. Is it plausible that the guy, or his competitors or colleagues, never published anything over the past decade? No. Is it plausible that someone else in the field might have come up with hypotheses similar to his? Absolutely. So was the AI working from where he started a decade ago, or from where the state of this field of research is today?The use of "co-scientist" research is mentioned today on BBC.
"AI cracks superbug problem in two days that took scientists years
A complex problem that took microbiologists a decade to get to the bottom of has been solved in just two days by a new artificial intelligence (AI) tool.
Professor José R Penadés and his team at Imperial College London had spent years working out and proving why some superbugs are immune to antibiotics.
He gave "co-scientist" - a tool made by Google - a short prompt asking it about the core problem he had been investigating and it reached the same conclusion in 48 hours."
www.bbc.co.uk/news/article...
Additionally, this is confirmation bias at work. "This tool told me what I am working on is valuable, therefore this is a good tool."Did the AI have access to his computer and unpublished work? No. Is it plausible that the guy, or his competitors or colleagues, never published anything over the past decade? No. Is it plausible that someone else in the field might have come up with hypotheses similar to his? Absolutely. So was the AI working from where he started a decade ago, or from where the state of this field of research is today?
If I told you a decade ago what computers today would be capable of, I'd be a visionary. If I told you today, I'd be restating what we already know. That's not as impressive as the article makes it sound.
And note that the AI didn't "solve" the issue. The AI simply suggested that the angle this guy was working might be right, but nothing has yet been proven. That's how hypotheses work.
The BBC article cited by MentalVerse above is a direct answer to this -- the Google Co-scientist delivered a team's research direction (among others) in days instead of years. User "caution live frogs" wonders if it cheated and stole ideas from the researcher's PC -- Google says no but who knows? "caution live frogs" raises other questions that might detract from the accomplishment, all valid things to be aware of and check out, and "GenericAnimeBoy" raises the very real concern of who owns/uses the tool's products. But does that mean you will never use the tool? What is wrong with using a tool that might yield benefits to the researcher? Scientific progress often relies on serendipity -- is using this new tool worse than leaving the petri dishes on the window sill?Wouldn't it be better to rate it vs someone formulating their own hypotheses?