Submitted by tmblweeds t3_zn0juq in MachineLearning
Seon9 t1_j0foo1j wrote
Reply to comment by Nowado in [P] Medical question-answering without hallucinating by tmblweeds
I think it's worse than Google or WebMD because people are biased toward believing detailed (richer) information sources, esp if it's masked behind AI. You might receive vague advice from a friend or WebMD but the information is weighted against the source's trustworthiness, perceived soundness, information richness, etc. I think this oversteps alot of that for most people and people might be more inclined to believe it.
It also overlooks that research ≠ clinical practice and that translating research into practice requires numerous hurdles. There can be a decade of research and numerous publication for a drug that then collapses in clinical testing. Or misattribution, that deaths diagnosed from shaken babies syndrome are prob due to prior non-shaking related trauma. I think access to research is a good thing but stripped of context like this isn't.
I also wonder if the hallucinations increase as the pool of information shrinks. Like I'm not gonna use this if I have a common cold but if I have rare S3 colorectal cancer then I might find this useful. Cool project but medical applications are always tough.
NotMyMain007 t1_j0i3nwz wrote
I do agree with you, but at the same time, there is some really shitty places on earth, it would be a good option to have instead of just laying on the ground and dying. Even more if this keep evolving.
Viewing a single comment thread. View all comments