Viewing a single comment thread. View all comments

goodDogsAndSam t1_isp1l07 wrote

There are a number of startups in this space (diagnostic ML), across a bunch of different health conditions and underlying datasets. The FDA has a procedure for getting clinical approval to sell/deploy ML systems in healthcare settings, and has greenlit a number of products:
https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices

2

Overall-Importance54 OP t1_isp9sz7 wrote

This is a cool comment, thank you. Maybe it's the field of opportunities it looks like. Like, early internet, let's make a website that has a directory of other websites kind of opportunity. Feels like all the low-hanging fruit is still on the tree.

1

goodDogsAndSam t1_ispbg92 wrote

IMO the biggest problem is not the ML side, it's the people-workflow-deployment side (as other commenters have pointed out) -- a radiologist has spent 18 years of schooling plus a residency, how can you position AI to help them do their job better rather than presumptively challenge their expertise? The cost of error (particularly false negatives) is orders of magnitude higher than the benefit of getting the vast majority of true-negatives correct, what's the right way to tune the model?

Also, based on my experience in the space, there's plenty of training data out there, but outside of routine preventive scans like mammograms, many diagnoses don't have a lot of "clean" negative examples, because doctors won't order a CT unless they have reason to suspect something is wrong. This is unlikely to change, since the radiation exposure from the scan poses individual harm to the patient without providing individual benefit.

1