Submitted by viertys t3_125xdrq in MachineLearning
BrotherAmazing t1_je7vj9v wrote
People saying get more than 100 images are right (all else being equal, yes, get more images!) but you likely can make good progress without as many images for your problem with clever augmentation and a smaller network.
Here’s why:
-
You only have to detect cavities. It’s not some 1,000-class semantic segmentation problem.
-
You should be working with single channel grayscale images, and not pixels that naturally come in 3-channel RGB color.
-
This is X-ray data just of teeth, so you don’t have nearly the amount of complex fine-detailed textures and patterns (with colors) that are exhibited in more general RGB optical datasets of all sorts of objects and environments.
Of course for a real operational system that you will use in commercial products you will need to get far more than 100 images. However, for a simple research problem or prototype demo, you can likely show good results and feasibility (without overfitting, yes) on your dataset with a smaller net and clever augmentation.
viertys OP t1_je9nno8 wrote
I didn't mention it in the post, but I'm using the albumentations module. I rotate, shift, rotate, blur, horizontal flip, downscale and use gauss noise. I get around 400 images after doing this. Is there anything you would suggest?
I have an accuracy of 98.50 and I have dice of around 0.30-0.65 in each image
And yes, the images are grayscale and they are cropped around the teeth area, so only that part of the X-ray remains.
Viewing a single comment thread. View all comments