top | item 17362095

Using Deep Learning to Help Pathologists Find Tumors

89 points| rusht | 7 years ago |research.baidu.com | reply

34 comments

order
[+] whafro|7 years ago|reply
I work in this field (not directly on the ML, for the company born out of the winner of Camelyon16) and the last two years of progress has been amazing to watch. Tumor detection has become incredibly accurate, across basically every tissue/tumor type, and we're now making real progress on the next major goal: determining the best therapy for a given patient.

It's a bit of a dirty secret in this space that pathologists have a pretty high error rate on a lot of these tasks — it's just tough work for human eyes to do literally hundreds of times every day. Applying computer vision techniques can not only improve accuracy and reproducibility over human assessment, but you can do types of analysis in seconds-to-minutes that would literally take years for a human. We're just scratching the surface.

There are lots of ML challenges here, but just as many general tech/engineering/design challenges. So if you're interested in working on bringing work like this to the masses, we'd love to talk at PathAI.

[+] ska|7 years ago|reply
That's a bit of an optimistic take. For example, I'm familiar with a number of areas with the imaging community is getting much better at academic challenges, but the resulting models all generalize quite poorly. This has a lot to do with the lack of sufficient labeled data, but it doesn't help that the understanding of tumor morphology is changing pretty rapidly.

It's true that screening is a particularly interesting application because of the issues of fatigue and low true positive rates. On the other hand, decades ago (i.e. well before deep learning approaches) we had clinically approved classifiers that did better than average radiologists for some of these tasks and the uptake still hasn't been that impressive. Lot's of non-technical issues around making stuff like this standard of care.

[+] poutrathor|7 years ago|reply
To give a counter point of view here, a close relative of mine is a pathologists. We talked a lot about this subject and so far he is unimpressed by the results. Moreover, it seems that the ability to provide fast (in seconds to minutes) analysis implies powerful computational ability and data network. Finally, the data acquisition right now takes more or less 20 to 45 minutes depending on the patient and the targeted area, which means the promised analysis quickness is not such a real pro.

I trust that science, research and ongoing work will end up providing interesting results, but many many startups will burn their cash before being able to provide real world usage services.

Like you said, there are LOTs of challenges here. Certainly not a "low hanging fruit", not that a profitable business either since most countries will squash costs anywhere they can because the health cost keeps growing, and a very difficult legal environment to deal with.

However, I am very thankful for your hard work and will to push toward a better future for health.

[+] ccarter84|7 years ago|reply
Do you think as we continue to get better about detection, that we'll find an increasing array of slow-growing tumors present in much of the body? I.e. ones that may not need a full round of radiation / chemo for perhaps 5 years and perhaps more targeted approaches can be deployed?

Just something i've been wondering about since I've got cancer on both sides of family and have been pondering doing full-body scans (which still seem quick excessive on the risk/reward)

[+] mkstowegnv|7 years ago|reply
I went to a talk by someone who had switched fields to one that involved analyzing portions of cells. After showing a series of slides with diverse, confusing blobs and lines, he said "when I first started this work I would look at a section and not see anything at all. But I have improved to the point that now I can look at a section and see anything I want to".
[+] toolslive|7 years ago|reply
I did a project like this early 2000s, and it's amazing how far you get by just combining frequency filtering and knn-clustering. Nothing fancy required. really.
[+] sacado2|7 years ago|reply
Another virtue of non-fancy methods is that they can provide a basic explanation of their reasoning, which is very important in the medical field. I think one of the main drawbacks of deep learning is its opacity.
[+] Gatsky|7 years ago|reply
At the moment, a big limitation of this approach is the input data. Images of tumours are generally very thin sections of a complex 3D tissue that is processed in a way that introduces artefacts and then stained with 2 colours.

To truly leverage the power of machine learning, an end to end solution where the tissue is processed in a more data rich manner would be better (eg spatially aware single cell assays, non destructive thick slice imaging). This would feasibly replace the current system entirely, as it truly would do something no human could do, not just do it more accurately.

[+] phonebucket|7 years ago|reply
While open sourcing the model is nice, it would be better still to open source the data set for the wider community to make more meaningful contributions.

Their GitHub repo states the following: "You need to apply for data access, and once it's approved, you can download from either Google Drive, or Baidu Pan."

[+] yil8|7 years ago|reply
We, Baidu Research, do not own the Camelyon16 Challenge dataset, and people need to apply on Camelyon16 Challenge website to download the original pathology slides. I guess my wording was bit confusing on github, which has been corrected, lol
[+] louden|7 years ago|reply
It would be nice to see the sensitivity and specificity of the technique and for humans. False positives and false negatives are not equal in medicine, so we should report in such a way that people can evaluate them.

In this type of cancer, a lower specificity is an acceptable trade off for a very high sensitivity.

[+] yil8|7 years ago|reply
There was indeed a professional pathologist involved in the Camelyon 16 Challenge, where s/he spent 30 hours reviewing 130 slides, and ended up 72.4% sensitivity with 0 false positives. Our algorithm achieves ~91% sensitive at 8 false positives per slides, seems a win according to your "a lower specificity is an acceptable trade off for a very high sensitivity."
[+] sooheon|7 years ago|reply
How is the "grid of patches" different from one more level of convolution?
[+] yorwba|7 years ago|reply
Using another level of convolution would produce outputs that are statistically independent if they are farther apart than the size of the convolution kernel. In a conditional random field, the dependence of outputs on each other can be modeled as well.

For example, a conditional random field could express "either these patches both contain a tumor or none of them does" (which is helpful when there's something suspicious on the patch boundary) and the consequences of committing to either possibility can propagate over the whole field. In contrast, a convolutional layer would have to make the decision independently for each local area.

[+] yil8|7 years ago|reply
You could use more levels of convolution with larger receptive field. But this corresponds to larger patches, e.g. 512x512 pixels, and larger patches sometimes may not just be pure tumor cells or pure normal cells. And if you are just predicting 1 label for larger patches, it sometimes confuses the learning. What we propose with CRF, is larger receptive field with dense predictions, e.g. predicting more than one labels, and we use CRF to model the correlation between labels.
[+] bitL|7 years ago|reply
Doesn't a grid of patches contain much more information than just one more level of convolution?
[+] gwenzek|7 years ago|reply
Maybe computation cost? Not sure.