(no title)
biotechbio | 7 months ago
- Sure, cancer can develop years before diagnosis. Pre-cancerous clones harboring somatic mutations can exist for decades before transformation into malignant disease.
- The eternal challenge in ctDNA is achieving a "useful" sensitivity and specificity. For example, imagine you take some of your blood, extract the DNA floating in the plasma, hybrid-capture enrich for DNA in cancer driver genes, sequence super deep, call variants, do some filtering to remove noise and whatnot, and then you find some low allelic fraction mutations in TP53. What can you do about this? I don't know. Many of us have background somatic mutations speckled throughout our body as we age. Over age ~50, most of us are liable to have some kind of pre-cancerous clones in the esophagus, prostate, or blood (due to CHIP). Many of the popular MCED tests (e.g. Grail's Galleri) use signals other than mutations (e.g. methylation status) to improve this sensitivity / specificity profile, but I'm not convinced its actually good enough to be useful at the population level.
- The cost-effectiveness of most follow on screening is not viable for the given sensitivity-specificity profile of MCED assays (Grail would disagree). To achieve this, we would need things like downstream screening to be drastically cheaper, or possibly a tiered non-invasive screening strategy with increasing specificity to be viable (e.g. Harbinger Health).
mapt|7 months ago
It COULD be used to craft a pipeline that dramatically improved everyone's health. It would take probably a decade or two of testing (an annual MRI, an annual sequencing effort, an annual very wide blood panel) in a longitudinal study with >10^6 people to start to show significant reductions in overall cancer mortality and improvements in diagnostics of serious illnesses. The diagnostic merit is almost certainly hiding in the data at high N.
The odds are that most of the useful things we would find from this are serendipitous - we wouldn't even know what we were looking at right now, first we need tons of training data thrown into a machine learning algorithm. We need to watch somebody who's going to be diagnosed with cancer 14 years from now, and see what their markers and imaging are like right now, and form a predictive model that differentiates between them and other people who don't end up with cancer 14 years from now. We [now] have the technology for picking through complex multidimensional data looking for signals exactly like this.
In the meantime, though, you have to deal with the fact that the system is set up exclusively for profitable care of well-progressed illnesses. It would be very expensive to run such a trial, over a long period of time, and the administrators would feel ethically bound to unblind and then report on every tiny incidentaloma, which completely fucks the training process.
This US is institutionally unable to run this study. The UK or China might, though.
aquafox|7 months ago
The child of a friend of mine has PTEN-Hamartom-Tumor-Syndrom, a tendency to develop tumors throughout life due to a mutation in the PTEN gene. The poor child gets whole body MRIs and other check-ups every half year. As someone in biological data science, I always tell the parents how difficult it will be to prevent false positives, because we don't have a lot of data on routine full body check-ups on healty people. We just know the huge spectrum on how healthy/ok tissue looks like.
spease|7 months ago
I wonder if our current research product is only considered the gold standard because doing things in a probabilistic way is the only way we can manage the complexity of the human body to date.
It’s like me running an application many, many times with many different configurations and datasets, while scanning some memory addresses at runtime before and after the test runs, to figure out whether a specific bug exists in a specific feature.
Wouldn’t it be a lot easier if I could look at the relevant function in the source code and understand its implementation to determine whether it was logically possible based on the implementation?
We currently don’t have the ability to decompile the human body, or understand the way it’s “implemented”, but that is something that tech is rapidly developing tools that could be used for such a thing. Either a way to corroborate enough information aggregated about the human body “in mind” than any person can in one lifetime and reason about it, or a way to simulate it with enough granularity to be meaningful.
Alternatively, the double-blindedness of a study might not be as necessary if you can continually objectively quantify the agreement of the results with the hypothesis.
Ie if your AI model is reporting low agreement while the researchers are reporting high agreement, that could be a signal that external investigation is warranted, or prompt the researchers to question their own biases where they would’ve previously succumbed to confidence bias.
All of this is fuzzy anyway - we likely will not ever understand everything at 100% or have perfect outcomes, but if you can cut the overhead of each study down by an order of magnitude, you can run more studies to fine-tune the results.
Alternatively, you can have an AI passively running studies to verify reproducibility and flag cases where it fails, whereas now the way the system values contributions makes it far less useful for a human author to invest the time, effort, and money. Ie improve recovery from a bad study a lot quicker rather than improve the accuracy.
EDIT: These are probably all ideas other people have had before, so sorry to anyone who reaches the end of my brainstorming and didn’t come out with anything new. :)
zaptheimpaler|7 months ago
Like if we had some kind of prophylactic cancer treatment that was easy/cheap/safe enough to recommend to people even on mild suspicion of cancer with false positives, we could offer it to positive tests. Maybe even just lifestyle interventions if those are proven to work. That's probably very difficult though, just dreaming out loud.
m463|7 months ago
It gives people the agency to alter their lifestyle trajectory.
I personally suspect that people get and cure cancer all the time.
I wonder if cancer is just damage to your body - either a lot of direct damage or interfering with the body's ability to manage/heal itself.
if someone was pre-cancer, would it help to exercise, cut out sugar, use the sauna, stop overeating? I'll bet it might make a difference
amy_petrik|7 months ago
the problem is you do the test for 7 billion people, say, 30 times over their life... 210000000000 tests. imagine how many false negatives and false positives, the cost of follow up testing only to find... false positive. the cost of telling someone they have cancer when they don't. the anger of telling someone they are free of cancer, only to find out they had it all along
this tech isn't that good, nowhere near it, more like a 1 in 100 or 10 in 100 rate of "being wrong". those numbers can get cheesed towards more false positives or false negatives.
as for grail, they tried to achieve this and printed OK numbers... ... .. but their test set was their training set. so the performance metrics went to shit when they rolled it out to production
pas|7 months ago
marcosdumay|7 months ago
edwardog|7 months ago
mbreese|7 months ago
So, more like — did the tumor come back? And if that does happen, with ctDNA, can you detect that there is a relapse before you would otherwise find it with standard imaging. Most studies I’ve seen have shown that this happens and ctDNA is a good biomarker for early detection of relapse.
The case for proactively looking for circulating tumor DNA without an initial diagnosis or underlying genetic condition is a bit dicier IMHO. For example, what if really like to know (I haven’t read this article, but I’m pretty familiar with the field) is how many people had a detectable cancer in their plasma (ctDNA), but didn’t receive a cancer diagnosis. It’s been known for a while that you can detect precancerous lesions well before a formal cancer diagnosis. But, what’s still an open question AFAIK, is how many people have precancerous lesions or positive ctDNA hits that don’t form a tumor?
(I’ve done a little work in this area)
refurb|7 months ago
And the question would be “do I believe the test when it tells me the cancer is gone?” When you know it’s not 100% accurate?
Or do you always do the adjuvant treatment considering the very small chance the test is wrong has a very high cost (death)?
tptacek|7 months ago
rscho|7 months ago
Spooky23|7 months ago
I lost my wife to melanoma that metastasized to her brain after cancerous mole and margin was removed 4 years earlier. They did due diligence and by all signs there was no evidence of recurrence, until there was. They think that the tumor appeared 2-3 months before symptoms (headaches) appeared, so it was unlikely that you’d discover it otherwise.
With something like this, maybe you could get lower dose immunotherapy that would help your body eradicate the cancer?
psadri|7 months ago
ethan_smith|7 months ago
im3w1l|7 months ago
aetherspawn|7 months ago
ada1981|7 months ago
eps|7 months ago
What is CHIP?
bglazer|7 months ago
It’s when bone marrow cells acquire mutations and expand to take up a noticeable proportion of all your bone marrow cells, but they’re not fully malignant, expanding out of control.
biotechbio|7 months ago
ajb|7 months ago
There are a lot of companies right now trying to apply AI to health, but what they are ignoring is that there are orders of magnitude less health data per person than there are cat pictures. (My phone probably contains 10^10 bits of cat pictures and my health record probably 10^3 bits, if that). But it's not wrong to try to apply AI, because we know that all processes leak information, including biological ones; and ML is a generic tool for extracting signal from noise, given sufficient data.
But our health information gathering systems are engineered to deal with individual very specific hypotheses generated by experts, which require high quality measurements of specific individual metrics that some expert, such as yourself, have figured may be relevant. So we get high quality data, in very small quantities -a few bits per measurement.
Suppose you invent a new cheap sensor for extracting large (10^7+ bits/day) quantities of information about human biochemistry, perhaps from excretions, or blood. You run a longitudinal study collecting this information from a cohort and start training a model to predict every health outcome.
What are the properties of the bits collected by such a sensor, that would make such a process likely to work out? The bits need to be "sufficiently heterogeneous" (but not necessarily independent) and their indexes need to be sufficiently stable (in some sense). What is not required if for specific individual data items to be measured with high quality. Because some information about the original that we're interested in (even though we don't know exactly what it is) will leak into the other measurements.
I predict that designs for such sensors, which cheaply perform large numbers of low quality measurements are would result in breakthroughs what in detection and treatment, by allowing ML to be applied to the problem effectively.
im3w1l|7 months ago
A chemosensor also sounds like a useful thing it should give concentration by time. Minimally invasive option would be to monitor breath, better signal in blood.
standingca|7 months ago
I'd be really curious to see how longitudinal results of sequencing + data banking, plus other routine bloodwork, could lead to early detection and better health outcomes.
rscho|7 months ago
gleenn|7 months ago
melagonster|7 months ago