←back to thread

279 points bookofjoe | 1 comments | | HN request time: 0s | source
Show context
biotechbio ◴[] No.44609723[source]
Some thoughts on this as someone working on circulating-tumor DNA for the last decade or so:

- Sure, cancer can develop years before diagnosis. Pre-cancerous clones harboring somatic mutations can exist for decades before transformation into malignant disease.

- The eternal challenge in ctDNA is achieving a "useful" sensitivity and specificity. For example, imagine you take some of your blood, extract the DNA floating in the plasma, hybrid-capture enrich for DNA in cancer driver genes, sequence super deep, call variants, do some filtering to remove noise and whatnot, and then you find some low allelic fraction mutations in TP53. What can you do about this? I don't know. Many of us have background somatic mutations speckled throughout our body as we age. Over age ~50, most of us are liable to have some kind of pre-cancerous clones in the esophagus, prostate, or blood (due to CHIP). Many of the popular MCED tests (e.g. Grail's Galleri) use signals other than mutations (e.g. methylation status) to improve this sensitivity / specificity profile, but I'm not convinced its actually good enough to be useful at the population level.

- The cost-effectiveness of most follow on screening is not viable for the given sensitivity-specificity profile of MCED assays (Grail would disagree). To achieve this, we would need things like downstream screening to be drastically cheaper, or possibly a tiered non-invasive screening strategy with increasing specificity to be viable (e.g. Harbinger Health).

replies(10): >>44610108 #>>44610191 #>>44610539 #>>44610565 #>>44610758 #>>44611096 #>>44611258 #>>44612058 #>>44612114 #>>44612576 #
ajb ◴[] No.44610758[source]
Here's what may seem like an unrelated question in response: how can we get 10^7+ bits of information out of the human body every day?

There are a lot of companies right now trying to apply AI to health, but what they are ignoring is that there are orders of magnitude less health data per person than there are cat pictures. (My phone probably contains 10^10 bits of cat pictures and my health record probably 10^3 bits, if that). But it's not wrong to try to apply AI, because we know that all processes leak information, including biological ones; and ML is a generic tool for extracting signal from noise, given sufficient data.

But our health information gathering systems are engineered to deal with individual very specific hypotheses generated by experts, which require high quality measurements of specific individual metrics that some expert, such as yourself, have figured may be relevant. So we get high quality data, in very small quantities -a few bits per measurement.

Suppose you invent a new cheap sensor for extracting large (10^7+ bits/day) quantities of information about human biochemistry, perhaps from excretions, or blood. You run a longitudinal study collecting this information from a cohort and start training a model to predict every health outcome.

What are the properties of the bits collected by such a sensor, that would make such a process likely to work out? The bits need to be "sufficiently heterogeneous" (but not necessarily independent) and their indexes need to be sufficiently stable (in some sense). What is not required if for specific individual data items to be measured with high quality. Because some information about the original that we're interested in (even though we don't know exactly what it is) will leak into the other measurements.

I predict that designs for such sensors, which cheaply perform large numbers of low quality measurements are would result in breakthroughs what in detection and treatment, by allowing ML to be applied to the problem effectively.

replies(5): >>44610815 #>>44610880 #>>44611051 #>>44612179 #>>44612833 #
rscho ◴[] No.44610880[source]
Last time someone tried to inject chips into the bloodstream, public opinion didn't handle it too well. It's the same as we would learn a lot by being more cruel to research animals. But most people have other priorities. Good or bad ? Who knows ? Research meets social constructs.
replies(2): >>44610914 #>>44610920 #
ajb ◴[] No.44610920[source]
I am not proposing injecting chips.
replies(1): >>44611015 #
rscho ◴[] No.44611015[source]
Apart from the likely technical infeasibility of your idea in today's society, this would require a humongous and diversified population sample to be meaningful (your 'heterogeneous bits'). This follows directly from the complexity of metabolic pathways you wish to analyze. Socially, you'll only be able to achieve that by not asking your sample for consent. Otherwise you'll have a highly biased sample, which could still be useful but for severely restricted research questions.
replies(1): >>44611222 #
1. ajb ◴[] No.44611222[source]
There are some pretty big longitudinal studies with consent ( "45 and up" are a quarter of a million people, for example - that's big enough that working predictions within the cohort would be a worthwhile health outcome).

There are nevertheless privacy issues, which I did not address as my first comment was already very long, especially for a tangent. Most obviously, people would be consenting to the collection of data whose significance they cannot reasonably forsee.

I do agree that most current AI companies are unlikely to be a good steward of such data, and the current rush to give away health records needs to stop. In a way it's a good thing that health records are currently so limited, since the costs will so obviously outweigh the benefits.