That’s LLM AI, but the type I’m talking about is the machine learning kind. I can envision a system that takes e.g. a sample’s test data and provides a summary, which is not far from what doctors do anyway. If you ever get a blood test’s results explained to you it’s “this value is high, which would be concerning except that this other value is not high, so you’re probably fine regarding X. However, I notice that this other value is low, and this can be an indicator of Y. I’m going to request a follow-up test regarding that.” Yes, I would trust an AI to give me that explanation, because those are very strict parameters to work with, and the input comes from a trusted source (lab results and medical training data) and not “Bob’s shrimping and hoola hoop dancing blog”.
That’s LLM AI, but the type I’m talking about is the machine learning kind. I can envision a system that takes e.g. a sample’s test data and provides a summary, which is not far from what doctors do anyway. If you ever get a blood test’s results explained to you it’s “this value is high, which would be concerning except that this other value is not high, so you’re probably fine regarding X. However, I notice that this other value is low, and this can be an indicator of Y. I’m going to request a follow-up test regarding that.” Yes, I would trust an AI to give me that explanation, because those are very strict parameters to work with, and the input comes from a trusted source (lab results and medical training data) and not “Bob’s shrimping and hoola hoop dancing blog”.