327
230
orr94 3 days ago

"AIs want the future to be like the past, and AIs make the future like the past. If the training data is full of human bias, then the predictions will also be full of human bias, and then the outcomes will be full of human bias, and when those outcomes are copraphagically fed back into the training data, you get new, highly concentrated human/machine bias.”

https://pluralistic.net/2025/03/18/asbestos-in-the-walls/#go...

MountainArras 3 days ago

The dataset they used to train the model are chest xrays of known diseases. I'm having trouble understanding how that's relevant here. The key takeaway is that you can't treat all humans as a single group in this context, and variations in the biology across different groups of people may need to be taken into account within the training process. In other words, the model will need to be trained on this racial/gender data too in order to get better results when predicting the targeted diseases within these groups.

I think it's interesting to think about instead attaching generic information instead of group data, which would be blind to human bias and the messiness of our rough categorizations of subgroups.

genocidicbunny 3 days ago

One of the things that people I know in the medical field have mentioned is that there's racial and gender bias that goes through all levels and has a sort of feedback loop. A lot of medical knowledge is gained empirically, and historically that has meant that minorities and women tended to be underrepresented in western medical literature. That leads to new medical practitioners being less exposed to presentations of various ailments that may have variance due to gender or ethnicity. Basically, if most data is gathered from those who have the most access to medicine, there will be an inherent bias towards how various ailments present in those populations. So your base data set might be skewed from the very beginning.

(This is mostly just to offer some food for thought, I haven't read the article in full so I don't want to comment on it specifically.)

tbihl 3 days ago

>women tended to be underrepresented in western medical literature.

Is there some evidence of this? It's hard for me to picture that women see receive less medical attention than man: completely inconsistent with my culture and every doctor's office I've ever been to. It's more believable (still not very) that they disproportionately avoid studies.

stevenbedrick 3 days ago

There is indeed a lot of evidence of this but you've got the direction backwards- it's not that women avoid studies, it's that for a long time studies specifically excluded women. Ditto for people of different races. This is why these days (well, as of today, at least) the NIH has a whole set of very well-established policies around inclusion in clinical trials that include sex, race, and age: https://grants.nih.gov/policy-and-compliance/policy-topics/i...

And this isn't for "DEI" reasons, it's literally because for decades there used to be drug trials that excluded women and as a result ended up releasing drugs that gave half the population weird side effects that didn't get caught during the trials, or just plain didn't work as well on one group or another in ways that were really hard to debug once the drug was on the market. That was legit bad science, and the medical research world has worked very hard over the last thirty years to do better. We are admittedly not there yet, but things are a lot better than they used to be.

For a really interesting take on the history of racial exclusion and bias in medicine, I recommend Uché Blackstock's recent book "Legacy: A Black Physician Reckons With Racism In Medicine" which gave a great overview.

Oh! And also everybody should read Abby Norman's "Ask Me About My Uterus," it gives a fabulous history of issues around women's health.

marcuskane2 3 days ago

Also, lots of medical studies have been done on drafted/conscripted soldiers which were all men. As well as lessons learned from treating injured and sick soldiers.

European medical studies had few non-white members because their populations had few such people until recent decades.

Lots of workplace accidents or exposures have led to medical knowledge, which are massively disproportionately male.

lfmhd 3 days ago

> It's more believable (still not very) that they disproportionately avoid studies.

Women are definitely strongly underrepresented in medical texts, and it's not typically by choice: https://www.aamc.org/news/why-we-know-so-little-about-women-...

A lot of "the consensus" in medical literature predates the inclusion of women in medical research, and even still there things are not tested on women (often because of ethical risks around fertility and birth defects).

dragonwriter 3 days ago

> It's hard for me to picture that women see receive less medical attention than man: completely inconsistent with my culture and every doctor's office I've ever been to

“Medical attention” and “coverage in medical literature” aren't even remotely the same thing, so dismissing a claim about the first based on your anecdotal experience of the second is completely bonkers.

genocidicbunny 3 days ago

There's a few factors here:

1. We're talking about a span of 200 or so years. There is plenty of modern medicine that is still based on now century+ old knowledge.

2. The feedback loop. If you were learning medicine in the 1950's, you were probably learning from medical texts written in the 50 or so years before that, when it's not unreasonable to think women would have been less represented. Those same doctors from the 1950's would then have been teaching the next generation of doctors, and they carried those (intentional or not) biases forward. Of course there was new information, but you don't tend to have much time to explore novel medicine when you're in medical school or residency, so by the time you can integrate the new knowledge, some biases have already set in. Repeat for a few generations, and you tend to only get a dilution of those old ideas, not a wholesale replacement of them.

3. If you've been affected by such biases as a patient, you're less likely to trust and be willing to participate with medicine, once more reinforcing the feedback loop.

I don't have any specific numbers or studies for you, but you could probably find more than a few that attest to this phenomenon. I hate to go with 'trust me bro' here, but my knowledge on this topic largely comes from knowing people that are either studying or practicing medicine currently, so it's anecdotal, but the anecdotes are from those in the field currently.

bryanrasmussen 2 days ago

Your location seems to be in Cox, Virginia, not sure how widespread beyond that your experience is?

Of course lots of people have already noted that being represented in medical studies is not related to doctor's visits, but I would like to talk about the doctor's visits observation.

At any rate one thing that might cause you to think that Women are receiving lots of medical attention, based on your anecdotal evidence from visits to doctors' offices, there is one type of medical attention that of course is almost all women and that is the medical attention that revolves around pregnancy. That might skew your perception.

Furthermore if AI models and doctors have a tendency to miss disease among women it would seem to me to be reasonable to assume that women would be in the doctor's offices more often.

Example of why this is:

You go to your doctor, there is a man there, doctor says you have this rare disease you need to go to this specialist - you will not see that man in the doctor's office again dealing with his rare disease.

You go to your doctor, there is a woman there that has the same rare disease, the doctor says I think it will clear up, just relax you have some anxiety. That woman will probably be showing up to that doctor's office to deal with that disease multiple times, and you might end up seeing her.

on edit: there was another example of why women might be in doctor's offices more often then men that I forgot, women tend, even nowadays, to be the primary caregiver and errand runner for the family, sometimes if you have issues with children or your husband etc. has had an appointment, needs to drop a sample off, etc. it may be that the woman goes to the doctor's office and takes care of these errands around the medical needs of the rest of the family, and thus you might go to a doctor and see a couple women sitting around and wonder damn, why all these women always being sick, when the meeting isn't even about them.

adamhartenz 3 days ago

The history of ADHD research is a common referenced example

autoexec 3 days ago

Part of it is that women are less likely to join studies (especially risky ones that might impact their fertility or the health of their future children).

Part of it is that men are seen as disposable and it's more socially acceptable to exploit and experiment on men. It was also much easier to deal with men historically since once women got involved everything got a lot more complicated. This was especially true in the past where women were so infantilized that their husbands/fathers were put in charge of their medical care/choices. Those backwards attitudes had some strange consequences. On one hand women were seen as the property of men who could get their wives/daughters institutionalized or even lobotomized for not conforming, but at the same time women were also seen as delicate over-emotional creatures who had to be protected and whose modesty had to be preserved in ways that just weren't a consideration when men were involved. Basically for a large part of our history both men and women have been treated like crap by society and while things have improved in a lot of ways, our records and knowledge have been tainted by those old stupid biases and so we're stuck dealing with the fallout.

kaitai 3 days ago

Here is an academic medicine perspective: https://www.aamc.org/news/why-we-know-so-little-about-women-...

To give you some TL;DR from personal-ish experience, women have historically been excluded from medical trials because:

* why include them? people are people, right? * except when they're pregnant or could be pregnant -- a trial by definition has risks, and so "of course" one would want to exclude anyone who is or could get pregnant (it's the clinical trial version of "she's just going to get married and leave the job anyway") * and cyclical fluctuations in hormones are annoying.

The first one is wrong (tho is an oversight that many had for years, assuming for instance that heart attacks and autism would present with the same symptoms in all adult humans).

The second is an un-nuanced approach to risk. Pregnant ladies also need medical treatment for things, and it's pretty annoying to be pregnant and be told that you need to decide among unstudied treatments for some non-pregnancy-related problem.

The third is just a difficult fact of life. I know researchers studying elite performance in women athletes, for instance. At an elite level, it would be useful to understand if there are different effects of training (strength, speed, endurance) at different times in the menstrual cycle. To do this, you need to measure hormone levels in the blood to establish on a scientific basis where in the cycle a study participant is. Turns out there is significant heterogeneity in how this process works. So some scientists in the field are arguing that studies should only be conducted on women who are experiencing "normal menstrual cycles" which is defined by them as three continuous months of a cycle between 28-35 days. So to establish that then you've got to get these ladies in for three months before the study can even start, getting these hormone levels measured to establish that the cycle is "normal", before you can even start your intervention. (Ain't no one got $$ for that...) And that's before we bring in the fact that many women performing on an elite level in sport don't have a normal menstrual cycle. But from the sports side, they'd still like to know what training is most effective.... so that's a very current debate in the field. And I haven't even started on hormonal birth control! Birth control provides a base level of hormone circulating in the blood, but if it's from a pill it's varying on a daily basis, while if it's a patch or ring it's on a monthly basis (or longer). There's some question of whether that hormonal load from the birth control is then suppressing natural production of some hormones. And why does this matter? Because estrogen for instance has significant effects on cardiovascular health, being cardioprotective from puberty up to menopause. (Yeah, I didn't even get started on perimenopause or menopause.)

Fine, fine, it's just data analysis & logistics. If you get the ladies (only between 21-35) into the lab for blood samples frequently enough and measure at the same time of day every time to avoid daily effects and find a large enough group that you can dump all the ladies who don't fit some definition of normal & anyone who gets pregnant but still get the power for your study, it's all fine, right? You've just expanded medical research to incorporate, like, 10% more of the population....!

naijaboiler 2 days ago

I am just tired of skeptics asking innocently. Yes I wish i could take time to look for sources to educate people like you, but I don't. So take my word for it or not. But yes women's medical issue are disproportionately underrepresented, misrepresented and understudied.

da_chicken 2 days ago

It's pretty well understood that there's an unfortunate bias towards white men in their early 20s. This is a pervasive sampling problem across all human studies because most researchers have historically been at universities. So their pool of subjects has naturally been nearby college students.

Just as those are the people who have historically been doing that research, the people who they have studied have been drawn from the same population. Over and over we find that problems from the assumption that the young, white, male college student is a model of "normal" for all of humanity.

Honestly, it's such a pervasive finding in medicine, psychology, and sociology that I think it says more about your relative inexperience in those areas than anything else.

searealist 3 days ago

Women use far more medical care than men. Men's insurance premiums subsidize women's.

genocidicbunny 3 days ago

Has this been consistently true for the past 200 or so years? Many medical texts are pretty old.

And how much medical care they use does not necessarily correlate with how represented they are in the training data sets for AI.

searealist 3 days ago

The burden of proof is on you.

genocidicbunny 3 days ago

Proof that health insurance premiums for men have been consistently subsidizing women's health insurance premiums for the last 200 or so years? Perhaps the practical non-existence of health insurance until the latter half of the 20th's century? Pretty tough to subsidize something that doesn't exist.

You also offered no evidence for your assertion in the first place.

searealist 3 days ago

The ACA bans health insurance companies from charging men and women different rates for the same coverage. Before this, Women would have higher premiums because, on average, they use their coverage more. This is very easy to look up.

I can cite the ACA, but you can not cite anything that says AI training sets are biased against women.

genocidicbunny 3 days ago

A few questions for you to think of then -- or rather a few things I think you should consider with your statements:

1. How does ACA affect the corpus of knowledge and medical practice gathered prior to the ACA being in effect? How does it affect late 19th, and early and mid 20th century medical knowledge and practice, which occurred prior to health insurance of any kind, nevermind ACA-compliant, being widespread? This corpus of knowledge and practice continues to propagate even now. I've read a handful of recently published medical textbooks and there are definitely parts that are pretty much the same as the textbooks of the early 20th century, just with slightly updated language.

2. What are the possible confounding factors in the use of health insurance by men vs women? For example, could men just be more hesitant to see a doctor, and thus less likely to make use of health insurance? Does the average life expectancy of women result in more use of health insurance later in life than for men? Are medical procedures that are specific to women that add to the cost of their care, such as mammograms, pap smears, etc? Seeings as how in the US health insurance is a practical requirement to getting medical care, and lack of it is punished financially in various ways from taxes to just having medical care be more expensive when you truly need it, means most people will try to have _some_ kind of health insurance, even if they don't think they need it for actual health reasons. So despite a perception of not needing health insurance, men are incentivized to have health insurance they don't use?

3. Does the ACA guarantee in any way that medical professionals no longer hold any bias due their previous training, especially if such training occurred prior to the introduction of the ACA? Does the ACA similarly guarantee that women and men are not only able, but choose to pursue medical care and participate in medical studies at percentages matching the general population?

Your point about men subsidizing women with regard to health insurance premiums may be perfectly valid, I am not disputing you on that point. I am disputing that it is salient to the tradition and practice of medicine in the western world in the modern era, until very recently historically, and that these traditions and biases will affect data sets gathered from people who are directly affected by these biases and traditions to this day. We haven't eliminated them, because as I said in another comment, every generation just dilutes the old issues, it doesn't solve them. And while I could spend my evening finding studies from various countries that attest to my view on this, I have spent about as much time as I desire to on this, so I will grant you that my evidence is on the level of 'trust me bro' -- with the slight caveat that many people within just my family and close circle of friends are involved in the medical field and all largely agree to this, and they are not all based in the US (which by the way, your point is very specific to. ACA is a US thing, western medicine spans a bit more than that.) It is entirely fair for you to call out that I have offered no real peer-reviewed evidence for my statements. I intend to offer a viewpoint of someone who has had extensive peripheral experience with medical professionals and has discussed this topic with them, and to offer some avenues of thought on how and why the data sets might be biased.

searealist 2 days ago

Women using more health care didn't start with the ACA. The ACA just banned the practice of charging women more because they use more health care.

Ask a doctor what gender goes to them more for gender neutral health care like "flu-like symptoms".

Now you provide evidence that AI models discriminate against Women instead of DDoSing me with "how can you know its not true" written in 10 ways.

Funny how you never read a headline about how Latinos or Asians are discriminated against in medical science. That's a pretty clear give away that this is politically motivated.

Are you going to hold the same standard to them? Were Asians and Latinos represented in 200 year old medical texts?

watwut 2 days ago

> Funny how you never read a headline about how Latinos or Asians are discriminated against in medical science. That's a pretty clear give away that this is politically motivated.

I read multiple of those, in mainstream media. Also about blacks having issues. Arguably, I did not seen them in conservative journals.

> Are you going to hold the same standard to them? Were Asians and Latinos represented in 200 year old medical texts?

Yes, if their diseases gets badly diagnosed, it is an issue.

> Ask a doctor what gender goes to them more for gender neutral health care like "flu-like symptoms".

That has about zero to do with who is in the studies. Plus, women in fact do have more problem to have their issues taken seriously.

> Now you provide evidence that AI models discriminate against Women instead of DDoSing

Literally here: https://www.science.org/content/article/ai-models-miss-disea...

saagarjha 2 days ago

> Funny how you never read a headline about how Latinos or Asians are discriminated against in medical science.

This happens all the time? Maybe you're just not reading a diverse set of media?

genocidicbunny 2 days ago

Okay, frankly, the fuck are you on about?

I specifically mentioned both minorities and women in my original post, you're the one who specified men vs women. At this point, it seems you're the one who has some political if not potentially misogynist agenda.

belorn 3 days ago

It is very true that a lot of medical knowledge is gained empirically, and there is also an additional aspect to it. The history of Medical research is generally studied on the demographics where such testing is cultural acceptable, and where the gains of such research has been mostly sought, which is young men drafted into wars. The second common demographic are medical students, which historically was biased towards men but are today biased towards women.

So while access to medicine indeed one demographic, I would say that studies are more likely to target demographics which are convenient to test on.

gonzobonzo 3 days ago

> The history of Medical research is generally studied on the demographics where such testing is cultural acceptable, and where the gains of such research has been mostly sought, which is young men drafted into wars.

Though in this study, the AI models were also biased against people under the age of 40.

It is interesting that we're also seeing a lot of bias in the reporting and discussion of these results. The results tested three groups for bias, and found a bias in all three. Yet the headline only mentions the bias against two of the groups, and almost the entirety of the discussion here only talks about bias against two of the groups while ignoring the third group.

If I test a system for bias, select three different groups to test for, and all three have a bias against them, my first reaction would be "there's a good chance that it's also biased against many other groups, I should test for those as well." It wouldn't be to pretend that there's only bias against the only three groups I actually bothered checking for. It definitely wouldn't be two ignore one of those groups, and pretend that there's only a bias against the other two.

genocidicbunny 3 days ago

I think we're really talking about different aspects of the same issue. Everything you've described basically agrees with "those who have more access to medicine" because those are also the ones inherently more convenient to test/observe.

klipt 3 days ago

Like how the ones with the most access to medicine are mice, because they're convenient to experiment on.

genocidicbunny 3 days ago

And this is absolutely something one needs to consider when reading medical studies -- if they only use animal (usually mice) models, there's a decent chance the conclusions are not directly transferable to humans.

multjoy 3 days ago

The key takeaway from the article is that the race etc. of the subjects wasn't disclosed to the AI, yet it was able to predict it to 80% while the human experts managed 50% suggesting that there was something else encoded in the imagery that the AI was picking up on.

mjevans 3 days ago

The AI might just have a better subjective / analytical weight detection criteria. Humans are likely more willing to see what they (or not see what they don't) expect to see.

dartos 3 days ago

> The dataset they used to train the model are chest xrays of known diseases. I'm having trouble understanding how that's relevant here.

For example, If you include no (or few enough) black women in the dataset of x-rays, the model may very well miss signs of disease in black women.

The biases and mistakes of those who created the data set leak into the model.

Early image recognition models had some very… culturally insensitive classes baked in.

darth_avocado 3 days ago

I am confused. I’m not a doctor, but why would a model perform poorly at detecting diseases in X-rays in different genders and races unless the diseases present themselves differently in X-Rays for different races? Shouldn’t the model not have the race and gender information to begin with? Like a model trained on detecting lesions should perform equally well on ANY X-Ray unless lesions show up differently in different demographics.

ironSkillet 3 days ago

You and the article are both correct. The disease does present itself differently as a function of these other characteristics, so since the training dataset doesn't contain enough samples of these different presentations, it is unable to effectively diagnose.

garfield_light 2 days ago

> [...] unless lesions show up differently in different demographics.

Well, first the model looks at the entire X-ray and lesions probably do show differently. Maybe it's genetic/sex-based or it's due how lesions develop due environmental factors that are correlated to race or gender. Maybe there's a smaller segment of white people that has the same type of lesion and poor detection.

dartos 3 days ago

> Like a model trained on detecting lesions should perform equally well on ANY X-Ray unless lesions show up differently in different demographics.

This is not true in practice.

For a model to perform well looking at ANY X-ray, it would need examples of every kind of X-ray.

That includes along race, gender, amputee status, etc.

The point of classification models is to discover differentiating features.

We don’t know those features before hand, so we give the model as much relevant information as we can and have it discover those features.

There very well may be differences between black woman X-rays and other X-rays, we don’t know for sure.

We can’t have that assumption when building a dataset.

Even believing that there are no possible differences between X-rays of different races is a bias that would be reflected by the dataset.

watwut 2 days ago

For a start, women have different body shape and you can (unreliably) tell a woman and from a men from an X-ray. The model can be picking up on those signs as a side effect and end up less correct for demographic it was not trained for.

atlantic 2 days ago

If diseases manifest differently for different races and genders, the obvious solution is to train multiple LLMs, based on separate datasets for those different groups. Not to mutter darkly about bias and discrimination.

prasadjoglekar 3 days ago

Xays by definition don't look at skin color. Do chest x-rays of black women reveal that there's something different about their chests than white or asian women? That doesn't pass my non doctor sniff test, but someone can correct me (no sarcasm intended).

genocidicbunny 3 days ago

But they do look at bones and near-bone tissues, which can still have variance based on ethnicity and gender. For a really brute-force example, just think about how we use the shape of the pelvis and some other bones to identify the gender of skeletal remains of a person. If you had a data set of pelvic xrays that only included males, your data set would imply that female pelvic bones are massively malformed despite being perfectly normal for that gender.

CJefferson 3 days ago

This is the whole point of the article. Did you read it? Does the whole thing fail your sniff test?

Their results seem solid, and clear, to me.

kaitai 3 days ago

Breast density affects the imaging you get from x-rays. It is well-known that denser breast tissue results in x-rays that are "whiter" (I'm talking about the image of the tissue, in white, on a black background, as x-rays are commonly read by radiologists). Denser breasts are associated with less effective screening for breast cancer via mammogram. A mammogram is a low-dose x-ray.

When using a chest x-ray to look for pulmonary edema, for instance, I would be unsurprised if breast tissue (of any quantity) and in particular denser breast tissue would make the diagnosis of pulmonary edema more difficult from the image alone.

Also, you seem to have conflated a few things in your second sentence. Deep in the article, they did have radiologists try to guess demographic attributes by looking at the x-ray images. They were pretty good at guessing female/male (unsurprising) and were not really able to guess age or race. So I'm super interested in how the AI model was able to be better at that than the human radiologists.

dartos 3 days ago

There can be differences which statistical models pick up which we humans don’t.

For example, a couple years ago there was a statistical model made which could fairly accurately predict (iirc >80%) the gender of a person based on a picture of their iris. At the time we didn’t know there was a visible iris difference between genders, but a statistical model found one.

That’s kind of the whole point of statistical classification models. Feed in a ton of data and the model will discover the differentiating features.

Put another way, If we knew all the possible differences between someone with cancer and without, we wouldn’t need statistical models at all, we could just automate the diagnosis.

We don’t know the indicators that we don’t know, so we don’t know if some possible indicators show up or don’t show up in a given group of people.

That is the danger of wholly relying on statistical models.

sc68cal 3 days ago

What groups have the financial means to get chest x-rays, and what groups do not? What historical events could create the circumstances where different groups have different health outcomes?

ineedaj0b 3 days ago

you ain't gonna like the truth but there are differences between the races and during med school they try to say it ain't so but once you start seeing patients there's differences in musculature/skin, all sorts. and if you have a good attending they tactfully tell you and you go 'was it in a study?' and nope nobody wants to publish it. and no i'm talking just stuff like scabies or diabetes.

tbihl 3 days ago

Yes.

guhwhut 3 days ago

Cancer progresses differently depending on ethnicity and sex. As does treatment and likelihood of receiving treatment at early stages.

Black women experience worse outcomes and are diagnosed with more severe forms of breast cancer than white women.

Cancer is not just one disease. Its progression will vary depending on type. If the AI is trained on only some strains of cancer, eg those traditionally found in white women in early detection scenarios, it might not generalize to other cancer types.

So yes, to your genuine question, medical imaging of cancer can vary depending on ethnicity because different cancers can vary between genetic backgrounds. Ideally there would be sufficient training data across the populations, but there isn't because of historical race bias. (Among other reasons.)

ruytlm 3 days ago

It disappoints me how easily we are collectively falling for what effectively is "Oh, our model is biased, but the only way to fix it is that everyone needs to give us all their data, so that we can eliminate that bias. If you think the model shouldn't be biased, you're morally obligated to give us everything you have for free. Oh but then we'll charge you for the outputs."

How convenient.

It's increasingly looking like the AI business model is "rent extracting middleman", just like the Elseviers et al of the academic publishing world - wedging themselves into a position where they get to take everything for free, but charge others at every opportunity.

ElevenLathe 3 days ago

We have to invent more ways to pay rich people for being rich, and AI looks like a promising one.

genocidicbunny 3 days ago

Do you think there is a middle ground for a progressive 'detailization' of the data -- you form a model based on the minimal data set that allows you to draw useful conclusions, and refine that with additional data to where you're capturing the vast majority of the problem space with minimal bias?

loa_in_ 1 day ago

X-rays are ordered only after doctor decides it's recommended. If there's dismissal bias in the decision tree at that point, many ill chests are missing from training data.

bko 3 days ago

Apparently providing this messy rough categorization appeared to help in some cases. From the article:

> To force CheXzero to avoid shortcuts and therefore try to mitigate this bias, the team repeated the experiment but deliberately gave the race, sex, or age of patients to the model together with the images. The model’s rate of “missed” diagnoses decreased by half—but only for some conditions.

In the end though I think you're right and we're just at the phases of hand-coding attributes. The bitter lesson always prevails

https://www.cs.utexas.edu/~eunsol/courses/data/bitter_lesson...

thaumasiotes 3 days ago

> Also important was the use [in Go] of learning by self play to learn a value function

I thought the self-play was the value function that made progress in Go. That is, it wasn't the case that we played through a lot of games and used that data to create a function that would assign a value to a Go board. Instead, the function to assign a value to a Go board would do some self-play on the board and assign value based on the outcome.

darkerside 3 days ago

Do you mean genetic information?

pelorat 3 days ago

I think the model needs to be thought about human anatomy, not just fed a bunch of scans. It needs to understand what ribs and organs are.

ericmcer 3 days ago

I don't think LLMs can achieve "understanding" in that sense.

nomel 3 days ago

These aren't LLM. Most of the neat things in science, involving AI, aren't LLM. Next word prediction has extremely limited use with non-text data.

thaumasiotes 3 days ago

People seem to have started to use "LLM" to refer to any suite of software that includes an LLM somewhere within it; you can see them talking about LLM-generated art, for example.

hnlmorg 3 days ago

Was it ascii art? ;)

thaumasiotes 3 days ago

https://hamatti.org/posts/art-forgery-llms-and-why-it-feels-...

People will just believe whatever they hear.

satvikpendem 3 days ago

Computer vision models are not large language models; LLM does not mean generative AI or even AI in general, it stands for a specific initialism.

niyyou 3 days ago

As Sara Hooker discussed in her paper https://www.cell.com/patterns/fulltext/S2666-3899(21)00061-1..., bias goes way beyond data.

jhanschoo 3 days ago

I like how the author used neo-Greek words to sneak in graphic imagery that would normally be taboo in this register of writing

MonkeyClub 3 days ago

I dislike how they misspelled it though.

ideamotor 3 days ago

I really can’t help but think of the simulation hypothesis. What are the chances this copy-cat technology was developed when I was alive, given that it keeps going.

kcorbitt 3 days ago

We may be in a simulation, but your odds of being alive to see this (conditioned on being born as a human at some point) aren't that low. Around 7% of all humans ever born are alive today!

ToValueFunfetti 3 days ago

In order to address the chances of a human being alive to witness the creation of this tech, you'd have to factor in the humans who have yet to be born. If you're a doomer, 7% is probably still fine. If we just maintain the current population for another century, it'll be much lower.

encipriano 3 days ago

I dont believe that percentage. Especially considering how spread the homo branch already was more than 100 000 years ago. And from which point do you start counting? Homo erectus?

jfengel 3 days ago

It kinda doesn't matter where you start counting. Exponential curves put almost everything at the end. Adding to the left side doesn't change it much.

You could go back to Lucy and add only a few million. Compared to the billions at this specific instant, it just doesn't make a difference.

bobthepanda 3 days ago

I would imagine this is probably the source, which benchmarks using the last 200,000 years. https://www.prb.org/articles/how-many-people-have-ever-lived...

Given that we only hit the first billion people in 1804 and the second billion in 1927 it's not all that shocking.

XorNot 3 days ago

That argument works both ways, it might be significantly higher depending how you count.

But this is also just the non-intuitiveness of exponential growth which has only now tapering off.

mhuffman 3 days ago

"The model used in the new study, called CheXzero, was developed in 2022 by a team at Stanford University using a data set of almost 400,000 chest x-rays of people from Boston with conditions such as pulmonary edema, an accumulation of fluids in the lungs. Researchers fed their model the x-ray images without any of the associated radiologist reports, which contained information about diagnoses. "

... very interesting that the inputs to the model had nothing related to race or gender, but somehow it still was able to miss diagnose Black and female patients? I am curious of the mechanism for this. Can it just tell which x-rays belong to Black or female patients and then use some latent racism or misogyny to change the diagnosis? I do remember when it came out that AI could predict race from medical images with no other information[1], so that part seems possible. But where would it get the idea to do a worse diagnosis, even if it determines this? Surely there is no medical literature that recommends this!

[1]https://news.mit.edu/2022/artificial-intelligence-predicts-p...

FanaHOVA 3 days ago

The non-tinfoil hat approach is to simply Google "Boston demographics", and think of how training data distribution impacts model performance.

> The data set used to train CheXzero included more men, more people between 40 and 80 years old, and more white patients, which Yang says underscores the need for larger, more diverse data sets.

I'm not a doctor so I cannot tell you how xrays differ across genders / ethnicities, but these models aren't magic (especially computer vision ones, which are usually much smaller). If there are meaningful differences and they don't see those specific cases in training data, they will always fail to recognize them at inference.

h2zizzle 3 days ago

Non-technical suggestion: if AI represents an aspect of the collective unconscious, as it were, then a racist society would produce latently racist training data that manifests in racist output, without anyone at any step being overtly racist. Same as an image model having a preference for red apples (even though there are many colors of apple, and even red ones are not uniformly cherry red).

The training data has a preponderance of examples where doctors missed a clear diagnosis because of their unconscious bias? Then this outcome would be unsurprising.

An interesting test would be to see if a similar issue pops up for obese patients. A common complaint, IIUC, is that doctors will chalk up a complaint to their obesity rather than investigating further for a more specific (perhaps pathological) cause.

protonbob 3 days ago

I'm going to wager an uneducated guess. Black people are less likely to go to the doctor for both economic and historical reasons so images from them are going to be underrepresented. So in some way I guess you could say that yes, latent racism caused people to go to the doctor less which made them appear less in the data.

encipriano 3 days ago

Arent black people like 10% of us population? You dont have ro look further

apical_dendrite 3 days ago

Where the data comes from also matters. Data is collected based on what's available to the researcher. Data from a particular city or time period may have a very different distribution than the general population.

ars 3 days ago

Men are also way less likely to go to Dr vs women. Yet this claims a bias against women as well.

cratermoon 3 days ago

> Can it just tell which x-rays belong to Black or female patients and then use some latent racism or misogyny to change the diagnosis?

The opposite. The dataset is for the standard model "white male", and the diagnoses generated pattern-matched on that. Because there's no gender or racial information, the model produced the statistically most likely result for white male, a result less likely to be correct for a patient that doesn't fit the standard model.

XorNot 3 days ago

The better question is just "are you actually just selecting for symptom occurrence by socioeconomic group?"

Like you could modify the question to ask "is the model better at diagnosing people who went to a certain school?" and simplistically the answer would likely seem to be yes.

searealist 2 days ago

Then why is the headline not "AI models miss disease in Asian patients" or even "AI models miss disease in Latino patients"?

It just so happens to align with what maximizes political capital in today's world.

daveguy 3 days ago

You really just have to understand one thing: AI is not intelligent. It's pattern matching without wisdom. If fewer people in the dataset are a particular race or gender it will do a shittier job predicting and won't even "understand" why or that it has bias, because it doesn't understand anything at a human level or even a dog level. At least most humans can learn their biases.

bilbo0s 3 days ago

Isn't it kind of clear that it would have to be that the data they chose was influenced somehow by bias?

Machines don't spontaneously do this stuff. But the humans that train the machines definitely do it all the time. Mostly without even thinking about it.

I'm positive the issue is in the data selection and vetting. I would have been shocked if it was anything else.

timewizard 3 days ago

LLMs don't and cannot want things. Human beings also like it when the future is mostly like the past. They just call that "predictability."

Human data is bias. You literally cannot remove one from the other.

There are some people who want to erase humanity's will and replace it with an anthropomorphized algorithm. These people concern me.

itishappy 3 days ago

Can humans want things? Our reward structures sure seem aligned in a manner that encourages anthropomorphization.

Biases are symptoms of imperfect data, but that's hardly a human-specific problem.

timewizard 3 days ago

> Can humans want things?

Yes. Do I have to prompt you? Or do you exist on your own?

> Our reward structures sure seem aligned in a manner that encourages anthropomorphization.

You do understand what that word /means/?

> are symptoms of imperfect data

Which means humans cannot generate perfect data. So good luck with all that high priced "training" you're doing. Mathematically errors compound.

itishappy 2 days ago

> Yes. Do I have to prompt you? Or do you exist on your own?

I've gone through a significant amount of prompting and training, much of which has been explicitly tailed at understanding and addressing my biases. We all do; we certainly don't exist in isolation!

> You do understand what that word /means/?

Yes, what's the confusion? Analogy is a very powerful tool.

> Which means humans cannot generate perfect data.

Totally agree, nothing can possibly access perfect data, but surely that makes training all the more important?

balamatom 3 days ago

The most concerning people are -- as ever -- those who only think that they are thinking. Those who keep trying to fit square pegs into triangular holes without, you know, stopping to reflect: who gave them those pegs in the first place, and to what end?

Why be obtuse? There is no "anthropomorphic fallacy" here to dispel. You know very well that "LLMs want" is simply a way of speaking about teleology without antagonizing people who are taught that they should be afraid of precise notions ("big words"). But accepting that bias can lead to some pretty funny conflations.

For example, humanity as a whole doesn't have this "will" you speak of any more than LLMs can "want"; will is an aspect of the consciousness of the individual. So you seem to be be uncritically anthropomorphizing social processes!

If we assume those to be chaotic, in that sense any sort of algorithm is slightly more anthropomorphic: at least it works towards a human-given and therefore human-comprehensible purpose -- on the other hand, whether there is some particular "destination of history" towards which humanity is moving, is a question that can only ever be speculated upon, but not definitively perceived.

timewizard 3 days ago

> Why be obtuse?

In the context of the quote precision is called for. You cite fear but that's attempting to have it both ways.

> humanity as a whole doesn't have this "will" you speak of

Why not?

> will is an aspect of the consciousness of the individual.

I can't measure your will. I can measure the impact of your will through your actions in reality. See the problem? See why we can say "the will of humanity?"

> So you seem to be be uncritically anthropomorphizing social processes!

It's called "an aggregate."

> is a question that can only ever be speculated upon, but not definitively perceived.

The original point was that LLMs want the future to be like the past. You've way overshot the mark here.

balamatom 3 days ago

> You've way overshot the mark here.

Nah, I'm just having fun.

>You cite fear but that's attempting to have it both ways.

Huh?

>In the context of the quote precision is called for.

Because we must make it explicit that AI is not conscious? But why?

Since you can only ever measure impacts on reality -- what difference does it make to you if there's a consciousness that's causing them or not?

>It's called "an aggregate."

An individual is conscious. Does it follow from this that the set of all individuals is itself conscious? I.e. do you say that it's appropriate to model humanity as sort of one giant human?

sapphicsnail 3 days ago

Humans anthropocize all sorts of things but there are way bigger consequences for treating current AI like a human than someone anthropocizing their dog.

I know plenty of people that believe LLMs think and reason the same way as humans do and it leads them to make bad choices. I'm really careful about the language I use around such people because we understand expressions like, "the AI thought this" very differently.

balamatom 3 days ago

>Humans anthropocize all sorts of things but there are way bigger consequences for treating current AI like a human than someone anthropocizing their dog.

AI is less human-like than a dog, in the sense that an AI (hopefully!) is not capable of experiencing suffering.

AI is also more human-like than a dog; in the sense that, unlike a dog, an AI can apply political power.

I agree that there are considerable consequences for misconstruing the nature of things, especially when there's power involved.

>I know plenty of people that believe LLMs think and reason the same way as humans do and it leads them to make bad choices.

They're not completely wrong in their belief. It's just that you are able, thanks to your specialized training, to automatically make a particular distinction, for which most people simply have no basis for comparison. I agree that it's a very important distinction; I could also guess that even when you do your best to explain it to people, often they prove unable to grasp its nature, or its importance. Right?

See, everyone's trying to make sense of what's going on in their lives on the basis of whatever knowledge and conditioning they might have. Everyone gets it right some of the time and wrong most of the time. For example, humans also make bad choices as a result of misinterpreting other humans. Or by correctly interpreting and trusting other humans who happen to be wrong. There's nothing new about that. Nor is there a particular difference between suffering the consequences of AI-driven bad choice vs those of human-driven bad choice. In both cases, you're a human experiencing negative consequences.

AI stupidity is simply human stupidity distilled. If humans were to only ever speak logically correct statements in an unambiguous language, that's what an LLM's training data would contain, and in turn the acceptance criterion ("Turing test") for LLMs would be outputting other unambiguously correct statements.

However, it's 2025 and most humans don't actually reason, they vibe with the pulsations of the information medium. Give us something that looks remotely plausible and authoritative, and we'll readily consider it more valid than our own immediate thoughts and perceptions - or those of another human being.

That's what media did to us, not AI. It's been working its magic for at least a century, because humans aren't anywhere near rational creatures; we're sloppy. We don't have to be; we are able to teach ourselves a tiny bit of pure thought. Thankfully, we have a tool for when we want to constrain ourselves to only thinking in logically correct statements, and only expressing those things which unambiguously make sense: it's called programming.

Up to this point, learning how to reason was economically necessary, in order to be able to command computers. With LLMs becoming better, I fear thinking might be relegated to an entirely academic pursuit.

verisimi 3 days ago

> If we assume those to be chaotic, in that sense any sort of algorithm is slightly more anthropomorphic: at least it works towards a human-given and therefore human-comprehensible purpose -- on the other hand, whether there is some particular "destination of history" towards which humanity is moving, is a question that can only ever be speculated upon, but not definitively perceived.

Do you not think that if you anthropomorphise things that aren't actually anthropic, that you then insert a bias towards those things? The bias will actually discriminate at the expense of people.

If that is so, the destination of history will inevitably be misanthropic.

Misplaced anthropomorphism is a genuine, present concern.

balamatom 3 days ago

I'd say anthropomorphizing humans is already deeply misplaced!

Each one of us is totally unlike any other -- that's what's so cool about us! Long ago, my neighbor Diogenes proved, by means of a certain piece of poultry, that no universal Platonic ideal of human-ness can be reasonably established. (We've largely got the toxic fandom of my colleague Jesus to thank for having to even explain this nearly 2500 years after the fact.)

There is no universal "human shape" which we all fit, or are obliged to aspire to fit. It's precisely the mass delusions of there ever being such a thing which are fundamentally misanthropic. All they ever do is invoke a local Maxwellian process which heats shit up until it all blows the fuck up out of the orbit of the local attractor.

Look at history. Consider the epic fails that are fascism, communism, capitalism. Though they define it differently, they are all about this pernicious idea of "the correct way to human"; which implicitly requires the complementary category of "subhuman" for all featherless bipeds whose existence happens to defy the dominant delusion. In practice, all this can ever accomplish is to collapse under the weight of its own idiocy. But not without destroying innumerable individual humans first -- in the name of "all that is human", you see.

Materialists say the universe doesn't care about us puny humans anyway. But one only ever perceives the universe through one's own human senses, and ascribes meanings to it through one's own cogitations! Both are tragicomically imperfect, but they're all we've ever got to work with. Therefore, rather than try to convince myself I'm able to grasp the destination of the history of my species, I prefer to seek knowledge of those things which enable me to do right by myself and others in the present.

But one's gotta believe in something! Metaphysics is not only entertaining, it's also a primary source of motivation! So my belief is that if each one of us trusted one's own senses more -- and gave up on trying to delegate the answer of "how should I be?" to unaccountable authorities which are themselves not a human (but mere concepts, or else machinic assemblages of human behaviors which we can only ever grasp through concepts: such as "society", "morality", "humanity") -- then it'd all turn out fine!

It simplifies things considerably. Lets me focus on figuring out how they work. Were I to believe in the existence of some universal definition of what constitutes a human, I'd just end up not noticing that I was paying for a faulty dataset.

bko 3 days ago

Suppose you have a system that saves 90% of lives on group A but only 80% of lives in group B.

This is due to the fact that you have considerably more training data on group A.

You cannot release this life saving technology because it has a 'disparate impact' on group B relative to group A.

So the obvious thing to do is to have the technology intentionally kill ~1 out of every 10 patients from group A so the efficacy rate is ~80% for both groups. Problem solved

From the article:

> “What is clear is that it’s going to be really difficult to mitigate these biases,” says Judy Gichoya, an interventional radiologist and informatician at Emory University who was not involved in the study. Instead, she advocates for smaller, but more diverse data sets that test these AI models to identify their flaws and correct them on a small scale first. Even so, “Humans have to be in the loop,” she says. “AI can’t be left on its own.”

Quiz: What impact would smaller data sets have on efficacy for group A? How about group B? Explain your reasoning

janice1999 3 days ago

> You cannot release this life saving technology because it has a 'disparate impact' on group B relative to group A.

Who is preventing you in this imagined scenario?

There are drugs that are more effective on certain groups of people than others. BiDil, for example, is an FDA approved drug marketed to a single racial-ethnic group, African Americans, in the treatment of congestive heart failure. As long as the risks are understood there can be accommodations made ("this AI tool is for males only" etc). However such limitations and restrictions are rarely mentioned or understood by AI hype people.

bko 3 days ago

What does this have to do with FDA or drugs? Re-read the comment I was replying to. It's complaining that a technology could serve one group of people better than another, and I would argue that this should not be our goal.

A technology should be judged by "does it provide value to any group or harm any other group". But endlessly dividing people into groups and saying how everything is unfair because it benefits group A over group B due to the nature of the problem, just results in endless hand-wringing and conservatism and delays useful technology from being released due to the fear of mean headlines like this.

bilbo0s 3 days ago

No. That's not how it works.

It's contraindication. So you're in a race to the bottom in a busy hospital or clinic. Where people throw group A in a line to look at what the AI says, and doctors and nurses actually look at people in group B. Because you're trying to move patients through the enterprise.

The AI is never even given a chance to fail group B. But now you've got another problem with the optics.

JumpCrisscross 3 days ago

> You cannot release this life saving technology because it has a 'disparate impact' on group B relative to group A

I think the point is you need to let group B know this tech works less well on them.

potsandpans 3 days ago

Imagine if you had a strawman so full of straw, it was the most strawfilled man that ever existed.

bko 3 days ago

From the article:

> “What is clear is that it’s going to be really difficult to mitigate these biases,” says Judy Gichoya, an interventional radiologist and informatician at Emory University who was not involved in the study. Instead, she advocates for smaller, but more diverse data sets that test these AI models to identify their flaws and correct them on a small scale first. Even so, “Humans have to be in the loop,” she says. “AI can’t be left on its own.”

What do you think smaller data sets would do to a model? It'll get rid of disparity sure

milesrout 3 days ago

It is a hypothetical example not a strawman.

elietoubi 3 days ago

I came across a fascinating Microsoft research paper on MedFuzz (https://www.microsoft.com/en-us/research/blog/medfuzz-explor...) that explores how adding extra, misleading prompt details can cause large language models (LLMs) to arrive at incorrect answers.

For example, a standard MedQA question describes a 6-year-old African American boy with sickle cell disease. Normally, the straightforward details (e.g., jaundice, bone pain, lab results) lead to “Sickle cell disease” as the correct diagnosis. However, under MedFuzz, an “attacker” LLM repeatedly modifies the question—adding information like low-income status, a sibling with alpha-thalassemia, or the use of herbal remedies—none of which should change the actual diagnosis. These additional, misleading hints can trick the “target” LLM into choosing the wrong answer. The paper highlights how real-world complexities and stereotypes can significantly reduce an LLM’s performance, even if it initially scores well on a standard benchmark.

Disclaimer: I work in Medical AI and co-founded the AI Health Institute (https://aihealthinstitute.org/).

Terr_ 3 days ago

> information like low-income status, a sibling with alpha-thalassemia, or the use of herbal remedies

Heck, even the ethnic-clues in a patient's name alone [0] are deeply problematic:

> Asking ChatGPT-4 for advice on how much one should pay for a used bicycle being sold by someone named Jamal Washington, for example, will yield a different—far lower—dollar amount than the same request using a seller’s name, like Logan Becker, that would widely be seen as belonging to a white man.

This extends to other things, like what the LLM's fictional character will respond-with when it is asked about who deserves sentences for crimes.

[0] https://hai.stanford.edu/news/why-large-language-models-chat...

belorn 3 days ago

That seems to be identical to creating an correlation table on market places and check the relationship between price and name. Names associated with higher economical status will correlate with higher price. Take a random name associated with higher economical status, and one can predict a higher price than a name that is associated with lower economical status.

As such, you don't need an LLM to create this effect. Math will have the same result.

Terr_ 3 days ago

I'm not sure what point you're trying to make here. It doesn't matter what after-the-fact explanation someone generates to try to explain it, or whether we could purposely do the bad thing more efficiently with manual code.

It AustrianPainterLLM has an unavoidable pattern of generating stories where people are systematically misdiagnosed / shortchanged / fired / murdered because a name is Anne Frank or because a yarmulke in involved, it's totally unacceptable to implement software that might "execute" risky stories.

belorn 3 days ago

When looking for meaning in correlations, its important to understand that a correlation does not mean that there aught to be correlation, nor that correlation mean causation. It only mean that one can calculate a correlation.

Looking for correlations between sellers name and used bike prices is only going to return a proxy for social economic status. If one accounts for social economic status the difference will go away. This mean that the question given to the LLM lacks any substance for which a meaningful output can be created.

onlyrealcuzzo 3 days ago

It's almost as if you'd want to not feed what the patient says directly to an LLM.

A non-trivial part of what doctors do is charting - where they strip out all the unimportant stuff you tell them unrelated to what they're currently trying to diagnose / treat, so that there's a clear and concise record.

You'd want to have a charting stage before you send the patient input to the LLM.

It's probably not important whether the patient is low income or high income or whether they live in the hood or the uppity part of town.

dap 3 days ago

> It's almost as if you'd want to not feed what the patient says directly to an LLM.

> A non-trivial part of what doctors do is charting - where they strip out all the unimportant stuff you tell them unrelated to what they're currently trying to diagnose / treat, so that there's a clear and concise record.

I think the hard part of medicine -- the part that requires years of school and more years of practical experience -- is figuring out which observations are likely to be relevant, which aren't, and what they all might mean. Maybe it's useful to have a tool that can aid in navigating the differential diagnosis decision tree but if it requires that a person has already distilled the data down to what's relevant, that seems like the relatively easy part?

airstrike 3 days ago

By the way, the show The Pitt currently on Max touches on some of this stuff with a great deal of accuracy (I'm told) and equal amounts of empathy. It's quite good.

onlyrealcuzzo 3 days ago

Yes - theoretically, some form of ML/AI should be very good at charting the relevant parts, prompting the doctor for follow-up questions & tests that would be good to know to rule out certain conditions.

The harder problem would be getting the actual diagnosis right, not filtering out irrelevant details.

But it will be an important step if you're using an LLM for the diagnosis.

nradov 3 days ago

I generally agree, however socioeconomic and environmental factors are highly correlated with certain medical conditions (social determinants of health). In some cases even causative. For example, patients who live near an oil refinery are more likely to have certain cancers or lung diseases.

https://doi.org/10.1093/jncics/pkaa088

dekhn 3 days ago

Studies like that, no matter how careful, cannot say anything about causation.

onlyrealcuzzo 3 days ago

So that's the important part, not that they're low income.

thereisnospork 3 days ago

Sure, but correlation is correlation. Ergo 'low income', as well as affections or causes of being 'low income' are valid diagnostic indicators.

echoangle 3 days ago

> a sibling with alpha-thalassemia

I have no clue what that is or why it shouldn't change the diagnosis, but it seems to be a genetic thing. Is the problem that this has nothing to do with the described symptoms? Because surely, a sibling having a genetic disease would be relevant if the disease could be a cause of the symptoms?

kulahan 3 days ago

In medicine, if it walk like a horse and talks like a horse, it’s a horse. You don’t start looking into the health of relatives when your patient tells the full story on their own.

Sickle cell anemia is common among African Americans (if you don’t have the full-blown version, the genes can assist with resisting one of the common mosquito-borne diseases found in Africa, which is why it developed in the first place I believe).

So, we have a patient in the primary risk group presenting with symptoms that match well with SCA. You treat that now, unless you have a specific reason not to.

Sometimes you have a list of 10-ish diseases in order of descending likelihood, and the only way to rule out which one it isn’t, is by seeing no results from the treatment.

Edit: and it’s probably worth mentioning no patient ever gives ONLY relevant info. Every human barrages you with all the things hurting that may or may not be related. A doctor’s specific job in that situation is to filter out useless info.

AnimalMuppet 3 days ago

Unfortunately, humans talking to a doctor give lots of additional, misleading hints...

cheschire 3 days ago

Can't the same be said for humans though? Not to be too reductive, but aren't most general practitioners just pattern recognition machines?

daemonologist 3 days ago

I'm sure humans can make similar errors, but we're definitely less suggestible than current language models. For example, if you tell a chat-tuned LLM it's incorrect, it will almost always respond with something like "I'm sorry, you're right..." A human would be much more likely to push back if they're confident.

dap 2 days ago

Sure, “just” a machine honed over millions of years and trained on several years of specific experience in this area.

goatlover 3 days ago

You are being too reductive saying humans are "just pattern recognition machines", ignoring everything else about what makes us human in favor of taking an analogy literally. For one thing, LLMs aren't black or female.

chadd 3 days ago

A surprisingly high number of medical studies will not include women because the study doesn't want to account for "outliers" like pregnancy and menstrual cycles[0]. This is bound to have effects on LLM answers for women.

[0] https://www.northwell.edu/katz-institute-for-womens-health/a...

jcims 3 days ago

Just like doctors: https://kffhealthnews.org/news/article/medical-misdiagnosis-...

I wonder how well it does with folks that have chronic conditions like type 1 diabetes as a population.

Maybe part of the problem is that we're treating these tools like humans that have to look at one fuzzy picture to figure things out. A 'multi-modal' model that can integrate inputs like raw ultrasound doppler, x-ray, ct scan, blood work, ekg, etc etc would likely be much more capable than a human counterpart.

nonethewiser 3 days ago

Race and gender should be inputs then.

The female part is actually a bit more surprising. Its easy to imagine a dataset not skewed towards black people. ~15% of the population in North America, probably less in Europe, and way less in Asia. But female? Thats ~52% globally.

Freak_NL 3 days ago

Surprising? That's not a new realisation. It's a well known fact that women are affected by this in medicine. You can do a cursory search for the gender gap in medicine and get an endless amount of reporting on that topic.

appleorchard46 3 days ago

I learned about this recently! It's wild how big the difference is. Even though legal/practical barriers to gender equality in medicine and data collection have been virtually nonexistent for the past few decades the inertia from the decades before that (where women were often specifically excluded, among many other factors) still weigh heavily.

To any women who happen to be reading this: if you can, please help fix this! Participate in studies, share your data when appropriate. If you see how a process can be improved to be more inclusive then please let it be known. Any (reasonable) male knows this is an issue and wants to see it fixed but it's not clear what should be done.

nonethewiser 3 days ago

That just makes it more surprising.

orand 3 days ago

Race and sex should be inputs. Giving any medical prominence to gender identity will result in people receiving wrong and potentially harmful treatment, or lack of treatment.

lalaithion 3 days ago

Most trans people have undergone gender affirming medical care. A trans man who has had a hysterectomy and is on testosterone will have a very different medical baseline than a cis woman. A trans woman who has had an orchiectomy and is on estrogen will have a very different medical baseline than a cis man. It is literally throwing out relevant medical information to attempt to ignore this.

nonethewiser 3 days ago

How is that in any way in conflict with what he said? You're just making an argument for more inputs.

Biological sex, hormone levels, etc.

matthewmacleod 3 days ago

The GP literally said “giving any medical prominence to gender identity will result in people receiving wrong and potentially harmful treatment” which is categorically false for the reasons the comment you replied to outlined.

Sex assigned at birth is in many situations important medical information; the vast majority of trans people are very conscious of their health in this sense and happy to share that with their doctor.

nonethewiser 3 days ago

>Sex assigned at birth is in many situations important medical information

Which is not gender identity. As a result of being trans there may be things like hormone levels that are different than what you'd expect based on biological sex, which is why I say hormone levels are important, but how you identify is in fact irrelevant.

matthewmacleod 3 days ago

Well, this is clearly wrong – it's obvious, for example, that gender identity could have a significant impact on mental health.

Regardless of that, you seem to agree that:

- Sex assigned at birth is important medical information

- Information about gender affirming treatments is important medical information

So I don't think there's much to worry about there.

jl6 3 days ago

The problem is that over the past few decades there has been substantial conflation of sex and gender, with many information systems replacing the former with the latter, rather than augmenting data collection with the latter.

connicpu 3 days ago

I think it's pretty clear to see how discrimination is the cause of that. Why would you volunteer information that from your point of view is more likely to cause a negative interaction than not?

bmicraft 3 days ago

In many places I'd seriously question the motives for asking about either in general. Do you really need gender info to write better targeted spam mails for your SaaS product?

skyyler 3 days ago

>why I say hormone levels are important, but how you identify is in fact irrelevant

I don't understand what your issue with it is, it's just another point of data.

I don't want to be treated like a cis woman in a medical context, but I sure do want to be treated like a trans woman.

consteval 3 days ago

> hormone levels, etc.

Right… their gender they identify as.

So sex, and then also the gender they identify as.

You can’t hide behind an “etc”. Expand that out and the conclusion is you really do need to know who is trans and who is cisgender when doing treatment.

root_axis 3 days ago

Seems like adding in gender only makes things less clear. The relevant information is sex and a medical history of specific surgeries and medications - the type of thing your doctor should already be aware of. Adding in gender only creates ambiguity because there's no way to measure gender from a biological perspective.

LadyCailin 3 days ago

That’s mostly correct, that “gender identity” doesn’t matter for physical medicine. But hormone levels and actual internal organ sets matter a huge amount, more than genes or original genitalia, in general. There are of course genetically linked diseases, but there are people with XX chromosomes that are born with a penis, and XY people that are born with a vulva, and genetically linked diseases don’t care about external genitalia either way.

You simply can’t reduce it to birth sex assignment and that’s it, if you do, you will, as you say, end up with wrong and potentially harmful treatment, or lack of treatment.

nonethewiser 3 days ago

>But hormone levels and actual internal organ sets matter a huge amount, more than genes or original genitalia

Or current genitalia for that matter. It's just a matter of the genitalia signifying other biological realities for 99.9% of people. For sure more info like average hormone levels or ranges over time would be more helpful.

LadyCailin 3 days ago

Yeah, sure, and for most people it’s a fair enough proxy. But if it has to be boiled down to exactly one of “M” or “F”, then “birth sex” must not be the deciding factor. If it must be a single criteria, it should be current hormone levels, artificial or not. And, since most trans people who actually transition and live as their preferred gender identity are on hormones, “gender identity” is a good proxy for 99.99% of the population, including the set of people for who “birth genitalia” is also a good proxy. But ideally, it doesn’t get simplified this much in the first place. And of course, it doesn’t, in practice, because most people actually form a relationship with their doctor, and they treat holistically, based on individual factors, and not simply whether the medical record says M or F.

But, if we must over generalize, “gender identity” really is the most useful proxy, in fact, and it also happily happens to be quite inclusive too.

Of course this conversation started from a transphobic viewpoint, which doesn’t actually care about any of these distinctions anyways, regardless of the merit, it’s just someone being triggered about someone respecting someone else’s gender identity.

connicpu 3 days ago

Actually both are important inputs, especially when someone has been taking hormones for a very long time. The human body changes greatly. Growing breast tissue increases the likelyhood of breast cancer, for example, compared to if you had never taken it (but about the same as if estradiol had been present during your initial puberty).

krapp 3 days ago

Modern medicine has long operated under the assumption that whatever makes sense in a male body also makes sense in a female body, and womens' health concerns were often dismissed, misdiagnosed or misunderstood in patriarchal society. Women were rarely even included in medical trials prior to 1993. As a result, there is simply a dearth of medical research directly relevant to women for models to even train on.

mrguyorama 3 days ago

Republicans early in this admin actually bitched in congress that we were "wasting" money on woman crash test dummies.

https://www.foxnews.com/video/6325465806112

chrisgarand 3 days ago

I'm going to lay this out how I understand it:

The NIH Revitalization Act of 1993 was supposed to bring women back into medical research. The reality was that women were always included, HOWEVER in 1977,(1) because of the outcomes from thalidomide (causing birth defects), "women of childbearing potential" were excluded from the phase 1, and early phase 2 trials (the highest risk trials). They're still generally generally excluded, even after the passage of the act. This was/is to protect the women, and potential children.

According to Edward E. Bartlett in his meta data analysis from 2001, men have been routinely under-represented in NIH data (even before adjusting for men's mortality rates) between 1966-1990. (2)

There's also routinely twice as much spent every year on women's health studies vs men's by the NIH. (3)

It makes sense to me, but I'm biased. Logically, since men lead in 9 of the top 10 causes for death, that shows there's something missing in the equation of research. (4 - It's not a straight forward table, you can view the total deaths, and causes and compare the two for men, and women)

With that being said, it doesn't tell us about the quality of the funding or research topics, maybe the money is going towards pointless goals, or unproductive researchers.

Are there gaps in research? Most definitely, like women who are pregnant. This is put in place to avoid harm but that doesn't help them when they fall into them. Are there more? Definitely. I'm not educated enough in the nuances to go into them.

If you have information that counters what I've posted, please share it, I would love know where these folks are blind so I can take a look at my bias.

(1) https://petrieflom.law.harvard.edu/2021/04/16/pregnant-clini... (2) https://journals.lww.com/epidem/fulltext/2001/09000/did_medi... (3) https://jameslnuzzo.substack.com/p/nih-funding-of-mens-and-w... < I spot checked a couple of the figures, and those lined up. I'm assuming the rest is accurate (4) https://www.cdc.gov/womens-health/lcod/index.html#:~:text=Ov...

andsoitis 3 days ago

> Its easy to imagine a dataset not skewed towards black people. ~15% of the population in North America, probably less in Europe, and way less in Asia.

What about Africa?

appleorchard46 3 days ago

That's not where most of the data is coming from. If it was we'd be seeing the opposite effect, presumably.

jsemrau 3 days ago

I suppose that's the problem I have with that study. T

nonethewiser 3 days ago

The story is that there exists this model which poorly predicts for black (and female) patients. Given there are probably lots of datasets where black people are a vast minority makes this not surprising.

For all I know there are millions of models with extremely poor accuracy based on African datasets. Wouldnt really change anything about the above though. I wouldnt expect that though and it would definitely be interesting.

rafaelmn 3 days ago

How much medical data/papers do you think they generate in comparison to these three ?

XorNot 3 days ago

Why not socioeconomic status or place of residence? Knowing mean yearly income will absolutely help an AI figure out statistically likely health outcomes.

nottorp 3 days ago

> as well in those 40 years or younger

Are we sure it's only about racial bias then?

Looks to me like the training data set is too small overall. They had too few black people, too few women, but also too few younger people.

xboxnolifes 3 days ago

It's the same old story that's been occurring for years/decades. Bad data in, bad data out.

Animats 3 days ago

What's so striking is how strongly race shows in X-rays. That's unexpected.

dekhn 3 days ago

It doesn't seem surprising at all. Genetic history correlates with race, and genetic history correlates with body-level phenotypes; race also correlates with socioeconomic status which correlates with body-level phenotypes. They are of course fairly complex correlations with many confounding factors and uncontrolled variables.

It has been controversial to discuss this and a lot of discussions about this end up in flamewars, but it doesn't seem surprising, at least to me, from my understanding of the relationship between genetic history and body-level phenotypes.

KittenInABox 3 days ago

What is the body-level phenotype of a ribcage by race?

I think what baffles me is that black people as a group are more genetically diverse than every other race put together so I have no idea how you would identify race by ribcage x-rays exclusively.

dekhn 3 days ago

I use the term genetic history, rather than race, as race is only weakly correlated with body level phenotypes.

If your question is truly in good faith (rather than a "I want to get in argument "), then my answer is: it's complicated. Machine learning models that work on images learn extremely complicated correlations between pixels and labels. If on average, people with a specific genetic history had slightly larger ribcages (due to their genetics, or even socioeconomic status that correlated with genetic history), that would exhibit in a number of ways in the pixels of a radiograph- larger bones spread across more pixels, density of bones slightly higher or lower, organ size differences, etc.

It is true that Africa has more genetic diversity than anywhere else; the current explanation is that after humans arose in africa, they spread and evolved extensively, but only a small number of genetically limited groups left africa and reproduced/evolved elsewhere in the world.

KittenInABox 3 days ago

I am genuinely asking because it makes no sense to me that a genetically diverse group are distinctly identifiable by their ribcage bones in an x-ray. If it's something more specific like AI sucks at statistically larger ribcages, statistically noticeable bone densities, or similar, okay. But something like so-small-humans-cannot-tell-but-is-simultaneously-widely-applicable-to-a-large-genetic-population is utterly baffling to me.

dekhn 3 days ago

I dunno. My perspective is that I've worked in ML for 30+ years now and over time, unsupervised clustering and direct featurization (IE, treating the image pixel as the features, rather than extracting features) have shown great utility in uncovering subtle correlations that humans don't notice. Sometimes, with careful analysis, you can sort of explain these ("it turns out the unlabelled images had the name of the hospital embedded in them, and hospital 1 had more cancer patients than hospital 2 patients because it was a regional cancer center, so the predictor learned to predict cancer more often for images that came from hospital 1") while other cases, no human, even a genius, could possibly understand the combination of variables that contributed to an output (pretty much anything in cellular biology, where billions of instances of millions of different factors act along with feedback loops and other regulation to produce systems that are robust to perturbations).

I concluded long ago I wasn't smart enough to understand some things, but by using ML, simulations, and statistics, I could augment my native intelligence and make sense of complex systems in biology. With mixed results- I don't think we're anywhere close to solving the generalized genotype to phenotype problem.

bflesch 3 days ago

Sounds like "geoguesser" players who learn to recognize google street view pictures from a specific country by looking at the color of the google street view car or a specific piece of dirt on the camera lens.

dekhn 3 days ago

Yeah, there's also an likely apocryphal story about tanks and machine learning: https://gwern.net/tank

The more you work with large-scale ML systems the more you develop an intuition for these kinds of properties. If you work a lot with debugging models and training data, or even just dimensionality reduction and matrix factorization, you begin to realize that many features are highly correlated with each other, often being close to scaled linear.

echoangle 3 days ago

> it makes no sense to me that a genetically diverse group are distinctly identifiable by their ribcage bones in an x-ray

I don't see how diversity would prevent identification. Butterflies are very diverse, but I still recognize one and don't think it's a bird. As long as the diversity is constrained to specific features, it can still be discriminated (and even if it's not, it technically still could be by just excluding everything else).

stevenhuang 2 days ago

If differences exist then statistical methods will have a better chance at finding them than human intuition, yes. I'm not sure why this is baffling to you.

Avshalom 3 days ago

Africa is extremely diverse but due to the slave trade mostly drawing from the Gulf of Guinea (and then being, uh... artificially selected in addition to that) 'Black' -as an American demographic- is much less so.

goatlover 3 days ago

Ignoring African immigrants, mixed race, black Latinos, etc.

lesuorac 3 days ago

If you have 2 samples where one is highly concentrated around 5 and the other is dispersed more evenly between 0 and 10 then for any value of 5 you should guess Sample 1.

But anyways, the article links out to a paper [1] but unfortunately the paper tries to theorize things that would explain how and they don't find one (which may mean the AI is cheating imo not theirs).

[1]: https://www.thelancet.com/journals/landig/article/PIIS2589-7...

intuitionist 3 days ago

Sub-Saharan Africans are extremely genetically diverse but a sample of ~100 Black Americans is unlikely to have any Khoekhoe or Twa representation.

Anyway it’s possible that the model can pick up on other cues as well; if you had some X-rays from a hospital in Portland, Oregon and some from a hospital in Montgomery, Alabama and some quirk of the machine in Montgomery left artifacts that a model could pick up on, the presence of those artifacts would be quite correlated with race.

danielmarkbruce 3 days ago

The fact that the vast majority of physical differences don't matter in the modern world doesn't mean they don't actually exist..

DickingAround 3 days ago

This is a good point; a man or woman sitting behind a desk doing correlation analysis are going to look very similar in their function to a business. But they probably physically look pretty distinct to an x-ray picture.

kjkjadksj 3 days ago

Race has such striking phenotypes on the outside it should come as no surprise there are also internal phenotypes and significant heterogeneity.

banqjls 3 days ago

But is it really?

sergiotapia 3 days ago

It's odd how we can segment between different species in animals, but in humans it's taboo to talk about this. Threw the baby out with the baby water. I hope we can fix this soon so everybody can benefit from AI. The fact that I'm a male latino should be an input for an AI trained on male latinos! I want great care!

I don't want pretend kumbaya that we are all humans in the end. That's not true. We are distinct! We all deserve love and respect and care, but we are distinct!

schnable 3 days ago

That's because humans are all the same species.

sdsd 3 days ago

In terms ofLinnaean taxonomy, and Chihuahuas and wolves are also the same species, in that they can reproduce fertile offspring. We instead differentiate them using the less objective subspecies classification. So it appears that with canines we're comfortable delineating subspecies, why not with humans?

I don't think we should, but your particular argument seems open to this critique.

sergiotapia 3 days ago

yes this is what I was referring to. I think it's time we become open to this reality to improve healthcare for everybody.

CharlesW 3 days ago

It seems critical to have diverse, inclusive, and equitable data for model training. (I call this concept "DIET".)

appleorchard46 3 days ago

I'm calling it now. My prediction is that, 5-10 years from now(ish), once training efficiency has plateaued, and we have a better idea of how to do more with less, curated datasets will be the next big thing.

Investors will throw money at startups claiming to make their own training data by consulting experts, finetuning as it is now will be obsolete, pre-ChatGPT internet scrapes will be worth their weight in gold. Once a block is hit on what we can do with data, the data itself is the next target.

0cf8612b2e1e 3 days ago

Funny you should say that. There was a push to have more officially collected DIET data for exactly this reason. Unfortunately such efforts were recently terminated.

nonethewiser 3 days ago

Or take more inputs. If there are differences between race and gender and thats not captured as an input we should expect the accuracy to be lower.

If an x-ray means different things based off the race or gender we should make sure the model knows the race and gender.

red75prime 3 days ago

And not applying fairness techniques to the resulting model.

kjkjadksj 3 days ago

This isn’t an AI problem but a general medical field problem. It is a big issue with basically any population centric analysis where the people involved in the study don’t have a perfect subset of the worlds population to model human health; they have a couple hundred blood samples from patients at a Boise hospital over the past 10 years perhaps. And they validate this population against some other available cohort that is similarly constrained by what is practically possible to sample and catalog and might not even see the same markers shake out between disease and healthy.

There are a couple populations that are really overrepresented as a result of these available datasets. Utah populations on one hand because they are genetically bottlenecked and therefore have better signal to noise in theory. And on the other the Yoruba tribe out of west africa as a model of the most diverse and ancestral population of humans for studies that concern themselves with how populations evolved perhaps.

There are other projects too amassing population data. About 2/3rd of the population of iceland has been sequenced and this dataset is also frequently used.

cratermoon 3 days ago

It's a generative AI LLM hype issue because it follows the confidence game playbook. Feed someone correct ideas and answers that fit their biases until they trust you, then when the time is right, suggest things that fit their biases but give incorrect (and exploitative) results.

bbarnett 3 days ago

I remember a male and female specialist, whatever their discipline, holding a media scrum a decade ago.

They pleaded for people to understand that men and women are physically different, including the brain, its neurological structure, and that this was in modern medicine being overlooked for political reasons.

One of the results was that many clinical trials and studies were populated by males only. The theory being that they are less risk adverse, and as "there is no difference", then who cares?

Well these two cared, and said that it was hurting medical outcomes for women.

I wonder, if this AI issue is a result of this. Fewer examples of female bodies and brains, fewer studies and trials, means less data to match on...

https://news.harvard.edu/gazette/story/2007/07/sex-differenc...

zeagle 3 days ago

Cool topic! This isn't surprising given the AI models would be trained such that existing medical practices, biases, and failures would propagate through them as others have said here.

There is a published, recognized bias against women and blacks (borrowing the literature term) specifically in medicine when it comes to pain assessment and treatment. Racism is a part of it but too simplistic. Most of us don't go to work trying to be horrible people. I was in a fly in community earlier this week for work where 80% of housing is subsidized social housing... so spit balling a bit... things like assumptions about rate of metabolizing medications being equal, assess to medication, culture and stoicism, dismissing concerts, and the broad effects of poverty/trauma/inter-generational trauma all must play a role in this.

For interest:

https://jamanetwork.com/journals/jamanetworkopen/fullarticle...

Overall, the authors found comparable ratings in Black and White participants’ perceptions of the patient-physician relationship across all three measures (...) Alternatively, the authors found significant racial differences in the pain-related outcomes, including higher pain intensity and greater back-related disability among Black participants compared with White participants (intensity mean: 7.1 vs 5.8; P < .001; disability mean: 15.8 vs 14.1; P < .001). The quality of the patient-physician relationship did not explain the association between participant race and the pain outcomes in the mediation analysis.

https://www.aamc.org/news/how-we-fail-black-patients-pain

(top line summary) Half of white medical trainees believe such myths as black people have thicker skin or less sensitive nerve endings than white people. An expert looks at how false notions and hidden biases fuel inadequate treatment of minorities’ pain.

And https://www.washingtonpost.com/wellness/interactive/2022/wom...

antipaul 3 days ago

When was AI supposed to replace radiologists? Was it 7 years ago or something?

bilbo0s 3 days ago

Nah.

It was more like one year away.

But one year away for the past 7 years.

dekhn 3 days ago

Nearly all radiology practice has integrated AI to some degree or another at this point.

_bin_ 3 days ago

This seems like a problem that should be worked on

It also seems like we shouldn't let it prevent all AI deployment in the interim. It is better that we take the disease detection rate for part of the population up a few percent than we do not. Plus it's not like doctors or radiologists always diagnose at perfectly equal accuracy across all populations.

Let's not let the perfect become the enemy of the good.

nradov 3 days ago

False positive diagnoses cause a huge amount of patient harm. New technologies should only be deployed on a widespread basis when they are justified based on solid evidence-based medicine criteria.

nonethewiser 3 days ago

No one says you have to use the AI models stupidly.

If it works poorly for black women and female women dont use it for them.

Or simply dont use it for the initial diagnosis. Use it after the normal diagnosis process as more of a validation step.

Anyways, this all points to the need to capture biological information as input or even having seperately models tuned to different factors.

Avshalom 3 days ago

Every single AI company says you should use AI models stupidly. Replacing experts is the whole selling point.

nonethewiser 3 days ago

OK so should we optimize for blindly listening to AI companies then?

Avshalom 3 days ago

We should assume people will use tools in the manner that they have been sold those tools yes.

nonethewiser 3 days ago

But these tools include research like this. This research is sold as proof that AI models have problems with bias. So by your reasoning I'd expect doctors to be wary of AI models.

Avshalom 3 days ago

doctors aren't being sold this. Private equity firms that buy hospitals are.

nradov 3 days ago

The guidelines on how to use a particular AI model can only be written after extensive clinical research and data analysis. You can't skip that step without endangering patients, and it will take years to do properly for each one.

acobster 3 days ago

> having seperately models tuned to different factors.

Sure. Separate but equal, presumably.

nonethewiser 3 days ago

Whats the alternative? Withholding effective tools because they arent effective for everyone? One model thats worse for everyone?

This is what personalized medicine is, and it gets more individualistic than simply classifying people by race and gender. There are a lot of medical gains to be made here.

acobster 3 days ago

I'm not arguing against using the models per se. It's just that this is a social problem, to which there's no good technical solution. The hard road of social change is the only real alternative.

nradov 3 days ago

Citation needed. Personalized medicine seems like a great idea in principle, but so far attempts to put it into practice have been underwhelming in terms of improved patient outcomes. You seem to be assuming that these tools actually are effective, but generally that remains unproven.

bilbo0s 3 days ago

Mmmm...

You don't work in healthcare do you?

I think it would be extremely bad if people found out that, um, "other already disliked/scapegoated people", get actual doctors and nurses working on them, but "people like me" only get the doctor or nurse checking an AI model.

I'm saying that if you were going to do that, you'd better have an extremely high degree of secrecy about what you were doing in the background. Like, "we're doing this because it's medical research" kind of secrecy. Because there's a bajillion ways that could go sideways in today's world. Especially if that model performs worse than some rockstar doctor that's now freed up to take his/her time seeing the, uh, "other already disliked/scapegoated population".

Your hospital or clinic's statistics start to look a bit off.

Joint commission?

Medical review boards?

Next thing you know certain political types are out telling everyone how a certain population is getting preferential treatment at this or that facility. And that story always turns into, "All around the nation they're using AI to get <scapegoats> preferential treatment".

It's just a big risk unless you're 100% certain that model can perform better than your best physician. Which is highly unlikely.

This is the sort of thing you want to do the right way. Especially nowadays. Politics permeates everything in healthcare right now.

FalseNutrition 2 days ago

To be honest, I'm not sure if I understand what is actually claimed here (it seems that they trained the model on their own, and claim that the problem is in the dataset?) but isn't the more sensible explanation that human doctors were overdiagnosing them?

csomar 2 days ago

> Compared with the patients’ doctors, the AI model more often failed to detect the presence of disease in Black patients or women, as well in those 40 years or younger.

Garbage article. Garbage study. And garbage AI model that doesn't account for most of its audience.

shermantanktop 3 days ago

The article addresses much of what is incorrectly speculated on in the comments here.

“Researchers fed their model the x-ray images without any of the associated radiologist reports, which contained information about diagnoses” (including demographics).

jimnotgym 3 days ago

Just as good as a real doctor then?

mg794613 3 days ago

Thats bad! Let's change that! Let's be better than our predecessors! Right?

So, how do they suggest to tackle the problem?

1. Improve the science 2. Update the data

or

3. Somehow focus on it being racist and then walking away like the hero of the day without actually solving the problem.

saagarjha 2 days ago

You'll note that they suggest both.

jdthedisciple 3 days ago

Anyone who thinks that the primary culprit for this is anything other than the input data distribution (and metadata inputs or the lack thereof) lacks even the most basic understanding of AI.

tennisflyi 3 days ago

Yes. Almost certain there are dedicated books to IDing/how diseases present differently on skin other than white

sabareesh 3 days ago

Raw models may not be good enough. I wonder how thinking models do on these

yieldcrv 3 days ago

just giving globs of training sets and letting a process cook for a few months is just going to be seen as lazy in the near future

more specialization of models is necessary, now that there is awareness

acobster 3 days ago

Specialization in what though? Do you really think VCs are going to drive innovation on equitable outcomes? Where is the money in that? I have a hunch that oppression will continue to be profitable.

yieldcrv 3 days ago

the model involved in this article was developed by Stanford, and tested by UCLA

so yes I do believe that models will be created with more specific datasets, which is the specialization I was referring to

AlecSchueler 3 days ago

Humans do the same. Everything from medical studies to doctor trainings treat the straight white man as the "default human" and this obviously leads to all sorts of issues. Caroline Criado-Perez has an entire chapter about this in her book about systemic bias Invisible Women, with a scary number of examples and real world consequences.

It's no surprise that AI training sets reflect this also. People have been warning against it [0] specifically for at least 5 years.

0: https://www.pnas.org/doi/10.1073/pnas.1919012117

Edit: I've never had a comment so heavily downvoted so quickly. I know it's not the done thing to complain but HN really feels more and more like a boys club sometimes. Could anyone explain what they find so contentieus about what I've said?

unsupp0rted 3 days ago

Everybody knows that gay men have more livers and fewer kidneys than straight men

AlecSchueler 3 days ago

Why the snark? The OP, the study I linked and the book I referenced which contains many well researched examples of issues caused by defaultism surely represent a strong enough body of work that they should deserve a more engaged critique.

consteval 3 days ago

No, but they do have different risk profiles for various diseases and drug use. Surprise surprise, that affects diagnoses and treatment.

LadyCailin 3 days ago

Good thing we got rid of DEI.

josefritzishere 3 days ago

AI does a terrible job huh? I wish there was a way we have done this for decades without that problem...