this post was submitted on 23 Nov 2024
453 points (99.3% liked)
Technology
59651 readers
2690 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
-Edit:
Adding Edit to the beginning to stop the replies from people who read the scenario for context and can't fight their compulsion to reply by nitpicking my completely made up list of "unbiased" metrics. To these peeps I say, "Fucking no. Bad dog. No!" I don't fucking care about your commentary to a quickly made up scenario. Whatever qualms you have, just fuckin change the imaginary scenario so it fits the purpose of what the purpose of the story is serving.
-Preface of actual comment:
Completely made up scenario to give context to my question. This is not me defending anything referenced to the article.
-Actual scenario with read, write, edit permissions to all users:
What if the court order the release of the AI code and training methods for this tenant analysis AI bot and found the metrics used were simply credit score, salary, employer and former rental references. No supplied data for race, name, background check or anything else that would tip the boy toward or away from any bias results. So this pure as it could be bot still produces the same results as seen in the article. Again, imaginary scenario that is likely no foundation of truth.
-My questions for the provided context:
Are there studies that compare methods of training LLMs with results showing differences in results ranging from less or no racist bias and more racist bias?
Are there ways of training LLMs to perform without bias or is the problem with the LLM's code and no matter how you train them there will always be a bias presence?
In the exact imaginary scenario, would the pure, unbiased angel version of rhe AI bot but producing equally racist results as biased trained AI bots see different court rulings that the AI that shows it's flawed design caused the biased results?
-I'm using bias over racist to reach broader area beyond race related issues. My driving purposes is:
To better understand how courts are handling AI related cases and if they give a fuck about the framework and design of the AI or if none of that matters and the courts are just looking at the results;
Wondering if there are ways to make or already made LLMs that aren't biased and what about their design makes them biased, is it the doing of the makers of the LLM or is it the training and implication of the LLM by the enduser/training party that is to blame?
The article is fake news. I suggest looking elsewhere for proper information.
As for your questions: LLMs were certainly not involved here. I can't guess what techniques were used.
Racial discrimination is often hard to nail down. Race is implicit in any number of facts. Place of birth, current address, school, ... You could infer race from such data. If you do not look at race at all but the end result still discriminates, then it's probably still racial discrimination. I say probably because you are free to do what you like and discriminate based on any number of factors, as long as it isn't race, sex, and the like. You certainly may discriminate based on education or wealth. Things being as they are, that will discriminate against minorities. They have systematically lower credit ratings, for example.
In the case of generative AI, bias is often not clearly defined. For example, you type "US President" into an image generator. All US presidents so far were male, and all but one white. But half of all people who are eligible for the presidency are female and (I think) a little less than half non-white. So what's the non-biased output?