this post was submitted on 18 Jul 2024
481 points (96.7% liked)

Technology

57895 readers
4511 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 7 points 1 month ago

Despite this article, I'm still not convinced that the algorithms aren't better. The policy states that people need to use their best judgement and can override the algorithms. The article argues that the algorithms are being over relied on. The article mentions in passing, however, that the statistics were worse before the algorithm was introduced.

The point of the matter is, best judgement can be shitty. Your average cop has no idea what questions to ask without a list and how important they are per research. Some suggestions are too continue using the tool but use things like psychologists to administer it. The only way you could reasonably have a psych on call for every police station is to make it a remote interview, which frankly doesn't seem better to me.

In the end, the unstated problem is resources and how best to utilize them to prevent the violence. I'm sure Spain's policy could be improved but shoring it up with an algorithm is a good practice.

[–] [email protected] 11 points 1 month ago

The computer response should be treated as just an indication and in all cases a human needs to decide to override that

Otherwise we’ll all become useless pieces of a simulation

I went to the bank to ask a loan and then it got rejected because the computer said I didn’t met the parameters by just 40 euro. Ah ok, I told the clerk, just lower the amount that I’m asking or spread it over a longer period. No, because after the quote is done and I signed the authorization for the algorithm to perform credit score, it can’t do it again in 3 months. What?? Call a supervisor and let them override it, 40 euro is so minimal that it’s not that big issue. No, impossible. So that means each single employee in the bank is just an interface to the computer and can be fired at will?

[–] [email protected] 4 points 1 month ago (1 children)

Can we stop having AI do.. anything?

[–] [email protected] 5 points 1 month ago* (last edited 1 month ago)

Algorithms aren't AI. They're standardization measures in cases like this. Hell you don't even need computers for many of them. We use tons in healthcare to classify risk, decide on treatment options, and even decide on how much medication to give. They're particularly present in psychiatric care.

[–] [email protected] 9 points 1 month ago (1 children)
[–] [email protected] 3 points 1 month ago

Cracked me up thanks!

[–] [email protected] 17 points 1 month ago (2 children)

About 20 new cases of gender violence arrive every day, each requiring investigation. Providing police protection for every victim would be impossible given staff sizes and budgets.

I think machine-learning is not the key part, the quote above is. All these 20 people a day come to the police for protection, a very small minority of them might be just paranoid, but I'm sure that most of them had some bad shit done to them by their partner already and (in an ideal world) would all deserve some protection. The algorithm's "success" in defined in the article as reducing probability of repeat attacks, especially the ones eventually leading to death.

The police are trying to focus on the ones who are deemed to be the most at risk. A well-trained algorithm can help reduce the risk vs the judgement of the possibly overworked or inexperienced human handling the complaint? I'll take that. But people are going to die anyway. Just, hopefully, a bit less of them and I don't think it's fair to say that it's the machine's fault when they do.

[–] [email protected] 7 points 1 month ago

Is it really too much to want enough resources to respond appropriately to all cases?

[–] [email protected] 8 points 1 month ago

Sounds like a triage situation. That really sucks for the women affected.

[–] [email protected] 7 points 1 month ago

More data for the algorithm then.

[–] [email protected] 23 points 1 month ago* (last edited 1 month ago)

Having worked in making software for almost 3 decades, including in Finance both before and after the 2008 Crash, this blind reliance on algorithms for law enforcement and victim protection scares the hell out of me.

An algorithm is just an encoding of whatever the people who made it think will happen: it's like using those actual people directly, only worse because by need an algorithm has a fixed set of input parameters and can't just ask more questions when something "smells fishy" as a person would.

Also making judgements by "entering something in a form" has a tendency to close people's thinking - instead of pondering on it and using their intuition to, for example, notice from the way people are talking that they're understating the gravity of the situation, people filling form tend to mindlessly do it like a box-ticking exercise - and that's not even going into the whole "As long as I just fill the form my ass is covered" effect when the responsability is delegated to the algorithm that leads people to play it safe and not dispute the results even when their instincts say otherwise.

For anybody who has experience with modelling, using computer algorithms within human processes and with how users actually treat such things (the "computer says" effect) this shit really is scary at many levels.

[–] [email protected] 22 points 1 month ago

Computers are only at fault when its convenient to blame them.

[–] [email protected] 19 points 1 month ago* (last edited 1 month ago) (1 children)

In the late 1970s (I was a kid) the computer is always right was a common sarcastic parody of all the people who actually believed it.

We'd discover in the 1980s it was possible to have missing data, insufficient data or erroneous data.

[–] [email protected] 4 points 1 month ago (1 children)

It's a sentiment at least as old as the first things that we now call computers.

On two occasions I have been asked, "Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?" … I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.

—Charles Babbage

[–] [email protected] 0 points 1 month ago

As if humans can magically make correct decisions with incorrect information lmao. So true.

[–] [email protected] 26 points 1 month ago (1 children)

The algorithm:

isSafe = random();

if isSafe >.5 println ("everything is fine\n");

[–] [email protected] 5 points 1 month ago (1 children)

per the article, it's rather better than that.

[–] [email protected] 18 points 1 month ago (1 children)

I admit to having grossly oversimplified things. Sorry.

[–] [email protected] 8 points 1 month ago

I thought that would have been obvious to everyone already.

[–] [email protected] 29 points 1 month ago (1 children)

I remember years ago when they said the value of our lives would be determined by a panel of people.

Now its by a machine.

[–] [email protected] 11 points 1 month ago (1 children)

Oh, it's far worse than that... the value of our lives have been determined by the (so-called) "free market" for a very long time now.

The machine is simply going to streamline the process.

[–] [email protected] 1 points 1 month ago

This comment is half a century late.

NASDAQ opened in 1971.

load more comments
view more: next ›