Sunday, October 13, 2024
Contact    |    RSS icon Twitter icon Facebook icon  
Unexplained Mysteries
You are viewing: Home > News > Science & Technology > News story
Welcome Guest ( Login or Register )  
All ▾
Search Submit

Science & Technology

AI judges - the future of justice or a disaster waiting to happen ?

July 12, 2023 · Comment icon 7 comments

Would you be happy with a verdict decided upon by an AI ? Image Credit: Pixabay / sergeitokmakov
With some judges using ChatGPT to help make decisions, should artificial intelligence really be used in court cases ?
Science fiction's visions of the future include many versions of artificial intelligence (AI), but relatively few examples where software replaces human judges. For once, the real world seems to be changing in ways that are not predicted in stories.

In February, a Colombian judge asked ChatGPT for guidance on how to decide an insurance case. Around the same time, a Pakistani judge used ChatGPT to confirm his decisions in two separate cases. There are also reports of judges in India and Bolivia seeking advice from ChatGPT.

These are unofficial experiments, but some systematic efforts at reform do involve AI. In China, judges are advised and assisted by AI, and this development is likely to continue. In a recent speech, the master of the rolls, Sir Geoffrey Vos - the second most senior judge in England and Wales - suggested that, as the legal system in that jurisdiction is digitised, AI might be used to decide some "less intensely personal disputes", such as commercial cases.

AI isn't really that smart

This might initially seem to be a good idea. The law is supposed to be applied impartially and objectively, "without fear or favour". Some say, what better way to achieve this than to use a computer program? AI doesn't need a lunch break, can't be bribed, and doesn't want a pay rise. AI justice can be applied more quickly and efficiently. Will we, therefore, see "robot judges" in courtrooms in the future?

There are four principal reasons why this might not be a good idea. The first is that, in practice, AI generally acts as an expert system or as a machine learning system. Expert systems involve encoding rules into a model of decisions and their consequences - called a decision tree - in software. These had their heyday in law in the 1980s. However, they ultimately proved unable to deliver good results on a large scale.

Machine learning is a form of AI that improves at what it does over time. It is often quite powerful, but no more so than a very educated guess. One strength is that it can find correlations and patterns in data that we don't have the capacity to calculate. However, one of its weaknesses is that it fails in ways that are different to the way people do, reaching conclusions that are obviously incorrect.

In a notable example, an AI was tricked into recognising a turtle as a gun. Facial recognition often has issues correctly identifying women, children and those with dark skin. So it's possible that AI could also erroneously place someone at a crime scene who wasn't there. It would be difficult to be confident in a legal system that produced outcomes that were clearly incorrect but also very difficult to review, as the reasoning behind machine learning is not transparent. It has outstripped our ability to understand its inner workings - a phenomenon known as the "black box problem".

When AI is used in legal processes, and it fails, the consequences can be severe. Large language models, the technology underlying AI chatbots such as ChatGPT, are known to write text that is completely untrue. This is known as an AI hallucination, even though it implies that the software is thinking rather than statistically determining what the next word in its output should be.

This year, it emerged that a New York lawyer had used ChatGPT to write submissions to a court, only to discover that it cited cases that do not exist. This indicates that these types of tools are not capable of replacing lawyers yet, and in fact, may never be.
Historical biases

Second, machine learning systems rely on historical data. In crime and law, these will often contain bias and prejudice. Marginalised communities will often feature more in records of arrests and convictions, so an AI system might draw the unwarranted conclusion that people from particular backgrounds are more likely to be guilty.

A prominent example of this is the Compas system, an AI algorithm used by US judges to make decisions on granting bail and sentencing. An investigation claimed that it generated "false positives" for people of colour and "false negatives" for white people. In other words, it suggested that people of colour would re-offend when they did not in fact do so, and suggested that white people would not re-offend when they did. However, the developer of the system challenges these claims.

Third, it is not clear that legal rules can be reliably converted into software rules. Individuals will interpret the same rule in different ways. When 52 programmers were assigned the task of automating the enforcement of speed limits, the programs that they wrote issued very different numbers of tickets for the same sample data.

Individual judges may have different interpretations of the law, but they do so in public and are subject to being overturned on appeal. This should reduce the amount of variation in judgments over time - at least in theory. But if a programmer is too strict or too lenient in their implementation of a rule, that may be very difficult to discover and correct.

Automated government systems fail at a scale and speed that's very difficult to recover from. The Dutch government used an automated system (SyRI) to detect benefits fraud, which wrongly accused many families, destroying lives in the process.

The Australian "Online Compliance Intervention" scheme is used to automatically assess debts from recipients of social welfare payments. It's commonly known as "Robodebt". The scheme overstepped its bounds, negatively affecting hundreds of thousands of people and was the subject of a Royal Commission in Australia. (Royal Commissions are investigations into matters of public importance in Australia.)

Finally, judging is not all that judges do. They have many other roles in the legal system, such as managing a courtroom, a caseload, and a team of staff, and those would be even more difficult to replace with software programs.

Rónán Kennedy, Associate Professor, University of Galway

This article is republished from The Conversation under a Creative Commons license.

Read the original article. The Conversation

Source: The Conversation | Comments (7)




Other news and articles
Recent comments on this story
Comment icon #1 Posted by Desertrat56 1 year ago
ChatGP is only as good as the data it has access to, there is not reasoning involved and those judges are making a mistake.   We have enough trouble with judges as it is, we don't need fake AI doing their job for them.
Comment icon #2 Posted by Alchopwn 1 year ago
I agree for the most part.  At some time in the future, it might be possible for far more advanced A.I. to fill the role of Judges, but we'd be placing a lot of trust in an algorithm and the ability to avoid letting it be hacked.  It is said that "He who judges, rules.  Now, clearly an A.I wouldn't rule, but what about the person who programs the A.I.?
Comment icon #3 Posted by Desertrat56 1 year ago
I have worked as a programmer in the state entities of Texas and even with the technology involved 15 years ago there was a judge and county clerk that were able to outsmart the judge wheel, they were taking bribes to guarantee certain people were seen only by that judge.   I was the one who uncovered how they did it because I love data and I can troll it looking for patterns.   The judge and county clerk did not get re-elected but the one who went to jail was the clerk's number one because he was the one who actually messed with the data.   Just think if someone with the skills I have now,... [More]
Comment icon #4 Posted by Abramelin 1 year ago
We created God/gods in our own image, and we're fkd ever since. Now we've created Artificial Intelligence, our latest baby that will very probably also grow out of proportion quite soon. No doubt soon those that question the decisions and guide lines dictated by AI will be imprisoned or worse. Technogically we have advanced. Psychologically we're still infants.  
Comment icon #5 Posted by Cho Jinn 1 year ago
Bad Idea Jeans
Comment icon #6 Posted by spartan max2 1 year ago
I don't see this as a good idea on the near future but possibly at some point. Idk, would really have to think about it. 
Comment icon #7 Posted by Amorlind 1 year ago
The terminology used currently for IA is in my opinion not appropriate...there is absolutely no Intelligence and no reflexion in current IAs...only exploitation of algorithms created by talented developpers and also then submitted to bias, etc... Judging is not only applying laws, its also interpreting them and making them evolve by human decisions taken in court...its a part of the legal system and we cant reduce that to a list of 0 and 1  Thats the job judges signed up for, if they are not happy with that, they have to find another job. The legal system is overwhelmed by requests of course,... [More]


Please Login or Register to post a comment.


Our new book is out now!
Book cover

The Unexplained Mysteries
Book of Weird News

 AVAILABLE NOW 

Take a walk on the weird side with this compilation of some of the weirdest stories ever to grace the pages of a newspaper.

Click here to learn more

We need your help!
Patreon logo

Support us on Patreon

 BONUS CONTENT 

For less than the cost of a cup of coffee, you can gain access to a wide range of exclusive perks including our popular 'Lost Ghost Stories' series.

Click here to learn more

Recent news and articles