We are currently updating our platform so you may experience some temporary disruptions or limited functionality
There will be many jobs AI can do better than humans. We might want humans to carry on doing them anyway.
Ask a lawyer about their biggest frustrations and, as long as the conversation’s private, one of them will be the inconsistency of judges; or, more specifically, the inconsistency between judges. Their individual rulings, perhaps driven by unconscious or other kinds of bias, may be sufficiently predictable for a law firm to decide its strategy – including whether to drop the case entirely – upon the appointment of a presiding judge. So predictable, in fact, that technology companies are already on to it. Algorithms now exist that are capable of reviewing witness statements for “sentiment” (principally language patterns and the use of key words) and matching these to the previous rulings by the judge to provide a predicted outcome and, by extension, a decision on whether or not the case should be taken to court, settled, or dropped. This is good business for the law firm, who can provide this analysis to their client. And clients, in general, prefer analysis to opinion.
All of which feels just a little depressing.
Instead of using AI (or in this case its younger sibling, Machine Learning) to predict the bias in the judicial system, what if it could be deployed instead to improve it? An equivalent analysis could be provided to judges themselves, reviewing comparable cases and rulings by other judges to provide a judgment “recommendation”. In this way greater consistency would be assured. And in any case, in areas such as sentencing, aren’t guidelines so confined that the opinions of judges are already largely removed from the equation?
This is causing major palpitations in the legal profession. How could such a system be accountable? Should there be a right to sue an algorithm? And if all judges were to follow algorithmic guidance, would this not itself create bias over time, as judgments “revert to the mean”?
The short answer is that AI can be used to inform, help, and guide humans, and perhaps protect them, but it cannot be allowed to make decisions over them, even if those decisions are likely to be better or fairer. AI will certainly have a major role to play in the legal system of the future, but the judgment of one fallible human being over another is one that shouldn’t be handed to robots.
There will be many jobs AI can do better than humans. We might want humans to carry on doing them anyway.
Russell Seekins
Editor Intermedia; Partner, Re:Strategy
We give innovators and regulators a forum in which to explore, debate and agree the best policies and regulatory frameworks for widest societal benefit.
Insight: Exchange: Influence
We give members a voice through conferences, symposiums and private meetings, as well as broad exposure of their differing viewpoints through articles, reports and interviews.
The new website will make it easier for you to gather fresh insights, exchange views with others and have a voice in the debate
Take a look Learn more about our updatesYou are seeing this because you are using a browser that is not supported. The International Institute of Communications website is built using modern technology and standards. We recommend upgrading your browser with one of the following to properly view our website:
Windows MacPlease note that this is not an exhaustive list of browsers. We also do not intend to recommend a particular manufacturer's browser over another's; only to suggest upgrading to a browser version that is compliant with current standards to give you the best and most secure browsing experience.