By Ronald Yu*
In his book Outliers, Malcolm Gladwell repeatedly mentions what has become known as the “10,000-hour rule”, which states that to become world-class in any field you must devote 10,000 hours of “deliberate practice”.
Whether or not you believe the 10,000-hour figure (Anders Ericsson, whose research Gladwell cites, disputes the number though he does support the notion that much practice is needed to achieve outstanding competence), many would acknowledge that to become an accomplished legal professional requires considerable legal (arguably in multiple practice areas), communications and, particularly in in-house environments, interpersonal skills that are often acquired after a tremendous amount of effort exerted over many years.
Enter the artificial intelligence (AI)
There has been much hoopla about artificial intelligence-based legal systems that some might have you believe will soon replace lawyers (no doubt causing a degree of anxiety among some legal professionals).
There is some misunderstanding among many lawyers and much of the public about what present AI systems are capable of. Can a legal AI, based on current technology, actually “think” like a lawyer? No. At best, today’s AI is an incomplete substitute for a human lawyer (though it could reduce the need for some lawyers and I’ll get to all that later).
However, something we should think seriously about right now is the long-term implication of the introduction of AI into the legal environment — notably the potential future loss of legal wisdom.
Why doesn’t AI think like a human?
First, let’s answer the question of why today’s AI doesn’t actually mimic the human brain. As an example, let’s look at automated translation systems such as those available from Google, Facebook or Microsoft. Such systems might appear to work the way human translators do, but what they actually do is match patterns derived from analyses of thousands, if not millions, of pages of text found on the web (employing a technology known as statistical machine translation).
For instance, if such a system wants to know how to translate the English greeting “hello” into French, it scans English and French translations on the web, statistically analyses the correlations between “hello” and various French greetings, then comes to the conclusion that the French equivalent of “hello” is “bonjour”.
Current AI is good at this kind of pattern matching, but less so at cognition and deductive reasoning. Consider that the human brain not only stores a large number of associations and accesses useful memories (sometimes quickly, sometimes not), it also transforms sensory and other information into generalisable representations invariant to unimportant changes, stores episodic memories and generalises learned examples into understanding. There are key capabilities yet to be solved by current AI technology, which involve generalisable knowledge representation and of cognition based on that representation.
Thus, while present AI-based legal systems might analyse judicial decisions, for example, to help litigators gain insights to a judge’s behaviour or a barrister’s track record, they do so by scrutinising existing data to reveal patterns and not by extrapolating from the content of those decisions the way an experienced human legal professional might.
The temptation to make redundant
But as AI systems become more capable, the temptation grows to use such systems not only to supplement but also to eliminate the need for some personnel.
While an AI system weak in cognition but strong in pattern matching probably could not replace an experienced professional in terms of drawing inferences, deductive reasoning or combining different practice areas to arrive at more comprehensive solutions, it could perform certain tasks such as searching for patterns of words in documents (for gathering evidence, for example) that have hitherto been delegated to lower level staff (such as paralegals, trainees, junior associates) — and do so better than any human could.
While one might argue that the introduction of AI systems will lighten the workload of legal professionals and thereby improve their quality of life, it also potentially diminishes the need for junior legal staff, which would only exacerbate the problem of oversupply in the legal profession.
Shrink now, suffer later?
If fewer junior legal professionals are hired, this means a smaller population of lower level staff and thus a smaller feeder pool for more senior positions. Not only that, as AI systems continue to improve and take more legal-related functions away from human staff, will this deprive junior legal professionals of opportunities to gain important experience — ie, get their 10,000 hours?
Will this result in fewer quality, experienced legal professionals in the future? In other words, if AI systems make more and more junior legal staff redundant, will that mean fewer great lawyers in the future?
And the future of legal AI?
There are yet two more (albeit related) things to think about.
First, the development and maintenance of a good AI system requires both technical and legal competency. Put another way, a legal AI system programmed by geeks ignorant in the law will be seriously, if not fatally, flawed. Thus, if we want to continue to develop more capable legal AI systems, good content providers — ie, good lawyers — will be needed.
Second, as laws and the legal, business and social environments in their respective jurisdictions evolve, developments that might not have been anticipated just a few years earlier will emerge. Only the very best legal and other minds will be able to cope with some of these developments — and update the relevant legal AI systems accordingly.
For example, when the US passed the Leahy-Smith America Invents Act (AIA) in 2011, it introduced new review procedures for existing patents with the intent of improving patent quality, but it also had the effect of introducing several unintended consequences (including the use of such procedures by hedge funds to invalidate patents for the purpose of affecting the stock price of the companies holding the patents, and the negative impact the AIA has had on inventors).
Updating an AI system to properly incorporate these developments requires not only a deep understanding of US patent law but also a perspective on patents, finance and the impact of patent policy and procedures on innovation — something that can only really be appreciated after years of experience. Moreover, this is something that could not have been programmed into an AI system half a decade ago and such content could probably not have been provided by a less capable, less experienced legal professional to an AI developer.
So what, if anything, can be done?
Sadly, there are no easy answers. Graduating fewer lawyers might alleviate the problem of oversupply but would also result in unemployment at educational institutions.
While the government (or government-backed NGO) could establish some sort of training centre for under-employed junior lawyers where these professionals could offer services pro bono to build their experience, this also smacks of government interference in the private practice market.
But we need to start thinking of solutions now because the introduction of AI into the legal profession and the potential prospect of putting more lawyers out of work could have profound implications for legal AI systems and the profession as a whole.
*Ronald Yu is a general counsel, a U.S. Patent Agent, digital forensics examiner and teaches classes on patent, intellectual property and information technology law. Ronald has published articles on online learning, digital forensics, intellectual property, information technology and constitutional law, authored/edited the books A Corporate Journey to the West, Intellectual Property for Business and the Hong Kong chapter for the 3rd edition of the book Electronic Evidence.