Navigating the benefits and risks of generative AI
Expanding use of in-house AI requires law firms to monitor themselves
Artificial intelligence is rapidly weaving itself into our personal and professional lives – and in applications that can leave us feeling empowered and overwhelmed in equal measures. As the release of ChatGPT in November 2022 made clear, generative artificial intelligence (AI) has impressive power to synthesize large amounts of information in seconds. Data-heavy organisations are seeing its potential to transform their work – even as they acknowledge the need to refine the technology and verify the accuracy of the results it produces.
Many law firms have been fine-tuning their own generative AI tools in the wake of ChatGPT’s launch, aiming to tap into the benefits of the technology while managing its potential risks. However, the adoption of technology to support legal work is nothing new for law firms. For years, firms have been using technology to automate tasks ranging from e-discovery to conflict checks. The adoption of generative AI is simply the next step in that process – but it does potentially introduce new complexities for firms, their clients and employees.
“Law firms have to consider a range of interests as they use generative AI in their work,” said Sharon Glynn, director and underwriter in the Bond & Specialty department at Travelers Europe. “Some clients may insist that firms use it to drive cost savings, while others may insist on firms not using it because they prefer the benefit of working with a human, or are concerned about the risks to their data. Firms must consider their workforce too. They must balance the needs of current and future employees who are eager to embrace generative AI with those employees who are less enthusiastic. In the process, they must also be mindful of the evolving tools they need to compete in the marketplace.”
Firms forge ahead with generative AI
A number of firms have already been navigating these opportunities and challenges in real time with their own generative AI tools. Allen & Overy, for example, released its “Harvey” model in early 2023 (see sidebar). Travers Smith developed a ChatGPT-like tool that it is using internally and also making available to other firms.
Specifically, the legal technology team at Travers Smith built a chatbot, YCNBot, as an alternative to ChatGPT. While the firm isn’t using YCNBot for client work yet, it is encouraging its lawyers to experiment with it internally and provide feedback. The firm made the code downloadable via GitHub and free through an open-source licence. This allows other firms to use the code to plug into the application programming interface of Microsoft and OpenAI, then tailor it to their own needs — including enhanced controls around compliance, security and data privacy.
Generative AI tools are triggering the creation of new roles within firms as well – and signalling a new commitment to the ongoing development and refinement of the technology.
In October, Eversheds Sutherland announced it had created a number of new positions to support AI. The firm appointed Nasser Ali Khasawneh, tech litigation partner at the firm, as its first Global Head of AI, a position that reports to Co-CEOs Lee Ranson and Mark Wasserman. In conjunction with that announcement, the firm said it had formed a global AI leadership team, a global AI task force, and a global generative AI skills development program.
Ranson and Wasserman described generative AI as a “catalyst for a fundamental shift in the legal profession” and “an imperative for the legal industry,” while also emphasising they would be taking a “people plus technology” approach.
Think of AI as amplified intelligence
The Eversheds Sutherland announcement demonstrates how firms are treating generative AI as a powerful tool that serves firms best when combined with human oversight and contributions. But it’s difficult to regulate a technology that is evolving faster than humans can understand it. Law firms, which are in the business of defining rights and responsibilities and protecting against risk, are in an important position when it comes to identifying the pitfalls of overreliance on AI.
AI impacts more functions of organisations than we may even realise, so firms will need to continuously recalibrate their efforts to monitor it. The regulation of AI applications isn’t simply about getting legal facts right. There are wider-ranging consequences that may impact how well a firm keeps its commitments to stakeholders and how the firm’s leadership and workforce evolve over time.
For instance, Amelia Kallman, a futurist who addressed Travelers legal clients at a conference earlier this year, says many of her client firms have set goals to be carbon neutral by 2030, but they haven’t figured AI’s carbon footprint into those goals – and its footprint is substantial.
“The more powerful AI is, the more energy it requires,” she said. “While data is still being collected, research indicates a generative AI search requires 4-5 times more energy than a regular search-engine query. The energy it took to train ChatGPT-3 with 175 billion parameters was equivalent to driving 123 gas vehicles for one year, generating 552 tons of CO2. Comparatively, GPT-4 was trained on over a trillion parameters. I believe that while it will inevitably contribute to our footprint, it is also the same technology that will help us to solve some of our greatest climate challenges. It’s up to us to be strategic and responsible today to ensure sustainability for tomorrow.”
Within firms, AI use also has the power to impact the makeup of the organisation’s leadership and overall workforce. A study by Revelio Labs has found that AI is set to disproportionately replace jobs held by women – largely because today women are less likely to hold leadership and decision-making positions. If women or other people in underrepresented groups aren’t at the table when firms are discussing the potential applications of AI across the business, their concerns are more likely to be overlooked.
Authored by Travelers