International law firm Allen and Overy, for instance, is rolling out an AI chatbot called Harvey as a legal assistant to its 3,500 lawyers. “[Built using OpenAI’s GPT technology,] Harvey can do contract analysis, due diligence and present draft arguments for cases. [With this,] human lawyers can spend more time with clients, think of a strategy for criminal investigation or a tax structure change, etc, as they no longer need to [do the same thing day in and out, such as] going over similar legal documents again and again looking for errors,” she shares.
She also gave the example of the Khan Academy piloting the use of OpenAI’s GPT-4 for its AI interface called Khanmigo to enable the underserved to receive quality education. According to the Khan Academy’s blog, Khanmigo is “like a virtual Socrates, guiding students through their educational journey.” It engages students in back-and-forth conversation instead of giving answers. As such, it can help students get help with math, learn computer programming or prep for examinations, to name a few.
With more organisations adopting AI, Khanna believes it is high time to think about AI ethics and regulations. “Before we get too excited about AI, what I really need you to think is to not trust the machine so much. I’m an AI engineer, so I know we make mistakes and that the data [supporting the AI] can be faulty. So in this new hybrid man-machine economy, humans are necessary to keep an eye on the machine because AI can hallucinate or take your corporate secrets,” she asserts.
For the latter, she referred to the recent case of Samsung Electronic banning the internal use of generative AI tools after it was discovered that employees accidentally leaked sensitive data through ChatGPT.
See also: How AI-powered coding is revolutionising software development
“Humans need to govern AI and to benefit ourselves and humanity. In this new world, you are the hero of this story, not AI,” she says.