Can Smaller Language Models Outshine Giants in Legal Tech?
legal AI, smaller models are showing they might punch above their weight. A study tests nine sub-10B models across key legal benchmarks, revealing that size isn't everything.
The legal industry, like many others, has been eyeing large language models for their potential. But there's a catch, we're talking about big costs, long wait times, and the looming shadow of data privacy. So, do we really need those gargantuan models with over 10 billion parameters to get the job done? Turns out, maybe not.
Testing the Little Guys
In an intriguing twist, a recent study put nine language models with under 10 billion parameters to the test. They were evaluated across three legal benchmarks: ContractNLI, CaseHOLD, and ECtHR. What's the twist? These models were challenged using five different prompting strategies, from direct to the dense RAG method.
Imagine running 405 experiments, each with three different setups, to see how these models stack up. One model, a Mixture-of-Experts activating just 3B parameters, went toe-to-toe with GPT-4o-mini on mean accuracy. Not only that, it actually bested its rival in legal holding identification. Clearly, the architecture and training quality take precedence over sheer size. Who knew brains could beat brawn data?
A Mixed Bag of Strategies
Not all strategies were created equal. Chain-of-thought prompting had its ups and downs, helping with contract entailment but falling short in multiple-choice legal reasoning. On the other hand, few-shot prompting emerged as the most reliable companion across tasks. It's like having a trusty friend who always has your back.
What about retrieval methods, you ask? When comparing BM25 and dense RAG, the results were nearly indistinguishable. It seems the real issue isnβt how well the context is retrieved. Rather, it's how the model makes use of what it gets. It's a classic case of having the tools but not using them well enough.
Affordable and Accessible
Here's the kicker. All these experiments were conducted via cloud inference APIs, and the total cost was just $62. That's right. You don't need a fancy setup or dedicated GPU infrastructure to conduct rigorous evaluations. For those venturing into legal AI, this should be an encouraging revelation.
So, why should we care? If smaller models can deliver comparable results without the hefty price tag or infrastructure demands, they might just be the key to democratizing AI in legal tech. The question is, will the industry adapt, or will it continue to chase the bigger is better mantra?
Get AI news in your inbox
Daily digest of what matters in AI.