Categories
Uncategorized

Why I joined Catylex. It’s the data, stupid.

I’ve had a twenty-year obsession with translating legal language – specifically, contract language – into data.  It’s a problem we started to solve at Exari, but it’s a big and complicated problem, so we didn’t quite finish the swing.  Nonetheless, it’s a problem worth solving, and it’s a solvable problem.   

The need is beyond question. Business and legal teams know that many of their contracts contain hidden risks, errors, gaps and surprises.  They just don’t know which ones and how much damage will be caused if those particular deals blow up.  As stressful as this may be, the cost of having experts read every page and track every risk is prohibitively high, so they suck it up and hope for the best.  Maybe that scary indemnity will never be triggered.  Maybe those uncapped sales contracts never come to light.  Maybe that rogue MFN clause never hits your revenue projections.  Maybe that missing amendment was totally harmless. 

If a machine can find all these risky contract terms quickly and at a fraction of the cost, everyone wins.  The lawyers don’t lose sleep waiting for bad contracts to blow up.  The business can mitigate and manage the risks that come to light.  And the weak contract haircut doesn’t spoil your financials. 

Easy to say.  Very hard to do.  

Many of you have heard about magic contract AI.  Some of you may have tried it.  Most of it has been disappointing.  This does not mean the idea was doomed from the start.  It just means it wasn’t fully baked yet.  It takes plenty of science and art to get this bread to rise.  

One secret to building a contract data machine that works is a team with deep experience, knowledge and skills, spanning the legal, business and technical domains.  This is probably the most important ingredient.  Many will claim to have this multi-disciplinary depth.  Catylex has it.  A rare bunch of legal-language-business-data obsessives who are solving complexity so that you don’t have to.  You can just eat the results. 

So, I’m excited to join Catylex, where the dream of building a fully functional contract data machine is a shared obsession, and where the promise of such a machine is now becoming reality.  This might just be the best thing since sliced bread. 

Categories
Uncategorized

The Clause Tracking Conundrum

I have to admit that “clause tracking” sounds good. Let’s analyze how many deals contain our preferred limitation of liability clause. Let’s track how many times we had to use fallback #2. It feels like these would be really useful things to know. Which is probably why so many contract managers ask for it.

Not Quite a Limitation of Liability Clause…

But as you peer behind the shiny facade of clause tracking, things start to look a little grimy. To explain why, I’m going to use a real life example that I recently plucked from the EDGAR database:

10.  Limitation of Liability. IN NO EVENT SHALL EITHER PARTY BE LIABLE TO THE OTHER FOR EXEMPLARY, CONSEQUENTIAL, SPECIAL OR PUNITIVE DAMAGES, ARISING FROM OR RELATING TO THE AGREEMENT, WHETHER IN CONTRACT, TORT (INCLUDING NEGLIGENCE) OR OTHERWISE, EVEN IF SUCH PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES; PROVIDED, HOWEVER, THAT THIS LIMITATION SHALL NOT APPLY TO EACH PARTY’S (A) BREACH OF SECTION 4 (CONFIDENTIALITY); (B) BREACH OF SECTION 7 (OWNERSHIP AND INVENTIONS); (C) INDEMNIFICATION OBLIGATIONS; OR (C) RECKLESSNESS, INTENTIONAL MISCONDUCT OR FRAUD. THE PARTIES FURTHER AGREE THAT THE MAXIMUM LIABILITY UNDER THIS AGREEMENT, REGARDLESS OF THE TYPE OF CLAIM OR NATURE OF DAMAGES, SHALL EXCEED THE TOTAL VALUE OF SERVICES PROVIDED BY DAVA TO AROG HEREUNDER.

The most glaring problem with this clause should be clear to an experienced contract negotiator fairly quickly. The word “NOT” is missing. Rather than saying “SHALL NOT EXCEED” it says “SHALL EXCEED”. And so what was probably intended as an aggregate cap on the supplier’s liability, instead becomes a strange declaration of liability exposure. Clearly, someone made a mistake. Quite a big one.

Which brings me to clause tracking. How do I build a machine that can spot this problem and flag it for remediation?

If my clause unit is the whole numbered paragraph 10 above (which seems reasonable on its face), then a clause tracking tool might use a statistical bag of words method to tell me that (because of the missing 3 letters) this is a 99% match to my preferred limitation of liability clause. But given the difficulty of perfectly matching in the grubby world of scanning errors, I am tempted to ignore 1% errors as immaterial. So this clause probably gets a green light.

This is the “perfect match” conundrum. In the real world of imperfect scans and OCR errors, I am almost never going to get a perfect match, so I treat 99% as OK. But in many cases, including the example above, a 99% match could actually contain a very high risk error.

If the other party produces the first draft (“third party paper”), your chances of clause matching go from bad to worse. Even if their template includes an acceptable version of the clauses you require, there’s a vanishingly small chance they used the same words as your clause library. So even the “matched” clauses will be well below 100% and you’ll have to double-check everything.

To be fair, a clause matching tool might redline the missing words and help a human negotiator to catch any mistakes. But redlining isn’t exactly ground-breaking. All it does it augment a skilled human process. We can unlock far more value by eliminating human tasks rather than augmenting them.

Another (arguably better) approach is to train a machine to understand the meaning of a “clause” and to define your risk playbook semantically, rather than by reference to precise word combinations. In this approach, you might split Section 10 into two semantic concepts. Sentence #1 is an exclusion of indirect loss, which is good. Sentence #2 is (or would be, but for the error) an aggregate cap on liability, which is also good (for the Supplier).

By training a machine with many examples of these two concepts, we can produce an AI service that assigns the correct semantic meaning to any sentence/clause processed, regardless of the precise words used. This offers far more value than precise word matching because it works not just for deals on our paper, but also for deals on third party paper.

We still need to solve the problem of training our machine to understand mistakes. Sentence #2 looks extremely similar to a liability cap. But it isn’t. So I need a “typo” training class to teach my AI that sloppy drafting like this should not be confused with the real thing.

Don’t get me wrong. There is some value to “perfect match” clause tracking if you are working with high fidelity documents that don’t suffer from OCR noise. But semantic machine learning analysis has far greater potential to automate contract review, especially if a large chunk of your deals are on third party paper.