Introducing the Open-Source Library for Testing NLP Models

While there’s a lot of work done on defining guidelines and policies for Responsible AI, there are far fewer that data scientists can apply today to build safe, fair, and robust models. This session introduces the open-source nlptest library, which provides a comprehensive solution to testing NLP models before taking them to production.

The library supports the full lifecycle of automatically generating tests, editing them, running them, evaluating pass/fail criteria, and generating augmented data to improve models. The nlptest library currently supports testing Spark NLP, Hugging Face, and spaCy models and is designed for extensibility for testing more NLP libraries and tasks.

This session will show you what problems the nlptest library solves, how to get things done, and how to extend it.

About the speaker

David Talby

CTO at John Snow Labs

David Talby is the Chief Technology Officer at John Snow Labs, helping companies apply artificial intelligence to solve real-world problems in healthcare and life science. David is the creator of Spark NLP – the world’s most widely used natural language processing library in the enterprise.

He has extensive experience building and running web-scale software platforms and teams – in startups, for Microsoft’s Bing in the US and Europe, and to scale Amazon’s financial systems in Seattle and the UK.

David holds a Ph.D. in Computer Science and Master’s degrees in both Computer Science and Business Administration. He was named USA CTO of the Year by the Global 100 Awards and GameChangers Awards in 2022. 



Online Event: April 4-5, 2023



Presented by