arXiv Analytics

Sign in

arXiv:2404.02408 [cs.CL]AbstractReferencesReviewsResources

CMULAB: An Open-Source Framework for Training and Deployment of Natural Language Processing Models

Zaid Sheikh, Antonios Anastasopoulos, Shruti Rijhwani, Lindia Tjuatja, Robbie Jimerson, Graham Neubig

Published 2024-04-03Version 1

Effectively using Natural Language Processing (NLP) tools in under-resourced languages requires a thorough understanding of the language itself, familiarity with the latest models and training methodologies, and technical expertise to deploy these models. This could present a significant obstacle for language community members and linguists to use NLP tools. This paper introduces the CMU Linguistic Annotation Backend, an open-source framework that simplifies model deployment and continuous human-in-the-loop fine-tuning of NLP models. CMULAB enables users to leverage the power of multilingual models to quickly adapt and extend existing tools for speech recognition, OCR, translation, and syntactic analysis to new languages, even with limited training data. We describe various tools and APIs that are currently available and how developers can easily add new models/functionality to the framework. Code is available at https://github.com/neulab/cmulab along with a live demo at https://cmulab.dev

Related articles: Most relevant | Search more
arXiv:2306.00168 [cs.CL] (Published 2023-05-31)
Measuring the Robustness of Natural Language Processing Models to Domain Shifts
arXiv:1910.03474 [cs.CL] (Published 2019-10-04)
Fine-grained Sentiment Classification using BERT
arXiv:1912.11078 [cs.CL] (Published 2019-11-09)
Predictive Biases in Natural Language Processing Models: A Conceptual Framework and Overview