The Company
Founded by leading experts in AI and NLP, this NYC-based company provides high-precision text reading for large data sets. Their clients include global healthcare and financial services alongside government contracts.
What You’ll Do
Map textual data to representations of meaning suitable for data aggregation
Design broad-coverage semantic representations, deep enough to capture linguistic phenomena relevant to data aggregation, then build statistical models to learn these representations
Work with large datasets to curate and enrich knowledge bases and package your model artifacts
Develop and improve entity recognition and entity linking using user feedback, dense embeddings, knowledge graph representations
Build and deliver models that bring value to clients and can be scaled efficiently over large data sets
What You Need to Succeed
3+ years of experience in prototyping, testing, and deploying NLP systems to production
Track record of publishing top-tier conferences
Experience working with large datasets
Experience working on text classification, entity recognition, entity disambiguation, relation extraction, interactive information extraction and/or knowledge base construction
Experience using tools for prototyping and deploying deep learning models, e.g. PyTorch, TensorFlow, spaCy, Hugging Face, scikit-learn.
Experience developing in OSX and Linux
Knowledge of state-of-the-art deep learning and NLP methods, e.g. CNNs, RNNS, transformers
Strong programming skills and experience with ML and deep learning libraries
Solid understanding of computer science fundamentals and software development
Ability to write and deliver code
What’s On Offer
Competitive compensation package
Bonus and Equity