A low-code dev platform for data-centric NLP

Designed with an open-core available on GitHub. Build from templates or build completely from scratch.

The Kern AI platform consists of four products


refinery is the editor for data-centric natural language processing. It combines training data and algorithms in a way that you can easily build NLP automations, e.g. to prototype ideas within an afternoon or to build quality assurance for your labeling workflow.

You can click on the below feature cards to jump into the documentation.

Manual labeling editor

refinery comes with a built-in editor (incl. role-based access) supporting classifications, span-extraction and text generation. Further, you can export data to other annotation tools like Labelstudio.

Best-in-class data management

Use our modular data management to find e.g. records with below 30% confidence and mismatching manual and automated labels, sorted by confidence. Assign that data either to an inhouse expert or a crowdlabeler.

Native large-language-model integration and finetuning

You love Hugging Face, GPT-X or cohere for their large language models? We do too. That is why we integrated them into refinery. You can use them for embeddings (and neural search), active transfer learning, or even to create the training data for finetuning these LLMs on your data.

Automate with heuristics

refinery is shipped with a Monaco editor, enabling you to write heuristics in plain Python. Use them for e.g. rules, API calls, regex, active transfer learning or zero-shot predictions.

Monitor your data quality

In the project dashboard, you can find distribution statistics and a confusion matrix showing you where your project needs improvement. Every analysis can be filtered down to atomic level.

Integrate into your workflow via our Python SDK

Use our Python SDK (also available for the CLI) to export and import data with ease. For instance, you can use it to batch-export data from refinery to your favorite data science framework or to batch-import data from your data sources into refinery.


Yes, you read that right. Our flagship product is open-sourced under the Apache 2.0 license. You can find the code on GitHub. We are also happy to accept contributions.

Why we believe low-code and data-centric will power the future of NLP

Speed and flexibility

You need some Python knowledge to build an application on our platform, but you do not need a PhD. You have the full flexibility, but can develop fast and easy.

Your intellectual data property

Data lives longer than code. With our data-centric approach, you build your intellectual data property, allowing you to stay flexible when it comes to your requirements.

Models change

Let's be honest here, models change. GPT-4 will follow GPT-3, and Huggingface releases new models incredibly fast. But your data is produced by your applications. Choose the right stack to build your data strategically.

Long-tail use cases

There are use cases like the [email protected] one, which is recurring quite often. But there are also use cases that are very specific to your business. We help you build both with our low-code, data-centric NLP approach.

Frequently asked questions

Common use cases

  • Email automation
  • GPT-like content
  • Data-centric NLP


  • Platform
  • Architecture
  • How it works
  • Labeling services


  • Changelog
  • refinery
  • bricks
  • gates
  • workflow


  • About
  • Blog
  • Careers
  • Contact
  • Pricing
  • Lean NLP Canvas


  • Imprint
  • Privacy policy
  • Security
  • Cookie settings
  • Public funding

Subscribe to our newsletter

The latest news, articles, and resources, sent to your inbox weekly.

© 2023 Kern AI GmbH. All rights reserved.