Contact us
Contact us
Back to projects

CaLICO: Building the world’s first multilingual foundation model to track online harmful content

Creating a model that can process harmful online content using AI - without generating more - is a challenge. Addressing it across multiple languages and the varying cultural sensitivities of different regions adds another layer of complexity. CaLICO is a large language model being developed by Textgain to overcome these challenges and detect and help mitigate harmful content in all official European languages.

Screenshot 2024-09-01 at 07.49.25
Screenshot 2024-09-01 at 07.49.25

Our approach

Large language models, especially commercial ones, refuse to process toxic language. This makes it almost impossible to use them to process harmful content. With CaLICO, we are developing a language model from scratch that can process harmful content content responsibly, without perpetuating it. As a winner of the prestigious Large AI Grand Challenge, CaLICO has received significant support, including the resources necessary to train a foundation model on the EU’s supercomputers.

What impact will it have?

We believe that CaLICO has the potential to revolutionize harmful content detection. By identifying harmful content across all EU languages, CaLICO will empower better and wider monitoring and processing of potentially dangerous content more effectively. The project’s ongoing development and our commitment to academic collaboration will ensure that CaLICO continues to evolve, providing unparalleled tools for processing online harmful content in the future.

Further Information

  • Winners Announced for the Large AI Grand Challenge (EU website)
  • New AI Language Model Can Detect Online Hate Speech in all EU Languages (TechNational)
  • The 4 startups that will be able to use European supercomputers for their AI (Wired)