Discover and explore top open-source AI tools and projects—updated daily.
Polish NLP resources: pre-trained models and language resources
Top 79.0% on SourcePulse
This repository provides a comprehensive collection of pre-trained models and language resources specifically for Natural Language Processing (NLP) tasks in Polish. It caters to researchers and developers working with the Polish language, offering a wide array of tools to enhance NLP applications.
How It Works
The project offers a diverse range of NLP models, including word embeddings (Word2Vec, FastText, GloVe, Wikipedia2Vec), language models (ELMo, RoBERTa, BART, GPT-2, Longformer), and text encoders for semantic similarity tasks. It also includes machine translation models, text correction utilities, and text ranking models for RAG pipelines. The resources are trained on extensive Polish corpora, leveraging various architectures and training methodologies to achieve high performance.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive