NLP tool for text classification, sequence labeling, and relation extraction
Top 85.2% on sourcepulse
Macadam is a Python NLP toolkit built on TensorFlow (Keras) and bert4keras, designed for text classification, sequence labeling, and relation extraction. It supports a wide array of embedding models and numerous algorithms for its target tasks, catering to researchers and practitioners in natural language processing.
How It Works
Macadam leverages the flexibility of TensorFlow/Keras and the advanced capabilities of bert4keras to provide a unified framework for various NLP tasks. It supports a diverse range of embedding strategies, from traditional Word2Vec and FastText to modern transformer-based models like BERT, ALBERT, and RoBERTa. The toolkit offers a modular design, allowing users to easily switch between different network architectures (e.g., TextCNN, Bi-LSTM-CRF) and embedding types for fine-tuning or experimentation.
Quick Start & Requirements
pip install Macadam
or pip install -i https://pypi.tuna.tsinghua.edu.cn/simple Macadam
Highlighted Details
Maintenance & Community
The project is authored by Yongzhuo Mo. Further community engagement channels or roadmap details are not explicitly mentioned in the README.
Licensing & Compatibility
The repository does not explicitly state a license. The provided BibTeX entry suggests it is a general GitHub project. Users should verify licensing for commercial or closed-source use.
Limitations & Caveats
The README indicates that relation extraction (RE) and specific model implementations like TextGCN for text classification and MRC for sequence labeling are still under TODO status. The project appears to be primarily focused on Chinese NLP tasks, with datasets like CLUE NER 2020 and People's Daily corpus mentioned.
2 years ago
1 day