notebooks to finetune `bert-small-amharic`, `bert-mini-amharic`, and `xlm-roberta-base` models using an Amharic text classification dataset and the transformers library
-
Updated
May 10, 2024 - Jupyter Notebook
notebooks to finetune `bert-small-amharic`, `bert-mini-amharic`, and `xlm-roberta-base` models using an Amharic text classification dataset and the transformers library
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
Проект в рамках ВКР под названием "Разработка программного модуля для анализа документов, подтверждающих индивидуальные достижения"
This project explores the foundational concepts of ML, NLP, and model optimization to develop an efficient and user-friendly healthcare solution.
🤖 A PyTorch library of curated Transformer models and their composable components
Trankit is a Light-Weight Transformer-based Python Toolkit for Multilingual Natural Language Processing
Currently running NLP project about political communication on Twitter. You can find more projects in my portfolio.
ZaBantu is a fleet of light-weight Masked Language Models for Southern Bantu Languages
Multilingual Deception Detection of GPT-generated Hotel Reviews
Punctuation Restoration for Khmer language
ML and Natual Language Processing
Fine tuned BERT, mBERT and XLMRoBERTa for Abusive Comments Detection in Telugu, Code-Mixed Telugu and Telugu-English.
This repository is a comprehensive project that leverages the XLM-Roberta model for intent detection. This repository is a valuable resource for developers looking to build and fine-tune intent detection models based on state-of-the-art techniques.
The project aims to identify the best model for the classification of texts derived from descriptions of assets subject to Italian judicial auctions. The employed models include both conventional models, such as Logistic Regression, Naive Bayes, SVM, and XGBoost, and neural network models, such as Fasttext and XLM-Roberta.
THREATENING_TEXT_DETECTION_USING_CNN_LSTM_BILSTM_XLMROBERTA
Fine-tuning Language Models
This repository contains a number of experiments with Multi Lingual Transformer models (Multi-Lingual BERT, DistilBERT, XLM-RoBERTa, mT5 and ByT5) focussed on the Dutch language.
Neural syntax annotator, supporting sequence labeling, lemmatization, and dependency parsing.
Add a description, image, and links to the xlm-roberta topic page so that developers can more easily learn about it.
To associate your repository with the xlm-roberta topic, visit your repo's landing page and select "manage topics."