CAMeL-Lab/CAMeLBERT
Code and models for "The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models". EACL 2021, WANLP.
This project offers pre-trained language models specifically designed for Arabic text. It takes raw Arabic text and can identify sentiments (positive/negative), determine the dialect, classify poetry, recognize named entities like people or places, and tag parts of speech. The models are ideal for linguists, researchers, or anyone working with large volumes of Arabic language data who needs to automatically categorize or extract information from text.
No commits in the last 6 months.
Use this if you need to perform advanced text analysis tasks on Arabic language content, such as understanding opinions, identifying geographic origins of text, or automating data extraction.
Not ideal if your primary need is for languages other than Arabic, or if you require a simple keyword search tool instead of a sophisticated linguistic analysis system.
Stars
55
Forks
13
Language
Python
License
MIT
Category
Last pushed
Jun 21, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/CAMeL-Lab/CAMeLBERT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
deepset-ai/FARM
:house_with_garden: Fast & easy transfer learning for NLP. Harvesting language models for the...
extreme-bert/extreme-bert
ExtremeBERT is a toolkit that accelerates the pretraining of customized language models on...
Grenzlinie/MgBERT_LLM_Classification_for_Materials_Science
Source code and result for Paper 'A Prompt-Engineered Large Language Model, Deep Learning...
UBC-NLP/marbert
UBC ARBERT and MARBERT Deep Bidirectional Transformers for Arabic
LazerLambda/modern-bert-score
Re-implementation of BERTScore for evaluation of generated text, leveraging vLLM and...