zhpmatrix/BERTem
论文实现(ACL2019):《Matching the Blanks: Distributional Similarity for Relation Learning》
This project helps natural language processing researchers or practitioners extract relationships between entities from text. You input raw text containing mentions of entities, and it outputs classifications of the semantic relationships between those entities (e.g., 'person works at organization'). This is useful for anyone building systems that need to understand factual connections hidden in unstructured text data.
153 stars. No commits in the last 6 months.
Use this if you are a researcher or NLP engineer looking to implement or experiment with neural relation extraction models, particularly those based on BERT embeddings.
Not ideal if you need a complete, pre-trained, production-ready model for relation extraction without needing to supply your own datasets or pre-trained models for fine-tuning.
Stars
153
Forks
24
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Dec 08, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/zhpmatrix/BERTem"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
davidsbatista/BREDS
"Bootstrapping Relationship Extractors with Distributional Semantics" (Batista et al., 2015) in...
davidsbatista/Snowball
Implementation with some extensions of the paper "Snowball: Extracting Relations from Large...
nicolay-r/AREkit
Document level Attitude and Relation Extraction toolkit (AREkit) for sampling and processing...
plkmo/BERT-Relation-Extraction
PyTorch implementation for "Matching the Blanks: Distributional Similarity for Relation Learning" paper
thunlp/FewRel
A Large-Scale Few-Shot Relation Extraction Dataset