XiaShan1227/Graphormer

Do Transformers Really Perform Bad for Graph Representation? [NIPS-2021]

31
/ 100
Emerging

This project helps machine learning researchers improve how they represent and analyze complex graph data using Transformer models. It takes graph structures, such as molecular graphs or social networks, and processes them to generate better numerical representations. The primary users are researchers and practitioners in machine learning and artificial intelligence working with graph-structured data.

No commits in the last 6 months.

Use this if you are a machine learning researcher exploring advanced graph representation learning techniques, particularly those interested in applying or improving Transformer architectures for graph data.

Not ideal if you are a business user looking for a plug-and-play solution for graph analytics without deep technical expertise in machine learning and model training.

graph-neural-networks machine-learning-research graph-representation-learning artificial-intelligence
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 8 / 25
Community 15 / 25

How are scores calculated?

Stars

61

Forks

10

Language

Python

License

Last pushed

Oct 28, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/XiaShan1227/Graphormer"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.