soobinseo/Attentive-Neural-Process
A Pytorch Implementation of Attentive Neural Process
This tool helps researchers and machine learning practitioners quickly experiment with and apply Attentive Neural Processes (ANPs). You provide a dataset with observed data points, and it generates predictions or fills in missing information for unobserved points, even with limited context. It's ideal for those working on predictive modeling or data imputation tasks.
No commits in the last 6 months.
Use this if you need to build models that can make robust predictions and quantify uncertainty from limited or irregularly sampled data, especially for tasks like data reconstruction or forecasting.
Not ideal if you are looking for a pre-trained, plug-and-play solution for a specific application without any coding or model training.
Stars
74
Forks
10
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
May 03, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/soobinseo/Attentive-Neural-Process"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
philipperemy/keras-attention
Keras Attention Layer (Luong and Bahdanau scores).
tatp22/linformer-pytorch
My take on a practical implementation of Linformer for Pytorch.
datalogue/keras-attention
Visualizing RNNs using the attention mechanism
ematvey/hierarchical-attention-networks
Document classification with Hierarchical Attention Networks in TensorFlow. WARNING: project is...
thushv89/attention_keras
Keras Layer implementation of Attention for Sequential models