dlb-rl/pulse-rl

Code for PulseRL: Enabling Offline Reinforcement Learning for Digital Marketing Systems via Conservative Q-Learning.

28
/ 100
Experimental

PulseRL helps digital marketing managers and strategists improve campaign performance using historical campaign data. It takes your past campaign interaction data (like clicks, views, and conversions) and uses it to recommend optimal strategies for future campaigns, aiming to maximize engagement and ROI. This tool is designed for professionals in digital advertising or e-commerce who want to refine their targeting and content delivery without running live, risky A/B tests.

No commits in the last 6 months.

Use this if you have a wealth of historical digital marketing data and want to develop more effective campaign strategies without directly experimenting on live audiences.

Not ideal if you need a tool for real-time campaign optimization or if you don't have a substantial dataset of past marketing interactions.

digital-marketing campaign-optimization e-commerce advertising-strategy marketing-analytics
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 4 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

8

Forks

1

Language

Python

License

MIT

Last pushed

Sep 08, 2021

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/dlb-rl/pulse-rl"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.