gersteinlab/Struc-Bench

[NAACL 2024] Struc-Bench: Are Large Language Models Good at Generating Complex Structured Tabular Data? https://aclanthology.org/2024.naacl-short.2/

30
/ 100
Emerging

This project helps researchers and developers evaluate how well large language models can generate complex, structured data in various formats. You provide test data in JSON, and it produces generated outputs (tables, HTML, or LaTeX) along with scores indicating generation quality. It's for anyone working with Large Language Models who needs to benchmark their ability to create structured text.

No commits in the last 6 months.

Use this if you are a machine learning researcher or developer evaluating the performance of Large Language Models (LLMs) in generating structured tabular data.

Not ideal if you are looking for a tool to generate production-ready structured data directly, as this is primarily an evaluation and benchmarking framework.

natural-language-processing large-language-models data-generation model-evaluation structured-data
No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 8 / 25
Maturity 8 / 25
Community 12 / 25

How are scores calculated?

Stars

55

Forks

7

Language

Python

License

Last pushed

Jul 31, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/gersteinlab/Struc-Bench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.