S2ORC CS Enriched: 1.1 Million Computer Science Papers with Structured Metadata

S2ORC CS Enriched: 1.1 Million Computer Science Papers with Structured Metadata

HuggingFace →

#datasets #scientific-papers #machine-learning

S2ORC CS Enriched: 1.1 Million Computer Science Papers with Structured Metadata

Allen AI's Semantic Scholar Open Research Corpus (S2ORC) is one of the largest open collections of academic papers, with full text, abstracts, citation graphs, and venue metadata across all fields. It is an extraordinary resource. It is also enormous and undifferentiated. If you want to study computer science specifically, or if you need structured information about what methods a paper uses, what models it trains, what compute it reports, or what its limitations are, you have to extract that yourself from raw text.

S2ORC CS Enriched is the result of doing that extraction at scale. We filtered S2ORC to its computer science subset and ran LLM-based enrichment across the full corpus, producing 1,117,706 rows with structured metadata that does not exist in the original dataset.

The full dataset is 54.7 GB across 1,118 parquet files, available on HuggingFace.

What Changed

The base S2ORC columns are preserved: corpus ID, parsed title, authors, abstract, full text, sections, figures, references, external identifiers (DOI, ArXiv, DBLP, ACL, PubMed), citation counts, venue metadata, publication date, open access status, and license information. Everything upstream provides, we keep.

On top of that, we added 14 enrichment columns derived from the paper text:

Column Description
summary A concise summary of the paper's contribution
classification Paper type (Original Research, Methods Paper, Application Paper, Survey, etc.)
methods Methods and techniques used
results Key findings and reported outcomes
models Models trained or evaluated
datasets Datasets used
metrics Evaluation metrics reported
limitations Stated or inferred limitations
explicit_gpu_hours GPU hours as reported in the paper
estimated_gpu_hours Estimated GPU hours where not explicitly stated
number_gpu Number of GPUs used
gpu_type GPU hardware (A100, V100, H100, etc.)
training_details Training configuration details
reasoning Chain-of-thought reasoning behind the enrichment

These columns turn unstructured paper text into queryable, filterable structured data.

Why This Matters

Raw paper text is hard to work with programmatically. If you want to answer questions like "which papers trained models on ImageNet using more than 8 GPUs" or "what fraction of NLP papers from 2023 report compute budgets," you need structured fields, not a full-text search over abstracts.

The enrichment makes several things practical that were previously expensive or manual:

Filtering by method or model. Find all papers that use a specific architecture, training technique, or evaluation protocol without reading abstracts.

Compute analysis. The GPU columns enable large-scale analysis of compute trends across CS research: who reports compute, how much they use, and how that has changed over time. This is relevant to understanding the resource dynamics of the field and to studying the economics of AI research.

Training data for research-aware models. The structured columns provide supervision signal for models that need to understand the internal structure of papers, not just their text. A model fine-tuned on this data could learn to extract methods, results, and limitations from new papers.

Benchmark and leaderboard construction. The models, datasets, and metrics columns, combined with results, provide the raw material for building structured benchmark comparisons across the literature.

Meta-research. The classification and limitation columns support large-scale bibliometric analysis: what kinds of papers dominate, what methods are trending, and where the field acknowledges its own gaps.

Scope and Coverage

The dataset covers 1,117,706 computer science papers. All rows are labeled Computer Science in the fieldofstudy_category column. This is not an ArXiv-only dataset: it includes papers from conferences, journals, and other venues indexed by Semantic Scholar. Of the full set, 44,053 rows have a non-null ArXiv identifier.

The enrichment was generated by running LLM extraction over the full text of each paper. Where a paper's text was empty or too short to extract meaningful information, the enrichment columns reflect that (typically with "None specified" values). The reasoning column preserves the chain-of-thought behind each enrichment, which is useful for auditing extraction quality and for understanding edge cases.

Limitations

The enrichment is LLM-generated and therefore imperfect. Extraction quality varies with paper length, writing style, and domain. Papers with unusual formatting, very short text, or heavily mathematical content may have lower-quality enrichment. The reasoning column exists partly to make these failure modes visible.

The compute columns (explicit_gpu_hours, estimated_gpu_hours, number_gpu, gpu_type) are only as good as what papers report. Most papers do not report compute at all. Where compute is not mentioned, these fields are empty or marked as unspecified. The dataset does not hallucinate compute figures; it extracts what is stated or makes conservative estimates where enough information exists.

The classification field is a coarse categorization. Papers that span multiple types (e.g., a methods paper that also includes an application study) are assigned a single label based on the primary contribution as assessed by the enrichment model.

Availability

The dataset is available now:

This is a strict subset of the companion S2ORC Computer Science parquet set by corpus_id, with the enrichment columns added. The parquet format supports efficient column-level access, so you can load only the columns you need without pulling the full 54.7 GB.