What It Means to Convert a PDF Paper into an ARA
A conventional research paper is usually distributed as a PDF. The PDF presents the research as a linear document. It includes motivation, related work, methods, results, figures, tables, and conclusions. This format is designed for human reading.
An Agent-Native Research Artifact, or ARA, treats the paper as a structured research object. The PDF may still exist, but it is no longer the only artifact. The ARA includes the claims, code, data references, experiment settings, outputs, and research history needed for an AI agent to inspect, reproduce, and extend the work.
In this sense, converting a PDF paper into an ARA means turning a document into an executable and inspectable research package.
A PDF might contain a sentence such as:We trained model X and obtained better results on benchmark Y.
An ARA would represent that claim with supporting structure:claim_id: C1 claim: "Model X improves performance on benchmark Y." metric: "accuracy" baseline: "Model B" effect_size: "+4.2 percentage points" evidence: - run_042 - table_2 - figure_4 status: "supported"
The difference is not only formatting. The ARA gives the agent objects it can refer to, execute, verify, and connect.
The problem with a PDF-only paper
A PDF paper compresses the research process into a narrative. That narrative usually contains the final method, the final results, and the final interpretation. It often excludes failed experiments, discarded hypotheses, intermediate configurations, debugging decisions, data-cleaning steps, and exact runtime details.
This compression works for publication, but it creates problems for agents. An agent cannot reliably reconstruct the research process from prose alone. It has to infer missing details, search for unstated assumptions, and guess how tables and figures were produced.
The ARA framing identifies two costs.
The first is the Storytelling Tax. Research is not usually linear, but the paper presents it as if it were. The path from question to result often involves branches, failures, reversals, and partial results. A PDF removes most of that structure.
The second is the Engineering Tax. A description that is sufficient for a reviewer may not be sufficient for reproduction. Phrases such as “standard settings,” “following prior work,” or “we use the default implementation” leave out details that an agent needs.
ARA addresses these costs by making the missing structure explicit.
What an ARA contains
An ARA can be understood as four layers.
The first layer is scientific logic. This contains the claims, assumptions, definitions, hypotheses, and relationships among them.
The second layer is executable code and specifications. This contains scripts, commands, configurations, environment files, dataset versions, model checkpoints, and test procedures.
The third layer is an exploration graph. This records the research path, including failed attempts, partial results, alternative branches, and decisions.
The fourth layer is evidence grounding. This links each claim to the experiments, logs, outputs, tables, and figures that support it.
A simple ARA directory could look like this:research-project/ paper.pdf paper.tex src/ notebooks/ data/ configs/ results/ logs/ ara/ ara.yaml claims.yaml assumptions.yaml evidence.yaml experiments.yaml exploration_graph.json reproduction_report.md
The PDF is still present, but it is one view of the research. The structured artifact is the object an agent can operate on.
Explicit claims
A normal paper often embeds claims inside paragraphs. An agent must identify which sentences are central, which are supporting, and which are background.
ARA makes claims explicit.
Instead of this:Our method improves robustness under distribution shift.
An ARA records this:claim_id: C3 claim: "Method A improves robustness on Dataset D under Shift S." metric: "accuracy" baseline: "Model B" effect_size: "+4.2 percentage points" evidence: - run_042 - table_2 - figure_4 status: "supported"
This gives the agent a unit of analysis. It can ask whether the claim has evidence, whether the metric is defined, whether the baseline is valid, and whether the result can be reproduced.
Exact experiment recipes
A paper may say that experiments were run using standard settings. That is not enough for an agent.
An ARA records the experiment as a reproducible recipe:experiment_id: E7 purpose: "reproduce Table 2" command: "python train.py --config configs/table2.yaml" dataset_version: "dataset-v2-2026-03-01" seed: 17 hardware: "8xA100-80GB" expected_output: "results/table2.json"
This gives the agent the command to run, the configuration to use, the dataset version to expect, the seed, the hardware assumptions, and the expected output.
The agent does not have to infer the procedure from prose. It can execute or inspect the procedure directly.
Evidence links
In a PDF, a claim may refer to a table or figure. That table or figure is usually a processed summary.
ARA records the chain behind the summary:claim → experiment → command → config → raw log → processed result → table or figure
This chain allows an agent to verify whether the claim is supported by the underlying evidence.
For example:evidence_id: EV12 supports_claim: C3 experiment: E7 raw_log: "logs/run_042.log" processed_result: "results/table2.json" rendered_output: "paper/table_2.tex"
The evidence link is important because it turns the paper from a static assertion into an inspectable object. The agent can trace a result back to its source.
Exploration graph
The exploration graph records the research path.
A PDF usually shows the final path. An ARA also records paths that did not become part of the final paper.Hypothesis H1 ├── Experiment E1: failed; learning rate unstable ├── Experiment E2: failed; data leakage found ├── Experiment E3: partial; worked only on small model └── Experiment E4: success; used in final paper
This is useful for an agent that tries to extend the work. It can see which branches were tested, which failures were encountered, and which decisions led to the final method.
The exploration graph also reduces repeated work. An agent does not need to retry an abandoned path without knowing that it was already tested.
Full specifications
A reviewer may understand a phrase such as “we follow prior work.” An agent needs the specific settings.
ARA replaces references and shorthand with specifications.
A weak specification is:We use the preprocessing procedure from Smith et al.
A stronger ARA-style specification is:preprocessing: source: "Smith et al." script: "src/preprocess.py" input_schema: "schemas/raw_data.schema.json" output_schema: "schemas/processed_data.schema.json" parameters: normalize: true token_limit: 2048 remove_duplicates: true
The ARA does not require that all prior work be copied into the artifact. It requires that the artifact contain enough detail for an agent to know what was done and where the operational details are.
Machine-readable definitions
Research papers define terms in prose. ARA makes central terms addressable.concept_id: K1 concept: "agent-native research artifact" definition: "A structured research package intended for agent inspection, execution, verification, and extension." related_claims: - C1 - C2 used_in_experiments: - E1 - E4
This allows an agent to track how a concept is used across the artifact. It also helps avoid ambiguity when a term appears in different sections.
Reproducibility checks
An ARA should support a direct question:Can the main result be reproduced from this artifact?
A reproduction report might include:reproduction: main_claims_checked: - C1 - C2 - C3 successful: - C1 - C2 partial: - C3 failed: [] missing_information: - "Dataset access token not provided" - "Checkpoint hash missing"
The purpose is not to guarantee that every result can be rerun on every machine. The purpose is to identify what is reproducible, what is partially reproducible, and what information is missing.
How a paper becomes agent-readable
A paper becomes more understandable to an agent when it provides:
Need
PDF-only form
ARA form
Main claim
Sentence in prose
Structured claim object
Method
Description
Command, config, code path
Data
Dataset name
Version, schema, access path, hash
Result
Table or figure
Raw output, processed result, rendered table
Support
Citation or paragraph
Claim-to-evidence link
Research history
Usually omitted
Exploration graph
Reproduction
Manual interpretation
Executable recipe and report
Definitions
Prose
Machine-readable concept objects
The main shift is from reading to operating. A PDF is a document to interpret. An ARA is a research object to inspect, run, check, and modify.
Summary
Converting a PDF paper into an ARA means extracting the hidden structure behind the paper and representing it in a form that an agent can use.
The important changes are:prose claims → structured claims method descriptions → executable commands tables and figures → linked raw outputs final story → exploration graph informal detail → explicit specifications paper review → reproducibility checks
The PDF remains useful for human reading. The ARA provides the operational layer.
In one sentence: ARA turns a paper from a document into a structured research package that an agent can inspect, execute, verify, and extend.





