Creating an Eval
How to create an eval on the Datalab.
Portex makes it easy to monetize your expert domain knowledge through custom evaluations or "evals" that you can offer to model builders on the Datalab. Follow this guide to get started with monetizing evals on the Datalab.
Creating an Eval Dataset Bundle
To get started, create a new dataset and choose Eval Dataset
- from here you will be prompted to upload your files. Eval datasets on Portex are uploaded as a bundle of 4 files (2 optional). This forms a Core Dataset which can be offered to sale to model builders in addition to per-eval runs. We walk through file in the Core Dataset each below.

Formatting your Eval Dataset Bundle
There are four files in an Eval Dataset Bundle. A Task List, Reference Files (Optional), Answer Key, and Knowledge Reference (Optional). Below is a description of each dataset and how to structure it.
🧩 Task List
A JSON file containing your tasks. Each record must include:
task_id
: unique identifiertask_prompt
: the question for the modelIf you’re referencing files (e.g. a 10-K PDF), include
task_reference_file
Optional fields (you may include other metadata fields, just make sure you cover the three above):
skills_tested
: metadata about the specific skills tested by the taskdifficulty
: an assessed difficulty of the questionquestion_type
: e.g. multiple choice, exact answer, summary
Example:
[
{
"task_id": "apple_net_margin_2024",
"task_prompt": "Using Apple’s FY 2024 10-K, compute its net profit margin (net income ÷ revenue).",
"task_reference_file": "apple_2024_10K.html"
}
]
This file will be downloadable by eval buyers so they can generate responses.
📂 Reference Files (Optional)
If your tasks refer to documents, images, or other supporting files, bundle them into a .gz archive. This is only needed if you have reference files that accompany your tasks list.
Example archive structure:
reference_files/
apple_2024_10K.html
revenue_breakdown_chart.png
Buyers will extract this archive to access the referenced files.
🧠 Answer Keys
A JSON file mapping task_id to the correct output. Must include:
task_id
answer
: the expected result (numerical, textual, set of objective criteria etc.)
Optional fields:
rationale
: explanation of how the answer was derivedanswer_type
: e.g. "percentage", "multiple_choice", "criteria"knowledge_reference_file
: filename from the Knowledge Reference archive that supports the answer e.g. expert notes or chain-of-thought steps
Example:
[
{
"task_id": "apple_net_margin_2024",
"answer": 0.2397,
"rationale": "Net income $93,736M ÷ Revenue $391,035M = 23.97%",
"answer_type": "percentage",
"knowledge_reference_file": "expert_notes.csv"
}
]
The answer key will be available for purchase in the core dataset price.
📘 Knowledge Reference (Optional)
A .gz archive containing additional files that support your reasoning or calculations (analyst notes, data tables, or expert commentary).
Example archive structure:
knowledge_reference/
expert_notes.csv
margin_calculation_factors.csv
Buyers can use these to validate or audit the answer logic, or even refine their models using reinforcement learning methods. This will be offered for sale within the core dataset pricing.
Creating an Eval Listing
Once you've uploaded your eval dataset, you can create a listing for it.
The first step will be setting two prices:
A per-eval price: a set price model builders pay each time they submit model responses and receive a performance report.
Core Dataset price and minimum bid: a fixed "buy now" price and minimum bid for model builders to access the Core Dataset (tasks and answers, as well as reference files and knowledge reference if applicable).

The next step is configuring your license and all other relevant details on the listing editor. Once you publish your listing, your eval will be ready for model builders to access.
Last updated