Artifacts Reference¶
OneEHR writes all experiment outputs to a structured run directory under {output.root}/{output.run_name}/.
Run Directory Layout¶
{output.root}/{output.run_name}/
manifest.json
preprocess/
binned.parquet
labels.parquet
split.json
static.parquet (when static.csv is provided)
feature_schema.json (code-to-column mapping)
obs_mask.parquet (observation mask before imputation)
fitted_pipeline.pt (fitted preprocessing pipeline)
train/
{model_name}/
checkpoint.ckpt
meta.json
test/
predictions.parquet
metrics.json
analyze/
comparison.json
feature_importance.json
fairness.json
calibration.json
statistical_tests.json
missing_data.json
figures/ (when `oneehr plot` is run)
roc.png / roc.pdf
pr.png / pr.pdf
...
manifest.json¶
The single source of truth for the run, written by oneehr preprocess and read by all downstream commands.
Key fields:
| Field | Description |
|---|---|
config |
Full experiment config snapshot |
feature_columns |
Dynamic feature column names from binning |
static_feature_columns |
Static feature column names (if static.csv provided) |
paths |
Relative paths to preprocess artifacts |
Preprocess Artifacts¶
Written by oneehr preprocess under preprocess/.
binned.parquet¶
Binned dynamic events in long format. Columns include patient_id, bin_time, and generated num__* / cat__* features.
labels.parquet¶
Processed labels.
- Patient mode:
patient_id,label - Time mode:
patient_id,bin_time,label,mask
split.json¶
Patient-level split definition.
Fields:
train_patientsval_patientstest_patients
static.parquet¶
Encoded static feature matrix keyed by patient_id. Only present when [dataset].static is provided.
feature_schema.json¶
Maps each original clinical code to its encoded feature columns. Each entry contains the code name, the list of generated column names (num__* or cat__*), and the number of columns (dim). Used by models that group features by code (e.g. SAFARI, PRISM).
obs_mask.parquet¶
Boolean observation mask with the same shape as the feature columns in binned.parquet, computed before any imputation. A 1.0 indicates the value was actually observed; 0.0 indicates it was missing. Used by models that need explicit missingness indicators (e.g. PAI).
fitted_pipeline.pt¶
Serialized FittedPipeline object saved via torch.save. Contains fitted statistics (means, stds, quantiles, etc.) from the train split only. Always created during preprocessing — empty when no pipeline steps are configured. Loaded and applied by oneehr train and oneehr test to transform data before modeling.
Train Artifacts¶
Written by oneehr train under train/{model_name}/.
checkpoint.ckpt¶
Serialized model checkpoint saved via torch.save (for all model types including tabular).
meta.json¶
Model metadata used to rebuild the model.
Key fields:
| Field | Description |
|---|---|
model_name |
Model config name (e.g. "xgboost", "gru") |
params |
Model hyperparameters from config |
train_metrics |
Metrics computed during training (includes history for DL models) |
feature_columns |
Feature columns the model was trained on |
For DL models, train_metrics.history contains a list of per-epoch dictionaries with train_loss, val_loss, and the monitored metric (e.g. val_auroc) when monitor != "val_loss".
Test Artifacts¶
Written by oneehr test under test/.
predictions.parquet¶
Unified predictions from all trained models and configured systems.
Columns:
| Column | Description |
|---|---|
system |
System/model name |
patient_id |
Patient identifier |
y_true |
Ground truth label |
y_pred |
Predicted value |
metrics.json¶
Aggregated test metrics per system.
Analyze Artifacts¶
Written by oneehr analyze under analyze/.
comparison.json¶
Cross-system comparison metrics. Contains per-system metric breakdowns computed from predictions.parquet.
Key fields:
| Field | Description |
|---|---|
module |
"comparison" |
task |
Task kind and prediction mode |
systems[] |
Per-system metrics (name, n, metrics dict) |
feature_importance.json¶
Feature importance results per trained model.
Key fields:
| Field | Description |
|---|---|
module |
"feature_importance" |
models |
Per-model importance (method, features, importances) |