Skip to content

Commit ffcdb00

Browse files
author
MarcelRobeer
committed
Replaced /
1 parent 5fc760e commit ffcdb00

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

paper/paper.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ bibliography: paper.bib
3838

3939
# Summary
4040

41-
Developed to meet the practical machine learning (ML) auditing requirements of the Netherlands National Police, `Explabox` is an open-source Python toolkit that implements a standardized four-step analysis workflow: *explore*, *examine*, *explain* and *expose*. The framework transforms models and data (*ingestibles*) into interpretable reports and visualizations (*digestibles*), covering everything from data statistics and performance metrics to local/global explanations, and sensitivity testing for fairness, robustness and security. Designed for developers, testers, and auditors, `Explabox` operationalizes the entire audit lifecycle in a reproducible manner. The initial release is focused on text classification and regression models, with plans for future expansion. Code and documentation are available open-source at [https://explabox.readthedocs.io](https://explabox.readthedocs.io/en/stable).
41+
Developed to meet the practical machine learning (ML) auditing requirements of the Netherlands National Police, `Explabox` is an open-source Python toolkit that implements a standardized four-step analysis workflow: *explore*, *examine*, *explain* and *expose*. The framework transforms models and data (*ingestibles*) into interpretable reports and visualizations (*digestibles*), covering everything from data statistics and performance metrics to local and global explanations, and sensitivity testing for fairness, robustness and security. Designed for developers, testers, and auditors, `Explabox` operationalizes the entire audit lifecycle in a reproducible manner. The initial release is focused on text classification and regression models, with plans for future expansion. Code and documentation are available open-source at [https://explabox.readthedocs.io](https://explabox.readthedocs.io/en/stable).
4242

4343
# Statement of need
4444

@@ -53,7 +53,7 @@ This fragmentation introduces significant challenges, particularly regarding *re
5353
`Explabox` transforms opaque *ingestibles* into transparent *digestibles* through four types of *analyses* to enhance explainability and aid fairness, robustness, and security audits.
5454

5555
## Ingestibles
56-
Ingestibles provide a unified model/data import interface, where layers abstract away access (\autoref{fig:layers}) to allow optimized processing. `Explabox` uses `instancelib` [@instancelib] for fast model/data encapsulation. The model can be any Python `Callable` containing a regression or (binary and multi-class) classification model. While this interface is model-agnostic, the current release provides data handling and analysis modules optimized specifically for text-based tasks. `scikit-learn` or `ONNX` models (e.g., `PyTorch`, `TensorFlow`, or `Keras`) import directly with optimizations and automatic input/output interpretation. Data can be automatically downloaded, extracted and loaded. Data inputs include `NumPy`, `Pandas`, `Hugging Face`, raw files (e.g., HDF5, CSV or TSV), and (compressed) file folders. Data can be subdivided into named splits (e.g., train-test-validation), and instance vectors and tokens can be precomputed and optionally saved for fast inferencing.
56+
Ingestibles provide a unified import interface for data and models, where layers abstract away access (\autoref{fig:layers}) to allow optimized processing. `Explabox` uses `instancelib` [@instancelib] for fast model and data encapsulation. The model can be any Python `Callable` containing a regression or (binary and multi-class) classification model. While this interface is model-agnostic, the current release provides data handling and analysis modules optimized specifically for text-based tasks. `scikit-learn` or `ONNX` models (e.g., `PyTorch`, `TensorFlow`, or `Keras`) import directly with optimizations and automatic input/output interpretation. Data can be automatically downloaded, extracted and loaded. Data inputs include `NumPy`, `Pandas`, `Hugging Face`, raw files (e.g., HDF5, CSV or TSV), and (compressed) file folders. Data can be subdivided into named splits (e.g., train-test-validation), and instance vectors and tokens can be precomputed and optionally saved for fast inferencing.
5757

5858
![Logical separation of `Explabox` into layers with interfaces.\label{fig:layers}](figure1.png){width=50%}
5959

@@ -71,7 +71,7 @@ Ingestibles provide a unified model/data import interface, where layers abstract
7171

7272
## Digestibles
7373

74-
Digestibles serve stakeholders—such as creators, auditors, applicants, end-users or clients [@Tomsett2018]—via a Jupyter/web UI (\autoref{fig:ui}) (using `plotly` [@plotly] visuals), integrated API, and static reporting.
74+
Digestibles serve stakeholders—such as creators, auditors, applicants, end-users or clients [@Tomsett2018]—via a Jupyter Notebook or Web UI (\autoref{fig:ui}) (using `plotly` [@plotly] visuals), integrated API, and static reporting.
7575

7676
![UI elements from the Jupyter Notebook interface, designed to present audit results to diverse stakeholders.\label{fig:ui}](figure2.png)
7777

0 commit comments

Comments
 (0)