Ready-to-use workflow template for evaluating Docker images submitted on Synapse.org
The model-to-data (m2d) workflow is ideal for scenarios where participants train their algorithm locally using provided training data. Participants will then submit their algorithm as a containerized Docker image to be run against the hold-out validation/test data (which they will have no access to) in order to generate a predictions file, which is then evaluated against the hidden groundtruth data.
- Customize evaluation logic: modify the scoring and validation scripts
within the
evaluationfolder - Configure workflow: adapt
workflow.cwl(andwriteup-workflow.cwl, if applicable) to define the inputs and steps specific to your challenge - Test your changes: use
cwltoolto test your CWL scripts within thestepsfolder
This template provides all necessary components for a full challenge pipeline:
.
├── evaluation // core scoring and validation scripts
├── README.md
├── scripts // scripts called by the individual CWL scripts in `steps`
├── steps // individual CWL scripts (called by the main workflow CWLs)
├── workflow.cwl // CWL workflow for evaluating submissions
└── writeup-workflow.cwl // CWL workflow to validate and archive writeup submissions
This template is built using CWL and Docker, and is designed to be handled by the SynapseWorkflowOrchestrator orchestration tool. For detailed information on utilizing these core technologies, please refer to their docs below:
- CWL: https://www.commonwl.org/user_guide/
- Docker: https://docs.docker.com/get-started/
- SynapseWorkflowOrchestrator: https://github.com/Sage-Bionetworks/SynapseWorkflowOrchestrator