This repository presents a deep learning-based tool to quantify subject motion in T1-weighted brain MRI.
The model used by this tool can be trained using our research code.
You will need an environment with at least Python 3.11. Then run:
pip install agitation
Alternatively, you can clone the repository and use:
python cli.py
instead of agitation
.
We use a TorchScript version of our best model.
All model checkpoints and the final TorchScript file are available on Zenodo.
The model will be downloaded automatically when needed. However, you can also manually download it with:
agitation manage check
The model is stored in your application data directory. You can retrieve the exact location using the check
command.
To remove all downloaded data:
agitation manage delete
Our model was trained on data preprocessed with Clinica's T1-linear pipeline.
While it may work with any T1-weighted MRI, we strongly recommend using the same preprocessing pipeline to ensure consistent results.
To quantify motion on a full dataset, use the command:
agitation dataset
-d, --dataset
: Path to the root of the dataset. It must be organized according to BIDS or CAPS (Clinica) standards and contain either ananat
folder ort1_linear
for CAPS.-f, --file
: Path to a CSV file describing the data to process. The file must contain at least adata
column specifying the path to each volume. Other columns will be copied to the output CSV.-g, --gpu
: Flag to enable GPU inference.--cuda
: Specify the GPU index to use (defaults to 0).-o, --output
: Path to the output CSV file.
agitation dataset --dataset <path_to_root> -g --output <path_to_output_file>
agitation dataset --file <path_to_csv>
To quantify motion at the subject level, use the command:
agitation inference
This entry point is recommended for use with pipeline tools like Nipoppy.
--bids_dir
: Path to the root of a BIDS dataset.--subject_id
: Subject identifier in BIDS format:sub-<label>
.--session_id
: Session identifier in BIDS format:ses-<label>
.-g, --gpu
: Flag to enable GPU inference.--cuda
: Specify the GPU index to use (defaults to 0).--output_dir
: Path to the output directory.
agitation inference --bids_dir tests/data/bids_sub_ses --subject_id sub-000103 --session_id ses-standard --output_dir ./
Our tool offers a Boutiques descriptor, available in the descriptors
folder, and a container image (container definition in containers
) available on Dockerhub.
A basic invocation looks like this :
{
"bids_dir": "tests/data/bids_sub_ses",
"output_dir": "./",
"subject_id": "sub-000103",
"session_id": "ses-headmotion2",
"gpu":true
}
Which would be called using Boutiques's bosh
command :
bosh exec launch descriptors/agitation.json <path_to_invocation>
To integrate our tool into a Nipoppy dataset, copy the content from descriptors/nipoppy
to your dataset's pipelines/agitation-<version>/
. The original descriptors/agitation.json
file works for Boutique's bosh
but not for Nipoppy.
You will also need to add the following pipeline in your global_config.json
:
{
"NAME": "agitation",
"VERSION": "0.0.2",
"CONTAINER_INFO": {
"FILE": "[[NIPOPPY_DPATH_CONTAINERS]]/[[PIPELINE_NAME]]_[[PIPELINE_VERSION]].sif",
"URI": "docker://chbricout/[[PIPELINE_NAME]]:[[PIPELINE_VERSION]]"
},
"CONTAINER_CONFIG": {
"ARGS": [
"--nv"
]
},
"STEPS": [
{
"INVOCATION_FILE": "[[NIPOPPY_DPATH_PIPELINES]]/[[PIPELINE_NAME]]-[[PIPELINE_VERSION]]/invocation.json",
"DESCRIPTOR_FILE": "[[NIPOPPY_DPATH_PIPELINES]]/[[PIPELINE_NAME]]-[[PIPELINE_VERSION]]/descriptor.json"
}
]
}
The --nv
is necessary only if you keep "gpu":true
in you invocation.json
it is used to allow access to Nvidia GPUs inside Apptainer.
The agitation
package can also be used as a library to include motion estimation in your projects.
To manually download the model within your code:
from agitation.data_manager import download_model
download_model()
To run inference on a dataloader:
from monai.data.dataset import Dataset
from torch.utils.data import DataLoader
from agitation.inference import estimate_motion_dl
from agitation.processing import LoadVolume
# Example usage
dataset = Dataset(<your_data_as_a_dict>, transform=LoadVolume())
dataloader = DataLoader(dataset)
estimate_motion_dl(dataloader, cuda=0)
To perform inference on a single batch:
import torch
from agitation.config import MODEL_PATH
from agitation.processing import SoftLabelToPred
# Dataloading, cropping, and normalization steps
model = torch.jit.load(
MODEL_PATH,
map_location="cuda:0" # If using CUDA
)
converter = SoftLabelToPred()
with torch.inference_mode():
prediction = model(data).cpu()
motions = converter(prediction)
Once the repository is cloned, install the development dependencies with:
pip install -r dev_requirements.txt
We use:
pytest
for unit testspytest-cov
for coverage reports (targeting 100% test coverage)
Run tests via:
pytest --cov
Other tools:
ruff
for linting and formatting (automatically applied viapre-commit
)- Additional code quality tools:
ssort
,pydocstyle
,mypy
, andpylint
All test data are extracted from MR-ART:
Nárai, Á., Hermann, P., Auer, T. et al. Movement-related artefacts (MR-ART) dataset of matched motion-corrupted and clean structural MRI brain scans. Sci Data 9, 630 (2022). https://doi.org/10.1038/s41597-022-01694-8
To fully deploy a new version, follow these steps in order:
- Build and deploy the PyPI package (used for the Apptainer image).
- Build the Apptainer image and publish it to Docker Hub.
- Publish any modifications to the Boutiques descriptor on Zenodo.
Build the package using:
python -m build
Deploy to PyPI with:
twine upload dist/*
Build the container using:
docker build -t chbricout/agitation:<version> .
You can do a test run with :
docker run --rm agitation:<version> agitation --help
Publish with:
docker push chbricout/agitation:<version>
Publish to Zenodo using:
bosh publish --sandbox -y --zenodo-token <ZENODO TOKEN> descriptors/agitation.json