Skip to content

Nnunet

nnunet #

Functions:

Name Description
generate_dataset_json

Generates a dataset.json file in the output folder

generate_nnunet_scripts

Creates two bash scripts:

generate_dataset_json #

generate_dataset_json(
    output_folder: pathlib.Path | str,
    channel_names: typing.Dict[str, str],
    labels: typing.Dict[str, int | list[int]],
    num_training_cases: int,
    file_ending: str,
    regions_class_order: (
        typing.Tuple[int, ...] | None
    ) = None,
    dataset_name: str | None = None,
    reference: str | None = None,
    release: str | None = None,
    usage_license: str = "hands off!",
    description: str | None = None,
    overwrite_image_reader_writer: str | None = None,
    **kwargs: object
) -> None

Generates a dataset.json file in the output folder

Code from: https://github.com/MIC-DKFZ/nnUNet/blob/master/nnunetv2/dataset_conversion/generate_dataset_json.py

channel_names: Channel names must map the index to the name of the channel, example: { 0: 'T1', 1: 'CT' } Note that the channel names may influence the normalization scheme!! Learn more in the documentation.

labels: This will tell nnU-Net what labels to expect. Important: This will also determine whether you use region-based training or not. Example regular labels: { 'background': 0, 'left atrium': 1, 'some other label': 2 } Example region-based training: { 'background': 0, 'whole tumor': (1, 2, 3), 'tumor core': (2, 3), 'enhancing tumor': 3 }

Remember that nnU-Net expects consecutive values for labels! nnU-Net also expects 0 to be background!

num_training_cases: is used to double check all cases are there!

file_ending: needed for finding the files correctly. IMPORTANT! File endings must match between images and segmentations!

dataset_name, reference, release, license, description: self-explanatory and not used by nnU-Net. Just for completeness and as a reminder that these would be great!

overwrite_image_reader_writer: If you need a special IO class for your dataset you can derive it from BaseReaderWriter, place it into nnunet.imageio and reference it here by name

kwargs: whatever you put here will be placed in the dataset.json as well

Source code in src/imgtools/utils/nnunet.py
def generate_dataset_json(
    output_folder: pathlib.Path | str,
    channel_names: Dict[str, str],
    labels: Dict[str, int | list[int]],
    num_training_cases: int,
    file_ending: str,
    regions_class_order: Tuple[int, ...] | None = None,
    dataset_name: str | None = None,
    reference: str | None = None,
    release: str | None = None,
    usage_license: str = "hands off!",
    description: str | None = None,
    overwrite_image_reader_writer: str | None = None,
    **kwargs: object,
) -> None:
    """
    Generates a dataset.json file in the output folder

    Code from: https://github.com/MIC-DKFZ/nnUNet/blob/master/nnunetv2/dataset_conversion/generate_dataset_json.py

    channel_names:
        Channel names must map the index to the name of the channel, example:
        {
            0: 'T1',
            1: 'CT'
        }
        Note that the channel names may influence the normalization scheme!! Learn more in the documentation.

    labels:
        This will tell nnU-Net what labels to expect. Important: This will also determine whether you use region-based training or not.
        Example regular labels:
        {
            'background': 0,
            'left atrium': 1,
            'some other label': 2
        }
        Example region-based training:
        {
            'background': 0,
            'whole tumor': (1, 2, 3),
            'tumor core': (2, 3),
            'enhancing tumor': 3
        }

        Remember that nnU-Net expects consecutive values for labels! nnU-Net also expects 0 to be background!

    num_training_cases: is used to double check all cases are there!

    file_ending: needed for finding the files correctly. IMPORTANT! File endings must match between images and
    segmentations!

    dataset_name, reference, release, license, description: self-explanatory and not used by nnU-Net. Just for
    completeness and as a reminder that these would be great!

    overwrite_image_reader_writer: If you need a special IO class for your dataset you can derive it from
    BaseReaderWriter, place it into nnunet.imageio and reference it here by name

    kwargs: whatever you put here will be placed in the dataset.json as well

    """

    has_regions: bool = any(
        [isinstance(i, (tuple, list)) and len(i) > 1 for i in labels.values()]
    )
    if has_regions:
        assert regions_class_order is not None, (
            "You have defined regions but regions_class_order is not set. "
            "You need that."
        )

    # Construct the dataset JSON structure
    dataset_json = {
        "channel_names": channel_names,
        "labels": labels,
        "numTraining": num_training_cases,
        "file_ending": file_ending,
        "name": dataset_name,
        "reference": reference,
        "release": release,
        "licence": usage_license,
        "description": description,
        "overwrite_image_reader_writer": overwrite_image_reader_writer,
        "regions_class_order": regions_class_order,
    }

    dataset_json = {k: v for k, v in dataset_json.items() if v is not None}

    dataset_json.update(kwargs)

    output_path = pathlib.Path(output_folder) / "dataset.json"
    with output_path.open("w") as f:
        json.dump(dataset_json, f, indent=4, sort_keys=False)

generate_nnunet_scripts #

generate_nnunet_scripts(
    output_directory: str | pathlib.Path, dataset_id: int
) -> None

Creates two bash scripts: 1. nnunet_preprocess.sh for running nnUNet preprocessing. 2. nnunet_train.sh for running nnUNet training.

Parameters: - output_directory (str): The directory where the output and subdirectories are located. - dataset_id (int): The ID of the dataset to be processed.

Source code in src/imgtools/utils/nnunet.py
def generate_nnunet_scripts(
    output_directory: str | pathlib.Path, dataset_id: int
) -> None:
    """
    Creates two bash scripts:
    1. `nnunet_preprocess.sh` for running nnUNet preprocessing.
    2. `nnunet_train.sh` for running nnUNet training.

    Parameters:
    - output_directory (str): The directory where the output and subdirectories are located.
    - dataset_id (int): The ID of the dataset to be processed.
    """
    # Define paths using pathlib
    output_directory = pathlib.Path(output_directory).resolve()

    # Paths for the scripts
    preprocess_shell_path = output_directory / "nnunet_preprocess.sh"
    train_shell_path = output_directory / "nnunet_train.sh"
    base_dir = output_directory.parent.parent

    # Remove any existing script files before creating new ones
    if preprocess_shell_path.exists():
        preprocess_shell_path.unlink()
    if train_shell_path.exists():
        train_shell_path.unlink()

    # Preprocessing script content
    preprocess_script_content = f"""#!/bin/bash
set -e

export nnUNet_raw="{base_dir}/nnUNet_raw"
export nnUNet_preprocessed="{base_dir}/nnUNet_preprocessed"
export nnUNet_results="{base_dir}/nnUNet_results"

# Preprocessing command for dataset {dataset_id}
nnUNetv2_plan_and_preprocess -d {dataset_id} --verify_dataset_integrity -c 3d_fullres
"""

    # Write the preprocessing script
    with preprocess_shell_path.open("w", newline="\n") as f:
        f.write(preprocess_script_content)

    # Make script executable
    preprocess_shell_path.chmod(preprocess_shell_path.stat().st_mode | 0o111)

    # Training script content
    train_script_content = f"""#!/bin/bash
set -e

export nnUNet_raw="{base_dir}/nnUNet_raw"
export nnUNet_preprocessed="{base_dir}/nnUNet_preprocessed"
export nnUNet_results="{base_dir}/nnUNet_results"

# Training loop
for (( i=0; i<5; i++ ))
do
    nnUNetv2_train {dataset_id} 3d_fullres $i
done
"""

    # Write the training script
    with train_shell_path.open("w", newline="\n") as f:
        f.write(train_script_content)

    # Make script executable
    train_shell_path.chmod(train_shell_path.stat().st_mode | 0o111)