Skip to main content
Version: v1.2.0

Experiment settings: Image object detection

The settings for an image object detection experiment are listed and described below.

General settings

Dataset

It defines the dataset for the experiment.

Problem type

Defines the problem type of the experiment, which also defines the settings H2O Hydrogen Torch displays for the experiment.

Note
  • The selected problem type and experience level determine the settings H2O Hydrogen Torch displays for the experiment
  • The From experiment option allows you to use the settings from a previously run experiment

Model type

It defines the model type for the experiment.

Note
  • For an image object detection experiment H2O Hydrogen Torch supports the following model types:
  • The selected problem type, experience level, and model type determine the settings H2O Hydrogen Torch displays for the experiment. Note that not all problem types support the specification of a model type; in that case, the selected problem type and experience level determine the settings H2O Hydrogen Torch displays for the experiment.

Efficientdet

EfficientDet models are among the most popular models to tackle image object detection. They are using EfficientNet models as a backbone and a weighted bi-directional feature pyramid network (BiFPN) as the feature network.

note

EfficientDet is the default model type for image object detection in H2O Hydrogen Torch. To learn more about EfficientDet, see EfficientDet: Scalable and efficient object detection.

Faster Rcnn

Faster Region-based Convolutional Neural Networks (FasterRCNN) is an advancement of classical Region-based Convolutional Neural Networks (RCNN) architectures, so-called region-based convolutional neural networks. The core idea is to apply selective search to extract regions of interest from an image, where each ROI might represent a bounding box of an object. Each region of interest (ROI) is fed through a neural network to produce output features used to classify the type of object. A FasterRCNN shares full-image convolutional features with the detection network and thus enables nearly cost-free region proposals, significantly improving the training and inference process compared to classical RCNN or Fast RCNN networks.

note

Fcos

Both EfficientDet and FasterRCNN are so-called anchor-based object detection models. A fully convolutional one-stage object detector (FCOS) is a fully convolutional one-stage object detector to solve object detection per pixel. Similar to how semantic segmentation models operate. FOCS is anchor box and proposal free.

note

Import config from YAML

Defines the .yml file that defines the experiment settings.

Note
  • H2O Hydrogen Torch supports a .yml file import and export functionality. You can download the config settings of finished experiments, make changes, and re-upload them when starting a new experiment in any instance of H2O Hydrogen Torch.

Experiment name

It defines the name of the experiment.

Dataset settings

Train dataframe

Defines a .csv or .pq file containing a dataframe with training records that H2O Hydrogen Torch will use to train the model.

note
  • The records will be combined into mini-batches when training the model.
  • If a validation dataframe is provided, a fold column is not needed in the train dataframe.

Data folder

Defines the folder location of the assets (e.g., images or audio clips) the model utilizes for training. H2O Hydrogen Torch loads assets from this folder during training.

Validation strategy

Specifies the validation strategy H2O Hydrogen Torch will use for the experiment.

note

To properly assess the performance of your trained models, it is common practice to evaluate it on separate holdout data that the model has not seen during training. H2O Hydrogen Torch allows you to specify different strategies for this task fitting your needs.

Options

  • K-fold cross validation

    Splits the data using the provided optional fold column in the train data or performs an automatic 5-fold cross-validation.

  • Grouped k-fold cross validation

    Allows to specify a group column based on which the data is split into folds.

  • Custom holdout validation:

    Specifies a separate holdout dataframe.

  • Automatic holdout validation

    Allows to specify a holdout validation sample size that is automatically generated.

Validation dataframe

Defines a .csv or .pq file containing a dataframe with validation records that H2O Hydrogen Torch will use to evaluate the model during training.

Note
  • To set a Validation dataframe requires the Validation strategy to be set to Custom holdout validation. In this case, H2O Hydrogen Torch will fully respect the choice of a separate validation dataframe and will not perform any internal cross-validation. In other words, the model is trained on the full provided train dataframe, and model performance is evaluated on the provided validation dataframe.
  • The validation dataframe should have the same format as the train dataframe but does not require a fold column.

Folds

Defines the validation folds in case of cross-validation; a separate model is trained for each value selected. Each model will use the corresponding part of the data as a holdout sample to assess performance while the model is fitted to the rest of the records from the training dataframe. As a result, folds estimate how the model will perform in general when used to make predictions on data not used during model training.

Note
  • If a column with the name fold is present in the train dataframe, H2O Hydrogen Torch will use the fold column values for folding; otherwise, a simple 5-fold (K-fold) will be applied.
  • H2O Hydrogen Torch allows running experiments on single folds for faster experimenting and multiple folds to gain more trust in the model's generalization and performance capabilities.
  • The Folds setting will only be available if Custom holdout validation is not selected as the Validation strategy.

Test dataframe

Defines a .csv or .pq file containing a dataframe with test records that H2O Hydrogen Torch will use to test the model.

note

The test dataframe should have the same format as the train dataframe but does not require a label column.

Data folder test

Defines the folder location of the assets (e.g., images or texts) H2O Hydrogen Torch will use to test the model. H2O Hydrogen Torch will load the assets from this folder when testing the model. This setting is only available if a test dataframe is selected.

Note

The Data Folder Test setting appears when you specify a test dataframe in the Test Dataframe setting.

Unlabeled dataframe

Defines a separate .csv or .pq file containing a dataframe with unlabeled records that H2O Hydrogen Torch uses to generate pseudo labels. H2O Hydrogen Torch first trains the model with the provided labeled data (Train dataframe). Right after, the model predicts pseudo labels for the data in the provided unlabeled dataframe before doing another training run that combines the original labels and pseudo labels.

note
  • Image regression | Image classification | Image object detection
    • The unlabeled dataframe just needs to contain a single image column
  • Text regression | Text classification
    • The unlabeled dataframe just needs to contain a single text column
  • Audio regression | Audio classification | Speech recognition
    • The unlabeled dataframe just needs to contain a single audio column
  • Image regression | Image classification | Image object detection | Audio regression | Audio classification | Speech recognition
    • Assets (e.g., images or audio) need to be located in the Data folder (setting)
  • The training time can significantly increase depending on the size of the unlabeled data
tip

As labeling can be expensive, having additional unlabeled data is quite common. You providing this unlabeled data in H2O Hydrogen Torch trains the model in a semi-supervised manner, potentially improving the model quality in contrast to only training on labeled data.

Class name column

Defines the dataset column containing a list of class names that H2O Hydrogen Torch will use for each instance mask.

X min column

Defines the dataset column containing a list of minimum X positions H2O Hydrogen Torch will use for each bounding box.

Y min column

Defines the dataset column containing a list of minimum Y positions H2O Hydrogen Torch will use for each bounding box.

X max column

Defines the dataset column containing a list of maximum X positions H2O Hydrogen Torch will use for each bounding box.

Y max column

Defines the dataset column containing a list of maximum Y positions H2O Hydrogen Torch will use for each bounding box.

Image column

Defines the dataframe column storing the names of images that H2O Hydrogen Torch will load from the data folder and data folder test when training and testing the model.

Data sample

Modifies the percentage of the data to use for the experiment. The default percentage is 100% (1).

note

Changing the default value can significantly increase the training speed. Still, it might lead to a substantially poor accuracy value. Using 100% of the data for final models is highly recommended.

Image settings

Image width

Defines the width H2O Hydrogen Torch will use to rescale the images for training and predictions.

note

Depending on the original image size, a bigger width can generate a higher accuracy value.

Image height

Defines the width H2O Hydrogen Torch will use to rescale the images for training and predictions.

note

Depending on the original image size, a bigger width can generate a higher accuracy value.

Image channels

Defines the number of channels the train images contain.

note
  • Typically images have three input channels (red, green, and blue (RGB)), but grayscale images have only 1. When you provide image data in a NumPy data format, any number of channels is allowed. For this reason, data scientists can specify the number of channels.
  • The defined number of channels will also refer to the provided validation and test datasets.

Image normalization

Grid search hyperparameter

Defines the transformer to normalize the image data before training the model.

note

Usually, state-of-the-art image models normalize the training images by scaling values of each of the input channels to predefined means and standard deviations.

Options

  • Channel

    Calculates mean and standard deviation per channel in all the images in the batch and then applies per channel normalization: subtracts mean and divides by standard deviation.

  • Image

    Calculates mean and standard deviation per image and then applies normalization.

  • ImageNet

    Divides input images by 255 and normalizes with mean and standard deviation equal to (0.485, 0.456, 0.406) and (0.229, 0.224, 0.225) per channel, respectively.

  • Inception

    Divides input images by 255 and normalizes with mean and standard deviation equal to 0.5.

  • Min_Max

    Calculates minimum and maximum values in all the images in the batch and then applies min-max normalization: subtracts min and divides by the max and min difference.

  • No

    No normalization is applied to the input images.

  • Simple

    Divides input images by 255.

note

Usually, state-of-the-art image models normalize the training images by scaling values of each of the input channels to predefined means and standard deviations.

Augmentation settings

Augmentations strategy

Grid search hyperparameter

Defines the augmentation strategy to apply to the input images. Soft, Medium, and Hard values correspond to the strength of the augmentations to apply.

Options

  • Soft: The Soft strategy applies image Resize and random HorizontalFlip during model training while applying image Resize during model inference.

  • Medium: The Medium strategy adds ShiftScaleRotate and CoarseDropout to the list of the train augmentations.

  • Hard: The Hard strategy applies RandomResizedCrop (instead of Resize) during model training while adding RandomBrightnessContrast to the list of train augmentations.

  • Custom: The Custom strategy allows users to use their own augmentations that can be defined in the following two settings:

note

Augmentations are ways to modify train images while keeping the target values valid, such as flipping the image or adding noise. Distorting training images do not influence the expected prediction of the model but enrich the training data. Augmentations help generalize the model better and improve its accuracy.

Mix image

Grid search hyperparameter

Defines the image mix augmentation to use during model training.

Options

  • Mixup

    Mixup overlays (mixes) two images one on another based on a random ratio. To learn more about this mix augmentation approach, refer to the following article: mixup: BEYOND EMPIRICAL RISK MINIMIZATION.

    note

    For an image object detection experiment using Mixup, H2O Hydrogen Torch uses the union of all the target boxes in mixed images.

  • Cutmix

    Cutmix replaces an image region with a patch from another image; the region size is based on a random ratio. To learn more about this mix augmentation approach, refer to the following article: SOLVING LINEAR SYSTEMS OVER TROPICAL SEMIRINGS THROUGH NORMALIZATION METHOD AND ITS APPLICATIONS.

    note

    For an image object detection experiment using Cutmix, H2O Hydrogen Torch uses the target boxes from the corresponding region from each image. Also, with Cutmix selected, H2O Hydrogen Torch cuts out and replaces only the corners of the images with a patch from another image.

  • Disabled

    No augmentation is applied.

Architecture settings

Pretrained

Defines whether the neural network should start with pre-trained weights. When this setting is On, the training of the neural network will start with a pre-trained model on a generic task. When turn Off, the initial weights of the neural network to train will be random.

Backbone

Grid search hyperparameter

Defines the backbone neural network architecture to train the model.

Note
  • Image regression | Image classification | Image metric learning | Audio regression | Audio classification
    • H2O Hydrogen Torch accepts backbone neural network architectures from the timm library (select or enter the architecture name).
  • Image object detection
    • H2O Hydrogen Torch provides several backbone state-of-the-art neural network architectures for model training. When you select Faster RCnn or Fcos as the model type for the experiment, you can input any architecture name from the timm library. When you select Efficientdet as the model type for the experiment, you can input any architecture name from the efficientdet-pytorch library.
  • Image semantic segmentation | Image instance segmentation
    • H2O Hydrogen Torch accepts backbone neural network architectures from the segmentation-models-pytorch library (select or enter the architecture name).
  • Text regression | Text classification | Text token classification | Text span prediction | Text sequence to sequence | Text metric learning
    • H2O Hydrogen Torch accepts backbone neural network architectures from the Hugging Face library (select or enter the architecture name).
  • Speech recognition
    • HuggingFace Wav2Vec2 CTC models are supported.
tip
  • Usually, it is good to use simpler architectures for quicker experiments and larger models when aiming for the highest accuracy.
  • Speech recognition
    • Leverage backbones that were pretrained as closely to your use-case if possible (e.g., noisy audio, casual speech etc).

Drop path rate

Defines the drop path rate for the Backbone to use during training. The drop path rate prevents co-adaptation of parallel paths in networks, similar to how dropout prevents co-adaption of activations. If set to Default, it will pick the default setting for the respective backbone.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Anchor num scales

Defines the number of anchor scales to use for each anchor box. You may want to change this to work with more fine-grained scales. Note that changing this setting will reset the head of the pre-trained model; in most use cases, it is recommended to use the default value.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Anchor scale

Defines the general scale factor for all anchor boxes; you may want to change this if your dataset contains a large amount of particularly small or large boxes.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Anchor aspect ratios

Defines the different anchor aspect ratios for anchor boxes; in the best case, the selected anchor aspect ratios should match the default shapes in the dataset. Note that changing this setting will reset the head of the pre-trained model: in most use cases, it is recommended to use the default value.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Anchor Iou match threshold

Defines the IoU threshold for matching anchor boxes. In particular, the IoU threshold is used to determine whether an anchor box matches a ground truth box.

example

If you set the Anchor IoU match threshold to 0.5, the anchor box will only match a ground truth box if the IoU is greater than 50%.

In other words, the IoU threshold determines positive labels for anchors.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Training settings

Optimizer

Grid search hyperparameter

Defines the algorithm or method (optimizer) to use for model training. The selected algorithm or method defines how the model should change the attributes of the neural network, such as weights and learning rate. Optimizers solve optimization problems and make more accurate updates to attributes to reduce learning losses.

Options

Learning rate

Grid search hyperparameter

Defines the learning rate H2O Hydrogen Torch will use when training the model, specifically when updating the neural network's weights. The learning rate is the speed at which the model updates its weights after processing each mini-batch of data.

note
  • Learning rate is an important setting to tune as it balances under- and overfitting.
  • The number of epochs highly impacts the optimal value of the learning rate.

Differential learning rate layers

Defines the learning rate to apply to certain layers of a model. H2O Hydrogen Torch applies the regular learning rate to layers without a specified learning rate.

Options

  • Backbone
    • H2O Hydrogen Torch applies a different learning rate to a body of the neural network architecture.
  • Head
    • H2O Hydrogen Torch applies a different learning rate to a head of the neural network architecture.
  • Neck
    • H2O Hydrogen Torch applies a different learning rate to a neck of the neural network architecture.
  • Loss
    • H2O Hydrogen Torch applies a different learning rate to an ArcFace block of the neural network architecture.
  • Encoder
    • H2O Hydrogen Torch applies a different learning rate to the encoder of the neural network architecture.
  • Decoder
    • H2O Hydrogen Torch applies a different learning rate to the decoder of the neural network architecture.
  • Segmentation head
    • H2O Hydrogen Torch applies a different learning rate to the head of the neural network architecture.
note

The options for an image object detection experiment are different based on the selected Model type (setting). Options:

  • If you select EfficientDet as the experiment's Model type (setting), the following options are available:

    Options

    • Backbone
      • H2O Hydrogen Torch applies a different learning rate to a body of the EfficientDet architecture.
    • FPN
      • H2O Hydrogen Torch applies a different learning rate to a Feature Pyramid Network (FPN) block of the EfficientDet architecture.
    • class_net
      • H2O Hydrogen Torch applies a different learning rate to a classification head of the EfficientDet architecture.
    • box_net
      • H2O Hydrogen Torch applies a different learning rate to a box regression head of the EfficientDet architecture.

  • If you select Faster R-CNN as the experiment's Model type (setting), the following options are available:

    Options

    • Body
      • H2O Hydrogen Torch applies a different learning rate to a body of the Faster R-CNN architecture.
    • FPN
      • H2O Hydrogen Torch applies a different learning rate to a Feature Pyramid Network (FPN) block in the Faster R-CNN architecture.
    • RPN
      • H2O Hydrogen Torch applies a different learning rate to a Region Proposal block of the Faster R-CNN architecture.
    • ROI heads
      • H2O Hydrogen Torch applies a different learning rate to the Faster R-CNN architecture proposal heads.

  • If you select FCOS as the experiment's Model type (setting), the following options are available:

    Options

    • Body
      • H2O Hydrogen Torch applies a different learning rate to a body of the FCOS architecture.
    • FPN
      • H2O Hydrogen Torch applies a different learning rate to a Feature Pyramid Network (FPN) block of the FCOS architecture.
    • classification_head
      • H2O Hydrogen Torch applies a different learning rate to the classification head of the FCOS architecture.
    • regression_head
      • H2O Hydrogen Torch applies a different learning rate to a box regression head of the FCOS architecture.

tip

A common strategy is to apply a lower learning rate to the backbone of a model for better convergence and training stability.

note

Different layers are available for different problem types.

Batch size

Grid search hyperparameter

Defines the number of training examples a mini-batch will use during an iteration of the training model to estimate the error gradient before updating the model weights. Batch size defines the batch size used per a single GPU.

note

During model training, the training data is packed into mini-batches of a fixed size.

Automatically adjust batch size

If this setting is turned On, H2O Hydrogen Torch will check whether the Batch size specified fits into the GPU memory. If a GPU out-of-memory (OOM) error occurs, H2O Hydrogen Torch will automatically decrease the Batch size by a factor of 2 units until it fits into the GPU memory or Batch size equals 1.

Drop last batch

H2O Hydrogen Torch drops the last incomplete batch during model training when this setting is turned On.

note

H2O Hydrogen Torch groups the train data into mini-batches of equal size during the training process, but the last batch can have fewer records than the others. Not dropping the last batch can lead to a less robust gradient estimation while causing a more volatile training step.

Epochs

Grid search hyperparameter

Defines the number of epochs to train the model. In other words, it specifies the number of times the learning algorithm will go through the entire training dataset.

note
  • The Epochs setting is an important setting to tune because it balances under- and overfitting.
  • The learning rate highly impacts the optimal value of the epochs.

Schedule

Grid search hyperparameter

Defines the learning rate schedule H2O Hydrogen Torch will use during model training. Specifying a learning rate schedule will prevent the learning rate from staying the same. Instead, a learning rate schedule will cause the learning rate to change over iterations, typically decreasing the learning rate to achieve a better model performance and training convergence.

Options

  • Constant

    H2O Hydrogen Torch applies a constant learning rate during the training process.

  • Cosine

    H2O Hydrogen Torch applies a cosine learning rate that follows the values of the cosine function.

  • Linear

    H2O Hydrogen Torch applies a linear learning rate that decreases the learning rate linearly.

Warmup epochs

Defines the number of epochs to warm up the learning rate where the learning rate should increase linearly from 0 to the desired learning rate.

Weight decay

Defines the weight decay that H2O Hydrogen Torch will use for the optimizer during model training.

note

Weight decay is a regularization technique that adds an L2 norm of all model weights to the loss function while increasing the probability of improving the model generalization.

Gradient clip

Defines the maximum norm of the gradients H2O Hydrogen Torch specifies during model training. Defaults to 0, no clipping. When a value greater than 0 is specified, H2O Hydrogen Torch will modify the gradients during model training. H2O Hydrogen Torch uses the specified value as an upper limit for the norm of the gradients, calculated using the Euclidean norm over all gradients per batch.

note

This setting can help model convergence when extreme gradient values cause high volatility of weight updates.

Grad accumulation

Defines the number of gradient accumulations before H2O Hydrogen Torch updates the neural network weights during model training.

note
  • Grad accumulation can be beneficial if only small batches are selected for training. With gradient accumulation, the loss and gradients are calculated after each batch, but it waits for the selected accumulations before updating the model weights. You can control the batch size through the Batch size setting.
  • Changing the default value of Grad Accumulation might require adjusting the learning rate and batch size.

Save best checkpoint

Determines if H2O Hydrogen Torch should save the model weights of the epoch exhibiting the best validation metric. When turned On, H2O Hydrogen Torch saves the model weights for the epoch exhibiting the best validation metric. When turned Off, H2O Hydrogen Torch saves the model weights after the last epoch is executed.

note
  • This setting should be turned On with care as it has the potential to lead to overfitting of the validation data.
  • The default goal should be to attempt to tune models so that the last or very last epoch is the best epoch.
  • Suppose an evident decline for later epochs is observed in logging. In that case, it is usually better to adjust hyperparameters, such as reducing the number of epochs or increasing regularization, instead of turning this setting On.

Evaluation epochs

Defines the number of epochs H2O Hydrogen Torch will use before each validation loop for model training. In other words, it determines the frequency (in a number of epochs) to run the model evaluation on the validation data.

note
  • Increasing the number of Evaluation Epochs can speed up an experiment.
  • The Evaluation epochs setting is available only if the following setting is turned Off: Save Best Checkpoint.

Calculate train metric

Determines whether the model metric should also be calculated for the training data at the end of the training. When On, the model metric will also be calculated for the training data. The resulting values will not indicate the true model performance because they will be based on H2O Hydrogen Torch's identical data records for model training but can give insights into over/underfitting.

Train validation data

Defines whether the model should use the entire train and validation dataset during model training. When turned On, H2O Hydrogen Torch will use the whole train dataset and validation data to train the model.

note
  • H2O Hydrogen Torch will also evaluate the model on the provided validation fold. Validation will always be only on the provided validation fold.
  • H2O Hydrogen Torch will use both datasets for model training if you provide a train and validation dataset.
    • To define a training dataset, use the Train dataframe setting. For more information, see Train dataframe.
    • To define a validation dataset, use the Validation dataframe setting. For more information, see Validation dataframe.
  • The Train validation data setting is only available if you turned the Save best checkpoint setting Off.
  • Turning On the Train validation data setting should produce a model that you can expect to perform better because H2O Hydrogen Torch trained the model on more data. Thought, also note that using the entire train dataset and out-of-fold validation dataset generally causes the model's accuracy to be overstated as information from the validation data is incorporated into the model during the training process.
    note

    If you have five folds and set fold 0 as validation, H2O Hydrogen Torch will usually train on folds 1-4 and report on fold 0. With Train validation data turned On, we can add fold 0 to the training, but H2O Hydrogen Torch will still report its accuracy. As a result, it will be overstated for fold 0 but should be better for any unseen (test) data/production scenarios. For that reason, you usually want to consider this setting after running your experiments and deciding on models.

Build scoring pipelines

Determines whether the experiment (model) automatically generates an H2O MLOps pipeline and Python scoring pipeline at the end of the experiment. If turned Off, you can still create scoring pipelines on demand when the experiment is complete (e.g., when you click Download soring or Download MLOps).

Box loss weight

Defines the weight of the box loss in EfficientDet (a type of object detection model); it is used to balance the loss of the bounding box regression and classification.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Focal Cls loss alpha

Defines the alpha hyperparameter value in the focal class loss function; for more information, refer to the following paper: Focal Loss for Dense Object Detection.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Focal Cls loss gamma

Defines the gamma hyperparameter value in the focal class loss function; for more information, refer to the following paper: Focal Loss for Dense Object Detection.

note

This setting is available when Efficientdet is selected as the model type for the experiment.

Prediction settings

Metric

Defines the metric to use to evaluate the model's performance.

note

Usually, the evaluation metric should reflect the quantitative way of assessing the model's value for the corresponding use case.

Metric Iou threshold

Defines the Intersection Over Union (IoU) threshold to calculate the selected metric for image object detection.

note

When calculating metrics, predicted bounding boxes with an IoU (with the true boxes) above the specified IoU threshold will be treated as true positives.

Nms Iou threshold

Defines the Intersection Over Union (IoU) threshold when calculating post-processing non-maximum suppression (NMS).

note

Non-maximum suppression (NMS) is a post-processing step that reduces the number of bounding boxes predicted by the model. The NMS algorithm will remove overlap boxes based on the selected IoU threshold. NMS will keep the higher scoring box.

Max det per image

Defines the maximum number of detections per image that the model will return.

Probability threshold

  • Image instance segmentation | Image semantic segmentation
    • Defines the probability threshold; a predicted pixel will be treated as positive if its probability is larger than the probability threshold.
  • Image object detection
    • Defines the probability threshold that the model utilizes to identify predicted bounding boxes with confidence larger than the defined probability threshold. Predicted bounding boxes above the defined probability threshold are added to the validation and test .csv files in the downloaded model predictions .zip file.
  • Audio classification | Image classification | Text classification
    • Define a threshold for threshold-dependent classification metrics (e.g. F1). For multi-class classification argmax will be used.
      note

      The defined threshold is used as a default threshold when displaying all other threshold-dependent metrics in the validation plots.

Environment settings

GPUs

Determines the list of GPUs H2O Hydrogen Torch can use for the experiment. GPUs are listed by name, referring to their system ID (starting from 1).

Number of GPUs per run

Defines the number of GPUs to use for a single run when training the model. A single run might represent a single fold or a single grid search run.

example

If 5 GPUs are available, it will be possible to run a 5-fold cross-validation in parallel using a single GPU per fold.

note
  • The available GPUs will be the ones that can be enabled using the GPUs setting.
  • If the number of GPUs is less than or equal to 1, this setting (Number of GPUs per run ) will not be available.

Mixed precision training

Determines whether to use mixed-precision during model training. When turned Off, H2O Hydrogen Torch will not use mixed-precision for training.

Note

Mixed-precision is a technique that helps decrease memory consumption and increases training speed.

Mixed precision inference

Determines whether to use mixed-precision during model inference.

note

Mixed-precision is a technique that helps decrease memory consumption and increases inference speed.

Sync batch normalization

Determines whether to synchronize batch normalization across GPUs in a distributed data-parallel (DDP) mode. In other words, when turned On, multi-GPU training is enabled to synchronize the batch normalization layers of the model across GPUs. In a nutshell, H2O Hydrogen Torch with multi GPU splits the batch across GPUs, and therefore, when a normalization layer wants to normalize data, it has access only to the part of the batch stored on the device. As a result, it will work out of the box but will give better results if the data in all GPUs is collected to normalize the data of the entire batch.

Note

When turned On, data scientists can expect the training speed to drop slightly while the model's accuracy improves. However, this rarely happens in practice and only occurs under specific problem types and defined batch sizes.

Number of workers

Defines the number of workers H2O Hydrogen Torch will use for the DataLoader. In other words, it defines the number of CPU processes to use when reading and loading data to GPUs during model training.

Seed

Defines the random seed value that H2O Hydrogen Torch will use during model training. It defaults to -1, an arbitrary value. When the value is modified (not -1), the random seed will allow results to be reproducible—defining a seed aids in obtaining predictable and repeatable results every time. Otherwise, not modifying the default seed value (-1) will lead to random numbers at every invocation.

Logging settings

Logger

Defines the logger type that H2O Hydrogen Torch will use for model training

Options

  • None

    H2O Hydrogen Torch does not use any logger.

  • Neptune

    H2O Hydrogen Torch will use Neptune as a logger to track the experiment. To use Neptune, you must specify a Neptune API token and a Neptune project.

Number of images

This setting defines the number of images to show in the experiment Insights tab.


Feedback