PyTorch One-Hot Labels

- 2 mins

Some interesting loss functions in deep learning require “one-hot” labels.

A “one-hot” label is simply a binary array of dimensions dim0...dimN, C, where C is the number of classes in your label. This differs from the corresponding integer label, which is an array of dim0...dimN, where the values of each element are in the range [0, C]. Storing labels in the integer format is more common, since it’s more compact.

When using PyTorch, the built in loss functions all accept integer label inputs (thanks to the devs for making our lives easy!).

However, if you implement your own loss functions, you may need one-hot labels. Converting between the two is easy and elegant in PyTorch, but may be a little unintuitive.

To do so, we rely on the torch.Tensor.scatter_() function, which fills the target tensor with values from the source along provided indices. See the documentation for details.

Below is a quick function I threw together to convert 2D integer labels to 2D one-hot labels, which can easily be altered for a different input/output dimensionality. See the Gist here.

Let’s see this in action.

>> labels = torch.LongTensor(4,4) % 3

2  1  0  0
1  0  0  0
2  0  0  1
2  0  0  1
[torch.LongTensor of size 4x4]

>> make_one_hot(labels)

(0 ,0 ,.,.) =
  0  0  1  1
  0  1  1  1
  0  1  1  0
  0  1  1  0

(0 ,1 ,.,.) =
  0  0  0  0
  1  0  0  0
  0  0  0  1
  0  0  0  1

(0 ,2 ,.,.) =
  1  1  0  0
  0  0  0  0
  1  0  0  0
  1  0  0  0
[torch.LongTensor of size 1x3x4x4]
Jacob C. Kimmel

Jacob C. Kimmel

Co-founder & Head of Research @ NewLimit. Interested in aging, genomics, imaging, & machine learning.

rss facebook twitter github youtube mail spotify instagram linkedin google pinterest medium vimeo gscholar