Noise and Saturation

When training a deep auto-encoder for the purposes of semantic hashing, Hinton adds noise to the input during the fine-tuning stage, because this forces the activities of the sigmoid units in the code to become bimodal (saturate). Why does adding noise have this effect?

1 Response to “Noise and Saturation”


  1. 1 Geoffroy MOURET April 21, 2013 at 20:36

    Considering sigmoid units h_i with an added noise \epsilon_i, the model will try to make the noise less significant compared to the primary input of the sigmoid unit.
    h_i = \text{sigmoid}(b_i + \sum_j W_{ij} z_j + \epsilon_i)
    This will increase b_i and W_{ij} to compensate the noise.


Leave a comment