Loading [MathJax]/extensions/Safe.js
Welcome to ShortScience.org!
RSS Feed Twitter Facebook

  • ShortScience.org is a platform for post-publication discussion aiming to improve accessibility and reproducibility of research ideas.
  • The website has 1583 public summaries, mostly in machine learning, written by the community and organized by paper, conference, and year.
  • Reading summaries of papers is useful to obtain the perspective and insight of another reader, why they liked or disliked it, and their attempt to demystify complicated sections.
  • Also, writing summaries is a good exercise to understand the content of a paper because you are forced to challenge your assumptions when explaining it.
  • Finally, you can keep up to date with the flood of research by reading the latest summaries on our Twitter and Facebook pages.
hide


[link]
Summary by Denny Britz 9 years ago

TLDR; The authors propose an importance-sampling approach to deal with large vocabularies in NMT models. During training, the corpus is partitioned, and for each partition only target words occurring in that partition are chosen. To improve decoding speed over the full vocabulary, the authors build a dictionary mapping from source sentence to potential target vocabulary. The authors evaluate their approach on standard MT tasks and perform better than the baseline models with smaller vocabulary.

Key Points:
  • Computing partition function is the bottleneck. Use sampling-based approach.
  • Dealing with large vocabulary during training is separate from dealing with large vocab during decoding. Training is handled with importance sampling. Decoding is handled with source-based candidate list.
  • Decoding with candidate list takes around 0.12s (0.05) per token on CPU (GPU). Without target list 0.8s (0.25s).
  • Issue: Candidate list is depended on source sentence, so it must be re-computed for each sentence.
  • Reshuffling the data set is expensive as new partitions need to be calculated (not necessary, but improved scores).
Notes:
  • How is the corpus partitioned? What's the effect of the partitioning strategy?
  • The authors say that they replace UNK tokens using "another word alignment model" but don't go into detail what this is. The results show that doing this results in much larger score bump than increasing the vocab does. (The authors do this for all comparison models though).
  • Reshuffling the dataset also results in a significant performance bump, but this operation is expensive. IMO the authors should take all these into account when reporting performance numbers. A single training update may be a lot faster, but the setup time increases. I'd would've like to see the authors assign a global time budget to train/test and then compare the models based on that.
  • The authors only briefly mentioned that re-building the target vocab for each source sentence is an issue and how they solve it, no details given.
more
[link]
Summary by Alexander Jung 7 years ago
  • They describe a model that upscales low resolution images to their high resolution equivalents ("Single Image Super Resolution").
  • Their model uses a deeper architecture than previous models and has a residual component.
How
  • Their model is a fully convolutional neural network.
  • Input of the model: The image to upscale, already upscaled to the desired size (but still blurry).
  • Output of the model: The upscaled image (without the blurriness).
  • They use 20 layers of padded 3x3 convolutions with size 64xHxW with ReLU activations. (No pooling.)
  • They have a residual component, i.e. the model only learns and outputs the change that has to be applied/added to the blurry input image (instead of outputting the full image). That change is applied to the blurry input image before using the loss function on it. (Note that this is a bit different from the currently used "residual learning".)
  • They use a MSE between the "correct" upscaling and the generated upscaled image (input image + residual).
  • They use SGD starting with a learning rate of 0.1 and decay it 3 times by a factor of 10.
  • They use weight decay of 0.0001.
  • During training they use a special gradient clipping adapted to the learning rate. Usually gradient clipping restricts the gradient values to [-t, t] (t is a hyperparameter). Their gradient clipping restricts the values to [-t/lr, t/lr] (where lr is the learning rate).
  • They argue that their special gradient clipping allows the use of significantly higher learning rates.
  • They train their model on multiple scales, e.g. 2x, 3x, 4x upscaling. (Not really clear how. They probably feed their upscaled image again into the network or something like that?)
Results
  • Higher accuracy upscaling than all previous methods.
  • Can handle well upscaling factors above 2x.
  • Residual network learns significantly faster than non-residual network.

Architecture

Architecture of the model.

Examples

Super-resolution quality of their model (top, bottom is a competing model).

more
[link]
Summary by David Stutz 6 years ago

Lee et al. propose a generative model for obtaining confidence-calibrated classifiers. Neural networks are known to be overconfident in their predictions – not only on examples from the task’s data distribution, but also on other examples taken from different distributions. The authors propose a GAN-based approach to force the classifier to predict uniform predictions on examples not taken from the data distribution. In particular, in addition to the target classifier, a generator and a discriminator are introduced. The generator generates “hard” out-of-distribution examples; ideally these examples are close to the in-distribution, i.e., the data distribution of the actual task. The discriminator is intended to distinguish between out- and in-distribution. The overall algorithm, including the necessary losses, is given in Algorithm 1. In experiments, the approach is shown to allow detecting out-distribution examples nearly perfectly. Examples of the generated “hard” out-of-distribution samples are given in Figure 1.

Algorithm 1: The proposed joint training scheme of out-distribution generator $G$, the in-/out-distribution discriminator $G$ and the original classifier providing $P_\theta$(y|x)$ with parameters $\theta$.

Figure 1: A comparison of a regular GAN (a and c) to the proposed framework (c and d). Clearly, the proposed approach generates out-of-distribution samples (i.e., no meaningful digits) close to the original data distribution.

more
[link]
Summary by David Stutz 6 years ago

Zhao et al. propose a generative adversarial network (GAN) based approach to generate meaningful and natural adversarial examples for images and text. With natural adversarial examples, the authors refer to meaningful changes in the image content instead of adding seemingly random/adversarial noise – as illustrated in Figure 1. These natural adversarial examples can be crafted by first learning a generative model of the data, e.g., using a GAN together with an inverter (similar to an encoder), see Figure 2. Then, given an image $x$ and its latent code $z$, adversarial examples $\tilde{z} = z + \delta$ can be found within the latent code. The hope is that these adversarial examples will correspond to meaningful, naturally looking adversarial examples in the image space.

Figure 1: Illustration of natural adversarial examples in comparison ot regular, FGSM adversarial examples.

Figure 2: Generative model (GAN) together with the required inverter.

In practice, e.g., on MNIST, any black-box classifier can be attacked by randomly sampling possible perturbations $\delta$ in the random space (with increasing norm) until an adversarial perturbation is found. Here, the inverted from Figure 2 is trained on top of the critic of the GAN (although specific details are missing in the paper).

Also find this summary at davidstutz.de.

more
[link]
Summary by José Manuel Rodríguez Sotelo 9 years ago

The main contribution of this paper is introducing a new transformation that the authors call Batch Normalization (BN). The need for BN comes from the fact that during the training of deep neural networks (DNNs) the distribution of each layer’s input change. This phenomenon is called internal covariate shift (ICS).

What is BN?

Normalize each (scalar) feature independently with respect to the mean and variance of the mini batch. Scale and shift the normalized values with two new parameters (per activation) that will be learned. The BN consists of making normalization part of the model architecture.

What do we gain?

According to the author, the use of BN provides a great speed up in the training of DNNs. In particular, the gains are greater when it is combined with higher learning rates. In addition, BN works as a regularizer for the model which allows to use less dropout or less L2 normalization. Furthermore, since the distribution of the inputs is normalized, it also allows to use sigmoids as activation functions without the saturation problem.

What follows?

This seems to be specially promising for training recurrent neural networks (RNNs). The vanishing and exploding gradient problems [journals/tnn/BengioSF94] have their origin in the iteration of transformation that scale up or down the activations in certain directions (eigenvectors). It seems that this regularization would be specially useful in this context since this would allow the gradient to flow more easily. When we unroll the RNNs, we usually have ultra deep networks.

Like
  • Simple idea that seems to improve training.
  • Makes training faster.
  • Simple to implement. Probably.
  • You can be less careful with initialization.
Dislike
  • Does not work with stochastic gradient descent (minibatch size = 1).
  • This could reduce the parallelism of the algorithm since now all the examples in a mini batch are tied.
  • Results on ensemble of networks for ImageNet makes it harder to evaluate the relevance of BN by itself. (Although they do mention the performance of a single model).
more

Send Feedback
ShortScience.org allows researchers to publish paper summaries that are voted on and ranked!
About

Sponsored by: