2

I noticed that there are many studies in recent years on how to train/update neural networks faster/quicker with equal or better performance. I find the following methods(except the chips arms race):

  1. using few-shot learning, for instance, pre-taining and etc.
  2. using the minimum viable dataset, for instance using (guided) progressive sampling.
  3. model compression, for instance, efficent transformers
  4. Data echoing, or simply put let the data pass multiple times in the graph(or GPU)

Is there a systematic structure on this topic and how can we update or train a model faster without loss of its capacity?

Lerner Zhang
  • 1,065
  • 1
  • 9
  • 22

0 Answers0