资源论文PADÉ ACTIVATION UNITS :E ND -TO -END LEARNINGOF FLEXIBLE ACTIVATION FUNCTIONS IN DEEP NET-WORKS

PADÉ ACTIVATION UNITS :E ND -TO -END LEARNINGOF FLEXIBLE ACTIVATION FUNCTIONS IN DEEP NET-WORKS

2020-01-02 | |  72 |   46 |   0

Abstract

The performance of deep network learning strongly depends on the choice of the non-linear activation function associated with each neuron. However, deciding on the best activation is non-trivial, and the choice depends on the architecture, hyper-parameters, and even on the dataset. Typically these activations are fixed by hand before training. Here, we demonstrate how to eliminate the reliance on first picking fixed activation functions by using flexible parametric rational functions instead. The resulting Padé Activation Units (PAUs) can both approximate common activation functions and also learn new ones while providing compact representations. Our empirical evidence shows that end-to-end learning deep networks with PAUs can increase the predictive performance. Moreover, PAUs pave the way to approximations with provable robustness.

上一篇:REINFORCEMENT LEARNING WITH COMPETITIVEE NSEMBLES OF INFORMATION -C ONSTRAINEDP RIMITIVES

下一篇:A LATENT MORPHOLOGY MODEL FOR OPEN-VOCABULARY NEURAL MACHINE TRANSL ATION

用户评价
全部评价

热门资源

  • The Variational S...

    Unlike traditional images which do not offer in...

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...