114

GitHub - kmkolasinski/deep-learning-notes: Experiments with Deep Learning

 5 years ago
source link: https://github.com/kmkolasinski/deep-learning-notes
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

README.md

deep-learning-notes

Experiments with Deep Learning and other resources:

  • keras-capsule-pooling - an after-hours experiment in which I try to implement Capsule pooling for images.
  • max-normed-optimizer - an experimental implementation of an interesting gradient descent optimizer which normalizes gradients according to their norms. Contains various experiments which show potential power of this method.
  • selu-regularization - a Keras Regularizer Layer which allows for forcing SELU like regularization on the model weights (Dense and Conv2D versions are provided). Selu was introduced as an activation function with special initialization method, those regularizers can be add to force the weight to preserve self normalizing property during the training.
  • tf-oversampling - example with how to implement oversampling with tf.data.Dataset API.

Seminars on Deep Learning and Machine Learning

Seminars - contains a bunch of presentations I have gave at our company.


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK