Commit Graph

217 Commits

Author SHA1 Message Date
Connor Olding a12121c16a rewrite readme and example code 2020-04-08 16:08:49 -07:00
Connor Olding a8871d4a67 comment on NLL implementation 2020-03-17 07:27:39 -07:00
Connor Olding 96dfc904b1 add ISRLU activation
needs testing
2020-03-17 07:27:03 -07:00
Connor Olding b6fe7f711e add DenseUnbiased layer 2020-03-17 07:26:45 -07:00
Connor Olding 5cd4e8d1c1 add preliminary Sparse layer 2020-03-17 07:26:31 -07:00
Connor Olding 4a5084df48 refine Adamlike 2020-03-17 07:26:05 -07:00
Connor Olding 06bbfc6340 fix a couple comments 2019-03-22 13:10:42 +01:00
Connor Olding bf4ec2ec94 add optimizer hacks 2019-03-22 12:59:33 +01:00
Connor Olding 2e80f8b1a7 add MSVAG optimizer 2019-03-22 12:58:41 +01:00
Connor Olding b3b82ca4f0 add aliases for optimizer names 2019-03-22 12:57:47 +01:00
Connor Olding 86491ad841 add Adamax optimizer 2019-03-22 12:57:05 +01:00
Connor Olding 2a4f92154d rewrite momentum optimizer
the original version wasn't very useful as it would typically diverge.
2019-03-22 12:55:13 +01:00
Connor Olding 5389ae7fca add LogCosh loss 2019-03-22 12:54:44 +01:00
Connor Olding 7f1d2afb02 add hinge losses 2019-03-22 12:54:20 +01:00
Connor Olding 285bf1d96a move Confidence to experimental 2019-03-22 12:53:30 +01:00
Connor Olding 0aba113cb7 update comments 2019-03-22 12:52:44 +01:00
Connor Olding e2530c17e5 clean up Ritual a little 2019-02-18 06:25:10 +01:00
Connor Olding 69e6ec3fad add a convenient generator method to Learner 2019-02-18 06:23:05 +01:00
Connor Olding 2b5798332d allow configuration of Neumann hyperparameters 2019-02-17 07:47:53 +01:00
Connor Olding c92082e07a fix momentum quantity in Neumann optimizer 2019-02-17 07:47:05 +01:00
Connor Olding 5a07cdac32 add mixup variant of batchize 2019-02-11 20:30:31 +01:00
Connor Olding 2cfcc9062e use the constant recommended by the paper 2019-02-06 18:41:55 +01:00
Connor Olding 594483c6d4 correct "frequency" to "period" 2019-02-06 18:41:38 +01:00
Connor Olding 50e0311051 allow passing model through Ritual init 2019-02-05 22:16:46 +01:00
Connor Olding fd2fc4bd76 begin rewriting Ritual 2019-02-05 06:09:35 +01:00
Connor Olding 2c921d34c2 add Yogi optimizer 2019-02-05 04:19:48 +01:00
Connor Olding 7deaa3c3f6 reword some comments 2019-02-05 04:19:14 +01:00
Connor Olding 7227559912 reset learning rates in optimizers 2019-02-05 04:15:28 +01:00
Connor Olding bd07d983be add polynomial features layer 2019-02-05 04:13:56 +01:00
Connor Olding 54ea41711b refactor gradient filtering 2019-02-03 15:10:43 +01:00
Connor Olding 0d28882ef0 remove YellowFin because it's not worth maintaining 2019-02-03 15:03:03 +01:00
Connor Olding 5fd2b7b546 remove old versions of optimizers 2019-02-03 14:43:04 +01:00
Connor Olding 1e6887cdbd comment on Softmax alpha term 2019-02-03 14:31:34 +01:00
Connor Olding b8c40d2e2f rewrite some comments 2019-02-03 14:30:58 +01:00
Connor Olding 94f27d6f2a add Adadelta optimizer 2019-02-03 14:30:47 +01:00
Connor Olding f60535aa01 generalize Adam-like optimizers 2019-02-03 14:30:03 +01:00
Connor Olding c6ebd02ea9 comply to PEP 8 2018-03-24 06:26:02 +01:00
Connor Olding a685db1489 add exact GELU activation 2018-03-22 12:02:17 +01:00
Connor Olding 7161f983ab fix __name__ being incorrect due to extra __all__
this fixes tracebacks and checks for __main__,
among other things.
2018-03-17 14:09:15 +01:00
Connor Olding ec71d4956e fix module leaking for real 2018-03-13 03:27:11 +01:00
Connor Olding 843524b80a move ActivityRegularizer to regularizer.py 2018-03-13 02:34:44 +01:00
Connor Olding b74e0941dc fix Swish inits 2018-03-12 02:17:28 +01:00
Connor Olding bd1e80b8de activation layers inherit a dummy class 2018-03-11 22:34:46 +01:00
Connor Olding cb4e6d3725 add Swish and SiLU activations 2018-03-11 03:34:00 +01:00
Connor Olding 91cdea3b26 fix inequalities in HardClip
oldest trick in the book
2018-03-10 05:03:26 +01:00
Connor Olding a6519f5455 improve notes on dependencies 2018-03-09 10:17:31 +01:00
Connor Olding bd4f2a9478 fix missing import for ActivityRegularizer 2018-03-09 10:09:50 +01:00
Connor Olding 9a45b26b7f add rough stratified k-folding utility class 2018-03-08 02:41:45 +01:00
Connor Olding 65bc9b8a6f optionally allow gradients to passthru Input layer 2018-03-08 02:40:56 +01:00
Connor Olding 4746103978 add HardClip activation layer 2018-03-08 02:40:42 +01:00