__init__.py
|
fix module leaking for real
|
2018-03-13 03:27:11 +01:00 |
activation.py
|
add ISRLU activation
|
2020-03-17 07:27:03 -07:00 |
experimental.py
|
move Confidence to experimental
|
2019-03-22 12:53:30 +01:00 |
float.py
|
add exact GELU activation
|
2018-03-22 12:02:17 +01:00 |
initialization.py
|
basic PEP 8 compliance
|
2018-01-22 19:40:36 +00:00 |
layer_base.py
|
basic PEP 8 compliance
|
2018-01-22 19:40:36 +00:00 |
learner.py
|
add a convenient generator method to Learner
|
2019-02-18 06:23:05 +01:00 |
loss.py
|
add LogCosh loss
|
2019-03-22 12:54:44 +01:00 |
math.py
|
basic PEP 8 compliance
|
2018-01-22 19:40:36 +00:00 |
nodal.py
|
basic PEP 8 compliance
|
2018-01-22 19:40:36 +00:00 |
optimizer.py
|
refine Adamlike
|
2020-03-17 07:26:05 -07:00 |
optimizer_base.py
|
reset learning rates in optimizers
|
2019-02-05 04:15:28 +01:00 |
parametric.py
|
add DenseUnbiased layer
|
2020-03-17 07:26:45 -07:00 |
ritual.py
|
begin rewriting Ritual
|
2019-02-05 06:09:35 +01:00 |
ritual_base.py
|
add optimizer hacks
|
2019-03-22 12:59:33 +01:00 |
utility.py
|
fix a couple comments
|
2019-03-22 13:10:42 +01:00 |
weight.py
|
basic PEP 8 compliance
|
2018-01-22 19:40:36 +00:00 |