- 01 Jul, 2017 1 commit
-
-
Davis King authored
between. This should make syncing more robust to sudden hardware failure that happens right when saving to disk.
-
- 22 Jan, 2017 1 commit
-
-
Davis King authored
early stopping based on observing the loss on held out data.
-
- 17 Dec, 2016 2 commits
-
-
Davis King authored
using train_one_step(), you will get the behavior of automatic disk syncs at the end of training.
-
Davis King authored
-
- 06 Nov, 2016 1 commit
-
-
Davis King authored
training_label_type instead of a single label_type. This way, the label type used for training can be distinct from the type output by the network. This change breaks backwards compatibility with the previous API.
-
- 10 Sep, 2016 1 commit
-
-
Davis King authored
training thread) out of the object into the calling code rather than terminating the application.
-
- 01 Sep, 2016 1 commit
-
-
jpblackburn authored
Add an overload of dnn_trainer::train_one_step that takes a pair of iterators rather than a std::vector.
-
- 30 May, 2016 2 commits
-
-
Davis King authored
-
Fm authored
-
- 15 May, 2016 3 commits
-
-
Davis King authored
progress" estimate. I also renamed the get/set functions for the shrink amount to have a consistent name and use the word "factor" instead of "amount".
-
Davis King authored
-
Davis King authored
object as an input. This allows the solvers to exhibit a more complex behavior that depends on the specific layer. It also removes the learning rate from the solver's parameter set and pushes it entirely into the core training code. This also removes the need for the separate "step size" which previously was multiplied with the output of the solvers. Most of the code is still the same, and in the core and trainer the step_size variables have just been renamed to learning_rate. The dnn_trainer's relevant member functions have also been renamed. The examples have been updated to reflect these API changes. I also cleaned up the resnet definition and added better downsampling.
-
- 19 Apr, 2016 1 commit
-
-
Davis King authored
-
- 16 Apr, 2016 1 commit
-
-
Davis King authored
- Made layer() recurse into repeat objects so that the index given to layer() does what you would expect. - Added an operator<< for network objects that prints the network architecture.
-
- 12 Apr, 2016 1 commit
-
-
Davis King authored
-
- 10 Apr, 2016 1 commit
-
-
Davis King authored
-
- 08 Apr, 2016 1 commit
-
-
Davis King authored
-
- 25 Mar, 2016 1 commit
-
-
Davis King authored
how much the current step size has converged.
-
- 27 Feb, 2016 1 commit
-
-
Davis King authored
automatic step size reduction code a little.
-
- 09 Jan, 2016 2 commits
-
-
Davis King authored
network (which would needlessly double VRAM usage). I also added a set_synchronization_file() method so you can tell it to automatically synchronize itself to disk every so often during training. This makes resuming an interrupted training session trivially easy.
-
Davis King authored
the loss isn't being reduced. Also, there is a stopping condition now based on how large the current learning rate is. That is, training stops when the learning rate gets small enough and it is clear that no progress is being made.
-
- 01 Jan, 2016 1 commit
-
-
Davis King authored
networks. This revolved mostly around removing really deep template recursions since that upsets the compiler when you make really deep networks.
-
- 13 Dec, 2015 1 commit
-
-
Davis King authored
-
- 29 Nov, 2015 1 commit
-
-
Davis King authored
train_one_step() member function. Also improved how the host to device transfers are overlapped with kernel computation.
-
- 15 Oct, 2015 1 commit
-
-
Davis King authored
-
- 14 Oct, 2015 1 commit
-
-
Davis King authored
-
- 03 Oct, 2015 1 commit
-
-
Davis King authored
-