1. 14 Sep, 2020 1 commit
  2. 07 Sep, 2020 1 commit
  3. 19 Aug, 2020 1 commit
  4. 14 Aug, 2020 1 commit
  5. 13 Aug, 2020 2 commits
  6. 12 Aug, 2020 4 commits
  7. 11 Aug, 2020 1 commit
  8. 30 Jul, 2020 1 commit
  9. 24 Jul, 2020 1 commit
  10. 01 Jul, 2020 1 commit
  11. 30 Jun, 2020 1 commit
    • Chi Song's avatar
      Reuse OpenPAI jobs to run multiple trials (#2521) · 0b9d6ce6
      Chi Song authored
      Designed new interface to support reusable training service, currently only applies to OpenPAI, and default disabled.
      
      Replace trial_keeper.py to trial_runner.py, trial_runner holds an environment, and receives commands from nni manager to run or stop an trial, and return events to nni manager.
      Add trial dispatcher, which inherits from original trianing service interface. It uses to share as many as possible code of all training service, and isolate with training services.
      Add EnvironmentService interface to manage environment, including start/stop an environment, refresh status of environments.
      Add command channel on both nni manager and trial runner parts, it supports different ways to pass messages between them. Current supported channels are file, web sockets. and supported commands from nni manager are start, kill trial, send new parameters; from runner are initialized(support some channel doesn't know which runner connected), trial end, stdout ((new type), including metric like before), version check (new type), gpu info (new type).
      Add storage service to wrapper a storage to standard file operations, like NFS, azure storage and so on.
      Partial support run multiple trials in parallel on runner side, but not supported by trial dispatcher side.
      Other minor changes,
      
      Add log_level to TS UT, so that UT can show debug level log.
      Expose platform to start info.
      Add RouterTrainingService to keep origianl OpenPAI training service, and support dynamic IOC binding.
      Add more GPU info for future usage, including GPU mem total/free/used, gpu type.
      Make some license information consistence.
      Fix async/await problems on Array.forEach, this method doesn't support async actually.
      Fix IT errors on download data, which causes by my #2484 .
      Accelerate some run loop pattern by reducing sleep seconds.
      0b9d6ce6
  12. 29 Jun, 2020 1 commit
  13. 23 Jun, 2020 1 commit
  14. 22 Jun, 2020 1 commit
  15. 12 Jun, 2020 1 commit
  16. 05 Jun, 2020 1 commit
    • Chi Song's avatar
      remove optimize_mode from curve fitting (#2471) · e75a9f5a
      Chi Song authored
      others
      
      1. fix failed curve fitting UTs, due to code changes.
      1. move all SDK UTs to tests folder, so that they can be run in default tests.
      1. fix some deprecated ut assert function calls.
      e75a9f5a
  17. 25 May, 2020 1 commit
  18. 19 May, 2020 1 commit
  19. 12 May, 2020 1 commit
  20. 06 May, 2020 1 commit
  21. 30 Apr, 2020 1 commit
  22. 26 Apr, 2020 1 commit
    • Chi Song's avatar
      fix #1578 and some improvements (#2370) · 1c6f1efa
      Chi Song authored
      Add shell support for ssh connection, so that remote script can be started with user environment.
      
      Minor fixes,
      
      1. Fix gpu_metrics_collector to support pyenv. As pyenv will create one more process, so that original pgrep code always got extra processes, and cannot start gpu_metrics_collector.
      2. Fix NASUI failure on dev-install-node-modules, to create subfolder every time.
      3. Fix MakeFile to reduce mis-created links, and other minor issues.
      4. Add node --watch for nni_manager for better dev experience.
      1c6f1efa
  23. 30 Mar, 2020 1 commit
  24. 27 Mar, 2020 1 commit
  25. 25 Mar, 2020 1 commit
  26. 18 Mar, 2020 1 commit
  27. 17 Mar, 2020 1 commit
  28. 05 Mar, 2020 1 commit
  29. 03 Mar, 2020 1 commit
  30. 02 Mar, 2020 1 commit
    • George Cheng's avatar
      DLTS integration (#1945) · 134368fa
      George Cheng authored
      
      
      * skeleton of dlts training service (#1844)
      
      * Hello, DLTS!
      
      * Revert version
      
      * Remove fs-extra
      
      * Add some default cluster config
      
      * schema
      
      * fix
      
      * Optional cluster (default to `.default`)
      
      Depends on DLWorkspace#837
      
      * fix
      
      * fix
      
      * optimize gpu type
      
      * No more copy
      
      * Format
      
      * Code clean up
      
      * Issue fix
      
      * Add optional fields in config
      
      * Issue fix
      
      * Lint
      
      * Lint
      
      * Validate email, password and team
      
      * Doc
      
      * Doc fix
      
      * Set TMPDIR
      
      * Use metadata instead of gpu_capacity
      
      * Cancel paused DLTS job
      
      * workaround lint rules
      
      * pylint
      
      * doc
      Co-authored-by: default avatarQuanluZhang <z.quanluzhang@gmail.com>
      134368fa
  31. 27 Feb, 2020 1 commit
  32. 14 Feb, 2020 1 commit
  33. 09 Feb, 2020 1 commit
  34. 07 Feb, 2020 2 commits
  35. 04 Feb, 2020 1 commit