Skip to content
Snippets Groups Projects
  1. Jun 10, 2024
    • Alexandru-Mihai GHERGHESCU's avatar
      Fix inference code · cb1a7974
      Alexandru-Mihai GHERGHESCU authored
      This should now work with any PyTorch model (Optimus is the example
      given in the source code), as well as any HuggingFace model (adjusted
      the code to be independent of any model source).
  2. Jun 04, 2024
  3. Jun 03, 2024
  4. Feb 16, 2024
  5. Feb 15, 2024
    • Vlad-Andrei BĂDOIU (78692)'s avatar
      Merge branch 'feature/fp16' into 'main' · 0faac554
      Vlad-Andrei BĂDOIU (78692) authored
      Add fp16 mixed precision training
      
      See merge request !17
      0faac554
    • Alexandru-Mihai GHERGHESCU's avatar
      Adjust optimizer epsilon value for AMP · 8579fc15
      Alexandru-Mihai GHERGHESCU authored
      Pick a better default as epsilon value. Although this value should never
      touch the fp16 gradients in mixed precision training (as the optimizer
      should only ever work on the master fp32 copy of the model), this value
      didn't need to be changed. However, in pure fp16 training, any epsilon
      value lower than 1e-7 would simply underflow to 0, causing it to become
      useless.
      
      Although the framework doesn't directly support the second case above,
      an epsilon value of 1e-7 seems like a better default for both AMP and
      normal training.
    • Alexandru-Mihai GHERGHESCU's avatar
      Add fp16 mixed precision training · 6db26eb1
      Alexandru-Mihai GHERGHESCU authored
      This should give training a theoretical 2x speedup in time (though in
      practice that's not usually the case), with close to no loss in
      performance.
      
      The interface allows the user to choose between mixed precision or no
      mixed precision training, which falls back to normal float32 precision.
      
      CPU support for training has been dropped, as it takes (with or without
      mixed precision) much much longer to train than on GPU's, and it's not
      really an alternative anyone considers. With the addition of mixed
      precision, supporting both CPU and GPU would complicate things too much,
      therefore CPU training support has been dropped.
  6. Jan 30, 2024
  7. Jan 29, 2024
  8. Jan 28, 2024
  9. Jan 26, 2024
  10. Jan 25, 2024
  11. Jan 24, 2024
Loading