-
Forward Automatic Differentiation
-- For cases when
$N_k$ is larger than$N_0$ in the computational graph - Quantization -- Technique to reduce computational load and fasten inferencing
- Automatic Mixed Precision -- To match appopriate datatype for each operation and reduce runtime
- Knowledge Distillation -- Method to transfer knowledge from larger to small network
- Profiling -- To assess time and memory consumption for a model's operation
- Pruning -- Method to reduce model parameters without affecting much of the performance
-
Notifications
You must be signed in to change notification settings - Fork 0
k-karna/optimising_pytorch_models
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
About
Experimentations with optimising techniques for pytorch modelling - Pruning, Automatic Mixed Precision, Knowledge Distillation, Quantization, Profiling, Forward Automatic Differentiation
Topics
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published