62 skills found · Page 1 of 3
Luolc / AdaBoundAn optimizer that trains as fast as Adam and as good as SGD.
JianGoForIt / YellowFinauto-tuning momentum SGD optimizer
JianGoForIt / YellowFin Pytorchauto-tuning momentum SGD optimizer
xswang / XflowDistributed LR、 FM model on Parameter Server. FTRL and SGD Optimization Algorithm.
kiddyboots216 / CommEfficientPyTorch for benchmarking communication-efficient distributed SGD optimization algorithms
Mrpatekful / SwatsUnofficial implementation of Switching from Adam to SGD optimization in PyTorch.
decile-team / JensenA C++ toolkit for Convex Optimization (Logistic Loss, SVM, SVR, Least Squares etc.), Convex Optimization algorithms (LBFGS, TRON, SGD, AdsGrad, CG, Nesterov etc.) and Classifiers/Regressors (Logistic Regression, SVMs, Least Squares Regression etc.)
ivallesp / Awesome OptimizersNeural Network optimizers implemented from scratch in numpy (Adam, Adadelta, RMSProp, SGD, etc.)
bharathsudharsan / ML MCUCode for IoT Journal paper 'ML-MCU: A Framework to Train ML Classifiers on MCU-based IoT Edge Devices'
v-iashin / CS231nPyTorch/Tensorflow solutions for Stanford's CS231n: "CNNs for Visual Recognition"
kozistr / AdaBound TensorflowAn optimizer that trains as fast as Adam and as good as SGD in Tensorflow
JYWa / Overlap Local SGDImplementation of (overlap) local SGD in Pytorch
khanmhmdi / Gradient Descent Optimizer VariationsThis repository contains implementation of stochastic gradient descent, SGD with momentum, Adagrad, RMSprop, Adam, Adamax optimizer from scratch using Python language.
bimalka98 / Computer Vision And Image ProcessingComputer Vision and Image Processing algorithms implemented using OpenCV, NumPy and MatPlotLib, for UOM's EN2550 Fundamentals of Image Processing and Machine Vision Module ❄
ParCIS / Ok TopkOk-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k communication volume which is asymptotically optimal) with the decentralized parallel Stochastic Gradient Descent (SGD) optimizer, and its convergence is proved theoretically and empirically.
Janus-Shiau / Lookahead TensorflowLookahead optimizer ("Lookahead Optimizer: k steps forward, 1 step back") for tensorflow
rdspring1 / Count Sketch OptimizersA compressed adaptive optimizer for training large-scale deep learning models using PyTorch
JYWa / MATCHACommunication-efficient decentralized SGD (Pytorch)
rzhu3 / YellowFin MXNetauto-tuning momentum SGD optimizer
jrios6 / Adam Vs SGD NumpyA 3-layer neural network with SGD and Adam optimizers built from scratch with numpy.