fast
Routine optimizer
A high-performance D library for optimizing everyday routines with minimal overhead.
A library for D that aims to provide the fastest possible implementation of some every day routines.
112 stars
9 watching
7 forks
Language: D
last commit: over 1 year ago
Linked from 1 awesome list
dlang
Related projects:
Repository | Description | Stars |
---|---|---|
dmlc/mxnet-memonger | A tool for optimizing deep learning models to reduce memory usage without sacrificing performance | 308 |
mapillary/inplace_abn | An optimization technique to reduce memory usage in deep neural networks during training | 1,321 |
mengrao/str | An optimized string class with an adaptive hash table for fast searching and comparisons | 125 |
dogada/fast-redux | An extension of Redux for improving its performance by dynamically importing and executing reducers in O(1) speed | 133 |
taolei87/sru | A recurrent neural network implementation optimized for speed and parallelizability | 31 |
dppalomar/sparseindextracking | A package for computing sparse portfolios of assets to track an index. | 50 |
uclaml/fedlinucb | An algorithm designed to optimize the selection of actions in multiple, distributed environments with feedback and context information. | 2 |
brml/climin | A framework for optimizing machine learning functions using gradient-based optimization methods. | 180 |
chrispenner/wc | An optimization project that uses Haskell to improve the performance of the Unix utility wc | 136 |
jiangoforit/yellowfin_pytorch | An optimizer that automatically tunes momentum and learning rate using local quadratic approximation. | 287 |
damirsvrtan/fasterer | Tools for optimizing Ruby code performance | 1,812 |
ziyuw/rembo | An optimization algorithm that uses Bayesian methods and random embeddings to solve complex problems in high-dimensional spaces. | 113 |
dwavesystems/dwave-neal | An implementation of a simulated annealing algorithm for approximate Boltzmann sampling or heuristic optimization. | 51 |
davisyoshida/lorax | A JAX transform that simplifies the training of large language models by reducing memory usage through low-rank adaptation. | 132 |
datacanvasio/hypergbm | An AutoML toolkit designed to automate the entire machine learning process pipeline for tabular data | 337 |