Mosaic ML
The goal of Mosaic ML is making ML training efficient, and to improve efficiency of neural network training with algorithmic methods that deliver speed, boost quality and reduce cost. Built by a mosaic of passionate technologists, including MIT principal investigator Michael Carbin, the team was brought together to discover new ways to make Machine Learning training better, faster, and more efficient in order to accelerate breakthroughs.
The MosaicML Composer is an open source deep learning library purpose-built to make it easy to add algorithmic methods and compose them together into novel recipes that speed up model training and improve model quality. This library includes 20 methods for computer vision and natural language processing in addition to standard models, datasets, and benchmarks.
Detailed information available from Mosaic ML.
Explore
New Method Could Increase LLM Training Efficiency
Adam Zewe | MIT News
By leveraging idle computing time, researchers can double the speed of model training while preserving accuracy.
AI Tool Generates High-Quality Images Faster Than State-of-the-Art Approaches
Adam Zewe | MIT News
Researchers fuse the best of two popular methods to create an image generator that uses less energy and can run locally on a laptop or smartphone.
New Security Protocol Shields Data From Attackers During Cloud-based Computation
Adam Zewe | MIT News
The technique leverages quantum properties of light to guarantee security while preserving the accuracy of a deep-learning model.




