MITB Banner

Google Introduces Model Search, An Open Source Platform To Automatically Find Optimal ML Models

Share

Google has released Model Search, an open-source platform for developing efficient and best machine learning models automatically. Model Search, built on TensorFlow, finds the best-fit architecture required for a given dataset and the problem while minimising the coding time and compute resources.

A neural network is as good as its ability to generalise for different tasks. Sure, popular generalisation techniques have cropped up in recent years, but the research community has not caught up with them. Further, the problem compounds in machine learning domains that require more in-depth knowledge.

The introduction of several AutoML algorithms and techniques such as neural architecture search (NAS) has helped researchers find the ‘right’ neural network automatically, without the need for manual intervention. 

However, these techniques require thousands of models to train before converging and are compute-heavy. They don’t fare too well when working with domains on which they haven’t been trained before. This is where domain-agnostic and flexible platforms like Model Search come in.

The Model Search system from Google runs both training and evaluation experiments for various ML models with different architectures and training techniques. The system contains:

  • Multiple trainers.
  • Algorithms for search and transfer learning.
  • A database to store the evaluated models. 

The Model Search system is highly collaborative; while each trainer conducts experiments independently, in the end, all the trainers share the observations and learnings from their individual experiments. The search algorithm checks all the completed trials at the beginning of every cycle. The algorithm decides what to try next using an approximate search strategy called beam search. The search algorithm then mutates over the best architectures found; the resulting model is assigned back to the trainer.

How Does It Work?

The Model Search system builds a neural network from a set of predefined blocks representing micro-architectures such as LSTM, Transformer layers or ResNet. With this approach of using pre-existing architectural components, Model Search can leverage ‘best knowledge’ from NAS research across domains, ensuring greater efficiency and reduction in the scale of search space.

Credit: Google AI Blog

The Model Search framework is built on TensorFlow, allowing blocks to implement all functions that take tensor as an input. The search algorithms implemented in Model Search are adaptive, greedy and incremental. This makes them converge faster than other Reinforcement Learning (RL) algorithms.

The search algorithm in the Model Search system imitates the ‘explore and exploit’ principle of RL algorithms. It separates the search for a good candidate (explore) and boosts the accuracy by ensembling the good candidates so discovered (exploit). The main search algorithm applies random changes to the architecture or the training techniques to the ‘best performing’ experiments.

Credit: Google AI Blog

Model Search uses either of the two methods — Knowledge distillation or weight sharing to further improve the efficiency and accuracy of transfer learning between internal experiments. Knowledge distillation is a process of transferring knowledge from a larger model to a smaller one. This method improves candidates’ accuracy by adding a loss term matching the high performing models’ prediction.

On the other hand, weight sharing is a method that allows robust feature detection while reducing the number of parameters. In the Model Search system, the Weight sharing method bootstraps some of the parameters from previously trained candidates. This is done by imitating suitable weights from previously trained models and by randomly initialising the rest. It is a faster training method and allows for better discovery of suitable architectures.

Wrapping Up

Google said the Model Search system improved on production models with very few iterations. It was particularly found useful for keyword spotting and language identification. During testing, the system was able to find a suitable architecture for image classification on an open source CIFAR-10 dataset.

“By building upon previous knowledge for a given domain, we believe that this framework is powerful enough to build models with the state-of-the-art performance on well-studied problems when provided with a search space composed of standard building blocks,” Hanna Mazzawi and Xavi Gonzalvo from Google Research said in a blog post.

Share
Picture of Shraddha Goled

Shraddha Goled

I am a technology journalist with AIM. I write stories focused on the AI landscape in India and around the world with a special interest in analysing its long term impact on individuals and societies. Reach out to me at shraddha.goled@analyticsindiamag.com.
Related Posts

CORPORATE TRAINING PROGRAMS ON GENERATIVE AI

Generative AI Skilling for Enterprises

Our customized corporate training program on Generative AI provides a unique opportunity to empower, retain, and advance your talent.

Upcoming Large format Conference

May 30 and 31, 2024 | 📍 Bangalore, India

Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.

AI Courses & Careers

Become a Certified Generative AI Engineer

AI Forum for India

Our Discord Community for AI Ecosystem, In collaboration with NVIDIA. 

Flagship Events

Rising 2024 | DE&I in Tech Summit

April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore

MachineCon GCC Summit 2024

June 28 2024 | 📍Bangalore, India

MachineCon USA 2024

26 July 2024 | 583 Park Avenue, New York

Cypher India 2024

September 25-27, 2024 | 📍Bangalore, India

Cypher USA 2024

Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA

Data Engineering Summit 2024

May 30 and 31, 2024 | 📍 Bangalore, India

Subscribe to Our Newsletter

The Belamy, our weekly Newsletter is a rage. Just enter your email below.