July 19th, 2019

AI for Smarter Manufacturing

RSS icon RSS Category: H2O Driverless AI, Manufacturing, Use Cases

Code 3

Manufacturing is a centuries old industry and has seen significant changes dating back to the first Industrial Revolution in the late 18th century. The use of conveyor belt assembly lines to replace assembly workers, newer precision robot technologies to further reduce manufacturing time, advances in ERP, historian databases, storage and computing technologies for efficient part ordering, plant monitoring and supply chain management are just a few examples of disruption that this industry has seen.

Over the recent decades, the manufacturing industry has evolved into a variety of different types depending on the vertical:

  • Electronics and device manufacturing
  • Food production, packaging and processing plants
  • Medical device manufacturing
  • Semiconductor test equipment manufacturing
  • Heavy manufacturing (automotive, etc).

Although each of the above are in very different markets, the challenges that each of them face are fairly similar. They all require the following at a minimum:

  • Failure predictability of machines
  • Optimization of the productivity of the equipment
  • In-depth product yield analysis
  • Optimal routing of raw material to the manufacturing plant, of finished goods to delivery locations, with minimal product waste in transit and at the least cost possible.
  • Achieve the perfect balance between supply and demand to reduce inventory cost
  • Customer usage statistics and market feedback on product quality

Now, to say that these are all difficult to achieve would be an understatement. Let’s take an example.

Predicting machine failure requires historical data points of past failures, near failures, maximum time between failures (MTBF), correlations between the various sensors deployed in a machine along with external data such as weather, maintenance logs and so on. Once you have the requisite data, a data scientist would need to accomplish a few key steps:


# Description of Step Examples
1 The data scientist would need a way to extract the right features or combine a few of these variables to generate a new feature that help with developing the model. A feature, in this case, could be the physical characteristics of the sensors embedded in the machinery (simpler), or a weighted combination of temperature variation and frequency of incoming sensor data (derived).
2 The data scientists would have to select a machine learning framework and pick a set of algorithms and models that might work best in this scenario, out of the several thousand models that the data science community has written thus far. PyTorch, XGBoost, TensorFlow, sklearn, Pandas and H2O are a few well-known libraries and frameworks (a.k.a. open source repositories) of ML models. Each library is a rich set of models developed by the community over the recent years.
3 They would then have to tune the parameters of the model to overcome over- or underfitting. The individual weights on statistical functions, depth of a decision tree, number of trees, to help predict machine failure in exactly the stipulated time.
4 This would then lead to a discussion on how the model should be deployed on the target. Deploying the model on the machine itself along with the runtime environment or a nearby gateway device where the data from other machines is also collected, or in the central datastore (in the cloud or on-premises data warehouse).


As you can imagine, a seemingly simple problem can easily be perceived as a very complex task to accomplish.

This is where a platform like the H2O Driverless AI can come handy.

H2O Driverless AI is a leading-edge platform that automates all of the above steps, resulting in drastic reduction in time taken from start of the project all the way to the point when a business can glean and review real-time actionable insights.

Again, let’s take an example set of steps to describe how the data scientist can accomplish all of this within the platform:

  • After they bring the relevant training data on the platform via a simple CSV upload or through one of our cloud or HDFS connectors, they would select the target column or variable to be predicted. In this case, this would perhaps be something like “Time to failure in less than 4 days”.
  • They would then select the accuracy, time and interpretability parameters that the platform should use in order to decide the extent of compute the user wants it to do before showing the final results.
  • Depending upon the dataset, the user also gets the flexibility to use GPUs for faster compute, instead of CPUs.
  • They would then ask Driverless AI to build the model. At that point, the platform will take a few minutes to iterate over potentially 1000s of models (not exaggerating here!!), tune the hyperparameters of the model to fit into the criteria of accuracy provided earlier and conclude on the importance of the variables in decreasing order. As the platform works on the dataset, all of this would be visible to the user on the UI itself.
  • Finally, once the model is ready, there are several deployment options to choose from. If they already have Java runtime libraries setup in the target environment, they could use the POJO or MOJO files that the platform provides as an output – particularly for on-premises or edge deployments. Alternatively, this can be deployed in the AWS Sagemaker, or just simply create a RESTful endpoint with an API key and a model ID for a web application to access it.

The benefits of using the Driverless AI platform go well beyond the above-mentioned descriptions. The algorithms and scorers are built by the H2O.ai’s data science community and curated by the Kaggle Grand Masters at H2O.ai. Moreover, with the new concept of BYO Recipes introduced in Driverless AI, the extent of flexibility and extensibility that the platform provides goes beyond the imagination.

For more information, please refer to the AI in Manufacturing solution brief, or check out some of successful customer stories in manufacturing here.

About the Author

vinod iyengar
Vinod Iyengar, VP of Products

Vinod is VP of Products at H2O.ai. He leads all product marketing efforts, new product development and integrations with partners. Vinod comes with over 10 years of Marketing & Data Science experience in multiple startups. He was the founding employee for his previous startup, Activehours (Earnin), where he helped build the product and bootstrap the user acquisition with growth hacking. He has worked to grow the user base for his companies from almost nothing to millions of customers. He’s built models to score leads, reduce churn, increase conversion, prevent fraud and many more use cases. He brings a strong analytical side and a metrics driven approach to marketing. When he is not busy hacking, Vinod loves painting and reading. He is a huge foodie and will eat anything that doesn’t crawl, swim or move.

Leave a Reply

Enhancing H2O Model Validation App with h2oGPT Integration

As machine learning practitioners, we’re always on the lookout for innovative ways to streamline and

May 17, 2023 - by Parul Pandey
Building a Manufacturing Product Defect Classification Model and Application using H2O Hydrogen Torch, H2O MLOps, and H2O Wave

Primary Authors: Nishaanthini Gnanavel and Genevieve Richards Effective product quality control is of utmost importance in

May 15, 2023 - by Shivam Bansal
AI for Good hackathon
Insights from AI for Good Hackathon: Using Machine Learning to Tackle Pollution

At H2O.ai, we believe technology can be a force for good, and we're committed to

May 10, 2023 - by Parul Pandey and Shivam Bansal
H2O democratizing LLMs
Democratization of LLMs

Every organization needs to own its GPT as simply as we need to own our

May 8, 2023 - by Sri Ambati
h2oGPT blog header
Building the World’s Best Open-Source Large Language Model: H2O.ai’s Journey

At H2O.ai, we pride ourselves on developing world-class Machine Learning, Deep Learning, and AI platforms.

May 3, 2023 - by Arno Candel
LLM blog header
Effortless Fine-Tuning of Large Language Models with Open-Source H2O LLM Studio

While the pace at which Large Language Models (LLMs) have been driving breakthroughs is remarkable,

May 1, 2023 - by Parul Pandey

Request a Demo

Explore how to Make, Operate and Innovate with the H2O AI Cloud today

Learn More