Customize and deploy open source AI models, create your own digital assistants and business GPTs.
Open weight small vision-language models for OCR and Document AI.
Industry and Use Case AI Apps
From Credit Scoring and Customer Churn to Anti-Money Laundering
From Clinical Workflow to Predicting ICU Transfers
From Claims Management to Fraud Mitigation
From Predictive Maintenance to Transportation Optimization
From Content Personalization to Lead Scoring
From Assortment Optimization to Pricing Optimization
From Predictive Customer Support to Predictive Fleet Maintenance
Gain expertise through engaging courses and earn certifications to thrive on your AI journey.
Get help and technology from the experts in H2O and access to Enterprise Team
Read the H2O.ai wiki for up-to-date resources about artificial intelligence and machine learning.
Learn the best practices for building responsible AI models and applications
By H2O.ai Team | minute read | June 07, 2013
Chocolate Cake (Wednesday, June 5, 2013)
You know how sometimes you have one bite of really good chocolate cake, or a really amazing peach and totally assume that you could eat another 30lbs of whatever without regard for good manners or physical limitations? Yeah. Decreasing marginal returns dictate that it almost always turns out that the last bite isn’t as good as the first one – having a little and having a lot are different.
Similarly, ingesting 1000 bytes of data and 1 byte are pretty different, and when you’re used to little bytes and start fooling around with the big ones the differences might not be immediately obvious or intuitive (maybe they are, and if that’s the case – awesome! Now go eat your cake).
When I’m trying to make sense of a problem I like to start with a small example and work through it to get a feel for the mechanics. With Big Data this gets a little weird, since we’re almost always mining, so we don’t always know well what to look for or expect, and because we need some intuition for how to get from the small to the big. To help that, I am trying to build some intuitive explanations. You can look at them topically under the posts beginning with header “Big vs. Little…”
Sometimes it is the case that using H2O to look at small data sets really makes no sense for whatever reason. In those cases we’ll talk about why, and I’ll use R for comparison. I’ll also provide you with relevant output for each (so that you can see how to get from one to the other). If you’re not familiar with R go here . Additionally, it’s worth mentioning that I’m tackling one set of assumptions at a time, so in general I’ll work as though we are going through some ad-hoc analysis instead of post-hoc analysis. There are some super cool differences between mucking vs. mining, but I want to talk about those separately.
At H2O.ai, democratizing AI isn’t just an idea. It’s a movement. And that means that it requires action. We started out as a group of like minded individuals in the open source community, collectively driven by the idea that there should be freedom around the creation and use of AI.
Today we have evolved into a global company built by people from a variety of different backgrounds and skill sets, all driven to be part of something greater than ourselves. Our partnerships now extend beyond the open-source community to include business customers, academia, and non-profit organizations.
Make data and AI deliver meaningful and significant value to your organization with our platform.