The 5 Key AI Takeaways for Today's C-Suite
Read the Full Transcript
Patrick Moran:Â Hello everybody. Thank you for joining us today. My name is Patrick Moran. Iâm from the marketing team at H2O.ai. And before we get started, weâd like to do a quick sound check. If you could please type into the questions tab if you are able to hear us clearly. Once we get a few responses, weâll go ahead and get started.
Ingrid Burton: Great. Well, thank you everybody. Thank you for joining us here this morning, this afternoon or this evening, depending on where youâre at. Iâm Ingrid Burton; Iâm the CMO and an AI transformation leader at H2O.ai. And I have with me today Vinod Iyengar â Iâm the lead of data transformation here at H2O.ai. Iâm really excited to be here. So what we wanted to do today is share with you some key takeaways for the C- Suite. Weâll share what weâve heard from a number of customers and people we talked to in the industry, as well as some key AI takeaways. So how do you embark upon an AI strategy if your company has not started? That seems to be a common theme today, as we move to seeing more AI and machine learning use across enterprises.
A lot of business leaders, um, from the CIO to the CTO to the chief analytics officer and even the CEOs have a lot of questions on âWhat do I do next?â So today, what we wanted to do is share with you some of the key challenges, the opportunities and just five things you should start looking at or thinking about. So with that, Iâm going to take it away.Â
In terms of the challenges today in AI, weâre at an all-time peak in terms of interest that we see in the industry. There are a lot of reports and news out there around AI. But the bottom line for many businesses today is that there are three key challenges in embarking upon an AI strategy. First of all, who? Where are you going to get the people? Thatâs the talent question.
Where are the data scientists? How do I embark upon this journey, if I donât even have a data scientist on my team. So thatâs a main question that reoccurs regularly; weâre going to talk about that today. The second thing is maybe you have a data scientist, or maybe you have a team. The element of time comes into play. So itâs really important to understand that itâs not a quick fix for your business. There isnât a fast way to just get an AI strategy and get results instantly. We would say that thereâs a time element, and we want to address that time element today and talk about how you get to faster results, because itâs all about getting business results. Itâs not about AI per se, itâs about what you can do with it.
So how do we cut the time to get into results? Time is a challenge today. Then there are a lot of questions around trust. How do I trust the models? How do I trust an AI? How do I explain this to regulators? And weâll have an opportunity to discuss that today as well, and we want to address that. So the talent, time, and trust challenges that exist today for businesses when they think about an AI strategy, are clearly top of mind. Weâve met with a lot of CEOs and a lot of chief analytics officers, chief data officers, and people who are in the Fortune 500 and Fortune 100. One of the things that we clearly have heard is that you have to start with a data culture.
When you start thinking about AI, you canât get to an AI result unless you have data, and you canât get to that unless you really embed data and a data culture within your company. We have regularly been looking at data for decades. A lot of you have gone through a data transformation, and data just exists everywhere. So how do you tap into that data? How do you get to a data transformation â a digital transformation â then harness it for AI? It really is a key challenge in order to not only think about it, but to understand who in your company is going to embark upon this AI transformation. We talk about data as a team sport, so why donât we elaborate on that a little bit?Â
Vinod Iyengar:Â Itâs a great point, right? We are living in a time where thereâs a lot of data being collected. People are collecting data from their marketing sources, from sales mechanisms, from the data products, from product launches, etc. You get customer analytics coming in through all these different sources. So we have the ability to collect data. So thatâs the focus, right? So building a data culture means you proactively start collecting data first and putting it in a format which is easily then accessible. To create a data culture, you need to be able to make it really easy for people to see the data and actually get insights from it. Now, how do you do that? To do that, you need to train the people on the team to be able to take advantage of data. Now you donât need PHD-level data scientists machine learning people from day one. When you think of a team sport, what we mean is that you need people of all different types of skills; you need to build a whole slew of people. You do of course want experts who can build really sophisticated models and regard them, but you also need people who can just look at the data and quickly get a sense of whatâs useful or not.
So when talking about a team sport, weâre talking about building people with all kinds of skills and also up-leveling existing people. So your existing marketers, your existing finance people, your existing product people â can they become data aware or data science-like? So how do you do that? Just giving them the basics of data science, like telling them how to make assumptions and ask the right questions â weâll come to that in a bit. But itâs about creating that sort of culture â up-leveling everyone to be data-aware, and also to start asking this question from your data.
Ingrid Burton:Â We were talking recently to a CEO and a CIO of a company in L.A. They were starting from the standpoint of the fact that they had data, but their company, their culture was data averse. And so they had to collect it all into a single place. They had to create dashboards. So they just started on their journey and now theyâre going into AI. But they really were at this point where the people that had been around for decades within the company said, âOh, I know my gut feel, I know whatâs going on in the market. I know my customers.â And when presented with data, it turned out that they were off by a few percentage points and more. So really embracing data as part of your culture is important. We look at the cultural transformation of getting to AI. There are a number of reports that say this now, is really a key driver.Â
You really have to understand that as a member of an executive team, you canât just mandate AI; you canât just mandate digital transformation â I think a lot of you know that. What you have to start embracing is, âWhat can you glean from the data? What can you find out from data that also augments that gut feel that we all have.?â I mean, weâve all been there, right? âI think I know best.â Data can tell me that Iâm a little off or that my gut feel is correct. So itâs good to know that. So when you think about whoâs on your team, we look at it as kind of a triumvirate, so to speak. Thereâs the business leader, right? The business leader is probably one of you in the c-suite; itâs the CMO, itâs the chief revenue officer, or itâs the officer of risk management at a bank. They are trying to solve a business problem. Weâll talk about that a little bit. Weâre not doing this because weâre just having fun. Weâre doing this because weâre trying to solve business problems, and to get answers to the business. They rely upon a data science team or an analytics team. So the chief data officer, or the chief analytics officer comes into play here. They want to know, âHow do I arrive at the best answer for the business?â So they have to work in concert together. And then of course IT is important; itâs so imperative, because you canât just drop new software or new processes into an existing system without IT really knowing about it. It becomes a real problem. So the CIO needs to be involved. This is the culture that you need to start creating. This is a team sport; people have to work together. They have to realize that theyâre deciding on outcomes. Weâre not trying to mandate AI, weâre not trying to mandate digital transformation, but weâre trying to look for answers so that a company can get to a competitive edge.Â
So the first key takeaway here is: build a data culture. Takeaway #1: Whoâs on the team? We also found that you might have the right talent within your organization already; you probably already have really smart, excellent people that want to be part of that team. This is something that we find to be true in almost every company that we talk with, from large telcos to media and entertainment companies, to banks. They have really smart people that want to become well-versed in becoming part of this data culture. And of course, as part of this data culture: data, data, data â from every facet.
Vinod Iyengar:Â Yes, and if you think about it from a tactical perspective, what weâve seen with some of the best mature organizations is that they didnât start a lot earlier; they started maybe a few years ahead of you. But all they did is create central banks of data science people who are data aware, and then build around them. So you can do it two ways. One is if you look at the trim rate that we talked about in the previous slide, you can have a cross-functional data science team that is basically permeating all over the organization. Our CEO always likes to say, put a data scientist in every office.Â
So what that means is that you have data-aware thinking permeating around the entire group because of this one person who might be an expert. And that way you can slowly build a mentor/mentee model and have more and more people surrounding them who start acquiring data science skills and become users of data and consumers of data, and then they start going one step further and are able to create insights from the data.
The second way to do this is to uplevel your existing teams. So you have people who are BI analysts, for example. These are people who have already been working with data. They know the structure of the data, and how different things can be combined together, so you can possibly send them to bootcamp. You can send them to one of our conferences, or send them to an online training course like Coursera. Give them the opportunity to uplevel, and youâll find that some of these people already working with data, and they will be able to pick this up really quick.
That way, you can sort of shortcut the problem by sort of up-leveling those in an existing workforce who are always trying to get to the next stage in their career. So there are different ways to do this; choose whatever works best for you.Â
Ingrid Burton: So the key takeaway is to build a data culture: figure out whoâs on the team, find the right talent within, and be very data- focused. The key takeaway #2 is that you really have to ask the right questions. If youâre a major bank or a retailer, what are they trying to solve for right now? Recommendations, offers, personalization, know their customer; they all have these questions. Every major bank wants to learn how to detect fraud faster.
How do we do KYC, which is know your customer? How do we thwart money laundering? How do we get to those questions? Yes, there are early adopters in many different industries, but there is so much low-hanging fruit in terms of just finding the answers to questions you already have. You already have these questions. How do I get that next customer? Who is that next customer? How do I optimize my supply chain? How do I cut out transportation costs in the supply chain? You can get down to very granular, detailed questions, so asking the right questions is imperative. So part two, when thinking about creating an AI strategy is that you need the right people. You need analytical people â mathematicians, statisticians; these are not computer scientists per se.
So this is a different skill set. These are problem-solvers. These could be business analysts that are ready to be part of an AI strategy within your company. I would contend, and I always like to put this in, I think youâve got to be really creative. You have to think outside the box. Youâve got to find people: you, yourself, and your company, have to think about creative solutions, meaning the same old way isnât the way forward, or maybe itâs part of the way forward, but youâve got to be very creative and open, really open minded and of course you really need be data-driven.Â
Your data is everywhere. Sometimes companies donât realize how much data theyâre generating. Look to the marketing department for sentiment analysis of whatâs going out on, on social. Look at the finance department, look at the transactional data that you have. Look at the engineering department or support functions. Look at the innovation functions within your team to hear feedback from your customers. Data is everywhere. And so you can harness it to answer the questions that you may have. Finding that next customer, making a recommendation â all of that is possible with machine learning. All of that is possible with AI in terms of getting closer to the pin, so to speak. You need to make sure that youâre spending your money wisely and getting those answers.Â
So weâre going to talk a little bit about use cases. If youâre embarking upon an AI strategy and youâre in one of these industries (by the way, AI and machine learning are now in almost every industry that we see), take a look at some examples here of what customers are doing today. Credit cards, new credit, retail banking, wholesale banking, hedge funds management, bond trading â the list goes on. Weâve got customers in almost every segment of financial services.Â
In healthcare and life sciences there are some really amazing results happening in terms of sepsis detection. Better patient outcomes are being had because of AI. I personally am very passionate about that â if weâre going to save somebodyâs life with AI, it was all worth it. Telcos â predictive maintenance, avoiding truck rolls. Next best customer, customer churn prediction, marketing and retail, etc. You have to have AI and machine learning. Weâre working with a number of retailers right now with digital marketing firms on next best customer, funnel predictions, personalization; itâs all very important.Â
Whatâs the point of all this? Save time, save money, and gain a competitive edge. Whatâs happening right now within our own customer base is if you arenât doing many of these use cases right now in your particular vertical, you may be left out, and we donât want that to happen. These are use cases, but they are problem solving, right? Weâre using our algorithms to provide you with the answers. You have a problem. Weâre giving you a prediction with our technology.Â
Vinod Iyengar:Â I know that itâs important to ask that question. Typically, when youâre framing a experiment, youâre basically asking you a specific question. You always want to talk about what the hypothesis is that youâre testing, and youâre either trying to prove the hypothesis or disprove the hypothesis, right? That basically boils down to data science. You typically ask a question and try to prove whether the answer is true or not. And so that means that you had asked the right question. For example, at one of the CEO summits, we heard someone jokingly say that management comes up and says, âHey, can you just throw AI on the problem?â and thatâs a good example of when you are not aware of what questions to ask, and you end up getting these naĂŻve questions.
We use these use cases to present the art of the possible. You can take fraud detection, for example. It can be applied across different verticals, right? But the question youâre asking is, is this transaction a fraudulent one or not? Is this person a fraudulent person or not?
If you frame the question in that sense, you know what data to collect, because the data you collect for a particular problem is very dependent on the question youâre asking. What if youâre trying to predict sales forecasts for the next 12 months? Then the data that you want to collect to ask that question is very specific. Youâre going to pull in all of your historical sales information, the customer information, the type of customers, and their interactions with your team. Once you have that information, then you can answer the question effectively, whether you can forecast the next 18 months or not, with some degree of conference. But again, the key is to frame the question. So how do you do that?Â
It goes back to the previous question. You have to keep educating your people to think about their high-level problems. As Ingrid said, every business wants to save money, make more money, and save time. So these are not different for any business, but how do you then âpeel the onion backâ to translate that high level goal to a question that you can ask of your data? So itâs having that culture of asking these questions and learning from what others have done. We just published 100 use cases, and those are just a way to start the conversation.
You might think, âOh, I have customer data, I have sales data. So what are the 10 questions you can ask to predict whether someoneâs going to buy a product, or whether someoneâs going to click on that path? How much time has someone spent on our web site? These are the questions that you could ask. Once you ask the question, you may then realize that you actually donât have all the data you need for that particular question. And that can trigger the question to the IT or data science person: âHow can I collect this data?â You might find that you need to do more, right? So asking the right questions is absolutely number one, because if you ask the right questions, that then determines what outcomes can be generated out of it. And then you can actually measure whether the experiments are successful or not. Because if you donât set the right outcomes, the right questions, your metrics may be all over the place. Worst case, they are probably wrong. So if you donât set the right metrics, the right successes, and the right outcomes, you are not going the right direction. You donât even know which direction youâre going in.
Ingrid Burton:Â The other thing that we hear a lot is, âTry a number of things. Ask a number of questions.â Donât just focus on just one thing, like, âHow do I add more customers?â Itâs almost too broad, right? So you need to ask a number of questions to get to the answers. So the key takeaway, number two, is to ask the right questions. Realize what problems are you trying to solve. Itâs not just about sprinkling a little AI on everything â thatâs not going to work. But what problem are you solving? Also determine outcomes. Measure your success and keep refining, right? You are never really done. Itâs just like everything else. Itâs a constant iterative approach. Determine outcomes. So if you want to minimize fraud in this part of our business, a good outcome would be to increase our detection rate by what, 2%, or 5% â put a measurement on it. Because then, youâre going to be able to measure that success. And youâre going to find out a lot once you start experimenting. So once your data science teams have the right data, and the business is asking the right questions, you can start to feed what youâre looking for. It starts to emerge very quickly in front of you. So asking the right questions is key.Â
So we have the also some other things that weâre weâre finding to be true, which is there are a lot of people that are in the same kind of situation where they are just starting, and they want to learn. Learning is all part of this. We make learning fun because machine learning and AI are just one facet. Asking questions is the other. But also learn from the community; send your people to be trained. There are so many online training options as we said, but there are ways to connect with the community online.Â
For instance, we have a huge meetup community. If you look up meetups and machine learning or you come to our website, h20.AI, you can learn a lot from other people. We share as much as information as we can. Itâs such a new field for many people. If you just share your knowledge and join a meetup or send your people to meetups and present, youâll find out thereâs great talent out there. Number one, itâs a great place to find talent, but youâll also learn so much from other peopleâs experiences. Whether itâs a mistake or success, weâre all going to learn. So what we encourage people to do is really join in the community, participate, and really become part of the movement as we see it. So key takeaway number three is connect with your local community. It doesnât matter where you are. There are community members in machine learning all over the world. We are seeing community involvement in Chile, Paraguay, Uruguay, London, Paris, Prague, all over the US and Canada, Mexico, Japan, Singapore, and India. So we want you to think about the fact that once you establish your teamâs data culture, you realize whoâs on the team. You want to ask all the right questions, share with others, participate with them, and learn from everyone. I think this is really critical for peopleâs success; the folks and the companies that share will actually benefit.Â
Vinod Iyengar:Â Iâm particular, Iâm really excited about this because this one is built on top of open source, and itâs very uniquely AI. In the last 10 years, most of the cutting-edge research has come from the academic community and from the open source community. So even large companies like Google, Facebook, or Microsoft- even with these largest companies, the latest cutting-edge research that comes out doesnât go into the product â it goes to open source; the community gets to try it out first. Thereâs a selfish reason for this because this space is so new; there are so many things that we are not aware of, especially when you get to some of these complex algorithms and technologies.Â
So you really want the community to do crowdsourcing and peer review and tell each other that this is really useful. What we then see are some amazing cross-disciplinary things coming out. For example, in the bioinformatics space, we are seeing some amazing research coming out there and in other sciences. So they are getting used in completely different verticals.Â
Ingrid Burton:Â So some of the algorithms from bioinformatics might get applied in marketing. Some of our customers are taking bioinformatics algorithms and applying it to marketing and sentiment analysis and into some of the banks. So itâs really interesting to see the cross-sectional kind of view of this.
Vinod Iyengar:Â Itâs really a fundamental function of what the data looks like and what the natural distributions look like in some of these industries. And thatâs really tough to know, even with experts, until you try different things and thatâs what the community allows you to do.Â
Thereâs also a second reason why you want to connect the community. Most of data scientists want to share and collaborate. If you try to build a closed community, youâll find it really tough to retain your data scientists; they really want to share, and thatâs how they learn and grow as well. I read this Tweet that said, âSpecific knowledge cannot be taught but it can be learned. Data science is that; itâs cutting-edge of research and itâs so new that you cannot really teach it in a university, but you can learn it from the community.
So if you close off your group, you donât connect the community, your peopleâs data science skills will atrophy. One great example was that five years ago, the algorithms that were popular were things like SPM. Today, there are a whole new batch of algorithms. If I came from school, learned the techniques but did not connect to the community, Iâd be stuck in a technique which was really old â that was barely five years ago. So you want to join the community. Keep in mind that the community can be multiple places â you can have meetups in diffent cities, you can sponsor a meetup.Â
Ingrid Burton:Â Thatâs right -you can get host meetups or even a hackathon in your office. A number of our customers are doing that and inviting others in. So theyâll do very public hackathons within their environment, and their data scientists and analytics people are all learning from people that are walking in. And you also getting students who are coming in. In terms of the talent issue, if you start holding these hackathons, youâll be bringing in top data scientists that are just about to graduate or could do an internship.Â
Vinod Iyengar:Â Unlike other spaces, interns actually might have more of the latest knowledge. You might have business knowledge, but an intern might know the latest algorithms. They can come and teach your team something that you might not be aware of. So connect with the local university, and maybe start an internship program when you have these data scientists coming in and interning with you.Â
Ingrid Burton:Â Every university out there is focused on this. We actually have a free academic program. Weâre giving out licenses to students, researchers, universities, and professors, and weâre working with them on creating curriculum. But also more importantly, what that means is weâre creating the talent pool. Helping people create the talent pool is really important to us. So if you connect to the community, youâre going to see more of these people out there, which is going to help you staff up regardless of who you are or where you are.Â
Thereâs also some great talent in the Midwest. I would say Latin Americaâs coming on strong as well, with a lot of data scientists. So the community is a wealth of information.
Letâs talk about that Kaggle Grandmasters briefly. Kaggle Grandmasters is hosted by Google. Kaggle holds regular competitions, and people vie for prize money and status, and they learn from each other. We were just talking to a Kaggle Grandmaster this morning who saids âI have a friend whoâs a Kaggle Grandmaster. Iâve actually never met him, but heâs in Melbourne, and he wants to work with us on XYZ.Â
As we all know, you go where the talent is, and Kaggle Grandmasters are in every part of the planet. So weâre really proud that we are supporting that. We actually have about 14 Kaggle Grandmasters at H20.ai that help customers get started. So back to connecting with the community; weâre here to help you, but also the Kaggle people themselves, a data scientists there, the universities want to get involved and also donât just look for data science curriculums. Look at people with math or physics degrees â our CTO here has a PhD in physics. Weâve gotten nuclear engineers here. These people are able to ask the right questions, be creative, think outside the box, and be fast and analytical. So connecting with the community is very important. Youâre going to find out that itâs not that you need to go hire a PhD in data science, or someone who has a Masterâs degree in data science.Â
By connecting with the community, youâre going to hear from such a broad section of people with diverse backgrounds, which is really important to get the answers. Weâre excited about the community. As Vinod said, open source is where we got our start. This community is unique in that itâs built on open source and we connect with everyone. We encourage you to do the same. So thatâs the key takeaway.
Weâre going to just jump right into takeaway number four. So in terms of technology considerations, how do you minimize the talent problem, the time problem, and the trust problem? There are technologies out there called automatic machine learning, auto ML. Thatâs what you should look for â automatic machine learning platforms. If youâre a CIO listening to this call or youâre in IT and youâre listening to this call, or youâre a business leader â you may not know where to begin or maybe you do. So there are a lot of different options out there. You have to think through, âAm I going to go open source? Do I have to run this solution in the cloud or on prem?â You can do both. Build or buy â do I build my own solution or do I buy one off the shelf? And of course, then youâve got to think about the data problem. So with that, Iâm going to turn it over to Vinod, who is the expert in this technology consideration.
Vinod Iyengar: There are a few different things to think about. If you are starting off fresh on this journey, the question is, should you go and buy a stack from a vendor or should you start with some open source libraries to build the knowledge? Thereâs no right or wrong. I highly recommend you do open source. Weâve cut our team on open source. Start with open source just to get a feel for how these are libraries work. The great news is that they are free. You can just download and try them and you can quickly get some value out of it and see that, okay, I can do some interesting things. Data scientists love them too because data scientists share their work through open source. They already probably know a Python library that they like, or they want to try out an R library, and itâs great.Â
But once you quickly get some wins, then the question is how do we know go to production? So the key with the dirty secret of data science is that 80% of most projects actually will never see the light of the day.Â
Ingrid Burton:Â 80% of the projects donât see the light of day. Thatâs because itâs an iterative process, right? Youâre never quite done.Â
Vinod Iyengar:Â Absolutely. But, but the other reason is technological, right? So oftentimes data scientists might pick a library which is deprecated or doesnât have support. When we say, âgoing into production,â what that means is your applications or your mission-critical systems are now relying on these models. When you reach that stage, your IT software engineers have a very different set of requirements that they are hoping for. So they are looking for liability, scale, performance, latency, and theyâre looking for support. So you start with open source, but youâll quickly realize that you might need a vendor to support me. So with my mission-critical applications, when thereâs an issue, thereâs someone I can call. And sometimes this sort of a maturity can be built internally, but oftentimes, youâll find that itâs better to just go with a vendor. For example, you can find a vendor who is making it out of an open source package.
For example, our H20.ai open source library gets used by over 18,000 organizations globally. But a small percentage of those, when they are ready to go to production and they have mission- critical applications running on them, they buy support from us. And what that gets them access to our data scientists, our support people, and basically 24/7 support if an issue happens on your production model.Â
The second thing to think about when it comes to prod and on- prem obviously is, how fast do you want to get started? So if you are coming into this journey really new, you might find that the cloud will get you running up and running very quickly, because you basically bypass or shotgun the whole DevOps journey of procuring hardware, putting software on top of it, maintaining that security infrastructure, and all those considerations.
So if youâre starting up fresh, cloud might be a good consideration to really jump the gun. On the other hand, if you already have a data center, you already have a reasonably mature DevOps practice, then you might just run it on prem. Obviously, you can save a lot of money by not going to the cloud because you already have your hardware, and itâs already amortized. So thatâs a consideration.Â
Often, we find that customers end up doing kind of like a hybrid study. After some point of time you might realize that you can guarantee that youâll have a certain amount of compute workloads for sure, but there are these spikes which you cannot model for. So you might use a hybrid strategy where you can jump to the cloud when you have spikes, but then use your on prem as much as possible. That way, you can optimize for your costs and also you donât have the heartburn of having to procure âx.âÂ
When it comes to build or buy, this is similar to the open source question. You do want to build some in-house expertise. And it often helps also hire people, because people want to contribute to packages and want to build new stuff. But at the same time, you want to have a good view on what the time value of money is. So the time to insight is critical for a lot of companies, especially in a competitive space. You often want to quickly get to an answer or solution quickly, and oftentimes, that means buying it. You can buy it, but also gain the expertise.
So you can basically work with the vendors like H20. With our customers, we donât just sell them software, we often teach their data scientists. We often run training programs; we give them tutorials, and we give them test drives. So they bought the product, but they also learn along with the using the product. And finally, data is of course critical, right? So, itâs about understanding how you save your data and how do you make it accessbile? So thereâs obviously a lot of conversations around data governance, data privacy, access, and lineage. So these are all critical pieces, especially with GDPR and the California privacy laws. You need to be very cautious about how you save your data and how you make your data accessible to your employees. So obviously, everyone should be able to look at the data and get insights out of it, but you should also have those guardrails to sort of put in place so that you donât accidentally leak data or you donât accidentally see something that youâre not supposed to.
Another thing is to understand how good your data is â the data quality piece comes into play. If your data has known issues or if there are missing values and a lot of errors, that might give you a bad results. Garbage in, garbage out, right? So itâs about creating processes so that you know good a particular set of data is that youâre going to be using for a particular set of models, and having a process to sort of keep checking that. So thatâd if you know that at some point the data is changing, you can sort of remodel or recalibrate your model.
Ingrid Burton:Â So these are all of the technology considerations that your data science and it teams are going to need to consider. As a business leader, you may or may not be thinking about it, but there are costs associated with it. Right now, nothing comes for free. If you want to get results, youâre going to have to invest. So you may already have all the infrastructure, you may already be on cloud, you may already have all of that. You may have all your data. But you still need to invest. And then, as Vinood said, one of the keys is that last mile, which is trust in the data and thinking about the data because of GDPR and the California â weâre going to see more of that across the US and across the world.
Itâs about the idea of trust in AI. So how do I trust the model? Whereâs the bias? Whatâs going on? You know, we hear itâs not fair. We hear we canât trust it. Weâre a regulated industry. So explainability is a big field right now in AI and, and I believe that h20.ai is at the forefront of leading the charge to really try and help people interpret and explain their models and give regulators the documentation that theyâre looking for. We could spend hours talking about fairness and bias in the data. The key point is humans. Youâre never going to take a human out of the loop. Thatâs, our assertion, right? You need, as itâs been said, garbage in, garbage out.
Humans put bias into data, right? This happens. Whatâs fair to you and I may be different to somebody else. So humans have to be in the mix to kind of oversee not only the AI but each other. And so when we talk about AI and data science and machine learning, we want the human element to be in there because humans still have the ability to look at an answer that a model might be giving them and say, wait a minute. Thatâs way out of balance from what Iâm thinking or that doesnât make any sense. Letâs talk about credit scoring. So letâs say you get a credit score, itâs phenomenal. And the AI says, âYouâre eligible to receive a $1 million loan right. Now, if you just let the AI go, that person might get $1 million, right? But you canât have that happen, right? The bank canât let that happen. They have to still check to see, âHow did we arrive at that model? Why did it indicate that about that particular individual?â So a human has to be in the loop in terms of looking at the models, understanding whatâs happening. You need to look at an auto ML platform (and we have one called Driverless AI), that helps you interpret that model, that documents the model, that provides reason codes to regulators and regulated industries, which include healthcare, banking, and other industries â fair lending, the Fair Housing Act. There are hundreds of regulations out there in the US and across Europe, and across the world.
You need to be looking at those regulations and be able to give a regulator or a lawmaker a reason why a particular model was scored the way it was. And thatâs a platform that does that automatically will be helpful, but also have a human in the loop making sure that youâre not missing something â youâve got to put some guardrails on it. So trusting an AI is a very complex subject, but itâs very simple when you think about it. Itâs âput a human in the loop from the beginning to the end of the process.â Put an automatic machine learning platform in there, and then really look at the documentation thatâs being generated. Look at the regulations that itâs adhering to. And really, the technology can help humans make better decisions.Â
Vinod Iyengar:Â Weâre talking about basically trust for humans, right? But really when you talk about trust, we are talking about human-centric trust. So what that means is, creating explanations that a human can understand and articulate. And that means that you have to document the results that youâre making, and then explain why those additions have been made. But this will often encourage really interesting things, like you might see that your data has bias in it. Just as an example, historically, say a certain group of people or a class of people may have not gotten credit because of different reasons. So if you start using the same model to build same data, to build future models, you will keep sort of reinforcing that bias and the data. This is where basically you might say, âIâm going to overrule the AIâs additions and other software and machine learning models additions because I know that the reason the model made the addition is because of past data. But I know that I want to sort of cover for the bias or sort of fix that part. A rule engine that does this is running the addition a machine learning model.Â
The other part of this goes back to the classic Occamâs Razor, right? Oftentimes the simplest explanation is often the right explanation. What that means in machine learning terms is that the simplest model that gets you the best accuracy is often the best model. So this is a great example that you see in Kaggle all the time. If you look at the Kaggle leaderboard for a competition, youâll find that the top sort of our 10% dialuup results are all within 0.1%. So they are all so close to each other, theyâre winning on like 0.01% decimals. Right? So, and at that point the question to ask is, âHey, is it worth the additional complexity to get an additional 0.01% lift in accuracy?â Most times, it is not. You go for the simplest model are there, which can get you 99% accuracy of the most complex model.
Another example is a Netflix competition about 10 years ago, when they put this whole massive data set out there and asked people to beat the recommendation engine. The winner built a really complex model which actually never got used. They actually picked the person who came in fifth or sixth, but they had a model which was much simpler, but good enough. So the concept of complexity can be significantly high, especially if you donât know what kind of edge cases that might be in your models and your data. Right? So itâs about picking simpler models which are more explainable, which have more constraints, which are a human-understandable and which can basically be overruled if needed. Thatâs what constitutes trust, and obviously this is a huge topic.Â
I recommend that you go to our website, which has tons of resources. Weâve done a lot of webinars on this topic.Â
Ingrid Burton:Â And we have a number of our customers have spoken on the topic at many of our events and weâll be talking about that even further. But go to our website and you can learn about explainable AI.Â
This is really important â youâve got to put it into your thinking when you embark upon an AI strategy. Because people are going to say, âI donât trust it. I donât know it. I donât get it.â Going back to the first thing â creating a data science or a data culture â youâve got to make sure that you know that that is addressed. Make sure that youâre asking the right questions, that youâre framing the questions appropriately.
Make sure youâre connecting to the community and being constantly learning. Think about the technology, and this is where the IT guys and gals, the data scientists, and the analytics people come in. Theyâve got to think through the long-term effect. Thereâs easy ways to get started in the cloud, but you have to decide if are youâre going to be in the cloud long term, or if youâre going to do a hybrid approach, etc. And then finally, trust in AI. Those are the five key takeaways. Really, youâve got to think about trust in AI. It goes back to actually trusting your team, being part of the culture, and saying, âYes, we get it.â
So this is what weâve learned over the last several months, talking to a number of CDOs, chief data officers, CIOs, chief analytics officers, business leaders, across many different industries, and across the world.Â
So where do you go from here? Whatâs your next move? You need get started. What problem are you solving? What technology do you need? Whoâs on the team? How do you get the support to get it done? If youâre the executive sponsor, youâve got to bring a team together around you. Itâs not just you in your office with your team; you have to bring a cross- functional team in. Weâve talked to CDOs that go and talk to their boards about this. This is a critical board issue right now in terms of making sure that theyâre getting the latest and greatest and keeping on the cutting edge.Â
So we want to continue the conversation. Weâd love to have you join us at an event or meetup â contact us! H20.ai is probably the fastest way to get ahold of us. Have your people try our platform. We have easy ways to get started. We have tutorials and free trials. We do POC pilots all the time. Weâve got data scientists that would love to share what they know. Weâve got our own customers who would love to talk to you about this. The more we all evangelize whatâs going on in the market today, the more weâre going to get more of a movement here and more people will be successful.
Â
So with that, there are some questions. excellent.Â
Â
Question:Â Do you find more customers coming to you to apply new tools to a well-defined business problem or more who are looking to see what things might be possible with new tools and techniques? It seems that the former is the much easier challenge to address.Â
Answer: Obviously the former would be low-hanging fruit, right? You already have a problem or question. If you have a question that you want to answer, then you have an equally good idea of what you want to do with it. So we do see that quite a bit. But on the other hand, we also see more of âIf you build it, they will come.â So buying licenses and asking us to come in and set up an infrastructure. And then giving it to data scientists. People who are data-aware think that âHey, weâve got H20 â letâs connect it to our data sources. We can also come in and do training in person, in webinars, or online training, and getting the teams at the ops level to say, âHey â itâs all connected. Itâs all set up. Now just go ahead and play with it.Â
Vinod Iyengar:Â And I would say itâs a mix, right? Itâs never 100% this way or that way. Some of the companies that weâre working with, they just come up with a new set of problems that theyâre trying to address that they havenât been able to solve in the past; that sometimes works out really well, too.Â
Ingrid Burton:Â Weâre willing also to come in and talk to your executive team. Our CEO talks to a number of boards and he gets invited all over the place to talk to boards and executive teams. Weâd love to see more adoption of AI, machine learning, and automatic machine learning. Weâre here to help. Thank you to everybody who joined us today. The presentation slides and recording will be made available through our Bright Talk channel.
Speaker Bios
Ingrid Burton:Â Ingrid Burton is CMO at H2O.ai, the open source leader in AI. She has several decades of experience leading global marketing teams to build brands, create demand, and engage and grow communities. She also serves as an independent director on the Aerohive board. Prior to H2O.ai she was CMO at Hortonworks, where she drove a brand and marketing transformation, and created ecosystem programs that positioned the company for growth. At SAP she co-created the Cloud strategy, led SAP HANA and Analytics marketing, and drove developer outreach.
She also served as CMO at Silver Spring Networks and Plantronics after spending almost 20 years at Sun Microsystems, where she was head of Sun marketing, led Java marketing to build out a thriving Java developer community, championed and led open source initiatives, and drove various product and strategic initiatives. A developer early in her career, Ingrid holds a BA in Math with a concentration in Computer Science from San Jose State University.`
Vinod Iyengar:Â Vinod is VP of marketing and technical alliances at H2O.ai. He leads all product marketing efforts, new product development and integrations with partners. Vinod comes with over 10 years of Marketing & Data Science experience in multiple startups. He was the founding employee for his previous startup, Activehours (Earnin), where he helped build the product and bootstrap the user acquisition with growth hacking. He has worked to grow the user base for his companies from almost nothing to millions of customers. Heâs built models to score leads, reduce churn, increase conversion, prevent fraud and many more use cases.
He brings a strong analytical side and a metrics driven approach to marketing. When he is not busy hacking, Vinod loves painting and reading. He is a huge foodie and will eat anything that doesnât crawl, swim or move.