- Community Home
- >
- Software
- >
- HPE Ezmeral: Uncut
- >
- How to accelerate model training and improve data ...
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark as New
- Mark as Read
- Bookmark
- Receive email notifications
- Printer Friendly Page
- Report Inappropriate Content
Re: How to accelerate model training and improve data science productivity with GPUs
Artificial Intelligence (AI) is todayโs hot topicโand hot technology.
From news headlines to trade shows to boardrooms, everyone is talking about how AI can be used to transform entire industries and deliver groundbreaking business innovations. The use cases may be as mundane as automating the interaction with customers during entry-level support requests, or as sophisticated as features that assist drivers in avoiding accidents on the road. For each of these use cases, software development and data science teams are at work behind the scenes, supported by operations teams and IT infrastructure.
Over the past few years, Iโve spoken with many data science teams. I always ask them about their challenges and what they could use to increase their productivity. Ultimately they want to operationalize their work more quickly, and thus connect it to business processes for various use cases to drive business outcomes. In this context, what they mean by operationalize is taking the components of a machine learning/deep learning or predictive analytics model (i.e. the code, scripts, libraries, and meta-data) and deploying it into a running state in โproduction.โ
Slow model training processes hinder productivity
The answers Iโve received are often specific to each teamโs unique circumstances and environment. But some consistent patterns come up repeatedly. In this blog, Iโm focusing on one common to all data science teams: frustration with slow model training processes. This is top of the list I suspect because itโs something that severely limits the amount of work that data science teams can take on, not to mention the quality of the outcomes they can deliver.
Model training processes can be slow for several reasonsโbut an important one is the need for access to the right compute resources. In an enterprise environment, the compute resources for model training must be able to process huge amounts of complex data (e.g. petabytes of images), which present punishing loads even for some of the best CPU-based systems which are commonly available to data science teams.
GPUs accelerate machine learning and deep learning
Consider this easily accessible solution to this: use specialized compute resources and accelerators that are tuned and optimized for just these scenarios. The most common examples are GPUs, which were originally designed to support high-quality, real-time graphics. Since then, GPUs have evolved to be the computing accelerator of choice for compute-intensive applications such as machine learning and deep learning. Running the same model through a GPU versus a traditional CPU can result in processing time going from hours to minutes.
If this is the case, then why donโt organizations have GPUs available to everyone that needs them?
The short answer is that they try to. But in large enterprises with multiple data science teams, itโs not always that easy. GPUs are in high demand, infrastructure procurement can take months, and thereโs no simple mechanism for sharing the existing GPU resources. The most common scenario is that the back-end infrastructure grows organically over time with siloed data science teams hoarding their personal GPU environments.
The problem with this is that due to the cyclical nature of a data scientistโs work and the binary model training (100% or 0% GPU utilization) workloads, you could end up with your most valuable assets sitting idle while other teams have projects that could benefit from the 100x GPU speed-up. The potential impact is significant: slower model development and reduced productivity from your data science teams due to underutilization of existing GPU investments.
So what can you do about this? Watch my educational talk below, where I discuss this scenario and how HPEโs BlueData software can help your enterprise get the most out of your GPU resources to accelerate model training processes and improve the productivity of your data science teams.
Matt Maccaux
Hewlett Packard Enterprise
twitter.com/HPE_AI
linkedin.com/showcase/hpe-servers-and-systems/
hpe.com/us/en/solutions/artificial-intelligence.html
- Back to Blog
- Newer Article
- Older Article
- Back to Blog
- Newer Article
- Older Article
- SFERRY on: What is machine learning?
- MTiempos on: HPE Ezmeral Container Platform is now HPE Ezmeral ...
- Arda Acar on: Analytic model deployment too slow? Accelerate dat...
- Jeroen_Kleen on: Introducing HPE Ezmeral Container Platform 5.1
- LWhitehouse on: Catch the next wave of HPE Discover Virtual Experi...
- jnewtonhp on: Bringing Trusted Computing to the Cloud
- Marty Poniatowski on: Leverage containers to maintain business continuit...
- Data Science training in hyderabad on: How to accelerate model training and improve data ...
- vanphongpham1 on: More enterprises are using containers; hereโs why.
- data science course on: Machine Learning Operationalization in the Enterpr...