Abstract
Training AI models is usually more demanding than training standard ML models because they are processing intensive and often the data sets involved are larger. That is why if you are serious about deep learning you have to have access to GPUs. In Azure there are a number of ways you can make use of GPUs, on single VMs or in orchestrated clusters of them. In this chapter, we summarize several of the most common methods available as well as the pros and cons of each. Then we expand on the code we wrote in Chapter 6, which used a VGG-like CNN to tackle the CIFAR10 data set using the DLVM as the computing environment. In this chapter, we extend to other training options such as Batch AI and Batch Shipyard, which can both be useful for scaling up or scaling out training. We finish by highlighting briefly some of the other methods of training AI models on Azure that are not as common but might be useful depending on the problem at hand.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Order and numbers are based on single precision FLOPS; cards with two chips are treated as individual GPUs.
- 2.
All the steps are detailed in the notebook Chapter_09_01.ipynb which can be found in the Chapter_09 folder http://bit.ly/CH09Notebooks.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2018 Mathew Salvaris, Danielle Dean, Wee Hyong Tok
About this chapter
Cite this chapter
Salvaris, M., Dean, D., Tok, W.H. (2018). Training AI Models. In: Deep Learning with Azure. Apress, Berkeley, CA. https://doi.org/10.1007/978-1-4842-3679-6_9
Download citation
DOI: https://doi.org/10.1007/978-1-4842-3679-6_9
Published:
Publisher Name: Apress, Berkeley, CA
Print ISBN: 978-1-4842-3678-9
Online ISBN: 978-1-4842-3679-6
eBook Packages: Professional and Applied ComputingApress Access BooksProfessional and Applied Computing (R0)