Fujitsu streamlines GPUs to double deep learning accuracy

Fujitsu Laboratories has streamlined the internal memory of GPUs to support neural networks in a development which has doubled the accuracy of machine learning.

Fujitsu streamlines GPUs to double deep learning accuracy

Recent years have seen a focus on technologies that use GPUs for high-speed machine learning to support the huge volume of calculations necessary for deep learning processing.

In order to make use of a GPU’s high-speed calculation ability, the data to be used in a series of calculations needs to be stored in the GPU’s internal memory. This, however, creates an issue where the scale of the neural network that could be built is limited by memory capacity.

Fujitsu Labs has now developed technology to improve memory efficiency, implementing and evaluating it in the Caffe open source deep learning framework software.

Upon commencement of learning, the technology analyzes the structure of the neural network, and optimizes the order of calculations and allocation of data to memory, so that memory space can be efficiently reused.

With AlexNet and VGGNet, image-recognition neural networks widely used in research, this technology was confirmed to enable the scale of learning of a neural network to be increased by up to roughly two times that of previous technology, thereby reducing the volume of internal GPU memory used by over 40%.

This technology makes it possible to expand the scale of a neural network that can be learned at high speed on one GPU, enabling the development of more accurate models.

Fujitsu Laboratories aims to commercialize this technology as part of Fujitsu Limited’s AI technology, Human Centric AI Zinrai, to work with customers in the use of AI.

In recent years, deep learning has been gaining attention as a machine learning method that emulates the structure of the human brain. In deep learning, the more layers there are in a neural network, the more accurately it performs tasks, such as recognition or categorization.

In order to increase accuracy, the scale of neural networks has been growing, but this lengthens learning times. Along with this, more attention is being placed on GPUs that execute computations with large volumes of data, and technology that accelerates the process by using multiple GPUs in parallel, as with supercomputers.

One method of increasing the scale of deep learning is to distribute a single neural network model across multiple computers and do the computations in parallel, but the volume of data that must be transmitted in exchanges between computers then becomes a bottleneck, greatly reducing learning speed. In order to take full advantage of the GPU’s high-speed calculation capability, to the extent possible the data to be used in series of calculations needs to be stored in the GPU’s internal memory.

However, as GPU memory is usually smaller than that of an ordinary computer, there had been the issue of limitations in scale of neural networks capable of high-speed learning.

Fujitsu Labs’ development streamlines memory efficiency to expand the scale of a neural network for computations with one GPU, without using parallelization methods that greatly reduce learning speed.

This technology reduces the volume of memory by enabling the reuse of memory resources; it takes advantage of the ability to independently execute both calculations to generate the intermediate error data from weighted data, and calculations to generate the weighted data error from intermediate data.

When learning begins, the structure of every layer of the neural network is analyzed, and the order of calculations is changed so that memory space in which larger data has been allocated can be reused.


Leave a Reply

Your email address will not be published. Required fields are marked *

*