Advertising

Monday 17 October 2016

Unknown

The Core Technologies of Deep Learning

When the movie The Terminator was released in 1984, the notion of computers becoming self-aware seemed so futuristic that it was almost difficult to fathom. But just 22 years later, computers are rapidly gaining the ability to autonomously learn, predict, and adapt through the analysis of massive datasets. And luckily for us, the result is not a nuclear holocaust as the movie predicted, but new levels of data-driven innovation and opportunities for competitive advantage for a variety of enterprises and industries.
HPC Core Technologies of Deep Learning
Artificial intelligence (AI) continues to play an expanding role in the future of high-performance computing (HPC). As machines increasingly become able to learn and even reason in ways similar to humans, we’re getting closer to solving the tremendously complex social problems that have always been beyond the realm of compute. Deep learning, a branch of machine learning, uses multi-layer artificial neural networks and data-intensive training techniques to refine algorithms as they are exposed to more data. This process emulates the decision-making abilities of the human brain, which until recently was the only network that could learn and adapt based on prior experiences.

Deep learning networks have grown so sophisticated they’ve begun to deliver even better performance than traditional machine learning approaches. One advantage of deep learning is that there is little need to "train" the system and define features that might be useful for modeling and prediction. With only basic labeling, machines can now learn these features independently as more data is introduced to the model. Deep learning has even begun to surpass the capabilities and speed of the human brain in many areas, including image, speech, or text classification, natural language processing, and pattern recognition.

HPC hardware platforms of Deep Learning

The core technologies required for deep learning are very similar to those necessary for data-intensive computing and HPC applications. Here are a few technologies that are well-positioned to support deep learning networks.

Multi-core processors:
Deep learning applications require substantial amounts of processing power, and a critical element to the success and usability of deep learning comes with the ability to reduce execution times. Multi-core processor architectures currently dominate the TOP500 list of the most powerful supercomputers available today, with 91% based on Intel processors. Multiple cores can run numerous instructions at the same time, increasing the overall processing speed for compute-intensive programs like deep learning, while reducing power requirements, increasing performance, and allowing for fault tolerance.

The Intel® Xeon Phi™ Processor, which features a whopping 72 cores, is geared specifically for high-level HPC and deep learning. These many-core processors can help data scientists significantly reduce training times and run a wider variety of workloads, something that is critical to the computing requirements of deep neural networks.

Software frameworks and toolkits:
There are various frameworks, libraries, and tools available today to help software developers train and deploy deep learning networks, such as Caffe, Theano, Torch, and the HPE Cognitive Computing Toolkit. Many of these tools are built as resources for those new to deep learning systems, and aim to make deep neural networks available to those that might be outside of the machine learning community. These tools can help data scientists significantly reduce model training times and accelerate time to value for their new deep learning applications.

Deep learning hardware platforms:
Not every server can efficiently handle the compute-intensive nature of deep learning environments. Hardware platforms that are purpose-built to handle these requirements will offer the highest levels of performance and efficiency. New HPE Apollo systems contain a high ratio of GPUs to CPUs in a dense 4U form factor, which enables scientists to run deep learning algorithms faster and more efficiently while controlling costs.

Enabling technologies for deep learning is ushering in a new era of cognitive computing that promises to help us solve the world’s greatest challenges with more efficiency and speed than ever before. As these technologies become faster, more available, and easier to implement, deep learning technologies will secure their place in real-world applications – not in science fiction.

Unknown

About Unknown -

High performance with simplicity and value - HPC Server Platforms designed to simplify HPC deployment and management. HPC, Scalable Solutions for Any Size Problem​ - for advanced technologies in any market with high-performance, high-density requirements. Highest performance for the lowest cost. You make the choice. HPC Appliance Aims to Boost Engineering Productivity. Get the Latest Industry News, Daily! Subscribe now to receive daily updates of the latest news, trends and features about the data center industry by email. It’s free and easy.

Subscribe to this Blog via Email :