GPU-Powered Deep Learning Emerges to Carry Big Data Torch Forward
The big data analytic ecosystem has matured over the last few years to the point where techniques such as machine learning don’t seem as exotic as they once were. In fact, consumers have even come to expect their products and services to be adaptive to their behavior, which is one of the hallmarks of machine learning. As we look to the future, the field of deep learning on GPUs is poised to carry the big data analytics torch forward into the realm of artificial intelligence.
While the theories behind deep learning are not new, the real-world application of deep learning in the modern sense is a relatively recent phenomenon that’s emerged in a big way over the past few years. Its rise is closely tied to the advent of modern GPUs, a fact that NVIDIA (NASDAQ: NVDA) co-founder and CEO Jen-Hsun Huang hammered home in his keynote at the GPU Technology Conference (GTC) last week.
“Five years ago the big bang of AI happened,” Huang said at GTC last Tuesday. “Amazing AI computer scientists discovered new algorithms that made it possible to achieve levels of results and perception using this technique called deep learning that nobody had ever imagined.”
At a high level, deep learning involves the use of layers of algorithms that “extract high-level, complex abstractions as data representations through a hierarchical learning process,” according to a 2015 article in the Journal of Big Data. “Complex abstractions are learnt at a given level based on relatively simpler abstractions formulated in the preceding level in the hierarchy.”
Various deep learning architectures–such as deep neural networks, convolutional deep neural networks, deep belief networks, and recurrent neural networks—are useful for tackling well-defined big data processing problems, including computer vision, speech recognition, text mining, natural language processing, audio recognition, and bioinformatics. But those aren’t the only ways that deep learning will be applied, which is what makes this field so interesting.
What makes deep learning so powerful, and what makes its future so incredibly bright, according to NVIDIA’s Huang, is how universally applicable the new algorithms are to a host of big data problems.
For example, consider AlphaGo, the program that Google’s DeepMind subsidiary created to play the ancient Chinese game of Go. Researchers “taught” AlphaGo how to play Go by using an artificial neural network that essentially to play millions and millions of rounds of Go. Eventually, the system figured out the best strategies to play the game (which features a nearly infinite number of possible moves) well enough to beat one of the best Go players in the world.
The interesting bit is that the Chinese Web giant Baidu (NASDAQ: BIDU) used the same basic approach to train its natural language processing service, called Deep Speech 2, how to recognize different languages. This is significant, according to Huang, because it shows that one approach is driving breakthroughs in different areas.
“Using one general architecture, one general algorithm, we’re now able to tackle one problem after another problem after another problem,” he said. “The fundamental difference is this. In the old approach, programs were written by domain experts and it took time and refinement. In this new approach, you have a general algorithm called deep learning and what you need is massive amounts of data and huge amounts of high performance computing horsepower. It’s a fundamental different approach for developing software.”
So far, deep learning has primarily been the domain of major tech firms, such as Google (NASDAQ: GOOG) and Baidu (NASDAQ: BIDU) which employ the algorithms on massive GPU-powered clusters to power various services they make available over the Web, such as image recognition or speech recognition. Other large companies like Amazon Web Services (NASDAQ: AMZN), Microsoft (NASDAQ: MSFT), and IBM (NYSE: IBM) have also made significant investments in deep learning. Facebook (NASDAQ: FB), which acquired AI startup MetaMind last week, is also exploring the area.
But like many other technologies that started out as research projects, deep learning is in the process of going mainstream. Thousands of AI startups have popped up in the last few years, and venture capitalists invested $5 billion in them, according to Huang. Underpinning many of these startups are investments in deep learning algorithms, which has emerged as the key technology driving the emerging artificial intelligence/cognitive computing industry that analysts say will be worth upwards of half a trillion dollars in the next few years.
Why AI Needs Big Data
This surge of interest into deep learning and AI is good business for CrowdFlower, the crowdsourced data labeling service that’s used by companies adopting machine learning. CrowdFlower founder and CEO Lukas Biewald says the fledgling AI industry is where the big data industry was several years ago.
“At this point, big data has settled into a place where people care about it and companies are using technologies related to it,” Biewald told Datanami in an interview earlier this year. “I think AI is a little less understood…Everybody’s trying to figure out what’s their AI strategy. If you’re a CIO or CTO at any company, you’re probably thinking what’s my AI strategy right now for 2016 and beyond.”
The advent deep learning techniques, when combined with the big data explosion and specialized processing environments such as NVIDIA’s GPUs, is giving people an incredible tool to make sense of the world, and people are trying to figure out how to use it.
“The coolest thing you can do with your data is AI, and AI needs big data to work,” Biewald says. “People are starting to realize that machine learning isn’t smarter than people. It’s just able to look at way more stuff than people. It’s attention span is a lot longer than a human’s…Now that companies are actually storing all these data, they can use it to train these algorithms and they’re way more successful than they were in the past.”
What Comes Next
So what comes next for deep learning? NVIDIA is certainly betting big on the technology driving huge demand for its products, including the new Tesla P100, a new chip it unveiled last week that features 15 billion transistors, or the DGX-1, which it bills as the world’s first deep learning supercomputer that’s able to churn through more than billion images per day.
Of particular interest is the intersection of deep learning and Internet of Things (IoT). The huge amounts of machine data emanating from sensors are perfect places for deep learning to go to work and boost the accuracy of predictions, which is critical for everything from self-driving cars to functional robots. But better and faster image recognition is just the start. We’re on the cusp of seeing new services that are powered by deep learning algorithms trained on mounds of big data that could change the way we live.
As of yet, we’ve yet to make full use of deep learning algorithms in the context of big data analytics, according to the writers of the aforementioned Journal of Big Data article “Deep learning applications and challenges in big data analytics.” “The ability of deep learning to extract high-level, complex abstractions and data representations from large volumes of data, especially unsupervised data, makes it attractive as a valuable tool for big data analytics,” they write.
At GTC, the Facebook AI Research team showed how programs backed by deep learning could extract important features from a corpus of landscape paintings by Romanticism-era artists and actually create a unique work that looks remarkably like an actual painting. Google is creating an email feature that will automatically reply to emails for you. And companies like Deep Genomics and Deep Instinct are using deep learning algorithms to accelerate the understanding of areas like genomics and cyber security, respectively.
“The beauty of deep learning is that it’s an approach that’s powerful enough to actually achieve super-human results without super humans to train them,” Huang said during his keynote. “It’s like Thor’s Hammer. It fell from the sky. It’s that big a deal.”
September 27, 2021
- TIBCO Delivers a Comprehensive, Connected Platform for the Adaptable Digital Business
- The World Economic Forum Welcomes Western Digital to Global Lighthouse Network
- LevaData Introduces New Suite of Supply Management Software
- KNIME Data Talks: Bringing Business and Data Science Together; Set for September 29
- BriefCam Introduces Video Analytics Enabled on Deep Learning Cameras from Axis Communications
September 24, 2021
- AWS Announces General Availability of Amazon QuickSight Q
- IDC’s 3rd Platform Industry Spending Guides Provide In-Depth Sub-Industry Forecasts for Technology Investments Across Nine Industries
- Scality Awarded US Patent for Hyperscale Data Protection
September 23, 2021
- AtScale Expands Semantic Layer Solution for Microsoft Excel
- CNCF End User Technology Radar Provides Insights into DevSecOps
- At Annual OCEANS 2021, Sofar Ocean Debuts First-of-Its-Kind Maritime Open Standard, Bristlemouth
- Elastic Announces the General Availability of Elastic App Search Web Crawler, New Features for Elastic Enterprise Search
- Securonix Achieves FedRAMP In-Process Authorization
- EDJX and Cubic Corporation Partner to Launch the Internet of Military Things Edge Platform
September 22, 2021
- GigaOm Names Moogsoft an Industry Leader in “Radar for AIOps Solutions” Report
- Clearsense Acquires Plug-and-Play AI Analytics Firm
- Purdue University Global Launches Master of Science in Data Analytics
- Dihuni OptiReady CognitX Deep Learning Servers and Workstations Powered by NVIDIA Ampere Architecture-based GPUs
- Scality Awarded New U.S. Patent for Breakthrough Technology in Hyper-Scale Data Protection
- MicroAI to Bring AI Training to Renesas MCUs
Most Read Features
- One on One with Google Cloud Product Director Irina Farooq
- Big Data File Formats Demystified
- What Is Data Science? A Turing Award Winner Shares His View
- Tabular Seeks to Remake Cloud Data Lakes in Iceberg’s Image
- What’s the Difference Between AI, ML, Deep Learning, and Active Learning?
- Who’s Winning In the $17B AIOps and Observability Market
- SambaNova Brings Custom Silicon To Bear on High-End AI Workloads
- How the Coronavirus Response Is Aided by Analytics
- In Search of the Modern Data Stack
- Rethinking Education in an AI-First World
- More Features…
Most Read News In Brief
- LinkedIn Open Sources Tech Behind 10,000-Node Hadoop Cluster
- Data and AI Salaries Continue Upward March, O’Reilly Says
- Data Prep Still Dominates Data Scientists’ Time, Survey Finds
- Gartner Shuffles the Technology Deck with Latest ‘Hype Cycle’ Report
- Who’s Winning in Open Source Data Tech
- Bigeye Observes $45 Million in Funding
- Why Is SAS Going Public?
- Hands-Off: Manual Data Integration Tasks Plummeting, Gartner Says
- Unstructured Data Growth Wearing Holes in IT Budgets
- Apollo CEO Bullish on GraphQL’s Potential in the Enterprise
- More News In Brief…
Most Read This Just In
- TIBCO NOW 2021 Showcases Limitless Power of Data
- Toloka Launches Data Research Grants, Announces First Eight Recipients
- Anaconda Announces Support for Pyston, Hiring Lead Developers Kevin Modzelewski and Marius Wachtler
- Kinetica Fuses Streaming and Contextual Analysis At Scale
- MariaDB Announces SIS Provider Campus Cloud Services Migration to MariaDB SkySQL
- Aporia Launches Self-Serve Machine Learning Platform Open to Public
- Snowflake Launches Financial Services Data Cloud
- Transaction Processing Performance Council (TPC) Launches an Artificial Intelligence Benchmark (TPCx-AI)
- DataRobot Launches “DataRobot AI Cloud” Platform
- OneTrust Enhances First-Party Data Solution to Strengthen Holistic Consent and Preference Management Platform
- More This Just In…
Sponsored Partner Content
October 5 - October 7
October 12 - October 14
October 19London United Kingdom
October 27 - October 28
November 29 - December 3
December 6 - December 10San Diego CA United States