Sears Rides Hadoop Up Retail Mountain
Falling behind Walmart and Target in retail store sales, Sears hopes to rebound by investing fairly heavily in Hadoop. Sears revenue had decreased from $50 billion in 2008 to $42 billion last year. However, smarter marketing as a result of being able to keep all their data and target customers individually has resulted in sizable growth over the last year, with sales over this year’s quarter ending on July 28 up 163% from the same quarter in 2011.
“With Hadoop we can keep everything, which is crucial because we don’t want to archive or delete meaningful data,” said Sears Chief Technology Officer Phil Shelley. Sears has seen their big data processing, especially with regard to evaluating marketing campaigns, quicken as a result of moving their data from Teradata and SAS onto Hadoop. According to Shelley, what took six weeks’ time now happens within a week on Hadoop. Their current 300-node cluster which contains 2PB allows the company to keep 100% of their data instead of a meager 10% according to Shelley.
Sears’s view and strategy regarding big data is an interesting one. Along with being Sears’s CTO and Executive VP, Shelley runs MetaScale, a Sears subsidiary whose goal it is to move into providing Hadoop services for other companies similar to Amazon and their Amazon Web Services.
It would seem that, in an effort to compete with Amazon, they would have a little catching up to do. On the other hand, Sears’s big data efforts currently surpass that of Walmart’s, who just started running ten Hadoop test nodes for experimental e-commerce analysis. Sears did that in 2010.
It is unfair to compare Sears, which makes it money historically from their physical stores, to online stores like Amazon. It may not even be fair to compare them to Target and Walmart, as Sears has more of an appliance focus while Target and Walmart are more general.
With that being said, Sears, and specifically MetaScale, wants to exist in the big data market. They have some interesting viewpoints regarding that. For example, Shelley sees little value in the modern era for ETL.
“ETL is an antiquated technique, and for large companies it’s inefficient and wasteful because you create multiple copies of data,” Shelley says. .“Everybody used ETL because they couldn’t put everything in one place, but that has changed with Hadoop, and now we copy data, as a matter of principle, only when we absolutely have to copy.”
Shelley’s principles are sound and may have led to the drastic reported reduction, $500,000, in their mainframe costs per year. Some, like Cloudera CEO Mike Olson, warn against the complete departure from ETL. But to Shelley the move is intuitive. “If in three years you come up with a new query or analysis, it doesn’t matter because there’s no schema,” Shelley says. “You just go get the raw data and transform it into any format you need.”
Related Articles
Cloudera CTO Reflects on Hadoop Underpinnings
Six Super-Scale Hadoop Deployments
May 10, 2024
May 9, 2024
- Lightning AI Launches Its Studio Platform in AWS Marketplace
- Precisely Continues to Expand Reach and Capabilities for Data Enrichment and Geo Addressing
- Databricks Enhances Enterprise AI with RAG Applications and Improved Model Serving
- StarTree Extends Real-Time Analytics Capabilities for Observability and Anomaly Detection Use Cases
- Snowflake Unveils the Future of AI and Data at 6th Annual Data Cloud Summit
May 8, 2024
- IBM and SAP Announce New Generative AI Capabilities to Transform Business Processes
- Alteryx: Businesses Embrace Generative AI, but General Public Remains Wary
- Atlan’s Control Plane to Power Data & AI Governance Attracts $105M
- Cisco Unlocks AI-Powered Intelligence for Self-Hosted Observability
- Boomi Accelerates Enterprise Data Connectivity with Advanced API Management Acquisitions
- Elastic Security Labs Releases Guidance to Avoid LLM Risks and Abuses
- Neo4j Empowers Syracuse University with $250K Grant to Tackle Misinformation in 2024 Elections
- Cloudera Unveils New Flow Management and Streaming Operators at Red Hat Summit
- DataForge Open-Sources First Functional Data Programming Framework
May 7, 2024
- Datadog Launches IT Event Management to Enhance AIOps Capabilities
- Panasas Rebrands as VDURA: A New Chapter in Software-Defined Storage for AI Applications
- Niobium Secures $5.5M in Venture Financing to Commercialize FHE Accelerator Chip
- Oracle Code Assist Can Help Developers Build Applications Faster with AI
Most Read Features
Sorry. No data so far.
Most Read News In Brief
Sorry. No data so far.
Most Read This Just In
Sorry. No data so far.
Sponsored Partner Content
-
Get your Data AI Ready – Celebrate One Year of Deep Dish Data Virtual Series!
-
Supercharge Your Data Lake with Spark 3.3
-
Learn How to Build a Custom Chatbot Using a RAG Workflow in Minutes [Hands-on Demo]
-
Overcome ETL Bottlenecks with Metadata-driven Integration for the AI Era [Free Guide]
-
Gartner® Hype Cycle™ for Analytics and Business Intelligence 2023
-
The Art of Mastering Data Quality for AI and Analytics
Sponsored Whitepapers
Contributors
Featured Events
-
AI & Big Data Expo North America 2024
June 5 - June 6Santa Clara CA United States -
CDAO Canada Public Sector 2024
June 18 - June 19 -
AI Hardware & Edge AI Summit Europe
June 18 - June 19London United Kingdom -
AI Hardware & Edge AI Summit 2024
September 10 - September 12San Jose CA United States -
CDAO Government 2024
September 18 - September 19Washington DC United States