Dremel Builder Gets $7M for SQL-Based Supertool
Big data startup Metanautix emerged from stealth mode today by announcing a $7-million round of venture funding to further development of a SQL-based power tool. Led by the former Google engineer who headed the development of Dremel, the company aims to dissolve product and technology barriers by “re-imagining” SQL at the heart of an emerging big data supply chain.
SQL is enjoying a renaissance as the big data boom continues to reverberate throughout the IT and business sectors. While emerging big data platforms like Hadoop and NoSQL database have introduced powerful new ways to store and access data, the resurgent momentum behind SQL is in many cases forcing the developers of these cutting-edge systems to rethink their omission of SQL–and in many cases, patch the systems with new SQL access points.
This SQL resurrection is not only leading Hadoop distributors like Cloudera, Hortonworks, and Pivotal to fund big SQL access programs like Impala, Stinger, and HAWQ, but it’s leading established business intelligence and data warehouse vendors like Oracle, Teradata, Microsoft, and IBM to put SQL front-and-center of their logical data warehouse strategies. Write one SQL query, the thinking goes, and let the system federate the query out to where the data resides.
Metanautix is taking a similar strategy, just without the legacy baggage of needing to support existing data warehouse customers. It stays as far away from the storage layer as it can, preferring instead to provide a thin, neutral layer of SQL oil that will lubricate the gears of the emerging “data supply chain.”
“What we’re trying to do is make it possible for analysts to reach into whatever data they want to reach into and not really have to care about all the details,” says Metanautix co-founder and CEO Theo Vassilakis, who led the a staff of 75 for the development of Dremel, the distributed query engine that powers Google’s BigQuery.
“We call it re-imagining SQL because is SQL is very old and standard,” Vassilakis continues. “It’s baked into many corporate applications. We think that one of the impediments to getting at all the data that’s out there is that analysts have to keep switching tools and using a lot of different systems. We wanted to make it so that you could do all the stages of analysis–a little ETL, a little ad hoc analysis, also some serving–and do it with one system with a standard language.”
The company’s eponymous product will use standard SQL to access any type of data wherever it resides, including relational databases like MySQL, NoSQL databases like Mongo, HDFS, object-data stores, Amazon S3, CIFS, NFS, and others. When it becomes generally available later this year (it’s currently in limited release), the software will perform a variety of functions–including ETL, ad-hoc queries, serving analytic dashboards, and even running machine learning algorithms–without requiring users to move among different tools.
It’s all about enabling the “data supply chain” to function as effortlessly as possible. “We’re using that phrase to refer to when one person feeds data into another person who feeds data into another person, and how do you aggregate that and provide a good sense of [what’s going on],” Vassilakis tells Datanami. “We want to fit into their existing environments as opposed to forcing them into a certain way.”
The product was inspired in part by Dremel, the insanely big and fast query tool that Google built to analyze petabytes of data and run in a distributed manner across thousands of nodes. Dremel has inspired other software developers, including those behind the Apache Drill project, which is developing an interactive query technology based on Dremel for the Hadoop environment.
Metanautix was also inspired in part by the work that Metanautix CTO and co-founder Apostolos Lerios did at Facebook. Lerios led the development of the photo-upload portion of the social media giant’s website. With more than 300 billion images, it’s currently the largest photo repository in the world. The combination of the two co-founder’s background led to Metanautix.
Metanautix supports all of the analytical aspects of the ANSI SQL standard; it doesn’t bother with the transaction oriented data types because that’s not the company’s focus. The software treats any data as a standard table, including structured data and unstructured data. For unstructured data types, such as JSON data and images, the company built its own functions.
“Our goal is basically, wherever you have your data, we want to be able to go read it,” Vassilakis tells says. “If it’s NoSQL, we’ll go read it from NoSQL. If it’s a server or if the Web or HDFS, we’ll go get it. Whatever we don’t support, we’ll use our extensibility mechanism….so people can plug in their legacy logic, but still make it visible to the user as plain SQL.”
“Metanautix is a distributed system. You can run one node or you can run 1,000-server configurations behind your Qlik product if that’s what you need,” Vassilakis says. “Part of the beauty of SQL is you don’t need to know the difference. We can say, hey is there big data? Just point your QlikTech instance to us, treat us as SQL, but we can go out and read a lot more data than whatever data you’ve allocated to Qlik and still have it be performant and integrated.”
In distributed and Hadoop environments, Metanautix software can replace the need for MapReduce coding. “Our view is Hadoop is great. You get lots of high-throughput storage and data formats and all of that. But it’s also challenging because for a lot of things, you have to write custom code or connectors,” Vassilakis says. “One of the areas customer are interested in is joining Hadoop data with non-Hadoop data. They say, Can I join Oracle and MySQL with Hadoop? Can I get next-gen data, such as GPS data or Internet of Things traces, into Hadoop, but then join it with more traditional accounts information database stuff, on the fly without writing new code or MapReduce?”
Metanautix will also run machine learning algorithms. The company has demonstrated how to implement the K-Means clustering algorithm in SQL. “People don’t ‘think of SQL that way,” Vassilakis says. “It turns out it’s not that hard. Eight queries we did it in. A lot of things that people felt are in the purview of some super specialized system that just does machine learning or clustering specifically is now at their fingertips. We’re basically trying to make it so just SQL and more queries is the answer. We treat all data like table and all questions as SQL and try to bring those two things together.”
The company wrote its software using a combination of C++ for performance and Java for extensibility. The software runs in virtualized Linux and Windows environments, and is designed to scale.
Today the company announced that it closed a $7 million Series A round financing, which was led by Sequoia Capital and includes investments from the Stanford University endowment fund and from Shiva Shivakumar, former vice president of engineering and distinguished entrepreneur at Google.
September 25, 2020
- PostgreSQL 13 Released: Performance Gains, Space Savings, Enhanced Security, Developer Experience
- WANdisco Announces Global Agreement with Infosys to De-Risk and Accelerate Data Lake Migration to the Cloud
- Matillion Partner Ecosystem Identifies Trends Driving Data Transformation Market
- TIBCO Simplifies Data Unification With TIBCO Any Data Hub
- Trifacta Named Leader in G2’s Fall Grid Report for Data Preparation
- Seagate’s New Solutions Equip Enterprises for the New Data Economy
September 24, 2020
- Spectra Logic Announces Industry’s First Tape Library to Store One Exabyte of Uncompressed Data Leveraging LTO-9 Technology
- QDA Miner 6 Powers Businesses with New Qualitative Analysis Capabilities
- Cambridge Semantics Appoints Brian D. Owen as Chief Executive Officer
- Exasol Dominates Its Peer Groups in BARC Data Management Survey 2020
- The Apache Software Foundation Announces Apache IoTDB as a Top-Level Project
- Sneak Peek of Breakout Sessions Announced for the In-Memory Computing Summit 2020 Virtual Worldwide Conference
September 23, 2020
- Elastic Announces ElasticON Global, Free Virtual User Conference to Take Place From October 13-15
- KIOXIA Bolsters NVMe-oF Ecosystem with Ethernet SSD Storage; Collaborates with Marvell, Foxconn-Ingrasys and Accton
- TIBCO Hyperconverged Analytics Dramatically Simplifies Analytics Experience
- NASA, ICIJ, ATPCO, Lyft and More Choose Neo4j for their Knowledge Graphs
September 22, 2020
- Qlik Expands Strategic Partnership With Google Cloud With Integrated Solution for SAP Data Analytics
- U.S. Food and Drug Administration Selects Cambridge Semantics for Data and Analytics Platform
- SAS accelerates development of analytics and data science talent with new academic program
- ThoughtSpot Launches SaaS Offering to Unlock the Value of Cloud Data Warehouses with Search & AI-Driven Analytics
Most Read Features
- How Facebook Accelerates SQL at Extreme Scale
- Microsoft Now Developing Its Own Hadoop
- Big Data File Formats Demystified
- 10 Big Data Statistics That Will Blow Your Mind
- VC Ben Horowitz Dishes on Hadoop, AI, and Data Culture
- How to Build a Better Machine Learning Pipeline
- R and Python: The Data Science Dynamic Duo
- How the Coronavirus Response Is Aided by Analytics
- Is Python Strangling R to Death?
- The Future of Labor in an AI World
- More Features…
Most Read News In Brief
- Snowflake to Make it SNOW on NYSE
- Aerospike Gives Legacy Infrastructure a Real-Time Boost
- Speech Recognition Gets an AutoML Training Tool
- A ‘Breakout Year’ for ModelOps, Forrester Says
- Google Joins the MLOps Crusade
- Snowflake Pops in ‘Largest Ever’ Software IPO
- Air Force Expands Predictive Maintenance
- Fivetran Launches Pay-As-You-Go Option for ETL
- New AI Tool Maps the Families of the Bible, A Song of Ice and Fire
- Cassandra Gets an Indexing Upgrade
- More News In Brief…
Most Read This Just In
- Monte Carlo Raises $16M to Build the World’s First Data Reliability Platform
- Talend Introduces Industry-First Measure of Data Health to Bring Clarity and Confidence to Every Business Decision
- Tabor Communications, Inc. Announces Expansion of the Editorial Team
- IBM Cognos Analytics-Based Business Transformation Going Strong
- Tamr Data Mastering Platform Now Available on Microsoft Azure
- Scality RING8 on All-Flash Delivers File and Object Storage Performance 10x Faster Than Competitive Solutions
- ScyllaDB Unveils One-Step Migration from Amazon DynamoDB to Scylla NoSQL Database
- Neo4j Delivers Integrated Graph Database Service on Google Cloud Platform
- Yugabyte Announces Speaker Lineup for Distributed SQL Summit 2020
- BigID Introduces Hyperscan for Speeding Unstructured File Scans at Scale
- More This Just In…