Technologies » Storage

Features

Tape Gets Second Wind as Big Data Mounts

Sep 16, 2014 |

Think tape is dead in our big data world? Think again. This week, the National Center for Supercomputing Applications (NCSA) announced that it bought 20PB of tape capacity to expand the world’s biggest data archive. Meanwhile, the LTO Program has plotted out a roadmap for the next decade that will eventually see a single LTO tape cartridge storing 120 TB of data. While tape is looked down upon in our speed-obsessed culture, the old standby has proven its relevance in Read more…

What’s Driving the Explosion of Government Data?

Sep 11, 2014 |

Most people already know that the world produces massive amounts of data every day, and government organizations are a huge source of that data. For instance, it is estimated that next year the Department of Energy will create 300 terabytes of data per day from analyzing light sources, and by 2020, the department will create 15 petabytes of data per year analyzing high-energy physics. That’s a lot of data. But where does all this data come from? The amount of Read more…

How a Web Analytics Firm Turbo-Charged Its Hadoop ETL

Sep 10, 2014 |

The Web analytics firm comScore knows a thing or two about managing big data. With tens of billions of data points added to its 400-node Hadoop cluster every day, the company is no stranger to scalability challenges. But there’s one ETL optimization trick in particular that helped comScore save petabytes of disk and improve data processing times in the process. ComScore is one of the biggest providers of Web analytics used by publishers, advertising firms, and their clients. If you Read more…

Who IBM’s Server Group Turns To for Machine Data Analytics

Aug 28, 2014 |

IBM’s engineering prowess is second to none, and its Systems and Technology Group builds the computers that run the world’s biggest companies. But when IBM’s STG unit went looking for a way to predict failures by analyzing log data returned by its customers’ servers and storage arrays, it looked externally to a little-known machine data analytics startup from Santa Clara. Glassbeam got its start five years ago, before the Internet of Things (IOT) became the industry’s hottest buzzword and out-hyped Read more…

How to Move 80PB Without Downtime

Aug 25, 2014 |

When the online photo company Shutterfly decided to move its entire data center recently, the possibility of downtime was a big issue. After all, the company had 80 petabytes of customer data spread across tens of thousands of spinning disks, and those disks wouldn’t be spinning while being physically moved. Months later, after the last deliver made its way to Shutterfly‘s new data center, not one piece of data was lost or even temporarily unavailable from the company’s website. How Read more…

News In Brief

Data Tools Help Researchers Tackle Pediatric Cancer

Sep 10, 2014 |

A genomic data analysis platform being installed at the National Cancer Institute will deliver rapid analysis of billions of data points required when sequencing human DNA and other genomic data. Translational Genomics Research (TGen) said Sept. 10 it is providing the National Cancer Institute with high performance computing and bioinformatics support along with specialized tools designed to support pediatric cancer research programs. Those efforts include personalized medicine trial for pediatric cancer patients being conducted by the Neuroblastoma and Medulloblastoma Translational Read more…

MapR Reports Accelerated OpenTSDB Performance

Sep 9, 2014 |

Eyeing new Internet of Things (IoT) applications, MapR Technologies said its open-source distribution of Apache Hadoop “ingested” more than 100 million data points per second. The performance benchmark for the MapR distribution with its in-Hadoop NoSQL database, MapR-DB, was achieved using only four nodes of a ten-node cluster. By accelerating its OpenTSDB software by a factor of 1,000 on a small cluster, MapR claimed the performance clears the way for managing huge amounts of data along with IoT and other Read more…

Are Data Lakes All Wet?

Aug 4, 2014 |

Enterprise data management platforms known as “data lakes” are being promoted as, among other things, a potential solution to “information siloes” by combining different managed collections of data in an unmanaged data lake. The theory is that data consolidation will increase use and sharing of information while reducing storage and server costs. However, a new market study dismisses most of those claims as a “fallacy,” arguing instead that enterprises still require secure data repositories, in other words, data warehouses. At Read more…

Big Data Survey Finds Growing Need for CEP

Jul 30, 2014 |

A global survey of software developers working on big data and advanced analytics projects found that the large majority of respondents require real-time, complex event processing for their applications. The survey, released July 29 by Evans Data Corp. of Santa Cruz, Calif., found that 71 percent of respondents said they require advanced processing more than half the time. “Hadoop’s batch processing model has worked well for several years,” Evans Data CEO Janel Garvin noted in a statement releasing the survey. Read more…

UK Regulators Seek to Balance Big Data, Privacy

Jul 29, 2014 |

As concerns grow over how companies collect and use personal data, a UK regulator has issued guidelines for how companies leveraging big data must operate under Britain’s data protection law. The UK Information Commissioner’s Office recently released a report on July 28 that seeks a balance between operating within Britain’s Data Protection Act of 1998 without slowing data innovation. The report acknowledges that big data analytics frequently focus on areas other than the collection and use of personal data. It Read more…

This Just In

Tarmin Joins Object Storage Alliance as a Founding Member

Sep 18, 2014 |

BOSTON, Mass., Sept. 18 — Tarmin Inc., a global pioneer of Data Defined Storage solutions, today announced that is has joined the Object Storage Alliance as a founding member. The Object Storage Alliance is a non-profit storage industry organization dedicated to promoting advanced technologies that allow organizations to drive informed decisions, glean insights and maximize the business value of information. Eighty percent of enterprise data is unstructured, and in 2013 Gartner predicted that enterprise data growth would exceed 800 percent by 2018. With Read more…

Seagate and Baidu Sign Strategic Cooperation Agreement

Sep 17, 2014 |

BEIJING, China, Sept. 17 — Seagate Technology, a worldwide leader in storage solutions, today announced it has signed a strategic cooperation agreement with Baidu, China’s largest web services firm, in the areas of low-cost online storage, archiving and Big Data analysis. Under the agreement, Baidu will give priority to Seagate products when considering components for all Baidu servers and storage facilities. In return Seagate will align with Baidu as a Strategic Customer and give priority to Baidu when providing enterprise storage Read more…

A3CUBE Introduces Fortissimo Foundation

Sep 16, 2014 |

SAN JOSE, Calif., Sept. 16 — A3CUBE Inc., a developer of revolutionary communication technology that extends the capability of standard PCIe interconnect to produce industry-leading networked computing and storage solutions, today introduced its Fortissimo Foundation building block software solution. Coupled with A3CUBE’s previously announced RONNIE Express platform, Fortissimo Foundation provides dramatic improvements in application performance and datacenter efficiency, particularly in unstructured, Big Data environments. Fortissimo Foundation works in concert with A3CUBE’s RONNIEE Express data path accelerator and RONNIEE NICs to Read more…

Research and Markets Adds Future of Big Data Analytics Report to Offering

Sep 16, 2014 |

Sept. 16 — Research and Markets has announced the addition of the “The Future of Big Data Analytics – Global Market and Technologies Forecast – 2015-2020″ report to their offering. Big data is data that exceeds the processing capacity of conventional database systems. The data is too big, moves too fast, or exceeds the limits of your database architectures. To gain value from this data, you must choose an alternate way to process it. Many argue that Big Data is all talk and no Read more…

MapR Announces Performance Results Using OpenTSDB on Distribution

Sep 9, 2014 |

SEATTLE, Wash., Sept. 9 — MapR Technologies, Inc., provider of the top-ranked distribution for Apache Hadoop, today announced at the Tableau Conference, breakthrough performance results achieved using standard open source software, OpenTSDB, running on the MapR Distribution. Using only four-nodes of a 10-node cluster, the MapR Distribution with its in-Hadoop NoSQL database, MapR-DB, ingested over 100 million data points per second. By accelerating OpenTSDB performance by 1,000 times on such a small cluster, MapR opens the doors to cost-effectively manage massive volumes Read more…