Dremio Jekyll

Aug 9, 2017

Looking Back At How We Exited Dremio From Stealth

Dremio recently made the leap out of stealth mode, and as we embark on the next phase of the company, I want to share some of the insights that we’ve picked up along the way. While in stealth, we built a customer-ready product, launched a beta, and learned a lot about what it means to reimagine data analytics in the modern world.

Read More

Jul 30, 2017

Summary of Dremio Launch Coverage

After two years of stealth, we launched Dremio on July 19, 2017. It was a long day that started at 5AM in Mountain View when the new site went live and Dremio first became available for download. A few hours later we had users on 6 continents and over two dozen countries. We celebrated with a party at the RedPoint offices in San Francisco:

Read More

Jul 19, 2017

Recognizing A New Tier

After nearly two years of R&D, I’m excited to announce the launch of Dremio today. Being able to share what we’ve been working gives me a great sense of pride for what this team has been able to build.

Read More

May 17, 2017

New Big Data Debt Calculator Helps Enterprise Organizations Measure Emerging Data Challenges

MOUNTAIN VIEW, CA–(Marketwired - Mar 8, 2017) - MOUNTAIN VIEW, Calif.–(BUSINESS WIRE)–Dremio, a stealth data analytics company, announced today a new way for enterprise organizations to quickly identify the amount of data debt that has been created by applications built to address strategic business initiatives. Dremio’s Big Data Debt Calculator is a free and simple way to estimate unplanned costs that arise from the use of non-relational data management technologies such as Hadoop, MongoDB, Elasticsearch, and Amazon S3, as well as data that is spread across many sources. The calculator also gives recommendations for minimizing debt, strategies for paying it down and ensuring it remains within acceptable bounds.

Read More

Apr 14, 2017

What Are Data Pipelines?

Data moves between systems. Data engineers create data pipelines to orchestrate the movement, transformation, validation, and loading of data, from source to final destination.

Read More

Mar 20, 2017

What is a Data Warehouse?

For over 30 years, we’ve approached data analytics the same way - copy the data, move it to a different system, spend a fortune, wait an eternity. This is building the data warehouse. We’re hard at work on a new alternative.

Read More

Mar 15, 2017

ETL Tools Explained

The ETL model has been in use for over 30 years - read data from different sources, apply transformations, then save the results in a different system for analytics. Modern hardware and distributed processing create new models for accessing data for analytics.

Read More

Mar 8, 2017

Open Source and Big Data Analytics Experts to Speak on Data Processing with Arrow and Parquet and Security in Hadoop at Strata+Hadoop World 2017

MOUNTAIN VIEW, CA–(Marketwired - Mar 8, 2017) - Dremio, a stealth data analytics company, announced today that its CTO and Co-Founder, Jacques Nadeau; Principal Architect, Julien Le Dem and Principal Software Engineer, Yuliya Feldman will be speakers at Strata+Hadoop World 2017 held at San Jose Convention Center, March 13-17. Strata + Hadoop World is a 4-day immersion in the most challenging problems, intriguing use cases, and enticing opportunities in data today. Leading experts in technology and business will take the stage to share their expertise and ideas, covering everything essential in data today.

Read More

Mar 8, 2017

What is Data Engineering?

There’s a relatively new role in many companies called Data Engineering. This team is responsible for making it easier for analysts, data scientists, and systems to access and analyze data. Because the role is new, there are lots of questions about what this team does, what tools they use, and how they work with data.

Read More

Jun 8, 2016

BI on Big Data: What are your options?

Deciding what combination of technologies will yield the best ‘BI on Big Data’ experience can be a major challenge for data professionals. This presentation, given by Dremio CEO Tomer Shiran at Strata + Hadoop World London, aims to shed some light on some of the solutions that are available in the space.

Read More

Feb 17, 2016

Introducing Apache Arrow: Columnar In-Memory Analytics

Apache Arrow establishes a de-facto standard for columnar in-memory analytics which will redefine the performance and interoperability of most Big Data technologies. The lead developers of 13 major open source Big Data projects have joined forces to create Arrow, and additional companies and projects are expected to adopt and leverage the technology in the coming months. Within the next few years, I expect the vast majority of all new data in the world to move through Arrow’s columnar in-memory layer.

Read More

Dec 13, 2015

Tuning Parquet file performance

Today I’d like to pursue a brief discussion about how changing the size of a Parquet file’s ‘row group’ to match a file system’s block size can effect the efficiency of read and write performance. This tweak can be especially important on HDFS environments in which I/O is intrinsically tied to network operations.

Read More