Here is the latest from Julia Computing

JuliaDB in Julia 1.0

27 Feb 2019 | Josh Day

In many data science applications, it is easy to run out of memory when working with data. An analyst working with big data has a few options available to them:

  1. Buy more RAM.
  2. Rent RAM with a cloud-based service.
  3. Use a sample of the dataset.
  4. Buy a SAS license.

None of these options are particularly good solutions.

Introducing JuliaDB

With JuliaDB, one can easily read big data, save it in an efficient binary format, and even run operations out-of-core. Analytics are available via OnlineStats integration, making statistical calculations on big data a breeze.

OnlineStats implements on-line (single-pass) algorithms for statistics and models, meaning you can run analyses like linear regression on data that is too big to fit in memory. Every statistic/model in OnlineStats also supports merging, enabling parallel processing. The combination of on-line updating/merging eliminates the need for the entire dataset to be loaded into RAM simultaneously, allowing analyses that would not be possible with traditional methods. Below is a visualization of how JuliaDB integrates with OnlineStats by scheduling the updating and merging operations:

fitting and merging


From Kaggle’s Huge Stock Market Dataset, there are over 7000 CSVs with historical price data (each stock’s history in a different file). JuliaDB can quickly load them into a distributed dataset and perform group-by operations:

using Distributed

@everywhere using JuliaDB, OnlineStats

# 7195 CSVs with 14,887,665 rows
files = glob("*.txt", "Stocks")

t = loadtable(files, filenamecol=:Stock)

groupreduce(Mean(), t, :Stock; select=:Volume)
Distributed Table with 7163 rows in 4 chunks:
Stock          Mean
""     Mean: n=4521 | value=3.9935e6
""    Mean: n=12074 | value=3.33776e6
""  Mean: n=505 | value=1.49989e5
""  Mean: n=5434 | value=2.3218e7
""   Mean: n=785 | value=2.20441e5
""   Mean: n=989 | value=1.00454e7
""  Mean: n=1211 | value=17644.4
""  Mean: n=2926 | value=6715.76
""   Mean: n=3201 | value=7.13972e5
""  Mean: n=1041 | value=7.94087e5
""  Mean: n=3201 | value=2.04258e5
""   Mean: n=3201 | value=1.22519e6
""  Mean: n=8364 | value=1.06642e8
""   Mean: n=1717 | value=2.15145e5
""   Mean: n=3177 | value=1.86901e5
""   Mean: n=3199 | value=4.28732e5
""  Mean: n=3162 | value=2.64112e5
""  Mean: n=3201 | value=1.40883e6
""    Mean: n=3201 | value=5.60349e5

Main Features

Just-in-Time Compiled

JuliaDB leverages Julia’s just-in-time compiler (JIT) so that table operations – even custom ones – are fast.

Compute in Parallel

Process data in parallel or even calculate statistical models out-of-core through integration with OnlineStats.jl.

Store Any Data Type

JuliaDB supports Strings, Dates, Float64… and any other Julia data type, whether built-in or defined by you.

Fast User-Defined Functions

JuliaDB is written 100% in Julia. That means user-defined functions are JIT compiled.

Fast CSV Parser

CSVs are loaded extremely fast! Many files can be read at the same time to create a single table.

Open Source

JuliaDB is released under the MIT License.

JuliaDB for Time Series

The ability to index (sort) on any number of columns and store any data type makes JuliaDB ideal for time series analysis. For a big data time series example, see the demo here.

Feature JuliaDB Pandas xts (R) TimeArrays
Distributed Computing
Data larger than memory
Multiple Indexes
Index Type(s) Any Built-ins Time Time
Value Type(s) Any Built-ins Built-ins Any
Compiled UDFs


Recent posts

Grokking Deep Learning with Julia
29 Jul 2020 | Deepak Suresh
Newsletter July 2020
17 Jul 2020 | Julia Computing
CSV Reader Benchmarks: Julia Reads CSVs 10-20x Faster than Python and R
22 Jun 2020 | Deepak Suresh
Get the latest news about Julia delivered to your inbox.