Making It Interactive


Synopsis: Now is the time to finalize the interactive part of the Social Network Benchmark (SNB). The benchmark must be both credible in a real social network setting and pose new challenges. There are many hard queries but not enough representation for what online systems in fact do. So, the workload mix must strike a balance between the practice and presenting new challenges.

It is about to be showtime for LDBC. The initial installment of the …

SNB Data Generator - Getting Started


In previous posts (this and this) we briefly introduced the design goals and philosophy behind DATAGEN, the data generator used in LDBC-SNB. In this post, I will explain how to use DATAGEN to generate the necessary datatsets to run LDBC-SNB. Of course, as DATAGEN is continuously under development, the instructions given in this tutorial might change in the future.

Getting and Configuring Hadoop

DATAGEN runs on top of hadoop 1.2.1 to be scale. …

The Day of Graph Analytics


Note: consider this post as a continuation of the
Making it interactive” post by Orri Erling.

I have now completed
the Virtuoso TPC-H work,
including scale out. Optimization possibilities extend to infinity but
the present level is good enough. TPC-H is the
classic of all analytics benchmarks and is difficult enough, I have
extensive commentary on this on my blog (In Hoc Signo Vinces series),
including experimental results. This …

During the past six months we (the OWLIM Team at Ontotext) have
integrated the LDBC Semantic Publishing Benchmark (LDBC-SPB) as a part of our development and
release process.

First thing we’ve started using the LDBC-SPB for is to monitor the
performance of our RDF Store when a new release is about to come out.

Initially we’ve decided to fix some of the benchmark parameters :

  • the dataset size - 50 million triples (LDBC-SPB50)
  • benchmark warmup and …