posted by user: rabl || 2514 views || tracked by 7 users: [display] 2013 : Fourth Workshop on Big Data Benchmarking


When Oct 9, 2013 - Oct 10, 2013
Where Santa Clara, California
Submission Deadline Sep 6, 2013
Notification Due Sep 20, 2013
Final Version Due Oct 18, 2013
Categories    big data   benchmarking   bigdata top100   bigbench

Call For Papers

[Apologies for cross-posting.]
***** 4th Workshop on Big Data Benchmarking *****

*** October 09-10, 2013 ***
*** Brocade Executive Briefing Center, San Jose, CA, USA ***

Workshop Description
The objective of this workshop is to explore issues in developing
industry-standard benchmarks for providing objective measures of
the effectiveness of hardware and software systems for big data
applications. While micro-benchmarks and functional benchmarks,
such as Terasort, play an important role, the workshop is especially
interested in application-level, end-to-end benchmarks. A successful
benchmark would be simple to implement and execute; cost effective,
so that the benefits of executing the benchmark justify its expense;
timely, with benchmark versions keeping pace with rapid changes in
the marketplace; and verifiable so that results of the benchmark can
be validated via independent means. Previous, independent workshops
on this topic have led to the development of two benchmark proposals:
one based on a Deep Analytics Pipeline for event processing and a
second, called BigBench, based on extending the TPC-DS benchmark with
semi-structured and unstructured data and new queries targeted at
those data.

Related research topics explore a broad range of characteristics that
define big data and big data applications, including:

O DATA FEATURES: New feature sets of data including, high-dimensional
data, sparse data, event-based data, and enormous data sizes.
O SYSTEN CHARACTERISTICS: System-level issues including, large-scale
and evolving system configurations, shifting loads, and
heterogeneous technologies for big data and cloud platforms.
O IMPLIMENTATION OPTIONS: Different implementation options such as
SQL, NoSQL, Hadoop software ecosystem, and different
implementations of HDFS.
O WORKLOAD: Representative big data business problems and
corresponding benchmark implementations. Specifying benchmark
applications that represent different modalities of big data,
including graphs, streams, scientific data, and document
O HARDWARE OPTIONS: Evaluating new options in hardware including
different types of HDD, SSD, and main memory, and large-memory
systems, and new platform options that include dedicated commodity
clusters and cloud platforms.
O DATA GENERATION: Models and procedures for generating large-scale
synthetic data with requisite properties.
O BENCHMARK EXECUTIONS RULES, e.g. data scale factors, benchmark
versioning to account for rapidly evolving workloads and system
configurations, benchmark metrics.
O METRICS FOR EFFICIENCY: Measuring the efficiency of the solution,
e.g., based on costs of acquisition, ownership, energy and/or other
O EVALUATION FRAMEWORKS: Tool chains, suites and frameworks for
evaluating big data systems.
O EARLY IMPLEMENTATIONS of the Deep Analytics Pipeline or BigBench,
or describing lessons learned in benchmarking big data applications
are solicited. Discussions of enhancements to these benchmarks are
also encouraged, for example, including more data genres (e.g.,
graphs) in the workload; considering a range of machine learning
and other algorithms, etc. Papers proposing other benchmarking
alternatives will also be considered.

Paper Submission, Review, and Publication
Papers should be formatted using the Springer LNCS Proceedings format.
Selected revised papers will be published in Lecture Notes in Computer
Science by Springer.

Full papers: max. 12 pages
Short papers: max. 6 pages

Please use the following submission system to submit your paper:

Submission Dates:
O September 6, 2013:
Due date for full workshop papers submission
O September 20, 2013:
Notification of paper acceptance to authors
O October 09-10, 2013:
O October 18, 2013:
Camera-ready of accepted papers

For questions please contact Chaitan Baru (baru[at]sdsc[dot]edu).

Previous instances of the workshops:
O May 2012 in San Jose, CA, USA (
O December 2012 in Pune, India (
O July 2013 in Xi'an, China (

Workshop Chair(s):
O Chaitan Baru (San Diego Supercomputer Center, UC San Diego)
O Milind Bhandarkar (Pivotal)
O Tilmann Rabl (Middleware Services Research Group, U of Toronto)

Publicity Chair(s):
O Florian Stegmaier (University of Passau, Germany)

Programm Chair(s):
O Dhruba Borthakur (Facebook)
O Yanpei Chen (Cloudera)
O John Galloway
O Ahmad Ghazal (Oracle)
O Boris Glavic (IIT Chicago)
O Bhasker Gowda (Intel)
O Eyal Gutkind (Mellanox)
O Songlin Hu (Chinese Academy of Science)
O Jian Li (IBM)
O DK Panda (Ohio State University)
O Scott Pearson (Brocade)
O Meikel Poess (Oracle)
O Francois Raab (InfoSizing)
O Kai Sachs (SAP)
O Jerry Zhao (Google)
O Jianfeng Zhan (Chinese Academy of Sciences)

Related Resources

ICBDB 2022   2022 4th International Conference on Big Data and Blockchain(ICBDB 2022)
SoCAV 2023   2023 International Symposium on Connected and Autonomous Vehicles (SoCAV 2023)
IEEE BigData 2022   2022 IEEE International Conference on Big Data
IEEE SSCI 2023   2023 IEEE Symposium Series on Computational Intelligence
MLDM 2023   18th International Conference on Machine Learning and Data Mining
AIAA 2022   12th International Conference on Artificial Intelligence, Soft Computing and Applications
AI-DH 2022   MDPI Big Data and Cognitive Computing - Special Issue on Artificial Intelligence in Digital Humanities
IOP, EI, Scopus-EMECS 2022   2022 International Conference on Electronics, Mechanical Engineering and Computer Science (EMECS 2022)-EI Compendex
SI PMABD 2023   Special Issue on Programming Models and Algorithms for Big Data
ICDM 2023   23th Industrial Conference on Data Mining