Ippokratis Pandis

Ippokratis Pandis

San Francisco Bay Area
8K followers 500 connections

Activity

Join now to see all activity

Experience

Publications

  • Memory-Efficient Hash Joins

    Proceedings of the VLDB Endowment (PVLDB), 2015.

    Other authors
  • Joins on Encoded and Partitioned Data

    VLDB 2014

    Compression has historically been used to reduce the cost of stor- age, I/Os from that storage, and buffer pool utilization, at the ex- pense of the CPU required to decompress data every time it is queried. However, significant additional CPU efficiencies can be achieved by deferring decompression as late in query processing as possible and performing query processing operations directly on the still-compressed data. In this paper, we investigate the benefits and challenges of performing joins…

    Compression has historically been used to reduce the cost of stor- age, I/Os from that storage, and buffer pool utilization, at the ex- pense of the CPU required to decompress data every time it is queried. However, significant additional CPU efficiencies can be achieved by deferring decompression as late in query processing as possible and performing query processing operations directly on the still-compressed data. In this paper, we investigate the benefits and challenges of performing joins on compressed (or encoded) data. We demonstrate the benefit of independently optimizing the compression scheme of each join column, even though join predi- cates relating values from multiple columns may require translation of the encoding of one join column into the encoding of the other. We also show the benefit of compressing “payload” data other than the join columns “on the fly,” to minimize the size of hash tables used in the join. By partitioning the domain of each column and defining separate dictionaries for each partition, we can achieve even better overall compression as well as increased flexibility in dealing with new values introduced by updates. Instead of decom- pressing both join columns participating in a join to resolve their different compression schemes, our system performs a light-weight mapping of only qualifying rows from one of the join columns to the encoding space of the other at run time. Consequently, join predicates can be applied directly on the compressed data. We call this procedure encoding translation. Two alternatives of encoding translation are developed and compared in the paper. We provide a comprehensive evaluation of these alternatives using product im- plementations of each on the TPC-H data set, and demonstrate that performing joins on encoded and partitioned data achieves both su- perior performance and excellent compression.

    Other authors
    See publication
  • DB2 with BLU Acceleration: So Much More than Just a Column Store

    39th International Conference on Very Large Data Bases (VLDB 2013)

    Want to know how BLU Acceleration REALLY works? Written by the researchers and developers that invented it, this highly technical overview is not really a white paper, but in fact a contribution to academic literature that was refereed by world experts and presented at the 39th International Conference on Very Large Data Bases (VLDB 2013) in Riva del Garda, Trento, Italy on 27 August 2013.

    Other authors
    See publication
  • Go Server Go: Parallel Computing with Moving Servers

    Proceedings of the ACM Symposium on Cloud Computing (SoCC), 2013

    Other authors
  • Business Analytics in (a) Blink

    IEEE Data Engineering Bulletin

    The Blink project’s ambitious goal is to answer all Business Intelligence (BI) queries in mere seconds,
    regardless of the database size, with an extremely low total cost of ownership...

    Other authors
    See publication
  • Scalability of write-ahead logging on multicore and multisocket hardware

    VLDB Journal

    Other authors
  • Blink: Not Your Father's Database!

    Proceedings of the 5th International Workshop on Business Intelligence for the Real-Time Enterprises (BIRTE)

    The Blink project’s ambitious goals are to answer all Business Intelligence (BI) queries in mere seconds, regardless of the database size, with an extremely low total cost of ownership. It takes a very innovative and counter-intuitive approach to processing BI queries, one that exploits several disruptive hardware and software technology trends. Specifically, it is a new, workload-optimized DBMS aimed primarily at BI query processing, and exploits scale-out of commodity multi-core processors…

    The Blink project’s ambitious goals are to answer all Business Intelligence (BI) queries in mere seconds, regardless of the database size, with an extremely low total cost of ownership. It takes a very innovative and counter-intuitive approach to processing BI queries, one that exploits several disruptive hardware and software technology trends. Specifically, it is a new, workload-optimized DBMS aimed primarily at BI query processing, and exploits scale-out of commodity multi-core processors and cheap DRAM to retain a (copy of a) data mart completely in main memory. Additionally, it exploits proprietary compression technology and cache-conscious algorithms that reduce memory bandwidth consumption and allow most SQL query processing to be performed on the compressed data. Ignoring the general wisdom of the last three decades that the only way to scalably search large databases is with indexes, Blink always performs simple, “brute force” scans of the entire data mart in parallel on all nodes, without using any indexes or materialized views, and without any query optimizer to choose among them. The Blink technology has thus far been incorporated into two products: (1) an accelerator appliance product for DB2 for z/OS (on the “mainframe”), called the IBM Smart Analytics Optimizer for DB2 for z/OS, V1.1, which was generally available in November 2010; and (2) the Informix Warehouse Accelerator (IWA), a software-only version that was generally available in March 2011. We are now working on the next generation of Blink, called BLink Ultra, or BLU, which will significantly expand the “sweet spot” of Blink technology to much larger, disk-based warehouses and allow BLU to “own” the data, rather than copies of it.

    Other authors
  • Aether: A Scalable Approach to Logging

    VLDB

    Other authors
  • Agent based middleware infrastructure for autonomous context-aware ubiquitous computing services

    Elsevier

    Middleware for ubiquitous and context-aware computing entails several challenges, including the need to balance between transparency and context-awareness and the requirement for a certain degree of autonomy. In this paper we outline most of these challenges, and highlight techniques for successfully confronting them. Accordingly, we present the design and implementation of a middleware infrastructure for ubiquitous computing services, which facilitates development of ubiquitous services…

    Middleware for ubiquitous and context-aware computing entails several challenges, including the need to balance between transparency and context-awareness and the requirement for a certain degree of autonomy. In this paper we outline most of these challenges, and highlight techniques for successfully confronting them. Accordingly, we present the design and implementation of a middleware infrastructure for ubiquitous computing services, which facilitates development of ubiquitous services, allowing the service developer to focus on the service logic rather than the middleware implementation. In particular, this infrastructure provides mechanisms for controlling
    sensors and actuators, dynamically registering and invoking resources and infrastructure elements, as well as modeling of composite contextual information. A core characteristic of this infrastructure is that it can exploit numerous perceptual components for context acquisition. The introduced middleware architecture has been implemented as a distributed multi-agent system. The various agents have been augmented with fault tolerance capabilities. This middleware infrastructure has been exploited in implementing a non-obtrusive ubiquitous computing service. The latter service resembles an intelligent non-intrusive human assistant for conferences, meetings and presentations and is illustrated as a manifestation of the benefits of the introduced infrastructure.

    Other authors
    See publication
  • Agent based middleware infrastructure for autonomous context-aware ubiquitous computing services

    Elsevier

    Middleware for ubiquitous and context-aware computing entails several challenges, including the need to balance between transparency and context-awareness and the requirement for a certain degree of autonomy. In this paper we outline most of these challenges, and highlight techniques for successfully confronting them. Accordingly, we present the design and implementation of a middleware infrastructure for ubiquitous computing services, which facilitates development of ubiquitous services…

    Middleware for ubiquitous and context-aware computing entails several challenges, including the need to balance between transparency and context-awareness and the requirement for a certain degree of autonomy. In this paper we outline most of these challenges, and highlight techniques for successfully confronting them. Accordingly, we present the design and implementation of a middleware infrastructure for ubiquitous computing services, which facilitates development of ubiquitous services, allowing the service developer to focus on the service logic rather than the middleware implementation. In particular, this infrastructure provides mechanisms for controlling
    sensors and actuators, dynamically registering and invoking resources and infrastructure elements, as well as modeling of composite contextual information. A core characteristic of this infrastructure is that it can exploit numerous perceptual components for context acquisition. The introduced middleware architecture has been implemented as a distributed multi-agent system. The various agents have been augmented with fault tolerance capabilities. This middleware infrastructure has been exploited in implementing a non-obtrusive ubiquitous computing service. The latter service resembles an intelligent non-intrusive human assistant for conferences, meetings and presentations and is illustrated as a manifestation of the benefits of the introduced infrastructure.

    Other authors
    See publication
  • Supporting the Provision of Specialized Taxonomic Hypermedia Services to Web Applications.

    In the International Workshop on architectures, models and infrastructures to generate semantics in Peer to Peer and Hypermedia Systems. ACM. Hypertext 2006 Conference.

    Other authors
  • Semantically annotated hypermedia services.

    In Proceedings of the sixteenth ACM conference on Hypertext and hypermedia (pp. 245-247). ACM.

    Other authors
    See publication
  • Developer Support in Open Hypermedia Systems: Towards a Hypermedia Service Discovery Mechanism.

    In Metainformatics (pp. 89-99). Springer Berlin Heidelberg.

    Other authors
    See publication
  • Babylon Bookmarks: A Taxonomic Approach to the Management of WWW Bookmarks

    In Metainformatics (pp. 42-48). Springer Berlin Heidelberg.

  • Increasing the usage of open hypermedia systems: a developer-side approach

    In Proceedings of the fourteenth ACM conference on Hypertext and hypermedia (HYPERTEXT '03). ACM, New York, NY, USA, 148-149.

  • Offering open hypermedia services to the WWW: a step-by-step approach for developers.

    In Proceedings of the 12th international conference on World Wide Web (WWW '03). ACM, New York, NY, USA, 482-489.

  • TPC-E vs. TPC-C: characterizing the new TPC-E benchmark via an I/O comparison study

    SIGMOD Record

    Other authors

Patents

  • Multi-level aggregation techniques for memory hierarchies

    Issued US 9195599

    Embodiments include method, system, and computer program product for providing aggregation hierarchy that is related memory hierarchies. In one embodiment, the method includes determining capacity of a first level memory of a memory hierarchy for processing data relating to completion of an aggregation process and generating a per thread local look-up table in said first level memory upon determining said capacity. Upon the first level memory reaching capacity, a plurality of per thread…

    Embodiments include method, system, and computer program product for providing aggregation hierarchy that is related memory hierarchies. In one embodiment, the method includes determining capacity of a first level memory of a memory hierarchy for processing data relating to completion of an aggregation process and generating a per thread local look-up table in said first level memory upon determining said capacity. Upon the first level memory reaching capacity, a plurality of per thread partitions to store remaining data to complete the aggregation process in a second level memory of the memory hierarchy is generated such that each of said per-thread partitions includes an identical amount of data portion on each thread. The method also includes storing the per thread partitions in said second level memory and providing a single global look up table for each of the identical data portions.

    Other inventors

Languages

  • German

    -

  • Greek

    -

More activity by Ippokratis

View Ippokratis’ full profile

  • See who you know in common
  • Get introduced
  • Contact Ippokratis directly
Join to view full profile

Other similar profiles

Explore collaborative articles

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Explore More

Add new skills with these courses