In an HBase cluster, the most computationally expensive portion of reading or writing operations happens when we apply server-side filters on scan results; although, this computation is very much specific to accessing the data. Similarly, with the coprocessor, we can move a part of the computation to where the data lives, like in the case of Hadoop, which works in a distributed way for data storage (HDFS), as well as data processing (MapReduce). Using HBase coprocessors, custom features such as secondary indexing, complex filtering and access control features can be developed.

HBase coprocessor-based code run in parallel across all RegionServers and convert the cluster from horizontally scalable storage to a highly capable, distributed, ...

Get HBase Essentials now with the O’Reilly learning platform.

O’Reilly members experience live online training, plus books, videos, and digital content from nearly 200 publishers.