Contents
The TIBCO StreamBase® Adapter for Apache HBase is implemented as a suite of five global Java operators, including the HBase Admin, Delete, Get, Put, and Scan operators.
This page describes the HBase Scan operator, which allows a StreamBase application to return the entire contents of, or a range of rows from, a specified HBase table. The operator uses property values in the project's server configuration file to set up the connection to the HBase database, as described in Configuration File Settings. Multiple HBase operators can share a single instance of an HBase connection by selecting the same HBase configuration setting ID.
This section describes the configuration for an HBase database connection instance that you specify in your project's sbconf file. This configuration is the same for all HBase operator types.
The <adapter-configurations>
element of a project's sbconf file, despite its name, is used to specify configuration value groups for either operators
or adapters.
The HBase configuration section of the sbconf file starts with an <adapter-configurations>
element that contains one <adapter-configuration name="hbase">
element. This element, in turn, contains one or more <setting>
elements.
Each <section name="hbase">
element must contain one element in the form <setting name="id" val="
, where HBaseConfigName
"/>HBaseConfigName
is the name you assign to a group of settings that uniquely define an individual HBase database connection. All other <setting>
elements are optional.
The example configuration below shows a basic configuration to connect to an HBase server. You can have as many configurations
as your application requires, but each configuration must have a unique id
.
Example 1. Example <adapter-configuration> Section for HBase
<adapter-configurations> <adapter-configuration name="hbase"> <section name="hbase"> <setting name="id" val="HBase Sample"/> <setting name="connectAtStartup" val="true" /> <!-- All values below are passed directly to the HBaseConfiguration class unmodified --/> <setting name="hbase.master" val="127.0.0.1:60000" /> <setting name="hbase.zookeeper.quorum" val="127.0.0.1" /> <setting name="hbase.zookeeper.property.clientPort" val="2181" /> <setting name="hbase.client.retries.number" val="5" /> <setting name="zookeeper.session.timeout" val="5000" /> <setting name="zookeeper.recovery.retry" val="5" /> </section> </adapter-configuration> </adapter-configurations>
Setting | Type | Description |
---|---|---|
id | string | The value of the id setting displays in the drop-down list in the adapter's Properties view, and is used to uniquely identify this section of
the configuration file.
|
connectAtStartup | true or false | If true, this operator instance connects to HBase on startup of this operator's containing module. |
*** | string | All other values are directly sent to the HBaseConfiguration class, which is responsible for setting up a connection to the HBase server. See the Apache HBase documentation for the available client configuration options and for further information on setting up a connection to HBase. |
This section describes the properties you can set for an HBase Scan operator, using the various tabs of the Properties view in StreamBase Studio.
Name: Use this field to specify or change the component's name, which must be unique in the application. The name must contain only alphabetic characters, numbers, and underscores, and no hyphens or other special characters. The first character must be alphabetic or an underscore.
Operator: A read-only field that shows the formal name of the operator.
Class: A field that shows the fully qualified class name that implements the functionality of this operator. Use this class name when loading the operator in StreamSQL programs with the APPLY JAVA statement. You can right-click this field and select Copy from the context menu to place the full class name in the system clipboard.
Start with application: If this field is set to Yes or to a module parameter that evaluates to true, an instance of this operator starts as part of the containing StreamBase Server. If this field is set to No or to a module parameter that evaluates to false, the adapter is loaded with the server, but does not start until you send an sbadmin resume command, or until you start the component with StreamBase Manager. With this option set to No or false, the operator does not start even if the application as a whole is suspended and later resumed. The recommended setting is selected by default.
Enable Error Output Port: Select this check box to add an Error Port to this component. In the EventFlow canvas, the Error Port shows as a red output port, always the last port for the component. See Using Error Ports and Error Streams to learn about Error Ports.
Description: Optionally enter text to briefly describe the component's purpose and function. In the EventFlow canvas, you can see the description by pressing Ctrl while the component's tooltip is displayed.
Property | Type | Description |
---|---|---|
HBase Config | drop-down list | The name of the HBase configuration to use with this operator. The value selected by this drop-down list determines the database connection this operator works against. The values that populate this list are stored in the project's sbconf file, as described in the Configuration File Settings section. |
Table Name | string | The HBase table that this operation is to be performed against. |
Max Versions | int | Specifies the maximum number of versions to fetch for each row from the database. If this value is 1 the output schema will represent a schema for a single version output. If however the version is greater than one the output schema will be modified to represent a list of values with their corresponding timestamp values. |
Enable Pass Through Fields | check box | If enabled, the fields passed into the operator are copied into a user defined field in the output schema. |
Pass Through Field Name | string | The name of the field in the output schema that is to hold the pass-through fields. |
Filter Factory Class | string | The fully qualified name of the class that implements com.streambase.sb.operator.hbase.IHBaseFilterFactory to create the org.apache.hadoop.hbase.filter.FilterList to perform filtering on the rows returned by this operation. Leave blank for no filtering. See the Filter Factory Interface
section for more on creating a FilterList class.
|
Output Tuple Per Row | check box | If enabled, the operator will send an output tuple for each row of the result set. If not enabled the operator will output a single tuple with a list of rows when the query is complete. |
Output Null Row Tuple On Scan Complete | check box | If enabled the operator will send an output tuple with a null value for the rowId when the scan operation has completed. |
Enable Status Port | check box | If enabled, a status port is made available for this operator instance, which will emit status tuples for various events from this operator. |
Log Level | Drop-down list | Controls the level of verbosity the adapter uses to issue informational traces to the console. This setting is independent of the containing application's overall log level. Available values, in increasing order of verbosity, are: OFF, ERROR, WARN, INFO, DEBUG, TRACE, and ALL. |
Property | Type | Description |
---|---|---|
Time Range Min Field Name | string | Specifies the field in the incoming schema to be used to specify the starting point timestamp of a time range. Only columns
and versions of columns within the specified time range are retrieved; a value of null or empty means do not use.
The default Max Versions setting to return is 1. If your time range spans more than one version and you want all versions within the time range returned, increase the Max Versions setting to a value greater than one. |
Time Range Max Field Name | string | Specifies the field in the incoming schema to be used to specify the end point timestamp of a time range. Only columns and
versions of columns within the specified time range are retrieved; a value of null or empty means do not use.
The default Max Versions setting to return is 1. If your time range spans more than one version and you want all versions within the range returned, increase the Max Versions setting to a value greater than one. |
Start Row Field Name | string | Specifies the field name in the inbound schema that represents the row to start fetching data from. If the field resolves to a null, the value will not be used. |
Stop Row Field Name | string | Specifies the field name in the inbound schema that represents the row at which to stop fetching data. If the field resolves to a null, the value will not be used. |
Property | Type | Description |
---|---|---|
Column Filter Type | "Range" or "Columns" | Indicates the type of column filtering to perform. |
Column Filter Min Range | string | The minimum value for the column range. If left blank, there is no lower bound. |
Column Filter Min Range Inclusive | check box | If enabled, the value of the minimum column itself is to be included in the range. |
Column Filter Max Range | string | The maximum value for the column range. If left blank, there is no upper bound. |
Column Filter Max Range Inclusive | check box | If enabled, the value of the maximum column itself is to be included in the range. |
Columns | field grid | The columns to return with this query. The Family column of the field grid must have a value, but the Column column can be blank, which directs the operator to include all columns of the specified Family. |
Property | Type | Description |
---|---|---|
Row Id Type | drop-down list | The data type used for the Row ID field. |
Convert Value To String | check box | This option only applies if no user-defined schema is provided. If enabled, this option tries to convert all cell values to a string; if not enabled, all values will be a blob. |
HBase Schema | schema grid | The schema used for scan data results. If no schema is provided, the schema is a list of tuples containing the family column value sets. |
Use the Concurrency tab to specify parallel regions for this instance of this component, or multiplicity options, or both. The Concurrency tab settings are described in Concurrency Options, and dispatch styles are described in Dispatch Styles.
Caution
Concurrency settings are not suitable for every application, and using these settings requires a thorough analysis of your application. For details, see Execution Order and Concurrency, which includes important guidelines for using the concurrency options.
The filter factory interface (com.streambase.sb.operator.hbase.IHBaseFilterFactory
) is used to create a FilterList for an HBase Scan or Get operation. The Filter list allows HBase to return only those rows
that match the criteria provided by the specified filters. The interface allows an interaction between the inbound tuple and
the Scan or Get operations. See the HBase API documentation for further information about the org.apache.hadoop.hbase.filter.FilterBase
class and its sub-classes, and for how-to information on filtering data.
When you have compiled a Java class to serve as a filter, provide the fully qualified name of your class in the Filter Factor Class field in the Operator Properties tab in your HBase Scan or Get operator's Properties view.
The next section provides an example class file that takes the inbound tuples and uses some of the fields of each tuple to create a filter to use for the read operation.
package com.streambase.sb.hbase.filterfactory; import org.apache.hadoop.hbase.filter.CompareFilter.CompareOp; import org.apache.hadoop.hbase.filter.FilterList; import org.apache.hadoop.hbase.filter.SingleColumnValueFilter; import org.apache.hadoop.hbase.filter.SubstringComparator; import org.apache.hadoop.hbase.util.Bytes; import com.streambase.sb.Tuple; import com.streambase.sb.operator.hbase.IHBaseFilterFactory; public class DemoFilterFactory implements IHBaseFilterFactory { @Override public FilterList createFilterList(Tuple tuple) throws Exception { FilterList filterList = new FilterList(FilterList.Operator.MUST_PASS_ALL); filterList.addFilter(new SingleColumnValueFilter(Bytes.toBytes(tuple.getString("family")), Bytes.toBytes(tuple.getString("column")), CompareOp.EQUAL, new SubstringComparator(tuple.getString("matchSubString")))); return filterList; } }