Contents
The TIBCO StreamBase® File System Adapter for HDFS performs various operations on the HDFS system as determined by the command value of the input tuple. HDFS is the well-known acronym for Apache's Hadoop Distributed File System.
The adapter shares a sample with one for the HDFS File Writer adapter, which is described in HDFS File Writer and File System Adapter Sample.
Note
When using the HDFS core libraries on Windows, you may see a console message
similar to Failed to locate the winutils binary in the hadoop
binary path
. The Apache Hadoop project acknowledges
that this is an error to be fixed in a future Hadoop release. In the meantime,
you can either ignore the error message as benign, or you can download, build, and
install the winutils executable as discussed on numerous Internet resources.
This section describes the properties you can set for this adapter, using the various tabs of the Properties view in StreamBase Studio.
Name: Use this field to specify or change the component's name, which must be unique in the application. The name must contain only alphabetic characters, numbers, and underscores, and no hyphens or other special characters. The first character must be alphabetic or an underscore.
Adapter: A read-only field that shows the formal name of the adapter.
Start with application: If this field is set to Yes or to a module parameter that evaluates to true, an instance of this adapter starts as part of the containing StreamBase Server. If this field is set to No or to a module parameter that evaluates to false, the adapter is loaded with the server, but does not start until you send an sbadmin resume command, or until you start the component with StreamBase Manager. With this option set to No or false, the adapter does not start even if the application as a whole is suspended and later resumed. The recommended setting is selected by default.
Enable Error Output Port: Select this check box to add an Error Port to this component. In the EventFlow canvas, the Error Port shows as a red output port, always the last port for the component. See Using Error Ports and Error Streams to learn about Error Ports.
Description: Optionally enter text to briefly describe the component's purpose and function. In the EventFlow canvas, you can see the description by pressing Ctrl while the component's tooltip is displayed.
Property | Data Type | Default | Description | StreamSQL Property |
---|---|---|---|---|
Enable Status Port | check box | Selected | If enabled, a status output port is present on which tuples are emitted to convey to important events within the adapter. | EnableStatusPort |
Enable HDFS Monitor Event Port | check box | Selected | If enabled, the startmonitor command may be used to start monitoring a path in the HDFS file system. When changes occur tuples will be emitted on this port. | EnableHDFSMonitorEventPort |
HDFS Configuration Files | Strings | Blank |
The HDFS configuration files to use when connecting to the HDFS file
system. For example, core-defaults.xml is the
standard file to use.
|
configurationFiles |
Log Level | drop-down list | INFO | Controls the level of verbosity the adapter uses to send notifications to the console. This setting can be higher than the containing application's log level. If set lower, the system log level will be used. Available values, in increasing order of verbosity, are: OFF, ERROR, WARN, INFO, DEBUG, TRACE, and ALL. | LogLevel |
Use the Concurrency tab to specify parallel regions for this instance of this component, or multiplicity options, or both. The Concurrency tab settings are described in Concurrency Options, and dispatch styles are described in Dispatch Styles.
Caution
Concurrency settings are not suitable for every application, and using these settings requires a thorough analysis of your application. For details, see Execution Order and Concurrency, which includes important guidelines for using the concurrency options.
The HDFS File System adapter's ports are used as follows:
-
Control (input): Tuples enqueued on this port cause the adapter to perform an operation on the HDFS file system. The schema for this port has the following fields:
-
Command, string. The command to perform on the HDFS file system. Valid commands are:
-
exists: Checks the HDFS file system and determines if a path exists. This works for both files and directories.
-
rename: Rename a file or directory in the HDFS file system; both Path and PathDestination fields are required.
-
listfiles: List files in a path. Use the Recursive flag to list all subdirectories recursively as well.
-
filestatus: Gets the file status for a path. This can be a file or a directory.
-
delete: Delete a path (file or directory).
-
status: Gets some system status information about the HDFS system (Capacity, Remaining, and Used).
-
startmonitor: Starts monitoring of an HDFS Path. Updates will be output on the monitor event port, which must be enabled.
-
stopmonitor: Stops monitoring of an HDFS Path.
-
-
Path, string or list<string>. The path(s) to use when performing an operation. This value is used for all commands but is option for status if no path is supplied for the status the status of the entire HDFS system is returned.
-
PathDestination, string. The destination path when performing a rename operation.
-
Recursive, boolean. Used during the delete and listfiles operations.
-
delete: If path is a directory and set to true, the directory is deleted and all its contents. If false and the directory is not empty an exception occurs.
-
listfiles: If the subdirectories need to be traversed recursively.
-
-
User, string. The user name to use when performing the operation. If null the user executing the
sbapp
is used. -
Permissions, tuple. A set of permissions to use when using the mkdirs operation.
-
User: The user permissions to set.
-
Group: The group permissions to set.
-
Other: The other permissions to set.
The value of the permissions string can be exactly one of the following:
-
NONE
-
EXECUTE
-
WRITE
-
WRITE_EXECUTE
-
READ
-
READ_EXECUTE
-
READ_WRITE
-
ALL
-
-
-
Data (output): The adapter emits tuples from this port when an input command required outputs:
-
Exists, boolean. Used for the exists command.
-
Files, list. Used for the listfiles and filestatus commands. This returns a detailed list of the files status.
-
Status, tuple. Used for the status command. This returns basic file system status.
-
PassThrough, tuple. Returns the input tuple that caused the output to occur.
-
-
Status (output): The adapter emits tuples from this port when significant events occur. The schema for this port has the following fields:
-
type, string. Returns the level type of this event (ERROR, WARN, INFO):
-
action, string. Returns an action associated with the event type which will be the command entered.
-
object, string. Returns an event type-specific value.
-
message, string. Returns a human-readable description of the event.
-
time, timestamp. The time the status event occurred.
-
inputTuple, tuple. Returns input tuple that caused the status message.
-
-
MonitorEvents (output): The adapter emits tuples from this port the startmonitor command is used and HDFS file system changes occur on the path being monitored:
-
Action, string. The type of action that occurred (append, close, create, metadata, rename, unlink, unknown).
-
Path, string. The path affected by this event.
-
PathDestination, string. If a rename operation, this is the destination path.
-
Length, long. The size of the file on close events.
-
CreationTime, timestamp. The time the path was created.
-
ModificationTime, timestamp. The time the path was changed.
-
Permissions, tuple. Returns a tuple containing the permissions if changed and the event is metadata.
-
The HDFS File System adapter uses typecheck messages to help you configure the adapter within your StreamBase application. In particular, the adapter generates typecheck messages for the following reasons:
-
The Control Input Port does not have exactly the correct fields with the correct type associated.
-
The adapter generates warning messages during runtime under various conditions.
When suspended, the adapter stops processing requests to perform operations on the file system.
When resumed, the adapter once again starts processing requests to perform operations on the file system.