Connecting TIBCO Data Science - Team Studio to Data Sources
Review and follow these steps to connect your installation of TIBCO Data Science - Team Studio to your data sources.
Perform this task on the computer where you have installed
TIBCO Data Science - Team Studio.
Before you begin Test network connectivity and configure the
TIBCO Data Science - Team Studio server.
- Procedure
- Enable web sockets.
Verify that Web Sockets are correctly enabled by using a Web Socket test.
- Access the cluster nodes, including the NameNode and DataNodes for Hadoop.
Verify that you can connect to them by using the command $ telnet hostname port.
- Enable read and write permissions for the appropriate directories, including
/tmp for Hadoop.
Verify this step by writing to a file in one of those directories and running a MapReduce job, if applicable.
- Ensure that the appropriate agent is enabled for your data source.
- Configure the necessary ports in $CHORUS_HOME/shared/ALPINE_DATA_REPOSITORY/configuration/alpine.conf.
- If you are using Spark, ensure the following.
The Spark host is added in $CHORUS_HOME/shared/ALPINE_DATA_REPOSITORY/configuration/alpine.conf.
alpine.spark.sparkAkka.akka.remote.netty.tcp.hostname = IP address for Team Studio Server
Full communication is open between the TIBCO Data Science - Team Studio server and all cluster nodes. - Ensure the
TIBCO Data Science - Team Studio server can access the LDAP server if applicable.
Verify that you can connect by using
$ telnet hostname port.
What to do next
Connect to either a Database data source, a Hadoop data source, or a TIBCO Data Virtualization data source.
Subtopics