Installation
Requirements
The software requirement for DataFibers are
Java 8
Apache Kafka >= 0.11.0
Apache Flink >= 1.3.0
MongoDB
The Java runtime version is at least version 1.8. Java installation varies by platform, so please check the JRE installation version for your platform before installing the DataFibers using following command:
Various installation options are listed in the sections below. Select the approach you prefer and follow the given instructions to install the Confluent Platform.
Operation System
Right now, DataFibers runs on Linux Ubuntu system. You can install it locally or in virtual machines, such as Oracle VirtualBox.
Environment Setup - Docker
To be added
Environment Setup - Vagrant
The pre-build scripts for vagrant setup is available in DF_DEMO repository. This pre-build VM has following main software installed automatically.
Java 1.8
Git 1.9.1
Apache Maven 3.0.5
Apache Hadoop 2.6
Apache Hive 1.2.1
Apache Flink 1.3.2
Confluent 3.3.0 with Apache Kafka 0.11.0.0
MongoDB 3.4
Please following steps below to get the VM up running.
Run below commands either in Linux or GitBash Console in Windows to download the setup scripts
To run default environment setup, run
Below are common folders used by the DataFibers package.
/opt : Dependent software, such as confluent/Kafka, Hive, Hadoop, Flink, are installed here.
/mnt/etc : The config used for dependent software.
/mnt/logs : Where to keep logs.
/mnt/dfs : Hadoop file system are mount here.
Install DF Packages
To install DataFibers packages as well as scripts tools, run the follow command in the folder where you want to setup as follows.
After running above installation script, following folders and artifacts are installed in the current folder.
conf/ : All configuration files are here.
lib/ : Where to keep certified connect and service jars.
repo/ : Where to keep source code.
bin/ : Where to keep operation and administration scripts.
In addition, the installation folder (current folder) is set as $DF_HOME variable in ~/.profile. $DF_HOME/bin is added to the system PATH. The df_ops or dfops alias is also added to the ~/.profile so the command can be accessed anywhere. For more details regarding to operation commands, please refer to this section.
Last updated