Hadoop Distributed File System
Hadoop Distributed File System (HDFS) is a file system that manages large data sets that can run on commodity hardware.
If you select Hadoop Distributed File System (HDFS) from the Storage section, configure the following details:
| Field | Description |
|---|---|
| Display name | Enter the name to be displayed. |
| Thrift URI | Enter the Thrift URI. |
| Thrift Port | Enter the Thrift port. |
| Kerberos authentication | Use the toggle switch to enable or disable Kerberos authentication. If enabled, enter the following information: a. Enter the following information: i. HDFS principal ii. Hive client principal iii. Hive server principal b. Upload the following files: i. Core site file (.xml) ii. HDFS site (.xml) iii. Kerberos config file (.config) iiii. HDFS keytab file (.keytab) iiiii. Hive keytab file (.keytab) |
| Upload core site file (.xml) | Upload core site file (.xml) |
| Upload HDFS site file (.xml) | Upload HDFS site file (.xml) |
| Associate catalog | Add a catalog for your storage. This catalog is associated with your storage and serves as your query interface with the data stored within. |
| Catalog type | The supported catalog is Apache Hive. |
| Catalog name | Enter the name of your catalog. |
| Associate | Click Associate to create the storage. |
Related API
For information on related API, see