Configuring access to ADLS

Hortonworks Data Platform (HDP) supports reading and writing block blobs and page blobs from and to Windows Azure Storage Blob (WASB) object store, as well as reading and writing files stored in an Azure Data Lake Storage (ADLS) account.

Azure Data Lake Store (ADLS) is an enterprise-wide hyper-scale repository for big data analytic workloads. ADLS is not supported as a default file system, but access to data in ADLS is possible via the adl connector.

These steps assume that you are using an HDP version that supports the adl cloud storage connector (HDP 2.6.1 or newer).

Prerequisites

If you would like to use ADLS to store your data, you must enable Azure subscription for Data Lake Store, and then create an Azure Data Lake Store storage account.

Configure access to ADLS

To configure authentication with ADLS using the client credential, you must register a new application with Active Directory service and then give your application access to your ADL account. After you've performed these steps, you can provide your application's information when creating a cluster.

Prerequisites

In one of the steps you will be required to assign the Owner role to your application. If you do not have sufficient permissions, the role assignment step may have to be performed by your Azure admin.

Steps

  1. Register an application and add it to the ADLS account, as described in Step 1 and Step 2 of How to configure authentication with ADLS.

    Do not perform the Step 3 described in this article. Cloudbreak automates this step.

  2. In Cloudbreak web UI, on the advanced Cloud Storage page of the create a cluster wizard, select Use existing ADLS storage.

  3. Provide the following parameters for your registered application:

    • ADLS Account Name: This is the ADL account that your application was assigned to.
    • Application ID: You can find it in your application's settings.
    • Key: This is the key that you generated for your application. If you did not copy the it, you must create a new key from the Keys page in your application's settings.

Once your cluster is in the running state, you will be able to access the Azure blob storage account from the cluster nodes.

Test access to ADLS

To tests access to ADLS, SSH to a cluster node and run a few hadoop fs shell commands against your existing ADLS account.

ADLS access path syntax is:

adl://account_name.azuredatalakestore.net/dir/file

For example, the following Hadoop FileSystem shell commands demonstrate access to a storage account named "myaccount":

hadoop fs -mkdir adl://myaccount.azuredatalakestore.net/testdir
hadoop fs -put testfile adl://myaccount.azuredatalakestore.net/testdir/testfile

To use DistCp against ADLS, use the following syntax:

hadoop distcp
    [-D hadoop.security.credential.provider.path=localjceks://file/home/user/adls.jceks]
    hdfs://namenode_hostname:9001/user/foo/007020615
    adl://myaccount.azuredatalakestore.net/testDir/

For more information about configuring the ADLS connector and working with data stored in ADLS, refer to Cloud Data Access documentation.

Related links
Cloud Data Access (Hortonworks)
How to configure authentication with ADLS (Hortonworks)
Azure Data Lake Store (External)
Create a storage account (External)
Get started with Azure Data Lake Store (External)

Configure ADLS storage locations

After configuring access to ADLS, you can optionally use that ADLS storage account as a base storage location; this storage location is mainly for the Hive Warehouse Directory (used for storing the table data for managed tables).

Steps

  1. When creating a cluster, on the Cloud Storage page in the advanced cluster wizard view, select Use existing ADLS storage and select the instance profile to use, as described in Configure access to ADLS.
  2. Under Storage Locations, enable Configure Storage Locations by clicking the On button.
  3. Provide your existing directory name under Base Storage Location.

    Make sure that the directory exists within the account.

  4. Under Path for Hive Warehouse Directory property (hive.metastore.warehouse.dir), Cloudbreak automatically suggests a location within the bucket. For example, if the directory that you specified is my-test-dir then the suggested location will be my-test-adls-account.azuredatalakestore.net/my-test-dir/apps/hive/warehouse. You may optionally update this path.

    Cloudbreak automatically creates this directory structure in your bucket.