Apache Hue Issues

Troubleshoot Apache Hue issues for Big Data Service clusters.

Configuring Node Name in HA Cluster in Case of Stand by Node Failure

HDFS name node services, active and standby, are run on the two master nodes of an HA cluster.

By default, the first master node (mn0) is configured as the active namenode in Apache Hue. However, any of the master nodes can act as the active namenode. If the standby node gets configured as the active node, then a standby node exception error is encountered.

Example:

Screen shot of Apache Ambari showing a name node failure error message.

Change the Hue config in the Ambari UI and restart the Hue server (in Ambari).

  1. Access Apache Ambari.
  2. From the side toolbar, under Services select Hue.
  3. Select the Configs tab to modify the Apache Hue configurations.
  4. Select Advanced.
  5. From the Advanced hue-hadoop-site section, change the HDFS HttpsFS Host field value from mn0 to mn1 or vice versa.
  6. Select Actions, and then select Restart All.
Running Spark Shell Workflow Using Oozie Workflow in Hue For Non-HA Cluster

Run Spark shell workflow using Oozie workflow in Hue for Big Data Service non-ha clusters.

The Spark jobs on yarn launched through shell action in Oozie workflow are always ran as yarn. This issue only occurs for the non-HA clusters.

If user "xyz" runs a Spark command from shell action, the resulting Spark job on Resource manager runs as the "yarn" user. In this case, the "yarn" user must have all relevant permissions that user "xyz" requires to run the Spark job. If user "xyz" writes to a location on HDFS, then the "yarn" user must also have write permission to that directory in HDFS.