5. Known Issues

Ambari 2.4 has the following known issues, scheduled for resolution in a future release. Also, refer to the Ambari Troubleshooting Guide for additional information.

Table 1.5. Ambari 2.4 Known Issues

Apache Jira

HWX Jira

Problem

Solution

N/A

BUG-65117

When installing or registering a Stack, there is an option for an HDP 2.3.ECS stack.

The HDP 2.3.ECS stack option is not supported.

The HDP 2.3.ECS stack option is not supported. Do not install or register the HDP 2.3.ECS stack. If you are looking to install the HDP 2.3 stack, choose the HDP 2.3 option and not HDP 2.3.ECS. This HDP 2.3.ECS stack option will be removed in a future Ambari maintenance release.

N/A

BUG-64966

When performing a rolling upgrade of a secure, HDP 2.4 cluster running Atlas on SLES11 sp3, the Atlas service fails to upgrade. You may see the following error:

resource_management.core.exceptions.Fail: Execution of 'ambari-sudo.sh /usr/bin/hdp-select set all
 `ambari-python-wrap /usr/bin/hdp-select versions | grep ^2.5.0.0-1238 | tail -1`' returned 1.
 symlink target /usr/hdp/current/atlas-client for atlas already exists and it is not a symlink.

For all environments, we recommend removing Atlas from the cluster before a performing a rolling upgrade, as described here.

Specific upgrade steps for a secure HDP 2.4 cluster running Atlas on SLES 11 sp3:

  1. Delete the Atlas service via the Ambari API2.

  2. Upgrade Ambari to version 2.4.03.

  3. Upgrade Ambari to version 2.4.04.

  4. Install HDP-2.5.x bits.

  5. Perform an Express Upgrade.

  6. Upgrade the stack.

  7. Reinstall Atlas.

  8. Restart all impacted services using Services > Restart All.

N/A

BUG-64959

If you have Storm in your cluster and have Kerberos enabled, after Ambari 2.4 upgrade, the Storm summary page in Ambari Web does not show information and exceptions are logged.

24 Aug 2016 14:19:38,107 ERROR [ambari-metrics-retrieval-service-thread-2738] MetricsRetrievalService:421
 - Unable to retrieve metrics from http://hcube1-1n02.eng.hortonworks.com:8744/api/v1/cluster/summary.
 Subsequent failures will be suppressed from the log for 20 minutes.
java.io.IOException: Server returned HTTP response code: 500 for URL: 
http://hcube1-1n02.eng.hortonworks.com:8744/api/v1/cluster/summary
 at sun.reflect.GeneratedConstructorAccessor228.newInstance(Unknown Source)
 at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)

You need to add the Ambari Server principal name to the Storm nimbus.admins property in Ambari Web > Services > Storm > Configs.

N/A

BUG-64947

Filtering by Role on Users page in Ambari Admin Interface is slow with 1000+ users.

If you have 1000+ users from LDAP synchronized into Ambari and you attempt to filtering by Role on the Users page in Ambari Admin Interface, the results will be slow (potential 20 seconds). Reducing the number of users you have synchronized into Ambari is a way to improve this performance.

N/A

BUG-64912

Apache Oozie requires a restart after an Atlas configuration update, but may not be included in the services marked as requiring restart in Ambari.

Select Oozie > Service Actions > Restart All to restart Oozie along with the other services.
N/A

BUG-64896

Upon adding Atlas to an HDP-2.5 cluster that is kerberized it reports error while trying to communicate with Kafka. You may see the following in the error :

"Could not login: the client is being asked for a password, but the Kafka client code
 does not currently support obtaining a password from the user.
Not available to garner authentication information from the user"

This is because the keytab and principal details may be missing from the Atlas config.

Add the following config properties to Atlas configuration type application-properties

Set atlas.jaas.KafkaClient.option.keyTab to the Atlas Service keytag

e.g. /etc/security/keytabs/atlas.service.keytabSet atlas.jaas.KafkaClient.option.principal to the principal name in the above keytab

e.g. atlas/atlas_server_host_name@EXAMPLE.COM

Then, restart Atlas Server.

N/A

BUG-64819

After performing a Rolling Upgrade, the “Percent JournalNodes Available” alert goes CRITICAL.

After performing a Rolling Upgrade, the “Percent JournalNodes Available” alert goes critical even though all the JournalNodes are started and healthy. You must disable and re-enable this alert to clear the CRITICAL status. In Ambari Web, browse to Alerts and find the “Percent JournalNodes Available” alert. Click to Disable the alert. Once disabled completes, re-enable and the alert status should be OK.

AMBARI-18177

BUG-64326

If you add or remove ZooKeeper servers and you have Atlas in your cluster, you must update some Atlas properties.

If you are running Atlas in your cluster and make changes to ZooKeeper (adding or removing ZooKeeper server components), you must update the Atlas properties list below to make sure they reflect the list of ZooKeeper servers in the cluster. You can modify the following properties from Ambari Web > Services > Atlas > Configs:

atlas.audit.hbase.zookeeper.quorum
Sample: node-test0.docker.nxa.io:2181,node-test2.docker.nxa.io:2181,node-test1.docker.nxa.io:2181

atlas.graph.index.search.solr.zookeeper-url
Sample: node-test1.docker.nxa.io:2181/infra-solr,node-test0.docker.nxa.io:2181/infra-solr,node-test2.docker.nxa.io:2181/infra-solr

atlas.graph.storage.hostname
Sample: node-test0.docker.nxa.io,node-test2.docker.nxa.io,node-test1.docker.nxa.io

atlas.kafka.zookeeper.connect
Sample: node-test0.docker.nxa.io,node-test2.docker.nxa.io,node-test1.docker.nxa.io 
N/A

BUG-62609

When upgrading Ambari 2.x to 2.4 on a secure, cluster running Storm on Ubuntu or SLES, the Storm service check fails to restart Nimbus server due to a randomized StormClient principal.

You must manually reset the original StormClient principal on the kerberos AD server.

N/A

BUG-57093

When toggling and saving the Hive service's "Enable Interactive Query" configuration, it's important to wait for the background operations associated with that change to complete before trying to toggle it once more. If you don't wait for background operations to complete and try to re-enable this feature it can cause some of the operations to fail resulting in configuration inconsistencies.

If such inconsistencies are observed then retry toggling and saving of the "Enabling Interactive Query" configuration and wait for all requests (background operations) to complete.

AMBARI-12436

BUG-40481

Falcon Service Check may fail when performing Rolling Upgrade or downgrade, with the following error:

2015-06-25 18:09:07,235 ERROR - [main:]
 ~ Failed to start ActiveMQ JMS Message Broker.
 Reason: java.io.IOException: Invalid location: 1:6763311, :
 java.lang.NegativeArraySizeException (BrokerService:528) 
 java.io.IOException: Invalid location: 1:6763311, :
 java.lang.NegativeArraySizeException
 at
 org.apache.kahadb.journal.DataFileAccessor.readRecord(DataFileAccessor.java:94)

This condition is rare.

When performing a Rolling Upgrade from HDP 2.2 to HDP 2.3 (or upgrade > dowgrade > upgrade) and Falcon Service Check fails with the above error, browse to the Falcon ActiveMQ data dir (specified in falcon properties file), remove the corrupted queues, and stop and start the Falcon Server.

cd <ACTIVEMQ_DATA_DIR>
rm -rf ./localhost
cd /usr/hdp/current/falcon-server 
su -l <FALCON_USER> 
./bin/falcon-stop
./bin/falcon-start
AMBARI-12283

BUG-40300

After adding or deleting ZooKeeper Servers to an existing cluster, Service Check fails.

After adding or deleting ZooKeeper Servers to an existing cluster, Service Check fails due to conflicting zk ids. Restart ZooKeeper service to clear the ids.

N/A

BUG-40694

The Slider view is not supported on a cluster with SSL (wire encryption) enabled. Only use the Slider view on clusters without wire encryption enabled. If it is required to run Slider on a cluster with wire encryption enabled, please contact Hortonworks support for further help.