I'm seeing a really weird issue with a 2.6.0 datanode (CDH 5.16). In short we have a cluster running HDFS 2.6.0 (CDH 5.16), a 'hdfs' CLI app running HDFS 3.0.0 (CDH 6.3), and an application using HDP 2.8.5. (CDH 6.x - not sure which version). If it matters the cluster uses Kerberos authentication and TLS encryption.
The hdfs CLI can do everything expected. This includes '-put' and '-get'. The app can do all datanode operations and can *read* existing files but it gives us 'no blocks available' when we try to *write* to a new file. This exception is seen in the datanode log. Finally the app works with a CDH 6.1.1 cluster. It has an identical configuration. I'm at a loss since it's clearly not an issue with the server configuration, client configuration, app implementation, etc. We're trying to write a file of < 20 bytes to a cluster with over 90 GB of free space. The only difference is the client library - and I would expect the opposite results if that were the cause. (Why would the newer version work when the older version failed?) Does anyone have any ideas? P.S., the last version of the app worked with a prior 5.x cluster. However we've rebuilt the cluster (including bumping to 5.16) and made changes to the app so we can't really use it as a data point. -- Bear Giles Sr. Software Engineer [email protected] Mobile: 720-749-7876 <http://www.snaplogic.com/about-us/jobs> *SnapLogic Inc | 1825 South Grant Street | San Mateo CA | USA * This message is confidential. It may also be privileged or otherwise protected by work product immunity or other legal rules. If you have received it by mistake, please let us know by e-mail reply and delete it from your system; you may not copy this message or disclose its contents to anyone. The integrity and security of this message cannot be guaranteed on the Internet.
