Open timoninmaxim opened 6 years ago
Hm, honestly I have no idea what might be going on here. Does walk
fail to descend into the directories?
Hello, I have found the bug. hdfs3 does not work correctly with ACL.
$ python -c 'from hdfs3 import HDFileSystem; print HDFileSystem().ls("/user")' [u'/user/cloudera', u'/user/history', u'/user/hive', u'/user/hue', u'/user/jenkins', u'/user/oozie', u'/user/root', u'/user/spark'] $ hadoop fs -setfacl -m user:jenkins:rwx /user/cloudera $ python -c 'from hdfs3 import HDFileSystem; print HDFileSystem().ls("/user")' [] $ python -c 'from hdfs3 import HDFileSystem; print HDFileSystem().ls("/user/cloudera")' Traceback (most recent call last): File "
", line 1, in File "/opt/miniconda/lib/python2.7/site-packages/hdfs3/core.py", line 380, in ls raise FileNotFoundError(path) IOError: /user/cloudera
I found similar issue for libhdfs3-downstream https://github.com/ContinuumIO/libhdfs3-downstream/issues/4
It is fixed for master branch, but conda-forge goes to concat branch for installing libhdfs3 https://github.com/conda-forge/libhdfs3-feedstock/blob/47aec11797cf29907738a02941690ef81de2fcfd/recipe/build.sh#L3
Yes, I agree that libhdfs3 should be rereleased, I can try to get to that soon.
Hello, I try to find out a reason of the behaviour. It looks like a bug
hdfs3 was installed from the conda-forge.
So with simple hadoop ls commands:
And this is with hdfs3 lib. It does not show listing of /smartdata, but does it for /smartdata/hive.
I saw similar issue here. But it seems it is not permission issue: https://stackoverflow.com/questions/40405527/python-hdfs3-fails-to-list-non-owned-files