It looks like '/user/test3' has owner '"hdfs" and denying the access while
performing operations via "shashi" user. One idea is to recursively set ACL
to sub-directories and files as follows:
hdfs dfs -setfacl -R -m default:user:shashi:rwx /user
-R, option can be used to apply operations to all files and
directories recursively.
Regards,
Rakesh
On Sun, Sep 18, 2016 at 8:53 PM, Shashi Vishwakarma <
[email protected]> wrote:
> I have following scenario. There is parent folder /user with five child
> folder as test1 , test2, test3 etc in HDFS.
>
> /user/test1
> /user/test2
> /user/test3
>
> I applied acl on parent folder to make sure user has automatically access
> to child folder.
>
> hdfs dfs -setfacl -m default:user:shashi:rwx /user
>
>
> but when i try to put some file , it is giving permission denied exception
>
> hadoop fs -put test.txt /user/test3
> put: Permission denied: user=shashi, access=WRITE,
> inode="/user/test3":hdfs:supergroup:drwxr-xr-x
>
> **getfacl output**
>
> hadoop fs -getfacl /user/test3
> # file: /user/test3
> # owner: hdfs
> # group: supergroup
> user::rwx
> group::r-x
> other::r-x
>
> Any pointers on this?
>
> Thanks
> Shashi
>