Thanks Rakesh. Just last question, is there any Java API available for recursively applying ACL or I need to iterate on all folders of dir and apply acl for each?
Thanks Shashi On 19 Sep 2016 9:56 am, "Rakesh Radhakrishnan" <[email protected]> wrote: > It looks like '/user/test3' has owner '"hdfs" and denying the > access while performing operations via "shashi" user. One idea is to > recursively set ACL to sub-directories and files as follows: > > hdfs dfs -setfacl -R -m default:user:shashi:rwx /user > > -R, option can be used to apply operations to all files and > directories recursively. > > Regards, > Rakesh > > On Sun, Sep 18, 2016 at 8:53 PM, Shashi Vishwakarma < > [email protected]> wrote: > >> I have following scenario. There is parent folder /user with five child >> folder as test1 , test2, test3 etc in HDFS. >> >> /user/test1 >> /user/test2 >> /user/test3 >> >> I applied acl on parent folder to make sure user has automatically access >> to child folder. >> >> hdfs dfs -setfacl -m default:user:shashi:rwx /user >> >> >> but when i try to put some file , it is giving permission denied exception >> >> hadoop fs -put test.txt /user/test3 >> put: Permission denied: user=shashi, access=WRITE, >> inode="/user/test3":hdfs:supergroup:drwxr-xr-x >> >> **getfacl output** >> >> hadoop fs -getfacl /user/test3 >> # file: /user/test3 >> # owner: hdfs >> # group: supergroup >> user::rwx >> group::r-x >> other::r-x >> >> Any pointers on this? >> >> Thanks >> Shashi >> > >
