When I ran some PySpark queries in Zeppelin against an EMR cluster I have in AWS, I got the following error:
org.apache.hadoop.security.AccessControlException: Permission denied: user=zeppelin, access=WRITE, inode="/":hdfs:hadoop:drwxr-xr-x
I guess I can just give a Write permission to this user (Zeppelin) but the user already has its own home directory in HDFS (/user/zeppelin). Why is it trying to write to the root directory ("/"). I am not so sure I am interpreting the error message properly though.
I ended up giving a write permission for the user "zeppelin" like this:
sudo su hdfs
hdfs dfs -chown zeppelin /