You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am unable to figure out how to point at a specific HDFS volume from the documentation associated with this project. I think it would simplify things to have a new option to support this and make the tool more useful "out of the box" for those who are new to hadoop/hdfs (such as myself) and are just looking to inspect the contents of a volume.
Here is the output from after the splash screen when running hadoopcli without any arguments:
java.lang.IllegalArgumentException: Wrong FS: hdfs://file:/user/joel, expected: file:///
at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:785)
at org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:89)
at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:645)
at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:939)
at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:640)
at org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:456)
at com.cloudera.utils.hadoop.hdfs.shell.command.HdfsCd.implementation(HdfsCd.java:109)
at com.cloudera.utils.hadoop.shell.command.AbstractCommand.execute(AbstractCommand.java:143)
at com.cloudera.utils.hadoop.shell.AbstractShell.processCommand(AbstractShell.java:355)
at com.cloudera.utils.hadoop.shell.AbstractShell.processInput(AbstractShell.java:269)
at com.cloudera.utils.hadoop.HadoopSession$1.call(HadoopSession.java:1122)
at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630)
at java.base/java.lang.Thread.run(Thread.java:832)
Wrong FS: hdfs://file:/user/joel, expected: file:///.
Attempted to set home directory. User home directory must exist.
If user is 'hdfs', consider using a proxy account for audit purposes.
Failed to Connect
Initialization Issue.
The text was updated successfully, but these errors were encountered:
I am unable to figure out how to point at a specific HDFS volume from the documentation associated with this project. I think it would simplify things to have a new option to support this and make the tool more useful "out of the box" for those who are new to hadoop/hdfs (such as myself) and are just looking to inspect the contents of a volume.
Here is the output from after the splash screen when running
hadoopcli
without any arguments:The text was updated successfully, but these errors were encountered: