hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Chris Nauroth (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-8973) DiskChecker cannot reliably detect an inaccessible disk on Windows with NTFS ACLs
Date Tue, 05 Mar 2013 00:07:13 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-8973?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13592830#comment-13592830

Chris Nauroth commented on HADOOP-8973:

That's interesting.  Yes, I believe this is a bug in the existing code for the other overload
of {{DiskChecker#checkDir}}.

For example, suppose a dfs.datanode.data.dir on the local file system with owner "foo" and
perms set to 700.  Now suppose we launch datanode as user "bar".  {{DiskChecker#checkDir}}
will just look for 700 and not consider the running user, so it will think that the directory
is usable.  Then, it would experience an I/O error later whenever the process first tries
to use that directory.

I'll file a separate jira for this.

> DiskChecker cannot reliably detect an inaccessible disk on Windows with NTFS ACLs
> ---------------------------------------------------------------------------------
>                 Key: HADOOP-8973
>                 URL: https://issues.apache.org/jira/browse/HADOOP-8973
>             Project: Hadoop Common
>          Issue Type: Bug
>          Components: util
>    Affects Versions: trunk-win
>            Reporter: Chris Nauroth
>            Assignee: Chris Nauroth
>         Attachments: HADOOP-8973-branch-trunk-win.patch
> DiskChecker.checkDir uses File.canRead, File.canWrite, and File.canExecute to check if
a directory is inaccessible.  These APIs are not reliable on Windows with NTFS ACLs due to
a known JVM bug.

This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

View raw message