Search in sources :

Example 1 with FSINFO3Response

use of org.apache.hadoop.nfs.nfs3.response.FSINFO3Response in project hadoop by apache.

the class FSINFO3Response method deserialize.

public static FSINFO3Response deserialize(XDR xdr) {
    int status = xdr.readInt();
    xdr.readBoolean();
    Nfs3FileAttributes postOpObjAttr = Nfs3FileAttributes.deserialize(xdr);
    int rtmax = 0;
    int rtpref = 0;
    int rtmult = 0;
    int wtmax = 0;
    int wtpref = 0;
    int wtmult = 0;
    int dtpref = 0;
    long maxFileSize = 0;
    NfsTime timeDelta = null;
    int properties = 0;
    if (status == Nfs3Status.NFS3_OK) {
        rtmax = xdr.readInt();
        rtpref = xdr.readInt();
        rtmult = xdr.readInt();
        wtmax = xdr.readInt();
        wtpref = xdr.readInt();
        wtmult = xdr.readInt();
        dtpref = xdr.readInt();
        maxFileSize = xdr.readHyper();
        timeDelta = NfsTime.deserialize(xdr);
        properties = xdr.readInt();
    }
    return new FSINFO3Response(status, postOpObjAttr, rtmax, rtpref, rtmult, wtmax, wtpref, wtmult, dtpref, maxFileSize, timeDelta, properties);
}
Also used : Nfs3FileAttributes(org.apache.hadoop.nfs.nfs3.Nfs3FileAttributes) NfsTime(org.apache.hadoop.nfs.NfsTime)

Example 2 with FSINFO3Response

use of org.apache.hadoop.nfs.nfs3.response.FSINFO3Response in project hadoop by apache.

the class RpcProgramNfs3 method fsinfo.

@VisibleForTesting
FSINFO3Response fsinfo(XDR xdr, SecurityHandler securityHandler, SocketAddress remoteAddress) {
    FSINFO3Response response = new FSINFO3Response(Nfs3Status.NFS3_OK);
    if (!checkAccessPrivilege(remoteAddress, AccessPrivilege.READ_ONLY)) {
        response.setStatus(Nfs3Status.NFS3ERR_ACCES);
        return response;
    }
    DFSClient dfsClient = clientCache.getDfsClient(securityHandler.getUser());
    if (dfsClient == null) {
        response.setStatus(Nfs3Status.NFS3ERR_SERVERFAULT);
        return response;
    }
    FSINFO3Request request;
    try {
        request = FSINFO3Request.deserialize(xdr);
    } catch (IOException e) {
        LOG.error("Invalid FSINFO request");
        return new FSINFO3Response(Nfs3Status.NFS3ERR_INVAL);
    }
    FileHandle handle = request.getHandle();
    if (LOG.isDebugEnabled()) {
        LOG.debug("NFS FSINFO fileId: " + handle.getFileId() + " client: " + remoteAddress);
    }
    try {
        int rtmax = config.getInt(NfsConfigKeys.DFS_NFS_MAX_READ_TRANSFER_SIZE_KEY, NfsConfigKeys.DFS_NFS_MAX_READ_TRANSFER_SIZE_DEFAULT);
        int wtmax = config.getInt(NfsConfigKeys.DFS_NFS_MAX_WRITE_TRANSFER_SIZE_KEY, NfsConfigKeys.DFS_NFS_MAX_WRITE_TRANSFER_SIZE_DEFAULT);
        int dtperf = config.getInt(NfsConfigKeys.DFS_NFS_MAX_READDIR_TRANSFER_SIZE_KEY, NfsConfigKeys.DFS_NFS_MAX_READDIR_TRANSFER_SIZE_DEFAULT);
        Nfs3FileAttributes attrs = Nfs3Utils.getFileAttr(dfsClient, Nfs3Utils.getFileIdPath(handle), iug);
        if (attrs == null) {
            LOG.info("Can't get path for fileId: " + handle.getFileId());
            return new FSINFO3Response(Nfs3Status.NFS3ERR_STALE);
        }
        int fsProperty = Nfs3Constant.FSF3_CANSETTIME | Nfs3Constant.FSF3_HOMOGENEOUS;
        return new FSINFO3Response(Nfs3Status.NFS3_OK, attrs, rtmax, rtmax, 1, wtmax, wtmax, 1, dtperf, Long.MAX_VALUE, new NfsTime(1), fsProperty);
    } catch (IOException e) {
        LOG.warn("Exception ", e);
        int status = mapErrorStatus(e);
        return new FSINFO3Response(status);
    }
}
Also used : DFSClient(org.apache.hadoop.hdfs.DFSClient) FSINFO3Response(org.apache.hadoop.nfs.nfs3.response.FSINFO3Response) FileHandle(org.apache.hadoop.nfs.nfs3.FileHandle) Nfs3FileAttributes(org.apache.hadoop.nfs.nfs3.Nfs3FileAttributes) NfsTime(org.apache.hadoop.nfs.NfsTime) FSINFO3Request(org.apache.hadoop.nfs.nfs3.request.FSINFO3Request) IOException(java.io.IOException) VisibleForTesting(com.google.common.annotations.VisibleForTesting)

Example 3 with FSINFO3Response

use of org.apache.hadoop.nfs.nfs3.response.FSINFO3Response in project hadoop by apache.

the class TestRpcProgramNfs3 method testFsinfo.

@Test(timeout = 60000)
public void testFsinfo() throws Exception {
    HdfsFileStatus status = nn.getRpcServer().getFileInfo("/tmp/bar");
    long dirId = status.getFileId();
    FileHandle handle = new FileHandle(dirId);
    XDR xdr_req = new XDR();
    FSINFO3Request req = new FSINFO3Request(handle);
    req.serialize(xdr_req);
    // Attempt by an unpriviledged user should fail.
    FSINFO3Response response1 = nfsd.fsinfo(xdr_req.asReadOnlyWrap(), securityHandlerUnpriviledged, new InetSocketAddress("localhost", 1234));
    assertEquals("Incorrect return code:", Nfs3Status.NFS3ERR_ACCES, response1.getStatus());
    // Attempt by a priviledged user should pass.
    FSINFO3Response response2 = nfsd.fsinfo(xdr_req.asReadOnlyWrap(), securityHandler, new InetSocketAddress("localhost", 1234));
    assertEquals("Incorrect return code:", Nfs3Status.NFS3_OK, response2.getStatus());
}
Also used : FSINFO3Response(org.apache.hadoop.nfs.nfs3.response.FSINFO3Response) FileHandle(org.apache.hadoop.nfs.nfs3.FileHandle) InetSocketAddress(java.net.InetSocketAddress) HdfsFileStatus(org.apache.hadoop.hdfs.protocol.HdfsFileStatus) XDR(org.apache.hadoop.oncrpc.XDR) FSINFO3Request(org.apache.hadoop.nfs.nfs3.request.FSINFO3Request) Test(org.junit.Test)

Aggregations

NfsTime (org.apache.hadoop.nfs.NfsTime)2 FileHandle (org.apache.hadoop.nfs.nfs3.FileHandle)2 Nfs3FileAttributes (org.apache.hadoop.nfs.nfs3.Nfs3FileAttributes)2 FSINFO3Request (org.apache.hadoop.nfs.nfs3.request.FSINFO3Request)2 FSINFO3Response (org.apache.hadoop.nfs.nfs3.response.FSINFO3Response)2 VisibleForTesting (com.google.common.annotations.VisibleForTesting)1 IOException (java.io.IOException)1 InetSocketAddress (java.net.InetSocketAddress)1 DFSClient (org.apache.hadoop.hdfs.DFSClient)1 HdfsFileStatus (org.apache.hadoop.hdfs.protocol.HdfsFileStatus)1 XDR (org.apache.hadoop.oncrpc.XDR)1 Test (org.junit.Test)1