hdfs cleanup and compile test against CDH 4.4.
This commit is contained in:
parent
bd1105aa5a
commit
c2fda55cfe
@ -289,7 +289,7 @@ if test "$USE_HDFS"; then
|
|||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
HDFS_CCFLAGS="$HDFS_CCFLAGS -I$JAVA_HOME/include -I$JAVA_HOME/include/linux -DUSE_HDFS"
|
HDFS_CCFLAGS="$HDFS_CCFLAGS -I$JAVA_HOME/include -I$JAVA_HOME/include/linux -DUSE_HDFS"
|
||||||
HDFS_LDFLAGS="$HDFS_LDFLAGS -Wl,--no-whole-archive hdfs/libhdfs.a -L$JAVA_HOME/jre/lib/amd64"
|
HDFS_LDFLAGS="$HDFS_LDFLAGS -Wl,--no-whole-archive -lhdfs -L$JAVA_HOME/jre/lib/amd64"
|
||||||
HDFS_LDFLAGS="$HDFS_LDFLAGS -L$JAVA_HOME/jre/lib/amd64/server -L$GLIBC_RUNTIME_PATH/lib"
|
HDFS_LDFLAGS="$HDFS_LDFLAGS -L$JAVA_HOME/jre/lib/amd64/server -L$GLIBC_RUNTIME_PATH/lib"
|
||||||
HDFS_LDFLAGS="$HDFS_LDFLAGS -ldl -lverify -ljava -ljvm"
|
HDFS_LDFLAGS="$HDFS_LDFLAGS -ldl -lverify -ljava -ljvm"
|
||||||
COMMON_FLAGS="$COMMON_FLAGS $HDFS_CCFLAGS"
|
COMMON_FLAGS="$COMMON_FLAGS $HDFS_CCFLAGS"
|
||||||
|
13
hdfs/README
13
hdfs/README
@ -1,19 +1,16 @@
|
|||||||
This directory contains the hdfs extensions needed to make rocksdb store
|
This directory contains the hdfs extensions needed to make rocksdb store
|
||||||
files in HDFS.
|
files in HDFS.
|
||||||
|
|
||||||
The hdfs.h file is copied from the Apache Hadoop 1.0 source code.
|
It has been compiled and testing against CDH 4.4 (2.0.0+1475-1.cdh4.4.0.p0.23~precise-cdh4.4.0).
|
||||||
It defines the libhdfs library
|
|
||||||
(http://hadoop.apache.org/common/docs/r0.20.2/libhdfs.html) to access
|
The configuration assumes that packages libhdfs0, libhdfs0-dev are
|
||||||
data in HDFS. The libhdfs.a is copied from the Apache Hadoop 1.0 build.
|
installed which basically means that hdfs.h is in /usr/include and libhdfs in /usr/lib
|
||||||
It implements the API defined in hdfs.h. If your hadoop cluster is running
|
|
||||||
a different hadoop release, then install these two files manually from your
|
|
||||||
hadoop distribution and then recompile rocksdb.
|
|
||||||
|
|
||||||
The env_hdfs.h file defines the rocksdb objects that are needed to talk to an
|
The env_hdfs.h file defines the rocksdb objects that are needed to talk to an
|
||||||
underlying filesystem.
|
underlying filesystem.
|
||||||
|
|
||||||
If you want to compile rocksdb with hdfs support, please set the following
|
If you want to compile rocksdb with hdfs support, please set the following
|
||||||
enviroment variables appropriately:
|
enviroment variables appropriately (also defined in setup.sh for convenience)
|
||||||
USE_HDFS=1
|
USE_HDFS=1
|
||||||
JAVA_HOME=/usr/local/jdk-6u22-64
|
JAVA_HOME=/usr/local/jdk-6u22-64
|
||||||
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/jdk-6u22-64/jre/lib/amd64/server:/usr/local/jdk-6u22-64/jre/lib/amd64/:./snappy/libs
|
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/local/jdk-6u22-64/jre/lib/amd64/server:/usr/local/jdk-6u22-64/jre/lib/amd64/:./snappy/libs
|
||||||
|
@ -14,7 +14,7 @@
|
|||||||
#include "rocksdb/status.h"
|
#include "rocksdb/status.h"
|
||||||
|
|
||||||
#ifdef USE_HDFS
|
#ifdef USE_HDFS
|
||||||
#include "hdfs/hdfs.h"
|
#include <hdfs.h>
|
||||||
|
|
||||||
namespace rocksdb {
|
namespace rocksdb {
|
||||||
|
|
||||||
|
477
hdfs/hdfs.h
477
hdfs/hdfs.h
@ -1,477 +0,0 @@
|
|||||||
// Copyright (c) 2013, Facebook, Inc. All rights reserved.
|
|
||||||
// This source code is licensed under the BSD-style license found in the
|
|
||||||
// LICENSE file in the root directory of this source tree. An additional grant
|
|
||||||
// of patent rights can be found in the PATENTS file in the same directory.
|
|
||||||
//
|
|
||||||
#ifndef LIBHDFS_HDFS_H
|
|
||||||
#define LIBHDFS_HDFS_H
|
|
||||||
|
|
||||||
#include <sys/types.h>
|
|
||||||
#include <sys/stat.h>
|
|
||||||
|
|
||||||
#include <fcntl.h>
|
|
||||||
#include <stdio.h>
|
|
||||||
#include <stdint.h>
|
|
||||||
#include <string.h>
|
|
||||||
#include <stdlib.h>
|
|
||||||
#include <time.h>
|
|
||||||
#include <errno.h>
|
|
||||||
|
|
||||||
#include <jni.h>
|
|
||||||
|
|
||||||
#ifndef O_RDONLY
|
|
||||||
#define O_RDONLY 1
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef O_WRONLY
|
|
||||||
#define O_WRONLY 2
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifndef EINTERNAL
|
|
||||||
#define EINTERNAL 255
|
|
||||||
#endif
|
|
||||||
|
|
||||||
|
|
||||||
/** All APIs set errno to meaningful values */
|
|
||||||
#ifdef __cplusplus
|
|
||||||
extern "C" {
|
|
||||||
#endif
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Some utility decls used in libhdfs.
|
|
||||||
*/
|
|
||||||
|
|
||||||
typedef int32_t tSize; /// size of data for read/write io ops
|
|
||||||
typedef time_t tTime; /// time type in seconds
|
|
||||||
typedef int64_t tOffset;/// offset within the file
|
|
||||||
typedef uint16_t tPort; /// port
|
|
||||||
typedef enum tObjectKind {
|
|
||||||
kObjectKindFile = 'F',
|
|
||||||
kObjectKindDirectory = 'D',
|
|
||||||
} tObjectKind;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The C reflection of org.apache.org.hadoop.FileSystem .
|
|
||||||
*/
|
|
||||||
typedef void* hdfsFS;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The C equivalent of org.apache.org.hadoop.FSData(Input|Output)Stream .
|
|
||||||
*/
|
|
||||||
enum hdfsStreamType
|
|
||||||
{
|
|
||||||
UNINITIALIZED = 0,
|
|
||||||
INPUT = 1,
|
|
||||||
OUTPUT = 2,
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* The 'file-handle' to a file in hdfs.
|
|
||||||
*/
|
|
||||||
struct hdfsFile_internal {
|
|
||||||
void* file;
|
|
||||||
enum hdfsStreamType type;
|
|
||||||
};
|
|
||||||
typedef struct hdfsFile_internal* hdfsFile;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsConnectAsUser - Connect to a hdfs file system as a specific user
|
|
||||||
* Connect to the hdfs.
|
|
||||||
* @param host A string containing either a host name, or an ip address
|
|
||||||
* of the namenode of a hdfs cluster. 'host' should be passed as NULL if
|
|
||||||
* you want to connect to local filesystem. 'host' should be passed as
|
|
||||||
* 'default' (and port as 0) to used the 'configured' filesystem
|
|
||||||
* (core-site/core-default.xml).
|
|
||||||
* @param port The port on which the server is listening.
|
|
||||||
* @param user the user name (this is hadoop domain user). Or NULL is equivelant to hhdfsConnect(host, port)
|
|
||||||
* @param groups the groups (these are hadoop domain groups)
|
|
||||||
* @return Returns a handle to the filesystem or NULL on error.
|
|
||||||
*/
|
|
||||||
hdfsFS hdfsConnectAsUser(const char* host, tPort port, const char *user , const char *groups[], int groups_size );
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsConnect - Connect to a hdfs file system.
|
|
||||||
* Connect to the hdfs.
|
|
||||||
* @param host A string containing either a host name, or an ip address
|
|
||||||
* of the namenode of a hdfs cluster. 'host' should be passed as NULL if
|
|
||||||
* you want to connect to local filesystem. 'host' should be passed as
|
|
||||||
* 'default' (and port as 0) to used the 'configured' filesystem
|
|
||||||
* (core-site/core-default.xml).
|
|
||||||
* @param port The port on which the server is listening.
|
|
||||||
* @return Returns a handle to the filesystem or NULL on error.
|
|
||||||
*/
|
|
||||||
hdfsFS hdfsConnect(const char* host, tPort port);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* This are the same as hdfsConnectAsUser except that every invocation returns a new FileSystem handle.
|
|
||||||
* Applications should call a hdfsDisconnect for every call to hdfsConnectAsUserNewInstance.
|
|
||||||
*/
|
|
||||||
hdfsFS hdfsConnectAsUserNewInstance(const char* host, tPort port, const char *user , const char *groups[], int groups_size );
|
|
||||||
hdfsFS hdfsConnectNewInstance(const char* host, tPort port);
|
|
||||||
hdfsFS hdfsConnectPath(const char* uri);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsDisconnect - Disconnect from the hdfs file system.
|
|
||||||
* Disconnect from hdfs.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsDisconnect(hdfsFS fs);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsOpenFile - Open a hdfs file in given mode.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The full path to the file.
|
|
||||||
* @param flags - an | of bits/fcntl.h file flags - supported flags are O_RDONLY, O_WRONLY (meaning create or overwrite i.e., implies O_TRUNCAT),
|
|
||||||
* O_WRONLY|O_APPEND. Other flags are generally ignored other than (O_RDWR || (O_EXCL & O_CREAT)) which return NULL and set errno equal ENOTSUP.
|
|
||||||
* @param bufferSize Size of buffer for read/write - pass 0 if you want
|
|
||||||
* to use the default configured values.
|
|
||||||
* @param replication Block replication - pass 0 if you want to use
|
|
||||||
* the default configured values.
|
|
||||||
* @param blocksize Size of block - pass 0 if you want to use the
|
|
||||||
* default configured values.
|
|
||||||
* @return Returns the handle to the open file or NULL on error.
|
|
||||||
*/
|
|
||||||
hdfsFile hdfsOpenFile(hdfsFS fs, const char* path, int flags,
|
|
||||||
int bufferSize, short replication, tSize blocksize);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsCloseFile - Close an open file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsCloseFile(hdfsFS fs, hdfsFile file);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsExists - Checks if a given path exsits on the filesystem
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path to look for
|
|
||||||
* @return Returns 0 on exists, 1 on non-exists, -1/-2 on error.
|
|
||||||
*/
|
|
||||||
int hdfsExists(hdfsFS fs, const char *path);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsSeek - Seek to given offset in file.
|
|
||||||
* This works only for files opened in read-only mode.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @param desiredPos Offset into the file to seek into.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsSeek(hdfsFS fs, hdfsFile file, tOffset desiredPos);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsTell - Get the current offset in the file, in bytes.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @return Current offset, -1 on error.
|
|
||||||
*/
|
|
||||||
tOffset hdfsTell(hdfsFS fs, hdfsFile file);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsRead - Read data from an open file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @param buffer The buffer to copy read bytes into.
|
|
||||||
* @param length The length of the buffer.
|
|
||||||
* @return Returns the number of bytes actually read, possibly less
|
|
||||||
* than than length;-1 on error.
|
|
||||||
*/
|
|
||||||
tSize hdfsRead(hdfsFS fs, hdfsFile file, void* buffer, tSize length);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsPread - Positional read of data from an open file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @param position Position from which to read
|
|
||||||
* @param buffer The buffer to copy read bytes into.
|
|
||||||
* @param length The length of the buffer.
|
|
||||||
* @return Returns the number of bytes actually read, possibly less than
|
|
||||||
* than length;-1 on error.
|
|
||||||
*/
|
|
||||||
tSize hdfsPread(hdfsFS fs, hdfsFile file, tOffset position,
|
|
||||||
void* buffer, tSize length);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsWrite - Write data into an open file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @param buffer The data.
|
|
||||||
* @param length The no. of bytes to write.
|
|
||||||
* @return Returns the number of bytes written, -1 on error.
|
|
||||||
*/
|
|
||||||
tSize hdfsWrite(hdfsFS fs, hdfsFile file, const void* buffer,
|
|
||||||
tSize length);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsWrite - Flush the data.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsFlush(hdfsFS fs, hdfsFile file);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsSync - Sync the data to persistent store.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsSync(hdfsFS fs, hdfsFile file);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetNumReplicasInPipeline - get number of remaining replicas in
|
|
||||||
* pipeline
|
|
||||||
* @param fs The configured filesystem handle
|
|
||||||
* @param file the file handle
|
|
||||||
* @return returns the # of datanodes in the write pipeline; -1 on error
|
|
||||||
*/
|
|
||||||
int hdfsGetNumCurrentReplicas(hdfsFS, hdfsFile file);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsAvailable - Number of bytes that can be read from this
|
|
||||||
* input stream without blocking.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param file The file handle.
|
|
||||||
* @return Returns available bytes; -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsAvailable(hdfsFS fs, hdfsFile file);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsCopy - Copy file from one filesystem to another.
|
|
||||||
* @param srcFS The handle to source filesystem.
|
|
||||||
* @param src The path of source file.
|
|
||||||
* @param dstFS The handle to destination filesystem.
|
|
||||||
* @param dst The path of destination file.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsCopy(hdfsFS srcFS, const char* src, hdfsFS dstFS, const char* dst);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsMove - Move file from one filesystem to another.
|
|
||||||
* @param srcFS The handle to source filesystem.
|
|
||||||
* @param src The path of source file.
|
|
||||||
* @param dstFS The handle to destination filesystem.
|
|
||||||
* @param dst The path of destination file.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsMove(hdfsFS srcFS, const char* src, hdfsFS dstFS, const char* dst);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsDelete - Delete file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the file.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsDelete(hdfsFS fs, const char* path);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsRename - Rename file.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param oldPath The path of the source file.
|
|
||||||
* @param newPath The path of the destination file.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsRename(hdfsFS fs, const char* oldPath, const char* newPath);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetWorkingDirectory - Get the current working directory for
|
|
||||||
* the given filesystem.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param buffer The user-buffer to copy path of cwd into.
|
|
||||||
* @param bufferSize The length of user-buffer.
|
|
||||||
* @return Returns buffer, NULL on error.
|
|
||||||
*/
|
|
||||||
char* hdfsGetWorkingDirectory(hdfsFS fs, char *buffer, size_t bufferSize);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsSetWorkingDirectory - Set the working directory. All relative
|
|
||||||
* paths will be resolved relative to it.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the new 'cwd'.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsSetWorkingDirectory(hdfsFS fs, const char* path);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsCreateDirectory - Make the given file and all non-existent
|
|
||||||
* parents into directories.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the directory.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsCreateDirectory(hdfsFS fs, const char* path);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsSetReplication - Set the replication of the specified
|
|
||||||
* file to the supplied value
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the file.
|
|
||||||
* @return Returns 0 on success, -1 on error.
|
|
||||||
*/
|
|
||||||
int hdfsSetReplication(hdfsFS fs, const char* path, int16_t replication);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsFileInfo - Information about a file/directory.
|
|
||||||
*/
|
|
||||||
typedef struct {
|
|
||||||
tObjectKind mKind; /* file or directory */
|
|
||||||
char *mName; /* the name of the file */
|
|
||||||
tTime mLastMod; /* the last modification time for the file in seconds */
|
|
||||||
tOffset mSize; /* the size of the file in bytes */
|
|
||||||
short mReplication; /* the count of replicas */
|
|
||||||
tOffset mBlockSize; /* the block size for the file */
|
|
||||||
char *mOwner; /* the owner of the file */
|
|
||||||
char *mGroup; /* the group associated with the file */
|
|
||||||
short mPermissions; /* the permissions associated with the file */
|
|
||||||
tTime mLastAccess; /* the last access time for the file in seconds */
|
|
||||||
} hdfsFileInfo;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsListDirectory - Get list of files/directories for a given
|
|
||||||
* directory-path. hdfsFreeFileInfo should be called to deallocate memory if
|
|
||||||
* the function returns non-NULL value.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the directory.
|
|
||||||
* @param numEntries Set to the number of files/directories in path.
|
|
||||||
* @return Returns a dynamically-allocated array of hdfsFileInfo
|
|
||||||
* objects; NULL if empty or on error.
|
|
||||||
* on error, numEntries will be -1.
|
|
||||||
*/
|
|
||||||
hdfsFileInfo *hdfsListDirectory(hdfsFS fs, const char* path,
|
|
||||||
int *numEntries);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetPathInfo - Get information about a path as a (dynamically
|
|
||||||
* allocated) single hdfsFileInfo struct. hdfsFreeFileInfo should be
|
|
||||||
* called when the pointer is no longer needed.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the file.
|
|
||||||
* @return Returns a dynamically-allocated hdfsFileInfo object;
|
|
||||||
* NULL on error.
|
|
||||||
*/
|
|
||||||
hdfsFileInfo *hdfsGetPathInfo(hdfsFS fs, const char* path);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsFreeFileInfo - Free up the hdfsFileInfo array (including fields)
|
|
||||||
* @param hdfsFileInfo The array of dynamically-allocated hdfsFileInfo
|
|
||||||
* objects.
|
|
||||||
* @param numEntries The size of the array.
|
|
||||||
*/
|
|
||||||
void hdfsFreeFileInfo(hdfsFileInfo *hdfsFileInfo, int numEntries);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetHosts - Get hostnames where a particular block (determined by
|
|
||||||
* pos & blocksize) of a file is stored. The last element in the array
|
|
||||||
* is NULL. Due to replication, a single block could be present on
|
|
||||||
* multiple hosts.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path The path of the file.
|
|
||||||
* @param start The start of the block.
|
|
||||||
* @param length The length of the block.
|
|
||||||
* @return Returns a dynamically-allocated 2-d array of blocks-hosts;
|
|
||||||
* NULL on error.
|
|
||||||
*/
|
|
||||||
char*** hdfsGetHosts(hdfsFS fs, const char* path,
|
|
||||||
tOffset start, tOffset length);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsFreeHosts - Free up the structure returned by hdfsGetHosts
|
|
||||||
* @param hdfsFileInfo The array of dynamically-allocated hdfsFileInfo
|
|
||||||
* objects.
|
|
||||||
* @param numEntries The size of the array.
|
|
||||||
*/
|
|
||||||
void hdfsFreeHosts(char ***blockHosts);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetDefaultBlockSize - Get the optimum blocksize.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @return Returns the blocksize; -1 on error.
|
|
||||||
*/
|
|
||||||
tOffset hdfsGetDefaultBlockSize(hdfsFS fs);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetCapacity - Return the raw capacity of the filesystem.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @return Returns the raw-capacity; -1 on error.
|
|
||||||
*/
|
|
||||||
tOffset hdfsGetCapacity(hdfsFS fs);
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsGetUsed - Return the total raw size of all files in the filesystem.
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @return Returns the total-size; -1 on error.
|
|
||||||
*/
|
|
||||||
tOffset hdfsGetUsed(hdfsFS fs);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsChown
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path the path to the file or directory
|
|
||||||
* @param owner this is a string in Hadoop land. Set to null or "" if only setting group
|
|
||||||
* @param group this is a string in Hadoop land. Set to null or "" if only setting user
|
|
||||||
* @return 0 on success else -1
|
|
||||||
*/
|
|
||||||
int hdfsChown(hdfsFS fs, const char* path, const char *owner, const char *group);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsChmod
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path the path to the file or directory
|
|
||||||
* @param mode the bitmask to set it to
|
|
||||||
* @return 0 on success else -1
|
|
||||||
*/
|
|
||||||
int hdfsChmod(hdfsFS fs, const char* path, short mode);
|
|
||||||
|
|
||||||
/**
|
|
||||||
* hdfsUtime
|
|
||||||
* @param fs The configured filesystem handle.
|
|
||||||
* @param path the path to the file or directory
|
|
||||||
* @param mtime new modification time or 0 for only set access time in seconds
|
|
||||||
* @param atime new access time or 0 for only set modification time in seconds
|
|
||||||
* @return 0 on success else -1
|
|
||||||
*/
|
|
||||||
int hdfsUtime(hdfsFS fs, const char* path, tTime mtime, tTime atime);
|
|
||||||
|
|
||||||
#ifdef __cplusplus
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#endif /*LIBHDFS_HDFS_H*/
|
|
||||||
|
|
||||||
/**
|
|
||||||
* vim: ts=4: sw=4: et
|
|
||||||
*/
|
|
BIN
hdfs/libhdfs.a
BIN
hdfs/libhdfs.a
Binary file not shown.
7
hdfs/setup.sh
Normal file
7
hdfs/setup.sh
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
export USE_HDFS=1
|
||||||
|
export LD_LIBRARY_PATH=$JAVA_HOME/jre/lib/amd64/server:$JAVA_HOME/jre/lib/amd64:/usr/lib/hadoop/lib/native
|
||||||
|
|
||||||
|
export CLASSPATH=
|
||||||
|
for f in `find /usr/lib/hadoop-hdfs | grep jar`; do export CLASSPATH=$CLASSPATH:$f; done
|
||||||
|
for f in `find /usr/lib/hadoop | grep jar`; do export CLASSPATH=$CLASSPATH:$f; done
|
||||||
|
for f in `find /usr/lib/hadoop/client | grep jar`; do export CLASSPATH=$CLASSPATH:$f; done
|
@ -15,11 +15,11 @@
|
|||||||
#include <sstream>
|
#include <sstream>
|
||||||
#include "rocksdb/env.h"
|
#include "rocksdb/env.h"
|
||||||
#include "rocksdb/status.h"
|
#include "rocksdb/status.h"
|
||||||
#include "hdfs/hdfs.h"
|
|
||||||
#include "hdfs/env_hdfs.h"
|
#include "hdfs/env_hdfs.h"
|
||||||
|
|
||||||
#define HDFS_EXISTS 0
|
#define HDFS_EXISTS 0
|
||||||
#define HDFS_DOESNT_EXIST 1
|
#define HDFS_DOESNT_EXIST -1
|
||||||
|
#define HDFS_SUCCESS 0
|
||||||
|
|
||||||
//
|
//
|
||||||
// This file defines an HDFS environment for rocksdb. It uses the libhdfs
|
// This file defines an HDFS environment for rocksdb. It uses the libhdfs
|
||||||
@ -223,7 +223,7 @@ class HdfsWritableFile: public WritableFile {
|
|||||||
if (hdfsFlush(fileSys_, hfile_) == -1) {
|
if (hdfsFlush(fileSys_, hfile_) == -1) {
|
||||||
return IOError(filename_, errno);
|
return IOError(filename_, errno);
|
||||||
}
|
}
|
||||||
if (hdfsSync(fileSys_, hfile_) == -1) {
|
if (hdfsHSync(fileSys_, hfile_) == -1) {
|
||||||
return IOError(filename_, errno);
|
return IOError(filename_, errno);
|
||||||
}
|
}
|
||||||
Log(mylog, "[hdfs] HdfsWritableFile Synced %s\n", filename_.c_str());
|
Log(mylog, "[hdfs] HdfsWritableFile Synced %s\n", filename_.c_str());
|
||||||
@ -398,12 +398,40 @@ Status HdfsEnv::NewRandomRWFile(const std::string& fname,
|
|||||||
return Status::NotSupported("NewRandomRWFile not supported on HdfsEnv");
|
return Status::NotSupported("NewRandomRWFile not supported on HdfsEnv");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
class HdfsDirectory : public Directory {
|
||||||
|
public:
|
||||||
|
explicit HdfsDirectory(int fd) : fd_(fd) {}
|
||||||
|
~HdfsDirectory() {
|
||||||
|
//close(fd_);
|
||||||
|
}
|
||||||
|
|
||||||
|
virtual Status Fsync() {
|
||||||
|
//if (fsync(fd_) == -1) {
|
||||||
|
// return IOError("directory", errno);
|
||||||
|
// }
|
||||||
|
return Status::OK();
|
||||||
|
}
|
||||||
|
|
||||||
|
private:
|
||||||
|
int fd_;
|
||||||
|
};
|
||||||
|
|
||||||
Status HdfsEnv::NewDirectory(const std::string& name,
|
Status HdfsEnv::NewDirectory(const std::string& name,
|
||||||
unique_ptr<Directory>* result) {
|
unique_ptr<Directory>* result) {
|
||||||
return Status::NotSupported("NewDirectory not supported on HdfsEnv");
|
|
||||||
|
int value = hdfsCreateDirectory(fileSys_, name.c_str());
|
||||||
|
result->reset(new HdfsDirectory(0));
|
||||||
|
switch (value) {
|
||||||
|
case HDFS_SUCCESS: // directory created
|
||||||
|
return Status::OK();
|
||||||
|
default:
|
||||||
|
Log(mylog, "directory already exists ");
|
||||||
|
return Status::OK();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bool HdfsEnv::FileExists(const std::string& fname) {
|
bool HdfsEnv::FileExists(const std::string& fname) {
|
||||||
|
|
||||||
int value = hdfsExists(fileSys_, fname.c_str());
|
int value = hdfsExists(fileSys_, fname.c_str());
|
||||||
switch (value) {
|
switch (value) {
|
||||||
case HDFS_EXISTS:
|
case HDFS_EXISTS:
|
||||||
@ -412,8 +440,8 @@ bool HdfsEnv::FileExists(const std::string& fname) {
|
|||||||
return false;
|
return false;
|
||||||
default: // anything else should be an error
|
default: // anything else should be an error
|
||||||
Log(mylog, "FileExists hdfsExists call failed");
|
Log(mylog, "FileExists hdfsExists call failed");
|
||||||
throw HdfsFatalException("hdfsExists call failed with error " +
|
throw HdfsFatalException("1. hdfsExists call failed with error " +
|
||||||
std::to_string(value) + ".\n");
|
std::to_string(value) + " on path " + fname + ".\n");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -449,13 +477,13 @@ Status HdfsEnv::GetChildren(const std::string& path,
|
|||||||
default: // anything else should be an error
|
default: // anything else should be an error
|
||||||
Log(mylog, "GetChildren hdfsExists call failed");
|
Log(mylog, "GetChildren hdfsExists call failed");
|
||||||
throw HdfsFatalException("hdfsExists call failed with error " +
|
throw HdfsFatalException("hdfsExists call failed with error " +
|
||||||
std::to_string(value) + ".\n");
|
std::to_string(value) + " on path " + path.c_str() + ".\n");
|
||||||
}
|
}
|
||||||
return Status::OK();
|
return Status::OK();
|
||||||
}
|
}
|
||||||
|
|
||||||
Status HdfsEnv::DeleteFile(const std::string& fname) {
|
Status HdfsEnv::DeleteFile(const std::string& fname) {
|
||||||
if (hdfsDelete(fileSys_, fname.c_str()) == 0) {
|
if (hdfsDelete(fileSys_, fname.c_str(),1) == 0) {
|
||||||
return Status::OK();
|
return Status::OK();
|
||||||
}
|
}
|
||||||
return IOError(fname, errno);
|
return IOError(fname, errno);
|
||||||
@ -478,7 +506,7 @@ Status HdfsEnv::CreateDirIfMissing(const std::string& name) {
|
|||||||
return CreateDir(name);
|
return CreateDir(name);
|
||||||
default: // anything else should be an error
|
default: // anything else should be an error
|
||||||
Log(mylog, "CreateDirIfMissing hdfsExists call failed");
|
Log(mylog, "CreateDirIfMissing hdfsExists call failed");
|
||||||
throw HdfsFatalException("hdfsExists call failed with error " +
|
throw HdfsFatalException("3. hdfsExists call failed with error " +
|
||||||
std::to_string(value) + ".\n");
|
std::to_string(value) + ".\n");
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
@ -514,7 +542,7 @@ Status HdfsEnv::GetFileModificationTime(const std::string& fname,
|
|||||||
// target already exists. So, we delete the target before attemting the
|
// target already exists. So, we delete the target before attemting the
|
||||||
// rename.
|
// rename.
|
||||||
Status HdfsEnv::RenameFile(const std::string& src, const std::string& target) {
|
Status HdfsEnv::RenameFile(const std::string& src, const std::string& target) {
|
||||||
hdfsDelete(fileSys_, target.c_str());
|
hdfsDelete(fileSys_, target.c_str(), 1);
|
||||||
if (hdfsRename(fileSys_, src.c_str(), target.c_str()) == 0) {
|
if (hdfsRename(fileSys_, src.c_str(), target.c_str()) == 0) {
|
||||||
return Status::OK();
|
return Status::OK();
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user