Hdfs3ReadStream(hdfsFS fs, hdfsFile file, uint64_t start_byte, uint64_t /* byte_count */) : fs_(fs), file_(file) { int err = hdfsSeek(fs_, file_, start_byte); die_unless(err == 0); }
ssize_t HdfsFile::read(void *ptr, size_t count) { size_t offset = tell(); tSize ret = hdfsRead(m_fs, m_file, ptr, count); // see comment in pread if( ret < 0 ) { reopen(); if (hdfsSeek(m_fs, m_file, offset) == 0) ret = hdfsRead(m_fs, m_file, ptr, count); } else if ((size_t) ret < count) // ret >= 0 { // retry int retryCount = 0; while (ret >= 0 && ret != (tSize) count && ++retryCount < 40) { ssize_t n = hdfsRead(m_fs, m_file, (char*)ptr+ret, count-ret); if (n >= 0) ret += n; else ret = n; } } if( IDBLogger::isEnabled() ) IDBLogger::logRW("read", m_fname, this, offset, count, ret); return ret; }
HDFSChunkReaderIterator::HDFSChunkReaderIterator(const ChunkID& chunk_id, unsigned& chunk_size,const unsigned& block_size) :ChunkReaderIterator(chunk_id,block_size,chunk_size){ block_buffer_=new Block(block_size_); fs_=hdfsConnect(Config::hdfs_master_ip.c_str(),Config::hdfs_master_port); hdfs_fd_=hdfsOpenFile(fs_,chunk_id.partition_id.getName().c_str(),O_RDONLY,0,0,0); if(!hdfs_fd_){ printf("fails to open HDFS file [%s]\n",chunk_id.partition_id.getName().c_str()); number_of_blocks_=0; } const unsigned start_pos=start_pos+CHUNK_SIZE*chunk_id_.chunk_off; if(hdfsSeek(fs_,hdfs_fd_,start_pos)==-1){ printf("fails to set the start offset %d for [%s]\n",start_pos,chunk_id.partition_id.getName().c_str()); number_of_blocks_=0; } hdfsFileInfo *file_info=hdfsGetPathInfo(fs_,"/imdb/");// to be refined after communicating with Zhang Lei if(start_pos+CHUNK_SIZE<file_info->mSize){ number_of_blocks_=CHUNK_SIZE/block_size_; } else{ number_of_blocks_=(file_info->mSize-start_pos)/block_size_; } hdfsFreeFileInfo(file_info,1); }
int MaprCopyingInputStream::Skip(int count){ CHECK(input_->is_open_); CHECK_GE(count, 0); tOffset start_pos = hdfsTell(input_->fs_, input_->file_); tOffset desired_pos = start_pos + count; //TODO(kheath): add optimization by avoiding extra "tell" if seek is successful hdfsSeek(input_->fs_, input_->file_, desired_pos); // success if returns 0 tOffset cur_pos = hdfsTell(input_->fs_, input_->file_); int actual_skipped = cur_pos - start_pos; return actual_skipped; }
int HDFSFileSplitter::read_block(const std::string& fn) { file_ = hdfsOpenFile(fs_, fn.c_str(), O_RDONLY, 0, 0, 0); assert(file_ != NULL); hdfsSeek(fs_, file_, offset_); size_t start = 0; size_t nbytes = 0; while (start < hdfs_block_size) { // only 128KB per hdfsRead nbytes = hdfsRead(fs_, file_, data_ + start, hdfs_block_size); start += nbytes; if (nbytes == 0) break; } return start; }
ssize_t HdfsFile::read(void *ptr, size_t count) { boost::mutex::scoped_lock lock(m_mutex); int savedErrno; tOffset offset = tell(); if (offset < 0) return offset; /* May get a performance boost by implementing read() s.t. it doesn't require a seek afterward, but probably not though. */ ssize_t numRead = real_pread(ptr, offset, count); savedErrno = errno; if (numRead > 0) hdfsSeek(m_fs, m_file, offset + numRead); errno = savedErrno; return numRead; }
int HdfsFile::seek(off64_t offset, int whence) { off_t mod_offset = offset; if( whence == SEEK_CUR ) { mod_offset = mod_offset + tell(); } else if( whence == SEEK_END ) { mod_offset = mod_offset + size(); } int ret = hdfsSeek(m_fs, m_file, mod_offset); if( IDBLogger::isEnabled() ) IDBLogger::logSeek(m_fname, this, offset, whence, ret); return ret; }
int HdfsFile::seek(off64_t offset, int whence) { boost::mutex::scoped_lock lock(m_mutex); int savedErrno; off_t mod_offset = offset; if( whence == SEEK_CUR ) { mod_offset = mod_offset + tell(); } else if( whence == SEEK_END ) { mod_offset = mod_offset + size(); } int ret = hdfsSeek(m_fs, m_file, mod_offset); savedErrno = errno; if( IDBLogger::isEnabled() ) IDBLogger::logSeek(m_fname, this, offset, whence, ret); errno = savedErrno; return ret; }
qioerr hdfs_seek(void* fl, off_t offset, int whence, off_t* offset_out, void* fs) { off_t got; qioerr err_out = 0; // We cannot seek unless we are in read mode! (HDFS restriction) if (to_hdfs_file(fl)->file->type != INPUT) QIO_RETURN_CONSTANT_ERROR(ENOSYS, "Seeking is not supported in write mode in HDFS"); STARTING_SLOW_SYSCALL; got = (off_t)hdfsSeek(to_hdfs_fs(fs)->hfs, to_hdfs_file(fl)->file, offset); if( got != (off_t) -1) { *offset_out = got; } else { *offset_out = got; } DONE_SLOW_SYSCALL; return err_out; }
static int vecsum_libhdfs(struct libhdfs_data *ldata, const struct options *opts) { int pass; ldata->buf = malloc(NORMAL_READ_CHUNK_SIZE); if (!ldata->buf) { fprintf(stderr, "failed to malloc buffer of size %d\n", NORMAL_READ_CHUNK_SIZE); return ENOMEM; } for (pass = 0; pass < opts->passes; ++pass) { int ret = vecsum_normal_loop(pass, ldata, opts); if (ret) { fprintf(stderr, "vecsum_normal_loop pass %d failed " "with error %d\n", pass, ret); return ret; } hdfsSeek(ldata->fs, ldata->file, 0); } return 0; }
int libhdfsconnector::streamFlatFileOffset(const char * filename, unsigned long seekPos, unsigned long readlen,unsigned long bufferSize, int maxretries) { hdfsFile readFile = hdfsOpenFile(fs, filename, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", filename); return EXIT_FAILURE; } if (hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", filename); return EXIT_FAILURE; } unsigned char buffer[bufferSize + 1]; unsigned long currentPos = seekPos; fprintf(stderr, "\n--Start piping: %ld--\n", currentPos); unsigned long bytesLeft = readlen; while (hdfsAvailable(fs, readFile) && bytesLeft > 0) { tSize num_read_bytes = hdfsRead(fs, readFile, buffer, bytesLeft < bufferSize ? bytesLeft : bufferSize); if (num_read_bytes <= 0) break; bytesLeft -= num_read_bytes; for (int i = 0; i < num_read_bytes; i++, currentPos++) fprintf(stdout, "%c", buffer[i]); } fprintf(stderr, "--\nStop Streaming: %ld--\n", currentPos); hdfsCloseFile(fs, readFile); return EXIT_SUCCESS; }
static int vecsum_zcr(struct libhdfs_data *ldata, const struct options *opts) { int ret, pass; struct hadoopRzOptions *zopts = NULL; zopts = hadoopRzOptionsAlloc(); if (!zopts) { fprintf(stderr, "hadoopRzOptionsAlloc failed.\n"); ret = ENOMEM; goto done; } if (hadoopRzOptionsSetSkipChecksum(zopts, 1)) { ret = errno; perror("hadoopRzOptionsSetSkipChecksum failed: "); goto done; } if (hadoopRzOptionsSetByteBufferPool(zopts, NULL)) { ret = errno; perror("hadoopRzOptionsSetByteBufferPool failed: "); goto done; } for (pass = 0; pass < opts->passes; ++pass) { ret = vecsum_zcr_loop(pass, ldata, zopts, opts); if (ret) { fprintf(stderr, "vecsum_zcr_loop pass %d failed " "with error %d\n", pass, ret); goto done; } hdfsSeek(ldata->fs, ldata->file, 0); } ret = 0; done: if (zopts) hadoopRzOptionsFree(zopts); return ret; }
// Caller takes ownership of returned object and must delete it when done google::protobuf::io::CodedInputStream* MaprInputCodedBlockFile::CreateCodedStream(uint64 position, uint64 length) { CHECK(is_open_); // Seek to requested position (relative to start of file). CHECK_LT(position, size_); CHECK_LE(position+length, size_); // Starting with MapR V2.1.1, sometimes seek fails and the hdfs connection // must be reset in order to try again. Not sure why this transient error // happens with MapR V2.1.1 but not earlier versions. bool success = false; for (int i=0; i < 10; ++i){ if (hdfsSeek(fs_, file_, position) == 0){ success = true; break; } //LOG(INFO) << "seek attempt failed: " << i; //LOG(INFO) << "path:" << path_ << "\n position: " << position << "\n length: " << length << "\n size: " << size_; // success if returns 0 CHECK_EQ(hdfsCloseFile(fs_, file_), 0); CHECK_EQ(hdfsDisconnect(fs_), 0); std::string host = "default"; fs_ = hdfsConnect(host.c_str(), 0); // use default config file settings CHECK(fs_) << "error connecting to maprfs"; file_ = hdfsOpenFile(fs_, path_.c_str(), O_RDONLY, 0, 0, 0); CHECK(file_ != NULL); sleep(2*i); } CHECK(success); // Create a coded stream (hold it in a scoped ptr to manage deleting). limiting_stream_.reset(NULL); // the destructor references the copying_stream_, so must destroy it before destroying it copying_stream_.reset(new google::protobuf::io::CopyingInputStreamAdaptor(copying_input_stream_.get())); limiting_stream_.reset(new google::protobuf::io::LimitingInputStream(copying_stream_.get(), length)); return new google::protobuf::io::CodedInputStream(limiting_stream_.get()); }
int main(int argc, char **argv) { hdfsFS fs = hdfsConnect("default", 0); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs!\n"); exit(-1); } hdfsFS lfs = hdfsConnect(NULL, 0); if(!lfs) { fprintf(stderr, "Oops! Failed to connect to 'local' hdfs!\n"); exit(-1); } const char* writePath = "/tmp/testfile.txt"; { //Write tests hdfsFile writeFile = hdfsOpenFile(fs, writePath, O_WRONLY|O_CREAT, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", writePath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", writePath); char* buffer = "Hello, World!"; tSize num_written_bytes = hdfsWrite(fs, writeFile, (void*)buffer, strlen(buffer)+1); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); tOffset currentPos = -1; if ((currentPos = hdfsTell(fs, writeFile)) == -1) { fprintf(stderr, "Failed to get current file position correctly! Got %ld!\n", currentPos); exit(-1); } fprintf(stderr, "Current position: %ld\n", currentPos); if (hdfsFlush(fs, writeFile)) { fprintf(stderr, "Failed to 'flush' %s\n", writePath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", writePath); hdfsCloseFile(fs, writeFile); } { //Read tests const char* readPath = "/tmp/testfile.txt"; int exists = hdfsExists(fs, readPath); if (exists) { fprintf(stderr, "Failed to validate existence of %s\n", readPath); exit(-1); } hdfsFile readFile = hdfsOpenFile(fs, readPath, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", readPath); exit(-1); } fprintf(stderr, "hdfsAvailable: %d\n", hdfsAvailable(fs, readFile)); tOffset seekPos = 1; if(hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", readPath); exit(-1); } tOffset currentPos = -1; if((currentPos = hdfsTell(fs, readFile)) != seekPos) { fprintf(stderr, "Failed to get current file position correctly! Got %ld!\n", currentPos); exit(-1); } fprintf(stderr, "Current position: %ld\n", currentPos); static char buffer[32]; tSize num_read_bytes = hdfsRead(fs, readFile, (void*)buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); num_read_bytes = hdfsPread(fs, readFile, 0, (void*)buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); hdfsCloseFile(fs, readFile); } int totalResult = 0; int result = 0; { //Generic file-system operations const char* srcPath = "/tmp/testfile.txt"; const char* dstPath = "/tmp/testfile2.txt"; fprintf(stderr, "hdfsCopy(remote-local): %s\n", ((result = hdfsCopy(fs, srcPath, lfs, srcPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsCopy(remote-remote): %s\n", ((result = hdfsCopy(fs, srcPath, fs, dstPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsMove(local-local): %s\n", ((result = hdfsMove(lfs, srcPath, lfs, dstPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsMove(remote-local): %s\n", ((result = hdfsMove(fs, srcPath, lfs, srcPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsRename: %s\n", ((result = hdfsRename(fs, dstPath, srcPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsCopy(remote-remote): %s\n", ((result = hdfsCopy(fs, srcPath, fs, dstPath)) ? "Failed!" : "Success!")); totalResult += result; const char* slashTmp = "/tmp"; const char* newDirectory = "/tmp/newdir"; fprintf(stderr, "hdfsCreateDirectory: %s\n", ((result = hdfsCreateDirectory(fs, newDirectory)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsSetReplication: %s\n", ((result = hdfsSetReplication(fs, srcPath, 2)) ? "Failed!" : "Success!")); totalResult += result; char buffer[256]; const char *resp; fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer, sizeof(buffer))) ? buffer : "Failed!")); totalResult += (resp ? 0 : 1); fprintf(stderr, "hdfsSetWorkingDirectory: %s\n", ((result = hdfsSetWorkingDirectory(fs, slashTmp)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer, sizeof(buffer))) ? buffer : "Failed!")); totalResult += (resp ? 0 : 1); fprintf(stderr, "hdfsGetDefaultBlockSize: %ld\n", hdfsGetDefaultBlockSize(fs)); fprintf(stderr, "hdfsGetCapacity: %ld\n", hdfsGetCapacity(fs)); fprintf(stderr, "hdfsGetUsed: %ld\n", hdfsGetUsed(fs)); hdfsFileInfo *fileInfo = NULL; if((fileInfo = hdfsGetPathInfo(fs, slashTmp)) != NULL) { fprintf(stderr, "hdfsGetPathInfo - SUCCESS!\n"); fprintf(stderr, "Name: %s, ", fileInfo->mName); fprintf(stderr, "Type: %c, ", (char)(fileInfo->mKind)); fprintf(stderr, "Replication: %d, ", fileInfo->mReplication); fprintf(stderr, "BlockSize: %ld, ", fileInfo->mBlockSize); fprintf(stderr, "Size: %ld, ", fileInfo->mSize); fprintf(stderr, "LastMod: %s", ctime(&fileInfo->mLastMod)); fprintf(stderr, "Owner: %s, ", fileInfo->mOwner); fprintf(stderr, "Group: %s, ", fileInfo->mGroup); char permissions[10]; permission_disp(fileInfo->mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileInfo->mPermissions, permissions); hdfsFreeFileInfo(fileInfo, 1); } else { totalResult++; fprintf(stderr, "waah! hdfsGetPathInfo for %s - FAILED!\n", slashTmp); } hdfsFileInfo *fileList = 0; int numEntries = 0; if((fileList = hdfsListDirectory(fs, slashTmp, &numEntries)) != NULL) { int i = 0; for(i=0; i < numEntries; ++i) { fprintf(stderr, "Name: %s, ", fileList[i].mName); fprintf(stderr, "Type: %c, ", (char)fileList[i].mKind); fprintf(stderr, "Replication: %d, ", fileList[i].mReplication); fprintf(stderr, "BlockSize: %ld, ", fileList[i].mBlockSize); fprintf(stderr, "Size: %ld, ", fileList[i].mSize); fprintf(stderr, "LastMod: %s", ctime(&fileList[i].mLastMod)); fprintf(stderr, "Owner: %s, ", fileList[i].mOwner); fprintf(stderr, "Group: %s, ", fileList[i].mGroup); char permissions[10]; permission_disp(fileList[i].mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileList[i].mPermissions, permissions); } hdfsFreeFileInfo(fileList, numEntries); } else { if (errno) { totalResult++; fprintf(stderr, "waah! hdfsListDirectory - FAILED!\n"); } else { fprintf(stderr, "Empty directory!\n"); } } char*** hosts = hdfsGetHosts(fs, srcPath, 0, 1); if(hosts) { fprintf(stderr, "hdfsGetHosts - SUCCESS! ... \n"); int i=0; while(hosts[i]) { int j = 0; while(hosts[i][j]) { fprintf(stderr, "\thosts[%d][%d] - %s\n", i, j, hosts[i][j]); ++j; } ++i; } } else { totalResult++; fprintf(stderr, "waah! hdfsGetHosts - FAILED!\n"); } char *newOwner = "root"; // setting tmp dir to 777 so later when connectAsUser nobody, we can write to it short newPerm = 0666; // chown write fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, writePath, NULL, "users")) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, writePath, newOwner, NULL)) ? "Failed!" : "Success!")); totalResult += result; // chmod write fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, writePath, newPerm)) ? "Failed!" : "Success!")); totalResult += result; sleep(2); tTime newMtime = time(NULL); tTime newAtime = time(NULL); // utime write fprintf(stderr, "hdfsUtime: %s\n", ((result = hdfsUtime(fs, writePath, newMtime, newAtime)) ? "Failed!" : "Success!")); totalResult += result; // chown/chmod/utime read hdfsFileInfo *finfo = hdfsGetPathInfo(fs, writePath); fprintf(stderr, "hdfsChown read: %s\n", ((result = (strcmp(finfo->mOwner, newOwner) != 0)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChmod read: %s\n", ((result = (finfo->mPermissions != newPerm)) ? "Failed!" : "Success!")); totalResult += result; // will later use /tmp/ as a different user so enable it fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, "/tmp/", 0777)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr,"newMTime=%ld\n",newMtime); fprintf(stderr,"curMTime=%ld\n",finfo->mLastMod); fprintf(stderr, "hdfsUtime read (mtime): %s\n", ((result = (finfo->mLastMod != newMtime)) ? "Failed!" : "Success!")); totalResult += result; // No easy way to turn on access times from hdfs_test right now // fprintf(stderr, "hdfsUtime read (atime): %s\n", ((result = (finfo->mLastAccess != newAtime)) ? "Failed!" : "Success!")); // totalResult += result; hdfsFreeFileInfo(finfo, 1); // Clean up fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, newDirectory)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, srcPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(lfs, srcPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(lfs, dstPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsExists: %s\n", ((result = hdfsExists(fs, newDirectory)) ? "Success!" : "Failed!")); totalResult += (result ? 0 : 1); } totalResult += (hdfsDisconnect(fs) != 0); { // // Now test as connecting as a specific user // This is only meant to test that we connected as that user, not to test // the actual fs user capabilities. Thus just create a file and read // the owner is correct. const char *tuser = "******"; const char* writePath = "/tmp/usertestfile.txt"; const char **groups = (const char**)malloc(sizeof(char*)* 2); groups[0] = "users"; groups[1] = "nobody"; fs = hdfsConnectAsUser("default", 0, tuser, groups, 2); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs as user %s!\n",tuser); exit(-1); } hdfsFile writeFile = hdfsOpenFile(fs, writePath, O_WRONLY|O_CREAT, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", writePath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", writePath); char* buffer = "Hello, World!"; tSize num_written_bytes = hdfsWrite(fs, writeFile, (void*)buffer, strlen(buffer)+1); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); if (hdfsFlush(fs, writeFile)) { fprintf(stderr, "Failed to 'flush' %s\n", writePath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", writePath); hdfsCloseFile(fs, writeFile); hdfsFileInfo *finfo = hdfsGetPathInfo(fs, writePath); fprintf(stderr, "hdfs new file user is correct: %s\n", ((result = (strcmp(finfo->mOwner, tuser) != 0)) ? "Failed!" : "Success!")); totalResult += result; } totalResult += (hdfsDisconnect(fs) != 0); if (totalResult != 0) { return -1; } else { return 0; } }
static int doTestZeroCopyReads(hdfsFS fs, const char *fileName) { hdfsFile file = NULL; struct hadoopRzOptions *opts = NULL; struct hadoopRzBuffer *buffer = NULL; uint8_t *block; file = hdfsOpenFile(fs, fileName, O_RDONLY, 0, 0, 0); EXPECT_NONNULL(file); opts = hadoopRzOptionsAlloc(); EXPECT_NONNULL(opts); EXPECT_ZERO(hadoopRzOptionsSetSkipChecksum(opts, 1)); /* haven't read anything yet */ EXPECT_ZERO(expectFileStats(file, 0LL, 0LL, 0LL, 0LL)); block = getZeroCopyBlockData(0); EXPECT_NONNULL(block); /* first read is half of a block. */ buffer = hadoopReadZero(file, opts, TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2); EXPECT_NONNULL(buffer); EXPECT_INT_EQ(TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2, hadoopRzBufferLength(buffer)); EXPECT_ZERO(memcmp(hadoopRzBufferGet(buffer), block, TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2)); hadoopRzBufferFree(file, buffer); /* read the next half of the block */ buffer = hadoopReadZero(file, opts, TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2); EXPECT_NONNULL(buffer); EXPECT_INT_EQ(TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2, hadoopRzBufferLength(buffer)); EXPECT_ZERO(memcmp(hadoopRzBufferGet(buffer), block + (TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2), TEST_ZEROCOPY_FULL_BLOCK_SIZE / 2)); hadoopRzBufferFree(file, buffer); free(block); EXPECT_ZERO(expectFileStats(file, TEST_ZEROCOPY_FULL_BLOCK_SIZE, TEST_ZEROCOPY_FULL_BLOCK_SIZE, TEST_ZEROCOPY_FULL_BLOCK_SIZE, TEST_ZEROCOPY_FULL_BLOCK_SIZE)); /* Now let's read just a few bytes. */ buffer = hadoopReadZero(file, opts, SMALL_READ_LEN); EXPECT_NONNULL(buffer); EXPECT_INT_EQ(SMALL_READ_LEN, hadoopRzBufferLength(buffer)); block = getZeroCopyBlockData(1); EXPECT_NONNULL(block); EXPECT_ZERO(memcmp(block, hadoopRzBufferGet(buffer), SMALL_READ_LEN)); hadoopRzBufferFree(file, buffer); EXPECT_INT64_EQ( (int64_t)TEST_ZEROCOPY_FULL_BLOCK_SIZE + (int64_t)SMALL_READ_LEN, hdfsTell(fs, file)); EXPECT_ZERO(expectFileStats(file, TEST_ZEROCOPY_FULL_BLOCK_SIZE + SMALL_READ_LEN, TEST_ZEROCOPY_FULL_BLOCK_SIZE + SMALL_READ_LEN, TEST_ZEROCOPY_FULL_BLOCK_SIZE + SMALL_READ_LEN, TEST_ZEROCOPY_FULL_BLOCK_SIZE + SMALL_READ_LEN)); /* Clear 'skip checksums' and test that we can't do zero-copy reads any * more. Since there is no ByteBufferPool set, we should fail with * EPROTONOSUPPORT. */ EXPECT_ZERO(hadoopRzOptionsSetSkipChecksum(opts, 0)); EXPECT_NULL(hadoopReadZero(file, opts, TEST_ZEROCOPY_FULL_BLOCK_SIZE)); EXPECT_INT_EQ(EPROTONOSUPPORT, errno); /* Verify that setting a NULL ByteBufferPool class works. */ EXPECT_ZERO(hadoopRzOptionsSetByteBufferPool(opts, NULL)); EXPECT_ZERO(hadoopRzOptionsSetSkipChecksum(opts, 0)); EXPECT_NULL(hadoopReadZero(file, opts, TEST_ZEROCOPY_FULL_BLOCK_SIZE)); EXPECT_INT_EQ(EPROTONOSUPPORT, errno); /* Now set a ByteBufferPool and try again. It should succeed this time. */ EXPECT_ZERO(hadoopRzOptionsSetByteBufferPool(opts, ELASTIC_BYTE_BUFFER_POOL_CLASS)); buffer = hadoopReadZero(file, opts, TEST_ZEROCOPY_FULL_BLOCK_SIZE); EXPECT_NONNULL(buffer); EXPECT_INT_EQ(TEST_ZEROCOPY_FULL_BLOCK_SIZE, hadoopRzBufferLength(buffer)); EXPECT_ZERO(expectFileStats(file, (2 * TEST_ZEROCOPY_FULL_BLOCK_SIZE) + SMALL_READ_LEN, (2 * TEST_ZEROCOPY_FULL_BLOCK_SIZE) + SMALL_READ_LEN, (2 * TEST_ZEROCOPY_FULL_BLOCK_SIZE) + SMALL_READ_LEN, TEST_ZEROCOPY_FULL_BLOCK_SIZE + SMALL_READ_LEN)); EXPECT_ZERO(memcmp(block + SMALL_READ_LEN, hadoopRzBufferGet(buffer), TEST_ZEROCOPY_FULL_BLOCK_SIZE - SMALL_READ_LEN)); free(block); block = getZeroCopyBlockData(2); EXPECT_NONNULL(block); EXPECT_ZERO(memcmp(block, (uint8_t*)hadoopRzBufferGet(buffer) + (TEST_ZEROCOPY_FULL_BLOCK_SIZE - SMALL_READ_LEN), SMALL_READ_LEN)); hadoopRzBufferFree(file, buffer); /* Check the result of a zero-length read. */ buffer = hadoopReadZero(file, opts, 0); EXPECT_NONNULL(buffer); EXPECT_NONNULL(hadoopRzBufferGet(buffer)); EXPECT_INT_EQ(0, hadoopRzBufferLength(buffer)); hadoopRzBufferFree(file, buffer); /* Check the result of reading past EOF */ EXPECT_INT_EQ(0, hdfsSeek(fs, file, TEST_ZEROCOPY_FILE_LEN)); buffer = hadoopReadZero(file, opts, 1); EXPECT_NONNULL(buffer); EXPECT_NULL(hadoopRzBufferGet(buffer)); hadoopRzBufferFree(file, buffer); /* Cleanup */ free(block); hadoopRzOptionsFree(opts); EXPECT_ZERO(hdfsCloseFile(fs, file)); return 0; }
int libhdfsconnector::readXMLOffset(const char * filename, unsigned long seekPos, unsigned long readlen, const char * rowTag, const char * headerText, const char * footerText, unsigned long bufferSize) { string xmlizedxpath; string elementname; string rootelement; xpath2xml(&xmlizedxpath, rowTag, true); getLastXPathElement(&elementname, rowTag); hdfsFile readFile = hdfsOpenFile(fs, filename, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", filename); return EXIT_FAILURE; } if (hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", filename); return EXIT_FAILURE; } unsigned char buffer[bufferSize + 1]; bool firstRowfound = false; string openRowTag("<"); openRowTag.append(elementname).append(1, '>'); string closeRowTag("</"); closeRowTag.append(elementname).append(1, '>'); string closeRootTag("</"); getLastXMLElement(&closeRootTag, footerText); closeRootTag.append(1, '>'); unsigned long currentPos = seekPos + openRowTag.size(); string currentTag(""); bool withinRecord = false; bool stopAtNextClosingTag = false; bool parsingTag = false; fprintf(stderr, "--Start looking <%s>: %ld--\n", elementname.c_str(), currentPos); fprintf(stdout, "%s", xmlizedxpath.c_str()); unsigned long bytesLeft = readlen; while (hdfsAvailable(fs, readFile) && bytesLeft > 0) { tSize numOfBytesRead = hdfsRead(fs, readFile, (void*) buffer, bufferSize); if (numOfBytesRead <= 0) { fprintf(stderr, "\n--Hard Stop at: %ld--\n", currentPos); break; } for (int buffIndex = 0; buffIndex < numOfBytesRead;) { char currChar = buffer[buffIndex]; if (currChar == '<' || parsingTag) { if (!parsingTag) currentTag.clear(); int tagpos = buffIndex; while (tagpos < numOfBytesRead) { currentTag.append(1, buffer[tagpos++]); if (buffer[tagpos - 1] == '>') break; } if (tagpos == numOfBytesRead && buffer[tagpos - 1] != '>') { fprintf(stderr, "\nTag accross buffer reads...\n"); currentPos += tagpos - buffIndex; bytesLeft -= tagpos - buffIndex; buffIndex = tagpos; parsingTag = true; if (bytesLeft <= 0) { bytesLeft = readlen; //not sure how much longer til next EOL read up readlen; stopAtNextClosingTag = true; } break; } else parsingTag = false; if (!firstRowfound) { firstRowfound = strcmp(currentTag.c_str(), openRowTag.c_str()) == 0; if (firstRowfound) fprintf(stderr, "--start piping tag %s at %lu--\n", currentTag.c_str(), currentPos); } if (strcmp(currentTag.c_str(), closeRootTag.c_str()) == 0) { bytesLeft = 0; break; } if (strcmp(currentTag.c_str(), openRowTag.c_str()) == 0) withinRecord = true; else if (strcmp(currentTag.c_str(), closeRowTag.c_str()) == 0) withinRecord = false; else if (firstRowfound && !withinRecord) { bytesLeft = 0; fprintf(stderr, "Unexpected Tag found: %s at position %lu\n", currentTag.c_str(), currentPos); break; } currentPos += tagpos - buffIndex; bytesLeft -= tagpos - buffIndex; buffIndex = tagpos; if (bytesLeft <= 0 && !withinRecord) stopAtNextClosingTag = true; if (stopAtNextClosingTag && strcmp(currentTag.c_str(), closeRowTag.c_str()) == 0) { fprintf(stdout, "%s", currentTag.c_str()); fprintf(stderr, "--stop piping at %s %lu--\n", currentTag.c_str(), currentPos); bytesLeft = 0; break; } if (firstRowfound) fprintf(stdout, "%s", currentTag.c_str()); else fprintf(stderr, "skipping tag %s\n", currentTag.c_str()); if (buffIndex < numOfBytesRead) currChar = buffer[buffIndex]; else break; } if (firstRowfound) fprintf(stdout, "%c", currChar); buffIndex++; currentPos++; bytesLeft--; if (bytesLeft <= 0) { if (withinRecord) { fprintf(stderr, "\n--Looking for last closing row tag: %ld--\n", currentPos); bytesLeft = readlen; //not sure how much longer til next EOL read up readlen; stopAtNextClosingTag = true; } else break; } } } xmlizedxpath.clear(); xpath2xml(&xmlizedxpath, rowTag, false); fprintf(stdout, "%s", xmlizedxpath.c_str()); return EXIT_SUCCESS; }
int main(int argc, char **argv) { const char *writePath = "/tmp/testfile.txt"; const char *fileContents = "Hello, World!"; const char *readPath = "/tmp/testfile.txt"; const char *srcPath = "/tmp/testfile.txt"; const char *dstPath = "/tmp/testfile2.txt"; const char *slashTmp = "/tmp"; const char *newDirectory = "/tmp/newdir"; const char *newOwner = "root"; const char *tuser = "******"; const char *appendPath = "/tmp/appends"; const char *userPath = "/tmp/usertestfile.txt"; char buffer[32], buffer2[256], rdbuffer[32]; tSize num_written_bytes, num_read_bytes; hdfsFS fs, lfs; hdfsFile writeFile, readFile, localFile, appendFile, userFile; tOffset currentPos, seekPos; int exists, totalResult, result, numEntries, i, j; const char *resp; hdfsFileInfo *fileInfo, *fileList, *finfo; char *buffer3; char permissions[10]; char ***hosts; short newPerm = 0666; tTime newMtime, newAtime; fs = hdfsConnectNewInstance("default", 0); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs!\n"); exit(-1); } lfs = hdfsConnectNewInstance(NULL, 0); if(!lfs) { fprintf(stderr, "Oops! Failed to connect to 'local' hdfs!\n"); exit(-1); } { //Write tests writeFile = hdfsOpenFile(fs, writePath, O_WRONLY|O_CREAT, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", writePath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", writePath); num_written_bytes = hdfsWrite(fs, writeFile, (void*)fileContents, (tSize)(strlen(fileContents)+1)); if (num_written_bytes != strlen(fileContents) + 1) { fprintf(stderr, "Failed to write correct number of bytes - expected %d, got %d\n", (int)(strlen(fileContents) + 1), (int)num_written_bytes); exit(-1); } fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); currentPos = -1; if ((currentPos = hdfsTell(fs, writeFile)) == -1) { fprintf(stderr, "Failed to get current file position correctly! Got %" PRId64 "!\n", currentPos); exit(-1); } fprintf(stderr, "Current position: %" PRId64 "\n", currentPos); if (hdfsFlush(fs, writeFile)) { fprintf(stderr, "Failed to 'flush' %s\n", writePath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", writePath); if (hdfsHFlush(fs, writeFile)) { fprintf(stderr, "Failed to 'hflush' %s\n", writePath); exit(-1); } fprintf(stderr, "HFlushed %s successfully!\n", writePath); hdfsCloseFile(fs, writeFile); } { //Read tests exists = hdfsExists(fs, readPath); if (exists) { fprintf(stderr, "Failed to validate existence of %s\n", readPath); exit(-1); } readFile = hdfsOpenFile(fs, readPath, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", readPath); exit(-1); } if (!hdfsFileIsOpenForRead(readFile)) { fprintf(stderr, "hdfsFileIsOpenForRead: we just opened a file " "with O_RDONLY, and it did not show up as 'open for " "read'\n"); exit(-1); } fprintf(stderr, "hdfsAvailable: %d\n", hdfsAvailable(fs, readFile)); seekPos = 1; if(hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", readPath); exit(-1); } currentPos = -1; if((currentPos = hdfsTell(fs, readFile)) != seekPos) { fprintf(stderr, "Failed to get current file position correctly! Got %" PRId64 "!\n", currentPos); exit(-1); } fprintf(stderr, "Current position: %" PRId64 "\n", currentPos); if (!hdfsFileUsesDirectRead(readFile)) { fprintf(stderr, "Direct read support incorrectly not detected " "for HDFS filesystem\n"); exit(-1); } fprintf(stderr, "Direct read support detected for HDFS\n"); // Test the direct read path if(hdfsSeek(fs, readFile, 0)) { fprintf(stderr, "Failed to seek %s for reading!\n", readPath); exit(-1); } memset(buffer, 0, sizeof(buffer)); num_read_bytes = hdfsRead(fs, readFile, (void*)buffer, sizeof(buffer)); if (strncmp(fileContents, buffer, strlen(fileContents)) != 0) { fprintf(stderr, "Failed to read (direct). Expected %s but got %s (%d bytes)\n", fileContents, buffer, num_read_bytes); exit(-1); } fprintf(stderr, "Read (direct) following %d bytes:\n%s\n", num_read_bytes, buffer); if (hdfsSeek(fs, readFile, 0L)) { fprintf(stderr, "Failed to seek to file start!\n"); exit(-1); } // Disable the direct read path so that we really go through the slow // read path hdfsFileDisableDirectRead(readFile); num_read_bytes = hdfsRead(fs, readFile, (void*)buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); memset(buffer, 0, strlen(fileContents + 1)); num_read_bytes = hdfsPread(fs, readFile, 0, (void*)buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); hdfsCloseFile(fs, readFile); // Test correct behaviour for unsupported filesystems localFile = hdfsOpenFile(lfs, writePath, O_WRONLY|O_CREAT, 0, 0, 0); if(!localFile) { fprintf(stderr, "Failed to open %s for writing!\n", writePath); exit(-1); } num_written_bytes = hdfsWrite(lfs, localFile, (void*)fileContents, (tSize)(strlen(fileContents) + 1)); hdfsCloseFile(lfs, localFile); localFile = hdfsOpenFile(lfs, writePath, O_RDONLY, 0, 0, 0); if (hdfsFileUsesDirectRead(localFile)) { fprintf(stderr, "Direct read support incorrectly detected for local " "filesystem\n"); exit(-1); } hdfsCloseFile(lfs, localFile); } totalResult = 0; result = 0; { //Generic file-system operations fprintf(stderr, "hdfsCopy(remote-local): %s\n", ((result = hdfsCopy(fs, srcPath, lfs, srcPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsCopy(remote-remote): %s\n", ((result = hdfsCopy(fs, srcPath, fs, dstPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsMove(local-local): %s\n", ((result = hdfsMove(lfs, srcPath, lfs, dstPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsMove(remote-local): %s\n", ((result = hdfsMove(fs, srcPath, lfs, srcPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsRename: %s\n", ((result = hdfsRename(fs, dstPath, srcPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsCopy(remote-remote): %s\n", ((result = hdfsCopy(fs, srcPath, fs, dstPath)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsCreateDirectory: %s\n", ((result = hdfsCreateDirectory(fs, newDirectory)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsSetReplication: %s\n", ((result = hdfsSetReplication(fs, srcPath, 2)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer2, sizeof(buffer2))) != 0 ? buffer2 : "Failed!")); totalResult += (resp ? 0 : 1); fprintf(stderr, "hdfsSetWorkingDirectory: %s\n", ((result = hdfsSetWorkingDirectory(fs, slashTmp)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer2, sizeof(buffer2))) != 0 ? buffer2 : "Failed!")); totalResult += (resp ? 0 : 1); fprintf(stderr, "hdfsGetDefaultBlockSize: %" PRId64 "\n", hdfsGetDefaultBlockSize(fs)); fprintf(stderr, "hdfsGetCapacity: %" PRId64 "\n", hdfsGetCapacity(fs)); fprintf(stderr, "hdfsGetUsed: %" PRId64 "\n", hdfsGetUsed(fs)); fileInfo = NULL; if((fileInfo = hdfsGetPathInfo(fs, slashTmp)) != NULL) { fprintf(stderr, "hdfsGetPathInfo - SUCCESS!\n"); fprintf(stderr, "Name: %s, ", fileInfo->mName); fprintf(stderr, "Type: %c, ", (char)(fileInfo->mKind)); fprintf(stderr, "Replication: %d, ", fileInfo->mReplication); fprintf(stderr, "BlockSize: %" PRId64 ", ", fileInfo->mBlockSize); fprintf(stderr, "Size: %" PRId64 ", ", fileInfo->mSize); fprintf(stderr, "LastMod: %s", ctime(&fileInfo->mLastMod)); fprintf(stderr, "Owner: %s, ", fileInfo->mOwner); fprintf(stderr, "Group: %s, ", fileInfo->mGroup); permission_disp(fileInfo->mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileInfo->mPermissions, permissions); hdfsFreeFileInfo(fileInfo, 1); } else { totalResult++; fprintf(stderr, "waah! hdfsGetPathInfo for %s - FAILED!\n", slashTmp); } fileList = 0; fileList = hdfsListDirectory(fs, newDirectory, &numEntries); if (!(fileList == NULL && numEntries == 0 && !errno)) { fprintf(stderr, "waah! hdfsListDirectory for empty %s - FAILED!\n", newDirectory); totalResult++; } else { fprintf(stderr, "hdfsListDirectory for empty %s - SUCCESS!\n", newDirectory); } fileList = 0; if((fileList = hdfsListDirectory(fs, slashTmp, &numEntries)) != NULL) { for(i=0; i < numEntries; ++i) { fprintf(stderr, "Name: %s, ", fileList[i].mName); fprintf(stderr, "Type: %c, ", (char)fileList[i].mKind); fprintf(stderr, "Replication: %d, ", fileList[i].mReplication); fprintf(stderr, "BlockSize: %" PRId64 ", ", fileList[i].mBlockSize); fprintf(stderr, "Size: %" PRId64 ", ", fileList[i].mSize); fprintf(stderr, "LastMod: %s", ctime(&fileList[i].mLastMod)); fprintf(stderr, "Owner: %s, ", fileList[i].mOwner); fprintf(stderr, "Group: %s, ", fileList[i].mGroup); permission_disp(fileList[i].mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileList[i].mPermissions, permissions); } hdfsFreeFileInfo(fileList, numEntries); } else { if (errno) { totalResult++; fprintf(stderr, "waah! hdfsListDirectory - FAILED!\n"); } else { fprintf(stderr, "Empty directory!\n"); } } hosts = hdfsGetHosts(fs, srcPath, 0, 1); if(hosts) { fprintf(stderr, "hdfsGetHosts - SUCCESS! ... \n"); i=0; while(hosts[i]) { j = 0; while(hosts[i][j]) { fprintf(stderr, "\thosts[%d][%d] - %s\n", i, j, hosts[i][j]); ++j; } ++i; } } else { totalResult++; fprintf(stderr, "waah! hdfsGetHosts - FAILED!\n"); } // setting tmp dir to 777 so later when connectAsUser nobody, we can write to it // chown write fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, writePath, NULL, "users")) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, writePath, newOwner, NULL)) != 0 ? "Failed!" : "Success!")); totalResult += result; // chmod write fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, writePath, newPerm)) != 0 ? "Failed!" : "Success!")); totalResult += result; sleep(2); newMtime = time(NULL); newAtime = time(NULL); // utime write fprintf(stderr, "hdfsUtime: %s\n", ((result = hdfsUtime(fs, writePath, newMtime, newAtime)) != 0 ? "Failed!" : "Success!")); totalResult += result; // chown/chmod/utime read finfo = hdfsGetPathInfo(fs, writePath); fprintf(stderr, "hdfsChown read: %s\n", ((result = (strcmp(finfo->mOwner, newOwner))) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChmod read: %s\n", ((result = (finfo->mPermissions != newPerm)) != 0 ? "Failed!" : "Success!")); totalResult += result; // will later use /tmp/ as a different user so enable it fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, "/tmp/", 0777)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr,"newMTime=%ld\n",newMtime); fprintf(stderr,"curMTime=%ld\n",finfo->mLastMod); fprintf(stderr, "hdfsUtime read (mtime): %s\n", ((result = (finfo->mLastMod != newMtime)) != 0 ? "Failed!" : "Success!")); totalResult += result; // No easy way to turn on access times from hdfs_test right now // fprintf(stderr, "hdfsUtime read (atime): %s\n", ((result = (finfo->mLastAccess != newAtime)) != 0 ? "Failed!" : "Success!")); // totalResult += result; hdfsFreeFileInfo(finfo, 1); // Clean up fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, newDirectory, 1)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, srcPath, 1)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(lfs, srcPath, 1)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(lfs, dstPath, 1)) != 0 ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsExists: %s\n", ((result = hdfsExists(fs, newDirectory)) != 0 ? "Success!" : "Failed!")); totalResult += (result ? 0 : 1); } { // TEST APPENDS // CREATE appendFile = hdfsOpenFile(fs, appendPath, O_WRONLY, 0, 0, 0); if(!appendFile) { fprintf(stderr, "Failed to open %s for writing!\n", appendPath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", appendPath); buffer3 = "Hello,"; num_written_bytes = hdfsWrite(fs, appendFile, (void*)buffer3, (tSize)strlen(buffer3)); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); if (hdfsFlush(fs, appendFile)) { fprintf(stderr, "Failed to 'flush' %s\n", appendPath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", appendPath); hdfsCloseFile(fs, appendFile); // RE-OPEN appendFile = hdfsOpenFile(fs, appendPath, O_WRONLY|O_APPEND, 0, 0, 0); if(!appendFile) { fprintf(stderr, "Failed to open %s for writing!\n", appendPath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", appendPath); buffer3 = " World"; num_written_bytes = hdfsWrite(fs, appendFile, (void*)buffer3, (tSize)(strlen(buffer3) + 1)); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); if (hdfsFlush(fs, appendFile)) { fprintf(stderr, "Failed to 'flush' %s\n", appendPath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", appendPath); hdfsCloseFile(fs, appendFile); // CHECK size finfo = hdfsGetPathInfo(fs, appendPath); fprintf(stderr, "fileinfo->mSize: == total %s\n", ((result = (finfo->mSize == (tOffset)(strlen("Hello, World") + 1))) == 1 ? "Success!" : "Failed!")); totalResult += (result ? 0 : 1); // READ and check data readFile = hdfsOpenFile(fs, appendPath, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", appendPath); exit(-1); } num_read_bytes = hdfsRead(fs, readFile, (void*)rdbuffer, sizeof(rdbuffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, rdbuffer); fprintf(stderr, "read == Hello, World %s\n", ((result = (strcmp(rdbuffer, "Hello, World"))) == 0 ? "Success!" : "Failed!")); hdfsCloseFile(fs, readFile); // DONE test appends } totalResult += (hdfsDisconnect(fs) != 0); { // // Now test as connecting as a specific user // This is only meant to test that we connected as that user, not to test // the actual fs user capabilities. Thus just create a file and read // the owner is correct. fs = hdfsConnectAsUserNewInstance("default", 0, tuser); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs as user %s!\n",tuser); exit(-1); } userFile = hdfsOpenFile(fs, userPath, O_WRONLY|O_CREAT, 0, 0, 0); if(!userFile) { fprintf(stderr, "Failed to open %s for writing!\n", userPath); exit(-1); } fprintf(stderr, "Opened %s for writing successfully...\n", userPath); num_written_bytes = hdfsWrite(fs, userFile, (void*)fileContents, (tSize)(strlen(fileContents)+1)); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); if (hdfsFlush(fs, userFile)) { fprintf(stderr, "Failed to 'flush' %s\n", userPath); exit(-1); } fprintf(stderr, "Flushed %s successfully!\n", userPath); hdfsCloseFile(fs, userFile); finfo = hdfsGetPathInfo(fs, userPath); fprintf(stderr, "hdfs new file user is correct: %s\n", ((result = (strcmp(finfo->mOwner, tuser))) != 0 ? "Failed!" : "Success!")); totalResult += result; } totalResult += (hdfsDisconnect(fs) != 0); if (totalResult != 0) { return -1; } else { return 0; } }
qioerr hdfs_preadv (void* file, const struct iovec *vector, int count, off_t offset, ssize_t* num_read_out, void* fs) { ssize_t got; ssize_t got_total; qioerr err_out = 0; int i; STARTING_SLOW_SYSCALL; #ifdef HDFS3 const hdfs_file orig_hfl = *to_hdfs_file(file); const hdfs_fs orig_hfs = *to_hdfs_fs(fs); hdfsFS hfs = hdfsConnect(orig_hfs.fs_name, orig_hfs.fs_port); hdfsFile hfl = hdfsOpenFile(hfs, orig_hfl.pathnm, O_RDONLY, 0, 0, 0); //assert connection CREATE_ERROR((hfs == NULL), err_out, ECONNREFUSED, "Unable to read HDFS file", error); if(hfl == NULL) { err_out = qio_mkerror_errno(); goto error; } #endif err_out = 0; got_total = 0; for(i = 0; i < count; i++) { #ifdef HDFS3 hdfsSeek(hfs, hfl, offset+got_total); got = hdfsRead(hfs, hfl, (void*)vector[i].iov_base, vector[i].iov_len); #else got = hdfsPread(to_hdfs_fs(fs)->hfs, to_hdfs_file(file)->file, offset + got_total, (void*)vector[i].iov_base, vector[i].iov_len); #endif if( got != -1 ) { got_total += got; } else { err_out = qio_mkerror_errno(); break; } if(got != (ssize_t)vector[i].iov_len ) { break; } } if( err_out == 0 && got_total == 0 && sys_iov_total_bytes(vector, count) != 0 ) err_out = qio_int_to_err(EEOF); *num_read_out = got_total; #ifdef HDFS3 got = hdfsCloseFile(hfs, hfl); if(got == -1) { err_out = qio_mkerror_errno(); } got = hdfsDisconnect(hfs); if(got == -1) { err_out = qio_mkerror_errno(); } #endif DONE_SLOW_SYSCALL; #ifdef HDFS3 error: #endif return err_out; }
int libhdfsconnector::streamCSVFileOffset(const char * filename, unsigned long seekPos, unsigned long readlen, const char * eolseq, unsigned long bufferSize, bool outputTerminator, unsigned long recLen, unsigned long maxLen, const char * quote, int maxretries) { fprintf(stderr, "CSV terminator: \'%s\' and quote: \'%c\'\n", eolseq, quote[0]); unsigned long recsFound = 0; hdfsFile readFile = hdfsOpenFile(fs, filename, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", filename); return EXIT_FAILURE; } unsigned eolseqlen = strlen(eolseq); if (seekPos > eolseqlen) seekPos -= eolseqlen; //read back sizeof(EOL) in case the seekpos happens to be a the first char after an EOL if (hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", filename); return EXIT_FAILURE; } bool withinQuote = false; unsigned char buffer[bufferSize + 1]; bool stopAtNextEOL = false; bool firstEOLfound = seekPos == 0 ? true : false; unsigned long currentPos = seekPos; fprintf(stderr, "--Start looking: %ld--\n", currentPos); unsigned long bytesLeft = readlen; while (hdfsAvailable(fs, readFile) && bytesLeft > 0) { tSize num_read_bytes = hdfsRead(fs, readFile, (void*) buffer, bufferSize); if (num_read_bytes <= 0) { fprintf(stderr, "\n--Hard Stop at: %ld--\n", currentPos); break; } for (int bufferIndex = 0; bufferIndex < num_read_bytes; bufferIndex++, currentPos++) { char currChar = buffer[bufferIndex]; if (currChar == EOF) break; if (currChar == quote[0]) { fprintf(stderr, "found quote char at pos: %ld\n", currentPos); withinQuote = !withinQuote; } if (currChar == eolseq[0] && !withinQuote) { bool eolfound = true; tSize extraNumOfBytesRead = 0; string tmpstr(""); if (eolseqlen > 1) { int eoli = bufferIndex; while (eoli < num_read_bytes && eoli - bufferIndex < eolseqlen) { tmpstr.append(1, buffer[eoli++]); } if (eoli == num_read_bytes && tmpstr.size() < eolseqlen) { //looks like we have to do a remote read, but before we do, let's make sure the substring matches if (strncmp(eolseq, tmpstr.c_str(), tmpstr.size()) == 0) { unsigned char tmpbuffer[eolseqlen - tmpstr.size() + 1]; //TODO have to make a read... of eolseqlen - tmpstr.size is it worth it? //extraNumOfBytesRead = hdfsRead(*fs, readFile, (void*) tmpbuffer, extraNumOfBytesRead = hdfsRead(fs, readFile, (void*) tmpbuffer, eolseqlen - tmpstr.size()); for (int y = 0; y < extraNumOfBytesRead; y++) tmpstr.append(1, tmpbuffer[y]); } } if (strcmp(tmpstr.c_str(), eolseq) != 0) eolfound = false; } if (eolfound) { if (!firstEOLfound) { bufferIndex = bufferIndex + eolseqlen - 1; currentPos = currentPos + eolseqlen - 1; bytesLeft = bytesLeft - eolseqlen; fprintf(stderr, "\n--Start reading: %ld--\n", currentPos); firstEOLfound = true; continue; } if (outputTerminator) { //if (currentPos > seekPos) //Don't output first EOL fprintf(stdout, "%s", eolseq); bufferIndex += eolseqlen; currentPos += eolseqlen; bytesLeft -= eolseqlen; } recsFound++; if (stopAtNextEOL) { fprintf(stderr, "\n--Stop piping: %ld--\n", currentPos); bytesLeft = 0; break; } if (bufferIndex < num_read_bytes) currChar = buffer[bufferIndex]; else break; } else if (extraNumOfBytesRead > 0) { if (hdfsSeek(fs, readFile, hdfsTell(fs, readFile) - extraNumOfBytesRead)) { fprintf(stderr, "Error while attempting to correct seek position\n"); return EXIT_FAILURE; } } } //don't pipe until we're beyond the first EOL (if offset = 0 start piping ASAP) if (firstEOLfound) { fprintf(stdout, "%c", currChar); bytesLeft--; } else { fprintf(stderr, "%c", currChar); bytesLeft--; if (maxLen > 0 && currentPos - seekPos > maxLen * 10) { fprintf(stderr, "\nFirst EOL was not found within the first %lu bytes", currentPos - seekPos); return EXIT_FAILURE; } } if (stopAtNextEOL) fprintf(stderr, "%c", currChar); // ok, so if bytesLeft <= 0 at this point, we need to keep piping // IF the last char read was not an EOL char if (bytesLeft <= 0 && currChar != eolseq[0]) { if (!firstEOLfound) { fprintf(stderr, "\n--Reached end of readlen before finding first record start at: %ld (breaking out)--\n", currentPos); break; } fprintf(stderr, "\n--Looking for Last EOL: %ld--\n", currentPos); bytesLeft = readlen; //not sure how much longer until next EOL read up readlen; stopAtNextEOL = true; } } } fprintf(stderr, "\nCurrentPos: %ld, RecsFound: %ld\n", currentPos, recsFound); hdfsCloseFile(fs, readFile); return EXIT_SUCCESS; }
int main(int argc, char **argv) { char buffer[32]; tSize num_written_bytes; const char* slashTmp = "/tmp"; int nnPort; char *rwTemplate, *rwTemplate2, *newDirTemplate, *appendTemplate, *userTemplate, *rwPath = NULL; const char* fileContents = "Hello, World!"; const char* nnHost = NULL; if (argc != 2) { fprintf(stderr, "usage: test_libwebhdfs_ops <username>\n"); exit(1); } struct NativeMiniDfsConf conf = { .doFormat = 1, .webhdfsEnabled = 1, .namenodeHttpPort = 50070, }; cluster = nmdCreate(&conf); if (!cluster) { fprintf(stderr, "Failed to create the NativeMiniDfsCluster.\n"); exit(1); } if (nmdWaitClusterUp(cluster)) { fprintf(stderr, "Error when waiting for cluster to be ready.\n"); exit(1); } if (nmdGetNameNodeHttpAddress(cluster, &nnPort, &nnHost)) { fprintf(stderr, "Error when retrieving namenode host address.\n"); exit(1); } hdfsFS fs = hdfsConnectAsUserNewInstance(nnHost, nnPort, argv[1]); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs!\n"); exit(-1); } { // Write tests rwTemplate = strdup("/tmp/helloWorldXXXXXX"); if (!rwTemplate) { fprintf(stderr, "Failed to create rwTemplate!\n"); exit(1); } rwPath = mktemp(rwTemplate); // hdfsOpenFile hdfsFile writeFile = hdfsOpenFile(fs, rwPath, O_WRONLY|O_CREAT, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", rwPath); exit(1); } fprintf(stderr, "Opened %s for writing successfully...\n", rwPath); // hdfsWrite num_written_bytes = hdfsWrite(fs, writeFile, (void*)fileContents, (int) strlen(fileContents) + 1); if (num_written_bytes != strlen(fileContents) + 1) { fprintf(stderr, "Failed to write correct number of bytes - " "expected %d, got %d\n", (int)(strlen(fileContents) + 1), (int) num_written_bytes); exit(1); } fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); // hdfsTell tOffset currentPos = -1; if ((currentPos = hdfsTell(fs, writeFile)) == -1) { fprintf(stderr, "Failed to get current file position correctly. Got %" PRId64 "!\n", currentPos); exit(1); } fprintf(stderr, "Current position: %" PRId64 "\n", currentPos); hdfsCloseFile(fs, writeFile); // Done test write } sleep(1); { //Read tests int available = 0, exists = 0; // hdfsExists exists = hdfsExists(fs, rwPath); if (exists) { fprintf(stderr, "Failed to validate existence of %s\n", rwPath); exists = hdfsExists(fs, rwPath); if (exists) { fprintf(stderr, "Still failed to validate existence of %s\n", rwPath); exit(1); } } hdfsFile readFile = hdfsOpenFile(fs, rwPath, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", rwPath); exit(1); } if (!hdfsFileIsOpenForRead(readFile)) { fprintf(stderr, "hdfsFileIsOpenForRead: we just opened a file " "with O_RDONLY, and it did not show up as 'open for " "read'\n"); exit(1); } available = hdfsAvailable(fs, readFile); fprintf(stderr, "hdfsAvailable: %d\n", available); // hdfsSeek, hdfsTell tOffset seekPos = 1; if(hdfsSeek(fs, readFile, seekPos)) { fprintf(stderr, "Failed to seek %s for reading!\n", rwPath); exit(1); } tOffset currentPos = -1; if((currentPos = hdfsTell(fs, readFile)) != seekPos) { fprintf(stderr, "Failed to get current file position correctly! Got %" PRId64 "!\n", currentPos); exit(1); } fprintf(stderr, "Current position: %" PRId64 "\n", currentPos); if(hdfsSeek(fs, readFile, 0)) { fprintf(stderr, "Failed to seek %s for reading!\n", rwPath); exit(1); } // hdfsRead memset(buffer, 0, sizeof(buffer)); tSize num_read_bytes = hdfsRead(fs, readFile, buffer, sizeof(buffer)); if (strncmp(fileContents, buffer, strlen(fileContents)) != 0) { fprintf(stderr, "Failed to read (direct). " "Expected %s but got %s (%d bytes)\n", fileContents, buffer, num_read_bytes); exit(1); } fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); if (hdfsSeek(fs, readFile, 0L)) { fprintf(stderr, "Failed to seek to file start!\n"); exit(1); } // hdfsPread memset(buffer, 0, strlen(fileContents + 1)); num_read_bytes = hdfsPread(fs, readFile, 0, buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); hdfsCloseFile(fs, readFile); // Done test read } int totalResult = 0; int result = 0; { //Generic file-system operations char *srcPath = rwPath; char buffer[256]; const char *resp; rwTemplate2 = strdup("/tmp/helloWorld2XXXXXX"); if (!rwTemplate2) { fprintf(stderr, "Failed to create rwTemplate2!\n"); exit(1); } char *dstPath = mktemp(rwTemplate2); newDirTemplate = strdup("/tmp/newdirXXXXXX"); if (!newDirTemplate) { fprintf(stderr, "Failed to create newDirTemplate!\n"); exit(1); } char *newDirectory = mktemp(newDirTemplate); // hdfsRename fprintf(stderr, "hdfsRename: %s\n", ((result = hdfsRename(fs, rwPath, dstPath)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsRename back: %s\n", ((result = hdfsRename(fs, dstPath, srcPath)) ? "Failed!" : "Success!")); totalResult += result; // hdfsCreateDirectory fprintf(stderr, "hdfsCreateDirectory: %s\n", ((result = hdfsCreateDirectory(fs, newDirectory)) ? "Failed!" : "Success!")); totalResult += result; // hdfsSetReplication fprintf(stderr, "hdfsSetReplication: %s\n", ((result = hdfsSetReplication(fs, srcPath, 1)) ? "Failed!" : "Success!")); totalResult += result; // hdfsGetWorkingDirectory, hdfsSetWorkingDirectory fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer, sizeof(buffer))) ? buffer : "Failed!")); totalResult += (resp ? 0 : 1); const char* path[] = {"/foo", "/foo/bar", "foobar", "//foo/bar//foobar", "foo//bar", "foo/bar///", "/", "////"}; int i; for (i = 0; i < 8; i++) { fprintf(stderr, "hdfsSetWorkingDirectory: %s, %s\n", ((result = hdfsSetWorkingDirectory(fs, path[i])) ? "Failed!" : "Success!"), hdfsGetWorkingDirectory(fs, buffer, sizeof(buffer))); totalResult += result; } fprintf(stderr, "hdfsSetWorkingDirectory: %s\n", ((result = hdfsSetWorkingDirectory(fs, slashTmp)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsGetWorkingDirectory: %s\n", ((resp = hdfsGetWorkingDirectory(fs, buffer, sizeof(buffer))) ? buffer : "Failed!")); totalResult += (resp ? 0 : 1); // hdfsGetPathInfo hdfsFileInfo *fileInfo = NULL; if((fileInfo = hdfsGetPathInfo(fs, slashTmp)) != NULL) { fprintf(stderr, "hdfsGetPathInfo - SUCCESS!\n"); fprintf(stderr, "Name: %s, ", fileInfo->mName); fprintf(stderr, "Type: %c, ", (char)(fileInfo->mKind)); fprintf(stderr, "Replication: %d, ", fileInfo->mReplication); fprintf(stderr, "BlockSize: %"PRId64", ", fileInfo->mBlockSize); fprintf(stderr, "Size: %"PRId64", ", fileInfo->mSize); fprintf(stderr, "LastMod: %s", ctime(&fileInfo->mLastMod)); fprintf(stderr, "Owner: %s, ", fileInfo->mOwner); fprintf(stderr, "Group: %s, ", fileInfo->mGroup); char permissions[10]; permission_disp(fileInfo->mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileInfo->mPermissions, permissions); hdfsFreeFileInfo(fileInfo, 1); } else { totalResult++; fprintf(stderr, "hdfsGetPathInfo for %s - FAILED!\n", slashTmp); } // hdfsListDirectory hdfsFileInfo *fileList = 0; int numEntries = 0; if((fileList = hdfsListDirectory(fs, slashTmp, &numEntries)) != NULL) { int i = 0; for(i=0; i < numEntries; ++i) { fprintf(stderr, "Name: %s, ", fileList[i].mName); fprintf(stderr, "Type: %c, ", (char)fileList[i].mKind); fprintf(stderr, "Replication: %d, ", fileList[i].mReplication); fprintf(stderr, "BlockSize: %"PRId64", ", fileList[i].mBlockSize); fprintf(stderr, "Size: %"PRId64", ", fileList[i].mSize); fprintf(stderr, "LastMod: %s", ctime(&fileList[i].mLastMod)); fprintf(stderr, "Owner: %s, ", fileList[i].mOwner); fprintf(stderr, "Group: %s, ", fileList[i].mGroup); char permissions[10]; permission_disp(fileList[i].mPermissions, permissions); fprintf(stderr, "Permissions: %d (%s)\n", fileList[i].mPermissions, permissions); } hdfsFreeFileInfo(fileList, numEntries); } else { if (errno) { totalResult++; fprintf(stderr, "waah! hdfsListDirectory - FAILED!\n"); } else { fprintf(stderr, "Empty directory!\n"); } } char *newOwner = "root"; // Setting tmp dir to 777 so later when connectAsUser nobody, // we can write to it short newPerm = 0666; // hdfsChown fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, rwPath, NULL, "users")) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChown: %s\n", ((result = hdfsChown(fs, rwPath, newOwner, NULL)) ? "Failed!" : "Success!")); totalResult += result; // hdfsChmod fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, rwPath, newPerm)) ? "Failed!" : "Success!")); totalResult += result; sleep(2); tTime newMtime = time(NULL); tTime newAtime = time(NULL); // utime write fprintf(stderr, "hdfsUtime: %s\n", ((result = hdfsUtime(fs, rwPath, newMtime, newAtime)) ? "Failed!" : "Success!")); totalResult += result; // chown/chmod/utime read hdfsFileInfo *finfo = hdfsGetPathInfo(fs, rwPath); fprintf(stderr, "hdfsChown read: %s\n", ((result = (strcmp(finfo->mOwner, newOwner) != 0)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsChmod read: %s\n", ((result = (finfo->mPermissions != newPerm)) ? "Failed!" : "Success!")); totalResult += result; // will later use /tmp/ as a different user so enable it fprintf(stderr, "hdfsChmod: %s\n", ((result = hdfsChmod(fs, slashTmp, 0777)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr,"newMTime=%ld\n",newMtime); fprintf(stderr,"curMTime=%ld\n",finfo->mLastMod); fprintf(stderr, "hdfsUtime read (mtime): %s\n", ((result = (finfo->mLastMod != newMtime / 1000)) ? "Failed!" : "Success!")); totalResult += result; // Clean up hdfsFreeFileInfo(finfo, 1); fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, newDirectory, 1)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, srcPath, 1)) ? "Failed!" : "Success!")); totalResult += result; fprintf(stderr, "hdfsExists: %s\n", ((result = hdfsExists(fs, newDirectory)) ? "Success!" : "Failed!")); totalResult += (result ? 0 : 1); // Done test generic operations } { // Test Appends appendTemplate = strdup("/tmp/appendsXXXXXX"); if (!appendTemplate) { fprintf(stderr, "Failed to create appendTemplate!\n"); exit(1); } char *appendPath = mktemp(appendTemplate); const char* helloBuffer = "Hello,"; hdfsFile writeFile = NULL; // Create writeFile = hdfsOpenFile(fs, appendPath, O_WRONLY, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", appendPath); exit(1); } fprintf(stderr, "Opened %s for writing successfully...\n", appendPath); num_written_bytes = hdfsWrite(fs, writeFile, helloBuffer, (int) strlen(helloBuffer)); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); hdfsCloseFile(fs, writeFile); fprintf(stderr, "hdfsSetReplication: %s\n", ((result = hdfsSetReplication(fs, appendPath, 1)) ? "Failed!" : "Success!")); totalResult += result; // Re-Open for Append writeFile = hdfsOpenFile(fs, appendPath, O_WRONLY | O_APPEND, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", appendPath); exit(1); } fprintf(stderr, "Opened %s for appending successfully...\n", appendPath); helloBuffer = " World"; num_written_bytes = hdfsWrite(fs, writeFile, helloBuffer, (int)strlen(helloBuffer) + 1); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); hdfsCloseFile(fs, writeFile); // Check size hdfsFileInfo *finfo = hdfsGetPathInfo(fs, appendPath); fprintf(stderr, "fileinfo->mSize: == total %s\n", ((result = (finfo->mSize == strlen("Hello, World") + 1)) ? "Success!" : "Failed!")); totalResult += (result ? 0 : 1); // Read and check data hdfsFile readFile = hdfsOpenFile(fs, appendPath, O_RDONLY, 0, 0, 0); if (!readFile) { fprintf(stderr, "Failed to open %s for reading!\n", appendPath); exit(1); } tSize num_read_bytes = hdfsRead(fs, readFile, buffer, sizeof(buffer)); fprintf(stderr, "Read following %d bytes:\n%s\n", num_read_bytes, buffer); fprintf(stderr, "read == Hello, World %s\n", (result = (strcmp(buffer, "Hello, World") == 0)) ? "Success!" : "Failed!"); hdfsCloseFile(fs, readFile); // Cleanup fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, appendPath, 1)) ? "Failed!" : "Success!")); totalResult += result; // Done test appends } totalResult += (hdfsDisconnect(fs) != 0); { // // Now test as connecting as a specific user // This only meant to test that we connected as that user, not to test // the actual fs user capabilities. Thus just create a file and read // the owner is correct. const char *tuser = "******"; userTemplate = strdup("/tmp/usertestXXXXXX"); if (!userTemplate) { fprintf(stderr, "Failed to create userTemplate!\n"); exit(1); } char* userWritePath = mktemp(userTemplate); hdfsFile writeFile = NULL; fs = hdfsConnectAsUserNewInstance("default", 50070, tuser); if(!fs) { fprintf(stderr, "Oops! Failed to connect to hdfs as user %s!\n",tuser); exit(1); } writeFile = hdfsOpenFile(fs, userWritePath, O_WRONLY|O_CREAT, 0, 0, 0); if(!writeFile) { fprintf(stderr, "Failed to open %s for writing!\n", userWritePath); exit(1); } fprintf(stderr, "Opened %s for writing successfully...\n", userWritePath); num_written_bytes = hdfsWrite(fs, writeFile, fileContents, (int)strlen(fileContents) + 1); fprintf(stderr, "Wrote %d bytes\n", num_written_bytes); hdfsCloseFile(fs, writeFile); hdfsFileInfo *finfo = hdfsGetPathInfo(fs, userWritePath); if (finfo) { fprintf(stderr, "hdfs new file user is correct: %s\n", ((result = (strcmp(finfo->mOwner, tuser) != 0)) ? "Failed!" : "Success!")); } else { fprintf(stderr, "hdfsFileInfo returned by hdfsGetPathInfo is NULL\n"); result = -1; } totalResult += result; // Cleanup fprintf(stderr, "hdfsDelete: %s\n", ((result = hdfsDelete(fs, userWritePath, 1)) ? "Failed!" : "Success!")); totalResult += result; // Done test specific user } totalResult += (hdfsDisconnect(fs) != 0); // Shutdown the native minidfscluster nmdShutdown(cluster); nmdFree(cluster); fprintf(stderr, "totalResult == %d\n", totalResult); if (totalResult != 0) { return -1; } else { return 0; } }