本文介绍了使用MiniDFSCluster单元测试hadoop hdfs着作的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我写了一个写给hadoop HDFS的类。
我正在使用版本1.2.1的hadoop jar。



我想测试这个类。
所以基于博客,如我写了我的代码:

pre $ private $ createSimulatedHdfs ();
// 100K blocksize
conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY,1024 * 100);
conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY,100);
conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY,1);
conf.setLong(DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY,DFS_REPLICATION_INTERVAL);
conf.setInt(DFSConfigKeys.DFS_NAMENODE_REPLICATION_INTERVAL_KEY,DFS_REPLICATION_INTERVAL);

try {
//模拟的HDFS
cluster = new MiniDFSCluster(conf,DATANODE_COUNT,true,null);
cluster.waitActive();
simulatedHdfs = cluster.getFileSystem();
} catch(IOException e){
Assert.fail(Could not create simulated HDFS+ e.getMessage());
}
}

但是当运行 new MiniDFSCluster 运行异常:

  java.lang.AssertionError:无法创建模拟的HDFS Can not运行程序du:CreateProcess error = 2,系统无法在com.taptica.hdfs.writer上的org.junit.Assert.fail(Assert.java:88)
处找到指定
的文件。 HdfsWriterUTest.createSimulatedHdfs(HdfsWriterUTest.java:101)
在com.taptica.hdfs.writer.HdfsWriterUTest.initJunitModeTest(HdfsWriterUTest.java:42)
在com.taptica.hdfs.writer.HdfsWriterUTest.writeTest( HdfsWriterUTest.java:50)
在sun.reflect.NativeMethodAccessorImpl.invoke0(本机方法)
在sun.reflect.NativeMethodAccessorImpl.invoke(来源不明)
在sun.reflect.DelegatingMethodAccessorImpl.invoke (未知源)
在java.lang.reflect.Method.invoke(未知源)
在org.junit.runners.model.FrameworkMethod $ 1.runReflectiv紧急呼叫(FrameworkMethod.java:47)
在org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
在org.junit.runners.model.FrameworkMethod.invokeExplosively( FrameworkMethod.java:44)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:在org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70 271)

在org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
。在组织.junit.runners.ParentRunner $ 3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner $ 1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner .runChildren(ParentRunner.java:236)
在org.junit.runners.ParentRunner.access $ 000(ParentRunner.java:53)
在org.junit.runners.ParentRunner $ 2.evaluate(ParentRunner.java :229)在org.junit.runners.ParentRunner上
。运行(ParentRunner.java:309)
在org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
在org.eclipse.jdt.internal.junit。 runner.TestExecution.run(TestExecution.java:38)
at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
at org.eclipse.jdt。 internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
在org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
。在组织。 eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)

我没有在我的本地环境中安装hadoop(我不打算这么做)。
我该如何解决这个问题?

解决方案

对于JUnit测试,本地文件系统可以在没有安装Hadoop的情况下使用第三方utils:

  LocalFileSystem fs = FileSystem.getLocal(new Configuration()); 

另外请看一下MRunit,MR测试的有用工具:


I wrote a class that writes to hadoop HDFS.I'm using hadoop jars of version 1.2.1.

I would like to test this class.So based on blogs such as this one I wrote my code:

private void createSimulatedHdfs() {
    conf = new Configuration();
    // 100K blocksize
    conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 1024 * 100);
    conf.setLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, 100);
    conf.setInt(DFSConfigKeys.DFS_BYTES_PER_CHECKSUM_KEY, 1);
    conf.setLong(DFSConfigKeys.DFS_HEARTBEAT_INTERVAL_KEY, DFS_REPLICATION_INTERVAL);
    conf.setInt(DFSConfigKeys.DFS_NAMENODE_REPLICATION_INTERVAL_KEY, DFS_REPLICATION_INTERVAL);

    try {
        // simulated HDFS
        cluster = new MiniDFSCluster(conf, DATANODE_COUNT, true, null);
        cluster.waitActive();
        simulatedHdfs = cluster.getFileSystem();
    } catch (IOException e) {
        Assert.fail("Could not create simulated HDFS " + e.getMessage());
    }
    }

But when running the new MiniDFSCluster running into an exception:

java.lang.AssertionError: Could not create simulated HDFS Cannot run program "du": CreateProcess error=2, The system cannot find the file specified
    at org.junit.Assert.fail(Assert.java:88)
    at com.taptica.hdfs.writer.HdfsWriterUTest.createSimulatedHdfs(HdfsWriterUTest.java:101)
    at com.taptica.hdfs.writer.HdfsWriterUTest.initJunitModeTest(HdfsWriterUTest.java:42)
    at com.taptica.hdfs.writer.HdfsWriterUTest.writeTest(HdfsWriterUTest.java:50)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
    at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
    at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
    at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
    at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
    at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
    at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
    at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
    at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
    at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
    at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
    at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
    at org.eclipse.jdt.internal.junit4.runner.JUnit4TestReference.run(JUnit4TestReference.java:50)
    at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
    at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
    at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
    at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
    at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)

I did not install hadoop on my local env (and I don't intend doing so).how can I get over this one?

解决方案

For JUnit testing local file system can be used without Hadoop installation and third-party utils:

        LocalFileSystem fs = FileSystem.getLocal(new Configuration());

Also take a look on MRunit, helpfull utility for MR testing: http://mrunit.apache.org/

这篇关于使用MiniDFSCluster单元测试hadoop hdfs着作的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持!

10-10 05:10