尝试运行日记节点时,它将失败。出现以下错误:

./hadoop-daemon.sh start journalnode

Error: Could not find or load main class org.apache.hadoop.hdfs.qjournal.server.JournalNode

可以是什么?

这是我的 core-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://hdfscluster</value>
    </property>
    <property>
        <name>io.native.lib.available</name>
        <value>True</value>
    </property>
    <property>
        <name>io.file.buffer.size</name>
        <value>65536</value>
    </property>
    <property>
        <name>fs.trash.interval</name>
        <value>60</value>
    </property>
</configuration>

这是 hdfs-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>

<configuration>
    <property>
        <name>dfs.namenode.name.dir</name>
        <value>file:///srv/node/d1/hdfs/nn,file:///srv/node/d2/hdfs/nn,file:///srv/node/d3/hdfs/nn</value>
        <final>true</final>
    </property>

    <property>
        <name>dfs.datanode.data.dir</name>
        <value>file:///srv/node/d1/hdfs/dn,file:///srv/node/d2/hdfs/dn,file:///srv/node/d3/hdfs/dn</value>
        <final>true</final>
    </property>

    <property>
        <name>dfs.namenode.checkpoint.dir</name>
        <value>file:///srv/node/d1/hdfs/snn,file:///srv/node/d2/hdfs/snn,file:///srv/node/d3/hdfs/snn</value>
        <final>true</final>
    </property>

    <property>
        <name>dfs.nameservices</name>
        <value>hdfscluster</value>
    </property>

    <property>
        <name>dfs.ha.namenodes.hdfscluster</name>
        <value>nn1,nn2</value>
    </property>

    <property>
        <name>dfs.namenode.rpc-address.hdfscluster.nn1</name>
        <value>192.168.57.101:8020</value>
    </property>

    <property>
        <name>dfs.namenode.http-address.hdfscluster.nn1</name>
        <value>192.168.57.101:50070</value>
    </property>
    <property>
        <name>dfs.namenode.rpc-address.hdfscluster.nn2</name>
        <value>192.168.57.102:8020</value>
    </property>

    <property>
        <name>dfs.namenode.http-address.hdfscluster.nn2</name>
        <value>192.168.57.102:50070</value>
    </property>

    <property>
        <name>dfs.journalnode.edits.dir</name>
        <value>/srv/node/d1/hdfs/journal</value>
        <final>true</final>
    </property>

    <property>
        <name>dfs.namenode.shared.edits.dir</name>
        <value>qjournal://192.168.57.101:8485;192.168.57.102:8485;192.168.57.103:8485/hdfscluster</value>
    </property>

    <property>
        <name>dfs.client.failover.proxy.provider.hdfscluster</name>
        <value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
    </property>

    <property>
        <name>dfs.ha.automatic-failover.enabled</name>
        <value>true</value>
    </property>

    <property>
        <name>ha.zookeeper.quorum</name>
        <value>192.168.57.101:2181,192.168.57.102:2181,192.168.57.103:2181</value>
    </property>

    <property>
        <name>dfs.ha.fencing.methods</name>
        <value>sshfence</value>
    </property>

    <property>
        <name>dfs.ha.fencing.ssh.private-key-files</name>
        <value>/home/hdfs/.ssh/id_dsa</value>
    </property>

    <property>
        <name>dfs.hosts</name>
        <value>/etc/hadoop/conf/dfs.hosts</value>
    </property>

    <property>
        <name>dfs.hosts.exclude</name>
        <value>/etc/hadoop/conf/dfs.hosts.exclude</value>
    </property>

    <property>
        <name>dfs.replication</name>
        <value>3</value>
    </property>
    <property>
        <name>dfs.permission</name>
        <value>False</value>
    </property>
    <property>
        <name>dfs.durable.sync</name>
        <value>True</value>
    </property>
    <property>
        <name>dfs.datanode.synconclose</name>
        <value>True</value>
    </property>
</configuration>

节点IP为192.168.57.1​​03
它必须运行journalnode和datanode

我正在使用Hadoop 2.8.0。是配置上的问题还是我只是错过了什么?

最佳答案

我不知道为什么,但是缺少/usr/lib/hadoop/share/hadoop/目录。我从头开始重新安装了hadoop,现在可以使用了。

关于hadoop - Hadoop-无法找到或加载主类org.apache.hadoop.hdfs.qjournal.server.JournalNode,我们在Stack Overflow上找到一个类似的问题:https://stackoverflow.com/questions/43576304/

10-16 22:42