Use Hadoop pesudo distributed.
Start Hadoop
Specify Env for Hadoop
Edit${HADOOP_HOME}/etc/hadoop/hadoop_env.sh
1export JAVA_HOME=/opt/jdkStart all processes
Format and start
123$> hadoop namenode -format // format file system$> start-all.sh // start all processes$> jps // check processesCheck file system by WebUI
1$> /etc/init.d/iptables stop // stop firewallOpen in browser:
http://localhost:50070
- Check HDFS file system 123$> hdfs dfs -ls /$> hdfs dfs -mkdir -p /usr/centos01/hadoop // make a directory$> hdfs dfs -lsr /
Hadoop Introduce
- Hadoop ports
50070 –> namenode http port
50075 –> datanode http port
50090 –> 2nd namenode http port
8020 –> namenode rpc port
50010 –> datanode rpc port - Hadoop 4 modules
- common + hdfs
NameNode, SecondaryNameNode, DataNode - mapred + yarn
NodeManager, ResourceManager
- common + hdfs
- Hadoop scripts
- ALL
start-all.sh
: start all nodes
stop-all.sh
: stop all nodes - HDFS
start-dfs.sh
: NameNode, DataNode, 2ndNameNode
stop-dfs.sh
: - YARN:
start-yarn.sh
: NodeManager, ResourceManager
stop-yarn.sh
:
- ALL