Hadoop Configuration
新增hadoopuser用户
[root@noc rou]# adduser
bash: adduser: command not found
[root@noc rou]# cd /usr/bin/
[root@noc bin]# ln -s /usr/sbin/adduser adduser
[root@noc bin]# adduser hadoopuser
passwd wpsop
修改系统允许打开的文件数
有时候在程序里面需要打开多个文件,进行分析,系统一般默认数量是1024,(用ulimit -n可以看到)对于正常使用是够了,但是对于程序来讲,就太少了。
修改办法:
重启就OK
修改2个文件。
1)/etc/security/limits.conf
vi /etc/security/limits.conf
加上:
* soft nofile 8192
* hard nofile 20480
2)./etc/pam.d/login
session required /lib/security/pam_limits.so
注意:要重启才能生效(也就是把putty关了再打开)
创建mysql用户kwps和密码kwps
grant all privileges on *.* to 'kwps'@'%' identified by 'kwps' ;
flush privileges ;
简化输入
sudo -s 切换到root
vi /usr/bin/wpsop 新建
#! /bin/bash
ssh s$1-opdev-wps.rdev.kingsoft.net -l hadoopuser 指定用户wpsop
更改hosts
1) sudo vi /etc/hosts
2) sudo vi /etc/sysconfig/network
3) hostname -v newhostname
SSH免密码公钥认证
1) mkdir .ssh
2) cd .ssh
sudo chmod 700 . //这一步很重要
3) ssh-keygen -t rsa
4) cat rsa_d.pub >> authorized_keys
当然也可以: cp rsa_d.pub authorized_keys
使用 scp向其他服务器发送,注意不要覆盖原有的文件!!
5) chmod 644 authorized_keys //这一步很重要
注意:要保证所有的结点间(包括自连接)都是免密码ssh连接的
解压Hadoop-0.19.1
tar -xvf Hadoop-0.19.1
Hadoop配置
Hadoop下载地址
http://apache.etoak.com/hadoop/core/
http://hadoop.apache.org/common/releases.html
本机环境:
版本:Hadoop-0.191
操作系统:CentOS
五台服务器:
S2 (namenode)
S5 (secondarynamenode datanode)
S6 (datanode)
S7 (datanode)
S8 (datanode)
S9 (datanode)
***/home/wps/hadoop-0.19.1/conf***
修改masters:
s5
修改slaves:
s5
s6
s7
s8
s9
修改log4j.propperties
hadoop.log.dir=/data/hadoop-0.19.1/logs
修改hadoop-env.sh
export JAVA_HOME=/opt/JDK-1.6.0.14
export HADOOP_HEAPSIZE=4000
修改hadoop-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
<name>fs.default.name</name>
<value>hdfs://s2-opdev-wps.rdev.kingsoft.net:9000/</value>
<description>The name of the default file system. Either the literal string "local" or a host:port for DFS.</description>
</property>
<property>
<name>mapred.job.tracker</name>
<value>s2-opdev-wps.rdev.kingsoft.net:9001</value>
<description>The host and port that the MapReduce job tracker runs at. If "local", then jobs are run in-process as a single map and reduce task.</description>
</property>
<property>
<name>dfs.name.dir</name>
<value>/data/hadoop-0.19.1/name</value>
<description>Determines where on the local filesystem the DFS name node should store the name table. If this is a comma-delimited list of directories then the name table is
replicated in all of the directories, for redundancy. </description>
</property>
<property>
<name>dfs.data.dir</name>
<value>/data/hadoop-0.19.1/dfsdata</value>
<description>Determines where on the local filesystem an DFS data node should store its blocks. If this is a comma-delimited list of directories, then data will be stored in
all named directories, typically on different devices. Directories that do not exist are ignored.</description>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/data/hadoop-0.19.1/tmp</value>
<description>A base for other temporary directories.</description>
</property>
<property>
<name>dfs.replication</name>
<value>3</value>
<description>Default block replication. The actual number of replications can be specified when the file is created. The default is used if replication is not specified in
create time.</description>
</property>
<property>
<name>fs.checkpoint.dir</name>
<value>/data/hadoop-0.19.1/namesecondary</value>
<description>Determines where on the local filesystem the DFS secondary
name node should store the temporary images to merge.
If this is a comma-delimited list of directories then the image is
replicated in all of the directories for redundancy.
</description>
</property>
<property>
<name>dfs.http.address</name>
<value>s2-opdev-wps.rdev.kingsoft.net:50070</value>
<description>
The address and the base port where the dfs namenode web ui will listen on.
If the port is 0 then the server will start on a free port.
</description>
</property>
<property>
<name>mapred.map.tasks</name>
<value>50</value>
<description>The default number of map tasks per job. Typically set
to a prime several times greater than number of available hosts.
Ignored when mapred.job.tracker is "local".
</description>
</property>
<property>
<name>mapred.reduce.tasks</name>
<value>7</value>
<description>The default number of reduce tasks per job. Typically set
to a prime close to the number of available hosts. Ignored when
mapred.job.tracker is "local".
</description>
</property>
启动hadoop
bin/hadoop namenode —format
&& Do not format a running Hadoop namenode ,this will cause all your data in the HDFS filesystem to be erased. &&
bin/start-all.sh
bin/stop-all.sh
查看文件目录:
bin/hadoop fs -ls /
查看数据块:
/home/wpsop/hadoop-0.19.1/running/dfsdata/current
Bin/hadoop fs -ls /data/user/hiveware
分享到:
相关推荐
hadoop配置,hadoop配置打包,hadoop配置。
hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件hadoop配置文件
Hadoop配置手册,详细介绍hadoop的配置 系统环境配置 假定有 3 台机器,机器上均按照了 Ubuntu 操作系统: 192.168.1.138 hadoop-master //作为 Hadoop 的主机用 192.168.1.140 hadoop-node1 //作为 Hadoop 的数据...
hadoop配置文件详解,core-site.xml是全局配置,hdfs-site.xml和mapred-site.xml分别是hdfs和mapred的局部配置。
hadoop 配置项的调优,列出大部分HADOOP系统参数的定义及与系统性能之间的关系。
hadoop配置属性hadoop配置属性hadoop配置属性hadoop配置属性hadoop配置属性
hadoop2.x配置文件,一共八个文件,能够配置hadoop yarn等框架
本文档列出了hadoop的常用配置说明,对hadoop的学习很有帮助。
经过多次反复试验,完全可用的hadoop配置,有0.19的版本,也有0.20的版本。并且有脚本可以在两个版本之间切换,无需格式化文件系统。Without "hadoop namenode -format" 目录结构: /data/hadoop 为工作目录,实际为...
hadoop2.2.0单机伪分布式配置文件实例。
使用SQLSERVER访问hadoop2.0http地址,通过powershell下载其配置文件(xml格式);在TSQL然后通过xml查询,解析出hadoop2.0配置详细信息,这样可以通过熟悉的TSQL检索查询hadoop配置,并归类之
调用保存文件的算子,需要配置Hadoop依赖 将文件夹中的 hadoop-3.0.0 解压到电脑任意位置 在Python代码中使用os模块配置:os.environ[‘HADOOP_HOME’] = ‘HADOOP解压文件夹路径’ winutils.exe,并放入Hadoop解压...
详细的hadoop单机配置步骤,非常好的一个文档。
史上最全的Hadoop配置完全手册;配置,安装,设置
hadoop 配置自我总结
4.Hadoop配置文件的更改 !!!!!!!!!!!!!!!!!!!! ps 请特别注意以下几个问题: 1.hostname 默认改成master ,如果不想修改,请在run.sh 中注释sudo ./change_hostname 2.集群的IP地址 默认为...
HADOOP配置文件-有注释版,里面附带源码,方便进行复制粘贴改动
不用积分,简单的HA配置参考,只是配置没有其他。为name迁移提供的配置参考https://blog.csdn.net/hxiaowang/article/details/89962226
hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书hadoop好书