cloudera cdh官方文档 Impala本地通过Parcel安装配置详解及什么是Parcel

使用cloudera manager 安装CDH5 -
- ITeye技术网站
博客分类:
使用cloudera manager安装cdh5
[root@hadoopone ~]# cat /etc/issue
CentOS release 6.5 (Final)
Kernel \r on an \m
[root@master01 .ssh]# getconf LONG_BIT
64
[root@master01 .ssh]#
1.修改主机名
vi /etc/sysconfig/network
[root@hadoopfive ~]# cat /etc/sysconfig/network
NETWORKING=yes
HOSTNAME=master01
[root@hadoopfive ~]#
分别对应关系
vi /etc/hosts
192.168.1.207 master01
192.168.1.208 master02
192.168.1.209 datanode01
192.168.1.213 datanode02
192.168.1.214 datanode03
2.配置信任关系
cd /root/.ssh
ssh-keygen -t rsa
[root@master01 .ssh]# cat id_rsa.pub &&authorized_keys
[root@master01 .ssh]# cat -n authorized_keys
&&&& 1 ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAqHUFyjOlpQtFN4gZy6kqQKSg26NciIG4L4yN5chs3lSweWUuui+0+B7ONQr2E6wCKeQXnzoD3ohht+YbOUkAg7mDWoBCEbsdGKIGzhbkfersajCMWv4h409cLW8WsCcGQkqhZn1h2u0bsbwyfyLsYH1BuWwA3xhu4lgqFp/KSsioDuKqysah41VeLUbxkj3FveLLS8rESBecv0YSqEkSLqKvKFXZ9CV7Txo7IzGSs9pnXgfZ818/wMSOXvL+5vzZyWXwuxhLsRwF+sJedd49eV2vXDO4Fb0b8W8zp+Mzeej7zGZ0zbtfXXu8l2HjX3bITMy61u6ezN5Kfg0+5C9g8w== root@master01
[root@master01 .ssh]# scp authorized_keys& root@192.168.1.208:/root/.ssh
The authenticity of host '192.168.1.208 (192.168.1.208)' can't be established.
RSA key fingerprint is e2:ce:9f:8c:d7:8d:5d:20:08:27:e4:22:64:73:92:75.
Are you sure you want to continue connecting (yes/no)? yes
Warning: Permanently added '192.168.1.208' (RSA) to the list of known hosts.
root@192.168.1.208's password:
authorized_keys&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&&& 100%& 395&&&& 0.4KB/s&& 00:00&&&
[root@master01 .ssh]#
要保证每一个都能直接ssh过去,所以先要都做一下操作
[root@master01 .ssh]# ssh root@192.168.1.213 date& 合计25次
3.关闭selinux及防火墙
sed -i -e 's/SELINUX=enforcing/SELINUX=disabled/g' /etc/sysconfig/selinux
sed -i -e 's/SELINUX=enforcing/SELINUX=disabled/g' /etc/selinux/config
chkconfig iptables off
修改完后要重启,并使用命令检查SELinux状态:
1、/usr/sbin/sestatus -v&&&&& ##如果SELinux status参数为enabled即为开启状态
SELinux status:&&&&&&&&&&&&&&&& enabled
2、getenforce&&&&&&&&&&&&&&&& ##也可以用这个命令检查
4.安装必要的包
yum install -y sysstat vim tree wget lrzsz screen gcc python-devel gcc-c++& ntpdate libyaml libyaml-devel python-setuptools ntp& libreoffice-core& libreoffice-ure
5.安装jdk
先卸载自带jdk
yum -y erase java
rpm -qa | grep jdk
rpm -e --nodeps java-*-openjdk*
安装jdk1.7版本 到/usr/java/jdk1.7
tar -xzvf jdk-7u67-linux-x64.tar.gz
mv jdk-7u67-linux-x64 /usr/java/jdk1.7
[root@master01 local]# source /etc/profile
export JAVA_HOME=/usr/java/jdk1.7
export JRE_HOME=$JAVA_HOME/jre
export PATH=$PATH:$JAVA_HOME/bin:$JRE_HOME/bin
export CLASSPATH=:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar:$JRE_HOME/lib
[root@master01 local]# java -version
java version "1.7.0_67"
Java(TM) SE Runtime Environment (build 1.7.0_67-b01)
Java HotSpot(TM) 64-Bit Server VM (build 24.65-b04, mixed mode)
6.所有需要的包:
cloudera-manager-agent-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
cloudera-manager-daemons-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
cloudera-manager-server-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
cloudera-manager-server-db-2-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
enterprise-debuginfo-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
jdk-6u31-linux-amd64.rpm&&&&&&&& (可选择不装,因为我们选择的版本已经不支持jdk1.6了,至少是1.7)
cloudera-manager-installer.bin (5.4.6)
CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel
CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha&&& (这个是从manifest.json中抠出来的对应版本的hash值)
manifest.json
7.安装cloudera manager
下载地址:
http://archive-/cm5/installer/5.4.6/&& --下载cloudera-manager-install.bin
http://archive-/cdh5/parcels/&&& --下载cdh
http://archive-/cm5/redhat/6/x86_64/cm/5.4.6/RPMS/x86_64/& --下载cloudera-manager的包
/cdh5/parcels/5.4.10/manifest.json& --查看对应cdh的hash
http://archive-/cm5/cm/5/& --下载gz包
说明:
cloudera-manager是5.4.6的
parcels都是用的5.4.10的
-rwxrwxrwx& 1 root root&&& 4781732 Jun& 4 18:02 cloudera-manager-agent-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx. 1 root root&
Jun& 4 17:00 cloudera-manager-daemons-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&&&& 8556 Jun& 4 18:02 cloudera-manager-server-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&&&& 9880 Jun& 4 18:02 cloudera-manager-server-db-2-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&& 980244 Jun& 4 18:02 enterprise-debuginfo-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&
Jun& 6 10:50 oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
master节点需要安装所有这6个
-rwxrwxrwx& 1 root root&&& 4781732 Jun& 4 18:02 cloudera-manager-agent-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx. 1 root root&
Jun& 4 17:00 cloudera-manager-daemons-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&&&& 8556 Jun& 4 18:02 cloudera-manager-server-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&&&& 9880 Jun& 4 18:02 cloudera-manager-server-db-2-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&&&& 980244 Jun& 4 18:02 enterprise-debuginfo-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx& 1 root root&
Jun& 6 10:50 oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
其他节点需要安装2个:
-rwxrwxrwx& 1 root root&&& 4781732 Jun& 4 18:02 cloudera-manager-agent-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rwxrwxrwx. 1 root root&
Jun& 4 17:00 cloudera-manager-daemons-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
安装命令:
yum localinstall --nogpgcheck& --skip-broken *.rpm
[root@master01 cm]# cd /opt/cloudera/parcel-repo/&&& --这个路径是安装cloudera manager的时候自动生成的,我们把下好的包直接拷贝进去,注意要root用户
下载完成后只上传到master节点即可。然后解压到/opt目录下,不能解压到其他地方,因为cdh5的源会默认在/opt/cloudera/parcel-repo寻找
[root@master01 parcel-repo]# ll
total 1233984
-rwxrwxrwx 1 root root
Jun& 7 11:03 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel
-rwxrwxrwx 1 root root&&&&&&&& 41 Jun& 7 11:00 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha
-rwxrwxrwx 1 root root&&&&& 43172 Jun& 7 12:59 manifest.json
[root@master01 parcel-repo]# cat CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha
2b9effbceeaaf3291308d
7.启动服务ntpd
service ntpd start
chkconfig ntpd on
编辑脚本,定时同步时间
*/1 * * * * root ntpdate -u ntpdate ntp.api.bz
补充:修改时区
vi /etc/sysconfig/clock
ZONE='Asia/Shanghai'
rm /etc/localtime
ln -sf /usr/share/zoneinfo/Asia/Shanghai /etc/localtime
reboot
8.关闭透明大内存页功能
echo never & /sys/kernel/mm/redhat_transparent_hugepage/defrag
echo "echo never & /sys/kernel/mm/redhat_transparent_hugepage/defrag" && /etc/rc.local
9.修改系统配置,防止hadoop运行出错
echo "* soft nofile 65536"&&/etc/security/limits.conf
echo "* hard nofile 65536"&&/etc/security/limits.conf
echo "root soft nofile 65536"&&/etc/security/limits.conf
echo "root hard nofile 65536"&&/etc/security/limits.conf
echo "* soft memlock unlimited"&&/etc/security/limits.conf
echo "* hard memlock unlimited"&&/etc/security/limits.conf
echo "root soft memlock unlimited"&&/etc/security/limits.conf
echo "root hard memlock unlimited"&&/etc/security/limits.conf
echo "* soft as unlimited"&&/etc/security/limits.conf
echo "* hard as umlimited"&&/etc/security/limits.conf
echo "root soft as unlimited"&&/etc/security/limits.conf
echo "root hard as unlimited"&&/etc/security/limits.conf
echo "vm.max_map_count = 131072"&&/etc/sysctl.conf
echo "vm.swappiness=0"&&/etc/sysctl.conf
sysctl -p& 使配置生效
10.需要挂载的磁盘先进行分区并格式化& /dev/sdb
parted /dev/sdb
mklabel gpt
mkpart primary %
mkfs.ext4 /dev/sdb1
11.将磁盘挂载到/data/disk01& /data/disk02等目录,并将磁盘挂载命令写入到/etc/fstab中,以便于重启后能生效
uuid=`ls /dev/disk/by-uuid/ -l |grep sdb1 |awk '{print $9}'`
echo "UUID=$uuid /data/disk01 ext4 default,noatime,nodiratime 0 0" &&/etc/fstab
12.启动服务:
Server
service cloudera-scm-server start
agent
service cloudera-scm-agent start& --必须要保证都是启动的
如果失败,可查看日志/var/log/cloudera-scm-agent/logs
失败处理:
service cloudera-scm-agent hard_stop_confirmed
yum remove -y 'cloudera-manager-*' hadoop hue-common 'bigtop-*'
rm -f /var/run/cloudera-scm-agent.pid
yum localinstall --nogpgcheck& --skip-broken *.rpm
service cloudera-scm-agent start
service cloudera-scm-agent status
13.开始安装manager(注意这里采用的离线方式,在线要求网速给力)
chmod +x cloudera-manager-installer.bin
./cloudera-manager-installer.bin
安装报错的话,可查看日志/var/log/cloudera-manager-installer/
这个步骤会在/etc/yum.repo.d/下生成cloudera-manager.repo,内容如下
[cloudera-manager]
name = Cloudera Manager, Version 5.4.6
baseurl = /cm5/redhat/6/x86_64/cm/5.4.6/
gpgkey = /redhat/cdh/RPM-GPG-KEY-cloudera
gpgcheck = 1
有时候会失败,依然会去网上下载,这个时候,我们可能需要重新安装那6个包,或者
./cloudera-manager-install.bin& --skip_repo_package=1&&& 让它不生成包,使用我们自己创建的包
如果需要重装,那么就要删除一下这个
rm -f& /etc/cloudera-scm-server/db.properties
cd /usr/share/cmf/
./uninstall-manager.bin
yum localinstall --nogpgcheck --skip-broken *.rpm
./cloude-manager-install.bin
还有时候会报错,找不到db.properties,我们可以在bin启动之后,自己创建一个放进去
[root@master01 opt]# /usr/share/cmf/schema/scm_prepare_database.sh mysql& -uroot -p --scm-host localhost scm scm scm123
Enter database password:
JAVA_HOME=/usr/java/jdk1.7.0
Verifying that we can write to /etc/cloudera-scm-server& --放到这个目录下了
Creating SCM configuration file in /etc/cloudera-scm-server
Executing:& /usr/java/jdk1.7.0/bin/java -cp /usr/share/java/mysql-connector-java.jar:/usr/share/java/oracle-connector-java.jar:/usr/share/cmf/schema/../lib/* com.cloudera.enterprise.dbutil.DbCommandExecutor /etc/cloudera-scm-server/db.properties com.cloudera.cmf.db.
[&&&&&&&&&&&&&&&&&&&&&&&&& main] DbCommandExecutor&&&&&&&&&&&&& INFO& Successfully connected to database.
All done, your SCM database is configured correctly!
[root@master01 opt]#
补充13说明,当然也可以不用./cloudera-manager-install.bin来安装,可以直接下载
cloudera-manager-el6-cm5.4.6_x86_64.tar.gz
解压到master节点/opt即可,给所有的节点添加cloudera-scm用户
useradd --system --home=/opt/cm-5.4.6/run/cloudera-scm-server --no-createhome --shell=/bin/false --comment "cloudera scm user" cloudera-scm
以上未经测试
安装成功,会看到:
Your browser should now open to http://192.168.1.207:7180/. Log in to Cloudera Manager with
the username and password set to 'admin' to continue installation.
14.进入管理页面配置
[root@master01 bin]# ./apachectl start&&& --起来后,要稍微等一会才行
httpd (pid 3473) already running
[root@master01 bin]#& service cloudera-scm-server start
cloudera-scm-server is already running
[root@master01 bin]#& service cloudera-scm-server-db start
Database is already running. Please stop it first., giving up
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180&& --确保这几个都正常启动了,才能打开页面
tcp&&&&&&& 0&&&&& 0 0.0.0.0:7180&&&&&&&&&&&&&&& 0.0.0.0:*&&&&&&&&&&&&&&&&&& LISTEN&&&&&
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55213&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55212&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55216&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55217&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55214&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55215&&&&&&&&& ESTABLISHED
页面地址:http://192.168.1.207:7180/
用户/密码:admin/admin
15.安装mysql库
安装:
groupadd mysql
useradd -s /sbin/nologin -g mysql mysql
tar -zxvf mysql-5.7.10-linux-glibc2.5-x86_64.tar.gz
mv mysql-5.7.10-linux-glibc2.5-x86_64 /usr/local/mysql
cd /usr/local/mysql/support-files
cp mysql.server /etc/init.d/mysql57
chkconfig --add mysql57
chkconfig --level 2345 mysql57 on
mkdir -p /usr/local/mysql/data
chown -R mysql:mysql /usr/local/mysql/
初始化:
cd /usr/local/mysql/
bin/mysqld --initialize --user=mysql --basedir=/usr/local/mysql --datadir=/usr/local/mysql/data
记下密码
修改配置
vi /f
[root@master01 parcel-repo]# cat /f
[mysqld]
port=3306
datadir=/usr/local/mysql/data
socket=/tmp/mysql.sock
[mysqld_safe]
log-error=/usr/local/mysql/mysqlerr.log
pid-file=/usr/local/mysql/mysqld.pid
启动服务
service mysql57 start
/usr/local/mysql/bin/mysql -uroot -p
set password=password('root');
创建所需要的库和用户
create database hive charset utf8;
grant all on *.* to hive@'%' identified by 'hive123';
create database monitor charset utf8;
grant all on *.* to monitor@'%' identified by 'monitor123';
create database oozie charset utf8;
grant all on *.* to oozie@'%' identified by 'oozie123';
create databa
grant all on *.* to amon@'%' identified by 'amon123';
create database oozie charset utf8;
grant all on *.* to hue@'%' identified by 'hue123';
16.拷贝hive连接mysql需要的jar包(这里hive使用的是mysql库作为元数据库)
unzip mysql-connector-java-5.1.30.zip
cp mysql-connector-java-5.1.30-bin.jar& /usr/java/mysql-connector-java.jar& --一定是要是这个名称
17.将Mysql的驱动包拷贝到/opt/cloudera/parcels/CDH-5.0.0-1.cdh5.0.0.p0.47/lib/hive/lib/目录下
18.页面开始安装,按照提示操作。
服务字母对应:
HBASE部分
M ---Master
HBTS --HBase Thrift Server
G& --Gateway
HBRS --HBase REST Server
RS --RegionServer
HDFS部分
FC --Failover Controller
SNN& --SecondaryNameNode
NFSC --NFS Gateway
HFS --HttpFS
NN --Namenode
G& --Gateway
JN --JournalNode
DN --DateNode
Hive部分
HMS --Hive Metastore Server
WHC --WebHCat Server
HS2 --HiveServer2
G --Gateway
Hue部分
HS --Hue Server
KTR kerberos Ticket Renewer
Oozie部分
OS --Oozie Server
YARN部分
G --Gateway
NM --NodeManager
Zookeeper
S --Server
测试环境各个软件对应的版本
Cloudera manager 5.4.6
CDH 5.4.10
hadoop 2.6.0
hive& 1.1.0
hbase 1.0.0
zookeeper& 3.4.5
sqoop 1.4.5
jdk 1.7.0_67
页面的基本步骤
1.选择免费版
2.选中所有主机
192.168.1.207,192.168.1.208,192.168.1.209,192.168.1.213,192.168.1.214
搜索全选
3.选择使用parcel安装集群
注意看CDH的版本,是否是我们下的那个CDH-5.4.10-1.cdh5.4.10.p0.16
可以配置JAVA加密(默认方式)
install oracle java se development kit
ssh数据root密码,同时安装的数量1
注释:
指定主机的 SSH 登录方式
一种通过root用户,密码需要一致,刚开始安装建议使用root。
也可以使用非root用户,但是需要保证sudo无密码访问
实现方法如下
给aboutyun用户设置无密码sudo权限:
chmod u+w /etc/sudoers
aboutyun ALL=(root)NOPASSWD:ALL
chmod u-w /etc/sudoers
测试:sudo ifconfig
4.正常读取的话,会显示(已下载100%)
然后就是分配、激活
5.主机检测,务必保证所有主机的配置都验证通过
jdk显示不适用,没关系
6.选择需要安装的服务组合
含HBase的内核
7.角色分配
将各个主机的角色进行均衡分配
8.数据库测试
hive库,monitor库等都选择mysql,适用我们刚刚创建的库名和用户名密码
测试连接全部successful再进行下一步
9.审核更改
修改目录,一般会在默认路径前面加/hadoop,方便统一管理
10.启动安装配置
全部运行成功后,基本就可以了
可以访问的一些网址
cloudera manager管理工具:http://192.168.1.207:7180&& admin& admin
hue网址:http://192.168.1.207:8888/accounts/login/?next=/&&& hue/hue123
job管理:http://192.168.1.207:8088/cluster&&
oozie页面:http://192.168.1.207:11000/oozie/
角色添加:
群集--选择需要添加的服务--实例--添加角色实例--选择主机-运行--启动
补充:oozie页面的配置
CDH刚配置好的时候,oozie页面会报错:
Oozie web console is disabled.
To enable Oozie web console install the Ext JS library.
Refer to Oozie Quick Start documentation for details.
安装这个操作即可:
/gplextras/misc/
下载
# mv ext-2.2.zip /var/lib/oozie/
# cd /var/lib/oozie
# unzip ext-2.2.zip
# chown -R oozie:oozie ext-2.2
刷新页面:http://192.168.1.207:11000/oozie/
果然ok了!
补充:简单的job测试:
[root@master01 hadoop-mapreduce]#& sudo -u hdfs hadoop jar /opt/cloudera/parcels/CDH/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar pi 10 100
Number of Maps& = 10
Samples per Map = 100
Wrote input for Map #0
Wrote input for Map #1
Wrote input for Map #2
Wrote input for Map #3
Wrote input for Map #4
Wrote input for Map #5
Wrote input for Map #6
Wrote input for Map #7
Wrote input for Map #8
Wrote input for Map #9
Starting Job
16/06/08 15:46:19 INFO client.RMProxy: Connecting to ResourceManager at master01/192.168.1.207:8032
16/06/08 15:46:20 INFO input.FileInputFormat: Total input paths to process : 10
16/06/08 15:46:20 INFO mapreduce.JobSubmitter: number of splits:10
16/06/08 15:46:20 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_0_0001
16/06/08 15:46:20 INFO impl.YarnClientImpl: Submitted application application_0_0001
16/06/08 15:46:20 INFO mapreduce.Job: The url to track the job: http://master01:8088/proxy/application_0_0001/
16/06/08 15:46:20 INFO mapreduce.Job: Running job: job_0_0001
16/06/08 15:46:25 INFO mapreduce.Job: Job job_0_0001 running in uber mode : false
16/06/08 15:46:25 INFO mapreduce.Job:& map 0% reduce 0%
16/06/08 15:46:30 INFO mapreduce.Job:& map 10% reduce 0%
16/06/08 15:46:31 INFO mapreduce.Job:& map 40% reduce 0%
16/06/08 15:46:34 INFO mapreduce.Job:& map 60% reduce 0%
16/06/08 15:46:35 INFO mapreduce.Job:& map 80% reduce 0%
16/06/08 15:46:37 INFO mapreduce.Job:& map 100% reduce 0%
16/06/08 15:46:40 INFO mapreduce.Job:& map 100% reduce 100%
16/06/08 15:46:40 INFO mapreduce.Job: Job job_0_0001 completed successfully
16/06/08 15:46:40 INFO mapreduce.Job: Counters: 49
File System Counters
FILE: Number of bytes read=99
FILE: Number of bytes written=1258850
FILE: Number of read operations=0
FILE: Number of large read operations=0
FILE: Number of write operations=0
HDFS: Number of bytes read=2620
HDFS: Number of bytes written=215
HDFS: Number of read operations=43
HDFS: Number of large read operations=0
HDFS: Number of write operations=3
Job Counters
Launched map tasks=10
Launched reduce tasks=1
Data-local map tasks=10
Total time spent by all maps in occupied slots (ms)=22270
Total time spent by all reduces in occupied slots (ms)=2090
Total time spent by all map tasks (ms)=22270
Total time spent by all reduce tasks (ms)=2090
Total vcore-seconds taken by all map tasks=22270
Total vcore-seconds taken by all reduce tasks=2090
Total megabyte-seconds taken by all map tasks=
Total megabyte-seconds taken by all reduce tasks=2140160
Map-Reduce Framework
Map input records=10
Map output records=20
Map output bytes=180
Map output materialized bytes=340
Input split bytes=1440
Combine input records=0
Combine output records=0
Reduce input groups=2
Reduce shuffle bytes=340
Reduce input records=20
Reduce output records=0
Spilled Records=40
Shuffled Maps =10
Failed Shuffles=0
Merged Map outputs=10
GC time elapsed (ms)=290
CPU time spent (ms)=4000
Physical memory (bytes) snapshot=
Virtual memory (bytes) snapshot=
Total committed heap usage (bytes)=
Shuffle Errors
CONNECTION=0
IO_ERROR=0
WRONG_LENGTH=0
WRONG_MAP=0
WRONG_REDUCE=0
File Input Format Counters
Bytes Read=1180
File Output Format Counters
Bytes Written=97
Job Finished in 20.497 seconds
Estimated value of Pi is 3.
成功运算
http://192.168.1.207:8088/cluster
可以看到各个运行成功的job
--------------------------------其他过程----------------------------
1)apache的网页目录访问位置
[root@master01 cloudera_manager]# pwd
/usr/local/apache/htdocs/cloudera_manager
[root@master01 cloudera_manager]# cd /opt/cm
[root@master01 cm]# ll
total 663364
-rw-r--r--& 1 root root&& 4781732 Jun& 4 18:02 cloudera-manager-agent-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rw-r--r--. 1 root root
Jun& 4 17:00 cloudera-manager-daemons-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rw-r--r--& 1 root root&&&&& 8556 Jun& 4 18:02 cloudera-manager-server-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rw-r--r--& 1 root root&&&&& 9880 Jun& 4 18:02 cloudera-manager-server-db-2-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rw-r--r--& 1 root root&&& 980244 Jun& 4 18:02 enterprise-debuginfo-5.4.6-1.cm546.p0.8.el6.x86_64.rpm
-rw-r--r--& 1 root root&& 2868261 Jun& 6 09:21 jdk-6u31-linux-amd64.rpm
-rw-r--r--& 1 root root&& 1334340 Jun& 4 18:02 oracle-j2sdk1.7-1.7.0+update67-1.x86_64.rpm
drwxr-xr-x& 2 root root&&&&& 4096 Jun& 6 09:40 repodata
[root@master01 cm]# ln -s& /opt/cm& /usr/local/apache/htdocs/cloudera_manager
之前尝试使用自己创建的目录,让其下载
[root@master01 htdocs]# rmdir cloudera_manager/
[root@master01 htdocs]# ln -s cloudera_manager/ /usr/local/apache/htdocs/
[root@master01 htdocs]#
/usr/local/apache/bin/apachectl start
http://192.168.1.207/cm即可访问
ln -s& /opt/cm& /usr/local/apache/htdocs/cm& --包存在问题
[root@master01 cloudera-manager-installer]# tail -30f 1.install-oracle-j2sdk1.7.log
Loaded plugins: fastestmirror, refresh-packagekit, security
Error: File contains no section headers.
file: file://///etc/yum.repos.d/cloudera-manager.repo, line: 1
'\xe3\x80\x90cloudera-manager\xe3\x80\x91\n'
^C
2)缺少依赖包:
Warning: RPMDB altered outside of yum.
** Found 3 pre-existing rpmdb problem(s), 'yum check' output follows:
1:libreoffice-core-4.0.4.2-9.el6.x86_64 has missing requires of libjawt.so()(64bit)
1:libreoffice-core-4.0.4.2-9.el6.x86_64 has missing requires of libjawt.so(SUNWprivate_1.1)(64bit)
1:libreoffice-ure-4.0.4.2-9.el6.x86_64 has missing requires of jre &= ('0', '1.5.0', None)
yum install -y libreoffice-core libreoffice-ure
yum localinstall --nogpgcheck& --skip-broken *.rpm
3)有关parcel.sha文件:
/opt/cloudera/parcel-repo
[root@master01 cloudera]# cd parcel-repo/
[root@master01 parcel-repo]# ll
total 9192
-rw-r----- 1 cloudera-scm cloudera-scm 9363456 Jun& 6 11:40 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.part
-rw-r--r-- 1 root&&&&&&&& root&&&&&&&&&&&&& 41 Jun& 6 11:46 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha1
-rw-r--r-- 1 root&&&&&&&& root&&&&&&&&&& 43173 Jun& 6 11:37 manifest.json
[root@master01 parcel-repo]# cat CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha1
2b9effbceeaaf3291308d
[root@master01 parcel-repo]# cat manifest.json |grep "hash"
&&&&&&&&&&& "hash": "706f7a40e95c4b9fe789e332d147"
&&&&&&&&&&& "hash": "dcdd3ca2c2dc1"
&&&&&&&&&&& "hash": "7b9ba6c2d6e73fa3f4e"
&&&&&&&&&&& "hash": "75837fadc1b7771febb6baf4df722"
&&&&&&&&&&& "hash": "bf4e3e872ec"
&&&&&&&&&&& "hash": "2b9effbceeaaf3291308d"
[root@master01 opt]# ps -ef|grep cloudera
root&&&&& & 0 Jun05 ?&&&&&&& 00:11:03 /usr/lib64/cmf/agent/build/env/bin/python /usr/lib64/cmf/agent/src/cmf/agent.py --package_dir /usr/lib64/cmf/service --agent_dir /var/run/cloudera-scm-agent --lib_dir /var/lib/cloudera-scm-agent --logfile /var/log/cloudera-scm-agent/cloudera-scm-agent.log
root&&&&& & 0 Jun05 ?&&&&&&& 00:00:00 /usr/lib64/cmf/agent/build/env/bin/python /usr/lib64/cmf/agent/src/cmf/supervisor_listener.py -l /var/log/cloudera-scm-agent/cmf_listener.log /var/run/cloudera-scm-agent/events
root&&&&& & 0 13:05 pts/1&&& 00:00:00 grep cloudera
496&&&&& 12076&&&& 1& 0 11:15 ?&&&&&&& 00:00:00 /usr/bin/postgres -D /var/lib/cloudera-scm-server-db/data
root&&&& 12113&&&& 1& 0 11:15 pts/0&&& 00:00:00 su cloudera-scm -s /bin/bash -c nohup /usr/sbin/cmf-server
496&&&&& & 2 11:15 ?&&&&&&& 00:02:18 /usr/java/jdk1.7.0_67-cloudera/bin/java -cp .:lib/*:/usr/share/java/mysql-connector-java.jar:/usr/share/java/oracle-connector-java.jar -server -Dlog4j.configuration=file:/etc/cloudera-scm-server/log4j.properties -Dfile.encoding=UTF-8 -Dcmf.root.logger=INFO,LOGFILE -Dcmf.log.dir=/var/log/cloudera-scm-server -Dcmf.log.file=cloudera-scm-server.log -Dcmf.jetty.threshhold=WARN -Dcmf.schema.dir=/usr/share/cmf/schema -Djava.awt.headless=true -Djava.net.preferIPv4Stack=true -Dpython.home=/usr/share/cmf/python -XX:+UseConcMarkSweepGC -XX:-CMSConcurrentMTEnabled -XX:+UseParNewGC -XX:+HeapDumpOnOutOfMemoryError -Xmx2G -XX:MaxPermSize=256m -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp -XX:OnOutOfMemoryError=kill -9 %p com.cloudera.server.cmf.Main
4)内存设置太小
& HDFS: Namenode 的 Java 堆栈大小(字节)
Java Heap Size of Namenode in Bytes is recommended to be at least 1GB for every million HDFS blocks.
Suggested minimum value:
页面修改为1G即可
5)停止scm-agent服务
service cloudera-scm-agent hard_stop_confirmed
Decommissioning requires the ResourceManager roles of service YARN (MR2 Included) to be running.
&
6)查看mysql-connect-java的jar包的默认位置
[root@master01 cloudera-scm-agent]# vi config.ini
[root@master01 cloudera-scm-agent]# pwd
/opt/cm-5.4.6/etc/cloudera-scm-agent
[root@master01 cloudera-scm-agent]#
[JDBC]
#cloudera_mysql_connector_jar=/usr/share/java/mysql-connector-java.jar
#cloudera_oracle_connector_jar=/usr/share/java/oracle-connector-java.jar
====================
1:安装rpm包
把需要的安装包拷贝到&& /home/hadoop/app目录下
yum localinstall --nogpgcheck& --skip-broken *.rpm(安装rpm包)
2:执行./cloudera-manager-installer.bin,如果遇到问题修改一下即可
agent节点(slave2~5)
1:把下面的包分发到agent节点上 /opt/hadoop/app 目录下
cloudera-manager-agent-5.6.0-1.cm560.p0.54.el7.x86_64.rpm
cloudera-manager-daemons-5.6.0-1.cm560.p0.54.el7.x86_64.rpm
2:agent安装
yum localinstall --nogpgcheck& --skip-broken *.rpm (安装rpm包)
rm -f& /etc/cloudera-scm-server/db.properties
3:启动服务
Server
service cloudera-scm-server start
agent
service cloudera-scm-agent start
[root@master01 opt]# chmod +x cloudera-manager-installer-5.4.10.bin
4.安装时可查看日志:
[root@master01 log]# cd /var/log/cloudera-manager-installer/
[root@master01 cloudera-manager-installer]# ll
total 20
-rw-r--r-- 1 root root&&& 0 Jun& 7 11:57 0.check-selinux.log
-rw-r--r-- 1 root root& 229 Jun& 7 11:57 1.install-oracle-j2sdk1.7.log
-rw-r--r-- 1 root root& 243 Jun& 7 11:57 2.install-cloudera-manager-server.log
-rw-r--r-- 1 root root& 248 Jun& 7 11:57 3.install-cloudera-manager-server-db-2.log
-rw-r--r-- 1 root root 1975 Jun& 7 11:58 4.start-embedded-db.log
-rw-r--r-- 1 root root&& 59 Jun& 7 11:58 5.start-scm-server.log
[root@master01 cloudera-manager-installer]#
5.跳过repo创建
[root@master01 opt]# ./cloudera-manager-install.bin& --skip_repo_package=1
[root@master01 opt]# cat /etc/yum.repos.d/cloudera-manager.repo&& --尝试使用自己创建的repo来,但好像不行
[cloudera-manager]
name=Cloudera Manager
#gpgcheck=1
baseurl=http://192.168.1.207/cm
enabled=1
gpgcheck=0
[root@master01 opt]#
还是可以直接用./cloudera-manager-install.bin
[root@master01 yum.repos.d]# ll
total 24
-rw-r--r--. 1 root root 1926 Nov 27& 2013 CentOS-Base.repo.bk
-rw-r--r--. 1 root root& 638 Nov 27& 2013 CentOS-Debuginfo.repo.bk
-rw-r--r--. 1 root root& 630 Nov 27& 2013 CentOS-Media.repo.bk
-rw-r--r--. 1 root root 3664 Nov 27& 2013 CentOS-Vault.repo
-rw-r--r--& 1 root root& 106 Jun& 7 11:56 cloudera-manager.repo
-rw-r--r--& 1 root root& 195 Sep 18& 2014 cloudera-manager.repo.rpmnew
[root@master01 yum.repos.d]# cat cloudera-manager.repo.rpmnew&& ---这个路径不对,会直接下载5.7.1最新版,但这是它自动生成的,有时候能找对
[cloudera-manager]
name=Cloudera Manager
baseurl=/cm5/redhat/6/x86_64/cm/5/
gpgkey = /cm5/redhat/6/x86_64/cm/RPM-GPG-KEY-cloudera
gpgcheck=1
[root@master01 yum.repos.d]#
[root@master01 yum.repos.d]#
6.找不到db.properties
[root@master01 opt]# /usr/share/cmf/schema/scm_prepare_database.sh mysql& -uroot -p --scm-host localhost scm scm scm123& --自己创建的db.properties
Enter database password:
JAVA_HOME=/usr/java/jdk1.7.0
Verifying that we can write to /etc/cloudera-scm-server
Creating SCM configuration file in /etc/cloudera-scm-server
Executing:& /usr/java/jdk1.7.0/bin/java -cp /usr/share/java/mysql-connector-java.jar:/usr/share/java/oracle-connector-java.jar:/usr/share/cmf/schema/../lib/* com.cloudera.enterprise.dbutil.DbCommandExecutor /etc/cloudera-scm-server/db.properties com.cloudera.cmf.db.
[&&&&&&&&&&&&&&&&&&&&&&&&& main] DbCommandExecutor&&&&&&&&&&&&& INFO& Successfully connected to database.
All done, your SCM database is configured correctly!
[root@master01 opt]#
可以在启动后,赶紧把这个拷贝进去就可以了!! mv db.properties& /etc/cloudera-manager-server/
7.卸载cloudera manager
如果要卸载的话,记得要删掉下面的目录:
然后删除/var/lib/cloudera-scm-server-db/目录,不然下次安装可能不成功。
rm -fr /var/lib/cloudera-scm-server-db/
rm -f /etc/cloudera-scm-server/db.properties
rm -fr /etc/cloudera-scm-server/*
rm -fr /var/run/cloudra-manager-server.pid&& --这个不删除很可能会网页打不开,service起不来
yum remove -y 'cloudera-manager-*' hadoop hue-common 'bigtop-*'
yum localinstall --nogpgcheck& --skip-broken *.rpm
8.确认7180端口可用
[root@master01 bin]# ./apachectl start&&& --起来后,要稍微等一会才行
httpd (pid 3473) already running
[root@master01 bin]#& service cloudera-scm-server start
cloudera-scm-server is already running
[root@master01 bin]#& service cloudera-scm-server-db start
Database is already running. Please stop it first., giving up
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
[root@master01 bin]# netstat -nat|grep 7180
tcp&&&&&&& 0&&&&& 0 0.0.0.0:7180&&&&&&&&&&&&&&& 0.0.0.0:*&&&&&&&&&&&&&&&&&& LISTEN&&&&&
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55213&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55212&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55216&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55217&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55214&&&&&&&&& ESTABLISHED
tcp&&&&&&& 0&&&&& 0 192.168.1.207:7180&&&&&&&&& 192.168.1.41:55215&&&&&&&&& ESTABLISHED
9.parcel-repo路径
[root@master01 parcel-repo]# ll
total 1233984
-rwxrwxrwx 1 root root
Jun& 7 11:03 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel
-rwxrwxrwx 1 root root&&&&&&&& 41 Jun& 7 11:00 CDH-5.4.10-1.cdh5.4.10.p0.16-el6.parcel.sha& ---改了属主后能读到了,但也不清楚是不是以为这个,反正很纠结
-rwxrwxrwx 1 root root&&&&& 43172 Jun& 7 12:59 manifest.json
10.scm-server启动报错
cloudera-scm-server dead but pid file exists
14:35:34
分类: Hadoop
1.查看内存是否占用过高
2.查看数据库是否能通
直接删除这个pid即可
rm -f /var/run/cloudera-scm-server.pid
11.apt-get方式安装jdk
apt-get -o Dpkg::Options::=--force-confdef -o Dpkg::Options::=--force-confold -y install oracle-j2sdk1.7
网上说jdk的默认安装路径为:/usr/lib/jvm/java-7-oracle-cloudera
12.远程parcel存储库url
/cdh5/parcels/{latest_supported}/
/cdh4/parcels/latest/
/impala/parcels/latest/
/search/parcels/latest/
/accumulo-c5/parcels/latest/
/spark/parcels/latest/
/accumulo/parcels/1.4/
/sqoop-connectors/parcels/latest/
/navigator-keytrustee5/parcels/latest/
/kafka/parcels/latest/
13.查看整个安装过程的日志:
[root@master01 cloudera-scm-server]# pwd
/var/log/cloudera-scm-server
[root@master01 cloudera-scm-server]# tail -100f cloudera-scm-server.log
14.反复安装,导致格式化失败
/hadoop/dfs/nn目录非空,所以没有执行hdfs格式化,因此我们需要改一下
Running in non-interactive mode, and data appears to exist in Storage Directory /hadoop/dfs/nn. Not formatting.
16/06/07 14:59:24 INFO util.ExitUtil: Exiting with status 1
16/06/07 14:59:24 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at master02/192.168.1.208
************************************************************/
清空/hadoop/dfs/nn
15.初始化mysql
[root@master02 mysql]# bin/mysqld --initialize --user=mysql --basedir=/usr/local/mysql --datadir=/usr/local/mysql/data
T07:14:54. [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
T07:14:55. [Warning] InnoDB: New log files created, LSN=45790
T07:14:55. [Warning] InnoDB: Creating foreign key constraint system tables.
T07:14:55. [Warning] No existing UUID has been found, so we assume that this is the first time that this server has been started. Generating a new UUID: c7f-11e6-a20f-f.
T07:14:55. [Warning] Gtid table is not ready to be used. Table 'mysql.gtid_executed' cannot be opened.
T07:14:55. [Note] A temporary password is generated for root@localhost: ?KRu6g3&jkZY
16.hive创建表失败,权限不足
[root@datanode03 local]# find / -name "hive-site.xml"
/var/run/cloudera-scm-agent/process/37-hive-metastore-create-tables/hive-site.xml
/var/run/cloudera-scm-agent/process/35-hive-metastore-create-tables/hive-site.xml
/var/run/cloudera-scm-agent/process/33-hive-metastore-create-tables/hive-site.xml
/var/run/cloudera-scm-agent/process/31-hive-metastore-create-tables/hive-site.xml
/var/run/cloudera-scm-agent/process/ccdeploy_hive-conf_etchiveconf.cloudera.hive_4822279/hive-conf/hive-site.xml
/etc/hive/conf.cloudera.hive/hive-site.xml
/opt/cloudera/parcels/CDH-5.4.10-1.cdh5.4.10.p0.16/etc/hive/conf.dist/hive-site.xml
[root@datanode03 local]# cd /var/run/cloudera-scm-agent
[root@datanode03 cloudera-scm-agent]# ll
total 8
drwxr-x--x.& 6 root root 4096 Jun& 7 11:06 cgroups
prw-------.& 1 root root&&& 0 Jun& 7 15:24 events
drwxr-x--x. 15 root root& 300 Jun& 7 15:24 process
drwxr-x--x.& 3 root root 4096 Jun& 7 13:02 supervisor
[root@datanode03 cloudera-scm-agent]# cd process/
[root@datanode03 process]# ll
total 0
drwxr-x--x. 3 hdfs& hdfs&& 340 Jun& 7 15:04 11-hdfs-DATANODE
drwxr-x--x. 3 hbase hbase& 360 Jun& 7 15:04 20-hbase-MASTER
drwxr-x--x. 3 hbase hbase& 360 Jun& 7 15:04 23-hbase-REGIONSERVER
drwxr-x--x. 3 yarn& hadoop 420 Jun& 7 15:05 28-yarn-NODEMANAGER
drwxr-x--x. 4 hive& hive&& 280 Jun& 7 15:17 31-hive-metastore-create-tables
drwxr-x--x. 4 hive& hive&& 280 Jun& 7 15:23 33-hive-metastore-create-tables
drwxr-x--x. 4 hive& hive&& 280 Jun& 7 15:24 35-hive-metastore-create-tables
drwxr-x--x. 4 hive& hive&& 260 Jun& 7 15:24 37-hive-metastore-create-tables
drwxr-x--x. 3 root& root&& 100 Jun& 7 14:32 3-cluster-host-inspector
drwxr-xr-x. 4 root& root&& 100 Jun& 7 14:58 ccdeploy_hadoop-conf_etchadoopconf.cloudera.hdfs_8952504
drwxr-xr-x. 4 root& root&& 100 Jun& 7 14:58 ccdeploy_hadoop-conf_etchadoopconf.cloudera.yarn_4652443
drwxr-xr-x. 4 root& root&& 100 Jun& 7 14:58 ccdeploy_hbase-conf_etchbaseconf.cloudera.hbase_6499928
drwxr-xr-x. 4 root& root&& 100 Jun& 7 14:58 ccdeploy_hive-conf_etchiveconf.cloudera.hive_4822279
mysql& create user 'hive'@'localhost' identified by '';
Query OK, 0 rows affected (0.00 sec)
mysql& grant all on *.* to 'hive'@'localhost';
Query OK, 0 rows affected (0.00 sec)
mysql&
Query OK, 0 rows affected (0.00 sec)
create user 'hive'@'localhost' identified by '';
grant all on *.* to 'hive'@'localhost' identified by '';
Tue Jun& 7 15:56:52 CST 2016
JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera
using /usr/java/jdk1.7.0_67-cloudera as JAVA_HOME
using 5 as CDH_VERSION
using /opt/cloudera/parcels/CDH-5.4.10-1.cdh5.4.10.p0.16/lib/hive as HIVE_HOME
using /var/run/cloudera-scm-agent/process/47-hive-metastore-create-tables as HIVE_CONF_DIR
using /opt/cloudera/parcels/CDH-5.4.10-1.cdh5.4.10.p0.16/lib/hadoop as HADOOP_HOME
using /var/run/cloudera-scm-agent/process/47-hive-metastore-create-tables/yarn-conf as HADOOP_CONF_DIR
Metastore connection URL:
jdbc:mysql://localhost:3306/metastore?useUnicode=true&characterEncoding=UTF-8
Metastore Connection Driver :
com.mysql.jdbc.Driver
Metastore connection User:
++ exec /opt/cloudera/parcels/CDH-5.4.10-1.cdh5.4.10.p0.16/lib/hadoop/bin/hadoop jar /opt/cloudera/parcels/CDH-5.4.10-1.cdh5.4.10.p0.16/lib/hive/lib/hive-cli-1.1.0-cdh5.4.10.jar org.apache.hive.beeline.HiveSchemaTool -verbose -dbType mysql -initSchema
org.apache.hadoop.hive.metastore.HiveMetaException: Failed to get schema version.
org.apache.hadoop.hive.metastore.HiveMetaException: Failed to get schema version.
at org.apache.hive.beeline.HiveSchemaHelper.getConnectionToMetastore(HiveSchemaHelper.java:77)
at org.apache.hive.beeline.HiveSchemaTool.getConnectionToMetastore(HiveSchemaTool.java:113)
at org.apache.hive.beeline.HiveSchemaTool.testConnectionToMetastore(HiveSchemaTool.java:159)
at org.apache.hive.beeline.HiveSchemaTool.doInit(HiveSchemaTool.java:257)
at org.apache.hive.beeline.HiveSchemaTool.doInit(HiveSchemaTool.java:243)
at org.apache.hive.beeline.HiveSchemaTool.main(HiveSchemaTool.java:473)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
Caused by: java.sql.SQLException: Access denied for user 'hive'@'localhost' (using password: YES)& --加了权限,试了好多次依然失败,最后重装了
at com.mysql.jdbc.SQLError.createSQLException(SQLError.java:1084)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4232)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:4164)
at com.mysql.jdbc.MysqlIO.checkErrorPacket(MysqlIO.java:926)
at com.mysql.jdbc.MysqlIO.proceedHandshakeWithPluggableAuthentication(MysqlIO.java:1748)
at com.mysql.jdbc.MysqlIO.doHandshake(MysqlIO.java:1288)
at com.mysql.jdbc.ConnectionImpl.coreConnect(ConnectionImpl.java:2506)
at com.mysql.jdbc.ConnectionImpl.connectOneTryOnly(ConnectionImpl.java:2539)
at com.mysql.jdbc.ConnectionImpl.createNewIO(ConnectionImpl.java:2321)
at com.mysql.jdbc.ConnectionImpl.&init&(ConnectionImpl.java:832)
at com.mysql.jdbc.JDBC4Connection.&init&(JDBC4Connection.java:46)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at com.mysql.jdbc.Util.handleNewInstance(Util.java:409)
at com.mysql.jdbc.ConnectionImpl.getInstance(ConnectionImpl.java:417)
at com.mysql.jdbc.NonRegisteringDriver.connect(NonRegisteringDriver.java:344)
at java.sql.DriverManager.getConnection(DriverManager.java:571)
at java.sql.DriverManager.getConnection(DriverManager.java:215)
at org.apache.hive.beeline.HiveSchemaHelper.getConnectionToMetastore(HiveSchemaHelper.java:73)
... 11 more
*** schemaTool failed ***
[root@master01 cloudera-manager-installer]# tail -30 5.start-scm-server.log
Starting cloudera-scm-server:&&&&&&&&&&&&&&&&&&&&&&&&&&&&& [FAILED]
17.安装agent报错:
&&[08/Jun/:00 + MainThread agent INFO Re-using pre-existing directory: /var/run/cloudera-scm-agent/process
&&[08/Jun/:00 + MainThread agent INFO Re-using pre-existing directory: /var/run/cloudera-scm-agent/supervisor
&&[08/Jun/:00 + MainThread agent INFO Re-using pre-existing directory: /var/run/cloudera-scm-agent/supervisor/include
&&[08/Jun/:00 + MainThread agent ERROR Failed to connect to previous supervisor.
提示:
安装失败。 无法接收 Agent 发出的检测信号。
请确保主机的名称已正确配置。
请确保端口 7182 可在 Cloudera Manager Server 上访问(检查防火墙规则)。
请确保正在添加的主机上的端口 9000 和 9001 空闲。
检查正在添加的主机上 /var/log/cloudera-scm-agent/ 中的代理日志(某些日志可在安装详细信息中找到)。
可以看到默认的JAVA_HOME的可选路径:
++ JAVA7_HOME_CANDIDATES=('/usr/java/jdk1.7' '/usr/java/jre1.7' '/usr/lib/jvm/j2sdk1.7-oracle' '/usr/lib/jvm/j2sdk1.7-oracle/jre' '/usr/lib/jvm/java-7-oracle')
18.报错:
java.io.IOException: All specified directories are failed to load.
at org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:473)
at org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1322)
at org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1292)
at org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:320)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:225)
at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:862)
at java.lang.Thread.run(Thread.java:745)
13:45:08,290 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Ending block pool service for: Block pool &registering& (Datanode Uuid unassigned) service to scm./103.231.66.62:8022
13:45:08,391 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Removed Block pool &registering& (Datanode Uuid unassigned)
13:45:10,392 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Exiting Datanode
13:45:10,393 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 0
13:45:10,395 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at /192.168.23.45
************************************************************/
解析:
以上红色字体部分已经说明, /data0/dfs/nn 目录初始化失败。其原因是应为& /data0/dfs/nn 是hdfs的 dfs.namenode.name.dir 目录,
而这里secondary namenode 节点已经启动了,这就说明/data0/dfs/nn 被当成secondary namenode的 检查点目录(fs.checkpoint.dir, dfs.namenode.checkpoint.di)
占用了。 而对于hdfs集群目录是不能和检查目录共用,所以节点1启动失败。
解决方法:
修改secondary namenode 的检查点目录,并重启hdfs集群就可以正常了。 --这是网上方法,但我们实际的问题是下面的
上面的问题其实是因为master和datanode之间的clusterid不一致导致的。
[root@master01 current]# cat VERSION& --master节点
#Wed Jun 08 10:15:25 CST 2016
namespaceID=
clusterID=cluster36&& --只需要这个保持一致
cTime=0
storageType=NAME_NODE
blockpoolID=BP--192.168.1.207-8
layoutVersion=-60
[root@master01 current]# pwd
/hadoop/dfs/nn/current
[root@master01 current]#
[root@datanode01 java]# cd /hadoop/dfs/dn/current/& ---datanode节点
[root@datanode01 current]# cat VERSION
#Wed Jun 08 11:01:15 CST 2016
storageID=DS-2f2d6cb1-2f78-4005-bca8-aca
clusterID=cluster18&& --改成跟上面一样的cluster36即可
cTime=0
datanodeUuid=1c1ed6c9-283b--c258af43ec95
storageType=DATA_NODE
layoutVersion=-56
[root@datanode01 current]#
18.zookeeper添加实例报错:
Exception in thread "main" org.apache.hadoop.HadoopIllegalArgumentException: HA is not enabled for this namenode.
at org.apache.hadoop.hdfs.tools.DFSZKFailoverController.create(DFSZKFailoverController.java:121)
at org.apache.hadoop.hdfs.tools.DFSZKFailoverController.main(DFSZKFailoverController.java:177)
当有一台机器正在跑zookeeper的时候,再添加其他的,就会报错如下:
Starting these new ZooKeeper Servers may cause the existing ZooKeeper Datastore to be lost. Try again after restarting any
existing ZooKeeper Servers with outdated configurations. If you do not want to preserve the existing Datastore, you can start
each ZooKeeper Server from its respective Status page.
将这个停止,然后再三台一起启动即可。
由于是装完后,才靠着记忆写下来的,因此可能有一些不对的地方,见谅
学习中,欢迎交流
qq:
tina
LJ你是唯一LT
浏览: 28473 次}

我要回帖

更多关于 cloudera impala 下载 的文章

更多推荐

版权声明:文章内容来源于网络,版权归原作者所有,如有侵权请点击这里与我们联系,我们将及时删除。

点击添加站长微信