zoukankan      html  css  js  c++  java
  • datanode报错Problem connecting to server

    从节点datanode启动成功,但是后台查看发现未连接namenode,查看日志提示如下:

    2019-02-26 13:53:16,307 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Problem connecting to server: server-21/192.168.0.21:9000
    2019-02-26 13:53:22,310 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
    2019-02-26 13:53:23,312 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 1 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
    2019-02-26 13:53:24,315 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 2 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
    2019-02-26 13:53:25,318 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 3 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
    2019-02-26 13:53:26,321 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 4 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
    2019-02-26 13:53:27,323 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: server-21/192.168.0.21:9000. Already tried 5 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

    解决方式一
    删除hadoop目录下数据,dfs,log,tmp
    重新格式化namenode,重启start-dfs.sh 试试

    删除hadoop目录下数据,dfs,log,tmp
    [root@server-22 hadoop-2.7.7]# ll
    total 112
    drwxr-xr-x 2 root root 194 Mar 23 2019 bin
    drwxr-xr-x 4 root root 30 Mar 23 2019 dfs
    drwxr-xr-x 3 root root 20 Mar 23 2019 etc
    drwxr-xr-x 2 root root 106 Mar 23 2019 include
    drwxr-xr-x 3 root root 20 Mar 23 2019 lib
    drwxr-xr-x 2 root root 239 Mar 23 2019 libexec
    -rw-r–r-- 1 root root 86424 Feb 26 13:36 LICENSE.txt
    drwxr-xr-x 2 root root 165 Feb 26 13:53 logs
    -rw-r–r-- 1 root root 14978 Feb 26 13:36 NOTICE.txt
    -rw-r–r-- 1 root root 1366 Feb 26 13:36 README.txt
    drwxr-xr-x 2 root root 4096 Mar 23 2019 sbin
    drwxr-xr-x 4 root root 31 Mar 23 2019 share
    drwxr-xr-x 3 root root 26 Feb 26 13:26 tmp
    [root@server-22 hadoop-2.7.7]#

    解决方式二
    查看主节点master监听的端口
    如下:nameNode监听的是本机的ip,也就是说其他IP不监听

    [root@server-21 ~]# netstat -an | grep 9000
    tcp 0 0 127.0.0.1:9000 0.0.0.0:* LISTEN
    tcp 0 0 127.0.0.1:9000 127.0.0.1:52430 ESTABLISHED
    tcp 0 0 127.0.0.1:52430 127.0.0.1:9000 ESTABLISHED

    修改/etc/hosts

    [root@server-21 ~]# cat /etc/hosts
    #127.0.0.1 localhost.localdomain localhost4 localhost4.localdomain4 server-21 qjw-01
    #::1 qjw-01 localhost localhost.localdomain localhost6 localhost6.localdomain6 server-21
    192.168.0.22 server-22
    192.168.0.21 server-21 qjw-01

    立即生效hosts

    [root@server-21 ~]# su
    [root@server-21 ~]# /etc/init.d/network restart

    启动hadoop,start-dfs.sh
    查看监听端口:如下,正常

    [root@server-21 ~]# netstat -an | grep 9000
    tcp 0 0 192.168.0.21:9000 0.0.0.0:* LISTEN
    tcp 0 0 192.168.0.21:48224 192.168.0.21:9000 TIME_WAIT
    tcp 0 0 192.168.0.21:9000 192.168.0.22:45798 ESTABLISHED
    tcp 0 0 192.168.0.21:9000 192.168.0.21:48196 ESTABLISHED
    tcp 0 0 192.168.0.21:48196 192.168.0.21:9000 ESTABLISHED
    [root@server-21 ~]#

    查看网页显示一切正常

    转载于:https://blog.csdn.net/bb23417274/article/details/87933725

  • 相关阅读:
    Remote Desktop manager 连接后无法自动登录
    .net framework 源码调试 与 问题解决
    SQLSever--基础学习--创建登录用户&创建数据库用户&分配权限
    sql server 2008 数据库管理系统使用SQL语句创建登录用户步骤详解
    python 脚本备份mssql数据库并删除数据库
    python 脚本bak文件还原mssql数据库
    python3 爬虫利用Requests 实现下载进度条
    MySQL的四种事务隔离级别
    查看mysql的版本号
    haproxy 配置文件详解 之 backend
  • 原文地址:https://www.cnblogs.com/it-deepinmind/p/14343626.html
Copyright © 2011-2022 走看看