Trackers not working


10 Answer(s)


hi Rajesh,

Looks like you've not set the proper permissions to the folders, hence you are getting "permission denied" . Check the logs of Namenode and mapreduce logs to see exact errors.

Datanode:

2015-10-06 22:02:39,327 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2015-10-06 22:02:39,333 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered. 2015-10-06 22:02:39,334 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2015-10-06 22:02:39,334 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2015-10-06 22:02:39,369 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,369 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,402 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,402 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,406 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi registered. 2015-10-06 22:02:39,409 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already exists! 2015-10-06 22:02:39,526 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,526 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:39,604 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Registered FSDatasetStatusMBean 2015-10-06 22:02:39,612 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened data transfer server at 50010 2015-10-06 22:02:39,614 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith is 1048576 bytes/s 2015-10-06 22:02:39,618 INFO org.apache.hadoop.util.NativeCodeLoader: Loaded the native-hadoop library 2015-10-06 22:02:40,208 INFO org.mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2015-10-06 22:02:40,244 INFO org.apache.hadoop.http.HttpServer: Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter) 2015-10-06 22:02:40,256 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dfs.webhdfs.enabled = false 2015-10-06 22:02:40,256 INFO org.apache.hadoop.http.HttpServer: Port returned by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the listener on 50075 2015-10-06 22:02:40,256 INFO org.apache.hadoop.http.HttpServer: listener.getLocalPort() returned 50075 webServer.getConnectors()[0].getLocalPort() returned 50075 2015-10-06 22:02:40,256 INFO org.apache.hadoop.http.HttpServer: Jetty bound to port 50075 2015-10-06 22:02:40,256 INFO org.mortbay.log: jetty-6.1.26 2015-10-06 22:02:40,445 INFO org.mortbay.log: Started SelectChannelConnector@0.0.0.0:50075 2015-10-06 22:02:40,449 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source jvm registered. 2015-10-06 22:02:40,450 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source DataNode registered. 2015-10-06 22:02:40,587 INFO org.apache.hadoop.ipc.Server: Starting SocketReader 2015-10-06 22:02:40,589 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source RpcDetailedActivityForPort50020 registered. 2015-10-06 22:02:40,589 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source RpcActivityForPort50020 registered. 2015-10-06 22:02:40,590 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration = DatanodeRegistration(localhost:50010, storageID=DS-749473128-127.0.0.1-50010-1444180521010, infoPort=50075, ipcPort=50020) 2015-10-06 22:02:40,591 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Finished generating blocks being written report for 1 volumes in 0 seconds 2015-10-06 22:02:40,593 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(127.0.0.1:50010, storageID=DS-749473128-127.0.0.1-50010-1444180521010, infoPort=50075, ipcPort=50020)In DataNode.run, data = FSDataset{dirpath='/tmp/hadoop-acrajesh/dfs/data/current'} 2015-10-06 22:02:40,593 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Finished asynchronous block report scan in 0ms 2015-10-06 22:02:40,593 INFO org.apache.hadoop.ipc.Server: IPC Server Responder: starting 2015-10-06 22:02:40,593 INFO org.apache.hadoop.ipc.Server: IPC Server listener on 50020: starting 2015-10-06 22:02:40,594 INFO org.apache.hadoop.ipc.Server: IPC Server handler 1 on 50020: starting 2015-10-06 22:02:40,594 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 50020: starting 2015-10-06 22:02:40,595 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: using BLOCKREPORT_INTERVAL of 3600000msec Initial delay: 0msec 2015-10-06 22:02:40,595 INFO org.apache.hadoop.ipc.Server: IPC Server handler 2 on 50020: starting 2015-10-06 22:02:40,596 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks took 0 msec to generate and 1 msecs for RPC and NN processing 2015-10-06 22:02:40,596 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic block scanner 2015-10-06 22:02:40,597 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Generated rough (lockless) block report in 0 ms

Jobracker:

Jul 22 15:23:09 PDT 2013 STARTUP_MSG: java = 1.8.0_60 ************************************************************/ 2015-10-06 22:02:45,923 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:45,923 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:45,987 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2015-10-06 22:02:45,995 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered. 2015-10-06 22:02:45,996 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2015-10-06 22:02:45,996 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: JobTracker metrics system started 2015-10-06 22:02:46,014 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:46,014 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:46,032 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:46,032 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:46,041 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source QueueMetrics,q=default registered. 2015-10-06 22:02:46,263 FATAL org.apache.hadoop.mapred.JobTracker: java.lang.IllegalArgumentException: Does not contain a valid host:port authority: local at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2131) at org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1869) at org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1689) at org.apache.hadoop.mapred.JobTracker.(JobTracker.java:1683) at org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:320) at org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:311) at org.apache.hadoop.mapred.JobTracker.startTracker(JobTracker.java:306) at org.apache.hadoop.mapred.JobTracker.main(JobTracker.java:4710) 2015-10-06 22:02:46,263 INFO org.apache.hadoop.mapred.JobTracker: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down JobTracker at localhost.localdomain/127.0.0.1 ************************************************************/

Task tracker:

*****/ 2015-10-06 22:02:52,105 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,105 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,160 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2015-10-06 22:02:52,210 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered. 2015-10-06 22:02:52,211 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2015-10-06 22:02:52,211 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics system started 2015-10-06 22:02:52,402 INFO org.apache.hadoop.util.NativeCodeLoader: Loaded the native-hadoop library 2015-10-06 22:02:52,433 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,433 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,437 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi registered. 2015-10-06 22:02:52,440 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already exists! 2015-10-06 22:02:52,485 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,485 WARN org.apache.hadoop.conf.Configuration: bad conf file: element not 2015-10-06 22:02:52,490 ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task tracker because java.lang.IllegalArgumentException: Does not contain a valid host:port authority: local at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2131) at org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1540) at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3937) 2015-10-06 22:02:52,490 INFO org.apache.hadoop.mapred.TaskTracker: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down TaskTracker at localhost.localdomain/127.0.0.1 ************************************************************/


Hey Rajesh,
I was also getting same error for the node which we added as part of upgrade and for which we didn't copied the authorization_Keys from NameNode to newly added node.
That was demonstration of production node where they may not shutdown the clusters but here everyday we do the Shutdown of VMs and looks like once we are doing the start NameNode is not getting authororized_Keys for that new Node added.

What I did is I copied the keys from NameNode server to that node using below command:

ssh-copy-id -i /home/hadoop/.ssh/id_rsa.pub hadoop@

It is working for me but I am not sure this shud be real solution for now it can be treated as Workaround.

Thanks!
Surendra

sorry command was not full, please replace the Ip from your node's IP

ssh-copy-id -i /home/hadoop/.ssh/id_rsa.pub hadoop@192.168.118.134

surendra
thanks, ur solution helpd in fixin the passwordless login issue but the original issue
remains

i am running hadoop 1..2.1 on fedora linux
I dont use VPN. Have only linux on my machine


Error message: .........Configuration: bad conf file: element not 2015-10-07 00:19:17,324 INFO ...................Can not start task tracker because java.lang.IllegalArgumentException: Does not contain a valid host:port authority: .........

[acrajesh@localhost hadoop-1.2.1]$ bin/start-mapred.sh
starting jobtracker, logging to /usr/lib/hadoop/hadoop-1.2.1/logs/hadoop-acrajesh-jobtracker-localhost.localdomain.out
localhost: starting tasktracker, logging to /usr/lib/hadoop/hadoop-1.2.1/logs/hadoop-acrajesh-tasktracker-localhost.localdomain.out
[acrajesh@localhost hadoop-1.2.1]$ bin/hadoop tasktracker
15/10/07 00:35:33 INFO mapred.TaskTracker: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting TaskTracker
STARTUP_MSG: host = localhost.localdomain/127.0.0.1
STARTUP_MSG: args = []
STARTUP_MSG: version = 1.2.1
STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 -r 1503152; compiled by 'mattf' on Mon Jul 22 15:23:09 PDT 2013
STARTUP_MSG: java = 1.8.0_60
************************************************************/
15/10/07 00:35:33 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:33 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:33 INFO impl.MetricsConfig: loaded properties from hadoop-metrics2.properties
15/10/07 00:35:33 INFO impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered.
15/10/07 00:35:33 INFO impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s).
15/10/07 00:35:33 INFO impl.MetricsSystemImpl: TaskTracker metrics system started
15/10/07 00:35:34 INFO util.NativeCodeLoader: Loaded the native-hadoop library
15/10/07 00:35:34 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:34 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:34 INFO impl.MetricsSourceAdapter: MBean for source ugi registered.
15/10/07 00:35:34 WARN impl.MetricsSystemImpl: Source name ugi already exists!
15/10/07 00:35:34 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:34 WARN conf.Configuration: bad conf file: element not
15/10/07 00:35:34 ERROR mapred.TaskTracker: Can not start task tracker because java.lang.IllegalArgumentException: Does not contain a valid host:port authority: local
at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130)
at org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2131)
at org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:1540)
at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3937)

15/10/07 00:35:34 INFO mapred.TaskTracker: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down TaskTracker at localhost.localdomain/127.0.0.1
************************************************************/

It would be helpful if you provide the hdfs-site.xml, mapred-site.xml and core-site.xml file details. Error is because of these configuration file. Please check whether these files are configured properly.

Working good now. I had not reproduced the config file contents properly. Thanks all

Rajesh,
Thanks for the update.