hadoop,cloudera,cloudera-managerRelated issues-Collection of common programming errors


  • cs_newbie
    java hadoop startup fatal-error
    i am starting my hadoop with 4 slaves and all works fine except for one machine. I have created them the exact same way.the error i receive when running ./start-all.sh is:xxxxx: starting tasktracker, logging to /xxxxx/xxxxx/hadoop/logs/hadoop-xxxxx-tasktracker-xxxxx.out xxxxx: /xxxxx/xxxxx/hadoop/hadoop-0.20/bin/hadoop: line 413: 7012 Aborted nohup $_JAVA_EXEC -Dproc_$COMMAND $JAVA_HEAP_MAX $HADOOP_OPTS -classpath “$CLASSPATH” $CLASS “$@” >”$_HADOOP_DAEMON_OUT” 2>&1 &l

  • j0k
    hadoop
    i am getting the following log on my namenode and its removing my datanode from execution2013-02-08 03:25:54,345 WARN namenode.NameNode (NameNodeRpcServer.java:errorReport(825)) – Fatal disk error on xxx.xxx.xxx.xxx:50010: DataNode failed volumes:/home/srikmvm/hadoop-0.23.0/tmp/current; 2013-02-08 03:25:54,349 INFO net.NetworkTopology (NetworkTopology.java:remove(367)) – Removing a node: /default-rack/xxx.xxx.xxx.xxx:50010Can anyone suggest how to rectify this ?Data Node Logs:2013-02-08 03:25:

  • ender
    hadoop cloudera sqoop
    When using the –incremental append flag in the sqoop import, the job will fail.ERROR tool.ImportTool: Imported Failed: Wrong FS: s3n://<api_key>:<api_secret>@bucket/folder/Here is the full command:sqoop import –connect jdbc:postgresql://someplace.net:12345/db –warehouse-dir s3n://<key>:<private>@bucket/folder/ –table my_table –hive-drop-import-delims –escaped-by “\\” –username root –password safepass -m 2 –split-by id –incremental append –check-column idThe ex

  • Nannie
    java hadoop cloudera oozie hue
    I’m new to Hadoop and I have the flowing problem:I keep getting a “java.lang.ClassNotFoundException” when I’m trying to run my oozzie work flow. I use the Cloudera quick start VM v 4.5. used oozzie settings:Jar name : sample.jar Main class : Driver Arguments : in_single/ outused java classpublic class Driver{ public static class TokenizerMapper extends Mapper<Object, Text, Text, Text> {@Overridepublic void map(final Object key, final Text value, final Mapper<Object, Text, Text, Text>

  • Alina
    java spring spring-mvc hadoop hive
    My application is combination of hadoop and rest service with spring framework. My aim is to provide results from hive table on request. But when I run application, after completion of mapreduce and hive job I get following error:java.lang.reflect.InvocationTargetExceptionat sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:

  • Peter Lawrey
    java linux hadoop mapreduce mapper
    Whether help to define mapper was executed and if it wasn’t executed, for what reason it could occur. I wrote an output of the read ways from a database to the text file of local file system on which mapper is executed. Here I give a codepackage org.myorg;import java.io.*; import java.util.*; import java.sql.Connection; import java.sql.DriverManager; import java.sql.ResultSet; import java.sql.SQLException; import java.sql.Statement; import java.util.logging.Level; import org.apache.hadoop.fs.*;

  • vpap
    hadoop streaming chaining
    This is a documentation on how to chain two or more streaming jobs, using Hadoop Streaming( currently 1.0.3) only and nothing more.In order to understand the final code that will do the chaining and be able to write any other chain job some preliminary but practical theory is required.First of all, what is a job in Hadoop? A Hadoop job is hadoopJob = Configuration + Executionwhere,Configuration : all the set up that makes Execution possible.Execution : the set of executable or script files t

  • ykesh
    hadoop apache-pig
    I installed hadoop-2.2.0 and can run MR jobs. Configured pig0.12 and trying to use the interactie grunt shell. But when I try to create data into records from input usingrecords = LOAD ‘/temp/temp.txt’ AS (year:chararray, temperature:int, quality:int);I get the following. Did not see this when I used pig0.12 earlier with hadoop-1.2.1 distribution.2013-11-27 11:11:37,225 [main] INFO org.apache.hadoop.conf.Configuration.deprecation – mapred.jobtracker.maxtasks.per.job is deprecated. Instead, use

  • user2304819
    hadoop hbase zookeeper
    I have recently started studying about NoSql and Big Data and decided to continue working on them. From couples of days I am trying to install and configure Hadoop and Hbase on my win2008 R2 64 bit machine. But unfortunately I have been unsuccessful, I am having different errors in each stage of installation. I am following below mentioned tutorials in this regards.for Hadoop= http://blog.sqltrainer.com/2012/01/installing-and-configuring-apache.html for Hbase= http://ics.upjs.sk/~novotnyr/blog/

  • yoonhok
    hadoop hive sqoop
    I read –> How to use Sqoop in Java Program?But it doesn’t work for me.I created a new java project.And I added Library Sqoop and Hadoop.(The Libraries are “hadoop-core-1.1.1.jar, sqoop-1.4.2.jar, etc…. “.)I added them to my new java project.And I tried below code.public class MySqoopDriver {public static void main(String[] args) {String[] str = { “export”, “–connect”, “jdbc:mysql://localhost/mytestdb”, “–hadoop-home”,”/home/yoonhok/development/hadoop-1.1.1″, “–table”, “tbl_1”, “–export-di

  • ender
    hadoop cloudera sqoop
    When using the –incremental append flag in the sqoop import, the job will fail.ERROR tool.ImportTool: Imported Failed: Wrong FS: s3n://<api_key>:<api_secret>@bucket/folder/Here is the full command:sqoop import –connect jdbc:postgresql://someplace.net:12345/db –warehouse-dir s3n://<key>:<private>@bucket/folder/ –table my_table –hive-drop-import-delims –escaped-by “\\” –username root –password safepass -m 2 –split-by id –incremental append –check-column idThe ex

  • Nannie
    java hadoop cloudera oozie hue
    I’m new to Hadoop and I have the flowing problem:I keep getting a “java.lang.ClassNotFoundException” when I’m trying to run my oozzie work flow. I use the Cloudera quick start VM v 4.5. used oozzie settings:Jar name : sample.jar Main class : Driver Arguments : in_single/ outused java classpublic class Driver{ public static class TokenizerMapper extends Mapper<Object, Text, Text, Text> {@Overridepublic void map(final Object key, final Text value, final Mapper<Object, Text, Text, Text>

  • Sumod
    hadoop hbase cloudera
    This is happening in pseudo-distributed as well as distributed mode. When I try to start HBase, initially all the 3 services – master, region and quorumpeer start. However within a minute, the master stops. In the logs, this is the trace -2013-05-06 20:10:25,525 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000. Already tried 0 time(s). 2013-05-06 20:10:26,528 INFO org.apache.hadoop.ipc.Client: Retrying connect to server: <master/master_ip>:9000.

  • user987339
    java hadoop hbase virtual-machine cloudera
    I’m trying to write a remote HBase client using Java. Here is the code for reference :package ttumdt.app.connector;import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.MasterNotRunningException; import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.client.HBaseAdmin; import org.apache.hadoop.hbase.client.HTable; import org.apache.hadoop.hbase.client.Result; import org.apache.hadoop.hba

  • phocion
    hadoop workflow hdfs cloudera oozie
    I installed a pseudo-distributed version of Cloudera on my Linux box, and ran some simple MapReduce examples with success. However, I’m trying to get Oozie to work, and am completely baffled by the errors I am receiving when attempting to execute a simple job workflow:tim@phocion:~$ oozie version Oozie client build version: 3.1.3-cdh4.0.1Copy the pre-packaged examples to HDFS and execute, per the documentation:tim@phocion:~$ oozie job -oozie http://phocion:11000/oozie -config /user/tim/examples

  • shanks_roux
    hadoop apache-pig bigdata piglatin cloudera
    I have a distant Hadoop cluster of machines (Cloudera CDH4). And I’m trying to run a pig script from my computer. Here is my JAVA code to do so:import org.apache.pig.ExecType; import org.apache.pig.PigServer; import org.apache.pig.backend.executionengine.ExecException; import org.apache.pig.data.Tuple;public class TestPig {public static void main(String args[]){PigServer pigServer;try {/** On définit les propriétés */Properties props = new Properties();props.setProperty(“fs.default.name”, “hdfs:

  • Vijay Ratnagiri
    hive cloudera hue beeswax
    I’m using:Cloudera Manager Free Edition: 4.5.1 Cloudera Hadoop Distro: CDH 4.2.0-1.cdh4.2.0.p0.10 (Parcel) Hive Metastore with cloudera manager embedded PostgreSQL database.My cloudera manager is running on a separate machine and it’s not part of the cluster.After setting up the cluster using cloudera manager, I started using hive through hue + beeswax.Everything was running fine for a while and then all of a suddden, whenever I ran any query against a particular table that had a large number of

  • josephmisiti
    hadoop mapreduce cloudera
    I am trying to run a simple example using a binary executable and the cached archive and it does not seem to be working:The example I am trying to run has a mapper which generates three random doubles and a key and the reducer will average those three numbers together and log the average. Very simple stuff. I wrote a simple EXE in c do generate the random numbers:#include <cstdio> #include <stdlib.h> #include <time.h> int main(int argc, char*argv[]) {srand ( time(NULL) );int ra

  • Sergey
    hadoop mapreduce cloudera yarn
    I’ve trying to overcome sudden problem. Befor that problem I’ve used old VM. I’ve downloaded the new one VM and still can’t make my job run. I get Java heap space error. I’ve alredy read this one post: out of Memory Error in HadoopHere is my configs from /etc/hadoop/conf:sudo vi hadoop-env.sh # Extra Java runtime options. Empty by default. #export HADOOP_OPTS=”-Djava.net.preferIPv4Stack=true $HADOOP_CLIENT_OPTS” export HADOOP_CLIENT_OPTS=”-Xmx256m $HADOOP_CLIENT_OPTS”Here is my mapred-site.xml&

  • Tucker
    hadoop amazon-ec2 cloudera java-api
    I’ve installed CDH4 on a ubuntu 12 LTS server successfully in the amazon cloud (1 server). I used Cloudera Manager free edition to install the software and had no errors). I have a program that uses the java API to load a file from my home computer to HDFS in the cloud. I would like to know why this program fails and how to fix it.Configuration conf = new Configuration(); conf.set(“fs.defaultFS”, “hdfs://node01:8020”); FileSystem fs = FileSystem.get(conf); Path targetPath = new

  • Nosk
    hadoop cloudera sqoop cloudera-manager
    I am running a cloudera cluster in a university. Some machine died so I had to remove them from the cluster. Now sqoop can’t start anymore. I have a log entry:16:20:36.635 ERROR org.apache.sqoop.core.SqoopServer Server startup failure org.apache.sqoop.common.SqoopException: JDBCREPO_0007:Unable to lease connectionat org.apache.sqoop.repository.JdbcRepositoryTransaction.begin(JdbcRepositoryTransaction.java:63)at org.apache.sqoop.repository.JdbcRepository.doWithConnection(JdbcRepository.ja

  • piyush
    hadoop hdfs cloudera-manager
    I am trying to installing CDH 4.6 in my cluster which is of 3 nodes. One data node out of this 3 is not able to start at all. Tried searching and solving this by all possible ways, but failed. Please help me in solving this. Below is the log. 5:49:10.708 PM FATAL org.apache.hadoop.hdfs.server.datanode.DataNode Exception in secureMain java.io.IOException: the path component: ‘/’ is world-writable. Its permissions are 0777. Please fix this or select a different socket path.at org.apache.hado

  • Vikas Hardia
    security hadoop kerberos cloudera cloudera-manager
    I have a Cloudera cluster of version 5 here i am facing a problem after enabling Hadoop security Hadoop services are not starting.I am getting error on YARN:PriviledgedActionException as:mapred/[email protected] (auth:KERBEROS) cause:java.io.IOException: Couldn’t setup connection for mapred/[email protected] to cdh4hdm/172.26.43.40:8020 3:23:49.298 PM INFO org.apache.hadoop.service.AbstractService Service org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager failed in state INITED; cause:

  • franklinsun
    cloudera-manager
    Just through the cloud manager 0.8.2 deployed cdh4, when i start mgmt1 through cm web console, the ServiceMonitor, ActivityMonitor, and HostMonitor fail to start, logs print :2014-03-11 19:50:36,373 FATAL org.hsqldb.cmdline.SqlFile: SQL Error at ‘UTF-8’ line 1: “ALTER TABLE PARTITION_INFO ADD COLUMN BEGIN_TS BIGINT(20)” Duplicate column name ‘BEGIN_TS’2014-03-11 19:50:36,374 ERROR com.cloudera.enterprise.dbutil.SqlFileRunner: Exception while executing ddl scripts.com.mysql.jdbc.exceptions.jdbc4.

  • zhutoulala
    hadoop cloudera cloudera-manager
    I downloaded CDH 4.5 quick start vm from here. Each service looks good except for below error was seen after I opened Activities tab to view mapreduce activities:The Activity Monitor server (activitymonitor (localhost)) is unavailable or not responding to connections.The problem remains after I tried to restart Activity Monitor service, then I found following error message in the log. Can anybody help take a look?11:24:35.862 PM WARN org.mortbay.log failed SelectChannelConnector@local

  • Naren
    hadoop hbase cloudera cloudera-manager
    In my setup, Because of Power interruption My VMs got down state after restarting all my VMs I cant able to start Hbase, all other services can restart/stop/start but in UI for CM in hbase ACTIONS tab all options were hidden..Following is my hbase master log,now its in exiting mode6:29:05.880 AM ERROR org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper ZooKeeper exists failed after 3 retries 6:29:05.880 AM WARN org.apache.hadoop.hbase.zookeeper.ZKUtil hconnection Unable to set w

Web site is in building