hadoop - Unable to run Pig Script in Psudo-distributed mode -
i trying write pig script.i have working in local mode, getting error when launch pig in psudo-distributed mode. here script: —————————————————————————————————————————————
register 'myfolder/target/myfunc-with-dependencies.jar'; set mapred.cache.files /tmp/scripts#scripts,/tmp/my_rules#my_rules; set mapred.create.symlink yes; %default input 'test.seq' %default output 'final.out' %default timeout_ms '180000' %default use_devshm 'true' data = load '$input' using com.pack.loaders.mytextloader('$timeout_ms'); store data '$output'
i following error: —————————————————————————————————————————————
2013-09-03 08:08:44,314 [main] warn org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - ooops! job has failed! specify -stop_on_failure if want pig stop on failure. 2013-09-03 08:08:44,316 [main] info org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - job job_201309030605_0019 has failed! stop running dependent jobs 2013-09-03 08:08:44,319 [main] info org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - 100% complete 2013-09-03 08:08:44,323 [main] error org.apache.pig.tools.pigstats.pigstatsutil - 1 map reduce job(s) failed! 2013-09-03 08:08:44,323 [main] warn org.apache.pig.tools.pigstats.scriptstate - unable read pigs manifest file 2013-09-03 08:08:44,325 [main] info org.apache.pig.tools.pigstats.simplepigstats - script statistics: hadoopversion pigversion userid startedat finishedat features 1.2.1 vmuser 2013-09-03 08:07:57 2013-09-03 08:08:44 unknown failed! failed jobs: jobid alias feature message outputs job_201309030605_0019 data map_only message: job failed! error - jobcleanup task failure, task: task_201309030605_0019_m_000001 hdfs://localhost:9000/user/vmuser/final.out, input(s): failed read data "hdfs://localhost:9000/user/vmuser/test.seq" output(s): failed produce result in "hdfs://localhost:9000/user/vmuser/final.out"
—————————————————————————————————————————————
the file exist on hdfs , script doesn't register jar , uses pigstorage() instead of udf working fine. there other way register jar , call udf? appreciated.
the contents of pig log file :
pig stack trace
error 2244: job failed, hadoop not return error message org.apache.pig.backend.executionengine.execexception: error 2244: job failed, hadoop not return error message @ org.apache.pig.tools.grunt.gruntparser.executebatch(gruntparser.java:145) @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:198) @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:170) @ org.apache.pig.tools.grunt.grunt.exec(grunt.java:84) @ org.apache.pig.main.run(main.java:604) @ org.apache.pig.main.main(main.java:157) @ sun.reflect.nativemethodaccessorimpl.invoke0(native method) @ sun.reflect.nativemethodaccessorimpl.invoke(nativemethodaccessorimpl.java:57) @ sun.reflect.delegatingmethodaccessorimpl.invoke(delegatingmethodaccessorimpl.java:43) @ java.lang.reflect.method.invoke(method.java:606) @ org.apache.hadoop.util.runjar.main(runjar.java:160)
might link helps you: invoking program through java jar used in pig , give complete path of jar file hadoop can locate anywhere.
Comments
Post a Comment