hadoop - Unable to run Pig Script in Psudo-distributed mode -


i trying write pig script.i have working in local mode, getting error when launch pig in psudo-distributed mode. here script: —————————————————————————————————————————————

register 'myfolder/target/myfunc-with-dependencies.jar'; set mapred.cache.files /tmp/scripts#scripts,/tmp/my_rules#my_rules; set mapred.create.symlink yes;  %default input 'test.seq' %default output 'final.out' %default timeout_ms '180000' %default use_devshm 'true'  data = load '$input' using com.pack.loaders.mytextloader('$timeout_ms'); store data '$output' 

i following error: —————————————————————————————————————————————

2013-09-03 08:08:44,314 [main] warn org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - ooops! job has failed! specify -stop_on_failure if want pig stop on failure. 2013-09-03 08:08:44,316 [main] info  org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - job job_201309030605_0019 has failed! stop running dependent jobs 2013-09-03 08:08:44,319 [main] info  org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - 100% complete 2013-09-03 08:08:44,323 [main] error org.apache.pig.tools.pigstats.pigstatsutil - 1 map reduce job(s) failed! 2013-09-03 08:08:44,323 [main] warn  org.apache.pig.tools.pigstats.scriptstate - unable read pigs manifest file 2013-09-03 08:08:44,325 [main] info  org.apache.pig.tools.pigstats.simplepigstats - script statistics:   hadoopversion   pigversion  userid  startedat   finishedat  features 1.2.1       vmuser  2013-09-03 08:07:57 2013-09-03 08:08:44 unknown  failed!  failed jobs: jobid   alias   feature message outputs job_201309030605_0019   data    map_only    message: job failed! error - jobcleanup task failure, task: task_201309030605_0019_m_000001 hdfs://localhost:9000/user/vmuser/final.out,  input(s): failed read data "hdfs://localhost:9000/user/vmuser/test.seq"  output(s): failed produce result in "hdfs://localhost:9000/user/vmuser/final.out" 

—————————————————————————————————————————————

the file exist on hdfs , script doesn't register jar , uses pigstorage() instead of udf working fine. there other way register jar , call udf? appreciated.

the contents of pig log file :

pig stack trace

error 2244: job failed, hadoop not return error message  org.apache.pig.backend.executionengine.execexception: error 2244: job failed, hadoop not return error message     @ org.apache.pig.tools.grunt.gruntparser.executebatch(gruntparser.java:145)     @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:198)     @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:170)     @ org.apache.pig.tools.grunt.grunt.exec(grunt.java:84)     @ org.apache.pig.main.run(main.java:604)     @ org.apache.pig.main.main(main.java:157)     @ sun.reflect.nativemethodaccessorimpl.invoke0(native method)     @ sun.reflect.nativemethodaccessorimpl.invoke(nativemethodaccessorimpl.java:57)     @ sun.reflect.delegatingmethodaccessorimpl.invoke(delegatingmethodaccessorimpl.java:43)     @ java.lang.reflect.method.invoke(method.java:606)     @ org.apache.hadoop.util.runjar.main(runjar.java:160) 

might link helps you: invoking program through java jar used in pig , give complete path of jar file hadoop can locate anywhere.


Comments

Popular posts from this blog

How to remove text and logo OR add Overflow on Android ActionBar using AppCompat on API 8? -

html - How to style widget with post count different than without post count -

url rewriting - How to redirect a http POST with urlrewritefilter -