hadoop - Unable to run Pig Script in Psudo-distributed mode -


i trying write pig script.i have working in local mode, getting error when launch pig in psudo-distributed mode. here script: —————————————————————————————————————————————

register 'myfolder/target/myfunc-with-dependencies.jar'; set mapred.cache.files /tmp/scripts#scripts,/tmp/my_rules#my_rules; set mapred.create.symlink yes;  %default input 'test.seq' %default output 'final.out' %default timeout_ms '180000' %default use_devshm 'true'  data = load '$input' using com.pack.loaders.mytextloader('$timeout_ms'); store data '$output' 

i following error: —————————————————————————————————————————————

2013-09-03 08:08:44,314 [main] warn org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - ooops! job has failed! specify -stop_on_failure if want pig stop on failure. 2013-09-03 08:08:44,316 [main] info  org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - job job_201309030605_0019 has failed! stop running dependent jobs 2013-09-03 08:08:44,319 [main] info  org.apache.pig.backend.hadoop.executionengine.mapreducelayer.mapreducelauncher - 100% complete 2013-09-03 08:08:44,323 [main] error org.apache.pig.tools.pigstats.pigstatsutil - 1 map reduce job(s) failed! 2013-09-03 08:08:44,323 [main] warn  org.apache.pig.tools.pigstats.scriptstate - unable read pigs manifest file 2013-09-03 08:08:44,325 [main] info  org.apache.pig.tools.pigstats.simplepigstats - script statistics:   hadoopversion   pigversion  userid  startedat   finishedat  features 1.2.1       vmuser  2013-09-03 08:07:57 2013-09-03 08:08:44 unknown  failed!  failed jobs: jobid   alias   feature message outputs job_201309030605_0019   data    map_only    message: job failed! error - jobcleanup task failure, task: task_201309030605_0019_m_000001 hdfs://localhost:9000/user/vmuser/final.out,  input(s): failed read data "hdfs://localhost:9000/user/vmuser/test.seq"  output(s): failed produce result in "hdfs://localhost:9000/user/vmuser/final.out" 

—————————————————————————————————————————————

the file exist on hdfs , script doesn't register jar , uses pigstorage() instead of udf working fine. there other way register jar , call udf? appreciated.

the contents of pig log file :

pig stack trace

error 2244: job failed, hadoop not return error message  org.apache.pig.backend.executionengine.execexception: error 2244: job failed, hadoop not return error message     @ org.apache.pig.tools.grunt.gruntparser.executebatch(gruntparser.java:145)     @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:198)     @ org.apache.pig.tools.grunt.gruntparser.parsestoponerror(gruntparser.java:170)     @ org.apache.pig.tools.grunt.grunt.exec(grunt.java:84)     @ org.apache.pig.main.run(main.java:604)     @ org.apache.pig.main.main(main.java:157)     @ sun.reflect.nativemethodaccessorimpl.invoke0(native method)     @ sun.reflect.nativemethodaccessorimpl.invoke(nativemethodaccessorimpl.java:57)     @ sun.reflect.delegatingmethodaccessorimpl.invoke(delegatingmethodaccessorimpl.java:43)     @ java.lang.reflect.method.invoke(method.java:606)     @ org.apache.hadoop.util.runjar.main(runjar.java:160) 

might link helps you: invoking program through java jar used in pig , give complete path of jar file hadoop can locate anywhere.


Comments

Popular posts from this blog

html - How to style widget with post count different than without post count -

How to remove text and logo OR add Overflow on Android ActionBar using AppCompat on API 8? -

javascript - storing input from prompt in array and displaying the array -