on 08-09-2016 4:53 PM
We are trying to connect Hadoop HDFS file system using BODS.
Hadoop System Version
Node versions : 2.7.1.2.3.4.0-3485
File System : hdfs
Operating System : RHEL 6.5
BODS Details
Version : BODS 4.1 SP07
Database : Microsoft SQL Server
Operating System : Windows 2012 Enterprise
Job Log
3188 5744 JOB 8/9/2016 7:49:11 AM Reading job <52be4348_a113_4768_9c20_ef9e09934999> from the repository; Server version is <14.2.6.1082>; Repository version is
3188 5744 JOB 8/9/2016 7:49:11 AM <14.2.6.0000>.
3188 5744 JOB 8/9/2016 7:49:11 AM Current directory of job <52be4348_a113_4768_9c20_ef9e09934999> is <C:\Program Files (x86)\SAP BusinessObjects\Data
3188 5744 JOB 8/9/2016 7:49:11 AM Services\bin>.
3188 5744 JOB 8/9/2016 7:49:12 AM Starting job on job server host <SAPBODS>, port <3500>.
3188 5744 JOB 8/9/2016 7:49:12 AM Job <HDFS_Test> of runid <2016080907491231885744> is initiated by user <Administrator>.
3188 5744 JOB 8/9/2016 7:49:12 AM Processing job <HDFS_Test>.
3188 5744 JOB 8/9/2016 7:49:12 AM Optimizing job <HDFS_Test>.
3188 5744 JOB 8/9/2016 7:49:12 AM Job <HDFS_Test> is started.
3372 5516 DATAFLOW 8/9/2016 7:49:13 AM Process to execute data flow <New_DataFlow8> is started.
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM The specified locale <eng_us.cp1252> has been coerced to <Unicode (UTF-16)> for data flow <New_DataFlow8> because the datastore
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM <SAP_HANA> obtains data in <Unicode (UTF-16)> codepage.
3372 5516 JOB 8/9/2016 7:49:14 AM Initializing transcoder for datastore <SAP_HANA> to transcode between engine codepage<Unicode (UTF-16)> and datastore codepage
3372 5516 JOB 8/9/2016 7:49:14 AM <<DEFAULT>>
3372 5516 JOB 8/9/2016 7:49:14 AM Initializing transcoder for datastore <File_Format_313> to transcode between engine codepage<Unicode (UTF-16)> and datastore
3372 5516 JOB 8/9/2016 7:49:14 AM codepage <UTF-8>
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM Data flow <New_DataFlow8> is started.
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM Cache statistics determined that data flow <New_DataFlow8> uses 0 caches with a total size of 0 bytes, which is less than (or
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM equal to) 3753902080 bytes available for caches in virtual memory. Data flow will use IN MEMORY cache type.
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM Data flow <New_DataFlow8> using IN MEMORY Cache.
3372 5516 BLKLOAD 8/9/2016 7:49:14 AM HANA table <TESTHADOOP>, type <Column store>, commit size <10000>, auto correct load <no>, update method <update>, update rows
3372 5516 BLKLOAD 8/9/2016 7:49:14 AM <no>, delete rows <no>.
3372 5516 BLKREAD 8/9/2016 7:49:14 AM Starting Pig. Command line used is <pig -f "C:/ProgramData/SAP BusinessObjects/Data
3372 5516 BLKREAD 8/9/2016 7:49:14 AM Services/log/hadoop/New_DataFlow8_3372_1/hdfsRead.ctl">.
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM Data flow <New_DataFlow8> is terminated due to error <50616>.
3372 5516 DATAFLOW 8/9/2016 7:49:14 AM Process to execute data flow <New_DataFlow8> is completed.
3188 5744 JOB 8/9/2016 7:49:14 AM Job <HDFS_Test> is terminated due to error <50616>.
Error Information
3372 5516 SYS-050616 8/9/2016 7:49:14 AM |Data flow New_DataFlow8
3372 5516 SYS-050616 8/9/2016 7:49:14 AM Cannot find the full path for file <pig>, due to error <2>.
3188 5744 SYS-050616 8/9/2016 7:49:15 AM |Data flow New_DataFlow8
3188 5744 SYS-050616 8/9/2016 7:49:15 AM Cannot find the full path for file <pig>, due to error <2>.
Could anyone, please help to resolve this error?
Thank you,
Karthic Sekar
Hi Karthic,
The file should be placed on the server where the job server resides. Please refer below link for more details:
Shared Directory access - Enterprise Information Management - SCN Wiki
Regards,
Santosh G.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
DS is not on Hadoop namenode \ one of the nodes in hadoop cluster and this is nto really a recommended setup however what's missing here is possibly some environment setup.
- HDFS and pig client need to be installed on Windows DS job server box.
- PIG_HOME and HDFS_HOME needs to be set
Are you able to view data\ data preview - most likely not.
Hi Aasarvari,
Thanks for your suggestion.
Yes, I am not able to preview the data. I have attached the error screenshot on my original post, please refer BODS_Dataflow.png
And may I please know where I can download HDFS and PIG client and where I should place it in order to set the path in windows server.
Thank you,
Karthic Sekar
Moving to Data Services Discussions.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
User | Count |
---|---|
83 | |
10 | |
10 | |
9 | |
7 | |
6 | |
5 | |
5 | |
4 | |
3 |
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.