[resolved] Exception in component tHDFSOutput_

One Star

[resolved] Exception in component tHDFSOutput_

Hi
I am new to Hadoop and Talend. I am trying to come up with a POC to see how Hadoop and Talend serve to our business needs.
We have set a VM with "Hortonworks Sandbox 2.1" and downloaded Talend Open Studio for BigData 5.6. In my job I have components "thdfsConnection" and "tfileInputJson"(to read a JSON file from my local Machine) and a tlogrow to see the output and thdfsoutput component to write file. Without HDFSOutput component,job is running fine and able to see JSON data. But not able to write this data into HDFS. Failing with the following error.
Here is the log.
Starting job jsonFileReader at 11:47 12/12/2014.
connecting to socket on port 3830
connected
: org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS
: org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
: org.apache.hadoop.util.Shell - Failed to locate the winutils binary in the hadoop binary path
java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries.

 at org.apache.hadoop.util.Shell.getQualifiedBinPath(Shell.java:318)
 at org.apache.hadoop.util.Shell.getWinUtilsPath(Shell.java:333)
 at org.apache.hadoop.util.Shell.<clinit>(Shell.java:326)
 at org.apache.hadoop.util.StringUtils.<clinit>(StringUtils.java:76)
 at org.apache.hadoop.security.Groups.parseStaticMapping(Groups.java:93)
 at org.apache.hadoop.security.Groups.<init>(Groups.java:77)
 at org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240)
 at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:256)
 at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:233)
 at org.apache.hadoop.security.UserGroupInformation.isAuthenticationMethodEnabled(UserGroupInformation.java:310)
 at org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:304)
 at org.apache.hadoop.ipc.RPC.getProtocolProxy(RPC.java:534)
 at org.apache.hadoop.hdfs.NameNodeProxies.createNNProxyWithClientProtocol(NameNodeProxies.java:348)
 at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:244)
 at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:144)
 at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:579)
 at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:524)
 at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:146)
 at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2397)
 at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:89)
 at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2431)
 at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2413)
 at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:368)
 at org.apache.hadoop.fs.FileSystem$1.run(FileSystem.java:157)
 at org.apache.hadoop.fs.FileSystem$1.run(FileSystem.java:154)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Unknown Source)
 at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
 at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:154)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.tFileInputJSON_1Process(jsonFileReader.java:760)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.tHDFSConnection_1Process(jsonFileReader.java:396)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.runJobInTOS(jsonFileReader.java:1526)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.main(jsonFileReader.java:1383)
Exception in component tHDFSOutput_1
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message tag had invalid wire type.; Host Details : local host is: "U90-CNADELLA/10.90.23.50"; destination host is: "10.90.22.112":8000;

 at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764)
 at org.apache.hadoop.ipc.Client.call(Client.java:1414)
 at org.apache.hadoop.ipc.Client.call(Client.java:1363)
 at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
 at com.sun.proxy.$Proxy7.create(Unknown Source)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
 at java.lang.reflect.Method.invoke(Unknown Source)
 at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190)
 at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103)
 at com.sun.proxy.$Proxy7.create(Unknown Source)
 at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:258)
 at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1600)
 at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1465)
 at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1390)
 at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:394)
 at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:390)
 at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:390)
 at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:334)
 at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906)
 at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:887)
 at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:784)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.tFileInputJSON_1Process(jsonFileReader.java:771)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.tHDFSConnection_1Process(jsonFileReader.java:396)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.runJobInTOS(jsonFileReader.java:1526)
 at bigdatademo.jsonfilereader_5_4.jsonFileReader.main(jsonFileReader.java:1383)
Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol message tag had invalid wire type.
 at com.google.protobuf.InvalidProtocolBufferException.invalidWireType(InvalidProtocolBufferException.java:99)
 at com.google.protobuf.UnknownFieldSet$Builder.mergeFieldFrom(UnknownFieldSet.java:498)
 at com.google.protobuf.GeneratedMessage.parseUnknownField(GeneratedMessage.java:193)
 at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>(RpcHeaderProtos.java:1404)
 at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.<init>(RpcHeaderProtos.java:1362)
 at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom(RpcHeaderProtos.java:1492)
 at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto$1.parsePartialFrom(RpcHeaderProtos.java:1487)
 at com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:200)
 at com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241)
 at com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253)
 at com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259)
 at com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49)
 at org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:2364)
 at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1055)
 at org.apache.hadoop.ipc.Client$Connection.run(Client.java:949)
disconnected
Job jsonFileReader ended at 11:47 12/12/2014.
Six Stars

Re: [resolved] Exception in component tHDFSOutput_

Two issues (well one really since the first is probably not really an issue):
": org.apache.hadoop.util.Shell - Failed to locate the winutils binary in the hadoop binary path
java.io.IOException: Could not locate executable null\bin\winutils.exe in the Hadoop binaries."
Is an upstream issue in the Hadoop code, but I don't believe it will actually hurt anything here (This happens any time org.apache.hadoop.util.Shell is called in Windows even as a library).
https://issues.apache.org/jira/browse/HADOOP-11003
The second:
"Exception in component tHDFSOutput_1
java.io.IOException: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message tag had invalid wire type.; Host Details : local host is: "U90-CNADELLA/10.90.23.50"; destination host is: "10.90.22.112":8000;"
This is a version mismatch. You either need to upgrade the JARs used by the Talend HDFS components to match your Hadoop distribution or select the correct Hadoop distribution with the current Talend components.