I have a Mule flow that reads files through a generic inbound endpoint from an ftp server, makes some modification on the data, and writes files through a generic outbound endpoint to an sftp server. Yesterday it processed 60 files. On 57 there were no errors, but on three the following traces appeared. Any suggestions are welcome.
Error writing data over SFTP service, error was: Failed to open local file
4: Failed to open local file
at com.jcraft.jsch.ChannelSftp.throwStatusError(ChannelSftp.java:2297)
at com.jcraft.jsch.ChannelSftp._put(ChannelSftp.java:484)
at com.jcraft.jsch.ChannelSftp.put(ChannelSftp.java:438)
at com.jcraft.jsch.ChannelSftp.put(ChannelSftp.java:405)
at org.mule.transport.sftp.SftpClient.storeFile(SftpClient.java:385)
at org.mule.transport.sftp.SftpMessageDispatcher.doDispatch(SftpMessageDispatcher.java:176)
at org.mule.transport.AbstractMessageDispatcher.process(AbstractMessageDispatcher.java:100)
at org.mule.transport.AbstractConnector$DispatcherMessageProcessor.process(AbstractConnector.java:2553)
at org.mule.processor.AbstractInterceptingMessageProcessorBase.processNext(AbstractInterceptingMessageProcessorBase.java:105)
at org.mule.interceptor.AbstractEnvelopeInterceptor.process(AbstractEnvelopeInterceptor.java:55)
at org.mule.processor.AsyncInterceptingMessageProcessor.processNextTimed(AsyncInterceptingMessageProcessor.java:111)
at org.mule.processor.AsyncInterceptingMessageProcessor$AsyncMessageProcessorWorker.doRun(AsyncInterceptingMessageProcessor.java:158)
at org.mule.work.AbstractMuleEventWork.run(AbstractMuleEventWork.java:43)
at org.mule.work.WorkerContext.run(WorkerContext.java:310)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
[ERROR] DispatchingLogger [Services].xmlSftpConnector.dispatcher.29 2014-06-09 15:55:07 Unexpected exception attempting to write file, message was: Failed to open local file
java.io.IOException: Failed to open local file
at org.mule.transport.sftp.SftpClient.storeFile(SftpClient.java:390)
at org.mule.transport.sftp.SftpMessageDispatcher.doDispatch(SftpMessageDispatcher.java:176)
at org.mule.transport.AbstractMessageDispatcher.process(AbstractMessageDispatcher.java:100)
at org.mule.transport.AbstractConnector$DispatcherMessageProcessor.process(AbstractConnector.java:2553)
at org.mule.processor.AbstractInterceptingMessageProcessorBase.processNext(AbstractInterceptingMessageProcessorBase.java:105)
at org.mule.interceptor.AbstractEnvelopeInterceptor.process(AbstractEnvelopeInterceptor.java:55)
at org.mule.processor.AsyncInterceptingMessageProcessor.processNextTimed(AsyncInterceptingMessageProcessor.java:111)
at org.mule.processor.AsyncInterceptingMessageProcessor$AsyncMessageProcessorWorker.doRun(AsyncInterceptingMessageProcessor.java:158)
at org.mule.work.AbstractMuleEventWork.run(AbstractMuleEventWork.java:43)
at org.mule.work.WorkerContext.run(WorkerContext.java:310)
at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:619)
[WARN] SftpUtil [Services].xmlSftpConnector.dispatcher.29 2014-06-09 15:55:07 Class java.io.ByteArrayInputStream did not implement the 'ErrorOccurred' decorator, errorOccured=true could not be set.
[ERROR] DispatchingLogger [Services].xmlSftpConnector.dispatcher.29 2014-06-09 15:55:07
********************************************************************************
Message : Failed to route event via endpoint: DefaultOutboundEndpoint{endpointUri=sftp://user:<password>#server.com/folder/test/upload, connector=SftpConnector
{
name=xmlSftpConnector
lifecycle=start
this=cfc9fac
numberOfConcurrentTransactedReceivers=4
createMultipleTransactedReceivers=true
connected=true
supportedProtocols=[sftp]
serviceOverrides=<none>
}
, name='endpoint.sftp.server.com.22.folder.test.upload', mep=ONE_WAY, properties={outputPattern=#[function:datestamp:dd-MM-yy]_#[function:systime].xml}, transactionConfig=Transaction{factory=null, action=INDIFFERENT, timeout=0}, deleteUnacceptedMessages=false, initialState=started, responseTimeout=10000, endpointEncoding=UTF-8, disableTransportTransformer=false}. Message payload is of type: String
Code : MULE_ERROR-42999
--------------------------------------------------------------------------------
Exception stack is:
1. Failed to open local file (java.io.IOException)
org.mule.transport.sftp.SftpClient:390 (null)
2. Failed to route event via endpoint: DefaultOutboundEndpoint{endpointUri=sftp://user:<password>#server.com/folder/test/upload, connector=SftpConnector
{
name=xmlSftpConnector
lifecycle=start
this=cfc9fac
numberOfConcurrentTransactedReceivers=4
createMultipleTransactedReceivers=true
connected=true
supportedProtocols=[sftp]
serviceOverrides=<none>
}
, name='endpoint.sftp.server.com.22.folder.test.upload', mep=ONE_WAY, properties={outputPattern=#[function:datestamp:dd-MM-yy]_#[function:systime].xml}, transactionConfig=Transaction{factory=null, action=INDIFFERENT, timeout=0}, deleteUnacceptedMessages=false, initialState=started, responseTimeout=10000, endpointEncoding=UTF-8, disableTransportTransformer=false}. Message payload is of type: String (org.mule.api.transport.DispatchException)
org.mule.transport.AbstractMessageDispatcher:109 (http://www.mulesoft.org/docs/site/current3/apidocs/org/mule/api/transport/DispatchException.html)
--------------------------------------------------------------------------------
Root Exception stack trace:
java.io.IOException: Failed to open local file
at org.mule.transport.sftp.SftpClient.storeFile(SftpClient.java:390)
at org.mule.transport.sftp.SftpMessageDispatcher.doDispatch(SftpMessageDispatcher.java:176)
at org.mule.transport.AbstractMessageDispatcher.process(AbstractMessageDispatcher.java:100)
+ 3 more (set debug level logging or '-Dmule.verbose.exceptions=true' for everything)
********************************************************************************
It looks like one of 3 reasons:
Permissions on the folder you're writing to.
Spaces on the file path (or name).
Wrong slash in the file path.
EDIT based on comment.
You can try configuring maxThreadsActive to limit the no of threads active at a time.
<dispatcher-threading-profile maxThreadsActive="5" maxThreadsIdle="5"/>
Related
I am very new to IBM Bluemix and Logstash.
My application is based on Spring Boot + Log4j which was deployed into IBM BlueMix.
Goal:
Reading entire java stack trace + BlueMix log and write into a file in local Linux Server.
I don't want to use elastic search & Kibana. I think it is overkill for my requirement. I don't want fancy GUI thing, but a basic text file which contains all log information.
What I did till now:
Installed & Setup logstash.
Able to write bluemix log into a file in my local server.
Below is my Logstash conf file. As I want entire log, I am not using any filter. (based on my 1 day old logstash knowledge)
input {
tcp {
port => 5000
type => syslog
}
}
filter {
}
output {
file {
path => "/app/uot0/cloud/logstash/logstash-2.3.4/tmp/access_log"
}
}
Problem:
Logstash output file "access_log" is NOT containing full stack trace from a Java Exception, it contains only name of that exception.
(part of) Original log from BlueMix console:
2016-07-15T19:31:34.184-0400[App/0]OUT23:31:34.183 [36m[http-nio-61430-exec-10][0;39m [39mDEBUG[0;39m [36mo.s.w.s.m.m.a.ExceptionHandlerExceptionResolver[0;39m [30m- Resolving exception from handler [public org.springframework.http.ResponseEntity<?> com.abc.xyx.rest.TaskController.getTasks(org.springframework.web.context.request.WebRequest)]: java.lang.IllegalArgumentException: Invalid enum name:Other in com.abc.xyx.service.task.TaskType
2016-07-15T19:31:34.187-0400[App/0]OUT[0;39m23:31:34.186 [36m[http-nio-61430-exec-10][0;39m [39mDEBUG[0;39m [36mo.s.b.f.s.DefaultListableBeanFactory[0;39m [30m- Returning cached instance of singleton bean 'exceptionHandlingAdvice'
2016-07-15T19:31:34.189-0400[App/0]OUT[0;39m23:31:34.189 [36m[http-nio-61430-exec-10][0;39m [39mDEBUG[0;39m [36mo.s.w.s.m.m.a.ExceptionHandlerExceptionResolver[0;39m [30m- Invoking #ExceptionHandler method: public void com.abc.xyx.rest.ExceptionHandlingAdvice.systemException(java.lang.Exception)
2016-07-15T19:31:34.193-0400[App/0]OUT[0;39m23:31:34.192 [36m[http-nio-61430-exec-10][0;39m [1;31mERROR[0;39m [36mc.r.t.r.ExceptionHandlingAdvice[0;39m [30m- Unexpected system exception
2016-07-15T19:31:34.193-0400[App/0]OUT[0;39mjava.lang.IllegalArgumentException: Invalid enum name:Other in com.abc.xyx.service.task.TaskType
2016-07-15T19:31:34.194-0400[App/0]OUT at com.abc.xyx.service.task.TaskUtils.getEnum(TaskUtils.java:30) ~[xyx-core-1.0-SNAPSHOT.jar:na]
2016-07-15T19:31:34.194-0400[App/0]OUT at com.abc.xyx.service.task.Task.init(Task.java:179) ~[xyx-core-1.0-SNAPSHOT.jar:na]
2016-07-15T19:31:34.194-0400[App/0]OUT at com.abc.xyx.service.task.TaskService.searchOnePageTasks(TaskService.java:558) ~[xyx-core-1.0-SNAPSHOT.jar:na]
2016-07-15T19:31:34.194-0400[App/0]OUT at com.abc.xyx.service.task.Task.<init>(Task.java:93) ~[xyx-core-1.0-SNAPSHOT.jar:na]
2016-07-15T19:31:34.194-0400[App/0]OUT at com.abc.xyx.service.task.TaskService$$FastClassBySpringCGLIB$$bb02ea04.invoke(<generated>) ~[xyx-core-1.0-SNAPSHOT.jar:na]
2016-07-15T19:31:34.194-0400[App/0]OUT at org.springframework.cglib.proxy.MethodProxy.invoke(MethodProxy.java:204) ~[spring-core-4.2.4.RELEASE.jar:4.2.4.RELEASE]
(entire) Equivalent log which was written into access.log by logstash is below:
{"message":"549 <14>1 2016-07-15T23:31:34.208666+00:00 loggregator c022c216-4373-418e-bb4d-fbde0b41d720 [App/0] - - \u001B[0;39m23:31:34.208 \u001B[36m[http-nio-61430-exec-10]\u001B[0;39m \u001B[39mDEBUG\u001B[0;39m \u001B[36mo.s.w.s.m.m.a.HttpEntityMethodProcessor\u001B[0;39m \u001B[30m- Written [{timestamp=Fri Jul 15 23:31:34 UTC 2016, status=500, error=Internal Server Error, exception=java.lang.IllegalArgumentException, message=System exception, path=/api/tasksearch}] as \"application/json;charset=UTF-8\" using
Question:
How can I get the full stack trace for Java Exceptions, so that it will be useful for debugging?
Any help will be appreciated...
I'm having trouble in mule service. Daily, the service crashes, about 4 times a day, making it necessary to redeploy my service.
Here's my stack trace.
ERROR 2016-02-04 10:30:38,063 [[app_service].NoSessionConnector.receiver.03] org.mule.exception.DefaultMessagingExceptionStrategy:
********************************************************************************
Message : Timeout for connection (java.net.SocketException). Message payload is of type: HttpResponse
Code : MULE_ERROR--2
--------------------------------------------------------------------------------
Exception stack is:
1. Timeout for connection (java.net.SocketException)
java.net.SocketOutputStream:-2 (null)
2. Timeout for connection (java.net.SocketException). Message payload is of type: HttpResponse (org.mule.execution.ResponseDispatchExcepti
on)
org.mule.transport.http.HttpMessageProcessTemplate:141 (http://www.mulesoft.org/docs/site/current3/apidocs/org/mule/execution/ResponseDispatc
hException.html)
--------------------------------------------------------------------------------
Root Exception stack trace:
java.net.SocketException: Tempo esgotado para conexão
at java.net.SocketOutputStream.socketWrite0(Native Method)
at java.net.SocketOutputStream.socketWrite(SocketOutputStream.java:113)
at java.net.SocketOutputStream.write(SocketOutputStream.java:159)
+ 3 more (set debug level logging or '-Dmule.verbose.exceptions=true' for everything)
********************************************************************************
I've tried changing the number of threads in maxThreadsActive and maxBufferSize, but is still occurring the same error
Seems like the tcp connection has not been usable for a long time.
Use the the Socket.setKeepAlive() or apply the heart beat mechanism.
I am using graylog GELFAppender in my log4j.properties.
In startup of application
this error occured:
log4j:ERROR Error during sending GELF message. Error code: -1.
2016-02-03 12:02:47,513 ERROR [caspian.banking.router.DefaultErrorHandler:33] ERROR in MessageListener:
weblogic.jms.common.JMSException: weblogic.messaging.dispatcher.DispatcherException: weblogic.rjvm.PeerGoneException: ; nested exception is:
weblogic.utils.net.SocketResetException
at weblogic.jms.dispatcher.DispatcherAdapter.convertToJMSExceptionAndThrow(DispatcherAdapter.java:116)
at weblogic.jms.dispatcher.DispatcherAdapter.dispatchSyncNoTran(DispatcherAdapter.java:61)
at weblogic.jms.client.JMSSession.receiveMessage(JMSSession.java:894)
....
java.io.IOException: An invalid argument was supplied
at sun.nio.ch.DatagramDispatcher.write0(Native Method)
at sun.nio.ch.DatagramDispatcher.write(DatagramDispatcher.java:51)
at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:93)
at sun.nio.ch.IOUtil.write(IOUtil.java:65)
at sun.nio.ch.DatagramChannelImpl.write(DatagramChannelImpl.java:616)
at org.graylog2.GelfUDPSender.sendDatagrams(GelfUDPSender.java:57)
at org.graylog2.GelfUDPSender.sendMessage(GelfUDPSender.java:41)
at org.graylog2.log.GelfAppender.append(GelfAppender.java:224)
...
Caused by: weblogic.utils.net.SocketResetException
at weblogic.socket.SocketMuxer.readReadySocketOnce(SocketMuxer.java:931)
at weblogic.socket.SocketMuxer.readReadySocket(SocketMuxer.java:887)
this is my log4j.properties file
## Graylog
log4j.appender.graylog2=org.graylog2.log.GelfAppender
log4j.appender.graylog2.graylogHost=192.168.243.23
#log4j.appender.graylog2.port=12201
log4j.appender.graylog2.originHost=localhost
log4j.appender.graylog2.layout=org.apache.log4j.PatternLayout
log4j.appender.graylog2.additionalFields={'environment': 'DEV', 'application': 'lotus-core'}
log4j.appender.graylog2.extractStacktrace=true
log4j.appender.graylog2.addExtendedInformation=true
log4j.appender.graylog2.facility=lotus-core
log4j.appender.graylog2.Threshold=WARN
I googled and nothing found.
please help
tnx
I am trying to run IBM MDM Server BatchProcessor ; on executing rubBatch.sh script I am getting following error:
ERROR - [Exception_CustomerReflectionDelegate_JNDINameLookup:] CDKBC2030E:Unable to locate a reference using the following parameters. JNDI = com/dwl/base/requestHandler/beans/DWLServiceController; URL = null; additional error message = Error getting WsnNameService properties
javax.naming.NamingException: Error getting WsnNameService properties [Root exception is org.omg.CORBA.TRANSIENT: initial and forwarded IOR inaccessible vmcid: IBM minor code: E07 completed: No]
at com.ibm.ws.naming.util.WsnInitCtxFactory.mergeWsnNSProperties(WsnInitCtxFactory.java:1482)
at com.ibm.ws.naming.util.WsnInitCtxFactory.getRootContextFromServer(WsnInitCtxFactory.java:989)
at com.ibm.ws.naming.util.WsnInitCtxFactory.getRootJndiContext(WsnInitCtxFactory.java:909)
at com.ibm.ws.naming.util.WsnInitCtxFactory.getInitialContextInternal(WsnInitCtxFactory.java:581)
at com.ibm.ws.naming.util.WsnInitCtx.getContext(WsnInitCtx.java:124)
at com.ibm.ws.naming.util.WsnInitCtx.getContextIfNull(WsnInitCtx.java:799)
at com.ibm.ws.naming.util.WsnInitCtx.lookup(WsnInitCtx.java:165)
at com.ibm.ws.naming.util.WsnInitCtx.lookup(WsnInitCtx.java:180)
at javax.naming.InitialContext.lookup(InitialContext.java:436)
at com.dwl.tcrm.delegate.CustomerReflectionDelegate.init(CustomerReflectionDelegate.java:382)
at com.dwl.tcrm.delegate.CustomerReflectionDelegate.<init>(CustomerReflectionDelegate.java:190)
at com.dwl.batchframework.delegate.Customer.init(Customer.java:122)
at com.dwl.batchframework.delegate.CustomerSingleton.init(CustomerSingleton.java:82)
at com.dwl.batchframework.BatchController.initServerConfiguration(BatchController.java:395)
at com.dwl.batchframework.BatchController.initialize(BatchController.java:360)
at com.dwl.batchframework.BatchController.runBatch(BatchController.java:273)
at com.dwl.batchframework.BatchController.main(BatchController.java:202)
Caused by: org.omg.CORBA.TRANSIENT: initial and forwarded IOR inaccessible vmcid: IBM minor code: E07 completed: No
at com.ibm.rmi.corba.ClientDelegate.createRequest(ClientDelegate.java:1153)
at com.ibm.CORBA.iiop.ClientDelegate.createRequest(ClientDelegate.java:1375)
at com.ibm.rmi.corba.ClientDelegate.createRequest(ClientDelegate.java:1042)
at com.ibm.CORBA.iiop.ClientDelegate.createRequest(ClientDelegate.java:1341)
at com.ibm.rmi.corba.ClientDelegate.request(ClientDelegate.java:1685)
at com.ibm.CORBA.iiop.ClientDelegate.request(ClientDelegate.java:1297)
at org.omg.CORBA.portable.ObjectImpl._request(ObjectImpl.java:458)
at com.ibm.WsnBootstrap._WsnNameServiceStub.getProperties(_WsnNameServiceStub.java:38)
at com.ibm.ws.naming.util.WsnInitCtxFactory.mergeWsnNSProperties(WsnInitCtxFactory.java:1479)
... 16 more
Caused by: org.omg.CORBA.COMM_FAILURE: CAUGHT_EXCEPTION_WHILE_CONFIGURING_SSL_CLIENT_SOCKET Exception=com.ibm.websphere.ssl.SSLException: CWPKI0315E: SSL configuration properties are null. Could be a problem parsing the SSL client configuration. vmcid: 0x49421000 minor code: 70 completed: No
at com.ibm.ws.security.orbssl.WSSSLClientSocketFactoryImpl.createSSLSocket(WSSSLClientSocketFactoryImpl.java:256)
at com.ibm.ws.orbimpl.transport.WSSSLTransportConnection.createSocket(WSSSLTransportConnection.java:236)
at com.ibm.CORBA.transport.TransportConnectionBase.connect(TransportConnectionBase.java:351)
at com.ibm.ws.orbimpl.transport.WSTransport$1.run(WSTransport.java:502)
at com.ibm.ws.security.util.AccessController.doPrivileged(AccessController.java:118)
at com.ibm.ws.orbimpl.transport.WSTransport.getConnection(WSTransport.java:499)
at com.ibm.CORBA.transport.TransportBase.getConnection(TransportBase.java:181)
at com.ibm.rmi.iiop.TransportManager.get(TransportManager.java:95)
at com.ibm.rmi.iiop.GIOPImpl.getConnection(GIOPImpl.java:131)
at com.ibm.rmi.iiop.GIOPImpl.locate(GIOPImpl.java:221)
at com.ibm.rmi.corba.ClientDelegate.locate(ClientDelegate.java:1769)
at com.ibm.rmi.corba.ClientDelegate._createRequest(ClientDelegate.java:1794)
at com.ibm.rmi.corba.ClientDelegate.createRequest(ClientDelegate.java:1064)
at com.ibm.rmi.corba.ClientDelegate.createRequest(ClientDelegate.java:1149)
... 24 more
An error occurred. See the log file for details. Additional error message = Error getting WsnNameService properties
Exception in thread "P=271190:O=0:CT" java.lang.NullPointerException
at com.dwl.batchframework.BatchController.closeIOqueues(BatchController.java:335)
at com.dwl.batchframework.BatchController.runBatch(BatchController.java:295)
at com.dwl.batchframework.BatchController.main(BatchController.java:202)
Patrick,
AS per the Exception "SSL configuration properties are null. Could be a problem parsing the SSL client configuration." .
Please check the path of the CLIENTSSL variable (ssl.client.props) in the runbatch.sh file. Looks like the script is unable to find ssl.client.props file.
I'm trying to connect to remote HDFS cluster. I've read some documentation and getting started's but didn't find a best solution how to do that.
Situation: I have HDFS on xxx-something.com. I can connect to it via SSH and everything works.
But what I'm trying to do, get the files from it to my local machine.
What I've done:
I've created core-site.xml in my conf folder (I'm creating Play! application). There I've changed fs.default.name config to hdfs://xxx-something.com:8020 (not sure about the port).
Then I'm trying to launch a simple test:
val conf = new Configuration()
conf.addResource(new Path("conf/core-site.xml"))
val fs = FileSystem.get(conf)
val status = fs.listStatus(new Path("/data/"))
And I'm getting errors:
13:56:09.012 [specs2.DefaultExecutionStrategy1] WARN org.apache.hadoop.conf.Configuration - conf/core-site.xml:a attempt to override final parameter: fs.trash.interval; Ignoring.
13:56:09.012 [specs2.DefaultExecutionStrategy1] WARN org.apache.hadoop.conf.Configuration - conf/core-site.xml:a attempt to override final parameter: hadoop.tmp.dir; Ignoring.
13:56:09.013 [specs2.DefaultExecutionStrategy1] WARN org.apache.hadoop.conf.Configuration - conf/core-site.xml:a attempt to override final parameter: fs.checkpoint.dir; Ignoring.
13:56:09.022 [specs2.DefaultExecutionStrategy1] DEBUG org.apache.hadoop.fs.FileSystem - Creating filesystem for hdfs://xxx-something.com:8021
13:56:09.059 [specs2.DefaultExecutionStrategy1] DEBUG org.apache.hadoop.conf.Configuration - java.io.IOException: config()
at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:226)
at org.apache.hadoop.conf.Configuration.<init>(Configuration.java:213)
at org.apache.hadoop.security.SecurityUtil.<clinit>(SecurityUtil.java:53)
at org.apache.hadoop.net.NetUtils.<clinit>(NetUtils.java:62)
Thanks in advance!
Update:
probably the port was wrong. Now I set it to 22, I'm still getting same errors, but after 3 times it does say:
14:01:01.877 [specs2.DefaultExecutionStrategy1] DEBUG org.apache.hadoop.ipc.Client - Connecting to xxx-something.com/someIp:22
14:01:02.187 [specs2.DefaultExecutionStrategy1] DEBUG org.apache.hadoop.ipc.Client - IPC Client (47) connection to xxx-something.com/someIp:22 from britva sending #0
14:01:02.188 [IPC Client (47) connection to xxx-something.com/someIp:22 from britva] DEBUG org.apache.hadoop.ipc.Client - IPC Client (47) connection to xxx-something.com/someIp:22 from britva: starting, having connections 1
14:01:02.422 [IPC Client (47) connection to xxx-something.com/someIp:22 from britva] DEBUG org.apache.hadoop.ipc.Client - IPC Client (47) connection to xxx-something.com/someIp:22 from britva got value #1397966893
And afterwards:
Call to xxx-something.com/someIp:22 failed on local exception: java.io.EOFException
java.io.IOException: Call to xxx-something.com/someIp:22 failed on local exception: java.io.EOFException
at org.apache.hadoop.ipc.Client.wrapException(Client.java:1103)
at org.apache.hadoop.ipc.Client.call(Client.java:1071)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
at com.sun.proxy.$Proxy1.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:118)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:222)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:187)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1328)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:65)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1346)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:244)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:122)
at HdfsSpec$$anonfun$1$$anonfun$apply$3.apply(HdfsSpec.scala:33)
at HdfsSpec$$anonfun$1$$anonfun$apply$3.apply(HdfsSpec.scala:17)
at testingSupport.specs2.MyNotifierRunner$$anon$2$$anon$1.executeBody(MyNotifierRunner.scala:16)
at testingSupport.specs2.MyNotifierRunner$$anon$2$$anon$1.execute(MyNotifierRunner.scala:16)
Caused by: java.io.EOFException
at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:807)
at org.apache.hadoop.ipc.Client$Connection.run(Client.java:745)
What does it mean?
You'll need to find the fs.default.name property in the $HADOOP_HOME/conf/core-site.xml on the server running the Name Node (HDFS master) to get the correct port. It might be 8020, or it could be something else. That's what you should use. Make sure there's no firewall between you and the server that disallows connections on the port.