Issue while configuring Hadoop 2.6.0 in Eclipse on Windows - java

I've been trying to configure Hadoop 2.6.0 in Eclipse on Windows using this tutorial - http://www.srccodes.com/p/article/47/build-install-configure-eclipse-plugin-apache-hadoop
I'm able to build Hadoop 2.6.0 plugin jar for Eclipse. My Hadoop cluster daemons are up and running on Windows. But when I try to connect eclipse to HDFS as per the tutorial nothing shows up.
I've also tried with Map/Reduce(V2) Master Port - 50070(namenode https port) and DFS Master Port - 8020(fs port) but no luck.
Any advise would be of great help.

Fixed this issue by running Eclipse as an administrator. Hadoop location would show '0' if there is no data in HDFS. At first instance it looks like, eclipse is not able to connect to HDFS though.

Related

Can't run Java EE project through WildFly 10

When I unzip the wildfly-10.1.0.Final.zip file on my computer at home. Then WildFly starts running automatically. I had verified this through going to localhost:8080. Because of this I can't run my Java EE project on Netbeans (I have added WildFly as server in Netbeans). In the logs I see:
Address localhost: 8080 is already in use
I also can't shutdown WildFly through the following command:
$ ./jboss-cli.sh --connect command=:shutdown
However I can shutdown WildFly by killing his process. But this still doesn't fixed my issue on Netbeans. Because I still get to see: Address localhost: 8080 is already in use.
At my work when I had unzipped the wildfly-10.1.0.Final.zip file. It didn't start automatically I also had no problems with running my project on WildFly. And I also can shutdown WildFly through the command line or Netbeans.
Anyone that maybe knows how I can fix my WildFly server problem on my computer at home?
Which version of NetBeans are you using ? Until 8.2 WildFly 10 isn't correctly identified.
You may try NetBeans 8.2 RC1 from https://netbeans.org/community/releases/82/ or a nightly build as this would fix your issue.
NetBeans checks if there the instance is already running before trying to start it, so you don't have to start it beforehand but you can :)

Connect remote hadoop server using eclipse

I am trying to run some map-reduce programs on a remote server of Hadoop 2.0.0, which is running on CentOS 6.4 using ssh.
I am using Eclipse LUNA on my windows 8 machine.
Is there a way to run the programs directly on my Eclipse without converting them to JAR files?
If hadoop is running on a linux machine, you can not connect directly from windows. The connection has to be SSH.
I hope you are looking for something like this:
https://help.eclipse.org/kepler/index.jsp?topic=%2Forg.eclipse.rse.doc.user%2Fgettingstarted%2Fgusing.html
The correct answer (similar) to this is here:
Work on a remote project with Eclipse via SSH

UnsatisfiedLinkError (NativeIO$Windows.access0) when submitting mapreduce job to hadoop 2.2 from windows to ubuntu

I submit my mapreduce jobs from a java application running on windows to the hadoop 2.2 cluster running on ubuntu. In hadoop 1.x this worked as expected but on hadoop 2.2 I get a strange Error:
java.lang.UnsatisfiedLinkError: org.apache.hadoop.io.nativeio.NativeIO$Windows.access0(Ljava/lang/String;I)Z
I compiled the necesary windows libraries (hadoop.dll and winutils.exe) and can access the hdfs via code and read the cluster information using hadoop API. Only the job submission does not work.
Any help is aprecciated.
Solution: I found it out myself, the path where the windows hadoop binaries can be found has to be added to the PATH variable of windows.
Get hadoop.dll (or libhadoop.so on *x). Make sure to match bitness (32- vs. 64-bit) with your JVM.
Make sure it is available via PATH or java.library.path.
Note that setting java.library.path overrides PATH. If you set java.library.path, make sure it is correct and contains the hadoop library.
This error generally occurs due to the mismatch in your binary files in your %HADOOP_HOME%\bin folder. So, what you need to do is to get hadoop.dll and winutils.exe specifically for your hadoop version.
Get hadoop.dll and winutils.exe for your specific hadoop version and copy them to your %HADOOP_HOME%\bin folder.
I have been having issues with my Windows 10 Hadoop installation since morning where the NameNode and DataNode were not starting due the mismatch in the binary files. The issues were resolved after I replaced the bin folder with the one that corresponds with the version of my Hadoop. Possibly, the bin folder I replaced with the one that came with the installation was for a different version, I don't know how it happened. If all your configurations are intact, you might want to replace the bin folder with a version that correspond with your Hadoop installation.

Do we need mahout and hadoop on same machine?

I am trying to build a recommendation engine using mahout, hadoop and java. It is my first time working with hadoop, I am getting my data sets from a server where hadoop is already installed which is a linux enviroment. My development environment is windows, now do I need to install mahout in my development environment or the server? If I need mahout on my development environment do I also need to install hadoop in it?
If you don't have Hadoop on your machine, Mahout will run in pseudo-distributed mode on the current machine.
Nonetheless, Windows and Hadoop don't really like each other, and depending on your Mahout version (more specifically the Hadoop dependency it has), you will most likely run into this issue (link). The issue is present from Hadoop 0.20.204 onwards (although I must admit that I don't know if it has been fixed on the latest version of Hadoop)

Configuring Eclipse (on windows) for Hadoop environment(on vm having SUSE Linux Enterprise Server 11 )

I had Install Hadoop in pseudo-distributed mode on a VMware having SUSE Linux Enterprise Server 11. I am able to run the hello world examples like word count. Also I used WinSCP to connect to that VM and uploaded several XML files into hadoop cluster.
My question is now how can I configure my eclipse which I am having on my local machine that is windows 7 to connect that VM and write some java code to play with data I had dumped in the cluster. I did some work and able to get Map/Reduce perspective in the eclipse but not able to figure out how to connect hadoop on VM from my local machine, write my java code (mapper,reducer classes) to play with data and save the result back in cluster.
If someone can help me with this that will be great. Thanks in advance.
Let me know if more information is needed.
I am using hadoop-0.20.2-cdh3u5 and eclipse europa 3.3.1
I am struggling with this as well at the moment. Maybe you will find these links helpful:
http://www.bigdatafarm.com/?p=4
http://developer.yahoo.com/hadoop/tutorial/module3.html
Cheers

Categories