I have successfully set up an Apache Juddi v3 installation (tomcat version) on my computer. What I want now is to publish a service whose WSDL is found at
http://localhost:8080/axis2/services/CmmdcService/wsdl
To achieve this, I created a standalone Java application (starting from the Juddi documentation) that publishes the service found at the above location.
The publish part looks ok, but then I want to query the juddi database for the service but a field that should contain the found services is always null (getServiceInfos()). I really don't know what is wrong and I didn't find any good documentation or tutorial about this on the internet.
Here you can find the sources of the program. Just unarchive it and go to the ./publish folder. The application is found there.
With out much Apache knowledge, It sounds as if getServiceInfos() function is trying to retrieve information from the wrong sub folder when you do a query. Try changing the location of the search Function so that it will search all folders/locations or a specific folder/location where the database is located.
I could be wrong ( I have limited skills with Apache ).
Good luck, sorry if this confused you or did not help.
Edit: Sorry, I misread the question. I'm not sure what search criteria you've specified, but the server didn't return any results.
When using the "approximateMatch" find qualifier, you really need to specify a wildcard character, such as % (any number of characters) or _ (a single character).
Long story short, this is probably a bug that has since been fixed. Try a newer version
Related
I've successfully set up an Apache Juddi v3 on my computer ans know I want to create a client that can publish and discover webservices using java. I have searched in google but i could not find a clear explanation of how to do that as am new in webservice and uddi.
any help please!!!
You want to start with the jUDDI examples. There's a bunch of them.
http://svn.apache.org/repos/asf/juddi/trunk/juddi-examples/
http://svn.apache.org/repos/asf/juddi/trunk/juddi-examples/more-uddi-samples/src/main/java/org/apache/juddi/samples/
Note: you didn't specify what version you're using, so some of that code may not compile against what you're using. Most solutions should be obvious. If not, try getting checking out the full source (the readme has directions)
I intend to make a niche search engine. I am using apache-nutch-1.6 as the crawler and apache-solr-3.6.2 as the searcher. I must say there is very less updated information on web about these technologies.
I followed this tutorial http://wiki.apache.org/nutch/NutchTutorial and have successfully installed apache and solr on my ubuntu system. I was also successful in injecting seed url to webdb and perform the crawl.
Using solr interface at http://localhost:8983/solr/admin, I can also query the crawled results. But this is the output I receive. .
Am I missing something here, the earlier apache-nutch-0.7 had a war which generated a clear html output like this. . How do I achieve this... Or if anyone could point me to a latest tutorial or guidebook, highly appreciated.
A couple of things:
If you are just starting, do not use Solr 3.6, go straight to latest 4.1+. A bunch of things have changed and a lot of new features are added.
You seem to be saying that you will expose Solr + UI directly to general web - that's a really bad idea, as Solr is completely unsecured and allows web-based delete queries. You really want a business layer in a middle.
With Solr 4.1, there is a pretty Admin UI and, also, there is a /browse page that shows how to use Velocity to do the pages backed by Solr. Or have a look at something like Project Blacklight for an example of how to get UI over Solr.
I found below link
http://cmusphinx.sourceforge.net/2012/06/building-a-java-application-with-apache-nutch-and-solr/
which answered my query.
I agree after reading the content available on above link, I felt very angry at me.
Solr package provides all the required objects to query solr.
Infact, the essential jars are just solr-solrj-3.4.0.jar, commons-httpclient-3.1.jar and slf4j-api-1.6.4.jar.
Anyone can build a java search engine using these objects to query the database and have a fancy UI.
Thanks again.
I am trying to find synonyms of some words(String type) in java using Wordnet java api. I have difficulties though in figuring out how it works.
I found this link http://lyle.smu.edu/~tspell/jaws/doc/edu/smu/tspell/wordnet/impl/file/ReferenceSynset.html#getTagCount%28java.lang.String%29 which I though it is useful, but still I don't know how to start. Do I have to create a ReferenceSynset object and then find its synonyms, and how can this be done? Or is there another easier way? Please help!
Thanks in advance!
JAWS - "Java API for WordNet Searching" has been created exactly for this purpose. It is possible that you haven't installed it right. In the same domain that you mention (smu.edu) there are instructions for installing JAWS.
First, you will need to download the executable and Wordnet itself.
Get the *.jar file
Download the wordnet database files to the appropriate directory (instructions here)
Once you have done this you should next try the example program. First, make sure that it works unmodified. If you get that working, you should see a bunch of synsets when you type:
java TestJAWS <your word here>
If that works, you can start modifying the code to suit your purpose.
Hope that helps.
I want to find a library that I can use from my Java application that will allow me to access specific Javadoc in the scope of my project (I specify where Javadocs are located). Just like in Netbeans, I want to potentially access the Javadoc from html files locally and remotely, and from source.
I expect that I could use code from Netbeans to achieve this, but I don't know how, and I can't easily digest their documentation.
Today I started thinking about the same thing.
From CI point of view, I could use #author annotation to send e-mail to someone, who wrote a test that is failing with error, not with a failure.
Google didn't help me (or I didn't google deep enough), so I started wondering how to do it on my own.
First thing that came to my mind is writing a little tool that will check all *.java files specified in a directory, bound file name to annotations and allow user to perform some actions on them.
Is that reasonable?
We have recently installed a Google Search Appliance in order to power our internal search (via the Java API), and all seems to be well, however I have a question regarding 'automatic' site-map generation that I'm hoping you guys may know the answer to.
We are aware of the GSA's ability to auto-generate site maps for each of its collections, however this process is rather manual, and considering that we have around 10 regional sites that need to be updated as often as possible, its not ideal to have to log into the admin interface on a regular basis in order to export them to the site root where search engines can find them.
Unfortunately there doesn't seem to be any API support for this, at least none that I can find, so I was wondering if anyone had any ideas for a solution/workaround or, if all else fails, the best alternative.
At present I'm thinking that if we can get the full index back from the API in the form of a list, then we can write an XML file out using that the old fashioned way using a chronjob or similar, however this seems like a bit of a clumsy solution - any better ideas.
You could try the GSA Admin Toolkit, or simply write some code yourself which just logs in on the administration page and then uses that session to invoke the sitemap export URL (which is basically what the Admin Toolkit does).