how to print an image using wifi/Bluetooth connection in android - java

Actually I have a problem that I have to print an image or text using android app using Wifi/Bluetooth connection.
But i am not found any resource for such type of printing.
I see the app printshare. i have to do such like that.
If some one have any solution then please help me

Some Bluetooth printers like Woosim has its own library and sample for Android. Refer to this link. I have used it for my project and was useful.

Related

Sceneform: Render WebRTC media stream in to ExternalTexture

I am trying to develop an AR app to help visual impaired people to make better their access conditions to a computer.
I am investigating on how AR can help HCI for visual disabilities, so, the application is using WebRTC to get computer Desktop to be magnified at AR environment using Sceneform.
I have successfully used the Sceneform example https://github.com/google-ar/sceneform-android-sdk/tree/master/samples/chromakeyvideo, but, I have no idea of how to render the WebRTC stream directly to a ExternalTexture. -> https://github.com/judicapo/MUITSS-ARBE/tree/master/SampleApps/ARCK
I already tried some Stackoverflow answers, but, have not found the clue.
Thank you all for your replies, hope some one has any idea.
Honestly I have not worked with this. but I could think of a way on how to approach this.
Instead of chromaKey rendering on a texture, why don't you try a 'ViewRenderable'? using this, you can place any android View to a node. You just need to place a VideoView and do your webRTC magic. let me know if this works
https://developers.google.com/ar/reference/java/sceneform/reference/com/google/ar/sceneform/rendering/ViewRenderable
https://developers.google.com/ar/develop/java/sceneform/create-renderables
AugmentedImage example - https://proandroiddev.com/arcore-sceneform-simple-video-playback-3fe2f909bfbc
ViewRenderable example - https://github.com/Hariofspades/ARExperiments/blob/master/AugmentedImages/app/src/main/java/com/hariofspades/augmentedimages/common/AugmentedImageNode.java

Implementing Clarifai API

I need some help with connecting hte Clarifai API into my android app. It's for a uni project where we tests different image recognition softwares. All it needs to do is take a picture on the phone and then run recognition on the pituces (which will be buildings).
I have no idea how to do this which is why i'm throwing this hail mary in the hope that someone would help me. I've put what i have in a drive here: https://drive.google.com/drive/folders/1LH79C0JtpBBpAMdKqNjKfYfj3KTvLgAh?usp=sharing
How can I solve this?
To use the Clarifai API on Android, you should use the Clarifai API Java client. Alternatively, if you wish not to install the library, you can do REST and execute HTTP requests directly using JSON. See the developer's guide for examples.
Note: There's also an Android SDK in works, which is beta at the time of this writing. See here on how to join the beta program.

How can I stream videos from a server to an Android app?

My local church hosts a weekly show and upload the videos on Youtube. And I've been asked to develop an app for them, so far all is well, however, I have to implement a feature that will allow me to stream the show's videos unto the app.
I honestly don't have much of a clue how to go about it (still new to android programming)! I've looked around here and seen that a few people have already attempted to do this, but unlike what most people have done or tried to do, the videos are released on that day (not live streaming). For example the 10/02/13 video will be released on 10/02/13.
This link gives me an idea on how to stream a video on Android, but that shows that I have to explicitly put in the link for the video.
Is there a way to do the same thing, but instead of adding the link myself, the app should retrieve the videos from a server??
Thank you in advance for your help!
Is there a way to get the link but instead of adding the link myself, the app should retrieve the videos from a server??
To achieve this you need to make the webservice which help to communicate mobile and web server. When you send the request from mobile to server then make the web service which give you response which have link of video then you can stream the video.
To stream the video.
Read here.
Inside the API DEMOS you have on example.
API DEMOS >> Media >> Mediaplayer >> Play Streaming Video
You need to downloaded using the SDK Manager, and then inside eclipse you go throw create new project form existing sample, then you read the code and copy it inside your app.

Android, use library or copy source code?

I hope my question is suitable here. I recently started learning android - making buttons, using google maps and other basic stuff. I saw a thread recently (can't find it) about this and I got interested in it: https://github.com/jackpal/Android-Terminal-Emulator/wiki
This is an open source android terminal emulator. It also seems to have a library called EmulatorView with which people can call methods etc. What I would like to do is use this app inside my own app. Instead of just looking at the file system on the android phone etc. my app would connect to the console of a serial device like a router, usb to serial. Something I usually use minicom for on my pc. As in there could be a button in my app and if I press it it opens the console of the attached router and I could issue administrative commands.
I was wondering what the easiest way to do this would be? Would it be to try and use this library and make my own terminal, seems quite hard, or would it be to copy their source into a project and try to edit it? Seems messy to organize everything, and could be hard too? Then use some library like this to connect to the serial device? http://slickdevlabs.com/slick-usb-2-serial-library/
What would the first steps be, make a terminal like the open source app has, use their library or paste in their code? Then try and edit it to do something over serial?
I have tried to use their example for using their library (jackpal.androidterm.sample.telnet), but it crashes when I click open shell, I was wondering if this open shell button in their example was what I needed?
I decided I would learn more by writing the app myself using the library.

how to convert j2me running lwuit to android?

How can i convert a J2ME application which is using lwuit library to an android apk ?
I tried to convert using mircoemulator but when I start the application it shows error and exits.
Read this article, LWUIT for Android. Here clearly telling about How to do LWUIT with Android. Follow the stpes carefully and change the your MIDlet code. Just checkout the LWUIT-Incubator repository and use this for android development with LWUIT. Also read the README_Android.txt on this repository.
There's no way to make a direct conversion. Strictly speaking, you can make a port, adapting the code, or use some emulator or wrapper to run the CLDC/MIPD classes.
The guys in Opera also used MicroEmulator for OperaMini.

Categories