[Q] linking two android tablets for reading - Android Software/Hacking General [Developers Only]

Because I have waited long enough for hardware music reader I thought about some other solution. I would like to know if this is possible to accomplish it with the software.
I want to make the ebook reader (pdf the most) that could work on two devices simultaneously. The software would use bluetooth link between two devices and make the turning pages commands from either device. The same document would be loaded and open on both tablets. One device would display even, the other the odd pages. Touching right screen would turn pages on both devices, left touch would turn back. With today's tablets prices and power (Le Pan III) that could make a double display with 1536X1024 resolution and 2 x 9.7inch screen size together quite interesting solution.
Could you tell me if it is doable or should I stop dreaming....

Related

Multitouch on a single touch screen.

This guy made it on a nokia...
http://30dbs.blogspot.com/2009/03/experimental-multi-touch-on-nokia-5800.html
i dont think this is real multitouch- just seems to remember the first pressed location and notices when pressure is gone
its not really multitouch.
i worked on a simular project (not a game, just tried to do something like virtual multitouch)
and its just like that.
example:
you press on the left and on the right side of our screen, wm thinks you were pressing just in the middle of both sides.
like this (X is your finger, O is the position where windows mobil thinks where your finger really is)
X-----O-----X
or
X----------
-----O-----
----------X
so i think this guy just made "invisible buttons" which you click when you press und 2 of his buttons at the same time.
of course this is just speculation because i cant test it, but that would be the easiest way.
Yep I think it's how it works... But it's not a dumb idea at all, and could be used in the dev of apps
freaksey said:
you press und the left and on the right side of our screen, wm thinks you were pressing just in the middle of both sides.
Click to expand...
Click to collapse
I agree with you.
This "trick" can simulate a multitouch but only with single clicks. I'm afraid that you can't simulate gestures such as the "iPhone-like" zoom.
Marshall
lpaso said:
Yep I think it's how it works... But it's not a dumb idea at all, and could be used in the dev of apps
Click to expand...
Click to collapse
Yeah, I realize it isn’t "real" multitouch to, but I also think this could be useful for a lot of cool apps on WM.
I was making some experiments and I noticed that it is really hard, if not impossible, to make it work with 3 simultaneous pressures.
Like freaksey said, pressing in 2 different points of the screen will make the OS think we are pressing in a point in the middle. But introducing a third point will not be detected if the point is in the same line.
Here is what I mean:
X----X----X (3 points)
X---------X (2 points)
X----O----X (What the OS thinks in both situations)
So I'm afraid that we can't make applications like "piano".
Marshall
Marshall07 said:
I was making some experiments and I noticed that it is really hard, if not impossible, to make it work with 3 simultaneous pressures.
Like freaksey said, pressing in 2 different points of the screen will make the OS think we are pressing in a point in the middle. But introducing a third point will not be detected if the point is in the same line.
Here is what I mean:
X----X----X (3 points)
X---------X (2 points)
X----O----X (What the OS thinks in both situations)
So I'm afraid that we can't make applications like "piano".
Marshall
Click to expand...
Click to collapse
Indeed, not like a piano, but for tiny games (like the nokia guy made) it could be fun
lpaso said:
Indeed, not like a piano, but for tiny games (like the nokia guy made) it could be fun
Click to expand...
Click to collapse
Of course
Marshall
I reckon u could simulate multitouch for zooming. And here is how i suggest to do it (im not a programmer, so someone come use it (Y)!)
When using a transitional touch screen, a user either presses a button and then releases it (taps it), and then taps another, or drags a stylus across the screen. When a user taps, there is a space of a few milliseconds of no input between the taps, so is recognisable.
When a user drags a stylus, the change in location is continuous (i.e. the point of contact moves to an adjacent sensor), so is recognisable.
If a user were to touch 2 points, the point of contact would appear to move straight to another point (not adjacent) immediately. This would then be recognisable as 2 points of contact, rather than just 2 taps.
This would allow 2 fingered gestures, even if the system wouldn't know the exact location of the 2 points of contact.
try this... here can you see where your phone thinks your finger is.
(to go into "singlepoint mode" draw at least 1 point on your screen and then you can change into it)
just copy to phone and run it
I am not a hardware engineer by any stretch of the imagination, but I feel this is more of a software limitation than anything else.
The fact that windows mobile can take two points of pressure and use the middle of the area as the true touch point, tells me that the hardware is understanding both touch points.
I believe the issue lies in Windows for not understanding two inputs at the same time. This is the same case with Windows PC's, as multi-touch is not possible on Vista or XP, but instead is being developed specifically for Window 7. Same thing with Windows Mobile, multi-touch is being developed (on a software level) only with WM 7.
Correct me if I am wrong, but the Android powered HTC phone has similar touch screen hardware as most current HTC devices, but a developer was able to (with some changes to the OS) allow for true multi touch recognition.
You are right, viridescent_zeal.
But I already tried what you are saying and if you press 2 point "at the same time" the hardware will immediatly get the point in the middle. It will not move immediatly but it will immediatly result as the point in the middle.
I'm afraid that it is an hardware limitation. If you want to realize the behaviour you are talking about, you have to press the two points with a little delay. This will cause the immediate movement and you can make a software believe you have touched two points in the screen.
Marshall
iservealot said:
I am not a hardware engineer by any stretch of the imagination, but I feel this is more of a software limitation than anything else.
The fact that windows mobile can take two points of pressure and use the middle of the area as the true touch point, tells me that the hardware is understanding both touch points.
I believe the issue lies in Windows for not understanding two inputs at the same time. This is the same case with Windows PC's, as multi-touch is not possible on Vista or XP, but instead is being developed specifically for Window 7. Same thing with Windows Mobile, multi-touch is being developed (on a software level) only with WM 7.
Correct me if I am wrong, but the Android powered HTC phone has similar touch screen hardware as most current HTC devices, but a developer was able to (with some changes to the OS) allow for true multi touch recognition.
Click to expand...
Click to collapse
It isn't a software limitation. This is just how resistive digitisers work.
here's two artciles from the last 24 hours showing multitouch in action:
http://www.fuzemobility.com/multitouch-exists-now-without-new-hardware-or-software/
http://www.fuzemobility.com/multitouch-in-action/
it can be implemented for games in a butchered form...it's not real multitouch but virtual key presses can work for games.
Very true. Still, be able to do that is better than nothing- if done right it could simulate multi-touch gestures providing we put 1 finger down first!

[IDEA] Writing on screen as quick launching

Dear All developers,
I have an idea for who interesting.
Have you ever seen Infill G4 Car pc? You can search it in google.
The thing that I mentioned it becasue its functunc.
You can write some letter of symbol on its touch screen then it will launch program automatically.
For Example, if you use your finger write "N" letter on the big screen during the screen is on, it will launch naviation software.
I think we adapt this concept on most every touch screen mobile device!
Don't you think, if we are on today screen of our WM phone, when we write "N" on screen, it's launching Igo or else that we preset. When we are in using some program, we write "O" on screen it back to home. We write "Z" it do something.
Do you think it might be more funny and functions?
And it might very useful on such a huge 4.3" screen device like we all own.
Something similar to your idea, but not using letters - currently only certain swipes. You should post your suggestion there I reckon.
http://forum.xda-developers.com/showthread.php?t=629263

Drop down menu with touch on surface, problems

Aloha all, Having trouble knowing how to ask this...bear with me.
I am having problems with the menus that are triggered by the 'hover' action of a mouse. Using the surface pro or even my lumia 920 I am frequently unable to operate many menus on sites.
officefootballpools.com tournamentpools.com and a host of other sites have the same type menus.
Normal action would be with a mouse and when you hover it drops down a menu that you click your selection. Its isn't a normal drop down but a .li. html type for the menu.
Im sure this has been discussed but I was looking for some vocabulary so I can search the right threads.
You cant interact with hover images via capacitative touch. The browser has no way to tell if you want to click or just show the drop down, it assumes click. Some sites (with limited success) I have managed to press and hold on a link to show its drop down, then if you ignore the usual right click popup then you can sometimes hit the correct icon, this is on a lumia 710.
Otherwise if you have a device with an active stylus (the surface pro for example) you can get rollover easily. If you notice when you hold the pen a tiny bit off of the screen you get a little dot appearing on the screen where the pen is pointing without any physical contact. This dot can trigger rollover events as in this case windows knows that you are not touching the screen, touch becomes a click, hovering the pen over the screen becomes mouse movement which is enough to trigger rollover.
Thank you for your reply. I understand and assumed the same. I have been working with an x61t for a few years so the jump to capacitive has thrown me a bit.
Okay, so what are web developers doing instead of this type of drop down? I recently read an article that says a lot of developers are moving to the new msn.com type format with nav bars on left or right with no hover...
Anyway thanks again. The volume of sites with this issue is pretty large. Yahoo.com/fantasy chokes a donkey, etc.
Side question, does Win8 have the on screen mouse that I used to have on my Lenovo? Or is that a Lenovo product probably?
Alot of web developers are doing nothing at all. In some cases clicking the link that causes the dropdown redirects to a page listing the other links in that dropdown, that's always handy. Usually most devs create mobile versions of sites which are normally touch friendly. I have seen 1 or 2 sites create iPhone versions before, these worked nicely on android so I would assume they are fine on the lumia and maybe the surface.
I haven't ever seen an on screen mouse before but if there is not one in windows 8 then there may be a 3rd party one somewhere. I am on my phone right now, otherwise I would have looked myself.
One advantage of using "Mobile" websites is that they should be designed with the limitation of touchscreens - specifically, the inability to track hover - in mind. That may help you out.
Alternatively, the Surface Pro uses an active Wacom digitizer (as well as a touchscreen) that can sense the pen at a distance. You can use the stylus pretty much perfectly as a mouse, with hover and right-click and everything.

Throw fruit, turn Kindle pages with eyes (5 dollar eye-tracking). Android SDK

**Integrating an eye tracker into the hardware**
“The Eye Tribe released its first eye-tracking product to developers in December -- a long, thin $99 module that attaches to a Windows laptop, computer or tablet. It sold out immediately and the company is now working on a second batch. But it also has a more exciting proposition in the pipeline -- a software development kit module for Android phones that it eventually wants to see integrated into the a wide range of mobile devices.
“Most of the requisite hardware is already built into phones. The Eye Tribe just needs to persuade companies to integrate the technology.
All that's required is a camera sensor with infrared capabilities. "What we know is that in Q4 this year, sensors are coming out that can switch between regular camera and infrared camera."”
Click to expand...
Click to collapse
wired/co/uk/news/archive/2014-02/25/eye-tribe-android
**Cost**
“OEM vendors could likely add this sensor to their handsets for just five dollars”
Click to expand...
Click to collapse
reviewscnet/eye-tribe-shows-off-working-eye-tracking-on-a-mobile-phone/
If modifying the device to add eye-tracking only adds 5 dollars to the manufacturing cost, then I’m sure that at least one of the smartphone, tablet, notebook, and laptop manufacturers will make the supposedly easy camera modification.
**See before touch**
I think that most of the time, a person will see a widget that they want to touch before they actually reach out, and physically touch it
(The only times where I’m not looking is when I press the Android Navigation Bar buttons that are near the bottom edge of the screen. Although, on a larger Nexus 10, I usually have to look at them first).
**Eyes + consecutively touching the same few buttons**
On certain tasks, it might be convenient and fast to have the option of touching “single tap where I’m looking”, and “swipe up where I’m looking” buttons. You would only need one or two buttons that are close to you (kind of like the Navigation Bar buttons at the bottom).
Look, touch an easy-to-reach spot, look, and then touch the same button again. You don’t have to keep changing your hand and finger positions between each tap.
“Looking at icons on a desktop instantly highlights them, and you can then tap anywhere on the screen to open up the selected app.”
Click to expand...
Click to collapse
stuff/tv/mwc-2014/eyes-eye-tribe-we-play-fruit-ninja-using-nothing-our-eyeballs/feature
I guess that in one of their demos, they temporarily made the entire screen a “tap where I’m looking” button.
Besides the three default buttons in the Navigation Bar, you could add “single tap where I’m looking”, and “swipe up where I’m looking” (to perhaps simulate a Page Down for reading) buttons, and those alone should allow you to do a lot of things).
Vertical touchscreen
If you have a vertically propped up tablet with an external keyboard, you could remap a keyboard button to be the “tap where I’m looking” button.
**Hands-free interaction**
Even without the above options, I still think the ability to have a page automatically scroll down when your eyes reach the bottom of the page, or have an e-book automatically turn the page when the gaze reaches the corner of the text would be pretty good features to have. They would be especially handy for computer interaction while cooking and eating, and interacting with a vertically set up touch device, or laptop that is more than an arms-length away while you do other stuff on the desktop.
(**Google eye tracking patents**
Notably, Google has an eye tracking patent that involves recording advertisement impressions through eye focus with pay-per-gaze, and another patent that demonstrates a method to unlock a device by having a sensor in a head-mounted accessory (probably something like Google Glass) track the patterns of the pupil.
It indicates that eye tracking could have even more backing in the near future).

Disabling Android system commands on bluetooth/usb Keyboard and Mouse buttons?

Hello XDA community!
I've done brief google-fu trying to work around issue I have found myself struggling with with nothing to show for it so I come to ask experts because surely, this being such a fundamental problem someone must've already tackled it, right?
So my problem is, I try to use my Android device (in this case POCO X3 NFC, but considering getting Android tablet specifically for this purpose if there is a workaround to issue I'm about to describe since my iPad is nothing more than expensive paper weight it seems) as a client device to access remote desktop and use BT mouse and keyboard to go along with it. Everything is sorta kinda working... except it doesn't as soon as you get to interact with it. ESC and few other keys prioritise Android shortcuts over application, same with right mouse button. I don't think I need to explain why right mouse button closing remote desktop app (Parsec and Moonlight) and booting me back to app launcher instead of opening the context menu on remote desktop is a MAJOR problem. I've found TeamViewer to work alrightish, it catches mouse clicks and button inputs and prevent app from closing itself... if it only wasn't a steaming pile of dog**** to work on - obviously desire here is to keep using VASTLY superior remote desktop experience that either Parsec or Moonlight provides with the benefit of using BT mouse and keyboard.
Is there ANY way that Android system button shortcuts could be completely disabled? I don't want my F-keys to adjust volume... Nor do I want my ESC key or mouse right click boot me back into app launcher...
Also, as a bonus if there's a way to significantly decrease mouse cursor sensitivity it would be greatly appreciated because holy **** moving mouse 1 inch on a desk on lowest setting Android allows for would scroll through several dozen screens worth of content...
I've found this to be a problem on iOS as well but unlike iOS being left at Apple's mercy I kinda expect Android to be more maliable by end-user.
Any takers?

Categories

Resources