ok, sorry, my bad again (it's getting late :-)). I messed up hexadecimal and binary notation here.
anyway, I'm stuck. from the code above, you can tell that packet[1] is responsible for the greatest part of the X coordinate value. when I put my pen in the top left corner, this packet contains a decimal 1. in the bottom left corner, a decimal 1 again. but somewhere in the middle, it gives me back a 2. in the top right corner, it's a 6, whereas in the bottom right corner it's a 5.
how can one ever convert such results into decent pointer coordinates?
I'm guessing something is missing, or going wrong, at the device initialisation. any hints from anyone?
ok, sorry, my bad again (it's getting late :-)). I messed up hexadecimal and binary notation here.
anyway, I'm stuck. from the code above, you can tell that packet[1] is responsible for the greatest part of the X coordinate value. when I put my pen in the top left corner, this packet contains a decimal 1. in the bottom left corner, a decimal 1 again. but somewhere in the middle, it gives me back a 2. in the top right corner, it's a 6, whereas in the bottom right corner it's a 5.
how can one ever convert such results into decent pointer coordinates?
I'm guessing something is missing, or going wrong, at the device initialisation. any hints from anyone?