Inconsistency with the Unity Remote and the new Input System

Hello!

At the moment it is very inconvient to debug a touch screen app with the unity remote app(android).

Multitouch on the new GUI and event system does not seem to be fully supported. At least I cannot press two uGUI buttons at the same time over the unity remote which is possible in full builds.

Also pointerIDs seem to be wrong. I am processing the IDs of pointerUp and pointerDown events from IPointerDown/UpHandler and with the Unity remote I get wrong IDs. IDs of pointer up and pointer down are not consistent.

Let me explain:
If I touch the screen with the first finger I get pointer ID 0 - fine. Second finger: pointer ID 1 - also fine.
Now I release the second finger and IPointerUp should return pointer ID 1 but it actually returns ID 0 as if I had released the first finger (which had the ID 0).
To be clear: This is resolved in full builds and works as intended…

I think it is really annoying that we cannot test our apps properly with the unity remote app and have to deploy a build each time because for some projects building the apk takes a few minutes … :-/
This makes debugging touch input based features a real pain.

same thing here. and the documentation is very poor on that aspect. it took me a while to realize that once on build it works fine