Posts: 2
Threads: 1
Joined: Mar 2024
Institution: ZJNU
03-19-2024, 10:38 AM
Hello!
I want to use the phyphox on the mobile phone to read sensor data and transmit it to the embedded device through Bluetooth for processing. I know that phyphox editor has the output tag interface, but I don't know if debugging is the same as input.
In addition, may I ask whether phyphox can read the saved images or real-time drawings on the mobile phone to obtain data.
I would appreciate your help!
Posts: 543
Threads: 24
Joined: Apr 2020
Institution: RWTH Aachen University
There is quite some information on Bluetooth Low Energy (BLE) communication in our wiki,
https://phyphox.org/wiki/index.php/Bluetooth_Low_Energy, and you also find on that page a link to our Arduino library phyphoxBLE,
https://github.com/phyphox/phyphox-arduino (now also being available for MicroPython on ESP32,
https://github.com/phyphox/phyphox-micropython, however, lagging a bit behind).
There has been some work done on the back channel for smartphone data in the latest update to phyphoxBLE, so you might start with a look at phyphoxBLE.
Phyphox can not yet interpret images/pictures. We are working on the camera being a future sensor input. I am not sure, though, if it would for your application. What exactly do you have in mind?
Posts: 2
Threads: 1
Joined: Mar 2024
Institution: ZJNU
(03-20-2024, 05:31 PM)Jens Noriʇzsɔɥ Wrote: There is quite some information on Bluetooth Low Energy (BLE) communication in our wiki, https://phyphox.org/wiki/index.php/Bluetooth_Low_Energy, and you also find on that page a link to our Arduino library phyphoxBLE, https://github.com/phyphox/phyphox-arduino (now also being available for MicroPython on ESP32, https://github.com/phyphox/phyphox-micropython, however, lagging a bit behind).
There has been some work done on the back channel for smartphone data in the latest update to phyphoxBLE, so you might start with a look at phyphoxBLE.
Phyphox can not yet interpret images/pictures. We are working on the camera being a future sensor input. I am not sure, though, if it would for your application. What exactly do you have in mind?
Thank you for your reply.
I am designing a device to demonstrate laser scanning, so I have to design my own software to transmit the scanning pattern to the device.
Because I don't want to edit a set of software and I found in the depth sensor experiment in phyphox that phyphox called the camera to measure, so I want to try whether I can use phyphox to transmit image data.
Posts: 543
Threads: 24
Joined: Apr 2020
Institution: RWTH Aachen University
The “camera” in the LiDAR experiment is just for the preview window in order to select a target. The data is taken from the LiDAR Scanner that provides us with a depth mesh. At the moment, only a processed depth value is accessible in the phyphox file format (nearest, average, weighted depth of a part of the mesh).