IMU data should be easy to get as soon as we have stable pipeline.
But we don’t have D435i to test its IMU, so I would ask you to test it when we’ll reach the point.
@robotanton , thing is we need the realsense in vvvv beta and don’t care much about how it is implemented in VL. it would be nice to have an easy way to test the thing instead of this 4 steps instruction inlcuding having to compile things and making references somewhere (which doesnt work).
Thanks Anton,
I tried with Alpha 64 & 32 but no menu in VL to go in Commandeline.
When I go to User…\Documents\vvvv\beta-preview\Packages> nuget install VL.Devices.RealSense -prerelease
Nuget is not recognize…? is this normal
In User…\Documents\vvvv\
there is only beta-preview…
Great work @robotanton ! Works out of the box and as aspected. At least the vvvv Gamma part. :)
The range for the D435 and D415 are specified at ~11m. Well… didn’t expected good results in this distance, but my cam (D435) seems so be limited already at a depth of 3,7m. Tried different settings and light-situation but nothing really changed. What did you exprienced @robotanton or @u7angel, what was your maximum z-depth with the RealSense?
Any chance, to get a “raw” depth image coresponding to the Kinect-depth image with floats in the R-channel for the distance in mm?
Searched fo it and tried several node-combinations and channel-settings for the depth-stream inside the “Config”-settings of the RealSense-node. But no success. Also the “Depth (Raw)”-node seems not work as expected.
Good news though: We managed to add the PointCloud-output as pos-buffer. Which works out of the box with DX11.Particlepack. So, fast homebrewn gpu-pointcloud-visualizer: ✔
But would be nice to also approach this texture-based, as described above. Any ideas?
@timpernagel Could you please provide an example or just explain the way to have a homebrewn gpu-pointcloud-visualizerGpu-Pos-Buffer in V4 with Particlepack ?
I just received my D435i, super tidy little camera! I have loaded up the demo patch, but not sure how to get it out into V4, and then how to pull that data into V4 to manipulate it, be that the pointcloud, or jhow to pull the video feed out into a dx11 renderer. Pretty exciting stuff!