-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Help with oadk #6
Comments
I personally haven't run this on humble, but I expect it should work without issue. I also have not used that particular sensor, but from a quick search it seems like it should work. A few tips to help debug your issue:
|
i just checked the camera callback is not called could you please tell me where to call the funtion |
It should automatically get called when images are received. Ensure that the |
I have not made any changes to the code i am running it exactly the way its mentioned in the repo first call the ros2 launch after that start reconstruction and then stop reconstruction |
And one more thing in the repos readme why is it just launch ut should be launch.xml right? |
You are correct, that is a mistake.
You might have to modify the service calls slightly for your application, the examples provided are intended to show you a template to follow, not explicit instructions to follow precisely. If your camera driver publishes the color and depth image on different topics you need to ensure you are specifying the correct topic names. Additionally, you need to make sure to give the correct tracking and relative frame names for your given robot set up. I would first advise you to ensure you can view the camera images as ROS topics in RVIZ or RQT. |
So i need to change the start reconstruction function i am new to ros but i am good with open3d and python |
Yes, change it to fit your application. Also the robot and camera drivers must be running and publishing information. These instructions assume that you already have robot and camera data publishing in your ROS 2 environment. It looks like a driver for your camera exists here. You might find some useful information for setting things up there. |
Wait so i need to run a robot i am working with ur10e so its mot just camera but even the robot has to be running? Could you pls explain it to me |
The robot needs to be publishing joint states in your ROS environment, so you need to connect over ethernet to the UR controller while the robot is on. This application assumes you have a camera mounted at the end of a robot and that the robot is moving through a trajectory while the camera is collecting data. The python code uses the camera's estimated position in conjunction with the depth and color images to stitch together a mesh using TSDF. In order for it to have access to all of that information you need data from both the robot and the camera, and a URDF defined with the camera mounted in the appropriate location at the end of the robot. You can find the ROS2 UR driver here. If you are completely new to ROS it might be worth going through our ROS-Industrial Training which walks you through creating a workspace and setting up URDFs.There's specifically an example where we walk through adding a UR robot to your URDF. |
Thank you soo much |
|
If the topics are being published and you run everything correctly you should at least see something print out. You'll need a robot to get it fully working though. Verify you can visualize the camera topics with RVIZ or RQT
You only need 1 camera
Yes, the camera needs to be defined in some frame that is moving |
|
This was primarily tested with intel realsense cameras, but I believe others have been successful with other cameras.
The Image and CameraInfo msg types are normal ros message types. If you are having trouble accessing these then either:
|
After following your steps i got this [industrial_reconstruction-1] 2023-08-18 10:28:14.439 [RTPS_TRANSPORT_SHM Error] Failed init_port fastrtps_port7414: open_and_lock_file failed -> Function open_port_internal |
so for rviz we are using depthai-example in rviz we are able to get the depth properly |
Failed to get transform: "world" passed to lookupTransform argument target_frame does not exist. i was able to publish the data to the cameracallback funtion |
after completing this i am willing to contribute to his repo by making a branch for oakd |
Sounds like potentially the world frame doesn't exist in your environment. Ensure in RVIZ you can see both the target frame and the reference frame in the TF display. |
So were able to generate the mesh we fixed the issue we were facing the data is getting saved in archive folder but when try to access the archive data using rviz command as sooh as i give ros2 service call /start_publishing std_srvs/srv/Trigger the rviz unexpectedly stops could pls help with it |
There could be an issue with that rviz config file, maybe something changed or maybe it improperly references something. You should be able to launch rviz separately and just add the display fields manually. The intent of the rviz parameter that loads that config file was just convenience, it is not required. |
Good Greetings
|
Look at discussions on #1 for how to improve mesh reconstruction results.
I think access to more mesh processing algorithms would be good. I don't want to make the call to the reconstruction process more complicated though. Perhaps additional services can be created specifically for modifying the mesh, or some other clean implementation of allowing processing without adding complexity to the process would be good. |
|
Good Greetings Everyone
I am using this repo with ros humble the camera that i am using is oakd the problem is i am not receiving any data the folders are also empty
when i try to log the triangle mesh it shows it has 0 vertices help
The text was updated successfully, but these errors were encountered: