I got introduced to your channel by a Master's Student whom I am supervising. This is the first video I have watched, and I am super impressed. I don't know if it comes to you easily, or if it is as a result of much research, but all I can say is please keep up the good work. I am now addicted to your channel :).
I am so thankful that I found you! Thanks you so much for so many helpful and easy tutorials for everything! I wish I could even support you some how! Keep going with these tutorials, you don't have much followers, but it is okay, because you help people who really want to be like you in future!
Seriously. I am definitely in the deep end of the pool; but thanks to your methodical outlining of what you’re doing and why; along with issues you see or even made yourself. Thanks a ton.
Very useful information, thanks! I would really appreciate it if you could add a small video using this camera and IMU data streamed can you build an RGBD SLAM or RTAB?
Thanks, it is definitely on my to-do list! I didn't want to make it a core part of the series as not everyone will have a depth camera, but it is top on the list of "extension projects".
I'm really impressed with the quality of your videos... can tell you work really hard in the editing. your RGB and video overview with showing the compression via the frames in the video we were watching was great. Very good at general presentation as well... thanks man :) hope you have a good day today!
man , im bout to cry rn it took me 3 months to find a solution for my robot which is working with depth cam( based only on 2 cams on chip) and i couldn't and once i saw ur vid i jumped from my seat , man ty sooo muchhh
Hey Nigel - I'll be honest here. My current process is basically: write a script for myself, follow the script for the camera, then turn the script into a blog post. It used to be the other way around (starting with the blog post) which was better, because now unless I get the blog post up within a couple of days of the video I find myself feeling the pressure to get the next video out and the blog post never gets published. So I do have an 80% finished blog post for this one which I definitely intend to publish *some time* but I'm not making any promises as to when. If there are particular questions you have, head over to the discussion thread for this video on the forums and I'll try to answer them. (Because the forums only went up last week there's not been much activity there yet). discourse.articulatedrobotics.xyz/t/discussion-depth-cameras-and-ros-making-a-mobile-robot-pt-10/35
hi how are you getting topics automatically? I am doing everything the same ,the only difference is i am using .sdf file and not xacro(in ros-noetic). during simulation i can see the image of object shown by camera rectangle, but no topic is created in rostopic list. you are writing depth and all depth topic got created, how?
I like your videos they are very informative and well structured ... can u make a tuto about camera calibration with ros2 , I would like to see how u gonna explain that .
Yeah I had hoped the same but so far I think not, as the depth seems to be a bit worse and it doesn't have an IMU. I think the OAK-D (rather than OAK-D Lite) is a closer competitor. I still have hope that it will be good for entry-level prototyping and simpler projects.
Hey, I'm having this problem with my depth camera. I'm using Hector quadrotor package that has ASUS Xtion PRO depth camera and I'm trying to use it for 3D mapping in custom Gazebo world (apartment). The problem is when I subscribe to /camera/rgb/image_raw in Rviz by Camera it can see only certain objects, but most of the objects for example walls, postboxes etc. are just gray. The depth /camera/depth/image_raw works just fine and I can see objects, but still only in shades of gray. I tried to switch format in urdf of a camera from R8B8G8 to B8R8G8 and RB8 but nothing changed... Do you have any idea what could be the problem?
Hey, I actually managed to fix this problem by running this in terminal: "$ export SVGA_VGPU10=0", right before launching my .launch file that starts Gazebo and Rviz. It seems that the color not being visualised is an error and didn't find the explanation for that. You can also echo it in your bashrc script so you don't need to write it every time -> $ echo "export SVGA_VGPU10=0" >> ~/.bashrc
Awesome video! Here's an Idea or a bunch of Ideas for up coming videos... I want to build a smart RV... I would want it to cook, clean, find my stuff and bring that stuff to me. I need to build a set of robotic arms that run along a track hung from the ceiling (It could fold itself up and be out of the way when not needed), I would like to use visual slam, to map and constantly update said map as it moves (obstacle avoidance, where is my cup of coffee, etc.) also has to do object recognition obviously. I have seen that using machine learning the robot ai could learn how to do tasks from watching RU-vid videos using pose estimations and then using ai to find the most efficient way to preform the task and be adaptable to new problems as they arise by running simulations of said task several times before performing the task, and learn as it does the task! I'm also not sure about the best way to run this... Computation (AI) is going to need a semi-decent gaming laptop at minimum, but the robot arm is only going to require a raspberry 5 with ssd. It would essentially just run the stepper motors and sensors (pressure sensors in the grippers, cameras, etc.). Then the main brain needs to communicate with the ros in the arm to preform tasks. I would want it to all run local with a G.U.I.
I'm excited to find your channel as I'm planning to develop a firefighting robot designed for navigating smoke-filled environments. This robot would: Generate a 3D map of the room to navigate through smoke. Detect cracks and assess structural integrity to predict potential collapses. Incorporate thermal imaging for identifying hotspots and victims. Utilize AI for autonomous decision-making and obstacle avoidance. Feature a water or foam dispenser for firefighting efforts. I'd greatly appreciate any guidance or collaboration on this project.
Does this work for ROS noetic as well? I went to the previous tutorial for a regular camera and had no issues, I however didnt see many of the other topics being published. nonetheless, i managed to get the image_raw topic functioning in rviz. the issue now is that when attempting to change the type = "depth" it does not display the topics as expected. I am able to display only when i change the filename="libgazebo_ros_depth_camera.so"> , under image topic: /depth/image_raw and it shows an image where far objects are white while near are black in the PointCloud2 however, i try to put the /points topic, but there is an error that says: Transform [sender=unknown_publisher] For frame []: Frame [] does not exist and in the terminal : [ WARN] [1716868128.356292091, 23.441000000]: Invalid argument passed to canTransform argument source_frame in tf2 frame_ids cannot be empty
Thank you so much for such a grear toturial video. Do you have any plan to create a video how can we use depth camera in collaborative robots and do some pick and place applications? if not, is there any sources you might introduce to me.. I am new to ROS 😊
Hi! i'm looking for a Stereo Camera to use for a heads-up project, and I was hoping you would have some recs for one that is compatible with the Raspberry pi 4 B with the newest release of the raspbian os .
Has someone figured out a way to know the size of an object using the depth camera from the gazebo? I am currently having trouble with it. Any help is appreciated.
It depends on the model. Mine is auto-focus and I have found the autofocus a bit iffy. I really haven't tried to fix it though so it may be a driver setting that can be tweaked. I believe you can also manually control the focus on that model, which I would probably do.
Thanks for this great tutorial. Unfortunately in my workspace some thing arenˋt quit working as in yours. For Example I don’t get to see the pointcloud in Rviz and I'm also not able to select the camera info topic. Do you have an idea what might be the problem?
Thank you so much, your video helps so much. Anyway, I meet some problem when I try to do the same thing with OAK-D-PRO-POE, I can only see the rgb_image, but can not echo anything from left/right camera. Do you have any idea about this problem?