Quantcast
Channel: ROS Answers: Open Source Q&A Forum - RSS feed
Viewing all 348 articles
Browse latest View live

How to use image data set to simulate a camera input for libviso2/ viso2_ros ?

$
0
0
Hi everyone I am new to ROS Indigo and am trying to work with the libviso2/ viso2_ros package for visual odometery. I have a fair enough idea as to how to use the package/node with a camera, but I have little idea as to how to make it work with pre-collected dataset of images. The datasets I am trying to experiment with are any of [these](http://www.cvlibs.net/datasets/karlsruhe_sequences/). Libviso2 looks for a rectified camera image. Please let me know if any other information is needed. Any help would be greatly appreciated.

USB Camera is not running - Book Learning ROS for Robotics Programming

$
0
0
I am using Ubuntu 15.04 and jade version of ROS. I am trying to use the usb camera by the method given in the book on page 187. My camera is already good and I checked it by the "cheese" method explained in there. Here is what I am doing $ sudo apt-get install ros-jade-usb-cam $roslaunch chapter5_tutorials usb_cam.launch view:=true here is what I am getting after doing this [ERROR] [1453172400.684518103]: Webcam: expected picture but didn't get it... [ERROR] [1453172400.684518103]: Webcam: expected picture but didn't get it...

do rviz need urdf file for viewing camera feed?

$
0
0
I am trying to create a stereo camera, I know that the feed can be viewed in `rviz` using `image` display type.But other than that i want to know how can we succesfully view camera feed using `camera` display type - i was successful in calibrating and publishing the `image_info` topic - but when try to view the camera using `camera` type i get warning status saying no image was received but i successfully gets the image via the `image` display type ![here i have attached image](/upfiles/14536873284636959.png) - so my suspicion is if I am missing something on the Fixed_frame or target_frame or something like that

depth_image_proc register issue

$
0
0
I'd like to use the depth_image_proc pipeline to create point cloud xyzrgb using a RGBD camera. I calibrated rgb and ir camera and used image_proc to rectify them. Now I would like to register depth image to the frame of the rgb camera. All input topics are up - depth image, depth camera info, rgb camera info and the required tf is also there. The topics depth_registered/camera_info and depth_registered/image_rect are generated, but they are empty. The question is why? ![image description](/upfiles/1453815158177028.png) ![image description](/upfiles/1453815173355060.png)

ar_sys camera receive problem

$
0
0
hi. i am beginner. i wanna run ar_sys package but i have a few problems. i have usb cam or can i do that with my laptop's camera ? **Firstly :** "Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist" probably my camera is not define to ROS. How can i do that ? **Secondly :** i try to run this line : rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100 But i have this error: Fixed Frame [world] does not exist **Thirdly :** Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)

ar_sys camera problem

$
0
0
hi. i am beginner. i wanna run ar_sys package but i have a few problems. i have usb cam or can i do that with my laptop's camera ? **Firstly :** "Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist" probably my camera is not define to ROS. How can i do that ? **Secondly :** i try to run this line : rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100 But i have this error: Fixed Frame [world] does not exist **Thirdly :** Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)

ar_sys camera problem

$
0
0
hi. i am beginner. i wanna run ar_sys package but i have a few problems. i have usb cam or can i do that with my laptop's camera ? **Firstly :** "Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist" probably my camera is not define to ROS. How can i do that ? **Secondly :** i try to run this line : rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100 But i have this error: Fixed Frame [world] does not exist **Thirdly :** Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)

How to visualize the data of "5-Megapixel camera" and "LED Texture Projector Triggered with Narrow-Angle Stereo Camera" in rviz?

$
0
0
I want to use rviz to confirm the functions of "5-Megapixel camera" and "LED Texture Projector Triggered with Narrow-Angle Stereo Camera" works well, but I cannot find the correct topic to visualize the camera data, does anyone know how to view in rviz? EDIT: I found [a page](http://pr2support.willowgarage.com/wiki/PR2%20Manual/Chapter9#Head_Cameras) that may be useful for this question. Now my understanding is, projector illuminates the objects. To visualize the data of "LED Texture Projector Triggered with Narrow-Angle Stereo Camera", I should start the projector first, then I can view narrow_stereo_textured data on rviz. Is this correct? Thanks in advance.

Failed to connect the usb camera

$
0
0
Hi everyone, I am having problem about using my Andoer USB 2.0 12 Megapixel HD Camera Web Cam. Once I plugged in the webcam, I tried the following: sudo apt-get install ros-indigo-usb-cam rosrun usb_cam usb_cam_node with a terminal running the master. I got the following error: [ INFO] [1456506319.466005157]: using default calibration URL [ INFO] [1456506319.466141992]: camera calibration URL: file:///home/ally/.ros/camera_info/head_camera.yaml [ INFO] [1456506319.466318445]: Unable to open camera calibration file [/home/ally/.ros/camera_info/head_camera.yaml] [ WARN] [1456506319.466353581]: Camera calibration file /home/ally/.ros/camera_info/head_camera.yaml not found. [ INFO] [1456506319.466388999]: Starting 'head_camera' (/dev/video0) at 640x480 via mmap (mjpeg) at 30 FPS [ WARN] [1456506319.517065779]: unknown control 'white_balance_temperature_auto' [ WARN] [1456506319.526420721]: unknown control 'focus_auto' [ERROR] [1456506324.536249248]: select timeout When I type: ls -ltr /dev/video* The output is: crwxrwxrwx+ 1 root video 81, 0 Feb 26 16:39 /dev/video0 Please help. Thanks!

Driver for HERO3 Camera?

$
0
0
I have a Logitech USB camera interfaced to ROS. But after adding a Hero3 camera to my 'bot, I'd like to use it for both ROS blob tracking and video capture. Is there a way to do this? There is a USB connector on the camera for updating from a PC. Alan KM6VV

ethzasl_ptam camera calibration from set of images

$
0
0
Hello, I successfully made my first calibration and rectification using ethzasl_ptam package. Could it be possible to use still images from another camera, lets say iphone6 camera? Here is what I am trying to do: I also run lsd_slam with a logitech c910 camera and got varying results, mostly due to I could not move the camera much. I want to capture a video with an iphone camera outdoors, then run lsd_slam on the video file. and for this I would need to calibrate the iphone6 lens, I could still take pictures of a checker pattern and then feed them to the calibration program, but how? Any ideas/help/recomendations appreciated. Best regards, C.

Intel RealSense depth camera on Ubuntu Arm

$
0
0
Hi all, Im going to develop some package with the [Intel RealSense depth camera](http://click.intel.com/intelrrealsensetm-developer-kit-featuring-sr300.html), but before i get that hardware, i wanna know if its possible to use it on the Ubuntu Arm version of ROS. Thank you so much!!

whycon cqamera localization help !

$
0
0
i am new to ros and am trying to localize swarm robots using normal webcam so i found whycon i am following the tutorial https://github.com/lrse/whycon/wiki/Tutorial but nothing happen and i dont know how to debug that iam sure that 1)my camera publish images on the topic /usb_cam/raw_image 2)the image processing node publishing image on topic /usb_cam/image_mono ....etc the whycon publishing the pose but when i am trying to echo the /robot_pose or the /image_out i have nothing on the screen no error message no output what should i do ?

creating a custom message type including an image

$
0
0
I'm trying to define and use a custom message type. This message will include a few primitive data types as well as a sensor_msgs/Image.msg type for transferring the image. I have not managed to work out how to define this last element yet. Do I need to update the package dependencies somehow? What type name would I use for this. I'm assuming this is even possible, if not would I need to make a copy of what's in the sensor_msgs/Image.msg instead. Thanks

no image received error in rviz

$
0
0
Hi, I am using Ubuntu 12.04 with ROS hydro . Currently ,i am stuck with a error in rviz which states that " no image received" in status field under camera. I am using iball usb_cam. I tried all the camera drivers like gscam,uvc_camera,usb_cam . But , this error still exists . [Note: The image window does not produce any error. but camera window only produce this error .] Please give me any suggestions, Thank you

camera status warn,image status warn

$
0
0
Hello everyone I am using Artags alvar but when I use command "rosrun rviz rviz -d rviz_artag_config.rviz" I get window but with Camera Status warn,Image status Warn no image captured by camera. I use camera usb logitech.Could you inform me what should I do now Thanks in advance Best Taher ---------- EDIT Hello I am using Logitech camera by `using usb_camera usb_camera.launch` file But I get error in Camera Calibration Waiting for service', '//set_camera_info,Service not found. and now in visualizing the Ar_tags_alvar.I sucessfully build alvar and using ROS_indigo packages. Best Taher

will i be able to use axis_camera package for all onvif compliant ptz camera

$
0
0
Currently my team is deciding on a make a model of ptz camera to attach on our robot. My colleague narrowed down to a model which is ONVIF Ver. 2.2 compliant. As Axis is one of the founding member of ONVIF, i was wondering if the AXIS_camera package in ros allows us to communicate and control other ONVIF compliant camera?

camera_aravis unable to read images

$
0
0
Hi, I'm using Basler aceA2040 camera. I'm able to use pylon to view the image, but when I run the node, then I get the following error. [ INFO] [1459415796.505253469]: Attached cameras: [ INFO] [1459415797.507343350]: # Interfaces: 2 [ INFO] [1459415797.507401522]: # Devices: 1 [ INFO] [1459415797.507439594]: Device0: Basler-21717060 [ INFO] [1459415797.508187757]: Opening: (any) [ INFO] [1459415797.649533133]: Opened: Basler-21717060 [ INFO] [1459415797.681021939]: Set gain = 36.000000 [ INFO] [1459415797.683831732]: Camera does not support FocusPos. [ INFO] [1459415799.696795557]: Using Camera Configuration: [ INFO] [1459415799.696895442]: --------------------------- [ INFO] [1459415799.697275209]: Vendor name = Basler [ INFO] [1459415799.697664075]: Model name = acA2040-25gc [ INFO] [1459415799.698057562]: Device id = 21717060 [ INFO] [1459415799.698113163]: Sensor width = 2048 [ INFO] [1459415799.698151466]: Sensor height = 2048 [ INFO] [1459415799.698195014]: ROI x,y,w,h = 0, 0, 2046, 2046 [ INFO] [1459415799.698259450]: Pixel format = yuv422packed [ INFO] [1459415799.698318762]: BytesPerPixel = 2 [ INFO] [1459415799.698932602]: Acquisition Mode = Continuous [ INFO] [1459415799.699541695]: Trigger Mode = Off [ INFO] [1459415799.700273893]: Trigger Source = Line1 [ INFO] [1459415799.700332053]: Can set FrameRate: True [ INFO] [1459415799.701400025]: AcquisitionFrameRate = 100 hz [ INFO] [1459415799.701450797]: Can set Exposure: True [ INFO] [1459415799.701492482]: Can set ExposureAuto: True [ INFO] [1459415799.701533147]: Exposure = 2000 us in range [24,1e+07] [ INFO] [1459415799.701571927]: Can set Gain: True [ INFO] [1459415799.701605875]: Can set GainAuto: True [ INFO] [1459415799.701642581]: Gain = 36.000000 % in range [36.000000,512.000000] [ INFO] [1459415799.701679707]: Can set FocusPos: False [ INFO] [1459415799.702172955]: Network mtu = 1500 [ INFO] [1459415799.702224649]: --------------------------- [ INFO] [1459415800.060123136]: using default calibration URL [ INFO] [1459415800.060201986]: camera calibration URL: file:///home/students/.ros/camera_info/21717060.yaml [ INFO] [1459415800.060287986]: Unable to open camera calibration file [/home/students/.ros/camera_info /21717060.yaml] [ WARN] [1459415800.060328945]: Camera calibration file /home/students/.ros/camera_info/21717060.yaml not found. [ WARN] [1459415800.804103698]: Frame error: ARV_BUFFER_STATUS_TIMEOUT I found a similar issue which suggested to generate a calibration file. I'm not sure how to do that. Could someone please suggest how to generate the correct calibration file needed by this node? ( I'm running the camera_aravis prackage). [Issue #4 Github](https://github.com/strawlab/camera_aravis/issues/4) Thanks

uEye camera in ROS

$
0
0
Hello, I am trying to setup an uEye USB camera in ROS. I already installed the official drivers, and the camera is working correctly. For ROS integration, I found two references: [http://ros.org/wiki/ueye](http://ros.org/wiki/ueye) It only supports IS_CM_BGR8_PACKED format, so no color images available. [http://www.ros.org/wiki/ueye_cam](http://www.ros.org/wiki/ueye_cam) Seems more complete than the previous one, but the pointer to the repository is missing and I am not able to find it. Can someone point me out where to find the package *ueye_cam* ? Thank you

Can't launch libuvc_camera

$
0
0
I'm trying to use [libuvc_camera](http://wiki.ros.org/libuvc_camera) to get my usb camera working on a nodelet to reduce CPU usage, but I'm having a lot of trouble. I've brute-forced my way through most of my problems, but now I'm stonewalled. Every time I try to boot the launch file I just get unsupported descriptor subtype: 3 unsupported descriptor subtype: 3 uvc_get_stream_ctrl_format_size: Invalid mode (-51) Here's my launch file: And here is the error: process[camera/mycam-1]: started with pid [22107] unsupported descriptor subtype: 3 unsupported descriptor subtype: 3 uvc_get_stream_ctrl_format_size: Invalid mode (-51) [ERROR] [1460040590.634224143]: check video_mode/width/height/frame_rate are available [camera/mycam-1] process has died [pid 22107, exit code -11, cmd /home/davelkan/locus/babbage/devel/lib/libuvc_camera/camera_node __name:=mycam __log:=/home/davelkan/.ros/log/48edeb92-fc35-11e5-9a05-0862664a2a25/camera-mycam-1.log]. log file: /home/davelkan/.ros/log/48edeb92-fc35-11e5-9a05-0862664a2a25/camera-mycam-1*.log all processes on machine have died, roslaunch will exit shutting down processing monitor... ... shutting down processing monitor complete done And here's the relevant data from lsusb -v Bus 004 Device 013: ID 2560:c110 Device Descriptor: bLength 18 bDescriptorType 1 bcdUSB 3.00 bDeviceClass 239 Miscellaneous Device bDeviceSubClass 2 ? bDeviceProtocol 1 Interface Association bMaxPacketSize0 9 idVendor 0x2560 idProduct 0xc110 bcdDevice 0.00 iManufacturer 1 e-con Systems iProduct 2 e-con's 1MP Monochrome Camera iSerial 3 2C0D5400 FAQS: - I have confirmed my device is uvc compliant - I have confirmed uvc is correctly installed - I have successfully gotten this camera working with [usb_cam](http://wiki.ros.org/usb_cam)
Viewing all 348 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>