Hi everyone
I am new to ROS Indigo and am trying to work with the libviso2/ viso2_ros package for visual odometery. I have a fair enough idea as to how to use the package/node with a camera, but I have little idea as to how to make it work with pre-collected dataset of images. The datasets I am trying to experiment with are any of [these](http://www.cvlibs.net/datasets/karlsruhe_sequences/). Libviso2 looks for a rectified camera image.
Please let me know if any other information is needed. Any help would be greatly appreciated.
↧
How to use image data set to simulate a camera input for libviso2/ viso2_ros ?
↧
USB Camera is not running - Book Learning ROS for Robotics Programming
I am using Ubuntu 15.04 and jade version of ROS. I am trying to use the usb camera by the method given in the book on page 187. My camera is already good and I checked it by the "cheese" method explained in there. Here is what I am doing
$ sudo apt-get install ros-jade-usb-cam
$roslaunch chapter5_tutorials usb_cam.launch view:=true
here is what I am getting after doing this
[ERROR] [1453172400.684518103]: Webcam: expected picture but didn't get it...
[ERROR] [1453172400.684518103]: Webcam: expected picture but didn't get it...
↧
↧
do rviz need urdf file for viewing camera feed?
I am trying to create a stereo camera, I know that the feed can be viewed in `rviz` using `image` display type.But other than that i want to know how can we succesfully view camera feed using `camera` display type
- i was successful in calibrating and publishing the `image_info` topic
- but when try to view the camera using `camera` type i get warning status saying no image was received but i successfully gets the image via the `image` display type 
- so my suspicion is if I am missing something on the Fixed_frame or target_frame or something like that
↧
depth_image_proc register issue
I'd like to use the depth_image_proc pipeline to create point cloud xyzrgb using a RGBD camera. I calibrated rgb and ir camera and used image_proc to rectify them. Now I would like to register depth image to the frame of the rgb camera. All input topics are up - depth image, depth camera info, rgb camera info and the required tf is also there. The topics depth_registered/camera_info and depth_registered/image_rect are generated, but they are empty.
The question is why?


↧
ar_sys camera receive problem
hi. i am beginner. i wanna run ar_sys package but i have a few problems.
i have usb cam or can i do that with my laptop's camera ?
**Firstly :**
"Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist"
probably my camera is not define to ROS. How can i do that ?
**Secondly :**
i try to run this line :
rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100
But i have this error:
Fixed Frame [world] does not exist
**Thirdly :**
Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)
↧
↧
ar_sys camera problem
hi. i am beginner. i wanna run ar_sys package but i have a few problems.
i have usb cam or can i do that with my laptop's camera ?
**Firstly :**
"Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist"
probably my camera is not define to ROS. How can i do that ?
**Secondly :**
i try to run this line :
rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100
But i have this error:
Fixed Frame [world] does not exist
**Thirdly :**
Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)
↧
ar_sys camera problem
hi. i am beginner. i wanna run ar_sys package but i have a few problems.
i have usb cam or can i do that with my laptop's camera ?
**Firstly :**
"Fixed Frame No tf data. Actual error: Fixed Frame [world] does not exist"
probably my camera is not define to ROS. How can i do that ?
**Secondly :**
i try to run this line :
rosrun tf static_transform_publisher 0.0 0.0 0.0 0.0 0.0 0.0 map my_frame 100
But i have this error:
Fixed Frame [world] does not exist
**Thirdly :**
Have i to use special camera for ar_sys package ? (like we have to use lidar scanner in hector slam.)
↧
How to visualize the data of "5-Megapixel camera" and "LED Texture Projector Triggered with Narrow-Angle Stereo Camera" in rviz?
I want to use rviz to confirm the functions of "5-Megapixel camera" and "LED Texture Projector Triggered with Narrow-Angle Stereo Camera" works well,
but I cannot find the correct topic to visualize the camera data, does anyone know how to view in rviz?
EDIT:
I found [a page](http://pr2support.willowgarage.com/wiki/PR2%20Manual/Chapter9#Head_Cameras) that may be useful for this question.
Now my understanding is, projector illuminates the objects. To visualize the data of "LED Texture Projector Triggered with Narrow-Angle Stereo Camera", I should start the projector first, then I can view narrow_stereo_textured data on rviz. Is this correct?
Thanks in advance.
↧
Failed to connect the usb camera
Hi everyone, I am having problem about using my Andoer USB 2.0 12 Megapixel HD Camera Web Cam. Once I plugged in the webcam, I tried the following:
sudo apt-get install ros-indigo-usb-cam
rosrun usb_cam usb_cam_node
with a terminal running the master. I got the following error:
[ INFO] [1456506319.466005157]: using default calibration URL
[ INFO] [1456506319.466141992]: camera calibration URL: file:///home/ally/.ros/camera_info/head_camera.yaml
[ INFO] [1456506319.466318445]: Unable to open camera calibration file [/home/ally/.ros/camera_info/head_camera.yaml]
[ WARN] [1456506319.466353581]: Camera calibration file /home/ally/.ros/camera_info/head_camera.yaml not found.
[ INFO] [1456506319.466388999]: Starting 'head_camera' (/dev/video0) at 640x480 via mmap (mjpeg) at 30 FPS
[ WARN] [1456506319.517065779]: unknown control 'white_balance_temperature_auto'
[ WARN] [1456506319.526420721]: unknown control 'focus_auto'
[ERROR] [1456506324.536249248]: select timeout
When I type:
ls -ltr /dev/video*
The output is:
crwxrwxrwx+ 1 root video 81, 0 Feb 26 16:39 /dev/video0
Please help. Thanks!
↧
↧
Driver for HERO3 Camera?
I have a Logitech USB camera interfaced to ROS. But after adding a Hero3 camera to my 'bot, I'd like to use it for both ROS blob tracking and video capture.
Is there a way to do this? There is a USB connector on the camera for updating from a PC.
Alan KM6VV
↧
ethzasl_ptam camera calibration from set of images
Hello,
I successfully made my first calibration and rectification using ethzasl_ptam package.
Could it be possible to use still images from another camera, lets say iphone6 camera?
Here is what I am trying to do: I also run lsd_slam with a logitech c910 camera and got varying results, mostly due to I could not move the camera much. I want to capture a video with an iphone camera outdoors, then run lsd_slam on the video file. and for this I would need to calibrate the iphone6 lens, I could still take pictures of a checker pattern and then feed them to the calibration program, but how?
Any ideas/help/recomendations appreciated.
Best regards,
C.
↧
Intel RealSense depth camera on Ubuntu Arm
Hi all,
Im going to develop some package with the [Intel RealSense depth camera](http://click.intel.com/intelrrealsensetm-developer-kit-featuring-sr300.html), but before i get that hardware, i wanna know if its possible to use it on the Ubuntu Arm version of ROS.
Thank you so much!!
↧
whycon cqamera localization help !
i am new to ros and am trying to localize swarm robots using normal webcam so i found whycon
i am following the tutorial
https://github.com/lrse/whycon/wiki/Tutorial
but nothing happen and i dont know how to debug that
iam sure that
1)my camera publish images on the topic /usb_cam/raw_image
2)the image processing node publishing image on topic /usb_cam/image_mono ....etc
the whycon publishing the pose
but when i am trying to echo the /robot_pose or the /image_out i have nothing on the screen no error message no output
what should i do ?
↧
↧
creating a custom message type including an image
I'm trying to define and use a custom message type. This message will include a few primitive data types as well as a sensor_msgs/Image.msg type for transferring the image. I have not managed to work out how to define this last element yet.
Do I need to update the package dependencies somehow? What type name would I use for this.
I'm assuming this is even possible, if not would I need to make a copy of what's in the sensor_msgs/Image.msg instead.
Thanks
↧
no image received error in rviz
Hi, I am using Ubuntu 12.04 with ROS hydro . Currently ,i am stuck with a error in rviz which states that " no image received" in status field under camera. I am using iball usb_cam.
I tried all the camera drivers like gscam,uvc_camera,usb_cam . But , this error still exists .
[Note: The image window does not produce any error. but camera window only produce this error .]
Please give me any suggestions,
Thank you
↧
camera status warn,image status warn
Hello everyone
I am using Artags alvar but when I use command
"rosrun rviz rviz -d rviz_artag_config.rviz"
I get window but with
Camera Status warn,Image status Warn no image captured by camera.
I use camera usb logitech.Could you inform me what should I do now
Thanks in advance
Best
Taher
----------
EDIT
Hello I am using Logitech camera by `using usb_camera usb_camera.launch` file
But I get error in Camera Calibration
Waiting for service', '//set_camera_info,Service not found.
and now in visualizing the Ar_tags_alvar.I sucessfully build alvar and using ROS_indigo packages.
Best Taher
↧
will i be able to use axis_camera package for all onvif compliant ptz camera
Currently my team is deciding on a make a model of ptz camera to attach on our robot. My colleague narrowed down to a model which is ONVIF Ver. 2.2 compliant. As Axis is one of the founding member of ONVIF, i was wondering if the AXIS_camera package in ros allows us to communicate and control other ONVIF compliant camera?
↧
↧
camera_aravis unable to read images
Hi,
I'm using Basler aceA2040 camera. I'm able to use pylon to view the image, but when I run the node, then I get the following error.
[ INFO] [1459415796.505253469]: Attached cameras:
[ INFO] [1459415797.507343350]: # Interfaces: 2
[ INFO] [1459415797.507401522]: # Devices: 1
[ INFO] [1459415797.507439594]: Device0: Basler-21717060
[ INFO] [1459415797.508187757]: Opening: (any)
[ INFO] [1459415797.649533133]: Opened: Basler-21717060
[ INFO] [1459415797.681021939]: Set gain = 36.000000
[ INFO] [1459415797.683831732]: Camera does not support FocusPos.
[ INFO] [1459415799.696795557]: Using Camera Configuration:
[ INFO] [1459415799.696895442]: ---------------------------
[ INFO] [1459415799.697275209]: Vendor name = Basler
[ INFO] [1459415799.697664075]: Model name = acA2040-25gc
[ INFO] [1459415799.698057562]: Device id = 21717060
[ INFO] [1459415799.698113163]: Sensor width = 2048
[ INFO] [1459415799.698151466]: Sensor height = 2048
[ INFO] [1459415799.698195014]: ROI x,y,w,h = 0, 0, 2046, 2046
[ INFO] [1459415799.698259450]: Pixel format = yuv422packed
[ INFO] [1459415799.698318762]: BytesPerPixel = 2
[ INFO] [1459415799.698932602]: Acquisition Mode = Continuous
[ INFO] [1459415799.699541695]: Trigger Mode = Off
[ INFO] [1459415799.700273893]: Trigger Source = Line1
[ INFO] [1459415799.700332053]: Can set FrameRate: True
[ INFO] [1459415799.701400025]: AcquisitionFrameRate = 100 hz
[ INFO] [1459415799.701450797]: Can set Exposure: True
[ INFO] [1459415799.701492482]: Can set ExposureAuto: True
[ INFO] [1459415799.701533147]: Exposure = 2000 us in range [24,1e+07]
[ INFO] [1459415799.701571927]: Can set Gain: True
[ INFO] [1459415799.701605875]: Can set GainAuto: True
[ INFO] [1459415799.701642581]: Gain = 36.000000 % in range [36.000000,512.000000]
[ INFO] [1459415799.701679707]: Can set FocusPos: False
[ INFO] [1459415799.702172955]: Network mtu = 1500
[ INFO] [1459415799.702224649]: ---------------------------
[ INFO] [1459415800.060123136]: using default calibration URL
[ INFO] [1459415800.060201986]: camera calibration URL: file:///home/students/.ros/camera_info/21717060.yaml
[ INFO] [1459415800.060287986]: Unable to open camera calibration file [/home/students/.ros/camera_info
/21717060.yaml]
[ WARN] [1459415800.060328945]: Camera calibration file /home/students/.ros/camera_info/21717060.yaml not found.
[ WARN] [1459415800.804103698]: Frame error: ARV_BUFFER_STATUS_TIMEOUT
I found a similar issue which suggested to generate a calibration file. I'm not sure how to do that. Could someone please suggest how to generate the correct calibration file needed by this node? ( I'm running the camera_aravis prackage).
[Issue #4 Github](https://github.com/strawlab/camera_aravis/issues/4)
Thanks
↧
uEye camera in ROS
Hello,
I am trying to setup an uEye USB camera in ROS. I already installed the official drivers, and the camera is working correctly.
For ROS integration, I found two references:
[http://ros.org/wiki/ueye](http://ros.org/wiki/ueye)
It only supports IS_CM_BGR8_PACKED format, so no color images available.
[http://www.ros.org/wiki/ueye_cam](http://www.ros.org/wiki/ueye_cam)
Seems more complete than the previous one, but the pointer to the repository is missing and I am not able to find it.
Can someone point me out where to find the package *ueye_cam* ?
Thank you
↧
Can't launch libuvc_camera
I'm trying to use [libuvc_camera](http://wiki.ros.org/libuvc_camera) to get my usb camera working on a nodelet to reduce CPU usage, but I'm having a lot of trouble. I've brute-forced my way through most of my problems, but now I'm stonewalled. Every time I try to boot the launch file I just get
unsupported descriptor subtype: 3
unsupported descriptor subtype: 3
uvc_get_stream_ctrl_format_size: Invalid mode (-51)
Here's my launch file:
And here is the error:
process[camera/mycam-1]: started with pid [22107]
unsupported descriptor subtype: 3
unsupported descriptor subtype: 3
uvc_get_stream_ctrl_format_size: Invalid mode (-51)
[ERROR] [1460040590.634224143]: check video_mode/width/height/frame_rate are available
[camera/mycam-1] process has died [pid 22107, exit code -11, cmd /home/davelkan/locus/babbage/devel/lib/libuvc_camera/camera_node __name:=mycam __log:=/home/davelkan/.ros/log/48edeb92-fc35-11e5-9a05-0862664a2a25/camera-mycam-1.log].
log file: /home/davelkan/.ros/log/48edeb92-fc35-11e5-9a05-0862664a2a25/camera-mycam-1*.log
all processes on machine have died, roslaunch will exit
shutting down processing monitor...
... shutting down processing monitor complete
done
And here's the relevant data from lsusb -v
Bus 004 Device 013: ID 2560:c110
Device Descriptor:
bLength 18
bDescriptorType 1
bcdUSB 3.00
bDeviceClass 239 Miscellaneous Device
bDeviceSubClass 2 ?
bDeviceProtocol 1 Interface Association
bMaxPacketSize0 9
idVendor 0x2560
idProduct 0xc110
bcdDevice 0.00
iManufacturer 1 e-con Systems
iProduct 2 e-con's 1MP Monochrome Camera
iSerial 3 2C0D5400
FAQS:
- I have confirmed my device is uvc compliant
- I have confirmed uvc is correctly installed
- I have successfully gotten this camera working with [usb_cam](http://wiki.ros.org/usb_cam)
↧