Browsing by Subject "Visual tracking"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
Item Multi-robot formation and cooperation using visual tracking(2013-02) Min, Hyeun JeongCamera-based multi-robot formation is a core and challenging field of research in distributed robotics. The goal is to manage multiple robots maintaining them in a group, capable of extending for various robotic missions. There has been a substantial body of work mostly based on the following assumptions: (i) The availability of inter-communication capability among the robots, (ii) the use of special sensors such as omni-directional cameras, laser scanners, etc., and (iii) the use of special (or fiduciary) markers on some objects of interest (e.g. the leader robot). Those who have pursued successful multi-robot formations have always used at least one of these assumptions. However, these assumptions limit the scope of applicability of the solutions that depend on them. These systems have the following representative limitations: (i) The size of robots should be large enough to be able to handle the large payload of the special sensors, (ii) the use of inter-communication requires that communication systems be either already installed or carried by robots, and (iii) robots are expected to have predictable motions. Many researchers have tried to address the vision-based formation problem using some or all of these assumptions, but none has attempted to go beyond these assumptions yet. In this thesis we approach the problem described above using a monocular camera to address the size limitation. This, in turn, allows us to use much smaller robots than in previous research. We also allow unpredictable motions of the robots, so that our solution can have broader applicability to several missions, such as, search, safety, surveillance, exploration, etc. In addition, because we impose fewer restrictions in the assumptions than in prior work, we are faced with various challenging issues when investigating object tracking and robot control. We present a new robotic formation methodology along with a unique technique for a group of robots based on visual tracking only. Instead of relying on prediction models about a target robot, our robots execute target tracking from consecutive images. For real-time processing, we propose a new object tracking method, limiting the pool of candidates and enhancing the matching performance. Our new multi-robot formation technique has great potential for many robotic applications. Accordingly, this technique motivates us to come up with a new solution to the multi-robot coverage problem. Specifically we deal with the challenging problem of determining the optimum set of robots and their optimized paths, given a set of locations of interest. In contrast, existing approaches to the coverage problem deal with a given set of robots and attempt to find optimized paths. Therefore, in this thesis, we have come up with a new approach, which addresses the coverage problem for a variety of real world applications.Item Pedestrian Control at Intersections (Phase I)(Minnesota Department of Transportation, 1996-10) Papanikolopoulos, Nikolaos P.; Masoud, Osama; Richards, Charles A.This report describes a real-time system for tracking pedestrians in sequences of grayscale images acquired by a stationary camera. The system outputs the spatio-temporal coordinates of each pedestrian during the period when the pedestrian is visible. Implemented on a Datacube MaxVideo 20 equipped with a Datacube Max 860, the system achieved a peak performance of over 30 framers per second. Experimental results based on indoor and outdoor scenes have shown that the system is robust under many difficult traffic situations. The system uses the "figure/ground" framework to accomplish the goal of pedestrian detection. The detection phase outputs the tracked blobs (regions), which in turn pass to the final level, the pedestrian level. The pedestrian level deals with pedestrian models and depends on the tracked blobs as the only source of input. By doing this, researchers avoid trying to infer information about pedestrians directly from raw images, a process that is highly sensitive to noise. The pedestrian level makes use of Kalman filtering to predict and estimate pedestrian attributes. The filtered attributes constitute the output of this level, which is the output of the system. This system was designed to be robust to high levels of noise and particularly to deal with difficult situations, such as partial or full occlusions of pedestrians. The report compares vision sensors with other types of possible sensors for the pedestrian control task and evaluates the use of active deformable models as an effective pedestrian tracking module.