Embedded Motion Control 2018 Group 2: Difference between revisions

From Control Systems Technology Group
Jump to navigation Jump to search
Line 153: Line 153:
The idea for this method of driving around was obtained from the Wiki page of Group 2 of Embedded Motion Control 2017.
The idea for this method of driving around was obtained from the Wiki page of Group 2 of Embedded Motion Control 2017.


Whenever the beam in front does not see any obstacles, PICO will drive forward. If an obstacle is seen in front, but the space to the right is free, then PICO will turn right. When the front and right are blocked, but left is free, then PICO will turn left. If the front, left and right are all blocked, PICO will turn around and drive back.
Whenever the beam in front does not see any obstacles, PICO will drive forward. If an obstacle is seen in front, but the space to the right is free, then PICO will turn right. After making its right turn, PICO will start driving forward again. When the front and right are blocked, but the left is free, then PICO will turn left. If the front, left and right are all blocked, PICO will turn around and drive back the way it came, searching for an exit somewhere else. A video of PICO driving around in a square room is shown in the video below.
 
[[File:Wall_follower_group2.gif|thumb|center|800px|PICO driving around a square room with a simple wall follower]]
 
After the shooting of this video, a simple repulsion algorithm was implemented alongside the wall follower, so PICO keeps a bigger distance from the walls.

Revision as of 17:37, 29 May 2018

Group Members

TU/e Number Name E-mail
0843128 Robbert (R.) Louwers r.louwers@student.tue.nl
1037038 Daniël (D.J.M.) Bos D.J.M.Bos@student.tue.nl
0895324 Lars (L.G.L.) Janssen l.g.l.janssen@student.tue.nl
1275801 Clara () Butt clara.butt@aiesec.net
0848638 Dorus (D.) van Dinther d.v.dinther@student.tue.nl










Relevant PDF Files

[|Initial Design Group 2]

Initial Design

Introduction:

The main goal of this course is to program the PICO robot so that it can map a `hospital' floor and find an object within it. The programming language used is C++ and the software is ran in Ubuntu 16.04, which is the same version of Ubuntu that is running on the robot. For the initial design, the requirements will be discussed as well as the functions and their specifications. Moreover, a list of possible components and a suiting interface is designed.

Note that this is an initial design towards the hospital competition which has overlap with the escape room competition. Some distinction will be indicated.

Requirements:

  • Execute all tasks autonomously.
  • Perform all tasks without bumping into a wall.
  • Perform all tasks without getting stuck in a loop.
  • Finish both challenges as fast as possible, but at most within 5 minutes.
  • Escape room competition:
    • Locate the exit and leave the room through this exit.
    • Keep a minimum distance of 20 cm to the walls or any other objects.
  • Final competition:
    • Explore and map the hospital rooms.
    • Use the map to get back to the starting position and park backwards into the wall behind it using both the map and the control effort.
    • Use a map to find and stop next to the object that is placed in one of the rooms after the first
  • Nice to have: Use the provided hint to be able to find the placed object directly without searching.

Functions:

he functions that are going to be implemented are subdivided in three different types of functionality. First of all there is the lowest level functionality, which will be referred to as the motion functions. These functions will be used to interact directly with the robot. This interaction will be specified more precisely in the function overview that can be found below. The mid level functionality, which will be referred to as the skill functions. As the name mid level functionality already suggests these functions act on a more abstract level and are built to be able to perform a specific action. All the actions that will be implemented in the skill functions can also be seen in the overview below. Finally there is the high level functionality, which will be referred to as task functions. These functions are the most abstract and are structured to carry out a specific task, which consists of multiple actions.

To give a full overview: in a task function a specific order of actions will be specified in order to be able to fulfill the task. The actions that the robot can perform, and are therefore available to be called by the task functions, are specified in the skill functions. The motion functions are used in the skill functions to be able to invoke physical behavior by the robot.

  • Motion functions
    • Actuation: Using the actuation function PICO is given the command to drive in a specific direction or to turn in a specific direction. The input to this function is a movement vector which specifies in which direction and how far the robot should move. The output is actual robot movement according to this vector and the control effort needed for this movement.
    • Obtain information: Using the obtain information function sensor readings can be requested. The input for this function is the request of a (sub)set of sensor data from a specific sensor(set), either the Laser Range Finder (LRF) or the odometry encoders.
  • Skill functions
    • Path planning: Using the path planning function a (set of) movement vectors is defined on the basis of the map to reach the defined goal. The input for this function is the current map and the goal that should be reached. The output of this function is a (set of) movement vectors.
    • Object avoidance: Using the object avoidance function it is ensured that PICO does not hit any objects. The input for this function is the LRF sensor data and the movement vector. If the movement vector prescribes a movement that collides with an object that is detected by the LRF the movement vector is adjusted such that PICO keeps a safe distance from the object. The output of this function is therefore an (adjusted) movement vector.
    • Mapping and localization: Using the mapping and localization skill a map of the environment is built up and PICO's location in this environment is determined. The input for this function is sensor information and actuation information that can be obtained from the motion functions. Using the LRF data PICO's environment can be defined in "blocked" and "free" area's, places in which obstacles are present and absent respectively. Using the odometry encoders and the control effort the location of PICO can be tracked in the environment. The output of this function is a visual map, that is expanded and updated every time this function is called.
    • Semantic mapping: Using the semantic mapping function extra information is added to the basic environment map. This function has as input the basic environment map. The surface of this map will be split into different sections which can have one of two different types of identifiers assigned to them: room or hallway. Both the splitting of the sections and the identifications of the sections will be based on shape fitting. If a rectangular shape can be fitted over (a piece of) the basic map, based on the LRF, this (piece of the) map will be defined a (separate) section with the identifier room. Note that each section with the identifier room will be given a separate number to be able to distinguish the different rooms from each other. This is necessary in order to be able to use the hint that will be given for the final challenge. The output of this function is a visual semantic map that is updated every time this function is called.
  • Task functions
    • Exit a mapped room: This is the escape room challenge. In this task all motion skills and all skill functions, except the semantic mapping skill, will be used.
    • Semantically map an entire floor and return to the beginning position: This is the first part of the final challenge. In this task all motion and skills functions will be called.
    • Find an item in the semantically mapped floor, preferably using the provided hint on the position of the item: This is the second part of the final challenge. In this task again all motion skills will be used and all the skill functions except the semantic mapping skill.

Components:

To be able to execute the challenges the PICO robot, already mentioned above, will be used. The following hardware components will be utilized:

  • Actuators:
    • Holonomic base with three omni-wheels.
  • Sensors:
    • Laser range finder (LRF): To detect the distances to objects in the environment.
      • Range: To be determined
      • Field of view: 270 degrees
      • Accuracy: To be determined
    • Wheel encoders: Combined with the control effort these encoders keep track of the robots position relative to its starting point. (This has to be double checked with the position in the map, since the encoders easily accumulate errors.)
      • Accuracy: To be determined
  • Computer
    • Running Ubuntu 16.04.

Specifications:

The specifications are based on the requirements.

  • The maximal transitional velocity of PICO in any direction is 0.5 [m/s].
  • The maximal rotational velocity of PICO is limited to 1.2 [rad/s]
  • PICO should not stand still or make no sensible movements for periods over 30 [seconds].
  • Escape room competition:
    • PICO has to finish the Escape Room within 5 [min].
    • The width of the corridor and the openings in the walls are between 0.5 [m] and 1.5 [m].
    • The shape of the room is rectangular.
    • The exit is located more than 3 [m] into the corridor.
  • Hospital competition:
    • PICO has to finish the Hospital within 5 [min].
    • The width of the corridor and the openings in the walls are between 0.5 [m] and 1.5 [m].

Interfaces:

The interfaces describe the data that is communicated between the several parts of code that will be developed. The functions that are discussed in the Chapter Functions are all part of the interface, however there is one other important piece in this interface: the world model. The world model will contain the most recent (semantic) environment map with the location of PICO included in it. The motion, skill and task functions can monitor and request the information that is stored in this world model and use it to properly execute their function. A graphic showing the interface between the world model and the tasks, skills and motions is shown below.

Figure 1: Software interfaces

Mapping

The necessity of mapping in the hospital challenge is of three fold. First of all a map is necessary to be able to detect the new object that will be placed in the environment after the initial mapping round. The new object can be found by comparing the map with the current environment. The second reason a map is needed is the navigation that needs to be performed. In the final challenge navigation through rooms and hallways is needed to be able to get to the newly placed object. The final reason for the need of mapping is the inaccuracy of the odometry of PICO. Due to this inaccuracy the odometry of PICO is useless for determining the position of PICO in the long term. By extending the determination of the position of the robot with the much more reliable laser data the position could be determined much more accurate. However in order to be able to do this a map of the environment is needed.

There are a lot of different ways in which it is possible to make an environment map. A well known method in the world of robotics to tackle this problem is called Simultaneous Localization And Mapping (SLAM). This method is used to construct and update a map of an unknown environment while simultaneously monitoring the location of the mapping entity, which is in this case a robot. Given the available resources there are still quite a few different possible SLAM algorithms that can be implemented. In this case specific research was done into Extended Kalman Filter SLAM (EKF SLAM) and the Gmapping functionality provided in ROS, which sits under the hood of the given software environment. Given the lack of knowledge on how to implement our own SLAM solution and the strict time limits it was chosen to pursue the SLAM solution that was already implemented in ROS.

Using this Gmapping functionality a 2-D occupancy grid map can be made using the laser data and the pose data (read position data) that are collected by the mapping entity. The input for this node, in this specific case, therefore exists of the laser data and the odometry data that are collected by PICO. However this is enough to be able to create the desired floor plan with occupancy information. What more is needed is the real-time transformation between different coordinate frames. This is necessary since the obtained data, the laser data and the odometry data, are collected in different frames. So in order for the Gmapping algorithm to be able to function properly two of these transforms, and therefore three coordinate frames, are needed: the transform between the initial position and the current position of the robot base and the transform between the position of the robot base and the position of the laser. Note that this last transform is static since the position of the laser will not change in relation to the position of the baselink. A graphical representation of this transform tree can be seen below.

Figure 2: Transform setup for Gmapping

In the given environment none of the needed transforms is performed, so these transforms have to be implemented. This will be done by subscribing to certain topics that are published, altering them if necessary, and publishing these transforms again over rostopics. This might sound simple, but is were the progress is currently stopping.

Space Recognition

A manner for PICO to move around without bumping into obstacles is designed by splitting all available data from the LRF into three beams. One to the left of PICO, one to the right, and one pointing to the front. A graphical representation of the beams is shown below.

The idea for this method of driving around was obtained from the Wiki page of Group 2 of Embedded Motion Control 2017.

Whenever the beam in front does not see any obstacles, PICO will drive forward. If an obstacle is seen in front, but the space to the right is free, then PICO will turn right. After making its right turn, PICO will start driving forward again. When the front and right are blocked, but the left is free, then PICO will turn left. If the front, left and right are all blocked, PICO will turn around and drive back the way it came, searching for an exit somewhere else. A video of PICO driving around in a square room is shown in the video below.

File:Wall follower group2.gif
PICO driving around a square room with a simple wall follower

After the shooting of this video, a simple repulsion algorithm was implemented alongside the wall follower, so PICO keeps a bigger distance from the walls.