progressing with the real robot

This commit is contained in:
Pablo Iñigo Blasco 2018-09-23 09:22:18 -04:00
Родитель 75b6926f87
Коммит 2641e0e6b6
8 изменённых файлов: 138 добавлений и 64 удалений

Просмотреть файл

@ -12,6 +12,6 @@
<node pkg="sorting_demo" type="gripper_action_server.py" name="gripper_action_server"/>
<include file="$(find sawyer_moveit_config)/launch/sawyer_moveit.launch">
<arg name="electric_gripper" value="true"/>
<arg name="electric_gripper" value="false"/>
</include>
</launch>

Просмотреть файл

@ -1,19 +1,24 @@
import re
from rospy_message_converter import message_converter
import demo_constants
class BlockState:
regex = re.compile(r'block(\d+)\.*')
def __init__(self, gazebo_id, pose):
def __init__(self, id, pose):
self.perception_id = None
self.gazebo_id = gazebo_id
self.gazebo_id = id
self.pose = pose
self.color = None
self.homogeneous_transform = None
search = BlockState.regex.search(gazebo_id)
if not demo_constants.REAL_ROBOT:
search = BlockState.regex.search(id)
self.num = int(search.group(1))
else:
self.num = int(id)
self.gazebo_pose = None
self.grasp_pose = None
self.place_pose = None

Просмотреть файл

@ -1,5 +1,6 @@
#!/usr/bin/python
import copy
import random
import re
import math
@ -13,7 +14,6 @@ from concepts.block import BlockState
from concepts.tray import TrayState
from concepts.table import Table
from cv_detection_head import CameraHelper, get_blobs_info
from cv_detection_right_hand import get_cubes_z_rotation
@ -51,6 +51,22 @@ class EnvironmentEstimation:
self.hand_camera_helper = CameraHelper("right_hand_camera", "base", TABLE_HEIGHT)
if demo_constants.REAL_ROBOT:
k = 3
for i in xrange(k):
for j in xrange(k):
q = tf.transformations.quaternion_from_euler(random.uniform(0, 2 * math.pi),
random.uniform(0, 2 * math.pi),
random.uniform(0, 2 * math.pi))
block = BlockState(id=str(len(self.gazebo_blocks)),
pose=Pose(position=Point(x=0.45 + j * 0.15 + random.uniform(-1, 1) * 0.03,
y=-0.15 + i * 0.15 + random.uniform(-1, 1) * 0.03,
z=0.7725),
orientation=Quaternion(x=q[0], y=q[1], z=q[2], w=q[3])))
self.gazebo_blocks.append(block)
def identify_block_from_aproximated_point(self, projected):
"""
:param projected:
@ -77,7 +93,7 @@ class EnvironmentEstimation:
return None
def compute_block_pose_estimation_from_arm_camera(self, CUBE_SIZE=150):
#get latest image from topic
# get latest image from topic
rospy.sleep(0.3)
# Take picture
img_data = self.hand_camera_helper.take_single_picture()
@ -143,7 +159,7 @@ class EnvironmentEstimation:
# select the other grasping
if not graspA and graspB:
rospy.logwarn("Swiching grasping orientation for current block (grasping clearance)")
if final_cube_yaw_angle >0:
if final_cube_yaw_angle > 0:
final_cube_yaw_angle -= math.pi / 2
else:
final_cube_yaw_angle += math.pi / 2
@ -155,16 +171,16 @@ class EnvironmentEstimation:
self.tf_broacaster.sendTransform(projected, poseq, rospy.Time(0), "estimated_cube_1", "base")
rospy.logwarn(projected)
#cv2.imshow("cube detection", cv_image)
#cv2.waitKey(0)
# cv2.imshow("cube detection", cv_image)
# cv2.waitKey(0)
return Pose(position=Point(x=projected[0], y=projected[1], z=projected[1]),
orientation=Quaternion(x=poseq[0], y=poseq[1], z=poseq[2], w=poseq[3])),graspA or graspB
orientation=Quaternion(x=poseq[0], y=poseq[1], z=poseq[2], w=poseq[3])), graspA or graspB
except Exception as ex:
rospy.logwarn("erroneous cube detection")
#cv2.imshow("erroneus cube detection", cv_image)
#cv2.waitKey(0)
# cv2.imshow("erroneus cube detection", cv_image)
# cv2.waitKey(0)
return None, False
def compute_block_pose_estimations_from_head_camera(self):
@ -180,7 +196,7 @@ class EnvironmentEstimation:
# Convert to OpenCV format
cv_image = self.bridge.imgmsg_to_cv2(img_data, "bgr8")
cv2.imwrite("/tmp/last_head_picture.jpg",cv_image)
cv2.imwrite("/tmp/last_head_picture.jpg", cv_image)
rospy.logwarn("processing head camera image to find blocks")
blobs_info = get_blobs_info(cv_image)
@ -278,7 +294,7 @@ class EnvironmentEstimation:
item = self.get_block_by_gazebo_id(name)
if item is None:
# rospy.logwarn("block create name: "+ name)
item = BlockState(gazebo_id=name, pose=pose)
item = BlockState(id=name, pose=pose)
item.color = demo_constants.BLOCK_COLOR_MAPPINGS[item.num]["material"].replace("Gazebo/", "")
else:
item.pose = pose
@ -288,7 +304,8 @@ class EnvironmentEstimation:
item = self.get_tray_by_gazebo_id(name)
# item = None
if item is None:
item = TrayState(gazebo_id=name, pose=pose, TRAY_SURFACE_THICKNESS= demo_constants.TRAY_SURFACE_THICKNESS)
item = TrayState(gazebo_id=name, pose=pose,
TRAY_SURFACE_THICKNESS=demo_constants.TRAY_SURFACE_THICKNESS)
item.color = demo_constants.TRAY_COLORS[item.num].replace("Gazebo/", "")
else:
item.pose = pose
@ -321,6 +338,7 @@ class EnvironmentEstimation:
blocks = []
trays = []
if not demo_constants.REAL_ROBOT:
# publish tfs
basehomopose = get_homo_matrix_from_pose_msg(self.gazebo_world_to_ros_transform, tag="base")
@ -420,7 +438,6 @@ class EnvironmentEstimation:
:param id:
:return:
"""
color = color.replace("Gazebo/", "")
rospy.logwarn("by color: " + str(color))
rospy.logwarn("by color: " + str(self.trays))

Просмотреть файл

@ -11,14 +11,15 @@ from control_msgs.msg import (
GripperCommandFeedback,
GripperCommandResult,
)
import real_gripper
from real_gripper import PSGGripper
import intera_interface
class GripperActionServer(object):
def __init__(self):
self._ns = "/robot/gripper_action_server"
self._gripper = intera_interface.Gripper()
self._gripper = PSGGripper() # intera_interface.Gripper()
# Action Server
self._server = actionlib.SimpleActionServer(

Просмотреть файл

@ -5,20 +5,24 @@ import sys
import rospy
import gazebo_models
from task_planner import TaskPlanner
import demo_constants
def main():
rospy.init_node("sorting_demo")
rospy.logwarn("Hello world")
if not demo_constants.REAL_ROBOT:
# Load Gazebo Models via Spawning Services
# Note that the models reference is the /world frame
# and the IK operates with respect to the /base frame
model_list, original_block_poses = gazebo_models.load_gazebo_models()
# Remove models from the scene on shutdown
rospy.on_shutdown(functools.partial(gazebo_models.delete_gazebo_models, model_list))
rospy.logwarn("Creating task planner")
task_planner = TaskPlanner()
task_facade = task_planner.get_task_facade()

Просмотреть файл

@ -0,0 +1,32 @@
class PSGGripper():
def __init__(self):
pass
def get_position(self):
return 0
def get_force(self):
return 0
def set_dead_zone(self, v):
return 0
def get_dead_zone(self):
return 0
def is_moving(self):
return 0
def has_error(self):
return False
def stop(self):
pass
def close(self):
pass
def open(self):
pass

Просмотреть файл

@ -8,6 +8,8 @@ from geometry_msgs.msg import (
Pose,
)
from sorting_demo import demo_constants
from real_gripper import PSGGripper
class SawyerRobotControl(object):
def __init__(self, trajectory_planner, limb="right", hover_distance=0.08, tip_name="right_gripper_tip"):
@ -21,7 +23,12 @@ class SawyerRobotControl(object):
self._tip_name = tip_name # string
self._hover_distance = hover_distance # in meters
self._limb = intera_interface.Limb(limb)
if demo_constants.REAL_ROBOT:
self._gripper =PSGGripper()
else:
self._gripper = intera_interface.Gripper()
self._robot_enable = intera_interface.RobotEnable()
# verify robot is enabled

Просмотреть файл

@ -29,10 +29,12 @@ class TaskPlanner:
def __init__(self):
"""
"""
rospy.logwarn("creating task planner")
limbname = 'right'
self._hover_distance = 0.1 # meters
self.place_hover_distance = 0.15
rospy.logwarn("creating environment estimation")
# subcomponents
self.environment_estimation = EnvironmentEstimation()
@ -44,6 +46,8 @@ class TaskPlanner:
self.ikservice = rospy.ServiceProxy("/sawyer_ik_5d_node/ik", moveit_msgs.srv.GetPositionIK)
rospy.logwarn("creating task planner")
self.cancel_signal = False
self.pause_flag = False
@ -64,6 +68,16 @@ class TaskPlanner:
self.joint_names = ["right_j0", "right_j1", "right_j2", "right_j3", "right_j4", "right_j5", "right_j6"]
self.gripper_state = GripperState()
self.head_pose_joints = {'right_j0': 0.0,
'right_j1': -numpy.pi / 2.0,
'right_j2': 0.0,
'right_j3': numpy.pi / 2.0,
'right_j4': 0.0,
'right_j5': 0.0,
'right_j6': 0.0}
self.starting_joint_angles = self.head_pose_joints
def scheduler_yield(self):
rospy.logwarn("scheduler yield")
if self.cancel_signal:
@ -314,13 +328,7 @@ class TaskPlanner:
oldceil = self.trajectory_planner.ceilheight
joint_angles = {'right_j0': 0.0,
'right_j1': -numpy.pi / 2.0,
'right_j2': 0.0,
'right_j3': numpy.pi / 2.0,
'right_j4': 0.0,
'right_j5': 0.0,
'right_j6': 0.0}
joint_angles = self.head_pose_joints
self.trajectory_planner.ceilheight = 2.0