This commit is contained in:
David Baumert 2021-01-27 16:42:30 -08:00 коммит произвёл GitHub
Родитель e44599ad73
Коммит 8f4cccf01b
Не найден ключ, соответствующий данной подписи
Идентификатор ключа GPG: 4AEE18F83AFDEB23
1 изменённых файлов: 1 добавлений и 34 удалений

Просмотреть файл

@ -18,44 +18,11 @@ https://github.com/microsoft/LabanotationSuite
The Gesture Authoring Tools repository provides an upper-torso human-gesture capture and editing solution. The system consists of compiled gesture capture applications using the Microsoft Kinect sensor device and a Windows 10 PC. Editing tools constructed in Python provide gesture trimming and movement analysis options with output in graphical Labanotation scores as well as movement data expressed in Labanotation and implemented in a JSON data format.
#### gestureBot Design Kit
https://github.com/davidbaumert/gestureBotDesignKit
https://github.com/microsoft/gestureBotDesignKit
With a Windows 10 PC, and optionally a 3D-printer and about $350(USD) in electronic servos and parts, the gestureBot Design Kit repository contains all the information needed to build both a virtual and physical desktop companion robot. It includes browser-based simulation and control software based on the Robotis XL series of servo motors. To construct a physical robot, it provides models for 3D-printable body-parts, a parts-list for electronic components, and step-by-step assembly instructions. No soldering is required, but some manual skill is needed to mate small electronic connectors, as well as manipulate small plastic rivets and miniature metal screws to assemble the body components.
#### Gesture Library
https://github.com/davidbaumert/GestureLibrary
The Gesture Library is a data-set comprised of upper-torso gestures expressed in Labanotation and implemented in a JSON format.
#### Gesture Service
https://github.com/davidbaumert/GestureService
The Gesture Service project is a software module constructed with Python and Google's neural network [word2vec](https://code.google.com/archive/p/word2vec/#!) that takes a text phrase as input and returns a corresponding gesture.
#### MSRAbot AirSim Design Kit
https://github.com/davidbaumert/MSRAbot_DesignKit
With a gamer-class Windows 10 PC, and optionally a 3D-printer and $800(USD) in electronic servos and parts, the MSRAbot AirSim Design Kit repository contains all the information needed to build both a virtual and physical mobile companion robot. It includes browser-based simulation and control software based on the [Futaba RS](https://futabausa.com/product/rs303mr/) series of servo motors. To construct a physical robot, it provides models for 3D-printable body-parts, a parts-list for electronic components, and step-by-step assembly instructions. Intermediate electronic assembly and soldering skills are required to build the physical robot. To place a virtual model of the robot in a 3D simulation environment, an implementation hosted in [**Microsoft AirSim**](https://github.com/microsoft/AirSim) is provided to support AI model training and robotics research projects.
***(the following projects and descriptions comprising our project roadmap ARE NOT included in public repositories)***
#### Full-Humanoid Gesture and Dance Capture
#### Full-Humanoid Gesture and Dance Performance
## Navigation
The field of robot navigation concerns solutions such as simultaneous-location-and-mapping (SLAM), path planning, and map management.
### Hololens Indoor Navigation
This project shows how a Hololens device can be placed on the head of [Softbank Robotics' Pepper robot](https://us.softbankrobotics.com/pepper) and provide a self-calibrating indoor navigation solution within a single room.
### Azure Kinect Indoor Navigation
This project shows how an Azure Kinect device can be placed on a mobile delivery robot and provide a self-calibrating indoor navigation solution within a large building.
## Manipulation
Common in industrial-robotics where actions are manually programmed and repeated, the field of robotic object manipulation is much more challenging in service-robotics scenarios. Our projects explore solutions where HRI and Navigation technologies can be leveraged to allow robots to learn from humans to perform manipulation tasks safely and effectively in dynamic residential, workplace, and public environments.
### Human-trained Object Manipulation
This project shows how a human who is expert in a given task can use HRI solutions and AI-based task-planning to teach a service robot to reliably perform the same task within a dynamic environment.