Celebrate the innovations in the fields of AI and robotics as we release the 2016-2017 Halcyon Dialogue series report, Shaping Robotics Policy for the 21st Century. Tune into the Facebook livestream on September 28 from 1-4pm as we roll out the report with exciting robot demos and captivating conversations.
Listen to leading experts as they discuss our future living with robots. Top industry experts, congressional leaders and staff, leading technology companies, and influential policymakers and media figures will take part in the event. Don’t miss out on the chance to hear what they have to say.
Join live on Facebook @HalcyonInspires.
Eager to learn more about the Robots visiting Halcyon? Check out our robotic guests:
Amazon Web Services will demonstrate how to use a fully managed IoT service in the cloud (AWS IoT) as well as an Image Recognition service (AWS Rekognition) to assist in emergency response scenarios. The demonstration will include an IoT connected search device which will be fitted with a camera and LED indicators. Conceptually these devices could be mounted on an unmanned vehicle (drone or land based) or an emergency response vehicle. AWS will have a web interface that allows operators to enter search terms (e.g. fire, person etc.) for the search device to detect. When the search device is activated it will transmit images and location to an image recognition system (running on AWS cloud); service will analyze the incoming images, extract data from it and compare against the list of search terms that were entered into the system. When the system finds a match, operator is notified via the web console to review the image. If the image is a positive match the operator can put the search device into a beacon mode where it will flash its LED’s to notify nearby emergency responders to investigate the area.
Calypso for Cozmo (“Calypso” for short) is a new robot intelligence framwork for the revolutionary Cozmo robot by Anki. Calypso allows anyone age 8-80 to program Cozmo using computer vision, speech recognition, and artificial intelligence algorithms. Calypso was developed by Professor David Touretzky of Carnegie Mellon University. During this live showcase, Dr. Touretzky will demonstrate how Calypso’s innovative user interface supports “transparent robot intelligence”. #Cozmo #CozmoMoments
EMIEW3 is a humanoid robot with an enhanced degree of autonomy based on EMIEW and EMIEW2. A “remote brain” consisting of control functions deployed on the cloud and a robot monitoring system, forms the robotics IT platform, which enables EMIEW3 to support customer and guidance services. #EMIEW3 #HitachiRobot #Hitachi
This exhibit will highlight robotic systems developed in the Laboratory for Computational Sensing and Robotics at Johns Hopkins University, including new microsurgical robot developed to assist surgeons in minimally invasive applications of otolaryngology, neurosurgery and similar critical fields. The system can eliminate hand tremor and enforce virtual safety barriers to allow surgeons to perform high-stress cases with improved confidence. We will demonstrate a research version of this robot. A clinical version of this robot is being developed by Galen Robotics, Inc.
The Perceptive Pixel (PPI) by Microsoft 55″ Touch Device is a touch-sensitive computer monitor capable of detecting and processing a virtually unlimited number of simultaneous on-screen touches. It has 1920 x 1080 resolution, adjustable brightness of up to 400 nits, a contrast ratio of up to 1000:1, and a display area of 47.6 x 26.8 inches. An advanced sensor distinguishes true touch from proximal motions of palms and arms, eliminating mistriggering and false starts. With optical bonding, the PPI by Microsoft 55″ Touch Device virtually eliminates parallax issues and exhibits superior brightness and contrast. And it has built-in color temperature settings to accomodate various environments and user preferences. #Surface #SurfaceFam #MagicWall
John Hopkins University Applied Physics Lab (#JHU or #JHUAPL) will present Intelligent Systems that can make decisions under uncertainty and take action when authorized. A legion of trusted, intelligent systems able to autonomously sense, think, decide and act, while interfacing with human teammates on critical applications. JHU/APL has developed Think/Decide algorithms that allow robots to operate “in the wild”. Specifically, their team has focused on enabling a heterogeneous team of robots to:
- perceive and describe novel objects they encounter
- collaborate autonomously on mapping, navigating, and manipulating a challenging environment
- interface with human teammates using natural language
Daniel Turner and Stephen Carter, the co-Founders of TRAXyL and two of Halcyon Incubator’s own Cohort 7 fellows, install optical fiber communications on the fly. Using their latest prototype, the FiberTRAXtor, to test their patented installation method, the final product blends in directly with the road surface and allows for digital communications without compromising road surface integrity.