Ali Shafti
Research Associate in Robotics and AI @ Imperial College London
News
02/2021: Talk + Poster accepted at VSS 2021, gaze intention decoding + autonomous driving w/ gaze attention.
02/2021: 2xPapers accepted at the IEEE NER 2021, on gaze interfaces for cognitive human-robot interaction.
01/2021: You can watch my talk below, here.
01/2021: Invited talk at the Imperial College London Dept. of Bioengineering Seminar. Come say Hi online!
01/2021: Happy new year - or at least congratulations on finishing 2020, let's hope for a better one!
12/2020: Happy holidays! Take a break... who'd have thought we'd ever want a break from staying home :-D
11/2020: Poster presentation on our study of Robotic Human Augmentation in Piano Playing at BioRob2020 WS
10/2020: IROS 2020 is free and on demand! Check it out, and our work too (you must sign up first).
09/2020: Back in the lab, socially distanced, but good to be back.
08/2020: Summer break!
07/2020: Our work on the EU Enhance project highlighted by the EU Innovation Radar!
07/2020: Organised seminar with Imperial College AI Network and Robotics forum on AI/Robotics in Healthcare.
07/2020: Paper accepted at IEEE/RSJ IROS'20 on real-world human-robot collaborative reinforcement learning.
06/2020: Invited talk at ICRA2020 WS on Human-Robot Handovers on Explainable Human-Robot Interaction
Latest research demo videos (more here):
Real-World Human-Robot Collaborative RL
A setup for real-world human-robot reinforcement learning of a fully collaborative motor task, in the form of a marble-maze game.
Gaze Prediction for Autonomous Driving
Prediction of human visual attention helps with the training of autonomous driving agents - attention masking helps the agent "see what matters".
Learning Explainable Robotic Manipulations
Hierarchical Reinforcement Learning is used to create more explainable representations of the manipulating agent's understanding of world dynamics.
Gaze-based HRI + Arm Inverse Kinematics
The system is aware of the human user's arm kinematics - this allows for full control of the human user's hand orientation, whilst keeping the interaction comfortable.
About me
I study physical collaboration and interaction between humans and robots. I look into making these intuitive and natural for increased synergy, and augmented capabilities on both sides. I am curious about achieving machine intelligence, while conserving the role of human intelligence as an essential part of the action/perception loop and the overall interaction. To this aim, my research involves human-robot collaboration through machine learning and human behaviour analytics.
My original training is in electronics and electrical engineering. During my BSc and MSc I was focused on microelectronics and analogue/digital circuits design. For my PhD I expanded my research into robotics, focusing on the electronics and computer science aspects, with human-robot collaboration as an area of application. I am now exploring machine intelligence and motor neuroscience and their application within the physical human-robot collaboration realm, as part of my PostDoc research.
For more details, please see my CV.
Collaboration
I am interested in collaborations within the above research topics, as well as other topics that fit in with my expertise. This can be in the form of academic or industrial collaborations, and as joint research projects or in the form of consulting. Do get in touch!