We are looking for master thesis students who are enthusiastic in applying deep learning techniques to address one of the most challenging questions in the virtual reality industry: how to realize natural interactions between humans and computers in virtual environments.
Natural interaction requires human-like perception and action, which poses complex yet exciting machine learning challenges that range from thorough understanding of 3D environments and objects, to understanding of avatar's kinematic structures from the 3D mesh, to solving inverse kinematics problems for arms and fingers, to generating human-like motion and animation, to analyzing human interactive behaviors in VR in the next-generation VR-powered training games.
Recently machine learning techniques that exploit the deep structure of neural networks have achieved significant progress towards many practical industrial problems. Deep neural networks (DNN) for 3D geometric data is also a highly emerging and challenging research area  with a wide range of potential applications for 3D shapes, including recognition and segmentation [2, 3] as well as encoding and synthesis [4-6]. The goal of this thesis is to exploit DNN to analyze the shape of 3D objects and facilitate the grasp animation of human hand. Specifically, the thesis project will include application of DNN for part-based object representation and will also involve a succeeding segmentation of the estimated parts.
 Introduction slides on 3D deep learning: http://ai.stanford.edu/~haosu/slides/IntroTo3DDL.pdf
 "O-CNN: Octree-based Convolutional Neural Networks for 3D Shape Analysis", Wang, P. S. et al. TOG, 2017
 “PointGrid: A Deep Network for 3D Shape Understanding”, Truc Le et al. CVPR, 2018
 “Learning Shape Abstractions by Assembling Volumetric Primitives”, Shubham Tulsiani et al. CVPR, 2017. Project page
 "CSGNet: Neural Shape Parser for Constructive Solid Geometry", Gopal Sharma et al. CVPR, 2018
 "GRASS: generative recursive autoencoders for shape structures", Li Jun et al. TOG, 2017
At Gleechi, we want to enable humans to interact naturally with the digital world, and robots to interact like humans. We are a Stockholm-based startup coming from robotics research and the first in the world to enable artificial hand movement and interaction in real-time. We collaborate with world-leading game and VR developers, we make it possible for stroke patients to interact with virtual worlds, and we enable robots to collaborate with people in hospitals. We're a small team based in central Stockholm that combines awarded entrepreneurs, top-ranked robotics researchers and experienced developers. The company was founded in the end of 2014 and since then we have been awarded the Super startup of the year by Veckans Affärer, won the European startup competition EIT Digital Idea Challenge and much else. We got a ridiculously exciting time ahead and we'd love to get more awesome people onboard!