{$ msg.text $}

Job Description

We are looking for master thesis students who are enthusiastic in applying deep learning techniques to address one of the most challenging questions in the virtual reality industry: how to realize natural interactions between humans and computers in virtual environments.

Natural interaction requires human-like perception and action, which poses complex yet exciting machine learning challenges that range from thorough understanding of 3D environments and objects, to understanding of avatar's kinematic structures from the 3D mesh, to solving inverse kinematics problems for arms and fingers, to generating human-like motion and animation, to analyzing human interactive behaviors in VR in the next-generation VR-powered training games.

Recently machine learning techniques that exploit the deep structure of neural networks have achieved significant progress towards many practical industrial problems. Deep neural networks (DNN) for 3D geometric data is also a highly emerging and challenging research area [1] with a wide range of potential applications for 3D shapes, including recognition and segmentation [2, 3] as well as encoding and synthesis [4-6]. The goal of this thesis is to exploit DNN to analyze the shape of 3D objects and facilitate the grasp animation of human hand. Specifically, the thesis project will include application of DNN for part-based object representation and will also involve a succeeding segmentation of the estimated parts.

Tasks:

  • Summarize state-of-the-art of deep learning study aimed at modeling and representing 3D object shape and segmentation of the shape.
  • Collect training database for avatar and hand representations.
  • Implement modeling and training of DNNs, preferrably using Caffe2 deep learning framework, in C++.
  • Test, optimize and evaluate the implemented process using the database.
  • Summarize and discuss the findings in a report / thesis.

References:

[1] Introduction slides on 3D deep learning: http://ai.stanford.edu/~haosu/slides/IntroTo3DDL.pdf
[2] "O-CNN: Octree-based Convolutional Neural Networks for 3D Shape Analysis", Wang, P. S. et al. TOG, 2017
[3] “PointGrid: A Deep Network for 3D Shape Understanding”, Truc Le et al. CVPR, 2018
[4] “Learning Shape Abstractions by Assembling Volumetric Primitives”, Shubham Tulsiani et al. CVPR, 2017. Project page
[5] "CSGNet: Neural Shape Parser for Constructive Solid Geometry", Gopal Sharma et al. CVPR, 2018
[6] "GRASS: generative recursive autoencoders for shape structures", Li Jun et al. TOG, 2017

Hiring Manager
Dan Song

About the company

At Gleechi, we want to enable humans to interact naturally with the digital world, and robots to interact like humans. We are a Stockholm-based startup coming from robotics research and the first in the world to enable artificial hand movement and interaction in real-time. We collaborate with world-leading game and VR developers, we make it possible for stroke patients to interact with virtual worlds, and we enable robots to collaborate with people in hospitals. We're a small team based in central Stockholm that combines awarded entrepreneurs, top-ranked robotics researchers and experienced developers. The company was founded in the end of 2014 and since then we have been awarded the Super startup of the year by Veckans Affärer, won the European startup competition EIT Digital Idea Challenge and much else. We got a ridiculously exciting time ahead and we'd love to get more awesome people onboard!

Visit website

Our Location

Stockholm

Follow Gleechi