Till KTH:s startsida Till KTH:s startsida

Visa version

Version skapad av Patric Jensfelt 2020-09-03 18:16

Visa < föregående | nästa >
Jämför < föregående | nästa >

Master thesis proposals - external

Volumental

Evaluate and improve camera pose estimation

Accurate camera pose estimation is a necessity in many practical applications such as VR, autonomous cars, 3D Reconstruction and so on. For some appli- cations accuracy is more important than real-time camera tracking. ARKit is Apple’s tool that can estimate among other things the camera pose using the camera and the IMU. The question is how good does this work? At Volumental we are currently developing an app for scanning feet and we rely on the camera pose that is given by ARKit. In the Master Thesis project we aim to evaluate ARKit and improve it by creating a novel benchmark and derive a new algorithm.

Volumental is a computer vision company from RPL, KTH active in 3D body scanning and product recommendation based on 3D measurements in footwear. Today we have our computer vision systems are deployed across 32 countries and scanning hundreds of thousands of people regularly, working with some of the world’s biggest brands. We are a relatively small but growing team of PhDs in computer vision and machine learning and are product RPL-alumni. We are almost half women and hail from 9 different countries, we are located centrally at T-Centralen.

PDF-version of this announcement

Contact: Erik Bylow, erik.bylow@volumental.com

Ericsson

In an automated warehouse where autonomous robots load trucks with products while sharing the same environment with humans, a proper safety analysis is performed to avoid the hazardous situations without compromising the productivity. We have a basic safety analysis mechanism using image processing for object identification and risk assessment and mitigation (using fuzzy logic, RL approaches) to provide safety. The tasks of this thesis will be to extend the safety model: (1) adding safety from communication perspective i.e. among robots and between the robot and the edge / cloud; (2) performing highly-computational processing in edge / cloud; (3) looking into the Explainability of the used AI (XAI) techniques for Trustworthiness perspective; (4) model compression to run complex deep learning architectures in devices with limited hardware.
The high-level task is to perform safety analysis and image processing in the cloud and locally at the attached hardware and comparing both with respect to safety, performance, energy usage in human-robot collaborative use case.

Required qualifications:  MSc studies in Computer Science, Electrical and Computer Engineering or similar area; Excellent programming skills in C/C++, or Python or Java or Matlab; Good knowledge of concepts in machine learning (e.g. deep learning), robotics, ROS, etc.; Experiences with machine learning libraries Tensor flow, Keras, PyTorch, sci-kit learn etc.; Knowledge of safety analysis techniques, XAI, communication protocols and technologies (edge/cloud, D2D) is a bonus; Like to build end to end prototypes and concepts.

Contact: Rafia Inam (rafia.inam@ericsson.com), Alberto Hata (alberto.hata@ericsson.com )

Full project description here 

H&M

These projects will be performed in collaboration with H&M and aim to provide efficient solutions for some of the problems that the company is facing. The successful applicant will have an opportunity to closely interact with experts from H&M, have access to the real-world data owned by the company, and receive guidance from KTH researchers.

Project 1: Quality prediction

Every day, H&M orders, transports, and sells thousands of items. Quality control is an important part of this process. When boxes with goods are delivered to one of the warehouses, random quality checks are performed. There are multiple factors that can potentially affect the quality of the products in a specific box: they come different suppliers overseas, are transported in different ways (by ships, trains, or planes), and indirect factors such as weather also may play a certain role.

In this project, we aim to develop a probabilistic model that will predict the quality of the products based on the information about the supplier, the means of transportation, etc. To train the model, H&M will provide the relevant data. If implemented successfully, the system could later be used to identify the boxes that are more likely to require quality control.

Required qualifications: proficiency in Machine Leaning and Data Science. Applicants are expected to have passed KTH courses such as Advanced Machine Learning, Project Course in Data Science, Probabilistic Graphical Models, or equivalent.

Contact person: Anastasiia Varava (KTH) varava@kth.se

Project 2: Carton fillrate optimization

To transport the products, H&M needs to make decisions on how to package them. Depending on the characteristics of the products being packed (type of product, size, material, fragility, etc), different types of cartons can be used, and different number of items are packed into each of them.

Previously, this decisions were made by humans on a case-to-case basis.

In this project, we aim to automate this process. Based on the data collected earlier at H&M (item descriptions and packaging specifications), we aim to develop a model that will find the best way to package novel products. The model will need to capture specific features of the products that can potentially affect these decisions, and thus can be trained on existing data that will be provided by the company.

Required qualifications: proficiency in Machine Leaning and Data Science. Applicants are expected to have passed KTH courses such as Advanced Machine Learning, Project Course in Data Science, Probabilistic Graphical Models, or equivalent.

Contact person: Anastasiia Varava (KTH) varava@kth.se

Project 3: Inbound leadtime prediction

H&M orders large amounts of goods from multiple different suppliers in various countries to markets all over the world. Each product is expected to be delivered by a specific date. However, delays in production and transportation are inevitable, especially when the products are transported by ship due to multiple factors such as weather conditions, scheduling imprecision, multiple connections between different routes, etc. The goal of this project is to develop a probabilistic model that will be trained on the historical data provided by the company to predict the expected time needed for the goods to be manufactured and delivered to markets.

Required qualifications: proficiency in Machine Leaning and Data Science. Applicants are expected to have passed KTH courses such as Advanced Machine Learning, Project Course in Data Science, Probabilistic Graphical Models, or equivalent.

Contact person: Anastasiia Varava (KTH) varava@kth.se

Scania

Reinforcement Learning at Flexible Alternating Current Transmission Systems

You will be part of the Business Unit Grid Integration, located in Västerås. FACTS (Flexible Alternating Current Transmission Systems) technologies provide more power and control in existing AC as well as green-field networks and have minimal environmental impact. With a complete portfolio and in-house manufacturing of key components, ABB is a reliable partner in shaping the grid of the future. Please find out more about our world leading technology at www.abb.com/facts.

Problem description

At FACTS you will aid in creating a more sustainable future. The increase of available measurements, growth of real time processing capacity and communication abilities is changing the opportunities in the power system landscape. There is a potential to utilize more measurement data, Reinforcement Learning and a FACTS device for optimization purposes.

The task would include: 
• Create suitable test networks
• Investigate and develop prototype algorithm based on latest advancements in deep Reinforcement Learning
• Train, tune and test the algorithms on the networks and demonstrate policy optimality

Requirements

We are looking for you who are studying a university master program within a relevant technical area along with an interest within artificial intelligence or Reinforcement Learning. Experience in Reinforcement Learning or programming is positive. We aim to start the thesis in early September.

Model-based Model-free Reinforcement Learning for Robotics Manipulation

Background
Recent advances in artificial intelligence has enabled machines to compete with humans even in the most difficult of domains. Google Deepmind's AlphaGo is a case in point. Similar approaches of reinforcement learning (RL) have been tried in the robotics community on problems of skill learning. By skill we mean a sensorimotor policy (control policy) that can perform a single continuous-time task. Numerous successes in skill learning have been reported for a variety of manipulation tasks that are otherwise difficult to program. Examples include, batting, pancake flipping, pouring, pole balancing etc. One of the most challenging class of manipulation tasks is assembly of mating parts. Not surprisingly, the capability to learn assembly skills is highly sought after. 

Problem description
RL can be divided into model-based and model-free methods. In model-based methods, the algorithm learns a dynamics model of the manipulation task and utilizes it to optimize the policy. Contrary to this, in model-free RL (policy search), the policy is often directly optimized without the intermediate step of model learning. The trade off here is between number of trials (sample efficiency) and model bias. While mode-based methods are sample efficient, model-free methods do not suffer from model bias. We propose a hybrid approach that has benefits of both methods in it. It employs a global black box optimization method called Bayesian optimization (BO) to learn the policy in a fundamentally model-free way, but at the same time uses a learned model to guide the process. We will exploit the fact that BO does not require a cost function for the learning process. Our application will be an assembly task in which an ABB YuMi robot will insert one part into another part.

Purpose and aims
The objective of this thesis is to develop a skill learning method under the framework of RL. The robot should be able to demonstrate the learning process by continuously trying to do the insertion while making incremental progress and finally achieve convergence by being able to complete the task successfully in a few consecutive trials.

The work will include the following tasks:

  1. Conduct literature review on RL based skill learning and BO.
  2. Formulate a strategy for utilizing a learned dynamics model for guiding the BO. Model learning algorithm can be assumed to be given.
  3. Set up either MuJoCo or Bullet simulation environment. Implement a simpler task of inverted pendulum and then the main insertion task.
  4. Develop a parameterized policy (not necessarily deep network) and implement the BO based RL algorithm including the results of Step 2.
  5. Evaluate the method on a real robot and draw conclusions about the hybrid method. 

We are searching for a highly motivated student from master programs such as Systems, Control and Robotics, or Machine learning, or a student with a similar background. Knowledge in modeling and control of robotics manipulator is highly advantageous. Any prior exposure to Gaussian process regression, RL or BO will be valued. A medium to high level of competency in either Python or Matlab is necessary. Masters level knowledge of linear algebra and probability theory is expected and general competence in machine learning will be highly appreciated.

The master student will gain competences within Robotics, Robot Control, Reinforcement learning, Bayesian optimization, Gaussian process, etc. Note that the student will work in ABB Corporate Research in Västerås and compensation plus accommodation will be provided by the company. This project is defined within the context of an ongoing PhD project and therefore, the student can expect a high level of research environment and support, including software and systems. Prospective PhD student will be given preference. It may also be possible to do this project at RPL but the decision will be taken on a case by case basis.

Contact:  Shahbaz Khader, +46725305968, shahbaz.khader@se.abb.com, ABB 

Online Planning Based Reinforcement Learning for Robotics Manipulation 

Background
Recent advances in artificial intelligence has enabled machines to compete with humans even in the most difficult of domains. Google Deepmind's AlphaGo is a case in point. Similar approaches of reinforcement learning (RL) have been tried in the robotics community on problems of skill learning. By skill we mean a sensorimotor policy (control policy) that can perform a single continuous-time task. Numerous successes in skill learning have been reported for a variety of manipulation tasks that are otherwise difficult to program. Examples include, batting, pancake flipping, pouring, pole balancing etc. One of the most challenging class of manipulation tasks is assembly of mating parts. Not surprisingly, the capability to learn assembly skills is highly sought after. 

Problem description
Most skill learning RL methods are of policy search type. In policy search methods, the optimal parameters of a parameterized policy is obtained from an optimization process. Computing a general policy that takes the best action in any possible state is a much harder problem than planning a sequence of actions from a single state. On the other hand, while the policy provides robustness to uncertainties, planning cannot cope with any deviations from the plan. Online planning or model predictive control (MPC) is a method in which the best of both worlds come together. Instead of computing a policy offline, a plan is computed in an online manner at every execution step. Only the first action is applied and the rest is discarded. The process is repeated at every time step. The drawback with the online planning method is the high computational cost of planning at every time step. When combined with dynamics model learning, the overall method becomes a reinforcement learning approach. Some of the challenges that we aim to tackle in this thesis are: trading off planning horizon versus computational cost, planning under uncertain dynamics model, and incorporating prior information of the task instead of completely relying on learning the dynamics. Our application will be an assembly task in which an ABB YuMi robot will insert one part into another part. 

Purpose and aims
The objective of this thesis is to develop a skill learning method under the framework of RL. The robot should be able to demonstrate the learning process by continuously trying to do the insertion while making incremental progress and finally achieve convergence by being able to complete the task successfully in a few consecutive trials.

The work will include the following tasks:

  1. Conduct literature review on RL based skill learning and MPC.
  2. Formulate a method for online planning that utilizes the uncertainties of the learned dynamics model. Model learning algorithm can be assumed to be given.
  3. Develop a strategy for combining offline learning from simulation and online planning.
  4. Evaluate the method on simulated tasks and also a real robot.

 

We are searching for a highly motivated student from master programs such as Systems, Control and Robotics, or Machine learning, or a student with a similar background. Knowledge in modeling and control of robotics manipulator is highly advantageous. Any prior exposure to optimal control, MPC, or RL will be valued. A medium to high level of competency in either Python or Matlab is necessary. Masters level knowledge of linear algebra and probability theory is expected and general competence in machine learning will be highly appreciated.

The master student will gain competences within Robotics, Robot Control, Reinforcement learning, Optimization, Optimal Control, etc. Note that the student will work in ABB Corporate Research in Västerås and compensation plus accommodation will be provided by the company. This project is defined within the context of an ongoing PhD project and therefore, the student can expect a high level of research environment and support, including software and systems. Prospective PhD student will be given preference. It may also be possible to do this project at RPL but the decision will be taken on a case by case basis.

Contact:  Shahbaz Khader, +46725305968, shahbaz.khader@se.abb.com, ABB Corporate Research