Topics for Theses

Available Topics

Image Captioning in computer vision. Image taken from [1].


Generating captions that describe the content of an image is a task emerging in computer vision. Lastly, Recurrent Neural Networks (RNN) in the form of Long Short-Term Memory (LSTM) networks have shown great success in generating captions matching an image's content. In contrast to traditional tasks like image classification or object detection, this task is more challenging. A model not only needs to identify a main class but also needs to recognize relationships between objects and describe them in a natural language like English. Recently, an encoder/decoder network presented by Vinyals et al. [1] won the Microsoft COCO 2015 captioning challenge.


Available Topics

Generating Image Captions for Unkown Objects

Examples of generated captions (LRCN) and improved sentences (LSTM-C) by using a copying mechanism that replaces unknown words by detected objects within the caption. Figure taken from [2].


As seen in [1], deep neural networks a capable of generating simple captions for given images when given a large dataset in the order of a million labeled image-caption pairs. However, as every data driven machine learning approach these models can’t create captions for objects never seen before, i.e. the dataset doesn’t contain images or captions of a certain type. Training such a model takes days to weeks on modern GPUs, but a requirement may be that this model can describe new object categories afterwards with little to no extra effort.


Ting et al. use a hybrid model that combines an object detection model with a captioning model. In figure 1 you can see a picture and its ground-truth caption. A standard captioning model (LRCN) may produce a caption that is not accurate (e.g. “a red fire hydrant is parked in the middele of a city”). Combined with an object detection pipeline, which detects a bus in this case this hybrid model can generate a much more accurate caption “a bus driving down a street next to a building”.


Your tasks will be:

  • Familiarize yourself with the  Tensorflow Framework and the  Show and Tell model
  • Integrate the LSTM copying mechanism introduced in [2] into a given Tensorflow Model (Show and Tell)
  • Train and evaluate your implementation on the  MSCOCO dataset by holding out 8 objects out of the dataset
  • Compare and analyze the performance of your implementation against the paper [2].

Python and Numpy knowledge is advantageous as Tensorflow models are implemented in Python.

If you are interested and want more information, please contact  Philipp Harzig.

Visual Question Answering

Examples of question/image input pairs and generated answers. Picture taken from [3].


Building on top of general image captioning another more challenging task in computer vision has come up recently. It is called visual question answering a question referencing some of the input image’s contents is part of the input. The model then tries to answer the question as accurate as possible. See figure 2 for image-question-answer examples. Most publications have agreed on one approach to tackle this problem. The question and image are both embedded in a vector representation, then combined in some way and the answer is generated as the most likely answer out of 3000 to 5000 possible answers. Therefore, the problem is modeled as a classification problem, i.e. all possible answers are assigned a probability and the most likely answer is selected out of all answers.


Your tasks will be:

  • Familiarize yourself with the  Tensorflow Framework and a classification pipeline already implemented in Tensorflow (e.g. the  Slim Framework)
  • Implement a model that embeds an image with a deep convolutional neural network (DCNN) like the ResNet or Inception network and a question with an LSTM network. This model should combine both representations and use a classification model that selects the most probable answer for this combination.
  • Train and evaluate your implementation on the  VQA-v2 dataset.
  • Compare and analyze the performance vs. state-of-the-art approaches.

Python and Numpy knowledge is advantageous as Tensorflow models are implemented in Python. If you are interested and want more information, please contact  Philipp Harzig.


Literature for Image Captioning

[1] Vinyals, Oriol, et al. "Show and tell: A neural image caption generator." Proceedings of the IEEE conference on computer vision and pattern recognition. 2015.

[2] Yao, Ting, et al. "Incorporating copying mechanism in image captioning for learning novel objects." 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017.

[3] Teney, Damien, et al. "Tips and Tricks for Visual Question Answering: Learnings from the 2017 Challenge." arXiv preprint arXiv:1708.02711 (2017).

Generating ground truth data can be very costly and time consuming. For instance, generating labels for a single image in semantic segmentation can take up to 1 hour. Alternatively, synthetic data can be generated automatically much easier and faster. However, neural networks trained on synthetic data show a poor generalization to real data. In this thesis we assume that we have a synthetic data set with labels and an unlabeled real dataset consisting only of images. The task is to set up methods that allow the model trained on synthetic data to generalize on real data as well. Various approaches can be used to solve this task using Generative Adversarial Networks (GANs) or a self-supervised training pipeline.

If you are interested and want more information, please contact Sebastian Scherer

In this work we assume the following szenario. We have a large amount of images, but only a small subset has annotated ground truth labels. Supervised approaches only allow the usage of the small subset of data. The question we ask in this work is, can we make use of all images? We will pre-train our models in a self-supervised task on a large amount of unlabeled data before adapting the model to the target task. Possible target taks may be semantic segmentation, human pose estimation or 3D object detection.

If you are interested and want more information, please contact Sebastian Scherer

Supervised training of deep neural networks require large labeled datasets. However, the label generation process can be very noisy/error-prone in the sense that some labels are labeled incorrectly. Additionally, there are self-supervised methods that generate pseudo labels on non-annotated data that are used for training afterwards. Training on noisy labels can yield to a poor performance. In this work, we will investigate the effect of wrong annotations in the training and design approaches that overcome this issue. Possible target taks may be image classification, semantic segmentation, human pose estimation or 3D object detection.

If you are interested and want more information, please contact Sebastian Scherer

Bottom-up Human Pose Estimation pipelines have the advantage that regardless of the number of people in the image, the runtime is nealy equal. On the contrary, it is complicated to group all the detected keypoints to person instances. The model HigherHRNet achieves great accuracy in keypoint detection, but needs a lot of time in this grouping postprocessing step compared to the model runtime. Furthermore, its grouping mechanism called Associative Embedding has some difficulties with occluded persons. Another promising possibility to perform the grouping task is based on Part Affinity Fields (PAF). In this thesis, the PAF grouping technique should be incorporated in the HigherHRNet model and compared to the current Associative Embedding technique regarding its runtime and grouping performance.

If you are interested and want more information, please contact Katja Ludwig

Semi-Supervised Learning is an active research field in computer vision with the goal to train neural networks with only a small labeled dataset and a lot of unlabeled data. For human pose estimation, this means that a large dataset with images from people is available, but only a small subset has annotated keypoints. Semi-supervised human pose estimation uses different techniques to train jointly on labeled and unlabeled images in order to improve the detection performance of the network. Popular methods are pseudo labels - the usage of network predictions as annotations - and teacher-student-approaches, where one network is enhanced by being trained by a second network.  


If you are interested and want more information, please contact  Katja Ludwig


Driven by the massive progress in 2D Human Pose Estimation and related detection-based tasks over the last years, active research is steadily advancing to the next logical step: the reconstruction of the human pose in 3D space. And while existing multi-view or RGB-D motion capture systems are perfectly capable of this task, current research focuses on the difficult and highly under-constrained case of single-view RGB images and videos. Reliably estimating the 3D pose of a human from a single consumer-grade camera opens up a vast area of practical applications.


Like in many computer vision topics, all current state-of-the-art methods in 3D human pose reconstruction (HPR) evolve around some form of convolutional neural network (CNN). The main differences come from

  • the specific task definition
  • the pose representation, especially within the CNN
  • the type of supervision (configuration and quantity of data and labels)
  • the runtime vs. fidelity trade-off

There is onging research in all these different topics, with year-to-year gains in precision, reliability and efficiency.


Thesis Topics

The research at our chair covers all the aforementioned concepts. We always have specific research questions that are suitabel for a Bachelor or Master thesis as well as practical courses and internships. With the speed of new developments and advancements in this field, the detailed topic for a thesis will be defined on-demand based on the current research at our chair and the prerequsites and interests of the student. Below are some topics that are suitabel for a potential thesis. If you are interested in the overall reserach field or one of the following topics, please contact Moritz Einfalt.


CNN Pose Representations for Multi-Person 3D HPR

Coming from the current state in 2D human pose estimation, the quasi-standard method to represent the detection targets for pose-defining human keypoints in CNN models are spatial 2D heatmaps. Retaining the spatial dimensions from input (image) to output (heatmaps, one per keypoint) is the currently best performing mode. The naive transfer of this concept to the 3D task (i.e. the detection of keypoints in 3D space) are volumentric 3D heatmaps. However, the additional dimension in the CNN output makes this representation very costly, especially when a high spatial resolution in the predicted heatmaps is required. And while the approach can be feasible for single-person 3D HPR on tight image crops, it completely breaks in the multi-person case on large images.


Current solutions try to factorize the 3D volume into smaller, more efficent 1D and 2D components [1]. This divides the learning task of the CNN into a detection part (2D heatmaps) and a regression part (e.g. numerical regression of the depth component), see figure 1. Other approaches use a learned compact representation of 3D heatmaps from an integrated autoencoder [2], see figure 2. Both variants have disadvantages and can lead to ambiguites in the encoding of 3D keypoints of tightly grouped people in the image. Potential topics for theses under this research question include the comparison of different existing representation methods and the development of new representations under the contraints of efficiency or precision.



Figure 1: Mixed pose representation: Spatial detection task with 2D heatmaps + sparse regression of 3D keypoint locations. Image taken from [1].


Figure 2: Reconstructed volumentic heatmap (summed over the z-axis for visualization) with the autoencoder from [2]. Image taken from the JTA dataset [3].


Real-Time 3D HPR on Edge Devices

The current state-of-the-art in monocular 3D HPR  is already at a level of precision and reliability where it can be intergrated into actual applications. This can range from analytical applications, where the motion of humans in 3D space is infered and evaluated, to interactive scenarios, where the human body is used as an input mode to control other agents (robots, virtual characters, ...). However, most of the current best-perfoming monocular 3D HPR methods rely on very deep CNNs, large spatial input and ouput sizes and sometimes even the combination of multiple CNN models for two-step person detection and pose estimation. Aside from the need of entire GPU servers for training, these models still require a dedicated high-end consumer or even professional GPU during inference (i.e. application) to reach real-time capabilities. This contraint massively hinders the development of new applications: It restricts the usage to stationary scenarios, where the recording device is connected to a powerful GPU machine. The true application potential lies in mobile applications, where the 3D HPR is performed direclty on the recording device (read: smartphone).


One highly relevant research questions is therefore the transfer of the current state-of-the-art in 3D HPR to less powerful edge devices like smartphones. Existing approaches focus on single-shot architectures [4] (  see figure 3),    low-resolution image crops or CNN model compression [5] (see figure 4). Potential topics for theses under this research question include benchmarking and adapting existing methods and developing new strategies in teacher-student model compression.

Figure 3: Single-shot multi-person 3D HPR with Pandanet. Image taken from [4].


Figure 4: Real-time 3D HPR directly on a smartphone via CNN model compression. Image taken from [5].



[1] Mehta, Dushyant, et al. "XNect: Real-time multi-person 3D motion capture with a single RGB camera." ACM Transactions on Graphics (TOG) 39.4 (2020): 82-1.


[2] Fabbri, Matteo, et al. "Compressed volumetric heatmaps for multi-person 3d pose estimation." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.


[3] Fabbri, Matteo, et al. "Learning to detect and track visible and occluded body joints in a virtual world." Proceedings of the European Conference on Computer Vision (ECCV). 2018.


[4] Benzine, Abdallah, et al. "Pandanet: Anchor-based single-shot multi-person 3d pose estimation." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.


[5] Hwang, Dong-Hyun, et al. "Lightweight 3D human pose estimation network training using teacher-student learning." Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision. 2020.