πŸš€ My Journey

Welcome to my journey. This page offers insights into my story and the adventures that have led me to where I am today.

Feel free to navigate through the timeline to discover more about my educational pursuits, career endeavors, and the valuable lessons I’ve gained along the way.

Let’s embark on this journey together! 🌟

πŸ’Ό Experience :

Master Thesis - Contact-Aware Visuomotor Policies for Robust Dexterous Object Manipulation

Company: Prehensio GmbH

Time: Feb. 2026 – Present

Description:
Currently assigned to the spin-out Prehensio GmbH as part of my ongoing research position at Fraunhofer IPA.

Skills:
Imitation Learning, Tactility, Flow Matching Models, PyTorch, Teleoperation, ROS2, Python, Computer Vision

Research Project - Dexterous Imitation Learning for Vision-Action Models

Company: Fraunhofer Institute IPA

Time: July. 2025 – Feb. 2026

Description:
- Developed a teleoperation and imitation learning pipeline using 3D camera and data glove in ROS2. - Trained Diffusion Policies to enable robust grasping capabilities in multi-object scenes. - Created a full MoveIt model for safe and reliable motion execution planning. - Applied the resulting policies to industrial automation scenarios.

Skills:
Imitation Learning, Diffusion Models, PyTorch, Teleoperation, ROS2, Python, Computer Vision

Research Assistant

Company: Fraunhofer Institute IPA

Time: Jan. 2025 – July. 2025

Description:
Focus on robots with dexterous hands for complex industrial processes: - Developed object pose estimation and tracking algorithms using a multi-camera setup - Integrated solutions into ROS2 and optimized performance with TensorRT and C++ - Improved robustness and accelerated existing model by 10x for industrial deployment

Skills:
PyTorch, Segmentation, ROS2, Python, Computer Vision, TensorRT, Docker

Bachelor Thesis: AI and Large Language Models

Company: SCHUNK – Hand in Hand for Tomorrow

Time: Mar. 2024 – Sept. 2024

Description:
Evaluation and implementation of a retrieval-augmented generation (RAG) approach for product data search using generative AI.

Skills:
Python, LangChain, Vector DBs, ChatGPT, PyTorch, Docker, Linux

Software Engineer (Working Student)

Company: IDS Imaging Development Systems GmbH

Time: Sept. 2023 – Mar. 2024

Description:
Worked on IDS lighthouse, a cloud-based AI vision studio: - Integrated an autolabelling service for easier and faster dataset creation. - Dockerised all services for easier development, testing and deployment. - Evaluated and integrated new SOTA detection models

Skills:
Python, Docker, Git, TensorFlow, Computer Vision, Segmentierung

Intern – Computer Vision

Company: IDS Imaging Development Systems GmbH

Time: Mar. 2023 – Sept. 2023

Description:
- Built multi-language examples for using the REST API of the NXT camera - Developed a cloud dashboard to monitor user performance - Trained and evaluated an image detection model for a client project - Created an AI-based service for image labeling and segmentation

Skills:
C/C++, Python, TypeScript, React, PyTorch, Docker, REST, TensorFlow

Test Engineer (Working Student)

Company: Bosch

Time: Mar. 2022 – Feb. 2023

Description:
Worked on engineering, testing, and development of hydraulic systems.

Skills:
Engineering, Testing, R&D

Student Assistant

Company: Heilbronn University – Center for Industrial AI

Time: Jun. 2022 – Dec. 2022

Description:
Assisted in hardware prototyping and embedded projects

Skills:
C, C++, Python, Arduino, Raspberry Pi

πŸŽ“ Education :

Master of Science – Autonomous Systems

Institution: University of Stuttgart

Time: Sept. 2024 – Sept. 2026

Description:
- Research project at Fraunhofer IPA on imitation learning of human grasping tasks with a dexterous robotic arm for industrial applications - Benchmark for evaluating SOTA VLMs on logical game related problem-solving challenges - Literature review on Visiomotor Policies and Vision-Language-Action Models for Robotic Manipulation - Literature review on the application of generative AI in large-scale codebases

Skills:
Foundation Models, Computer Vision, Robotics, Deep Learning, Reinforcement Learning, Artificial Intelligence, LLM

Bachelor of Engineering – Mechatronics and Robotics

Institution: Heilbronn University

Time: Nov. 2020 – Aug. 2024

Description:
Bachelor's thesis GPA: 4.0/4.0 Topic: Retrieval-augmented generation (RAG) for product data search with generative AI Seminar: Time series prediction of chaotic double pendulum using neural networks Projects: - Traffic sign recognition with CNN - Chatbot using LSTMs

Skills:
TensorFlow, Machine Learning, Python, Fusion 360, C++, PyTorch, Git, Time Series Analysis, Image Processing, Robotics, MATLAB, CATIA

πŸ’‘ Personal Projects

I built a complete teleoperation setup to control a robot hand and arm, using this to collect demonstrations and train an adapted 3D Diffusion Policy. My focus was on specific industrial scenarios involving the handling of multiple identical parts. This is normally a nightmare for visuomotor policies because the model cannot distinguish between identical objects based solely on the global features of the vision encoder. Current solutions relying on textual conditioning are insufficient for the industrial context.In addition to conducting experiments on the behaviour of the policy with unknown objects and areas, I also found a method to train the model to indicate when it has finished the task so it can reliably switch back to a classical motion planner.

Features :

  • Developed a teleoperation and imitation learning pipeline using 3D camera and data glove in ROS2.
  • Trained Diffusion Policies to enable robust grasping capabilities in multi-object scenes.
  • Created a full MoveIt model for safe and reliable motion execution planning.
  • Applied the resulting policies to industrial automation scenarios.

Tech Stack :

python diffusion ROS2 pytorch

Traffic Sign Recognition with YOLOv8

Developed as part of a university project at Heilbronn University, this real-time traffic sign recognition system uses YOLOv8n for fast and robust detection. It handles challenging conditions like occlusion, poor lighting, and complex backgrounds by leveraging a custom synthetic dataset, multi-stage classification, and real-time frame filtering.

Features :

  • Real-time traffic sign detection using YOLOv8n (Nano version for speed and efficiency).
  • Custom synthetic dataset generation with COCO backgrounds and heavy augmentation.
  • Two-stage classification specifically for speed limit signs.
  • Frame caching logic to reduce false positives during inference.
  • Visualization via UI overlay: persistent speed sign display + rotating multi-sign view.
  • Trained on 3000+ synthetic images and validated with GTSDB and dashcam footage.
  • Fast inference: ~0.06–0.09 seconds/frame.

Tech Stack :

python opencv yolov8 matplotlib computer-vision

Stabled Grounding SAM

Stabled Grounding SAM is a powerful tool for generating synthetic datasets with pre-segmented images. It combines Stable Diffusion, Grounding DINO, and Segment Anything to create annotated datasets from just a single input image and a label file.

Features :

  • Generates synthetic images from a single input image using Stable Diffusion's `img2img`.
  • Automatically detects and labels objects using Grounding DINO.
  • Refines segmentations using Meta’s Segment Anything model.
  • Outputs datasets in YOLO format for easy training integration.
  • Great for quickly building vision datasets without manual labeling.

Tech Stack :

python stable-diffusion grounding-dino segment-anything yolov5

Get in Touch! πŸ‘‹

Hello there, fellow tech enthusiast! πŸš€

I’m thrilled that you’ve dropped by. Whether you have a burning question, an exciting project idea, or just want to chat about all things code, I’m all ears.

Let’s make this the start of a fantastic conversation. Feel free to reach out, and I’ll get back to you as soon as I can.

Cheers to the future collaborations and coding adventures! 🌟

🌐 Connect with me online:

Discord Linkedin Github

Looking forward to hearing from you soon! πŸ“©