Martin Förtsch

keynote

Martin Förtsch is an IT consultant at TNG Technology Consulting GmbH, located in Unterföhring near Munich. He holds a degree in computer sciences. Professionally, he specializes in Agile Development, primarily in Java, Search Engine Technologies, Information Retrieval, and Databases. As an Intel Software Innovator and Intel Black Belt Software Developer, Martin plays an active role in the creation of innovative tech-demos and showcases alongside the Innovation Hacking team.In the past, Martin developed software for gesture-controlled quadrocopters using Intel RealSense technology and designed an Augmented Reality wearable prototype device. He frequently presents at both national and international conferences on topics such as Artificial Intelligence, Internet of Things, Augmented Reality, and Test Driven Development. His prowess as a speaker has earned him multiple accolades, including the Oracle JavaOne Rockstar award.

remote

Thomas Endres

keynote

Thomas Endres is a Managing Partner for TNG Technology Consulting in Munich. Besides his normal work for the company and the management of the customers he is creating various prototypes, he and the the Innovation Hacking Team develop various prototypes, including an Augmented Reality application that displays the world from an artist's perspective, real-time deepfakes, or an AI for generating presentations. He works on applications in the field of AR/VR, AI, and gesture control to e.g. autonomously fly quadcopters or control them without physical contact. Furthermore, he is involved in various Open Source projects. Thomas studied computer science at TU Munich and is a passionate software developer. As an Intel Software Innovator and Black Belt, he presents new technologies such as AI, AR/VR, and robotics worldwide. For this, he has received awards including a JavaOne Rockstar Award and several Best Speaker Awards.

AI'll Be Back: Exploring Diffusion Models for Image, Audio, and Video

This talk introduces you to the world of generative AI with a focus on Text-to-Image, Text-to-Audio and Text-to-Video for creating images, music and short videos. We explain how neural networks can generate various output formats from short text inputs using diffusion models and so-called Transformer architectures.We focus on advanced technologies such as Sora or Midjourney. The techniques used, such as Latent Diffusion Models, allow us to generate and edit images and videos by combining text understanding through attention mechanisms and transformers with denoising processes.A detailed examination of the video generation process with Sora shows how it compresses visual data, breaks it into patches, and then reconstructs it into the final video. In addition to Sora, we also discuss alternative methods and tools like RunwayML or SunoAI, to present a broad spectrum of tools for image, audio and video generation.By the end of this talk, you will have a basic understanding of diffusion models, an overview of tools for image, audio and video generation, and a deeper understanding of the functionality. Practical examples and demos round off the presentation.

Starting from: $536

Renew Your Mind at LambdaConf 2025

Buy tickets