Deep generative models are a category of machine learning models that utilizes deep neural networks to model data distributions and generate new samples. In this seminar, we first introduce our proposed framework to learn a generative model via Schrödinger Bridge, as a stochastic differential equation (SDE)-based generative model. The generative learning task can be formulated as interpolating between a reference distribution and a target distribution based on the Kullback-Leibler divergence, which can be characterized via an SDE on [0, 1] with a time-varying drift term. However, although SDE-based generative models have achieved state-of-the-art performance, they have a less efficient sampling procedure compared with other models such as generative adversarial networks. In the next part, we will discuss feasible ways to solve this problem.

2 May 2022
4:00pm - 5:00pm
Where
https://hkust.zoom.us/j/97557961147 (Passcode: 672570)
Speakers/Performers
Mr. Gefei WANG
Organizer(S)
Department of Mathematics
Contact/Enquiries
Payment Details
Audience
Alumni, Faculty and staff, PG students, UG students
Language(s)
English
Other Events
22 Nov 2024
Seminar, Lecture, Talk
IAS / School of Science Joint Lecture - Leveraging Protein Dynamics Memory with Machine Learning to Advance Drug Design: From Antibiotics to Targeted Protein Degradation
Abstract Protein dynamics are fundamental to protein function and encode complex biomolecular mechanisms. Although Markov state models have made it possible to capture long-timescale protein co...
8 Nov 2024
Seminar, Lecture, Talk
IAS / School of Science Joint Lecture - Some Theorems in the Representation Theory of Classical Lie Groups
Abstract After introducing some basic notions in the representation theory of classical Lie groups, the speaker will explain three results in this theory: the multiplicity one theorem for classical...