Deep generative models are a category of machine learning models that utilizes deep neural networks to model data distributions and generate new samples. In this seminar, we first introduce our proposed framework to learn a generative model via Schrödinger Bridge, as a stochastic differential equation (SDE)-based generative model. The generative learning task can be formulated as interpolating between a reference distribution and a target distribution based on the Kullback-Leibler divergence, which can be characterized via an SDE on [0, 1] with a time-varying drift term. However, although SDE-based generative models have achieved state-of-the-art performance, they have a less efficient sampling procedure compared with other models such as generative adversarial networks. In the next part, we will discuss feasible ways to solve this problem.

2 May 2022
4:00pm - 5:00pm
Where
https://hkust.zoom.us/j/97557961147 (Passcode: 672570)
Speakers/Performers
Mr. Gefei WANG
Organizer(S)
Department of Mathematics
Contact/Enquiries
Payment Details
Audience
Alumni, Faculty and staff, PG students, UG students
Language(s)
English
Other Events
5 Dec 2025
Seminar, Lecture, Talk
IAS / School of Science Joint Lecture - Human B Cell Receptor-Epitope Selection for Pan-Sarbecovirus Neutralization
Abstract The induction of broadly neutralizing antibodies (bnAbs) against viruses requires the specific activation of human B cell receptors (BCRs) by viral epitopes. Following BCR activation, ...
10 Oct 2025
Seminar, Lecture, Talk
IAS / School of Science Joint Lecture - Use of Large Animal Models to Investigate Brain Diseases
Abstract Genetically modified animal models have been extensively used to investigate the pathogenesis of age-dependent neurodegenerative diseases, such as Alzheimer (AD), Parkinson (PD), Hunti...