报告题目 (Title):Variational Model based Attention/Transformer Mechanisms for Image Inverse Problem (基于变分法的图像反问题的注意力/Transformer机制)
报告人 (Speaker):刘君 副教授(北京师范大学)
报告时间 (Time):2023年11月24日(周五) 10:00
报告地点 (Place):腾讯会议 533326207
邀请人(Inviter):彭亚新
报告摘要:Features extracted by the deep convolution neural networks (DCNN) are always complicated and difficult to model. We developed a method to integrate the features prior into the DCNN architectures by a variational method. It is built upon the universal approximation property of the probability density functions for the mixture distributions. By considering the duality of the maximum likelihood estimation for the deep features in high dimension space, several mechanisms such as learnable fidelity, regularizer, segmentation with geometry prior would be proposed. It partly reveals the connections between the variational methods and some popular DCNN architectures in image processing. For example, weighted norm and attention, nonlocal regularization and transformer, dual and translation, multi-grid and encoder-decoder U-net.