세미나 및 콜로퀴엄

2020-09
Sun Mon Tue Wed Thu Fri Sat
    1 2 3 4 5
6 7 8 9 10 11 12
13 14 15 16 17 1 18 1 19
20 21 22 23 24 1 25 26
27 28 29 30      
2020-10
Sun Mon Tue Wed Thu Fri Sat
        1 2 3
4 5 6 7 8 1 9 10
11 12 13 1 14 1 15 16 17
18 19 20 21 22 23 24
25 26 27 28 29 30 31

구글 Calendar나 iPhone 등에서 구독하면 세미나 시작 전에 알림을 받을 수 있습니다.

Recently, with the enormous development of deep learning techniques, solving underdetermined linear systems (more unknowns than equations) have become one of major concerns in medical imaging. Typical examples include undersampled MRI, local tomography, and sparse view CT, where deep learning techniques have shown excellent performance. Although deep learning methods appear to overcome limitations of existing mathematical methods in handling various underdetermined problems, there is a lack of rigorous mathematical foundations which would allow us to understand reasons why deep learning methods perform that well. This talk deals with this learning causal relationship about structure of training data suitable for deep learning to solve highly underdetermined inverse problems. We examine whether or not a desired reconstruction map can be learnable from the training data and the underdetermined system. Most problems of solving underdetermined linear systems in medical imaging are highly non-linear.
Host: 이창옥     미정     2020-09-09 11:03:00
In this talk, we introduce an idea of producing rational torsion points on J_0(N), which is well-known. Conjecturally, the points constructed in this way exhaust all the rational torsion points on J_0(N). So, we briefly explain how to compute the orders of such points, and prove the conjecture up to finitely many primes. (If you would like to join this online seminar, please contact Bo-Hae Im to get the Zoom link.)
Host: Bo-Hae Im     미정     2020-09-16 13:54:13
The essential dimension quantifies the algebraic-geometric complexity of a class of algebraic objects (such as, but not necessarily, the class of Galois extensions with a given group): roughly speaking, it is the minimal number of parameters required to describe all objects in this class (over all fields containing a given field K). We introduce and discuss arithmetic-geometric and local analogues of this notion. These are supposed to quantify the difference in complexity between the local and global Galois theory of a given group over a given number field K. In particular, we show that the "local dimension" of a finite group is bounded by 2 - whereas arithmetic dimension remains mysterious in general. We give an application concerning solution of Grunwald problems.
Host: 백상훈     영어     2020-09-25 16:09:30
In this talk, I will explain first what is the representation theory with an easy example from finite groups. Next, I will introduce certain algebraic objects called the quantum affine algebras and the quiver Hecke algebras, and explain a connection between those objects via representation theory of associative algebras.
한국어     2020-09-09 11:00:37
Hessian operators are, roughly speaking, the ones that depend on the eigenvalues of the Hessian matrix. Classical examples include the Laplacian and the real and complex Monge-Amp\`ere operator. Typically discussion of Hessian equations is restricted to subfamilies of functions, so that the problem becomes (degenerate) elliptic. In my talk I will discuss the basics of general Hessian equations and explain its links to problems arising in geometric analysis. If time permits I will focus on more specific examples admitting a richer theory.
영어     2020-09-14 10:38:36
Deep neural networks usually act on fixed dimensional items. However, many real-world problems are formulated as learning mappings from sets of items to outputs. Such problems include multiple-instance learning, visual scene understandings, few-shot classifications, and even generic Bayesian inference procedures. Recently, several methods have been proposed to construct neural networks taking sets as inputs. The key properties required for those neural networks are permutation invariance and equivariance, meaning that intermediate outputs and final values of a network should remain unchanged with respect to the processing order of items in sets. This talk discusses recent advances in permutation invariant and equivariant neural networks, and discuss their theoretical properties, especially their universalities. The later part of the talk will also introduce interesting applications of the permutation invariant/equivariant neural networks.
Host: 전현호     미정     2020-09-09 10:58:40