AbstractTransformer is a powerful architecture that achieves superior performance on various sequence learning tasks, including neural machine translation, language understanding, and so on. As the core of the architecture, the self-attention mechanism is a kind of kernel smoothing method, or "local model" by the speaker's word. The whole architecure also could be seen as a sequence model of me...
Math+ML+X Seminar SeriesOrganizer:Angelica Aviles-RiveroSpeaker:Shizheng Wen (ETH Zürich)Time:Fri., 16:00 , Mar. 20, 2026Online:Voov (Tencent): 201-467-303Title:Geometry Aware Operator Transformer as an Efficient and Accurate Neural Surrogate for PDEs on Arbitrary DomainsAbstract:Neural operators have emerged as promising surrogates for PDE solvers, yet applying them to domains with comple...