Events

Filter by:

Limit to events where the first date of the event:
Date range
Limit to events where the first date of the event:
Limit to events where the title matches:
Limit to events where the type is one or more of:
Limit to events tagged with one or more of:
Limit to events where the audience is one or more of:

Please note: This seminar will take place in DC 1302.

Roswitha Rissner, Department of Mathematics
Alpen-Adria-Universität Klagenfurt, Austria

Given a square matrix B' over a (commutative) ring S, the null ideal N_0(B') is the ideal consisting of all polynomials f in S[X] for which f(B')=0. In the case that S=R/J is the residue class ring of a ring R modulo an ideal J, we can equivalently study the so-called J-ideals

N_J(B) =  { f in  R[X]  |  f(B) in M_n(J) }

Please note: This master’s thesis presentation will take place in DC 1304 and virtually.

Benjamin Thérien, Master’s candidate
David R. Cheriton School of Computer Science

Supervisor: Professor Krzysztof Czarnecki

Please note: This master’s thesis presentation will take place in DC 3317 and virtually over Zoom.

Eva Feng, Master’s candidate
David R. Cheriton School of Computer Science

Supervisor: Professor David Toman

Monday, April 17, 2023 3:00 pm - 4:00 pm EDT (GMT -04:00)

PhD Seminar • Computer Graphics • A Projective Drawing System

Please note: This PhD seminar will take place online.

Greg Philbrick, PhD candidate
David R. Cheriton School of Computer Science

Supervisor: Professor Craig Kaplan

This paper treats the subject of pseudo-3D modeling (via drawing in projective coordinates). I'll talk about the authors’ methods, as well as my own exploration of pseudo-3D drawing techniques.

Please note: This master’s thesis presentation will take place online.

Odunayo Ogundepo, Master’s candidate
David. R. Cheriton School of Computer Science

Supervisor: Professor Jimmy Lin

Thursday, April 20, 2023 1:00 pm - 2:00 pm EDT (GMT -04:00)

Seminar • Machine Learning • Backpropagation Beyond the Gradient

Please note: This seminar will take place in DC 2585.

Felix Dangel, Postdoctoral Researcher
Vector Institute for Artificial Intelligence

Popular deep learning frameworks prioritize computing the average mini-batch gradient. Yet, other quantities such as its variance or many approximations to the Hessian can be computed efficiently, and at the same time as the gradient mean. They are of great interest to researchers and practitioners, but implementing them is often burdensome or inefficient.