Events

Filter by:

Limit to events where the first date of the event:
Date range
Limit to events where the first date of the event:
Limit to events where the title matches:
Limit to events where the type is one or more of:
Limit to events tagged with one or more of:
Limit to events where the audience is one or more of:
Friday, May 4, 2018 2:00 pm - 2:00 pm EDT (GMT -04:00)

PhD Seminar: Regularized Losses for Weakly-supervised CNN Segmentation

Speaker: Meng Tang, PhD candidate

Minimization of regularized losses is a principled approach to weak supervision well established in deep learning, in general. However, it is largely overlooked in semantic segmentation currently dominated by methods mimicking full supervision via "fake" fully-labeled training masks (proposals) generated from available partial input. To obtain such full masks the typical methods explicitly use standard regularization techniques for "shallow" segmentation, e.g., graph cuts or dense CRFs. In contrast, we integrate such standard regularizers and clustering criteria directly into the loss functions over partial input. This approach simplifies weakly-supervised training by avoiding extra MRF/CRF inference steps or layers explicitly generating full masks, while improving both the quality and efficiency of training. 

Speaker: Ivana Kajić, PhD candidate

The representation of semantic knowledge poses a central modelling decision in many models of cognitive phenomena. However, not all such representations reflect properties observed in human semantic networks. Here, we evaluate the psychological plausibility of two distributional semantic models widely used in natural language processing: word2vec and GloVe. We use these models to construct directed and undirected semantic networks and compare them to networks of human association norms using a set of graph-theoretic analyses. 

Tuesday, June 5, 2018 12:00 pm - 12:00 pm EDT (GMT -04:00)

AI Seminar: Making Fair and Efficient Collective Decisions

Nisarg Shah, Department of Computer Science
University of Toronto

Algorithms are increasingly making decisions that affect humans. The field of computational social choice deals with algorithms for eliciting individual preferences and making collective decisions. Everyday examples of such decisions include citizens electing their representatives, roommates dividing collectively purchased items, or residents voting over allocation of city's budget. Making reasonable collective decisions requires viewing the problem through the lenses of elicitation, fairness, efficiency, incentives, and ethics.

Priyank Jaini, PhD candidate
David R. Cheriton School of Computer Science

At their core, many unsupervised learning models provide a compact representation of homogeneous density mixtures, but their similarities and differences are not always clearly understood. In this work, we formally establish the relationships among latent tree graphical models (including special cases such as hidden Markov models and tensorial mixture models), hierarchical tensor formats and sum-product networks.

Thursday, June 21, 2018 9:00 am - 9:00 am EDT (GMT -04:00)

PhD Seminar: Computer Vision on Web Pages: A Study of Man-Made Images

Michael Cormier, PhD candidate

This thesis is focused on the development of computer vision techniques for parsing web pages using an image of the rendered page as evidence, and on understanding this under-explored class of images from the perspective of computer vision. This project is divided into two tracks — applied and theoretical — which complement each other. Our practical motivation is the application of improved web page parsing to assistive technology, such as screenreaders for visually impaired users or the ability to declutter the presentation of a web page for those with cognitive deficit. From a more theoretical standpoint, images of rendered web pages have interesting properties from a computer vision perspective; in particular, low-level assumptions can be made in this domain, but the most important cues are often subtle and can be highly non-local. The parsing system developed in this thesis is a principled Bayesian segmentation-classification pipeline, using innovative techniques to produce valuable results in this challenging domain. The thesis includes both implementation and evaluation solutions.

Abdullah Rashwan, PhD candidate

Sum-product networks have recently emerged as an attractive representation due to their dual view as a special type of deep neural network with clear semantics and a special type of probabilistic graphical model for which inference is always tractable. Those properties follow from some conditions (i.e., completeness and decomposability) that must be respected by the structure of the network. 

Tuesday, July 24, 2018 2:00 pm - 2:00 pm EDT (GMT -04:00)

PhD Seminar: Gradient-based Filter Design for the Dual-tree Wavelet Transform

Daniel Recoskie, PhD candidate
David R. Cheriton School of Computer Science

The wavelet transform has seen success when incorporated into neural network architectures, such as in wavelet scattering networks. More recently, it has been shown that the dual-tree complex wavelet transform can provide better representations than the standard transform.

Thursday, December 13, 2018 4:00 pm - 4:00 pm EST (GMT -05:00)

PhD Seminar: Beyond LIF: The Computational Power of Passive Dendritic Trees

Andreas Stöckel, PhD candidate
David R. Cheriton School of Computer Science

The artificial neurons typically employed in machine learning and computational neuroscience bear little resemblance to biological neurons. They are often derived from the “leaky integrate and fire” (LIF) model, neglect spatial extent, and assume a linear combination of input variables. It is well known that these simplifications have a profound impact on the family of functions that can be computed in a single-layer neural network. 

Friday, December 14, 2018 3:00 pm - 3:00 pm EST (GMT -05:00)

PhD Seminar: Progressive Memory Banks for Incremental Domain Adaptation

Nabiha Asghar, PhD candidate
David R. Cheriton School of Computer Science

We address the problem of incremental domain adaptation (IDA). We assume each domain comes one after another, and that we could only access data in the current domain. The goal of IDA is to build a unified model performing well on all the domains that we have encountered. We propose to augment a recurrent neural network (RNN) with a directly parameterized memory bank, which is retrieved by an attention mechanism at each step of RNN transition. The memory bank provides a natural way of IDA: when adapting our model to a new domain, we progressively add new slots to the memory bank, which increases the number of parameters, and thus the model capacity.