Events

Departmental events are listed below. Please see the left column for other special event pages.

October 29th, 2024 - FM Power Hour - Dr. Ryan Ferguson's Talk

Time: October 29th, 2024, 1:30 PM - 2:30 PM
Location: Western Science Centre 248
Speaker: Dr. Ryan Ferguson

Ryan is Founder and CEO at Riskfuel, a capital markets focused AI startup. Previously, Ryan was Managing Director and Head of Securitization, Credit Derivatives and XVA at Scotiabank. Prior roles have included credit correlation trading and managing the equity derivatives trading desk. Ryan began his career with positions in risk management and financial engineering. Ryan has a PhD in Physics from Imperial College, and a BASc and MASc in Electrical Engineering from the University of Waterloo.

Title: Riskfuel: Accelerating valuation and risk sensitivity calculations in the capital markets - Replacing slow numerical solvers with fast neural network inferencing

Abstract: Riskfuel uses traditional numerical solvers to train deep neural networks via the creation of large datasets. Once trained, a network can return accurate results millions of times faster than the traditional approach. This talk will cover the motivation for fast pricing and the various approaches available to accelerate pricing. We will show how neural network inferencing leverages many of these approaches simultaneously, including function approximation, hardware specialization (GPUs), and adjoint algorithmic differentiation. Examples from exotic derivatives pricing and portfolio optimization will be used to demonstrate the technique.

October 17th, 2024 - Ana Carolina da Cruz's PhD thesis proposal public lecture

Supervisors: Dr. Camila de Souza
Time: October 17th, 2024, 10:00 AM - 10:45 AM

Location: Middlesex College 204

Title: Bayesian methods for clustering change-point data and functional data analysis

Abstract: Technological advancements have made high-dimensional data, such as multiple sequences of change-point data and functional data, increasingly available. However, the complexity of such data presents significant challenges in the data analysis, necessitating efficient and reliable methodologies. Bayesian methods, which integrate prior knowledge and manage model complexity, are commonly used in data analysis. Variational inference methods, in particular, are becoming increasingly popular for estimating Bayesian models due to their efficiency and low computational cost. In this thesis proposal, I introduce my research contribution towards three novel Bayesian methods addressing different types of high-dimensional data. In the first project,  I introduce a nonparametric Bayesian model via a Dirichlet process prior to cluster sequences of observations based on their constant-wise change-point profiles via a Gibbs sampler. In the second project, I develop a variational EM algorithm for basis function selection in functional data representation accounting for within-curve correlation. Finally, the third project focuses on variable selection in functional regression, particularly scalar-on-function regression. I propose a variational Bayes algorithm for SoFR using Bernoulli latent variables to enable variable selection.

October 8th, 2024 - FM Power Hour - Dr. Sebastian Ferrando's Talk

Time: October 8th, 2024, 1:30 PM - 2:30 PM
Location: Western Science Centre 248
Speaker: Dr. Sebastian Ferrando - Professor, Department of Mathematics, Toronto Metropolitan University.

Title: Agent-Based Models for Two Stocks with Superhedging

Abstract: An agent-based modelling methodology for the joint price evolution of two stocks is introduced. The method models future multidimensional price trajectories reflecting how a class of agents rebalance their portfolios in an operational way by reacting to how stocks’ charts unfold. Prices are expressed in units of a third stock that acts as numeraire. The methodology is robust, in particular, it does not depend on any prior probability or analytical assumptions and it is based on constructing scenarios/trajectories. A main ingredient is a superhedging interpretation that provides relative superhedging prices between the two modelled stocks. The operational nature of the methodology gives objective conditions for the validity of the model and so implies realistic risk-rewards profiles for the agent’s operations. Superhedging computations are performed with a dynamic programming algorithm deployed on a graph data structure. The superhedging algorithm handles null sets in a rigorous and intuitive way.

September 24th, 2024 - FM Power Hour - Dr. Letitia Golubitsky's Talk

Time: September 24th, 2024, 11:30 AM - 12:30 PM
Location: Western Science Centre 248
Speaker: Dr. Letitia Golubitsky
 
Letitia-Golubitsky.jpg
 
Dr. Letitia Golubitsky is a mathematician by training and an experienced professional in the financial industry holding a PhD in Mathematics from Western University, a Master of Science in Mathematics from Queen's University and a Master of Science in Financial Mathematics from McMaster University. Letitia started her career as a quantitative developer in Model Development at TD bank developing in-house mathematical models for Counterparty Credit Risk. Later, she joined CIBC Model Development as a senior quantitative developer building in-house commodity models for Counterparty Credit Risk and Market Risk. Letitia has worked for the past seven years in Model Validation at Bank of Montreal and  Scotiabank as a senior specialist and senior manager responsible for vetting commodity models and  pricing derivatives used by the trading desks.  Dr. Letitia Golubitsky is the Chair of the Canadian Artificial Intelligence 2022 conference leading the Industry track. Dr. Letitia was elected to the Queen's University Council Executive Committee for a four year term.

Title: Introduction to Counterparty Credit Risk Modelling

Abstract: Counterparty risk is traditionally thought of as credit risk between derivatives counterparties. Since the credit risk crisis of 2008 and the failure of major institutions such as Lehman Brothers, Fannie Mae and Freddie Mac, counterparty risk has been considered by most market participants to be the key financial risk. In this talk I will introduce various counterparty exposure metrics with an emphasis on Monte Carlo simulations of risk factors models based on stochastic differential equations. Model calibration for most of the models relies on sophisticated optimization methods and an accurate estimation of the model parameters is key to estimating the exposure metrics with the counterparty.

September 20th, 2024 - Chengqian Xian's PhD public lecture

Supervisors: Dr. Camila de Souza, Dr. Wenqing He, and Dr. Felipe Rodrigues
Time: September 20th, 2024, 1:30 PM - 2:30 PM

Location: ZOOM (Waiting room activated. No passcode. After the public lecture, except the examination committee, all other participants are asked to leave the room so that the examination can start.)

Title: Variational Bayesian inference for functional data clustering and survival data analysis

Abstract: Variational Bayesian inference is a method to approximate the posterior distribution under a Bayesian model analytically. As an alternative to Markov Chain Monte Carlo (MCMC) methods, variational inference (VI) produces an analytical solution to an approximation of the posterior but have a lower computational cost compared to MCMC methods. The main challenge of applying VI comes from deriving the equations used to update the approximated posterior parameters iteratively, especially when dealing with complex data. In this thesis, we apply the VI to the context of functional data clustering and survival data analysis. The main objective is to develop novel VI algorithms and investigate their performance under these complex statistical models.

In functional data analysis, clustering aims to identify underlying groups of curves without prior group membership information. The first project in this thesis presents a novel variational Bayes (VB) algorithm for simultaneous clustering and smoothing of functional data using a B-spline regression mixture model with random intercepts. The deviance information criterion is employed to select the optimal number of clusters.

The second project shifts focus to survival data analysis, proposing a novel mean-field VB algorithm to infer parameters of the log-logistic accelerated failure time (AFT) model. To address intractable calculations, we propose and incorporate a piecewise approximation technique into the VB algorithm, achieving Bayesian conjugacy.

The third project is motivated by invasive mechanical ventilation data from intensive care units (ICUs) in Ontario, Canada, which form multiple clusters. We assume that patients within the same ICU cluster are correlated. Extending the second project's methodology, a shared frailty log-logistic AFT model is introduced to account for intra-cluster correlation through a cluster-specific random intercept. A novel and fast VB algorithm for model parameter inference is presented.

Extensive simulation studies assess the performance of the proposed VB algorithms, comparing them with other methods, including MCMC algorithms. Applications to real data, such as ICU ventilation data from Ontario, illustrate the methodologies' practical use. The proposed VB algorithms demonstrate excellent performance in clustering functional data and analyzing survival data, while significantly reducing computational cost compared to MCMC methods.

September 17th, 2024 - Yu Shi's PhD public lecture

Supervisors: Dr. Grace Yi
Time: September 17th, 2024, 12:30 PM - 1:30 PM

Location: Zoom (Waiting room activated. No passcode. After the public lecture, except the examination committee, all other participants are asked to leave the room so that the examination can start.)

Title: Conditional Dependence Learning of Noisy Data under Graphical Models

Abstract: Graphical models are useful tools for characterizing the conditional dependence among variables with complex structures. While many methods have been developed under graphical models, their validity is vulnerable to the quality of data. A fundamental assumption associated with most available methods is that the variables need to be precisely measured. This assumption is, however, commonly violated in reality. In addition, the frequent occurrence of missingness in data exacerbates the difficulties of estimation within the context of graphical models. Ignoring either mismeasurement or missingness effects in estimation procedures can yield biased results, and it is imperative to accommodate these effects when conducting inferences under graphical models. In this thesis, we address challenges arising from noisy data with measurement error or missing observations within the framework of graphical models for conditional dependence learning.

The first project addresses mixed graphical models applied to data involving mismeasurement in discrete and continuous variables. We propose a mixed latent Gaussian copula graphical measurement error model to describe error-contaminated data with mixed continuous and discrete variables. To estimate the model parameters, we develop a simulation-based expectation-maximization method that incorporates the measurement error effects. Furthermore, we devise a computationally efficient procedure to implement the proposed method. The asymptotic properties of the proposed estimator are established, and the finite sample performance of the proposed method is evaluated by numerical studies.

In contrast to analyzing error-prone variables in the first project, we further examine variables that are susceptible to not only mismeasurement but also missingness. In the second project, we examine noisy data that are subject to both error-contamination and incompleteness, in which we focus on the Ising model designed for learning the conditional dependence structure among binary variables. We extend the conventional Ising model using additional layers of modeling to describe data with both misclassification and missingness. To estimate the model parameters with the misclassification and missingness effects accommodated simultaneously, we develop a new inferential procedure by utilizing the strength of the insertion correction strategy and the inverse probability weighted method. To facilitate the sparsity of the graphical model, we further employ the regularization technique, and accommodate for a class of penalty functions, including widely-used penalty functions such as Lasso, SCAD, MCP, and HT penalties. To broaden the applicability scope, we investigate settings with both fixed and diverging dimensions of the variables, and moreover, we rigorously establish the asymptotic properties of the proposed estimators, with associated regularity conditions identified.

The third project deepens the second one by accommodating mixed variables subject to both mismeasurement and missingness. Unlike the first two projects that focus on a single dataset, we consider the availability of auxiliary datasets from related studies, along with the target study dataset, where data are subjected to missingness. From the measurement error perspective, the target and auxiliary datasets can be regarded as accurate and error-contaminated measurements for the variables of interest, respectively. To describe the conditional dependence relationships among variables, we explore mixed graphical models characterized by the exponential family distributions. Moreover, leveraging the transfer learning strategy, we propose an inferential procedure that accommodates missingness effects to enhance the estimation of the model parameters pertinent to the target study using the information from auxiliary datasets. We rigorously establish theoretical properties for the proposed estimators and evaluate the finite sample performance of the proposed methods through numerical studies.

This thesis contributes new methodologies to address challenges arising from the presence of noisy data with mismeasurement or missing values. The proposed methods broaden the application of graphical models for learning complex conditional dependency among variables of various nature.

September 16th, 2024 - Wei Li Fan's PhD thesis proposal public lecture

Supervisors: Dr. Marcos Escobar-Anel
Time: September 16th, 2024, 12:15 PM - 1:00 PM

Location: Western Science Centre 187

Title: Enhancing portfolio investment strategies through CEV-related frameworks

Abstract: This thesis explores portfolio optimization through four related models: our newly proposed LVO-CEV and SEV-SV models, as well as the established CEV and M-CEV models. The study starts by introducing and investigating a new type of Constant Elasticity of Volatility (CEV) model, titled LVO-CEV. We prove that the stochastic differential equations of the LVO-CEV model could exhibit strong or weak solutions contingent upon the elasticity parameter. Additionally, the model offers closed-form solutions under hyperbolic absolute risk aversion (HARA) utilities, enhancing tailored investment strategies. Empirical comparisons with other models validate its efficacy in real-world scenarios. Subsequently, the thesis addresses ambiguity aversion among utility-maximizing investors, employing both the LVO-CEV and standard CEV models. This investigation is ongoing and will be completed for the final defense. Furthermore, the thesis extends its examination to include ambiguity aversion within the framework of a Modified Constant-Elasticity-of-Volatility (M-CEV) model, on the underlying asset. Through this exploration, we derive closed-form solutions of a non-affine nature for the optimal asset allocation and the value function, leveraging a Cauchy problem approach. This analysis represents a significant advancement, as we extend existing research to the presence of ambiguity while also accommodating Hyperbolic Absolute Risk Aversion utility (HARA). Lastly, we introduce and analyze a very general and novel family of diffusion models for stock prices, with direct applications in portfolio optimization. This innovative model (SEV-SV) integrates stochastic elasticity of volatility (SEV) with stochastic volatility (SV). Emphasis is placed on the SEV component, driven by an Ornstein-Uhlenbeck process with two distinct functional choices, while the SV component utilizes the 4/2 model. This endeavor yields closed-form solutions for optimal strategy, value function, and optimal wealth process, elucidating two distinct scenarios regarding prices of risk associated with the stock and highlighting the models' applicability in real-world investment contexts.

September 12th, 2024 - Pingbo Hu's PhD public lecture

Supervisors: Dr. Grace Yi
Time: September 12th, 2024, 9:00 AM - 10:00 AM

Location: Western Science Centre 248


Title: Statistical Learning of Noisy Data: Classification and Causal Inference with Measurement Error and Missingness


Abstract: Causal inference and statistical learning have made significant advancements in various fields, including healthcare, epidemiology, computer vision, information retrieval, and language processing. Despite numerous methods, research gaps still remain, particularly regarding noisy data with features such as missing data, censoring, and measurement errors, etc. Addressing the challenges presented by noisy data is crucial to reduce bias and enhance statistical learning of such data. This thesis tackles several issues in causal inference and statistical learning that are related to noisy data.

The first project addresses causal inference about longitudinal studies with bivariate responses, focusing on data with missingness and censoring. We decompose the overall treatment effect into two separable effects, each mediated through different causal pathways. Furthermore, we establish identification conditions for estimating these separable treatment effects using observed data. Subsequently, we employ the likelihood method to estimate these effects and derive hypothesis testing procedures for their comparison.

In the second project, we tackle the problem of detecting cause-effect relationships between two sets of variables, formed as two vectors. Although this problem can be framed as a binary classification task, it is prone to mislabeling of causal relationships for paired vectors under the study - an inherent challenge in causation studies. We quantify the effects of mislabeled outputs on training results and introduce metrics to characterize these effects. Furthermore, we develop valid learning methods that account for mislabeling effects and provide theoretical justification for their validity. Our contributions present reliable learning methods designed to handle real-world data, which commonly involve label noise.

The third project extends the research in the second project by exploring binary classification with noisy data in the general framework. To scrutinize the impact of different types of label noise, we introduce a sensible way to categorize noisy labels into three types: instance-dependent, semi-instance-independent, and instance-independent noisy labels. We theoretically assess the impact of each noise type on learning. In particular, we quantify an upper bound of bias when ignoring the effects of instance-dependent noisy labels and identify conditions under which ignoring semi-instance-independent noisy labels is acceptable. Moreover, we propose correction methods for each type of noisy label.

Contrasting with the third project that focuses on classification with label noise, the fourth project examines binary classification with mismeasured inputs. We begin by theoretically analyzing the bias induced by ignoring measurement error effects and identify a scenario where such an ignorance is acceptable. We then propose three correction methods to address the mismeasured input effects, including methods leveraging validation data and modifications to the loss function using regression calibration and conditional expectation. Finally, we establish theoretical results for each proposed method.

In summary, this thesis explores several interesting problems in causal inference and statistical learning concerning noisy data. We contribute new findings and methods to enhance our understanding of the complexities induced by noisy data and provide solutions to address them.

September 10th, 2024 - Dr. Laura Cowen's Talk

Speaker: Dr. Laura Cowen - Professor, The University of Victoria
Time: September 10th, 2024, 2:30 PM - 3:30 PM

Location: Western Science Centre 248

Title: Disease analytic models with applications to estimating undetected COVID-19 cases

Abstract: Even with daily case counts, the true scope of the COVID-19 pandemic in Canada is unknown due to undetected cases. We develop a novel multi-site disease analytics model which estimates undetected cases using discrete-valued multivariate time series in the framework of Bayesian hidden Markov modelling techniques. We apply our multi-site model to estimate the pandemic scope using publicly available disease count data including detected cases, recoveries among detected cases, and total deaths. These counts are used to estimate the case detection probability, the probability of recovery, and several important population parameters including the rate of spread, and importation of external cases. We estimate the total number of active COVID-19 cases per region of Canada for each reporting interval. We applied this multi-site model Canada-wide to all provinces and territories, providing an estimate of the total COVID-19 burden for the 90 weeks from 23 Apr 2020 to 10 Feb 2022. We also applied this model to the five Health Authority regions of British Columbia, Canada, describing the pandemic in B.C. over the 31 weeks from 2 Apr 2020 to 30 Oct 2020.

August 28th, 2024 - Duo Xu's PhD thesis proposal public lecture

Supervisors: Dr. Shu Li
Time: August 28th, 2024, 8:30 AM - 9:10 AM
Location: Western Science Centre 248

Title: Drawdown-dependent surplus analysis and applications

Abstract: This thesis mainly focus on the drawdown-dependent surplus analysis and its applications in various espects. In Section 2, we proposed a new fee structure for the drawdown insurance and analyze the fair market premium under the fee structre, we further examine the optimal terminations for the policyholder when a cancellation feature is allowed in the insurance contract, and make the comparison between the two premium structures. In Section 3, we extend our analysis to encompass the joint Laplace transform concerning the first-passage time, occupation time, and local time within the drawdown process. Under spectral negative L\'evy process, we provided expressions for the joint Laplace transforms of these three components in scenarios involving both two-sided exit and one-sided exit. Sections 4 and 5 delve into the realm of equity-linked insurance, with a focus on valuation and associated considerations. In Section 4, we present valuation outcomes concerning guaranteed minimum death benefit (GMDB) and guaranteed minimum mature benefit (GMMB) under a state-dependent fee structure. In section 5, we delve into the surrender option available to policyholders, allowing them to opt out at any point before maturity. We proceed by addressing the fair valuation problem associated with equity-linked insurance in light of this surrender option. We then tackle the optimization of surrender decisions by establishing specific surrender criteria and employing Hamilton-Jacobi-Bellman (HJB) equations to derive optimal solutions.

August 26th, 2024 - Yuan Bian's PhD public lecture

Supervisors: Dr. Grace Y. Yi & Dr. Wenqing He
Time: August 26th, 2024, 9:15 AM - 10:00 AM
Location: Western Science Centre 248

Title: Statistical inference and learning with incomplete data

Abstract:

Incomplete data commonly arise in applications, and research on this topic has received extensive attention over the past few decades. Numerous inference methods have been developed to address various issues related to incomplete data, such as different types of missing observations and distinct missing data mechanisms, which are often classified as missing completely at random, missing at random, and missing not at random. However, research gaps still remain.

Assessing a plausible missing data mechanism is typically difficult due to the lack of validation data, and the presence of spurious variables in covariates further complicates the challenge. Prediction in the presence of incomplete data is another area worth exploring. By utilizing newly emerging techniques, we explore new avenues in the analysis of incomplete data. This thesis aims to contribute fresh insights into statistical inference within the context of incomplete data and provide valid methods to address a few existing research gaps.

Focusing on missingness in the response variable, the first project proposes a unified framework to address the effects of missing data. By leveraging the generalized linear model to facilitate the dependence of the response on associated covariates, we develop concurrent estimation and variable selection procedures using regularized likelihood. We rigorously establish the asymptotic properties of the resultant estimators. The proposed methods offer flexibility and generality, eliminating the need to assume a specific missing data mechanism -- a requirement in most available methods. Empirical studies demonstrate the satisfactory performance of the proposed methods in finite sample settings. Furthermore, the project outlines extensions to accommodate missingness in both the response and covariates.

The second problem of interest approaches missing data from a different perspective by placing it within the framework of statistical machine learning, with a specific emphasis on exploring boosting techniques; two projects are generated accordingly. Despite the increasing attention gained by boosting, many advancements in this area have primarily focused on numerical implementation procedures, with relatively limited theoretical work. Moreover, existing boosting approaches are predominantly designed to handle datasets with complete observations, and their validity is hampered by the presence of missing data. In this thesis, we employ semiparametric estimation approaches to develop unbiased boosting estimation methods for data with missing responses. We investigate several strategies to account for the missingness effects. The proposed methods are implemented using the functional gradient descent algorithm and are justified by the establishment of theoretical properties, including convergence and consistency of the proposed estimators. Numerical studies confirm the satisfactory performance of the proposed methods in finite sample settings.

The third topic further explores different boosting procedures in the context of interval censored data, where the exact observed value for the response variable is unavailable but only known to fall within an interval. Such data commonly arise in survival analysis and fields involving time-to-events, and they present a unique challenge in data analysis. In this project, we develop boosting methods for both regression and classification problems with interval censored data. We address the censoring effects by adjusting the loss functions or imputing transformed responses. The proposed methods are implemented using a functional gradient descent algorithm, and we rigorously establish their theoretical properties, including mean squared error tradeoffs and the optimality of the proposed estimators. Numerical studies are conducted to assess the performance of the proposed methods in finite sample settings.

August 26th, 2024 - Dr. John Braun's Talk

Speaker: Dr. John Braun - Professor, The University of British Columbia
Time: August 26th, 2024, 3:00 PM - 4:00 PM
Location: Western Science Centre 248

Title: Iterated Data Sharpening for Local Polynomial Regression

Abstract: Data sharpening in kernel regression has been shown to be an effective method of reducing
 bias while having minimal effects on variance. Earlier efforts to iterate the data sharpening 
procedure have been less effective, due to the employment of an inappropriate sharpening
 transformation. In the present talk, an iterated data sharpening algorithm is described which
 reduces the asymptotic bias at each iteration, while having modest effects on the variance. The efficacy of the iterative approach is demonstrated
 theoretically and via a simulation study. Boundary effects persist and the affected region successively
 grows when the iteration is applied to local 
 constant regression. By contrast, boundary bias successively decreases for each iteration step when applied to local linear regression. After iteration, the resulting estimates
 are less sensitive to bandwidth choice, and a further simulation study demonstrates that iterated
 data sharpening with data-driven
 bandwidth selection via cross-validation can lead to more accurate regression function estimation. 
 Examples with real data are used to illustrate the scope of change made possible
 by using iterated data sharpening and to also identify its limitations.

(Based on joint work with Hanxiao Chen of Boston University and Xiaoping Shi of UBC)

August 26th, 2024 - Dr. Severien Nkurunziza's Talk

Speaker: Dr. Severien Nkurunziza - Professor, The University of Windsor
Time: August 26th, 2024, 4:00 PM - 5:00 PM
Location: Western Science Centre 248

Title: On Robust Inference In Some Mean-Reverting Processes With Change-Points

Abstract: In this talk, we consider inference problem concerning the drift parameter in generalized mean-reverting processes with unknown change-points. We also consider the scenario where the target parameter is suspected to satisfy some restrictions. We generalize some recent findings in five ways. First, the established method incorporates the uncertain prior knowledge. Second, we derive the unrestricted estimator (UE) and the restricted estimator (RE) as well as their asymptotic properties. Third, we propose a test for testing the hypothesized restrictions and we establish its asymptotic power. Fourth, we construct a class of shrinkage estimators (SEs) which includes as special cases the UE, RE, and classical SEs. Fifth, we study the asymptotic risk performance of the proposed class of SEs, and we prove that James-Stein type estimators dominate the UE. On the top of these interesting findings, the additional novelty of the derived results consists in the fact that the dimensions of the proposed estimators are random. Because of that, the asymptotic power of the proposed test and the asymptotic risk analysis do not follow from classical results in statistical literature. To overcome this problem, we establish an asymptotic result which is useful in its own.