Loading Events

« All Events

  • This event has passed.

Doctoral Dissertation Oral Proposal, Qinqing Liu

April 22 @ 10:00 am - 11:00 am EDT

Title: Enhancing Attention-Based Models by Learning Dynamics with Static Content
Ph.D. Candidate: Qinqing Liu
Major Advisor: Dr. Jinbo Bi
Associate Advisors: Dr. Dongjin Song, Dr. Derek Aguiar

Date/Time: Monday, April 22, 2024, 10:00 AM
Location: HBL 1947 room
Meeting link:
https://uconn-cmr.webex.com/uconn-cmr/j.php?MTID=m0751d982e1ea32cf52c8578f29bdb9be

Meeting number: 2630 589 7453
Password: h6NFqr

Abstract:
In the rapidly evolving field of data-driven machine learning, learning the underlying patterns within temporal and sequential data is essential for a plethora of applications, ranging from time series analysis to natural language processing. Yet, the challenge lies in effectively distinguishing and leveraging the dynamics intrinsic to these static features not being confounded by the dynamics. The static features could serve as a foundational bedrock, powerfully modulating the dynamic features’ behavior. The dependence of dynamic properties on their static counterparts is far from trivial, suggesting a profound interdependence that warrants deeper examination. This dissertation explores the novel paradigm of “Learning Dynamics with Static Context,” which seeks to enhance the learning capabilities of attention-based models by anchoring the dynamic features with static context.

Building upon this paradigm, our research delves into the fusion of static features with modern machine learning architectures for sequence learning to improve performance and generalization. The introduction of attention and shortcut connections to the Long Short-Term Memory (LSTM) model constitutes our first study, where the LSTM backbone captures dynamic time series patterns, while the static variables provide attention weights to aggregate the embeddings of each time step and also build a shortcut connections between single embedding and final representation. The different stages of dynamics could be distinguished according to the static features and the shortcut connection ensures the information in early stages could be preserved in final representation, thus refining predictions. The second study scales this concept to the Transformer architecture, revolutionizing the treatment of position encodings. Instead of the conventional method of sharing positional encodings among all sequences, we proposed to apply the static variables to inform Customized Positional Encoding that interacts with the self-attention through a Partially Linearized Attention Module. This innovation reconstructs the paradigm of positional awareness in attention-based models, catering to the static variables that underlie the temporal features and proves to be robust. Our forthcoming third study continues this thread, targeting the consistent elements within input and output sequences in text-related tasks. Focusing on seq2seq denoising task, we introduce the Follow-On Tokenizer to output string, which includes a mapping mechanism to identify and preserve consistent tokens between input and output thus minimizing the ‘noise’ or dynamic aspects of the data, thereby enabling the model to focus on rectifying errors such as typographical and phonetic anomalies. This facilitates more precise, context-aware spell correction and suggests broader implications for error remediation in textual data.

Details

Date:
April 22
Time:
10:00 am - 11:00 am EDT
Website:
https://uconn-cmr.webex.com/uconn-cmr/j.php?MTID=m0751d982e1ea32cf52c8578f29bdb9be

Venue

HBL Class of 1947 Conference Room
UConn Library, 369 Fairfield Way, Unit 1005
Storrs, CT 06269 United States
+ Google Map
Phone
(860) 486-2518
View Venue Website

Connect With Us