ML

Text transformer for context-aware encoding

This project focuses on training transformer encoders whose representations incorporate information about higher-order context, i.e., characteristics of the author and/or the pragmatic context. We feed models a target sequence and a number of ‘context’ sequences (i.e., text from the same author, or from the same subreddit) as a single example, and train models on a variant of MLM where the MLM head is fed the combination of token-level representations of the input sequence and an aggregate representation of context sequences.

We experiment with three DistilBERT-inspired architectures: a bi-encoder (where context and target are fed to two separate encoders), a ‘batch’ encoder (single encoder with added context aggregation and target-context combination layers) and a hierarchical encoder (applying attention across [CLS] tokens in between standard transformer layers to integrate information across contexts and target sequence). The benefits of this training protocol are evaluated both by comparing their MLM performance with no-context MLM training and to random-context training, as well as on extrinsic tasks.

This project is still in progress.

NLP transformers DistilBERT TensorFlow huggingface ML

Understanding cognitive dimensions of political identity using NLP

I am currently part of an interdisciplinary consortium working on investigating factors shaping political identity at the national and transnational level. As a postdoc at the Interacting Minds Centre at Aarhus University, I am currently working on understanding which cognitive representations underlie people’s attachment to national and transnational institutions using multilingual corpora of social media data and open-text survey data.
applied NLP computational social science data science ML