Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page Not Found
Page not found. Your pixels are in another canvas.
About
About me
Page not in menu
This is a page not in th emain menu
Future Blog Post
Published:
This post will show up by default. To disable scheduling of future posts, edit config.yml
and set future: false
.
Blog Post number 4
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 3
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 2
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Blog Post number 1
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Machine Reading of Historical Events
Paper Codeby Or Honovich*, Lucas Torroba Hennigen*, Omri Abend, Shay B. Cohen
in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, July 2020
We introduce the historical event ordering (HEO) task, where a series of short textual descriptions of historical events, potentially alongside some additional information, are ordered chronologically. We compile two datasets for this task, and compare the performance of two models in it.
Intrinsic Probing Through Dimension Selection
Paper Codeby Lucas Torroba Hennigen, Adina Williams, Ryan Cotterell
in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, November 2020
We introduce a novel framework for intrinsic probing that leverages a decomposable multivariate Gaussian probe. We run experiments on 36 languages from the Universal Dependencies treebanks, and find that fastText concentrates its linguistic structure more than BERT.
Classifying Dyads for Militarized Conflict Analysis
Paper Codeby Niklas Stoehr, Lucas Torroba Hennigen, Samin Ahbab, Robert West, Ryan Cotterell
in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, November 2021
We compare dyadic and systemic correlates of conflict in terms of their ability to infer if two entities are allies or enemies. Our results suggests that our systemic features appear to be more correlated.
Probing as Quantifying the Inductive Bias of Pre-trained Representations
Paper Codeby Alexander Immer*, Lucas Torroba Hennigen*, Vincent Fortuin, Ryan Cotterell
in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, May 2022
We propose a new method to evaluate the inductive biases of pre-trained NLP representations which addresses limitations of previous work in probing. Our results suggest that fastText may offer a better inductive bias than BERT in certain multilingual morphosyntactic tasks.
Same Neurons, Different Languages: Probing Morphosyntax in Multilingual Pre-trained Models
Paper Codeby Karolina Stańczak, Edoardo Ponti, Lucas Torroba Hennigen, Ryan Cotterell, Isabelle Augenstein
in Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, July 2022
We compare the neurons that are most informative of particular morphosyntactic categories in multilingual representations, and find significant overlap across languages.
A Latent-Variable Model for Intrinsic Probing
Paper Codeby Karolina Stańczak*, Lucas Torroba Hennigen*, Adina Williams, Ryan Cotterell, Isabelle Augenstein
in Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, February 2023
We introduce a novel latent-variable formulation of intrinsic probing which yields tighter mutual information estimates than previously proposed methods.
An Ordinal Latent Variable Model of Conflict Intensity
Paper Codeby Niklas Stoehr, Lucas Torroba Hennigen, Josef Valvoda, Robert West, Ryan Cotterell, Aaron Schein
in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, July 2023
We introduce a latent variable model for that models conflict intensity as an ordinal latent variable.
Generalizing Backpropagation for Gradient-based Interpretability
Paper Codeby Kevin Du, Lucas Torroba Hennigen, Niklas Stoehr, Alexander Warstadt, Ryan Cotterell
in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, July 2023
By viewing backpropagation as a particular instance of a semiring algorithm, we explore its generalization to other semirings as a tool for interpretability.
A Measure-Theoretic Characterization of Tight Language Models
Paperby Li Du, Lucas Torroba Hennigen, Tiago Pimentel, Clara Meister, Jason Eisner, Ryan Cotterell
in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, July 2023
We provide a measure-theoretic characterization of language models and use it to characterize the notion of tightness more precisely.
Deriving Language Models from Masked Language Models
Paper Codeby Lucas Torroba Hennigen, Yoon Kim
in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, July 2023
We explore and benchmark ways of deriving a joint distribution from the unary conditionals specified by a masked language model.
Towards Verifiable Text Generation with Symbolic References
Paperby Lucas Torroba Hennigen*, Shannon Shen*, Aniruddha Nrusimha, Bernhard Gapp, David Sontag, Yoon Kim
preprint on arXiv, November 2023
We develop a method for more verifiable text generation by prompting LLMs to generate their output using symbolic references into some source data.
Principled Gradient-based Markov Chain Monte Carlo for Text Generation
Paperby Li Du, Afra Amini, Lucas Torroba Hennigen, Xinyan Velocity Yu, Jason Eisner, Holden Lee, Ryan Cotterell
preprint on arXiv, December 2023
We show that previous gradient-based sampling methods for text generation are unfaithful to the true target distribution, and propose faithful alternatives.
Talk 1 on Relevant Topic in Your Field
Published:
This is a description of your talk, which is a markdown files that can be all markdown-ified like any other post. Yay markdown!
Conference Proceeding talk 3 on Relevant Topic in Your Field
Published:
This is a description of your conference proceedings talk, note the different field in type. You can put anything in this field.
Teaching experience 1
Undergraduate course, University 1, Department, 2014
This is a description of a teaching experience. You can use markdown like any other post.
Teaching experience 2
Workshop, University 1, Department, 2015
This is a description of a teaching experience. You can use markdown like any other post.