Projects
Reading
People
Chat
SU\G
(𝔸)
/K·U
Projects
Reading
People
Chat
Sign Up
Light
Dark
System
Djamé Seddah
Follow
Share
Generating author description...
All published works
Action
Title
Year
Authors
+
PDF
Chat
Beyond Dataset Creation: Critical View of Annotation Variation and Bias Probing of a Dataset for Online Radical Content Detection
2024
Arij Riabi
Virginie Mouilleron
Menel Mahamdi
Wissam Antoun
Djamé Seddah
+
PDF
Chat
Common Ground, Diverse Roots: The Difficulty of Classifying Common Examples in Spanish Varieties
2024
Javier A. Lopetegui
Arij Riabi
Djamé Seddah
+
PDF
Chat
CamemBERT 2.0: A Smarter French Language Model Aged to Perfection
2024
Wissam Antoun
Francis Kulumba
Rian Touchent
Éric Villemonte de la Clergerie
Benoît Sagot
Djamé Seddah
+
Cloaked Classifiers: Pseudonymization Strategies on Sensitive Classification Tasks
2024
Arij Riabi
Menel Mahamdi
Virginie Mouilleron
Djamé Seddah
+
PDF
Chat
On the Granularity of Explanations in Model Agnostic NLP Interpretability
2023
Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+
PDF
Chat
Data-Efficient French Language Modeling with CamemBERTa
2023
Wissam Antoun
Benoît Sagot
Djamé Seddah
+
Data-Efficient French Language Modeling with CamemBERTa
2023
Wissam Antoun
Benoît Sagot
Djamé Seddah
+
Towards a Robust Detection of Language Model Generated Text: Is ChatGPT that Easy to Detect?
2023
Wissam Antoun
Virginie Mouilleron
Benoît Sagot
Djamé Seddah
+
Enriching the NArabizi Treebank: A Multifaceted Approach to Supporting an Under-Resourced Language
2023
Arij Riabi
Menel Mahamdi
Djamé Seddah
+
From Text to Source: Results in Detecting Large Language Model-Generated Content
2023
Wissam Antoun
Benoît Sagot
Djamé Seddah
+
PDF
Chat
Treebanking user-generated content: a UD based overview of guidelines, corpora and unified recommendations
2022
Manuela Sanguinetti
Cristina Bosco
Lauren C. Cassidy
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
+
PDF
Chat
Towards Unsupervised Content Disentanglement in Sentence Representations via Syntactic Roles
2022
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
PDF
Chat
Can Character-based Language Models Improve Downstream Task Performance in Low-Resource and Noisy Language Scenarios?
2022
Arij Riabi
Benoît Sagot
Djamé Seddah
+
Exploiting Inductive Bias in Transformers for Unsupervised Disentanglement of Syntax and Semantics with VAEs
2022
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
Towards Unsupervised Content Disentanglement in Sentence Representations via Syntactic Roles
2022
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
PDF
Chat
Exploiting Inductive Bias in Transformers for Unsupervised Disentanglement of Syntax and Semantics with VAEs
2022
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
Multilingual Auxiliary Tasks Training: Bridging the Gap between Languages for Zero-Shot Transfer of Hate Speech Detection Models
2022
Syrielle Montariol
Arij Riabi
Djamé Seddah
+
PDF
Chat
Multilingual Auxiliary Tasks Training: Bridging the Gap between Languages for Zero-Shot Transfer of Hate Speech Detection Models
2022
Syrielle Montariol
Arij Riabi
Djamé Seddah
+
PDF
Chat
First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT
2021
Benjamin Müller
Yanai Elazar
Benoît Sagot
Djamé Seddah
+
PDF
Chat
Can Multilingual Language Models Transfer to an Unseen Dialect? A Case Study on North African Arabizi
2021
Benjamin Müller
Benoît Sagot
Djamé Seddah
+
PDF
Chat
When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models
2021
Benjamin Müller
Antonios Anastasopoulos
Benoît Sagot
Djamé Seddah
+
First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT
2021
Benjamin Müller
Yanai Elazar
Benoît Sagot
Djamé Seddah
+
Challenging the Semi-Supervised VAE Framework for Text Classification
2021
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
PAGnol: An Extra-Large French Generative Model
2021
Julien Launay
E. L. Tommasone
Baptiste Pannier
François Boniface
Amélie Chatelain
Alessandro Cappelli
Iacopo Poli
Djamé Seddah
+
Noisy UGC Translation at the Character Level: Revisiting Open-Vocabulary Capabilities and Robustness of Char-Based Models
2021
José Carlos Rosales Núñez
Guillaume Wisniewski
Djamé Seddah
+
Understanding the Impact of UGC Specificities on Translation Quality
2021
José Carlos Rosales Núñez
Djamé Seddah
Guillaume Wisniewski
+
PDF
Chat
Synthetic Data Augmentation for Zero-Shot Cross-Lingual Question Answering
2021
Arij Riabi
Thomas Scialom
Rachel Keraron
Benoît Sagot
Djamé Seddah
Jacopo Staiano
+
Simple, Interpretable and Stable Method for Detecting Words with Usage Change across Corpora
2021
Hila Gonen
Ganesh Jawahar
Djamé Seddah
Yoav Goldberg
+
Can Character-based Language Models Improve Downstream Task Performance in Low-Resource and Noisy Language Scenarios?
2021
Arij Riabi
Benoît Sagot
Djamé Seddah
+
Noisy UGC Translation at the Character Level: Revisiting Open-Vocabulary Capabilities and Robustness of Char-Based Models
2021
José Carlos Rosales Núñez
Guillaume Wisniewski
Djamé Seddah
+
Challenging the Semi-Supervised VAE Framework for Text Classification
2021
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
Sentence-Based Model Agnostic NLP Interpretability.
2020
Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+
Controlling the Interaction Between Generation and Inference in Semi-Supervised Variational Autoencoders Using Importance Weighting.
2020
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
CamemBERT: a Tasty French Language Model
2020
Louis Martin
Benjamin Müller
Pedro Ortiz Suarez
Yoann Dupont
Laurent Romary
Éric Villemonte de la Clergerie
Djamé Seddah
Benoît Sagot
+
Simple, Interpretable and Stable Method for Detecting Words with Usage Change across Corpora
2020
Hila Gonen
Ganesh Jawahar
Djamé Seddah
Yoav Goldberg
+
Disentangling semantics in language through VAEs and a certain architectural choice
2020
Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+
QUACKIE: A NLP Classification Task With Ground Truth Explanations
2020
Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+
Treebanking User-Generated Content: a UD Based Overview of Guidelines, Corpora and Unified Recommendations
2020
Manuela Sanguinetti
Lauren C. Cassidy
Cristina Bosco
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
+
Synthetic Data Augmentation for Zero-Shot Cross-Lingual Question Answering
2020
Arij Riabi
Thomas Scialom
Rachel Keraron
Benoît Sagot
Djamé Seddah
Jacopo Staiano
+
Controlling the Interaction Between Generation and Inference in Semi-Supervised Variational Autoencoders Using Importance Weighting
2020
Ghazi Felhi
Joseph Leroux
Djamé Seddah
+
On the Granularity of Explanations in Model Agnostic NLP Interpretability
2020
Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+
Can Multilingual Language Models Transfer to an Unseen Dialect? A Case Study on North African Arabizi
2020
Benjamin Müller
Benoît Sagot
Djamé Seddah
Common Coauthors
Coauthor
Papers Together
Benoît Sagot
15
Arij Riabi
8
Joseph Le Roux
8
Ghazi Felhi
7
Wissam Antoun
6
Xavier Renard
4
Yves Rychener
4
Pascal Frossard
4
Marcin Detyniecki
4
Benjamin Müller
3
Menel Mahamdi
3
José Carlos Rosales Núñez
3
Virginie Mouilleron
3
Benjamin Müller
3
Guillaume Wisniewski
3
Cristina Bosco
2
Josef Ruppenhofer
2
Teresa Lynn
2
Yoav Goldberg
2
Éric Villemonte de la Clergerie
2
Alessandra Teresa Cignarella
2
Arij Riabi
2
Ghazi Felhi
2
Özlem Çetinoğlu
2
Rachel Keraron
2
Ganesh Jawahar
2
Syrielle Montariol
2
Ines Rehbein
2
Lauren C. Cassidy
2
Manuela Sanguinetti
2
Yanai Elazar
2
Jacopo Staiano
2
Hila Gonen
2
Amir Zeldes
2
Thomas Scialom
2
E. L. Tommasone
1
Francis Kulumba
1
Jesús A. González
1
Yoann Dupont
1
Louis Martin
1
Amélie Chatelain
1
Alessandro Cappelli
1
Joseph Leroux
1
Javier A. Lopetegui
1
Julien Launay
1
François Boniface
1
Pedro Ortiz Suarez
1
Iacopo Poli
1
Laurent Romary
1
Rian Touchent
1
Commonly Cited References
Action
Title
Year
Authors
# of times referenced
+
RoBERTa: A Robustly Optimized BERT Pretraining Approach
2019
Yinhan Liu
Myle Ott
Naman Goyal
Jingfei Du
Mandar Joshi
Danqi Chen
Omer Levy
Mike Lewis
Luke Zettlemoyer
Veselin Stoyanov
6
+
CamemBERT: a Tasty French Language Model
2020
Louis Martin
Benjamin Müller
Pedro Ortiz Suarez
Yoann Dupont
Laurent Romary
Éric Villemonte de la Clergerie
Djamé Seddah
Benoît Sagot
6
+
Attention is All you Need
2017
Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin
6
+
Unsupervised Cross-lingual Representation Learning at Scale
2020
Alexis Conneau
Kartikay Khandelwal
Naman Goyal
Vishrav Chaudhary
Guillaume Wenzek
Francisco Guzmán
Édouard Grave
Myle Ott
Luke Zettlemoyer
Veselin Stoyanov
6
+
Adam: A Method for Stochastic Optimization
2014
Diederik P. Kingma
Jimmy Ba
5
+
PDF
Chat
SQuAD: 100,000+ Questions for Machine Comprehension of Text
2016
Pranav Rajpurkar
Jian Zhang
Konstantin Lopyrev
Percy Liang
5
+
Deep Biaffine Attention for Neural Dependency Parsing
2016
Timothy Dozat
Christopher D. Manning
5
+
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
2018
Jacob Devlin
Ming‐Wei Chang
Kenton Lee
Kristina Toutanova
4
+
Deep Contextualized Word Representations
2018
Matthew E. Peters
Mark E Neumann
Mohit Iyyer
Matt Gardner
Christopher Clark
Kenton Lee
Luke Zettlemoyer
4
+
Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
2019
Colin Raffel
Noam Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
4
+
Generating Sentences from a Continuous Space
2016
Samuel R. Bowman
Luke Vilnis
Oriol Vinyals
Andrew M. Dai
Rafał Józefowicz
Samy Bengio
4
+
PDF
Chat
BERT Rediscovers the Classical NLP Pipeline
2019
Ian Tenney
Dipanjan Das
Ellie Pavlick
4
+
PDF
Chat
A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations
2019
Mingda Chen
Qingming Tang
Sam Wiseman
Kevin Gimpel
3
+
Disentangled Representation Learning for Non-Parallel Text Style Transfer
2019
Vineet John
Lili Mou
Hareesh Bahuleyan
Olga Vechtomova
3
+
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
2020
Mike Lewis
Yinhan Liu
Naman Goyal
Marjan Ghazvininejad
Abdelrahman Mohamed
Omer Levy
Veselin Stoyanov
Luke Zettlemoyer
3
+
Attention Is All You Need
2017
Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin
3
+
Syntax-Infused Variational Autoencoder for Text Generation
2019
Xinyuan Zhang
Yi Yang
Siyang Yuan
Dinghan Shen
Lawrence Carin
3
+
Know What You Don’t Know: Unanswerable Questions for SQuAD
2018
Pranav Rajpurkar
Robin Jia
Percy Liang
3
+
Distributed Representations of Words and Phrases and their Compositionality
2013
Tomáš Mikolov
Ilya Sutskever
Kai Chen
Greg S. Corrado
Jeffrey Dean
3
+
SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing
2018
Taku Kudo
John T. E. Richardson
3
+
The State and Fate of Linguistic Diversity and Inclusion in the NLP World
2020
Pratik Joshi
Sebastin Santy
Amar Budhiraja
Kalika Bali
Monojit Choudhury
3
+
Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection
2020
Joakim Nivre
Marie-Catherine de Marneffe
Filip Ginter
Jan Hajič
Christopher D. Manning
Sampo Pyysalo
Sebastian Schuster
Francis M. Tyers
Daniel Zeman
3
+
Language Models are Few-Shot Learners
2020
T. B. Brown
Benjamin Mann
Nick Ryder
Melanie Subbiah
Jared Kaplan
Prafulla Dhariwal
Arvind Neelakantan
Pranav Shyam
Girish Sastry
Amanda Askell
3
+
Cross-lingual Language Model Pretraining
2019
Guillaume Lample
Alexis Conneau
3
+
Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT
2019
Shijie Wu
Mark Dredze
3
+
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
2019
Zhenzhong Lan
Mingda Chen
Sebastian Goodman
Kevin Gimpel
Piyush Sharma
Radu Soricut
3
+
On the importance of pre-training data volume for compact language models
2020
Vincent Micheli
Martin d’Hoffschmidt
François Fleuret
3
+
Treebanking User-Generated Content: a UD Based Overview of Guidelines, Corpora and Unified Recommendations
2020
Manuela Sanguinetti
Lauren C. Cassidy
Cristina Bosco
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
3
+
Isolating Sources of Disentanglement in Variational Autoencoders
2018
Ricky T. Q. Chen
Xuechen Li
Roger Grosse
David Duvenaud
2
+
PDF
Chat
MTNT: A Testbed for Machine Translation of Noisy Text
2018
Paul Michel
Graham Neubig
2
+
FastText.zip: Compressing text classification models
2016
Armand Joulin
Édouard Grave
Piotr Bojanowski
Matthijs Douze
Hervé Jeǵou
Tomáš Mikolov
2
+
OpenNMT: Neural Machine Translation Toolkit
2018
Guillaume Klein
Yoon Kim
Yuntian Deng
Vincent Nguyen
Jean Sénellart
Alexander M. Rush
2
+
Massively Multilingual Transfer for NER
2019
Afshin Rahimi
Li Yan Yuan
Trevor Cohn
2
+
PDF
Chat
Subword Regularization: Improving Neural Network Translation Models with Multiple Subword Candidates
2018
Taku Kudo
2
+
Deep Biaffine Attention for Neural Dependency Parsing
2016
Timothy Dozat
Christopher D. Manning
2
+
PDF
Chat
Universal Adversarial Perturbations
2017
Seyed-Mohsen Moosavi-Dezfooli
Alhussein Fawzi
Omar Fawzi
Pascal Frossard
2
+
Bag of Tricks for Efficient Text Classification
2017
Armand Joulin
Édouard Grave
Piotr Bojanowski
Tomáš Mikolov
2
+
Variational Pretraining for Semi-supervised Text Classification
2019
Suchin Gururangan
Tam Dang
Dallas Card
Noah A. Smith
2
+
PDF
Chat
Variational Autoencoders Pursue PCA Directions (by Accident)
2019
Michal Rolínek
Dominik Zietlow
Georg Martius
2
+
PDF
Chat
Fully Character-Level Neural Machine Translation without Explicit Segmentation
2017
Jason D. Lee
Kyunghyun Cho
Thomas Hofmann
2
+
PDF
Chat
Recurrent Neural Network Grammars
2016
Chris Dyer
Adhiguna Kuncoro
Miguel Ballesteros
Noah A. Smith
2
+
PDF
Chat
Rationalizing Neural Predictions
2016
Tao Leí
Regina Barzilay
Tommi Jaakkola
2
+
XNMT: The eXtensible Neural Machine Translation Toolkit
2018
Graham Neubig
Matthias Sperber
Xinyi Wang
Matthieu Felix
Austin Matthews
Sarguna Janani Padmanabhan
Qi Ye
Devendra Singh Sachan
Philip Arthur
Pierre Godard
2
+
Fine-tuned Language Models for Text Classification.
2018
Jeremy Howard
Sebastian Ruder
2
+
PDF
Chat
Enriching Word Vectors with Subword Information
2017
Piotr Bojanowski
Édouard Grave
Armand Joulin
Tomáš Mikolov
2
+
PDF
Chat
Deep neural networks are easily fooled: High confidence predictions for unrecognizable images
2015
Anh‐Tu Nguyen
Jason Yosinski
Jeff Clune
2
+
OpenNMT: Neural Machine Translation Toolkit
2018
Guillaume Klein
Yoon Kim
Yuntian Deng
Vincent Nguyen
Jean Sénellart
Alexander M. Rush
2
+
PDF
Chat
A Structured Variational Autoencoder for Contextual Morphological Inflection
2018
Lawrence Wolf-Sonkin
Jason Naradowsky
Sebastian J. Mielke
Ryan Cotterell
2
+
Efficient Estimation of Word Representations in Vector Space
2013
Tomáš Mikolov
Kai Chen
Greg S. Corrado
Jay B. Dean
2
+
Synthetic and Natural Noise Both Break Neural Machine Translation
2017
Yonatan Belinkov
Yonatan Bisk
2