Djamé Seddah

Follow

Generating author description...

All published works
Action Title Year Authors
+ PDF Chat Beyond Dataset Creation: Critical View of Annotation Variation and Bias Probing of a Dataset for Online Radical Content Detection 2024 Arij Riabi
Virginie Mouilleron
Menel Mahamdi
Wissam Antoun
Djamé Seddah
+ PDF Chat Common Ground, Diverse Roots: The Difficulty of Classifying Common Examples in Spanish Varieties 2024 Javier A. Lopetegui
Arij Riabi
Djamé Seddah
+ PDF Chat CamemBERT 2.0: A Smarter French Language Model Aged to Perfection 2024 Wissam Antoun
Francis Kulumba
Rian Touchent
Éric Villemonte de la Clergerie
Benoît Sagot
Djamé Seddah
+ Cloaked Classifiers: Pseudonymization Strategies on Sensitive Classification Tasks 2024 Arij Riabi
Menel Mahamdi
Virginie Mouilleron
Djamé Seddah
+ PDF Chat On the Granularity of Explanations in Model Agnostic NLP Interpretability 2023 Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+ PDF Chat Data-Efficient French Language Modeling with CamemBERTa 2023 Wissam Antoun
Benoît Sagot
Djamé Seddah
+ Data-Efficient French Language Modeling with CamemBERTa 2023 Wissam Antoun
Benoît Sagot
Djamé Seddah
+ Towards a Robust Detection of Language Model Generated Text: Is ChatGPT that Easy to Detect? 2023 Wissam Antoun
Virginie Mouilleron
Benoît Sagot
Djamé Seddah
+ Enriching the NArabizi Treebank: A Multifaceted Approach to Supporting an Under-Resourced Language 2023 Arij Riabi
Menel Mahamdi
Djamé Seddah
+ From Text to Source: Results in Detecting Large Language Model-Generated Content 2023 Wissam Antoun
Benoît Sagot
Djamé Seddah
+ PDF Chat Treebanking user-generated content: a UD based overview of guidelines, corpora and unified recommendations 2022 Manuela Sanguinetti
Cristina Bosco
Lauren C. Cassidy
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
+ PDF Chat Towards Unsupervised Content Disentanglement in Sentence Representations via Syntactic Roles 2022 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ PDF Chat Can Character-based Language Models Improve Downstream Task Performance in Low-Resource and Noisy Language Scenarios? 2022 Arij Riabi
Benoît Sagot
Djamé Seddah
+ Exploiting Inductive Bias in Transformers for Unsupervised Disentanglement of Syntax and Semantics with VAEs 2022 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ Towards Unsupervised Content Disentanglement in Sentence Representations via Syntactic Roles 2022 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ PDF Chat Exploiting Inductive Bias in Transformers for Unsupervised Disentanglement of Syntax and Semantics with VAEs 2022 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ Multilingual Auxiliary Tasks Training: Bridging the Gap between Languages for Zero-Shot Transfer of Hate Speech Detection Models 2022 Syrielle Montariol
Arij Riabi
Djamé Seddah
+ PDF Chat Multilingual Auxiliary Tasks Training: Bridging the Gap between Languages for Zero-Shot Transfer of Hate Speech Detection Models 2022 Syrielle Montariol
Arij Riabi
Djamé Seddah
+ PDF Chat First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT 2021 Benjamin Müller
Yanai Elazar
Benoît Sagot
Djamé Seddah
+ PDF Chat Can Multilingual Language Models Transfer to an Unseen Dialect? A Case Study on North African Arabizi 2021 Benjamin Müller
Benoît Sagot
Djamé Seddah
+ PDF Chat When Being Unseen from mBERT is just the Beginning: Handling New Languages With Multilingual Language Models 2021 Benjamin Müller
Antonios Anastasopoulos
Benoît Sagot
Djamé Seddah
+ First Align, then Predict: Understanding the Cross-Lingual Ability of Multilingual BERT 2021 Benjamin Müller
Yanai Elazar
Benoît Sagot
Djamé Seddah
+ Challenging the Semi-Supervised VAE Framework for Text Classification 2021 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ PAGnol: An Extra-Large French Generative Model 2021 Julien Launay
E. L. Tommasone
Baptiste Pannier
François Boniface
Amélie Chatelain
Alessandro Cappelli
Iacopo Poli
Djamé Seddah
+ Noisy UGC Translation at the Character Level: Revisiting Open-Vocabulary Capabilities and Robustness of Char-Based Models 2021 José Carlos Rosales Núñez
Guillaume Wisniewski
Djamé Seddah
+ Understanding the Impact of UGC Specificities on Translation Quality 2021 José Carlos Rosales Núñez
Djamé Seddah
Guillaume Wisniewski
+ PDF Chat Synthetic Data Augmentation for Zero-Shot Cross-Lingual Question Answering 2021 Arij Riabi
Thomas Scialom
Rachel Keraron
Benoît Sagot
Djamé Seddah
Jacopo Staiano
+ Simple, Interpretable and Stable Method for Detecting Words with Usage Change across Corpora 2021 Hila Gonen
Ganesh Jawahar
Djamé Seddah
Yoav Goldberg
+ Can Character-based Language Models Improve Downstream Task Performance in Low-Resource and Noisy Language Scenarios? 2021 Arij Riabi
Benoît Sagot
Djamé Seddah
+ Noisy UGC Translation at the Character Level: Revisiting Open-Vocabulary Capabilities and Robustness of Char-Based Models 2021 José Carlos Rosales Núñez
Guillaume Wisniewski
Djamé Seddah
+ Challenging the Semi-Supervised VAE Framework for Text Classification 2021 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ Sentence-Based Model Agnostic NLP Interpretability. 2020 Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+ Controlling the Interaction Between Generation and Inference in Semi-Supervised Variational Autoencoders Using Importance Weighting. 2020 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ CamemBERT: a Tasty French Language Model 2020 Louis Martin
Benjamin Müller
Pedro Ortiz Suarez
Yoann Dupont
Laurent Romary
Éric Villemonte de la Clergerie
Djamé Seddah
Benoît Sagot
+ Simple, Interpretable and Stable Method for Detecting Words with Usage Change across Corpora 2020 Hila Gonen
Ganesh Jawahar
Djamé Seddah
Yoav Goldberg
+ Disentangling semantics in language through VAEs and a certain architectural choice 2020 Ghazi Felhi
Joseph Le Roux
Djamé Seddah
+ QUACKIE: A NLP Classification Task With Ground Truth Explanations 2020 Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+ Treebanking User-Generated Content: a UD Based Overview of Guidelines, Corpora and Unified Recommendations 2020 Manuela Sanguinetti
Lauren C. Cassidy
Cristina Bosco
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
+ Synthetic Data Augmentation for Zero-Shot Cross-Lingual Question Answering 2020 Arij Riabi
Thomas Scialom
Rachel Keraron
Benoît Sagot
Djamé Seddah
Jacopo Staiano
+ Controlling the Interaction Between Generation and Inference in Semi-Supervised Variational Autoencoders Using Importance Weighting 2020 Ghazi Felhi
Joseph Leroux
Djamé Seddah
+ On the Granularity of Explanations in Model Agnostic NLP Interpretability 2020 Yves Rychener
Xavier Renard
Djamé Seddah
Pascal Frossard
Marcin Detyniecki
+ Can Multilingual Language Models Transfer to an Unseen Dialect? A Case Study on North African Arabizi 2020 Benjamin Müller
Benoît Sagot
Djamé Seddah
Common Coauthors
Commonly Cited References
Action Title Year Authors # of times referenced
+ RoBERTa: A Robustly Optimized BERT Pretraining Approach 2019 Yinhan Liu
Myle Ott
Naman Goyal
Jingfei Du
Mandar Joshi
Danqi Chen
Omer Levy
Mike Lewis
Luke Zettlemoyer
Veselin Stoyanov
6
+ CamemBERT: a Tasty French Language Model 2020 Louis Martin
Benjamin Müller
Pedro Ortiz Suarez
Yoann Dupont
Laurent Romary
Éric Villemonte de la Clergerie
Djamé Seddah
Benoît Sagot
6
+ Attention is All you Need 2017 Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin
6
+ Unsupervised Cross-lingual Representation Learning at Scale 2020 Alexis Conneau
Kartikay Khandelwal
Naman Goyal
Vishrav Chaudhary
Guillaume Wenzek
Francisco Guzmán
Édouard Grave
Myle Ott
Luke Zettlemoyer
Veselin Stoyanov
6
+ Adam: A Method for Stochastic Optimization 2014 Diederik P. Kingma
Jimmy Ba
5
+ PDF Chat SQuAD: 100,000+ Questions for Machine Comprehension of Text 2016 Pranav Rajpurkar
Jian Zhang
Konstantin Lopyrev
Percy Liang
5
+ Deep Biaffine Attention for Neural Dependency Parsing 2016 Timothy Dozat
Christopher D. Manning
5
+ BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding 2018 Jacob Devlin
Ming‐Wei Chang
Kenton Lee
Kristina Toutanova
4
+ Deep Contextualized Word Representations 2018 Matthew E. Peters
Mark E Neumann
Mohit Iyyer
Matt Gardner
Christopher Clark
Kenton Lee
Luke Zettlemoyer
4
+ Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer 2019 Colin Raffel
Noam Shazeer
Adam Roberts
Katherine Lee
Sharan Narang
Michael Matena
Yanqi Zhou
Wei Li
Peter J. Liu
4
+ Generating Sentences from a Continuous Space 2016 Samuel R. Bowman
Luke Vilnis
Oriol Vinyals
Andrew M. Dai
Rafał Józefowicz
Samy Bengio
4
+ PDF Chat BERT Rediscovers the Classical NLP Pipeline 2019 Ian Tenney
Dipanjan Das
Ellie Pavlick
4
+ PDF Chat A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations 2019 Mingda Chen
Qingming Tang
Sam Wiseman
Kevin Gimpel
3
+ Disentangled Representation Learning for Non-Parallel Text Style Transfer 2019 Vineet John
Lili Mou
Hareesh Bahuleyan
Olga Vechtomova
3
+ BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension 2020 Mike Lewis
Yinhan Liu
Naman Goyal
Marjan Ghazvininejad
Abdelrahman Mohamed
Omer Levy
Veselin Stoyanov
Luke Zettlemoyer
3
+ Attention Is All You Need 2017 Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin
3
+ Syntax-Infused Variational Autoencoder for Text Generation 2019 Xinyuan Zhang
Yi Yang
Siyang Yuan
Dinghan Shen
Lawrence Carin
3
+ Know What You Don’t Know: Unanswerable Questions for SQuAD 2018 Pranav Rajpurkar
Robin Jia
Percy Liang
3
+ Distributed Representations of Words and Phrases and their Compositionality 2013 Tomáš Mikolov
Ilya Sutskever
Kai Chen
Greg S. Corrado
Jeffrey Dean
3
+ SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing 2018 Taku Kudo
John T. E. Richardson
3
+ The State and Fate of Linguistic Diversity and Inclusion in the NLP World 2020 Pratik Joshi
Sebastin Santy
Amar Budhiraja
Kalika Bali
Monojit Choudhury
3
+ Universal Dependencies v2: An Evergrowing Multilingual Treebank Collection 2020 Joakim Nivre
Marie-Catherine de Marneffe
Filip Ginter
Jan Hajič
Christopher D. Manning
Sampo Pyysalo
Sebastian Schuster
Francis M. Tyers
Daniel Zeman
3
+ Language Models are Few-Shot Learners 2020 T. B. Brown
Benjamin Mann
Nick Ryder
Melanie Subbiah
Jared Kaplan
Prafulla Dhariwal
Arvind Neelakantan
Pranav Shyam
Girish Sastry
Amanda Askell
3
+ Cross-lingual Language Model Pretraining 2019 Guillaume Lample
Alexis Conneau
3
+ Beto, Bentz, Becas: The Surprising Cross-Lingual Effectiveness of BERT 2019 Shijie Wu
Mark Dredze
3
+ ALBERT: A Lite BERT for Self-supervised Learning of Language Representations 2019 Zhenzhong Lan
Mingda Chen
Sebastian Goodman
Kevin Gimpel
Piyush Sharma
Radu Soricut
3
+ On the importance of pre-training data volume for compact language models 2020 Vincent Micheli
Martin d’Hoffschmidt
François Fleuret
3
+ Treebanking User-Generated Content: a UD Based Overview of Guidelines, Corpora and Unified Recommendations 2020 Manuela Sanguinetti
Lauren C. Cassidy
Cristina Bosco
Özlem Çetinoğlu
Alessandra Teresa Cignarella
Teresa Lynn
Ines Rehbein
Josef Ruppenhofer
Djamé Seddah
Amir Zeldes
3
+ Isolating Sources of Disentanglement in Variational Autoencoders 2018 Ricky T. Q. Chen
Xuechen Li
Roger Grosse
David Duvenaud
2
+ PDF Chat MTNT: A Testbed for Machine Translation of Noisy Text 2018 Paul Michel
Graham Neubig
2
+ FastText.zip: Compressing text classification models 2016 Armand Joulin
Édouard Grave
Piotr Bojanowski
Matthijs Douze
Hervé Jeǵou
Tomáš Mikolov
2
+ OpenNMT: Neural Machine Translation Toolkit 2018 Guillaume Klein
Yoon Kim
Yuntian Deng
Vincent Nguyen
Jean Sénellart
Alexander M. Rush
2
+ Massively Multilingual Transfer for NER 2019 Afshin Rahimi
Li Yan Yuan
Trevor Cohn
2
+ PDF Chat Subword Regularization: Improving Neural Network Translation Models with Multiple Subword Candidates 2018 Taku Kudo
2
+ Deep Biaffine Attention for Neural Dependency Parsing 2016 Timothy Dozat
Christopher D. Manning
2
+ PDF Chat Universal Adversarial Perturbations 2017 Seyed-Mohsen Moosavi-Dezfooli
Alhussein Fawzi
Omar Fawzi
Pascal Frossard
2
+ Bag of Tricks for Efficient Text Classification 2017 Armand Joulin
Édouard Grave
Piotr Bojanowski
Tomáš Mikolov
2
+ Variational Pretraining for Semi-supervised Text Classification 2019 Suchin Gururangan
Tam Dang
Dallas Card
Noah A. Smith
2
+ PDF Chat Variational Autoencoders Pursue PCA Directions (by Accident) 2019 Michal Rolínek
Dominik Zietlow
Georg Martius
2
+ PDF Chat Fully Character-Level Neural Machine Translation without Explicit Segmentation 2017 Jason D. Lee
Kyunghyun Cho
Thomas Hofmann
2
+ PDF Chat Recurrent Neural Network Grammars 2016 Chris Dyer
Adhiguna Kuncoro
Miguel Ballesteros
Noah A. Smith
2
+ PDF Chat Rationalizing Neural Predictions 2016 Tao Leí
Regina Barzilay
Tommi Jaakkola
2
+ XNMT: The eXtensible Neural Machine Translation Toolkit 2018 Graham Neubig
Matthias Sperber
Xinyi Wang
Matthieu Felix
Austin Matthews
Sarguna Janani Padmanabhan
Qi Ye
Devendra Singh Sachan
Philip Arthur
Pierre Godard
2
+ Fine-tuned Language Models for Text Classification. 2018 Jeremy Howard
Sebastian Ruder
2
+ PDF Chat Enriching Word Vectors with Subword Information 2017 Piotr Bojanowski
Édouard Grave
Armand Joulin
Tomáš Mikolov
2
+ PDF Chat Deep neural networks are easily fooled: High confidence predictions for unrecognizable images 2015 Anh‐Tu Nguyen
Jason Yosinski
Jeff Clune
2
+ OpenNMT: Neural Machine Translation Toolkit 2018 Guillaume Klein
Yoon Kim
Yuntian Deng
Vincent Nguyen
Jean Sénellart
Alexander M. Rush
2
+ PDF Chat A Structured Variational Autoencoder for Contextual Morphological Inflection 2018 Lawrence Wolf-Sonkin
Jason Naradowsky
Sebastian J. Mielke
Ryan Cotterell
2
+ Efficient Estimation of Word Representations in Vector Space 2013 Tomáš Mikolov
Kai Chen
Greg S. Corrado
Jay B. Dean
2
+ Synthetic and Natural Noise Both Break Neural Machine Translation 2017 Yonatan Belinkov
Yonatan Bisk
2