no code implementations • WMT (EMNLP) 2020 • Lucia Specia, Zhenhao Li, Juan Pino, Vishrav Chaudhary, Francisco Guzmán, Graham Neubig, Nadir Durrani, Yonatan Belinkov, Philipp Koehn, Hassan Sajjad, Paul Michel, Xian Li
We report the findings of the second edition of the shared task on improving robustness in Machine Translation (MT).
no code implementations • IWSLT 2016 • Nadir Durrani, Fahim Dalvi, Hassan Sajjad, Stephan Vogel
This paper describes QCRI’s machine translation systems for the IWSLT 2016 evaluation campaign.
1 code implementation • 23 May 2024 • Basel Mousi, Nadir Durrani, Fahim Dalvi, Majd Hawasly, Ahmed Abdelali
Our analysis focuses on quantifying the \textit{alignment} and \textit{overlap} of these concepts across various languages within the latent space.
no code implementations • 18 Apr 2024 • Xuemin Yu, Fahim Dalvi, Nadir Durrani, Hassan Sajjad
Therefore, given a word in context, the latent space derived from our training process reflects a specific facet of that word.
1 code implementation • 20 Aug 2023 • Majd Hawasly, Fahim Dalvi, Nadir Durrani
Despite the revolution caused by deep NLP models, they remain black boxes, necessitating research to understand their decision-making processes.
1 code implementation • 9 Aug 2023 • Fahim Dalvi, Maram Hasanain, Sabri Boughorbel, Basel Mousi, Samir Abdaljalil, Nizi Nazar, Ahmed Abdelali, Shammur Absar Chowdhury, Hamdy Mubarak, Ahmed Ali, Majd Hawasly, Nadir Durrani, Firoj Alam
In this study, we introduce the LLMeBench framework, which can be seamlessly customized to evaluate LLMs for any NLP task, regardless of language.
1 code implementation • 26 May 2023 • Fahim Dalvi, Hassan Sajjad, Nadir Durrani
The Python toolkit is available at https://www. github. com/fdalvi/NeuroX.
no code implementations • 24 May 2023 • Ahmed Abdelali, Hamdy Mubarak, Shammur Absar Chowdhury, Maram Hasanain, Basel Mousi, Sabri Boughorbel, Yassine El Kheir, Daniel Izham, Fahim Dalvi, Majd Hawasly, Nizi Nazar, Yousseif Elshahawy, Ahmed Ali, Nadir Durrani, Natasa Milic-Frayling, Firoj Alam
Our findings provide valuable insights into the applicability of LLMs for Arabic NLP and speech processing tasks.
no code implementations • 22 May 2023 • Basel Mousi, Nadir Durrani, Fahim Dalvi
We propose using a large language model, ChatGPT, as an annotator to enable fine-grained interpretation analysis of pre-trained language models.
no code implementations • 6 Mar 2023 • Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Tamim Jaban, Musab Husaini, Ummar Abbas
NxPlain discovers latent concepts learned in a deep NLP model, provides an interpretation of the knowledge learned in the model, and explains its predictions based on the used concepts.
no code implementations • 12 Nov 2022 • Firoj Alam, Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Abdul Rafae Khan, Jia Xu
We use an unsupervised method to discover concepts learned in these models and enable a graphical interface for humans to generate explanations for the concepts.
no code implementations • 23 Oct 2022 • Nadir Durrani, Hassan Sajjad, Fahim Dalvi, Firoj Alam
We study the evolution of latent space in fine-tuned NLP models.
no code implementations • 18 Oct 2022 • Ahmed Abdelali, Nadir Durrani, Fahim Dalvi, Hassan Sajjad
Given the success of pre-trained language models, many transformer models trained on Arabic and its dialects have surfaced.
1 code implementation • NAACL 2022 • Hassan Sajjad, Nadir Durrani, Fahim Dalvi, Firoj Alam, Abdul Rafae Khan, Jia Xu
We propose a novel framework ConceptX, to analyze how latent concepts are encoded in representations learned within pre-trained language models.
no code implementations • 27 Jun 2022 • Nadir Durrani, Fahim Dalvi, Hassan Sajjad
Our data-driven, quantitative analysis illuminates interesting findings: (i) we found small subsets of neurons that can predict different linguistic tasks, ii) with neurons capturing basic lexical information (such as suffixation) localized in lower most layers, iii) while those learning complex concepts (such as syntactic role) predominantly in middle and higher layers, iii) that salient linguistic neurons are relocated from higher to lower layers during transfer learning, as the network preserve the higher layers for task specific information, iv) we found interesting differences across pre-trained models, with respect to how linguistic information is preserved within, and v) we found that concept exhibit similar neuron distribution across different languages in the multilingual transformer models.
no code implementations • 15 Jun 2022 • Ahmed Abdelali, Nadir Durrani, Cenk Demiroglu, Fahim Dalvi, Hamdy Mubarak, Kareem Darwish
We concatenated Tacotron1 with the WaveRNN vocoder, Tacotron2 with the WaveGlow vocoder and ESPnet transformer with the parallel wavegan vocoder to synthesize waveforms from the spectrograms.
no code implementations • ICLR 2022 • Fahim Dalvi, Abdul Rafae Khan, Firoj Alam, Nadir Durrani, Jia Xu, Hassan Sajjad
We address this limitation by discovering and analyzing latent concepts learned in neural network models in an unsupervised fashion and provide interpretations from the model's perspective.
no code implementations • 19 Jan 2022 • Ahmed Abdelali, Nadir Durrani, Fahim Dalvi, Hassan Sajjad
Arabic is a Semitic language which is widely spoken with many dialects.
no code implementations • 30 Aug 2021 • Hassan Sajjad, Nadir Durrani, Fahim Dalvi
The proliferation of deep neural networks in various domains has seen an increased need for interpretability of these models.
no code implementations • 1 Jul 2021 • Shammur Absar Chowdhury, Nadir Durrani, Ahmed Ali
In our study, we conduct a post-hoc functional interpretability analysis of pretrained speech models using the probing framework [1].
no code implementations • Findings (ACL) 2021 • Nadir Durrani, Hassan Sajjad, Fahim Dalvi
The pattern varies across architectures, with BERT retaining linguistic information relatively deeper in the network compared to RoBERTa and XLNet, where it is predominantly delegated to the lower layers.
no code implementations • NAACL 2021 • Hassan Sajjad, Narine Kokhlikyan, Fahim Dalvi, Nadir Durrani
This paper is a write-up for the tutorial on "Fine-grained Interpretation and Causation Analysis in Deep NLP Models" that we are presenting at NAACL 2021.
no code implementations • COLING 2022 • Hassan Sajjad, Firoj Alam, Fahim Dalvi, Nadir Durrani
However, post-processing for contextualized embeddings is an under-studied problem.
no code implementations • COLING 2020 • Hassan Sajjad, Ahmed Abdelali, Nadir Durrani, Fahim Dalvi
The evaluation suite and the dialectal system are publicly available for research purposes.
1 code implementation • EMNLP 2020 • Nadir Durrani, Hassan Sajjad, Fahim Dalvi, Yonatan Belinkov
We found small subsets of neurons to predict linguistic tasks, with lower level tasks (such as morphology) localized in fewer neurons, compared to higher level task of predicting syntax.
1 code implementation • 15 Jul 2020 • Firoj Alam, Fahim Dalvi, Shaden Shaar, Nadir Durrani, Hamdy Mubarak, Alex Nikolov, Giovanni Da San Martino, Ahmed Abdelali, Hassan Sajjad, Kareem Darwish, Preslav Nakov
With the outbreak of the COVID-19 pandemic, people turned to social media to read and to share timely information including statistics, warnings, advice, and inspirational stories.
no code implementations • WS 2020 • Ebrahim Ansari, Amittai Axelrod, Nguyen Bach, Ond{\v{r}}ej Bojar, Roldano Cattoni, Fahim Dalvi, Nadir Durrani, Marcello Federico, Christian Federmann, Jiatao Gu, Fei Huang, Kevin Knight, Xutai Ma, Ajay Nagesh, Matteo Negri, Jan Niehues, Juan Pino, Elizabeth Salesky, Xing Shi, Sebastian St{\"u}ker, Marco Turchi, Alex Waibel, er, Changhan Wang
The evaluation campaign of the International Conference on Spoken Language Translation (IWSLT 2020) featured this year six challenge tracks: (i) Simultaneous speech translation, (ii) Video speech translation, (iii) Offline speech translation, (iv) Conversational speech translation, (v) Open domain translation, and (vi) Non-native speech translation.
1 code implementation • ACL 2020 • John M. Wu, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, James Glass
We use existing and novel similarity measures that aim to gauge the level of localization of information in the deep models, and facilitate the investigation of which design factors affect model similarity, without requiring any external linguistic annotation.
2 code implementations • Findings (EMNLP) 2021 • Firoj Alam, Shaden Shaar, Fahim Dalvi, Hassan Sajjad, Alex Nikolov, Hamdy Mubarak, Giovanni Da San Martino, Ahmed Abdelali, Nadir Durrani, Kareem Darwish, Abdulaziz Al-Homaid, Wajdi Zaghouani, Tommaso Caselli, Gijs Danoe, Friso Stolk, Britt Bruntink, Preslav Nakov
With the emergence of the COVID-19 pandemic, the political and the medical aspects of disinformation merged as the problem got elevated to a whole new level to become the first global infodemic.
4 code implementations • 8 Apr 2020 • Hassan Sajjad, Fahim Dalvi, Nadir Durrani, Preslav Nakov
Transformer-based NLP models are trained using hundreds of millions or even billions of parameters, limiting their applicability in computationally constrained environments.
1 code implementation • EMNLP 2020 • Fahim Dalvi, Hassan Sajjad, Nadir Durrani, Yonatan Belinkov
Transformer-based deep NLP models are trained using hundreds of millions of parameters, limiting their applicability in computationally constrained environments.
no code implementations • CL 2020 • Yonatan Belinkov, Nadir Durrani, Fahim Dalvi, Hassan Sajjad, James Glass
(iii) Do the representations capture lexical semantics?
1 code implementation • WS 2019 • Xi-An Li, Paul Michel, Antonios Anastasopoulos, Yonatan Belinkov, Nadir Durrani, Orhan Firat, Philipp Koehn, Graham Neubig, Juan Pino, Hassan Sajjad
We share the findings of the first shared task on improving robustness of Machine Translation (MT).
no code implementations • NAACL 2019 • Nadir Durrani, Fahim Dalvi, Hassan Sajjad, Yonatan Belinkov, Preslav Nakov
Recent work has shown that contextualized word representations derived from neural machine translation are a viable alternative to such from simple word predictions tasks.
2 code implementations • 21 Dec 2018 • Fahim Dalvi, Avery Nortonsmith, D. Anthony Bau, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, James Glass
We present a toolkit to facilitate the interpretation and understanding of neural network models.
1 code implementation • 21 Dec 2018 • Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, Anthony Bau, James Glass
We further present a comprehensive analysis of neurons with the aim to address the following questions: i) how localized or distributed are different linguistic properties in the models?
no code implementations • ICLR 2019 • Anthony Bau, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, James Glass
Neural machine translation (NMT) models learn representations containing substantial linguistic information.
no code implementations • NAACL 2018 • Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Stephan Vogel
We address the problem of simultaneous translation by modifying the Neural MT decoder to operate with dynamically built encoder and attention.
no code implementations • IWSLT 2017 • Nadir Durrani, Fahim Dalvi
We also observed improvements compared to the systems that used POS tags and word clusters to train these models.
1 code implementation • IJCNLP 2017 • Yonatan Belinkov, Lluís Màrquez, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, James Glass
In this paper, we investigate the representations learned at different layers of NMT encoders.
no code implementations • IJCNLP 2017 • Fahim Dalvi, Nadir Durrani, Hassan Sajjad, Yonatan Belinkov, Stephan Vogel
End-to-end training makes the neural machine translation (NMT) architecture simpler, yet elegant compared to traditional statistical machine translation (SMT).
no code implementations • ACL 2017 • Hassan Sajjad, Fahim Dalvi, Nadir Durrani, Ahmed Abdelali, Yonatan Belinkov, Stephan Vogel
Word segmentation plays a pivotal role in improving any Arabic NLP application.
no code implementations • IWSLT 2017 • Hassan Sajjad, Nadir Durrani, Fahim Dalvi, Yonatan Belinkov, Stephan Vogel
Model stacking works best when training begins with the furthest out-of-domain data and the model is incrementally fine-tuned with the next furthest domain and so on.
1 code implementation • ACL 2017 • Yonatan Belinkov, Nadir Durrani, Fahim Dalvi, Hassan Sajjad, James Glass
Neural machine translation (MT) models obtain state-of-the-art performance while maintaining a simple, end-to-end architecture.
no code implementations • EACL 2017 • Renars Liepins, Ulrich Germann, Guntis Barzdins, Alex Birch, ra, Steve Renals, Susanne Weber, Peggy van der Kreeft, Herv{\'e} Bourlard, Jo{\~a}o Prieto, Ond{\v{r}}ej Klejch, Peter Bell, Alex Lazaridis, ros, Alfonso Mendes, Sebastian Riedel, Mariana S. C. Almeida, Pedro Balage, Shay B. Cohen, Tomasz Dwojak, Philip N. Garner, Andreas Giefer, Marcin Junczys-Dowmunt, Hina Imran, David Nogueira, Ahmed Ali, Mir, Sebasti{\~a}o a, Andrei Popescu-Belis, Lesly Miculicich Werlen, Nikos Papasarantopoulos, Abiola Obamuyide, Clive Jones, Fahim Dalvi, Andreas Vlachos, Yang Wang, Sibo Tong, Rico Sennrich, Nikolaos Pappas, Shashi Narayan, Marco Damonte, Nadir Durrani, Sameer Khurana, Ahmed Abdelali, Hassan Sajjad, Stephan Vogel, David Sheppey, Chris Hernon, Jeff Mitchell
We present the first prototype of the SUMMA Platform: an integrated platform for multilingual media monitoring.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +5
no code implementations • EACL 2017 • Fahim Dalvi, Yifan Zhang, Sameer Khurana, Nadir Durrani, Hassan Sajjad, Ahmed Abdelali, Hamdy Mubarak, Ahmed Ali, Stephan Vogel
This paper presents QCRI{'}s Arabic-to-English live speech translation system.
no code implementations • 16 Jan 2017 • Nadeem Jadoon Khan, Waqas Anwar, Nadir Durrani
We report the performance of baseline systems translating from Indian languages (Bengali, Guajarati, Hindi, Malayalam, Punjabi, Tamil, Telugu and Urdu) into English with average 10% accurate results for all the language pairs.
no code implementations • 14 Jan 2017 • Nadir Durrani, Fahim Dalvi, Hassan Sajjad, Stephan Vogel
This paper describes QCRI's machine translation systems for the IWSLT 2016 evaluation campaign.
no code implementations • COLING 2016 • Nadir Durrani, Hassan Sajjad, Shafiq Joty, Ahmed Abdelali
We present a novel fusion model for domain adaptation in Statistical Machine Translation.
no code implementations • 9 Oct 2016 • Ahmad Musleh, Nadir Durrani, Irina Temnikova, Preslav Nakov, Stephan Vogel, Osama Alsaad
We present research towards bridging the language gap between migrant workers in Qatar and medical staff.
no code implementations • 18 Jun 2016 • Hassan Sajjad, Nadir Durrani, Francisco Guzman, Preslav Nakov, Ahmed Abdelali, Stephan Vogel, Wael Salloum, Ahmed El Kholy, Nizar Habash
The competition focused on informal dialectal Arabic, as used in SMS, chat, and speech.