Theses and Dissertations
Permanent URI for this collectionhttp://ir.daiict.ac.in/handle/123456789/1
Browse
2 results
Search Results
Item Open Access Sentence detection(2020) Shah, Pushya; Mitra, Suman K.Sentence detection is a very important task for any natural language processing (NLP) application. Accuracy and performance of all other downstream natural language processing (NLP) task like Sentiment, Text Classification, named entity recognition (NER), Relation, etc depends on the accuracy of correctly detected sentence boundary. Clinical domain is very different compare to general domain of languages. Clinical sentence structure and vocabulary are different from general English. That’s why available sentence boundary detector tools are not performing well on clinical domain and we required a specific sentence detection model for clinical documents. ezDI Solutions (India) LLP have developed such system that can detect the sentence boundary. We examined Bidirectional Encoder Representations from Transformers (BERT) and Bidirectional Long Short-Term Memory (BiLSTM) algorithm and used BiLSTM-BERT hybrid model for sentence boundary detection on medical corpora.Item Open Access What does BERT learn about questions(2020) Tyagi, Akansha; Majumder, PrasenjitRecent research in Question Answering is highly motivated by the introduction of the BERT [5] model. This model has gained considerable attention since the researcher of Google AI Language has claimed state-of-the-art results over various NLP tasks, including QA. On one side, where the introduction of end-to-end pipeline models consisting of an IR and an RC model has opened the scope of research in two different areas, new BERT representations alone show a significant improvement in the performance of a QA system. In this study, we have covered several pipeline models like R3: Reinforced Ranker-Reader [15], Re-Ranker Model [16], and Interactive Retriever-Reader Model [4] along with the transformer-based QA system i.e., BERT. The motivation of this work is to deeply understand the black-box BERT model and try to identify the BERT’s learning about the question to predict the correct answer for it from a given context. We will discuss all the experiments that we have performed to understand BERT’s behavior from a different perspective. We have performed all the experiments using the SQuAD dataset. We have also used the LRP [3] technique to get a better understanding and for a better analysis of the experiment results. Along with the study about what the model learns, we have also tried to find what the model does not learn. For this, we have analyzed various examples from the dataset to determine the types of questions for whom the model predicts an incorrect answer. Finally, we have presented the overall findings of the BERT model in the conclusion section.