International Journal of Science and Research (IJSR)

International Journal of Science and Research (IJSR)
Call for Papers | Fully Refereed | Open Access | Double Blind Peer Reviewed

ISSN: 2319-7064

Downloads: 0 | Views: 3

Research Paper | Computer Technology | India | Volume 10 Issue 10, October 2021 | Rating: 3.1 / 10


Weak Labelers for Iteratively Improving Models Faster

Ashish Bansal [2]


Abstract: Deep neural networks are becoming omnipresent in natural language applications (NLP). How - ever, they require large amounts of labeled training data, which is often only available for English. This is a big challenge for many languages and domains where labeled data is limited. In recent years, a variety of methods have been proposed to tackle this situation. This paper gives an overview of these approaches that help you train NLP models in resource - lean scenarios. This includes both ideas to increase the amount of labeled data as well as methods following the popular pre - train and ?ne - tune paradigm. Supervised learning techniques construct predictive models by learning from a large number of training examples, where each training example has a label indicating its ground - truth output. Though current techniques have achieved great success, it is noteworthy that in many tasks it is di?cult to get strong supervision information like fully ground - truth labels due to the high cost of the data - labeling process. Thus, it is desirable for machine - learning techniques to work with weak supervision. This paper outlines the advantages of weakly supervised learning in collecting more robust data fastly and using less resource, focusing on three typical types of weak supervision: incomplete supervision, where only a subset of training data is given with labels; inexact supervision, where the training data are given with only coarse - grained labels; and inaccurate supervision, where the given labels are not always ground - truth. The main focus will be on the weak supervision technique where we will explain how a smaller dataset is used to train a classi?er model and then that model is used to label the new data having weak labels which might be accurately predicting those labels to some extent. This method in - volves human - in - loop where human would reviews those predicted labels and correct the wrong predictions which create an additional data points to train a new weak labeler model. Using this technique iteratively it helped researchers in creating more ground truth data that can be used to train better performing models very fast.


Keywords: machine learning, weakly supervised learning, supervised learning, NLP


Edition: Volume 10 Issue 10, October 2021,


Pages: 1640 - 1643

Rate this Article


Select Rating (Lowest: 1, Highest: 10)

5

Your Comments

Characters: 0


Type Your Registered Email Address below to Rate the Article


Verification Code will appear in 2 Seconds ... Wait

Top