Published September 30, 2023 | Version CC BY-NC-ND 4.0
Journal article Open

Reinforcement Learning based NLP

  • 1. B Tech, Department of Computer Science and Engineering, Lovely Professional University, Phagwara, Punjab, India.

Description

In the field of Natural Language Processing (NLP), reinforcement learning (RL) has drawn attention as a viable method for training models. An agent is trained to interact with a linguistic environment in order to carry out a given task using RL- based NLP, and the agent learns from feedback in the form of rewards or penalties. This method has been effectively used for a variety of linguistic problems, including text summarization, conversation systems, and machine translation. Sequence-to- sequence Two common methods used in RL-based NLP are reinforcement learning and deep reinforcement learning. Sequence-to-sequence While deep reinforcement learning includes training a neural network to discover the optimum strategy for a language challenge, reinforcement learning (RL) trains a model to create a series of words or characters that most closely matches a goal sequence. In several linguistic challenges, RL-based NLP has demonstrated promising results and attained cutting-edge performance. There are still issues to be solved, such as the need for more effective exploration tactics, data scarcity, and sample efficiency. In summary, RL-based NLP represents a potential line of inquiry for NLP research in the future. This method outperforms more established NLP strategies in a variety of language problems and has the added benefit of being able to improve over time with user feedback. To further enhance RL-based NLP's effectiveness and increase its applicability to real-world settings, future research should concentrate on resolving the difficulties associated with this approach.

Notes

Published By: Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP) © Copyright: All rights reserved.

Files

J047610101023.pdf

Files (260.1 kB)

Name Size Download all
md5:bab58d9916d8d91770aa09bebcb7a94e
260.1 kB Preview Download

Additional details

Related works

Is cited by
Journal article: 2231-2307 (ISSN)

References

  • Shu, R., Yoo, K., & Ha, J. (2021). Reward Optimization for Neural Machine Translation with Learned Metrics. ArXiv, abs/2104.07541.

Subjects

ISSN: 2231-2307 (Online)
https://portal.issn.org/resource/ISSN/2231-2307#
Retrieval Number: 100.1/ijsce.J047610101023
https://www.ijsce.org/portfolio-item/J047610101023/
Journal Website: www.ijsce.org
https://www.ijsce.org/
Publisher: Blue Eyes Intelligence Engineering and Sciences Publication (BEIESP)
https://www.blueeyesintelligence.org/