Mathbert: A pre-trained model for mathematical formula understanding

S Peng, K Yuan, L Gao, Z Tang - arxiv preprint arxiv:2105.00377, 2021 - arxiv.org
Large-scale pre-trained models like BERT, have obtained a great success in various Natural
Language Processing (NLP) tasks, while it is still a challenge to adapt them to the math …

Does entity abstraction help generative Transformers reason?

N Gontier, S Reddy, C Pal - arxiv preprint arxiv:2201.01787, 2022 - arxiv.org
We study the utility of incorporating entity type abstractions into pre-trained Transformers
and test these methods on four NLP tasks requiring different forms of logical reasoning:(1) …

Impact of Grammar on Language Model Comprehension

K Ameri, M Hempel, H Sharif, J Lopez… - 2023 International …, 2023 - ieeexplore.ieee.org
Machine Learning and Natural Language Processing are playing an increasingly vital role
in many different areas, including cybersecurity in Information Technology and Operational …

[PDF][PDF] Compositional linguistic generalization in artificial neural networks

N Kim - 2021 - jscholarship.library.jhu.edu
Compositionality---the principle that the meaning of a complex expression is built from the
meanings of its parts---is considered a central property of human language. This dissertation …

Exploring the Impact of Syntactic Structure Information on Unknown Entity Recognition in Transformer-based Natural Language Understanding

R Yang, F Javar, K Wakabayashi… - 2023 14th IIAI …, 2023 - ieeexplore.ieee.org
Natural language understanding (NLU) of user utterances is an essential task in dialogue
systems. However, existing machine learning models suffer from low accuracy in …

Investigating learning in deep neural networks using layer-wise weight change

AM Agrawal, A Tendle, H Sikka, S Singh… - … : Proceedings of the 2021 …, 2021 - Springer
Understanding the learning dynamics of deep neural networks is of significant interest to the
research community as it can provide insights into the black box nature of neural nets. In this …

[PDF][PDF] Occupation coding using a pretrained language model by integrating domain knowledge

VRPV Karanam - 2022 - wwwiti.cs.uni-magdeburg.de
Generally, surveys are considered one of the popular mechanisms to collect data regarding
a problem associated with any field. The researchers use the collected data to conduct …

Applications of Deep Representation Learning to Natural Language Processing and Satellite Imagery

G Wang - 2020 - search.proquest.com
Deep representation learning has shown its effectiveness in many tasks such as text
classification and image processing. Many researches have been done to directly improve …

WeightScale: Interpreting Weight Change in Neural Networks

AM Agrawal, A Tendle, H Sikka, S Singh - arxiv preprint arxiv:2107.07005, 2021 - arxiv.org
Interpreting the learning dynamics of neural networks can provide useful insights into how
networks learn and the development of better training and design approaches. We present …

Scalable syntactic inductive biases for neural language models

AS Kuncoro - 2022 - ora.ox.ac.uk
Natural language has a sequential surface form, although its underlying structure has been
argued to be hierarchical and tree-structured in nature, whereby smaller linguistic units like …