The landscape and challenges of HPC research and LLMs

L Chen, NK Ahmed, A Dutta, A Bhattacharjee… - arxiv preprint arxiv …, 2024 - arxiv.org
Recently, language models (LMs), especially large language models (LLMs), have
revolutionized the field of deep learning. Both encoder-decoder models and prompt-based …

Ompgpt: A generative pre-trained transformer model for openmp

L Chen, A Bhattacharjee, N Ahmed, N Hasabnis… - … Conference on Parallel …, 2024 - Springer
Large language models (LLMs) such as ChatGPT have significantly advanced the field of
Natural Language Processing (NLP). This trend led to the development of code-based large …

Compile: A large ir dataset from production sources

A Grossman, L Paehler, K Parasyris, T Ben-Nun… - arxiv preprint arxiv …, 2023 - arxiv.org
Code is increasingly becoming a core data modality of modern machine learning research
impacting not only the way we write code with conversational agents like OpenAI's ChatGPT …

Mpirigen: Mpi code generation through domain-specific language models

N Schneider, N Hasabnis, VA Vo, T Kadosh… - Proceedings of the …, 2024 - dl.acm.org
The imperative need to scale computation across numerous nodes highlights the
significance of efficient parallel computing, particularly in the realm of Message Passing …

OMPar: Automatic Parallelization with AI-Driven Source-to-Source Compilation

T Kadosh, N Hasabnis, P Soundararajan, VA Vo… - arxiv preprint arxiv …, 2024 - arxiv.org
Manual parallelization of code remains a significant challenge due to the complexities of
modern software systems and the widespread adoption of multi-core architectures. This …

[PDF][PDF] PcMINER: Mining Performance Related Commits at Scale

MAK Azad, M Alexerder, M Alexender, SSM Tariq… - sc24.supercomputing.org
Performance inefficiencies challenge software development, degrading application
performance and wasting computational resources. Application developers invest …