Deepseekmath: Pushing the limits of mathematical reasoning in open language models
Mathematical reasoning poses a significant challenge for language models due to its
complex and structured nature. In this paper, we introduce DeepSeekMath 7B, which …
complex and structured nature. In this paper, we introduce DeepSeekMath 7B, which …
Math-shepherd: Verify and reinforce llms step-by-step without human annotations
In this paper, we present an innovative process-oriented math process reward model called
Math-shepherd, which assigns a reward score to each step of math problem solutions. The …
Math-shepherd, which assigns a reward score to each step of math problem solutions. The …
Metamath: Bootstrap your own mathematical questions for large language models
Large language models (LLMs) have pushed the limits of natural language understanding
and exhibited excellent problem-solving ability. Despite the great success, most existing …
and exhibited excellent problem-solving ability. Despite the great success, most existing …
Statistical rejection sampling improves preference optimization
Improving the alignment of language models with human preferences remains an active
research challenge. Previous approaches have primarily utilized Reinforcement Learning …
research challenge. Previous approaches have primarily utilized Reinforcement Learning …
Mathematical language models: A survey
In recent years, there has been remarkable progress in leveraging Language Models (LMs),
encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models …
encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models …
Open-Ethical AI: Advancements in Open-Source Human-Centric Neural Language Models
This survey summarises the most recent methods for building and assessing helpful, honest,
and harmless neural language models, considering small, medium, and large-size models …
and harmless neural language models, considering small, medium, and large-size models …
Amortizing intractable inference in large language models
Autoregressive large language models (LLMs) compress knowledge from their training data
through next-token conditional distributions. This limits tractable querying of this knowledge …
through next-token conditional distributions. This limits tractable querying of this knowledge …
When Do Program-of-Thought Works for Reasoning?
The reasoning capabilities of large language models (LLMs) play a pivotal role in the realm
of embodied artificial intelligence. Although there are effective methods like program-of …
of embodied artificial intelligence. Although there are effective methods like program-of …
Knowledgeable preference alignment for llms in domain-specific question answering
Deploying large language models (LLMs) to real scenarios for domain-specific question
answering (QA) is a key thrust for LLM applications, which poses numerous challenges …
answering (QA) is a key thrust for LLM applications, which poses numerous challenges …
Mmicl: Empowering vision-language model with multi-modal in-context learning
Starting from the resurgence of deep learning, vision-language models (VLMs) benefiting
from large language models (LLMs) have never been so popular. However, while LLMs can …
from large language models (LLMs) have never been so popular. However, while LLMs can …