#大语言模型#ToRA is a series of Tool-integrated Reasoning LLM Agents designed to solve challenging mathematical reasoning problems by interacting with tools [ICLR'24].
#计算机科学#Resources of deep learning for mathematical reasoning (DL4MATH).
#自然语言处理#[NeurIPS 2024] Code for the paper "Diffusion of Thoughts: Chain-of-Thought Reasoning in Diffusion Language Models"
🚀ReVisual-R1 is a 7B open-source multimodal language model that follows a three-stage curriculum—cold-start pre-training, multimodal reinforcement learning, and text-only reinforcement learning—to ac...
#大语言模型#Small and Efficient Mathematical Reasoning LLMs
[ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"
#大语言模型#[EMNLP '23] Discriminator-Guided Chain-of-Thought Reasoning
[NeurIPS 2024] Code and Data Repo for Paper "Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning"
Evaluation Code Repo for Paper "PolyMath: Evaluating Mathematical Reasoning in Multilingual Contexts"
#大语言模型#[ACL 2024 Findings] The official repo for "ConceptMath: A Bilingual Concept-wise Benchmark for Measuring Mathematical Reasoning of Large Language Models".
The lecture notes for my discrete mathematics classes.
#大语言模型#🧠 Train your own DeepSeek-R1 style reasoning model on Mac! First MLX implementation of GRPO - the breakthrough technique behind R1's o1-matching performance. Build mathematical reasoning AI without e...
#计算机科学#Beyond Accuracy: Dissecting Mathematical Reasoning for LLMs Under Reinforcement Learning
MathPrompter Implementation: This repository hosts an implementation based on the 'MathPrompter: Mathematical Reasoning Using Large Language Models' paper by Microsoft Research. The code replicates th...
[AAAI 2025] Assessing the Creativity of LLMs in Proposing Novel Solutions to Mathematical Problems
#自然语言处理#Fuzzy reasoning of Generalized Quantifiers (EMNLP 2024)
GSM8K-Consistency is a benchmark database for analyzing the consistency of Arithmetic Reasoning on GSM8K.
#大语言模型#MLX-GRPO allows you to train your own DeepSeek-R1 models directly on your Mac. This implementation simplifies the process of building advanced reasoning AI, making it accessible for developers. 🐙🌟
#大语言模型#We implement and benchmark various prompting techniques for LLMs (i.e. PAL, CoT, PoT, etc.) on a specialized math reasoning dataset (on elementary school grade).
#大语言模型#This dataset aggregates carefully selected elementary-level math problems from various existing resources, providing an optimal mix for testing and enhancing math-solving chatbots for young learners.