<img height="1" width="1" style="display:none" src="https://www.facebook.com/tr?id=145304570664993&amp;ev=PageView&amp;noscript=1">

秘色传媒

Flying papers

Jun 10, 2025

May Papers: Parallel scaling, Evolving code, Understanding LLM reasoning

Written By:

Luke Prince, Luka Ribar, Paul Balanca, Luke Hudlass-Galley

We're Hiring

Join us and build the next generation AI stack - including silicon, hardware and software - the worldwide standard for AI compute

Join our team

Hurtling past the NeurIPS submission deadline into the summer months, we switch from huddling around server rooms to keep warm to babysitting experiments whilst basking in the sun. We鈥檝e had a bumper month of papers to sift through and once again we offer summaries of a few of our favourites.

First, proposes a novel method of scaling compute with language models inspired by classifier-free guidance that finetunes a model to run multiple forward passes with different learned vector prefixes. We also looked into , an evolutionary algorithm from Google DeepMind that generates and refine prompts for Gemini that can advance the state-of-the-art in algorithm design.

Since it has been a particularly exciting month for contributions on LLM reasoning, we picked two papers to dive into deeper. In the authors attempt to improve on prior work sampling continuous token embeddings rather than discrete tokens during reasoning phases of text generation. Finally, in they find that even rewarding random answers can improve reasoning ability, potentially forcing us to reconsider how we understand post-training techniques to improve the use of test-time compute.

We hope you enjoy this month鈥檚 papers as much as we did! If you have thoughts or questions, please reach out to us at .

Here鈥檚 our summary of this month鈥檚 chosen papers:

Authors: Mouxiang Chen et al. (Zhejiang University, Qwen Team, Alibaba Group)

The key idea

Researches at Qwen introduce a new dimension of scaling: parallel forward passes. Their method, PARSCALE, runs 饾憙 parallel copies of a model, each with a different learned prefix. They find that running 饾憙 parallel passes is equivalent to scaling the model parameters by 饾憘(log 饾憙).

three-scaling-approaches (1)

Three parallel scaling approaches: Parameter, Inference time, and Parallel computation.

Background

The approach comes from a practical inference bottleneck: for large models, single batch inference can be memory-bound, especially on resource constrained edge devices. Rather than increasing model size or generating more reasoning steps, PARSCALE aims to scale a new axis, parallel computation, to keep model size approximately constant while improving performance.

Inspired by techniques like Classifier-Free Guidance (CFG), PARSCALE hypothesizes:

Scaling parallel computation (while maintaining the nearly constant parameters) enhances the model鈥檚 capability, with similar effects as scaling parameters.

Methodology

PARSCALE executes 饾憙 forward passes in parallel, each conditioned with a unique learned prefix (implemented via prefix tuning). Outputs of the different streams are combined using a learned aggregation MLP.

Unlike inference-time tricks (e.g., beam search or self-consistency), PARSCALE learns the aggregation during training, leading to more effective use of parallel compute. Conceptually this is similar to ensembling, but with almost complete parameter sharing between the members.

Training Strategy

To reduce training costs, they propose a two-stage approach:

  • Stage 1: Standard pre-training (1T tokens)
  • Stage 2: Add PARSCALE (20B tokens, 2% overhead)

Dramatically reduces cost of parallel scaling training (which requires 饾憙 forward passes) only applied to the final 20B tokens, not the full 1T.

Results

parscale-loss-contours (1)

PARSCALE results: Parameter, Inference time, and Parallel computation.

Coding Tasks (Stack-V2-Python)

Chart 1 (1)

General Tasks (Pile)

Chart 2 (1)

For a 1.6B model, scaling to 饾憙=8 parallel streams achieves performance comparable with a 4.4B model on coding tasks. These efficiency gains are most pronounced at small batch sizes (鈮8) where inference is memory-bound. This makes PARSCALE most suitable for edge deployment scenarios.

  • 22x less memory increase compared to parameter scaling.
  • 6x lower latency.
  • 8x increase (linear with 饾憙) KV cache size.

Dynamic Parallel Scaling

PARSCALE remains effective with frozen main parameters for different values of P. This enables dynamic parallel scaling: switching P to dynamically adapt model capabilities during inference.

Takeaways

PARSCALE provides a new axis in which to boost model capability, particuarly in resource constrained single-batch inference. However KV cache grows linearly with the number of parallel streams (饾憙) so effectiveness may diminish beyond 饾憙 = 8 (the largest tested configuration). It is an open question as to whether 饾憘(log 饾憙) scaling holds for 饾憙鈮8.

Full paper:

Authors: Emilien Dupont, et al. (Google DeepMind)

AlphaEvolve, evolves (no pun intended) the seminal method introduced in late 2023. Powered by a frontier model rather than a smaller LLM, it leverages evolutionary algorithms to successively prompt Gemini to finding novel solutions with respect to a multi-objective target. The results are quite compelling, improving upon various mathematical results relating to matrix multiplication and even reducing Google Cloud鈥檚 costs with 0.7%. This is a staggering saving considering the scale of Google.

The key idea

AlphaEvolve鈥檚 workflow begins with an initial program and an automated evaluation function. Marked code blocks are evolved through LLM-proposed diffs, evaluated for quality, and selectively retained in a program database. Prompts are assembled from high-performing ancestors and injected with stochastic formatting, context, and feedback.

The use of powerful Gemini 2.0 models (Flash for speed, Pro for breakthroughs) ensures a mix of exploration and high-quality suggestions. The evolution loop is fully asynchronous and can evaluate expensive programs in parallel. AlphaEvolve can optimize for multiple metrics and adapt across abstraction levels: from evolving functions to full search algorithms.

By exploiting the capacity of LLMs and engineering a feedback loop, it can effectively help Gemini to act as an 鈥渁rbitrary鈥 format optimiser for very vague tasks, as long as the problem can be expressed in some form of syntax paired with an evaluation criteria.

Fig1 (8) (1)

Figure 1. AlphaEvolve feedback loop design.

Their method

AlphaEvolve works by evolving full programs through an LLM-driven, feedback-grounded loop. Candidate solutions are generated as diffs from existing programs, scored via an automated evaluation function, and stored in a database to seed further generations. It supports evolving entire files across multiple languages and can simultaneously optimize for multiple metrics.

To understand what makes AlphaEvolve effective, the authors performed ablation studies on two tasks: matrix multiplication and the kissing number problem. The key findings:

  1. No evolution: Removing the evolutionary loop (i.e., re-prompting the initial program) significantly degrades performance.
  2. No context: Stripping rich context from prompts leads to worse solutions, confirming that prompt engineering matters.
  3. Small LLMs only: Using only lightweight models reduces result quality. Strong base models like Gemini Pro make a difference.
  4. No full-file evolution: Restricting changes to a single function (vs. whole files) limits AlphaEvolve鈥檚 power and flexibility.
  5. No meta-prompt evolution : Removing the co-evolution of prompts results in slower progress, showing prompt quality co-evolution is a key driver.

Together, these ablations show that AlphaEvolve鈥檚 strength comes from multiple interacting components; especially full-code evolution, high-quality LLMs, and contextual prompting.

Fig2 (8) (1)

Figure 2. AlphaEvolve ablation study results.

Results

AlphaEvolve discovered a new algorithm to multiply 4脳4 complex matrices using 48 scalar multiplications鈥攂eating Strassen鈥檚 49 from 1969. It improved tensor decomposition methods for matrix multiplication, set new state-of-the-art results on the Erd艖s minimum overlap and kissing number problems, and evolved efficient scheduling heuristics that saved 0.7% of Google鈥檚 data center compute.

They further demonstrated that AlphaEvolve also able to carry out assembly level code to optimise kernels for Gemini鈥檚 attention layer, yielding a 23% performance improvement, with a 1% decrease in wall time.

Takeaways

Some observations here is that frontier models can evidently be used as 鈥済eneral-purpose鈥 optimisers assuming a well engineered feedback loop. This can likely be generalised to a product using an arbitrary frontier model, and may possibly add another venue for the agentic-LLM community to explore.

Full paper:

Authors: Zhen Zhang et al. (University of California, Purdue University, LMSYS Org, Microsoft)

The key idea

Conventional reasoning models generate long reasoning traces, which are typically constrained to the expressivity of the model鈥檚 vocabulary. The discretisation of reasoning makes it hard to explore multiple paths or ideas, and limits the model鈥檚 ability to 鈥渢hink鈥 about abstract concepts, as it is always forced to express its thoughts in natural language.

Recent works have looked into latent thoughts, such as and , in which the thoughts are not necessarily discretised tokens, but rather continuous tokens in some latent space. However, training these methods is non-trivial and scaling model size can be very challenging.

In Soft Thinking, the authors propose a training-free approach to latent reasoning, in which the 鈥渃oncept tokens鈥 are a probability-weighted mixture of the token embeddings.

soft-thinking-schematic (1)

Figure 2: Soft Thinking replaces discrete tokens with soft, abstract concept tokens, enabling reasoning in continuous concept space.

Their method

Typically, reasoning models employ standard LLM inference techniques for generating their reasoning traces: each forward pass 饾憱 generates a probability distribution over the vocabulary, from which a token 饾憽饾憱 is sampled from. This token is then embedded using the embedding matrix 饾悇
 and injected into the model鈥檚 input.

Mathematically, this can be expressed as

\(e_{i+1} = \mathbf{E}[t_i]\)

such that

\(t_i \sim p_i = \mathrm{LLM}(e_1, \cdots,  e_{i-1})\)

where 饾憹饾憱 is the probability distribution for the 饾憱th forward pass, and LLM is the model.

The sampling operation of LLM inference discretises the model鈥檚 output, limiting its expressivity. In contrast, Soft Thinking proposes taking a probability-weighted mixture of the input token embeddings, making a so-called concept token. This means the next input token can be expressed as

\(e_{i+1} = \sum_{k=1}^{|V|}p_i[k] \cdot E[k]\)

This approach means that the input embedding layer and output head do not need to be weight-tied, which can cause issues for other continuous reasoning approaches such as .

As the model no longer injects conventional tokens into the model as part of its reasoning trace, over time the model will be in an out-of-distribution regime. To mitigate this, the authors suggest a cold stop mechanism, which measures the entropy of the concept token, and if it falls below a threshold 饾湉 for some number of consecutive iterations, then a </think> token is injected into the sequence to terminate the reasoning trace and commence answer generation. This prevents the model from becoming overconfident, and provides a simple stopping condition for the model to exit latent-thought generation.

Results

The authors examine Soft Thinking over a number of mathematical and coding tasks, on three different models: QwQ-32B, DeepSeek-R1-DistillQwen-32B, and DeepSeek-R1-Distill-Llama-70B. They find that across all models and tasks, they see an improvement in task performance, and very often a reduction in sequence length, indicating that Soft Thinking enables richer concepts for a given token.

results-table-1 (1)

Table 1: Comparison of Soft Thinking and various baseline methods on accuracy and generation length across mathematical datasets. Best results are highlighted in bold.

results-table-2 (1)

Table 2: Comparison of Soft Thinking and various baseline methods on accuracy and generation length across coding datasets. Best results are highlighted in bold.

One concern surrounding latent reasoning is difficulty in interpeting the reasoning trace. While questions the validity of traces to the actual reasoning itself, the Soft Thinking authors are still able to generate legible reasoning traces, simply by examining the highest-probability (discrete) token after each forward pass.

probability-distribution (1)

Figure 4: An example illustrating the probability distribution of our proposed Soft Thinking method. At each step, top-饾憳 token candidates and their probabilities are shown. Red boxes indicate the selected tokens that form the final generated sequence for readability and interpretability.

Takeaways

Soft Thinking offers a viable way to imbue pre-trained reasoning models with latent reasoning capabilities which permit abstract concepts in a continuous space, without requiring any additional fine-tuning. As their results demonstrate, this offers the opportunity for greater task performance with shorter sequence lengths. While this work doesn鈥檛 investigate how we can train models to best use the concept space, it does indicate that research in this direction is likely to bear promising results.

Full paper:

Authors: Rulin Shao, Shuyue Stella Li, Rui Xin and Scott Geng (University of Washington, Allen Institute for AI and University of California, Berkeley)

The key idea

The authors (claim to) boost significantly performance via reinforcement learning (RL) with random, or even spurious, rewards on the MATH-500 benchmark. This hints at models鈥 already possessing reasoning abilities, which the RL procedure 鈥渃oaxes out鈥, rather than developing.

figure_1 (4)

Figure 1. MATH-500 accuracy after 150 steps of RLVR on various training signals. We show that even "spurious rewards" (e.g., rewarding *incorrect* labels or with completely random rewards) can yield strong MATH-500 gains on Qwen models. Notably, these reward signals do not work for other models like Llama3 and OLMo2, which have different reasoning priors.

NB. The baseline evaluations in this paper, and related ones, have been . They an under-reporting of ~15 percentage points (49.4% vs 64.3%) here. This is raised in a on the original . Still, some of the claimed gain is not absorbed by correct baseline evaluation.

With all this in mind, the claims in the current paper (outlined below), and from similar papers, should be taken with a pinch of salt.

Background

Reinforcement Learning with Verifiable Rewards (RLVR) has become a standard approach for enhancing reasoning in LLMs. The conventional wisdom that high-quality supervision signals are essential has been challenged recently, from to or all improving maths capabilities.

The current paper asks and addresses the question,

鈥淲hat is a minimum requirement for rewards to provide meaningful RLVR training signals? Do rewards even need to be correctly aligned?鈥

Their method

RLVR experiments with Qwen2.5-Math-7B, the de facto model that has been widely used by the community for reinforcement learning, were performed using different ground-truth supervision across multiple benchmarks.

Somewhat surprisingly, it is possible to almost match the ground-truth with incorrect or random rewards. This finding directly challenges the prevailing understanding of reinforcement learning鈥檚 role in improving AI reasoning capabilities.

Results

The MATH-500 benchmark is the focus. The baseline model was tested (no RLVR), alongside four RLVR approaches (including the genuine truth), with results tabulated below.

Math chart (2)

*鈥淎djusted鈥 here is an estimation of the boost after removing ~15%, as mentioned above (and ).

Whilst none of the other options perform as well as RLVR with the true labels, random and incorrect rewards get pretty close. Interestingly, other models analysed (LLaMA 3.2B Instruct and OLMo 2 7B) did not show such gains.

Takeaways

One interpretation is that the model can already reason (fairly well), and that RLVR is more teasing out this capability, rather than improving it. This agrees with :

鈥淩LVR with spurious rewards can serve as a mechanism to amplify and surface useful reasoning representations learned during pre-training.鈥

NB. Remember, though, that the baseline evaluations have been seriously questioned. This, in turn, reduces the claimed gains, almost wiping out several of them.

Full paper:

Reviews by: , , ,

Discover more on the , and subscribe to the Papers of the Month newsletter.