YouZum

News

News

Distilling the Essence: Efficient Reasoning Distillation via Sequence Truncation

arXiv:2512.21002v2 Announce Type: replace Abstract: Distilling the capabilities from a large reasoning model (LRM) to...

Distilling Multilingual Vision-Language Models: When Smaller Models Stay Multilingual

arXiv:2510.26271v1 Announce Type: new Abstract: Vision-language models (VLMs) exhibit uneven performance across languages, a problem...

Disco-RAG: Discourse-Aware Retrieval-Augmented Generation

arXiv:2601.04377v2 Announce Type: replace Abstract: Retrieval-Augmented Generation (RAG) has emerged as an important means of...

Digging for clues about the North Pole’s past

In the past, even with an icebreaker and during peak melt season, getting to the...

DiFlow-TTS: Discrete Flow Matching with Factorized Speech Tokens for Low-Latency Zero-Shot Text-To-Speech

arXiv:2509.09631v1 Announce Type: cross Abstract: Zero-shot Text-to-Speech (TTS) aims to synthesize high-quality speech that mimics...

Diffusion Language Models Know the Answer Before Decoding

arXiv:2508.19982v1 Announce Type: new Abstract: Diffusion language models (DLMs) have recently emerged as an alternative...

Diffuse Thinking: Exploring Diffusion Language Models as Efficient Thought Proposers for Reasoning

arXiv:2510.27469v1 Announce Type: new Abstract: In recent years, large language models (LLMs) have witnessed remarkable...

Did You Forget What I Asked? Prospective Memory Failures in Large Language Models

arXiv:2603.23530v1 Announce Type: new Abstract: Large language models often fail to satisfy formatting instructions when...

Did I Faithfully Say What I Thought? Bridging the Gap Between Neural Activity and Self-Explanations in Large Language Models

arXiv:2506.09277v2 Announce Type: replace Abstract: Large Language Models (LLM) have demonstrated the capability of generating...

DICE: Structured Reasoning in LLMs through SLM-Guided Chain-of-Thought Correction

arXiv:2510.09211v2 Announce Type: replace Abstract: When performing reasoning tasks with user-specific requirements, such as strict...

DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

arXiv:2603.03321v1 Announce Type: new Abstract: Evaluating instruction following in Large Language Models requires decomposing instructions...

DiaBlo: Diagonal Blocks Are Sufficient For Finetuning

arXiv:2506.03230v1 Announce Type: cross Abstract: Finetuning is a critical step for adapting large language models...

We use cookies to improve your experience and performance on our website. You can learn more at Privacy Policy and manage your privacy settings by clicking Settings.

Privacy Preferences

You can choose your cookie settings by turning on/off each type of cookie as you wish, except for essential cookies.

Allow All
Manage Consent Preferences
  • Always Active

Save
en_US