YouZum

News

News

FinCoT: Grounding Chain-of-Thought in Expert Financial Reasoning

arXiv:2506.16123v2 Announce Type: replace Abstract: This paper presents FinCoT, a structured chain-of-thought (CoT) prompting framework...

FinBERT-QA: Financial Question Answering with pre-trained BERT Language Models

arXiv:2505.00725v1 Announce Type: new Abstract: Motivated by the emerging demand in the financial industry for...

FHSTP@EXIST 2025 Benchmark: Sexism Detection with Transparent Speech Concept Bottleneck Models

arXiv:2507.20924v1 Announce Type: new Abstract: Sexism has become widespread on social media and in online...

FELLE: Autoregressive Speech Synthesis with Token-Wise Coarse-to-Fine Flow Matching

arXiv:2502.11128v2 Announce Type: replace Abstract: To advance continuous-valued token modeling and temporal-coherence enforcement, we propose...

FEEDER: A Pre-Selection Framework for Efficient Demonstration Selection in LLMs

LLMs have demonstrated exceptional performance across multiple tasks by utilizing few-shot inference, also known as...

Feedback Indicators: The Alignment between Llama and a Teacher in Language Learning

arXiv:2508.11364v1 Announce Type: new Abstract: Automated feedback generation has the potential to enhance students’ learning...

Faster and Better LLMs via Latency-Aware Test-Time Scaling

arXiv:2505.19634v4 Announce Type: replace Abstract: Test-Time Scaling (TTS) has proven effective in improving the performance...

Fast, Slow, and Tool-augmented Thinking for LLMs: A Review

arXiv:2508.12265v1 Announce Type: new Abstract: Large Language Models (LLMs) have demonstrated remarkable progress in reasoning...

FalseReject: A Resource for Improving Contextual Safety and Mitigating Over-Refusals in LLMs via Structured Reasoning

arXiv:2505.08054v1 Announce Type: new Abstract: Safety alignment approaches in large language models (LLMs) often lead...

False Sense of Security: Why Probing-based Malicious Input Detection Fails to Generalize

arXiv:2509.03888v1 Announce Type: new Abstract: Large Language Models (LLMs) can comply with harmful instructions, raising...

Fair-GPTQ: Bias-Aware Quantization for Large Language Models

arXiv:2509.15206v1 Announce Type: new Abstract: High memory demands of generative language models have drawn attention...

Failure by Interference: Language Models Make Balanced Parentheses Errors When Faulty Mechanisms Overshadow Sound Ones

arXiv:2507.00322v1 Announce Type: new Abstract: Despite remarkable advances in coding capabilities, language models (LMs) still...

We use cookies to improve your experience and performance on our website. You can learn more at Privacy Policy and manage your privacy settings by clicking Settings.

Privacy Preferences

You can choose your cookie settings by turning on/off each type of cookie as you wish, except for essential cookies.

Allow All
Manage Consent Preferences
  • Always Active

Save
en_US