YouZum

ニュース

ニュース

Learning to Interpret Weight Differences in Language Models

arXiv:2510.05092v3 Announce Type: replace-cross Abstract: Finetuning (pretrained) language models is a standard approach for updating...

Learning to Align, Aligning to Learn: A Unified Approach for Self-Optimized Alignment

arXiv:2508.07750v1 Announce Type: cross Abstract: Alignment methodologies have emerged as a critical pathway for enhancing...

Learning the Wrong Lessons: Syntactic-Domain Spurious Correlations in Language Models

arXiv:2509.21155v2 Announce Type: replace Abstract: For an LLM to correctly respond to an instruction it...

Learning Human-Perceived Fakeness in AI-Generated Videos via Multimodal LLMs

arXiv:2509.22646v1 Announce Type: cross Abstract: Can humans identify AI-generated (fake) videos and provide grounded reasons?...

Learn and Unlearn: Addressing Misinformation in Multilingual LLMs

arXiv:2406.13748v3 Announce Type: replace Abstract: This paper investigates the propagation of harmful information in multilingual...

LayerNorm and RMS Norm in Transformer Models

This post is divided into five parts; they are: • Why Normalization is Needed in...

Layered Insights: Generalizable Analysis of Authorial Style by Leveraging All Transformer Layers

arXiv:2503.00958v2 Announce Type: replace Abstract: We propose a new approach for the authorship attribution task...

Layer-Wise Perturbations via Sparse Autoencoders for Adversarial Text Generation

arXiv:2508.10404v1 Announce Type: new Abstract: With the rapid proliferation of Natural Language Processing (NLP), especially...

Layer-wise Minimal Pair Probing Reveals Contextual Grammatical-Conceptual Hierarchy in Speech Representations

arXiv:2509.15655v1 Announce Type: new Abstract: Transformer-based speech language models (SLMs) have significantly improved neural speech...

Latent Visual Reasoning

arXiv:2509.24251v1 Announce Type: cross Abstract: Multimodal Large Language Models (MLLMs) have achieved notable gains in...

Latent Fusion Jailbreak: Blending Harmful and Harmless Representations to Elicit Unsafe LLM Outputs

arXiv:2508.10029v1 Announce Type: new Abstract: Large language models (LLMs) demonstrate impressive capabilities in various language...

Large Language Models Show Signs of Alignment with Human Neurocognition During Abstract Reasoning

arXiv:2508.10057v1 Announce Type: cross Abstract: This study investigates whether large language models (LLMs) mirror human...

We use cookies to improve your experience and performance on our website. You can learn more at プライバシーポリシー and manage your privacy settings by clicking Settings.

Privacy Preferences

You can choose your cookie settings by turning on/off each type of cookie as you wish, except for essential cookies.

Allow All
Manage Consent Preferences
  • Always Active

Save
ja