YouZum

Noticias

Noticias

MSCoRe: A Benchmark for Multi-Stage Collaborative Reasoning in LLM Agents

arXiv:2509.17628v1 Announce Type: new Abstract: Large Language Models (LLMs) have excelled in question-answering (QA) tasks...

MOSAIC: Multi-agent Orchestration for Task-Intelligent Scientific Coding

arXiv:2510.08804v1 Announce Type: new Abstract: We present MOSAIC, a multi-agent Large Language Model (LLM) framework...

MoonshotAI Released Checkpoint-Engine: A Simple Middleware to Update Model Weights in LLM Inference Engines, Effective for Reinforcement Learning

MoonshotAI has open-sourced checkpoint-engine, a lightweight middleware aimed at solving one of the key bottlenecks...

Moonshot AI’s Kimi K2 outperforms GPT-4 in key benchmarks — and it’s free

Chinese AI startup Moonshot releases open-source Kimi K2 model that outperforms OpenAI and Anthropic on...

Moonshot AI Releases Kimi K2: A Trillion-Parameter MoE Model Focused on Long Context, Code, Reasoning, and Agentic Behavior

Kimi K2, launched by Moonshot AI in July 2025, is a purpose-built, open-source Mixture-of-Experts (MoE)...

MoE Architecture Comparison: Qwen3 30B-A3B vs. GPT-OSS 20B

This article provides a technical comparison between two recently released Mixture-of-Experts (MoE) transformer models: Alibaba’s...

Model Interpretability and Rationale Extraction by Input Mask Optimization

arXiv:2508.11388v1 Announce Type: new Abstract: Concurrent to the rapid progress in the development of neural-network...

Model Context Protocol: A promising AI integration layer, but not a standard (yet)

Enterprises should experiment with MCP where it adds value, isolate dependencies and prepare for a...

MIT’s LEGO: A Compiler for AI Chips that Auto-Generates Fast, Efficient Spatial Accelerators

Table of contents Hardware Generation without Templates Input IR: Affine, Relation-Centric Semantics (Deconstruct) Front End:...

MIT Researchers Enhanced Artificial Intelligence (AI) 64x Better at Planning, Achieving 94% Accuracy

Can a 8B-parameter language model produce provably valid multi-step plans instead of plausible guesses? MIT...

MIT Researchers Develop Methods to Control Transformer Sensitivity with Provable Lipschitz Bounds and Muon

Training large-scale transformers stably has been a longstanding challenge in deep learning, particularly as models...

We use cookies to improve your experience and performance on our website. You can learn more at Política de privacidad and manage your privacy settings by clicking Settings.

Privacy Preferences

You can choose your cookie settings by turning on/off each type of cookie as you wish, except for essential cookies.

Allow All
Manage Consent Preferences
  • Always Active

Save
es_ES