ATACompressor: Adaptive Task-Aware Compression for Efficient Long-Context Processing in LLMs
arXiv:2602.03226v1 Announce Type: cross Abstract: Long-context inputs in large language models (LLMs) often suffer from the “lost in the middle”…
arXiv:2602.03226v1 Announce Type: cross Abstract: Long-context inputs in large language models (LLMs) often suffer from the “lost in the middle”…
arXiv:2602.02230v2 Announce Type: replace-cross Abstract: Telemetry streams from large-scale Internet-connected systems (e.g., IoT deployments and online platforms) naturally form an…
arXiv:2602.00053v1 Announce Type: new Abstract: Efficient and scalable deployment of machine learning (ML) models is a prerequisite for modern production…
arXiv:2601.23261v2 Announce Type: replace-cross Abstract: The Muon optimizer has demonstrated strong empirical performance in pre-training large language models by performing…
arXiv:2602.01567v1 Announce Type: cross Abstract: Social media engagement prediction is a central challenge in computational social science, particularly for understanding…
arXiv:2602.01561v1 Announce Type: cross Abstract: Commonsense reasoning in multimodal contexts remains a foundational challenge in artificial intelligence. We introduce Multimodal…
arXiv:2601.23039v2 Announce Type: replace-cross Abstract: Differentiable matching layers, often implemented via entropy-regularized Optimal Transport, serve as a critical approximate inference…
arXiv:2601.22269v1 Announce Type: new Abstract: Judge agents are fundamental to agentic AI frameworks: they provide automated evaluation, and enable iterative…
arXiv:2601.21969v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) often hallucinate, generating content inconsistent with the input. Retrieval-Augmented Generation (RAG)…
arXiv:2601.23059v1 Announce Type: cross Abstract: Large Language Models (LLMs) are increasingly relevant in Software Engineering research and practice, with Automated…