3

Breaking the Bottleneck with DiffuApriel: High-Throughput Diffusion LMs with Mamba Backbone
Diffusion-based language models have recently emerged as a promising alternative to autoregressive generation, yet their reliance on …
Automatic Data Augmentation Learning using Bilevel Optimization for Histopathological Images
One of the main challenges faced when training a deep learning based model to classify histopathological images is the color and shape …
FM2DS: Few-Shot Multimodal Multihop Data Synthesis with Knowledge Distillation for Question Answering
Multimodal multihop question answering is a complex task that requires reasoning over multiple sources of information, such as images …
InCoRo: In-Context Learning for Robotics Control with Feedback Loops
One of the challenges in robotics is to enable robotic units with the reasoning capability that would be robust enough to execute …
Language Decision Transformers with Exponential Tilt for Interactive Text Environments
Text-based game environments are challenging because agents must deal with long sequences of text, execute compositional actions using …
Layer-Wise Quantization: A Pragmatic and Effective Method for Quantizing LLMs Beyond Integer Bit-Levels
We present a simple meta quantization approach that quantizes different layers of a large language model (LLM) at different bit levels, …
Layered gradient accumulation and modular pipeline parallelism: fast and efficient training of large language models
The advent of the transformer has sparked a quick growth in the size of language models, far outpacing hardware improvements. (Dense) …
LitLLM: A Toolkit for Scientific Literature Review
Literature reviews are an essential component of scientific research. We explore the zero-shot abilities of recent large language …
M-RewardBench: Evaluating Reward Models in Multilingual Settings
Reward models (RMs) have driven the state-of-the-art performance of LLMs today by enabling the integration of human feedback into the …