LLMs


Large language models (LLMs) are computational models notable for their ability to achieve general-purpose language generation and other natural language processing tasks such as classification. Based on language models, LLMs acquire these abilities by learning statistical relationships from vast amounts of text during a computationally intensive self-supervised and semi-supervised training process.

Weight Tying Biases Token Embeddings Towards the Output Space

Add code
Mar 27, 2026
Viaarxiv icon

Sustainability Is Not Linear: Quantifying Performance, Energy, and Privacy Trade-offs in On-Device Intelligence

Add code
Mar 27, 2026
Viaarxiv icon

AMALIA Technical Report: A Fully Open Source Large Language Model for European Portuguese

Add code
Mar 27, 2026
Viaarxiv icon

Automating Clinical Information Retrieval from Finnish Electronic Health Records Using Large Language Models

Add code
Mar 27, 2026
Viaarxiv icon

A Formal Framework for Uncertainty Analysis of Text Generation with Large Language Models

Add code
Mar 27, 2026
Viaarxiv icon

Automating Domain-Driven Design: Experience with a Prompting Framework

Add code
Mar 27, 2026
Viaarxiv icon

A Universal Vibe? Finding and Controlling Language-Agnostic Informal Register with SAEs

Add code
Mar 27, 2026
Viaarxiv icon

ClinicalAgents: Multi-Agent Orchestration for Clinical Decision Making with Dual-Memory

Add code
Mar 27, 2026
Viaarxiv icon

DataFlex: A Unified Framework for Data-Centric Dynamic Training of Large Language Models

Add code
Mar 27, 2026
Viaarxiv icon

Clash of the models: Comparing performance of BERT-based variants for generic news frame detection

Add code
Mar 27, 2026
Viaarxiv icon