Results for "subspace focus"
Learning physical parameters from data.
Mechanism that computes context-aware mixtures of representations; scales well and captures long-range dependencies.
Architecture based on self-attention and feedforward layers; foundation of modern LLMs and many multimodal models.
Samples from the k highest-probability tokens to limit unlikely outputs.
Hidden behavior activated by specific triggers, causing targeted mispredictions or undesired outputs.
Reconstructing a model or its capabilities via API queries or leaked artifacts.
Methods to protect model/data during inference (e.g., trusted execution environments) from operators/attackers.
A system that perceives state, selects actions, and pursues goals—often combining LLM reasoning with tools and memory.
A single attention mechanism within multi-head attention.
Techniques to handle longer documents without quadratic cost.
Separates planning from execution in agent architectures.
Models that learn to generate samples resembling training data.
GNN using attention to weight neighbor contributions dynamically.
Decomposing goals into sub-tasks.
Attention between different modalities.
Assigning a role or identity to the model.
Detects trigger phrases in audio streams.
Using markers to isolate context segments.
Distributed agents producing emergent intelligence.