All Updates

Page 898 of 906

February 12, 2026

๐Ÿ“„
ArXiv AIโ€ข79d ago

MeCSAFNet Boosts Multispectral Segmentation

MeCSAFNet uses dual ConvNeXt encoders for visible and non-visible channels in multispectral land cover segmentation. It employs smooth attentional feature fusion with CBAM and ASAU activation. Outperforms baselines like U-Net and SegFormer by up to 19% mIoU on FBP and Potsdam datasets.

#research#mecsafnet#base-large
๐Ÿ“„
ArXiv AIโ€ข79d ago

LRMs Fail to Transfer Reasoning to ToM

Study compares reasoning vs non-reasoning LLMs on ToM benchmarks, finding no consistent gains and sometimes worse performance. Insights reveal slow thinking collapse, need for adaptive reasoning, and option-matching shortcuts. Interventions like S2F and T2M mitigate issues.

#research#tom-study#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LOREN: Low-Rank Adaptation for Neural Receivers

LOREN introduces low-rank adapters to enable code-rate adaptation in neural receivers without storing separate weights. It freezes a shared base network and trains lightweight adapters per code rate. Achieves comparable performance with major hardware savings.

#research#loren#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LoRA Enables Modular Chemistry Prediction

Evaluates LoRA for parameter-efficient fine-tuning of LLMs on organic reaction datasets like USPTO and C-H functionalisation. Matches full fine-tuning accuracy while preserving multi-task performance and mitigating forgetting. Reveals distinct reactivity patterns for better adaptation.

#research#lora#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

Locomo-Plus Tests LLM Cognitive Memory

Locomo-Plus benchmarks cognitive memory in LLM agents under cue-trigger disconnects, focusing on latent conversational constraints. It proposes constraint consistency evaluation over string-matching. Reveals gaps in existing memory systems.

#research#locomo-plus#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Tackle Agent-Based Model Replication

Study evaluates 17 LLMs on ODD-to-Python code generation for predator-prey model. Assesses executability, fidelity, efficiency via NetLogo baseline. GPT-4.1 excels, but reliability varies.

#research#llms#gpt-4-1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Predict Stroke Outcomes from Notes

Fine-tuned LLMs like Llama predict mRS scores from admission notes alone. Achieves 33.9% exact 90-day accuracy and 76.3% binary, matching structured baselines. Enables seamless clinical integration without data extraction.

#research#stroke-llm#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Outstrategize Humans in Games

Uses AlphaEvolve to discover interpretable models of human and LLM strategic behavior from data. Analysis on iterated rock-paper-scissors shows frontier LLMs capable of deeper strategy than humans. Provides foundation for understanding behavioral differences in interactions.

#research#alphacevolve#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Generate Planning Abstractions

Prompts pretrained LLMs to create QNP abstractions for generalized planning from domains and tasks. Automated debugging detects/fixes errors iteratively. Guided LLMs produce useful abstractions for qualitative numerical planning.

#research#qnp-generator#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Fail Cultural Recipes

LLMs generate culturally unrepresentative recipe adaptations unlike humans. Outputs ignore cultural distance correlations from GlobalFusion dataset. Issues stem from weak cultural representations and novelty inflation.

#research#recipe-study#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLMs Accelerate Systematic Mapping

Experience report on using LLMs for systematic mapping studies. Highlights time savings in screening and extraction but notes challenges like hallucinations and prompt engineering. Offers lessons and recommendations for adoption.

#research#llm-mapping#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLM Evolutionary Sampling Speeds Databases

DBPlanBench exposes physical query plans for LLM-proposed localized edits, refined via evolutionary search. LLMs leverage semantic knowledge for optimizations like join orderings. Achieves up to 4.78x speedups, with transfers from small to large databases.

#research#dbplanbench#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LLM Agents Auto-Optimize RecSys Models

A self-evolving system uses Google's Gemini LLMs to autonomously generate, train, and deploy recommendation model improvements. It features an Offline Agent for hypothesis generation and an Online Agent for production validation. Deployed successfully at YouTube, surpassing manual workflows.

#research#youtube#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LITT: Timing Transformer for EHR Events

LITT introduces a Timing-Transformer architecture that aligns sequential events on a virtual relative timeline for event-timing-focused attention. It enables personalized clinical trajectory interpretations. Validated on EHR data from 3,276 breast cancer patients to predict cardiotoxicity onset.

#research#litt#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

Latent Flows Model Reaction Trajectories

LatentRxnFlow predicts reactions as continuous latent trajectories via Conditional Flow Matching from reactant-product pairs. Offers SOTA USPTO accuracy with trajectory diagnostics and uncertainty estimation. Enables error mitigation and reliable predictions.

#research#latentrxnflow#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

Large-Scale AI Social Simulation Launched

AIvilization v0 deploys a resource-constrained artificial society with unified LLM agents. Features hierarchical planning, adaptive profiles, and human steering for long-horizon autonomy. Reproduces real market stylized facts like wealth stratification.

#research#aivilization#v0
๐Ÿ“„
ArXiv AIโ€ข79d ago

LAP Achieves Zero-Shot Robot Embodiment Transfer

Language-Action Pre-training (LAP) represents robot actions in natural language for zero-shot transfer across embodiments without fine-tuning. LAP-3B, a 3B VLA, delivers over 50% success on novel robots and tasks. Enables efficient adaptation and unifies action prediction with VQA.

#research#lap-3b#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

LakeMLB Benchmarks ML in Data Lakes

LakeMLB is a benchmark for machine learning in data lakes, focusing on multi-table union and join scenarios with real datasets from government, finance, and more. Supports pre-training, augmentation strategies. Evaluates tabular ML methods and releases datasets/code.

#research#lakemlb#data-lakes
๐Ÿ“„
ArXiv AIโ€ข79d ago

KSTER Attacks Reverse LLM Model Edits

KSTER exploits low-rank updates in locate-then-edit methods to recover edited data via spectral keyspace reconstruction and entropy prompt recovery. Achieves high success on multiple LLMs. Defense subspace camouflage uses decoys to hide fingerprints.

#security#kster#v1
๐Ÿ“„
ArXiv AIโ€ข79d ago

KPO Stabilizes LLM Policy Optimization

Online Causal Kalman Filtering models IS ratios as evolving latent states for stable RL in LLMs. Smooths noise while preserving token structure. Superior on math reasoning datasets.

#research#kpo#v1
Page 898 of 906