TLDR: This paper introduces an innovative Electronic Health Record (EHR) foundation model that integrates Polygenic Risk Scores (PRS) as a foundational data modality, moving beyond traditional EHR-only approaches to build more holistic health...
TLDR: RoGBot is a novel bot detection framework that overcomes the limitation of relying on follower-following data. It combines semantic features from tweets (using BERT) with user metadata to construct a user-user interaction graph based on behavioral and...
TLDR: LRQK (Low Rank Query and Key attention) is a novel two-stage framework designed to make Large Language Models (LLMs) more memory-efficient for long-context inference. It achieves this by jointly decomposing query and key matrices into compact low-rank...
TLDR: A new research paper introduces two zero-shot, inference-time debiasing methods, Static and Dynamic, that intervene at the final logits layer of Large Language Models (LLMs). Unlike unstable hidden-layer interventions that cause generative collapse, these methods leverage Logit...
TLDR: CLP (Continuous Layer Pruning) is a new framework for compressing large language models by automatically identifying and removing contiguous blocks of layers. It uses a differentiable concave gating algorithm for precise pruning and a "cutoff endpoint tuning"...
TLDR: A new framework called Spatiotemporally Autocorrelated Error Adjustment (SAEA) improves traffic forecasting by explicitly modeling and adjusting for prediction errors that are correlated across time and space. Unlike traditional methods assuming random errors, SAEA uses a vector...
TLDR: A new research paper introduces ELBO-KTO, a novel framework for aligning Diffusion Language Models (dLLMs) with human preferences using unpaired feedback. Unlike traditional methods that require costly paired data and struggle with dLLMs' intractable log-likelihoods, ELBO-KTO combines...
TLDR: Optimize Any Topology (OAT) is a new foundation model for structural topology optimization that predicts minimum-compliance layouts for designs with arbitrary aspect ratios, resolutions, volume fractions, loads, and fixtures. Trained on the new 2.2 million-sample OpenTO dataset,...
TLDR: A new hybrid model for traffic flow forecasting combines Seasonal-Trend decomposition using Loess (STL) with LSTM, ARIMA, and XGBoost. STL breaks traffic data into trend, seasonal, and residual components, which are then predicted by LSTM, ARIMA, and...
TLDR: A research paper by Agarwala, Agarwal, and Rana investigates the real-world drivers of generative AI adoption by analyzing millions of Claude AI interactions mapped to O*NET tasks. The study found that AI usage is highly concentrated, with...
TLDR: Mixture of Experts (MoE) models, crucial for large language models, are often poorly understood mechanistically. This research introduces "network sparsity" as a key characteristic, showing that MoEs exhibit greater "monosemanticity" (less feature overlap) compared to dense networks,...
TLDR: The research paper "MCPGuard: Automatically Detecting Vulnerabilities in MCP Servers" systematically analyzes the security landscape of Model Context Protocol (MCP) based systems, which enable Large Language Models (LLMs) to interact with external tools. It identifies three primary...
TLDR: RefleXGen is a novel method that significantly enhances the security of code generated by large language models (LLMs). It integrates Retrieval-Augmented Generation (RAG) with guided self-reflection, enabling LLMs to iteratively assess and optimize code for security without...