Research Daily: Top AI papers of the day

Get these updates on email!

All of our top picks

Controlled Low-Rank Adaptation with Subspace Regularization for Continued Training on Large Language Models

October 23, 2024

LLM Training, Catastrophic Forgetting, CLoRA, Model Efficiency

Addresses catastrophic forgetting in LLMs. CLoRA improves performance on new and old tasks, impacting LLM training.

Can Large Language Models Invent Algorithms to Improve Themselves?

October 23, 2024

Self-Improving LLMs, Autonomous AI, Algorithm Invention, AI Development

Groundbreaking self-developing framework lets LLMs create self-improvement algorithms, showcasing potential for autonomous AI.

LLMScan: Causal Scan for LLM Misbehavior Detection

October 23, 2024

LLM Safety, Causal Analysis, Misbehavior Detection, AI Safety

Novel causal analysis for LLM safety, detecting various misbehaviors. Comprehensive solution to LLM safety concerns.

Steering Large Language Models using Conceptors: Improving Addition-Based Activation Engineering

October 23, 2024

LLM Control, Conceptors, Activation Engineering, AI Safety

Introduces 'conceptors' for precise LLM output control, advancing LLM control and safety. High impact.

Verification of Neural Control Barrier Functions with Symbolic Derivative Bounds Propagation

October 23, 2024

Neural CBFs, Safety Verification, Robotics, Control Systems

Novel verification framework for neural CBFs enhancing safety and efficiency in robotics and AI safety. High impact.

Deep Domain Isolation and Sample Clustered Federated Learning for Semantic Segmentation

October 22, 2024

Federated Learning, Semantic Segmentation, Domain Isolation, Covariate Shift

Introduces Deep Domain Isolation and Sample Clustered Federated Learning for improved performance in Non-IID settings.

QuAILoRA: Quantization-Aware Initialization for LoRA

October 22, 2024

Quantization, LoRA, LLMs, Memory Efficiency

Introduces a quantization-aware initialization for LoRA to mitigate quantization errors and improve model performance.

RepoGraph: Enhancing AI Software Engineering with Repository-level Code Graph

October 22, 2024

AI Software Engineering, Code Graph, Repository-level, LLMs

Enhances AI software engineering by introducing RepoGraph, a plug-in module for repository-level code understanding.

Rethinking VLMs and LLMs for Image Classification

October 22, 2024

Visual Language Models, LLMs, Image Classification, Model Routing

Challenges the conventional wisdom on VLMs and LLMs in image classification, proposing a lightweight fix for improved efficiency.

Transformers are Efficient Compilers, Provably

October 22, 2024

Transformers, Compilers, Expressive Power, Theoretical Computer Science

Provides a formal investigation of transformers as compilers, showing logarithmic parameter scaling with input length.

Articulate-Anything: Automatic Modeling of Articulated Objects via a Vision-Language Foundation Model

October 21, 2024

3D Modeling, Vision-Language Models, Articulated Objects, Digital Twins

Automates creation of interactive 3D objects; uses vision-language models; state-of-the-art performance.

Boosting LLM Translation Skills without General Ability Loss via Rationale Distillation

October 21, 2024

Large Language Models, Machine Translation, Rationale Distillation, Catastrophic Forgetting

Novel approach to improve translation without losing general abilities; uses rationales; enhances translation performance.

COOL: Efficient and Reliable Chain-Oriented Objective Logic with Neural Networks Feedback Control for Program Synthesis

October 21, 2024

Program Synthesis, Neural Networks, Feedback Control, Domain-Specific Language

Novel program synthesis approach; improves efficiency and reliability; addresses limitations of existing methods.

Automatically Interpreting Millions of Features in Large Language Models

October 21, 2024

Large Language Models, Interpretability, Sparse Autoencoders, Natural Language Explanations

Automated pipeline for interpreting LLM features; uses sparse autoencoders and LLMs; novel explanation scoring techniques.

S4ST: A Strong, Self-transferable, faSt, and Simple Scale Transformation for Transferable Targeted Attack

October 21, 2024

Adversarial Attacks, Deep Neural Networks, Transferable Attacks, Image Transformations

Highly efficient transferable targeted attacks; addresses gradient vanishing; state-of-the-art performance.