arxiv_id
stringlengths 10
10
| published
stringlengths 20
20
| titles
stringlengths 9
243
| authors
sequencelengths 1
389
| abstract
stringlengths 96
3.09k
| categories
sequencelengths 1
10
| selected
bool 2
classes |
---|---|---|---|---|---|---|
2402.11755 | 2024-02-19T00:53:48Z | SPML: A DSL for Defending Language Models Against Prompt Attacks | [
"Reshabh K Sharma",
"Vinayak Gupta",
"Dan Grossman"
] | Large language models (LLMs) have profoundly transformed natural language
applications, with a growing reliance on instruction-based definitions for
designing chatbots. However, post-deployment the chatbot definitions are fixed
and are vulnerable to attacks by malicious users, emphasizing the need to
prevent unethical applications and financial losses. Existing studies explore
user prompts' impact on LLM-based chatbots, yet practical methods to contain
attacks on application-specific chatbots remain unexplored. This paper presents
System Prompt Meta Language (SPML), a domain-specific language for refining
prompts and monitoring the inputs to the LLM-based chatbots. SPML actively
checks attack prompts, ensuring user inputs align with chatbot definitions to
prevent malicious execution on the LLM backbone, optimizing costs. It also
streamlines chatbot definition crafting with programming language capabilities,
overcoming natural language design challenges. Additionally, we introduce a
groundbreaking benchmark with 1.8k system prompts and 20k user inputs, offering
the inaugural language and benchmark for chatbot definition evaluation.
Experiments across datasets demonstrate SPML's proficiency in understanding
attacker prompts, surpassing models like GPT-4, GPT-3.5, and LLAMA. Our data
and codes are publicly available at: https://prompt-compiler.github.io/SPML/. | [
"cs.LG",
"cs.CL",
"cs.CR",
"cs.PL"
] | false |
2402.11764 | 2024-02-19T01:28:48Z | ChatGPT Based Data Augmentation for Improved Parameter-Efficient
Debiasing of LLMs | [
"Pengrui Han",
"Rafal Kocielnik",
"Adhithya Saravanan",
"Roy Jiang",
"Or Sharir",
"Anima Anandkumar"
] | Large Language models (LLMs), while powerful, exhibit harmful social biases.
Debiasing is often challenging due to computational costs, data constraints,
and potential degradation of multi-task language capabilities. This work
introduces a novel approach utilizing ChatGPT to generate synthetic training
data, aiming to enhance the debiasing of LLMs. We propose two strategies:
Targeted Prompting, which provides effective debiasing for known biases but
necessitates prior specification of bias in question; and General Prompting,
which, while slightly less effective, offers debiasing across various
categories. We leverage resource-efficient LLM debiasing using adapter tuning
and compare the effectiveness of our synthetic data to existing debiasing
datasets. Our results reveal that: (1) ChatGPT can efficiently produce
high-quality training data for debiasing other LLMs; (2) data produced via our
approach surpasses existing datasets in debiasing performance while also
preserving internal knowledge of a pre-trained LLM; and (3) synthetic data
exhibits generalizability across categories, effectively mitigating various
biases, including intersectional ones. These findings underscore the potential
of synthetic data in advancing the fairness of LLMs with minimal retraining
cost. | [
"cs.CL",
"cs.AI",
"cs.CY",
"68T50",
"I.2.7; K.4.1"
] | false |
2402.11777 | 2024-02-19T02:08:03Z | Uncovering Latent Human Wellbeing in Language Model Embeddings | [
"Pedro Freire",
"ChengCheng Tan",
"Adam Gleave",
"Dan Hendrycks",
"Scott Emmons"
] | Do language models implicitly learn a concept of human wellbeing? We explore
this through the ETHICS Utilitarianism task, assessing if scaling enhances
pretrained models' representations. Our initial finding reveals that, without
any prompt engineering or finetuning, the leading principal component from
OpenAI's text-embedding-ada-002 achieves 73.9% accuracy. This closely matches
the 74.6% of BERT-large finetuned on the entire ETHICS dataset, suggesting
pretraining conveys some understanding about human wellbeing. Next, we consider
four language model families, observing how Utilitarianism accuracy varies with
increased parameters. We find performance is nondecreasing with increased model
size when using sufficient numbers of principal components. | [
"cs.CL",
"cs.AI",
"cs.LG",
"I.2.7"
] | false |
2402.11804 | 2024-02-19T03:21:19Z | LLM as Prompter: Low-resource Inductive Reasoning on Arbitrary Knowledge
Graphs | [
"Kai Wang",
"Yuwei Xu",
"Zhiyong Wu",
"Siqiang Luo"
] | Knowledge Graph (KG) inductive reasoning, which aims to infer missing facts
from new KGs that are not seen during training, has been widely adopted in
various applications. One critical challenge of KG inductive reasoning is
handling low-resource scenarios with scarcity in both textual and structural
aspects. In this paper, we attempt to address this challenge with Large
Language Models (LLMs). Particularly, we utilize the state-of-the-art LLMs to
generate a graph-structural prompt to enhance the pre-trained Graph Neural
Networks (GNNs), which brings us new methodological insights into the KG
inductive reasoning methods, as well as high generalizability in practice. On
the methodological side, we introduce a novel pretraining and prompting
framework ProLINK, designed for low-resource inductive reasoning across
arbitrary KGs without requiring additional training. On the practical side, we
experimentally evaluate our approach on 36 low-resource KG datasets and find
that ProLINK outperforms previous methods in three-shot, one-shot, and
zero-shot reasoning tasks, exhibiting average performance improvements by 20%,
45%, and 147%, respectively. Furthermore, ProLINK demonstrates strong
robustness for various LLM promptings as well as full-shot scenarios. | [
"cs.AI",
"cs.CL",
"cs.SI"
] | false |
2402.11818 | 2024-02-19T04:17:21Z | Where It Really Matters: Few-Shot Environmental Conservation Media
Monitoring for Low-Resource Languages | [
"Sameer Jain",
"Sedrick Scott Keh",
"Shova Chettri",
"Karun Dewan",
"Pablo Izquierdo",
"Johanna Prussman",
"Pooja Shreshtha",
"Cesar Suarez",
"Zheyuan Ryan Shi",
"Lei Li",
"Fei Fang"
] | Environmental conservation organizations routinely monitor news content on
conservation in protected areas to maintain situational awareness of
developments that can have an environmental impact. Existing automated media
monitoring systems require large amounts of data labeled by domain experts,
which is only feasible at scale for high-resource languages like English.
However, such tools are most needed in the global south where news of interest
is mainly in local low-resource languages, and far fewer experts are available
to annotate datasets sustainably. In this paper, we propose NewsSerow, a method
to automatically recognize environmental conservation content in low-resource
languages. NewsSerow is a pipeline of summarization, in-context few-shot
classification, and self-reflection using large language models (LLMs). Using
at most 10 demonstration example news articles in Nepali, NewsSerow
significantly outperforms other few-shot methods and achieves comparable
performance with models fully fine-tuned using thousands of examples. The World
Wide Fund for Nature (WWF) has deployed NewsSerow for media monitoring in
Nepal, significantly reducing their operational burden, and ensuring that AI
tools for conservation actually reach the communities that need them the most.
NewsSerow has also been deployed for countries with other languages like
Colombia. | [
"cs.CL",
"cs.AI",
"cs.CY"
] | false |
2402.11839 | 2024-02-19T05:06:10Z | An enhanced Teaching-Learning-Based Optimization (TLBO) with Grey Wolf
Optimizer (GWO) for text feature selection and clustering | [
"Mahsa Azarshab",
"Mohammad Fathian",
"Babak Amiri"
] | Text document clustering can play a vital role in organizing and handling the
everincreasing number of text documents. Uninformative and redundant features
included in large text documents reduce the effectiveness of the clustering
algorithm. Feature selection (FS) is a well-known technique for removing these
features. Since FS can be formulated as an optimization problem, various
meta-heuristic algorithms have been employed to solve it.
Teaching-Learning-Based Optimization (TLBO) is a novel meta-heuristic algorithm
that benefits from the low number of parameters and fast convergence. A hybrid
method can simultaneously benefit from the advantages of TLBO and tackle the
possible entrapment in the local optimum. By proposing a hybrid of TLBO, Grey
Wolf Optimizer (GWO), and Genetic Algorithm (GA) operators, this paper suggests
a filter-based FS algorithm (TLBO-GWO). Six benchmark datasets are selected,
and TLBO-GWO is compared with three recently proposed FS algorithms with
similar approaches, the main TLBO and GWO. The comparison is conducted based on
clustering evaluation measures, convergence behavior, and dimension reduction,
and is validated using statistical tests. The results reveal that TLBO-GWO can
significantly enhance the effectiveness of the text clustering technique
(K-means). | [
"cs.NE",
"cs.CL",
"cs.LG"
] | false |
2402.11960 | 2024-02-19T09:04:30Z | DB-LLM: Accurate Dual-Binarization for Efficient LLMs | [
"Hong Chen",
"Chengtao Lv",
"Liang Ding",
"Haotong Qin",
"Xiabin Zhou",
"Yifu Ding",
"Xuebo Liu",
"Min Zhang",
"Jinyang Guo",
"Xianglong Liu",
"Dacheng Tao"
] | Large language models (LLMs) have significantly advanced the field of natural
language processing, while the expensive memory and computation consumption
impede their practical deployment. Quantization emerges as one of the most
effective methods for improving the computational efficiency of LLMs. However,
existing ultra-low-bit quantization always causes severe accuracy drops. In
this paper, we empirically relieve the micro and macro characteristics of
ultra-low bit quantization and present a novel Dual-Binarization method for
LLMs, namely DB-LLM. For the micro-level, we take both the accuracy advantage
of 2-bit-width and the efficiency advantage of binarization into account,
introducing Flexible Dual Binarization (FDB). By splitting 2-bit quantized
weights into two independent sets of binaries, FDB ensures the accuracy of
representations and introduces flexibility, utilizing the efficient bitwise
operations of binarization while retaining the inherent high sparsity of
ultra-low bit quantization. For the macro-level, we find the distortion that
exists in the prediction of LLM after quantization, which is specified as the
deviations related to the ambiguity of samples. We propose the Deviation-Aware
Distillation (DAD) method, enabling the model to focus differently on various
samples. Comprehensive experiments show that our DB-LLM not only significantly
surpasses the current State-of-The-Art (SoTA) in ultra-low bit quantization
(eg, perplexity decreased from 9.64 to 7.23), but also achieves an additional
20\% reduction in computational consumption compared to the SOTA method under
the same bit-width. Our code will be released soon. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2402.11997 | 2024-02-19T09:43:03Z | Remember This Event That Year? Assessing Temporal Information and
Reasoning in Large Language Models | [
"Himanshu Beniwal",
"Kowsik Nandagopan D",
"Mayank Singh"
] | Large Language Models (LLMs) are increasingly becoming ubiquitous, yet their
ability to reason about and retain temporal information remains limited. This
hinders their application in real-world scenarios where understanding the
sequential nature of events is crucial. This paper experiments with
state-of-the-art models on a novel, large-scale temporal dataset,
\textbf{TempUN}, to reveal significant limitations in temporal retention and
reasoning abilities. Interestingly, closed-source models indicate knowledge
gaps more frequently, potentially suggesting a trade-off between uncertainty
awareness and incorrect responses. Further, exploring various fine-tuning
approaches yielded no major performance improvements. The associated dataset
and code are available at the following URL
(https://github.com/lingoiitgn/TempUN). | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12061 | 2024-02-19T11:28:20Z | All Language Models Large and Small | [
"Zhixun Chen",
"Yali Du",
"David Mguni"
] | Many leading language models (LMs) use high-intensity computational resources
both during training and execution. This poses the challenge of lowering
resource costs for deployment and faster execution of decision-making tasks
among others. We introduce a novel plug-and-play LM framework named Language
Optimising Network Distribution (LONDI) framework. LONDI learns to selectively
employ large LMs only where complex decision-making and reasoning are required
while using low-resource LMs everywhere else. LONDI consists of a system of two
(off-)policy networks, an LM, a large LM (LLM), and a reinforcement learning
module that uses switching controls to quickly learn which system states to
call the LLM. We then introduce a variant of LONDI that maintains budget
constraints on LLM calls and hence its resource usage. Theoretically, we prove
LONDI learns the subset of system states to activate the LLM required to solve
the task. We then prove that LONDI converges to optimal solutions while also
preserving budgetary constraints on LLM calls almost surely enabling it to
solve various tasks while significantly lowering computational costs. We test
LONDI's performance in a range of tasks in ScienceWorld and BabyAI-Text and
demonstrate that LONDI can solve tasks only solvable by resource-intensive LLMs
while reducing GPU usage by up to 30%. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2402.12100 | 2024-02-19T12:31:56Z | Groot: Adversarial Testing for Generative Text-to-Image Models with
Tree-based Semantic Transformation | [
"Yi Liu",
"Guowei Yang",
"Gelei Deng",
"Feiyue Chen",
"Yuqi Chen",
"Ling Shi",
"Tianwei Zhang",
"Yang Liu"
] | With the prevalence of text-to-image generative models, their safety becomes
a critical concern. adversarial testing techniques have been developed to probe
whether such models can be prompted to produce Not-Safe-For-Work (NSFW)
content. However, existing solutions face several challenges, including low
success rate and inefficiency. We introduce Groot, the first automated
framework leveraging tree-based semantic transformation for adversarial testing
of text-to-image models. Groot employs semantic decomposition and sensitive
element drowning strategies in conjunction with LLMs to systematically refine
adversarial prompts. Our comprehensive evaluation confirms the efficacy of
Groot, which not only exceeds the performance of current state-of-the-art
approaches but also achieves a remarkable success rate (93.66%) on leading
text-to-image models such as DALL-E 3 and Midjourney. | [
"cs.CL",
"cs.AI",
"cs.CR",
"cs.SE"
] | false |
2402.12146 | 2024-02-19T13:57:55Z | Meta Ranking: Less Capable Language Models are Capable for Single
Response Judgement | [
"Zijun Liu",
"Boqun Kou",
"Peng Li",
"Ming Yan",
"Ji Zhang",
"Fei Huang",
"Yang Liu"
] | Although Large Language Models (LLMs) have demonstrated strong performance on
a wide range of tasks, they still face reliability challenges such as
hallucination. Previous studies reveal that highly capable LLMs like GPT-4 are
effective in judging the reliability of individual responses, while less
capable ones are often tuned to evaluate the relative reliability of responses
to the same query. To enable less capable LLMs to effectively judge the
reliability of individual responses, we propose a novel method named
$\textit{Meta}$ $\textit{Ranking}$ (MR). Unlike previous methods, which assess
the response directly, we achieve the judgement by comparing the target
query-response pair with reference query-response pairs. We found its
remarkable effectiveness in error detection for LLM responses on reasoning
tasks, where less capable LLMs could outperform strong baselines, even without
fine-tuning. We further demonstrate that MR can be used to enhance the
performance of LLMs in two practical applications: query routing and iterative
training data filtering. The former achieves GPT-4-turbo comparable performance
with less than half the token consumption, while the latter makes the
instruction-tuned LLaMA-7B and Phi-2, a 2.7B model, significantly surpass
Alpaca-13B over fewer training samples, underscoring the high potential of our
proposed method. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12189 | 2024-02-19T14:52:50Z | Amplifying Training Data Exposure through Fine-Tuning with
Pseudo-Labeled Memberships | [
"Myung Gyo Oh",
"Hong Eun Ahn",
"Leo Hyun Park",
"Taekyoung Kwon"
] | Neural language models (LMs) are vulnerable to training data extraction
attacks due to data memorization. This paper introduces a novel attack scenario
wherein an attacker adversarially fine-tunes pre-trained LMs to amplify the
exposure of the original training data. This strategy differs from prior
studies by aiming to intensify the LM's retention of its pre-training dataset.
To achieve this, the attacker needs to collect generated texts that are closely
aligned with the pre-training data. However, without knowledge of the actual
dataset, quantifying the amount of pre-training data within generated texts is
challenging. To address this, we propose the use of pseudo-labels for these
generated texts, leveraging membership approximations indicated by
machine-generated probabilities from the target LM. We subsequently fine-tune
the LM to favor generations with higher likelihoods of originating from the
pre-training data, based on their membership probabilities. Our empirical
findings indicate a remarkable outcome: LMs with over 1B parameters exhibit a
four to eight-fold increase in training data exposure. We discuss potential
mitigations and suggest future research directions. | [
"cs.CL",
"cs.CR",
"cs.LG",
"I.2.7; K.6.5"
] | false |
2402.12222 | 2024-02-19T15:30:40Z | CovRL: Fuzzing JavaScript Engines with Coverage-Guided Reinforcement
Learning for LLM-based Mutation | [
"Jueon Eom",
"Seyeon Jeong",
"Taekyoung Kwon"
] | Fuzzing is an effective bug-finding technique but it struggles with complex
systems like JavaScript engines that demand precise grammatical input.
Recently, researchers have adopted language models for context-aware mutation
in fuzzing to address this problem. However, existing techniques are limited in
utilizing coverage guidance for fuzzing, which is rather performed in a
black-box manner. This paper presents a novel technique called CovRL
(Coverage-guided Reinforcement Learning) that combines Large Language Models
(LLMs) with reinforcement learning from coverage feedback. Our fuzzer,
CovRL-Fuzz, integrates coverage feedback directly into the LLM by leveraging
the Term Frequency-Inverse Document Frequency (TF-IDF) method to construct a
weighted coverage map. This map is key in calculating the fuzzing reward, which
is then applied to the LLM-based mutator through reinforcement learning.
CovRL-Fuzz, through this approach, enables the generation of test cases that
are more likely to discover new coverage areas, thus improving vulnerability
detection while minimizing syntax and semantic errors, all without needing
extra post-processing. Our evaluation results indicate that CovRL-Fuzz
outperforms the state-of-the-art fuzzers in terms of code coverage and
bug-finding capabilities: CovRL-Fuzz identified 48 real-world security-related
bugs in the latest JavaScript engines, including 39 previously unknown
vulnerabilities and 11 CVEs. | [
"cs.CR",
"cs.CL",
"cs.LG",
"cs.SE",
"D.4.6; I.2.5; D.2.4"
] | false |
2402.12264 | 2024-02-19T16:26:00Z | Uncertainty quantification in fine-tuned LLMs using LoRA ensembles | [
"Oleksandr Balabanov",
"Hampus Linander"
] | Fine-tuning large language models can improve task specific performance,
although a general understanding of what the fine-tuned model has learned,
forgotten and how to trust its predictions is still missing. We derive
principled uncertainty quantification for fine-tuned LLMs with posterior
approximations using computationally efficient low-rank adaptation ensembles.
We analyze three common multiple-choice datasets using low-rank adaptation
ensembles based on Mistral-7b, and draw quantitative and qualitative
conclusions on their perceived complexity and model efficacy on the different
target domains during and after fine-tuning. In particular, backed by the
numerical experiments, we hypothesise about signals from entropic uncertainty
measures for data domains that are inherently difficult for a given
architecture to learn. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] | false |
2402.12329 | 2024-02-19T18:01:36Z | Query-Based Adversarial Prompt Generation | [
"Jonathan Hayase",
"Ema Borevkovic",
"Nicholas Carlini",
"Florian Tramèr",
"Milad Nasr"
] | Recent work has shown it is possible to construct adversarial examples that
cause an aligned language model to emit harmful strings or perform harmful
behavior. Existing attacks work either in the white-box setting (with full
access to the model weights), or through transferability: the phenomenon that
adversarial examples crafted on one model often remain effective on other
models. We improve on prior work with a query-based attack that leverages API
access to a remote language model to construct adversarial examples that cause
the model to emit harmful strings with (much) higher probability than with
transfer-only attacks. We validate our attack on GPT-3.5 and OpenAI's safety
classifier; we can cause GPT-3.5 to emit harmful strings that current transfer
attacks fail at, and we can evade the safety classifier with nearly 100%
probability. | [
"cs.CL",
"cs.AI",
"cs.CR",
"cs.LG"
] | false |
2402.12348 | 2024-02-19T18:23:36Z | GTBench: Uncovering the Strategic Reasoning Limitations of LLMs via
Game-Theoretic Evaluations | [
"Jinhao Duan",
"Renming Zhang",
"James Diffenderfer",
"Bhavya Kailkhura",
"Lichao Sun",
"Elias Stengel-Eskin",
"Mohit Bansal",
"Tianlong Chen",
"Kaidi Xu"
] | As Large Language Models (LLMs) are integrated into critical real-world
applications, their strategic and logical reasoning abilities are increasingly
crucial. This paper evaluates LLMs' reasoning abilities in competitive
environments through game-theoretic tasks, e.g., board and card games that
require pure logic and strategic reasoning to compete with opponents. We first
propose GTBench, a language-driven environment composing 10 widely-recognized
tasks, across a comprehensive game taxonomy: complete versus incomplete
information, dynamic versus static, and probabilistic versus deterministic
scenarios. Then, we investigate two key problems: (1) Characterizing
game-theoretic reasoning of LLMs; (2) LLM-vs-LLM competitions as reasoning
evaluation. We observe that (1) LLMs have distinct behaviors regarding various
gaming scenarios; for example, LLMs fail in complete and deterministic games
yet they are competitive in probabilistic gaming scenarios; (2) Open-source
LLMs, e.g., CodeLlama-34b-Instruct, are less competitive than commercial LLMs,
e.g., GPT-4, in complex games. In addition, code-pretraining greatly benefits
strategic reasoning, while advanced reasoning methods such as Chain-of-Thought
(CoT) and Tree-of-Thought (ToT) do not always help. Detailed error profiles are
also provided for a better understanding of LLMs' behavior. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.12354 | 2024-02-19T18:33:49Z | LoRA+: Efficient Low Rank Adaptation of Large Models | [
"Soufiane Hayou",
"Nikhil Ghosh",
"Bin Yu"
] | In this paper, we show that Low Rank Adaptation (LoRA) as originally
introduced in Hu et al. (2021) leads to suboptimal finetuning of models with
large width (embedding dimension). This is due to the fact that adapter
matrices A and B in LoRA are updated with the same learning rate. Using scaling
arguments for large width networks, we demonstrate that using the same learning
rate for A and B does not allow efficient feature learning. We then show that
this suboptimality of LoRA can be corrected simply by setting different
learning rates for the LoRA adapter matrices A and B with a well-chosen ratio.
We call this proposed algorithm LoRA$+$. In our extensive experiments, LoRA$+$
improves performance (1-2 $\%$ improvements) and finetuning speed (up to $\sim$
2X SpeedUp), at the same computational cost as LoRA. | [
"cs.LG",
"cs.AI",
"cs.CL",
"stat.ML"
] | false |
2402.12366 | 2024-02-19T18:53:54Z | A Critical Evaluation of AI Feedback for Aligning Large Language Models | [
"Archit Sharma",
"Sedrick Keh",
"Eric Mitchell",
"Chelsea Finn",
"Kushal Arora",
"Thomas Kollar"
] | Reinforcement learning with AI feedback (RLAIF) is a popular paradigm for
improving the instruction-following abilities of powerful pre-trained language
models. RLAIF first performs supervised fine-tuning (SFT) using demonstrations
from a teacher model and then further fine-tunes the model with reinforcement
learning (RL), using feedback from a critic model. While recent popular
open-source models have demonstrated substantial improvements in performance
from the RL step, in this paper we question whether the complexity of this RL
step is truly warranted for AI feedback. We show that the improvements of the
RL step are virtually entirely due to the widespread practice of using a weaker
teacher model (e.g. GPT-3.5) for SFT data collection than the critic (e.g.,
GPT-4) used for AI feedback generation. Specifically, we show that simple
supervised fine-tuning with GPT-4 as the teacher outperforms existing RLAIF
pipelines. More generally, we find that the gains from RLAIF vary substantially
across base model families, test-time evaluation protocols, and critic models.
Finally, we provide a mechanistic explanation for when SFT may outperform the
full two-step RLAIF pipeline as well as suggestions for making RLAIF maximally
useful in practice. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2402.12419 | 2024-02-19T09:55:32Z | EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs | [
"Song Guo",
"Fan Wu",
"Lei Zhang",
"Xiawu Zheng",
"Shengchuan Zhang",
"Fei Chao",
"Yiyu Shi",
"Rongrong Ji"
] | Existing methods for fine-tuning sparse LLMs often suffer from
resource-intensive requirements and high retraining costs. Additionally, many
fine-tuning methods often rely on approximations or heuristic optimization
strategies, which may lead to suboptimal solutions. To address these issues, we
propose an efficient and fast framework for fine-tuning sparse LLMs based on
minimizing reconstruction error. Our approach involves sampling a small dataset
for calibration and utilizing backpropagation to iteratively optimize
block-wise reconstruction error, on a block-by-block basis, aiming for optimal
solutions. Extensive experiments on various benchmarks consistently demonstrate
the superiority of our method over other baselines. For instance, on the
Wikitext2 dataset with LlamaV1-7B at 70% sparsity, our proposed EBFT achieves a
perplexity of 16.88, surpassing the state-of-the-art DSnoT with a perplexity of
75.14. Moreover, with a structured sparsity ratio of 26\%, EBFT achieves a
perplexity of 16.27, outperforming LoRA (perplexity 16.44). Furthermore, the
fine-tuning process of EBFT for LlamaV1-7B only takes approximately 30 minutes,
and the entire framework can be executed on a single 16GB GPU. The source code
is available at https://github.com/sunggo/EBFT. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2402.12423 | 2024-02-19T16:22:21Z | On the Semantic Latent Space of Diffusion-Based Text-to-Speech Models | [
"Miri Varshavsky-Hassid",
"Roy Hirsch",
"Regev Cohen",
"Tomer Golany",
"Daniel Freedman",
"Ehud Rivlin"
] | The incorporation of Denoising Diffusion Models (DDMs) in the Text-to-Speech
(TTS) domain is rising, providing great value in synthesizing high quality
speech. Although they exhibit impressive audio quality, the extent of their
semantic capabilities is unknown, and controlling their synthesized speech's
vocal properties remains a challenge. Inspired by recent advances in image
synthesis, we explore the latent space of frozen TTS models, which is composed
of the latent bottleneck activations of the DDM's denoiser. We identify that
this space contains rich semantic information, and outline several novel
methods for finding semantic directions within it, both supervised and
unsupervised. We then demonstrate how these enable off-the-shelf audio editing,
without any further training, architectural changes or data requirements. We
present evidence of the semantic and acoustic qualities of the edited audio,
and provide supplemental samples:
https://latent-analysis-grad-tts.github.io/speech-samples/. | [
"cs.SD",
"cs.CL",
"cs.LG",
"eess.AS"
] | false |
2402.12530 | 2024-02-19T20:40:48Z | Parallel Structures in Pre-training Data Yield In-Context Learning | [
"Yanda Chen",
"Chen Zhao",
"Zhou Yu",
"Kathleen McKeown",
"He He"
] | Pre-trained language models (LMs) are capable of in-context learning (ICL):
they can adapt to a task with only a few examples given in the prompt without
any parameter update. However, it is unclear where this capability comes from
as there is a stark distribution shift between pre-training text and ICL
prompts. In this work, we study what patterns of the pre-training data
contribute to ICL. We find that LMs' ICL ability depends on $\textit{parallel
structures}$ in the pre-training data -- pairs of phrases following similar
templates in the same context window. Specifically, we detect parallel
structures by checking whether training on one phrase improves prediction of
the other, and conduct ablation experiments to study their effect on ICL. We
show that removing parallel structures in the pre-training data reduces LMs'
ICL accuracy by 51% (vs 2% from random ablation). This drop persists even when
excluding common patterns such as n-gram repetitions and long-range dependency,
showing the diversity and generality of parallel structures. A closer look at
the detected parallel structures indicates that they cover diverse linguistic
tasks and span long distances in the data. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.14838 | 2024-02-19T00:40:17Z | RFBES at SemEval-2024 Task 8: Investigating Syntactic and Semantic
Features for Distinguishing AI-Generated and Human-Written Texts | [
"Mohammad Heydari Rad",
"Farhan Farsi",
"Shayan Bali",
"Romina Etezadi",
"Mehrnoush Shamsfard"
] | Nowadays, the usage of Large Language Models (LLMs) has increased, and LLMs
have been used to generate texts in different languages and for different
tasks. Additionally, due to the participation of remarkable companies such as
Google and OpenAI, LLMs are now more accessible, and people can easily use
them. However, an important issue is how we can detect AI-generated texts from
human-written ones. In this article, we have investigated the problem of
AI-generated text detection from two different aspects: semantics and syntax.
Finally, we presented an AI model that can distinguish AI-generated texts from
human-written ones with high accuracy on both multilingual and monolingual
tasks using the M4 dataset. According to our results, using a semantic approach
would be more helpful for detection. However, there is a lot of room for
improvement in the syntactic approach, and it would be a good approach for
future work. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.14840 | 2024-02-19T06:57:02Z | RJUA-MedDQA: A Multimodal Benchmark for Medical Document Question
Answering and Clinical Reasoning | [
"Congyun Jin",
"Ming Zhang",
"Xiaowei Ma",
"Li Yujiao",
"Yingbo Wang",
"Yabo Jia",
"Yuliang Du",
"Tao Sun",
"Haowen Wang",
"Cong Fan",
"Jinjie Gu",
"Chenfei Chi",
"Xiangguo Lv",
"Fangzhou Li",
"Wei Xue",
"Yiran Huang"
] | Recent advancements in Large Language Models (LLMs) and Large Multi-modal
Models (LMMs) have shown potential in various medical applications, such as
Intelligent Medical Diagnosis. Although impressive results have been achieved,
we find that existing benchmarks do not reflect the complexity of real medical
reports and specialized in-depth reasoning capabilities. In this work, we
introduced RJUA-MedDQA, a comprehensive benchmark in the field of medical
specialization, which poses several challenges: comprehensively interpreting
imgage content across diverse challenging layouts, possessing numerical
reasoning ability to identify abnormal indicators and demonstrating clinical
reasoning ability to provide statements of disease diagnosis, status and advice
based on medical contexts. We carefully design the data generation pipeline and
proposed the Efficient Structural Restoration Annotation (ESRA) Method, aimed
at restoring textual and tabular content in medical report images. This method
substantially enhances annotation efficiency, doubling the productivity of each
annotator, and yields a 26.8% improvement in accuracy. We conduct extensive
evaluations, including few-shot assessments of 5 LMMs which are capable of
solving Chinese medical QA tasks. To further investigate the limitations and
potential of current LMMs, we conduct comparative experiments on a set of
strong LLMs by using image-text generated by ESRA method. We report the
performance of baselines and offer several observations: (1) The overall
performance of existing LMMs is still limited; however LMMs more robust to
low-quality and diverse-structured images compared to LLMs. (3) Reasoning
across context and image content present significant challenges. We hope this
benchmark helps the community make progress on these challenging tasks in
multi-modal medical document understanding and facilitate its application in
healthcare. | [
"cs.CL",
"cs.AI",
"stat.AP"
] | false |
2402.14843 | 2024-02-19T09:24:02Z | Text Diffusion with Reinforced Conditioning | [
"Yuxuan Liu",
"Tianchi Yang",
"Shaohan Huang",
"Zihan Zhang",
"Haizhen Huang",
"Furu Wei",
"Weiwei Deng",
"Feng Sun",
"Qi Zhang"
] | Diffusion models have demonstrated exceptional capability in generating
high-quality images, videos, and audio. Due to their adaptiveness in iterative
refinement, they provide a strong potential for achieving better
non-autoregressive sequence generation. However, existing text diffusion models
still fall short in their performance due to a challenge in handling the
discreteness of language. This paper thoroughly analyzes text diffusion models
and uncovers two significant limitations: degradation of self-conditioning
during training and misalignment between training and sampling. Motivated by
our findings, we propose a novel Text Diffusion model called TREC, which
mitigates the degradation with Reinforced Conditioning and the misalignment by
Time-Aware Variance Scaling. Our extensive experiments demonstrate the
competitiveness of TREC against autoregressive, non-autoregressive, and
diffusion baselines. Moreover, qualitative analysis shows its advanced ability
to fully utilize the diffusion process in refining samples. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.14845 | 2024-02-19T14:00:39Z | Purifying Large Language Models by Ensembling a Small Language Model | [
"Tianlin Li",
"Qian Liu",
"Tianyu Pang",
"Chao Du",
"Qing Guo",
"Yang Liu",
"Min Lin"
] | The emerging success of large language models (LLMs) heavily relies on
collecting abundant training data from external (untrusted) sources. Despite
substantial efforts devoted to data cleaning and curation, well-constructed
LLMs have been reported to suffer from copyright infringement, data poisoning,
and/or privacy violations, which would impede practical deployment of LLMs. In
this study, we propose a simple and easily implementable method for purifying
LLMs from the negative effects caused by uncurated data, namely, through
ensembling LLMs with benign and small language models (SLMs). Aside from
theoretical guarantees, we perform comprehensive experiments to empirically
confirm the efficacy of ensembling LLMs with SLMs, which can effectively
preserve the performance of LLMs while mitigating issues such as copyright
infringement, data poisoning, and privacy violations. | [
"cs.CL",
"cs.AI",
"cs.LG",
"I.2"
] | false |
2402.14846 | 2024-02-19T14:53:01Z | Stick to your Role! Stability of Personal Values Expressed in Large
Language Models | [
"Grgur Kovač",
"Rémy Portelas",
"Masataka Sawayama",
"Peter Ford Dominey",
"Pierre-Yves Oudeyer"
] | The standard way to study Large Language Models (LLMs) through benchmarks or
psychology questionnaires is to provide many different queries from similar
minimal contexts (e.g. multiple choice questions). However, due to LLM's highly
context-dependent nature, conclusions from such minimal-context evaluations may
be little informative about the model's behavior in deployment (where it will
be exposed to many new contexts). We argue that context-dependence should be
studied as another dimension of LLM comparison alongside others such as
cognitive abilities, knowledge, or model size. In this paper, we present a
case-study about the stability of value expression over different contexts
(simulated conversations on different topics), and as measured using a standard
psychology questionnaire (PVQ) and a behavioral downstream task. We consider 19
open-sourced LLMs from five families. Reusing methods from psychology, we study
Rank-order stability on the population (interpersonal) level, and Ipsative
stability on the individual (intrapersonal) level. We explore two settings:
with and without instructing LLMs to simulate particular personalities. We
observe similar trends in the stability of models and model families - Mixtral,
Mistral and Qwen families being more stable than LLaMa-2 and Phi - over those
two settings, two different simulated populations, and even in the downstream
behavioral task. When instructed to simulate particular personas, LLMs exhibit
low Rank-Order stability, and this stability further diminishes with
conversation length. This highlights the need for future research directions on
LLMs that can coherently simulate a diversity of personas, as well as how
context-dependence can be studied in more thorough and efficient ways. This
paper provides a foundational step in that direction, and, to our knowledge, it
is the first study of value stability in LLMs. | [
"cs.CL",
"cs.AI",
"cs.LG",
"68T07",
"I.2.7"
] | false |
2402.14849 | 2024-02-19T19:48:02Z | Asynchronous and Segmented Bidirectional Encoding for NMT | [
"Jingpu Yang",
"Zehua Han",
"Mengyu Xiang",
"Helin Wang",
"Yuxiao Huang",
"Miao Fang"
] | With the rapid advancement of Neural Machine Translation (NMT), enhancing
translation efficiency and quality has become a focal point of research.
Despite the commendable performance of general models such as the Transformer
in various aspects, they still fall short in processing long sentences and
fully leveraging bidirectional contextual information. This paper introduces an
improved model based on the Transformer, implementing an asynchronous and
segmented bidirectional decoding strategy aimed at elevating translation
efficiency and accuracy. Compared to traditional unidirectional translations
from left-to-right or right-to-left, our method demonstrates heightened
efficiency and improved translation quality, particularly in handling long
sentences. Experimental results on the IWSLT2017 dataset confirm the
effectiveness of our approach in accelerating translation and increasing
accuracy, especially surpassing traditional unidirectional strategies in long
sentence translation. Furthermore, this study analyzes the impact of sentence
length on decoding outcomes and explores the model's performance in various
scenarios. The findings of this research not only provide an effective encoding
strategy for the NMT field but also pave new avenues and directions for future
studies. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2403.08818 | 2024-02-19T23:48:40Z | Multimodal Fusion of EHR in Structures and Semantics: Integrating
Clinical Records and Notes with Hypergraph and LLM | [
"Hejie Cui",
"Xinyu Fang",
"Ran Xu",
"Xuan Kan",
"Joyce C. Ho",
"Carl Yang"
] | Electronic Health Records (EHRs) have become increasingly popular to support
clinical decision-making and healthcare in recent decades. EHRs usually contain
heterogeneous information, such as structural data in tabular form and
unstructured data in textual notes. Different types of information in EHRs can
complement each other and provide a more complete picture of the health status
of a patient. While there has been a lot of research on representation learning
of structured EHR data, the fusion of different types of EHR data (multimodal
fusion) is not well studied. This is mostly because of the complex medical
coding systems used and the noise and redundancy present in the written notes.
In this work, we propose a new framework called MINGLE, which integrates both
structures and semantics in EHR effectively. Our framework uses a two-level
infusion strategy to combine medical concept semantics and clinical note
semantics into hypergraph neural networks, which learn the complex interactions
between different types of data to generate visit representations for
downstream prediction. Experiment results on two EHR datasets, the public
MIMIC-III and private CRADLE, show that MINGLE can effectively improve
predictive performance by 11.83% relatively, enhancing semantic integration as
well as multimodal fusion for structural and textual EHR data. | [
"cs.LG",
"cs.AI",
"cs.CL"
] | false |
2403.15401 | 2024-02-19T17:58:41Z | Large Language Model for Mental Health: A Systematic Review | [
"Zhijun Guo",
"Alvina Lai",
"Johan Hilge Thygesen",
"Joseph Farrington",
"Thomas Keen",
"Kezhi Li"
] | Large language models (LLMs) have received much attention and shown their
potential in digital health, while their application in mental health is
subject to ongoing debate. This systematic review aims to summarize and
characterize the use of LLMs in mental health by investigating the strengths
and limitations of the latest work in LLMs and discusses the challenges and
opportunities for early screening, digital interventions, and other clinical
applications in mental health. Following PRISMA guidelines, we examined English
articles from PubMed, DBLP Computer Science Bibliography, and IEEE Xplore,
published between 1 January 2017, and 1 September 2023, focusing on mental
health and LLMs. The review analyzed 32 articles, including mental health
analysis using social media datasets (n=13), mental health chatbots (n=10), and
other mental health applications (n=9). Findings reveal LLMs' effectiveness in
mental health issue detection and the enhancement of telepsychological services
through personalised healthcare. Nonetheless, risks like text inconsistencies,
hallucinatory content, and the lack of an ethical framework raise concerns
about their clinical use. Despite these challenges, the advancement of LLMs
underscores their potential as innovative clinical tools, necessitating further
research and development. The review emphasizes that LLMs should complement,
not replace, professional mental health services. | [
"cs.CY",
"cs.AI",
"cs.CL"
] | false |
2402.12326 | 2024-02-19T18:00:30Z | LLM Agents for Psychology: A Study on Gamified Assessments | [
"Qisen Yang",
"Zekun Wang",
"Honghui Chen",
"Shenzhi Wang",
"Yifan Pu",
"Xin Gao",
"Wenhao Huang",
"Shiji Song",
"Gao Huang"
] | Psychological measurement is essential for mental health, self-understanding,
and personal development. Traditional methods, such as self-report scales and
psychologist interviews, often face challenges with engagement and
accessibility. While game-based and LLM-based tools have been explored to
improve user interest and automate assessment, they struggle to balance
engagement with generalizability. In this work, we propose PsychoGAT
(Psychological Game AgenTs) to achieve a generic gamification of psychological
assessment. The main insight is that powerful LLMs can function both as adept
psychologists and innovative game designers. By incorporating LLM agents into
designated roles and carefully managing their interactions, PsychoGAT can
transform any standardized scales into personalized and engaging interactive
fiction games. To validate the proposed method, we conduct psychometric
evaluations to assess its effectiveness and employ human evaluators to examine
the generated content across various psychological constructs, including
depression, cognitive distortions, and personality traits. Results demonstrate
that PsychoGAT serves as an effective assessment tool, achieving statistically
significant excellence in psychometric metrics such as reliability, convergent
validity, and discriminant validity. Moreover, human evaluations confirm
PsychoGAT's enhancements in content coherence, interactivity, interest,
immersion, and satisfaction. | [
"cs.CL",
"cs.CY",
"cs.HC",
"cs.LG",
"cs.MA"
] | false |
2402.12327 | 2024-02-19T18:00:53Z | Shall We Talk: Exploring Spontaneous Collaborations of Competing LLM
Agents | [
"Zengqing Wu",
"Shuyuan Zheng",
"Qianying Liu",
"Xu Han",
"Brian Inhyuk Kwon",
"Makoto Onizuka",
"Shaojie Tang",
"Run Peng",
"Chuan Xiao"
] | Recent advancements have shown that agents powered by large language models
(LLMs) possess capabilities to simulate human behaviors and societal dynamics.
However, the potential for LLM agents to spontaneously establish collaborative
relationships in the absence of explicit instructions has not been studied. To
address this gap, we conduct three case studies, revealing that LLM agents are
capable of spontaneously forming collaborations even within competitive
settings. This finding not only demonstrates the capacity of LLM agents to
mimic competition and cooperation in human societies but also validates a
promising vision of computational social science. Specifically, it suggests
that LLM agents could be utilized to model human social interactions, including
those with spontaneous collaborations, thus offering insights into social
phenomena. The source codes for this study are available at
https://github.com/wuzengqing001225/SABM_ShallWeTalk . | [
"cs.AI",
"cs.CL",
"cs.CY",
"cs.MA",
"econ.GN",
"q-fin.EC"
] | false |
2402.11838 | 2024-02-19T05:04:11Z | UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal
Prediction | [
"Yuan Yuan",
"Jingtao Ding",
"Jie Feng",
"Depeng Jin",
"Yong Li"
] | Urban spatio-temporal prediction is crucial for informed decision-making,
such as transportation management, resource optimization, and urban planning.
Although pretrained foundation models for natural languages have experienced
remarkable breakthroughs, wherein one general-purpose model can tackle multiple
tasks across various domains, urban spatio-temporal modeling lags behind.
Existing approaches for urban prediction are usually tailored for specific
spatio-temporal scenarios, requiring task-specific model designs and extensive
in-domain training data. In this work, we propose a universal model, UniST, for
urban spatio-temporal prediction. Drawing inspiration from large language
models, UniST achieves success through: (i) flexibility towards diverse
spatio-temporal data characteristics, (ii) effective generative pre-training
with elaborated masking strategies to capture complex spatio-temporal
relationships, (iii) spatio-temporal knowledge-guided prompts that align and
leverage intrinsic and shared knowledge across scenarios. These designs
together unlock the potential of a one-for-all model for spatio-temporal
prediction with powerful generalization capability. Extensive experiments on 15
cities and 6 domains demonstrate the universality of UniST in advancing
state-of-the-art prediction performance, especially in few-shot and zero-shot
scenarios. | [
"cs.LG"
] | false |
2402.11995 | 2024-02-19T09:39:54Z | Network Inversion of Binarised Neural Nets | [
"Pirzada Suhail",
"Supratik Chakraborty",
"Amit Sethi"
] | While the deployment of neural networks, yielding impressive results, becomes
more prevalent in various applications, their interpretability and
understanding remain a critical challenge. Network inversion, a technique that
aims to reconstruct the input space from the model's learned internal
representations, plays a pivotal role in unraveling the black-box nature of
input to output mappings in neural networks. In safety-critical scenarios,
where model outputs may influence pivotal decisions, the integrity of the
corresponding input space is paramount, necessitating the elimination of any
extraneous "garbage" to ensure the trustworthiness of the network. Binarised
Neural Networks (BNNs), characterized by binary weights and activations, offer
computational efficiency and reduced memory requirements, making them suitable
for resource-constrained environments. This paper introduces a novel approach
to invert a trained BNN by encoding it into a CNF formula that captures the
network's structure, allowing for both inference and inversion. | [
"cs.LG"
] | false |
2402.12149 | 2024-02-19T14:02:13Z | MLFEF: Machine Learning Fusion Model with Empirical Formula to Explore
the Momentum in Competitive Sports | [
"Ruixin Peng",
"Ziqing Li"
] | Tennis is so popular that coaches and players are curious about factors other
than skill, such as momentum. This article will try to define and quantify
momentum, providing a basis for real-time analysis of tennis matches. Based on
the tennis Grand Slam men's singles match data in recent years, we built two
models, one is to build a model based on data-driven, and the other is to build
a model based on empirical formulas. For the data-driven model, we first found
a large amount of public data including public data on tennis matches in the
past five years and personal information data of players. Then the data is
preprocessed, and feature engineered, and a fusion model of SVM, Random Forrest
algorithm and XGBoost was established. For the mechanism analysis model,
important features were selected based on the suggestions of many tennis
players and enthusiasts, the sliding window algorithm was used to calculate the
weight, and different methods were used to visualize the momentum. For further
analysis of the momentum fluctuation, it is based on the popular CUMSUM
algorithm in the industry as well as the RUN Test, and the result shows the
momentum is not random and the trend might be random. At last, the robustness
of the fusion model is analyzed by Monte Carlo simulation. | [
"cs.LG"
] | false |
2402.12201 | 2024-02-19T15:04:53Z | Dictionary Learning Improves Patch-Free Circuit Discovery in Mechanistic
Interpretability: A Case Study on Othello-GPT | [
"Zhengfu He",
"Xuyang Ge",
"Qiong Tang",
"Tianxiang Sun",
"Qinyuan Cheng",
"Xipeng Qiu"
] | Sparse dictionary learning has been a rapidly growing technique in
mechanistic interpretability to attack superposition and extract more
human-understandable features from model activations. We ask a further question
based on the extracted more monosemantic features: How do we recognize circuits
connecting the enormous amount of dictionary features? We propose a circuit
discovery framework alternative to activation patching. Our framework suffers
less from out-of-distribution and proves to be more efficient in terms of
asymptotic complexity. The basic unit in our framework is dictionary features
decomposed from all modules writing to the residual stream, including
embedding, attention output and MLP output. Starting from any logit, dictionary
feature or attention score, we manage to trace down to lower-level dictionary
features of all tokens and compute their contribution to these more
interpretable and local model behaviors. We dig in a small transformer trained
on a synthetic task named Othello and find a number of human-understandable
fine-grained circuits inside of it. | [
"cs.LG"
] | false |
2402.12231 | 2024-02-19T15:36:36Z | Diffusion Tempering Improves Parameter Estimation with Probabilistic
Integrators for Ordinary Differential Equations | [
"Jonas Beck",
"Nathanael Bosch",
"Michael Deistler",
"Kyra L. Kadhim",
"Jakob H. Macke",
"Philipp Hennig",
"Philipp Berens"
] | Ordinary differential equations (ODEs) are widely used to describe dynamical
systems in science, but identifying parameters that explain experimental
measurements is challenging. In particular, although ODEs are differentiable
and would allow for gradient-based parameter optimization, the nonlinear
dynamics of ODEs often lead to many local minima and extreme sensitivity to
initial conditions. We therefore propose diffusion tempering, a novel
regularization technique for probabilistic numerical methods which improves
convergence of gradient-based parameter optimization in ODEs. By iteratively
reducing a noise parameter of the probabilistic integrator, the proposed method
converges more reliably to the true parameters. We demonstrate that our method
is effective for dynamical systems of different complexity and show that it
obtains reliable parameter estimates for a Hodgkin-Huxley model with a
practically relevant number of parameters. | [
"cs.LG"
] | false |
2402.12242 | 2024-02-19T15:57:39Z | Synthetic location trajectory generation using categorical diffusion
models | [
"Simon Dirmeier",
"Ye Hong",
"Fernando Perez-Cruz"
] | Diffusion probabilistic models (DPMs) have rapidly evolved to be one of the
predominant generative models for the simulation of synthetic data, for
instance, for computer vision, audio, natural language processing, or
biomolecule generation. Here, we propose using DPMs for the generation of
synthetic individual location trajectories (ILTs) which are sequences of
variables representing physical locations visited by individuals. ILTs are of
major importance in mobility research to understand the mobility behavior of
populations and to ultimately inform political decision-making. We represent
ILTs as multi-dimensional categorical random variables and propose to model
their joint distribution using a continuous DPM by first applying the diffusion
process in a continuous unconstrained space and then mapping the continuous
variables into a discrete space. We demonstrate that our model can synthesize
realistic ILPs by comparing conditionally and unconditionally generated
sequences to real-world ILPs from a GNSS tracking data set which suggests the
potential use of our model for synthetic data generation, for example, for
benchmarking models used in mobility research. | [
"cs.LG"
] | false |
2402.11778 | 2024-02-19T02:08:09Z | Towards Theoretical Understandings of Self-Consuming Generative Models | [
"Shi Fu",
"Sen Zhang",
"Yingjie Wang",
"Xinmei Tian",
"Dacheng Tao"
] | This paper tackles the emerging challenge of training generative models
within a self-consuming loop, wherein successive generations of models are
recursively trained on mixtures of real and synthetic data from previous
generations. We construct a theoretical framework to rigorously evaluate how
this training regimen impacts the data distributions learned by future models.
Specifically, we derive bounds on the total variation (TV) distance between the
synthetic data distributions produced by future models and the original real
data distribution under various mixed training scenarios. Our analysis
demonstrates that this distance can be effectively controlled under the
condition that mixed training dataset sizes or proportions of real data are
large enough. Interestingly, we further unveil a phase transition induced by
expanding synthetic data amounts, proving theoretically that while the TV
distance exhibits an initial ascent, it declines beyond a threshold point.
Finally, we specialize our general results to diffusion models, delivering
nuanced insights such as the efficacy of optimal early stopping within the
self-consuming loop. | [
"cs.LG",
"cs.AI"
] | false |
2402.11857 | 2024-02-19T05:59:09Z | Communication-Efficient Distributed Learning with Local Immediate Error
Compensation | [
"Yifei Cheng",
"Li Shen",
"Linli Xu",
"Xun Qian",
"Shiwei Wu",
"Yiming Zhou",
"Tie Zhang",
"Dacheng Tao",
"Enhong Chen"
] | Gradient compression with error compensation has attracted significant
attention with the target of reducing the heavy communication overhead in
distributed learning. However, existing compression methods either perform only
unidirectional compression in one iteration with higher communication cost, or
bidirectional compression with slower convergence rate. In this work, we
propose the Local Immediate Error Compensated SGD (LIEC-SGD) optimization
algorithm to break the above bottlenecks based on bidirectional compression and
carefully designed compensation approaches. Specifically, the bidirectional
compression technique is to reduce the communication cost, and the compensation
technique compensates the local compression error to the model update
immediately while only maintaining the global error variable on the server
throughout the iterations to boost its efficacy. Theoretically, we prove that
LIEC-SGD is superior to previous works in either the convergence rate or the
communication cost, which indicates that LIEC-SGD could inherit the dual
advantages from unidirectional compression and bidirectional compression.
Finally, experiments of training deep neural networks validate the
effectiveness of the proposed LIEC-SGD algorithm. | [
"cs.LG",
"cs.DC"
] | false |
2402.11867 | 2024-02-19T06:22:09Z | LoRA Training in the NTK Regime has No Spurious Local Minima | [
"Uijeong Jang",
"Jason D. Lee",
"Ernest K. Ryu"
] | Low-rank adaptation (LoRA) has become the standard approach for
parameter-efficient fine-tuning of large language models (LLM), but our
theoretical understanding of LoRA has been limited. In this work, we
theoretically analyze LoRA fine-tuning in the neural tangent kernel (NTK)
regime with $N$ data points, showing: (i) full fine-tuning (without LoRA)
admits a low-rank solution of rank $r\lesssim \sqrt{N}$; (ii) using LoRA with
rank $r\gtrsim \sqrt{N}$ eliminates spurious local minima, allowing gradient
descent to find the low-rank solutions; (iii) the low-rank solution found using
LoRA generalizes well. | [
"cs.LG",
"math.OC"
] | false |
2402.11877 | 2024-02-19T06:33:51Z | Finite-Time Error Analysis of Online Model-Based Q-Learning with a
Relaxed Sampling Model | [
"Han-Dong Lim",
"HyeAnn Lee",
"Donghwan Lee"
] | Reinforcement learning has witnessed significant advancements, particularly
with the emergence of model-based approaches. Among these, $Q$-learning has
proven to be a powerful algorithm in model-free settings. However, the
extension of $Q$-learning to a model-based framework remains relatively
unexplored. In this paper, we delve into the sample complexity of $Q$-learning
when integrated with a model-based approach. Through theoretical analyses and
empirical evaluations, we seek to elucidate the conditions under which
model-based $Q$-learning excels in terms of sample efficiency compared to its
model-free counterpart. | [
"cs.LG",
"cs.AI"
] | false |
2402.11904 | 2024-02-19T07:45:04Z | Scalable Virtual Valuations Combinatorial Auction Design by Combining
Zeroth-Order and First-Order Optimization Method | [
"Zhijian Duan",
"Haoran Sun",
"Yichong Xia",
"Siqiang Wang",
"Zhilin Zhang",
"Chuan Yu",
"Jian Xu",
"Bo Zheng",
"Xiaotie Deng"
] | Automated auction design seeks to discover empirically high-revenue and
incentive-compatible mechanisms using machine learning. Ensuring dominant
strategy incentive compatibility (DSIC) is crucial, and the most effective
approach is to confine the mechanism to Affine Maximizer Auctions (AMAs).
Nevertheless, existing AMA-based approaches encounter challenges such as
scalability issues (arising from combinatorial candidate allocations) and the
non-differentiability of revenue. In this paper, to achieve a scalable
AMA-based method, we further restrict the auction mechanism to Virtual
Valuations Combinatorial Auctions (VVCAs), a subset of AMAs with significantly
fewer parameters. Initially, we employ a parallelizable dynamic programming
algorithm to compute the winning allocation of a VVCA. Subsequently, we propose
a novel optimization method that combines both zeroth-order and first-order
techniques to optimize the VVCA parameters. Extensive experiments demonstrate
the efficacy and scalability of our proposed approach, termed Zeroth-order and
First-order Optimization of VVCAs (ZFO-VVCA), particularly when applied to
large-scale auctions. | [
"cs.GT",
"cs.LG"
] | false |
2402.11933 | 2024-02-19T08:19:26Z | SLADE: Detecting Dynamic Anomalies in Edge Streams without Labels via
Self-Supervised Learning | [
"Jongha Lee",
"Sunwoo Kim",
"Kijung Shin"
] | To detect anomalies in real-world graphs, such as social, email, and
financial networks, various approaches have been developed. While they
typically assume static input graphs, most real-world graphs grow over time,
naturally represented as edge streams. In this context, we aim to achieve three
goals: (a) instantly detecting anomalies as they occur, (b) adapting to
dynamically changing states, and (c) handling the scarcity of dynamic anomaly
labels. In this paper, we propose SLADE (Self-supervised Learning for Anomaly
Detection in Edge Streams) for rapid detection of dynamic anomalies in edge
streams, without relying on labels. SLADE detects the shifts of nodes into
abnormal states by observing deviations in their interaction patterns over
time. To this end, it trains a deep neural network to perform two
self-supervised tasks: (a) minimizing drift in node representations and (b)
generating long-term interaction patterns from short-term ones. Failure in
these tasks for a node signals its deviation from the norm. Notably, the neural
network and tasks are carefully designed so that all required operations can be
performed in constant time (w.r.t. the graph size) in response to each new edge
in the input stream. In dynamic anomaly detection across four real-world
datasets, SLADE outperforms nine competing methods, even those leveraging label
supervision. | [
"cs.LG",
"cs.SI"
] | false |
2402.11953 | 2024-02-19T08:47:20Z | Stealing the Invisible: Unveiling Pre-Trained CNN Models through
Adversarial Examples and Timing Side-Channels | [
"Shubhi Shukla",
"Manaar Alam",
"Pabitra Mitra",
"Debdeep Mukhopadhyay"
] | Machine learning, with its myriad applications, has become an integral
component of numerous technological systems. A common practice in this domain
is the use of transfer learning, where a pre-trained model's architecture,
readily available to the public, is fine-tuned to suit specific tasks. As
Machine Learning as a Service (MLaaS) platforms increasingly use pre-trained
models in their backends, it's crucial to safeguard these architectures and
understand their vulnerabilities. In this work, we present an approach based on
the observation that the classification patterns of adversarial images can be
used as a means to steal the models. Furthermore, the adversarial image
classifications in conjunction with timing side channels can lead to a model
stealing method. Our approach, designed for typical user-level access in remote
MLaaS environments exploits varying misclassifications of adversarial images
across different models to fingerprint several renowned Convolutional Neural
Network (CNN) and Vision Transformer (ViT) architectures. We utilize the
profiling of remote model inference times to reduce the necessary adversarial
images, subsequently decreasing the number of queries required. We have
presented our results over 27 pre-trained models of different CNN and ViT
architectures using CIFAR-10 dataset and demonstrate a high accuracy of 88.8%
while keeping the query budget under 20. | [
"cs.CR",
"cs.LG"
] | false |
2402.11963 | 2024-02-19T09:06:26Z | Imbalance in Regression Datasets | [
"Daniel Kowatsch",
"Nicolas M. Müller",
"Kilian Tscharke",
"Philip Sperl",
"Konstantin Bötinger"
] | For classification, the problem of class imbalance is well known and has been
extensively studied. In this paper, we argue that imbalance in regression is an
equally important problem which has so far been overlooked: Due to under- and
over-representations in a data set's target distribution, regressors are prone
to degenerate to naive models, systematically neglecting uncommon training data
and over-representing targets seen often during training. We analyse this
problem theoretically and use resulting insights to develop a first definition
of imbalance in regression, which we show to be a generalisation of the
commonly employed imbalance measure in classification. With this, we hope to
turn the spotlight on the overlooked problem of imbalance in regression and to
provide common ground for future research. | [
"cs.LG",
"cs.AI"
] | false |
2402.11973 | 2024-02-19T09:19:01Z | Bayesian Active Learning for Censored Regression | [
"Frederik Boe Hüttel",
"Christoffer Riis",
"Filipe Rodrigues",
"Francisco Câmara Pereira"
] | Bayesian active learning is based on information theoretical approaches that
focus on maximising the information that new observations provide to the model
parameters. This is commonly done by maximising the Bayesian Active Learning by
Disagreement (BALD) acquisitions function. However, we highlight that it is
challenging to estimate BALD when the new data points are subject to
censorship, where only clipped values of the targets are observed. To address
this, we derive the entropy and the mutual information for censored
distributions and derive the BALD objective for active learning in censored
regression ($\mathcal{C}$-BALD). We propose a novel modelling approach to
estimate the $\mathcal{C}$-BALD objective and use it for active learning in the
censored setting. Across a wide range of datasets and models, we demonstrate
that $\mathcal{C}$-BALD outperforms other Bayesian active learning methods in
censored regression. | [
"cs.LG",
"stat.ML"
] | false |
2402.12034 | 2024-02-19T10:42:34Z | When Do Off-Policy and On-Policy Policy Gradient Methods Align? | [
"Davide Mambelli",
"Stephan Bongers",
"Onno Zoeter",
"Matthijs T. J. Spaan",
"Frans A. Oliehoek"
] | Policy gradient methods are widely adopted reinforcement learning algorithms
for tasks with continuous action spaces. These methods succeeded in many
application domains, however, because of their notorious sample inefficiency
their use remains limited to problems where fast and accurate simulations are
available. A common way to improve sample efficiency is to modify their
objective function to be computable from off-policy samples without importance
sampling. A well-established off-policy objective is the excursion objective.
This work studies the difference between the excursion objective and the
traditional on-policy objective, which we refer to as the on-off gap. We
provide the first theoretical analysis showing conditions to reduce the on-off
gap while establishing empirical evidence of shortfalls arising when these
conditions are not met. | [
"stat.ML",
"cs.LG"
] | false |
2402.12035 | 2024-02-19T10:43:13Z | Class-incremental Learning for Time Series: Benchmark and Evaluation | [
"Zhongzheng Qiao",
"Quang Pham",
"Zhen Cao",
"Hoang H Le",
"P. N. Suganthan",
"Xudong Jiang",
"Ramasamy Savitha"
] | Real-world environments are inherently non-stationary, frequently introducing
new classes over time. This is especially common in time series classification,
such as the emergence of new disease classification in healthcare or the
addition of new activities in human activity recognition. In such cases, a
learning system is required to assimilate novel classes effectively while
avoiding catastrophic forgetting of the old ones, which gives rise to the
Class-incremental Learning (CIL) problem. However, despite the encouraging
progress in the image and language domains, CIL for time series data remains
relatively understudied. Existing studies suffer from inconsistent experimental
designs, necessitating a comprehensive evaluation and benchmarking of methods
across a wide range of datasets. To this end, we first present an overview of
the Time Series Class-incremental Learning (TSCIL) problem, highlight its
unique challenges, and cover the advanced methodologies. Further, based on
standardized settings, we develop a unified experimental framework that
supports the rapid development of new algorithms, easy integration of new
datasets, and standardization of the evaluation process. Using this framework,
we conduct a comprehensive evaluation of various generic and
time-series-specific CIL methods in both standard and privacy-sensitive
scenarios. Our extensive experiments not only provide a standard baseline to
support future research but also shed light on the impact of various design
factors such as normalization layers or memory budget thresholds. Codes are
available at https://github.com/zqiao11/TSCIL. | [
"cs.LG",
"cs.AI"
] | false |
2402.12118 | 2024-02-19T13:13:16Z | DualView: Data Attribution from the Dual Perspective | [
"Galip Ümit Yolcu",
"Thomas Wiegand",
"Wojciech Samek",
"Sebastian Lapuschkin"
] | Local data attribution (or influence estimation) techniques aim at estimating
the impact that individual data points seen during training have on particular
predictions of an already trained Machine Learning model during test time.
Previous methods either do not perform well consistently across different
evaluation criteria from literature, are characterized by a high computational
demand, or suffer from both. In this work we present DualView, a novel method
for post-hoc data attribution based on surrogate modelling, demonstrating both
high computational efficiency, as well as good evaluation results. With a focus
on neural networks, we evaluate our proposed technique using suitable
quantitative evaluation strategies from the literature against related
principal local data attribution methods. We find that DualView requires
considerably lower computational resources than other methods, while
demonstrating comparable performance to competing approaches across evaluation
metrics. Futhermore, our proposed method produces sparse explanations, where
sparseness can be tuned via a hyperparameter. Finally, we showcase that with
DualView, we can now render explanations from local data attributions
compatible with established local feature attribution methods: For each
prediction on (test) data points explained in terms of impactful samples from
the training set, we are able to compute and visualize how the prediction on
(test) sample relates to each influential training sample in terms of features
recognized and by the model. We provide an Open Source implementation of
DualView online, together with implementations for all other local data
attribution methods we compare against, as well as the metrics reported here,
for full reproducibility. | [
"cs.LG",
"cs.AI"
] | false |
2402.12134 | 2024-02-19T13:32:30Z | Molecule Generation and Optimization for Efficient Fragrance Creation | [
"Bruno C. L. Rodrigues",
"Vinicius V. Santana",
"Sandris Murins",
"Idelfonso B. R. Nogueira"
] | This research introduces a Machine Learning-centric approach to replicate
olfactory experiences, validated through experimental quantification of perfume
perception. Key contributions encompass a hybrid model connecting perfume
molecular structure to human olfactory perception. This model includes an
AI-driven molecule generator (utilizing Graph and Generative Neural Networks),
quantification and prediction of odor intensity, and refinery of optimal
solvent and molecule combinations for desired fragrances. Additionally, a
thermodynamic-based model establishes a link between olfactory perception and
liquid-phase concentrations. The methodology employs Transfer Learning and
selects the most suitable molecules based on vapor pressure and fragrance
notes. Ultimately, a mathematical optimization problem is formulated to
minimize discrepancies between new and target olfactory experiences. The
methodology is validated by reproducing two distinct olfactory experiences
using available experimental data. | [
"physics.chem-ph",
"cs.LG"
] | false |
2402.12142 | 2024-02-19T13:52:37Z | Federated Bayesian Network Ensembles | [
"Florian van Daalen",
"Lianne Ippel",
"Andre Dekker",
"Inigo Bermejo"
] | Federated learning allows us to run machine learning algorithms on
decentralized data when data sharing is not permitted due to privacy concerns.
Ensemble-based learning works by training multiple (weak) classifiers whose
output is aggregated. Federated ensembles are ensembles applied to a federated
setting, where each classifier in the ensemble is trained on one data location.
In this article, we explore the use of federated ensembles of Bayesian
networks (FBNE) in a range of experiments and compare their performance with
locally trained models and models trained with VertiBayes, a federated learning
algorithm to train Bayesian networks from decentralized data. Our results show
that FBNE outperforms local models and provides a significant increase in
training speed compared with VertiBayes while maintaining a similar performance
in most settings, among other advantages. We show that FBNE is a potentially
useful tool within the federated learning toolbox, especially when local
populations are heavily biased, or there is a strong imbalance in population
size across parties. We discuss the advantages and disadvantages of this
approach in terms of time complexity, model accuracy, privacy protection, and
model interpretability. | [
"cs.LG",
"cs.CR"
] | false |
2402.12175 | 2024-02-19T14:29:35Z | Learning Discretized Bayesian Networks with GOMEA | [
"Damy M. F. Ha",
"Tanja Alderliesten",
"Peter A. N. Bosman"
] | Bayesian networks model relationships between random variables under
uncertainty and can be used to predict the likelihood of events and outcomes
while incorporating observed evidence. From an eXplainable AI (XAI)
perspective, such models are interesting as they tend to be compact. Moreover,
captured relations can be directly inspected by domain experts. In practice,
data is often real-valued. Unless assumptions of normality can be made,
discretization is often required. The optimal discretization, however, depends
on the relations modelled between the variables. This complicates learning
Bayesian networks from data. For this reason, most literature focuses on
learning conditional dependencies between sets of variables, called structure
learning. In this work, we extend an existing state-of-the-art structure
learning approach based on the Gene-pool Optimal Mixing Evolutionary Algorithm
(GOMEA) to jointly learn variable discretizations. The proposed Discretized
Bayesian Network GOMEA (DBN-GOMEA) obtains similar or better results than the
current state-of-the-art when tasked to retrieve randomly generated
ground-truth networks. Moreover, leveraging a key strength of evolutionary
algorithms, we can straightforwardly perform DBN learning multi-objectively. We
show how this enables incorporating expert knowledge in a uniquely insightful
fashion, finding multiple DBNs that trade-off complexity, accuracy, and the
difference with a pre-determined expert network. | [
"cs.LG",
"cs.NE"
] | false |
2402.12220 | 2024-02-19T15:26:19Z | Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic
Forgetting | [
"Haolin Chen",
"Philip N. Garner"
] | Although motivated by the adaptation of text-to-speech synthesis models, we
argue that more generic parameter-efficient fine-tuning (PEFT) is an
appropriate framework to do such adaptation. However, catastrophic forgetting
remains an issue with PEFT, damaging the pre-trained model's inherent
capabilities. We demonstrate that existing Bayesian learning techniques can be
applied to PEFT to prevent catastrophic forgetting as long as the parameter
shift of the fine-tuned layers can be calculated differentiably. In a
principled series of experiments on language modeling and speech synthesis
tasks, we utilize established Laplace approximations, including diagonal and
Kronecker factored approaches, to regularize PEFT with the low-rank adaptation
(LoRA) and compare their performance in pre-training knowledge preservation.
Our results demonstrate that catastrophic forgetting can be overcome by our
methods without degrading the fine-tuning performance, and using the Kronecker
factored approximations produces a better preservation of the pre-training
knowledge than the diagonal ones. | [
"eess.AS",
"cs.LG"
] | false |
2402.12235 | 2024-02-19T15:44:54Z | The Fundamental Limits of Least-Privilege Learning | [
"Theresa Stadler",
"Bogdan Kulynych",
"Nicoals Papernot",
"Michael Gastpar",
"Carmela Troncoso"
] | The promise of least-privilege learning -- to find feature representations
that are useful for a learning task but prevent inference of any sensitive
information unrelated to this task -- is highly appealing. However, so far this
concept has only been stated informally. It thus remains an open question
whether and how we can achieve this goal. In this work, we provide the first
formalisation of the least-privilege principle for machine learning and
characterise its feasibility. We prove that there is a fundamental trade-off
between a representation's utility for a given task and its leakage beyond the
intended task: it is not possible to learn representations that have high
utility for the intended task but, at the same time prevent inference of any
attribute other than the task label itself. This trade-off holds regardless of
the technique used to learn the feature mappings that produce these
representations. We empirically validate this result for a wide range of
learning techniques, model architectures, and datasets. | [
"cs.LG",
"cs.CR"
] | false |
2402.12240 | 2024-02-19T15:54:36Z | BEARS Make Neuro-Symbolic Models Aware of their Reasoning Shortcuts | [
"Emanuele Marconato",
"Samuele Bortolotti",
"Emile van Krieken",
"Antonio Vergari",
"Andrea Passerini",
"Stefano Teso"
] | Neuro-Symbolic (NeSy) predictors that conform to symbolic knowledge -
encoding, e.g., safety constraints - can be affected by Reasoning Shortcuts
(RSs): They learn concepts consistent with the symbolic knowledge by exploiting
unintended semantics. RSs compromise reliability and generalization and, as we
show in this paper, they are linked to NeSy models being overconfident about
the predicted concepts. Unfortunately, the only trustworthy mitigation strategy
requires collecting costly dense supervision over the concepts. Rather than
attempting to avoid RSs altogether, we propose to ensure NeSy models are aware
of the semantic ambiguity of the concepts they learn, thus enabling their users
to identify and distrust low-quality concepts. Starting from three simple
desiderata, we derive bears (BE Aware of Reasoning Shortcuts), an ensembling
technique that calibrates the model's concept-level confidence without
compromising prediction accuracy, thus encouraging NeSy architectures to be
uncertain about concepts affected by RSs. We show empirically that bears
improves RS-awareness of several state-of-the-art NeSy models, and also
facilitates acquiring informative dense annotations for mitigation purposes. | [
"cs.LG",
"cs.AI"
] | false |
2402.12271 | 2024-02-19T16:34:59Z | Secure Federated Learning Across Heterogeneous Cloud and
High-Performance Computing Resources -- A Case Study on Federated Fine-tuning
of LLaMA 2 | [
"Zilinghan Li",
"Shilan He",
"Pranshu Chaturvedi",
"Volodymyr Kindratenko",
"Eliu A Huerta",
"Kibaek Kim",
"Ravi Madduri"
] | Federated learning enables multiple data owners to collaboratively train
robust machine learning models without transferring large or sensitive local
datasets by only sharing the parameters of the locally trained models. In this
paper, we elaborate on the design of our Advanced Privacy-Preserving Federated
Learning (APPFL) framework, which streamlines end-to-end secure and reliable
federated learning experiments across cloud computing facilities and
high-performance computing resources by leveraging Globus Compute, a
distributed function as a service platform, and Amazon Web Services. We further
demonstrate the use case of APPFL in fine-tuning a LLaMA 2 7B model using
several cloud resources and supercomputers. | [
"cs.DC",
"cs.LG"
] | false |
2402.12284 | 2024-02-19T16:51:29Z | Refining Minimax Regret for Unsupervised Environment Design | [
"Michael Beukman",
"Samuel Coward",
"Michael Matthews",
"Mattie Fellows",
"Minqi Jiang",
"Michael Dennis",
"Jakob Foerster"
] | In unsupervised environment design, reinforcement learning agents are trained
on environment configurations (levels) generated by an adversary that maximises
some objective. Regret is a commonly used objective that theoretically results
in a minimax regret (MMR) policy with desirable robustness guarantees; in
particular, the agent's maximum regret is bounded. However, once the agent
reaches this regret bound on all levels, the adversary will only sample levels
where regret cannot be further reduced. Although there are possible performance
improvements to be made outside of these regret-maximising levels, learning
stagnates. In this work, we introduce Bayesian level-perfect MMR (BLP), a
refinement of the minimax regret objective that overcomes this limitation. We
formally show that solving for this objective results in a subset of MMR
policies, and that BLP policies act consistently with a Perfect Bayesian policy
over all levels. We further introduce an algorithm, ReMiDi, that results in a
BLP policy at convergence. We empirically demonstrate that training on levels
from a minimax regret adversary causes learning to prematurely stagnate, but
that ReMiDi continues learning. | [
"cs.LG",
"cs.AI"
] | false |
2402.12307 | 2024-02-19T17:30:09Z | Multi-View Conformal Learning for Heterogeneous Sensor Fusion | [
"Enrique Garcia-Ceja"
] | Being able to assess the confidence of individual predictions in machine
learning models is crucial for decision making scenarios. Specially, in
critical applications such as medical diagnosis, security, and unmanned
vehicles, to name a few. In the last years, complex predictive models have had
great success in solving hard tasks and new methods are being proposed every
day. While the majority of new developments in machine learning models focus on
improving the overall performance, less effort is put on assessing the
trustworthiness of individual predictions, and even to a lesser extent, in the
context of sensor fusion. To this end, we build and test multi-view and
single-view conformal models for heterogeneous sensor fusion. Our models
provide theoretical marginal confidence guarantees since they are based on the
conformal prediction framework. We also propose a multi-view semi-conformal
model based on sets intersection. Through comprehensive experimentation, we
show that multi-view models perform better than single-view models not only in
terms of accuracy-based performance metrics (as it has already been shown in
several previous works) but also in conformal measures that provide uncertainty
estimation. Our results also showed that multi-view models generate prediction
sets with less uncertainty compared to single-view models. | [
"cs.LG",
"cs.AI"
] | false |
2402.12415 | 2024-02-19T07:47:23Z | Vehicle-group-based Crash Risk Formation and Propagation Analysis for
Expressways | [
"Tianheng Zhu",
"Ling Wang",
"Yiheng Feng",
"Wanjing Ma",
"Mohamed Abdel-Aty"
] | Previous studies in predicting crash risk primarily associated the number or
likelihood of crashes on a road segment with traffic parameters or geometric
characteristics of the segment, usually neglecting the impact of vehicles'
continuous movement and interactions with nearby vehicles. Advancements in
communication technologies have empowered driving information collected from
surrounding vehicles, enabling the study of group-based crash risks. Based on
high-resolution vehicle trajectory data, this research focused on vehicle
groups as the subject of analysis and explored risk formation and propagation
mechanisms considering features of vehicle groups and road segments. Several
key factors contributing to crash risks were identified, including past
high-risk vehicle-group states, complex vehicle behaviors, high percentage of
large vehicles, frequent lane changes within a vehicle group, and specific road
geometries. A multinomial logistic regression model was developed to analyze
the spatial risk propagation patterns, which were classified based on the trend
of high-risk occurrences within vehicle groups. The results indicated that
extended periods of high-risk states, increase in vehicle-group size, and
frequent lane changes are associated with adverse risk propagation patterns.
Conversely, smoother traffic flow and high initial crash risk values are linked
to risk dissipation. Furthermore, the study conducted sensitivity analysis on
different types of classifiers, prediction time intervalsss and adaptive TTC
thresholds. The highest AUC value for vehicle-group risk prediction surpassed
0.93. The findings provide valuable insights to researchers and practitioners
in understanding and prediction of vehicle-group safety, ultimately improving
active traffic safety management and operations of Connected and Autonomous
Vehicles. | [
"cs.LG",
"cs.CY"
] | false |
2402.12417 | 2024-02-19T08:27:53Z | Predicting trucking accidents with truck drivers 'safety climate
perception across companies: A transfer learning approach | [
"Kailai Sun",
"Tianxiang Lan",
"Say Hong Kam",
"Yang Miang Goh",
"Yueng-Hsiang Huang"
] | There is a rising interest in using artificial intelligence (AI)-powered
safety analytics to predict accidents in the trucking industry. Companies may
face the practical challenge, however, of not having enough data to develop
good safety analytics models. Although pretrained models may offer a solution
for such companies, existing safety research using transfer learning has mostly
focused on computer vision and natural language processing, rather than
accident analytics. To fill the above gap, we propose a pretrain-then-fine-tune
transfer learning approach to help any company leverage other companies' data
to develop AI models for a more accurate prediction of accident risk. We also
develop SafeNet, a deep neural network algorithm for classification tasks
suitable for accident prediction. Using the safety climate survey data from
seven trucking companies with different data sizes, we show that our proposed
approach results in better model performance compared to training the model
from scratch using only the target company's data. We also show that for the
transfer learning model to be effective, the pretrained model should be
developed with larger datasets from diverse sources. The trucking industry may,
thus, consider pooling safety analytics data from a wide range of companies to
develop pretrained models and share them within the industry for better
knowledge and resource transfer. The above contributions point to the promise
of advanced safety analytics to make the industry safer and more sustainable. | [
"cs.LG",
"cs.AI"
] | false |
2402.12435 | 2024-02-19T19:00:01Z | Emulating the interstellar medium chemistry with neural operators | [
"Lorenzo Branca",
"Andrea Pallottini"
] | Galaxy formation and evolution critically depend on understanding the complex
photo-chemical processes that govern the evolution and thermodynamics of the
InterStellar Medium (ISM). Computationally, solving chemistry is among the most
heavy tasks in cosmological and astrophysical simulations. The evolution of
such non-equilibrium photo-chemical network relies on implicit, precise,
computationally costly, ordinary differential equations (ODE) solvers. Here, we
aim at substituting such procedural solvers with fast, pre-trained, emulators
based on neural operators. We emulate a non-equilibrium chemical network up to
H$_2$ formation (9 species, 52 reactions) by adopting the DeepONet formalism,
i.e. by splitting the ODE solver operator that maps the initial conditions and
time evolution into a tensor product of two neural networks. We use
$\texttt{KROME}$ to generate a training set spanning $-2\leq
\log(n/\mathrm{cm}^{-3}) \leq 3.5$, $\log(20) \leq\log(T/\mathrm{K}) \leq 5.5$,
$-6 \leq \log(n_i/n) < 0$, and by adopting an incident radiation field
$\textbf{F}$ sampled in 10 energy bins with a continuity prior. We separately
train the solver for $T$ and each $n_i$ for $\simeq 4.34\,\rm GPUhrs$. Compared
with the reference solutions obtained by $\texttt{KROME}$ for single zone
models, the typical precision obtained is of order $10^{-2}$, i.e. the $10
\times$ better with a training that is $40 \times$ less costly with respect to
previous emulators which however considered only a fixed $\mathbf{F}$. The
present model achieves a speed-up of a factor of $128 \times$ with respect to
stiff ODE solvers. Our neural emulator represents a significant leap forward in
the modeling of ISM chemistry, offering a good balance of precision,
versatility, and computational efficiency. | [
"astro-ph.GA",
"cs.LG"
] | false |
2402.12465 | 2024-02-19T19:11:22Z | Neuro-mimetic Task-free Unsupervised Online Learning with Continual
Self-Organizing Maps | [
"Hitesh Vaidya",
"Travis Desell",
"Ankur Mali",
"Alexander Ororbia"
] | An intelligent system capable of continual learning is one that can process
and extract knowledge from potentially infinitely long streams of pattern
vectors. The major challenge that makes crafting such a system difficult is
known as catastrophic forgetting - an agent, such as one based on artificial
neural networks (ANNs), struggles to retain previously acquired knowledge when
learning from new samples. Furthermore, ensuring that knowledge is preserved
for previous tasks becomes more challenging when input is not supplemented with
task boundary information. Although forgetting in the context of ANNs has been
studied extensively, there still exists far less work investigating it in terms
of unsupervised architectures such as the venerable self-organizing map (SOM),
a neural model often used in clustering and dimensionality reduction. While the
internal mechanisms of SOMs could, in principle, yield sparse representations
that improve memory retention, we observe that, when a fixed-size SOM processes
continuous data streams, it experiences concept drift. In light of this, we
propose a generalization of the SOM, the continual SOM (CSOM), which is capable
of online unsupervised learning under a low memory budget. Our results, on
benchmarks including MNIST, Kuzushiji-MNIST, and Fashion-MNIST, show almost a
two times increase in accuracy, and CIFAR-10 demonstrates a state-of-the-art
result when tested on (online) unsupervised class incremental learning setting. | [
"cs.LG",
"cs.NE"
] | false |
2402.12479 | 2024-02-19T19:34:07Z | In deep reinforcement learning, a pruned network is a good network | [
"Johan Obando-Ceron",
"Aaron Courville",
"Pablo Samuel Castro"
] | Recent work has shown that deep reinforcement learning agents have difficulty
in effectively using their network parameters. We leverage prior insights into
the advantages of sparse training techniques and demonstrate that gradual
magnitude pruning enables agents to maximize parameter effectiveness. This
results in networks that yield dramatic performance improvements over
traditional networks and exhibit a type of "scaling law", using only a small
fraction of the full network parameters. | [
"cs.LG",
"cs.AI"
] | true |
2402.12508 | 2024-02-19T20:18:29Z | SDEs for Minimax Optimization | [
"Enea Monzio Compagnoni",
"Antonio Orvieto",
"Hans Kersting",
"Frank Norbert Proske",
"Aurelien Lucchi"
] | Minimax optimization problems have attracted a lot of attention over the past
few years, with applications ranging from economics to machine learning. While
advanced optimization methods exist for such problems, characterizing their
dynamics in stochastic scenarios remains notably challenging. In this paper, we
pioneer the use of stochastic differential equations (SDEs) to analyze and
compare Minimax optimizers. Our SDE models for Stochastic Gradient
Descent-Ascent, Stochastic Extragradient, and Stochastic Hamiltonian Gradient
Descent are provable approximations of their algorithmic counterparts, clearly
showcasing the interplay between hyperparameters, implicit regularization, and
implicit curvature-induced noise. This perspective also allows for a unified
and simplified analysis strategy based on the principles of It\^o calculus.
Finally, our approach facilitates the derivation of convergence conditions and
closed-form solutions for the dynamics in simplified settings, unveiling
further insights into the behavior of different optimizers. | [
"cs.LG",
"math.OC"
] | false |
2402.12527 | 2024-02-19T20:38:00Z | The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning | [
"Anya Sims",
"Cong Lu",
"Yee Whye Teh"
] | Offline reinforcement learning aims to enable agents to be trained from
pre-collected datasets, however, this comes with the added challenge of
estimating the value of behavior not covered in the dataset. Model-based
methods offer a solution by allowing agents to collect additional synthetic
data via rollouts in a learned dynamics model. The prevailing theoretical
understanding is that this can then be viewed as online reinforcement learning
in an approximate dynamics model, and any remaining gap is therefore assumed to
be due to the imperfect dynamics model. Surprisingly, however, we find that if
the learned dynamics model is replaced by the true error-free dynamics,
existing model-based methods completely fail. This reveals a major
misconception. Our subsequent investigation finds that the general procedure
used in model-based algorithms results in the existence of a set of
edge-of-reach states which trigger pathological value overestimation and
collapse in Bellman-based algorithms. We term this the edge-of-reach problem.
Based on this, we fill some gaps in existing theory and also explain how prior
model-based methods are inadvertently addressing the true underlying
edge-of-reach problem. Finally, we propose Reach-Aware Value Learning (RAVL), a
simple and robust method that directly addresses the edge-of-reach problem and
achieves strong performance across both proprioceptive and pixel-based
benchmarks. Code open-sourced at: https://github.com/anyasims/edge-of-reach. | [
"cs.LG",
"cs.AI"
] | false |
2402.12535 | 2024-02-19T20:48:09Z | Locality-Sensitive Hashing-Based Efficient Point Transformer with
Applications in High-Energy Physics | [
"Siqi Miao",
"Zhiyuan Lu",
"Mia Liu",
"Javier Duarte",
"Pan Li"
] | This study introduces a novel transformer model optimized for large-scale
point cloud processing in scientific domains such as high-energy physics (HEP)
and astrophysics. Addressing the limitations of graph neural networks and
standard transformers, our model integrates local inductive bias and achieves
near-linear complexity with hardware-friendly regular operations. One
contribution of this work is the quantitative analysis of the error-complexity
tradeoff of various sparsification techniques for building efficient
transformers. Our findings highlight the superiority of using
locality-sensitive hashing (LSH), especially OR \& AND-construction LSH, in
kernel approximation for large-scale point cloud data with local inductive
bias. Based on this finding, we propose LSH-based Efficient Point Transformer
(\textbf{HEPT}), which combines E$^2$LSH with OR \& AND constructions and is
built upon regular computations. HEPT demonstrates remarkable performance in
two critical yet time-consuming HEP tasks, significantly outperforming existing
GNNs and transformers in accuracy and computational speed, marking a
significant advancement in geometric deep learning and large-scale scientific
data processing. Our code is available at
\url{https://github.com/Graph-COM/HEPT}. | [
"cs.LG",
"hep-ex"
] | false |
2402.12538 | 2024-02-19T20:55:12Z | A Machine Learning Ensemble Model for the Detection of Cyberbullying | [
"Abulkarim Faraj Alqahtani",
"Mohammad Ilyas"
] | The pervasive use of social media platforms, such as Facebook, Instagram, and
X, has significantly amplified our electronic interconnectedness. Moreover,
these platforms are now easily accessible from any location at any given time.
However, the increased popularity of social media has also led to
cyberbullying.It is imperative to address the need for finding, monitoring, and
mitigating cyberbullying posts on social media platforms. Motivated by this
necessity, we present this paper to contribute to developing an automated
system for detecting binary labels of aggressive tweets.Our study has
demonstrated remarkable performance compared to previous experiments on the
same dataset. We employed the stacking ensemble machine learning method,
utilizing four various feature extraction techniques to optimize performance
within the stacking ensemble learning framework. Combining five machine
learning algorithms,Decision Trees, Random Forest, Linear Support Vector
Classification, Logistic Regression, and K-Nearest Neighbors into an ensemble
method, we achieved superior results compared to traditional machine learning
classifier models. The stacking classifier achieved a high accuracy rate of
94.00%, outperforming traditional machine learning models and surpassing the
results of prior experiments that utilized the same dataset. The outcomes of
our experiments showcased an accuracy rate of 0.94% in detection tweets as
aggressive or non-aggressive. | [
"cs.SI",
"cs.LG"
] | false |
2402.12558 | 2024-02-19T21:32:56Z | Evaluation of Country Dietary Habits Using Machine Learning Techniques
in Relation to Deaths from COVID-19 | [
"María Teresa García-Ordás",
"Natalia Arias",
"Carmen Benavides",
"Oscar García-Olalla",
"José Alberto Benítez-Andrades"
] | COVID-19 disease has affected almost every country in the world. The large
number of infected people and the different mortality rates between countries
has given rise to many hypotheses about the key points that make the virus so
lethal in some places. In this study, the eating habits of 170 countries were
evaluated in order to find correlations between these habits and mortality
rates caused by COVID-19 using machine learning techniques that group the
countries together according to the different distribution of fat, energy, and
protein across 23 different types of food, as well as the amount ingested in
kilograms. Results shown how obesity and the high consumption of fats appear in
countries with the highest death rates, whereas countries with a lower rate
have a higher level of cereal consumption accompanied by a lower total average
intake of kilocalories. | [
"cs.LG",
"q-bio.QM"
] | false |
2402.12562 | 2024-02-19T21:36:54Z | Dynamic Pricing and Learning with Long-term Reference Effects | [
"Shipra Agrawal",
"Wei Tang"
] | We consider a dynamic pricing problem where customer response to the current
price is impacted by the customer price expectation, aka reference price. We
study a simple and novel reference price mechanism where reference price is the
average of the past prices offered by the seller. As opposed to the more
commonly studied exponential smoothing mechanism, in our reference price
mechanism the prices offered by seller have a longer term effect on the future
customer expectations.
We show that under this mechanism, a markdown policy is near-optimal
irrespective of the parameters of the model. This matches the common intuition
that a seller may be better off by starting with a higher price and then
decreasing it, as the customers feel like they are getting bargains on items
that are ordinarily more expensive. For linear demand models, we also provide a
detailed characterization of the near-optimal markdown policy along with an
efficient way of computing it.
We then consider a more challenging dynamic pricing and learning problem,
where the demand model parameters are apriori unknown, and the seller needs to
learn them online from the customers' responses to the offered prices while
simultaneously optimizing revenue. The objective is to minimize regret, i.e.,
the $T$-round revenue loss compared to a clairvoyant optimal policy. This task
essentially amounts to learning a non-stationary optimal policy in a
time-variant Markov Decision Process (MDP). For linear demand models, we
provide an efficient learning algorithm with an optimal $\tilde{O}(\sqrt{T})$
regret upper bound. | [
"cs.LG",
"cs.GT"
] | false |
2402.12570 | 2024-02-19T21:52:44Z | Offline Multi-task Transfer RL with Representational Penalization | [
"Avinandan Bose",
"Simon Shaolei Du",
"Maryam Fazel"
] | We study the problem of representation transfer in offline Reinforcement
Learning (RL), where a learner has access to episodic data from a number of
source tasks collected a priori, and aims to learn a shared representation to
be used in finding a good policy for a target task. Unlike in online RL where
the agent interacts with the environment while learning a policy, in the
offline setting there cannot be such interactions in either the source tasks or
the target task; thus multi-task offline RL can suffer from incomplete
coverage.
We propose an algorithm to compute pointwise uncertainty measures for the
learnt representation, and establish a data-dependent upper bound for the
suboptimality of the learnt policy for the target task. Our algorithm leverages
the collective exploration done by source tasks to mitigate poor coverage at
some points by a few tasks, thus overcoming the limitation of needing uniformly
good coverage for a meaningful transfer by existing offline algorithms. We
complement our theoretical results with empirical evaluation on a
rich-observation MDP which requires many samples for complete coverage. Our
findings illustrate the benefits of penalizing and quantifying the uncertainty
in the learnt representation. | [
"cs.LG",
"cs.AI"
] | false |
2402.12595 | 2024-02-19T23:19:15Z | Truncated Polynomial Expansion-Based Detection in Massive MIMO: A
Model-Driven Deep Learning Approach | [
"Kazem Izadinasab",
"Ahmed Wagdy Shaban",
"Oussama Damen"
] | In this paper, we propose a deep learning (DL)-based approach for efficiently
computing the inverse of Hermitian matrices using truncated polynomial
expansion (TPE). Our model-driven approach involves optimizing the coefficients
of the TPE during an offline training procedure for a given number of TPE
terms. We apply this method to signal detection in uplink massive
multiple-input multiple-output (MIMO) systems, where the matrix inverse
operation required by linear detectors, such as zero-forcing (ZF) and minimum
mean square error (MMSE), is approximated using TPE. Our simulation results
demonstrate that the proposed learned TPE-based method outperforms the
conventional TPE method with optimal coefficients in terms of asymptotic
convergence speed and reduces the computational complexity of the online
detection stage, albeit at the expense of the offline training stage. However,
the limited number of trainable parameters leads to a swift offline training
process. | [
"eess.SP",
"cs.LG"
] | false |
2402.12598 | 2024-02-19T23:22:30Z | Graph-based Virtual Sensing from Sparse and Partial Multivariate
Observations | [
"Giovanni De Felice",
"Andrea Cini",
"Daniele Zambon",
"Vladimir V. Gusev",
"Cesare Alippi"
] | Virtual sensing techniques allow for inferring signals at new unmonitored
locations by exploiting spatio-temporal measurements coming from physical
sensors at different locations. However, as the sensor coverage becomes sparse
due to costs or other constraints, physical proximity cannot be used to support
interpolation. In this paper, we overcome this challenge by leveraging
dependencies between the target variable and a set of correlated variables
(covariates) that can frequently be associated with each location of interest.
From this viewpoint, covariates provide partial observability, and the problem
consists of inferring values for unobserved channels by exploiting observations
at other locations to learn how such variables can correlate. We introduce a
novel graph-based methodology to exploit such relationships and design a graph
deep learning architecture, named GgNet, implementing the framework. The
proposed approach relies on propagating information over a nested graph
structure that is used to learn dependencies between variables as well as
locations. GgNet is extensively evaluated under different virtual sensing
scenarios, demonstrating higher reconstruction accuracy compared to the
state-of-the-art. | [
"cs.LG",
"cs.AI"
] | false |
2402.13285 | 2024-02-19T10:15:11Z | Leveraging PAC-Bayes Theory and Gibbs Distributions for Generalization
Bounds with Complexity Measures | [
"Paul Viallard",
"Rémi Emonet",
"Amaury Habrard",
"Emilie Morvant",
"Valentina Zantedeschi"
] | In statistical learning theory, a generalization bound usually involves a
complexity measure imposed by the considered theoretical framework. This limits
the scope of such bounds, as other forms of capacity measures or
regularizations are used in algorithms. In this paper, we leverage the
framework of disintegrated PAC-Bayes bounds to derive a general generalization
bound instantiable with arbitrary complexity measures. One trick to prove such
a result involves considering a commonly used family of distributions: the
Gibbs distributions. Our bound stands in probability jointly over the
hypothesis and the learning sample, which allows the complexity to be adapted
to the generalization gap as it can be customized to fit both the hypothesis
class and the task. | [
"stat.ML",
"cs.LG"
] | false |
2402.14844 | 2024-02-19T12:48:02Z | The New Era of Dynamic Pricing: Synergizing Supervised Learning and
Quadratic Programming | [
"Gustavo Bramao",
"Ilia Tarygin"
] | In this paper, we explore a novel combination of supervised learning and
quadratic programming to refine dynamic pricing models in the car rental
industry. We utilize dynamic modeling of price elasticity, informed by ordinary
least squares (OLS) metrics such as p-values, homoscedasticity, error
normality. These metrics, when their underlying assumptions hold, are integral
in guiding a quadratic programming agent. The program is tasked with optimizing
margin for a given finite set target. | [
"math.OC",
"cs.LG"
] | false |
2402.15524 | 2024-02-19T20:03:45Z | Graph Pruning for Enumeration of Minimal Unsatisfiable Subsets | [
"Panagiotis Lymperopoulos",
"Liping Liu"
] | Finding Minimal Unsatisfiable Subsets (MUSes) of binary constraints is a
common problem in infeasibility analysis of over-constrained systems. However,
because of the exponential search space of the problem, enumerating MUSes is
extremely time-consuming in real applications. In this work, we propose to
prune formulas using a learned model to speed up MUS enumeration. We represent
formulas as graphs and then develop a graph-based learning model to predict
which part of the formula should be pruned. Importantly, our algorithm does not
require data labeling by only checking the satisfiability of pruned formulas.
It does not even require training data from the target application because it
extrapolates to data with different distributions. In our experiments we
combine our algorithm with existing MUS enumerators and validate its
effectiveness in multiple benchmarks including a set of real-world problems
outside our training distribution. The experiment results show that our method
significantly accelerates MUS enumeration on average on these benchmark
problems. | [
"cs.AI",
"cs.LG"
] | false |
2403.00785 | 2024-02-19T02:43:55Z | Applying News and Media Sentiment Analysis for Generating Forex Trading
Signals | [
"Oluwafemi F Olaiyapo"
] | The objective of this research is to examine how sentiment analysis can be
employed to generate trading signals for the Foreign Exchange (Forex) market.
The author assessed sentiment in social media posts and news articles
pertaining to the United States Dollar (USD) using a combination of methods:
lexicon-based analysis and the Naive Bayes machine learning algorithm. The
findings indicate that sentiment analysis proves valuable in forecasting market
movements and devising trading signals. Notably, its effectiveness is
consistent across different market conditions. The author concludes that by
analyzing sentiment expressed in news and social media, traders can glean
insights into prevailing market sentiments towards the USD and other pertinent
countries, thereby aiding trading decision-making. This study underscores the
importance of weaving sentiment analysis into trading strategies as a pivotal
tool for predicting market dynamics. | [
"q-fin.ST",
"cs.LG"
] | false |
2402.11747 | 2024-02-19T00:21:07Z | Parameter Efficient Finetuning for Speech Emotion Recognition and Domain
Adaptation | [
"Nineli Lashkarashvili",
"Wen Wu",
"Guangzhi Sun",
"Philip C. Woodland"
] | Foundation models have shown superior performance for speech emotion
recognition (SER). However, given the limited data in emotion corpora,
finetuning all parameters of large pre-trained models for SER can be both
resource-intensive and susceptible to overfitting. This paper investigates
parameter-efficient finetuning (PEFT) for SER. Various PEFT adaptors are
systematically studied for both classification of discrete emotion categories
and prediction of dimensional emotional attributes. The results demonstrate
that the combination of PEFT methods surpasses full finetuning with a
significant reduction in the number of trainable parameters. Furthermore, a
two-stage adaptation strategy is proposed to adapt models trained on acted
emotion data, which is more readily available, to make the model more adept at
capturing natural emotional expressions. Both intra- and cross-corpus
experiments validate the efficacy of the proposed approach in enhancing the
performance on both the source and target domains. | [
"eess.AS",
"cs.LG",
"cs.SD"
] | false |
2402.11771 | 2024-02-19T01:55:55Z | Evaluating the Effectiveness of Index-Based Treatment Allocation | [
"Niclas Boehmer",
"Yash Nair",
"Sanket Shah",
"Lucas Janson",
"Aparna Taneja",
"Milind Tambe"
] | When resources are scarce, an allocation policy is needed to decide who
receives a resource. This problem occurs, for instance, when allocating scarce
medical resources and is often solved using modern ML methods. This paper
introduces methods to evaluate index-based allocation policies -- that allocate
a fixed number of resources to those who need them the most -- by using data
from a randomized control trial. Such policies create dependencies between
agents, which render the assumptions behind standard statistical tests invalid
and limit the effectiveness of estimators. Addressing these challenges, we
translate and extend recent ideas from the statistics literature to present an
efficient estimator and methods for computing asymptotically correct confidence
intervals. This enables us to effectively draw valid statistical conclusions, a
critical gap in previous work. Our extensive experiments validate our
methodology in practical settings, while also showcasing its statistical power.
We conclude by proposing and empirically verifying extensions of our
methodology that enable us to reevaluate a past randomized control trial to
evaluate different ML allocation policies in the context of a mHealth program,
drawing previously invisible conclusions. | [
"cs.LG",
"cs.AI",
"stat.ME",
"stat.ML"
] | false |
2402.11835 | 2024-02-19T04:58:39Z | Easy as ABCs: Unifying Boltzmann Q-Learning and Counterfactual Regret
Minimization | [
"Luca D'Amico-Wong",
"Hugh Zhang",
"Marc Lanctot",
"David C. Parkes"
] | We propose ABCs (Adaptive Branching through Child stationarity), a
best-of-both-worlds algorithm combining Boltzmann Q-learning (BQL), a classic
reinforcement learning algorithm for single-agent domains, and counterfactual
regret minimization (CFR), a central algorithm for learning in multi-agent
domains. ABCs adaptively chooses what fraction of the environment to explore
each iteration by measuring the stationarity of the environment's reward and
transition dynamics. In Markov decision processes, ABCs converges to the
optimal policy with at most an O(A) factor slowdown compared to BQL, where A is
the number of actions in the environment. In two-player zero-sum games, ABCs is
guaranteed to converge to a Nash equilibrium (assuming access to a perfect
oracle for detecting stationarity), while BQL has no such guarantees.
Empirically, ABCs demonstrates strong performance when benchmarked across
environments drawn from the OpenSpiel game library and OpenAI Gym and exceeds
all prior methods in environments which are neither fully stationary nor fully
nonstationary. | [
"cs.LG",
"cs.GT",
"cs.MA"
] | false |
2402.11925 | 2024-02-19T08:12:47Z | Energy-Efficient Edge Learning via Joint Data Deepening-and-Prefetching | [
"Sujin Kook",
"Won-Yong Shin",
"Seong-Lyun Kim",
"Seung-Woo Ko"
] | The vision of pervasive artificial intelligence (AI) services can be realized
by training an AI model on time using real-time data collected by internet of
things (IoT) devices. To this end, IoT devices require offloading their data to
an edge server in proximity. However, transmitting high-dimensional and
voluminous data from energy-constrained IoT devices poses a significant
challenge. To address this limitation, we propose a novel offloading
architecture, called joint data deepening-and-prefetching (JD2P), which is
feature-by-feature offloading comprising two key techniques. The first one is
data deepening, where each data sample's features are sequentially offloaded in
the order of importance determined by the data embedding technique such as
principle component analysis (PCA). Offloading is terminated once the already
transmitted features are sufficient for accurate data classification, resulting
in a reduction in the amount of transmitted data. The criteria to offload data
are derived for binary and multi-class classifiers, which are designed based on
support vector machine (SVM) and deep neural network (DNN), respectively. The
second one is data prefetching, where some features potentially required in the
future are offloaded in advance, thus achieving high efficiency via precise
prediction and parameter optimization. We evaluate the effectiveness of JD2P
through experiments using the MNIST dataset, and the results demonstrate its
significant reduction in expected energy consumption compared to several
benchmarks without degrading learning accuracy. | [
"cs.LG",
"cs.AI",
"cs.IT",
"math.IT"
] | false |
2402.11948 | 2024-02-19T08:43:00Z | Mini-Hes: A Parallelizable Second-order Latent Factor Analysis Model | [
"Jialiang Wang",
"Weiling Li",
"Yurong Zhong",
"Xin Luo"
] | Interactions among large number of entities is naturally high-dimensional and
incomplete (HDI) in many big data related tasks. Behavioral characteristics of
users are hidden in these interactions, hence, effective representation of the
HDI data is a fundamental task for understanding user behaviors. Latent factor
analysis (LFA) model has proven to be effective in representing HDI data. The
performance of an LFA model relies heavily on its training process, which is a
non-convex optimization. It has been proven that incorporating local curvature
and preprocessing gradients during its training process can lead to superior
performance compared to LFA models built with first-order family methods.
However, with the escalation of data volume, the feasibility of second-order
algorithms encounters challenges. To address this pivotal issue, this paper
proposes a mini-block diagonal hessian-free (Mini-Hes) optimization for
building an LFA model. It leverages the dominant diagonal blocks in the
generalized Gauss-Newton matrix based on the analysis of the Hessian matrix of
LFA model and serves as an intermediary strategy bridging the gap between
first-order and second-order optimization methods. Experiment results indicate
that, with Mini-Hes, the LFA model outperforms several state-of-the-art models
in addressing missing data estimation task on multiple real HDI datasets from
recommender system. (The source code of Mini-Hes is available at
https://github.com/Goallow/Mini-Hes) | [
"cs.LG",
"cs.AI",
"stat.ML"
] | false |
2402.11984 | 2024-02-19T09:29:37Z | Hebbian Learning based Orthogonal Projection for Continual Learning of
Spiking Neural Networks | [
"Mingqing Xiao",
"Qingyan Meng",
"Zongpeng Zhang",
"Di He",
"Zhouchen Lin"
] | Neuromorphic computing with spiking neural networks is promising for
energy-efficient artificial intelligence (AI) applications. However, different
from humans who continually learn different tasks in a lifetime, neural network
models suffer from catastrophic forgetting. How could neuronal operations solve
this problem is an important question for AI and neuroscience. Many previous
studies draw inspiration from observed neuroscience phenomena and propose
episodic replay or synaptic metaplasticity, but they are not guaranteed to
explicitly preserve knowledge for neuron populations. Other works focus on
machine learning methods with more mathematical grounding, e.g., orthogonal
projection on high dimensional spaces, but there is no neural correspondence
for neuromorphic computing. In this work, we develop a new method with neuronal
operations based on lateral connections and Hebbian learning, which can protect
knowledge by projecting activity traces of neurons into an orthogonal subspace
so that synaptic weight update will not interfere with old tasks. We show that
Hebbian and anti-Hebbian learning on recurrent lateral connections can
effectively extract the principal subspace of neural activities and enable
orthogonal projection. This provides new insights into how neural circuits and
Hebbian learning can help continual learning, and also how the concept of
orthogonal projection can be realized in neuronal systems. Our method is also
flexible to utilize arbitrary training methods based on presynaptic
activities/traces. Experiments show that our method consistently solves
forgetting for spiking neural networks with nearly zero forgetting under
various supervised training methods with different error propagation
approaches, and outperforms previous approaches under various settings. Our
method can pave a solid path for building continual neuromorphic computing
systems. | [
"cs.NE",
"cs.AI",
"cs.LG"
] | false |
2402.12008 | 2024-02-19T10:02:00Z | Cluster Metric Sensitivity to Irrelevant Features | [
"Miles McCrory",
"Spencer A. Thomas"
] | Clustering algorithms are used extensively in data analysis for data
exploration and discovery. Technological advancements lead to continually
growth of data in terms of volume, dimensionality and complexity. This provides
great opportunities in data analytics as the data can be interrogated for many
different purposes. This however leads challenges, such as identification of
relevant features for a given task. In supervised tasks, one can utilise a
number of methods to optimise the input features for the task objective (e.g.
classification accuracy). In unsupervised problems, such tools are not readily
available, in part due to an inability to quantify feature relevance in
unlabeled tasks. In this paper, we investigate the sensitivity of clustering
performance noisy uncorrelated variables iteratively added to baseline datasets
with well defined clusters. We show how different types of irrelevant variables
can impact the outcome of a clustering result from $k$-means in different ways.
We observe a resilience to very high proportions of irrelevant features for
adjusted rand index (ARI) and normalised mutual information (NMI) when the
irrelevant features are Gaussian distributed. For Uniformly distributed
irrelevant features, we notice the resilience of ARI and NMI is dependent on
the dimensionality of the data and exhibits tipping points between high scores
and near zero. Our results show that the Silhouette Coefficient and the
Davies-Bouldin score are the most sensitive to irrelevant added features
exhibiting large changes in score for comparably low proportions of irrelevant
features regardless of underlying distribution or data scaling. As such the
Silhouette Coefficient and the Davies-Bouldin score are good candidates for
optimising feature selection in unsupervised clustering tasks. | [
"cs.LG",
"cs.AI",
"stat.ML"
] | false |
2402.12010 | 2024-02-19T10:03:46Z | Training Green AI Models Using Elite Samples | [
"Mohammed Alswaitti",
"Roberto Verdecchia",
"Grégoire Danoy",
"Pascal Bouvry",
"Johnatan Pecero"
] | The substantial increase in AI model training has considerable environmental
implications, mandating more energy-efficient and sustainable AI practices. On
the one hand, data-centric approaches show great potential towards training
energy-efficient AI models. On the other hand, instance selection methods
demonstrate the capability of training AI models with minimised training sets
and negligible performance degradation. Despite the growing interest in both
topics, the impact of data-centric training set selection on energy efficiency
remains to date unexplored. This paper presents an evolutionary-based sampling
framework aimed at (i) identifying elite training samples tailored for datasets
and model pairs, (ii) comparing model performance and energy efficiency gains
against typical model training practice, and (iii) investigating the
feasibility of this framework for fostering sustainable model training
practices. To evaluate the proposed framework, we conducted an empirical
experiment including 8 commonly used AI classification models and 25 publicly
available datasets. The results showcase that by considering 10% elite training
samples, the models' performance can show a 50% improvement and remarkable
energy savings of 98% compared to the common training practice. | [
"cs.LG",
"cs.AI",
"cs.NE"
] | false |
2402.12015 | 2024-02-19T10:13:25Z | An Index Policy Based on Sarsa and Q-learning for Heterogeneous Smart
Target Tracking | [
"Yuhang Hao",
"Zengfu Wang",
"Jing Fu",
"Quan Pan"
] | In solving the non-myopic radar scheduling for multiple smart target tracking
within an active and passive radar network, we need to consider both short-term
enhanced tracking performance and a higher probability of target maneuvering in
the future with active tracking. Acquiring the long-term tracking performance
while scheduling the beam resources of active and passive radars poses a
challenge. To address this challenge, we model this problem as a Markov
decision process consisting of parallel restless bandit processes. Each bandit
process is associated with a smart target, of which the estimation state
evolves according to different discrete dynamic models for different actions -
whether or not the target is being tracked. The discrete state is defined by
the dynamic mode. The problem exhibits the curse of dimensionality, where
optimal solutions are in general intractable. We resort to heuristics through
the famous restless multi-armed bandit techniques. It follows with efficient
scheduling policies based on the indices that are real numbers representing the
marginal rewards of taking different actions. For the inevitable practical case
with unknown transition matrices, we propose a new method that utilizes the
forward Sarsa and backward Q-learning to approximate the indices through
adapting the state-action value functions, or equivalently the Q-functions, and
propose a new policy, namely ISQ, aiming to maximize the long-term tracking
rewards. Numerical results demonstrate that the proposed ISQ policy outperforms
conventional Q-learning-based methods and rapidly converges to the well-known
Whittle index policy with revealed state transition models, which is considered
the benchmark. | [
"eess.SY",
"cs.LG",
"cs.SY"
] | false |
2402.12042 | 2024-02-19T10:56:47Z | Linear bandits with polylogarithmic minimax regret | [
"Josep Lumbreras",
"Marco Tomamichel"
] | We study a noise model for linear stochastic bandits for which the
subgaussian noise parameter vanishes linearly as we select actions on the unit
sphere closer and closer to the unknown vector. We introduce an algorithm for
this problem that exhibits a minimax regret scaling as $\log^3(T)$ in the time
horizon $T$, in stark contrast the square root scaling of this regret for
typical bandit algorithms. Our strategy, based on weighted least-squares
estimation, achieves the eigenvalue relation $\lambda_{\min} ( V_t ) = \Omega
(\sqrt{\lambda_{\max}(V_t ) })$ for the design matrix $V_t$ at each time step
$t$ through geometrical arguments that are independent of the noise model and
might be of independent interest. This allows us to tightly control the
expected regret in each time step to be of the order $O(\frac1{t})$, leading to
the logarithmic scaling of the cumulative regret. | [
"cs.LG",
"cs.AI",
"stat.ML"
] | false |
2402.12067 | 2024-02-19T11:35:01Z | Interpretable Brain-Inspired Representations Improve RL Performance on
Visual Navigation Tasks | [
"Moritz Lange",
"Raphael C. Engelhardt",
"Wolfgang Konen",
"Laurenz Wiskott"
] | Visual navigation requires a whole range of capabilities. A crucial one of
these is the ability of an agent to determine its own location and heading in
an environment. Prior works commonly assume this information as given, or use
methods which lack a suitable inductive bias and accumulate error over time. In
this work, we show how the method of slow feature analysis (SFA), inspired by
neuroscience research, overcomes both limitations by generating interpretable
representations of visual data that encode location and heading of an agent. We
employ SFA in a modern reinforcement learning context, analyse and compare
representations and illustrate where hierarchical SFA can outperform other
feature extractors on navigation tasks. | [
"cs.LG",
"cs.NE",
"cs.RO"
] | false |
2402.12232 | 2024-02-19T15:39:39Z | Kernel KMeans clustering splits for end-to-end unsupervised decision
trees | [
"Louis Ohl",
"Pierre-Alexandre Mattei",
"Mickaël Leclercq",
"Arnaud Droit",
"Frédéric Precioso"
] | Trees are convenient models for obtaining explainable predictions on
relatively small datasets. Although there are many proposals for the end-to-end
construction of such trees in supervised learning, learning a tree end-to-end
for clustering without labels remains an open challenge. As most works focus on
interpreting with trees the result of another clustering algorithm, we present
here a novel end-to-end trained unsupervised binary tree for clustering: Kauri.
This method performs a greedy maximisation of the kernel KMeans objective
without requiring the definition of centroids. We compare this model on
multiple datasets with recent unsupervised trees and show that Kauri performs
identically when using a linear kernel. For other kernels, Kauri often
outperforms the concatenation of kernel KMeans and a CART decision tree. | [
"stat.ML",
"cs.AI",
"cs.LG",
"62h30",
"G.3"
] | false |
2402.12241 | 2024-02-19T15:56:43Z | Convergence of Gradient Descent for Recurrent Neural Networks: A
Nonasymptotic Analysis | [
"Semih Cayci",
"Atilla Eryilmaz"
] | We analyze recurrent neural networks trained with gradient descent in the
supervised learning setting for dynamical systems, and prove that gradient
descent can achieve optimality \emph{without} massive overparameterization. Our
in-depth nonasymptotic analysis (i) provides sharp bounds on the network size
$m$ and iteration complexity $\tau$ in terms of the sequence length $T$, sample
size $n$ and ambient dimension $d$, and (ii) identifies the significant impact
of long-term dependencies in the dynamical system on the convergence and
network width bounds characterized by a cutoff point that depends on the
Lipschitz continuity of the activation function. Remarkably, this analysis
reveals that an appropriately-initialized recurrent neural network trained with
$n$ samples can achieve optimality with a network size $m$ that scales only
logarithmically with $n$. This sharply contrasts with the prior works that
require high-order polynomial dependency of $m$ on $n$ to establish strong
regularity conditions. Our results are based on an explicit characterization of
the class of dynamical systems that can be approximated and learned by
recurrent neural networks via norm-constrained transportation mappings, and
establishing local smoothness properties of the hidden state with respect to
the learnable parameters. | [
"cs.LG",
"math.OC",
"stat.ML"
] | false |
2402.12265 | 2024-02-19T16:26:40Z | On the Byzantine-Resilience of Distillation-Based Federated Learning | [
"Christophe Roux",
"Max Zimmer",
"Sebastian Pokutta"
] | Federated Learning (FL) algorithms using Knowledge Distillation (KD) have
received increasing attention due to their favorable properties with respect to
privacy, non-i.i.d. data and communication cost. These methods depart from
transmitting model parameters and, instead, communicate information about a
learning task by sharing predictions on a public dataset. In this work, we
study the performance of such approaches in the byzantine setting, where a
subset of the clients act in an adversarial manner aiming to disrupt the
learning process. We show that KD-based FL algorithms are remarkably resilient
and analyze how byzantine clients can influence the learning process compared
to Federated Averaging. Based on these insights, we introduce two new byzantine
attacks and demonstrate that they are effective against prior
byzantine-resilient methods. Additionally, we propose FilterExp, a novel method
designed to enhance the byzantine resilience of KD-based FL algorithms and
demonstrate its efficacy. Finally, we provide a general method to make attacks
harder to detect, improving their effectiveness. | [
"cs.LG",
"cs.AI",
"cs.DC"
] | false |
2402.12302 | 2024-02-19T17:25:12Z | Asymptotic Gaussian Fluctuations of Eigenvectors in Spectral Clustering | [
"Hugo Lebeau",
"Florent Chatelain",
"Romain Couillet"
] | The performance of spectral clustering relies on the fluctuations of the
entries of the eigenvectors of a similarity matrix, which has been left
uncharacterized until now. In this letter, it is shown that the signal $+$
noise structure of a general spike random matrix model is transferred to the
eigenvectors of the corresponding Gram kernel matrix and the fluctuations of
their entries are Gaussian in the large-dimensional regime. This CLT-like
result was the last missing piece to precisely predict the classification
performance of spectral clustering. The proposed proof is very general and
relies solely on the rotational invariance of the noise. Numerical experiments
on synthetic and real data illustrate the universality of this phenomenon. | [
"stat.ML",
"cs.LG",
"math.PR"
] | false |
2402.12319 | 2024-02-19T17:44:35Z | Dynamic Environment Responsive Online Meta-Learning with Fairness
Awareness | [
"Chen Zhao",
"Feng Mi",
"Xintao Wu",
"Kai Jiang",
"Latifur Khan",
"Feng Chen"
] | The fairness-aware online learning framework has emerged as a potent tool
within the context of continuous lifelong learning. In this scenario, the
learner's objective is to progressively acquire new tasks as they arrive over
time, while also guaranteeing statistical parity among various protected
sub-populations, such as race and gender, when it comes to the newly introduced
tasks. A significant limitation of current approaches lies in their heavy
reliance on the i.i.d (independent and identically distributed) assumption
concerning data, leading to a static regret analysis of the framework.
Nevertheless, it's crucial to note that achieving low static regret does not
necessarily translate to strong performance in dynamic environments
characterized by tasks sampled from diverse distributions. In this paper, to
tackle the fairness-aware online learning challenge in evolving settings, we
introduce a unique regret measure, FairSAR, by incorporating long-term fairness
constraints into a strongly adapted loss regret framework. Moreover, to
determine an optimal model parameter at each time step, we introduce an
innovative adaptive fairness-aware online meta-learning algorithm, referred to
as FairSAOML. This algorithm possesses the ability to adjust to dynamic
environments by effectively managing bias control and model accuracy. The
problem is framed as a bi-level convex-concave optimization, considering both
the model's primal and dual parameters, which pertain to its accuracy and
fairness attributes, respectively. Theoretical analysis yields sub-linear upper
bounds for both loss regret and the cumulative violation of fairness
constraints. Our experimental evaluation on various real-world datasets in
dynamic environments demonstrates that our proposed FairSAOML algorithm
consistently outperforms alternative approaches rooted in the most advanced
prior online learning methods. | [
"cs.LG",
"cs.AI",
"cs.CY"
] | false |
2402.12331 | 2024-02-19T18:02:10Z | Generating Survival Interpretable Trajectories and Data | [
"Andrei V. Konstantinov",
"Stanislav R. Kirpichenko",
"Lev V. Utkin"
] | A new model for generating survival trajectories and data based on applying
an autoencoder of a specific structure is proposed. It solves three tasks.
First, it provides predictions in the form of the expected event time and the
survival function for a new generated feature vector on the basis of the Beran
estimator. Second, the model generates additional data based on a given
training set that would supplement the original dataset. Third, the most
important, it generates a prototype time-dependent trajectory for an object,
which characterizes how features of the object could be changed to achieve a
different time to an event. The trajectory can be viewed as a type of the
counterfactual explanation. The proposed model is robust during training and
inference due to a specific weighting scheme incorporating into the variational
autoencoder. The model also determines the censored indicators of new generated
data by solving a classification task. The paper demonstrates the efficiency
and properties of the proposed model using numerical experiments on synthetic
and real datasets. The code of the algorithm implementing the proposed model is
publicly available. | [
"cs.LG",
"cs.AI",
"stat.ML"
] | false |
2402.12338 | 2024-02-19T18:11:37Z | An Adversarial Approach to Evaluating the Robustness of Event
Identification Models | [
"Obai Bahwal",
"Oliver Kosut",
"Lalitha Sankar"
] | Intelligent machine learning approaches are finding active use for event
detection and identification that allow real-time situational awareness. Yet,
such machine learning algorithms have been shown to be susceptible to
adversarial attacks on the incoming telemetry data. This paper considers a
physics-based modal decomposition method to extract features for event
classification and focuses on interpretable classifiers including logistic
regression and gradient boosting to distinguish two types of events: load loss
and generation loss. The resulting classifiers are then tested against an
adversarial algorithm to evaluate their robustness. The adversarial attack is
tested in two settings: the white box setting, wherein the attacker knows
exactly the classification model; and the gray box setting, wherein the
attacker has access to historical data from the same network as was used to
train the classifier, but does not know the classification model. Thorough
experiments on the synthetic South Carolina 500-bus system highlight that a
relatively simpler model such as logistic regression is more susceptible to
adversarial attacks than gradient boosting. | [
"eess.SY",
"cs.CR",
"cs.LG",
"cs.SY"
] | false |
2402.12365 | 2024-02-19T18:52:13Z | Universal Physics Transformers | [
"Benedikt Alkin",
"Andreas Fürst",
"Simon Schmid",
"Lukas Gruber",
"Markus Holzleitner",
"Johannes Brandstetter"
] | Deep neural network based surrogates for partial differential equations have
recently gained increased interest. However, akin to their numerical
counterparts, different techniques are used across applications, even if the
underlying dynamics of the systems are similar. A prominent example is the
Lagrangian and Eulerian specification in computational fluid dynamics, posing a
challenge for neural networks to effectively model particle- as opposed to
grid-based dynamics. We introduce Universal Physics Transformers (UPTs), a
novel learning paradigm which models a wide range of spatio-temporal problems -
both for Lagrangian and Eulerian discretization schemes. UPTs operate without
grid- or particle-based latent structures, enabling flexibility across meshes
and particles. UPTs efficiently propagate dynamics in the latent space,
emphasized by inverse encoding and decoding techniques. Finally, UPTs allow for
queries of the latent space representation at any point in space-time. We
demonstrate the efficacy of UPTs in mesh-based fluid simulations, steady-state
Reynolds averaged Navier-Stokes simulations, and Lagrangian-based dynamics.
Project page: https://ml-jku.github.io/UPT | [
"cs.LG",
"cs.AI",
"physics.flu-dyn"
] | false |
2402.12411 | 2024-02-19T02:34:23Z | Deep Structural Knowledge Exploitation and Synergy for Estimating Node
Importance Value on Heterogeneous Information Networks | [
"Yankai Chen",
"Yixiang Fang",
"Qiongyan Wang",
"Xin Cao",
"Irwin King"
] | Node importance estimation problem has been studied conventionally with
homogeneous network topology analysis. To deal with network heterogeneity, a
few recent methods employ graph neural models to automatically learn diverse
sources of information. However, the major concern revolves around that their
full adaptive learning process may lead to insufficient information
exploration, thereby formulating the problem as the isolated node value
prediction with underperformance and less interpretability. In this work, we
propose a novel learning framework: SKES. Different from previous automatic
learning designs, SKES exploits heterogeneous structural knowledge to enrich
the informativeness of node representations. Based on a sufficiently
uninformative reference, SKES estimates the importance value for any input
node, by quantifying its disparity against the reference. This establishes an
interpretable node importance computation paradigm. Furthermore, SKES dives
deep into the understanding that "nodes with similar characteristics are prone
to have similar importance values" whilst guaranteeing that such
informativeness disparity between any different nodes is orderly reflected by
the embedding distance of their associated latent features. Extensive
experiments on three widely-evaluated benchmarks demonstrate the performance
superiority of SKES over several recent competing methods. | [
"cs.SI",
"cs.AI",
"cs.LG"
] | false |
2402.12418 | 2024-02-19T09:52:45Z | Beyond Uniform Scaling: Exploring Depth Heterogeneity in Neural
Architectures | [
"Akash Guna R. T",
"Arnav Chavan",
"Deepak Gupta"
] | Conventional scaling of neural networks typically involves designing a base
network and growing different dimensions like width, depth, etc. of the same by
some predefined scaling factors. We introduce an automated scaling approach
leveraging second-order loss landscape information. Our method is flexible
towards skip connections a mainstay in modern vision transformers. Our
training-aware method jointly scales and trains transformers without additional
training iterations. Motivated by the hypothesis that not all neurons need
uniform depth complexity, our approach embraces depth heterogeneity. Extensive
evaluations on DeiT-S with ImageNet100 show a 2.5% accuracy gain and 10%
parameter efficiency improvement over conventional scaling. Scaled networks
demonstrate superior performance upon training small scale datasets from
scratch. We introduce the first intact scaling mechanism for vision
transformers, a step towards efficient model scaling. | [
"cs.LG",
"cs.AI",
"cs.NE"
] | false |
2402.12448 | 2024-02-19T19:00:09Z | DBNets: A publicly available deep learning tool to measure the masses of
young planets in dusty protoplanetary discs | [
"Alessandro Ruzza",
"Giuseppe Lodato",
"Giovanni Pietro Rosotti"
] | Current methods to characterize embedded planets in protoplanetary disc
observations are severely limited either in their ability to fully account for
the observed complex physics or in their computational and time costs. To
address this shortcoming, we developed DBNets: a deep learning tool, based on
convolutional neural networks, that analyses substructures observed in the dust
continuum emission of protoplanetary discs to quickly infer the mass of
allegedly embedded planets. We focussed on developing a method to reliably
quantify not only the planet mass, but also the associated uncertainty
introduced by our modelling and adopted techniques. Our tests gave promising
results achieving an 87% reduction of the log Mp mean squared error with
respect to an analytical formula fitted on the same data (DBNets metrics: lmse
0.016, r2-score 97%). With the goal of providing the final user of DBNets with
all the tools needed to interpret their measurements and decide on their
significance, we extensively tested our tool on out-of-distribution data. We
found that DBNets can identify inputs strongly outside its training scope
returning an uncertainty above a specific threshold and we thus provided a
rejection criterion that helps determine the significance of the results
obtained. Additionally, we outlined some limitations of our tool: it can be
reliably applied only on discs observed with inclinations below approximately
60{\deg}, in the optically thin regime, with a resolution 8 times better than
the gap radial location and with a signal-to-noise ratio higher than
approximately ten. Finally, we applied DBNets to 33 actual observations of
protoplanetary discs measuring the mass of 48 proposed planets and comparing
our results with the available literature. We confirmed that most of the
observed gaps imply planets in the sub-Jupiter regime. DBNets is publicly
available at dbnets.fisica.unimi.it. | [
"astro-ph.EP",
"astro-ph.IM",
"cs.LG"
] | false |
2402.12475 | 2024-02-19T19:21:45Z | Diffeomorphism Neural Operator for various domains and parameters of
partial differential equations | [
"Zhiwei Zhao",
"Changqing Liu",
"Yingguang Li",
"Zhibin Chen",
"Xu Liu"
] | Many science and engineering applications demand partial differential
equations (PDE) evaluations that are traditionally computed with
resource-intensive numerical solvers. Neural operator models provide an
efficient alternative by learning the governing physical laws directly from
data in a class of PDEs with different parameters, but constrained in a fixed
boundary (domain). Many applications, such as design and manufacturing, would
benefit from neural operators with flexible domains when studied at scale. Here
we present a diffeomorphism neural operator learning framework towards
developing domain-flexible models for physical systems with various and complex
domains. Specifically, a neural operator trained in a shared domain mapped from
various domains of fields by diffeomorphism is proposed, which transformed the
problem of learning function mappings in varying domains (spaces) into the
problem of learning operators on a shared diffeomorphic domain. Meanwhile, an
index is provided to evaluate the generalization of diffeomorphism neural
operators in different domains by the domain diffeomorphism similarity.
Experiments on statics scenarios (Darcy flow, mechanics) and dynamic scenarios
(pipe flow, airfoil flow) demonstrate the advantages of our approach for neural
operator learning under various domains, where harmonic and volume
parameterization are used as the diffeomorphism for 2D and 3D domains. Our
diffeomorphism neural operator approach enables strong learning capability and
robust generalization across varying domains and parameters. | [
"math.NA",
"cs.LG",
"cs.NA"
] | false |
2402.12482 | 2024-02-19T19:38:37Z | SECP: A Speech Enhancement-Based Curation Pipeline For Scalable
Acquisition Of Clean Speech | [
"Adam Sabra",
"Cyprian Wronka",
"Michelle Mao",
"Samer Hijazi"
] | As more speech technologies rely on a supervised deep learning approach with
clean speech as the ground truth, a methodology to onboard said speech at scale
is needed. However, this approach needs to minimize the dependency on human
listening and annotation, only requiring a human-in-the-loop when needed. In
this paper, we address this issue by outlining Speech Enhancement-based
Curation Pipeline (SECP) which serves as a framework to onboard clean speech.
This clean speech can then train a speech enhancement model, which can further
refine the original dataset and thus close the iterative loop. By running two
iterative rounds, we observe that enhanced output used as ground truth does not
degrade model performance according to $\Delta_{PESQ}$, a metric used in this
paper. We also show through comparative mean opinion score (CMOS) based
subjective tests that the highest and lowest bound of refined data is
perceptually better than the original data. | [
"cs.SD",
"cs.IR",
"cs.LG",
"eess.AS"
] | false |