arxiv_id
stringlengths 10
10
| published
stringlengths 20
20
| titles
stringlengths 9
243
| authors
sequencelengths 1
389
| abstract
stringlengths 96
3.09k
| categories
sequencelengths 1
10
| selected
bool 2
classes |
---|---|---|---|---|---|---|
2402.11167 | 2024-02-17T02:25:57Z | Token-Ensemble Text Generation: On Attacking the Automatic AI-Generated
Text Detection | [
"Fan Huang",
"Haewoon Kwak",
"Jisun An"
] | The robustness of AI-content detection models against cultivated attacks
(e.g., paraphrasing or word switching) remains a significant concern. This
study proposes a novel token-ensemble generation strategy to challenge the
robustness of current AI-content detection approaches. We explore the ensemble
attack strategy by completing the prompt with the next token generated from
random candidate LLMs. We find the token-ensemble approach significantly drops
the performance of AI-content detection models (The code and test sets will be
released). Our findings reveal that token-ensemble generation poses a vital
challenge to current detection models and underlines the need for advancing
detection technologies to counter sophisticated adversarial strategies. | [
"cs.CL",
"cs.AI"
] | false |
2402.11177 | 2024-02-17T02:55:35Z | A Question Answering Based Pipeline for Comprehensive Chinese EHR
Information Extraction | [
"Huaiyuan Ying",
"Sheng Yu"
] | Electronic health records (EHRs) hold significant value for research and
applications. As a new way of information extraction, question answering (QA)
can extract more flexible information than conventional methods and is more
accessible to clinical researchers, but its progress is impeded by the scarcity
of annotated data. In this paper, we propose a novel approach that
automatically generates training data for transfer learning of QA models. Our
pipeline incorporates a preprocessing module to handle challenges posed by
extraction types that are not readily compatible with extractive QA frameworks,
including cases with discontinuous answers and many-to-one relationships. The
obtained QA model exhibits excellent performance on subtasks of information
extraction in EHRs, and it can effectively handle few-shot or zero-shot
settings involving yes-no questions. Case studies and ablation studies
demonstrate the necessity of each component in our design, and the resulting
model is deemed suitable for practical use. | [
"cs.CL",
"cs.IR"
] | false |
2402.11187 | 2024-02-17T04:16:30Z | LaCo: Large Language Model Pruning via Layer Collapse | [
"Yifei Yang",
"Zouying Cao",
"Hai Zhao"
] | Large language models (LLMs) based on transformer are witnessing a notable
trend of size expansion, which brings considerable costs to both model training
and inference. However, existing methods such as model quantization, knowledge
distillation, and model pruning are constrained by various issues, including
hardware support limitations, the need for extensive training, and alterations
to the internal structure of the model. In this paper, we propose a concise
layer-wise pruning method called \textit{Layer Collapse (LaCo)}, in which rear
model layers collapse into a prior layer, enabling a rapid reduction in model
size while preserving the model structure. Comprehensive experiments show that
our method maintains an average task performance of over 80\% at pruning ratios
of 25-30\%, significantly outperforming existing state-of-the-art structured
pruning methods. We also conduct post-training experiments to confirm that the
proposed pruning method effectively inherits the parameters of the original
model. Finally, we discuss our motivation from the perspective of layer-wise
similarity and evaluate the performance of the pruned LLMs across various
pruning ratios. | [
"cs.CL",
"cs.AI"
] | false |
2402.11192 | 2024-02-17T05:05:31Z | I Learn Better If You Speak My Language: Enhancing Large Language Model
Fine-Tuning with Style-Aligned Response Adjustments | [
"Xuan Ren",
"Biao Wu",
"Lingqiao Liu"
] | Fine-tuning large language models (LLMs) with a small data set for particular
tasks is a widely encountered yet complex challenge. The potential for
overfitting on a limited number of examples can negatively impact the model's
ability to generalize and retain its original skills. Our research explores the
impact of the style of ground-truth responses during the fine-tuning process.
We found that matching the ground-truth response style with the LLM's inherent
style results in better learning outcomes. Building on this insight, we
developed a method that minimally alters the LLM's pre-existing responses to
correct errors, using these adjusted responses as training targets. This
technique enables precise corrections in line with the model's native response
style, safeguarding the model's core capabilities and thus avoid overfitting.
Our findings show that this approach not only improves the LLM's task-specific
accuracy but also crucially maintains its original competencies and
effectiveness. | [
"cs.CL",
"cs.AI"
] | false |
2402.11243 | 2024-02-17T10:37:51Z | Can Large Language Models perform Relation-based Argument Mining? | [
"Deniz Gorur",
"Antonio Rago",
"Francesca Toni"
] | Argument mining (AM) is the process of automatically extracting arguments,
their components and/or relations amongst arguments and components from text.
As the number of platforms supporting online debate increases, the need for AM
becomes ever more urgent, especially in support of downstream tasks.
Relation-based AM (RbAM) is a form of AM focusing on identifying agreement
(support) and disagreement (attack) relations amongst arguments. RbAM is a
challenging classification task, with existing methods failing to perform
satisfactorily. In this paper, we show that general-purpose Large Language
Models (LLMs), appropriately primed and prompted, can significantly outperform
the best performing (RoBERTa-based) baseline. Specifically, we experiment with
two open-source LLMs (Llama-2 and Mistral) with ten datasets. | [
"cs.CL",
"cs.AI",
"I.2.7"
] | false |
2402.11260 | 2024-02-17T12:25:31Z | MoRAL: MoE Augmented LoRA for LLMs' Lifelong Learning | [
"Shu Yang",
"Muhammad Asif Ali",
"Cheng-Long Wang",
"Lijie Hu",
"Di Wang"
] | Adapting large language models (LLMs) to new domains/tasks and enabling them
to be efficient lifelong learners is a pivotal challenge. In this paper, we
propose MoRAL, i.e., Mixture-of-Experts augmented Low-Rank Adaptation for
Lifelong Learning. MoRAL combines the multi-tasking abilities of MoE with the
fine-tuning abilities of LoRA for effective life-long learning of LLMs. In
contrast to the conventional approaches that use factual triplets as inputs
MoRAL relies on simple question-answer pairs, which is a more practical and
effective strategy for robust and efficient learning. Owing to new data
settings, we introduce a new evaluation benchmark namely: Life Long Learning of
LLM (5L-bench) encompassing a newly curated dataset of question-answer pairs,
and a set of evaluation metrics for rigorous evaluation of MoRAL in open-book
and closed-book settings. Experimental evaluation shows (i) LLMs learn fast in
open-book settings with up to 30.15% improvement in "RA" for Phi-2-2.7B
compared to closed-book (for models fine-tuned with MoRAL); (ii) MoRAL shows
higher performance improvement for models with a greater number of parameters;
(iii) MoRAL is robust to catastrophic forgetting offering better knowledge
retention compared to baselines. | [
"cs.CL",
"cs.AI"
] | false |
2402.11279 | 2024-02-17T13:37:39Z | Multi-Perspective Consistency Enhances Confidence Estimation in Large
Language Models | [
"Pei Wang",
"Yejie Wang",
"Muxi Diao",
"Keqing He",
"Guanting Dong",
"Weiran Xu"
] | In the deployment of large language models (LLMs), accurate confidence
estimation is critical for assessing the credibility of model predictions.
However, existing methods often fail to overcome the issue of overconfidence on
incorrect answers. In this work, we focus on improving the confidence
estimation of large language models. Considering the fragility of
self-awareness in language models, we introduce a Multi-Perspective Consistency
(MPC) method. We leverage complementary insights from different perspectives
within models (MPC-Internal) and across different models (MPC-Across) to
mitigate the issue of overconfidence arising from a singular viewpoint. The
experimental results on eight publicly available datasets show that our MPC
achieves state-of-the-art performance. Further analyses indicate that MPC can
mitigate the problem of overconfidence and is effectively scalable to other
models. | [
"cs.CL",
"cs.AI"
] | false |
2402.11291 | 2024-02-17T14:19:38Z | Puzzle Solving using Reasoning of Large Language Models: A Survey | [
"Panagiotis Giadikiaroglou",
"Maria Lymperaiou",
"Giorgos Filandrianos",
"Giorgos Stamou"
] | Exploring the capabilities of Large Language Models (LLMs) in puzzle solving
unveils critical insights into their potential and challenges in artificial
intelligence, marking a significant step towards understanding their
applicability in complex reasoning tasks. This survey leverages a unique
taxonomy -- dividing puzzles into rule-based and rule-less categories -- to
critically assess LLMs through various methodologies, including prompting
techniques, neuro-symbolic approaches, and fine-tuning. Through a critical
review of relevant datasets and benchmarks, we assess LLMs' performance,
identifying significant challenges in complex puzzle scenarios. Our findings
highlight the disparity between LLM capabilities and human-like reasoning,
particularly in those requiring advanced logical inference. The survey
underscores the necessity for novel strategies and richer datasets to advance
LLMs' puzzle-solving proficiency and contribute to AI's logical reasoning and
creative problem-solving advancements. | [
"cs.CL",
"cs.AI"
] | false |
2402.11296 | 2024-02-17T14:34:31Z | Dissecting Human and LLM Preferences | [
"Junlong Li",
"Fan Zhou",
"Shichao Sun",
"Yikai Zhang",
"Hai Zhao",
"Pengfei Liu"
] | As a relative quality comparison of model responses, human and Large Language
Model (LLM) preferences serve as common alignment goals in model fine-tuning
and criteria in evaluation. Yet, these preferences merely reflect broad
tendencies, resulting in less explainable and controllable models with
potential safety risks. In this work, we dissect the preferences of human and
32 different LLMs to understand their quantitative composition, using
annotations from real-world user-model conversations for a fine-grained,
scenario-wise analysis. We find that humans are less sensitive to errors, favor
responses that support their stances, and show clear dislike when models admit
their limits. On the contrary, advanced LLMs like GPT-4-Turbo emphasize
correctness, clarity, and harmlessness more. Additionally, LLMs of similar
sizes tend to exhibit similar preferences, regardless of their training
methods, and fine-tuning for alignment does not significantly alter the
preferences of pretrained-only LLMs. Finally, we show that preference-based
evaluation can be intentionally manipulated. In both training-free and
training-based settings, aligning a model with the preferences of judges boosts
scores, while injecting the least preferred properties lowers them. This
results in notable score shifts: up to 0.59 on MT-Bench (1-10 scale) and 31.94
on AlpacaEval 2.0 (0-100 scale), highlighting the significant impact of this
strategic adaptation. Interactive Demo:
https://huggingface.co/spaces/GAIR/Preference-Dissection-Visualization Dataset:
https://huggingface.co/datasets/GAIR/preference-dissection Code:
https://github.com/GAIR-NLP/Preference-Dissection | [
"cs.CL",
"cs.AI"
] | false |
2402.11349 | 2024-02-17T17:52:24Z | Tasks That Language Models Don't Learn | [
"Bruce W. Lee",
"JaeHyuk Lim"
] | We argue that there are certain properties of language that our current large
language models (LLMs) don't learn. We present an empirical investigation of
visual-auditory properties of language through a series of tasks, termed
H-TEST. This benchmark highlights a fundamental gap between human linguistic
comprehension, which naturally integrates sensory experiences, and the
sensory-deprived processing capabilities of LLMs. In support of our hypothesis,
1. deliberate reasoning (Chain-of-Thought), 2. few-shot examples, or 3.
stronger LLM from the same model family (LLaMA 2 13B -> LLaMA 2 70B) do not
trivially bring improvements in H-TEST performance. Therefore, we make a
particular connection to the philosophical case of Mary, who learns about the
world in a sensory-deprived environment (Jackson, 1986). Our experiments show
that some of the strongest proprietary LLMs stay near random chance baseline
accuracy of 50%, highlighting the limitations of knowledge acquired in the
absence of sensory experience. | [
"cs.CL",
"cs.AI"
] | false |
2402.11359 | 2024-02-17T18:31:21Z | Training Language Model Agents without Modifying Language Models | [
"Shaokun Zhang",
"Jieyu Zhang",
"Jiale Liu",
"Linxin Song",
"Chi Wang",
"Ranjay Krishna",
"Qingyun Wu"
] | Researchers and practitioners have recently reframed powerful Large Language
Models (LLMs) as agents, enabling them to automate complex tasks largely via
the use of specialized functions. To facilitate the development of LLM agents,
we present a novel paradigm of training LLM agents without modifying the LLM
weights, which is particularly useful when the LLMs are difficult or
inaccessible for modifications. Inspired by how humans continuously forge tools
to adapt to real-world tasks, rather than change our biological structure to
fit a static set of tools, we propose to progressively forge agent's functions
to better solve the downstream tasks instead of modifying the LLM weights. By
treating the functions as learnable `agent parameters' and leveraging the
fundamental idea of model training in artificial intelligence, we develop
AgentOptimizer that employs the LLM to update agents' functions and devise an
agent training algorithm with two strategies, roll-back, and early-stop, to
streamline the training process. With extensive experiments, we showcase that
the agent training paradigm could significantly improve the performance of
representative LLM agents in various downstream tasks. We also study the
behavior of the agent training regarding aspects like the learning curve and
domain transferability. | [
"cs.AI",
"cs.CL"
] | false |
2403.05561 | 2024-02-17T10:32:43Z | Detecting a Proxy for Potential Comorbid ADHD in People Reporting
Anxiety Symptoms from Social Media Data | [
"Claire S. Lee",
"Noelle Lim",
"Michael Guerzhoy"
] | We present a novel task that can elucidate the connection between anxiety and
ADHD; use Transformers to make progress toward solving a task that is not
solvable by keyword-based classifiers; and discuss a method for visualization
of our classifier illuminating the connection between anxiety and ADHD
presentations.
Up to approximately 50% of adults with ADHD may also have an anxiety disorder
and approximately 30\% of adults with anxiety may also have ADHD. Patients
presenting with anxiety may be treated for anxiety without ADHD ever being
considered, possibly affecting treatment. We show how data that bears on ADHD
that is comorbid with anxiety can be obtained from social media data, and show
that Transformers can be used to detect a proxy for possible comorbid ADHD in
people with anxiety symptoms.
We collected data from anxiety and ADHD online forums (subreddits). We
identified posters who first started posting in the Anxiety subreddit and later
started posting in the ADHD subreddit as well. We use this subset of the
posters as a proxy for people who presented with anxiety symptoms and then
became aware that they might have ADHD. We fine-tune a Transformer
architecture-based classifier to classify people who started posting in the
Anxiety subreddit and then started posting in the ADHD subreddit vs. people who
posted in the Anxiety subreddit without later posting in the ADHD subreddit. We
show that a Transformer architecture is capable of achieving reasonable results
(76% correct for RoBERTa vs. under 60% correct for the best keyword-based
model, both with 50% base rate). | [
"cs.CY",
"cs.CL"
] | false |
2402.11138 | 2024-02-17T00:09:32Z | Contrastive Instruction Tuning | [
"Tianyi Yan",
"Fei Wang",
"James Y. Huang",
"Wenxuan Zhou",
"Fan Yin",
"Aram Galstyan",
"Wenpeng Yin",
"Muhao Chen"
] | Instruction tuning has been used as a promising approach to improve the
performance of large language models (LLMs) on unseen tasks. However, current
LLMs exhibit limited robustness to unseen instructions, generating inconsistent
outputs when the same instruction is phrased with slightly varied forms or
language styles. This behavior indicates LLMs' lack of robustness to textual
variations and generalizability to unseen instructions, potentially leading to
trustworthiness issues. Accordingly, we propose Contrastive Instruction Tuning,
which maximizes the similarity between the hidden representations of
semantically equivalent instruction-instance pairs while minimizing the
similarity between semantically different ones. To facilitate this approach, we
augment the existing FLAN collection by paraphrasing task instructions.
Experiments on the PromptBench benchmark show that CoIN consistently improves
LLMs' robustness to unseen instructions with variations across character, word,
sentence, and semantic levels by an average of +2.5% in accuracy. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.11140 | 2024-02-17T00:13:36Z | Boosting of Thoughts: Trial-and-Error Problem Solving with Large
Language Models | [
"Sijia Chen",
"Baochun Li",
"Di Niu"
] | The reasoning performance of Large Language Models (LLMs) on a wide range of
problems critically relies on chain-of-thought prompting, which involves
providing a few chain of thought demonstrations as exemplars in prompts. Recent
work, e.g., Tree of Thoughts, has pointed out the importance of exploration and
self-evaluation in reasoning step selection for complex problem solving. In
this paper, we present Boosting of Thoughts (BoT), an automated prompting
framework for problem solving with LLMs by iteratively exploring and
self-evaluating many trees of thoughts in order to acquire an ensemble of
trial-and-error reasoning experiences, which will serve as a new form of
prompting to solve the complex problem. Starting from a simple prompt without
requiring examples, BoT iteratively explores and evaluates a large collection
of reasoning steps, and more importantly, uses error analysis obtained from the
LLM on them to explicitly revise prompting, which in turn enhances reasoning
step generation, until a final answer is attained. Our experiments with GPT-4
and Llama2 across extensive complex mathematical problems demonstrate that BoT
consistently achieves higher or comparable problem-solving rates than other
advanced prompting approaches. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.11203 | 2024-02-17T05:44:40Z | Exploring ChatGPT for Next-generation Information Retrieval:
Opportunities and Challenges | [
"Yizheng Huang",
"Jimmy Huang"
] | The rapid advancement of artificial intelligence (AI) has highlighted ChatGPT
as a pivotal technology in the field of information retrieval (IR).
Distinguished from its predecessors, ChatGPT offers significant benefits that
have attracted the attention of both the industry and academic communities.
While some view ChatGPT as a groundbreaking innovation, others attribute its
success to the effective integration of product development and market
strategies. The emergence of ChatGPT, alongside GPT-4, marks a new phase in
Generative AI, generating content that is distinct from training examples and
exceeding the capabilities of the prior GPT-3 model by OpenAI. Unlike the
traditional supervised learning approach in IR tasks, ChatGPT challenges
existing paradigms, bringing forth new challenges and opportunities regarding
text quality assurance, model bias, and efficiency. This paper seeks to examine
the impact of ChatGPT on IR tasks and offer insights into its potential future
developments. | [
"cs.IR",
"cs.AI",
"cs.CL",
"cs.LG"
] | false |
2402.11208 | 2024-02-17T06:48:45Z | Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based
Agents | [
"Wenkai Yang",
"Xiaohan Bi",
"Yankai Lin",
"Sishuo Chen",
"Jie Zhou",
"Xu Sun"
] | Leveraging the rapid development of Large Language Models LLMs, LLM-based
agents have been developed to handle various real-world applications, including
finance, healthcare, and shopping, etc. It is crucial to ensure the reliability
and security of LLM-based agents during applications. However, the safety
issues of LLM-based agents are currently under-explored. In this work, we take
the first step to investigate one of the typical safety threats, backdoor
attack, to LLM-based agents. We first formulate a general framework of agent
backdoor attacks, then we present a thorough analysis on the different forms of
agent backdoor attacks. Specifically, from the perspective of the final
attacking outcomes, the attacker can either choose to manipulate the final
output distribution, or only introduce malicious behavior in the intermediate
reasoning process, while keeping the final output correct. Furthermore, the
former category can be divided into two subcategories based on trigger
locations: the backdoor trigger can be hidden either in the user query or in an
intermediate observation returned by the external environment. We propose the
corresponding data poisoning mechanisms to implement the above variations of
agent backdoor attacks on two typical agent tasks, web shopping and tool
utilization. Extensive experiments show that LLM-based agents suffer severely
from backdoor attacks, indicating an urgent need for further research on the
development of defenses against backdoor attacks on LLM-based agents. Warning:
This paper may contain biased content. | [
"cs.CR",
"cs.AI",
"cs.CL"
] | false |
2402.11353 | 2024-02-17T18:05:53Z | Understanding the Impact of Long-Term Memory on Self-Disclosure with
Large Language Model-Driven Chatbots for Public Health Intervention | [
"Eunkyung Jo",
"Yuin Jeong",
"SoHyun Park",
"Daniel A. Epstein",
"Young-Ho Kim"
] | Recent large language models (LLMs) offer the potential to support public
health monitoring by facilitating health disclosure through open-ended
conversations but rarely preserve the knowledge gained about individuals across
repeated interactions. Augmenting LLMs with long-term memory (LTM) presents an
opportunity to improve engagement and self-disclosure, but we lack an
understanding of how LTM impacts people's interaction with LLM-driven chatbots
in public health interventions. We examine the case of CareCall -- an
LLM-driven voice chatbot with LTM -- through the analysis of 1,252 call logs
and interviews with nine users. We found that LTM enhanced health disclosure
and fostered positive perceptions of the chatbot by offering familiarity.
However, we also observed challenges in promoting self-disclosure through LTM,
particularly around addressing chronic health conditions and privacy concerns.
We discuss considerations for LTM integration in LLM-driven chatbots for public
health monitoring, including carefully deciding what topics need to be
remembered in light of public health goals. | [
"cs.HC",
"cs.AI",
"cs.CL",
"H.5.2; I.2.7"
] | false |
2402.11355 | 2024-02-17T18:12:02Z | What Changed? Converting Representational Interventions to Natural
Language | [
"Matan Avitan",
"Ryan Cotterell",
"Yoav Goldberg",
"Shauli Ravfogel"
] | Interventions targeting the representation space of language models (LMs)
have emerged as effective means to influence model behavior. These methods are
employed, for example, to eliminate or alter the encoding of demographic
information such as gender within the model's representations, creating a
counterfactual representation. However, since the intervention operates within
the representation space, understanding precisely which features it modifies
poses a challenge. We show that representation-space counterfactuals can be
converted into natural language counterfactuals. We demonstrate that this
approach enables us to analyze the linguistic alterations corresponding to a
given representation-space intervention and to interpret the features utilized
for encoding a specific concept. Moreover, the resulting counterfactuals can be
used to mitigate bias in classification. | [
"cs.CL",
"cs.CY",
"cs.LG"
] | false |
2402.11399 | 2024-02-17T22:50:38Z | k-SemStamp: A Clustering-Based Semantic Watermark for Detection of
Machine-Generated Text | [
"Abe Bohan Hou",
"Jingyu Zhang",
"Yichen Wang",
"Daniel Khashabi",
"Tianxing He"
] | Recent watermarked generation algorithms inject detectable signatures during
language generation to facilitate post-hoc detection. While token-level
watermarks are vulnerable to paraphrase attacks, SemStamp (Hou et al., 2023)
applies watermark on the semantic representation of sentences and demonstrates
promising robustness. SemStamp employs locality-sensitive hashing (LSH) to
partition the semantic space with arbitrary hyperplanes, which results in a
suboptimal tradeoff between robustness and speed. We propose k-SemStamp, a
simple yet effective enhancement of SemStamp, utilizing k-means clustering as
an alternative of LSH to partition the embedding space with awareness of
inherent semantic structure. Experimental results indicate that k-SemStamp
saliently improves its robustness and sampling efficiency while preserving the
generation quality, advancing a more effective tool for machine-generated text
detection. | [
"cs.CL",
"cs.CR",
"cs.CY",
"cs.LG"
] | false |
2402.14833 | 2024-02-17T22:37:17Z | CliqueParcel: An Approach For Batching LLM Prompts That Jointly
Optimizes Efficiency And Faithfulness | [
"Jiayi Liu",
"Tinghan Yang",
"Jennifer Neville"
] | Large language models (LLMs) have become pivotal in recent research. However,
during the inference process, LLMs still require substantial resources. In this
paper, we propose CliqueParcel, a method designed to improve the efficiency of
LLMs via prompt batching. Existing strategies to optimize inference efficiency
often compromise on output quality, leading to a discounted output problem.
This issue might result in reduced accuracy or outputs that are less detailed.
CliqueParcel is our answer to this challenge. While ensuring accuracy and
minimizing deviations from the original outputs (i.e., faithfulness), our
method significantly improves efficiency during inference.
To lay the groundwork, we first redefine efficiency measurements by excluding
the reduction in running time due to shorter lengths. Then, we provide a
comprehensive trade-off between efficiency and faithfulness to clarify the
nature of the 'discounted output' problem. Within the CliqueParcel framework,
we suggest multiple batching sub-methods and discuss the specific scenarios in
which they can be applied. During evaluation, CliqueParcel is tested on eight
widely recognized datasets, which can be classified into three types: reading
comprehension, open-source question-answering, and reasoning. Our experiments
explore the performance of CliqueParcel, including efficiency, faithfulness,
and the trade-off between them. This work provides novel insights into
inference efficiency and demonstrates promising performance. | [
"cs.CL",
"cs.AI",
"cs.LG"
] | false |
2402.11153 | 2024-02-17T00:40:12Z | Beyond Generalization: A Survey of Out-Of-Distribution Adaptation on
Graphs | [
"Shuhan Liu",
"Kaize Ding"
] | Distribution shifts on graphs -- the data distribution discrepancies between
training and testing a graph machine learning model, are often ubiquitous and
unavoidable in real-world scenarios. Such shifts may severely deteriorate the
performance of the model, posing significant challenges for reliable graph
machine learning. Consequently, there has been a surge in research on graph
Out-Of-Distribution (OOD) adaptation methods that aim to mitigate the
distribution shifts and adapt the knowledge from one distribution to another.
In our survey, we provide an up-to-date and forward-looking review of graph OOD
adaptation methods, covering two main problem scenarios including training-time
as well as test-time graph OOD adaptation. We start by formally formulating the
two problems and then discuss different types of distribution shifts on graphs.
Based on our proposed taxonomy for graph OOD adaptation, we systematically
categorize the existing methods according to their learning paradigm and
investigate the techniques behind them. Finally, we point out promising
research directions and the corresponding challenges. We also provide a
continuously updated reading list at
https://github.com/kaize0409/Awesome-Graph-OOD-Adaptation.git | [
"cs.LG"
] | false |
2402.11223 | 2024-02-17T08:41:37Z | HEAL: Brain-inspired Hyperdimensional Efficient Active Learning | [
"Yang Ni",
"Zhuowen Zou",
"Wenjun Huang",
"Hanning Chen",
"William Youngwoo Chung",
"Samuel Cho",
"Ranganath Krishnan",
"Pietro Mercati",
"Mohsen Imani"
] | Drawing inspiration from the outstanding learning capability of our human
brains, Hyperdimensional Computing (HDC) emerges as a novel computing paradigm,
and it leverages high-dimensional vector presentation and operations for
brain-like lightweight Machine Learning (ML). Practical deployments of HDC have
significantly enhanced the learning efficiency compared to current deep ML
methods on a broad spectrum of applications. However, boosting the data
efficiency of HDC classifiers in supervised learning remains an open question.
In this paper, we introduce Hyperdimensional Efficient Active Learning (HEAL),
a novel Active Learning (AL) framework tailored for HDC classification. HEAL
proactively annotates unlabeled data points via uncertainty and
diversity-guided acquisition, leading to a more efficient dataset annotation
and lowering labor costs. Unlike conventional AL methods that only support
classifiers built upon deep neural networks (DNN), HEAL operates without the
need for gradient or probabilistic computations. This allows it to be
effortlessly integrated with any existing HDC classifier architecture. The key
design of HEAL is a novel approach for uncertainty estimation in HDC
classifiers through a lightweight HDC ensemble with prior hypervectors.
Additionally, by exploiting hypervectors as prototypes (i.e., compact
representations), we develop an extra metric for HEAL to select diverse samples
within each batch for annotation. Our evaluation shows that HEAL surpasses a
diverse set of baselines in AL quality and achieves notably faster acquisition
than many BNN-powered or diversity-guided AL methods, recording 11 times to
40,000 times speedup in acquisition runtime per batch. | [
"cs.LG"
] | false |
2402.11235 | 2024-02-17T09:52:43Z | ZeroG: Investigating Cross-dataset Zero-shot Transferability in Graphs | [
"Yuhan Li",
"Peisong Wang",
"Zhixun Li",
"Jeffrey Xu Yu",
"Jia Li"
] | With the development of foundation models such as large language models,
zero-shot transfer learning has become increasingly significant. This is
highlighted by the generative capabilities of NLP models like GPT-4, and the
retrieval-based approaches of CV models like CLIP, both of which effectively
bridge the gap between seen and unseen data. In the realm of graph learning,
the continuous emergence of new graphs and the challenges of human labeling
also amplify the necessity for zero-shot transfer learning, driving the
exploration of approaches that can generalize across diverse graph data without
necessitating dataset-specific and label-specific fine-tuning. In this study,
we extend such paradigms to zero-shot transferability in graphs by introducing
ZeroG, a new framework tailored to enable cross-dataset generalization.
Addressing the inherent challenges such as feature misalignment, mismatched
label spaces, and negative transfer, we leverage a language model to encode
both node attributes and class semantics, ensuring consistent feature
dimensions across datasets. We also propose a prompt-based subgraph sampling
module that enriches the semantic information and structure information of
extracted subgraphs using prompting nodes and neighborhood aggregation,
respectively. We further adopt a lightweight fine-tuning strategy that reduces
the risk of overfitting and maintains the zero-shot learning efficacy of the
language model. The results underscore the effectiveness of our model in
achieving significant cross-dataset zero-shot transferability, opening pathways
for the development of graph foundation models. Especially, ZeroG, as a
zero-shot method, can even achieve results comparable to those of
semi-supervised learning on Pubmed. | [
"cs.LG"
] | false |
2402.11367 | 2024-02-17T19:49:00Z | Multi Task Inverse Reinforcement Learning for Common Sense Reward | [
"Neta Glazer",
"Aviv Navon",
"Aviv Shamsian",
"Ethan Fetaya"
] | One of the challenges in applying reinforcement learning in a complex
real-world environment lies in providing the agent with a sufficiently detailed
reward function. Any misalignment between the reward and the desired behavior
can result in unwanted outcomes. This may lead to issues like "reward hacking"
where the agent maximizes rewards by unintended behavior. In this work, we
propose to disentangle the reward into two distinct parts. A simple
task-specific reward, outlining the particulars of the task at hand, and an
unknown common-sense reward, indicating the expected behavior of the agent
within the environment. We then explore how this common-sense reward can be
learned from expert demonstrations. We first show that inverse reinforcement
learning, even when it succeeds in training an agent, does not learn a useful
reward function. That is, training a new agent with the learned reward does not
impair the desired behaviors. We then demonstrate that this problem can be
solved by training simultaneously on multiple tasks. That is, multi-task
inverse reinforcement learning can be applied to learn a useful reward
function. | [
"cs.LG"
] | false |
2402.12398 | 2024-02-17T05:39:48Z | Primary and Secondary Factor Consistency as Domain Knowledge to Guide
Happiness Computing in Online Assessment | [
"Xiaohua Wu",
"Lin Li",
"Xiaohui Tao",
"Frank Xing",
"Jingling Yuan"
] | Happiness computing based on large-scale online web data and machine learning
methods is an emerging research topic that underpins a range of issues, from
personal growth to social stability. Many advanced Machine Learning (ML) models
with explanations are used to compute the happiness online assessment while
maintaining high accuracy of results. However, domain knowledge constraints,
such as the primary and secondary relations of happiness factors, are absent
from these models, which limits the association between computing results and
the right reasons for why they occurred. This article attempts to provide new
insights into the explanation consistency from an empirical study perspective.
Then we study how to represent and introduce domain knowledge constraints to
make ML models more trustworthy. We achieve this through: (1) proving that
multiple prediction models with additive factor attributions will have the
desirable property of primary and secondary relations consistency, and (2)
showing that factor relations with quantity can be represented as an importance
distribution for encoding domain knowledge. Factor explanation difference is
penalized by the Kullback-Leibler divergence-based loss among computing models.
Experimental results using two online web datasets show that domain knowledge
of stable factor relations exists. Using this knowledge not only improves
happiness computing accuracy but also reveals more significative happiness
factors for assisting decisions well. | [
"cs.LG"
] | false |
2402.14027 | 2024-02-17T22:26:56Z | Learning causation event conjunction sequences | [
"Thomas E. Portegys"
] | This is an examination of some methods that learn causations in event
sequences. A causation is defined as a conjunction of one or more cause events
occurring in an arbitrary order, with possible intervening non-causal events,
that lead to an effect. The methods include recurrent and non-recurrent
artificial neural networks (ANNs), as well as a histogram-based algorithm. An
attention recurrent ANN performed the best of the ANNs, while the histogram
algorithm was significantly superior to all the ANNs. | [
"cs.LG"
] | false |
2402.11139 | 2024-02-17T00:10:33Z | LiGNN: Graph Neural Networks at LinkedIn | [
"Fedor Borisyuk",
"Shihai He",
"Yunbo Ouyang",
"Morteza Ramezani",
"Peng Du",
"Xiaochen Hou",
"Chengming Jiang",
"Nitin Pasumarthy",
"Priya Bannur",
"Birjodh Tiwana",
"Ping Liu",
"Siddharth Dangi",
"Daqi Sun",
"Zhoutao Pei",
"Xiao Shi",
"Sirou Zhu",
"Qianqi Shen",
"Kuang-Hsuan Lee",
"David Stein",
"Baolei Li",
"Haichao Wei",
"Amol Ghoting",
"Souvik Ghosh"
] | In this paper, we present LiGNN, a deployed large-scale Graph Neural Networks
(GNNs) Framework. We share our insight on developing and deployment of GNNs at
large scale at LinkedIn. We present a set of algorithmic improvements to the
quality of GNN representation learning including temporal graph architectures
with long term losses, effective cold start solutions via graph densification,
ID embeddings and multi-hop neighbor sampling. We explain how we built and sped
up by 7x our large-scale training on LinkedIn graphs with adaptive sampling of
neighbors, grouping and slicing of training data batches, specialized
shared-memory queue and local gradient optimization. We summarize our
deployment lessons and learnings gathered from A/B test experiments. The
techniques presented in this work have contributed to an approximate relative
improvements of 1% of Job application hearing back rate, 2% Ads CTR lift, 0.5%
of Feed engaged daily active users, 0.2% session lift and 0.1% weekly active
user lift from people recommendation. We believe that this work can provide
practical solutions and insights for engineers who are interested in applying
Graph neural networks at large scale. | [
"cs.LG",
"cs.AI"
] | false |
2402.11156 | 2024-02-17T00:51:29Z | Efficient Low-Rank Matrix Estimation, Experimental Design, and
Arm-Set-Dependent Low-Rank Bandits | [
"Kyoungseok Jang",
"Chicheng Zhang",
"Kwang-Sung Jun"
] | We study low-rank matrix trace regression and the related problem of low-rank
matrix bandits. Assuming access to the distribution of the covariates, we
propose a novel low-rank matrix estimation method called LowPopArt and provide
its recovery guarantee that depends on a novel quantity denoted by B(Q) that
characterizes the hardness of the problem, where Q is the covariance matrix of
the measurement distribution. We show that our method can provide tighter
recovery guarantees than classical nuclear norm penalized least squares
(Koltchinskii et al., 2011) in several problems. To perform efficient
estimation with a limited number of measurements from an arbitrarily given
measurement set A, we also propose a novel experimental design criterion that
minimizes B(Q) with computational efficiency. We leverage our novel estimator
and design of experiments to derive two low-rank linear bandit algorithms for
general arm sets that enjoy improved regret upper bounds. This improves over
previous works on low-rank bandits, which make somewhat restrictive assumptions
that the arm set is the unit ball or that an efficient exploration distribution
is given. To our knowledge, our experimental design criterion is the first one
tailored to low-rank matrix estimation beyond the naive reduction to linear
regression, which can be of independent interest. | [
"stat.ML",
"cs.LG"
] | false |
2402.11185 | 2024-02-17T04:05:01Z | Minimally Supervised Topological Projections of Self-Organizing Maps for
Phase of Flight Identification | [
"Zimeng Lyu",
"Pujan Thapa",
"Travis Desell"
] | Identifying phases of flight is important in the field of general aviation,
as knowing which phase of flight data is collected from aircraft flight data
recorders can aid in the more effective detection of safety or hazardous
events. General aviation flight data for phase of flight identification is
usually per-second data, comes on a large scale, and is class imbalanced. It is
expensive to manually label the data and training classification models usually
faces class imbalance problems. This work investigates the use of a novel
method for minimally supervised self-organizing maps (MS-SOMs) which utilize
nearest neighbor majority votes in the SOM U-matrix for class estimation.
Results show that the proposed method can reach or exceed a naive SOM approach
which utilized a full data file of labeled data, with only 30 labeled
datapoints per class. Additionally, the minimally supervised SOM is
significantly more robust to the class imbalance of the phase of flight data.
These results highlight how little data is required for effective phase of
flight identification. | [
"cs.LG",
"cs.NE"
] | false |
2402.11196 | 2024-02-17T05:14:47Z | Maintaining Adversarial Robustness in Continuous Learning | [
"Xiaolei Ru",
"Xiaowei Cao",
"Zijia Liu",
"Jack Murdoch Moore",
"Xin-Ya Zhang",
"Xia Zhu",
"Wenjia Wei",
"Gang Yan"
] | Adversarial robustness is essential for security and reliability of machine
learning systems. However, the adversarial robustness gained by sophisticated
defense algorithms is easily erased as the neural network evolves to learn new
tasks. This vulnerability can be addressed by fostering a novel capability for
neural networks, termed continual robust learning, which focuses on both the
(classification) performance and adversarial robustness on previous tasks
during continuous learning. To achieve continuous robust learning, we propose
an approach called Double Gradient Projection that projects the gradients for
weight updates orthogonally onto two crucial subspaces -- one for stabilizing
the smoothed sample gradients and another for stabilizing the final outputs of
the neural network. The experimental results on four benchmarks demonstrate
that the proposed approach effectively maintains continuous robustness against
strong adversarial attacks, outperforming the baselines formed by combining the
existing defense strategies and continual learning methods. | [
"cs.LG",
"cs.AI"
] | false |
2402.11198 | 2024-02-17T05:22:46Z | Achieving Linear Speedup in Asynchronous Federated Learning with
Heterogeneous Clients | [
"Xiaolu Wang",
"Zijian Li",
"Shi Jin",
"Jun Zhang"
] | Federated learning (FL) is an emerging distributed training paradigm that
aims to learn a common global model without exchanging or transferring the data
that are stored locally at different clients. The Federated Averaging
(FedAvg)-based algorithms have gained substantial popularity in FL to reduce
the communication overhead, where each client conducts multiple localized
iterations before communicating with a central server. In this paper, we focus
on FL where the clients have diverse computation and/or communication
capabilities. Under this circumstance, FedAvg can be less efficient since it
requires all clients that participate in the global aggregation in a round to
initiate iterations from the latest global model, and thus the synchronization
among fast clients and straggler clients can severely slow down the overall
training process. To address this issue, we propose an efficient asynchronous
federated learning (AFL) framework called Delayed Federated Averaging
(DeFedAvg). In DeFedAvg, the clients are allowed to perform local training with
different stale global models at their own paces. Theoretical analyses
demonstrate that DeFedAvg achieves asymptotic convergence rates that are on par
with the results of FedAvg for solving nonconvex problems. More importantly,
DeFedAvg is the first AFL algorithm that provably achieves the desirable linear
speedup property, which indicates its high scalability. Additionally, we carry
out extensive numerical experiments using real datasets to validate the
efficiency and scalability of our approach when training deep neural networks. | [
"cs.LG",
"cs.DC"
] | false |
2402.11224 | 2024-02-17T08:54:25Z | Neural Networks with (Low-Precision) Polynomial Approximations: New
Insights and Techniques for Accuracy Improvement | [
"Chi Zhang",
"Man Ho Au",
"Siu Ming Yiu"
] | Replacing non-polynomial functions (e.g., non-linear activation functions
such as ReLU) in a neural network with their polynomial approximations is a
standard practice in privacy-preserving machine learning. The resulting neural
network, called polynomial approximation of neural network (PANN) in this
paper, is compatible with advanced cryptosystems to enable privacy-preserving
model inference. Using ``highly precise'' approximation, state-of-the-art PANN
offers similar inference accuracy as the underlying backbone model. However,
little is known about the effect of approximation, and existing literature
often determined the required approximation precision empirically. In this
paper, we initiate the investigation of PANN as a standalone object.
Specifically, our contribution is two-fold. Firstly, we provide an explanation
on the effect of approximate error in PANN. In particular, we discovered that
(1) PANN is susceptible to some type of perturbations; and (2) weight
regularisation significantly reduces PANN's accuracy. We support our
explanation with experiments. Secondly, based on the insights from our
investigations, we propose solutions to increase inference accuracy for PANN.
Experiments showed that combination of our solutions is very effective: at the
same precision, our PANN is 10% to 50% more accurate than state-of-the-arts;
and at the same accuracy, our PANN only requires a precision of $2^{-9}$ while
state-of-the-art solution requires a precision of $2^{-12}$ using the ResNet-20
model on CIFAR-10 dataset. | [
"cs.LG",
"cs.CR"
] | false |
2402.11227 | 2024-02-17T09:10:05Z | On the Role of Similarity in Detecting Masquerading Files | [
"Jonathan Oliver",
"Jue Mo",
"Susmit Yenkar",
"Raghav Batta",
"Sekhar Josyoula"
] | Similarity has been applied to a wide range of security applications,
typically used in machine learning models. We examine the problem posed by
masquerading samples; that is samples crafted by bad actors to be similar or
near identical to legitimate samples. We find that these samples potentially
create significant problems for machine learning solutions. The primary problem
being that bad actors can circumvent machine learning solutions by using
masquerading samples.
We then examine the interplay between digital signatures and machine learning
solutions. In particular, we focus on executable files and code signing. We
offer a taxonomy for masquerading files. We use a combination of similarity and
clustering to find masquerading files. We use the insights gathered in this
process to offer improvements to similarity based and machine learning security
solutions. | [
"cs.CR",
"cs.LG"
] | false |
2402.11242 | 2024-02-17T10:34:53Z | Learning with Imbalanced Noisy Data by Preventing Bias in Sample
Selection | [
"Huafeng Liu",
"Mengmeng Sheng",
"Zeren Sun",
"Yazhou Yao",
"Xian-Sheng Hua",
"Heng-Tao Shen"
] | Learning with noisy labels has gained increasing attention because the
inevitable imperfect labels in real-world scenarios can substantially hurt the
deep model performance. Recent studies tend to regard low-loss samples as clean
ones and discard high-loss ones to alleviate the negative impact of noisy
labels. However, real-world datasets contain not only noisy labels but also
class imbalance. The imbalance issue is prone to causing failure in the
loss-based sample selection since the under-learning of tail classes also leans
to produce high losses. To this end, we propose a simple yet effective method
to address noisy labels in imbalanced datasets. Specifically, we propose
Class-Balance-based sample Selection (CBS) to prevent the tail class samples
from being neglected during training. We propose Confidence-based Sample
Augmentation (CSA) for the chosen clean samples to enhance their reliability in
the training process. To exploit selected noisy samples, we resort to
prediction history to rectify labels of noisy samples. Moreover, we introduce
the Average Confidence Margin (ACM) metric to measure the quality of corrected
labels by leveraging the model's evolving training dynamics, thereby ensuring
that low-quality corrected noisy samples are appropriately masked out. Lastly,
consistency regularization is imposed on filtered label-corrected noisy samples
to boost model performance. Comprehensive experimental results on synthetic and
real-world datasets demonstrate the effectiveness and superiority of our
proposed method, especially in imbalanced scenarios. Comprehensive experimental
results on synthetic and real-world datasets demonstrate the effectiveness and
superiority of our proposed method, especially in imbalanced scenarios. | [
"cs.LG",
"cs.AI"
] | false |
2402.11262 | 2024-02-17T12:27:30Z | Mirror Gradient: Towards Robust Multimodal Recommender Systems via
Exploring Flat Local Minima | [
"Shanshan Zhong",
"Zhongzhan Huang",
"Daifeng Li",
"Wushao Wen",
"Jinghui Qin",
"Liang Lin"
] | Multimodal recommender systems utilize various types of information to model
user preferences and item features, helping users discover items aligned with
their interests. The integration of multimodal information mitigates the
inherent challenges in recommender systems, e.g., the data sparsity problem and
cold-start issues. However, it simultaneously magnifies certain risks from
multimodal information inputs, such as information adjustment risk and inherent
noise risk. These risks pose crucial challenges to the robustness of
recommendation models. In this paper, we analyze multimodal recommender systems
from the novel perspective of flat local minima and propose a concise yet
effective gradient strategy called Mirror Gradient (MG). This strategy can
implicitly enhance the model's robustness during the optimization process,
mitigating instability risks arising from multimodal information inputs. We
also provide strong theoretical evidence and conduct extensive empirical
experiments to show the superiority of MG across various multimodal
recommendation models and benchmarks. Furthermore, we find that the proposed MG
can complement existing robust training methods and be easily extended to
diverse advanced recommendation models, making it a promising new and
fundamental paradigm for training multimodal recommender systems. The code is
released at https://github.com/Qrange-group/Mirror-Gradient. | [
"cs.IR",
"cs.LG"
] | false |
2402.11317 | 2024-02-17T16:03:35Z | Debiased Offline Representation Learning for Fast Online Adaptation in
Non-stationary Dynamics | [
"Xinyu Zhang",
"Wenjie Qiu",
"Yi-Chen Li",
"Lei Yuan",
"Chengxing Jia",
"Zongzhang Zhang",
"Yang Yu"
] | Developing policies that can adjust to non-stationary environments is
essential for real-world reinforcement learning applications. However, learning
such adaptable policies in offline settings, with only a limited set of
pre-collected trajectories, presents significant challenges. A key difficulty
arises because the limited offline data makes it hard for the context encoder
to differentiate between changes in the environment dynamics and shifts in the
behavior policy, often leading to context misassociations. To address this
issue, we introduce a novel approach called Debiased Offline Representation for
fast online Adaptation (DORA). DORA incorporates an information bottleneck
principle that maximizes mutual information between the dynamics encoding and
the environmental data, while minimizing mutual information between the
dynamics encoding and the actions of the behavior policy. We present a
practical implementation of DORA, leveraging tractable bounds of the
information bottleneck principle. Our experimental evaluation across six
benchmark MuJoCo tasks with variable parameters demonstrates that DORA not only
achieves a more precise dynamics encoding but also significantly outperforms
existing baselines in terms of performance. | [
"cs.LG",
"cs.AI"
] | false |
2402.11339 | 2024-02-17T17:13:41Z | Expressive Higher-Order Link Prediction through Hypergraph Symmetry
Breaking | [
"Simon Zhang",
"Cheng Xin",
"Tamal K. Dey"
] | A hypergraph consists of a set of nodes along with a collection of subsets of
the nodes called hyperedges. Higher-order link prediction is the task of
predicting the existence of a missing hyperedge in a hypergraph. A hyperedge
representation learned for higher order link prediction is fully expressive
when it does not lose distinguishing power up to an isomorphism. Many existing
hypergraph representation learners, are bounded in expressive power by the
Generalized Weisfeiler Lehman-1 (GWL-1) algorithm, a generalization of the
Weisfeiler Lehman-1 algorithm. However, GWL-1 has limited expressive power. In
fact, induced subhypergraphs with identical GWL-1 valued nodes are
indistinguishable. Furthermore, message passing on hypergraphs can already be
computationally expensive, especially on GPU memory. To address these
limitations, we devise a preprocessing algorithm that can identify certain
regular subhypergraphs exhibiting symmetry. Our preprocessing algorithm runs
once with complexity the size of the input hypergraph. During training, we
randomly replace subhypergraphs identified by the algorithm with covering
hyperedges to break symmetry. We show that our method improves the expressivity
of GWL-1. Our extensive experiments also demonstrate the effectiveness of our
approach for higher-order link prediction on both graph and hypergraph datasets
with negligible change in computation. | [
"cs.LG",
"stat.ML"
] | false |
2402.11342 | 2024-02-17T17:31:48Z | Ransomware detection using stacked autoencoder for feature selection | [
"Mike Nkongolo",
"Mahmut Tokmak"
] | The aim of this study is to propose and evaluate an advanced ransomware
detection and classification method that combines a Stacked Autoencoder (SAE)
for precise feature selection with a Long Short Term Memory (LSTM) classifier
to enhance ransomware stratification accuracy. The proposed approach involves
thorough pre processing of the UGRansome dataset and training an unsupervised
SAE for optimal feature selection or fine tuning via supervised learning to
elevate the LSTM model's classification capabilities. The study meticulously
analyzes the autoencoder's learned weights and activations to identify
essential features for distinguishing ransomware families from other malware
and creates a streamlined feature set for precise classification. Extensive
experiments, including up to 400 epochs and varying learning rates, are
conducted to optimize the model's performance. The results demonstrate the
outstanding performance of the SAE-LSTM model across all ransomware families,
boasting high precision, recall, and F1 score values that underscore its robust
classification capabilities. Furthermore, balanced average scores affirm the
proposed model's ability to generalize effectively across various malware
types. The proposed model achieves an exceptional 99% accuracy in ransomware
classification, surpassing the Extreme Gradient Boosting (XGBoost) algorithm
primarily due to its effective SAE feature selection mechanism. The model also
demonstrates outstanding performance in identifying signature attacks,
achieving a 98% accuracy rate. | [
"cs.LG",
"cs.CR"
] | false |
2402.12397 | 2024-02-17T00:22:29Z | Multi-class Temporal Logic Neural Networks | [
"Danyang Li",
"Roberto Tron"
] | Time-series data can represent the behaviors of autonomous systems, such as
drones and self-driving cars. The problem of binary and multi-class
classification has received a lot of attention in this field. Neural networks
represent a popular approach to classifying data; However, they lack
interpretability, which poses a significant challenge in extracting meaningful
information from them. Signal Temporal Logic (STL) is a formalism to describe
the properties of timed behaviors. We propose a method that combines all of the
above: neural networks that represent STL specifications for multi-class
classification of time-series data. We offer two key contributions: 1) We
introduce a notion of margin for multi-class classification, and 2) we
introduce the use of STL-based attributes for enhancing the interpretability of
the results. We evaluate our method on two datasets and compare with
state-of-the-art baselines. | [
"stat.ML",
"cs.LG"
] | false |
2402.12400 | 2024-02-17T20:16:41Z | Estimating the age-conditioned average treatment effects curves: An
application for assessing load-management strategies in the NBA | [
"Shinpei Nakamura-Sakai",
"Laura Forastiere",
"Brian Macdonald"
] | In the realm of competitive sports, understanding the performance dynamics of
athletes, represented by the age curve (showing progression, peak, and
decline), is vital. Our research introduces a novel framework for quantifying
age-specific treatment effects, enhancing the granularity of performance
trajectory analysis. Firstly, we propose a methodology for estimating the age
curve using game-level data, diverging from traditional season-level data
approaches, and tackling its inherent complexities with a meta-learner
framework that leverages advanced machine learning models. This approach
uncovers intricate non-linear patterns missed by existing methods. Secondly,
our framework enables the identification of causal effects, allowing for a
detailed examination of age curves under various conditions. By defining the
Age-Conditioned Treatment Effect (ACTE), we facilitate the exploration of
causal relationships regarding treatment impacts at specific ages. Finally,
applying this methodology to study the effects of rest days on performance
metrics, particularly across different ages, offers valuable insights into load
management strategies' effectiveness. Our findings underscore the importance of
tailored rest periods, highlighting their positive impact on athlete
performance and suggesting a reevaluation of current management practices for
optimizing athlete performance. | [
"stat.AP",
"cs.LG"
] | false |
2402.17772 | 2024-02-17T05:22:41Z | EEG2Rep: Enhancing Self-supervised EEG Representation Through
Informative Masked Inputs | [
"Navid Mohammadi Foumani",
"Geoffrey Mackellar",
"Soheila Ghane",
"Saad Irtza",
"Nam Nguyen",
"Mahsa Salehi"
] | Self-supervised approaches for electroencephalography (EEG) representation
learning face three specific challenges inherent to EEG data: (1) The low
signal-to-noise ratio which challenges the quality of the representation
learned, (2) The wide range of amplitudes from very small to relatively large
due to factors such as the inter-subject variability, risks the models to be
dominated by higher amplitude ranges, and (3) The absence of explicit
segmentation in the continuous-valued sequences which can result in less
informative representations. To address these challenges, we introduce EEG2Rep,
a self-prediction approach for self-supervised representation learning from
EEG. Two core novel components of EEG2Rep are as follows: 1) Instead of
learning to predict the masked input from raw EEG, EEG2Rep learns to predict
masked input in latent representation space, and 2) Instead of conventional
masking methods, EEG2Rep uses a new semantic subsequence preserving (SSP)
method which provides informative masked inputs to guide EEG2Rep to generate
rich semantic representations. In experiments on 6 diverse EEG tasks with
subject variability, EEG2Rep significantly outperforms state-of-the-art
methods. We show that our semantic subsequence preserving improves the existing
masking methods in self-prediction literature and find that preserving 50\% of
EEG recordings will result in the most accurate results on all 6 tasks on
average. Finally, we show that EEG2Rep is robust to noise addressing a
significant challenge that exists in EEG data. Models and code are available
at: https://github.com/Navidfoumani/EEG2Rep | [
"eess.SP",
"cs.LG"
] | false |
2402.17773 | 2024-02-17T20:03:02Z | SINR-Aware Deep Reinforcement Learning for Distributed Dynamic Channel
Allocation in Cognitive Interference Networks | [
"Yaniv Cohen",
"Tomer Gafni",
"Ronen Greenberg",
"Kobi Cohen"
] | We consider the problem of dynamic channel allocation (DCA) in cognitive
communication networks with the goal of maximizing a global
signal-to-interference-plus-noise ratio (SINR) measure under a specified target
quality of service (QoS)-SINR for each network. The shared bandwidth is
partitioned into K channels with frequency separation. In contrast to the
majority of existing studies that assume perfect orthogonality or a one- to-one
user-channel allocation mapping, this paper focuses on real-world systems
experiencing inter-carrier interference (ICI) and channel reuse by multiple
large-scale networks. This realistic scenario significantly increases the
problem dimension, rendering existing algorithms inefficient. We propose a
novel multi-agent reinforcement learning (RL) framework for distributed DCA,
named Channel Allocation RL To Overlapped Networks (CARLTON). The CARLTON
framework is based on the Centralized Training with Decentralized Execution
(CTDE) paradigm, utilizing the DeepMellow value-based RL algorithm. To ensure
robust performance in the interference-laden environment we address, CARLTON
employs a low-dimensional representation of observations, generating a QoS-type
measure while maximizing a global SINR measure and ensuring the target QoS-SINR
for each network. Our results demonstrate exceptional performance and robust
generalization, showcasing superior efficiency compared to alternative
state-of-the-art methods, while achieving a marginally diminished performance
relative to a fully centralized approach. | [
"eess.SP",
"cs.LG"
] | false |
2403.00780 | 2024-02-17T15:00:45Z | Empirical and Experimental Insights into Data Mining Techniques for
Crime Prediction: A Comprehensive Survey | [
"Kamal Taha"
] | This survey paper presents a comprehensive analysis of crime prediction
methodologies, exploring the various techniques and technologies utilized in
this area. The paper covers the statistical methods, machine learning
algorithms, and deep learning techniques employed to analyze crime data, while
also examining their effectiveness and limitations. We propose a methodological
taxonomy that classifies crime prediction algorithms into specific techniques.
This taxonomy is structured into four tiers, including methodology category,
methodology sub-category, methodology techniques, and methodology
sub-techniques. Empirical and experimental evaluations are provided to rank the
different techniques. The empirical evaluation assesses the crime prediction
techniques based on four criteria, while the experimental evaluation ranks the
algorithms that employ the same sub-technique, the different sub-techniques
that employ the same technique, the different techniques that employ the same
methodology sub-category, the different methodology sub-categories within the
same category, and the different methodology categories. The combination of
methodological taxonomy, empirical evaluations, and experimental comparisons
allows for a nuanced and comprehensive understanding of crime prediction
algorithms, aiding researchers in making informed decisions. Finally, the paper
provides a glimpse into the future of crime prediction techniques, highlighting
potential advancements and opportunities for further research in this field | [
"cs.LG",
"cs.AI"
] | false |
2403.18930 | 2024-02-17T16:36:01Z | Optimizing Wireless Networks with Deep Unfolding: Comparative Study on
Two Deep Unfolding Mechanisms | [
"Abuzar B. M. Adam",
"Mohammed A. M. Elhassan",
"Elhadj Moustapha Diallo"
] | In this work, we conduct a comparative study on two deep unfolding mechanisms
to efficiently perform power control in the next generation wireless networks.
The power control problem is formulated as energy efficiency over multiple
interference links. The problem is nonconvex. We employ fractional programming
transformation to design two solutions for the problem. The first solution is a
numerical solution while the second solution is a closed-form solution. Based
on the first solution, we design a semi-unfolding deep learning model where we
combine the domain knowledge of the wireless communications and the recent
advances in the data-driven deep learning. Moreover, on the highlights of the
closed-form solution, fully deep unfolded deep learning model is designed in
which we fully leveraged the expressive closed-form power control solution and
deep learning advances. In the simulation results, we compare the performance
of the proposed deep learning models and the iterative solutions in terms of
accuracy and inference speed to show their suitability for the real-time
application in next generation networks. | [
"cs.NI",
"cs.LG"
] | false |
2404.07209 | 2024-02-17T04:12:09Z | Deep Reinforcement Learning Based Toolpath Generation for Thermal
Uniformity in Laser Powder Bed Fusion Process | [
"Mian Qin",
"Junhao Ding",
"Shuo Qu",
"Xu Song",
"Charlie C. L. Wang",
"Wei-Hsin Liao"
] | Laser powder bed fusion (LPBF) is a widely used metal additive manufacturing
technology. However, the accumulation of internal residual stress during
printing can cause significant distortion and potential failure. Although
various scan patterns have been studied to reduce possible accumulated stress,
such as zigzag scanning vectors with changing directions or a chessboard-based
scan pattern with divided small islands, most conventional scan patterns cannot
significantly reduce residual stress. The proposed adaptive toolpath generation
(ATG) algorithms, aiming to minimize the thermal gradients, may result in
extremely accumulated temperature fields in some cases. To address these
issues, we developed a deep reinforcement learning (DRL)-based toolpath
generation framework, with the goal of achieving uniformly distributed heat and
avoiding extremely thermal accumulation regions during the LPBF process. We
first developed an overall pipeline for the DRL-based toolpath generation
framework, which includes uniformly sampling, agent moving and environment
observation, action selection, moving constraints, rewards calculation, and the
training process. To accelerate the training process, we simplified the
data-intensive numerical model by considering the turning angles on the
toolpath. We designed the action spaces with three options, including the
minimum temperature value, the smoothest path, and the second smoothest path.
The reward function was designed to minimize energy density to ensure the
temperature field remains relatively stable. To verify the effectiveness of the
proposed DRL-based toolpath generation framework, we performed numerical
simulations of polygon shape printing domains. In addition, four groups of thin
plate samples with different scan patterns were compared using the LPBF
process. | [
"cs.CE",
"cs.LG"
] | false |
2402.11173 | 2024-02-17T02:42:56Z | How to Make the Gradients Small Privately: Improved Rates for
Differentially Private Non-Convex Optimization | [
"Andrew Lowy",
"Jonathan Ullman",
"Stephen J. Wright"
] | We provide a simple and flexible framework for designing differentially
private algorithms to find approximate stationary points of non-convex loss
functions. Our framework is based on using a private approximate risk minimizer
to "warm start" another private algorithm for finding stationary points. We use
this framework to obtain improved, and sometimes optimal, rates for several
classes of non-convex loss functions. First, we obtain improved rates for
finding stationary points of smooth non-convex empirical loss functions.
Second, we specialize to quasar-convex functions, which generalize star-convex
functions and arise in learning dynamical systems and training some neural
nets. We achieve the optimal rate for this class. Third, we give an optimal
algorithm for finding stationary points of functions satisfying the
Kurdyka-Lojasiewicz (KL) condition. For example, over-parameterized neural
networks often satisfy this condition. Fourth, we provide new state-of-the-art
rates for stationary points of non-convex population loss functions. Fifth, we
obtain improved rates for non-convex generalized linear models. A modification
of our algorithm achieves nearly the same rates for second-order stationary
points of functions with Lipschitz Hessian, improving over the previous
state-of-the-art for each of the above problems. | [
"cs.LG",
"cs.CR",
"math.OC"
] | false |
2402.11179 | 2024-02-17T03:19:23Z | Uncertainty Quantification of Graph Convolution Neural Network Models of
Evolving Processes | [
"Jeremiah Hauth",
"Cosmin Safta",
"Xun Huan",
"Ravi G. Patel",
"Reese E. Jones"
] | The application of neural network models to scientific machine learning tasks
has proliferated in recent years. In particular, neural network models have
proved to be adept at modeling processes with spatial-temporal complexity.
Nevertheless, these highly parameterized models have garnered skepticism in
their ability to produce outputs with quantified error bounds over the regimes
of interest. Hence there is a need to find uncertainty quantification methods
that are suitable for neural networks. In this work we present comparisons of
the parametric uncertainty quantification of neural networks modeling complex
spatial-temporal processes with Hamiltonian Monte Carlo and Stein variational
gradient descent and its projected variant. Specifically we apply these methods
to graph convolutional neural network models of evolving systems modeled with
recurrent neural network and neural ordinary differential equations
architectures. We show that Stein variational inference is a viable alternative
to Monte Carlo methods with some clear advantages for complex neural network
models. For our exemplars, Stein variational interference gave similar
uncertainty profiles through time compared to Hamiltonian Monte Carlo, albeit
with generally more generous variance.Projected Stein variational gradient
descent also produced similar uncertainty profiles to the non-projected
counterpart, but large reductions in the active weight space were confounded by
the stability of the neural network predictions and the convoluted likelihood
landscape. | [
"cs.LG",
"math.ST",
"physics.comp-ph",
"stat.TH"
] | false |
2402.11215 | 2024-02-17T07:49:50Z | AdAdaGrad: Adaptive Batch Size Schemes for Adaptive Gradient Methods | [
"Tim Tsz-Kit Lau",
"Han Liu",
"Mladen Kolar"
] | The choice of batch sizes in stochastic gradient optimizers is critical for
model training. However, the practice of varying batch sizes throughout the
training process is less explored compared to other hyperparameters. We
investigate adaptive batch size strategies derived from adaptive sampling
methods, traditionally applied only in stochastic gradient descent. Given the
significant interplay between learning rates and batch sizes, and considering
the prevalence of adaptive gradient methods in deep learning, we emphasize the
need for adaptive batch size strategies in these contexts. We introduce
AdAdaGrad and its scalar variant AdAdaGradNorm, which incrementally increase
batch sizes during training, while model updates are performed using AdaGrad
and AdaGradNorm. We prove that AdaGradNorm converges with high probability at a
rate of $\mathscr{O}(1/K)$ for finding a first-order stationary point of smooth
nonconvex functions within $K$ iterations. AdaGrad also demonstrates similar
convergence properties when integrated with a novel coordinate-wise variant of
our adaptive batch size strategies. Our theoretical claims are supported by
numerical experiments on various image classification tasks, highlighting the
enhanced adaptability of progressive batching protocols in deep learning and
the potential of such adaptive batch size strategies with adaptive gradient
optimizers in large-scale model training. | [
"cs.LG",
"math.OC",
"stat.ML"
] | false |
2402.11285 | 2024-02-17T13:57:20Z | Fair Resource Allocation in Virtualized O-RAN Platforms | [
"Fatih Aslan",
"George Iosifidis",
"Jose A. Ayala-Romero",
"Andres Garcia-Saavedra",
"Xavier Costa-Perez"
] | O-RAN systems and their deployment in virtualized general-purpose computing
platforms (O-Cloud) constitute a paradigm shift expected to bring unprecedented
performance gains. However, these architectures raise new implementation
challenges and threaten to worsen the already-high energy consumption of mobile
networks. This paper presents first a series of experiments which assess the
O-Cloud's energy costs and their dependency on the servers' hardware, capacity
and data traffic properties which, typically, change over time. Next, it
proposes a compute policy for assigning the base station data loads to O-Cloud
servers in an energy-efficient fashion; and a radio policy that determines at
near-real-time the minimum transmission block size for each user so as to avoid
unnecessary energy costs. The policies balance energy savings with performance,
and ensure that both of them are dispersed fairly across the servers and users,
respectively. To cater for the unknown and time-varying parameters affecting
the policies, we develop a novel online learning framework with fairness
guarantees that apply to the entire operation horizon of the system (long-term
fairness). The policies are evaluated using trace-driven simulations and are
fully implemented in an O-RAN compatible system where we measure the energy
costs and throughput in realistic scenarios. | [
"cs.NI",
"cs.AI",
"cs.LG"
] | false |
2402.11318 | 2024-02-17T16:16:24Z | BiasBuster: a Neural Approach for Accurate Estimation of Population
Statistics using Biased Location Data | [
"Sepanta Zeighami",
"Cyrus Shahabi"
] | While extremely useful (e.g., for COVID-19 forecasting and policy-making,
urban mobility analysis and marketing, and obtaining business insights),
location data collected from mobile devices often contain data from a biased
population subset, with some communities over or underrepresented in the
collected datasets. As a result, aggregate statistics calculated from such
datasets (as is done by various companies including Safegraph, Google, and
Facebook), while ignoring the bias, leads to an inaccurate representation of
population statistics. Such statistics will not only be generally inaccurate,
but the error will disproportionately impact different population subgroups
(e.g., because they ignore the underrepresented communities). This has dire
consequences, as these datasets are used for sensitive decision-making such as
COVID-19 policymaking. This paper tackles the problem of providing accurate
population statistics using such biased datasets. We show that statistical
debiasing, although in some cases useful, often fails to improve accuracy. We
then propose BiasBuster, a neural network approach that utilizes the
correlations between population statistics and location characteristics to
provide accurate estimates of population statistics. Extensive experiments on
real-world data show that BiasBuster improves accuracy by up to 2 times in
general and up to 3 times for underrepresented populations. | [
"cs.LG",
"cs.CY",
"cs.DB"
] | false |
2402.11338 | 2024-02-17T17:09:19Z | Fair Classification with Partial Feedback: An Exploration-Based
Data-Collection Approach | [
"Vijay Keswani",
"Anay Mehrotra",
"L. Elisa Celis"
] | In many predictive contexts (e.g., credit lending), true outcomes are only
observed for samples that were positively classified in the past. These past
observations, in turn, form training datasets for classifiers that make future
predictions. However, such training datasets lack information about the
outcomes of samples that were (incorrectly) negatively classified in the past
and can lead to erroneous classifiers. We present an approach that trains a
classifier using available data and comes with a family of exploration
strategies to collect outcome data about subpopulations that otherwise would
have been ignored. For any exploration strategy, the approach comes with
guarantees that (1) all sub-populations are explored, (2) the fraction of false
positives is bounded, and (3) the trained classifier converges to a "desired"
classifier. The right exploration strategy is context-dependent; it can be
chosen to improve learning guarantees and encode context-specific group
fairness properties. Evaluation on real-world datasets shows that this approach
consistently boosts the quality of collected outcome data and improves the
fraction of true positives for all groups, with only a small reduction in
predictive utility. | [
"cs.LG",
"cs.AI",
"cs.CY",
"stat.ML"
] | false |
2402.11345 | 2024-02-17T17:37:53Z | Variational Entropy Search for Adjusting Expected Improvement | [
"Nuojin Cheng",
"Stephen Becker"
] | Bayesian optimization is a widely used technique for optimizing black-box
functions, with Expected Improvement (EI) being the most commonly utilized
acquisition function in this domain. While EI is often viewed as distinct from
other information-theoretic acquisition functions, such as entropy search (ES)
and max-value entropy search (MES), our work reveals that EI can be considered
a special case of MES when approached through variational inference (VI). In
this context, we have developed the Variational Entropy Search (VES)
methodology and the VES-Gamma algorithm, which adapts EI by incorporating
principles from information-theoretic concepts. The efficacy of VES-Gamma is
demonstrated across a variety of test functions and read datasets, highlighting
its theoretical and practical utilities in Bayesian optimization scenarios. | [
"stat.ML",
"cs.LG",
"math.OC"
] | false |
2402.11365 | 2024-02-17T19:30:33Z | Data-Driven Stochastic AC-OPF using Gaussian Processes | [
"Mile Mitrovic"
] | The thesis focuses on developing a data-driven algorithm, based on machine
learning, to solve the stochastic alternating current (AC) chance-constrained
(CC) Optimal Power Flow (OPF) problem. Although the AC CC-OPF problem has been
successful in academic circles, it is highly nonlinear and computationally
demanding, which limits its practical impact. The proposed approach aims to
address this limitation and demonstrate its empirical efficiency through
applications to multiple IEEE test cases. To solve the non-convex and
computationally challenging CC AC-OPF problem, the proposed approach relies on
a machine learning Gaussian process regression (GPR) model. The full Gaussian
process (GP) approach is capable of learning a simple yet non-convex
data-driven approximation to the AC power flow equations that can incorporate
uncertain inputs. The proposed approach uses various approximations for
GP-uncertainty propagation. The full GP CC-OPF approach exhibits highly
competitive and promising results, outperforming the state-of-the-art
sample-based chance constraint approaches. To further improve the robustness
and complexity/accuracy trade-off of the full GP CC-OPF, a fast data-driven
setup is proposed. This setup relies on the sparse and hybrid Gaussian
processes (GP) framework to model the power flow equations with input
uncertainty. | [
"cs.LG",
"math.OC",
"stat.ML"
] | false |
2402.11384 | 2024-02-17T21:35:13Z | Reinforcement learning to maximise wind turbine energy generation | [
"Daniel Soler",
"Oscar Mariño",
"David Huergo",
"Martín de Frutos",
"Esteban Ferrer"
] | We propose a reinforcement learning strategy to control wind turbine energy
generation by actively changing the rotor speed, the rotor yaw angle and the
blade pitch angle. A double deep Q-learning with a prioritized experience
replay agent is coupled with a blade element momentum model and is trained to
allow control for changing winds. The agent is trained to decide the best
control (speed, yaw, pitch) for simple steady winds and is subsequently
challenged with real dynamic turbulent winds, showing good performance. The
double deep Q- learning is compared with a classic value iteration
reinforcement learning control and both strategies outperform a classic PID
control in all environments. Furthermore, the reinforcement learning approach
is well suited to changing environments including turbulent/gusty winds,
showing great adaptability. Finally, we compare all control strategies with
real winds and compute the annual energy production. In this case, the double
deep Q-learning algorithm also outperforms classic methodologies. | [
"cs.LG",
"math-ph",
"math.MP",
"math.OC"
] | false |
2402.11397 | 2024-02-17T22:40:22Z | Random Projection Neural Networks of Best Approximation: Convergence
theory and practical applications | [
"Gianluca Fabiani"
] | We investigate the concept of Best Approximation for Feedforward Neural
Networks (FNN) and explore their convergence properties through the lens of
Random Projection (RPNNs). RPNNs have predetermined and fixed, once and for
all, internal weights and biases, offering computational efficiency. We
demonstrate that there exists a choice of external weights, for any family of
such RPNNs, with non-polynomial infinitely differentiable activation functions,
that exhibit an exponential convergence rate when approximating any infinitely
differentiable function. For illustration purposes, we test the proposed
RPNN-based function approximation, with parsimoniously chosen basis functions,
across five benchmark function approximation problems. Results show that RPNNs
achieve comparable performance to established methods such as Legendre
Polynomials, highlighting their potential for efficient and accurate function
approximation. | [
"cs.LG",
"cs.NA",
"math.NA",
"41A25, 41A30, 41A46, 41A50, 41A52, 65D15, 65Y20, 68W20"
] | false |
2402.11228 | 2024-02-17T09:10:40Z | Adaptive Split Balancing for Optimal Random Forest | [
"Yuqian Zhang",
"Weijie Ji",
"Jelena Bradic"
] | While random forests are commonly used for regression problems, existing
methods often lack adaptability in complex situations or lose optimality under
simple, smooth scenarios. In this study, we introduce the adaptive split
balancing forest (ASBF), capable of learning tree representations from data
while simultaneously achieving minimax optimality under the Lipschitz class. To
exploit higher-order smoothness levels, we further propose a localized version
that attains the minimax rate under the H\"older class $\mathcal{H}^{q,\beta}$
for any $q\in\mathbb{N}$ and $\beta\in(0,1]$. Rather than relying on the
widely-used random feature selection, we consider a balanced modification to
existing approaches. Our results indicate that an over-reliance on auxiliary
randomness may compromise the approximation power of tree models, leading to
suboptimal results. Conversely, a less random, more balanced approach
demonstrates optimality. Additionally, we establish uniform upper bounds and
explore the application of random forests in average treatment effect
estimation problems. Through simulation studies and real-data applications, we
demonstrate the superior empirical performance of the proposed methods over
existing random forests. | [
"stat.ML",
"cs.LG",
"math.ST",
"stat.ME",
"stat.TH"
] | false |
2402.11435 | 2024-02-18T03:04:38Z | Momentor: Advancing Video Large Language Model with Fine-Grained
Temporal Reasoning | [
"Long Qian",
"Juncheng Li",
"Yu Wu",
"Yaobo Ye",
"Hao Fei",
"Tat-Seng Chua",
"Yueting Zhuang",
"Siliang Tang"
] | Large Language Models (LLMs) demonstrate remarkable proficiency in
comprehending and handling text-based tasks. Many efforts are being made to
transfer these attributes to video modality, which are termed Video-LLMs.
However, existing Video-LLMs can only capture the coarse-grained semantics and
are unable to effectively handle tasks related to comprehension or localization
of specific video segments. In light of these challenges, we propose Momentor,
a Video-LLM capable of accomplishing fine-grained temporal understanding tasks.
To support the training of Momentor, we design an automatic data generation
engine to construct Moment-10M, a large-scale video instruction dataset with
segment-level instruction data. We train Momentor on Moment-10M, enabling it to
perform segment-level reasoning and localization. Zero-shot evaluations on
several tasks demonstrate that Momentor excels in fine-grained temporally
grounded comprehension and localization. | [
"cs.CV"
] | false |
2402.11458 | 2024-02-18T04:59:23Z | Key Patch Proposer: Key Patches Contain Rich Information | [
"Jing Xu",
"Beiwen Tian",
"Hao Zhao"
] | In this paper, we introduce a novel algorithm named Key Patch Proposer (KPP)
designed to select key patches in an image without additional training. Our
experiments showcase KPP's robust capacity to capture semantic information by
both reconstruction and classification tasks. The efficacy of KPP suggests its
potential application in active learning for semantic segmentation. Our source
code is publicly available at https://github.com/CA-TT-AC/key-patch-proposer. | [
"cs.CV"
] | false |
2402.11473 | 2024-02-18T06:31:05Z | Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery
Detection | [
"Jiawei Liang",
"Siyuan Liang",
"Aishan Liu",
"Xiaojun Jia",
"Junhao Kuang",
"Xiaochun Cao"
] | The proliferation of face forgery techniques has raised significant concerns
within society, thereby motivating the development of face forgery detection
methods. These methods aim to distinguish forged faces from genuine ones and
have proven effective in practical applications. However, this paper introduces
a novel and previously unrecognized threat in face forgery detection scenarios
caused by backdoor attack. By embedding backdoors into models and incorporating
specific trigger patterns into the input, attackers can deceive detectors into
producing erroneous predictions for forged faces. To achieve this goal, this
paper proposes \emph{Poisoned Forgery Face} framework, which enables
clean-label backdoor attacks on face forgery detectors. Our approach involves
constructing a scalable trigger generator and utilizing a novel convolving
process to generate translation-sensitive trigger patterns. Moreover, we employ
a relative embedding method based on landmark-based regions to enhance the
stealthiness of the poisoned samples. Consequently, detectors trained on our
poisoned samples are embedded with backdoors. Notably, our approach surpasses
SoTA backdoor baselines with a significant improvement in attack success rate
(+16.39\% BD-AUC) and reduction in visibility (-12.65\% $L_\infty$).
Furthermore, our attack exhibits promising performance against backdoor
defenses. We anticipate that this paper will draw greater attention to the
potential threats posed by backdoor attacks in face forgery detection
scenarios. Our codes will be made available at
\url{https://github.com/JWLiang007/PFF} | [
"cs.CV"
] | false |
2402.11476 | 2024-02-18T06:54:51Z | EndoOOD: Uncertainty-aware Out-of-distribution Detection in Capsule
Endoscopy Diagnosis | [
"Qiaozhi Tan",
"Long Bai",
"Guankun Wang",
"Mobarakol Islam",
"Hongliang Ren"
] | Wireless capsule endoscopy (WCE) is a non-invasive diagnostic procedure that
enables visualization of the gastrointestinal (GI) tract. Deep learning-based
methods have shown effectiveness in disease screening using WCE data,
alleviating the burden on healthcare professionals. However, existing capsule
endoscopy classification methods mostly rely on pre-defined categories, making
it challenging to identify and classify out-of-distribution (OOD) data, such as
undefined categories or anatomical landmarks. To address this issue, we propose
the Endoscopy Out-of-Distribution (EndoOOD) framework, which aims to
effectively handle the OOD detection challenge in WCE diagnosis. The proposed
framework focuses on improving the robustness and reliability of WCE diagnostic
capabilities by incorporating uncertainty-aware mixup training and long-tailed
in-distribution (ID) data calibration techniques. Additionally, virtual-logit
matching is employed to accurately distinguish between OOD and ID data while
minimizing information loss. To assess the performance of our proposed
solution, we conduct evaluations and comparisons with 12 state-of-the-art
(SOTA) methods using two publicly available datasets. The results demonstrate
the effectiveness of the proposed framework in enhancing diagnostic accuracy
and supporting clinical decision-making. | [
"cs.CV"
] | false |
2402.11487 | 2024-02-18T07:28:37Z | Visual Concept-driven Image Generation with Text-to-Image Diffusion
Model | [
"Tanzila Rahman",
"Shweta Mahajan",
"Hsin-Ying Lee",
"Jian Ren",
"Sergey Tulyakov",
"Leonid Sigal"
] | Text-to-image (TTI) diffusion models have demonstrated impressive results in
generating high-resolution images of complex and imaginative scenes. Recent
approaches have further extended these methods with personalization techniques
that allow them to integrate user-illustrated concepts (e.g., the user
him/herself) using a few sample image illustrations. However, the ability to
generate images with multiple interacting concepts, such as human subjects, as
well as concepts that may be entangled in one, or across multiple, image
illustrations remains illusive. In this work, we propose a concept-driven TTI
personalization framework that addresses these core challenges. We build on
existing works that learn custom tokens for user-illustrated concepts, allowing
those to interact with existing text tokens in the TTI model. However,
importantly, to disentangle and better learn the concepts in question, we
jointly learn (latent) segmentation masks that disentangle these concepts in
user-provided image illustrations. We do so by introducing an Expectation
Maximization (EM)-like optimization procedure where we alternate between
learning the custom tokens and estimating masks encompassing corresponding
concepts in user-supplied images. We obtain these masks based on
cross-attention, from within the U-Net parameterized latent diffusion model and
subsequent Dense CRF optimization. We illustrate that such joint alternating
refinement leads to the learning of better tokens for concepts and, as a
bi-product, latent masks. We illustrate the benefits of the proposed approach
qualitatively and quantitatively (through user studies) with a number of
examples and use cases that can combine up to three entangled concepts. | [
"cs.CV"
] | false |
2402.11497 | 2024-02-18T07:56:29Z | Thyroid ultrasound diagnosis improvement via multi-view self-supervised
learning and two-stage pre-training | [
"Jian Wang",
"Xin Yang",
"Xiaohong Jia",
"Wufeng Xue",
"Rusi Chen",
"Yanlin Chen",
"Xiliang Zhu",
"Lian Liu",
"Yan Cao",
"Jianqiao Zhou",
"Dong Ni",
"Ning Gu"
] | Thyroid nodule classification and segmentation in ultrasound images are
crucial for computer-aided diagnosis; however, they face limitations owing to
insufficient labeled data. In this study, we proposed a multi-view contrastive
self-supervised method to improve thyroid nodule classification and
segmentation performance with limited manual labels. Our method aligns the
transverse and longitudinal views of the same nodule, thereby enabling the
model to focus more on the nodule area. We designed an adaptive loss function
that eliminates the limitations of the paired data. Additionally, we adopted a
two-stage pre-training to exploit the pre-training on ImageNet and thyroid
ultrasound images. Extensive experiments were conducted on a large-scale
dataset collected from multiple centers. The results showed that the proposed
method significantly improves nodule classification and segmentation
performance with limited manual labels and outperforms state-of-the-art
self-supervised methods. The two-stage pre-training also significantly exceeded
ImageNet pre-training. | [
"cs.CV"
] | false |
2402.11530 | 2024-02-18T10:09:10Z | Efficient Multimodal Learning from Data-centric Perspective | [
"Muyang He",
"Yexin Liu",
"Boya Wu",
"Jianhao Yuan",
"Yueze Wang",
"Tiejun Huang",
"Bo Zhao"
] | Multimodal Large Language Models (MLLMs) have demonstrated notable
capabilities in general visual understanding and reasoning tasks. However,
their deployment is hindered by substantial computational costs in both
training and inference, limiting accessibility to the broader research and user
communities. A straightforward solution is to leverage smaller pre-trained
vision and language models, which inevitably causes significant performance
drop. In this paper, we demonstrate the possibility to beat the scaling law and
train a smaller but better MLLM by exploring more informative training data.
Specifically, we introduce Bunny, a family of lightweight MLLMs with flexible
vision and language backbones for efficient multimodal learning from condensed
training data. Remarkably, our Bunny-3B outperforms the state-of-the-art large
MLLMs, especially LLaVA-v1.5-13B, on multiple benchmarks. The code, models and
data can be found in https://github.com/BAAI-DCAI/Bunny. | [
"cs.CV"
] | false |
2402.11540 | 2024-02-18T10:43:53Z | CPN: Complementary Proposal Network for Unconstrained Text Detection | [
"Longhuang Wu",
"Shangxuan Tian",
"Youxin Wang",
"Pengfei Xiong"
] | Existing methods for scene text detection can be divided into two paradigms:
segmentation-based and anchor-based. While Segmentation-based methods are
well-suited for irregular shapes, they struggle with compact or overlapping
layouts. Conversely, anchor-based approaches excel for complex layouts but
suffer from irregular shapes. To strengthen their merits and overcome their
respective demerits, we propose a Complementary Proposal Network (CPN) that
seamlessly and parallelly integrates semantic and geometric information for
superior performance. The CPN comprises two efficient networks for proposal
generation: the Deformable Morphology Semantic Network, which generates
semantic proposals employing an innovative deformable morphological operator,
and the Balanced Region Proposal Network, which produces geometric proposals
with pre-defined anchors. To further enhance the complementarity, we introduce
an Interleaved Feature Attention module that enables semantic and geometric
features to interact deeply before proposal generation. By leveraging both
complementary proposals and features, CPN outperforms state-of-the-art
approaches with significant margins under comparable computation cost.
Specifically, our approach achieves improvements of 3.6%, 1.3% and 1.0% on
challenging benchmarks ICDAR19-ArT, IC15, and MSRA-TD500, respectively. Code
for our method will be released. | [
"cs.CV"
] | false |
2402.11413 | 2024-02-18T01:01:13Z | A Multispectral Automated Transfer Technique (MATT) for machine-driven
image labeling utilizing the Segment Anything Model (SAM) | [
"James E. Gallagher",
"Aryav Gogia",
"Edward J. Oughton"
] | Segment Anything Model (SAM) is drastically accelerating the speed and
accuracy of automatically segmenting and labeling large Red-Green-Blue (RGB)
imagery datasets. However, SAM is unable to segment and label images outside of
the visible light spectrum, for example, for multispectral or hyperspectral
imagery. Therefore, this paper outlines a method we call the Multispectral
Automated Transfer Technique (MATT). By transposing SAM segmentation masks from
RGB images we can automatically segment and label multispectral imagery with
high precision and efficiency. For example, the results demonstrate that
segmenting and labeling a 2,400-image dataset utilizing MATT achieves a time
reduction of 87.8% in developing a trained model, reducing roughly 20 hours of
manual labeling, to only 2.4 hours. This efficiency gain is associated with
only a 6.7% decrease in overall mean average precision (mAP) when training
multispectral models via MATT, compared to a manually labeled dataset. We
consider this an acceptable level of precision loss when considering the time
saved during training, especially for rapidly prototyping experimental modeling
methods. This research greatly contributes to the study of multispectral object
detection by providing a novel and open-source method to rapidly segment,
label, and train multispectral object detection models with minimal human
interaction. Future research needs to focus on applying these methods to (i)
space-based multispectral, and (ii) drone-based hyperspectral imagery. | [
"cs.CV",
"cs.LG"
] | false |
2402.11424 | 2024-02-18T01:54:28Z | Data Distribution Distilled Generative Model for Generalized Zero-Shot
Recognition | [
"Yijie Wang",
"Mingjian Hong",
"Luwen Huangfu",
"Sheng Huang"
] | In the realm of Zero-Shot Learning (ZSL), we address biases in Generalized
Zero-Shot Learning (GZSL) models, which favor seen data. To counter this, we
introduce an end-to-end generative GZSL framework called D$^3$GZSL. This
framework respects seen and synthesized unseen data as in-distribution and
out-of-distribution data, respectively, for a more balanced model. D$^3$GZSL
comprises two core modules: in-distribution dual space distillation (ID$^2$SD)
and out-of-distribution batch distillation (O$^2$DBD). ID$^2$SD aligns
teacher-student outcomes in embedding and label spaces, enhancing learning
coherence. O$^2$DBD introduces low-dimensional out-of-distribution
representations per batch sample, capturing shared structures between seen and
unseen categories. Our approach demonstrates its effectiveness across
established GZSL benchmarks, seamlessly integrating into mainstream generative
frameworks. Extensive experiments consistently showcase that D$^3$GZSL elevates
the performance of existing generative GZSL methods, underscoring its potential
to refine zero-shot learning practices.The code is available at:
https://github.com/PJBQ/D3GZSL.git | [
"cs.CV",
"cs.AI"
] | false |
2402.11507 | 2024-02-18T08:34:15Z | MAL: Motion-Aware Loss with Temporal and Distillation Hints for
Self-Supervised Depth Estimation | [
"Yup-Jiang Dong",
"Fang-Lue Zhang",
"Song-Hai Zhang"
] | Depth perception is crucial for a wide range of robotic applications.
Multi-frame self-supervised depth estimation methods have gained research
interest due to their ability to leverage large-scale, unlabeled real-world
data. However, the self-supervised methods often rely on the assumption of a
static scene and their performance tends to degrade in dynamic environments. To
address this issue, we present Motion-Aware Loss, which leverages the temporal
relation among consecutive input frames and a novel distillation scheme between
the teacher and student networks in the multi-frame self-supervised depth
estimation methods. Specifically, we associate the spatial locations of moving
objects with the temporal order of input frames to eliminate errors induced by
object motion. Meanwhile, we enhance the original distillation scheme in
multi-frame methods to better exploit the knowledge from a teacher network. MAL
is a novel, plug-and-play module designed for seamless integration into
multi-frame self-supervised monocular depth estimation methods. Adding MAL into
previous state-of-the-art methods leads to a reduction in depth estimation
errors by up to 4.2% and 10.8% on KITTI and CityScapes benchmarks,
respectively. | [
"cs.CV",
"cs.RO"
] | false |
2402.11510 | 2024-02-18T08:51:29Z | Underestimation of lung regions on chest X-ray segmentation masks
assessed by comparison with total lung volume evaluated on computed
tomography | [
"Przemysław Bombiński",
"Patryk Szatkowski",
"Bartłomiej Sobieski",
"Tymoteusz Kwieciński",
"Szymon Płotka",
"Mariusz Adamek",
"Marcin Banasiuk",
"Mariusz I. Furmanek",
"Przemysław Biecek"
] | Lung mask creation lacks well-defined criteria and standardized guidelines,
leading to a high degree of subjectivity between annotators. In this study, we
assess the underestimation of lung regions on chest X-ray segmentation masks
created according to the current state-of-the-art method, by comparison with
total lung volume evaluated on computed tomography (CT). We show, that lung
X-ray masks created by following the contours of the heart, mediastinum, and
diaphragm significantly underestimate lung regions and exclude substantial
portions of the lungs from further assessment, which may result in numerous
clinical errors. | [
"eess.IV",
"cs.CV"
] | false |
2402.11520 | 2024-02-18T09:22:58Z | Cross-Attention Fusion of Visual and Geometric Features for Large
Vocabulary Arabic Lipreading | [
"Samar Daou",
"Ahmed Rekik",
"Achraf Ben-Hamadou",
"Abdelaziz Kallel"
] | Lipreading involves using visual data to recognize spoken words by analyzing
the movements of the lips and surrounding area. It is a hot research topic with
many potential applications, such as human-machine interaction and enhancing
audio speech recognition. Recent deep-learning based works aim to integrate
visual features extracted from the mouth region with landmark points on the lip
contours. However, employing a simple combination method such as concatenation
may not be the most effective approach to get the optimal feature vector. To
address this challenge, firstly, we propose a cross-attention fusion-based
approach for large lexicon Arabic vocabulary to predict spoken words in videos.
Our method leverages the power of cross-attention networks to efficiently
integrate visual and geometric features computed on the mouth region. Secondly,
we introduce the first large-scale Lip Reading in the Wild for Arabic (LRW-AR)
dataset containing 20,000 videos for 100-word classes, uttered by 36 speakers.
The experimental results obtained on LRW-AR and ArabicVisual databases showed
the effectiveness and robustness of the proposed approach in recognizing Arabic
words. Our work provides insights into the feasibility and effectiveness of
applying lipreading techniques to the Arabic language, opening doors for
further research in this field. Link to the project page:
https://crns-smartvision.github.io/lrwar | [
"cs.CV",
"cs.MM"
] | false |
2402.11557 | 2024-02-18T11:57:01Z | Evaluating Adversarial Robustness of Low dose CT Recovery | [
"Kanchana Vaishnavi Gandikota",
"Paramanand Chandramouli",
"Hannah Droege",
"Michael Moeller"
] | Low dose computed tomography (CT) acquisition using reduced radiation or
sparse angle measurements is recommended to decrease the harmful effects of
X-ray radiation. Recent works successfully apply deep networks to the problem
of low dose CT recovery on bench-mark datasets. However, their robustness needs
a thorough evaluation before use in clinical settings. In this work, we
evaluate the robustness of different deep learning approaches and classical
methods for CT recovery. We show that deep networks, including model-based
networks encouraging data consistency, are more susceptible to untargeted
attacks. Surprisingly, we observe that data consistency is not heavily affected
even for these poor quality reconstructions, motivating the need for better
regularization for the networks. We demonstrate the feasibility of universal
attacks and study attack transferability across different methods. We analyze
robustness to attacks causing localized changes in clinically relevant regions.
Both classical approaches and deep networks are affected by such attacks
leading to changes in the visual appearance of localized lesions, for extremely
small perturbations. As the resulting reconstructions have high data
consistency with the original measurements, these localized attacks can be used
to explore the solution space of the CT recovery problem. | [
"eess.IV",
"cs.CV"
] | false |
2402.11574 | 2024-02-18T12:43:38Z | Visual In-Context Learning for Large Vision-Language Models | [
"Yucheng Zhou",
"Xiang Li",
"Qianning Wang",
"Jianbing Shen"
] | In Large Visual Language Models (LVLMs), the efficacy of In-Context Learning
(ICL) remains limited by challenges in cross-modal interactions and
representation disparities. To overcome these challenges, we introduce a novel
Visual In-Context Learning (VICL) method comprising Visual Demonstration
Retrieval, Intent-Oriented Image Summarization, and Intent-Oriented
Demonstration Composition. Our approach retrieves images via ''Retrieval &
Rerank'' paradigm, summarises images with task intent and task-specific visual
parsing, and composes language-based demonstrations that reduce token count and
alleviate cross-modal interaction problem. Experimental evaluations on five
visual reasoning datasets demonstrate the effectiveness of our method.
Moreover, our extensive experiments leverage information flow analysis to
elucidate the effectiveness of our method, and investigate the impact of length
and position of demonstrations for LVLM. The use of in-context unlearning
further shows promise in resetting specific model knowledge without retraining. | [
"cs.CV",
"cs.CL"
] | false |
2402.11627 | 2024-02-18T16:01:28Z | Interactive Garment Recommendation with User in the Loop | [
"Federico Becattini",
"Xiaolin Chen",
"Andrea Puccia",
"Haokun Wen",
"Xuemeng Song",
"Liqiang Nie",
"Alberto Del Bimbo"
] | Recommending fashion items often leverages rich user profiles and makes
targeted suggestions based on past history and previous purchases. In this
paper, we work under the assumption that no prior knowledge is given about a
user. We propose to build a user profile on the fly by integrating user
reactions as we recommend complementary items to compose an outfit. We present
a reinforcement learning agent capable of suggesting appropriate garments and
ingesting user feedback so to improve its recommendations and maximize user
satisfaction. To train such a model, we resort to a proxy model to be able to
simulate having user feedback in the training loop. We experiment on the
IQON3000 fashion dataset and we find that a reinforcement learning-based agent
becomes capable of improving its recommendations by taking into account
personal preferences. Furthermore, such task demonstrated to be hard for
non-reinforcement models, that cannot exploit exploration during training. | [
"cs.CV",
"cs.IR"
] | false |
2402.11631 | 2024-02-18T16:17:25Z | Neuromorphic Face Analysis: a Survey | [
"Federico Becattini",
"Lorenzo Berlincioni",
"Luca Cultrera",
"Alberto Del Bimbo"
] | Neuromorphic sensors, also known as event cameras, are a class of imaging
devices mimicking the function of biological visual systems. Unlike traditional
frame-based cameras, which capture fixed images at discrete intervals,
neuromorphic sensors continuously generate events that represent changes in
light intensity or motion in the visual field with high temporal resolution and
low latency. These properties have proven to be interesting in modeling human
faces, both from an effectiveness and a privacy-preserving point of view.
Neuromorphic face analysis however is still a raw and unstructured field of
research, with several attempts at addressing different tasks with no clear
standard or benchmark. This survey paper presents a comprehensive overview of
capabilities, challenges and emerging applications in the domain of
neuromorphic face analysis, to outline promising directions and open issues.
After discussing the fundamental working principles of neuromorphic vision and
presenting an in-depth overview of the related research, we explore the current
state of available data, standard data representations, emerging challenges,
and limitations that require further investigation. This paper aims to
highlight the recent process in this evolving field to provide to both
experienced and newly come researchers an all-encompassing analysis of the
state of the art along with its problems and shortcomings. | [
"cs.CV",
"cs.ET"
] | false |
2402.11670 | 2024-02-18T18:16:43Z | Challenging the Black Box: A Comprehensive Evaluation of Attribution
Maps of CNN Applications in Agriculture and Forestry | [
"Lars Nieradzik",
"Henrike Stephani",
"Jördis Sieburg-Rockel",
"Stephanie Helmling",
"Andrea Olbrich",
"Janis Keuper"
] | In this study, we explore the explainability of neural networks in
agriculture and forestry, specifically in fertilizer treatment classification
and wood identification. The opaque nature of these models, often considered
'black boxes', is addressed through an extensive evaluation of state-of-the-art
Attribution Maps (AMs), also known as class activation maps (CAMs) or saliency
maps. Our comprehensive qualitative and quantitative analysis of these AMs
uncovers critical practical limitations. Findings reveal that AMs frequently
fail to consistently highlight crucial features and often misalign with the
features considered important by domain experts. These discrepancies raise
substantial questions about the utility of AMs in understanding the
decision-making process of neural networks. Our study provides critical
insights into the trustworthiness and practicality of AMs within the
agriculture and forestry sectors, thus facilitating a better understanding of
neural networks in these application areas. | [
"cs.CV",
"cs.LG"
] | false |
2402.11682 | 2024-02-18T19:12:18Z | Learning Conditional Invariances through Non-Commutativity | [
"Abhra Chaudhuri",
"Serban Georgescu",
"Anjan Dutta"
] | Invariance learning algorithms that conditionally filter out domain-specific
random variables as distractors, do so based only on the data semantics, and
not the target domain under evaluation. We show that a provably optimal and
sample-efficient way of learning conditional invariances is by relaxing the
invariance criterion to be non-commutatively directed towards the target
domain. Under domain asymmetry, i.e., when the target domain contains
semantically relevant information absent in the source, the risk of the encoder
$\varphi^*$ that is optimal on average across domains is strictly lower-bounded
by the risk of the target-specific optimal encoder $\Phi^*_\tau$. We prove that
non-commutativity steers the optimization towards $\Phi^*_\tau$ instead of
$\varphi^*$, bringing the $\mathcal{H}$-divergence between domains down to
zero, leading to a stricter bound on the target risk. Both our theory and
experiments demonstrate that non-commutative invariance (NCI) can leverage
source domain samples to meet the sample complexity needs of learning
$\Phi^*_\tau$, surpassing SOTA invariance learning algorithms for domain
adaptation, at times by over $2\%$, approaching the performance of an oracle.
Implementation is available at https://github.com/abhrac/nci. | [
"cs.LG",
"cs.CV"
] | false |
2402.11690 | 2024-02-18T19:38:44Z | Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning | [
"Zhiyang Xu",
"Chao Feng",
"Rulin Shao",
"Trevor Ashby",
"Ying Shen",
"Di Jin",
"Yu Cheng",
"Qifan Wang",
"Lifu Huang"
] | Despite vision-language models' (VLMs) remarkable capabilities as versatile
visual assistants, two substantial challenges persist within the existing VLM
frameworks: (1) lacking task diversity in pretraining and visual instruction
tuning, and (2) annotation error and bias in GPT-4 synthesized instruction
tuning data. Both challenges lead to issues such as poor generalizability,
hallucination, and catastrophic forgetting. To address these challenges, we
construct Vision-Flan, the most diverse publicly available visual instruction
tuning dataset to date, comprising 187 diverse tasks and 1,664,261 instances
sourced from academic datasets, and each task is accompanied by an
expert-written instruction. In addition, we propose a two-stage instruction
tuning framework, in which VLMs are firstly finetuned on Vision-Flan and
further tuned on GPT-4 synthesized data. We find this two-stage tuning
framework significantly outperforms the traditional single-stage visual
instruction tuning framework and achieves the state-of-the-art performance
across a wide range of multi-modal evaluation benchmarks. Finally, we conduct
in-depth analyses to understand visual instruction tuning and our findings
reveal that: (1) GPT-4 synthesized data does not substantially enhance VLMs'
capabilities but rather modulates the model's responses to human-preferred
formats; (2) A minimal quantity (e.g., 1,000) of GPT-4 synthesized data can
effectively align VLM responses with human-preference; (3) Visual instruction
tuning mainly helps large-language models (LLMs) to understand visual features. | [
"cs.CL",
"cs.CV"
] | true |
2402.11735 | 2024-02-18T23:29:28Z | LiRaFusion: Deep Adaptive LiDAR-Radar Fusion for 3D Object Detection | [
"Jingyu Song",
"Lingjun Zhao",
"Katherine A. Skinner"
] | We propose LiRaFusion to tackle LiDAR-radar fusion for 3D object detection to
fill the performance gap of existing LiDAR-radar detectors. To improve the
feature extraction capabilities from these two modalities, we design an early
fusion module for joint voxel feature encoding, and a middle fusion module to
adaptively fuse feature maps via a gated network. We perform extensive
evaluation on nuScenes to demonstrate that LiRaFusion leverages the
complementary information of LiDAR and radar effectively and achieves notable
improvement over existing methods. | [
"cs.RO",
"cs.CV"
] | false |
2402.11411 | 2024-02-18T00:56:16Z | Aligning Modalities in Vision Large Language Models via Preference
Fine-tuning | [
"Yiyang Zhou",
"Chenhang Cui",
"Rafael Rafailov",
"Chelsea Finn",
"Huaxiu Yao"
] | Instruction-following Vision Large Language Models (VLLMs) have achieved
significant progress recently on a variety of tasks. These approaches merge
strong pre-trained vision models and large language models (LLMs). Since these
components are trained separately, the learned representations need to be
aligned with joint training on additional image-language pairs. This procedure
is not perfect and can cause the model to hallucinate - provide answers that do
not accurately reflect the image, even when the core LLM is highly factual and
the vision backbone has sufficiently complete representations. In this work, we
frame the hallucination problem as an alignment issue, tackle it with
preference tuning. Specifically, we propose POVID to generate feedback data
with AI models. We use ground-truth instructions as the preferred response and
a two-stage approach to generate dispreferred data. First, we prompt GPT-4V to
inject plausible hallucinations into the correct answer. Second, we distort the
image to trigger the inherent hallucination behavior of the VLLM. This is an
automated approach, which does not rely on human data generation or require a
perfect expert, which makes it easily scalable. Finally, both of these
generation strategies are integrated into an RLHF pipeline via Direct
Preference Optimization. In experiments across broad benchmarks, we show that
we can not only reduce hallucinations, but improve model performance across
standard benchmarks, outperforming prior approaches. Our data and code are
available at https://github.com/YiyangZhou/POVID. | [
"cs.LG",
"cs.CL",
"cs.CV"
] | false |
2402.11622 | 2024-02-18T15:28:39Z | Logical Closed Loop: Uncovering Object Hallucinations in Large
Vision-Language Models | [
"Junfei Wu",
"Qiang Liu",
"Ding Wang",
"Jinghao Zhang",
"Shu Wu",
"Liang Wang",
"Tieniu Tan"
] | Object hallucination has been an Achilles' heel which hinders the broader
applications of large vision-language models (LVLMs). Object hallucination
refers to the phenomenon that the LVLMs claim non-existent objects in the
image. To mitigate the object hallucinations, instruction tuning and external
model-based detection methods have been proposed, which either require
large-scare computational resources or depend on the detection result of
external models. However, there remains an under-explored field to utilize the
LVLM itself to alleviate object hallucinations. In this work, we adopt the
intuition that the LVLM tends to respond logically consistently for existent
objects but inconsistently for hallucinated objects. Therefore, we propose a
Logical Closed Loop-based framework for Object Hallucination Detection and
Mitigation, namely LogicCheckGPT. In specific, we devise logical consistency
probing to raise questions with logical correlations, inquiring about
attributes from objects and vice versa. Whether their responses can form a
logical closed loop serves as an indicator of object hallucination. As a
plug-and-play method, it can be seamlessly applied to all existing LVLMs.
Comprehensive experiments conducted on three benchmarks across four LVLMs have
demonstrated significant improvements brought by our method, indicating its
effectiveness and generality. | [
"cs.CV",
"cs.AI",
"cs.CL",
"cs.LG"
] | false |
2402.11680 | 2024-02-18T19:08:19Z | 3D Point Cloud Compression with Recurrent Neural Network and Image
Compression Methods | [
"Till Beemelmanns",
"Yuchen Tao",
"Bastian Lampe",
"Lennart Reiher",
"Raphael van Kempen",
"Timo Woopen",
"Lutz Eckstein"
] | Storing and transmitting LiDAR point cloud data is essential for many AV
applications, such as training data collection, remote control, cloud services
or SLAM. However, due to the sparsity and unordered structure of the data, it
is difficult to compress point cloud data to a low volume. Transforming the raw
point cloud data into a dense 2D matrix structure is a promising way for
applying compression algorithms. We propose a new lossless and calibrated
3D-to-2D transformation which allows compression algorithms to efficiently
exploit spatial correlations within the 2D representation. To compress the
structured representation, we use common image compression methods and also a
self-supervised deep compression approach using a recurrent neural network. We
also rearrange the LiDAR's intensity measurements to a dense 2D representation
and propose a new metric to evaluate the compression performance of the
intensity. Compared to approaches that are based on generic octree point cloud
compression or based on raw point cloud data compression, our approach achieves
the best quantitative and visual performance. Source code and dataset are
available at https://github.com/ika-rwth-aachen/Point-Cloud-Compression. | [
"cs.CV",
"cs.AI",
"eess.IV"
] | false |
2402.11733 | 2024-02-18T23:14:40Z | The Effectiveness of Random Forgetting for Robust Generalization | [
"Vijaya Raghavan T Ramkumar",
"Bahram Zonooz",
"Elahe Arani"
] | Deep neural networks are susceptible to adversarial attacks, which can
compromise their performance and accuracy. Adversarial Training (AT) has
emerged as a popular approach for protecting neural networks against such
attacks. However, a key challenge of AT is robust overfitting, where the
network's robust performance on test data deteriorates with further training,
thus hindering generalization. Motivated by the concept of active forgetting in
the brain, we introduce a novel learning paradigm called "Forget to Mitigate
Overfitting (FOMO)". FOMO alternates between the forgetting phase, which
randomly forgets a subset of weights and regulates the model's information
through weight reinitialization, and the relearning phase, which emphasizes
learning generalizable features. Our experiments on benchmark datasets and
adversarial attacks show that FOMO alleviates robust overfitting by
significantly reducing the gap between the best and last robust test accuracy
while improving the state-of-the-art robustness. Furthermore, FOMO provides a
better trade-off between standard and robust accuracy, outperforming baseline
adversarial methods. Finally, our framework is robust to AutoAttacks and
increases generalization in many real-world scenarios. | [
"cs.LG",
"cs.AI",
"cs.CV"
] | false |
2402.12406 | 2024-02-18T08:13:57Z | Teacher as a Lenient Expert: Teacher-Agnostic Data-Free Knowledge
Distillation | [
"Hyunjune Shin",
"Dong-Wan Choi"
] | Data-free knowledge distillation (DFKD) aims to distill pretrained knowledge
to a student model with the help of a generator without using original data. In
such data-free scenarios, achieving stable performance of DFKD is essential due
to the unavailability of validation data. Unfortunately, this paper has
discovered that existing DFKD methods are quite sensitive to different teacher
models, occasionally showing catastrophic failures of distillation, even when
using well-trained teacher models. Our observation is that the generator in
DFKD is not always guaranteed to produce precise yet diverse samples using the
existing representative strategy of minimizing both class-prior and adversarial
losses. Through our empirical study, we focus on the fact that class-prior not
only decreases the diversity of generated samples, but also cannot completely
address the problem of generating unexpectedly low-quality samples depending on
teacher models. In this paper, we propose the teacher-agnostic data-free
knowledge distillation (TA-DFKD) method, with the goal of more robust and
stable performance regardless of teacher models. Our basic idea is to assign
the teacher model a lenient expert role for evaluating samples, rather than a
strict supervisor that enforces its class-prior on the generator. Specifically,
we design a sample selection approach that takes only clean samples verified by
the teacher model without imposing restrictions on the power of generating
diverse samples. Through extensive experiments, we show that our method
successfully achieves both robustness and training stability across various
teacher models, while outperforming the existing DFKD methods. | [
"cs.LG",
"cs.AI",
"cs.CV"
] | false |
2402.12407 | 2024-02-18T10:49:23Z | Accelerating local laplacian filters on FPGAs | [
"Shashwat Khandelwal",
"Ziaul Choudhury",
"Shashwat Shrivastava",
"Suresh Purini"
] | Images when processed using various enhancement techniques often lead to edge
degradation and other unwanted artifacts such as halos. These artifacts pose a
major problem for photographic applications where they can denude the quality
of an image. There is a plethora of edge-aware techniques proposed in the field
of image processing. However, these require the application of complex
optimization or post-processing methods. Local Laplacian Filtering is an
edge-aware image processing technique that involves the construction of simple
Gaussian and Laplacian pyramids. This technique can be successfully applied for
detail smoothing, detail enhancement, tone mapping and inverse tone mapping of
an image while keeping it artifact-free. The problem though with this approach
is that it is computationally expensive. Hence, parallelization schemes using
multi-core CPUs and GPUs have been proposed. As is well known, they are not
power-efficient, and a well-designed hardware architecture on an FPGA can do
better on the performance per watt metric. In this paper, we propose a hardware
accelerator, which exploits fully the available parallelism in the Local
Laplacian Filtering algorithm, while minimizing the utilization of on-chip FPGA
resources. On Virtex-7 FPGA, we obtain a 7.5x speed-up to process a 1 MB image
when compared to an optimized baseline CPU implementation. To the best of our
knowledge, we are not aware of any other hardware accelerators proposed in the
research literature for the Local Laplacian Filtering problem. | [
"eess.IV",
"cs.CV",
"cs.GR",
"eess.SP"
] | false |
2402.11414 | 2024-02-18T01:03:25Z | Fine-grained and Explainable Factuality Evaluation for Multimodal
Summarization | [
"Liqiang Jing",
"Jingxuan Zuo",
"Yue Zhang"
] | Multimodal summarization aims to generate a concise summary based on the
input text and image. However, the existing methods potentially suffer from
unfactual output. To evaluate the factuality of multimodal summarization
models, we propose two fine-grained and explainable evaluation frameworks
(FALLACIOUS) for different application scenarios, i.e. reference-based
factuality evaluation framework and reference-free factuality evaluation
framework. Notably, the reference-free factuality evaluation framework doesn't
need ground truth and hence it has a wider application scenario. To evaluate
the effectiveness of the proposed frameworks, we compute the correlation
between our frameworks and the other metrics. The experimental results show the
effectiveness of our proposed method. We will release our code and dataset via
github. | [
"cs.CL"
] | false |
2402.11420 | 2024-02-18T01:40:34Z | Rethinking the Roles of Large Language Models in Chinese Grammatical
Error Correction | [
"Yinghui Li",
"Shang Qin",
"Jingheng Ye",
"Shirong Ma",
"Yangning Li",
"Libo Qin",
"Xuming Hu",
"Wenhao Jiang",
"Hai-Tao Zheng",
"Philip S. Yu"
] | Recently, Large Language Models (LLMs) have been widely studied by
researchers for their roles in various downstream NLP tasks. As a fundamental
task in the NLP field, Chinese Grammatical Error Correction (CGEC) aims to
correct all potential grammatical errors in the input sentences. Previous
studies have shown that LLMs' performance as correctors on CGEC remains
unsatisfactory due to its challenging task focus. To promote the CGEC field to
better adapt to the era of LLMs, we rethink the roles of LLMs in the CGEC task
so that they can be better utilized and explored in CGEC. Considering the rich
grammatical knowledge stored in LLMs and their powerful semantic understanding
capabilities, we utilize LLMs as explainers to provide explanation information
for the CGEC small models during error correction to enhance performance. We
also use LLMs as evaluators to bring more reasonable CGEC evaluations, thus
alleviating the troubles caused by the subjectivity of the CGEC task. In
particular, our work is also an active exploration of how LLMs and small models
better collaborate in downstream tasks. Extensive experiments and detailed
analyses on widely used datasets verify the effectiveness of our thinking
intuition and the proposed methods. | [
"cs.CL"
] | false |
2402.11422 | 2024-02-18T01:46:46Z | Mitigating Catastrophic Forgetting in Multi-domain Chinese Spelling
Correction by Multi-stage Knowledge Transfer Framework | [
"Peng Xing",
"Yinghui Li",
"Shirong Ma",
"Xinnian Liang",
"Haojing Huang",
"Yangning Li",
"Hai-Tao Zheng",
"Wenhao Jiang",
"Ying Shen"
] | Chinese Spelling Correction (CSC) aims to detect and correct spelling errors
in given sentences. Recently, multi-domain CSC has gradually attracted the
attention of researchers because it is more practicable. In this paper, we
focus on the key flaw of the CSC model when adapting to multi-domain scenarios:
the tendency to forget previously acquired knowledge upon learning new
domain-specific knowledge (i.e., catastrophic forgetting). To address this, we
propose a novel model-agnostic Multi-stage Knowledge Transfer (MKT) framework,
which utilizes a continuously evolving teacher model for knowledge transfer in
each domain, rather than focusing solely on new domain knowledge. It deserves
to be mentioned that we are the first to apply continual learning methods to
the multi-domain CSC task. Experiments prove the effectiveness of our proposed
method, and further analyses demonstrate the importance of overcoming
catastrophic forgetting for improving the model performance. | [
"cs.CL"
] | false |
2402.11430 | 2024-02-18T02:41:06Z | EventRL: Enhancing Event Extraction with Outcome Supervision for Large
Language Models | [
"Jun Gao",
"Huan Zhao",
"Wei Wang",
"Changlong Yu",
"Ruifeng Xu"
] | In this study, we present EventRL, a reinforcement learning approach
developed to enhance event extraction for large language models (LLMs). EventRL
utilizes outcome supervision with specific reward functions to tackle prevalent
challenges in LLMs, such as instruction following and hallucination, manifested
as the mismatch of event structure and the generation of undefined event types.
We evaluate EventRL against existing methods like Few-Shot Prompting (FSP)
(based on GPT4) and Supervised Fine-Tuning (SFT) across various LLMs, including
GPT-4, LLaMa, and CodeLLaMa models. Our findings show that EventRL
significantly outperforms these conventional approaches by improving the
performance in identifying and structuring events, particularly in handling
novel event types. The study emphasizes the critical role of reward function
selection and demonstrates the benefits of incorporating code data for better
event extraction. While increasing model size leads to higher accuracy,
maintaining the ability to generalize is essential to avoid overfitting. | [
"cs.CL"
] | false |
2402.11432 | 2024-02-18T02:52:54Z | Can Deception Detection Go Deeper? Dataset, Evaluation, and Benchmark
for Deception Reasoning | [
"Kang Chen",
"Zheng Lian",
"Haiyang Sun",
"Bin Liu",
"Jianhua Tao"
] | Deception detection has attracted increasing attention due to its importance
in many practical scenarios. Currently, data scarcity harms the development of
this field. On the one hand, it is costly to hire participants to simulate
deception scenarios. On the other hand, it is difficult to collect videos
containing deceptive behaviors on the Internet. To address data scarcity, this
paper proposes a new data collection pipeline. Specifically, we use GPT-4 to
simulate a role-play between a suspect and a police officer. During
interrogation, the suspect lies to the police officer to evade responsibility
for the crime, while the police officer uncovers the truth and gathers
evidence. Compared with previous datasets, this strategy reduces data
collection costs, providing a promising way to increase the dataset size.
Meanwhile, we extend the traditional deception detection task to deception
reasoning, further providing evidence for deceptive parts. This dataset can
also be used to evaluate the complex reasoning capability of current large
language models and serve as a reasoning benchmark for further research. | [
"cs.CL"
] | false |
2402.11442 | 2024-02-18T03:38:51Z | Can LLMs Reason with Rules? Logic Scaffolding for Stress-Testing and
Improving LLMs | [
"Siyuan Wang",
"Zhongyu Wei",
"Yejin Choi",
"Xiang Ren"
] | Large language models (LLMs) have achieved impressive human-like performance
across various reasoning tasks. However, their mastery of underlying
inferential rules still falls short of human capabilities. To investigate this,
we propose a logic scaffolding inferential rule generation framework, to
construct an inferential rule base, ULogic, comprising both primitive and
compositional rules across five domains. Our analysis of GPT-series models over
a rule subset reveals significant gaps in LLMs' logic understanding compared to
human performance, especially in compositional and structural complex rules
with certain bias patterns. We further distill these rules into a smaller-scale
inference engine for flexible rule generation and enhancing downstream
reasoning. Through a multi-judger evaluation, our inference engine proves
effective in generating accurate, complex and abstract conclusions and
premises, and improve various commonsense reasoning tasks. Overall, our work
sheds light on LLMs' limitations in grasping inferential rule and suggests ways
to enhance their logical reasoning abilities~\footnote{Code and data are
available at \url{https://github.com/SiyuanWangw/ULogic}.}. | [
"cs.CL"
] | false |
2402.11443 | 2024-02-18T03:40:06Z | Benchmark Self-Evolving: A Multi-Agent Framework for Dynamic LLM
Evaluation | [
"Siyuan Wang",
"Zhuohan Long",
"Zhihao Fan",
"Zhongyu Wei",
"Xuanjing Huang"
] | This paper presents a benchmark self-evolving framework to dynamically
evaluate rapidly advancing Large Language Models (LLMs), aiming for a more
accurate assessment of their capabilities and limitations. We utilize a
multi-agent system to manipulate the context or question of original instances,
reframing new evolving instances with high confidence that dynamically extend
existing benchmarks. Towards a more scalable, robust and fine-grained
evaluation, we implement six reframing operations to construct evolving
instances testing LLMs against diverse queries, data noise and probing their
problem-solving sub-abilities. With this framework, we extend benchmark
datasets of four tasks. Experimental results show a general performance decline
in most LLMs against their original results. This decline under our scalable
and robust evaluations, alongside our fine-grained evaluation, more accurately
reflect models' capabilities. Besides, our framework widens performance
discrepancies both between different models and within the same model across
various tasks, facilitating more informed model selection for specific tasks
(Code and data are available at
https://github.com/NanshineLoong/Self-Evolving-Benchmark). | [
"cs.CL"
] | false |
2402.11447 | 2024-02-18T04:08:10Z | In-Context Example Ordering Guided by Label Distributions | [
"Zhichao Xu",
"Daniel Cohen",
"Bei Wang",
"Vivek Srikumar"
] | By allowing models to predict without task-specific training, in-context
learning (ICL) with pretrained LLMs has enormous potential in NLP. However, a
number of problems persist in ICL. In particular, its performance is sensitive
to the choice and order of in-context examples. Given the same set of
in-context examples with different orderings, model performance may vary
between near random to near state-of-the-art. In this work, we formulate
in-context example ordering as an optimization problem. We examine three
problem settings that differ in the assumptions they make about what is known
about the task. Inspired by the idea of learning from label proportions, we
propose two principles for in-context example ordering guided by model's
probability predictions. We apply our proposed principles to thirteen text
classification datasets and nine different autoregressive LLMs with 700M to 13B
parameters. We demonstrate that our approach outperforms the baselines by
improving the classification accuracy, reducing model miscalibration, and also
by selecting better in-context examples. | [
"cs.CL"
] | false |
2402.11452 | 2024-02-18T04:28:16Z | AutoPRM: Automating Procedural Supervision for Multi-Step Reasoning via
Controllable Question Decomposition | [
"Zhaorun Chen",
"Zhuokai Zhao",
"Zhihong Zhu",
"Ruiqi Zhang",
"Xiang Li",
"Bhiksha Raj",
"Huaxiu Yao"
] | Recent advancements in large language models (LLMs) have shown promise in
multi-step reasoning tasks, yet their reliance on extensive manual labeling to
provide procedural feedback remains a significant impediment. To address this
challenge, in this paper, we propose a novel self-supervised framework AutoPRM
that efficiently enhances the fine-tuning of LLMs for intricate reasoning
challenges. Specifically, AutoPRM first decomposes complex problems into more
manageable subquestions with a controllable granularity switch, then
sequentially apply reinforcement learning to iteratively improve the
subquestion solver. Additionally, we propose context-guided-decoding to avoid
reward tampering and guide the subquestion solver towards the solution of the
holistic problem. Extensive experiments show that AutoPRM significantly
improves performance on mathematical and commonsense reasoning tasks over SOTA.
More encouragingly, AutoPRM can be easily integrated with other orthogonal
reasoning pipelines. | [
"cs.CL"
] | false |
2402.11455 | 2024-02-18T04:41:25Z | LoRA-Flow: Dynamic LoRA Fusion for Large Language Models in Generative
Tasks | [
"Hanqing Wang",
"Bowen Ping",
"Shuo Wang",
"Xu Han",
"Yun Chen",
"Zhiyuan Liu",
"Maosong Sun"
] | LoRA employs lightweight modules to customize large language models (LLMs)
for each downstream task or domain, where different learned additional modules
represent diverse skills. Combining existing LoRAs to address new tasks can
enhance the reusability of learned LoRAs, particularly beneficial for tasks
with limited annotated data. Most prior works on LoRA combination primarily
rely on task-level weights for each involved LoRA, making different examples
and tokens share the same LoRA weights. However, in generative tasks, different
tokens may necessitate diverse skills to manage. Taking the Chinese math task
as an example, understanding the problem description may depend more on the
Chinese LoRA, while the calculation part may rely more on the math LoRA. To
this end, we propose LoRA-Flow, which utilizes dynamic weights to adjust the
impact of different LoRAs. The weights at each step are determined by a fusion
gate with extremely few parameters, which can be learned with only 200 training
examples. Experiments across six generative tasks demonstrate that our method
consistently outperforms baselines with task-level fusion weights. This
underscores the necessity of introducing dynamic fusion weights for LoRA
combination. | [
"cs.CL"
] | false |
2402.11456 | 2024-02-18T04:45:01Z | FactPICO: Factuality Evaluation for Plain Language Summarization of
Medical Evidence | [
"Sebastian Antony Joseph",
"Lily Chen",
"Jan Trienes",
"Hannah Louisa Göke",
"Monika Coers",
"Wei Xu",
"Byron C Wallace",
"Junyi Jessy Li"
] | Plain language summarization with LLMs can be useful for improving textual
accessibility of technical content. But how factual are these summaries in a
high-stakes domain like medicine? This paper presents FactPICO, a factuality
benchmark for plain language summarization of medical texts describing
randomized controlled trials (RCTs), which are the basis of evidence-based
medicine and can directly inform patient treatment. FactPICO consists of 345
plain language summaries of RCT abstracts generated from three LLMs (i.e.,
GPT-4, Llama-2, and Alpaca), with fine-grained evaluation and natural language
rationales from experts. We assess the factuality of critical elements of RCTs
in those summaries: Populations, Interventions, Comparators, Outcomes (PICO),
as well as the reported findings concerning these. We also evaluate the
correctness of the extra information (e.g., explanations) added by LLMs. Using
FactPICO, we benchmark a range of existing factuality metrics, including the
newly devised ones based on LLMs. We find that plain language summarization of
medical evidence is still challenging, especially when balancing between
simplicity and factuality, and that existing metrics correlate poorly with
expert judgments on the instance level. | [
"cs.CL"
] | false |
2402.11457 | 2024-02-18T04:57:19Z | When Do LLMs Need Retrieval Augmentation? Mitigating LLMs'
Overconfidence Helps Retrieval Augmentation | [
"Shiyu Ni",
"Keping Bi",
"Jiafeng Guo",
"Xueqi Cheng"
] | Large Language Models (LLMs) have been found to have difficulty knowing they
do not possess certain knowledge and tend to provide specious answers in such
cases. Retrieval Augmentation (RA) has been extensively studied to mitigate
LLMs' hallucinations. However, due to the extra overhead and unassured quality
of retrieval, it may not be optimal to conduct RA all the time. A
straightforward idea is to only conduct retrieval when LLMs are uncertain about
a question. This motivates us to enhance the LLMs' ability to perceive their
knowledge boundaries to help RA. In this paper, we first quantitatively measure
LLMs' such ability and confirm their overconfidence. Then, we study how LLMs'
certainty about a question correlates with their dependence on external
retrieved information. We propose several methods to enhance LLMs' perception
of knowledge boundaries and show that they are effective in reducing
overconfidence. Additionally, equipped with these methods, LLMs can achieve
comparable or even better performance of RA with much fewer retrieval calls. | [
"cs.CL"
] | false |
2402.11481 | 2024-02-18T07:10:02Z | DictLLM: Harnessing Key-Value Data Structures with Large Language Models
for Enhanced Medical Diagnostics | [
"YiQiu Guo",
"Yuchen Yang",
"Ya Zhang",
"Yu Wang",
"Yanfeng Wang"
] | Structured data offers a sophisticated mechanism for the organization of
information. Existing methodologies for the text-serialization of structured
data in the context of large language models fail to adequately address the
heterogeneity inherent in key-value structured data. These methods are not
ideal and frequently result in larger input sizes and poor adaptability to
input changes. In this paper, we introduce DictLLM, an innovative framework
designed to improve the modeling of key-value structured data, like medical
laboratory reports, for generating medical diagnoses. DictLLM integrates three
key components: (1) group positional encoding to maintain permutation
invariance, (2) hierarchical attention bias to capture the inherent bias in
structured data, and (3) an optimal transport alignment layer that aligns the
embedding generated by the dictionary encoder with the LLM, thereby producing a
sequence of fixed-length virtual tokens. We carry out experiments using various
LLM models on a comprehensive real-world medical laboratory report dataset for
automatic diagnosis generation, our findings illustrate that DictLLM
significantly outperforms established baseline methods and few-shot GPT-4
implementations in terms of both Rouge-L and Knowledge F1 scores. Furthermore,
our evaluation of the framework's scalability and robustness, through a series
of experiments, underscores its exceptional capability in accurately modeling
the complex key-value data structure of medical dictionary data. | [
"cs.CL"
] | false |
2402.11489 | 2024-02-18T07:42:49Z | What's the Plan? Evaluating and Developing Planning-Aware Techniques for
LLMs | [
"Eran Hirsch",
"Guy Uziel",
"Ateret Anaby-Tavor"
] | Planning is a fundamental task in artificial intelligence that involves
finding a sequence of actions that achieve a specified goal in a given
environment. Large language models (LLMs) are increasingly used for
applications that require planning capabilities, such as web or embodied
agents. In line with recent studies, we demonstrate through experimentation
that LLMs lack necessary skills required for planning. Based on these
observations, we advocate for the potential of a hybrid approach that combines
LLMs with classical planning methodology. Then, we introduce SimPlan, a novel
hybrid-method, and evaluate its performance in a new challenging setup. Our
extensive experiments across various planning domains demonstrate that SimPlan
significantly outperforms existing LLM-based planners. | [
"cs.CL"
] | false |
2402.11493 | 2024-02-18T07:48:15Z | Benchmarking Knowledge Boundary for Large Language Model: A Different
Perspective on Model Evaluation | [
"Xunjian Yin",
"Xu Zhang",
"Jie Ruan",
"Xiaojun Wan"
] | In recent years, substantial advancements have been made in the development
of large language models, achieving remarkable performance across diverse
tasks. To evaluate the knowledge ability of language models, previous studies
have proposed lots of benchmarks based on question-answering pairs. We argue
that it is not reliable and comprehensive to evaluate language models with a
fixed question or limited paraphrases as the query, since language models are
sensitive to prompt. Therefore, we introduce a novel concept named knowledge
boundary to encompass both prompt-agnostic and prompt-sensitive knowledge
within language models. Knowledge boundary avoids prompt sensitivity in
language model evaluations, rendering them more dependable and robust. To
explore the knowledge boundary for a given model, we propose projected gradient
descent method with semantic constraints, a new algorithm designed to identify
the optimal prompt for each piece of knowledge. Experiments demonstrate a
superior performance of our algorithm in computing the knowledge boundary
compared to existing methods. Furthermore, we evaluate the ability of multiple
language models in several domains with knowledge boundary. | [
"cs.CL"
] | false |
2402.11532 | 2024-02-18T10:10:40Z | Chain-of-Instructions: Compositional Instruction Tuning on Large
Language Models | [
"Shirley Anugrah Hayati",
"Taehee Jung",
"Tristan Bodding-Long",
"Sudipta Kar",
"Abhinav Sethy",
"Joo-Kyung Kim",
"Dongyeop Kang"
] | Fine-tuning large language models (LLMs) with a collection of large and
diverse instructions has improved the model's generalization to different
tasks, even for unseen tasks. However, most existing instruction datasets
include only single instructions, and they struggle to follow complex
instructions composed of multiple subtasks (Wang et al., 2023a). In this work,
we propose a novel concept of compositional instructions called
chain-of-instructions (CoI), where the output of one instruction becomes an
input for the next like a chain. Unlike the conventional practice of solving
single instruction tasks, our proposed method encourages a model to solve each
subtask step by step until the final answer is reached. CoI-tuning (i.e.,
fine-tuning with CoI instructions) improves the model's ability to handle
instructions composed of multiple subtasks. CoI-tuned models also outperformed
baseline models on multilingual summarization, demonstrating the
generalizability of CoI models on unseen composite downstream tasks. | [
"cs.CL"
] | false |
2402.11548 | 2024-02-18T11:41:07Z | KMMLU: Measuring Massive Multitask Language Understanding in Korean | [
"Guijin Son",
"Hanwool Lee",
"Sungdong Kim",
"Seungone Kim",
"Niklas Muennighoff",
"Taekyoon Choi",
"Cheonbok Park",
"Kang Min Yoo",
"Stella Biderman"
] | We propose KMMLU, a new Korean benchmark with 35,030 expert-level
multiple-choice questions across 45 subjects ranging from humanities to STEM.
Unlike previous Korean benchmarks that are translated from existing English
benchmarks, KMMLU is collected from original Korean exams, capturing linguistic
and cultural aspects of the Korean language. We test 26 publically available
and proprietary LLMs, identifying significant room for improvement. The best
publicly available model achieves 50.54% on KMMLU, far below the average human
performance of 62.6%. This model was primarily trained for English and Chinese,
not Korean. Current LLMs tailored to Korean, such as Polyglot-Ko, perform far
worse. Surprisingly, even the most capable proprietary LLMs, e.g., GPT-4 and
HyperCLOVA X, achieve 59.95% and 53.40%, respectively. This suggests that
further work is needed to improve Korean LLMs, and KMMLU offers the right tool
to track this progress. We make our dataset publicly available on the Hugging
Face Hub and integrate the benchmark into EleutherAI's Language Model
Evaluation Harness. | [
"cs.CL"
] | false |