A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness

B Pecher, I Srba, M Bielikova - ACM Computing Surveys, 2024 - dl.acm.org
Learning with limited labelled data, such as prompting, in-context learning, fine-tuning, meta-
learning, or few-shot learning, aims to effectively train a model using only a small amount of …

On the effectiveness of parameter-efficient fine-tuning

Z Fu, H Yang, AMC So, W Lam, L Bing… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …

Video understanding with large language models: A survey

Y Tang, J Bi, S Xu, L Song, S Liang, T Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
With the burgeoning growth of online video platforms and the escalating volume of video
content, the demand for proficient video understanding tools has intensified markedly. Given …

Finematch: Aspect-based fine-grained image and text mismatch detection and correction

H Hua, J Shi, K Kafle, S Jenni, D Zhang… - … on Computer Vision, 2025 - Springer
Recent progress in large-scale pre-training has led to the development of advanced vision-
language models (VLMs) with remarkable proficiency in comprehending and generating …

[HTML][HTML] Introducing NBEATSX to realized volatility forecasting

HG Souto, A Moradi - Expert Systems with Applications, 2024 - Elsevier
This paper investigates the application of neural basis expansion analysis with exogenous
variables (NBEATSx) in the prediction of daily stock realized volatility for various time steps …

Generalization in graph neural networks: Improved pac-bayesian bounds on graph diffusion

H Ju, D Li, A Sharma, HR Zhang - … Conference on Artificial …, 2023 - proceedings.mlr.press
Graph neural networks are widely used tools for graph prediction tasks. Motivated by their
empirical performance, prior works have developed generalization bounds for graph neural …

Videoxum: Cross-modal visual and textural summarization of videos

J Lin, H Hua, M Chen, Y Li, J Hsiao… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Video summarization aims to distill the most important information from a source video into
either an abridged video clip or a textual narrative. Existing methods often treat the …

Tuning Stable Rank Shrinkage: Aiming at the Overlooked Structural Risk in Fine-tuning

S Shen, Y Zhou, B Wei, EI Chang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Existing fine-tuning methods for computer vision tasks primarily focus on re-weighting the
knowledge learned from the source domain during pre-training. They aim to retain beneficial …

Improving pretrained language model fine-tuning with noise stability regularization

H Hua, X Li, D Dou, CZ Xu, J Luo - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
The advent of large-scale pretrained language models (PLMs) has contributed greatly to the
progress in natural language processing (NLP). Despite its recent success and wide …

V2xum-llm: Cross-modal video summarization with temporal prompt instruction tuning

H Hua, Y Tang, C Xu, J Luo - arXiv preprint arXiv:2404.12353, 2024 - arxiv.org
Video summarization aims to create short, accurate, and cohesive summaries of longer
videos. Despite the existence of various video summarization datasets, a notable limitation …