Bayesian Methodology for Adaptive Sparsity and Shrinkage in Regression
TBD, 2024
a data-adaptive method that performs model selection, covering the spectrum where the model collection spans from a sparse to a dense one
TBD, 2024
a data-adaptive method that performs model selection, covering the spectrum where the model collection spans from a sparse to a dense one
TBD, 2024
a review of deep-learning based SSMs for sequence modeling
TBD, 2025
generalized Bayesian method for high-dimensional robust regression
TBD, 2025
neural network-based method for detecting change points for multivariate time series
TBD, 2025
The conditional density characterizes the distribution of a response variable y given other predictor x, and plays a key role in many statistical tasks, including classification and outlier detection. Although there has been abundant work on the problem of Conditional Density Estimation (CDE) for a low-dimensional response in the presence of a high-dimensional predictor, little work has been done for a high-dimensional response such as images. The promising performance of normalizing flow (NF) neural networks in unconditional density estimation acts a motivating starting point. In this work, we extend NF neural networks when external $x$ is present. Specifically, they use the NF to parameterize a one-to-one transform between a high-dimensional y and a latent z that comprises two components $([z_P,z_N])$. The zP component is a low-dimensional subvector obtained from the posterior distribution of an elementary predictive model for x, such as logistic/linear regression. The zN component is a high-dimensional independent Gaussian vector, which explains the variations in y not or less related to $x$. Unlike existing CDE methods, the proposed approach, coined Augmented Posterior CDE (AP-CDE), only requires a simple modification on the common normalizing flow framework, while significantly improving the interpretation of the latent component, since zP represents a supervised dimension reduction. In image analytics applications, AP-CDE shows good separation of x-related variations due to factors such as lighting condition and subject id, from the other random variations. Further, the experiments show that an unconditional NF neural network, based on an unsupervised model of z, such as Gaussian mixture, fails to generate interpretable results.
TBD, 2025
Parameter-efficient fine-tuning (PEFT) methods, such as LoRA, offer compact and effective alternatives to full model fine-tuning by introducing low-rank updates to pre-trained weights. However, most existing approaches rely on global low rank structures, which can overlook spatial patterns spread across the parameter space. In this work, we propose Localized LoRA, a generalized framework that models weight updates as a composition of low-rank matrices applied to structured blocks of the weight matrix. This formulation enables dense, localized updates throughout the parameter space without increasing the total number of trainable parameters. We provide a formal comparison between global, diagonal-local, and fully localized low-rank approximations, and show that our method consistently achieves lower approximation error under matched parameter budgets. Experiments on both synthetic and practical settings demonstrate that Localized LoRA offers a more expressive and adaptable alternative to existing methods, enabling efficient fine-tuning with improved performance.
TBD, 2025
Panel vector auto-regressive (VAR) models are widely used to capture the dynamics of multivariate time series across different subpopulations, where each subpopulation shares a common set of variables. In this work, we propose a panel VAR model with a shared low-rank structure, modulated by subpopulation-specific weights, and complemented by idiosyncratic sparse components. To ensure parameter identifiability, we impose structural constraints that lead to a nonsmooth, nonconvex optimization problem. We develop a multi-block Alternating Direction Method of Multipliers (ADMM) algorithm for parameter estimation and establish its convergence under mild regularity conditions. Furthermore, we derive consistency guarantees for the proposed estimators under high-dimensional scaling. The effectiveness of the proposed modeling framework and estimators is demonstrated through experiments on both synthetic data and a real-world neuroscience data set.
IEEE Transactions on Signal Processing, 2022
a consensus-based distributed adaptive moment estimation method (DAdam) for online optimization over a decentralized network
Journal of Machine Learning Research, 2023
community detection and network modeling via a probablistic network model
Journal of the American Statistical Association, 2023
detecting and locating change points in high-dimensional Vector Autoregressive (VAR) models
Annals of Applied Statistics, 2023
an efficient algorithm for DAG estimation in the presence of partial ordering
Journal of Machine Learning Research, 2023
an algorithm for change point detection in the presence of covariate shift
Statistica Sinica, 2023
vector autoregressive model for high-dimensional mixed frequency data under the sparsity assumption
Annals of Applied Statistics, 2023
joint estimation of mixed frequency data via a Bayesian approach
Journal of Machine Learning Research, 2023
estimating a vector autoregressive (VAR) model assuming that Granger causal graph has a tree structure
International Journal of Forecasting, 2023
DNN method for mixed-frequency data prediction
Technometrics, 2023
a general framework for modeling network autoregressive processes
Journal of Computational and Graphical Statistics, 2024
an ADMM algorithm for high dimensional structural VAR estimation in the presence of partial ordering
Transactions on Machine Learning Research (TMLR), 2024
DNN method for estimating the Granger causality of multiple dynamical systems via a VAE-based formulation
Statistica Sinica, 2024
modeling for multivariate time series where node values depend on its own past and its neighbors
Journal of Machine Learning Research, 2024
flow-based decomposition of effect under a causal DAG
Journal of Machine Learning Research, 2024
Ising model with peer-effect from the network and the effect from covariates
Annals of Applied Statistics, 2024
explore connectivity patterns amongst US financial institutions based on debiased Lasso
Automatica, 2024
system identification of multiple related time-invariant linear dynamical systems
Automatica, 2025
a distributed algorithm for bilevel-optimization
Transactions on Machine Learning Research (TMLR), 2025
estimate covariate-dependent graphical models using DNN with PAC guarantees