Abstract

Sparse learning problems, known as feature selection problems or variable selection problems, are a popular branch in the field of statistical learning. When faced with a dataset with only a few observations but a large number of features, we are interested in extracting the most useful features automatically by solving an optimization problem. In this dissertation, we start by introducing a novel penalty function as well as an iterative reweighted algorithm to solve the group sparsity problem, a special type of feature selection problems. The penalty function, named group LOG, shows a better ability to recover the ground-truth compared to other existing penalty functions in some datasets. Then, we propose a generalized formulation for a non-overlapping group sparsity problem that can be reduced to many existing works and provide the related algorithm by applying the alternating direction method of multipliers (ADMM) framework. In the end, we investigate a class of difference-of-convex (DC) programs that includes feature selection problems. Not only do we exhibit an exact algorithm for such problems, but we also design an inexact algorithm, which allows early termination in each iteration and justifies the stopping criteria in practice.

Degree Date

Spring 2023

Document Type

Thesis

Degree Name

Ph.D.

Department

Operations Research and Engineering Management

Advisor

Miju Ahn

Subject Area

Statistics

Number of Pages

112

Format

.pdf

Creative Commons License

Creative Commons Attribution-Noncommercial 4.0 License
This work is licensed under a Creative Commons Attribution-Noncommercial 4.0 License

Available for download on Wednesday, May 01, 2024

Share

COinS