A fast stochastic gradient algorithm: Maximal use of sparsification benefits under computational constraints

Masahiro Yukawa, Wolfgang Utschick

Research output: Contribution to journalArticlepeer-review

3 Citations (Scopus)

Abstract

In this paper, we propose a novel stochastic gradient algorithm for efficient adaptive filtering. The basic idea is to sparsify the initial error vector and maximize the benefits from the sparsification under computational constraints. To this end, we formulate the task of algorithmdesign as a constrained optimization problem and derive its (non-trivial) closed-form solution. The computational constraints are formed by focusing on the fact that the energy of the sparsified error vector concentrates at the first few components. The numerical examples demonstrate that the proposed algorithm achieves the convergence as fast as the computationally expensive method based on the optimization without the computational constraints.

Original languageEnglish
Pages (from-to)467-475
Number of pages9
JournalIEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences
VolumeE93-A
Issue number2
DOIs
Publication statusPublished - 2010 Feb
Externally publishedYes

Keywords

  • Adaptive filter
  • Proportionate adaptive filtering
  • Stochastic gradient algorithm

ASJC Scopus subject areas

  • Signal Processing
  • Computer Graphics and Computer-Aided Design
  • Electrical and Electronic Engineering
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'A fast stochastic gradient algorithm: Maximal use of sparsification benefits under computational constraints'. Together they form a unique fingerprint.

Cite this