跳到主要导航 跳到搜索 跳到主要内容

Revisiting linear machine learning through the perspective of inverse problems

  • Shuang Liu
  • , Sergey Kabanikhin
  • , Sergei Strijhak*
  • , Ying Ao Wang
  • , Ye Zhang
  • *此作品的通讯作者
  • Novosibirsk State University
  • RAS - Sobolev Institute of Mathematics, Siberian Branch
  • Russian Academy of Sciences
  • Beijing Institute of Technology
  • Shenzhen MSU-BIT University

科研成果: 期刊稿件文章同行评审

摘要

In this paper, we revisit Linear Neural Networks (LNNs) with single-output neurons performing linear operations. The study focuses on constructing an optimal regularized weight matrix Q from training pairs { G, H } {\{G,H\}}, reformulating the LNNs framework as matrix equations, and addressing it as a linear inverse problem. The ill-posedness of linear machine learning problems is analyzed through the lens of inverse problems. Furthermore, classical and modern regularization techniques from both the machine learning and inverse problems communities are reviewed. The effectiveness of LNNs is demonstrated through a real-world application in blood test classification, highlighting their practical value in solving real-life problems.

源语言英语
页(从-至)281-303
页数23
期刊Journal of Inverse and Ill-Posed Problems
33
2
DOI
出版状态已出版 - 1 4月 2025
已对外发布

指纹

探究 'Revisiting linear machine learning through the perspective of inverse problems' 的科研主题。它们共同构成独一无二的指纹。

引用此