Paris Machine Learning ( MLParis.org ) **@Meetup.com (6774 members) | @archives | @LinkedIn (1975) | @Google+(522) | @Facebook (388) | @Twitter (2312 followers) ** |
### A Neural Architecture for Bayesian CompressiveSensing over the Simplex via Laplace Techniques
Dear Igor,
I’m a long-time reader of your blog and wanted to share our recent paper on a relation between compressed sensing and neural network architectures. The paper introduces a new network construction based on the Laplace transform that results in activations such as ReLU and gating/threshold functions. It would be great if you could distribute the link on nuit-blanche.
The paper/preprint is here:https://ieeexplore.ieee.org/document/8478823https://www.netit.tu-berlin.de/fileadmin/fg314/limmer/LimSta18.pdf
Many thanks and best regards,Steffen
Dipl.-Ing. Univ. Steffen LimmerRaum HFT-TA 412Technische Universität BerlinInstitut für TelekommunikationssystemeFachgebiet Netzwerk-InformationstheorieEinsteinufer 25, 10587 Berlin
This paper presents a theoretical and conceptual framework to design neural architectures for Bayesian compressive sensing of simplex-constrained sparse stochastic vectors. First we recast the problem of MMSE estimation (w.r.t. a pre-defined uniform input distribution over the simplex) as the problem of computing the centroid of a polytope that is equal to the intersection of the simplex and an affine subspace determined by compressive measurements. Then we use multidimensional Laplace techniques to obtain a closed-form solution to this computation problem, and we show how to map this solution to a neural architecture comprising threshold functions, rectified linear (ReLU) and rectified polynomial (ReP) activation functions. In the proposed architecture, the number of layers is equal to the number of measurements which allows for faster solutions in the low-measurement regime when compared to the integration by domain decomposition or Monte-Carlo approximation. We also show by simulation that the proposed solution is robust to small model mismatches; furthermore, the proposed architecture yields superior approximations with less parameters when compared to a standard ReLU architecture in a supervised learning setting.
Printfriendly
Nuit Blanche community:@Google+ , @Facebook and @RedditCompressive Sensing community @LinkedInAdvanced Matrix Factorization community @Linkedin
Paris Machine Learning @Meetup.com, @archives, @LinkedIn , @Facebook,@Google+, @Twitter
Search Nuit Blanche
Subscribe by E-MAIL to Nuit Blanche
Contact me:
Popular Posts
Pages
Subscribe to the LinkedIn Matrix Factorization Group
Subscribe to the LinkedIn Compressive Sensing group
Nuit Blanche QR code
Readership Statistics
and another set of and
More than 838 readers receive every entries in their mailboxes while more than
There are more detailed information in the following blog entries.
So far, this site has seen more than 5,400,000 pageviews since a counter was installed in 2007.
Nuit Blanche Referenced in the Dead Tree World!
If You Like It, Link To It
Another Blog List
Search Pages I Link To
Previous Entries
Books Wish List
Books Wish List
Start-ups we like