LenslessMic
					Collection
				
Models and Datasets from "LenslessMic: Audio Encryption and Authentication via Lensless Computational Imaging"
					β’ 
				4 items
				β’ 
				Updated
					
				
Reconstruction algoritms from the "LenslessMic: Audio Encryption and Authentication via Lensless Computational Imaging" paper.
To download the models and work with them, use our official repository.
The models are saved in the following format:
.
βββ checkpoint_tag
    βββ checkpoint_name.pth # PyTorch checkpoint with model state dict under 'state_dict' key.
    βββ config.yaml # Hydra config used to train the model
Checkpoint tag is represented in the following format:
{latent_size}_{training_dataset}_{loss_functions_used}_{reconstruction_algorithm}
latent_size is either 16x16 or 32x32, depends on the neural audio codec used in the dataset.random or librispeech. For librispeech, a groupped version can be used, tagged as
 group_n_m_r_c (see LenslessMic Version of Librispeech
 (with 288x288 after group if the sensor image size is not the default 256x256). The version of the model, which is
 fine-tuned using train-other, is tagged as librispeech_other and _ft at the end.loss_function is usually MSE, SSIM, and Raw SSIM, as in the paper. We also provide checkpoints with only MSE,
MSE and SSIM, and all three with L1 waveform or Mel Losses.PSF_Unet4M_U5_Unet4M is the Learned and R-Learned methods from the paper.
Unet8M is the NoPSF method.If you use these models, please cite it as follows:
@article{grinberg2025lenslessmic,
  title = {LenslessMic: Audio Encryption and Authentication via Lensless Computational Imaging},
  author = {Grinberg, Petr and Bezzam, Eric and Prandoni, Paolo and Vetterli, Martin},
  journal = {arXiv preprint arXiv:2509.16418},
  year = {2025},
}