Bahareh Tolooshams
Bahareh Tolooshams

PhD Candidate, Electrical Engineering
Harvard University

News
  • 03/2023: I gave a talk at the COSYNE main meeting.
  • 02/2023: I defended my PhD.
  • 01/2023: Submitted a paper to ICML 2023.
  • 03/2022: I have received the GSAS Student Council Spring Conference Grant at Harvard University.
  • 02/2022: I joined InTouch, a peer-to-peer support network to build community and provide support for graduate students.
  • 05/2021: I joined the Applied Sciences Group at Microsoft as a Research Intern.
  • 07/2019: I received AWS Machine Learning Research Awards (MLRA).
  • 04/2019: I joined Amazon AI as ML/DSP Research Intern.
  • 04/2019: I received QBio Student Fellowship.
  • 08/2018: I received QBio Student Award Competition Fellowship.
  • 07/2018: I received a travel grant for MLSP2018.
  • 09/2017: I started my graduate studies at CRISP Group at Harvard University.
  • 06/2017: I received my B.ASc. degree with distinction in Electrical Engineering from the University of Waterloo.

I received B.ASc. degree with distinction in Electrical Engineering from the University of Waterloo in 2017. I am a PhD Candidate, advised by Demba Ba, at Harvard University. I spent summer 2019 at Amazon AI as an Applied Scientist Intern and joined the Applied Sciences Group at Microsoft as a Research Intern for summer 2021. I am the recipient of the awards including the Machine Learning Research Award from AWS and the QBio Student Fellowship from Harvard University.

Mentorship and community building: I actively mentor Harvard College students through the Women in STEM Mentorship program. I am also a mentor at InTouch, a peer-to-peer support network to build community and provide support for graduate students.

Research

My research interests are at the intersection of machine learning, optimization, statistical learning, and computational neuroscience. I offer an optimization-based signal processing perspective to design and analyze deep learning architectures. Particularly, I am interested in representation learning and probabilistic generative models to develop deep interpretable, and efficient neural architectures. My research is related to a class of machine learning algorithms referred to as unrolled learning in the litearture.

My PhD research is divided into three tracks:

  • Deep learning theory for model recovery and interpretability: I use a model-based optimization approach to improve the theoretical rigor of deep learning. This approach allows the designing of deep-learning-based provable algorithms. Moreover, it offers interpretability and mathematical reasons for representations of a new test example and extracts similar/dissimilar data from the training set. For deep sparse coding-based networks, I have shown that backpropagation not only accelerates learning but also guarantees a better model recovery. Check out this work for the interpretability of deep unrolled networks as well as their theoretical properties for model recovery.
  • Advancement of inverse problems in engineering: Inverse problems are conventionally solved by slow and unscalable optimization techniques. Deep learning exhibits a superior performance in solving inverse problems at scale. However, inverse problems often suffer from data scarcity. Hence, a question arises concerning the imposition of an inductive bias on deep architectures to enhance their generalization in unsupervised or data-scarce inverse problems. My research addresses this question by designing structured deep networks that optimize a statistical model, demonstrating a superior performance in solving inverse problems with accelerated inference. In my recent work, I highlighted a superior generalization in data-limited regime in radar sensing. In my ICML publication, I have demonstrated the efficiency and state-of-the-art competitive performance of my approach for Poisson image denoising.
  • Computational neuroscience: Deep learning can capture neural population dynamics in computational neuroscience. The black-box nature of deep learning, however, limits the unsupervised identification of factors driving neural activity. My research addresses this consequential drawback using interpretable learning; I associate the hidden network representations with a human-understandable model, linking them directly to stimuli and neural activity. My framework has deconvolved, for the first time, the single-trial activity of dopamine neurons into interpretable components in this abstract. Overall, this track aims to enable deep-learning applications to help answer scientific questions in computational neuroscience. I closely collaborate with Paul Masset from Murthy Lab.

During my two fantastic research internship experiences, I worked on speech enhancement. Here, I proposed channel-attention to improve multichannel speech enhancement. In another publication, a joint work with Kazuhito Koishida from Microsoft, I proposed a training framework for perceptual enhancement of stereo speech.

Publications
Journal and Conference Proceedings

  1. Bayesian unrolling: scalable, inverse-free maximum likelihood estimation of latent Gaussian models
    A. Lin, B. Tolooshams, Y. Atchadé, and D. Ba
    Submitted 2023
  2. Learning filter-based compressed blind-deconvolution
    B. Tolooshams*, S. Mulleti*, D. Ba, and Y. C. Eldar
    Submitted to IEEE Transactions on Signal Processing 2022 [paper]
  3. Discriminative reconstruction via simultaneous dense and sparse coding
    A. Tasissa, E. Theodosis, B. Tolooshams, and D. Ba
    Submitted to Information and Inference: A Journal of the IMA 2022 [paper] [code]
  4. Stable and interpretable unrolled dictionary learning
    B. Tolooshams and D. Ba
    TMLR 2022 [paper] [code]
  5. A training framework for stereo-aware speech enhancement using deep neural networks
    B. Tolooshams and K. Koishida
    IEEE ICASSP 2022 [paper] [slides] [poster]
  6. On the convergence of group-sparse autoencoders
    E. Theodosis, B. Tolooshams*, P. Tankala*, A. Tasissa, and D. Ba
    arXiv 2021 [paper]
  7. Gaussian process convolutional dictionary learning
    A. H. Song, B. Tolooshams, and D. Ba
    IEEE Signal Processing Letters 2021 [paper]
  8. Unfolding neural networks for compressive multichannel blind deconvolution
    B. Tolooshams*, S. Mulleti*, D. Ba, and Y. C. Eldar
    IEEE ICASSP 2021 [paper] [slides] [poster]
  9. Deep residual autoencoders for expectation maximization-inspired dictionary learning
    B. Tolooshams, S. Dey, and D. Ba
    IEEE Transactions on Neural Networks and Learning Systems 2021 [paper] [code]
  10. Convolutional dictionary learning based auto-encoders for natural exponential-family distributions
    B. Tolooshams*, A. H. Song*, S. Temereanca, and D. Ba
    ICML 2020 [paper] [code] [slides]
  11. Channel-attention dense u-net for multichannel speech enhancement
    B. Tolooshams, R. Giri, A. H. Song, U. Isik, and A. Krishnaswamy
    IEEE ICASSP 2020 [paper]
  12. Convolutional dictionary learning in hierarchical networks
    J. Zazo, B. Tolooshams, and D. Ba
    IEEE CAMSAP 2019 [paper]
  13. RandNet: deep learning with compressed measurements of images
    T. Chang*, B. Tolooshams*, and D. Ba
    IEEE MLSP 2019 [paper] [code] [poster]
  14. Scalable convolutional dictionary learning with constrained recurrent sparse auto-encoders
    B. Tolooshams, S. Dey, and D. Ba
    IEEE MLSP 2018 [paper] [code]
  15. Robustness of frequency division technique for online myoelectric pattern recognition against contraction-level variation
    B. Tolooshams and N. Jiang
    Frontiers in Bioengineering and Biotechnology 2017 [paper]
Abstracts

  1. Interpretable deep learning for deconvolution of multiplexed neural signals
    B. Tolooshams, S. Matias, H. Wu, N. Uchida, V. N. Murthy, P. Masset, and D. Ba
    COSYNE 2023 [paper]
  2. Interpretable unrolled dictionary learning networks
    B. Tolooshams and D. Ba
    DeepMath 2022 [paper] [slides]
  3. Unsupervised sparse deconvolutional learning of features driving neural activity
    B. Tolooshams, H. Wu, N. Uchida, V. N. Murthy, P. Masset, and D. Ba
    COSYNE 2022 [paper] [poster]
  4. Unsupervised learning of a dictionary of neural impulse responses from spiking data
    B. Tolooshams, H. Wu, P. Masset, V. N. Murthy, and D. Ba
    COSYNE 2021 [paper] [poster]
  5. Convolutional dictionary learning of stimulus from spiking data
    A. H. Song*, B. Tolooshams*, S. Temereanca, and D. Ba
    COSYNE 2020 [paper] [poster]
Talks & Workshops
  1. Interpretable deep learning for deconvolution of multiplexed neural signals
    Talk, Computational and Systems Neuroscience (COSYNE), 2023
  2. Deep learning for inverse problems in engineering and science
    Talk, Healthy ML Group at MIT, 2023
  3. Deep learning for inverse problems in engineering and science
    Talk, My PhD Defense at Harvard University, 2023
  4. Interpretable multimodal deconvolutional representation learning
    Talk, MURI annual meetings at ARL, 2023
  5. Deep representation learning for computational neuroscience
    Talk, DiCarlo Lab at MIT, 2022
  6. Interpretable unrolled dictionary learning networks
    Talk, Conference on the Mathematical Theory of Deep Neural Networks (DeepMath), 2022
  7. Design interpretable and efficient neural architectures for science and engineering
    Talk, Anima AI + Science Lab at Caltech, 2022
  8. Deconvolution of multiplexed neural signals using interpretable deep learning
    Talk, Uchida Lab at Harvard University, 2022
  9. Deep unrolled learning using bilevel optimizations
    Talk, Vector Institute, 2022
  10. Deep unrolling for inverse problems
    Talk, Poggio Lab at MIT, 2021
  11. Unfolded neural networks for implicit acceleration of dictionary learning
    Talk, Amirkabir Artificial Intelligence Student Summit (AAISS) at Amirkabir University of Technology, 2021
  12. Perceptual stereo speech enhancement
    Presentation, Applied Sciences Group at Microsoft, 2021
  13. Model-based deep learning
    Tutorial, IEEE ICASSP Conference, 2021
  14. Introduction to deep learning for computational neuroscience
    Workshop, Neurosur, 2021 [github]
  15. Dictionary learning based autoencoders for inverse problems
    Talk, Decision Theory - APMTH 231 at Harvard University, 2021
  16. On the relationship between dictionary learning and sparse autoencoders
    Talk, Computational and Applied Math Seminar at Tufts University, 2020
  17. On the relationship between dictionary learning and sparse autoencoders
    Talk, Pierre E. Jacob's Group at Harvard University, 2020
  18. Multichannel end-to-end neural architectures for speech enhancement
    Presentation, Amazon AI - AWS, 2019
  19. Autoencoders for unsupervised source separation
    Talk, Decision Theory - APMTH 231 at Harvard University, 2019
  20. State-space models and deep deconvolutional networks
    Talk, CRISP Group at Harvard University, 2018