I am a PhD student at the The Rachel and Selim Benin School of Computer Science and Engineering at the Hebrew University of Jerusalem. My advisor is Prof. Amnon Shashua. My main research interests are Neural Networks, Graphical Models and Machine Learning in general. I have recieved my B.Sc. degree in Physics, Mathematics and Computer Science from the Hebrew University. Click here for my complete CV.
If you are looking for my class notes, you can find them here (website and notes are in Hebrew).
- or.sharir cs.huji.ac.il
- Room B503, Rothberg Buildings, Givat Ram Campus.
- We released a paper bridging many-body quantum physics and deep learning methods, in which we present theoretical results for both the analysis of deep learning architectures, as well as for the capacity of deep networks to represent quantum states.—March 28th, 2018.
- Two of our papers, “On the Expressive Power of Overlapping Architectures of Deep Learning” and “Benefits of Depth for Long-Term Memory of Recurrent Networks”, were accepted for presentation at ICLR 2018, the former on the conference track and the latter on the workshop track.—January 30th, 2018.
- Our paper “Sum-Product-Quotient Networks” was accepted to AISTATS 2018.—December 22nd, 2017.
- We released a paper theoretically analyzing the benefits of depth to deep RNNs.—October 25th, 2017.
- We released a paper proposing an extension to Sum-Product Networks that exponentially boosts their expressive power.—October 12th, 2017.
- We released a paper theoretically analyzing the importance of overlapping receptive fields.—March 6th, 2017.
- We released a paper on merging our tensorial analysis of ConvNets with Sum-Product Networks.—October 13th, 2016.
- Our paper “On the Expressive Power of Deep Learning: A Tensor Analysis” was accepted to COLT 2016.—April 26th, 2016.
- Our paper “Deep Simnets” was accepted to CVPR 2016.—Feburary 29th, 2016.
- Github-MathJax (Chrome Extension) — a chrome extension for rendering LaTeX equations in Github repositories. Very useful for documenting research projects.
- reprochart — A usefuly python script for easily reproducible charts.
- Generative ConvACs (Experiments) — scripts for reproducing our experiments on Generative ConvACs.
- SimNets (Caffe Fork) — our fork of Caffe with the original implementation of the SimNets architecture.
Bridging Many-Body Quantum Physics and Deep Learning via Tensor Networks. Preprint.
Originally released on arXiv on March 26th, 2018.
Benefits of Depth for Long-Term Memory of Recurrent Networks. ICLR 2018 Workshop.
Originally released on arXiv on October 25th, 2017.
Sum-Product-Quotient Networks. AISTATS 2018.
Originally released on arXiv on October 12th, 2017.
On the Expressive Power of Overlapping Architectures of Deep Learning. ICLR 2018.
Originally released on arXiv on March 6th, 2017.
Tensorial Mixture Models. Preprint. (Source Code)
Originally released on arXiv on October 13th, 2016.
On the Expressive Power of Deep Learning: A Tensor Analysis. COLT 2016.
Originally released on arXiv on September 16th, 2015.
Deep SimNets. CVPR 2016. (Source Code)
Originally released on arXiv on June 9th, 2015.