This reading group will go on from mid August till first week of december. We have approximately 14 weeks which means we can hold 28 presentations in total.
- (webpage link)
- Sign up sheet (spreadsheet)
- Unanswered questions doc (doc)
Contributors
Name | Role |
---|---|
Aditya Desai | TBD |
Zhenghui Guo | R2 |
Gaurav Gupta | R3 |
Masa Maksimovic | R2 |
Atishay Jain | R1 |
Sanya Garg | R2 |
Benjamin Meisburger | R2 |
Apoorv Walia | R1 |
Jonah Yi | R1 |
Name | Title | Suggested to be read by | date | Unanswered questions updated? | slides |
---|---|---|---|---|---|
Aditya Desai | Moments, deviations, chernoff/Hoeffding | All | 08/28/2023 | Yes | N.A |
Tony Zhang | Near neighbor graph, inner product transformation, inner product search | All | 09/06/2023 | Yes | Slides |
Jonah Yi | Learning to Route | All | 09/06/2023 | Yes | Slides |
Ben Meisburger | Deep Gradient Compression | All | 09/11/2023 | Yes | Slides |
Masa Maksimovic | Convergence in ML | All | 09/11/2023 | Yes | ❗ |
Zhenghui Guo(Kevin) | KV cache inference | All | 09/18/2023 | Yes | Slides |
Atishay Jain | Knowledge neurons | All | 09/18/2023 | ❗ | ❗ |
Gaurav | Vector Databases | All | 09/25/2023 | ❗ | ❗ |
Apoorv | Randomized projection based compression | All | 10/16/2023 | ❗ | ❗ |
Aditya Desai | Recent results on projection based compression | All | 10/16/2023 | ❗ | ❗ |
Zhenghui Guo(Kevin) | You Only Sample (Almost) Once: Linear Cost Self-Attention Via Bernoulli Sampling | All | 10/23/23 | ❗ | [Slides] |
Jonah Yi | Accelerating Large-Scale Inference with Anisotropic Vector Quantization | All | 10/30/23 | ❗ | Slides |
Benjamin Meisburger | ROSE: Robust Caching for Amazon Search | 11/13/23 | Slides | ||
Jonah Yi | ANN-Benchmarks: A Benchmarking Tool for Approximate Nearest Neighbor Algorithms | All | 11/20/23 | ❗ | Slides |
- ANN-Benchmarks: A Benchmarking Tool for Approximate Nearest Neighbor Algorithms: https://arxiv.org/pdf/1807.05614.pdf
- A Comprehensive Survey and Experimental Comparison of Graph-Based Approximate Nearest Neighbor Search: https://arxiv.org/pdf/2101.12631.pdf
- Robust Caching for Amazon Search: https://assets.amazon.science/dc/5e/b919974a4abdba7a3cd82c1bc86f/rose-robust-caches-for-amazon-product-search.pdf
- Accelerating Large-Scale Inference with Anisotropic Vector Quantization: https://arxiv.org/pdf/1908.10396.pdf
- You Only Sample (Almost) Once: Linear Cost Self-Attention Via Bernoulli Sampling: https://proceedings.mlr.press/v139/zeng21a/zeng21a.pdf
- Learning without Forgetting for Vision-Language Models: https://arxiv.org/pdf/2305.19270.pdf
- Introduction to Parameter Sharing Methods (Papers - Roast: Efficient Model Compression with Random Operation Access Specific Tile Hashing - https://arxiv.org/pdf/2207.10702.pdf)
- Robe: Random Offset Block Embedding Array - https://arxiv.org/pdf/2108.02191.pdf
- Compressing Neural Networks with the Hashing Trick - https://arxiv.org/pdf/1504.04788.pdf
- Results from "In defense of parameter sharing for model compression"
- Introduction to Vector databases (paper- Milvus: A purpose-built vector data management system, https://dl.acm.org/doi/abs/10.1145/3448016.3457550)
- Knowledge Neurons in Pretrained Transformers: https://arxiv.org/abs/2104.08696
- Locating and Editing Factual Associations in GPT: https://arxiv.org/abs/2202.05262
- Efficient Memory Management for Large Language Model Serving with PagedAttention: https://arxiv.org/pdf/2309.06180.pdf
- Deep Gradient Compression: https://openreview.net/pdf?id=SkhQHMW0W
- Convergence proofs for some simple settings: https://akyrillidis.github.io/comp414-514/schedule/?fbclid=IwAR1_ImKSLRQFgEGENLVBJD5stwFKf7fogHwse-w-NBrHxFVdMHl7iLRMUVA
- A convergence theory for deep learning: https://arxiv.org/abs/1811.03962v1
- Hierarchical Navigable Small World (HNSW) data structure for approximate near neighbor search: https://arxiv.org/abs/1603.09320
- Vector transformation for making them more amenable to inner product search: https://arxiv.org/abs/1405.5869
- IP-NSW for maximum inner product search: https://proceedings.neurips.cc/paper/2018/hash/229754d7799160502a143a72f6789927-Abstract.html
Learning to Route (LTR) in Similarity Graphs: https://arxiv.org/abs/1905.10987
We have confirmed booking for DH1049 every monday (except Oct 16 where we will book a library room)
I will aim to cover parts of chapter 3 and chapter 4 from http://lib.ysu.am/open_books/413311.pdf
We have the following roles in this edition.
This is focused research presentations. You will have to declare a particular problem that you are looking at and make three presentations:
- The first presentation will be motivating the problem (signifiance, impact, urgency, etc) and some related work that people have done. (Example : Embedding table compression is important, why, how have people solved this in the past, what problems remain, etc)
- Fundamentals of the approach. This is a "teaching presentation" where you explain the fundamental toolkit that you are using. (Example: I am going to use random projection based approach. what are random projections, how to analyse their usefulness, what are hash functions, why they are important, analysing hash functions)
- Your proposed method and results . (Example, the recipe to apply random projection based compression in embedding tables, results - quality, efficiency, etc)
This is more general research presentations. You will pick an area of research (for example, training efficiency, model compression, ml on edge, etc). You ll be making 2 presentations. This is similar to how things ran in summer edition of reading group. You ll keep reading papers on a single topic and summarize your learnings in two presentations.
This is a role of "moderating presentations". You ll not be presenting yourself. But you ll shadow readings of others if possible and actively discuss while they are presenting. The goal of this role is to keep the presentations engaging and to get the best out of it.
This is a role of "general audience and shadow". You ll read papers from one of the broad topics shadowing other presenters, listen in to presentations and participate in discussions.
❗ Add your name to the table above about roles.
- If you are participating in the reading group, send me your github id at [email protected] . I will add you to the collaborators so that you can change the readme and thus update the webpage.
- It will a collaborative effort to keep this webpage updated. So make sure to get yourself added to collaborators and update portions that are relevant to you over the period of entire edition.