david-shavin commited on
Commit
8eddf52
·
verified ·
1 Parent(s): d77e344

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -3
README.md CHANGED
@@ -1,3 +1,56 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+
5
+ # Model Card for Splat and Distill (SnD)
6
+
7
+ **Splat and Distill (SnD)** is a framework that imparts 3D awareness into 2D Vision Foundation Models (VFMs) by augmenting a teacher network with a feed-forward 3D reconstruction pipeline. It uses 3D Gaussian Splatting (3DGS) to supervise a student model with geometrically consistent features across novel views.
8
+
9
+
10
+
11
+ ## Model Details
12
+
13
+ ### Model Description
14
+
15
+ SnD bridges the gap between 2D representation and 3D understanding. It lifts 2D features from a teacher model into a 3D feature field using a feed-forward reconstruction model. These features are then "splatted" onto target views to provide a 3D-consistent supervisory signal for the student.
16
+
17
+ - **Developed by:** David Shavin, Sagie Benaim
18
+ - **Model type:** 3D-Aware Vision Foundation Model (Distillation Framework)
19
+ - **Conference:** ICLR 2026
20
+ - **License:** MIT
21
+ - **Finetuned from model:** DINOv2
22
+
23
+ ### Model Sources
24
+
25
+ - **Repository:** [https://github.com/davidshavin4/Splat-and-Distill](https://github.com/davidshavin4/Splat-and-Distill)
26
+ - **Paper:** [https://arxiv.org/abs/2602.06032](https://arxiv.org/abs/2602.06032)
27
+ - **Project Page:** [https://davidshavin4.github.io/Splat-and-Distill/](https://davidshavin4.github.io/Splat-and-Distill/)
28
+ - **Blog Post:** [Medium | Splat and Distill](https://medium.com/@davidshavin4/splat-and-distill-augmenting-teachers-with-feed-forward-3d-reconstruction-for-3d-aware-1f2c5e778399)
29
+
30
+ ## Uses
31
+
32
+ ### Direct Use
33
+
34
+ This model provides 3D-aware semantic features. There are two primary versions available depending on your downstream application:
35
+
36
+ * **With Blending:** Optimized for **single-view dense estimation tasks**. Use this version for tasks like semantic segmentation, depth estimation, and surface normal estimation.
37
+ * **Without Blending:** Optimized for tasks requiring **multi-view correspondence**. Use this version for geometric matching or tasks that rely on consistent feature tracking across different perspectives.
38
+
39
+ ## Bias, Risks, and Limitations
40
+
41
+ * **Data Bias:** The model was trained using the **ScanNet++** dataset. Consequently, the performance and geometric priors are primarily representative of indoor scene distributions found within that dataset.
42
+
43
+ ## Citation
44
+
45
+ **BibTeX:**
46
+
47
+ ```bibtex
48
+ @misc{shavin2026splatdistillaugmentingteachers,
49
+ title={Splat and Distill: Augmenting Teachers with Feed-Forward 3D Reconstruction For 3D-Aware Distillation},
50
+ author={David Shavin and Sagie Benaim},
51
+ year={2026},
52
+ eprint={2602.06032},
53
+ archivePrefix={arXiv},
54
+ primaryClass={cs.CV},
55
+ url={[https://arxiv.org/abs/2602.06032](https://arxiv.org/abs/2602.06032)},
56
+ }