📝 Triple-View Knowledge Distillation for Semi-Supervised Semantic Segmentation 🔭
"TriKD utilizes the knowledge distillation skill to learn the complementary semantics among these encoders with two different architectures for semi-supervised semantic segmentation by using a few labeled images and a large amount of unlabeled images." [gal30b+] 🤖 #CV
🔗 https://arxiv.org/abs/2309.12557v1 #arxiv
https://creative.ai/system/media_attachments/files/111/127/401/388/026/932/original/76847f773d831b5f.jpg
https://creative.ai/system/media_attachments/files/111/127/401/447/703/859/original/b46081764a8fc94c.jpg
https://creative.ai/system/media_attachments/files/111/127/401/523/940/257/original/e2133caa7b2b7d97.jpg