posted on 2025-11-02, 10:16authored byChengyao Qian
Deep learning has significantly advanced computer vision but remains limited by heavy computation, slow inference, and privacy concerns. This thesis introduces resource-efficient solutions: (1) three knowledge-distillation methods—ℓ₁-regularized teacher training, teacher fine-tuning, and language-guided re-ranking—to close the student–teacher performance gap; (2) a parallelized diffusion sampling algorithm that eliminates sequential dependencies, accelerating high-quality image synthesis; and (3) a training-free machine unlearning mechanism that removes targeted concepts by excising corresponding subspaces in the weight space, rendering models blind to specified content. Collectively, these techniques reduce model size, speed up inference, and enforce data privacy while maintaining state-of-the-art accuracy.<p></p>
History
Principal supervisor
Mehrtash Harandi
Additional supervisor 1
Trung Le
Year of Award
2025
Department, School or Centre
Electrical and Computer Systems Engineering
Course
Doctor of Philosophy
Degree Type
DOCTORATE
Faculty
Faculty of Engineering
Rights Statement
The author retains copyright of this thesis. It must only be used for personal non-commercial research, education and study. It must not be used for any other purposes and may not be transmitted or shared with others without prior permission. For further terms use the In Copyright link under the License field.