Skip to content

bunnysocks/DAv2-Finetuning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 

Repository files navigation

Domain-Adaptive Depth Estimation for Transparent and Reflective Environments

Overview

Monocular depth estimation models often fail when dealing with transparent, translucent, or mirrored objects — surfaces that violate normal visual depth cues.
This project proposes a Domain-Adaptive Depth Estimation framework that fine-tunes the Depth Anything v2 model to handle these complex visual conditions.

Our goal is to make monocular depth estimation more reliable for environments where robots, drones, or autonomous systems encounter materials like glass, metal, and reflective plastics.


Project Objective

Train a depth estimation model that:

  • Adapts to transparent and reflective domains without requiring ground-truth depth labels.
  • Maintains consistent depth prediction under style and lighting variations.
  • Outperforms general-purpose models like Depth Anything v2 on this specialized domain.

Core Idea

We introduce a Style Consistency Loss that ensures the model predicts similar depth maps even after stylistic or illumination transformations of the same image.
This enforces geometry-focused learning, making the model robust to misleading visual textures such as reflections or glare.


Method Summary

  1. Use Depth Anything v2 (Small) as a pretrained base model.
  2. Perform unsupervised domain adaptation with unlabeled images of transparent/mirrored scenes.
  3. Train a lightweight adapter and decoder using:
    • Style Consistency Loss
    • Edge-Aware Smoothness Loss
    • (Optional) Domain Adversarial Loss
  4. Evaluate on datasets like ClearGrasp.

Key Features

  • Domain adaptation without labeled depth data
  • Style-invariant depth learning using Style Consistency Loss
  • Specialized performance in reflective and transparent scenes

Future Extensions

  • Add uncertainty estimation to handle ambiguous surfaces.
  • Integrate physics-aware priors (refraction modeling).
  • Expand to multi-view or stereo adaptation.

About

Domain-Adaptive Depth Estimation for Transparent and Reflective Environments

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published