pcuenq's picture
pcuenq HF staff
Remove large image
8771831
metadata
tags:
  - depth-estimation
library_name: coreml
license: apache-2.0

Depth Anything Core ML Models

Depth Anything model was introduced in the paper Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data by Lihe Yang et al. and first released in this repository.

Model description

Depth Anything leverages the DPT architecture with a DINOv2 backbone.

The model is trained on ~62 million images, obtaining state-of-the-art results for both relative and absolute depth estimation.

drawing

Depth Anything overview. Taken from the original paper.

Evaluation - Variants

Variant Parameters Size (MB) Weight precision Act. precision abs-rel error abs-rel reference
small-original (PyTorch) 24.8M 99.2 Float32 Float32
DepthAnythingSmallF32 24.8M 99.0 Float32 Float32 0.0073 small-original
DepthAnythingSmallF16 24.8M 45.8 Float16 Float16 0.0077 small-original

Evaluation - Inference time

The following results use the small-float16 variant.

Device OS Inference time (ms) Dominant compute unit
iPhone 12 Pro Max 18.0 31.10 Neural Engine
iPhone 15 Pro Max 17.4 33.90 Neural Engine
MacBook Pro (M1 Max) 15.0 32.80 Neural Engine
MacBook Pro (M3 Max) 15.0 24.58 Neural Engine

Download

Install huggingface-cli

brew install huggingface-cli

To download one of the .mlpackage folders to the models directory:

huggingface-cli download \
  --local-dir models --local-dir-use-symlinks False \
  apple/coreml-depth-anything-small \
  --include "DepthAnythingSmallF16.mlpackage/*"

To download everything, skip the --include argument.

Integrate in Swift apps

The huggingface/coreml-examples repository contains sample Swift code for coreml-depth-anything-small and other models. See the instructions there to build the demo app, which shows how to use the model in your own Swift apps.