You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This project explores generating high-quality images using depth maps and conditioning techniques like Canny edges, leveraging Stable Diffusion and ControlNet models. It focuses on optimizing image generation with different aspect ratios, inference steps to balance speed and quality.
Depth map estimation tool using Depth-Anything-V2. Generate accurate depth maps from images with support for both relative and metric depth measurements.
A visualization tool for analyzing preprocessing results of WAI/Nerfstudio format datasets containing RGB images and corresponding depth maps to generate interactive 3D visualizations.
This repository features a customized Detection Transformer (DETR) model integrated with the NYU Depth V2 dataset, forming an end-to-end semantic segmentation pipeline. Optimized for AR applications, it utilizes depth data to enhance segmentation accuracy and enable sophisticated depth-aware scene understanding.
I'm starting this repository to explore image depth analysis using various models. The ultimate goal is to implement Visual SLAM on the robot I previously built. This space will serve as a sandbox for experimentation—where I'll test out different depth estimation models and work on automating the robot's navigation in new environments.
3D-Image-Toolbox is a Python-based tool that transforms images and videos into immersive spatial experiences. Using depth-anything-v2 model, it generates depth maps from standard 2D media and converts them into side-by-side 3D formats. From spacial photos (heic format) it uses the contained depth map.
This repository contains code and experiments for monocular depth estimation from single RGB images using deep convolutional neural networks. It leverages synthetic and real-world datasets, advanced data augmentation, and state-of-the-art architectures to predict depth maps for applications in robotics, autonomous driving, and augmented reality.