| | import os |
| | import torch |
| | import safetensors.torch |
| | import matplotlib |
| | import numpy as np |
| | import torch.nn.functional as F |
| | from torchvision import transforms as tt |
| | from omegaconf import OmegaConf |
| | from PIL import Image |
| | from transformers import Blip2Processor, Blip2ForConditionalGeneration |
| |
|
| | from PrimeDepth_inference.ldm.util import instantiate_from_config |
| |
|
| |
|
| | def get_state_dict(d): |
| | return d.get('state_dict', d) |
| |
|
| |
|
| | def load_state_dict(ckpt_path, location='cpu'): |
| | _, extension = os.path.splitext(ckpt_path) |
| | if extension.lower() == ".safetensors": |
| | state_dict = safetensors.torch.load_file(ckpt_path, device=location) |
| | else: |
| | state_dict = get_state_dict( |
| | torch.load( |
| | ckpt_path, map_location=torch.device(location) |
| | ) |
| | ) |
| | state_dict = get_state_dict(state_dict) |
| | print(f'Loaded state_dict from [{ckpt_path}]') |
| | return state_dict |
| |
|
| |
|
| | def create_model(config_path, device="cpu"): |
| | config = OmegaConf.load(config_path) |
| | model = instantiate_from_config(config.model).to(device) |
| | print(f'Loaded model config from [{config_path}]') |
| | return model |
| |
|
| |
|
| | def get_image(path): |
| | image = Image.open(path) |
| | if not image.mode == "RGB": |
| | image = image.convert("RGB") |
| | return image |
| |
|
| |
|
| | class InferenceEngine: |
| | """ |
| | Utility class for obtaining PrimeDepth predictions |
| | |
| | """ |
| |
|
| | def __init__(self, pd_config_path, blip2_cache_dir=None, cmap="Spectral", device="cuda"): |
| | """ |
| | pd_config_path : str |
| | Path to the model configuration file |
| | blip2_cache_dir : str, optional |
| | Path to the cache directory for the BLIP2 model, by default None |
| | cmap: str, optional |
| | Matplotlib colormap name, by default "Spectral" |
| | device : str, optional |
| | Device to run the model on, by default "cuda" |
| | |
| | """ |
| |
|
| | self.pd = create_model(pd_config_path, device) |
| | self.processor, self.model = self.load_BLIP2(blip2_cache_dir, device) |
| | self.cm = matplotlib.colormaps[cmap] |
| |
|
| | ''' |
| | def load_BLIP2(self, cache_dir=None, device="cuda"): |
| | processor = Blip2Processor.from_pretrained("Salesforce/blip2-opt-2.7b", cache_dir=cache_dir) |
| | model = Blip2ForConditionalGeneration.from_pretrained( |
| | "Salesforce/blip2-opt-2.7b", torch_dtype=torch.float16, cache_dir=cache_dir |
| | ) |
| | model = model.to(device) |
| | return processor, model |
| | ''' |
| |
|
| | def load_BLIP2(self, cache_dir=None, device="cuda"): |
| | import os |
| | from transformers import AutoProcessor, AutoModelForCausalLM |
| | import torch |
| | |
| | |
| | model_name = "microsoft/git-base" |
| | |
| | try: |
| | print(f"Loading alternative model: {model_name}") |
| | processor = AutoProcessor.from_pretrained(model_name) |
| | model = AutoModelForCausalLM.from_pretrained(model_name) |
| | model = model.to(device) |
| | print("Model loaded successfully!") |
| | return processor, model |
| | |
| | except Exception as e: |
| | print(f"Error loading model: {e}") |
| | raise |
| |
|
| | def captionize(self, image): |
| | |
| | inputs = self.processor(images=image, return_tensors="pt") |
| | |
| | inputs = {k: v.to(device='cuda', dtype=torch.float16) if torch.is_tensor(v) else v |
| | for k, v in inputs.items()} |
| |
|
| | |
| | generated_ids = self.model.generate( |
| | **inputs, |
| | max_length=30, |
| | num_beams=4, |
| | num_return_sequences=1 |
| | ) |
| | |
| | |
| | caption = self.processor.batch_decode(generated_ids, skip_special_tokens=True)[0] |
| | return caption |
| |
|
| | ''' |
| | def captionize(self, image): |
| | inputs = self.processor(images=image, return_tensors="pt").to('cuda', torch.float16) |
| | generated_ids = self.model.generate(**inputs) |
| | caption = self.processor.batch_decode(generated_ids, skip_special_tokens=True)[0].strip() |
| | return caption |
| | ''' |
| |
|
| |
|
| | def predict(self, image_path, max_size=1024): |
| | """ |
| | Parameters |
| | ---------- |
| | image_path : str |
| | Path to the image |
| | max_size : int, optional |
| | Maximal processing size of the longer image edge, by default 1024 |
| | |
| | Returns |
| | ------- |
| | depth_ssi : np.ndarray |
| | Scale and shift invariant depth map prediction |
| | depth_color : PIL.Image |
| | Colorized depth map prediction |
| | |
| | """ |
| |
|
| | image = get_image(image_path) |
| | caption = self.captionize(image) |
| | w, h = image.size |
| |
|
| | if max_size is not None and max(h, w) > max_size: |
| | if h == w: |
| | image = tt.Resize(max_size)(image) |
| | else: |
| | image = tt.Resize(max_size-1, max_size=max_size)(image) |
| |
|
| | with torch.no_grad(): |
| | labels = self.pd.get_label_from_image(image, prompt=caption) |
| | depth_ssi = labels['depth'].clone().mean(dim=1) |
| |
|
| | if max_size is not None and max(h, w) > max_size: |
| | depth_ssi = F.interpolate(depth_ssi[None], (h, w), mode='bilinear', align_corners=False)[0, 0] |
| | else: |
| | depth_ssi = depth_ssi[0] |
| |
|
| | depth_ssi = depth_ssi.cpu().numpy().astype(np.float32) |
| |
|
| | depth_color = depth_ssi.copy() |
| | depth_color = (depth_color - depth_color.min()) / (depth_color.max() - depth_color.min()) |
| | depth_color = self.cm(depth_color)[:, :, :3] |
| | depth_color = (depth_color * 255).astype(np.uint8) |
| | depth_color = Image.fromarray(depth_color) |
| |
|
| | return depth_ssi, depth_color |
| |
|