|
|
import torch |
|
|
import torch.nn.functional as F |
|
|
|
|
|
def total_variation_loss(x): |
|
|
"""Total variation regularization""" |
|
|
batch_size = x.size(0) |
|
|
h_tv = torch.abs(x[:, :, 1:, :] - x[:, :, :-1, :]).sum() |
|
|
w_tv = torch.abs(x[:, :, :, 1:] - x[:, :, :, :-1]).sum() |
|
|
return (h_tv + w_tv) / batch_size |
|
|
|
|
|
def gradient_loss(x): |
|
|
"""Sobel gradient loss""" |
|
|
sobel_x = torch.tensor([[-1, 0, 1], [-2, 0, 2], [-1, 0, 1]], dtype=torch.float32, device=x.device).view(1, 1, 3, 3) |
|
|
sobel_y = torch.tensor([[-1, -2, -1], [0, 0, 0], [1, 2, 1]], dtype=torch.float32, device=x.device).view(1, 1, 3, 3) |
|
|
|
|
|
grad_x = F.conv2d(x, sobel_x.repeat(x.size(1), 1, 1, 1), padding=1, groups=x.size(1)) |
|
|
grad_y = F.conv2d(x, sobel_y.repeat(x.size(1), 1, 1, 1), padding=1, groups=x.size(1)) |
|
|
|
|
|
return torch.mean(grad_x**2 + grad_y**2) |
|
|
|
|
|
def diffusion_loss(model, x0, t, noise_scheduler, config): |
|
|
xt, noise = noise_scheduler.apply_noise(x0, t) |
|
|
pred_noise = model(xt, t) |
|
|
|
|
|
|
|
|
mse_loss = F.mse_loss(pred_noise, noise) |
|
|
|
|
|
|
|
|
tv_loss = total_variation_loss(xt) |
|
|
grad_loss = gradient_loss(xt) |
|
|
|
|
|
|
|
|
total_loss = mse_loss + config.tv_weight * tv_loss + 0.001 * grad_loss |
|
|
|
|
|
|
|
|
if torch.isnan(total_loss) or total_loss > 1e6: |
|
|
print(f"WARNING: Extreme loss detected!") |
|
|
print(f"MSE: {mse_loss.item():.4f}, TV: {tv_loss.item():.4f}, Grad: {grad_loss.item():.4f}") |
|
|
print(f"Noise range: [{noise.min().item():.4f}, {noise.max().item():.4f}]") |
|
|
print(f"Pred range: [{pred_noise.min().item():.4f}, {pred_noise.max().item():.4f}]") |
|
|
|
|
|
return total_loss |