legoandmars commited on
Commit
8663a23
·
1 Parent(s): eb80da0

more accurate masking

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -129,7 +129,7 @@ def inpaint(input_img, prompt):
129
 
130
  input_img_with_mask_64 = input_img_with_mask_converted.resize((64, 64), resample=Image.BICUBIC)
131
  # TODO: make 256x256 mask more accurate when upscaling?
132
- # input_img_with_mask_256 = input_img_with_mask_converted.resize((256, 256), resample=Image.BICUBIC)
133
  # return input_img, input_img_with_mask_64
134
  # Source image we are inpainting
135
  source_image_256 = pil_to_numpy(input_img_256)
@@ -139,18 +139,18 @@ def inpaint(input_img, prompt):
139
  # Assuming that all black pixels are meant for inpainting.
140
  # input_img_with_mask_64 = input_img_with_mask.convert('L').resize((64, 64), resample=Image.BICUBIC)
141
  gray_scale_source_image_64 = image_to_tensor(input_img_with_mask_64)
142
- # gray_scale_source_image_256 = image_to_tensor(input_img_with_mask_256)
143
 
144
  source_mask_64 = (gray_scale_source_image_64!=0).float()
145
- # source_mask_256 = (gray_scale_source_image_256!=0).float()
146
 
147
  source_mask_64_img = tensor_to_image(source_mask_64)
148
 
149
  # The mask should always be a boolean 64x64 mask, and then we
150
  # can upsample it for the second stage.
151
  source_mask_64 = source_mask_64.unsqueeze(0)
152
- # source_mask_256 = source_mask_256.unsqueeze(0)
153
- source_mask_256 = F.interpolate(source_mask_64, (256, 256), mode='nearest')
154
 
155
 
156
  ##############################
 
129
 
130
  input_img_with_mask_64 = input_img_with_mask_converted.resize((64, 64), resample=Image.BICUBIC)
131
  # TODO: make 256x256 mask more accurate when upscaling?
132
+ input_img_with_mask_256 = input_img_with_mask_converted.resize((256, 256), resample=Image.BICUBIC)
133
  # return input_img, input_img_with_mask_64
134
  # Source image we are inpainting
135
  source_image_256 = pil_to_numpy(input_img_256)
 
139
  # Assuming that all black pixels are meant for inpainting.
140
  # input_img_with_mask_64 = input_img_with_mask.convert('L').resize((64, 64), resample=Image.BICUBIC)
141
  gray_scale_source_image_64 = image_to_tensor(input_img_with_mask_64)
142
+ gray_scale_source_image_256 = image_to_tensor(input_img_with_mask_256)
143
 
144
  source_mask_64 = (gray_scale_source_image_64!=0).float()
145
+ source_mask_256 = (gray_scale_source_image_256!=0).float()
146
 
147
  source_mask_64_img = tensor_to_image(source_mask_64)
148
 
149
  # The mask should always be a boolean 64x64 mask, and then we
150
  # can upsample it for the second stage.
151
  source_mask_64 = source_mask_64.unsqueeze(0)
152
+ source_mask_256 = source_mask_256.unsqueeze(0)
153
+ # source_mask_256 = F.interpolate(source_mask_64, (256, 256), mode='nearest')
154
 
155
 
156
  ##############################