wooj0216 commited on
Commit
9117c71
·
1 Parent(s): 6898ce9

change examples

Browse files
app.py CHANGED
@@ -90,13 +90,11 @@ def process_example(image_path):
90
  model_type = determine_model_type(image_path)
91
  return Image.open(image_path), model_type
92
 
93
-
94
- example_images = [
95
- "examples/fake/facial.jpg",
96
- "examples/fake/general.jpg",
97
- "examples/real/facial.jpg",
98
- "examples/real/general.jpg",
99
- ]
100
 
101
  with gr.Blocks() as demo:
102
 
@@ -105,22 +103,34 @@ with gr.Blocks() as demo:
105
  input_type = gr.Radio(["Image", "Video"], label="Choose Input Type", value="Image")
106
 
107
  model_type = gr.Radio(["Facial", "General"], label="Choose Model Type", value="General")
108
-
109
- image_input = gr.Image(type="pil", label="Upload Image", visible=True)
110
- video_input = gr.Video(label="Upload Video", visible=False)
111
-
 
112
  process_button = gr.Button("Run Model")
113
 
114
- pred_score_output = gr.Textbox(label="Prediction Score")
115
- attn_map_output = gr.Image(type="pil", label="Attention Map")
116
 
117
  # Example Images 추가
118
  gr.Examples(
119
- examples=example_images,
 
 
 
 
 
 
 
 
 
120
  inputs=[image_input],
121
  outputs=[image_input, model_type],
122
  fn=process_example,
123
- cache_examples=False
 
 
124
  )
125
 
126
  input_type.change(fn=change_input, inputs=[input_type], outputs=[image_input, video_input])
 
90
  model_type = determine_model_type(image_path)
91
  return Image.open(image_path), model_type
92
 
93
+ fake_examples, real_examples = [], []
94
+ for example in os.listdir("examples/fake"):
95
+ fake_examples.append(os.path.join("examples/fake", example))
96
+ for example in os.listdir("examples/real"):
97
+ real_examples.append(os.path.join("examples/real", example))
 
 
98
 
99
  with gr.Blocks() as demo:
100
 
 
103
  input_type = gr.Radio(["Image", "Video"], label="Choose Input Type", value="Image")
104
 
105
  model_type = gr.Radio(["Facial", "General"], label="Choose Model Type", value="General")
106
+
107
+ H, W = 300, 300
108
+ image_input = gr.Image(type="pil", label="Upload Image", visible=True, height=H, width=W)
109
+ video_input = gr.Video(label="Upload Video", visible=False, height=H, width=W)
110
+
111
  process_button = gr.Button("Run Model")
112
 
113
+ pred_score_output = gr.Textbox(label="Prediction Score : 0 - REAL, 1 - FAKE")
114
+ attn_map_output = gr.Image(type="pil", label="Attention Map", height=H, width=W)
115
 
116
  # Example Images 추가
117
  gr.Examples(
118
+ examples=fake_examples,
119
+ inputs=[image_input],
120
+ outputs=[image_input, model_type],
121
+ fn=process_example,
122
+ cache_examples=False,
123
+ examples_per_page=10,
124
+ label="Fake Examples"
125
+ )
126
+ gr.Examples(
127
+ examples=real_examples,
128
  inputs=[image_input],
129
  outputs=[image_input, model_type],
130
  fn=process_example,
131
+ cache_examples=False,
132
+ examples_per_page=10,
133
+ label="Real Examples"
134
  )
135
 
136
  input_type.change(fn=change_input, inputs=[input_type], outputs=[image_input, video_input])
examples/fake/facial.jpg DELETED
Binary file (127 kB)
 
examples/fake/facial1.png ADDED
examples/fake/facial2.jpg ADDED
examples/fake/{general.jpg → general1.jpg} RENAMED
File without changes
examples/fake/general2.jpg ADDED
examples/real/facial.jpg DELETED
Binary file (66.4 kB)
 
examples/real/facial1.jpg ADDED
examples/real/facial2.jpg ADDED
examples/real/{general.jpg → general1.jpg} RENAMED
File without changes
examples/real/general2.jpg ADDED