Spaces:
Sleeping
Sleeping
| import numpy as np | |
| import gradio as gr | |
| import glob | |
| import cv2 | |
| import matplotlib.pyplot as plt | |
| import insightface | |
| from insightface.app import FaceAnalysis | |
| from insightface.data import get_image as ins_get_image | |
| def predict(image_in_video, image_in_img): | |
| if image_in_video == None and image_in_img == None: | |
| raise gr.Error("Please capture an image using the webcam or upload an image.") | |
| image = image_in_video or image_in_img | |
| return swapi(image) | |
| app = FaceAnalysis(name='buffalo_l') | |
| app.prepare(ctx_id=0, det_size=(640, 640)) | |
| swapper = insightface.model_zoo.get_model('inswapper_128.onnx') | |
| def swapi(imagen): | |
| # Use the uploaded image to extract features | |
| img_user = cv2.imread(imagen) | |
| faces_user = app.get(img_user) | |
| # Use another image "background1" for modifications | |
| img_background = cv2.imread('background1.jpg') | |
| faces_background = app.get(img_background) | |
| # Assuming the user image has a face and we are using its features | |
| source_face = faces_user[0] | |
| # Apply modifications to the "background" image | |
| res = img_background.copy() | |
| for face in faces_background: | |
| res = swapper.get(res, face, source_face, paste_back=True) | |
| # Convert from BGR to RGB | |
| res_rgb = cv2.cvtColor(res, cv2.COLOR_BGR2RGB) | |
| return res_rgb | |
| with gr.Blocks() as blocks: | |
| gr.Markdown("### Capture Image Using WebCam or Upload") | |
| with gr.Row(): | |
| with gr.Column(): | |
| image_or_file_opt = gr.Radio(["webcam", "file"], value="webcam", | |
| label="How would you like to upload your image?") | |
| image_in_video = gr.Image(source="webcam", type="filepath") | |
| image_in_img = gr.Image(source="upload", visible=False, type="filepath") | |
| # Update visibility based on selection | |
| def toggle(choice): | |
| if choice == "webcam": | |
| return gr.update(visible=True, value=None), gr.update(visible=False, value=None) | |
| else: | |
| return gr.update(visible=False, value=None), gr.update(visible=True, value=None) | |
| image_or_file_opt.change(fn=toggle, inputs=[image_or_file_opt], | |
| outputs=[image_in_video, image_in_img], queue=False, show_progress=False) | |
| with gr.Column(): | |
| image_out = gr.Image() | |
| run_btn = gr.Button("Run") | |
| run_btn.click(fn=predict, inputs=[image_in_img, image_in_video], outputs=[image_out]) | |
| gr.Examples(fn=predict, examples=[], inputs=[image_in_img, image_in_video], outputs=[image_out]) | |
| blocks.queue() | |
| blocks.launch(debug=True) |