remove white borders around edges in video

77 Views Asked by At

i am trying to replace background of a video using open-cv. my problem is a white borders that are appearing in the edges of a person.enter image description here. this is a frame of video i am trying to replace background frame by frame in a video. u guys can test this code and give me results too, heres the full code:

import cv2
from rembg import remove
from PIL import Image, ImageFilter
import os
import numpy as np
from moviepy.editor import VideoFileClip, ImageSequenceClip

def smooth_edges(image):
    # Apply Gaussian blur to the alpha channel for smoothing
    alpha_channel = image[:, :, 3]
    blurred_alpha = cv2.GaussianBlur(alpha_channel, (15, 15), 0)

    # Apply dilation to the alpha channel to reduce white borders
    kernel = np.ones((5, 5), np.uint8)
    dilated_alpha = cv2.dilate(alpha_channel, kernel, iterations=1)

    # Update the alpha channel with the smoothed and dilated alpha
    image[:, :, 3] = dilated_alpha

    return image

def process_video_with_audio(input_video_path, output_video_path='output_video.mp4', frame_limit=40):
    # Check if the video file exists
    if not os.path.exists(input_video_path):
        print(f"Video file '{input_video_path}' not found.")
        return None

    # Open the video capture
    video_capture = cv2.VideoCapture(input_video_path)

    # Check if the video capture is open
    if not video_capture.isOpened():
        print(f"Failed to open the video '{input_video_path}'.")
        return None

    frame_count = 0  # Initialize a frame count
    frame_list = []  # List to store processed frames

    while frame_count < frame_limit:
        # Read one frame from the video
        ret, frame = video_capture.read()

        # Check if the frame was successfully read
        if not ret:
            print(f"Failed to read frame {frame_count} from the video.")
            break  # Exit the loop if there are no more frames

        # Specify the output path for the processed frame
        frame_name = f'frame{frame_count}.png'
        frame_path = 'masked/' + frame_name

        # Save the frame as an image
        cv2.imwrite(frame_path, frame)

        # Output path for the processed image
        output_image_path = f'masked/processed_frame{frame_count}.png'

        # Save the frame as an image before processing it
        cv2.imwrite(output_image_path, frame)

        with open(output_image_path, 'rb') as f:
            input_image = f.read()
            subject = remove(input_image, alpha_matting=True, alpha_matting_foreground_threshold=50)
            with open(f'masked/background{frame_count}.png', 'wb') as output_file:
                output_file.write(subject)

        background_img_path = 'Cover.jpg'

        if os.path.exists(background_img_path):
            background_img = Image.open(background_img_path)

            # Load the processed frame as the foreground image
            foreground_img = Image.open(f'masked/background{frame_count}.png').convert("RGBA")

            # Resize the background image to match the dimensions of the video frame
            background_img = background_img.resize((foreground_img.width, foreground_img.height))

            # Ensure the foreground image has an 'RGBA' mode with an alpha channel
            foreground_img = foreground_img.convert("RGBA")

            # Apply Gaussian blur to the alpha channel for smoothing
            blurred_alpha = foreground_img.split()[3].filter(ImageFilter.GaussianBlur(radius=5))

            # Apply dilation to the alpha channel to reduce white borders
            dilated_alpha = np.array(blurred_alpha)
            kernel = np.ones((5, 5), np.uint8)
            dilated_alpha = cv2.dilate(dilated_alpha, kernel, iterations=1)
            blurred_alpha = Image.fromarray(dilated_alpha)

            # Composite the images with the smoothed and dilated alpha channel
            composite_img = Image.merge("RGBA", foreground_img.split()[:3] + (blurred_alpha,))

            # Save the final composite image as PNG (supports transparency)
            result = Image.alpha_composite(background_img.convert("RGBA"), composite_img)

            # Append the processed frame to the list
            frame_list.append(np.array(result))  # Convert to NumPy array

            frame_count += 1  # Increment the frame count

            # DEBUG: Print statement for debugging
            print(f"Processing frame {frame_count}...")

        else:
            print(f"Background image '{background_img_path}' not found")

    # Check if there are processed frames to create a video
    if frame_list:
        # Get the audio from the original video
        original_clip = VideoFileClip(input_video_path)
        audio = original_clip.audio

        # Create an ImageSequenceClip from the processed frames
        processed_clip = ImageSequenceClip(frame_list, fps=30)

        # Set the audio of the processed video to the original audio
        processed_clip = processed_clip.set_audio(audio)

        # Set the duration of the processed video to match the specified number of frames
        processed_clip = processed_clip.subclip(0, processed_clip.duration)

        # Save the final video with audio using MoviePy
        processed_clip.write_videofile(output_video_path, codec='libx264', audio_codec='aac')

        print(f"Video created at '{output_video_path}' with audio.")
        return output_video_path
    else:
        print("No processed frames to create a video.")
        return None

# Example usage:
video_path = 'alex.mp4'  # Change this to the actual path of your video
output_path = process_video_with_audio(video_path)

# You can use the 'output_path' variable as needed.

already tried dilation and gaussian blur but it didnt work

0

There are 0 best solutions below