From 330f86a4e474b8823d1394c2db9fbc37c44ec1fa Mon Sep 17 00:00:00 2001 From: henryruhs Date: Sat, 18 Jan 2025 15:00:57 +0100 Subject: [PATCH] Use 255.0 everywhere --- facefusion/face_classifier.py | 2 +- facefusion/face_masker.py | 4 ++-- facefusion/processors/modules/frame_enhancer.py | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/facefusion/face_classifier.py b/facefusion/face_classifier.py index 186a4ae..ce093cf 100644 --- a/facefusion/face_classifier.py +++ b/facefusion/face_classifier.py @@ -67,7 +67,7 @@ def classify_face(temp_vision_frame : VisionFrame, face_landmark_5 : FaceLandmar model_mean = get_model_options().get('mean') model_standard_deviation = get_model_options().get('standard_deviation') crop_vision_frame, _ = warp_face_by_face_landmark_5(temp_vision_frame, face_landmark_5, model_template, model_size) - crop_vision_frame = crop_vision_frame.astype(numpy.float32)[:, :, ::-1] / 255 + crop_vision_frame = crop_vision_frame.astype(numpy.float32)[:, :, ::-1] / 255.0 crop_vision_frame -= model_mean crop_vision_frame /= model_standard_deviation crop_vision_frame = crop_vision_frame.transpose(2, 0, 1) diff --git a/facefusion/face_masker.py b/facefusion/face_masker.py index 50c541b..362e253 100755 --- a/facefusion/face_masker.py +++ b/facefusion/face_masker.py @@ -171,7 +171,7 @@ def create_occlusion_mask(crop_vision_frame : VisionFrame) -> Mask: face_occluder_model = state_manager.get_item('face_occluder_model') model_size = create_static_model_set('full').get(face_occluder_model).get('size') prepare_vision_frame = cv2.resize(crop_vision_frame, model_size) - prepare_vision_frame = numpy.expand_dims(prepare_vision_frame, axis = 0).astype(numpy.float32) / 255 + prepare_vision_frame = numpy.expand_dims(prepare_vision_frame, axis = 0).astype(numpy.float32) / 255.0 prepare_vision_frame = prepare_vision_frame.transpose(0, 1, 2, 3) occlusion_mask = forward_occlude_face(prepare_vision_frame) occlusion_mask = occlusion_mask.transpose(0, 1, 2).clip(0, 1).astype(numpy.float32) @@ -184,7 +184,7 @@ def create_region_mask(crop_vision_frame : VisionFrame, face_mask_regions : List face_parser_model = state_manager.get_item('face_parser_model') model_size = create_static_model_set('full').get(face_parser_model).get('size') prepare_vision_frame = cv2.resize(crop_vision_frame, model_size) - prepare_vision_frame = prepare_vision_frame[:, :, ::-1].astype(numpy.float32) / 255 + prepare_vision_frame = prepare_vision_frame[:, :, ::-1].astype(numpy.float32) / 255.0 prepare_vision_frame = numpy.subtract(prepare_vision_frame, numpy.array([ 0.485, 0.456, 0.406 ]).astype(numpy.float32)) prepare_vision_frame = numpy.divide(prepare_vision_frame, numpy.array([ 0.229, 0.224, 0.225 ]).astype(numpy.float32)) prepare_vision_frame = numpy.expand_dims(prepare_vision_frame, axis = 0) diff --git a/facefusion/processors/modules/frame_enhancer.py b/facefusion/processors/modules/frame_enhancer.py index 1c8abe4..1c84db8 100644 --- a/facefusion/processors/modules/frame_enhancer.py +++ b/facefusion/processors/modules/frame_enhancer.py @@ -479,7 +479,7 @@ def forward(tile_vision_frame : VisionFrame) -> VisionFrame: def prepare_tile_frame(vision_tile_frame : VisionFrame) -> VisionFrame: vision_tile_frame = numpy.expand_dims(vision_tile_frame[:, :, ::-1], axis = 0) vision_tile_frame = vision_tile_frame.transpose(0, 3, 1, 2) - vision_tile_frame = vision_tile_frame.astype(numpy.float32) / 255 + vision_tile_frame = vision_tile_frame.astype(numpy.float32) / 255.0 return vision_tile_frame