* Cleanup after age modifier PR * Cleanup after age modifier PR * Use OpenVino 2024.2.0 for installer * Prepare 3.0.0 for installer * Fix benchmark suite, Introduce sync_item() for state manager * Fix lint * Render slide preview also in lower res * Lower thread and queue count to avoid false usage * Fix spacing * Feat/jobs UI (#627) * Jobs UI part1 * Change naming * Jobs UI part2 * Jobs UI part3 * Jobs UI part4 * Jobs UI part4 * Jobs UI part5 * Jobs UI part6 * Jobs UI part7 * Jobs UI part8 * Jobs UI part9 * Jobs UI part10 * Jobs UI part11 * Jobs UI part12 * Fix rebase * Jobs UI part13 * Jobs UI part14 * Jobs UI part15 * changes (#626) * Remove useless ui registration * Remove useless ui registration * move job_list.py replace [0] with get_first() * optimize imports * fix date None problem add test job list * Jobs UI part16 * Jobs UI part17 * Jobs UI part18 * Jobs UI part19 * Jobs UI part20 * Jobs UI part21 * Jobs UI part22 * move job_list_options * Add label to job status checkbox group * changes * changes --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Update some dependencies * UI helper to convert 'none' * validate job (#628) * changes * changes * add test * changes * changes * Minor adjustments * Replace is_json with is_file * Handle empty and invalid json in job_list * Handle empty and invalid json in job_list * Handle empty and invalid json in job_list * Work on the job manager UI * Cosmetic changes on common helper * Just make it work for now * Just make it work for now * Just make it work for now * Streamline the step index lookups * Hide footer * Simplify instant runner * Simplify instant runner UI and job manager UI * Fix empty step choices * Fix empty step choices * Fix none values in UI * Rework on benchmark (add warmup) and job list * Improve ValueAndUnit * Add step 1 of x output * Cosmetic changes on the UI * Fix invalid job file names * Update preview * Introducing has_step() and sorting out insert behaviour * Introducing has_step() and sorting out insert behaviour * Add [ none ] to some job id dropdowns * Make updated dropdown values kinda perfect * Make updated dropdown values kinda perfect * Fix testing * Minor improvement on UI * Fix false config lookup * Remove TensorRT as our models are not made for it * Feat/cli commands second try rev2 (#640) * Refactor CLI to commands * Refactor CLI to commands part2 * Refactor CLI to commands part3 * Refactor CLI to commands part4 * Rename everything to facefusion.py * Refactor CLI to commands part5 * Refactor CLI to commands part6 * Adjust testing * Fix lint * Fix lint * Fix lint * Refactor CLI to commands part7 * Extend State typing * Fix false config lookup, adjust logical orders * Move away from passing program part1 * Move away from passing program part2 * Move away from passing program part3 * Fix lint * Move away from passing program part4 * ui-args update * ui-args update * ui-args update * temporary type fix * Move away from passing program part5 * remove unused * creates args.py * Move away from passing program part6 * Move away from passing program part7 --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Minor optimizations * Update commands in README * Fix job-retry command * Fix multi runs via UI * add more job keys * Cleanup codebase * One method to create inference session (#641) * One method to create inference session * Remove warnings, as there are none * Remember job id during processing * Fix face masker config block * Change wording * Prevent age modifier from using CoreML * add expression restorer (#642) * add expression restorer * fix import * fix lint * changes * changes * changes * Host the final model for expression restorer * Insert step on the given index * UI workover (#644) * UI workover part1 * Introduce ComponentOptions * Only set Media components to None when visibility changes * Clear static faces and reference faces between step processing * Minor changes * Minor changes * Fix testing * Enable test_sanitize_path_for_windows (#646) * Dynamic download during job processing (#647) * Fix face masker UI * Rename run-headless to headless-run * Feat/split frame processor UI (#649) * Split frame processor UI * Split frame processor UI part3, Refactor get_model_initializer * Split frame processor UI part4 * Feat/rename frame processors (#651) * Rename frame processors * Rename frame processors part2 * Fix imports Conflicts: facefusion/uis/layouts/benchmark.py facefusion/uis/layouts/default.py * Fix imports * Cosmetic changes * Fix multi threading for ROCm * Change temp frames pattern * Adjust terminal help * remove expression restorer (#653) * Expression restorer as processor (#655) * add expression restorer * changes * Cleanup code * Add TensorRT support back * Add TensorRT support back * Add TensorRT support back * changes (#656) * Change minor wording * Fix face enhancer slider * Add more typing * Fix expression-restorer when using trim (#659) * changes * changes * Rework/model and inference pool part2 (#660) * Rework on model and inference pool * Introduce inference sources and pools part1 * Introduce inference sources and pools part2 * Introduce inference sources and pools part3 * Introduce inference sources and pools part4 * Introduce inference sources and pools part5 * Introduce inference sources and pools part6 * Introduce inference sources and pools part6 * Introduce inference sources and pools part6 * Introduce inference sources and pools part7 * Introduce inference sources and pools part7 * Introduce inference sources and pools part8 * Introduce inference sources and pools part9 * Introduce inference sources and pools part10 * Introduce inference sources and pools part11 * Introduce inference sources and pools part11 * Introduce inference sources and pools part11 * Introduce inference sources and pools part12 * Reorganize the face masker UI * Fix trim in UI * Feat/hashed sources (#668) * Introduce source helper * Remove post_check() and just use process_manager * Remove post_check() part2 * Add hash based downloads * Add hash based downloads part2 * Add hash based downloads part3 * Add hash based downloads part4 * Add hash based downloads part5 * Add hash based downloads part6 * Add hash based downloads part7 * Add hash based downloads part7 * Add hash based downloads part8 * Remove print * Prepare 3.0.0 release * Fix UI * Release the check when really done * Update inputs for live portrait * Update to 3.0.0 releases, extend download postfix * Move files to the right place * Logging for the hash and source validation * Changing logic to handle corrupt sources * Fix typo * Use names over get_inputs(), Remove set_options() call * Age modifier now works for CoreML too * Update age_modifier.py * Add video encoder h264_videotoolbox and hevc_videotoolbox * Face editor add eye gaze & remove open factor sliders (#670) * changes * add eye gaze * changes * cleanup * add eyebrow control * changes * changes * Feat/terminal UI (#671) * Introduce terminal to the UI * Introduce terminal to the UI part2 * Introduce terminal to the UI part2 * Introduce terminal to the UI part2 * Calc range step to avoid weird values * Use Sequence for ranges * Use Sequence for ranges * changes (#673) * Use Sequence for ranges * Finalize terminal UI * Finalize terminal UI * Webcam cosmetics, Fix normalize fps to accept int * Cosmetic changes * Finalize terminal UI * Rename leftover typings * Fix wording * Fix rounding in metavar * Fix rounding in metavar * Rename to face classifier * Face editor lip moves (#677) * changes * changes * changes * Fix rounding in metavar * Rename to face classifier * changes * changes * update naming --------- Co-authored-by: henryruhs <info@henryruhs.com> * Fix wording * Feat/many landmarker + face analyser breakdown (#678) * Basic multi landmarker integration * Simplify some method names * Break into face_detector and face_landmarker * Fix cosmetics * Fix testing * Break into face_attributor and face_recognizer * Clear them all * Clear them all * Rename to face classifier * Rename to face classifier * Fix testing * Fix stuff * Add face landmarker model to UI * Add face landmarker model to UI part2 * Split the config * Split the UI * Improvement from code review * Improvement from code review * Validate args also for sub parsers * Remove clear of processors in process step * Allow finder control for the face editor * Fix lint * Improve testing performance * Remove unused file, Clear processors from the UI before job runs * Update the installer * Uniform set handler for swapper and detector in the UI * Fix example urls * Feat/inference manager (#684) * Introduce inference manager * Migrate all to inference manager * clean ini * Introduce app context based inference pools * Fix lint * Fix typing * Adjust layout * Less border radius * Rename app context names * Fix/live portrait directml (#691) * changes (#690) * Adjust naming * Use our assets release * Adjust naming --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Add caches to gitignore * Update dependencies and drop CUDA 11.8 support (#693) * Update dependencies and drop CUDA 11.8 support * Play save and keep numpy 1.x.x * Improve TensorRT optimization * changes * changes * changes * changes * changes * changes * changes * changes * changes * Reuse inference sessions (#696) * Fix force-download command * Refactor processors to forward() (#698) * Install tensorrt when selecting cuda * Minor changes * Use latest numpy * Fix limit system memory * Implement forward() for every inference (#699) * Implement forward() for every inference * Implement forward() for every inference * Implement forward() for every inference * Implement forward() for every inference * changes * changes * changes * changes * Feat/fairface (#710) * Replace gender_age model with fair face (#709) * changes * changes * changes * age dropdown to range-slider * Cleanup code * Cleanup code --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Extend installer to set library paths for cuda and tensorrt (#707) * Extend installer to set library paths for cuda and tensorrt * Add refresh of conda env * Remove invalid commands * Set the conda env according to operating system * Update for ROCm 6.2 * fix installer * Aktualisieren von installer.py * Add missing face selector keys * Try to keep original LD_LIBRARY_PATH * windows support installer * Final touch to the installer * Remove spaces * Simplidy collect_model_downloads() * Fix force download for once and forever * Housekeeping (#715) * changes * changes * changes * Fix performance part1 * Fix mixed states (#689) * Fix mixed states * Add missing sync for job args * Move UnionStateXXX to base typing * Undo * Remove UnionStateXXX * Fix app context performance lookup (#717) * Restore performance for inswapper * Mover upper() to the logger * Undo debugging * Move TensorRT installation to docs * Sort out log level typing, Add log level UI dropdown (#719) * Fix inference pool part1 * Validate conda library paths existence * Default face selector order to large-small * Fix inference pool context according to execution provider (#720) * Fix app context under Windows * CUDA and TensorRT update for the installer * Remove concept of static processor modules * Revert false commit * Change event order makes a difference * Fix multi model context in inference pool (#721) * Fix multi model context in inference pool * Fix multi model context in inference pool part2 * Use latest gradio to avoid fastapi bug * Rework on the Windows Installer * Use embedding converter (#724) * changes (#723) * Upload models to official assets repo --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Rework on the Windows Installer part2 * Resolve subprocess calls (#726) * Experiment * Resolve subprocess calls to cover edge cases like broken PATH * Adjust wording * Simplify code * Rework on the Windows Installer part3 * Rework on the Windows Installer part4 * Numpy fix for older onnxruntime * changes (#729) * Add space * Add MacOS installer * Use favicon * Fix disabled logger * Layout polishing (#731) * Update dependencies, Adjust many face landmarker logic * Cosmetics changes * Should be button * Introduce randomized action button * Fix update of lip syncer and expression restorer * Stop sharing inference session this prevents flushing VRAM * Fix test * Fix urls * Prepare release * Vanish inquirer * Sticky preview does not work on portrait images * Sticky preview only for landscape images and videos * remove gradio tunnel env * Change wording and deeplinks * increase peppa landmark score offset * Change wording * Graceful exit install.py * Just adding a required * Cannot use the exit_helper * Rename our model * Change color of face-landmark-68/5 * Limit liveportrait (#739) * changes * changes * changes * Cleanup * Cleanup --------- Co-authored-by: harisreedhar <h4harisreedhar.s.s@gmail.com> Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * limit expression restorer * change expression restorer 0-100 range * Use 256x icon * changes * changes * changes * changes * Limit face editor rotation (#745) * changes (#743) * Finish euler methods --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> * Use different coveralls badge * Move about wording * Shorten scope in the logger * changes * changes * Shorten scope in the logger * fix typo * Simplify the arcface converter names * Update preview --------- Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com> Co-authored-by: harisreedhar <h4harisreedhar.s.s@gmail.com>
271 lines
12 KiB
Python
Executable File
271 lines
12 KiB
Python
Executable File
from argparse import ArgumentParser
|
|
from typing import List
|
|
|
|
import cv2
|
|
import numpy
|
|
|
|
import facefusion.jobs.job_manager
|
|
import facefusion.jobs.job_store
|
|
import facefusion.processors.core as processors
|
|
from facefusion import config, content_analyser, face_classifier, face_detector, face_landmarker, face_masker, face_recognizer, inference_manager, logger, process_manager, state_manager, voice_extractor, wording
|
|
from facefusion.audio import create_empty_audio_frame, get_voice_frame, read_static_voice
|
|
from facefusion.common_helper import get_first
|
|
from facefusion.download import conditional_download_hashes, conditional_download_sources
|
|
from facefusion.face_analyser import get_many_faces, get_one_face
|
|
from facefusion.face_helper import create_bounding_box, paste_back, warp_face_by_bounding_box, warp_face_by_face_landmark_5
|
|
from facefusion.face_masker import create_mouth_mask, create_occlusion_mask, create_static_box_mask
|
|
from facefusion.face_selector import find_similar_faces, sort_and_filter_faces
|
|
from facefusion.face_store import get_reference_faces
|
|
from facefusion.filesystem import filter_audio_paths, has_audio, in_directory, is_image, is_video, resolve_relative_path, same_file_extension
|
|
from facefusion.processors import choices as processors_choices
|
|
from facefusion.processors.typing import LipSyncerInputs
|
|
from facefusion.program_helper import find_argument_group
|
|
from facefusion.thread_helper import conditional_thread_semaphore
|
|
from facefusion.typing import ApplyStateItem, Args, AudioFrame, Face, InferencePool, ModelOptions, ModelSet, ProcessMode, QueuePayload, UpdateProgress, VisionFrame
|
|
from facefusion.vision import read_image, read_static_image, restrict_video_fps, write_image
|
|
|
|
MODEL_SET : ModelSet =\
|
|
{
|
|
'wav2lip':
|
|
{
|
|
'hashes':
|
|
{
|
|
'lip_syncer':
|
|
{
|
|
'url': 'https://github.com/facefusion/facefusion-assets/releases/download/models-3.0.0/wav2lip.hash',
|
|
'path': resolve_relative_path('../.assets/models/wav2lip.hash')
|
|
}
|
|
},
|
|
'sources':
|
|
{
|
|
'lip_syncer':
|
|
{
|
|
'url': 'https://github.com/facefusion/facefusion-assets/releases/download/models-3.0.0/wav2lip.onnx',
|
|
'path': resolve_relative_path('../.assets/models/wav2lip.onnx')
|
|
}
|
|
},
|
|
'size': (96, 96)
|
|
},
|
|
'wav2lip_gan':
|
|
{
|
|
'hashes':
|
|
{
|
|
'lip_syncer':
|
|
{
|
|
'url': 'https://github.com/facefusion/facefusion-assets/releases/download/models-3.0.0/wav2lip_gan.hash',
|
|
'path': resolve_relative_path('../.assets/models/wav2lip_gan.hash')
|
|
}
|
|
},
|
|
'sources':
|
|
{
|
|
'lip_syncer':
|
|
{
|
|
'url': 'https://github.com/facefusion/facefusion-assets/releases/download/models-3.0.0/wav2lip_gan.onnx',
|
|
'path': resolve_relative_path('../.assets/models/wav2lip_gan.onnx')
|
|
}
|
|
},
|
|
'size': (96, 96)
|
|
}
|
|
}
|
|
|
|
|
|
def get_inference_pool() -> InferencePool:
|
|
model_sources = get_model_options().get('sources')
|
|
model_context = __name__ + '.' + state_manager.get_item('lip_syncer_model')
|
|
return inference_manager.get_inference_pool(model_context, model_sources)
|
|
|
|
|
|
def clear_inference_pool() -> None:
|
|
model_context = __name__ + '.' + state_manager.get_item('lip_syncer_model')
|
|
inference_manager.clear_inference_pool(model_context)
|
|
|
|
|
|
def get_model_options() -> ModelOptions:
|
|
lip_syncer_model = state_manager.get_item('lip_syncer_model')
|
|
return MODEL_SET.get(lip_syncer_model)
|
|
|
|
|
|
def register_args(program : ArgumentParser) -> None:
|
|
group_processors = find_argument_group(program, 'processors')
|
|
if group_processors:
|
|
group_processors.add_argument('--lip-syncer-model', help = wording.get('help.lip_syncer_model'), default = config.get_str_value('processors.lip_syncer_model', 'wav2lip_gan'), choices = processors_choices.lip_syncer_models)
|
|
facefusion.jobs.job_store.register_step_keys([ 'lip_syncer_model' ])
|
|
|
|
|
|
def apply_args(args : Args, apply_state_item : ApplyStateItem) -> None:
|
|
apply_state_item('lip_syncer_model', args.get('lip_syncer_model'))
|
|
|
|
|
|
def pre_check() -> bool:
|
|
download_directory_path = resolve_relative_path('../.assets/models')
|
|
model_hashes = get_model_options().get('hashes')
|
|
model_sources = get_model_options().get('sources')
|
|
|
|
return conditional_download_hashes(download_directory_path, model_hashes) and conditional_download_sources(download_directory_path, model_sources)
|
|
|
|
|
|
def pre_process(mode : ProcessMode) -> bool:
|
|
if not has_audio(state_manager.get_item('source_paths')):
|
|
logger.error(wording.get('choose_audio_source') + wording.get('exclamation_mark'), __name__)
|
|
return False
|
|
if mode in [ 'output', 'preview' ] and not is_image(state_manager.get_item('target_path')) and not is_video(state_manager.get_item('target_path')):
|
|
logger.error(wording.get('choose_image_or_video_target') + wording.get('exclamation_mark'), __name__)
|
|
return False
|
|
if mode == 'output' and not in_directory(state_manager.get_item('output_path')):
|
|
logger.error(wording.get('specify_image_or_video_output') + wording.get('exclamation_mark'), __name__)
|
|
return False
|
|
if mode == 'output' and not same_file_extension([ state_manager.get_item('target_path'), state_manager.get_item('output_path') ]):
|
|
logger.error(wording.get('match_target_and_output_extension') + wording.get('exclamation_mark'), __name__)
|
|
return False
|
|
return True
|
|
|
|
|
|
def post_process() -> None:
|
|
read_static_image.cache_clear()
|
|
read_static_voice.cache_clear()
|
|
if state_manager.get_item('video_memory_strategy') in [ 'strict', 'moderate' ]:
|
|
clear_inference_pool()
|
|
if state_manager.get_item('video_memory_strategy') == 'strict':
|
|
content_analyser.clear_inference_pool()
|
|
face_classifier.clear_inference_pool()
|
|
face_detector.clear_inference_pool()
|
|
face_landmarker.clear_inference_pool()
|
|
face_masker.clear_inference_pool()
|
|
face_recognizer.clear_inference_pool()
|
|
voice_extractor.clear_inference_pool()
|
|
|
|
|
|
def sync_lip(target_face : Face, temp_audio_frame : AudioFrame, temp_vision_frame : VisionFrame) -> VisionFrame:
|
|
model_size = get_model_options().get('size')
|
|
temp_audio_frame = prepare_audio_frame(temp_audio_frame)
|
|
crop_vision_frame, affine_matrix = warp_face_by_face_landmark_5(temp_vision_frame, target_face.landmark_set.get('5/68'), 'ffhq_512', (512, 512))
|
|
face_landmark_68 = cv2.transform(target_face.landmark_set.get('68').reshape(1, -1, 2), affine_matrix).reshape(-1, 2)
|
|
bounding_box = create_bounding_box(face_landmark_68)
|
|
bounding_box[1] -= numpy.abs(bounding_box[3] - bounding_box[1]) * 0.125
|
|
mouth_mask = create_mouth_mask(face_landmark_68)
|
|
box_mask = create_static_box_mask(crop_vision_frame.shape[:2][::-1], state_manager.get_item('face_mask_blur'), state_manager.get_item('face_mask_padding'))
|
|
crop_masks =\
|
|
[
|
|
mouth_mask,
|
|
box_mask
|
|
]
|
|
|
|
if 'occlusion' in state_manager.get_item('face_mask_types'):
|
|
occlusion_mask = create_occlusion_mask(crop_vision_frame)
|
|
crop_masks.append(occlusion_mask)
|
|
|
|
close_vision_frame, close_matrix = warp_face_by_bounding_box(crop_vision_frame, bounding_box, model_size)
|
|
close_vision_frame = prepare_crop_frame(close_vision_frame)
|
|
close_vision_frame = forward(temp_audio_frame, close_vision_frame)
|
|
close_vision_frame = normalize_close_frame(close_vision_frame)
|
|
crop_vision_frame = cv2.warpAffine(close_vision_frame, cv2.invertAffineTransform(close_matrix), (512, 512), borderMode = cv2.BORDER_REPLICATE)
|
|
crop_mask = numpy.minimum.reduce(crop_masks)
|
|
paste_vision_frame = paste_back(temp_vision_frame, crop_vision_frame, crop_mask, affine_matrix)
|
|
return paste_vision_frame
|
|
|
|
|
|
def forward(temp_audio_frame : AudioFrame, close_vision_frame : VisionFrame) -> VisionFrame:
|
|
lip_syncer = get_inference_pool().get('lip_syncer')
|
|
|
|
with conditional_thread_semaphore():
|
|
close_vision_frame = lip_syncer.run(None,
|
|
{
|
|
'source': temp_audio_frame,
|
|
'target': close_vision_frame
|
|
})[0]
|
|
|
|
return close_vision_frame
|
|
|
|
|
|
def prepare_audio_frame(temp_audio_frame : AudioFrame) -> AudioFrame:
|
|
temp_audio_frame = numpy.maximum(numpy.exp(-5 * numpy.log(10)), temp_audio_frame)
|
|
temp_audio_frame = numpy.log10(temp_audio_frame) * 1.6 + 3.2
|
|
temp_audio_frame = temp_audio_frame.clip(-4, 4).astype(numpy.float32)
|
|
temp_audio_frame = numpy.expand_dims(temp_audio_frame, axis = (0, 1))
|
|
return temp_audio_frame
|
|
|
|
|
|
def prepare_crop_frame(crop_vision_frame : VisionFrame) -> VisionFrame:
|
|
crop_vision_frame = numpy.expand_dims(crop_vision_frame, axis = 0)
|
|
prepare_vision_frame = crop_vision_frame.copy()
|
|
prepare_vision_frame[:, 48:] = 0
|
|
crop_vision_frame = numpy.concatenate((prepare_vision_frame, crop_vision_frame), axis = 3)
|
|
crop_vision_frame = crop_vision_frame.transpose(0, 3, 1, 2).astype('float32') / 255.0
|
|
return crop_vision_frame
|
|
|
|
|
|
def normalize_close_frame(crop_vision_frame : VisionFrame) -> VisionFrame:
|
|
crop_vision_frame = crop_vision_frame[0].transpose(1, 2, 0)
|
|
crop_vision_frame = crop_vision_frame.clip(0, 1) * 255
|
|
crop_vision_frame = crop_vision_frame.astype(numpy.uint8)
|
|
return crop_vision_frame
|
|
|
|
|
|
def get_reference_frame(source_face : Face, target_face : Face, temp_vision_frame : VisionFrame) -> VisionFrame:
|
|
pass
|
|
|
|
|
|
def process_frame(inputs : LipSyncerInputs) -> VisionFrame:
|
|
reference_faces = inputs.get('reference_faces')
|
|
source_audio_frame = inputs.get('source_audio_frame')
|
|
target_vision_frame = inputs.get('target_vision_frame')
|
|
many_faces = sort_and_filter_faces(get_many_faces([ target_vision_frame ]))
|
|
|
|
if state_manager.get_item('face_selector_mode') == 'many':
|
|
if many_faces:
|
|
for target_face in many_faces:
|
|
target_vision_frame = sync_lip(target_face, source_audio_frame, target_vision_frame)
|
|
if state_manager.get_item('face_selector_mode') == 'one':
|
|
target_face = get_one_face(many_faces)
|
|
if target_face:
|
|
target_vision_frame = sync_lip(target_face, source_audio_frame, target_vision_frame)
|
|
if state_manager.get_item('face_selector_mode') == 'reference':
|
|
similar_faces = find_similar_faces(many_faces, reference_faces, state_manager.get_item('reference_face_distance'))
|
|
if similar_faces:
|
|
for similar_face in similar_faces:
|
|
target_vision_frame = sync_lip(similar_face, source_audio_frame, target_vision_frame)
|
|
return target_vision_frame
|
|
|
|
|
|
def process_frames(source_paths : List[str], queue_payloads : List[QueuePayload], update_progress : UpdateProgress) -> None:
|
|
reference_faces = get_reference_faces() if 'reference' in state_manager.get_item('face_selector_mode') else None
|
|
source_audio_path = get_first(filter_audio_paths(source_paths))
|
|
temp_video_fps = restrict_video_fps(state_manager.get_item('target_path'), state_manager.get_item('output_video_fps'))
|
|
|
|
for queue_payload in process_manager.manage(queue_payloads):
|
|
frame_number = queue_payload.get('frame_number')
|
|
target_vision_path = queue_payload.get('frame_path')
|
|
source_audio_frame = get_voice_frame(source_audio_path, temp_video_fps, frame_number)
|
|
if not numpy.any(source_audio_frame):
|
|
source_audio_frame = create_empty_audio_frame()
|
|
target_vision_frame = read_image(target_vision_path)
|
|
output_vision_frame = process_frame(
|
|
{
|
|
'reference_faces': reference_faces,
|
|
'source_audio_frame': source_audio_frame,
|
|
'target_vision_frame': target_vision_frame
|
|
})
|
|
write_image(target_vision_path, output_vision_frame)
|
|
update_progress(1)
|
|
|
|
|
|
def process_image(source_paths : List[str], target_path : str, output_path : str) -> None:
|
|
reference_faces = get_reference_faces() if 'reference' in state_manager.get_item('face_selector_mode') else None
|
|
source_audio_frame = create_empty_audio_frame()
|
|
target_vision_frame = read_static_image(target_path)
|
|
output_vision_frame = process_frame(
|
|
{
|
|
'reference_faces': reference_faces,
|
|
'source_audio_frame': source_audio_frame,
|
|
'target_vision_frame': target_vision_frame
|
|
})
|
|
write_image(output_path, output_vision_frame)
|
|
|
|
|
|
def process_video(source_paths : List[str], temp_frame_paths : List[str]) -> None:
|
|
source_audio_paths = filter_audio_paths(state_manager.get_item('source_paths'))
|
|
temp_video_fps = restrict_video_fps(state_manager.get_item('target_path'), state_manager.get_item('output_video_fps'))
|
|
for source_audio_path in source_audio_paths:
|
|
read_static_voice(source_audio_path, temp_video_fps)
|
|
processors.multi_process_frames(source_paths, temp_frame_paths, process_frames)
|