* Mark as NEXT * Reduce caching to avoid RAM explosion * Reduce caching to avoid RAM explosion * Update dependencies * add face-detector-pad-factor * update facefusion.ini * fix test * change pad to margin * fix order * add prepare margin * use 50% max margin * Minor fixes part2 * Minor fixes part3 * Minor fixes part4 * Minor fixes part1 * Downgrade onnxruntime as of BiRefNet broken on CPU add test update update facefusion.ini add birefnet * rename models add more models * Fix versions * Add .claude to gitignore * add normalize color add 4 channel add colors * worflows * cleanup * cleanup * cleanup * cleanup * add more models (#961) * Fix naming * changes * Fix style and mock Gradio * Fix style and mock Gradio * Fix style and mock Gradio * apply clamp * remove clamp * Add normalizer test * Introduce sanitizer for the rescue (#963) * Introduce sanitizer for the rescue * Introduce sanitizer for the rescue * Introduce sanitizer for the rescue * prepare ffmpeg for alpha support * Some cleanup * Some cleanup * Fix CI * List as TypeAlias is not allowed (#967) * List as TypeAlias is not allowed * List as TypeAlias is not allowed * List as TypeAlias is not allowed * List as TypeAlias is not allowed * Add mpeg and mxf support (#968) * Add mpeg support * Add mxf support * Adjust fix_xxx_encoder for the new formats * Extend output pattern for batch-run (#969) * Extend output pattern for batch-run * Add {target_extension} to allowed mixed files * Catch invalid output pattern keys * alpha support * cleanup * cleanup * add ProcessorOutputs type * fix preview and streamer, support alpha for background_remover * Refactor/open close processors (#972) * Introduce open/close processors * Add locales for translator * Introduce __autoload__ for translator * More cleanup * Fix import issues * Resolve the scope situation for locals * Fix installer by not using translator * Fixes after merge * Fixes after merge * Fix translator keys in ui * Use LOCALS in installer * Update and partial fix DirectML * Use latest onnxruntime * Fix performance * Fix lint issues * fix mask * fix lint * fix lint * Remove default from translator.get() * remove 'framerate=' * fix test * Rename and reorder models * Align naming * add alpha preview * fix frame-by-frame * Add alpha effect via css * preview support alpha channel * fix preview modes * Use official assets repositories * Add support for u2net_cloth * fix naming * Add more models * Add vendor, license and year direct to the models * Add vendor, license and year direct to the models * Update dependencies, Minor CSS adjustment * Ready for 3.5.0 * Fix naming * Update about messages * Fix return * Use groups to show/hide * Update preview * Conditional merge mask * Conditional merge mask * Fix import order --------- Co-authored-by: harisreedhar <h4harisreedhar.s.s@gmail.com> Co-authored-by: Harisreedhar <46858047+harisreedhar@users.noreply.github.com>
175 lines
6.4 KiB
Python
175 lines
6.4 KiB
Python
import os
|
|
import subprocess
|
|
from functools import lru_cache
|
|
from typing import List, Optional, Tuple
|
|
from urllib.parse import urlparse
|
|
|
|
from tqdm import tqdm
|
|
|
|
import facefusion.choices
|
|
from facefusion import curl_builder, logger, process_manager, state_manager, translator
|
|
from facefusion.filesystem import get_file_name, get_file_size, is_file, remove_file
|
|
from facefusion.hash_helper import validate_hash
|
|
from facefusion.types import Command, DownloadProvider, DownloadSet
|
|
|
|
|
|
def open_curl(commands : List[Command]) -> subprocess.Popen[bytes]:
|
|
commands = curl_builder.run(commands)
|
|
return subprocess.Popen(commands, stdin = subprocess.PIPE, stdout = subprocess.PIPE)
|
|
|
|
|
|
def conditional_download(download_directory_path : str, urls : List[str]) -> None:
|
|
for url in urls:
|
|
download_file_name = os.path.basename(urlparse(url).path)
|
|
download_file_path = os.path.join(download_directory_path, download_file_name)
|
|
initial_size = get_file_size(download_file_path)
|
|
download_size = get_static_download_size(url)
|
|
|
|
if initial_size < download_size:
|
|
with tqdm(total = download_size, initial = initial_size, desc = translator.get('downloading'), unit = 'B', unit_scale = True, unit_divisor = 1024, ascii = ' =', disable = state_manager.get_item('log_level') in [ 'warn', 'error' ]) as progress:
|
|
commands = curl_builder.chain(
|
|
curl_builder.download(url, download_file_path),
|
|
curl_builder.set_timeout(5)
|
|
)
|
|
open_curl(commands)
|
|
current_size = initial_size
|
|
progress.set_postfix(download_providers = state_manager.get_item('download_providers'), file_name = download_file_name)
|
|
|
|
while current_size < download_size:
|
|
if is_file(download_file_path):
|
|
current_size = get_file_size(download_file_path)
|
|
progress.update(current_size - progress.n)
|
|
|
|
|
|
@lru_cache(maxsize = 64)
|
|
def get_static_download_size(url : str) -> int:
|
|
commands = curl_builder.chain(
|
|
curl_builder.head(url),
|
|
curl_builder.set_timeout(5)
|
|
)
|
|
process = open_curl(commands)
|
|
lines = reversed(process.stdout.readlines())
|
|
|
|
for line in lines:
|
|
__line__ = line.decode().lower()
|
|
if 'content-length:' in __line__:
|
|
_, content_length = __line__.split('content-length:')
|
|
return int(content_length)
|
|
|
|
return 0
|
|
|
|
|
|
@lru_cache(maxsize = 64)
|
|
def ping_static_url(url : str) -> bool:
|
|
commands = curl_builder.chain(
|
|
curl_builder.head(url),
|
|
curl_builder.set_timeout(5)
|
|
)
|
|
process = open_curl(commands)
|
|
process.communicate()
|
|
return process.returncode == 0
|
|
|
|
|
|
def conditional_download_hashes(hash_set : DownloadSet) -> bool:
|
|
hash_paths = [ hash_set.get(hash_key).get('path') for hash_key in hash_set.keys() ]
|
|
|
|
process_manager.check()
|
|
_, invalid_hash_paths = validate_hash_paths(hash_paths)
|
|
if invalid_hash_paths:
|
|
for index in hash_set:
|
|
if hash_set.get(index).get('path') in invalid_hash_paths:
|
|
invalid_hash_url = hash_set.get(index).get('url')
|
|
if invalid_hash_url:
|
|
download_directory_path = os.path.dirname(hash_set.get(index).get('path'))
|
|
conditional_download(download_directory_path, [ invalid_hash_url ])
|
|
|
|
valid_hash_paths, invalid_hash_paths = validate_hash_paths(hash_paths)
|
|
|
|
for valid_hash_path in valid_hash_paths:
|
|
valid_hash_file_name = get_file_name(valid_hash_path)
|
|
logger.debug(translator.get('validating_hash_succeeded').format(hash_file_name = valid_hash_file_name), __name__)
|
|
for invalid_hash_path in invalid_hash_paths:
|
|
invalid_hash_file_name = get_file_name(invalid_hash_path)
|
|
logger.error(translator.get('validating_hash_failed').format(hash_file_name = invalid_hash_file_name), __name__)
|
|
|
|
if not invalid_hash_paths:
|
|
process_manager.end()
|
|
return not invalid_hash_paths
|
|
|
|
|
|
def conditional_download_sources(source_set : DownloadSet) -> bool:
|
|
source_paths = [ source_set.get(source_key).get('path') for source_key in source_set.keys() ]
|
|
|
|
process_manager.check()
|
|
_, invalid_source_paths = validate_source_paths(source_paths)
|
|
if invalid_source_paths:
|
|
for index in source_set:
|
|
if source_set.get(index).get('path') in invalid_source_paths:
|
|
invalid_source_url = source_set.get(index).get('url')
|
|
if invalid_source_url:
|
|
download_directory_path = os.path.dirname(source_set.get(index).get('path'))
|
|
conditional_download(download_directory_path, [ invalid_source_url ])
|
|
|
|
valid_source_paths, invalid_source_paths = validate_source_paths(source_paths)
|
|
|
|
for valid_source_path in valid_source_paths:
|
|
valid_source_file_name = get_file_name(valid_source_path)
|
|
logger.debug(translator.get('validating_source_succeeded').format(source_file_name = valid_source_file_name), __name__)
|
|
for invalid_source_path in invalid_source_paths:
|
|
invalid_source_file_name = get_file_name(invalid_source_path)
|
|
logger.error(translator.get('validating_source_failed').format(source_file_name = invalid_source_file_name), __name__)
|
|
|
|
if remove_file(invalid_source_path):
|
|
logger.error(translator.get('deleting_corrupt_source').format(source_file_name = invalid_source_file_name), __name__)
|
|
|
|
if not invalid_source_paths:
|
|
process_manager.end()
|
|
return not invalid_source_paths
|
|
|
|
|
|
def validate_hash_paths(hash_paths : List[str]) -> Tuple[List[str], List[str]]:
|
|
valid_hash_paths = []
|
|
invalid_hash_paths = []
|
|
|
|
for hash_path in hash_paths:
|
|
if is_file(hash_path):
|
|
valid_hash_paths.append(hash_path)
|
|
else:
|
|
invalid_hash_paths.append(hash_path)
|
|
|
|
return valid_hash_paths, invalid_hash_paths
|
|
|
|
|
|
def validate_source_paths(source_paths : List[str]) -> Tuple[List[str], List[str]]:
|
|
valid_source_paths = []
|
|
invalid_source_paths = []
|
|
|
|
for source_path in source_paths:
|
|
if validate_hash(source_path):
|
|
valid_source_paths.append(source_path)
|
|
else:
|
|
invalid_source_paths.append(source_path)
|
|
|
|
return valid_source_paths, invalid_source_paths
|
|
|
|
|
|
def resolve_download_url(base_name : str, file_name : str) -> Optional[str]:
|
|
download_providers = state_manager.get_item('download_providers')
|
|
|
|
for download_provider in download_providers:
|
|
download_url = resolve_download_url_by_provider(download_provider, base_name, file_name)
|
|
if download_url:
|
|
return download_url
|
|
|
|
return None
|
|
|
|
|
|
def resolve_download_url_by_provider(download_provider : DownloadProvider, base_name : str, file_name : str) -> Optional[str]:
|
|
download_provider_value = facefusion.choices.download_provider_set.get(download_provider)
|
|
|
|
for download_provider_url in download_provider_value.get('urls'):
|
|
if ping_static_url(download_provider_url):
|
|
return download_provider_url + download_provider_value.get('path').format(base_name = base_name, file_name = file_name)
|
|
|
|
return None
|