Asset Pipeline: Improve Performance #235
@ -20,7 +20,7 @@
|
|||||||
|
|
||||||
import logging
|
import logging
|
||||||
from typing import List, Tuple
|
from typing import List, Tuple
|
||||||
from . import prefs
|
from . import prefs, constants
|
||||||
|
|
||||||
|
|
||||||
def get_logger(name="asset_pipeline"):
|
def get_logger(name="asset_pipeline"):
|
||||||
@ -45,3 +45,89 @@ def get_logger(name="asset_pipeline"):
|
|||||||
# add ch to logger
|
# add ch to logger
|
||||||
logger.addHandler(ch)
|
logger.addHandler(ch)
|
||||||
return logger
|
return logger
|
||||||
|
|
||||||
|
|
||||||
|
PROFILE_KEYS = {
|
||||||
|
"IMPORT": "To import Collection & add suffixes",
|
||||||
|
"MAPPING": "To create Asset Mapping",
|
||||||
|
"TRANSFER_DATA": "To apply all Transferable Data",
|
||||||
|
"OBJECTS": "To remap all Obejcts",
|
||||||
|
"INDEXES": "To restore Active Indexes on all Objects",
|
||||||
|
"COLLECTIONS": "To remap all Collections",
|
||||||
|
"SHARED_IDS": "To remap all Shared IDs",
|
||||||
|
"MERGE": "To complete entire merge process",
|
||||||
|
"TOTAL": "Total time to sync this direction",
|
||||||
|
}
|
||||||
|
|
||||||
|
TD_KEYS = [type for type in constants.TRANSFER_DATA_TYPES]
|
||||||
|
|
||||||
|
INFO_KEYS = ["TOTAL"] # Profile Keys to print in the logger's info mode
|
||||||
|
|
||||||
|
_profiler_instance = None
|
||||||
|
|
||||||
|
|
||||||
|
def get_profiler():
|
||||||
|
global _profiler_instance
|
||||||
|
if not _profiler_instance:
|
||||||
|
_profiler_instance = Profiler()
|
||||||
|
return _profiler_instance
|
||||||
|
|
||||||
|
|
||||||
|
class Profiler:
|
||||||
|
|
||||||
|
def __init__(self) -> None:
|
||||||
|
self.pull_profiles = {}
|
||||||
|
self.push_profiles = {}
|
||||||
|
self._logger = get_logger()
|
||||||
|
|
||||||
|
def add(self, elapsed_time: int, key: str):
|
||||||
|
if self._is_push:
|
||||||
|
profiles = self.push_profiles
|
||||||
|
else: # is pull
|
||||||
|
profiles = self.pull_profiles
|
||||||
|
|
||||||
|
if key not in profiles:
|
||||||
|
profiles[key] = elapsed_time
|
||||||
|
else:
|
||||||
|
profiles[key] += elapsed_time
|
||||||
|
|
||||||
|
def log_all(self):
|
||||||
|
self.log_profiles("PULL", self.pull_profiles)
|
||||||
|
self.log_profiles("PUSH", self.push_profiles)
|
||||||
|
|
||||||
|
def log_profiles(self, direction: str, profiles: dict):
|
||||||
|
if profiles == {}:
|
||||||
|
return
|
||||||
|
for key, value in profiles.items():
|
||||||
|
seconds = self.get_non_scientific_number(value)
|
||||||
|
# Special case for transfer data keys
|
||||||
|
if key in TD_KEYS:
|
||||||
|
name = constants.TRANSFER_DATA_TYPES[key][0]
|
||||||
|
self._logger.debug(
|
||||||
|
f"{direction} TD: {name.upper()} - {seconds} seconds to transfer {name} data for all objects"
|
||||||
|
)
|
||||||
|
continue
|
||||||
|
msg = f"{direction} {key} - {seconds} seconds {PROFILE_KEYS[key]}"
|
||||||
|
if key in INFO_KEYS:
|
||||||
|
self._logger.info(msg)
|
||||||
|
else:
|
||||||
|
self._logger.debug(msg)
|
||||||
|
|
||||||
|
def get_non_scientific_number(self, x: float):
|
||||||
|
float_str = f'{x:.64f}'.rstrip('0')
|
||||||
|
|
||||||
|
significant_digits = 0
|
||||||
|
for index, c in enumerate(float_str):
|
||||||
|
if significant_digits == 3:
|
||||||
|
return float_str[:index:]
|
||||||
|
|
||||||
|
if c != "0" and c != ".":
|
||||||
|
significant_digits += 1
|
||||||
|
|
||||||
|
def reset(self):
|
||||||
|
self.pull_profiles = {}
|
||||||
|
self._is_push = False
|
||||||
|
self._logger = get_logger()
|
||||||
|
|
||||||
|
def set_push(self, is_push=True):
|
||||||
|
self._is_push = is_push
|
||||||
|
@ -169,26 +169,12 @@ class AssetTransferMapping:
|
|||||||
|
|
||||||
return coll_map
|
return coll_map
|
||||||
|
|
||||||
def _transfer_data_get_map_content(self, obj, target_obj, transfer_data_item):
|
def _get_transfer_data_dict(self, transfer_data_item):
|
||||||
temp_transfer_data = bpy.context.scene.asset_pipeline.temp_transfer_data
|
return {
|
||||||
temp_transfer_data_item_index = len(temp_transfer_data)
|
'name': transfer_data_item.name,
|
||||||
temp_transfer_data_item = transfer_data_add_entry(
|
"owner": transfer_data_item.owner,
|
||||||
transfer_data=temp_transfer_data,
|
"surrender": transfer_data_item.surrender,
|
||||||
name=transfer_data_item.name,
|
|
||||||
td_type_key=transfer_data_item.type,
|
|
||||||
task_layer_name=transfer_data_item.owner,
|
|
||||||
surrender=transfer_data_item.surrender,
|
|
||||||
)
|
|
||||||
|
|
||||||
map_item = {
|
|
||||||
'transfer_data_item_index': temp_transfer_data_item_index,
|
|
||||||
'source_obj': obj,
|
|
||||||
'target_obj': target_obj,
|
|
||||||
}
|
}
|
||||||
# Names of each map item need to be unique
|
|
||||||
# below name avoids name conflicts between different types
|
|
||||||
name = transfer_data_item.name + '_' + transfer_data_item.type + obj.name
|
|
||||||
return name, map_item
|
|
||||||
|
|
||||||
def _transfer_data_pair_not_local(self, td_1, td_2):
|
def _transfer_data_pair_not_local(self, td_1, td_2):
|
||||||
# Returns true if neither owners are local to current file
|
# Returns true if neither owners are local to current file
|
||||||
@ -243,11 +229,21 @@ class AssetTransferMapping:
|
|||||||
"""Adds item to Transfer Data Map"""
|
"""Adds item to Transfer Data Map"""
|
||||||
if self._transfer_data_is_surrendered(transfer_data_item):
|
if self._transfer_data_is_surrendered(transfer_data_item):
|
||||||
return
|
return
|
||||||
|
td_type_key = transfer_data_item.type
|
||||||
|
transfer_data_dict = self._get_transfer_data_dict(transfer_data_item)
|
||||||
|
|
||||||
name, map_item = self._transfer_data_get_map_content(
|
if not source_obj in self.transfer_data_map:
|
||||||
source_obj, target_obj, transfer_data_item
|
self.transfer_data_map[source_obj] = {
|
||||||
)
|
"target_obj": target_obj,
|
||||||
self.transfer_data_map[name] = map_item
|
"td_types": {td_type_key: [transfer_data_dict]},
|
||||||
|
}
|
||||||
|
return
|
||||||
|
|
||||||
|
if not td_type_key in self.transfer_data_map[source_obj]["td_types"]:
|
||||||
|
self.transfer_data_map[source_obj]["td_types"][td_type_key] = [transfer_data_dict]
|
||||||
|
return
|
||||||
|
else:
|
||||||
|
self.transfer_data_map[source_obj]["td_types"][td_type_key].append(transfer_data_dict)
|
||||||
|
|
||||||
def _transfer_data_map_item(self, source_obj, target_obj, transfer_data_item):
|
def _transfer_data_map_item(self, source_obj, target_obj, transfer_data_item):
|
||||||
"""Verifies if Transfer Data Item is valid/can be mapped"""
|
"""Verifies if Transfer Data Item is valid/can be mapped"""
|
||||||
@ -268,12 +264,11 @@ class AssetTransferMapping:
|
|||||||
|
|
||||||
def _gen_transfer_data_map(self):
|
def _gen_transfer_data_map(self):
|
||||||
# Generate Mapping for Transfer Data Items
|
# Generate Mapping for Transfer Data Items
|
||||||
temp_transfer_data = bpy.context.scene.asset_pipeline.temp_transfer_data
|
|
||||||
temp_transfer_data.clear()
|
|
||||||
for objs in self.object_map.items():
|
for objs in self.object_map.items():
|
||||||
source_obj, target_obj = objs
|
_, target_obj = objs
|
||||||
for obj in objs:
|
for obj in objs:
|
||||||
# Must execute for both objs in map (so we map external and local TD)
|
# Must execute for both objs in map (so we map external and local TD)
|
||||||
|
# Must include maps even if obj==target_obj to preserve exisiting local TD entry
|
||||||
for transfer_data_item in obj.transfer_data_ownership:
|
for transfer_data_item in obj.transfer_data_ownership:
|
||||||
if self._transfer_data_check_conflict(obj, transfer_data_item):
|
if self._transfer_data_check_conflict(obj, transfer_data_item):
|
||||||
continue
|
continue
|
||||||
@ -284,19 +279,20 @@ class AssetTransferMapping:
|
|||||||
# Generate a Map of Indexes that need to be set post merge
|
# Generate a Map of Indexes that need to be set post merge
|
||||||
# Stores active_uv & active_color_attribute
|
# Stores active_uv & active_color_attribute
|
||||||
index_map = {}
|
index_map = {}
|
||||||
for _, item in self.transfer_data_map.items():
|
|
||||||
temp_transfer_data = bpy.context.scene.asset_pipeline.temp_transfer_data
|
|
||||||
temp_transfer_data_item = temp_transfer_data[item.get('transfer_data_item_index')]
|
|
||||||
source_obj = item.get('source_obj')
|
|
||||||
target_obj = item.get('target_obj')
|
|
||||||
|
|
||||||
if temp_transfer_data_item.type != constants.MATERIAL_SLOT_KEY:
|
for source_obj in self.transfer_data_map:
|
||||||
|
target_obj = self.transfer_data_map[source_obj]["target_obj"]
|
||||||
|
td_types = self.transfer_data_map[source_obj]["td_types"]
|
||||||
|
for td_type_key, _ in td_types.items():
|
||||||
|
if td_type_key != constants.MATERIAL_SLOT_KEY:
|
||||||
continue
|
continue
|
||||||
if source_obj.type != 'MESH':
|
if source_obj.type != 'MESH':
|
||||||
continue
|
continue
|
||||||
|
|
||||||
active_uv_name = (
|
active_uv_name = (
|
||||||
source_obj.data.uv_layers.active.name if source_obj.data.uv_layers.active else ''
|
source_obj.data.uv_layers.active.name
|
||||||
|
if source_obj.data.uv_layers.active
|
||||||
|
else ''
|
||||||
)
|
)
|
||||||
active_color_attribute_name = source_obj.data.color_attributes.active_color_name
|
active_color_attribute_name = source_obj.data.color_attributes.active_color_name
|
||||||
index_map[source_obj] = {
|
index_map[source_obj] = {
|
||||||
|
@ -21,6 +21,7 @@ from .transfer_data.transfer_functions.transfer_function_util.active_indexes imp
|
|||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Dict
|
from typing import Dict
|
||||||
from .. import constants, logging
|
from .. import constants, logging
|
||||||
|
import time
|
||||||
|
|
||||||
|
|
||||||
def ownership_transfer_data_cleanup(
|
def ownership_transfer_data_cleanup(
|
||||||
@ -175,6 +176,11 @@ def merge_task_layer(
|
|||||||
local_tls: (list[str]): list of task layers that are local to the current file
|
local_tls: (list[str]): list of task layers that are local to the current file
|
||||||
external_file (Path): external file to pull data into the current file from
|
external_file (Path): external file to pull data into the current file from
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
logger = logging.get_logger()
|
||||||
|
profiles = logging.get_profiler()
|
||||||
|
|
||||||
|
start_time = time.time()
|
||||||
local_col = context.scene.asset_pipeline.asset_collection
|
local_col = context.scene.asset_pipeline.asset_collection
|
||||||
if not local_col:
|
if not local_col:
|
||||||
return "Unable to find Asset Collection"
|
return "Unable to find Asset Collection"
|
||||||
@ -185,6 +191,8 @@ def merge_task_layer(
|
|||||||
|
|
||||||
appended_col = import_data_from_lib(external_file, "collections", col_base_name)
|
appended_col = import_data_from_lib(external_file, "collections", col_base_name)
|
||||||
merge_add_suffix_to_hierarchy(appended_col, external_suffix)
|
merge_add_suffix_to_hierarchy(appended_col, external_suffix)
|
||||||
|
imported_time = time.time()
|
||||||
|
profiles.add((imported_time - start_time), "IMPORT")
|
||||||
|
|
||||||
local_col = bpy.data.collections[f"{col_base_name}.{local_suffix}"]
|
local_col = bpy.data.collections[f"{col_base_name}.{local_suffix}"]
|
||||||
external_col = bpy.data.collections[f"{col_base_name}.{external_suffix}"]
|
external_col = bpy.data.collections[f"{col_base_name}.{external_suffix}"]
|
||||||
@ -204,6 +212,8 @@ def merge_task_layer(
|
|||||||
type_name = get_id_type_name(type(conflict_obj))
|
type_name = get_id_type_name(type(conflict_obj))
|
||||||
error_msg += f"Ownership conflict found for {type_name}: '{conflict_obj.name}'\n"
|
error_msg += f"Ownership conflict found for {type_name}: '{conflict_obj.name}'\n"
|
||||||
return error_msg
|
return error_msg
|
||||||
|
mapped_time = time.time()
|
||||||
|
profiles.add((mapped_time - imported_time), "MAPPING")
|
||||||
|
|
||||||
# Remove all Transferable Data from target objects
|
# Remove all Transferable Data from target objects
|
||||||
for source_obj in map.object_map:
|
for source_obj in map.object_map:
|
||||||
@ -211,11 +221,15 @@ def merge_task_layer(
|
|||||||
target_obj.transfer_data_ownership.clear()
|
target_obj.transfer_data_ownership.clear()
|
||||||
|
|
||||||
apply_transfer_data(context, map.transfer_data_map)
|
apply_transfer_data(context, map.transfer_data_map)
|
||||||
|
apply_td_time = time.time()
|
||||||
|
profiles.add((apply_td_time - mapped_time), "TRANSFER_DATA")
|
||||||
|
|
||||||
for source_obj in map.object_map:
|
for source_obj in map.object_map:
|
||||||
target_obj = map.object_map[source_obj]
|
target_obj = map.object_map[source_obj]
|
||||||
remap_user(source_obj, target_obj)
|
remap_user(source_obj, target_obj)
|
||||||
transfer_data_clean(target_obj)
|
transfer_data_clean(target_obj)
|
||||||
|
obj_remap_time = time.time()
|
||||||
|
profiles.add((obj_remap_time - apply_td_time), "OBJECTS")
|
||||||
|
|
||||||
# Restore Active UV Layer and Active Color Attributes
|
# Restore Active UV Layer and Active Color Attributes
|
||||||
for _, index_map_item in map.index_map.items():
|
for _, index_map_item in map.index_map.items():
|
||||||
@ -224,6 +238,8 @@ def merge_task_layer(
|
|||||||
transfer_active_color_attribute_index(
|
transfer_active_color_attribute_index(
|
||||||
target_obj, index_map_item.get('active_color_attribute_name')
|
target_obj, index_map_item.get('active_color_attribute_name')
|
||||||
)
|
)
|
||||||
|
index_time = time.time()
|
||||||
|
profiles.add((index_time - obj_remap_time), "INDEXES")
|
||||||
|
|
||||||
for col in map.collection_map:
|
for col in map.collection_map:
|
||||||
remap_user(col, map.collection_map[col])
|
remap_user(col, map.collection_map[col])
|
||||||
@ -233,13 +249,17 @@ def merge_task_layer(
|
|||||||
|
|
||||||
for col in map.external_col_to_remove:
|
for col in map.external_col_to_remove:
|
||||||
local_col.children.unlink(col)
|
local_col.children.unlink(col)
|
||||||
|
col_remap_time = time.time()
|
||||||
|
profiles.add((col_remap_time - index_time), "COLLECTIONS")
|
||||||
|
|
||||||
for id in map.shared_id_map:
|
for id in map.shared_id_map:
|
||||||
remap_user(id, map.shared_id_map[id])
|
remap_user(id, map.shared_id_map[id])
|
||||||
|
shared_id_remap_time = time.time()
|
||||||
|
profiles.add((shared_id_remap_time - col_remap_time), "SHARED_IDS")
|
||||||
|
|
||||||
bpy.ops.outliner.orphans_purge(do_local_ids=True, do_linked_ids=False, do_recursive=True)
|
bpy.ops.outliner.orphans_purge(do_local_ids=True, do_linked_ids=False, do_recursive=True)
|
||||||
merge_remove_suffix_from_hierarchy(local_col)
|
merge_remove_suffix_from_hierarchy(local_col)
|
||||||
|
profiles.add((time.time() - start_time), "MERGE")
|
||||||
|
|
||||||
def import_data_from_lib(
|
def import_data_from_lib(
|
||||||
libpath: Path,
|
libpath: Path,
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
import bpy
|
import bpy
|
||||||
|
import time
|
||||||
from .transfer_functions import (
|
from .transfer_functions import (
|
||||||
attributes,
|
attributes,
|
||||||
constraints,
|
constraints,
|
||||||
@ -9,15 +9,20 @@ from .transfer_functions import (
|
|||||||
vertex_groups,
|
vertex_groups,
|
||||||
materials,
|
materials,
|
||||||
)
|
)
|
||||||
|
from typing import List
|
||||||
from ... import constants, logging
|
from ... import constants, logging
|
||||||
|
|
||||||
from .transfer_util import (
|
from .transfer_util import (
|
||||||
transfer_data_add_entry,
|
transfer_data_add_entry,
|
||||||
check_transfer_data_entry,
|
check_transfer_data_entry,
|
||||||
|
link_objs_to_collection,
|
||||||
|
isolate_collection,
|
||||||
)
|
)
|
||||||
|
|
||||||
def copy_transfer_data_ownership(transfer_data_item, target_obj: bpy.types.Object) -> None:
|
|
||||||
|
# TODO use logging module here
|
||||||
|
def copy_transfer_data_ownership(
|
||||||
|
td_type_key: str, target_obj: bpy.types.Object, transfer_data_dict: dict
|
||||||
|
) -> None:
|
||||||
"""Copy Transferable Data item to object if non entry exists
|
"""Copy Transferable Data item to object if non entry exists
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@ -27,16 +32,16 @@ def copy_transfer_data_ownership(transfer_data_item, target_obj: bpy.types.Objec
|
|||||||
transfer_data = target_obj.transfer_data_ownership
|
transfer_data = target_obj.transfer_data_ownership
|
||||||
matches = check_transfer_data_entry(
|
matches = check_transfer_data_entry(
|
||||||
transfer_data,
|
transfer_data,
|
||||||
transfer_data_item.name,
|
transfer_data_dict["name"],
|
||||||
transfer_data_item.type,
|
td_type_key,
|
||||||
)
|
)
|
||||||
if len(matches) == 0:
|
if len(matches) == 0:
|
||||||
transfer_data_add_entry(
|
transfer_data_add_entry(
|
||||||
transfer_data,
|
transfer_data,
|
||||||
transfer_data_item.name,
|
transfer_data_dict["name"],
|
||||||
transfer_data_item.type,
|
td_type_key,
|
||||||
transfer_data_item.owner,
|
transfer_data_dict["owner"],
|
||||||
transfer_data_item.surrender,
|
transfer_data_dict["surrender"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -96,6 +101,93 @@ def init_transfer_data(
|
|||||||
attributes.init_attributes(scene, obj)
|
attributes.init_attributes(scene, obj)
|
||||||
|
|
||||||
|
|
||||||
|
def apply_transfer_data_items(
|
||||||
|
context,
|
||||||
|
source_obj: bpy.types.Object,
|
||||||
|
target_obj: bpy.types.Object,
|
||||||
|
td_type_key: str,
|
||||||
|
transfer_data_dicts: List[dict],
|
||||||
|
):
|
||||||
|
logger = logging.get_logger()
|
||||||
|
# Get source/target from first item in list, because all items in list are same object/type
|
||||||
|
if target_obj is None:
|
||||||
|
logger.warning(f"Failed to Transfer {td_type_key.title()} from {source_obj.name}")
|
||||||
|
return
|
||||||
|
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
copy_transfer_data_ownership(td_type_key, target_obj, transfer_data_dict)
|
||||||
|
|
||||||
|
# if TD Source is Target, restore the ownership data but don't transfer anything
|
||||||
|
if source_obj == target_obj:
|
||||||
|
return
|
||||||
|
|
||||||
|
if td_type_key == constants.VERTEX_GROUP_KEY:
|
||||||
|
# Transfer All Vertex Groups in one go
|
||||||
|
logger.debug(f"Transferring All Vertex Groups from {source_obj.name} to {target_obj.name}.")
|
||||||
|
vertex_groups.transfer_vertex_groups(
|
||||||
|
vertex_group_names=[item["name"] for item in transfer_data_dicts],
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
)
|
||||||
|
if td_type_key == constants.MODIFIER_KEY:
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
logger.debug(
|
||||||
|
f"Transferring Modifier {transfer_data_dict['name']} from {source_obj.name} to {target_obj.name}."
|
||||||
|
)
|
||||||
|
modifers.transfer_modifier(
|
||||||
|
modifier_name=transfer_data_dict["name"],
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
)
|
||||||
|
if td_type_key == constants.CONSTRAINT_KEY:
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
logger.debug(
|
||||||
|
f"Transferring Constraint {transfer_data_dict['name']} from {source_obj.name} to {target_obj.name}."
|
||||||
|
)
|
||||||
|
constraints.transfer_constraint(
|
||||||
|
constraint_name=transfer_data_dict["name"],
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
)
|
||||||
|
if td_type_key == constants.MATERIAL_SLOT_KEY:
|
||||||
|
logger.debug(f"Transferring Materials from {source_obj.name} to {target_obj.name}.")
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
materials.transfer_materials(
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
)
|
||||||
|
if td_type_key == constants.SHAPE_KEY_KEY:
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
logger.debug(
|
||||||
|
f"Transferring Shape Key {transfer_data_dict['name']} from {source_obj.name} to {target_obj.name}."
|
||||||
|
)
|
||||||
|
shape_keys.transfer_shape_key(
|
||||||
|
context=context,
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
shape_key_name=transfer_data_dict["name"],
|
||||||
|
)
|
||||||
|
if td_type_key == constants.ATTRIBUTE_KEY:
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
logger.debug(
|
||||||
|
f"Transferring Attribute {transfer_data_dict['name']} from {source_obj.name} to {target_obj.name}."
|
||||||
|
)
|
||||||
|
attributes.transfer_attribute(
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
attribute_name=transfer_data_dict["name"],
|
||||||
|
)
|
||||||
|
if td_type_key == constants.PARENT_KEY:
|
||||||
|
for transfer_data_dict in transfer_data_dicts:
|
||||||
|
logger.debug(
|
||||||
|
f"Transferring Parent Relationship from {source_obj.name} to {target_obj.name}."
|
||||||
|
)
|
||||||
|
parent.transfer_parent(
|
||||||
|
target_obj=target_obj,
|
||||||
|
source_obj=source_obj,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def apply_transfer_data(context: bpy.types.Context, transfer_data_map) -> None:
|
def apply_transfer_data(context: bpy.types.Context, transfer_data_map) -> None:
|
||||||
"""Apply all Transferable Data from Transferable Data map onto objects.
|
"""Apply all Transferable Data from Transferable Data map onto objects.
|
||||||
Copies any Transferable Data owned by local layer onto objects owned by external layers.
|
Copies any Transferable Data owned by local layer onto objects owned by external layers.
|
||||||
@ -107,84 +199,19 @@ def apply_transfer_data(context: bpy.types.Context, transfer_data_map) -> None:
|
|||||||
context (bpy.types.Context): context of .blend file
|
context (bpy.types.Context): context of .blend file
|
||||||
transfer_data_map: Map generated by class AssetTransferMapping
|
transfer_data_map: Map generated by class AssetTransferMapping
|
||||||
"""
|
"""
|
||||||
logger = logging.get_logger()
|
# Create/isolate tmp collection to reduce depsgraph update time
|
||||||
for name in transfer_data_map:
|
profiler = logging.get_profiler()
|
||||||
temp_transfer_data = context.scene.asset_pipeline.temp_transfer_data
|
td_col = bpy.data.collections.new("ISO_COL_TEMP")
|
||||||
transfer_data = transfer_data_map[name]
|
with isolate_collection(td_col):
|
||||||
transfer_data_item = temp_transfer_data[transfer_data.get('transfer_data_item_index')]
|
# Loop over objects in Transfer data map
|
||||||
target_obj = transfer_data.get('target_obj')
|
for source_obj in transfer_data_map:
|
||||||
source_obj = transfer_data.get('source_obj')
|
target_obj = transfer_data_map[source_obj]["target_obj"]
|
||||||
if target_obj is None:
|
td_types = transfer_data_map[source_obj]["td_types"]
|
||||||
logger.warning(f"Failed to Transfer Data for {transfer_data_item.id_data.name}")
|
with link_objs_to_collection(set([target_obj, source_obj]), td_col):
|
||||||
continue
|
for td_type_key, td_dicts in td_types.items():
|
||||||
if transfer_data_item is None:
|
start_time = time.time()
|
||||||
continue
|
apply_transfer_data_items(
|
||||||
if source_obj != target_obj:
|
context, source_obj, target_obj, td_type_key, td_dicts
|
||||||
if transfer_data_item.type == constants.VERTEX_GROUP_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Vertex Group {transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
vertex_groups.transfer_vertex_group(
|
|
||||||
context=context,
|
|
||||||
vertex_group_name=transfer_data_item.name,
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.MODIFIER_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Modifier{transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
modifers.transfer_modifier(
|
|
||||||
modifier_name=transfer_data_item.name,
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.CONSTRAINT_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Constraint {transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
constraints.transfer_constraint(
|
|
||||||
constraint_name=transfer_data_item.name,
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.MATERIAL_SLOT_KEY:
|
|
||||||
logger.debug(f"Transferring Materiald from {source_obj.name} to {target_obj.name}.")
|
|
||||||
materials.transfer_materials(
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.SHAPE_KEY_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Shape Key {transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
shape_keys.transfer_shape_key(
|
|
||||||
context=context,
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
shape_key_name=transfer_data_item.name,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.ATTRIBUTE_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Attribute {transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
attributes.transfer_attribute(
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
attribute_name=transfer_data_item.name,
|
|
||||||
)
|
|
||||||
if transfer_data_item.type == constants.PARENT_KEY:
|
|
||||||
logger.debug(
|
|
||||||
f"Transferring Parent Relationship from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
parent.transfer_parent(
|
|
||||||
target_obj=target_obj,
|
|
||||||
source_obj=source_obj,
|
|
||||||
)
|
|
||||||
logger.debug(
|
|
||||||
f"Copying Ownership Data for {transfer_data_item.name} from {source_obj.name} to {target_obj.name}."
|
|
||||||
)
|
|
||||||
copy_transfer_data_ownership(
|
|
||||||
transfer_data_item=transfer_data_item,
|
|
||||||
target_obj=target_obj,
|
|
||||||
)
|
)
|
||||||
|
profiler.add(time.time() - start_time, td_type_key)
|
||||||
|
bpy.data.collections.remove(td_col)
|
||||||
|
@ -35,33 +35,27 @@ def init_vertex_groups(scene, obj):
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def transfer_vertex_group(
|
def transfer_vertex_groups(
|
||||||
context,
|
vertex_group_names: List[str],
|
||||||
vertex_group_name: str,
|
|
||||||
target_obj: bpy.types.Object,
|
target_obj: bpy.types.Object,
|
||||||
source_obj: bpy.types.Object,
|
source_obj: bpy.types.Object,
|
||||||
):
|
):
|
||||||
logger = logging.get_logger()
|
logger = logging.get_logger()
|
||||||
if target_obj == source_obj:
|
for vertex_group_name in vertex_group_names:
|
||||||
return
|
|
||||||
|
|
||||||
if not source_obj.vertex_groups.get(vertex_group_name):
|
if not source_obj.vertex_groups.get(vertex_group_name):
|
||||||
logger.error(f"Vertex Group {vertex_group_name} not found in {source_obj.name}")
|
logger.error(f"Vertex Group {vertex_group_name} not found in {source_obj.name}")
|
||||||
return
|
return
|
||||||
|
|
||||||
# If topology matches transfer directly, otherwise use vertex proximity
|
# If topology matches transfer directly, otherwise use vertex proximity
|
||||||
if is_obdata_identical(source_obj, target_obj):
|
if is_obdata_identical(source_obj, target_obj):
|
||||||
transfer_single_vgroup_by_topology(
|
for vertex_group_name in vertex_group_names:
|
||||||
source_obj, target_obj, vertex_group_name
|
transfer_single_vgroup_by_topology(source_obj, target_obj, vertex_group_name)
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
precalc_and_transfer_single_group(
|
precalc_and_transfer_multiple_groups(source_obj, target_obj, vertex_group_names, expand=2)
|
||||||
source_obj, target_obj, vertex_group_name, expand=2
|
|
||||||
)
|
|
||||||
|
|
||||||
def transfer_single_vgroup_by_topology(source_obj, target_obj, vgroup_name):
|
def transfer_single_vgroup_by_topology(source_obj, target_obj, vgroup_name):
|
||||||
""" Function to quickly transfer single vertex group between mesh objects in case of matching topology.
|
"""Function to quickly transfer single vertex group between mesh objects in case of matching topology."""
|
||||||
"""
|
|
||||||
|
|
||||||
# Remove group from the target obj if it already exists. TODO: de-duplicate
|
# Remove group from the target obj if it already exists. TODO: de-duplicate
|
||||||
tgt_vg = target_obj.vertex_groups.get(vgroup_name)
|
tgt_vg = target_obj.vertex_groups.get(vgroup_name)
|
||||||
@ -75,6 +69,33 @@ def transfer_single_vgroup_by_topology(source_obj, target_obj, vgroup_name):
|
|||||||
if vgroup_src.index in [g.group for g in v.groups]:
|
if vgroup_src.index in [g.group for g in v.groups]:
|
||||||
vgroup_tgt.add([v.index], vgroup_src.weight(v.index), 'REPLACE')
|
vgroup_tgt.add([v.index], vgroup_src.weight(v.index), 'REPLACE')
|
||||||
|
|
||||||
|
|
||||||
|
def precalc_and_transfer_multiple_groups(source_obj, target_obj, vgroup_names, expand=2):
|
||||||
|
"""Convenience function to transfer a single group. For transferring multiple groups,
|
||||||
|
this is very inefficient and shouldn't be used.
|
||||||
|
|
||||||
|
Instead, you should:
|
||||||
|
- build_kd_tree ONCE per source mesh.
|
||||||
|
- build_vert_influence_map and transfer_vertex_groups ONCE per object pair.
|
||||||
|
"""
|
||||||
|
|
||||||
|
# Remove group from the target obj if it already exists. TODO: de-duplicate
|
||||||
|
vgroups = [source_obj.vertex_groups[name] for name in vgroup_names]
|
||||||
|
for vgroup_name in vgroup_names:
|
||||||
|
target_vgroup = target_obj.vertex_groups.get(vgroup_name)
|
||||||
|
if target_vgroup:
|
||||||
|
target_obj.vertex_groups.remove(target_vgroup)
|
||||||
|
|
||||||
|
kd_tree = build_kdtree(source_obj.data)
|
||||||
|
vert_influence_map = build_vert_influence_map(source_obj, target_obj, kd_tree, expand)
|
||||||
|
transfer_multiple_vertex_groups(
|
||||||
|
source_obj,
|
||||||
|
target_obj,
|
||||||
|
vert_influence_map,
|
||||||
|
vgroups,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def precalc_and_transfer_single_group(source_obj, target_obj, vgroup_name, expand=2):
|
def precalc_and_transfer_single_group(source_obj, target_obj, vgroup_name, expand=2):
|
||||||
"""Convenience function to transfer a single group. For transferring multiple groups,
|
"""Convenience function to transfer a single group. For transferring multiple groups,
|
||||||
this is very inefficient and shouldn't be used.
|
this is very inefficient and shouldn't be used.
|
||||||
@ -90,10 +111,9 @@ def precalc_and_transfer_single_group(source_obj, target_obj, vgroup_name, expan
|
|||||||
target_obj.vertex_groups.remove(tgt_vg)
|
target_obj.vertex_groups.remove(tgt_vg)
|
||||||
|
|
||||||
kd_tree = build_kdtree(source_obj.data)
|
kd_tree = build_kdtree(source_obj.data)
|
||||||
vert_influence_map = build_vert_influence_map(
|
vert_influence_map = build_vert_influence_map(source_obj, target_obj, kd_tree, expand)
|
||||||
source_obj, target_obj, kd_tree, expand
|
|
||||||
)
|
transfer_multiple_vertex_groups(
|
||||||
transfer_vertex_groups(
|
|
||||||
source_obj,
|
source_obj,
|
||||||
target_obj,
|
target_obj,
|
||||||
vert_influence_map,
|
vert_influence_map,
|
||||||
@ -110,9 +130,7 @@ def build_kdtree(mesh):
|
|||||||
|
|
||||||
|
|
||||||
def build_vert_influence_map(obj_from, obj_to, kd_tree, expand=2):
|
def build_vert_influence_map(obj_from, obj_to, kd_tree, expand=2):
|
||||||
verts_of_edge = {
|
verts_of_edge = {i: (e.vertices[0], e.vertices[1]) for i, e in enumerate(obj_from.data.edges)}
|
||||||
i: (e.vertices[0], e.vertices[1]) for i, e in enumerate(obj_from.data.edges)
|
|
||||||
}
|
|
||||||
|
|
||||||
edges_of_vert: Dict[int, List[int]] = {}
|
edges_of_vert: Dict[int, List[int]] = {}
|
||||||
for edge_idx, edge in enumerate(obj_from.data.edges):
|
for edge_idx, edge in enumerate(obj_from.data.edges):
|
||||||
@ -166,29 +184,24 @@ def get_source_vert_influences(
|
|||||||
parts_sum = sum(parts)
|
parts_sum = sum(parts)
|
||||||
|
|
||||||
influences = [
|
influences = [
|
||||||
(idx, 1 if dist == 0 else part / parts_sum)
|
(idx, 1 if dist == 0 else part / parts_sum) for part, dist in zip(parts, distances)
|
||||||
for part, dist in zip(parts, distances)
|
|
||||||
]
|
]
|
||||||
|
|
||||||
return influences
|
return influences
|
||||||
|
|
||||||
|
|
||||||
def get_nearest_vert(
|
def get_nearest_vert(coords: Vector, kd_tree: kdtree.KDTree) -> Tuple[Vector, int, float]:
|
||||||
coords: Vector, kd_tree: kdtree.KDTree
|
|
||||||
) -> Tuple[Vector, int, float]:
|
|
||||||
"""Return coordinate, index, and distance of nearest vert to coords in kd_tree."""
|
"""Return coordinate, index, and distance of nearest vert to coords in kd_tree."""
|
||||||
return kd_tree.find(coords)
|
return kd_tree.find(coords)
|
||||||
|
|
||||||
|
|
||||||
def other_vert_of_edge(
|
def other_vert_of_edge(edge: int, vert: int, verts_of_edge: Dict[int, Tuple[int, int]]) -> int:
|
||||||
edge: int, vert: int, verts_of_edge: Dict[int, Tuple[int, int]]
|
|
||||||
) -> int:
|
|
||||||
verts = verts_of_edge[edge]
|
verts = verts_of_edge[edge]
|
||||||
assert vert in verts, f"Vert {vert} not part of edge {edge}."
|
assert vert in verts, f"Vert {vert} not part of edge {edge}."
|
||||||
return verts[0] if vert == verts[1] else verts[1]
|
return verts[0] if vert == verts[1] else verts[1]
|
||||||
|
|
||||||
|
|
||||||
def transfer_vertex_groups(obj_from, obj_to, vert_influence_map, src_vgroups):
|
def transfer_multiple_vertex_groups(obj_from, obj_to, vert_influence_map, src_vgroups):
|
||||||
"""Transfer src_vgroups in obj_from to obj_to using a pre-calculated vert_influence_map."""
|
"""Transfer src_vgroups in obj_from to obj_to using a pre-calculated vert_influence_map."""
|
||||||
|
|
||||||
for src_vg in src_vgroups:
|
for src_vg in src_vgroups:
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
import bpy
|
import bpy
|
||||||
from ..naming import merge_get_basename
|
from ..naming import merge_get_basename
|
||||||
from ..task_layer import get_transfer_data_owner
|
from ..task_layer import get_transfer_data_owner
|
||||||
|
import contextlib
|
||||||
|
|
||||||
def check_transfer_data_entry(
|
def check_transfer_data_entry(
|
||||||
transfer_data: bpy.types.CollectionProperty, key: str, td_type_key: str
|
transfer_data: bpy.types.CollectionProperty, key: str, td_type_key: str
|
||||||
@ -30,6 +30,8 @@ def transfer_data_add_entry(
|
|||||||
td_type_key: str,
|
td_type_key: str,
|
||||||
task_layer_name: str,
|
task_layer_name: str,
|
||||||
surrender: bool,
|
surrender: bool,
|
||||||
|
target_obj: bpy.types.Object = None,
|
||||||
|
obj: bpy.types.Object = None,
|
||||||
):
|
):
|
||||||
"""Add entry to Transferable Data ownership
|
"""Add entry to Transferable Data ownership
|
||||||
|
|
||||||
@ -44,6 +46,10 @@ def transfer_data_add_entry(
|
|||||||
transfer_data_item.owner = task_layer_name
|
transfer_data_item.owner = task_layer_name
|
||||||
transfer_data_item.type = td_type_key
|
transfer_data_item.type = td_type_key
|
||||||
transfer_data_item.surrender = surrender
|
transfer_data_item.surrender = surrender
|
||||||
|
if target_obj:
|
||||||
|
transfer_data_item.target_obj = target_obj
|
||||||
|
if obj:
|
||||||
|
transfer_data_item.obj = obj
|
||||||
return transfer_data_item
|
return transfer_data_item
|
||||||
|
|
||||||
|
|
||||||
@ -125,3 +131,36 @@ def transfer_data_item_init(
|
|||||||
obj_name=obj.name,
|
obj_name=obj.name,
|
||||||
surrender=auto_surrender,
|
surrender=auto_surrender,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@contextlib.contextmanager
|
||||||
|
def isolate_collection(iso_col: bpy.types.Collection):
|
||||||
|
col_exclude = {}
|
||||||
|
view_layer_col = bpy.context.view_layer.layer_collection
|
||||||
|
view_layer_col.collection.children.link(iso_col)
|
||||||
|
for col in view_layer_col.children:
|
||||||
|
col_exclude[col.name] = col.exclude
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Exclude all collections that are not iso collection
|
||||||
|
for col in view_layer_col.children:
|
||||||
|
col.exclude = col.name != iso_col.name
|
||||||
|
yield
|
||||||
|
|
||||||
|
finally:
|
||||||
|
for col in view_layer_col.children:
|
||||||
|
col.exclude = col_exclude[col.name]
|
||||||
|
view_layer_col.collection.children.unlink(iso_col)
|
||||||
|
|
||||||
|
|
||||||
|
@contextlib.contextmanager
|
||||||
|
def link_objs_to_collection(objs: set, col: bpy.types.Collection):
|
||||||
|
...
|
||||||
|
try:
|
||||||
|
for obj in objs:
|
||||||
|
col.objects.link(obj)
|
||||||
|
yield
|
||||||
|
|
||||||
|
finally:
|
||||||
|
for obj in objs:
|
||||||
|
col.objects.unlink(obj)
|
||||||
|
@ -5,7 +5,7 @@ from pathlib import Path
|
|||||||
from typing import Set
|
from typing import Set
|
||||||
import os
|
import os
|
||||||
|
|
||||||
from . import constants, config, opscore
|
from . import constants, config, opscore, logging
|
||||||
from .asset_catalog import get_asset_catalog_items, get_asset_id
|
from .asset_catalog import get_asset_catalog_items, get_asset_id
|
||||||
from .config import verify_task_layer_json_data
|
from .config import verify_task_layer_json_data
|
||||||
from .hooks import Hooks, get_production_hook_dir, get_asset_hook_dir
|
from .hooks import Hooks, get_production_hook_dir, get_asset_hook_dir
|
||||||
@ -309,6 +309,8 @@ class ASSETPIPE_OT_sync_pull(bpy.types.Operator):
|
|||||||
opscore.sync_draw(self, context)
|
opscore.sync_draw(self, context)
|
||||||
|
|
||||||
def execute(self, context: bpy.types.Context):
|
def execute(self, context: bpy.types.Context):
|
||||||
|
profiler = logging.get_profiler()
|
||||||
|
profiler.reset()
|
||||||
asset_col = context.scene.asset_pipeline.asset_collection
|
asset_col = context.scene.asset_pipeline.asset_collection
|
||||||
if self.save:
|
if self.save:
|
||||||
save_images()
|
save_images()
|
||||||
@ -324,6 +326,7 @@ class ASSETPIPE_OT_sync_pull(bpy.types.Operator):
|
|||||||
|
|
||||||
hooks_instance.execute_hooks(merge_mode="pull", merge_status='post', asset_col=asset_col)
|
hooks_instance.execute_hooks(merge_mode="pull", merge_status='post', asset_col=asset_col)
|
||||||
self.report({'INFO'}, "Asset Pull Complete")
|
self.report({'INFO'}, "Asset Pull Complete")
|
||||||
|
profiler.log_all()
|
||||||
return {'FINISHED'}
|
return {'FINISHED'}
|
||||||
|
|
||||||
|
|
||||||
@ -370,12 +373,15 @@ class ASSETPIPE_OT_sync_push(bpy.types.Operator):
|
|||||||
opscore.sync_draw(self, context)
|
opscore.sync_draw(self, context)
|
||||||
|
|
||||||
def execute(self, context: bpy.types.Context):
|
def execute(self, context: bpy.types.Context):
|
||||||
|
profiler = logging.get_profiler()
|
||||||
|
profiler.reset()
|
||||||
asset_col = context.scene.asset_pipeline.asset_collection
|
asset_col = context.scene.asset_pipeline.asset_collection
|
||||||
hooks_instance = Hooks()
|
hooks_instance = Hooks()
|
||||||
hooks_instance.load_hooks(context)
|
hooks_instance.load_hooks(context)
|
||||||
save_images()
|
save_images()
|
||||||
bpy.ops.wm.save_mainfile()
|
bpy.ops.wm.save_mainfile()
|
||||||
|
|
||||||
|
# Seperate if statement so hook can execute before updating ownership/prep sync
|
||||||
if self.pull:
|
if self.pull:
|
||||||
hooks_instance.execute_hooks(merge_mode="pull", merge_status='pre', asset_col=asset_col)
|
hooks_instance.execute_hooks(merge_mode="pull", merge_status='pre', asset_col=asset_col)
|
||||||
# Find current task Layer
|
# Find current task Layer
|
||||||
@ -387,6 +393,8 @@ class ASSETPIPE_OT_sync_push(bpy.types.Operator):
|
|||||||
hooks_instance.execute_hooks(
|
hooks_instance.execute_hooks(
|
||||||
merge_mode="pull", merge_status='post', asset_col=asset_col
|
merge_mode="pull", merge_status='post', asset_col=asset_col
|
||||||
)
|
)
|
||||||
|
|
||||||
|
profiler.set_push()
|
||||||
hooks_instance.execute_hooks(merge_mode="push", merge_status='pre', asset_col=asset_col)
|
hooks_instance.execute_hooks(merge_mode="push", merge_status='pre', asset_col=asset_col)
|
||||||
bpy.ops.wm.save_mainfile(filepath=self._current_file.__str__())
|
bpy.ops.wm.save_mainfile(filepath=self._current_file.__str__())
|
||||||
|
|
||||||
@ -395,6 +403,7 @@ class ASSETPIPE_OT_sync_push(bpy.types.Operator):
|
|||||||
self.report({'INFO'}, "Asset Sync Complete")
|
self.report({'INFO'}, "Asset Sync Complete")
|
||||||
else:
|
else:
|
||||||
self.report({'INFO'}, "Asset Force Push Complete")
|
self.report({'INFO'}, "Asset Force Push Complete")
|
||||||
|
profiler.log_all()
|
||||||
return {'FINISHED'}
|
return {'FINISHED'}
|
||||||
|
|
||||||
|
|
||||||
|
@ -1,4 +1,5 @@
|
|||||||
import bpy
|
import bpy
|
||||||
|
import time
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from .merge.publish import (
|
from .merge.publish import (
|
||||||
find_sync_target,
|
find_sync_target,
|
||||||
@ -20,6 +21,7 @@ from .merge.task_layer import draw_task_layer_selection
|
|||||||
from .asset_catalog import get_asset_id
|
from .asset_catalog import get_asset_id
|
||||||
from . import prefs
|
from . import prefs
|
||||||
|
|
||||||
|
|
||||||
def sync_poll(cls, context):
|
def sync_poll(cls, context):
|
||||||
if any([img.is_dirty for img in bpy.data.images]):
|
if any([img.is_dirty for img in bpy.data.images]):
|
||||||
cls.poll_message_set("Please save unsaved Images")
|
cls.poll_message_set("Please save unsaved Images")
|
||||||
@ -134,6 +136,8 @@ def update_temp_file_paths(self, context, temp_file_path):
|
|||||||
|
|
||||||
|
|
||||||
def sync_execute_pull(self, context):
|
def sync_execute_pull(self, context):
|
||||||
|
start_time = time.time()
|
||||||
|
profiler = logging.get_profiler()
|
||||||
logger = logging.get_logger()
|
logger = logging.get_logger()
|
||||||
logger.info("Pulling Asset")
|
logger.info("Pulling Asset")
|
||||||
temp_file_path = create_temp_file_backup(self, context)
|
temp_file_path = create_temp_file_backup(self, context)
|
||||||
@ -160,9 +164,12 @@ def sync_execute_pull(self, context):
|
|||||||
context.scene.asset_pipeline.sync_error = True
|
context.scene.asset_pipeline.sync_error = True
|
||||||
self.report({'ERROR'}, error_msg)
|
self.report({'ERROR'}, error_msg)
|
||||||
return {'CANCELLED'}
|
return {'CANCELLED'}
|
||||||
|
profiler.add(time.time() - start_time, "TOTAL")
|
||||||
|
|
||||||
|
|
||||||
def sync_execute_push(self, context):
|
def sync_execute_push(self, context):
|
||||||
|
start_time = time.time()
|
||||||
|
profiler = logging.get_profiler()
|
||||||
logger = logging.get_logger()
|
logger = logging.get_logger()
|
||||||
logger.info("Pushing Asset")
|
logger.info("Pushing Asset")
|
||||||
_catalog_id = None
|
_catalog_id = None
|
||||||
@ -203,3 +210,4 @@ def sync_execute_push(self, context):
|
|||||||
|
|
||||||
bpy.ops.wm.save_as_mainfile(filepath=file_path)
|
bpy.ops.wm.save_as_mainfile(filepath=file_path)
|
||||||
bpy.ops.wm.open_mainfile(filepath=self._current_file.__str__())
|
bpy.ops.wm.open_mainfile(filepath=self._current_file.__str__())
|
||||||
|
profiler.add(time.time() - start_time, "TOTAL")
|
||||||
|
@ -57,7 +57,18 @@ class AssetTransferDataTemp(bpy.types.PropertyGroup):
|
|||||||
items=constants.TRANSFER_DATA_TYPES_ENUM_ITEMS,
|
items=constants.TRANSFER_DATA_TYPES_ENUM_ITEMS,
|
||||||
)
|
)
|
||||||
surrender: bpy.props.BoolProperty(name="Surrender Ownership", default=False)
|
surrender: bpy.props.BoolProperty(name="Surrender Ownership", default=False)
|
||||||
obj_name: bpy.props.StringProperty(name="Object Name")
|
obj_name: bpy.props.StringProperty(name="Object Name", default="")
|
||||||
|
|
||||||
|
def check_transfer_data_entry(self) -> set:
|
||||||
|
"""
|
||||||
|
Verifies if Transferable Data entry exists
|
||||||
|
"""
|
||||||
|
existing_items = [
|
||||||
|
transfer_data_item.name
|
||||||
|
for transfer_data_item in self.target_obj.transfer_data_ownership
|
||||||
|
if transfer_data_item.type == self.type
|
||||||
|
]
|
||||||
|
return set([self.name]).intersection(set(existing_items))
|
||||||
|
|
||||||
|
|
||||||
class TaskLayerSettings(bpy.types.PropertyGroup):
|
class TaskLayerSettings(bpy.types.PropertyGroup):
|
||||||
@ -113,12 +124,12 @@ class AssetPipeline(bpy.types.PropertyGroup):
|
|||||||
|
|
||||||
temp_transfer_data: bpy.props.CollectionProperty(type=AssetTransferDataTemp)
|
temp_transfer_data: bpy.props.CollectionProperty(type=AssetTransferDataTemp)
|
||||||
|
|
||||||
def add_temp_transfer_data(self, name, owner, type, obj_name, surrender):
|
def add_temp_transfer_data(self, name, owner, type_key, obj_name, surrender):
|
||||||
new_transfer_data = self.temp_transfer_data
|
new_transfer_data = self.temp_transfer_data
|
||||||
transfer_data_item = new_transfer_data.add()
|
transfer_data_item = new_transfer_data.add()
|
||||||
transfer_data_item.name = name
|
transfer_data_item.name = name
|
||||||
transfer_data_item.owner = owner
|
transfer_data_item.owner = owner
|
||||||
transfer_data_item.type = type
|
transfer_data_item.type = type_key
|
||||||
transfer_data_item.obj_name = obj_name
|
transfer_data_item.obj_name = obj_name
|
||||||
transfer_data_item.surrender = surrender
|
transfer_data_item.surrender = surrender
|
||||||
|
|
||||||
@ -201,6 +212,7 @@ class AssetPipeline(bpy.types.PropertyGroup):
|
|||||||
description="Select Asset Library Catalog for the current Asset, this value will be updated each time you Push to an 'Active' Publish",
|
description="Select Asset Library Catalog for the current Asset, this value will be updated each time you Push to an 'Active' Publish",
|
||||||
) # type: ignore
|
) # type: ignore
|
||||||
|
|
||||||
|
|
||||||
@bpy.app.handlers.persistent
|
@bpy.app.handlers.persistent
|
||||||
def set_asset_collection_name_post_file_load(_):
|
def set_asset_collection_name_post_file_load(_):
|
||||||
# Version the PointerProperty to the StringProperty, and the left-over pointer.
|
# Version the PointerProperty to the StringProperty, and the left-over pointer.
|
||||||
|
Loading…
Reference in New Issue
Block a user