split packing utils
This commit is contained in:
643
packer/blendfile_path_walker.py
Normal file
643
packer/blendfile_path_walker.py
Normal file
@@ -0,0 +1,643 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
# ***** BEGIN GPL LICENSE BLOCK *****
|
||||||
|
#
|
||||||
|
# This program is free software; you can redistribute it and/or
|
||||||
|
# modify it under the terms of the GNU General Public License
|
||||||
|
# as published by the Free Software Foundation; either version 2
|
||||||
|
# of the License, or (at your option) any later version.
|
||||||
|
#
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU General Public License for more details.
|
||||||
|
#
|
||||||
|
# You should have received a copy of the GNU General Public License
|
||||||
|
# along with this program; if not, write to the Free Software Foundation,
|
||||||
|
# Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
||||||
|
#
|
||||||
|
# ***** END GPL LICENCE BLOCK *****
|
||||||
|
|
||||||
|
VERBOSE = True
|
||||||
|
TIMEIT = True
|
||||||
|
|
||||||
|
class C_defs:
|
||||||
|
__slots__ = ()
|
||||||
|
def __new__(cls, *args, **kwargs):
|
||||||
|
raise RuntimeError("%s should not be instantiated" % cls)
|
||||||
|
|
||||||
|
# DNA_sequence_types.h (Sequence.type)
|
||||||
|
SEQ_TYPE_IMAGE = 0
|
||||||
|
SEQ_TYPE_META = 1
|
||||||
|
SEQ_TYPE_SCENE = 2
|
||||||
|
SEQ_TYPE_MOVIE = 3
|
||||||
|
SEQ_TYPE_SOUND_RAM = 4
|
||||||
|
SEQ_TYPE_SOUND_HD = 5
|
||||||
|
SEQ_TYPE_MOVIECLIP = 6
|
||||||
|
SEQ_TYPE_MASK = 7
|
||||||
|
SEQ_TYPE_EFFECT = 8
|
||||||
|
|
||||||
|
|
||||||
|
if VERBOSE:
|
||||||
|
_A = open("/tmp/a.log", 'w')
|
||||||
|
class log_deps:
|
||||||
|
@staticmethod
|
||||||
|
def info(msg):
|
||||||
|
_A.write(msg)
|
||||||
|
_A.write("\n")
|
||||||
|
|
||||||
|
def set_as_str(s):
|
||||||
|
if s is None:
|
||||||
|
return "None"
|
||||||
|
else:
|
||||||
|
return (", ".join(sorted(i.decode('ascii') for i in sorted(s))))
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
class FPElem:
|
||||||
|
"""
|
||||||
|
Tiny filepath class to hide blendfile.
|
||||||
|
"""
|
||||||
|
|
||||||
|
__slots__ = (
|
||||||
|
"basedir",
|
||||||
|
|
||||||
|
# library link level
|
||||||
|
"level",
|
||||||
|
|
||||||
|
"userdata",
|
||||||
|
)
|
||||||
|
|
||||||
|
def __init__(self, basedir, level,
|
||||||
|
# subclasses get/set functions should use
|
||||||
|
userdata):
|
||||||
|
self.basedir = basedir
|
||||||
|
self.level = level
|
||||||
|
|
||||||
|
# subclass must call
|
||||||
|
self.userdata = userdata
|
||||||
|
|
||||||
|
# --------
|
||||||
|
# filepath
|
||||||
|
|
||||||
|
@property
|
||||||
|
def filepath(self):
|
||||||
|
return self._get_cb()
|
||||||
|
|
||||||
|
@filepath.setter
|
||||||
|
def filepath(self, filepath):
|
||||||
|
self._set_cb(filepath)
|
||||||
|
|
||||||
|
|
||||||
|
class FPElem_block_path(FPElem):
|
||||||
|
"""
|
||||||
|
Simple block-path:
|
||||||
|
userdata = (block, path)
|
||||||
|
"""
|
||||||
|
__slots__ = ()
|
||||||
|
|
||||||
|
def _get_cb(self):
|
||||||
|
block, path = self.userdata
|
||||||
|
return block[path]
|
||||||
|
|
||||||
|
def _set_cb(self, filepath):
|
||||||
|
block, path = self.userdata
|
||||||
|
block[path] = filepath
|
||||||
|
|
||||||
|
|
||||||
|
class FPElem_sequence_single(FPElem):
|
||||||
|
"""
|
||||||
|
Movie sequence
|
||||||
|
userdata = (block, path)
|
||||||
|
"""
|
||||||
|
__slots__ = ()
|
||||||
|
|
||||||
|
def _get_cb(self):
|
||||||
|
block, path, sub_block, sub_path = self.userdata
|
||||||
|
return block[path] + sub_block[sub_path]
|
||||||
|
|
||||||
|
def _set_cb(self, filepath):
|
||||||
|
block, path, sub_block, sub_path = self.userdata
|
||||||
|
|
||||||
|
# TODO, os.sep
|
||||||
|
a, b = filepath.rsplit(b'/', 1)
|
||||||
|
|
||||||
|
block[path] = a + b'/'
|
||||||
|
sub_block[sub_path] = b
|
||||||
|
|
||||||
|
|
||||||
|
class FilePath:
|
||||||
|
__slots__ = ()
|
||||||
|
|
||||||
|
def __new__(cls, *args, **kwargs):
|
||||||
|
raise RuntimeError("%s should not be instantiated" % cls)
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------
|
||||||
|
# Main function to visit paths
|
||||||
|
@staticmethod
|
||||||
|
def visit_from_blend(
|
||||||
|
filepath,
|
||||||
|
|
||||||
|
# never modify the blend
|
||||||
|
readonly=True,
|
||||||
|
# callback that creates a temp file and returns its path.
|
||||||
|
temp_remap_cb=None,
|
||||||
|
|
||||||
|
# recursive options
|
||||||
|
recursive=False,
|
||||||
|
# list of ID block names we want to load, or None to load all
|
||||||
|
block_codes=None,
|
||||||
|
# root when we're loading libs indirectly
|
||||||
|
rootdir=None,
|
||||||
|
level=0,
|
||||||
|
# dict of id's used so we don't follow these links again
|
||||||
|
# prevents cyclic references too!
|
||||||
|
# {lib_path: set([block id's ...])}
|
||||||
|
lib_visit=None,
|
||||||
|
):
|
||||||
|
# print(level, block_codes)
|
||||||
|
import os
|
||||||
|
|
||||||
|
if VERBOSE:
|
||||||
|
indent_str = " " * level
|
||||||
|
# print(indent_str + "Opening:", filepath)
|
||||||
|
# print(indent_str + "... blocks:", block_codes)
|
||||||
|
|
||||||
|
|
||||||
|
log_deps.info("~")
|
||||||
|
log_deps.info("%s%s" % (indent_str, filepath.decode('utf-8')))
|
||||||
|
log_deps.info("%s%s" % (indent_str, set_as_str(block_codes)))
|
||||||
|
|
||||||
|
|
||||||
|
basedir = os.path.dirname(os.path.abspath(filepath))
|
||||||
|
if rootdir is None:
|
||||||
|
rootdir = basedir
|
||||||
|
|
||||||
|
if recursive and (level > 0) and (block_codes is not None):
|
||||||
|
# prevent from expanding the
|
||||||
|
# same datablock more then once
|
||||||
|
expand_codes = set()
|
||||||
|
# {lib_id: {block_ids... }}
|
||||||
|
expand_codes_idlib = {}
|
||||||
|
|
||||||
|
# libraries used by this blend
|
||||||
|
block_codes_idlib = set()
|
||||||
|
|
||||||
|
# only for this block
|
||||||
|
def _expand_codes_add_test(block, code):
|
||||||
|
# return True, if the ID should be searched further
|
||||||
|
#
|
||||||
|
# we could investigate a better way...
|
||||||
|
# Not to be accessing ID blocks at this point. but its harmless
|
||||||
|
if code == b'ID':
|
||||||
|
assert(code == block.code)
|
||||||
|
if recursive:
|
||||||
|
expand_codes_idlib.setdefault(block[b'lib'], set()).add(block[b'name'])
|
||||||
|
return False
|
||||||
|
else:
|
||||||
|
len_prev = len(expand_codes)
|
||||||
|
expand_codes.add(block[b'id.name'])
|
||||||
|
return (len_prev != len(expand_codes))
|
||||||
|
|
||||||
|
def block_expand(block, code):
|
||||||
|
assert(block.code == code)
|
||||||
|
if _expand_codes_add_test(block, code):
|
||||||
|
yield block
|
||||||
|
|
||||||
|
assert(block.code == code)
|
||||||
|
fn = ExpandID.expand_funcs.get(code)
|
||||||
|
if fn is not None:
|
||||||
|
for sub_block in fn(block):
|
||||||
|
if sub_block is not None:
|
||||||
|
yield from block_expand(sub_block, sub_block.code)
|
||||||
|
else:
|
||||||
|
yield block
|
||||||
|
else:
|
||||||
|
expand_codes = None
|
||||||
|
|
||||||
|
# set below
|
||||||
|
expand_codes_idlib = None
|
||||||
|
|
||||||
|
# never set
|
||||||
|
block_codes_idlib = None
|
||||||
|
|
||||||
|
def block_expand(block, code):
|
||||||
|
assert(block.code == code)
|
||||||
|
yield block
|
||||||
|
|
||||||
|
# ------
|
||||||
|
# Define
|
||||||
|
#
|
||||||
|
# - iter_blocks_id(code)
|
||||||
|
# - iter_blocks_idlib()
|
||||||
|
if block_codes is None:
|
||||||
|
def iter_blocks_id(code):
|
||||||
|
return blend.find_blocks_from_code(code)
|
||||||
|
|
||||||
|
def iter_blocks_idlib():
|
||||||
|
return blend.find_blocks_from_code(b'LI')
|
||||||
|
else:
|
||||||
|
def iter_blocks_id(code):
|
||||||
|
for block in blend.find_blocks_from_code(code):
|
||||||
|
if block[b'id.name'] in block_codes:
|
||||||
|
yield from block_expand(block, code)
|
||||||
|
|
||||||
|
if block_codes_idlib is not None:
|
||||||
|
def iter_blocks_idlib():
|
||||||
|
for block in blend.find_blocks_from_code(b'LI'):
|
||||||
|
if block[b'name'] in block_codes_idlib:
|
||||||
|
yield from block_expand(block, b'LI')
|
||||||
|
else:
|
||||||
|
def iter_blocks_idlib():
|
||||||
|
return blend.find_blocks_from_code(b'LI')
|
||||||
|
|
||||||
|
if temp_remap_cb is not None:
|
||||||
|
filepath_tmp = temp_remap_cb(filepath, level)
|
||||||
|
else:
|
||||||
|
filepath_tmp = filepath
|
||||||
|
|
||||||
|
# store info to pass along with each iteration
|
||||||
|
extra_info = rootdir, os.path.basename(filepath)
|
||||||
|
|
||||||
|
import blendfile
|
||||||
|
blend = blendfile.open_blend(filepath_tmp, "rb" if readonly else "r+b")
|
||||||
|
|
||||||
|
for code in blend.code_index.keys():
|
||||||
|
# handle library blocks as special case
|
||||||
|
if ((len(code) != 2) or
|
||||||
|
(code in {
|
||||||
|
# libraries handled below
|
||||||
|
b'LI',
|
||||||
|
b'ID',
|
||||||
|
# unneeded
|
||||||
|
b'WM',
|
||||||
|
b'SN', # bScreen
|
||||||
|
})):
|
||||||
|
|
||||||
|
continue
|
||||||
|
|
||||||
|
# if VERBOSE:
|
||||||
|
# print(" Scanning", code)
|
||||||
|
|
||||||
|
for block in iter_blocks_id(code):
|
||||||
|
yield from FilePath.from_block(block, basedir, extra_info, level)
|
||||||
|
|
||||||
|
# print("A:", expand_codes)
|
||||||
|
# print("B:", block_codes)
|
||||||
|
if VERBOSE:
|
||||||
|
log_deps.info("%s%s" % (indent_str, set_as_str(expand_codes)))
|
||||||
|
|
||||||
|
if recursive:
|
||||||
|
|
||||||
|
if expand_codes_idlib is None:
|
||||||
|
expand_codes_idlib = {}
|
||||||
|
for block in blend.find_blocks_from_code(b'ID'):
|
||||||
|
expand_codes_idlib.setdefault(block[b'lib'], set()).add(block[b'name'])
|
||||||
|
|
||||||
|
# look into libraries
|
||||||
|
lib_all = []
|
||||||
|
|
||||||
|
for lib_id, lib_block_codes in sorted(expand_codes_idlib.items()):
|
||||||
|
lib = blend.find_block_from_offset(lib_id)
|
||||||
|
lib_path = lib[b'name']
|
||||||
|
|
||||||
|
# get all data needed to read the blend files here (it will be freed!)
|
||||||
|
# lib is an address at the moment, we only use as a way to group
|
||||||
|
|
||||||
|
lib_all.append((lib_path, lib_block_codes))
|
||||||
|
# import IPython; IPython.embed()
|
||||||
|
|
||||||
|
# ensure we expand indirect linked libs
|
||||||
|
if block_codes_idlib is not None:
|
||||||
|
block_codes_idlib.add(lib_path)
|
||||||
|
|
||||||
|
# do this after, incase we mangle names above
|
||||||
|
for block in iter_blocks_idlib():
|
||||||
|
yield from FilePath.from_block(block, basedir, extra_info, level)
|
||||||
|
|
||||||
|
blend.close()
|
||||||
|
|
||||||
|
# ----------------
|
||||||
|
# Handle Recursive
|
||||||
|
if recursive:
|
||||||
|
# now we've closed the file, loop on other files
|
||||||
|
|
||||||
|
# note, sorting - isn't needed, it just gives predictable load-order.
|
||||||
|
for lib_path, lib_block_codes in lib_all:
|
||||||
|
lib_path_abs = os.path.normpath(utils.compatpath(utils.abspath(lib_path, basedir)))
|
||||||
|
|
||||||
|
# if we visited this before,
|
||||||
|
# check we don't follow the same links more than once
|
||||||
|
lib_block_codes_existing = lib_visit.setdefault(lib_path_abs, set())
|
||||||
|
lib_block_codes -= lib_block_codes_existing
|
||||||
|
# don't touch them again
|
||||||
|
lib_block_codes_existing.update(lib_block_codes)
|
||||||
|
|
||||||
|
# print("looking for", lib_block_codes)
|
||||||
|
|
||||||
|
# import IPython; IPython.embed()
|
||||||
|
if VERBOSE:
|
||||||
|
print((indent_str + " "), "Library: ", filepath, " -> ", lib_path_abs, sep="")
|
||||||
|
# print((indent_str + " "), lib_block_codes)
|
||||||
|
yield from FilePath.visit_from_blend(
|
||||||
|
lib_path_abs,
|
||||||
|
readonly=readonly,
|
||||||
|
temp_remap_cb=temp_remap_cb,
|
||||||
|
recursive=True,
|
||||||
|
block_codes=lib_block_codes,
|
||||||
|
rootdir=rootdir,
|
||||||
|
level=level + 1,
|
||||||
|
lib_visit=lib_visit,
|
||||||
|
)
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------
|
||||||
|
# Direct filepaths from Blocks
|
||||||
|
#
|
||||||
|
# (no expanding or following references)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def from_block(block, basedir, extra_info, level):
|
||||||
|
assert(block.code != b'DATA')
|
||||||
|
fn = FilePath._from_block_dict.get(block.code)
|
||||||
|
if fn is not None:
|
||||||
|
yield from fn(block, basedir, extra_info, level)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_MC(block, basedir, extra_info, level):
|
||||||
|
# TODO, image sequence
|
||||||
|
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_IM(block, basedir, extra_info, level):
|
||||||
|
# (IMA_SRC_FILE, IMA_SRC_SEQUENCE, IMA_SRC_MOVIE)
|
||||||
|
if block[b'source'] not in {1, 2, 3}:
|
||||||
|
return
|
||||||
|
if block[b'packedfile']:
|
||||||
|
return
|
||||||
|
|
||||||
|
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_VF(block, basedir, extra_info, level):
|
||||||
|
if block[b'packedfile']:
|
||||||
|
return
|
||||||
|
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_SO(block, basedir, extra_info, level):
|
||||||
|
if block[b'packedfile']:
|
||||||
|
return
|
||||||
|
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_ME(block, basedir, extra_info, level):
|
||||||
|
block_external = block.get_pointer(b'ldata.external')
|
||||||
|
if block_external is not None:
|
||||||
|
yield FPElem_block_path(basedir, level, (block_external, b'filename')), extra_info
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_SC(block, basedir, extra_info, level):
|
||||||
|
block_ed = block.get_pointer(b'ed')
|
||||||
|
if block_ed is not None:
|
||||||
|
sdna_index_Sequence = block.file.sdna_index_from_id[b'Sequence']
|
||||||
|
|
||||||
|
def seqbase(someseq):
|
||||||
|
for item in someseq:
|
||||||
|
item_type = item.get(b'type', sdna_index_refine=sdna_index_Sequence)
|
||||||
|
|
||||||
|
if item_type >= C_defs.SEQ_TYPE_EFFECT:
|
||||||
|
pass
|
||||||
|
elif item_type == C_defs.SEQ_TYPE_META:
|
||||||
|
yield from seqbase(bf_utils.iter_ListBase(item.get_pointer(b'seqbase.first', sdna_index_refine=sdna_index_Sequence)))
|
||||||
|
else:
|
||||||
|
item_strip = item.get_pointer(b'strip', sdna_index_refine=sdna_index_Sequence)
|
||||||
|
if item_strip is None: # unlikely!
|
||||||
|
continue
|
||||||
|
item_stripdata = item_strip.get_pointer(b'stripdata')
|
||||||
|
|
||||||
|
if item_type == C_defs.SEQ_TYPE_IMAGE:
|
||||||
|
# TODO, multiple images
|
||||||
|
yield FPElem_sequence_single(basedir, level, (item_strip, b'dir', item_stripdata, b'name')), extra_info
|
||||||
|
elif item_type == C_defs.SEQ_TYPE_MOVIE:
|
||||||
|
yield FPElem_sequence_single(basedir, level, (item_strip, b'dir', item_stripdata, b'name')), extra_info
|
||||||
|
elif item_type == C_defs.SEQ_TYPE_SOUND_RAM:
|
||||||
|
pass
|
||||||
|
elif item_type == C_defs.SEQ_TYPE_SOUND_HD:
|
||||||
|
pass
|
||||||
|
|
||||||
|
yield from seqbase(bf_utils.iter_ListBase(block_ed.get_pointer(b'seqbase.first')))
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _from_block_LI(block, basedir, extra_info, level):
|
||||||
|
if block.get(b'packedfile', None):
|
||||||
|
return
|
||||||
|
|
||||||
|
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
||||||
|
|
||||||
|
# _from_block_IM --> {b'IM': _from_block_IM, ...}
|
||||||
|
_from_block_dict = {
|
||||||
|
k.rpartition("_")[2].encode('ascii'): s_fn.__func__ for k, s_fn in locals().items()
|
||||||
|
if isinstance(s_fn, staticmethod)
|
||||||
|
if k.startswith("_from_block_")
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class bf_utils:
|
||||||
|
@staticmethod
|
||||||
|
def iter_ListBase(block):
|
||||||
|
while block:
|
||||||
|
yield block
|
||||||
|
block = block.file.find_block_from_offset(block[b'next'])
|
||||||
|
|
||||||
|
def iter_array(block, length=-1):
|
||||||
|
assert(block.code == b'DATA')
|
||||||
|
import blendfile
|
||||||
|
import os
|
||||||
|
handle = block.file.handle
|
||||||
|
header = block.file.header
|
||||||
|
|
||||||
|
for i in range(length):
|
||||||
|
block.file.handle.seek(block.file_offset + (header.pointer_size * i), os.SEEK_SET)
|
||||||
|
offset = blendfile.DNA_IO.read_pointer(handle, header)
|
||||||
|
sub_block = block.file.find_block_from_offset(offset)
|
||||||
|
yield sub_block
|
||||||
|
|
||||||
|
|
||||||
|
# -----------------------------------------------------------------------------
|
||||||
|
# ID Expand
|
||||||
|
|
||||||
|
class ExpandID:
|
||||||
|
# fake module
|
||||||
|
#
|
||||||
|
# TODO:
|
||||||
|
#
|
||||||
|
# Array lookups here are _WAY_ too complicated,
|
||||||
|
# we need some nicer way to represent pointer indirection (easy like in C!)
|
||||||
|
# but for now, use what we have.
|
||||||
|
#
|
||||||
|
__slots__ = ()
|
||||||
|
|
||||||
|
def __new__(cls, *args, **kwargs):
|
||||||
|
raise RuntimeError("%s should not be instantiated" % cls)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _expand_generic_material(block):
|
||||||
|
array_len = block.get(b'totcol')
|
||||||
|
if array_len != 0:
|
||||||
|
array = block.get_pointer(b'mat')
|
||||||
|
for sub_block in bf_utils.iter_array(array, array_len):
|
||||||
|
yield sub_block
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _expand_generic_mtex(block):
|
||||||
|
field = block.dna_type.field_from_name[b'mtex']
|
||||||
|
array_len = field.dna_size // block.file.header.pointer_size
|
||||||
|
|
||||||
|
for i in range(array_len):
|
||||||
|
path = ('mtex[%d]' % i).encode('ascii')
|
||||||
|
item = block.get_pointer(path)
|
||||||
|
if item:
|
||||||
|
yield item.get_pointer(b'tex')
|
||||||
|
yield item.get_pointer(b'object')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _expand_generic_nodetree(block):
|
||||||
|
assert(block.dna_type.dna_type_id == b'bNodeTree')
|
||||||
|
|
||||||
|
sdna_index_bNode = block.file.sdna_index_from_id[b'bNode']
|
||||||
|
for item in bf_utils.iter_ListBase(block.get_pointer(b'nodes.first')):
|
||||||
|
item_type = item.get(b'type', sdna_index_refine=sdna_index_bNode)
|
||||||
|
|
||||||
|
if item_type != 221: # CMP_NODE_R_LAYERS
|
||||||
|
yield item.get_pointer(b'id', sdna_index_refine=sdna_index_bNode)
|
||||||
|
|
||||||
|
def _expand_generic_nodetree_id(block):
|
||||||
|
block_ntree = block.get_pointer(b'nodetree')
|
||||||
|
if block_ntree is not None:
|
||||||
|
yield from ExpandID._expand_generic_nodetree(block_ntree)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def _expand_generic_animdata(block):
|
||||||
|
block_adt = block.get_pointer(b'adt')
|
||||||
|
if block_adt:
|
||||||
|
yield block_adt.get_pointer(b'action')
|
||||||
|
# TODO, NLA
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_OB(block): # 'Object'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield block.get_pointer(b'data')
|
||||||
|
yield block.get_pointer(b'dup_group')
|
||||||
|
|
||||||
|
yield block.get_pointer(b'proxy')
|
||||||
|
yield block.get_pointer(b'proxy_group')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_ME(block): # 'Mesh'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_material(block)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_CU(block): # 'Curve'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_material(block)
|
||||||
|
|
||||||
|
sub_block = block.get_pointer(b'vfont')
|
||||||
|
if sub_block is not None:
|
||||||
|
yield sub_block
|
||||||
|
yield block.get_pointer(b'vfontb')
|
||||||
|
yield block.get_pointer(b'vfonti')
|
||||||
|
yield block.get_pointer(b'vfontbi')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_MB(block): # 'MBall'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_material(block)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_LA(block): # 'Lamp'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree_id(block)
|
||||||
|
yield from ExpandID._expand_generic_mtex(block)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_MA(block): # 'Material'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree_id(block)
|
||||||
|
yield from ExpandID._expand_generic_mtex(block)
|
||||||
|
|
||||||
|
yield block.get_pointer(b'group')
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_TE(block): # 'Tex'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree_id(block)
|
||||||
|
yield block.get_pointer(b'ima')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_WO(block): # 'World'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree_id(block)
|
||||||
|
yield from ExpandID._expand_generic_mtex(block)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_NT(block): # 'bNodeTree'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree(block)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_SC(block): # 'Scene'
|
||||||
|
yield from ExpandID._expand_generic_animdata(block)
|
||||||
|
yield from ExpandID._expand_generic_nodetree_id(block)
|
||||||
|
yield block.get_pointer(b'world')
|
||||||
|
|
||||||
|
sdna_index_Base = block.file.sdna_index_from_id[b'Base']
|
||||||
|
for item in bf_utils.iter_ListBase(block.get_pointer(b'base.first')):
|
||||||
|
yield item.get_pointer(b'object', sdna_index_refine=sdna_index_Base)
|
||||||
|
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def expand_GR(block): # 'Group'
|
||||||
|
sdna_index_GroupObject = block.file.sdna_index_from_id[b'GroupObject']
|
||||||
|
for item in bf_utils.iter_ListBase(block.get_pointer(b'gobject.first')):
|
||||||
|
yield item.get_pointer(b'ob', sdna_index_refine=sdna_index_GroupObject)
|
||||||
|
|
||||||
|
# expand_GR --> {b'GR': expand_GR, ...}
|
||||||
|
expand_funcs = {
|
||||||
|
k.rpartition("_")[2].encode('ascii'): s_fn.__func__ for k, s_fn in locals().items()
|
||||||
|
if isinstance(s_fn, staticmethod)
|
||||||
|
if k.startswith("expand_")
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
# -----------------------------------------------------------------------------
|
||||||
|
# Packing Utility
|
||||||
|
|
||||||
|
|
||||||
|
class utils:
|
||||||
|
# fake module
|
||||||
|
__slots__ = ()
|
||||||
|
|
||||||
|
def __new__(cls, *args, **kwargs):
|
||||||
|
raise RuntimeError("%s should not be instantiated" % cls)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def abspath(path, start, library=None):
|
||||||
|
import os
|
||||||
|
if path.startswith(b'//'):
|
||||||
|
# if library:
|
||||||
|
# start = os.path.dirname(abspath(library.filepath))
|
||||||
|
return os.path.join(start, path[2:])
|
||||||
|
return path
|
||||||
|
|
||||||
|
if __import__("os").sep == '/':
|
||||||
|
@staticmethod
|
||||||
|
def compatpath(path):
|
||||||
|
return path.replace(b'\\', b'/')
|
||||||
|
else:
|
||||||
|
@staticmethod
|
||||||
|
def compatpath(path):
|
||||||
|
# keep '//'
|
||||||
|
return path[:2] + path[2:].replace(b'/', b'\\')
|
||||||
|
|
681
packer/packer.py
681
packer/packer.py
@@ -18,636 +18,17 @@
|
|||||||
#
|
#
|
||||||
# ***** END GPL LICENCE BLOCK *****
|
# ***** END GPL LICENCE BLOCK *****
|
||||||
|
|
||||||
VERBOSE = True
|
import blendfile_path_walker
|
||||||
|
|
||||||
TIMEIT = True
|
TIMEIT = True
|
||||||
|
|
||||||
class C_defs:
|
|
||||||
__slots__ = ()
|
|
||||||
def __new__(cls, *args, **kwargs):
|
|
||||||
raise RuntimeError("%s should not be instantiated" % cls)
|
|
||||||
|
|
||||||
# DNA_sequence_types.h (Sequence.type)
|
def pack(blendfile_src, blendfile_dst, mode='FILE', pathmap=None):
|
||||||
SEQ_TYPE_IMAGE = 0
|
|
||||||
SEQ_TYPE_META = 1
|
|
||||||
SEQ_TYPE_SCENE = 2
|
|
||||||
SEQ_TYPE_MOVIE = 3
|
|
||||||
SEQ_TYPE_SOUND_RAM = 4
|
|
||||||
SEQ_TYPE_SOUND_HD = 5
|
|
||||||
SEQ_TYPE_MOVIECLIP = 6
|
|
||||||
SEQ_TYPE_MASK = 7
|
|
||||||
SEQ_TYPE_EFFECT = 8
|
|
||||||
|
|
||||||
|
|
||||||
if VERBOSE:
|
|
||||||
_A = open("/tmp/a.log", 'w')
|
|
||||||
class log_deps:
|
|
||||||
@staticmethod
|
|
||||||
def info(msg):
|
|
||||||
_A.write(msg)
|
|
||||||
_A.write("\n")
|
|
||||||
|
|
||||||
def set_as_str(s):
|
|
||||||
if s is None:
|
|
||||||
return "None"
|
|
||||||
else:
|
|
||||||
return (", ".join(sorted(i.decode('ascii') for i in sorted(s))))
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class FPElem:
|
|
||||||
"""
|
"""
|
||||||
Tiny filepath class to hide blendfile.
|
:param pathmap: Store path pathmap info as follows.
|
||||||
"""
|
|
||||||
|
|
||||||
__slots__ = (
|
|
||||||
"basedir",
|
|
||||||
|
|
||||||
# library link level
|
|
||||||
"level",
|
|
||||||
|
|
||||||
"userdata",
|
|
||||||
)
|
|
||||||
|
|
||||||
def __init__(self, basedir, level,
|
|
||||||
# subclasses get/set functions should use
|
|
||||||
userdata):
|
|
||||||
self.basedir = basedir
|
|
||||||
self.level = level
|
|
||||||
|
|
||||||
# subclass must call
|
|
||||||
self.userdata = userdata
|
|
||||||
|
|
||||||
# --------
|
|
||||||
# filepath
|
|
||||||
|
|
||||||
@property
|
|
||||||
def filepath(self):
|
|
||||||
return self._get_cb()
|
|
||||||
|
|
||||||
@filepath.setter
|
|
||||||
def filepath(self, filepath):
|
|
||||||
self._set_cb(filepath)
|
|
||||||
|
|
||||||
|
|
||||||
class FPElem_block_path(FPElem):
|
|
||||||
"""
|
|
||||||
Simple block-path:
|
|
||||||
userdata = (block, path)
|
|
||||||
"""
|
|
||||||
__slots__ = ()
|
|
||||||
|
|
||||||
def _get_cb(self):
|
|
||||||
block, path = self.userdata
|
|
||||||
return block[path]
|
|
||||||
|
|
||||||
def _set_cb(self, filepath):
|
|
||||||
block, path = self.userdata
|
|
||||||
block[path] = filepath
|
|
||||||
|
|
||||||
|
|
||||||
class FPElem_sequence_single(FPElem):
|
|
||||||
"""
|
|
||||||
Movie sequence
|
|
||||||
userdata = (block, path)
|
|
||||||
"""
|
|
||||||
__slots__ = ()
|
|
||||||
|
|
||||||
def _get_cb(self):
|
|
||||||
block, path, sub_block, sub_path = self.userdata
|
|
||||||
return block[path] + sub_block[sub_path]
|
|
||||||
|
|
||||||
def _set_cb(self, filepath):
|
|
||||||
block, path, sub_block, sub_path = self.userdata
|
|
||||||
|
|
||||||
# TODO, os.sep
|
|
||||||
a, b = filepath.rsplit(b'/', 1)
|
|
||||||
|
|
||||||
block[path] = a + b'/'
|
|
||||||
sub_block[sub_path] = b
|
|
||||||
|
|
||||||
|
|
||||||
class FilePath:
|
|
||||||
__slots__ = ()
|
|
||||||
|
|
||||||
def __new__(cls, *args, **kwargs):
|
|
||||||
raise RuntimeError("%s should not be instantiated" % cls)
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------
|
|
||||||
# Main function to visit paths
|
|
||||||
@staticmethod
|
|
||||||
def visit_from_blend(
|
|
||||||
filepath,
|
|
||||||
|
|
||||||
# never modify the blend
|
|
||||||
readonly=True,
|
|
||||||
# callback that creates a temp file and returns its path.
|
|
||||||
temp_remap_cb=None,
|
|
||||||
|
|
||||||
# recursive options
|
|
||||||
recursive=False,
|
|
||||||
# list of ID block names we want to load, or None to load all
|
|
||||||
block_codes=None,
|
|
||||||
# root when we're loading libs indirectly
|
|
||||||
rootdir=None,
|
|
||||||
level=0,
|
|
||||||
# dict of id's used so we don't follow these links again
|
|
||||||
# prevents cyclic references too!
|
|
||||||
# {lib_path: set([block id's ...])}
|
|
||||||
lib_visit=None,
|
|
||||||
):
|
|
||||||
# print(level, block_codes)
|
|
||||||
import os
|
|
||||||
|
|
||||||
if VERBOSE:
|
|
||||||
indent_str = " " * level
|
|
||||||
# print(indent_str + "Opening:", filepath)
|
|
||||||
# print(indent_str + "... blocks:", block_codes)
|
|
||||||
|
|
||||||
|
|
||||||
log_deps.info("~")
|
|
||||||
log_deps.info("%s%s" % (indent_str, filepath.decode('utf-8')))
|
|
||||||
log_deps.info("%s%s" % (indent_str, set_as_str(block_codes)))
|
|
||||||
|
|
||||||
|
|
||||||
basedir = os.path.dirname(os.path.abspath(filepath))
|
|
||||||
if rootdir is None:
|
|
||||||
rootdir = basedir
|
|
||||||
|
|
||||||
# store info to pass along with each iteration
|
|
||||||
extra_info = rootdir, os.path.basename(filepath)
|
|
||||||
|
|
||||||
if recursive and (level > 0) and (block_codes is not None):
|
|
||||||
# prevent from expanding the
|
|
||||||
# same datablock more then once
|
|
||||||
expand_codes = set()
|
|
||||||
# {lib_id: {block_ids... }}
|
|
||||||
expand_codes_idlib = {}
|
|
||||||
|
|
||||||
# libraries used by this blend
|
|
||||||
block_codes_idlib = set()
|
|
||||||
|
|
||||||
# only for this block
|
|
||||||
def _expand_codes_add_test(block, code):
|
|
||||||
# return True, if the ID should be searched further
|
|
||||||
#
|
|
||||||
# we could investigate a better way...
|
|
||||||
# Not to be accessing ID blocks at this point. but its harmless
|
|
||||||
if code == b'ID':
|
|
||||||
assert(code == block.code)
|
|
||||||
if recursive:
|
|
||||||
expand_codes_idlib.setdefault(block[b'lib'], set()).add(block[b'name'])
|
|
||||||
return False
|
|
||||||
else:
|
|
||||||
len_prev = len(expand_codes)
|
|
||||||
expand_codes.add(block[b'id.name'])
|
|
||||||
return (len_prev != len(expand_codes))
|
|
||||||
|
|
||||||
def block_expand(block, code):
|
|
||||||
assert(block.code == code)
|
|
||||||
if _expand_codes_add_test(block, code):
|
|
||||||
yield block
|
|
||||||
|
|
||||||
assert(block.code == code)
|
|
||||||
fn = ExpandID.expand_funcs.get(code)
|
|
||||||
if fn is not None:
|
|
||||||
for sub_block in fn(block):
|
|
||||||
if sub_block is not None:
|
|
||||||
yield from block_expand(sub_block, sub_block.code)
|
|
||||||
else:
|
|
||||||
yield block
|
|
||||||
else:
|
|
||||||
expand_codes = None
|
|
||||||
|
|
||||||
# set below
|
|
||||||
expand_codes_idlib = None
|
|
||||||
|
|
||||||
# never set
|
|
||||||
block_codes_idlib = None
|
|
||||||
|
|
||||||
def block_expand(block, code):
|
|
||||||
assert(block.code == code)
|
|
||||||
yield block
|
|
||||||
|
|
||||||
# ------
|
|
||||||
# Define
|
|
||||||
#
|
|
||||||
# - iter_blocks_id(code)
|
|
||||||
# - iter_blocks_idlib()
|
|
||||||
if block_codes is None:
|
|
||||||
def iter_blocks_id(code):
|
|
||||||
return blend.find_blocks_from_code(code)
|
|
||||||
|
|
||||||
def iter_blocks_idlib():
|
|
||||||
return blend.find_blocks_from_code(b'LI')
|
|
||||||
else:
|
|
||||||
def iter_blocks_id(code):
|
|
||||||
for block in blend.find_blocks_from_code(code):
|
|
||||||
if block[b'id.name'] in block_codes:
|
|
||||||
yield from block_expand(block, code)
|
|
||||||
|
|
||||||
if block_codes_idlib is not None:
|
|
||||||
def iter_blocks_idlib():
|
|
||||||
for block in blend.find_blocks_from_code(b'LI'):
|
|
||||||
if block[b'name'] in block_codes_idlib:
|
|
||||||
yield from block_expand(block, b'LI')
|
|
||||||
else:
|
|
||||||
def iter_blocks_idlib():
|
|
||||||
return blend.find_blocks_from_code(b'LI')
|
|
||||||
|
|
||||||
if temp_remap_cb is not None:
|
|
||||||
filepath_tmp = temp_remap_cb(filepath, level)
|
|
||||||
else:
|
|
||||||
filepath_tmp = filepath
|
|
||||||
|
|
||||||
import blendfile
|
|
||||||
blend = blendfile.open_blend(filepath_tmp, "rb" if readonly else "r+b")
|
|
||||||
|
|
||||||
for code in blend.code_index.keys():
|
|
||||||
# handle library blocks as special case
|
|
||||||
if ((len(code) != 2) or
|
|
||||||
(code in {
|
|
||||||
# libraries handled below
|
|
||||||
b'LI',
|
|
||||||
b'ID',
|
|
||||||
# unneeded
|
|
||||||
b'WM',
|
|
||||||
b'SN', # bScreen
|
|
||||||
})):
|
|
||||||
|
|
||||||
continue
|
|
||||||
|
|
||||||
# if VERBOSE:
|
|
||||||
# print(" Scanning", code)
|
|
||||||
|
|
||||||
for block in iter_blocks_id(code):
|
|
||||||
yield from FilePath.from_block(block, basedir, extra_info, level)
|
|
||||||
|
|
||||||
# print("A:", expand_codes)
|
|
||||||
# print("B:", block_codes)
|
|
||||||
if VERBOSE:
|
|
||||||
log_deps.info("%s%s" % (indent_str, set_as_str(expand_codes)))
|
|
||||||
|
|
||||||
if recursive:
|
|
||||||
|
|
||||||
if expand_codes_idlib is None:
|
|
||||||
expand_codes_idlib = {}
|
|
||||||
for block in blend.find_blocks_from_code(b'ID'):
|
|
||||||
expand_codes_idlib.setdefault(block[b'lib'], set()).add(block[b'name'])
|
|
||||||
|
|
||||||
# look into libraries
|
|
||||||
lib_all = []
|
|
||||||
|
|
||||||
for lib_id, lib_block_codes in sorted(expand_codes_idlib.items()):
|
|
||||||
lib = blend.find_block_from_offset(lib_id)
|
|
||||||
lib_path = lib[b'name']
|
|
||||||
|
|
||||||
# get all data needed to read the blend files here (it will be freed!)
|
|
||||||
# lib is an address at the moment, we only use as a way to group
|
|
||||||
|
|
||||||
lib_all.append((lib_path, lib_block_codes))
|
|
||||||
# import IPython; IPython.embed()
|
|
||||||
|
|
||||||
# ensure we expand indirect linked libs
|
|
||||||
if block_codes_idlib is not None:
|
|
||||||
block_codes_idlib.add(lib_path)
|
|
||||||
|
|
||||||
# do this after, incase we mangle names above
|
|
||||||
for block in iter_blocks_idlib():
|
|
||||||
yield from FilePath.from_block(block, basedir, extra_info, level)
|
|
||||||
|
|
||||||
blend.close()
|
|
||||||
|
|
||||||
# ----------------
|
|
||||||
# Handle Recursive
|
|
||||||
if recursive:
|
|
||||||
# now we've closed the file, loop on other files
|
|
||||||
|
|
||||||
# note, sorting - isn't needed, it just gives predictable load-order.
|
|
||||||
for lib_path, lib_block_codes in lib_all:
|
|
||||||
lib_path_abs = os.path.normpath(utils.compatpath(utils.abspath(lib_path, basedir)))
|
|
||||||
|
|
||||||
# if we visited this before,
|
|
||||||
# check we don't follow the same links more than once
|
|
||||||
lib_block_codes_existing = lib_visit.setdefault(lib_path_abs, set())
|
|
||||||
lib_block_codes -= lib_block_codes_existing
|
|
||||||
# don't touch them again
|
|
||||||
lib_block_codes_existing.update(lib_block_codes)
|
|
||||||
|
|
||||||
# print("looking for", lib_block_codes)
|
|
||||||
|
|
||||||
# import IPython; IPython.embed()
|
|
||||||
if VERBOSE:
|
|
||||||
print((indent_str + " "), "Library: ", filepath, " -> ", lib_path_abs, sep="")
|
|
||||||
# print((indent_str + " "), lib_block_codes)
|
|
||||||
yield from FilePath.visit_from_blend(
|
|
||||||
lib_path_abs,
|
|
||||||
readonly=readonly,
|
|
||||||
temp_remap_cb=temp_remap_cb,
|
|
||||||
recursive=True,
|
|
||||||
block_codes=lib_block_codes,
|
|
||||||
rootdir=rootdir,
|
|
||||||
level=level + 1,
|
|
||||||
lib_visit=lib_visit,
|
|
||||||
)
|
|
||||||
|
|
||||||
# ------------------------------------------------------------------------
|
|
||||||
# Direct filepaths from Blocks
|
|
||||||
#
|
|
||||||
# (no expanding or following references)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def from_block(block, basedir, extra_info, level):
|
|
||||||
assert(block.code != b'DATA')
|
|
||||||
fn = FilePath._from_block_dict.get(block.code)
|
|
||||||
if fn is not None:
|
|
||||||
yield from fn(block, basedir, extra_info, level)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_MC(block, basedir, extra_info, level):
|
|
||||||
# TODO, image sequence
|
|
||||||
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_IM(block, basedir, extra_info, level):
|
|
||||||
# (IMA_SRC_FILE, IMA_SRC_SEQUENCE, IMA_SRC_MOVIE)
|
|
||||||
if block[b'source'] not in {1, 2, 3}:
|
|
||||||
return
|
|
||||||
if block[b'packedfile']:
|
|
||||||
return
|
|
||||||
|
|
||||||
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_VF(block, basedir, extra_info, level):
|
|
||||||
if block[b'packedfile']:
|
|
||||||
return
|
|
||||||
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_SO(block, basedir, extra_info, level):
|
|
||||||
if block[b'packedfile']:
|
|
||||||
return
|
|
||||||
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_ME(block, basedir, extra_info, level):
|
|
||||||
block_external = block.get_pointer(b'ldata.external')
|
|
||||||
if block_external is not None:
|
|
||||||
yield FPElem_block_path(basedir, level, (block_external, b'filename')), extra_info
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_SC(block, basedir, extra_info, level):
|
|
||||||
block_ed = block.get_pointer(b'ed')
|
|
||||||
if block_ed is not None:
|
|
||||||
sdna_index_Sequence = block.file.sdna_index_from_id[b'Sequence']
|
|
||||||
|
|
||||||
def seqbase(someseq):
|
|
||||||
for item in someseq:
|
|
||||||
item_type = item.get(b'type', sdna_index_refine=sdna_index_Sequence)
|
|
||||||
|
|
||||||
if item_type >= C_defs.SEQ_TYPE_EFFECT:
|
|
||||||
pass
|
|
||||||
elif item_type == C_defs.SEQ_TYPE_META:
|
|
||||||
yield from seqbase(bf_utils.iter_ListBase(item.get_pointer(b'seqbase.first', sdna_index_refine=sdna_index_Sequence)))
|
|
||||||
else:
|
|
||||||
item_strip = item.get_pointer(b'strip', sdna_index_refine=sdna_index_Sequence)
|
|
||||||
if item_strip is None: # unlikely!
|
|
||||||
continue
|
|
||||||
item_stripdata = item_strip.get_pointer(b'stripdata')
|
|
||||||
|
|
||||||
if item_type == C_defs.SEQ_TYPE_IMAGE:
|
|
||||||
# TODO, multiple images
|
|
||||||
yield FPElem_sequence_single(basedir, level, (item_strip, b'dir', item_stripdata, b'name')), extra_info
|
|
||||||
elif item_type == C_defs.SEQ_TYPE_MOVIE:
|
|
||||||
yield FPElem_sequence_single(basedir, level, (item_strip, b'dir', item_stripdata, b'name')), extra_info
|
|
||||||
elif item_type == C_defs.SEQ_TYPE_SOUND_RAM:
|
|
||||||
pass
|
|
||||||
elif item_type == C_defs.SEQ_TYPE_SOUND_HD:
|
|
||||||
pass
|
|
||||||
|
|
||||||
yield from seqbase(bf_utils.iter_ListBase(block_ed.get_pointer(b'seqbase.first')))
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _from_block_LI(block, basedir, extra_info, level):
|
|
||||||
if block.get(b'packedfile', None):
|
|
||||||
return
|
|
||||||
|
|
||||||
yield FPElem_block_path(basedir, level, (block, b'name')), extra_info
|
|
||||||
|
|
||||||
# _from_block_IM --> {b'IM': _from_block_IM, ...}
|
|
||||||
_from_block_dict = {
|
|
||||||
k.rpartition("_")[2].encode('ascii'): s_fn.__func__ for k, s_fn in locals().items()
|
|
||||||
if isinstance(s_fn, staticmethod)
|
|
||||||
if k.startswith("_from_block_")
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
class bf_utils:
|
|
||||||
@staticmethod
|
|
||||||
def iter_ListBase(block):
|
|
||||||
while block:
|
|
||||||
yield block
|
|
||||||
block = block.file.find_block_from_offset(block[b'next'])
|
|
||||||
|
|
||||||
def iter_array(block, length=-1):
|
|
||||||
assert(block.code == b'DATA')
|
|
||||||
import blendfile
|
|
||||||
import os
|
|
||||||
handle = block.file.handle
|
|
||||||
header = block.file.header
|
|
||||||
|
|
||||||
for i in range(length):
|
|
||||||
block.file.handle.seek(block.file_offset + (header.pointer_size * i), os.SEEK_SET)
|
|
||||||
offset = blendfile.DNA_IO.read_pointer(handle, header)
|
|
||||||
sub_block = block.file.find_block_from_offset(offset)
|
|
||||||
yield sub_block
|
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
# ID Expand
|
|
||||||
|
|
||||||
class ExpandID:
|
|
||||||
# fake module
|
|
||||||
#
|
|
||||||
# TODO:
|
|
||||||
#
|
|
||||||
# Array lookups here are _WAY_ too complicated,
|
|
||||||
# we need some nicer way to represent pointer indirection (easy like in C!)
|
|
||||||
# but for now, use what we have.
|
|
||||||
#
|
|
||||||
__slots__ = ()
|
|
||||||
|
|
||||||
def __new__(cls, *args, **kwargs):
|
|
||||||
raise RuntimeError("%s should not be instantiated" % cls)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _expand_generic_material(block):
|
|
||||||
array_len = block.get(b'totcol')
|
|
||||||
if array_len != 0:
|
|
||||||
array = block.get_pointer(b'mat')
|
|
||||||
for sub_block in bf_utils.iter_array(array, array_len):
|
|
||||||
yield sub_block
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _expand_generic_mtex(block):
|
|
||||||
field = block.dna_type.field_from_name[b'mtex']
|
|
||||||
array_len = field.dna_size // block.file.header.pointer_size
|
|
||||||
|
|
||||||
for i in range(array_len):
|
|
||||||
path = ('mtex[%d]' % i).encode('ascii')
|
|
||||||
item = block.get_pointer(path)
|
|
||||||
if item:
|
|
||||||
yield item.get_pointer(b'tex')
|
|
||||||
yield item.get_pointer(b'object')
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _expand_generic_nodetree(block):
|
|
||||||
assert(block.dna_type.dna_type_id == b'bNodeTree')
|
|
||||||
|
|
||||||
sdna_index_bNode = block.file.sdna_index_from_id[b'bNode']
|
|
||||||
for item in bf_utils.iter_ListBase(block.get_pointer(b'nodes.first')):
|
|
||||||
item_type = item.get(b'type', sdna_index_refine=sdna_index_bNode)
|
|
||||||
|
|
||||||
if item_type != 221: # CMP_NODE_R_LAYERS
|
|
||||||
yield item.get_pointer(b'id', sdna_index_refine=sdna_index_bNode)
|
|
||||||
|
|
||||||
def _expand_generic_nodetree_id(block):
|
|
||||||
block_ntree = block.get_pointer(b'nodetree')
|
|
||||||
if block_ntree is not None:
|
|
||||||
yield from ExpandID._expand_generic_nodetree(block_ntree)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def _expand_generic_animdata(block):
|
|
||||||
block_adt = block.get_pointer(b'adt')
|
|
||||||
if block_adt:
|
|
||||||
yield block_adt.get_pointer(b'action')
|
|
||||||
# TODO, NLA
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_OB(block): # 'Object'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield block.get_pointer(b'data')
|
|
||||||
yield block.get_pointer(b'dup_group')
|
|
||||||
|
|
||||||
yield block.get_pointer(b'proxy')
|
|
||||||
yield block.get_pointer(b'proxy_group')
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_ME(block): # 'Mesh'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_material(block)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_CU(block): # 'Curve'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_material(block)
|
|
||||||
|
|
||||||
sub_block = block.get_pointer(b'vfont')
|
|
||||||
if sub_block is not None:
|
|
||||||
yield sub_block
|
|
||||||
yield block.get_pointer(b'vfontb')
|
|
||||||
yield block.get_pointer(b'vfonti')
|
|
||||||
yield block.get_pointer(b'vfontbi')
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_MB(block): # 'MBall'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_material(block)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_LA(block): # 'Lamp'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree_id(block)
|
|
||||||
yield from ExpandID._expand_generic_mtex(block)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_MA(block): # 'Material'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree_id(block)
|
|
||||||
yield from ExpandID._expand_generic_mtex(block)
|
|
||||||
|
|
||||||
yield block.get_pointer(b'group')
|
|
||||||
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_TE(block): # 'Tex'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree_id(block)
|
|
||||||
yield block.get_pointer(b'ima')
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_WO(block): # 'World'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree_id(block)
|
|
||||||
yield from ExpandID._expand_generic_mtex(block)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_NT(block): # 'bNodeTree'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree(block)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_SC(block): # 'Scene'
|
|
||||||
yield from ExpandID._expand_generic_animdata(block)
|
|
||||||
yield from ExpandID._expand_generic_nodetree_id(block)
|
|
||||||
yield block.get_pointer(b'world')
|
|
||||||
|
|
||||||
sdna_index_Base = block.file.sdna_index_from_id[b'Base']
|
|
||||||
for item in bf_utils.iter_ListBase(block.get_pointer(b'base.first')):
|
|
||||||
yield item.get_pointer(b'object', sdna_index_refine=sdna_index_Base)
|
|
||||||
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def expand_GR(block): # 'Group'
|
|
||||||
sdna_index_GroupObject = block.file.sdna_index_from_id[b'GroupObject']
|
|
||||||
for item in bf_utils.iter_ListBase(block.get_pointer(b'gobject.first')):
|
|
||||||
yield item.get_pointer(b'ob', sdna_index_refine=sdna_index_GroupObject)
|
|
||||||
|
|
||||||
# expand_GR --> {b'GR': expand_GR, ...}
|
|
||||||
expand_funcs = {
|
|
||||||
k.rpartition("_")[2].encode('ascii'): s_fn.__func__ for k, s_fn in locals().items()
|
|
||||||
if isinstance(s_fn, staticmethod)
|
|
||||||
if k.startswith("expand_")
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
# -----------------------------------------------------------------------------
|
|
||||||
# Packing Utility
|
|
||||||
|
|
||||||
|
|
||||||
class utils:
|
|
||||||
# fake module
|
|
||||||
__slots__ = ()
|
|
||||||
|
|
||||||
def __new__(cls, *args, **kwargs):
|
|
||||||
raise RuntimeError("%s should not be instantiated" % cls)
|
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def abspath(path, start, library=None):
|
|
||||||
import os
|
|
||||||
if path.startswith(b'//'):
|
|
||||||
# if library:
|
|
||||||
# start = os.path.dirname(abspath(library.filepath))
|
|
||||||
return os.path.join(start, path[2:])
|
|
||||||
return path
|
|
||||||
|
|
||||||
if __import__("os").sep == '/':
|
|
||||||
@staticmethod
|
|
||||||
def compatpath(path):
|
|
||||||
return path.replace(b'\\', b'/')
|
|
||||||
else:
|
|
||||||
@staticmethod
|
|
||||||
def compatpath(path):
|
|
||||||
# keep '//'
|
|
||||||
return path[:2] + path[2:].replace(b'/', b'\\')
|
|
||||||
|
|
||||||
|
|
||||||
def pack(blendfile_src, blendfile_dst, mode='FILE', remap=None):
|
|
||||||
"""
|
|
||||||
:param remap: Store path remapping info as follows.
|
|
||||||
{"file.blend": {"path_new": "path_old", ...}, ...}
|
{"file.blend": {"path_new": "path_old", ...}, ...}
|
||||||
|
|
||||||
:type remap: dict or None
|
:type pathmap: dict or None
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# Internal details:
|
# Internal details:
|
||||||
@@ -678,7 +59,7 @@ def pack(blendfile_src, blendfile_dst, mode='FILE', remap=None):
|
|||||||
"""
|
"""
|
||||||
Create temp files in the destination path.
|
Create temp files in the destination path.
|
||||||
"""
|
"""
|
||||||
filepath = utils.compatpath(filepath)
|
filepath = blendfile_path_walker.utils.compatpath(filepath)
|
||||||
|
|
||||||
if level == 0:
|
if level == 0:
|
||||||
filepath_tmp = os.path.join(base_dir_dst, os.path.basename(filepath)) + TEMP_SUFFIX
|
filepath_tmp = os.path.join(base_dir_dst, os.path.basename(filepath)) + TEMP_SUFFIX
|
||||||
@@ -687,7 +68,7 @@ def pack(blendfile_src, blendfile_dst, mode='FILE', remap=None):
|
|||||||
|
|
||||||
filepath_tmp = os.path.normpath(filepath_tmp)
|
filepath_tmp = os.path.normpath(filepath_tmp)
|
||||||
|
|
||||||
# only overwrite once (allows us to )
|
# only overwrite once (so we can write into a path already containing files)
|
||||||
if filepath_tmp not in path_temp_files:
|
if filepath_tmp not in path_temp_files:
|
||||||
shutil.copy(filepath, filepath_tmp)
|
shutil.copy(filepath, filepath_tmp)
|
||||||
path_temp_files.add(filepath_tmp)
|
path_temp_files.add(filepath_tmp)
|
||||||
@@ -702,23 +83,25 @@ def pack(blendfile_src, blendfile_dst, mode='FILE', remap=None):
|
|||||||
|
|
||||||
lib_visit = {}
|
lib_visit = {}
|
||||||
|
|
||||||
for fp, (rootdir, fp_blend_basename) in FilePath.visit_from_blend(
|
for fp, (rootdir, fp_blend_basename) in blendfile_path_walker.FilePath.visit_from_blend(
|
||||||
blendfile_src,
|
blendfile_src,
|
||||||
readonly=False,
|
readonly=False,
|
||||||
temp_remap_cb=temp_remap_cb,
|
temp_remap_cb=temp_remap_cb,
|
||||||
recursive=True,
|
recursive=True,
|
||||||
lib_visit=lib_visit):
|
lib_visit=lib_visit,
|
||||||
|
):
|
||||||
|
|
||||||
# assume the path might be relative
|
# assume the path might be relative
|
||||||
path_rel = utils.compatpath(fp.filepath)
|
path_src_orig = fp.filepath
|
||||||
|
path_rel = blendfile_path_walker.utils.compatpath(path_src_orig)
|
||||||
path_base = path_rel.split(os.sep.encode('ascii'))[-1]
|
path_base = path_rel.split(os.sep.encode('ascii'))[-1]
|
||||||
path_src = utils.abspath(path_rel, fp.basedir)
|
path_src = blendfile_path_walker.utils.abspath(path_rel, fp.basedir)
|
||||||
|
|
||||||
# rename in the blend
|
# rename in the blend
|
||||||
path_dst = os.path.join(base_dir_dst_subdir, path_base)
|
path_dst = os.path.join(base_dir_dst_subdir, path_base)
|
||||||
|
|
||||||
if fp.level == 0:
|
if fp.level == 0:
|
||||||
path_dst_final =b"//" + os.path.join(SUBDIR, path_base)
|
path_dst_final = b"//" + os.path.join(SUBDIR, path_base)
|
||||||
else:
|
else:
|
||||||
path_dst_final = b'//' + path_base
|
path_dst_final = b'//' + path_base
|
||||||
|
|
||||||
@@ -726,25 +109,40 @@ def pack(blendfile_src, blendfile_dst, mode='FILE', remap=None):
|
|||||||
|
|
||||||
# add to copy-list
|
# add to copy-list
|
||||||
# never copy libs (handled separately)
|
# never copy libs (handled separately)
|
||||||
if not isinstance(fp, FPElem_block_path) or fp.userdata[0].code != b'LI':
|
if not isinstance(fp, blendfile_path_walker.FPElem_block_path) or fp.userdata[0].code != b'LI':
|
||||||
path_copy_files.add((path_src, path_dst))
|
path_copy_files.add((path_src, path_dst))
|
||||||
|
|
||||||
if remap is not None:
|
if pathmap is not None:
|
||||||
# this needs to become JSON later... ugh, need to use strings
|
# this needs to become JSON later... ugh, need to use strings
|
||||||
remap.setdefault(
|
pathmap.setdefault(
|
||||||
fp_blend_basename.decode('utf-8'),
|
fp_blend_basename.decode('utf-8'),
|
||||||
{})[path_dst_final.decode('utf-8')] = path_src.decode('utf-8')
|
{})[path_dst_final.decode('utf-8')] = path_src_orig.decode('utf-8')
|
||||||
|
|
||||||
del lib_visit
|
del lib_visit
|
||||||
|
|
||||||
if TIMEIT:
|
if TIMEIT:
|
||||||
print(" Time: %.4f\n" % (time.time() - t))
|
print(" Time: %.4f\n" % (time.time() - t))
|
||||||
|
|
||||||
|
# handle pathmap and file renaming
|
||||||
|
if pathmap is not None:
|
||||||
|
blendfile_src_basename = os.path.basename(blendfile_src).decode('utf-8')
|
||||||
|
blendfile_dst_basename = os.path.basename(blendfile_dst).decode('utf-8')
|
||||||
|
print("AAAAA")
|
||||||
|
print(blendfile_src_basename, blendfile_dst_basename)
|
||||||
|
|
||||||
|
if blendfile_src_basename != blendfile_dst_basename:
|
||||||
|
pathmap[blendfile_dst_basename] = pathmap[blendfile_src_basename]
|
||||||
|
del pathmap[blendfile_src_basename]
|
||||||
|
# del blendfile_src_basename, blendfile_dst_basename
|
||||||
|
# print(pathmap)
|
||||||
|
|
||||||
|
|
||||||
# --------------------
|
# --------------------
|
||||||
# Handle File Copy/Zip
|
# Handle File Copy/Zip
|
||||||
|
|
||||||
if mode == 'FILE':
|
if mode == 'FILE':
|
||||||
blendfile_dst_tmp = temp_remap_cb(blendfile_src, 0)
|
blendfile_dst_tmp = temp_remap_cb(blendfile_src, 0)
|
||||||
|
|
||||||
shutil.move(blendfile_dst_tmp, blendfile_dst)
|
shutil.move(blendfile_dst_tmp, blendfile_dst)
|
||||||
path_temp_files.remove(blendfile_dst_tmp)
|
path_temp_files.remove(blendfile_dst_tmp)
|
||||||
|
|
||||||
@@ -805,7 +203,7 @@ def create_argparse():
|
|||||||
"-i", "--input", dest="path_src", metavar='FILE', required=True,
|
"-i", "--input", dest="path_src", metavar='FILE', required=True,
|
||||||
help="Input path(s) or a wildcard to glob many files")
|
help="Input path(s) or a wildcard to glob many files")
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-o", "--output", dest="path_dst", metavar='FILE', required=True,
|
"-o", "--output", dest="path_dst", metavar='DIR', required=True,
|
||||||
help="Output file or a directory when multiple inputs are passed")
|
help="Output file or a directory when multiple inputs are passed")
|
||||||
parser.add_argument(
|
parser.add_argument(
|
||||||
"-m", "--mode", dest="mode", metavar='MODE', required=False,
|
"-m", "--mode", dest="mode", metavar='MODE', required=False,
|
||||||
@@ -827,23 +225,22 @@ def main():
|
|||||||
encoding = sys.getfilesystemencoding()
|
encoding = sys.getfilesystemencoding()
|
||||||
|
|
||||||
if args.path_remap:
|
if args.path_remap:
|
||||||
remap = {}
|
pathmap = {}
|
||||||
else:
|
else:
|
||||||
remap = None
|
pathmap = None
|
||||||
|
|
||||||
pack(args.path_src.encode(encoding),
|
pack(args.path_src.encode(encoding),
|
||||||
args.path_dst.encode(encoding),
|
args.path_dst.encode(encoding),
|
||||||
args.mode,
|
args.mode,
|
||||||
remap,
|
pathmap,
|
||||||
)
|
)
|
||||||
|
|
||||||
if remap is not None:
|
if pathmap is not None:
|
||||||
import json
|
import json
|
||||||
|
|
||||||
with open(args.path_remap, 'w', encoding='utf-8') as f:
|
with open(args.path_remap, 'w', encoding='utf-8') as f:
|
||||||
|
|
||||||
json.dump(
|
json.dump(
|
||||||
remap, f, ensure_ascii=False,
|
pathmap, f, ensure_ascii=False,
|
||||||
# optional (pretty)
|
# optional (pretty)
|
||||||
sort_keys=True, indent=4, separators=(',', ': '),
|
sort_keys=True, indent=4, separators=(',', ': '),
|
||||||
)
|
)
|
||||||
|
131
packer/packer_restore.py
Executable file
131
packer/packer_restore.py
Executable file
@@ -0,0 +1,131 @@
|
|||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
# ***** BEGIN GPL LICENSE BLOCK *****
|
||||||
|
#
|
||||||
|
# This program is free software; you can redistribute it and/or
|
||||||
|
# modify it under the terms of the GNU General Public License
|
||||||
|
# as published by the Free Software Foundation; either version 2
|
||||||
|
# of the License, or (at your option) any later version.
|
||||||
|
#
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU General Public License for more details.
|
||||||
|
#
|
||||||
|
# You should have received a copy of the GNU General Public License
|
||||||
|
# along with this program; if not, write to the Free Software Foundation,
|
||||||
|
# Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
||||||
|
#
|
||||||
|
# ***** END GPL LICENCE BLOCK *****
|
||||||
|
|
||||||
|
"""
|
||||||
|
This script takes Blend-File and remaps their paths to the original locations.
|
||||||
|
|
||||||
|
(needed for uploading to the server)
|
||||||
|
"""
|
||||||
|
|
||||||
|
VERBOSE = 1
|
||||||
|
|
||||||
|
import blendfile_path_walker
|
||||||
|
|
||||||
|
|
||||||
|
def blendfile_remap(blendfile_src, blendpath_dst, path_remap):
|
||||||
|
import os
|
||||||
|
|
||||||
|
def temp_remap_cb(filepath, level):
|
||||||
|
"""
|
||||||
|
Simply point to the output dir.
|
||||||
|
"""
|
||||||
|
basename = os.path.basename(blendfile_src)
|
||||||
|
filepath_tmp = os.path.join(blendpath_dst, basename)
|
||||||
|
|
||||||
|
# ideally we could avoid copying _ALL_ blends
|
||||||
|
# TODO(cam)
|
||||||
|
import shutil
|
||||||
|
shutil.copy(filepath, filepath_tmp)
|
||||||
|
|
||||||
|
return filepath_tmp
|
||||||
|
|
||||||
|
|
||||||
|
for fp, (rootdir, fp_blend_basename) in blendfile_path_walker.FilePath.visit_from_blend(
|
||||||
|
blendfile_src,
|
||||||
|
readonly=False,
|
||||||
|
temp_remap_cb=temp_remap_cb,
|
||||||
|
recursive=False,
|
||||||
|
):
|
||||||
|
|
||||||
|
# path_dst_final - current path in blend.
|
||||||
|
# path_src_orig - original path from JSON.
|
||||||
|
|
||||||
|
path_dst_final = fp.filepath.decode('utf-8')
|
||||||
|
path_src_orig = path_remap.get(path_dst_final)
|
||||||
|
if path_src_orig is not None:
|
||||||
|
fp.filepath = path_src_orig.encode('utf-8')
|
||||||
|
if VERBOSE:
|
||||||
|
print(" Remapping:", path_dst_final, "->", path_src_orig)
|
||||||
|
|
||||||
|
|
||||||
|
def pack_restore(blendfile_dir_src, blendfile_dir_dst, pathmap):
|
||||||
|
import os
|
||||||
|
|
||||||
|
for dirpath, dirnames, filenames in os.walk(blendfile_dir_src):
|
||||||
|
if dirpath.startswith(b"."):
|
||||||
|
continue
|
||||||
|
|
||||||
|
for filename in filenames:
|
||||||
|
if os.path.splitext(filename)[1].lower() == b".blend":
|
||||||
|
remap = pathmap.get(filename.decode('utf-8'))
|
||||||
|
if remap is not None:
|
||||||
|
filepath = os.path.join(dirpath, filename)
|
||||||
|
|
||||||
|
# main function call
|
||||||
|
blendfile_remap(filepath, blendfile_dir_dst, remap)
|
||||||
|
|
||||||
|
|
||||||
|
def create_argparse():
|
||||||
|
import os
|
||||||
|
import argparse
|
||||||
|
|
||||||
|
usage_text = (
|
||||||
|
"Run this script to remap blend-file(s) paths using a JSON file created by 'packer.py':" +
|
||||||
|
os.path.basename(__file__) +
|
||||||
|
"--input=DIR --remap=JSON [options]")
|
||||||
|
|
||||||
|
parser = argparse.ArgumentParser(description=usage_text)
|
||||||
|
|
||||||
|
# for main_render() only, but validate args.
|
||||||
|
parser.add_argument(
|
||||||
|
"-i", "--input", dest="path_src", metavar='DIR', required=True,
|
||||||
|
help="Input path(s) or a wildcard to glob many files")
|
||||||
|
parser.add_argument(
|
||||||
|
"-o", "--output", dest="path_dst", metavar='DIR', required=True,
|
||||||
|
help="Output directory ")
|
||||||
|
parser.add_argument(
|
||||||
|
"-r", "--remap", dest="path_remap", metavar='JSON', required=True,
|
||||||
|
help="JSON file containing the path remapping info")
|
||||||
|
|
||||||
|
return parser
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
import sys
|
||||||
|
import json
|
||||||
|
|
||||||
|
parser = create_argparse()
|
||||||
|
args = parser.parse_args(sys.argv[1:])
|
||||||
|
|
||||||
|
encoding = sys.getfilesystemencoding()
|
||||||
|
|
||||||
|
with open(args.path_remap, 'r', encoding='utf-8') as f:
|
||||||
|
pathmap = json.load(f)
|
||||||
|
|
||||||
|
pack_restore(
|
||||||
|
args.path_src.encode(encoding),
|
||||||
|
args.path_dst.encode(encoding),
|
||||||
|
pathmap,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
|
|
Reference in New Issue
Block a user