Merge branch 'develop' into 45-vse-support

This commit is contained in:
Swann 2020-11-04 22:42:05 +01:00
commit c00a7184ff
No known key found for this signature in database
GPG Key ID: E1D3641A7C43AACB
8 changed files with 45 additions and 57 deletions

View File

@ -44,7 +44,7 @@ from . import environment
DEPENDENCIES = { DEPENDENCIES = {
("replication", '0.1.7'), ("replication", '0.1.8'),
} }

View File

@ -42,7 +42,7 @@ KEYFRAME = [
] ]
def dump_fcurve(fcurve: bpy.types.FCurve, use_numpy:bool =True) -> dict: def dump_fcurve(fcurve: bpy.types.FCurve, use_numpy: bool = True) -> dict:
""" Dump a sigle curve to a dict """ Dump a sigle curve to a dict
:arg fcurve: fcurve to dump :arg fcurve: fcurve to dump
@ -59,7 +59,7 @@ def dump_fcurve(fcurve: bpy.types.FCurve, use_numpy:bool =True) -> dict:
if use_numpy: if use_numpy:
points = fcurve.keyframe_points points = fcurve.keyframe_points
fcurve_data['keyframes_count'] = len(fcurve.keyframe_points) fcurve_data['keyframes_count'] = len(fcurve.keyframe_points)
fcurve_data['keyframe_points'] = np_dump_collection(points, KEYFRAME) fcurve_data['keyframe_points'] = np_dump_collection(points, KEYFRAME)
else: # Legacy method else: # Legacy method
@ -92,7 +92,8 @@ def load_fcurve(fcurve_data, fcurve):
if use_numpy: if use_numpy:
keyframe_points.add(fcurve_data['keyframes_count']) keyframe_points.add(fcurve_data['keyframes_count'])
np_load_collection(fcurve_data["keyframe_points"], keyframe_points, KEYFRAME) np_load_collection(
fcurve_data["keyframe_points"], keyframe_points, KEYFRAME)
else: else:
# paste dumped keyframes # paste dumped keyframes
@ -153,8 +154,11 @@ class BlAction(BlDatablock):
dumped_data_path, index=dumped_array_index) dumped_data_path, index=dumped_array_index)
load_fcurve(dumped_fcurve, fcurve) load_fcurve(dumped_fcurve, fcurve)
if data['id_root']:
target.id_root = data['id_root'] id_root = data.get('id_root')
if id_root:
target.id_root = id_root
def _dump_implementation(self, data, instance=None): def _dump_implementation(self, data, instance=None):
dumper = Dumper() dumper = Dumper()

View File

@ -48,12 +48,15 @@ class BlCamera(BlDatablock):
background_images = data.get('background_images') background_images = data.get('background_images')
target.background_images.clear()
if background_images: if background_images:
target.background_images.clear()
for img_name, img_data in background_images.items(): for img_name, img_data in background_images.items():
target_img = target.background_images.new() img_id = img_data.get('image')
target_img.image = bpy.data.images[img_name] if img_id:
loader.load(target_img, img_data) target_img = target.background_images.new()
target_img.image = bpy.data.images[img_id]
loader.load(target_img, img_data)
def _dump_implementation(self, data, instance=None): def _dump_implementation(self, data, instance=None):
assert(instance) assert(instance)

View File

@ -37,28 +37,28 @@ def load_node(node_data, node_tree):
""" """
loader = Loader() loader = Loader()
target_node = node_tree.nodes.new(type=node_data["bl_idname"]) target_node = node_tree.nodes.new(type=node_data["bl_idname"])
target_node.select = False
loader.load(target_node, node_data) loader.load(target_node, node_data)
image_uuid = node_data.get('image_uuid', None) image_uuid = node_data.get('image_uuid', None)
if image_uuid and not target_node.image: if image_uuid and not target_node.image:
target_node.image = get_datablock_from_uuid(image_uuid, None) target_node.image = get_datablock_from_uuid(image_uuid, None)
for input in node_data["inputs"]: for idx, inpt in enumerate(node_data["inputs"]):
if hasattr(target_node.inputs[input], "default_value"): if hasattr(target_node.inputs[idx], "default_value"):
try: try:
target_node.inputs[input].default_value = node_data["inputs"][input]["default_value"] target_node.inputs[idx].default_value = inpt["default_value"]
except: except:
logging.error( logging.error(
f"Material {input} parameter not supported, skipping") f"Material {inpt.keys()} parameter not supported, skipping")
for output in node_data["outputs"]: for idx, output in enumerate(node_data["outputs"]):
if hasattr(target_node.outputs[output], "default_value"): if hasattr(target_node.outputs[idx], "default_value"):
try: try:
target_node.outputs[output].default_value = node_data["outputs"][output]["default_value"] target_node.outputs[idx].default_value = output["default_value"]
except: except:
logging.error( logging.error(
f"Material {output} parameter not supported, skipping") f"Material {output.keys()} parameter not supported, skipping")
def load_links(links_data, node_tree): def load_links(links_data, node_tree):
@ -142,24 +142,20 @@ def dump_node(node):
dumped_node = node_dumper.dump(node) dumped_node = node_dumper.dump(node)
if hasattr(node, 'inputs'): if hasattr(node, 'inputs'):
dumped_node['inputs'] = {} dumped_node['inputs'] = []
for i in node.inputs: io_dumper = Dumper()
input_dumper = Dumper() io_dumper.depth = 2
input_dumper.depth = 2 io_dumper.include_filter = ["default_value"]
input_dumper.include_filter = ["default_value"]
if hasattr(i, 'default_value'): for idx, inpt in enumerate(node.inputs):
dumped_node['inputs'][i.name] = input_dumper.dump(i) if hasattr(inpt, 'default_value'):
dumped_node['inputs'].append(io_dumper.dump(inpt))
dumped_node['outputs'] = {} dumped_node['outputs'] = []
for i in node.outputs: for idx, output in enumerate(node.outputs):
output_dumper = Dumper() if hasattr(output, 'default_value'):
output_dumper.depth = 2 dumped_node['outputs'].append(io_dumper.dump(output))
output_dumper.include_filter = ["default_value"]
if hasattr(i, 'default_value'):
dumped_node['outputs'][i.name] = output_dumper.dump(i)
if hasattr(node, 'color_ramp'): if hasattr(node, 'color_ramp'):
ramp_dumper = Dumper() ramp_dumper = Dumper()

View File

@ -172,12 +172,7 @@ class BlMesh(BlDatablock):
data['vertex_colors'][color_map.name]['data'] = np_dump_collection_primitive(color_map.data, 'color') data['vertex_colors'][color_map.name]['data'] = np_dump_collection_primitive(color_map.data, 'color')
# Fix material index # Fix material index
m_list = [] data['material_list'] = [(m.uuid, m.name) for m in instance.materials if m]
for material in instance.materials:
if material:
m_list.append((material.uuid,material.name))
data['material_list'] = m_list
return data return data

View File

@ -24,7 +24,6 @@ from replication.exception import ContextError
from .bl_datablock import BlDatablock, get_datablock_from_uuid from .bl_datablock import BlDatablock, get_datablock_from_uuid
from .dump_anything import Dumper, Loader from .dump_anything import Dumper, Loader
from replication.exception import ReparentException
def load_pose(target_bone, data): def load_pose(target_bone, data):
@ -120,9 +119,7 @@ class BlObject(BlDatablock):
data_uuid = data.get("data_uuid") data_uuid = data.get("data_uuid")
data_id = data.get("data") data_id = data.get("data")
if target.type != data['type']: if target.data and (target.data.name != data_id):
raise ReparentException()
elif target.data and (target.data.name != data_id):
target.data = get_datablock_from_uuid(data_uuid, find_data_from_name(data_id), ignore=['images']) target.data = get_datablock_from_uuid(data_uuid, find_data_from_name(data_id), ignore=['images'])
# vertex groups # vertex groups
@ -191,10 +188,10 @@ class BlObject(BlDatablock):
target_bone.bone_group = target.pose.bone_group[bone_data['bone_group_index']] target_bone.bone_group = target.pose.bone_group[bone_data['bone_group_index']]
# TODO: find another way... # TODO: find another way...
if target.type == 'EMPTY': if target.empty_display_type == "IMAGE":
img_uuid = data.get('data_uuid') img_uuid = data.get('data_uuid')
if target.data is None and img_uuid: if target.data is None and img_uuid:
target.data = get_datablock_from_uuid(img_uuid, None)#bpy.data.images.get(img_key, None) target.data = get_datablock_from_uuid(img_uuid, None)
def _dump_implementation(self, data, instance=None): def _dump_implementation(self, data, instance=None):
assert(instance) assert(instance)

View File

@ -36,8 +36,7 @@ from replication.constants import (FETCHED,
STATE_ACTIVE, STATE_ACTIVE,
STATE_SYNCING, STATE_SYNCING,
STATE_LOBBY, STATE_LOBBY,
STATE_SRV_SYNC, STATE_SRV_SYNC)
REPARENT)
from replication.interface import session from replication.interface import session
from replication.exception import NonAuthorizedOperationError from replication.exception import NonAuthorizedOperationError
@ -122,15 +121,6 @@ class ApplyTimer(Timer):
session.apply(node) session.apply(node)
except Exception as e: except Exception as e:
logging.error(f"Fail to apply {node_ref.uuid}: {e}") logging.error(f"Fail to apply {node_ref.uuid}: {e}")
elif node_ref.state == REPARENT:
# Reload the node
node_ref.remove_instance()
node_ref.resolve()
session.apply(node)
for parent in session._graph.find_parents(node):
logging.info(f"Applying parent {parent}")
session.apply(parent, force=True)
node_ref.state = UP
class DynamicRightSelectTimer(Timer): class DynamicRightSelectTimer(Timer):

View File

@ -62,6 +62,9 @@ def install_package(name, version):
del env["PIP_REQUIRE_VIRTUALENV"] del env["PIP_REQUIRE_VIRTUALENV"]
subprocess.run([str(PYTHON_PATH), "-m", "pip", "install", f"{name}=={version}"], env=env) subprocess.run([str(PYTHON_PATH), "-m", "pip", "install", f"{name}=={version}"], env=env)
if name in sys.modules:
del sys.modules[name]
def check_package_version(name, required_version): def check_package_version(name, required_version):
logging.info(f"Checking {name} version...") logging.info(f"Checking {name} version...")
out = subprocess.run([str(PYTHON_PATH), "-m", "pip", "show", name], capture_output=True) out = subprocess.run([str(PYTHON_PATH), "-m", "pip", "show", name], capture_output=True)