This commit is contained in:
Mark Moissette 2024-03-20 16:58:21 +00:00 committed by GitHub
commit 3c20464e93
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
11 changed files with 216 additions and 87 deletions

View File

@ -14,3 +14,7 @@ pub struct Animations {
/// this is for convenience, because currently , Bevy's gltf parsing inserts `AnimationPlayers` "one level down" /// this is for convenience, because currently , Bevy's gltf parsing inserts `AnimationPlayers` "one level down"
/// ie armature/root for animated models, which means more complex queries to trigger animations that we want to avoid /// ie armature/root for animated models, which means more complex queries to trigger animations that we want to avoid
pub struct AnimationPlayerLink(pub Entity); pub struct AnimationPlayerLink(pub Entity);
#[derive(Component, Reflect, Default, Debug)]
#[reflect(Component)]
pub struct Animated;

View File

@ -120,6 +120,7 @@ impl Plugin for BlueprintsPlugin {
.register_type::<MaterialInfo>() .register_type::<MaterialInfo>()
.register_type::<SpawnHere>() .register_type::<SpawnHere>()
.register_type::<Animations>() .register_type::<Animations>()
.register_type::<Animated>()
.register_type::<BlueprintsList>() .register_type::<BlueprintsList>()
.register_type::<Vec<String>>() .register_type::<Vec<String>>()
.register_type::<HashMap<String, Vec<String>>>() .register_type::<HashMap<String, Vec<String>>>()

View File

@ -14,3 +14,4 @@ bevy_rapier3d = { version = "0.25.0", features = ["serde-serialize", "debug-rend
bevy_asset_loader = { version = "0.20", features = ["standard_dynamic_assets"] } bevy_asset_loader = { version = "0.20", features = ["standard_dynamic_assets"] }
bevy_editor_pls = { version = "0.8" } bevy_editor_pls = { version = "0.8" }
rand = "0.8.5" rand = "0.8.5"
json-writer ="0.3"

View File

@ -3516,6 +3516,17 @@
"type": "object", "type": "object",
"typeInfo": "Struct" "typeInfo": "Struct"
}, },
"bevy_gltf_blueprints::animation::Animated": {
"additionalProperties": false,
"isComponent": true,
"isResource": false,
"properties": {},
"required": [],
"short_name": "Animated",
"title": "bevy_gltf_blueprints::animation::Animated",
"type": "object",
"typeInfo": "Struct"
},
"bevy_gltf_blueprints::animation::Animations": { "bevy_gltf_blueprints::animation::Animations": {
"additionalProperties": false, "additionalProperties": false,
"isComponent": true, "isComponent": true,

View File

@ -1,24 +1,30 @@
pub mod in_game; pub mod in_game;
use std::{ use std::{
fs::{self}, collections::HashMap, fs, time::Duration
time::Duration,
}; };
use bevy_gltf_blueprints::{AnimationPlayerLink, BlueprintName, BlueprintsList}; use bevy_gltf_blueprints::{AnimationPlayerLink, Animations, BlueprintName, BlueprintsList};
pub use in_game::*; pub use in_game::*;
use bevy::{ use bevy::{
prelude::*, render::view::screenshot::ScreenshotManager, time::common_conditions::on_timer, ecs::query, prelude::*, render::view::screenshot::ScreenshotManager, time::common_conditions::on_timer, window::PrimaryWindow
window::PrimaryWindow,
}; };
use bevy_gltf_worlflow_examples_common_rapier::{AppState, GameState}; use bevy_gltf_worlflow_examples_common_rapier::{AppState, GameState};
use crate::{TupleTestF32, UnitTest}; use crate::{TupleTestF32, UnitTest};
use json_writer::to_json_string;
fn start_game(mut next_app_state: ResMut<NextState<AppState>>) { fn start_game(mut next_app_state: ResMut<NextState<AppState>>) {
next_app_state.set(AppState::AppLoading); next_app_state.set(AppState::AppLoading);
} }
#[derive(Debug)]
struct Node {
value: String,
children: Vec<String>,
}
// if the export from Blender worked correctly, we should have animations (simplified here by using AnimationPlayerLink) // if the export from Blender worked correctly, we should have animations (simplified here by using AnimationPlayerLink)
// if the export from Blender worked correctly, we should have an Entity called "Cylinder" that has two components: UnitTest, TupleTestF32 // if the export from Blender worked correctly, we should have an Entity called "Cylinder" that has two components: UnitTest, TupleTestF32
// if the export from Blender worked correctly, we should have an Entity called "Blueprint4_nested" that has a child called "Blueprint3" that has a "BlueprintName" component with value Blueprint3 // if the export from Blender worked correctly, we should have an Entity called "Blueprint4_nested" that has a child called "Blueprint3" that has a "BlueprintName" component with value Blueprint3
@ -34,19 +40,10 @@ fn validate_export(
empties_candidates: Query<(Entity, &Name, &GlobalTransform)>, empties_candidates: Query<(Entity, &Name, &GlobalTransform)>,
blueprints_list: Query<(Entity, &BlueprintsList)>, blueprints_list: Query<(Entity, &BlueprintsList)>,
root: Query<(Entity, &Name, &Children), (Without<Parent>, With<Children>)>
) { ) {
let animations_found = !animation_player_links.is_empty(); let animations_found = !animation_player_links.is_empty();
let mut cylinder_found = false;
if let Ok(nested_cylinder) = exported_cylinder.get_single() {
let parent_name = names
.get(parents.get(nested_cylinder.0).unwrap().get())
.unwrap();
cylinder_found = parent_name.to_string() == *"Cube.001"
&& nested_cylinder.1.to_string() == *"Cylinder"
&& nested_cylinder.3 .0 == 75.1;
}
let mut nested_blueprint_found = false; let mut nested_blueprint_found = false;
for (entity, name, blueprint_name) in blueprints.iter() { for (entity, name, blueprint_name) in blueprints.iter() {
if name.to_string() == *"Blueprint4_nested" && blueprint_name.0 == *"Blueprint4_nested" { if name.to_string() == *"Blueprint4_nested" && blueprint_name.0 == *"Blueprint4_nested" {
@ -71,14 +68,46 @@ fn validate_export(
break; break;
} }
} }
// check if there are blueprints_list components
let blueprints_list_found = !blueprints_list.is_empty(); let blueprints_list_found = !blueprints_list.is_empty();
// there should be no entity named xxx____bak as it means an error in the Blender side export process
let mut exported_names_correct = true;
for name in names.iter() {
if name.to_string().ends_with("___bak") {
exported_names_correct = false;
break;
}
}
// generate parent/child tree
if !root.is_empty() {
let root = root.single();
let mut tree: HashMap<String, Vec<String>> = HashMap::new();
// println!("root {}", root.1);
for child in children.iter_descendants(root.0) {
let child_name:String = names.get(child).map_or(String::from("no_name"), |e| e.to_string() ); //|e| e.to_string(), || "no_name".to_string());
//println!(" child {}", child_name);
let parent = parents.get(child).unwrap();
let parent_name:String = names.get(parent.get()).map_or(String::from("no_name"), |e| e.to_string() ); //|e| e.to_string(), || "no_name".to_string());
tree.entry(parent_name).or_default().push(child_name.clone());
}
let hierarchy = to_json_string(&tree);
fs::write(
"bevy_hierarchy.json",
hierarchy
)
.expect("unable to write hierarchy file")
}
fs::write( fs::write(
"bevy_diagnostics.json", "bevy_diagnostics.json",
format!( format!(
"{{ \"animations\": {}, \"cylinder_found\": {} , \"nested_blueprint_found\": {}, \"empty_found\": {}, \"blueprints_list_found\": {} }}", "{{ \"animations\": {}, \"nested_blueprint_found\": {}, \"empty_found\": {}, \"blueprints_list_found\": {}, \"exported_names_correct\": {} }}",
animations_found, cylinder_found, nested_blueprint_found, empty_found, blueprints_list_found animations_found, nested_blueprint_found, empty_found, blueprints_list_found, exported_names_correct
), ),
) )
.expect("Unable to write file"); .expect("Unable to write file");
@ -97,6 +126,35 @@ fn exit_game(mut app_exit_events: ResMut<Events<bevy::app::AppExit>>) {
app_exit_events.send(bevy::app::AppExit); app_exit_events.send(bevy::app::AppExit);
} }
/*
#[derive(Resource)]
struct Animations(Vec<Handle<AnimationClip>>);
*/
fn animations(
foo:Query<(Entity, &Name, &AnimationPlayer),(Added<AnimationPlayer>)>,
asset_server: Res<AssetServer>,
mut commands: Commands,
) {
for bla in foo.iter() {
let mut counter = 0;
counter +=1;
println!("found some animations {} {}", counter, bla.1);
if bla.1.to_string() == "Collection".to_string(){
/*commands.insert_resource(Animations(vec![
asset_server.load("models/World.glb#Animation0"),
asset_server.load("models/World.glb#Animation1"),
]));*/
/*commands.entity(bla.0).insert(Animations {
named_animations:
})*/
}
}
}
pub struct GamePlugin; pub struct GamePlugin;
impl Plugin for GamePlugin { impl Plugin for GamePlugin {
fn build(&self, app: &mut App) { fn build(&self, app: &mut App) {
@ -104,6 +162,7 @@ impl Plugin for GamePlugin {
.add_systems(Update, validate_export) .add_systems(Update, validate_export)
.add_systems(OnEnter(AppState::MenuRunning), start_game) .add_systems(OnEnter(AppState::MenuRunning), start_game)
.add_systems(OnEnter(AppState::AppRunning), setup_game) .add_systems(OnEnter(AppState::AppRunning), setup_game)
.add_systems(Update, animations)
.add_systems(Update, generate_screenshot.run_if(on_timer(Duration::from_secs_f32(0.2)))) // TODO: run once .add_systems(Update, generate_screenshot.run_if(on_timer(Duration::from_secs_f32(0.2)))) // TODO: run once
.add_systems( .add_systems(
Update, Update,

View File

@ -0,0 +1,6 @@
import bpy
def get_standard_exporter_settings():
for scene in bpy.data.scenes:
if 'glTF2ExportSettings' in scene:
print("standard exporter settings", scene['glTF2ExportSettings'])

View File

@ -20,49 +20,41 @@ def remove_unwanted_custom_properties(object):
for component_name in object.keys(): for component_name in object.keys():
if not is_component_valid(object, component_name): if not is_component_valid(object, component_name):
to_remove.append(component_name) to_remove.append(component_name)
for cp in custom_properties_to_filter_out + to_remove: for cp in custom_properties_to_filter_out + to_remove:
if cp in object: if cp in object:
del object[cp] del object[cp]
def duplicate_object(object): # TODO: rename actions ?
obj_copy = object.copy() def copy_animation_data(source, target):
if object.data: """if source.data:
data = object.data.copy() data = source.data.copy()
obj_copy.data = data target.data = data"""
if object.animation_data and object.animation_data.action: if source.animation_data and source.animation_data:
obj_copy.animation_data.action = object.animation_data.action.copy() #print("copying animation data from", source.name, "to", target.name)
return obj_copy """print("I have animation data")
ad = source.animation_data
if ad.action:
print(source.name,'uses',ad.action.name)
for t in ad.nla_tracks:
for s in t.strips:
print(source.name,'uses',s.action.name)"""
#also removes unwanted custom_properties for all objects in hiearchy """if target.animation_data == None:
def duplicate_object_recursive(object, parent, collection): target.animation_data_create()
original_name = object.name target.animation_data.action = source.animation_data.action.copy()"""
object.name = original_name + "____bak" # alternative method, using the build in link animation operator
copy = duplicate_object(object) with bpy.context.temp_override(active_object=source, selected_editable_objects=[target]):
copy.name = original_name bpy.ops.object.make_links_data(type='ANIMATION')
collection.objects.link(copy) # we add an "animated" flag component
target['Animated'] = '()'
remove_unwanted_custom_properties(copy) """print("copying animation data for", source.name, target.animation_data)
properties = [p.identifier for p in source.animation_data.bl_rna.properties if not p.is_readonly]
if parent: for prop in properties:
copy.parent = parent print("copying stuff", prop)
setattr(target.animation_data, prop, getattr(source.animation_data, prop))"""
for child in object.children:
duplicate_object_recursive(child, copy, collection)
return copy
# copies the contents of a collection into another one while replacing library instances with empties
def copy_hollowed_collection_into(source_collection, destination_collection, parent_empty=None, filter=None, library_collections=[], addon_prefs={}):
collection_instances_combine_mode = getattr(addon_prefs, "collection_instances_combine_mode")
legacy_mode = getattr(addon_prefs, "export_legacy_mode")
collection_instances_combine_mode= collection_instances_combine_mode
for object in source_collection.objects:
if filter is not None and filter(object) is False:
continue
#check if a specific collection instance does not have an ovveride for combine_mode
combine_mode = object['_combine'] if '_combine' in object else collection_instances_combine_mode
def duplicate_object(object, parent, combine_mode, destination_collection, library_collections, legacy_mode, nester=""):
copy = None
if object.instance_type == 'COLLECTION' and (combine_mode == 'Split' or (combine_mode == 'EmbedExternal' and (object.instance_collection.name in library_collections)) ): if object.instance_type == 'COLLECTION' and (combine_mode == 'Split' or (combine_mode == 'EmbedExternal' and (object.instance_collection.name in library_collections)) ):
#print("creating empty for", object.name, object.instance_collection.name, library_collections, combine_mode) #print("creating empty for", object.name, object.instance_collection.name, library_collections, combine_mode)
collection_name = object.instance_collection.name collection_name = object.instance_collection.name
@ -79,28 +71,62 @@ def copy_hollowed_collection_into(source_collection, destination_collection, par
root_node = CollectionNode() root_node = CollectionNode()
root_node.name = "root" root_node.name = "root"
children_per_collection = {} children_per_collection = {}
print("collection stuff", original_name)
get_sub_collections([object.instance_collection], root_node, children_per_collection) get_sub_collections([object.instance_collection], root_node, children_per_collection)
empty_obj["BlueprintsList"] = f"({json.dumps(dict(children_per_collection))})" empty_obj["BlueprintsList"] = f"({json.dumps(dict(children_per_collection))})"
#empty_obj["Assets"] = {"Animations": [], "Materials": [], "Models":[], "Textures":[], "Audio":[], "Other":[]} #empty_obj["Assets"] = {"Animations": [], "Materials": [], "Models":[], "Textures":[], "Audio":[], "Other":[]}
# we copy custom properties over from our original object to our empty # we copy custom properties over from our original object to our empty
for component_name, component_value in object.items(): for component_name, component_value in object.items():
if component_name not in custom_properties_to_filter_out and is_component_valid(object, component_name): #copy only valid properties if component_name not in custom_properties_to_filter_out and is_component_valid(object, component_name): #copy only valid properties
empty_obj[component_name] = component_value empty_obj[component_name] = component_value
if parent_empty is not None: copy = empty_obj
empty_obj.parent = parent_empty
else: else:
# for objects which are NOT collection instances
# we create a copy of our object and its children, to leave the original one as it is # we create a copy of our object and its children, to leave the original one as it is
if object.parent == None: original_name = object.name
copy = duplicate_object_recursive(object, None, destination_collection) object.name = original_name + "____bak"
copy = object.copy()
copy.name = original_name
# FIXME: orphan data comes from this one, not even sure if this copying is needed at all
"""if object.data:
data = object.data.copy()
obj_copy.data = data"""
destination_collection.objects.link(copy)
"""if object.parent == None:
if parent_empty is not None: if parent_empty is not None:
copy.parent = parent_empty copy.parent = parent_empty
if object.animation_data:
copy['Animated'] = '()'"""
# for every sub-collection of the source, copy its content into a new sub-collection of the destination print(nester, "copy", copy)
# do this both for empty replacements & normal copies
if parent is not None:
copy.parent = parent
remove_unwanted_custom_properties(copy)
copy_animation_data(object, copy)
for child in object.children:
duplicate_object(child, copy, combine_mode, destination_collection, library_collections, legacy_mode, nester+" ")
# copies the contents of a collection into another one while replacing library instances with empties
def copy_hollowed_collection_into(source_collection, destination_collection, parent_empty=None, filter=None, library_collections=[], addon_prefs={}):
collection_instances_combine_mode = getattr(addon_prefs, "collection_instances_combine_mode")
legacy_mode = getattr(addon_prefs, "export_legacy_mode")
collection_instances_combine_mode= collection_instances_combine_mode
for object in source_collection.objects:
if object.name.endswith("____bak"): # some objects could already have been handled, ignore them
continue
if filter is not None and filter(object) is False:
continue
#check if a specific collection instance does not have an ovveride for combine_mode
combine_mode = object['_combine'] if '_combine' in object else collection_instances_combine_mode
parent = parent_empty
duplicate_object(object, parent, combine_mode, destination_collection, library_collections, legacy_mode)
# for every child-collection of the source, copy its content into a new sub-collection of the destination
for collection in source_collection.children: for collection in source_collection.children:
original_name = collection.name original_name = collection.name
collection.name = original_name + "____bak" collection.name = original_name + "____bak"
@ -108,7 +134,6 @@ def copy_hollowed_collection_into(source_collection, destination_collection, par
if parent_empty is not None: if parent_empty is not None:
collection_placeholder.parent = parent_empty collection_placeholder.parent = parent_empty
copy_hollowed_collection_into( copy_hollowed_collection_into(
source_collection = collection, source_collection = collection,
destination_collection = destination_collection, destination_collection = destination_collection,
@ -118,6 +143,8 @@ def copy_hollowed_collection_into(source_collection, destination_collection, par
addon_prefs=addon_prefs addon_prefs=addon_prefs
) )
return {} return {}
# clear & remove "hollow scene" # clear & remove "hollow scene"
@ -138,14 +165,14 @@ def clear_hollow_scene(temp_scene, original_root_collection):
# reset original names # reset original names
restore_original_names(original_root_collection) restore_original_names(original_root_collection)
# remove empties (only needed when we go via ops ????) # remove any data we created
temp_root_collection = temp_scene.collection temp_root_collection = temp_scene.collection
temp_scene_objects = [o for o in temp_root_collection.objects] temp_scene_objects = [o for o in temp_root_collection.all_objects]
for object in temp_scene_objects: for object in temp_scene_objects:
print("removing", object.name)
bpy.data.objects.remove(object, do_unlink=True) bpy.data.objects.remove(object, do_unlink=True)
# remove the temporary scene # remove the temporary scene
bpy.data.scenes.remove(temp_scene) bpy.data.scenes.remove(temp_scene, do_unlink=True)
# convenience utility to get lists of scenes # convenience utility to get lists of scenes
def get_scenes(addon_prefs): def get_scenes(addon_prefs):

View File

@ -0,0 +1 @@
{"b_Tail02_013":["b_Tail03_014"],"Blueprint4_nested.001":["Blueprint3"],"Collection 2 1":["Empty_in_sub_collection"],"b_Root_00":["b_Hip_01"],"b_LeftForeArm_010":["b_LeftHand_011"],"b_Spine01_02":["b_Spine02_03"],"Blueprint7_hierarchy.001":["Blueprint4_nested.001","Cube.001"],"b_RightLeg01_019":["b_RightLeg02_020"],"b_LeftUpperArm_09":["b_LeftForeArm_010"],"no_name":["Parent_Object","lighting_components_World","assets_list_World_components","Collection","Collection 2"],"Blueprint3":["Blueprint3_mesh","Blueprint3_mesh"],"world":["no_name"],"Parent_Object":["Cube.003","Blueprint1","Cylinder.001"],"Light":["Light","DirectionalLight Gizmo"],"Blueprint1.001":["Blueprint1_mesh"],"Blueprint7_hierarchy":["Cube.001"],"Spot":["Spot"],"b_Hip_01":["b_Spine01_02","b_Tail01_012","b_LeftLeg01_015","b_RightLeg01_019"],"Cylinder":["Cylinder.001","Cylinder.001"],"Collection 2":["Collection 2 1","Empty_in_collection","Spot"],"b_RightForeArm_07":["b_RightHand_08"],"Blueprint3_mesh":["Cylinder","Cylinder"],"Blueprint4_nested":["Blueprint3"],"Fox_mesh":["fox1"],"b_LeftLeg01_015":["b_LeftLeg02_016"],"b_Neck_04":["b_Head_05"],"b_RightFoot01_021":["b_RightFoot02_022"],"Blueprint1_mesh":["Cube.001","Cube.001"],"b_Tail01_012":["b_Tail02_013"],"Fox":["Fox_mesh","_rootJoint"],"Collection":["Blueprint1.001","Blueprint4_nested","Blueprint6_animated","Blueprint7_hierarchy","Camera","Cube","Empty","Light","Plane"],"Cube":["Cube"],"_rootJoint":["b_Root_00"],"b_RightLeg02_020":["b_RightFoot01_021"],"b_RightUpperArm_06":["b_RightForeArm_07"],"Plane":["Plane"],"Camera":["Camera Gizmo"],"Blueprint6_animated":["Fox"],"b_Spine02_03":["b_Neck_04","b_RightUpperArm_06","b_LeftUpperArm_09"],"b_LeftLeg02_016":["b_LeftFoot01_017"],"b_LeftFoot01_017":["b_LeftFoot02_018"],"Cube.001":["Cube.002","Cylinder","Cube.002","Cylinder"],"Cylinder.001":["Cylinder.002","Blueprint7_hierarchy.001","Empty_as_child"],"Blueprint1":["Blueprint1_mesh"]}

View File

@ -19,7 +19,6 @@ def setup_data(request):
def finalizer(): def finalizer():
print("\nPerforming teardown...") print("\nPerforming teardown...")
get_orphan_data()
if os.path.exists(models_path): if os.path.exists(models_path):
shutil.rmtree(models_path) shutil.rmtree(models_path)
@ -38,7 +37,10 @@ def setup_data(request):
def get_orphan_data(): def get_orphan_data():
orphan_meshes = [m.name for m in bpy.data.meshes if m.users == 0] orphan_meshes = [m.name for m in bpy.data.meshes if m.users == 0]
# print("orphan meshes before", orphan_meshes) orphan_objects = [m.name for m in bpy.data.objects if m.users == 0]
#print("orphan meshes before", orphan_meshes)
return orphan_meshes + orphan_objects
def test_export_do_not_export_blueprints(setup_data): def test_export_do_not_export_blueprints(setup_data):
auto_export_operator = bpy.ops.export_scenes.auto_gltf auto_export_operator = bpy.ops.export_scenes.auto_gltf
@ -61,6 +63,9 @@ def test_export_do_not_export_blueprints(setup_data):
) )
assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == False assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == False
orphan_data = get_orphan_data()
assert len(orphan_data) == 0
def test_export_custom_blueprints_path(setup_data): def test_export_custom_blueprints_path(setup_data):
auto_export_operator = bpy.ops.export_scenes.auto_gltf auto_export_operator = bpy.ops.export_scenes.auto_gltf
@ -83,6 +88,7 @@ def test_export_custom_blueprints_path(setup_data):
) )
assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "another_library_path", "Blueprint1.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "another_library_path", "Blueprint1.glb")) == True
assert len(get_orphan_data()) == 0
def test_export_materials_library(setup_data): def test_export_materials_library(setup_data):
auto_export_operator = bpy.ops.export_scenes.auto_gltf auto_export_operator = bpy.ops.export_scenes.auto_gltf
@ -107,7 +113,7 @@ def test_export_materials_library(setup_data):
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == True
assert os.path.exists(os.path.join(setup_data["materials_path"], "testing_materials_library.glb")) == True assert os.path.exists(os.path.join(setup_data["materials_path"], "testing_materials_library.glb")) == True
assert len(get_orphan_data()) == 0
def test_export_materials_library_custom_path(setup_data): def test_export_materials_library_custom_path(setup_data):
auto_export_operator = bpy.ops.export_scenes.auto_gltf auto_export_operator = bpy.ops.export_scenes.auto_gltf
@ -134,6 +140,7 @@ def test_export_materials_library_custom_path(setup_data):
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == True
assert os.path.exists(os.path.join(setup_data["materials_path"], "testing_materials_library.glb")) == False assert os.path.exists(os.path.join(setup_data["materials_path"], "testing_materials_library.glb")) == False
assert os.path.exists(os.path.join(setup_data["other_materials_path"], "testing_materials_library.glb")) == True assert os.path.exists(os.path.join(setup_data["other_materials_path"], "testing_materials_library.glb")) == True
assert len(get_orphan_data()) == 0
def test_export_collection_instances_combine_mode(setup_data): # TODO: change & check this def test_export_collection_instances_combine_mode(setup_data): # TODO: change & check this
auto_export_operator = bpy.ops.export_scenes.auto_gltf auto_export_operator = bpy.ops.export_scenes.auto_gltf
@ -160,6 +167,7 @@ def test_export_collection_instances_combine_mode(setup_data): # TODO: change &
assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "World_dynamic.glb")) == False assert os.path.exists(os.path.join(setup_data["models_path"], "World_dynamic.glb")) == False
assert len(get_orphan_data()) == 0
def test_export_do_not_export_marked_assets(setup_data): def test_export_do_not_export_marked_assets(setup_data):
@ -188,6 +196,7 @@ def test_export_do_not_export_marked_assets(setup_data):
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint3.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint3.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint4_nested.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint4_nested.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint5.glb")) == False assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint5.glb")) == False
assert len(get_orphan_data()) == 0
def test_export_separate_dynamic_and_static_objects(setup_data): def test_export_separate_dynamic_and_static_objects(setup_data):
@ -216,6 +225,7 @@ def test_export_separate_dynamic_and_static_objects(setup_data):
assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "World_dynamic.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World_dynamic.glb")) == True
assert len(get_orphan_data()) == 0
def test_export_should_not_generate_orphan_data(setup_data): def test_export_should_not_generate_orphan_data(setup_data):
@ -239,4 +249,5 @@ def test_export_should_not_generate_orphan_data(setup_data):
) )
assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True assert os.path.exists(os.path.join(setup_data["models_path"], "World.glb")) == True
assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == False assert os.path.exists(os.path.join(setup_data["models_path"], "library", "Blueprint1.glb")) == False
assert len(get_orphan_data()) == 0

View File

@ -5,6 +5,7 @@ import json
import pytest import pytest
import shutil import shutil
import filecmp
from PIL import Image from PIL import Image
from pixelmatch.contrib.PIL import pixelmatch from pixelmatch.contrib.PIL import pixelmatch
@ -56,7 +57,8 @@ def test_export_complex(setup_data):
# we use the global settings for that # we use the global settings for that
export_props = { export_props = {
"main_scene_names" : ['World'], "main_scene_names" : ['World'],
"library_scene_names": ['Library'] "library_scene_names": ['Library'],
# "export_format":'GLTF_SEPARATE'
} }
stored_settings = bpy.data.texts[".gltf_auto_export_settings"] if ".gltf_auto_export_settings" in bpy.data.texts else bpy.data.texts.new(".gltf_auto_export_settings") stored_settings = bpy.data.texts[".gltf_auto_export_settings"] if ".gltf_auto_export_settings" in bpy.data.texts else bpy.data.texts.new(".gltf_auto_export_settings")
stored_settings.clear() stored_settings.clear()
@ -107,9 +109,15 @@ def test_export_complex(setup_data):
diagnostics = json.load(diagnostics_file) diagnostics = json.load(diagnostics_file)
print("diagnostics", diagnostics) print("diagnostics", diagnostics)
assert diagnostics["animations"] == True assert diagnostics["animations"] == True
assert diagnostics["cylinder_found"] == True
assert diagnostics["empty_found"] == True assert diagnostics["empty_found"] == True
assert diagnostics["blueprints_list_found"] == True assert diagnostics["blueprints_list_found"] == True
assert diagnostics["exported_names_correct"] == True
with open(os.path.join(root_path, "bevy_hierarchy.json")) as hierarchy_file:
with open(os.path.join(os.path.dirname(__file__), "expected_bevy_hierarchy.json")) as expexted_hierarchy_file:
hierarchy = json.load(hierarchy_file)
expected = json.load(expexted_hierarchy_file)
assert sorted(hierarchy.items()) == sorted(expected.items())
# last but not least, do a visual compare # last but not least, do a visual compare
screenshot_expected_path = os.path.join(root_path, "expected_screenshot.png") screenshot_expected_path = os.path.join(root_path, "expected_screenshot.png")