main #2
@ -1 +0,0 @@
|
|||||||
# armorcore_bin
|
|
14
leenkx.py
14
leenkx.py
@ -182,7 +182,7 @@ class LeenkxAddonPreferences(AddonPreferences):
|
|||||||
description=(
|
description=(
|
||||||
"Whether to use OpenCL processing to generate radiance maps with CMFT."
|
"Whether to use OpenCL processing to generate radiance maps with CMFT."
|
||||||
" If you experience extremely long build times caused by CMFT, try disabling this option."
|
" If you experience extremely long build times caused by CMFT, try disabling this option."
|
||||||
" For more information see https://github.com/leenkx3d/leenkx/issues/2760"
|
" For more information see https://dev.leenkx.com/LeenkxTeam/LNXSDK/issues/2760"
|
||||||
))
|
))
|
||||||
legacy_shaders: BoolProperty(name="Legacy Shaders", description="Attempt to compile shaders runnable on older hardware, use this for WebGL1 or GLES2 support in mobile render path", default=False)
|
legacy_shaders: BoolProperty(name="Legacy Shaders", description="Attempt to compile shaders runnable on older hardware, use this for WebGL1 or GLES2 support in mobile render path", default=False)
|
||||||
relative_paths: BoolProperty(name="Generate Relative Paths", description="Write relative paths in khafile", default=False)
|
relative_paths: BoolProperty(name="Generate Relative Paths", description="Write relative paths in khafile", default=False)
|
||||||
@ -617,8 +617,6 @@ def try_os_call(func: Callable, *args, **kwargs) -> bool:
|
|||||||
|
|
||||||
|
|
||||||
def git_clone(done: Callable[[bool], None], rootdir: str, gitn: str, subdir: str, recursive=False):
|
def git_clone(done: Callable[[bool], None], rootdir: str, gitn: str, subdir: str, recursive=False):
|
||||||
print('Download SDK manually from https://leenkx.com...')
|
|
||||||
return False
|
|
||||||
rootdir = os.path.normpath(rootdir)
|
rootdir = os.path.normpath(rootdir)
|
||||||
path = rootdir + '/' + subdir if subdir != '' else rootdir
|
path = rootdir + '/' + subdir if subdir != '' else rootdir
|
||||||
|
|
||||||
@ -629,14 +627,12 @@ def git_clone(done: Callable[[bool], None], rootdir: str, gitn: str, subdir: str
|
|||||||
shutil.rmtree(path, onerror=remove_readonly)
|
shutil.rmtree(path, onerror=remove_readonly)
|
||||||
|
|
||||||
if recursive:
|
if recursive:
|
||||||
run_proc(['git', 'clone', '--recursive', 'https://github.com/' + gitn, path, '--depth', '1', '--shallow-submodules', '--jobs', '4'], done)
|
run_proc(['git', 'clone', '--recursive', 'https://dev.leenkx.com/' + gitn, path, '--depth', '1', '--shallow-submodules', '--jobs', '4'], done)
|
||||||
else:
|
else:
|
||||||
run_proc(['git', 'clone', 'https://github.com/' + gitn, path, '--depth', '1'], done)
|
run_proc(['git', 'clone', 'https://dev.leenkx.com/' + gitn, path, '--depth', '1'], done)
|
||||||
|
|
||||||
|
|
||||||
def git_test(self: bpy.types.Operator, required_version=None):
|
def git_test(self: bpy.types.Operator, required_version=None):
|
||||||
print('Download SDK manually from https://leenkx.com...')
|
|
||||||
return False
|
|
||||||
try:
|
try:
|
||||||
p = subprocess.Popen(['git', '--version'], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
p = subprocess.Popen(['git', '--version'], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
||||||
output, _ = p.communicate()
|
output, _ = p.communicate()
|
||||||
@ -789,7 +785,7 @@ def download_sdk(self: bpy.types.Operator, context):
|
|||||||
else:
|
else:
|
||||||
done(False)
|
done(False)
|
||||||
|
|
||||||
git_clone(done_clone, sdk_path, 'leenkx3d/lnxsdk', '', recursive=True)
|
git_clone(done_clone, sdk_path, 'LeenkxTeam/LNXSDK', '', recursive=True)
|
||||||
|
|
||||||
|
|
||||||
class LnxAddonRestoreButton(bpy.types.Operator):
|
class LnxAddonRestoreButton(bpy.types.Operator):
|
||||||
@ -812,7 +808,7 @@ class LnxAddonHelpButton(bpy.types.Operator):
|
|||||||
"""Updater help"""
|
"""Updater help"""
|
||||||
bl_idname = "lnx_addon.help"
|
bl_idname = "lnx_addon.help"
|
||||||
bl_label = "Help"
|
bl_label = "Help"
|
||||||
bl_description = "Leenkx Updater is currently disabled. Download SDK manually from https://leenkx.com"
|
bl_description = "Leenkx Updater to get the latest SDK from https://dev.leenkx.com/LeenkxTeam/LNXSDK"
|
||||||
|
|
||||||
def execute(self, context):
|
def execute(self, context):
|
||||||
webbrowser.open('https://leenkx.com/support')
|
webbrowser.open('https://leenkx.com/support')
|
||||||
|
@ -1,7 +0,0 @@
|
|||||||
# The zlib/libpng License
|
|
||||||
|
|
||||||
This software is provided 'as-is', without any express or implied warranty. In no event will the authors be held liable for any damages arising from the use of this software.
|
|
||||||
Permission is granted to anyone to use this software for any purpose, including commercial applications, and to alter it and redistribute it freely, subject to the following restrictions:
|
|
||||||
The origin of this software must not be misrepresented; you must not claim that you wrote the original software. If you use this software in a product, an acknowledgment in the product documentation would be appreciated but is not required.
|
|
||||||
Altered source versions must be plainly marked as such, and must not be misrepresented as being the original software.
|
|
||||||
This notice may not be removed or altered from any source distribution.
|
|
@ -1,8 +0,0 @@
|
|||||||
![](https://leenkx.com/git/iron.jpg)
|
|
||||||
|
|
||||||
iron
|
|
||||||
==============
|
|
||||||
|
|
||||||
[Wiki](https://github.com/leenkx3d/iron/wiki) - [Examples](https://github.com/leenkx3d/iron_examples)
|
|
||||||
|
|
||||||
Iron is a 3D engine core library built with [Kha](https://github.com/Kode/Kha) and [Haxe](https://github.com/HaxeFoundation/haxe). Powering [Leenkx](https://leenkx.com).
|
|
@ -1,7 +0,0 @@
|
|||||||
|
|
||||||
* 2019-11-04: `comma` key is now `,`, `period` key is now `.`
|
|
||||||
* 2018-08-28: `LampObject` and `LampData` is now `LightObject` and `LightData`
|
|
||||||
* 2018-08-28: Use `new Vec4(0,0,1)` (or `Vec4.zAxis()`) instead of `Vec4.up()`
|
|
||||||
* 2018-08-28: `Vec4.lerp` is no longer static, use `var v = new Vec4(); v.lerp(from, to, s)`
|
|
||||||
* 2018-08-21: `Vec4.lerp(to, from, s)` is now `Vec4.lerp(from, to, s)`
|
|
||||||
* 2017-12-10: Use `iron.math.Vec4.distance()` instead of `iron.math.Vec4.distance3d()`.
|
|
@ -1,253 +0,0 @@
|
|||||||
{
|
|
||||||
"defaultSeverity": "INFO",
|
|
||||||
"checks": [
|
|
||||||
{
|
|
||||||
"type": "CodeSimilarity"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "DefaultComesLast"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "DocCommentStyle"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "ERegLiteral"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"tokens": [
|
|
||||||
"CLASS_DEF",
|
|
||||||
"ENUM_DEF",
|
|
||||||
"ABSTRACT_DEF",
|
|
||||||
"TYPEDEF_DEF",
|
|
||||||
"INTERFACE_DEF",
|
|
||||||
"OBJECT_DECL",
|
|
||||||
"FUNCTION",
|
|
||||||
"FOR",
|
|
||||||
"IF",
|
|
||||||
"WHILE",
|
|
||||||
"SWITCH",
|
|
||||||
"TRY",
|
|
||||||
"CATCH"
|
|
||||||
],
|
|
||||||
"option": "empty"
|
|
||||||
},
|
|
||||||
"type": "EmptyBlock"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"max": 1
|
|
||||||
},
|
|
||||||
"type": "EmptyLines"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"option": "lowerCase"
|
|
||||||
},
|
|
||||||
"type": "HexadecimalLiteral"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "InnerAssignment"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"modifiers": [
|
|
||||||
"MACRO",
|
|
||||||
"OVERRIDE",
|
|
||||||
"PUBLIC_PRIVATE",
|
|
||||||
"STATIC",
|
|
||||||
"INLINE",
|
|
||||||
"DYNAMIC"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"type": "ModifierOrder"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "MultipleVariableDeclarations"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"allowSingleLineStatement": true,
|
|
||||||
"tokens": [
|
|
||||||
"FOR",
|
|
||||||
"IF",
|
|
||||||
"ELSE_IF",
|
|
||||||
"WHILE",
|
|
||||||
"DO_WHILE"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"type": "NeedBraces"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"assignOpPolicy": "around",
|
|
||||||
"unaryOpPolicy": "none",
|
|
||||||
"ternaryOpPolicy": "around",
|
|
||||||
"arithmeticOpPolicy": "around",
|
|
||||||
"compareOpPolicy": "around",
|
|
||||||
"bitwiseOpPolicy": "around",
|
|
||||||
"boolOpPolicy": "around",
|
|
||||||
"intervalOpPolicy": "none",
|
|
||||||
"arrowPolicy": "none",
|
|
||||||
"oldFunctionTypePolicy": "none",
|
|
||||||
"newFunctionTypePolicy": "none",
|
|
||||||
"arrowFunctionPolicy": "around"
|
|
||||||
},
|
|
||||||
"type": "OperatorWhitespace"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"tokens": [
|
|
||||||
"=",
|
|
||||||
"*",
|
|
||||||
"/",
|
|
||||||
"%",
|
|
||||||
">",
|
|
||||||
"<",
|
|
||||||
">=",
|
|
||||||
"<=",
|
|
||||||
"==",
|
|
||||||
"!=",
|
|
||||||
"&",
|
|
||||||
"|",
|
|
||||||
"^",
|
|
||||||
"<<",
|
|
||||||
">>",
|
|
||||||
">>>",
|
|
||||||
"+=",
|
|
||||||
"-=",
|
|
||||||
"*=",
|
|
||||||
"/=",
|
|
||||||
"%=",
|
|
||||||
"<<=",
|
|
||||||
">>=",
|
|
||||||
">>>=",
|
|
||||||
"|=",
|
|
||||||
"&=",
|
|
||||||
"^=",
|
|
||||||
"...",
|
|
||||||
"=>",
|
|
||||||
"++",
|
|
||||||
"--",
|
|
||||||
"+",
|
|
||||||
"-",
|
|
||||||
"&&",
|
|
||||||
"||"
|
|
||||||
],
|
|
||||||
"option": "eol"
|
|
||||||
},
|
|
||||||
"type": "OperatorWrap"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "RedundantModifier"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "RedundantAllowMeta"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "RedundantAccessMeta"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"allowEmptyReturn": true,
|
|
||||||
"enforceReturnType": false
|
|
||||||
},
|
|
||||||
"type": "Return"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"dotPolicy": "none",
|
|
||||||
"commaPolicy": "after",
|
|
||||||
"semicolonPolicy": "after"
|
|
||||||
},
|
|
||||||
"type": "SeparatorWhitespace"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"tokens": [
|
|
||||||
","
|
|
||||||
],
|
|
||||||
"option": "eol"
|
|
||||||
},
|
|
||||||
"type": "SeparatorWrap"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"spaceIfCondition": "should",
|
|
||||||
"spaceAroundBinop": true,
|
|
||||||
"spaceForLoop": "should",
|
|
||||||
"ignoreRangeOperator": true,
|
|
||||||
"spaceWhileLoop": "should",
|
|
||||||
"spaceCatch": "should",
|
|
||||||
"spaceSwitchCase": "should",
|
|
||||||
"noSpaceAroundUnop": true
|
|
||||||
},
|
|
||||||
"type": "Spacing"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"allowException": true,
|
|
||||||
"policy": "doubleAndInterpolation"
|
|
||||||
},
|
|
||||||
"type": "StringLiteral"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "TrailingWhitespace"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "UnusedImport"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "UnusedLocalVar"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"tokens": [
|
|
||||||
",",
|
|
||||||
";",
|
|
||||||
":"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"type": "WhitespaceAfter"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"props": {
|
|
||||||
"tokens": [
|
|
||||||
"=",
|
|
||||||
"+",
|
|
||||||
"-",
|
|
||||||
"*",
|
|
||||||
"/",
|
|
||||||
"%",
|
|
||||||
">",
|
|
||||||
"<",
|
|
||||||
">=",
|
|
||||||
"<=",
|
|
||||||
"==",
|
|
||||||
"!=",
|
|
||||||
"&",
|
|
||||||
"|",
|
|
||||||
"^",
|
|
||||||
"&&",
|
|
||||||
"||",
|
|
||||||
"<<",
|
|
||||||
">>",
|
|
||||||
">>>",
|
|
||||||
"+=",
|
|
||||||
"-=",
|
|
||||||
"*=",
|
|
||||||
"/=",
|
|
||||||
"%=",
|
|
||||||
"<<=",
|
|
||||||
">>=",
|
|
||||||
">>>=",
|
|
||||||
"|=",
|
|
||||||
"&=",
|
|
||||||
"^=",
|
|
||||||
"=>"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"type": "WhitespaceAround"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
@ -1,861 +0,0 @@
|
|||||||
"""Leenkx Mesh Exporter"""
|
|
||||||
#
|
|
||||||
# https://leenkx.com/
|
|
||||||
#
|
|
||||||
# Based on Open Game Engine Exchange
|
|
||||||
# https://opengex.org/
|
|
||||||
# Export plugin for Blender by Eric Lengyel
|
|
||||||
# Copyright 2015, Terathon Software LLC
|
|
||||||
#
|
|
||||||
# This software is licensed under the Creative Commons
|
|
||||||
# Attribution-ShareAlike 3.0 Unported License:
|
|
||||||
# http://creativecommons.org/licenses/by-sa/3.0/deed.en_US
|
|
||||||
|
|
||||||
import io
|
|
||||||
import os
|
|
||||||
import struct
|
|
||||||
import time
|
|
||||||
import bpy
|
|
||||||
from bpy_extras.io_utils import ExportHelper
|
|
||||||
from mathutils import Vector
|
|
||||||
import numpy as np
|
|
||||||
|
|
||||||
bl_info = {
|
|
||||||
"name": "Leenkx Mesh Exporter",
|
|
||||||
"category": "Import-Export",
|
|
||||||
"location": "File -> Export",
|
|
||||||
"description": "Leenkx mesh data",
|
|
||||||
"author": "Leenkx3D.org",
|
|
||||||
"version": (2022, 12, 0),
|
|
||||||
"blender": (3, 3, 0),
|
|
||||||
"doc_url": "https://github.com/leenkx3d/iron/wiki",
|
|
||||||
"tracker_url": "https://github.com/leenkx3d/iron/issues",
|
|
||||||
}
|
|
||||||
|
|
||||||
NodeTypeBone = 1
|
|
||||||
NodeTypeMesh = 2
|
|
||||||
structIdentifier = ["object", "bone_object", "mesh_object"]
|
|
||||||
|
|
||||||
|
|
||||||
class LeenkxExporter(bpy.types.Operator, ExportHelper):
|
|
||||||
"""Export to Leenkx format"""
|
|
||||||
|
|
||||||
bl_idname = "export_scene.lnx"
|
|
||||||
bl_label = "Export Leenkx"
|
|
||||||
filename_ext = ".lnx"
|
|
||||||
|
|
||||||
def execute(self, context):
|
|
||||||
profile_time = time.time()
|
|
||||||
current_frame = context.scene.frame_current
|
|
||||||
current_subframe = context.scene.frame_subframe
|
|
||||||
self.scene = context.scene
|
|
||||||
self.output = {}
|
|
||||||
self.bobjectArray = {}
|
|
||||||
self.bobjectBoneArray = {}
|
|
||||||
self.meshArray = {}
|
|
||||||
self.boneParentArray = {}
|
|
||||||
self.bone_tracks = []
|
|
||||||
self.depsgraph = context.evaluated_depsgraph_get()
|
|
||||||
scene_objects = self.scene.collection.all_objects
|
|
||||||
|
|
||||||
for bobject in scene_objects:
|
|
||||||
if not bobject.parent:
|
|
||||||
self.process_bobject(bobject)
|
|
||||||
|
|
||||||
self.process_skinned_meshes()
|
|
||||||
|
|
||||||
self.output["name"] = self.scene.name
|
|
||||||
self.output["objects"] = []
|
|
||||||
for bo in scene_objects:
|
|
||||||
if not bo.parent:
|
|
||||||
self.export_object(bo, self.scene)
|
|
||||||
|
|
||||||
self.output["mesh_datas"] = []
|
|
||||||
for o in self.meshArray.items():
|
|
||||||
self.export_mesh(o)
|
|
||||||
|
|
||||||
self.write_lnx(self.filepath, self.output)
|
|
||||||
self.scene.frame_set(current_frame, subframe=current_subframe)
|
|
||||||
|
|
||||||
print(f"Scene exported in {str(time.time() - profile_time)}")
|
|
||||||
|
|
||||||
return {"FINISHED"}
|
|
||||||
|
|
||||||
def write_lnx(self, filepath, output):
|
|
||||||
with open(filepath, "wb") as f:
|
|
||||||
f.write(packb(output))
|
|
||||||
|
|
||||||
def write_matrix(self, matrix):
|
|
||||||
return [
|
|
||||||
matrix[0][0],
|
|
||||||
matrix[0][1],
|
|
||||||
matrix[0][2],
|
|
||||||
matrix[0][3],
|
|
||||||
matrix[1][0],
|
|
||||||
matrix[1][1],
|
|
||||||
matrix[1][2],
|
|
||||||
matrix[1][3],
|
|
||||||
matrix[2][0],
|
|
||||||
matrix[2][1],
|
|
||||||
matrix[2][2],
|
|
||||||
matrix[2][3],
|
|
||||||
matrix[3][0],
|
|
||||||
matrix[3][1],
|
|
||||||
matrix[3][2],
|
|
||||||
matrix[3][3],
|
|
||||||
]
|
|
||||||
|
|
||||||
def find_bone(self, name):
|
|
||||||
return next(
|
|
||||||
(
|
|
||||||
bobject_ref
|
|
||||||
for bobject_ref in self.bobjectBoneArray.items()
|
|
||||||
if bobject_ref[0].name == name
|
|
||||||
),
|
|
||||||
None,
|
|
||||||
)
|
|
||||||
|
|
||||||
def collect_bone_animation(self, armature, name):
|
|
||||||
path = 'pose.bones["' + name + '"].'
|
|
||||||
curve_array = []
|
|
||||||
if armature.animation_data:
|
|
||||||
if action := armature.animation_data.action:
|
|
||||||
curve_array.extend(
|
|
||||||
fcurve
|
|
||||||
for fcurve in action.fcurves
|
|
||||||
if fcurve.data_path.startswith(path)
|
|
||||||
)
|
|
||||||
return curve_array
|
|
||||||
|
|
||||||
def export_bone(self, armature, bone, scene, o, action):
|
|
||||||
if bobjectRef := self.bobjectBoneArray.get(bone):
|
|
||||||
o["type"] = structIdentifier[bobjectRef["objectType"]]
|
|
||||||
o["name"] = bobjectRef["structName"]
|
|
||||||
self.export_bone_transform(armature, bone, o, action)
|
|
||||||
o["children"] = []
|
|
||||||
for subbobject in bone.children:
|
|
||||||
so = {}
|
|
||||||
self.export_bone(armature, subbobject, scene, so, action)
|
|
||||||
o["children"].append(so)
|
|
||||||
|
|
||||||
def export_pose_markers(self, oanim, action):
|
|
||||||
if action.pose_markers is None or len(action.pose_markers) == 0:
|
|
||||||
return
|
|
||||||
oanim["marker_frames"] = []
|
|
||||||
oanim["marker_names"] = []
|
|
||||||
for m in action.pose_markers:
|
|
||||||
oanim["marker_frames"].append(int(m.frame))
|
|
||||||
oanim["marker_names"].append(m.name)
|
|
||||||
|
|
||||||
def process_bone(self, bone):
|
|
||||||
self.bobjectBoneArray[bone] = {
|
|
||||||
"objectType": NodeTypeBone,
|
|
||||||
"structName": bone.name,
|
|
||||||
}
|
|
||||||
for subbobject in bone.children:
|
|
||||||
self.process_bone(subbobject)
|
|
||||||
|
|
||||||
def process_bobject(self, bobject):
|
|
||||||
if bobject.type not in ["MESH", "ARMATURE"]:
|
|
||||||
return
|
|
||||||
|
|
||||||
btype = NodeTypeMesh if bobject.type == "MESH" else 0
|
|
||||||
self.bobjectArray[bobject] = {"objectType": btype, "structName": bobject.name}
|
|
||||||
|
|
||||||
if bobject.type == "ARMATURE":
|
|
||||||
if skeleton := bobject.data:
|
|
||||||
for bone in skeleton.bones:
|
|
||||||
if not bone.parent:
|
|
||||||
self.process_bone(bone)
|
|
||||||
|
|
||||||
for subbobject in bobject.children:
|
|
||||||
self.process_bobject(subbobject)
|
|
||||||
|
|
||||||
def process_skinned_meshes(self):
|
|
||||||
for bobjectRef in self.bobjectArray.items():
|
|
||||||
if bobjectRef[1]["objectType"] == NodeTypeMesh:
|
|
||||||
if armature := bobjectRef[0].find_armature():
|
|
||||||
for bone in armature.data.bones:
|
|
||||||
boneRef = self.find_bone(bone.name)
|
|
||||||
if boneRef:
|
|
||||||
boneRef[1]["objectType"] = NodeTypeBone
|
|
||||||
|
|
||||||
def export_bone_transform(self, armature, bone, o, action):
|
|
||||||
pose_bone = armature.pose.bones.get(bone.name)
|
|
||||||
transform = bone.matrix_local.copy()
|
|
||||||
if bone.parent is not None:
|
|
||||||
transform = bone.parent.matrix_local.inverted_safe() @ transform
|
|
||||||
|
|
||||||
o["transform"] = {}
|
|
||||||
o["transform"]["values"] = self.write_matrix(transform)
|
|
||||||
|
|
||||||
curve_array = self.collect_bone_animation(armature, bone.name)
|
|
||||||
animation = len(curve_array) != 0
|
|
||||||
|
|
||||||
if animation and pose_bone:
|
|
||||||
begin_frame = int(action.frame_range[0])
|
|
||||||
end_frame = int(action.frame_range[1])
|
|
||||||
tracko = {}
|
|
||||||
o["anim"] = {"tracks": [tracko]}
|
|
||||||
tracko["target"] = "transform"
|
|
||||||
tracko["frames"] = [
|
|
||||||
i - begin_frame for i in range(begin_frame, end_frame + 1)
|
|
||||||
]
|
|
||||||
tracko["values"] = []
|
|
||||||
self.bone_tracks.append((tracko["values"], pose_bone))
|
|
||||||
|
|
||||||
def write_bone_matrices(self, scene, action):
|
|
||||||
if len(self.bone_tracks) > 0:
|
|
||||||
begin_frame = int(action.frame_range[0])
|
|
||||||
end_frame = int(action.frame_range[1])
|
|
||||||
for i in range(begin_frame, end_frame + 1):
|
|
||||||
scene.frame_set(i)
|
|
||||||
for track in self.bone_tracks:
|
|
||||||
values, pose_bone = track[0], track[1]
|
|
||||||
if parent := pose_bone.parent:
|
|
||||||
values += self.write_matrix(
|
|
||||||
(parent.matrix.inverted_safe() @ pose_bone.matrix)
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
values += self.write_matrix(pose_bone.matrix)
|
|
||||||
|
|
||||||
def export_object(self, bobject, scene, parento=None):
|
|
||||||
if bobjectRef := self.bobjectArray.get(bobject):
|
|
||||||
o = {
|
|
||||||
"type": structIdentifier[bobjectRef["objectType"]],
|
|
||||||
"name": bobjectRef["structName"],
|
|
||||||
}
|
|
||||||
if bobject.parent_type == "BONE":
|
|
||||||
o["parent_bone"] = bobject.parent_bone
|
|
||||||
|
|
||||||
if bobjectRef["objectType"] == NodeTypeMesh:
|
|
||||||
objref = bobject.data
|
|
||||||
if objref not in self.meshArray:
|
|
||||||
self.meshArray[objref] = {
|
|
||||||
"structName": objref.name,
|
|
||||||
"objectTable": [bobject],
|
|
||||||
}
|
|
||||||
else:
|
|
||||||
self.meshArray[objref]["objectTable"].append(bobject)
|
|
||||||
oid = self.meshArray[objref]["structName"]
|
|
||||||
o["data_ref"] = oid
|
|
||||||
o["dimensions"] = self.calc_aabb(bobject)
|
|
||||||
|
|
||||||
o["transform"] = {}
|
|
||||||
o["transform"]["values"] = self.write_matrix(bobject.matrix_local)
|
|
||||||
|
|
||||||
# If the object is parented to a bone and is not relative, undo the
|
|
||||||
# bone's transform
|
|
||||||
if bobject.parent_type == "BONE":
|
|
||||||
armature = bobject.parent.data
|
|
||||||
bone = armature.bones[bobject.parent_bone]
|
|
||||||
o["parent_bone_connected"] = bone.use_connect
|
|
||||||
if bone.use_connect:
|
|
||||||
bone_translation = Vector((0, bone.length, 0)) + bone.head
|
|
||||||
o["parent_bone_tail"] = [
|
|
||||||
bone_translation[0],
|
|
||||||
bone_translation[1],
|
|
||||||
bone_translation[2],
|
|
||||||
]
|
|
||||||
else:
|
|
||||||
bone_translation = bone.tail - bone.head
|
|
||||||
o["parent_bone_tail"] = [
|
|
||||||
bone_translation[0],
|
|
||||||
bone_translation[1],
|
|
||||||
bone_translation[2],
|
|
||||||
]
|
|
||||||
pose_bone = bobject.parent.pose.bones[bobject.parent_bone]
|
|
||||||
bone_translation_pose = pose_bone.tail - pose_bone.head
|
|
||||||
o["parent_bone_tail_pose"] = [
|
|
||||||
bone_translation_pose[0],
|
|
||||||
bone_translation_pose[1],
|
|
||||||
bone_translation_pose[2],
|
|
||||||
]
|
|
||||||
|
|
||||||
if bobject.type == "ARMATURE" and bobject.data is not None:
|
|
||||||
bdata = bobject.data
|
|
||||||
action = None
|
|
||||||
adata = bobject.animation_data
|
|
||||||
|
|
||||||
# Active action
|
|
||||||
if adata is not None:
|
|
||||||
action = adata.action
|
|
||||||
if action is None:
|
|
||||||
bobject.animation_data_create()
|
|
||||||
actions = bpy.data.actions
|
|
||||||
action = actions.get("leenkx_pose")
|
|
||||||
if action is None:
|
|
||||||
action = actions.new(name="leenkx_pose")
|
|
||||||
|
|
||||||
# Collect export actions
|
|
||||||
export_actions = [action]
|
|
||||||
if hasattr(adata, "nla_tracks") and adata.nla_tracks is not None:
|
|
||||||
for track in adata.nla_tracks:
|
|
||||||
if track.strips is None:
|
|
||||||
continue
|
|
||||||
for strip in track.strips:
|
|
||||||
if strip.action is None:
|
|
||||||
continue
|
|
||||||
if strip.action.name == action.name:
|
|
||||||
continue
|
|
||||||
export_actions.append(strip.action)
|
|
||||||
|
|
||||||
basename = os.path.basename(self.filepath)[:-4]
|
|
||||||
o["bone_actions"] = []
|
|
||||||
for action in export_actions:
|
|
||||||
o["bone_actions"].append(basename + "_" + action.name)
|
|
||||||
|
|
||||||
orig_action = bobject.animation_data.action
|
|
||||||
for action in export_actions:
|
|
||||||
bobject.animation_data.action = action
|
|
||||||
bones = []
|
|
||||||
self.bone_tracks = []
|
|
||||||
for bone in bdata.bones:
|
|
||||||
if not bone.parent:
|
|
||||||
boneo = {}
|
|
||||||
self.export_bone(bobject, bone, scene, boneo, action)
|
|
||||||
bones.append(boneo)
|
|
||||||
self.write_bone_matrices(scene, action)
|
|
||||||
if len(bones) > 0 and "anim" in bones[0]:
|
|
||||||
self.export_pose_markers(bones[0]["anim"], action)
|
|
||||||
# Save action separately
|
|
||||||
action_obj = {}
|
|
||||||
action_obj["name"] = action.name
|
|
||||||
action_obj["objects"] = bones
|
|
||||||
self.write_lnx(
|
|
||||||
self.filepath[:-4] + "_" + action.name + ".lnx", action_obj
|
|
||||||
)
|
|
||||||
|
|
||||||
bobject.animation_data.action = orig_action
|
|
||||||
|
|
||||||
if parento is None:
|
|
||||||
self.output["objects"].append(o)
|
|
||||||
else:
|
|
||||||
parento["children"].append(o)
|
|
||||||
|
|
||||||
if not hasattr(o, "children") and len(bobject.children) > 0:
|
|
||||||
o["children"] = []
|
|
||||||
|
|
||||||
for subbobject in bobject.children:
|
|
||||||
self.export_object(subbobject, scene, o)
|
|
||||||
|
|
||||||
def export_skin(self, bobject, armature, exportMesh, o):
|
|
||||||
# This function exports all skinning data, which includes the skeleton
|
|
||||||
# and per-vertex bone influence data
|
|
||||||
oskin = {}
|
|
||||||
o["skin"] = oskin
|
|
||||||
|
|
||||||
# Write the skin bind pose transform
|
|
||||||
otrans = {}
|
|
||||||
oskin["transform"] = otrans
|
|
||||||
otrans["values"] = self.write_matrix(bobject.matrix_world)
|
|
||||||
|
|
||||||
bone_array = armature.data.bones
|
|
||||||
bone_count = len(bone_array)
|
|
||||||
max_bones = 128
|
|
||||||
bone_count = min(bone_count, max_bones)
|
|
||||||
|
|
||||||
# Write the bone object reference array
|
|
||||||
oskin["bone_ref_array"] = np.empty(bone_count, dtype=object)
|
|
||||||
oskin["bone_len_array"] = np.empty(bone_count, dtype="<f4")
|
|
||||||
|
|
||||||
for i in range(bone_count):
|
|
||||||
if boneRef := self.find_bone(bone_array[i].name):
|
|
||||||
oskin["bone_ref_array"][i] = boneRef[1]["structName"]
|
|
||||||
oskin["bone_len_array"][i] = bone_array[i].length
|
|
||||||
else:
|
|
||||||
oskin["bone_ref_array"][i] = ""
|
|
||||||
oskin["bone_len_array"][i] = 0.0
|
|
||||||
|
|
||||||
# Write the bind pose transform array
|
|
||||||
oskin["transformsI"] = []
|
|
||||||
for i in range(bone_count):
|
|
||||||
skeletonI = (
|
|
||||||
armature.matrix_world @ bone_array[i].matrix_local
|
|
||||||
).inverted_safe()
|
|
||||||
skeletonI = skeletonI @ bobject.matrix_world
|
|
||||||
oskin["transformsI"].append(self.write_matrix(skeletonI))
|
|
||||||
|
|
||||||
# Export the per-vertex bone influence data
|
|
||||||
group_remap = []
|
|
||||||
for group in bobject.vertex_groups:
|
|
||||||
for i in range(bone_count):
|
|
||||||
if bone_array[i].name == group.name:
|
|
||||||
group_remap.append(i)
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
group_remap.append(-1)
|
|
||||||
|
|
||||||
bone_count_array = np.empty(len(exportMesh.loops), dtype="<i2")
|
|
||||||
bone_index_array = np.empty(len(exportMesh.loops) * 4, dtype="<i2")
|
|
||||||
bone_weight_array = np.empty(len(exportMesh.loops) * 4, dtype="<f4")
|
|
||||||
|
|
||||||
vertices = bobject.data.vertices
|
|
||||||
count = 0
|
|
||||||
for index, l in enumerate(exportMesh.loops):
|
|
||||||
bone_count = 0
|
|
||||||
total_weight = 0.0
|
|
||||||
bone_values = []
|
|
||||||
for g in vertices[l.vertex_index].groups:
|
|
||||||
bone_index = group_remap[g.group]
|
|
||||||
bone_weight = g.weight
|
|
||||||
if bone_index >= 0: # and bone_weight != 0.0:
|
|
||||||
bone_values.append((bone_weight, bone_index))
|
|
||||||
total_weight += bone_weight
|
|
||||||
bone_count += 1
|
|
||||||
|
|
||||||
if bone_count > 4:
|
|
||||||
bone_count = 4
|
|
||||||
bone_values.sort(reverse=True)
|
|
||||||
bone_values = bone_values[:4]
|
|
||||||
|
|
||||||
bone_count_array[index] = bone_count
|
|
||||||
for bv in bone_values:
|
|
||||||
bone_weight_array[count] = bv[0]
|
|
||||||
bone_index_array[count] = bv[1]
|
|
||||||
count += 1
|
|
||||||
|
|
||||||
if total_weight not in (0.0, 1.0):
|
|
||||||
normalizer = 1.0 / total_weight
|
|
||||||
for i in range(bone_count):
|
|
||||||
bone_weight_array[count - i - 1] *= normalizer
|
|
||||||
|
|
||||||
bone_index_array = bone_index_array[:count]
|
|
||||||
bone_weight_array = bone_weight_array[:count]
|
|
||||||
bone_weight_array *= 32767
|
|
||||||
bone_weight_array = np.array(bone_weight_array, dtype="<i2")
|
|
||||||
|
|
||||||
oskin["bone_count_array"] = bone_count_array
|
|
||||||
oskin["bone_index_array"] = bone_index_array
|
|
||||||
oskin["bone_weight_array"] = bone_weight_array
|
|
||||||
|
|
||||||
def calc_aabb(self, bobject):
|
|
||||||
aabb_center = 0.125 * sum((Vector(b) for b in bobject.bound_box), Vector())
|
|
||||||
return [
|
|
||||||
abs(
|
|
||||||
(bobject.bound_box[6][0] - bobject.bound_box[0][0]) / 2
|
|
||||||
+ abs(aabb_center[0])
|
|
||||||
)
|
|
||||||
* 2,
|
|
||||||
abs(
|
|
||||||
(bobject.bound_box[6][1] - bobject.bound_box[0][1]) / 2
|
|
||||||
+ abs(aabb_center[1])
|
|
||||||
)
|
|
||||||
* 2,
|
|
||||||
abs(
|
|
||||||
(bobject.bound_box[6][2] - bobject.bound_box[0][2]) / 2
|
|
||||||
+ abs(aabb_center[2])
|
|
||||||
)
|
|
||||||
* 2,
|
|
||||||
]
|
|
||||||
|
|
||||||
def export_mesh_data(self, exportMesh, bobject, o, has_armature=False):
|
|
||||||
exportMesh.calc_normals_split()
|
|
||||||
exportMesh.calc_loop_triangles()
|
|
||||||
|
|
||||||
loops = exportMesh.loops
|
|
||||||
num_verts = len(loops)
|
|
||||||
num_uv_layers = len(exportMesh.uv_layers)
|
|
||||||
num_colors = len(exportMesh.vertex_colors)
|
|
||||||
has_tex = num_uv_layers > 0
|
|
||||||
has_tex1 = num_uv_layers > 1
|
|
||||||
has_col = num_colors > 0
|
|
||||||
has_tang = False
|
|
||||||
|
|
||||||
pdata = np.empty(num_verts * 4, dtype="<f4") # p.xyz, n.z
|
|
||||||
ndata = np.empty(num_verts * 2, dtype="<f4") # n.xy
|
|
||||||
if has_tex:
|
|
||||||
t0map = 0 # Get active uvmap
|
|
||||||
t0data = np.empty(num_verts * 2, dtype="<f4")
|
|
||||||
uv_layers = exportMesh.uv_layers
|
|
||||||
if uv_layers is not None:
|
|
||||||
for i in range(0, len(uv_layers)):
|
|
||||||
if uv_layers[i].active_render:
|
|
||||||
t0map = i
|
|
||||||
break
|
|
||||||
if has_tex1:
|
|
||||||
t1map = 1 if t0map == 0 else 0
|
|
||||||
t1data = np.empty(num_verts * 2, dtype="<f4")
|
|
||||||
# Scale for packed coords
|
|
||||||
maxdim = 1.0
|
|
||||||
lay0 = uv_layers[t0map]
|
|
||||||
for v in lay0.data:
|
|
||||||
if abs(v.uv[0]) > maxdim:
|
|
||||||
maxdim = abs(v.uv[0])
|
|
||||||
if abs(v.uv[1]) > maxdim:
|
|
||||||
maxdim = abs(v.uv[1])
|
|
||||||
if has_tex1:
|
|
||||||
lay1 = uv_layers[t1map]
|
|
||||||
for v in lay1.data:
|
|
||||||
if abs(v.uv[0]) > maxdim:
|
|
||||||
maxdim = abs(v.uv[0])
|
|
||||||
if abs(v.uv[1]) > maxdim:
|
|
||||||
maxdim = abs(v.uv[1])
|
|
||||||
if maxdim > 1:
|
|
||||||
o["scale_tex"] = maxdim
|
|
||||||
invscale_tex = (1 / o["scale_tex"]) * 32767
|
|
||||||
else:
|
|
||||||
invscale_tex = 1 * 32767
|
|
||||||
if has_tang:
|
|
||||||
exportMesh.calc_tangents(uvmap=lay0.name)
|
|
||||||
tangdata = np.empty(num_verts * 3, dtype="<f4")
|
|
||||||
if has_col:
|
|
||||||
cdata = np.empty(num_verts * 3, dtype="<f4")
|
|
||||||
|
|
||||||
# Scale for packed coords
|
|
||||||
aabb = self.calc_aabb(bobject)
|
|
||||||
maxdim = max(aabb[0], max(aabb[1], aabb[2]))
|
|
||||||
if maxdim > 2:
|
|
||||||
o["scale_pos"] = maxdim / 2
|
|
||||||
else:
|
|
||||||
o["scale_pos"] = 1.0
|
|
||||||
if has_armature: # Allow up to 2x bigger bounds for skinned mesh
|
|
||||||
o["scale_pos"] *= 2.0
|
|
||||||
|
|
||||||
scale_pos = o["scale_pos"]
|
|
||||||
invscale_pos = (1 / scale_pos) * 32767
|
|
||||||
|
|
||||||
verts = exportMesh.vertices
|
|
||||||
if has_tex:
|
|
||||||
lay0 = exportMesh.uv_layers[t0map]
|
|
||||||
if has_tex1:
|
|
||||||
lay1 = exportMesh.uv_layers[t1map]
|
|
||||||
if has_col:
|
|
||||||
vcol0 = exportMesh.vertex_colors[0].data
|
|
||||||
|
|
||||||
for i, loop in enumerate(loops):
|
|
||||||
v = verts[loop.vertex_index]
|
|
||||||
co = v.co
|
|
||||||
normal = loop.normal
|
|
||||||
tang = loop.tangent
|
|
||||||
|
|
||||||
i4 = i * 4
|
|
||||||
i2 = i * 2
|
|
||||||
pdata[i4] = co[0]
|
|
||||||
pdata[i4 + 1] = co[1]
|
|
||||||
pdata[i4 + 2] = co[2]
|
|
||||||
pdata[i4 + 3] = normal[2] * scale_pos # Cancel scale
|
|
||||||
ndata[i2] = normal[0]
|
|
||||||
ndata[i2 + 1] = normal[1]
|
|
||||||
if has_tex:
|
|
||||||
uv = lay0.data[loop.index].uv
|
|
||||||
t0data[i2] = uv[0]
|
|
||||||
t0data[i2 + 1] = 1.0 - uv[1] # Reverse Y
|
|
||||||
if has_tex1:
|
|
||||||
uv = lay1.data[loop.index].uv
|
|
||||||
t1data[i2] = uv[0]
|
|
||||||
t1data[i2 + 1] = 1.0 - uv[1]
|
|
||||||
if has_tang:
|
|
||||||
i3 = i * 3
|
|
||||||
tangdata[i3] = tang[0]
|
|
||||||
tangdata[i3 + 1] = tang[1]
|
|
||||||
tangdata[i3 + 2] = tang[2]
|
|
||||||
if has_col:
|
|
||||||
col = vcol0[loop.index].color
|
|
||||||
i3 = i * 3
|
|
||||||
cdata[i3] = col[0]
|
|
||||||
cdata[i3 + 1] = col[1]
|
|
||||||
cdata[i3 + 2] = col[2]
|
|
||||||
|
|
||||||
mats = exportMesh.materials
|
|
||||||
poly_map = []
|
|
||||||
for i in range(max(len(mats), 1)):
|
|
||||||
poly_map.append([])
|
|
||||||
for poly in exportMesh.polygons:
|
|
||||||
poly_map[poly.material_index].append(poly)
|
|
||||||
|
|
||||||
o["index_arrays"] = []
|
|
||||||
|
|
||||||
# map polygon indices to triangle loops
|
|
||||||
tri_loops = {}
|
|
||||||
for loop in exportMesh.loop_triangles:
|
|
||||||
if loop.polygon_index not in tri_loops:
|
|
||||||
tri_loops[loop.polygon_index] = []
|
|
||||||
tri_loops[loop.polygon_index].append(loop)
|
|
||||||
|
|
||||||
for index, polys in enumerate(poly_map):
|
|
||||||
tris = 0
|
|
||||||
for poly in polys:
|
|
||||||
tris += poly.loop_total - 2
|
|
||||||
if tris == 0: # No face assigned
|
|
||||||
continue
|
|
||||||
prim = np.empty(tris * 3, dtype="<i4")
|
|
||||||
|
|
||||||
i = 0
|
|
||||||
for poly in polys:
|
|
||||||
for loop in tri_loops[poly.index]:
|
|
||||||
prim[i] = loops[loop.loops[0]].index
|
|
||||||
prim[i + 1] = loops[loop.loops[1]].index
|
|
||||||
prim[i + 2] = loops[loop.loops[2]].index
|
|
||||||
i += 3
|
|
||||||
|
|
||||||
ia = {}
|
|
||||||
ia["values"] = prim
|
|
||||||
ia["material"] = 0
|
|
||||||
if len(mats) > 1:
|
|
||||||
for i in range(len(mats)): # Multi-mat mesh
|
|
||||||
if mats[i] == mats[index]: # Default material for empty slots
|
|
||||||
ia["material"] = i
|
|
||||||
break
|
|
||||||
o["index_arrays"].append(ia)
|
|
||||||
|
|
||||||
# Pack
|
|
||||||
pdata *= invscale_pos
|
|
||||||
ndata *= 32767
|
|
||||||
pdata = np.array(pdata, dtype="<i2")
|
|
||||||
ndata = np.array(ndata, dtype="<i2")
|
|
||||||
if has_tex:
|
|
||||||
t0data *= invscale_tex
|
|
||||||
t0data = np.array(t0data, dtype="<i2")
|
|
||||||
if has_tex1:
|
|
||||||
t1data *= invscale_tex
|
|
||||||
t1data = np.array(t1data, dtype="<i2")
|
|
||||||
if has_col:
|
|
||||||
cdata *= 32767
|
|
||||||
cdata = np.array(cdata, dtype="<i2")
|
|
||||||
if has_tang:
|
|
||||||
tangdata *= 32767
|
|
||||||
tangdata = np.array(tangdata, dtype="<i2")
|
|
||||||
|
|
||||||
# Output
|
|
||||||
o["vertex_arrays"] = []
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{"attrib": "pos", "values": pdata, "data": "short4norm"}
|
|
||||||
)
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{"attrib": "nor", "values": ndata, "data": "short2norm"}
|
|
||||||
)
|
|
||||||
if has_tex:
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{"attrib": "tex", "values": t0data, "data": "short2norm"}
|
|
||||||
)
|
|
||||||
if has_tex1:
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{"attrib": "tex1", "values": t1data, "data": "short2norm"}
|
|
||||||
)
|
|
||||||
if has_col:
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{"attrib": "col", "values": cdata, "data": "short4norm", "padding": 1}
|
|
||||||
)
|
|
||||||
if has_tang:
|
|
||||||
o["vertex_arrays"].append(
|
|
||||||
{
|
|
||||||
"attrib": "tang",
|
|
||||||
"values": tangdata,
|
|
||||||
"data": "short4norm",
|
|
||||||
"padding": 1,
|
|
||||||
}
|
|
||||||
)
|
|
||||||
|
|
||||||
def export_mesh(self, objectRef):
|
|
||||||
# This function exports a single mesh object
|
|
||||||
table = objectRef[1]["objectTable"]
|
|
||||||
bobject = table[0]
|
|
||||||
oid = objectRef[1]["structName"]
|
|
||||||
o = {}
|
|
||||||
o["name"] = oid
|
|
||||||
|
|
||||||
armature = bobject.find_armature()
|
|
||||||
apply_modifiers = not armature
|
|
||||||
|
|
||||||
bobject_eval = (
|
|
||||||
bobject.evaluated_get(self.depsgraph) if apply_modifiers else bobject
|
|
||||||
)
|
|
||||||
exportMesh = bobject_eval.to_mesh()
|
|
||||||
|
|
||||||
self.export_mesh_data(exportMesh, bobject, o, has_armature=armature is not None)
|
|
||||||
if armature:
|
|
||||||
self.export_skin(bobject, armature, exportMesh, o)
|
|
||||||
|
|
||||||
self.output["mesh_datas"].append(o)
|
|
||||||
bobject_eval.to_mesh_clear()
|
|
||||||
|
|
||||||
|
|
||||||
def menu_func(self, context):
|
|
||||||
self.layout.operator(LeenkxExporter.bl_idname, text="Leenkx (.lnx)")
|
|
||||||
|
|
||||||
|
|
||||||
def register():
|
|
||||||
bpy.utils.register_class(LeenkxExporter)
|
|
||||||
bpy.types.TOPBAR_MT_file_export.append(menu_func)
|
|
||||||
|
|
||||||
|
|
||||||
def unregister():
|
|
||||||
bpy.types.TOPBAR_MT_file_export.remove(menu_func)
|
|
||||||
bpy.utils.unregister_class(LeenkxExporter)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
register()
|
|
||||||
|
|
||||||
# Msgpack parser with typed arrays
|
|
||||||
# Based on u-msgpack-python v2.4.1 - v at sergeev.io
|
|
||||||
# https://github.com/vsergeev/u-msgpack-python
|
|
||||||
#
|
|
||||||
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
||||||
# of this software and associated documentation files (the "Software"), to deal
|
|
||||||
# in the Software without restriction, including without limitation the rights
|
|
||||||
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
||||||
# copies of the Software, and to permit persons to whom the Software is
|
|
||||||
# furnished to do so, subject to the following conditions:
|
|
||||||
#
|
|
||||||
# The above copyright notice and this permission notice shall be included in
|
|
||||||
# all copies or substantial portions of the Software.
|
|
||||||
#
|
|
||||||
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
||||||
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
||||||
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
||||||
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
||||||
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
||||||
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
||||||
# THE SOFTWARE.
|
|
||||||
#
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_integer(obj, fp):
|
|
||||||
if obj < 0:
|
|
||||||
if obj >= -32:
|
|
||||||
fp.write(struct.pack("b", obj))
|
|
||||||
elif obj >= -(2 ** (8 - 1)):
|
|
||||||
fp.write(b"\xd0" + struct.pack("b", obj))
|
|
||||||
elif obj >= -(2 ** (16 - 1)):
|
|
||||||
fp.write(b"\xd1" + struct.pack("<h", obj))
|
|
||||||
elif obj >= -(2 ** (32 - 1)):
|
|
||||||
fp.write(b"\xd2" + struct.pack("<i", obj))
|
|
||||||
elif obj >= -(2 ** (64 - 1)):
|
|
||||||
fp.write(b"\xd3" + struct.pack("<q", obj))
|
|
||||||
else:
|
|
||||||
raise Exception("huge signed int")
|
|
||||||
else:
|
|
||||||
if obj <= 127:
|
|
||||||
fp.write(struct.pack("B", obj))
|
|
||||||
elif obj <= 2**8 - 1:
|
|
||||||
fp.write(b"\xcc" + struct.pack("B", obj))
|
|
||||||
elif obj <= 2**16 - 1:
|
|
||||||
fp.write(b"\xcd" + struct.pack("<H", obj))
|
|
||||||
elif obj <= 2**32 - 1:
|
|
||||||
fp.write(b"\xce" + struct.pack("<I", obj))
|
|
||||||
elif obj <= 2**64 - 1:
|
|
||||||
fp.write(b"\xcf" + struct.pack("<Q", obj))
|
|
||||||
else:
|
|
||||||
raise Exception("huge unsigned int")
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_nil(fp):
|
|
||||||
fp.write(b"\xc0")
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_boolean(obj, fp):
|
|
||||||
fp.write(b"\xc3" if obj else b"\xc2")
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_float(obj, fp):
|
|
||||||
# NOTE: forced 32-bit floats for Leenkx
|
|
||||||
# fp.write(b"\xcb" + struct.pack("<d", obj)) # Double
|
|
||||||
fp.write(b"\xca" + struct.pack("<f", obj))
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_string(obj, fp):
|
|
||||||
obj = obj.encode("utf-8")
|
|
||||||
if len(obj) <= 31:
|
|
||||||
fp.write(struct.pack("B", 0xA0 | len(obj)) + obj)
|
|
||||||
elif len(obj) <= 2**8 - 1:
|
|
||||||
fp.write(b"\xd9" + struct.pack("B", len(obj)) + obj)
|
|
||||||
elif len(obj) <= 2**16 - 1:
|
|
||||||
fp.write(b"\xda" + struct.pack("<H", len(obj)) + obj)
|
|
||||||
elif len(obj) <= 2**32 - 1:
|
|
||||||
fp.write(b"\xdb" + struct.pack("<I", len(obj)) + obj)
|
|
||||||
else:
|
|
||||||
raise Exception("huge string")
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_binary(obj, fp):
|
|
||||||
if len(obj) <= 2**8 - 1:
|
|
||||||
fp.write(b"\xc4" + struct.pack("B", len(obj)) + obj)
|
|
||||||
elif len(obj) <= 2**16 - 1:
|
|
||||||
fp.write(b"\xc5" + struct.pack("<H", len(obj)) + obj)
|
|
||||||
elif len(obj) <= 2**32 - 1:
|
|
||||||
fp.write(b"\xc6" + struct.pack("<I", len(obj)) + obj)
|
|
||||||
else:
|
|
||||||
raise Exception("huge binary string")
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_array(obj, fp):
|
|
||||||
if len(obj) <= 15:
|
|
||||||
fp.write(struct.pack("B", 0x90 | len(obj)))
|
|
||||||
elif len(obj) <= 2**16 - 1:
|
|
||||||
fp.write(b"\xdc" + struct.pack("<H", len(obj)))
|
|
||||||
elif len(obj) <= 2**32 - 1:
|
|
||||||
fp.write(b"\xdd" + struct.pack("<I", len(obj)))
|
|
||||||
else:
|
|
||||||
raise Exception("huge array")
|
|
||||||
|
|
||||||
if len(obj) > 0 and isinstance(obj[0], float):
|
|
||||||
fp.write(b"\xca")
|
|
||||||
for e in obj:
|
|
||||||
fp.write(struct.pack("<f", e))
|
|
||||||
elif len(obj) > 0 and isinstance(obj[0], bool):
|
|
||||||
for e in obj:
|
|
||||||
pack(e, fp)
|
|
||||||
elif len(obj) > 0 and isinstance(obj[0], int):
|
|
||||||
fp.write(b"\xd2")
|
|
||||||
for e in obj:
|
|
||||||
fp.write(struct.pack("<i", e))
|
|
||||||
# Float32
|
|
||||||
elif len(obj) > 0 and isinstance(obj[0], np.float32):
|
|
||||||
fp.write(b"\xca")
|
|
||||||
fp.write(obj.tobytes())
|
|
||||||
# Int32
|
|
||||||
elif len(obj) > 0 and isinstance(obj[0], np.int32):
|
|
||||||
fp.write(b"\xd2")
|
|
||||||
fp.write(obj.tobytes())
|
|
||||||
# Int16
|
|
||||||
elif len(obj) > 0 and isinstance(obj[0], np.int16):
|
|
||||||
fp.write(b"\xd1")
|
|
||||||
fp.write(obj.tobytes())
|
|
||||||
# Regular
|
|
||||||
else:
|
|
||||||
for e in obj:
|
|
||||||
pack(e, fp)
|
|
||||||
|
|
||||||
|
|
||||||
def _pack_map(obj, fp):
|
|
||||||
if len(obj) <= 15:
|
|
||||||
fp.write(struct.pack("B", 0x80 | len(obj)))
|
|
||||||
elif len(obj) <= 2**16 - 1:
|
|
||||||
fp.write(b"\xde" + struct.pack("<H", len(obj)))
|
|
||||||
elif len(obj) <= 2**32 - 1:
|
|
||||||
fp.write(b"\xdf" + struct.pack("<I", len(obj)))
|
|
||||||
else:
|
|
||||||
raise Exception("huge array")
|
|
||||||
|
|
||||||
for k, v in obj.items():
|
|
||||||
pack(k, fp)
|
|
||||||
pack(v, fp)
|
|
||||||
|
|
||||||
|
|
||||||
def pack(obj, fp):
|
|
||||||
if obj is None:
|
|
||||||
_pack_nil(fp)
|
|
||||||
elif isinstance(obj, bool):
|
|
||||||
_pack_boolean(obj, fp)
|
|
||||||
elif isinstance(obj, int):
|
|
||||||
_pack_integer(obj, fp)
|
|
||||||
elif isinstance(obj, float):
|
|
||||||
_pack_float(obj, fp)
|
|
||||||
elif isinstance(obj, str):
|
|
||||||
_pack_string(obj, fp)
|
|
||||||
elif isinstance(obj, bytes):
|
|
||||||
_pack_binary(obj, fp)
|
|
||||||
elif isinstance(obj, (list, np.ndarray, tuple)):
|
|
||||||
_pack_array(obj, fp)
|
|
||||||
elif isinstance(obj, dict):
|
|
||||||
_pack_map(obj, fp)
|
|
||||||
else:
|
|
||||||
raise Exception(f"unsupported type: {str(type(obj))}")
|
|
||||||
|
|
||||||
|
|
||||||
def packb(obj):
|
|
||||||
fp = io.BytesIO()
|
|
||||||
pack(obj, fp)
|
|
||||||
return fp.getvalue()
|
|
@ -1,7 +1,6 @@
|
|||||||
package leenkx.logicnode;
|
package leenkx.logicnode;
|
||||||
import iron.data.Data;
|
import iron.data.Data;
|
||||||
import aura.Aura;
|
import aura.Aura;
|
||||||
import aura.Assets;
|
|
||||||
|
|
||||||
class AudioLoadNode extends LogicNode {
|
class AudioLoadNode extends LogicNode {
|
||||||
public var property1: String;
|
public var property1: String;
|
||||||
@ -25,53 +24,18 @@ class AudioLoadNode extends LogicNode {
|
|||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|
||||||
//var sound = aura.Assets.Sound(name, Uncompress);
|
//var loadConfig: AuraLoadConfig = {
|
||||||
//var anotherSound = aura.Assets.Sound("AnotherSoundFile", KeepCompressed);
|
// uncompressed: [name],
|
||||||
//var hrtf = aura.Assets.HRTF("myHRTF_mhr");
|
// compressed: [name],
|
||||||
|
// hrtf: ["myHRTF_mhr"],
|
||||||
|
//};
|
||||||
|
|
||||||
|
|
||||||
Data.getSound(name, function (data) {
|
Data.getSound(name, function (data) {
|
||||||
//audio = Aura.createUncompBufferChannel(data,inputs[2].get(),Aura.mixChannels[inputs[0].get()]);
|
audio = Aura.createUncompBufferChannel(data,inputs[2].get(),Aura.mixChannels[inputs[0].get()]);
|
||||||
var assetList = [
|
|
||||||
data,
|
|
||||||
//anotherSound,
|
|
||||||
//hrtf,
|
|
||||||
];
|
|
||||||
|
|
||||||
aura.Assets.startLoading(assetList,
|
|
||||||
|
|
||||||
(asset: aura.Assets.Asset, numLoaded: Int, numTotalAssets: Int) -> {
|
|
||||||
trace('Loaded $numLoadedAssets of $totalNumAssets: ${asset.name}');
|
|
||||||
audio = Aura.createUncompBufferChannel(asset,inputs[2].get(),Aura.mixChannels[inputs[0].get()]);
|
|
||||||
if (numLoaded == totalNumAssets) {
|
|
||||||
trace("Loaded all assets");
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
(asset: aura.Assets.Asset, error: kha.AssetError) -> {
|
|
||||||
trace('Failed to load asset ${asset.name}. Reason: $error');
|
|
||||||
return AbortLoading;
|
|
||||||
}
|
|
||||||
);
|
|
||||||
|
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
//Data.getSound(name, function (data) {
|
|
||||||
// audio = Aura.createUncompBufferChannel(data,inputs[2].get(),Aura.mixChannels[inputs[0].get()]);
|
|
||||||
//});
|
|
||||||
|
|
||||||
return audio;
|
return audio;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
In addition to the above change:
|
|
||||||
- `aura.types.HRTF` was renamed to `aura.types.HRTFData`
|
|
||||||
- `aura.dsp.panner.HRTFPanner.new()` now expects an `aura.Assets.HRTF` object instead of an `aura.types.HRTFData` object as its second parameter
|
|
||||||
- `Aura.getSound()` now returns `Null<aura.Assets.Sound>` instead of `Null<kha.Sound>`
|
|
||||||
- `Aura.getHRTF()` now returns `Null<aura.Assets.HRTF>` instead of `Null<aura.types.HRTFData>`
|
|
||||||
- `Aura.createUncompBufferChannel()` and `Aura.createCompBufferChannel()` now take an `aura.Assets.Sound` as their first parameter instead of a `kha.Sound`
|
|
Loading…
x
Reference in New Issue
Block a user