Using the Rigify animation rig

From OniGalore
Jump to navigation Jump to search

Introduction

This article is intended to explain the workflow of using the Rigify-based animation rig for Blender (made by geyser) for the purpose of making Oni animations, and also how to use it, and what are the Oni-specific additions to it that we've made.

Prerequisite tutorials

I highly recommend watching the tutorials below if you feel you're lacking knowledge in any of these subjects:

General Blender basics:


Practical and intuitive explanation of what Inverse and Forward Kinematics (IK/FK) are - this is explained in Maya, but same principles apply in every 3D editing program including Blender:


Basics of using Rigify - all you need to know about animating with Rigify is stored in the video below. There is no point in making a tutorial on how to use the rig controllers themselves here - it's all in the video below:


(OPTIONAL) Tutorial on how to create the rig described here - this is strictly optional; the only practical reason to make your own rig is for the purpose of adjusting its controllers to a specific character, however, the advantage that you get out of that is purely visual and a marginal one at that - if you've built your rig for Konoko, there's nothing stopping you from using it on a TCTF SWAT - the limb controllers will be only slightly out of place, which doesn't impact animating in any way. So ultimately you may want to go through this if you want to learn how this rig works, find this rig confusing, or even just for the bunch of great Youtube tutorials linked in that tutorial.

Also, it's worth noting that a huge part of this article is in fact copy-pasted from this tutorial.

Required tools and commands

There are several things worth knowing before you start creating the rig, tools to get and tutorials to have ready to reference for commands.

Tools and relevant tutorials

Screenshot showing Cmder's capabilities within the context of modding Oni
  • Current version of OniSplit. This is the tool needed to import and export assets out of Oni. DO NOT USE OniSplit GUI or Vago for importing Oni assets into Blender - neither of these weren't updated in a long time, and thus they don't support OniSplit's v0.9.99.2 -blender option. You can still use them for other purposes though, such as sounds and converting .oni files to XMLs, etc.
  • Cmder (Windows only) - because OniSplit is a command line tool, it is highly recommended to get any upgrade to Windows' Command Prompt. As shown in the screenshot on the right, Cmder allows you to start it from the context menu in any selected folder, and it also remembers your most recently used commands, vastly improving your workflow when you're forced to use any command line tools.
  • Rigify - it's a plugin for Blender designed to automate a lot of rigging work. The rig described here is generated using Rigify.
  • Oni-Blender tutorial by EdT. Please read this in entirety to know likely-to-happen issues and refer to this as your guide for OniSplit commands relating to Blender.
  • Brief overview on creating TRAMs by EdT - while this was written with XSI in mind, this is still relevant as the process for preparing the XML files for Oni is still the same. Also the next post in that thread, called Brief walk through on modifying a TRAM, is an example of that overview put into practice.

General workflow of animating for Oni using Rigify

The general workflow of making animations using this rig is the following, assuming you want to make a single animation (i.e. you don't want to make a throw animation):

  1. Open up a Blender scene containing the rig.
  2. Delete either the model in Pose1 or Pose2.
  3. (OPTIONAL) Assuming you want to animate a different character, delete the T-posed model.
    1. Import your desired character with textures using -noanim.
  4. Import the animation you are using as a starting point. For more detailed information on how to import an Oni animation to Blender, see the next header.
    1. If you did point 3, apply the textures on the animated model.
  5. Using the RigBoneConstraintTargetSetter script available HERE, set the targets of the rig's bone constraints to be the body parts of the animation you've imported.
  6. Using the Pose Matching functionality, constrain the rig to the imported animation (Set the IK-FK sliders on rig limbs to 1, as Pose Matching works only through FK controllers)
  7. For each frame that you want to remain in your animation, while having all the bones selected (or the ones you want), press Ctrl+A, select Apply Visual Transform to Pose and keyframe it.
  8. Disable bone constraints on the rig using the BoneConstraintEnabler script available [Blender#Script_for_enabling.2Fdisabling_bone_constraints_in_the_selected_armature HERE.]
  9. Make your animation.
  10. Once your animation is ready, use the Visual Transformer script available HERE, to bake the rig keyframes into the character model.
  11. Using the ObjectConstraintEnabler script available HERE, disable the object constraints on the character model. The character model should now be animated and have the needed rotations and locations.
  12. Export the animation as a DAE.

Importing Oni animations to Blender

This header describes how to import Oni animations to Blender.

For detailed explanation of the required OniSplit commands, please refer to the Oni-Blender Tutorial by EdT listed in the Tools and relevant tutorials.

  1. Assuming the character you want to animate isn't in your rig file, then using OniSplit, export any character you want as a DAE (-extract:dae) using -noanim and -blender arguments.
    1. As per EdT's Oni-Blender Tutorial, you should get AnimationDaeWriter: custom axis conversion in OniSplit output if you've used the -blender argument. If you didn't get that on the output, it means something most likely went wrong and you won't be able to import the model into Blender (or you will be able to import it but it will be wrong).
    2. Assuming you wanted a textured model and thus you've exported an ONCC, you should now get a DAE file and an images folder containing the textures for it.
  2. Using OniSplit, export the animation you want as an XML (-extract:xml) using -anim-body (lets you specify the character you want) and -blender arguments.
    1. You should get one DAE and one XML file for the animation.
  3. Open up a Blender scene containing the rig.
  4. Delete either the model either in Pose1 or Pose2 collection.
  5. Import the animation into Blender (MAKE SURE YOU CHECK THE Import Unit BOX EACH TIME, otherwise you will import the model with arbitrary units which will break everything and will be basically unadjustable later))
  6. Import the -noanim model into Blender (ALSO MAKE SURE YOU CHECK THE Import Unit BOX)
  7. Move the animated model to either Pose1 or Pose2 collection,.
  8. If you've imported the textured model, apply textures to the animated models and fix the alpha transparency issue.

Added functionalities

This header describes functionalities added to the rig by geyser that are not normally present in Rigify.

Pose matching

The way we do animations is we almost universally start by copying the first or last frame of the preceeding animation. For this reason, we needed functionality that would allow us to snap the rig to an animated Oni character. The way it works, it that it uses bone constraints in rig controller bones targeting the character model in order to snap the rig to the model. The influence of the bone constraints is controlled through the Z location of Pose1 and Pose2 bones in the Pose Matching layer, which is done through drivers.

To snap the rig to the animation in one of the Pose collections, simply move the appropriate Pose bone above the XY plane while in Pose mode.

Pose1 and Pose2 below the XY plane Rig snapped to the animation in Pose 1 collection Rig snapped to the animation in Pose 2 collection
Unposed.png
Pose1.png
Pose2.png

Wrist snapping

Often when using the IK handles for the arms you will find out that most of the time, you will need the hand to have the same rotation as the forearm. Normally this would mean that you would have to rotate the hands all the time, because IK handles are mainly moved around, and rotated later. For this reason, the wrist snapping functionality was implemented - you can simply copy the rotation of the forearm to the hand by pressing the "IK wrist" button.

Both of the hands after just moving them around... Effects of using the IK Wrist button
IK Wrist Example1.png
IK Wrist Example2.png