COMPUTER ANIMATION AND VIRTUAL WORLDS Comp. Anim. Virtual Worlds 2004; 15: 259–268 (DOI: 10.1002/cav.28) * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Tai Chi synthesizer: a motion synthesis framework based on key-postures and motion instructions By Shih-Pin Chao*, Chih-Yi Chiu, Shi-Nine Yang and Tsang-Gang Lin * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Human spontaneous motions such as walking and runing are seldom described in detail. However, some well designed motions such as conditioning and rehabilitation exercises are usually described in detail and edited into exercise manuals for training purpose. This paper presents a novel framework for synthesizing new motions based on given motion manuals and corresponding motion capture examples. First, using text analysis method, a set of basic motion texts is extracted from exercise manuals and converted into text normal form. Then we introduce an annotation method to build the mapping table between basic motion texts and their corresponding motion clips in the given examples. To facilitate the search for proper motion clips, a multi-dimensional index structure based on posture parameters is proposed. Then a new motion with given textual description can be synthesized by converting the description sentences into a sequence of text normal forms and then concatenating the corresponding motion clips to form the desired motion. Moreover, to realize the smooth concatenation, we show that it can be achieved by finding an appropriate path in the proposed multi-dimensional index space. Several experimental examples are given to demonstrate the proposed method is effective in synthesizing desired motions according to given descriptions. Copyright # 2004 John Wiley & Sons, Ltd. KEY WORDS:

animation from motion/video data; animation of articulated figures: animation system; animation techniques; motion capture and retargeting; language parsing and understanding

Introduction and Related Work Many exercise instruction manuals describe human exercises in texts and depict key-posture pictures. For example, Tai Chi Chuan (TCC), a traditional Chinese martial art, has been handed down by texts augmented with drawings or pictures for several hundred years. Additionally, TCC is a conditioning exercise. Many studies1–3 found that TCC practitioners had significantly better postural control and a reduced risk of falls. This study is an attempt to propose a framework (See Figure 1) for synthesizing complex human exercises based on high-level motion instructions, i.e. texts and pictures. It is motivated by the request of physicians

*Correspondence to: Shin-Pin Chao, Department of Computer Science, National Tsing Hua University, Hsinchu, 300 Taiwan. E-mail: [email protected]

looking for a text-based 3D animation authoring tool to develop a special rehabilitation workout for patients suffering different degrees of disability. It is known that TCC has been developed for several hundred years and it owns a great volume of texts and pictures to specify TCC exercise. Therefore, we use TCC to demonstrate the effectiveness of the proposed framework. In computer animation, Badler et al.4 have introduced an architecture, NL2PAR, to allow avatar control through natural language instructions. Cassell et al.5 developed a text-to-speech system, BEAT, which can output synthesized speech with appropriate and synchronized nonverbal behaviors. In fact, these systems are procedure-based, therefore the synthesized motions are less realistic. Besides, people are so used to see human motion that any artifact in the synthesized human motion can be easily noticed. Recently, the motion capture method has been notable for generating very realistic motions. It records the detail and nuance of live motion from human directly. However, motion capture

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

S.-P. CHAO ET AL. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Figure 1. Framework. (1) Motion database; (2) BMT-normal form directory; (3) User Interfaces; (4) Motion synthesis. process is labor-intensive and time-consuming. Besides, if the existing captured data does not fit the new requirements, then there will be many efforts for further modifications. This drawback hinders the motion capture method from being popular. To overcome this difficulty, many approaches have been proposed, for examples, motion signal processing6,7 and constraintbased motion editing.8,9 Most of these techniques deal primarily with modifying specific motion clips, not the generation of a continuous stream of motion. Although, several methods have recently been proposed for synthesizing continuous streams of motion,10 the connection between high-level motion control and lowlevel mathematical models developed by these systems is unclear.11 Besides, they either suffer from long preprocessing time or produce unexpected transitions.11 For consideration of high-level motion control, Kim et al.12 propose a rhythm-driven motion synthesis scheme. Given an input sound signal, their scheme online synthesizes a novel motion by traversing the motion transition graph, which is synchronized with the input sound signal and also satisfies kinematic constraints given explicitly and implicitly. However, their scheme only applies to rhythmic exercises such as dances and aerobics. There are motions such as TCC that are inherently not driven and synchronized by music or beats. Arikan et al.13 presented a framework that allows a user to synthesize lifelike human motion capture data by painting a timeline with annotated actions such as running, picking up, walking, etc. Among these examples, the transition between actions can occur at any point according to a timeline specification. However, it is hard for TCC to specify a transition at an arbitrary time, since all TCC basic motions have to be performed completely and their durations are different. In general,

a common approach for writing exercise instructions in texts is through key-posturing, that is, to describe the body movements from key-posture to key-posture. For brevity, a set of sentences involves only two keypostures, namely, the start-posture and the end-posture is called a basic motion text (BMT). Based on this observation, we propose a framework to synthesize new motions based on given instruction manuals and motion examples. First, a set of affine invariant posture parameters is proposed. Based on proposed posture parameters, a motion index table (MIT) is introduced to segment given examples into indexed motion clips (Moclips). Then, we show that the MIT can speed up the example search during motion annotation and facilitate motion synthesis process afterwards. Next, we use text analysis method to build a directory of normal forms for BMTs (BMT-normal form directory) from given exercise manuals. Then through dynamic time warping (DTW)14 which is a similarity comparison algorithm, Moclips corresponding to these BMTs are identified and extracted from given examples. Then, a new motion with given textual description can be synthesized by converting the given sentences into BMT-normal forms in sequence and concatenating the corresponding Moclips. Moreover, we show that a smooth motion concatenations of two clips can be realized by finding an appropriate path between the end-posture index of the former clip and the start-posture index of the latter clip in multidimensional index space. In order to specify keyposture from pictures, which are often provided in exercise manuals, a graphical user interface based on key-posture pictures is provided. This 2D to 3D reconstruction algorithm is based on Taylor’s paper.15 The organization of the proposed framework is shown as in Figure 1. It is composed of four

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

260

Comp. Anim. Virtual Worlds 2004; 15: 259–268

TAI CHI SYNTHESIZER * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

components: (1) Motion database, (2) BMT-normal form directory, (3) User interfaces, and (4) Motion synthesis. Based on this framework, the paper is structured as follows. In the next section, we propose a set of posture parameters to construct MIT. The third section describes the building of BMT-normal form directory. The fourth section introduces user interfaces with text-based and image-based key-posture specification to facilitate highlevel motion control. Then the fifth section proposes a path finding algorithm through MIT to satisfy specified constraints. Several experimental examples are given in the sixth section. The final section provides a conclusion and future work.

Construction of Motion Index Table This section introduces an efficient index function for constructing motion database. Since raw materials of our motion annotation and motion synthesis are from given motion examples, finding appropriate motion data efficiently is essential. Moreover, motion transition is reliable only if the motions are similar and the posture similarity is an important factor in considering motion similarity.10,11,13,16 In addition, basic motions can be characterized by their key-postures. Therefore, we propose a set of posture parameters for indexing a posture and roughly segmenting motion data. First, we assume that a motion example is a continuous long stream of 3D motion captured by a motion capture device. In other words, it is a sequence of motion frames such that each frame can be regarded as a 3D posture. In this paper, we use motion frame or its corresponding posture interchangeably. The basic notion of motion index table construction is to segment the long sequence of motion into short motion clips (Moclips) according to the variation of limb postures. For simplicity, we assume two postures are equivalent if there is an affine transformation between them. Now, we introduce a set of affine invariant posture parameters. Assume that an arm (leg) is composed of the lower limb and the upper limb of a posture (motion frame). In order to reduce the number of dimensionality and to be affine invariant during body movement, the representation of the limb is defined by transforming their 3D Cartesian coordinates to 2D spherical coordinates as follows (See Figure 2). Let ! ov be a limb vector, and ! or be equivalent to the root orientation vector. Suppose  is the plane passing through the joint o and parallel to the

Figure 2. Features of a single limb vector. ! floor. Let the projection of ! ov and ! or on  be the ov0 and !0 or respectively. Then  and , the spherical coordinates ! of v on , are measured in angular degrees from ov0 to !0 ! or and from ov to N respectively, where N is the normal vector of  at o. Consequently, an arm is represented as 4D tuples ð; ; b; bÞ, where ð; Þ and ðb; bÞ are extracted from the lower and upper limb, respectively. The biped posture is likewise represented as 8D tuples. All these posture parameters form a 16-dimension feature vector. In the following, we describe the multidimensional indexing structure for motion database according to posture feature vectors. For each motion frame, its posture feature is a 16dimension vector. This feature vector codes posture parameters of arms and biped. We quantize this posture feature vector to form its index. For illustration purpose, we use a single arm projection map as an example. Suppose the i-th frame feature of the arm is ð; ; b; bÞ, then its index can be computed using the following truncation function H:     b b Hð; ; b; bÞ ¼ b c; b c; b c; b c ð1Þ a b c d where a, b, c and d are step sizes of angular degrees, and b c denotes the floor function. For expository purpose, we illustrate the indexing process of a single arm in Figure 3. It is clear that some consecutive motion frames may have the same index due to temporal coherence. For convenience, these consecutive frame are grouped into an indexed Moclip. Moreover, if there are multiple Moclips sharing the same index I, then these Moclips are stored as a link list indexed by I (See Figure 3). Now, suppose the step size of each parameter is set to 30 angular degrees. Then there are 12 and 6 steps along the -axis and -axis, respectively. It requires an infeasibly large memory to store total 128 68 indices in 16D. Therefore, we use k-d tree, a well-known spatial data structure for storing indices. This is because its memory

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

261

Comp. Anim. Virtual Worlds 2004; 15: 259–268

S.-P. CHAO ET AL. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Figure 4. Normal form structure. Figure 3. Indexing of a single arm. Now, consider the following typical BMT example: requirement is proportional to actual data size. According to our experiment, we have a sequence of TCC motion capture data with 23,130 frames, but only 671 indexed Moclips are distributed over 527 indices.

BMT-Normal Form Directory This section describes the building of the directory of normal form of basic motion texts (BMTs). BMT is a set of sentences that involves only two key-postures, namely, the start-posture and the end-posture. We use Cheng’s Tai Chi Chuan17,18 as the example, though the same consideration may apply to many other exercise domains. Since the same motion can be described in many different ways, it is obvious that different exercise instruction manuals such as Tai Chi Chuan (TCC) textbooks may use different sentences to describe the same motion. For convenience, we call these TCC textbooks TCC manuals (TCCMs). A typical Cheng’s TCCM has 37 forms, which are composed of a sequence of basic motions. In order to allow our textual input subsystem to accept different descriptions about the same motion, we use the notion of thesaurus of natural language process to construct the BMT-normal form directory. For example, ‘move downward’ is used as the preferred term for ‘downward’, ‘move down’, and ‘go down’. To achieve this goal, we type in three popular Cheng’s TCCMs and apply Chinese word segmentation package ‘CCL-Tagger’ to segment all sentences in these TCCMs into words and terms with part of speech tagged. After some statistical procedures, we have a frequency table for TCC key words in Chinese. Then a thesaurus for these key words can be constructed according to the general thesaurus for Chinese natural language processing.

Sink the body; Shift most of the weight to the right leg; Lift right hand to chest; Bend two knees slightly. Then after word segmentation processing and part of speech tagging, we can formalize above sentences into following three basic tags: (1) Motion Object [MO]: such as body and legs; (2) Motion [M]: verbs such as sink, shift, and lift; (3) Motion Space (Start-End Position/Motion Measurement) [MS]: It specifies the spatial extent of the motion such as ‘to chest’ or ‘about two to three inches’. Then these key words are replaced by their preferred terms in the thesaurus and arranged them into the following normal form structure (See Figure 4): {[Body]: (MO) (M) (MS)][Arm]: (MO) (M) (MS)] [Leg]: (MO) (M) (MS)] [Misc]: (MO) (M) (MS)]} Moreover, consecutive sentences may describe simultaneous motions of different limbs, for example, Drop the left leg back to the left-rear corner. At the same time, drop the left hand and raise the right hand to meet the left hand. Two hands press straight forward and the waist/ hip joint turns to the left foot. Our thesaurus uses ‘simultaneously’ as the preferred term for following terms: ‘and’, ‘at the same time’, ‘while’, ‘meanwhile’. Then above text descriptions will be converted into following two normal forms. *

*

Normal Form 1: {[Body: nil] [Arm1: (MO: left hand) (M: drop) (MS: nil)] [Arm2: (MO: right hand) (M: raise) (MS: meet left hand)] [Leg: (MO: left leg) (M: drop back) (MS: left-rear corner)] [Misc: nil]} Normal Form 2: {[Body: (MO: waist/hip joint) (M: turn) (MS: left foot)] [Arm: (MO: two hands) (M: press) (MS: straight forward)] [Leg: nil] [Misc: nil]}

The advantage of this formalization is that every motion description can be easily coded in XML

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

262

Comp. Anim. Virtual Worlds 2004; 15: 259–268

TAI CHI SYNTHESIZER * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

(Extensible Markup Language) format to achieve portability and generality. To summarize, we have collected a set of BMTs, describing stepwise motion of body and limbs from TCCMs and explained how to convert a BMT into its corresponding normal form, namely, BMT-normal form.

Motion Annotation In order to synthesize motion through text description, we need to annotate motion data using BMT-normal forms. In general, manually annotating motion data is time consuming and labor intensive. Therefore, Arikan et al.13 presented a semi-automatic annotation method to reduce the effort. In their system, users give a few Moclips and use support vector machine (SVM) classifiers to classify (annotate) the rest of motion collection. In order to obtain well-trained SVM classifiers, users need to interactively correct misclassification during annotation. In this study, an alternative method for motion annotation is proposed. Let M be an annotated Moclip, we use the start frame and end frame of M to find candidate Moclips, which are similar to M. With the aid of MIT, we can find possible candidates efficiently. Figure 5 illustrates the process of finding candidate frames. Suppose red (blue) frames have the same index as that of fstart (fend ). Then according to the start and end ordering in the timeline, candidate Moclips can be extracted (See candidates in Figure 5). Finally, a dynamic time warping (DTW)14 is applied to

compute the similarity between M and each candidate Moclip. Figure 6(a) shows a given annotated example and Figure 6(b–e) are similar Moclips that are retrieved from our motion data of TCC. In these examples, a human skeleton is sketched every five frames. Red and green lines indicate the trajectories of hands and feet. These similar Moclips can be described by the text descriptions of the given example. According to our experiments, the proposed method indeed facilitates the annotation work. Besides the annotation of a BMT, if some combination of BMTs repeatedly appears in the timeline of annotated motion data, we perform the union of the time intervals from these annotated Moclips to form a consecutive Moclip and annotate a higher level descriptor called subform, for example, ‘Roll back, press, and push’. Again, based on DTW, we automatically annotate these subforms. Similarly, we can annotate even higher level descriptors called forms. For example, the well-known ‘White Crane Spreads Its Wings’. motion is a form. After annotation process, we build the TCC motion hierarchy, i.e. TCC BMT-normal form directory. Moreover, this hierarchical directory maps each BMT, subform, or form to corresponding Moclips (See Figure 7). The advantage of this hierarchical directory is that the higher level descriptor requires less cost of motion transition than the lower level descriptor. For example, three BMT input texts ‘Roll back’, ‘Press’, and ‘Push’ requires two motion transitions. One is from ‘Roll back’ to ‘Press’ and the other is from ‘Press’ to ‘Push’. However, the input text ‘Roll back, press, and push’, which is a subform, does not require any

Figure 5. Search similar Moclip. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

263

Comp. Anim. Virtual Worlds 2004; 15: 259–268

S.-P. CHAO ET AL. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Figure 7. TCC motion hierarchy.

*

Figure 6. (a) Given annotated example; (b)–(e) Retrieved similar Moclips. motion transition because the Moclip corresponding to this text is the union of time intervals from ‘Roll back’, ‘Press’, and ‘Push’.

Motion Synthesis by Path Finding Algorithm

High-level Motion Control *

Key-posturing through Texts: In order to specify a new exercise through textual interface, users have to

describe start and end motions in text. Then the textual input subsystem will parse the high-level description into a sequence of BMT-normal forms. As a BMT-normal form can be considered a sequence of ordered tags, it inspires us to apply dynamic time warping (DTW)14 to search for the most similar one in the BMT-normal form directory. Then, the corresponding start Moclip Mstart and end Moclip Mend can be obtained. The last frame of Mstart and the first frame of Mend are chosen as fstart and fend , respectively. The procedure to synthesize a motion with given fstart and fend will be introduced in the next section. Key-posturing through Images: To facilitate keyposture specification, we introduce a graphical user interface (GUI) for users to load a 2D image with a posture and to mark required joints of the posture. Besides, users have to provide visual precedence between two consecutive joints. Then we apply Taylor’s posture reconstruction method15 to obtain its 3D posture. Moreover, users can also use the mouse to adjust the reconstructed 3D posture interactively. To synthesize a motion, two posture drawings or pictures are chosen as the start and the end postures, respectively. Then the system will reconstruct the corresponding 3D postures, fstart and fend , and synthesize the required motion. An illustration example will be given in the Experiment Results section.

With given start and end postures, a continue motion can be obtained by directly applying motion transition techniques, if they are similar. However, if the given two

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

264

Comp. Anim. Virtual Worlds 2004; 15: 259–268

TAI CHI SYNTHESIZER * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Figure 8. Motion synthesis by path finding.

postures are quite different, then some intermediate key-postures are required to synthesize a smooth continue motion. This is because the motion transition is reliable only if the given two motions are similar in posture. In this section, we introduce an algorithm for finding intermediate key-postures automatically based on our proposed motion index table (MIT). It is clear that postures belong to adjacent indices are more similar than those in non-adjacent indices. Let us consider the adjacent graph induced by the MIT, that is, each entry of the table is considered as a node and two adjacent entries are connected by an edge. Clearly, it is a multidimensional grid graph and can be regarded as a kind of motion transition graph10 (See Figure 8). Now, if two given postures are mapped far apart in the graph, then the motion synthesis can be realized by finding a proper path in the graph. Although motion synthesis through motion transition graph may produce unexpected transitions, this problem can be alleviated in our method. Because traveling along a path in the graph induced by MIT implies the trajectory of moving limbs, we can confine the trajectory of moving limbs toward the desired posture to avoid wandering. Based on this notion, we propose a path finding algorithm in the following subsection.

Path Finding Let fstart and fend be the start frame and the end frame respectively. Then their corresponding posture parameters can be extracted and mapped to the correspond-

ing indices Istart and Iend respectively by equation (1). The synthesized motion from fstart and fend can be realized by finding an appropriate path P from Istart to Iend and each edge along P is chosen according to the next best algorithm (See Figure 8). In other words, the synthesized motion is obtained by incrementally selecting the best Moclip among neighbors and provides proper transitions. For selecting proper Moclips during the incremental Moclip concatenations, we apply a motion similarity function, modified from that in10 to find the smoothest transition. Note that our similarity function is a weighted sum of the posture similarity and the velocity similarity. To avoid unexpected transitions from wandering in the graph, we propose a heuristic strategy as follows. Assume Ii is the current visiting index and S is the set of valid adjacent indices of Ii , that is, indices pointing to at least one Moclip. For each Ij 2S, ! we define the direction cosine ij of the vector Ii Ij as follows: ! ! Ii Iend  Ii Ij ! ! kIi Iend kk Ii Ij k

ð2Þ

! By forcing ij > 0, the direction cosine of the vector Ii Ij ! is confined towards Iend by the homing vector Ii Iend . Based on this notion, the path finding algorithm is proposed as follows and Figure 8 depicts an example path from Istart to Iend . *

Algorithm: Input: A start index Istart and an end index Iend . Output: A Path P.

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

265

Comp. Anim. Virtual Worlds 2004; 15: 259–268

S.-P. CHAO ET AL. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Step 1: Initialize a path P to be an empty list. Step 2: Initialize Ii ¼ Istart and Mi ¼ Mstart . Where Mi is the selected Moclip in Ii and Mstart is a Moclip at Istart chosen by fstart . Step 3: Add Ii and Mi to the path P. Step 4: For each index Ij in S, compute its homing direction cosine ij by equation (2) and remove Ij from S if ij < 0, where S is the set of valid adjacent indices of Ii . Step 5: For each Ij in S, find the Moclip Mj which has the maximum motion similarity according to Mi . Step 6: Find K, the index of the global maximum of motion similarity, by arg maxðWj Þ: j Step 7: Add IK and MK to P. Step 8: If IK ¼ Iend , then stop. Else set Ii ¼ IK and Mi ¼ MK , then go to Step 3 to search the next index in P. After obtaining a path P, it is necessary to provide proper transitions for concatenating selected Moclips in P. Considering the smooth motion transition without violating constraints of the original motion (e.g. feet skating, penetrating the floor, or being suspended in midair), we utilize the motion blending process in10 and in footskate cleanup method19 to perform our constraint-based motion transition.

Experiment Results The motion data used in our experiment is a 10 minutes TCC captured at 30 Hz from a professional coach. It took only 11.55 seconds for posture extracting and indexing of 23,130 motion frames on an Intel Pentium II 650 MHz computer with 256 MB memory. In our experiment, more than 60 frames can be synthesized per second. Our pre-processing time takes only a few seconds, and our synthesis speed can achieve an interactive rate. Figure 9 shows the results of motion synthesis between two given images. Figure 9(a) is the user-specified start frame (left) and end frame (right) and red markers are user labeled joints. Figure 9(b) and 9(c) are the perspective view, the front view, and the side view from 3D skeletons of start-posture and end-posture, respectively. Figure 9(d) shows frames as the results that satisfied the soft constraint of Figure 9(a). Red and green lines in these figures indicate the trajectories of hands and feet. Figure 10 demonstrates an example of text-to-animation. We input TCC text descriptions, ‘Two hands press

Figure 9. Example of images to animation. straight forward and the waist/hip joint turns to the left foot. The right handing hand opens, relaxes, and sinks; the left hand also relaxes, sinks, and moves slightly to the right. Drop the right arm and leg back in toward the body following the waist’ and Figure 10(a–e) are shots of the synthesized motion according to the given text descriptions. Figure 10a shows the motion that satisfies the first sentence, Figure 10(b) is the synthesized motion from the endposture of the first sentence to the start-posture of the second sentence, Figure 10(c) shows the motion that satisfies the second sentence, Figure 10(d) is the synthesized motion from the end-posture of the second sentence to the start-posture of the third sentence, and Figure 10(e) is the motion that satisfies the third sentence. The above results show that the proposed method can effectively generate desired motions.

Conclusion and Future Work In this study, a new framework for synthesizing motions based on given motion manuals and corresponding

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

266

Comp. Anim. Virtual Worlds 2004; 15: 259–268

TAI CHI SYNTHESIZER * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

meters of motion quality introduced in Laban Movement Analysis20 will be involved, so that the system can provide more flexible motion control through texts.

ACKNOWLEDGEMENTS

This study was supported partially by MOE Program for Promoting Academic Excellence of Universities under the grant number 89-E-FA04-1-4 and NSC grant number 92-2213E-007-081.

References 1. Lan C, Lai JS, Wong MK, Yu ML. The cardiorespiratory function, flexibility, and body composition among geriatric Tai Chi Chuan practitioners, Archives of Physical and Medical Rehabilitation 1996; 77(6): 612–616. 2. Li JX, Hong Y, Chan KM. Tai Chi: physiological characteristics and beneficial effects on health. British Journal of Sports Medicine 2001; 35(3): 148–156. 3. Wang JS, Lan C, Wong MK. Tai Chi Chuan training to enhance microcirculatory function in healthy elderly men. Archives of Physical and Medical Rehabilitation 2001; 82(9): 1176–1180. 4. Bindiganavale R, Schuler W, Allbeck J, Badler N, Joshi A, Palmer M. Dynamically altering agent behaviors using natural language instructions. Autonomous Agents 2000; 293–300. 5. Cassell J, Vilhja´lmsson HH, Bickmore T. BEAT: the Behavior Expression Animation Toolkit. ACM SIGGRAPH 2001 Conference Proceedings 2001; 477–486. 6. Bruderlin A, Williams L. Motion signal processing. ACM SIGGRAPH ’95 Conference Proceedings 1995; 97–104. 7. Unuma M, Anjyo K, Takeuchi R. Fourier principles for emotion-based human figure animation. ACM SIGGRAPH ’95 Conference Proceedings 1995; 91–96. 8. Choi KJ, Ko HS. Online motion retargetting. The Journal of Visualization and Computer Animation 2000; 11(5): 223–235. 9. Gleicher M, Litwinowicz P. Constraint-based motion adaptation. The Journal of Visualization and Computer Animation 1998; 9(2): 65–94. 10. Kovar L, Gleicher M, Pighin F. Motion graphs. ACM Trans. Graph. ACM Press 2002; 21(3): 473–482. 11. Gleicher M, Shin HJ, Kovar L, Jepsen A. Snap-together motion: assembling run-time animations. ACM Symposium on Interactive 3D Graphics 2003; 181–188. 12. Kim TH, Park SI, Shin SY. Rhythmic-motion synthesis based on motion-beat analysis. ACM Trans. Graph., ACM Press 2003; 22(3): 392–401. 13. Arikan O, Forsyth DA, O’Brien JF. Motion synthesis from annotations. ACM Trans. Graph., ACM Press 2003; 22(3): 402–408. 14. Parsons TW. Voice and Speech Processing. McGraw-Hill: New York; 1986.

Figure 10. Example of text description to animation.

motion capture examples is introduced. A motion with a given textual description can be synthesized by converting the given sentences into BMT-normal forms and concatenating the corresponding motion clips. Moreover, we introduce a motion index table to facilitate the example search during motion annotation and smooth motion concatenations. Several experimental examples are given to demonstrate the effectiveness of the proposed method. Finally, we note, in this work, the motion modifiers, such as fast, slow, strong, and light, of basic motion texts cannot be changed, since they are tightly bound to the motion data. For future work, the para-

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

267

Comp. Anim. Virtual Worlds 2004; 15: 259–268

S.-P. CHAO ET AL. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

18. McFarlane S. The Complete book of T’ai Chi. Owl Publishing House, a division of Cite Publishing Ltd: 2002. 19. Kovar L, Schreiner J, Gleicher M. Footskate cleanup for motion capture editing. ACM Symposium on Computer Animation 2002; 97–104. 20. Chi D, Costa M, Zhao L, Badler N. The EMOTE model for effort and shape. ACM SIGGRAPH ’00 Conference Proceedings 2000; 173–182.

15. Taylor CJ. Reconstruction of articulated objects from point correspondences in a single uncalibrated image. Computer Vision and Image Understanding 2000; 80(3): 349–363. 16. Kovar L, Gleicher M. Flexible automatic motion blending with registration curves. ACM Symposium on Computer Animation 2003; 214–224. 17. Cheng MC. Cheng Man-Chings Advanced Tai-Chi Form Instructions, Wile D (ed.). Sweet Chii Press: 1986.

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

Copyright # 2004 John Wiley & Sons, Ltd.

268

Comp. Anim. Virtual Worlds 2004; 15: 259–268

Tai Chi synthesizer: a motion synthesis framework based on key ...

presents a novel framework for synthesizing new motions based on given motion manuals .... exercise manuals, a graphical user interface based on ..... the table is considered as a node and two adjacent .... Lan C, Lai JS, Wong MK, Yu ML.
Missing:

387KB Sizes 0 Downloads 279 Views

Recommend Documents

Tai Chi synthesizer: a motion synthesis framework ...
system; animation techniques; motion capture and retargeting; language parsing and ..... (Extensible Markup Language) format to achieve port- .... Then we apply. Taylor's posture reconstruction method15 to obtain its 3D posture. Moreover, users can a

Tai-Chi Chuan Motion Tracking by Integrating Inertial ...
The accuracy of the proposed method is verified using a mocap system. Zhu and Zhou [15] propose a real time tracking method using sensor modules to track human motion. Each sensor module consists of a tri-axis accelerometer, a tri-axis rate-gyro and

Tai-Chi Chuan Motion Tracking by Integrating Inertial ...
aided Tai-Chi learning system is demanding a less- restrictive ... sensors, cameras, magnetics and ultrasound systems ... location information for posture tracking. ..... 1+bi,z bi,x. 0. −bi,xbi,y. 1+bi,z. 1 − b2 i,y. 1+bi,z bi,y. 0. −bi,x. −

tai chi 24 forms pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. tai chi 24 forms ...

tai chi fa jin pdf
Page 1 of 1. File: Tai chi fa jin pdf. Download now. Click here if your download doesn't start automatically. Page 1 of 1. tai chi fa jin pdf. tai chi fa jin pdf. Open. Extract. Open with. Sign In. Main menu. Displaying tai chi fa jin pdf. Page 1 of

tai chi yang pdf
Page 1 of 1. File: Tai chi yang pdf. Download now. Click here if your download doesn't start automatically. Page 1 of 1. tai chi yang pdf. tai chi yang pdf. Open. Extract. Open with. Sign In. Main menu. Displaying tai chi yang pdf. Page 1 of 1.

On Robust Key Agreement Based on Public Key Authentication
explicitly specify a digital signature scheme. ... applies to all signature-based PK-AKE protocols. ..... protocol design and meanwhile achieve good efficiency.

On Robust Key Agreement Based on Public Key ... - Semantic Scholar
in practice. For example, a mobile user and the desktop computer may hold .... require roughly 1.5L multiplications which include L square operations and 0.5L.

MCGP: A Software Synthesis Tool Based on Model ... - Semantic Scholar
whether) a candidate solution program satisfies a property. The main ... software, a natural challenge is to generate automatically correct-by-design pro- grams.

MCGP: A Software Synthesis Tool Based on Model ... - Semantic Scholar
Department of Computer Science, Bar Ilan University. Ramat Gan 52900 .... by following the best generated programs, and by navigating through the chain.

MCGP: A Software Synthesis Tool Based on Model ...
candidate program from a correct solution. The tool allows the user to control various parameters, such as the syntactic building blocks, the structure of the programs, and the fitness function, and to follow their effect on the convergence of the sy

A Synthesis Instance Pruning Approach Based on ... - Semantic Scholar
Department of Computer Science, Ocean University of China, Qingdao 266100, China;. 2. Department of ..... the loss of dynamic coverage after deleting the unit's leaf that ..... uniform instances is in some degree made up by virtual non-uniform ...

A Multiple Layered Spatial Data Integration Framework Based on ...
JAVA. RMI. Access. Local. Remote. NetBIOS. TCP/IP. Figure 5 shows another portion of ontology of the Persistence category. Its related concepts are Format, Access, .... Middleware. Figure 6 shows the process for data search/query/integration from the

A Multiple Layered Spatial Data Integration Framework Based on ...
modeling, data mining, textual analysis, and statistical techniques. ... model which can be converted to database schema, or a result of a data mining .... Excel. Interface. POP3. Distributed. API. WebService. ODBC. DCOM. CORBA. JAVA. RMI. Access. Lo

G2 Remoting: A Cycle Stealing Framework based on ...
A dedicated server machine is used as the physical manifestation of a virtual parallel machine on which remote objects logically reside from the programmers ... remote objects are not, however, physically created on this server machine; ..... example

A Framework to Optimize Biodiversity Restoration Efforts Based on ...
tinguishing ecological systems that are capable to recover by autogenic processes from ... chances of restoration success, defined as the best cost/benefit outcome. ..... tivity with little biological data (Minor and Urban 2007) or even considering .

A Modular Verification Framework Based on Finite ...
strongly connect components that we will talk about from now on. ..... 0: no push button; 1: using push button1; 2: using push buttons 1 &. 2; using push buttons 1 ...

a framework based on gaussian mixture models and ...
Sep 24, 2010 - Since the invention of the CCD and digital imaging, digital image processing has ...... Infrared and ultraviolet radiation signature first appears c. .... software can be constantly upgraded and improved to add new features to an.

RefaX: A Refactoring Framework Based on XML
system to improve its internal quality and preserving its external .... technologies, such as XSLT [29], or as an API for extracting or .... converting the original source code files selected by the user into the ..... http://beautyj.berlios.de/beaut

eBook Download Learning Tai Chi Chuan Full Download
Book Synopsis. This book represents the legacy of one of the world s foremost teachers and proponents of Yang Style Tai. Chi Chuan in the modern era,. Master ...