Introduction
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Motion Enriching Using Humanoide Captured Motions
MASTER THESIS: MOTION ENRICHING USING HUMANOIDE CAPTURED MOTIONS STUDENT: SINAN MUTLU ADVISOR : A NTONIO SUSÌN SÀNCHEZ SEPTEMBER, 8TH 2010 COURSE: MASTER IN COMPUTING LSI DEPERTMANT POLYTECNIC UNIVERSITY OF CATALUNYA 1 Abstract Animated humanoid characters are a delight to watch. Nowadays they are extensively used in simulators. In military applications animated characters are used for training soldiers, in medical they are used for studying to detect the problems in the joints of a patient, moreover they can be used for instructing people for an event(such as weather forecasts or giving a lecture in virtual environment). In addition to these environments computer games and 3D animation movies are taking the benefit of animated characters to be more realistic. For all of these mediums motion capture data has a great impact because of its speed and robustness and the ability to capture various motions. Motion capture method can be reused to blend various motion styles. Furthermore we can generate more motions from a single motion data by processing each joint data individually if a motion is cyclic. If the motion is cyclic it is highly probable that each joint is defined by combinations of different signals. On the other hand, irrespective of method selected, creating animation by hand is a time consuming and costly process for people who are working in the art side. For these reasons we can use the databases which are open to everyone such as Computer Graphics Laboratory of Carnegie Mellon University. Creating a new motion from scratch by hand by using some spatial tools (such as 3DS Max, Maya, Natural Motion Endorphin or Blender) or by reusing motion captured data has some difficulties. -
Automated Staging for Virtual Cinematography Amaury Louarn, Marc Christie, Fabrice Lamarche
Automated Staging for Virtual Cinematography Amaury Louarn, Marc Christie, Fabrice Lamarche To cite this version: Amaury Louarn, Marc Christie, Fabrice Lamarche. Automated Staging for Virtual Cinematography. MIG 2018 - 11th annual conference on Motion, Interaction and Games, Nov 2018, Limassol, Cyprus. pp.1-10, 10.1145/3274247.3274500. hal-01883808 HAL Id: hal-01883808 https://hal.inria.fr/hal-01883808 Submitted on 28 Sep 2018 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Automated Staging for Virtual Cinematography Amaury Louarn Marc Christie Fabrice Lamarche IRISA / INRIA IRISA / INRIA IRISA / INRIA Rennes, France Rennes, France Rennes, France [email protected] [email protected] [email protected] Scene 1: Camera 1, CU on George front screencenter and Marty 3/4 backright screenleft. George and Marty are in chair and near bar. (a) Scene specification (b) Automated placement of camera and characters (c) Resulting shot enforcing the scene specification Figure 1: Automated staging for a simple scene, from a high-level language specification (a) to the resulting shot (c). Our system places both actors and camera in the scene. Three constraints are displayed in (b): Close-up on George (green), George seen from the front (blue), and George screencenter, Marty screenleft (red). -
Lightweight Procedural Animation with Believable Physical Interactions
Proceedings of the Fourth Artificial Intelligence and Interactive Digital Entertainment Conference Lightweight Procedural Animation with Believable Physical Interactions Ian Horswill Northwestern University, Departments of EECS and Radio/Television/Film 2133 Sheridan Road, Evanston IL 60208 [email protected] Abstract matics system. Posture control is performed by applying I describe a procedural animation system that uses tech- simulated forces and torques to the torso and pelvis. niques from behavior-based robot control, combined with a Interestingly, the use of a dynamic simulation actually minimalist physical simulation, to produce believable cha- simplifies control, allowing the use of relatively crude con- racter motions in a dynamic world. Although less realistic trol signals, which are then smoothed by the passive dy- than motion capture or full biomechanical simulation, the namics of the character body and body-environment inte- system produces compelling, responsive character behavior. raction; similar results have been found in both human and It is also fast, supports believable physical interactions be- robot motor control (Williamson, 2003). tween characters such as hugging, and makes it easy to au- Twig shows that surprisingly simple techniques can gen- thor new behaviors. erate believable2 motions and interactions. Much of the focus of this paper will be on ways in which Twig is able to Overview1 cheat to avoid doing complicated modeling or control, while still maintaining believability. This work is indebted Versatile procedural animation is a necessary component to the work of Jakobsen (Jakobsen, 2001) and Perlin (Per- for applications such as interactive drama, in which charac- lin, 1995, 2003; Perlin & Goldberg, 1996), both for their ters participate in complex interactions that cannot be pre- general approaches of using simple techniques to generate planned at authoring time. -
Inverse Kinematics Last Time? Today Keyframing Procedural Animation Physically-Based Animation
Last Time? • Navier-Stokes Equations • Conservation of Inverse Kinematics Momentum & Mass • Incompressible Flow Today Keyframing • How do we animate? • Use spline curves to automate the in betweening – Keyframing – Good control – Less tedious than drawing every frame – Procedural Animation • Creating a good animation still requires considerable – Physically-Based Animation skill and talent – Forward and Inverse Kinematics – Motion Capture • Rigid Body Dynamics • Finite Element Method ACM © 1987 “Principles of traditional animation applied to 3D computer animation” Procedural Animation Physically-Based Animation • Describes the motion algorithmically, • Assign physical properties to objects as a function of small number of (masses, forces, inertial properties) parameters • Example: a clock with second, minute • Simulate physics by solving equations and hour hands • Realistic but difficult to control – express the clock motions in terms of a “seconds” variable – the clock is animated by varying the v0 seconds parameter mg • Example: A bouncing ball -kt –Abs(sin(ωt+θ0))*e 1 Articulated Models Skeleton Hierarchy • Articulated models: • Each bone transformation – rigid parts described relative xyzhhhhhh,,,,,qf s – connected by joints to the parent in hips the hierarchy: qfttt,, s • They can be animated by specifying the joint left-leg ... angles as functions of time. r-thigh qc qi qi ()t r-calf y vs qfff, x r-foot z t1 t2 t1 t2 1 DOF: knee 2 DOF: wrist 3 DOF: arm Forward Kinematics Inverse Kinematics (IK) • Given skeleton xyzhhhhhh,,,,,qf -
Ipi Soft Motion Capture Software Inspires Feature Length Machinima
iPi Soft Motion Capture Software Inspires Feature Length Machinima Horror Movie Ticket Zer0 Director Ian Chisholm Uses End-to-End Markerless Performance Capture Pipeline For Character Animation; Film Now Airing on YouTube _____ MOSCOW, RUSSIA – September 25, 2018 – For his most ambitious project to date, the feature length horror film Ticket Zer0, motion capture director/writer Ian Chisholm once again leaned heavily on iPi Motion Capture, the advanced markerless motion capture technology from iPi Soft. The 66-minute animated machinima cinematic production is currently available for streaming on YouTube or Blu-Ray here. Ticket Zer0 is a follow-up to Chisholm’s Clear Skies trilogy series (viewed by hundreds of thousands on YouTube), which also used iPi Mocap throughout for performance capture. Ticket Zer0 tells the story of the night shift at a deserted industrial facility. When a troubleshooter is sent in to fix a mysterious problem, something far more sinister emerges than any of them could have imagined. Chisholm notes production for Ticket Zer0 took almost five years to complete and used iPi Motion Capture for recording movement in every single frame for all the characters in the film. “iPi Soft technology has proven to be an invaluable creative resource for me as a filmmaker during motion capture sessions,” he says. “There wasn't anything on screen that wasn't recorded by iPi Mocap. The ability to perform the motion capture was exactly what I needed to expand and improve the immersive quality of my film.” Workflow Details: For motion capture production, Chisholm used iPi Mocap in conjunction with six Sony PS3 Eye cameras along with two PlayStation Move controllers to track hand movements. -
Procedural Animation
Procedural Animation Computer Graphics Seminar 2017 Spring Andreas Sepp Coming up today ● Keyframe animation ● Procedural animation for basic character animation ○ Procedural animation with keyframe animation ○ Inverse kinematics ● Procedural animation as a broader field ○ Artificial life animation ○ Physics based modelling and animation Basic keyframe animation ● Animator draws or models the starting and ending points of a transition, called keyframes, and sets their position in time ● The remaining frames inbetween 2 keyframes are interpolated from them ● The animator is in control of everything at every point in time key key key Basic keyframe animation problems ● Transitions in interactive setting ○ Blend walking and running animation? ○ Create a walk ↔ run transition animation? ■ 15 animations - 105 blends required ● Unrealistic movement ○ No proper feedback from the environment ○ Not acceptable anymore Procedural Animation ● a type of computer animation, used to automatically generate animation in real-time to allow for a more diverse series of actions than could otherwise be created using predefined animations ● Animator not in control of everything anymore ● a) Integrated with keyframe animation ○ Roughly follows keyframes ○ Change dynamically ■ e.g. getting hit while running, going up the stairs ● b) Fully procedural animation ○ Initial parameters and some sort of input parameters are provided to control the animation ■ Initial position; forces, torques in time Keyframe + procedural animation ● Dynamic combining of multiple animations ○ Lower body running ○ Upper body swinging a sword ○ Body recoiling to a blow ● What you can achieve with just 14 keyframes and procedural animation: ○ http://www.gamasutra.com/view/news/216973/Video_An_indie_a pproach_to_procedural_animation.php [4:00-10:00] To actually feel connected to the world.. -
Application of Performance Motion Capture Technology in Film and Television Performance Animation
Proceedings of the 2nd International Symposium on Computer, Communication, Control and Automation (ISCCCA-13) Application of Performance Motion Capture Technology in Film and Television Performance Animation ZHANG Manyu Dept. Animation of Art, Tianjin Academy of Fine Arts Tianjin , CHINA E-mail: [email protected] Abstract—In today’s animation films, virtual character tissue simulation technique” for the first time, so that acting animation is a research field developing rapidly in computer skills of actors can be delivered to viewers maximally. graphics, and motion capture is the most significant What’s more, people realize that “motion capture component, which has brought about revolutionary change for technology” shall be called as “performance capture 3D animation production technology. It makes that animation technology” increasingly. Orangutan Caesar presented in producers are able to drive animation image models directly with the use of actors’ performance actions and expressions, Rise of the Apes through “motion capture” technology is which has simplified animation manufacturing operation very classical. Various complicated and subtle expressions greatly and enhanced quality for animation production. and body languages of Caesar’s not vanished bestiality and initially emerged human nature have been presented Keywords- film and television films, motion capture, naturally and freshly in the film, which has achieved the technology excellent realm of mixing the false with the genuine. When Caesar stands in front of Doctor Will in equal gesture, its I. INTRODUCTION independence and arbitrariness in its expression with dignity With the development of modern computer technology, of orangutan has undoubtedly let viewers understand computer animation technology has developed rapidly. -
Motion Capture Assisted Animation: Texturing and Synthesis
Motion Capture Assisted Animation: Texturing and Synthesis Katherine Pullen Christoph Bregler Stanford University Stanford University Abstract Keyframed Data 2 We discuss a method for creating animations that allows the anima- 0 (a) tor to sketch an animation by setting a small number of keyframes −2 on a fraction of the possible degrees of freedom. Motion capture −4 data is then used to enhance the animation. Detail is added to de- −6 grees of freedom that were keyframed, a process we call texturing. −8 Degrees of freedom that were not keyframed are synthesized. The −10 method takes advantage of the fact that joint motions of an artic- −12 ulated figure are often correlated, so that given an incomplete data 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 set, the missing degrees of freedom can be predicted from those that Motion Capture Data are present. 1 CR Categories: I.3.7 [Computer Graphics]: Three-Dimensional 0 (b) Graphics and Realism—Animation; J.5 [Arts and Humantities]: −1 performing arts −2 Keywords: animation, motion capture, motion texture, motion −3 synthesis −4 translation in inches −5 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 time in seconds 1 Introduction Figure 1: Comparison of keyframed data and motion capture As the availability of motion capture data has increased, there has data for root y translation for walking. (a) keyframed data, with been more and more interest in using it as a basis for creating com- keyframes indicated by red dots (b) motion capture data. -
Creating 3DCG Animation with a Network of 3D Depth Sensors Ning
2016 International Conference on Sustainable Energy, Environment and Information Engineering (SEEIE 2016) ISBN: 978-1-60595-337-3 Creating 3DCG Animation with a Network of 3D Depth Sensors Ning-ping SUN1,*, Yohei YAMASHITA2 and Ryosuke OKUMURA1 1Dept. of Human-Oriented Information System Engineering 2Advanced Electronics and Information Systems Engineering Course, National Institute of Technology, Kumamoto College, 2659-2 Suya, Koshi, Kumamoto 861-1102, Japan *Corresponding author Keywords: Depth sensor, Network, 3D computer graphics, Animation production. Abstract. The motion capture is a useful technique for creating animations instead of manual labor, but the expensive camera based devices restrict its practical application. The emerging of inexpensive small size of depth-sensing devices provides us opportunities to reconstitute motion capture system. We introduced a network of 3D depth sensors into the production of real time 3DCG animation. Different to the existing system, the new system works well for data communication and anime rendering. We discuss below server how communicate with a group of sensors, and how we operate a group of sensors in the rendering of animation. Some results of verification experiments are provided as well. Introduction The presentation of character’s movement is very important for animation productions. In general, there are two kinds of technique to present movement in an anime. First one is a traditional method, in which animators draw a series of frames of character’s movements individually, and show these frames continuously in a short time interval. Fps, i.e., frames per second is involved the quality of an animation. The more the fps is, the better the quality becomes. -
384 Cinemacraft: Immersive Live Machinima As an Empathetic
Cinemacraft: Immersive Live Machinima as an Empathetic Musical Storytelling Platform Siddharth Narayanan Ivica Ico Bukvic Electrical and Computer Engineering Institue for Creativity, Arts & Technology Virginia Tech Virginia Tech [email protected] [email protected] ABSTRACT or joystick manipulation for various body motions, rang- ing from simple actions such as walking or jumping to In the following paper we present Cinemacraft, a technology- more elaborate tasks like opening doors or pulling levers. mediated immersive machinima platform for collaborative Such an approach can often result in non-natural and po- performance and musical human-computer interaction. To tentially limiting interactions. Such interactions also lead achieve this, Cinemacraft innovates upon a reverse-engineered to profoundly different bodily experiences for the user and version of Minecraft, offering a unique collection of live may detract from the sense of immersion [10]. Modern day machinima production tools and a newly introduced Kinect game avatars are also loaded with body, posture, and ani- HD module that allows for embodied interaction, includ- mation details in an unending quest for realism and com- ing posture, arm movement, facial expressions, and a lip pelling human representations that are often stylized and syncing based on captured voice input. The result is a mal- limited due to the aforesaid limited forms of interaction. leable and accessible sensory fusion platform capable of These challenges lead to the uncanny valley problem that delivering compelling live immersive and empathetic mu- has been explored through the human likeness of digitally sical storytelling that through the use of low fidelity avatars created faces [11], and the effects of varying degrees of re- also successfully sidesteps the uncanny valley. -
Use Style: Paper Title
INTERNATIONAL CONFERENCE ON INFORMATICS AND CREATIVE MULTIMEDIA 2013 (ICICM’13) UTM, KUALA LUMPUR. SEPTEMBER 3-6, 2013, pp.104,109, 4-6 Sept. 2013 doi:10.1109/ICICM.2013.25 IEEEXplore Expression driven Trignometric based Procedural Animation of Quadrupeds Zeeshan Bhatti, Asadullah Shah, Mustafa Karabasi and Waheed Mahesar Khulliyyah of Information and Communication Technology International Islamic University Malaysia, Kuala Lumpur e-mail: [email protected], [email protected], [email protected], [email protected] Abstract— This research paper addresses the problem of riggers need [3]. So a character animator normally ends up generating involuntary and precise animation of quadrupeds building a custom skeletal rig for the ease of animation with automatic rigging system of various character types. The [3][4]. The process is also known as Character Rigging. We technique proposed through this research is based on a two have used MAYA as the basic development and simulation tier animation control curve with base simulation being driven tool with motion equations implemented as Maya Embedded through dynamic mathematical model using procedural Language (MEL) code written as expressions that gets algorithm and the top layer with a custom user controlled executed at every frame controlling and driving the various animation provided with intuitive Graphical User Interface body part of the character rig. (GUI). The character rig is based on forward and inverse kinematics driven through trigonometric based motion II. LITERATURE REVIEW equations. The User is provided with various manipulators and Quadruped motion has always been an integral part of attributes to control and handle the locomotion gaits of the characters and choose between various types of simulated character animation and simulation. -
A Software Pipeline for 3D Animation Generation Using Mocap Data and Commercial Shape Models
A Software Pipeline for 3D Animation Generation using Mocap Data and Commercial Shape Models Xin Zhang, David S. Biswas and Guoliang Fan School of Electrical and Computer Engineering Oklahoma State University Stillwater, Oklahoma, United States {xin.zhang, david.s.bisaws, guoliang.fan}@okstate.edu ABSTRACT Additionally, high-quality motion capture (mocap) data We propose a software pipeline to generate 3D animations by and realistic human shape models are two important compo- using the motion capture (mocap) data and human shape nents for animation generation, each of which involves a spe- models. The proposed pipeline integrates two animation cific skeleton definition including the number of joints, the software tools, Maya and MotionBuilder in one flow. Specif- naming convention, hierarchical relationship and underlying ically, we address the issue of skeleton incompatibility among physical meaning of each joint. Ideally, animation software the mocap data, shape models, and animation software. Our can drive a human shape model to move and articulate ac- objective is to generate both realistic and accurate motion- cording to the given mocap data and optimize the deforma- specific animation sequences. Our method is tested by three tion of the body surface with natural smoothness, as shown mocap data sets of various motion types and five commer- in Fig. 1. With the development of computer graphics and cial human shape models, and it demonstrates better visual the mocap technology, there are plenty of mocap data and realisticness and kinematic accuracy when compared with 3D human models available for various research activities. three other animation generation methods. However, due to their different sources, there is a major gap between the mocap data, shape models and animation soft- ware, which often makes animation generation a challenging Categories and Subject Descriptors task.