IR Film Review: ALITA – BATTLE ANGEL [20th Century Fox]

The mixture of James Cameron and Robert Rodriguez definitely has a great angle to it. “Alita: Battle Angel” was a title heard years ago and moved around as myth actually for a time. Harrison Ford was attached per se at one point. This story of stories that Cameron had developed for years actually was something Rodriguez told Inside Reel in a Fest Track interview at SxSW in March 2018. Reading the first 3 volumes before talking with producer Jon Landau as well as leads Rosa Salazar & Keaan Johnson in Austin (see their Fest Track interview here), gave a good perception of the structure but what is one to say between an anime/graphic novel that was written many years ago versus the ideals of the actual script (which having been co-written by James Cameron definitely should retain his story sense). What “Alita: Battle Angel” does very well is keep itself focused. The one true balance that stays pretty crisp and clear throughout the film is Rosa Salazar as Alita. Many may think that it is simply a computer performance but that could not put the sense of innocence, anger and breathe in what is seen here. Granted it is not Andy Serkis but who can compete on that level. What Rosa brings is a soul to this girl who was originally built as a killing machine. Salazar has been missing in part from many of the media rounds per se (in large part) but that might be better so the character simply exists on her own. Rodriguez’s touch is here for sure but it is sometimes lost in the bigger sequences. Oddly enough, this reviewer kept seeing “Speed Racer” in the race sequences per se. They are good but at a certain point are more video game oriented.

The character build even though it takes a while in the beginning does the film correctly but there is no “a-ha” moment. The scene though where Alita first tries her new body with fighting moves shows a path to identity and the sequence inside a bar (a very visceral scene in the graphic novel) definitely comes to life. The reason why is that all the characters in there are so unique. It makes one think of “From Dusk Till Dawn”. What seems to be missing is some of Robert’s camera tricks and stylistic touches although to be fair Rodriguez did mention in that same interview that this was not him doing a Robert Rodriguez film but instead doing a Jim Cameron film. So in that respect it does work, the script is tight, the visuals are fluid and it does its job. It is fun to watch but it is not spectacular. There is never quite a moment where Alita becomes the chosen one or that her love against her own life will ring out. One scene inside the apartment of Hugo (played by Johnson) comes close and really makes the CG of Rosa as a cyborg really key into the story. The climax, like most, has to serve a story point and that is understandable. Christoph Waltz does an admirable job as the Doc and Jennifer Connelly & Mahershala Ali do their part within the structure but Ed Skrein as a competitor is the only one who brings an edge to the proceedings. Here is hoping “Alita” connects to the audience because unlike many recent popcorn films, it understands the concept of a beginning, middle and end within a true story arc. But it is in the silent moments, when you can hear the acting, that make the most impact. One simple act of Alita laying her head on her father’s shoulder has almost more power than a large action sequence. But that said, one does not exist in the large scale, big budget film without the other.


By Tim Wassberg


Avatar – Theatrical Review

“Avatar” has long been hyped as the next age of cinema and on many levels in terms of sheer mastery of the medium, it is. It blends some of the schmaltz of “Titanic” with some of the balls-out rendition of “Aliens”. Now while it might be too early to render its true appraisal since the gee whiz factor enters so much into the equation, the third act of the picture, which takes the essential element of the battle scenes and sheer bad ass intensity while still staying within the PG-13 realm, is quite impressive. The photo real propensity makes you almost forget that the blue skinned “avatars” are not real in real life. Their adherence through performance capture especially with Sigourney Weaver’s Grace is so spot on since the image of Weaver (specifically in Cameron’s “Aliens”) is so ingrained to people’s minds.

Another lipnus test that was on supreme importance was interactivity of characters. Even though the blue alien love scene is a bit overplayed, the physical intermingling (like kissing) which I found problematic in terms of the technology back in 99 with “Final Fantasy: The Spirits Within” is corrected here. The format “Avatar” was seen, which most of the theaters cannot do, is 3D IMAX. The digital negative seems to have been shot for the format with a 70mm (like “Aliens”) which means the image should be barely cropped. The sense of scale especially with this movie is necessary so the angle is important. At one specific point when the natives are going up to meld with a bird (takes some explaining), the straight down shot can induce vertigo if you want it to. The battle sequences especially towards the end in the air are as great as anything Cameron has done or better.

In terms of the story on first glance, the aspect of the natives seems almost like pat storytelling but in further examination, despite the details, becomes very mythic. Like “Star Wars”, alot of the elements come from Tolkien which Cameron admits inspired him. Actually it was Peter Jackson’s work on Gollum speaking to himself in “Rings” that made Cameron believe that “Avatar” could finally be pulled off after all these years. The irony of course is there. However what “Avatar” reminds me of more in structure is Frank Herbert’s “Dune”. Like Paul Atreides, Jake (played here by Sam Worthington) is an outsider from another planet brought in to help secure a mineral resource. Here, in true vague fashion, this material is called “Unobtanium” whereas “Dune” had “The Spice”. Jake is run from his people (in Paul’s case, royalty – in Jake’s case – The Marines) and ends up embedding himself with the supposed savages (in “Dune”, they were the Fremen). With their care and guidance, Jake must learn their ways and eventually takes a mate (like Paul did with Chani in “Dune”).

In “Dune”, Paul must tame a massive worm and then transmute the dangerous Water Of Life. Jake must tame a wild bird and then transcend his Avatar. Very similar but everything is in reverse with certain elements. The planet here is green with life whereas Arrakis is barren. The Earth the military comes from in “Avatar” is one where everything in terms of living has been destroyed. It is a dying world.

Despite these close similarities, Cameron is a student of cinema. Most of his films have been franchise based. This is no different. You need a universal theme which this does deliver. Granted despite the bad ass elements, there seems to be a tree-hugger friendly message but that is the biggest criticism. You just expect Cameron to say “fuck it all” and take it into the fire. Messages seem to be following him as he grows older which might be his recent interest with documentaries and education.

All said “Avatar” is pretty amazing to watch. Not a perfect movie by far but with a wonder of elements going for it. In ambition definitely a good follow up from “Titanic” doing something original but presenting a marketing challenge at times which is understandable when creating a new world. Out of 5, I give it a 3 1/2.

Capping The Image: 2009 Produced By Conference – Los Angeles – Part II


After the essence of pre-vis, seeing things in advance is very helpful especially when planning or in some cases, adding material for a project, because of an ability to accomplish something more with what is available. At the 2009 Produced By Conference on its second day at Sony Studios, the aspects of Mo Cap (or Motion Capture), which is being optimized by the top people in the business from Bob Zemeckis to David Fincher to James Cameron, was brought into focus.

Performance/Motion Capture Production Technology Performance basis of creatures is motivated ultimately by camera angle. Certain examples of what was done in the past to emphasize mo cap was always personified by the dotted faces of actors used to create the computer grids. To personify this point, footage is shown of Bill Nighy paying the slithery captain in “Pirates Of The Caribbean: Dead Man’s Chest”. Nighy called his getup when he was playing the creature: his “funky computer pajamas. The thing is that mo cap still can’t capture directly around the eye sockets and the actual eyes. Steve Sullivan, Sr. Technology Officer at ILM, breaks down the different perspectives of motion capture. “Facial Mo-Cap” involves a “zillion” dots on the face and uses the same fundamental techniques as “Body Mo-Cap”. Make up can be sometimes used to make the dots. Sometimes it works but it depends on the lighting. When you start working on facial re-targeting where you actually align a new face with a different performance, it becomes a subjective artistic endeavor. These re-targeting exercises however can now be done in real time. Sullivan then shows a test using a live low res example running off an XBOX 360 console and not one of their super computers. The result is still fairly comprehensive.

The realization is that once you do the actual mo cap and get the data, then you can do camera coverage. After that, you just go back and key out any problematic frames. Sullivan then showed an “Indiana Jones” video game engine optimizing mo cap which was made in two hours. It wasn’t done to show the possibilities of FX but more to see how the actual story could play out. This kind of low tech Mo-Cap with facial re-targeting is effective in his mind for actor freedom and director control. It also allows on-demand capture for quick experimentation and real time documenting for timing.

Rob Readow and Debbie Denise from Sony Imageworks began the ball rolling on Mo-Cap back when director Bob Zemeckis first came in with “Polar Express”. They went about with different tests. The second one they did for “Polar” at Imageworks was green screen with full digital environments like “300”. They suggested the InMotion Mo-Cap system they had been working on to Zemeckis. Zemeckis was hesitant but it was Tom Hanks who said that it could allow him to play both the boy and the conductor in the same movie which seemingly ultimately swayed the director. Despite what anyone thinks, according to Readow and Denise, these movies are still mounted as big pictures. The box that they first worked on for “Polar” was only 10 feet by 10 feet which is the most the computer could do at the time. On “Monster House” they could have the ability of 200 cameras shooting on once. And by the time “Beowulf” came around, they could capture horses running across the soundstage.

The importance aspect to remember is that it is still a live action shoot of sorts. The actors just don’t have to hit their marks. The “Beowulf” shoot itself took 38-40 days. Denise says Hanks said that it was the hardest he has ever worked because you never go back to your trailer. Readow follows up Denise’s observations talking about using the Sony DXC-M3A camera as a virtual camera rig where you are using its systems to shoot in a virtual world. That is how they got the realistic camera movements in “Surf’s Up”. By comparison when they worked on some of the elements of “Watchmen” specifically Dr. Manhattan with Billy Crudup, the scanning was done low impact on set. After the initial scan done in live action with the dot structure, the body fabrication was scanned off a bodybuilder with the head and partial torso of Crudup melded together. The simulation software even worked to cause the muscles to ripple. However the camera still racked Crudup in 2D space. Even in the small sequence where Manhattan is forming, the dyanmic simulation of the circulatory systems was done on set with Mo-Cap with the final VFX adding the other different elements such as heat, static and luminescence.

Greg LaSalle, Founder of Mova, works in the essence of the balance between real and photoreal. He worked on “The Curious Case Of Benjamin Button” using aspects of their proprietary Mo-Cap system but wanted to let Oscar winner for “Button”: Steve Preeg talk about that show. Greg talks about Mova in a different light saying that his company works with effects companies but is not actually one themself. Their specific system is called Contour Digital Capture. It records the precise motion that is needed to make something photo real like Brad Pitt’s older head replacement in “Benjamin Button”. When this information is outputted from their system it is like a raw reconstruction. The result is alot of reference video. It becomes what is known as Data Driven Mesh (DDM). From this information you can create FACS which is normal face shape computer systems but ones that operate in real time. This way the expressions are more natural and you are able to get subtle and accurate skin, bone and muscle positions. They have also recently integarted invisible make up which lets the pores and ridges of the face shine through on the scan. Mova is a subsisdiary on On Live which is a new video game avatar system that will be able to be optimized in the future. The test we saw has divisive possibilities in many different sectors. This is not a rendered scan operation but instead it would have scans of people filtered through an engine running in literal real time.

From that, Steve Preeg, Oscar winner as an animation supervisor for “Benjamin Button” last year from Digital Domain, talked about what Mo-Cap allowed them to do on that specific movie as a workshop example. In “Benjamin Button”, they had to do 325 body double head replacements in 325 shots over 52 minutes. The big obstacles and goals was locking Pitt’s performance to the body double’s head motion, making sure the body motion adhered to the dialogue timing and then that the eye line was right. They tended to use a blendshape route which allows for linear transformations between individual shapes. They originally thought about doing emotional study and creating some elements ala Andy Serkis in “King Kong” but that was out of and beyond the budget. What they did was strap Pitt into the Mova system. Unlike the dot system, Pitt’s face was covered with a green paste which allowed a more infinite scan of his skin. Preeg said this allowed much more pore detail to come through. The problem can be stablization because nothing at any time in this process is stable. They actually had to build a plug in for Maya to deal with this issue.


The fear was that at some times the wrong part moved in simulation or if Pitt couldn’t hit that particular point. The other concern is that there is no data captured right around the eyes. They had to create an eye rig to match that. The hardest part of the eye to recreate, according to Preeg, was the goopy part where there is an angle of reflection. Add these elements to the onset data capture where you have to track 3-4 cameras per shot. You shoot the material on set in layers to get positional information but then you have to track the head markers through the other elements. The onset was the first part. Now several months later you need to get Pitt’s take on Ben. Pitt was strapped into the HD Viper Cam rig. He basically had to angle into the audio keys (what Preeg calls almost “visual ADR”). Add to that equation, “Image Analysis” which brings the timings closer together. The thing is no matter what you do, the computer still can’t get across intent. This is a creative endeavor but there always needs to be an artist behind the notion. This technology will never replace that kind of talent in terms of the actor. A big point was made on this.

Patrick Runyon, Product Specialist at Xsens, brought a real world example of Mo-Cap with a portable system that is used in the industry. Specifically cited was Third Floor who was represented in the pre-vis panel. Their system called MVN uses the continuing trend of flexible capture. It doesn’t require cameras but uses motion trackers and wire frames inside a suit along with scopes and magnetic parameters. Biomechanics eventually comes into play for the aspect of the precise measurements. You just need a laptop, the case with the hardware and the suit to make it work. The onscreen motion set within a 3D grid was real time and showed the practical application of scanning real world movements on the fly in a virtual setting.

In the essence of wrap up, several questions were posed. Rob Readow at Imageworks spoke about future technologies specifically the aspect of “passive digital” where the scan doesn’t need to be in “line of sight”. He said that the aerospace arena is leveraging the data but that currently optical is still the highest fidelity. Readow was also asked about adding stereoscopic elements in post in terms of Mo-Cap and also animation. He says that it is fairly easy to add an extra eye for Stereo 3D as long as they have the other data from the opposite eye complete. He brings up the point that “Polar Express” was originally not in 3D. They were told 3 or 4 months ahead of release that this was a new angle. They got it done but that was not the plan from the beginning.

Denise from Imageworks also was asked about a rig that was supposedly used on Spielberg’s upcoming “Tin Tin” movie which apparently incorporated mounted head cams with the body suits instead of dots. She says that it is just another way to capture all the facial tracking markers. It is simple image recognization. James Cameron’s “Avatar” is using this kind of rig as well. It simply uses a different marker set.

Mo-Cap like Pre-Vis offers more complicated and infinitely fulfilling ways in which to realize many possibilities that couldn’t be done before but also make the realization process a little smoother by allowing virtual worlds to come alive in a more organix way.

For more info on Mo-Cap, visit The Motion Capture Society.