Star Wars Battlefront 2 Visual Effects Community Transmission


The latest community transmission explores Star Wars Battlefront 2 visual effects, the ways in which they impact the game, and how they are achieved. this community transmission came from Global Community Lead for Star Wars Battlefront 2 on Reddit, F8RGE.

From explosions and blasters, to lighting and rain, visual effects, VFX for short, are present in every video game. How they fit into a title isn’t always explored, but with a game such as Star Wars Battlefront 2 it’s impossible not to be in awe of the incredible visuals they create. The Star Wars franchise in particular has some of the best VFX in the industry, though they’re a double-edged sword at times. On one hand, the films provide a visual reference for how the games need to look, meaning that the true work is in replicating the movie experience, not making a new one. On the other hand, the games need to replicate the films, meaning there is little room for deviation from what is already established in this beloved franchise.

Sometimes it’s hard to pin down where certain VFX begin and and, but luckily Anders Egleus, Senior VFX Artist on Star Wars Battlefront 2, has given some insight into them. He explained, “Most people associate VFX with sprites, i.e. flat planes with transparent textures. But modern game engines like Frostbite have become so flexible and powerful that we can now use our VFX tools to create a lot of things that traditionally would have to be created by other parts of the game team, e.g. animation, lighting, technical art, even code. From birds and leaves on Naboo, to full scenes of background fighting with vehicles and hundreds of soldiers on Geonosis, a lot of people don’t even think of these elements as effects. Point being, our job is anything but repetitive, and because we dabble in many different areas of the game, we get to collaborate with a lot of talented artists, designers and coders too, which is super rewarding.

Egleus went on to describe the roles that VFX play in certain parts of each game.

Gameplay Effects

This type of effects are there to communicate gameplay mechanics and give feedback to the player when you do something, or something happens. Examples are weapon impacts, muzzle flashes, blaster bolts, destruction elements, lightsabers, character/vehicle abilities and UI-like effects. These all have to work well in many different situations and lighting conditions. Since visual fidelity is such an important aspect of Star Wars, and by extension Battlefront II, they also need to stay true to the source material, look great, and feel “Star Wars-y.” Creating these types of effects often involves many iterations until all of the above requirements are satisfied.

Environment Effects

Most Levels (maps) are embellished by different types of environment effects; Fog, fire, smoke, rain, snow, swirling leaves, and more all help the world come alive, but also give visual cues to help you understand your surroundings.

Narrative Effects

Cinematics, and many Battle Beyond scenes, often use the same types of effects as the other two categories, but the story telling aspect is the most important part, and the setup is more similar to traditional keyframe animation in that effects are triggered by a timed sequence rather than resulting from triggered events.

Capital Supremacy contains many cinematic and narrative elements. Needless to say, many effects fall into more than one, even all three categories, like the dust storm on Geonosis or the fires in the scene below. There’s a gameplay aspect in that they provide cover and sight reduction, risk/reward decisions (i.e. Do I take a shortcut and take some damage or do I go around?); they certainly enhance the environment art; and they tell a story about a battle that has been raging before you got there.

What all these types of effects also have in common is, of course, to help immerse the player into the world and support the artistic vision of the game.

Visual Targets

Before starting work on an effect, it’s important to have a good reference of the desired end result. This helps identifying details that would otherwise be missed and facilitates communication with the art director, designers, and other stakeholders.

Elaborating more on this point, Egleus went on to say, “The Star Wars franchise is extremely gratifying because you almost always have access to good reference material, and because the visual style is so distinct. There’s of course a flipside to that: You’re expected to nail the look of the movies with every effect. Other IPs might not come with the same baggage of what things should look like, but I prefer that any day to an unclear vision. Most of the time it’s easy to find the reference we need, like in the case of the Droideka’s shield which appears in full glory at the start of Episode I. Other times we take help from our in-house Star Wars gurus like CJ and Guillaume to dig up just the right scene from this or that episode of The Clone Wars.

The collaboration with Lucasfilm is also crucial to the result. They often suggest things we would never have thought of. They provide us with the needed reference material. They’ve even sent us raw footage from practical on-set explosions and other special effects, and we constantly rely on their expertise for feedback and advice.

Another source of reference is the work of our talented concept art team. The concept art establishes a clear vision and allows us to start thinking about potential challenges early on. We can look at a concept image and go ‘yeah, no, that’s never gonna fit into our performance budget’, but then you go back and start thinking about the challenge and eventually find a way to cram it all in there.

Just Like the Simulations

So how are VFX made in a game like Star Wars Battlefront II? Unlike other artists who use programs like Photoshop and Maya to create their assets, the main tool of the VFX team is the game engine itself. The most common part of an effect is the particle system, or emitter. A particle is just a point in space with properties like size, rotation, color, and transparency. Unlike traditional keyframe animation, particles are then usually simulated, which (grossly oversimplified) could be described as the game engine applying basic physics to them as they evolve over time. The artist can then vary these physical properties (e.g. gravity, air resistance, etc.) to get the desired result.

The final effect is typically made up of multiple emitters, lights, and other components. Often, many different effects are needed to make up a gameplay feature, environment. or cinematic event.

Lighting Effects

When it comes to making effects fit into the rest of the game world, there’s probably nothing more important than their interaction with lighting. For this to work, and be fast enough to calculate for all target platforms, the engine combines cheap per-vertex lighting with approximated texture based volumetric lighting. A simplified forward scattering model is also used to get nicely backlit particles. To give the right look to fire and other self-illuminating materials, a so-called blackbody calculation converts grey scale images into believable fire colors and intensities.

New Technology

Star Wars Battlefront 2 was the fist Frostbite game to ship with a new GPU particle technology, in addition to the standard particles Egleus has covered so far. Recalling the moment when he and his team were approached to collaborate on developing this new system he exclaimed, “Happy to be your guinea pigs! A lot of the effects in the game simply couldn’t have been made without it. GPU particles are much cheaper than their old CPU counter parts so you can have lots and lots more of them. Also they’re programmable, so that allows for much more complexity and control. The downside of that though is that they’re a lot harder to make because we have to build everything from scratch, so we only use them where the old system can’t deliver what we need (e.g. background crowds). Some other examples of effects created with the new GPU particle system are rain, snow, sparks, embers, pebbles, metal debris, leaves, insects (and Whisties!), lightning/electricity, birds, blaster bolts, and, of course, lightsaber blades.

Constructing a New Lightsaber

Lightsabers, and specifically lightsaber blades, are some of the most iconic effects in the Star Wars universe. In Star Wars Battlefront, the shape and colors of the blade are procedural, which means that instead of using a texture to define the blade, different math functions are combined in a so-called “pixel shader” which gets called (i.e. activated) when the blade is drawn to the screen (i.e. ignited). This allows for visual tricks like making them look like they have a volume even though they’re drawn on flat planes.

However, Egleus went on to explain why creating lightsabers in Star Wars Battlefront 2 wasn’t quite as straightforward as the team first thought. “We used basically the same technique in both games, but it turned out to be a lot harder to get it right in the sequel. One reason for this was simply time constraint. There were more lightsaber wielders in the second game (the first one had only Luke and Vader), and they were more diverse. We weren’t 100% happy with the look, but it kind of worked so we figured let’s move on and build all the other effects needed for the game.

Another reason was that the differences between light settings were much bigger in the sequel, making it more difficult to keep consistency across all lighting conditions. Finally, stretch bloom (horizontal glow), which was introduced to mimic the look of The Force Awakens, became very exaggerated with very bright objects. As a result, the intensities of the blades had to be toned down a lot, since glow is applied equally to everything on screen based on the intensities of the pixels.” Egleus concluded that, “Ironically, once we decided to remove stretch bloom in a post launch patch, it paved the way for a look that was more true to the movies.

The motion fan of lightsabers also proved problematic to Egleus and the team.


Traditionally you would rely on the built-in motion blur of the game engine, which we tried during the early production of the first game. This proved problematic, however, as the blade would have to be a solid object (like Phasma’s staff) instead of the pulsing cylindrical volume we all know and love. Moreover, the framerate in a game is much higher than in a film and can be even higher on powerful PCs. Therefore, the amount of motion blur would be much less than expected and didn’t look right. The answer was to connect planes, which, much like an accordion, would stretch from the current position to a fixed time offset (say 1/60 seconds ago).

An additional problem was that, before the June update, the back rotation was stored in world space, so sometimes you’d get a lot more motion blur than you’d expect, e.g. when rotating with your character. The blade didn’t move at all in camera space, but rotated quite a lot in world space. Now we store it in camera space which is closer to how real camera motion blur works. It’s a bit painful that it’s taken so long, we tend to prioritize new content over polishing existing assets, but it’s nice to finally be at a point (no pun intended) were we’re happy with the lightsaber blades. We will of course continue to improve lightsabers and many other effects in future updates.

This is where the community transmission ended what is a fairly comprehensive look into the VFX of Star Wars Battlefront 2, and just how they’re implemented.

Since the game launched, Star Wars Battlefront 2 has faced backlash from critics and players alike. Microtransactions almost ruined it, and a lack of vision seemed to leave it dead in the water. However, the game has a clear roadmap with content still ahead of it, and those players who are still enjoying it every day are very much looking forward to getting a new map this fall.

Source: Reddit

Previous Post

Next Post

Top Games and Upcoming Releases