This page has not yet been reviewed for compatibility with version 0.9.0. Details may now be incorrect.
What is performance?
Lets get one thing clear: Indigo is not the fastest 2D game engine out there. Not even close. We haven't checked but it might even be the slowest - who knows! Nonetheless, here's a little demo merrily chugging along at 60 frames per second (performance will no doubt vary by system spec):
Performance is relative. When people talk about game engine performance, they are usually talking about how many things it can do at once.
The way raw engine performance is typically achieved is by utilizing clever data structures, mutable data stores, reusing allocated memory to keep usage constant, insisting on never creating new objects, and utilizing multi-threading.
All of that can lead to higher performance, but it takes careful management since your guarantees around runtime behavior are reduced, and bugs can occur.
In order to bring performance up, Indigo itself does all this too. Where Indigo and normal game engines diverge is that while these techniques are used by Indigo under the covers, it is strongly discouraged at the API level. Why is that?
What if you could build games with an expectation of no bugs? Games that you know work before you play them because you can bring to bear an advanced type system, refinement types, property based testing, TDD (without excessive mocking), and any other code quality tool you can think of?
Couldn't developer productivity be considered a type of performance too? The design of Indigo is to take an engine performance hit at the API level, in order to increase developer performance (easy to reason about, easy to test) and game production efficiency.
Please note: You're still going to have to play your game at some point. No automated test suite in the world will tell you if you've made something fun or not!
Is it fast enough?
Having establishing that you're never going to get a billion triangles though Indigo - what can you expect? Is the performance "good enough"?
Well, two slightly disappointing observations that we've made about 2D games during Indigo's development:
- Most 2D games on the market don't actually have that many elements on screen at any one time - there's only so much screen real estate available for use when you have no real perspective or depth. (Naturally there are exceptions)
- Most game logic is quite straight forward, and revolve around a few core mechanics. The complexity is in how the elements interact.
With that in mind then: Indigo is fast enough that you should not have to do anything clever or special to be able to hit a respectable frame rate for most games.
Start on the assumption that it will be fast enough
In general, you should start by assuming Indigo will be fast enough. Build your game using good functional programming practices, and write your tests.
You should only consider changing things if it really isn't fast enough when you play / performance test it. Remember that it's always easier to make clean, well written code go faster, than trying to improve the performance of something messy.
One weird trick to grind Indigo to a halt
Two sort of similar features that Indigo does not have are:
- GPU rendered particle systems;
- Shader based effects like light trails.
Both of these cases can only be simulated on the CPU side, and attempting to do so with very heavy numbers of "particles" will slow Indigo right down, particularly if you use a naive solution.
How to get more speed
Depending on what kind of things you're doing, you can make Indigo do more in less time.
Your main enemy in the quest for performance is memory allocation and subsequent garbage collection pauses.
The near universal mantra of game developers is to never allocate memory, i.e. create new objects, during a frame. Ever. Reuse, reuse, reuse. Unfortunately, Indigo is built on Scala offering up an immutable API, so we're going to be making new objects all the time.
Generally your performance will suffer the more things you add to the screen. Our goal is to get the same effect you get by adding lots of things, but by doing less work. Here are a few tips for squeezing out some extra juice.
Measure twice, cut once
Be Aware: Running profiling tooling is, itself, expensive! You're game might do 58 FPS normally and 47 under profiling. Try plugging in the FPS Counter
SubSystemfor another view.
Modern browsers have amazing profiling tools built into them these days. Get to know them by recording the performance of your running game, and then look for evidence of where you're creating things like GC pauses or large numbers of allocations, and try to track down the culprit code.
If you aren't improving the slowest bottleneck of your game code, you aren't improving anything.
Tip: In Firefox make sure you check the "Record Allocations" setting in the performance tab.
If you can't see it, don't draw it
Normally when you create a game element like a sprite in a conventional game engine, they hang around until you remove them. Indigo only draws what your
SceneUpdateFragment describes right now. As such, Indigo assumes you know what you're doing and doesn't try to do any clever interpretation of your scene*.
*Culling at the render level will occur, but that won't prevent the processing of your scene elements.
Normal practice is to perform "Viewing-frustum culling" to remove things that aren't visible on the screen i.e it's "off camera".
Render batch size
WebGL 2.0 only!
There is a batch size option in the main game config that you can tweak, it relates to the maximum number of renderable items that are bundled up before a draw call is issued. It's a trade off: Bigger batches mean more memory usage but fewer calls to the graphics card.
In general fiddling will only affect games with a lot of scene elements.
Automata are great fun, and a very convenient way to manage and program short lived little entities with simple life cycles e.g. the points that appear above a characters head when they pick up a coin.
However, they can allocate relatively large amounts of memory because the structures they use to make them fun to use, are not the most light weight way to solve the problem. As a rule of thumb: A few hundred are probably fine.
Tweak the pool size
A simple way to improve performance of automata - depending on the circumstances - is to reduce the default pool size (
None by default) that controls how many automaton instances can be alive at any one time. If your effect isn't mission critical, try limiting the pool size.
SubSystems vs Automata
The Automata system is just a SubSystem. If you need simple particles in high volumes and are prepared to manage the state yourself, you can write much leaner code by hand crafting a
SubSystem, especially in combination with clones.
WebGL 2.0 only!
Clones are Indigo's version of what is known as "instancing". Say you want to render lots (lots and lots and lots!) of things that are more or less identical, such as blades of grass or tiles in a tile map, you should consider using clones.
With clones, you set up a reference object (that you can update as normal), and then tell indigo to render many more of them. This shortcuts a lot of processing in the pipeline and allows for vast numbers of copies to be drawn at the cost of a lack of variety, although a limited set of properties can be changed.
Clone batches can also be declared static for even more performance, if they never change.
A recent test ran 10,000 animated clones at 60fps alongside other screen elements, which would not be possible with other types of primitives.
May seem obvious but some values are just expensive to work out. Object boundaries for example, particularly text boundaries or groups with many elements, can be quite expensive to calculate. If you (as the game programmer) know that a value is going to be used a lot but is never going to need to be recalculated, consider storing it and looking it up next time. In the example of boundaries, a good place to store it might be as part of the view model.
Consider the cost of different primitives
Different types of primitives have different costs, here they are ranked from cheapest to (potentially) most expensive:
- Clone - Unintelligent copies of things.
- Graphic - bounds require no calculation, they have no clever inner workings and process no events.
- Sprite - processes events, has animation state to update, bounds must be recalculated each time.
- Text - no state, but processes events which are based on ver expensive bounds calculations.
- Group - no state, no events, but bounds calculation is based on the contents - so can be big.
One way to reduce the cost of animated elements is to reuse them!
Example scenario: You have a number of background elements, and it's acceptable for all of them to be on the same animation loop at the exact same frame position (e.g. you have a forest and 25% of the trees are identical), but you want to be able to interact with them individually (thus ruling out clones).
- Establish a master sprite;
- Give all of the sprites the same animation key;
- Only update the animation of the master sprite;
- All the others will be animated identically but with no additional event or animation processing cost.
Manually cache groups
There is an open issue to look into formal ways to speed up static groups.
If you have a big
Group of static things, consider caching it in your view model and reusing it.
Manually accept interactions on Graphics
Graphics do not accept events / interactions in the same way that Sprites and Text do. This is to reduce the processing cost of graphics, since event handling can be ignored for
Graphics during the scene graph traversal, allowing you to have a lot more of them on the screen.
For one off interactive elements though, you can avoid using a Sprite by mimicking the event handler behavior manually. For example, you can check the mouse position and whether or not it was clicked within graphic node's very cheap to calculate bounds during the view presentation function.
In terms of view processing, memory allocations are typically the problem, but you can also experience CPU bottlenecks depending on the kind of game you're building.
Use appropriate data structures
In many cases, a simple array of things to process will do. However, just as an example, if you're processing something spacial like collision detection or available moves on large playing grid, you should look for data structures that can efficiently ignore / avoid processing irrelevant areas, such as a BSP or Quadtree structures.
Split expensive calculation work over multiple frames.
You're writing a farming game and have a massive grid of crops to update. You can certainly reduce the drawing overhead using some of the techniques previously mentioned, but what about updating them all?
To help with the time element, you could use a
TimeVaryingValue that will automatically advance the value of the crops for a given time. However, updating all the elements is expensive - perhaps your calculation has to decide what's happening to the crop based on many environmental conditions?
If you wrap your crops up in an
UpdateList you can specific an update pattern, for example you could update 25% of the crops this frame, and 25% on each of the next three subsequent frames until they're all done. As long as your calculation is time based, and the accuracy of when you need to know the crops are ready is acceptable to be within 4 frames of the actual completion time, you can quarter your per frame processing costs.
Note that you're still allocating for the whole grid in this scenario! Consider combining these with better data structures as discussed above for further gains.
Don't fear mutability
Your last angle of attack is to use mutable data and imperative programming techniques.
Scala is an impure functional language and there is nothing wrong with taking advantage of that. You should always pull for purity and immutability first, but games by their very nature are always pushing the resource constraints of their system in one way or another, and finding game performance is about making trade-offs. Sometimes the trade off is your purely functional sensibilities!
If you've measured and identified an area of your code that is causing a bottleneck, sometimes the best solution is to roll up your sleeves and use a more imperative solution to do a bit of specific, localized, optimization. Never ever do this without profiling your code first, or you're probably wasting your time.
Remember, a function is pure and referentially transparent as long as for a given set of arguments you always get the same result - there are no limits on how you make that happen. Using a while loop and a var is not against the rules! Just write good tests, and use strong encapsulation.