I am very excited about C++26 reflection.
I am also obsessed by having my code compile as quickly as possible. Fast compilation times are extremely valuable to keep iteration times low, productivity and motivation high, and to quickly see the impact of your changes. 1
With time and experience, I’ve realized that C++ can be an extremely fast-to-compile language. Language features like templates are not the issue – the Standard Library is.
My fork of SFML uses almost no Standard Library at all, and I can recompile the entire thing from scratch in ~4.3s. That’s around ~900 TUs, including external dependencies, tests, and examples.2 Incremental builds are, for all intents and purposes, instantaneous. I love it.
I would love to live in a world where C++26 reflection is purely a lightweight language feature, however that ship has sailed (thank you, Jonathan Müller, for trying).
In this article, I’ll try to provide some early expectations about the compile-time impact of C++26 reflection.
an update and an apology
Shortly after publishing the initial version of this article, a reader pointed out a critical issue in my benchmarks: the third-party Docker container I originally used for GCC 16 was compiled with internal compiler assertions enabled3, as that is the default build configuration.
This unfairly penalized GCC, making the compiler look slower than it actually is, and skewed the benchmark data.
I sincerely apologize to the GCC team and my readers for my mistake –
I should have verified the container’s GCC build flags before
publishing, as I had assumed that GCC had been built in release mode. I
have since re-run all benchmarks on a proper release build (using
fedora:44 as
suggested by Jonathan Wakely with
--enable-checking=release), and the article below has been
fully updated to reflect the accurate (faster) numbers.
To ensure previous readers are informed about the new data, I will manually go through every comment on every platform and alert users about any incorrect number/conclusions.
I have also uploaded all benchmark files on GitHub, so that you can easily reproduce my findings on your own setups.
let’s measure!
For the benchmarks below, I used a Fedora 44 Docker container with
GCC 16, the first version that supports reflection. To get reasonably
stable measurements, I used the hyperfine
command-line benchmarking tool.
These are my specs:
- CPU: 13th Gen Intel Core i9-13900K
- RAM: 32GB (2x16GB) DDR5-6400 CL32
- OS: Fedora 44 (with the official Docker image)
- Compiler: gcc version 16.0.1 20260305 (Red Hat 16.0.1-0)
- Flags:
-std=c++26 -freflection
My test scenarios (available on GitHub) were as follows:
Baseline test. Just a
int main() { }.Header inclusion test. Same as above, but with
#include <meta>.Basic reflection over a struct’s fields:
template <typename T> void reflect_struct(const T& obj) { template for (constexpr std::meta::info field : std::define_static_array(std::meta::nonstatic_data_members_of( ^^T, std::meta::access_context::current()))) { use(std::meta::identifier_of(field)); use(obj.[:field:]); } } struct User { std::string_view name; int age; bool active; }; int main() { reflect_struct(User{.name = "Alice", .age = 30, .active = true}); }Barry Revzin’s AoS to SoA transformation example, from his blog post.4
I’ve also tested using precompiled headers (PCHs) for
<meta> and other large dependencies.
⚠️ DISCLAIMER: please take these benchmark results with a grain of salt. ⚠️
- My measurements are not that rigorous and that the compiler I used is still work-in-progress.
- Also note that my specs are quite beefy – YMMV.
- Finally, remember that these measurements are for a single translation unit – in a real project, you’d have to multiply the compile time overhead by the number of affected TUs.
benchmark results
| # | Scenario | Code | Precompiled Header (PCH) | Compile Time (Mean) |
|---|---|---|---|---|
| 1 | Baseline (No Reflection Flag) | int main() only |
None | 33.2 ms |
| 2 | Baseline +
-freflection |
int main() only |
None | 33.9 ms |
| 3 | <meta> Header
Inclusion |
int main() +
#include <meta> |
None | 187.2 ms |
| 4 | Basic Struct Reflection (1 type) | reflect_struct with
User |
None | 200.3 ms |
| 5 | Basic Struct Reflection (10 types) | reflect_struct with
User<N> |
None | 210.0 ms |
| 6 | Basic Struct Reflection (20 types) | reflect_struct with
User<N> |
None | 227.4 ms |
| 7 | Basic Struct Reflection (1 type + PCH) | reflect_struct with
User |
<meta> |
91.7 ms |
| 8 | Basic Struct Reflection (10 types + PCH) | reflect_struct with
User<N> |
<meta> |
96.9 ms |
| 9 | Basic Struct Reflection (20 types + PCH) | reflect_struct with
User<N> |
<meta> |
109.3 ms |
| 10 | AoS to SoA (Original) | Barry Revzin’s Unedited Code | None | 818.9 ms |
| 11 | AoS to SoA (No Print) | Removed <print> |
None | 310.2 ms |
| 12 | AoS to SoA (No Print/Ranges) | Removed <print>,
<ranges> |
None | 224.4 ms |
| 13 | AoS to SoA (Original + PCH) | Barry Revzin’s Unedited Code | <meta>,
<ranges>, <print> |
628.0 ms |
| 14 | AoS to SoA (No Print + PCH) | Removed <print> |
<meta>,
<ranges> |
137.4 ms |
| 15 | AoS to SoA (No Print/Ranges + PCH) | Removed <print>,
<ranges> |
<meta> |
113.7 ms |
A few clarifications:
The number of types in the “Basic Struct Reflection” example was adjusted by instantiating unique “clones” of the test struct type:
template <int> struct User { std::string_view name; int age; bool active; }; reflect_struct(User<0>{.name = "Alice", .age = 30, .active = true}); reflect_struct(User<1>{.name = "Alice", .age = 30, .active = true}); reflect_struct(User<2>{.name = "Alice", .age = 30, .active = true}); // ...“Removed
<print>” implies removing all of the formatting/printing code from Barry’s example.“Removed
<ranges>” implies rewriting range-based code likestd::views::transformorstd::views::iotato good old boomer loops™.
insights
- The reflection feature flag itself is free.
- Simply turning on
-freflectionadds 0 ms of overhead.
- Simply turning on
- Basic reflection seems to scale reasonably well.
- Base cost of reflecting 1 struct: 200.3 ms (but
~187 ms of this is just including
<meta>). The actual reflection logic only costs ~13.1 ms. - Cost to reflect 9 extra types: +9.7 ms (~1.1 ms per type).
- Cost to reflect 10 more types (20 total): +17.4 ms (~1.7 ms per type).
- While these numbers are promising, remember that my example is extremely basic, and that a large project can have hundreds (if not thousands) of types that will be reflected.
- Base cost of reflecting 1 struct: 200.3 ms (but
~187 ms of this is just including
- Standard Library Headers are a big bottleneck.
- The massive compile times in modern C++ don’t come from your metaprogramming logic, but from parsing standard library headers.
- Pulling in
<meta>adds ~155.2 ms of pure parsing time over the baseline. - Pulling in
<print>and its formatting logic adds an astronomical ~508.7 ms (Scenario 10 vs 11).
- Precompiled Headers (PCH) are mandatory for
scaling.
- Caching
<meta>and avoiding heavy dependencies cuts the AoS compile time down to 113.7 ms (scenario 15). - Caching
<meta>+<ranges>drops the time from 310.2 ms to 137.4 ms (scenario 11 vs 14). - Interestingly, while caching
<print>helps, it still leaves the compile time uncomfortably high (628.0 ms).
- Caching
what about modules?
I ran some more measurements using import std; with a
properly-built std module that includes reflection.
Firstly, I created the module via:
g++ -std=c++26 -fmodules -freflection -fsearch-include-path -fmodule-only -c bits/std.ccAnd then benchmarked with:
hyperfine "g++ -std=c++26 -fmodules -freflection ./main.cpp"No #include was used – only import std.
These are the results (mean compilation time):
| Scenario | With Modules | With PCH |
|---|---|---|
| Basic Struct Reflection (1 type) | 279.5 ms | 91.7 ms |
| AoS to SoA (No Print/Ranges) | 301.9 ms | 113.7 ms |
| AoS to SoA (Original) | 605.7 ms | 628.0 ms |
At the current moment in time it seems that PCH drastically
outperforms modules for smaller headers like <meta>,
while pretty much ties modules with large includes such as
<print>.
conclusion
Reflection is going to bring a lot of power to C++26. New libraries
that heavily rely on reflection are going to become widespread. Every
single TU including one of those libraries will virally include
<meta> and any other used dependencies.
Assuming the usage of <meta> +
<ranges> becomes widespread, we’re looking at a bare
minimum of ~310ms compilation overhead per TU. Using
PCHs (or modules) will become pretty much mandatory, especially in large
projects.
I really, really wish that Jonathan Müller’s paper (P3429: <meta> should
minimize standard library dependencies) was given more thought and
support.
I also really wish that a game-changing feature such as reflection wasn’t so closely tied to the Standard Library. The less often I use the Standard Library, the more enjoyable and productive I find C++ as a language – insanely fast compilation times are a large part of that.
Hopefully, as reflection implementations are relatively new, things will only get better from here.
shameless self-promotion
I offer training, mentoring, and consulting services. If you are interested, check out romeo.training, alternatively you can reach out at
mail (at) vittorioromeo (dot) comor on Twitter.Check out my newly-released game on Steam: BubbleByte – it’s only $3.99 and one of those games that you can either play actively as a timewaster or more passively to keep you company in the background while you do something else.
My book “Embracing Modern C++ Safely” is available from all major resellers.
- For more information, read this interview: “Why 4 Bloomberg engineers wrote another C++ book”
If you enjoy fast-paced open-source arcade games with user-created content, check out Open Hexagon, my VRSFML-powered game available on Steam and on itch.io.
- Open Hexagon is a community-driven spiritual successor to Terry Cavanagh’s critically acclaimed Super Hexagon.
appendix: previous benchmark results
For transparency, and as a cautionary tale for using third-party Docker images, here is a direct comparison between my original (flawed) measurements and the new ones.
These are the benchmark results from the previous version of the article, where I accidentally used a version of GCC 16 compiled with internal assertions enabled.
| Scenario | Old Time (Checking) | New Time (Release) | Speedup (Difference) |
|---|---|---|---|
Baseline +
-freflection |
43.1 ms | 32.0 ms | -11.1 ms (~25% faster) |
<meta> Header
Inclusion |
310.4 ms | 187.2 ms | -123.2 ms (~40% faster) |
| Basic Struct Reflection (1 type) | 331.2 ms | 200.3 ms | -130.9 ms (~40% faster) |
| Basic Struct Reflection (10 types) | 388.6 ms | 210.0 ms | -178.6 ms (~45% faster) |
| Basic Struct Reflection (20 types) | 410.9 ms | 227.4 ms | -183.5 ms (~45% faster) |
| AoS to SoA (Original) | 1,622.0 ms | 818.9 ms | -803.1 ms (~50% faster) |
| AoS to SoA (No Print) | 540.1 ms | 310.2 ms | -229.9 ms (~42% faster) |
| AoS to SoA (No Print/Ranges) | 391.4 ms | 224.4 ms | -167.0 ms (~42% faster) |
| AoS to SoA (Original + PCH) | 1,265.0 ms | 628.0 ms | -637.0 ms (~50% faster) |
| AoS to SoA (No Print + PCH) | 229.7 ms | 137.4 ms | -92.3 ms (~40% faster) |
| AoS to SoA (No Print/Ranges + PCH) | 181.9 ms | 113.7 ms | -68.2 ms (~37% faster) |
The checking assertions absolutely crippled C++23 module performance in the original run:
- Modules (Basic 1 type): from 352.8 ms to 279.5 ms (-73.3 ms)
- Modules (AoS Original): from 1,077.0 ms to 605.7 ms (-471.3 ms, ~43% faster)
The current bottlenecks are (by far)
cpptraceandnlohmann::json, used only in a few TUs (translation units). Without those, I could easily get a sub 1s build. I could get even lower with unity builds and perhaps some well-chosen PCHs.↩︎I opened an issue to inform the authors of the Docker image about that: https://github.com/SourceMation/images/issues/242↩︎
If you haven’t already, check out his great CppCon 2025 talk: “Practical Reflection With C++26”, and perhaps my “More Speed & Simplicity: Practical Data-Oriented Design in C++” keynote as well!↩︎