Intel’s New Adaptive Boost Technology for Core i9-K/KF

Taken from our news item

To say that Intel’s turbo levels are complicated to understand is somewhat of an understatement. Trying to teach the difference between the turbo levels to those new to measuring processor performance is an art form in of itself. But here’s our handy guide, taken from our article on the subject.

Adaptive Boost Technology is now the fifth frequency metric Intel uses on its high-end enthusiast grade processors, and another element in Intel’s ever complex ‘Turbo’ family of features. Here’s the list, in case we forget one:

Intel Frequency Levels
Base Frequency - The frequency at which the processor is guaranteed to run under warranty conditions with a power consumption no higher than the TDP rating of the processor.
Turbo Boost 2.0 TB2 When in a turbo mode, this is the defined frequency the cores will run at. TB2 varies with how many cores are being used.
Turbo Boost Max 3.0 TBM3
'Favored Core'
When in a turbo mode, for the best cores on the processor (usually one or two), these will get extra frequency when they are the only cores in use.
Thermally Velocity Boost TVB When in a turbo mode, if the peak thermal temperature detected on the processor is below a given value (70ºC on desktops), then the whole processor will get a frequency boost of +100 MHz. This follows the TB2 frequency tables depending on core loading.
Adaptive Boost Technology ABT
'floating turbo'
When in a turbo mode, if 3 or more cores are active, the processor will attempt to provide the best frequency within the power budget, regardless of the TB2 frequency table. The limit of this frequency is given by TB2 in 2-core mode. ABT overrides TVB when 3 or more cores are active.
*Turbo mode is limited by the turbo power level (PL2) and timing (Tau) of the system. Intel offers recommended guidelines for this, but those guidelines can be overridden (and are routinely ignored) by motherboard manufacturers. Most gaming motherboards will implement an effective ‘infinite’ turbo mode. In this mode, the peak power observed will be the PL2 value. It is worth noting that the 70ºC requirement for TVB is also often ignored, and TVB will be applied whatever the temperature.

Intel provided a slide trying to describe the new ABT, however the diagram is a bit of a mess and doesn’t explain it that well. Here’s the handy AnandTech version.

First up is the Core i7-11700K that AnandTech has already reviewed. This processor has TB2, TBM3, but not TVB or ABT.

The official specifications show that when one to four cores are loaded, when in turbo mode, it will boost to 4.9 GHz. If it is under two cores, the OS will shift the threads onto the favored cores and Turbo Boost Max 3.0 will kick in for 5.0 GHz. More than four core loading will be distributed as above.

On the Core i9-11900, the non-overclocking version, we also get Thermal Velocity Boost which adds another +100 MHz onto every core max turbo, but only if the processor is below 70ºC.

We can see here that the first two cores get both TBM3 (favored core) as well as TVB, which makes those two cores give a bigger jump. In this case, if all eight cores are loaded, the turbo is 4.6 GHz, unless the CPU is under 70ºC, then we get an all-core turbo of 4.7 GHz.

Now move up to the Core i9-11900K or Core i9-11900KF, which are the only two processors with the new floating turbo / Adaptive Boost Technology. Everything beyond two cores changes and TVB no longer applies.

Here we see what looks like a 5.1 GHz all-core turbo, from three cores to eight cores loaded. This is +300 MHz above TVB when all eight cores are loaded. But the reason why I’m calling this a floating turbo is because it is opportunistic.

What this means is that, if all 8 cores are loaded, TB2 means that it will run at 4.7 GHz. If there is power budget and thermal budget, it will attempt 4.8 GHz. If there is more power budget and thermal budget available, it will go to 4.9 GHz, then 5.0 GHz, then 5.1 GHz. The frequency will float as long as it has enough of those budgets to play with, and it will increase/decrease as necessary. This is important as different instructions cause different amounts of power draw and such.

If this sounds familiar, you are not wrong. AMD does the same thing, and they call it Precision Boost 2, and it was introduced in April 2018 with Zen+. AMD applies its floating turbo to all of its processors – Intel is currently limiting floating turbo to only the Core i9-K and Core i9-KF in Core 11th Gen Rocket Lake.

One of the things that we noticed with AMD however is that this floating turbo does increase power draw, especially with AVX/AVX2 workloads. Intel is likely going to see similar increases in power draw. What might be a small saving grace here is that Intel’s frequency jumps are still limited to full 100 MHz steps, whereas AMD can do it on the 25 MHz boundary. This means that Intel has to manage larger steps, and will likely only cross that boundary if it knows it can be maintained for a fixed amount of time. It will be interesting to see if Intel gives the user the ability to change those entry/exit points for Adaptive Boost Technology.

There will be some users who are already familiar with Multi-Core Enhancement / Multi-Core Turbo. This is a feature from some motherboard vendors have, and often enable at default, which lets a processor reach an all-core turbo equal to the single core turbo. That is somewhat similar to ABT, but that was more of a fixed frequency, whereas ABT is a floating turbo design. That being said, some motherboard vendors might still have Multi-Core Enhancement as part of their design anyway, bypassing ABT.

Overall, it’s a performance plus. It makes sense for the users that can also manage the thermals. AMD caught a wind with the feature when it moved to TSMC’s 7nm. I have a feeling that Intel will have to shift to a new manufacturing node to get the best out of ABT, and then we might see the feature on the more mainstream CPUs, as well as becoming default as standard.

Motherboards and Overclocking Support Power Consumption: Caution on Core i9
Comments Locked

279 Comments

View All Comments

  • blppt - Tuesday, March 30, 2021 - link

    I disagree. I had a 9590 (which shipped WITH a small AIO cooler!) and the thing was shaky at best for stability, easily topping 90c at stock settings.

    Not the mobo fault either, I had the top end ASUS CHVF-Z 990FX, which was such a mature chipset it practically had grey hairs.
  • TheinsanegamerN - Wednesday, March 31, 2021 - link

    the 9000 series all had stability issues. Backing off 1 clock bin or tinkering with voltage would usually fix them.

    Bulldozer didnt have the thermal density issues modern CPUs have. If you had the cooling, it would work. Bulldozer's issue was the sheer amount of heat being being generated would overwhelm many CPU coolers of the time, which were built aroudn the more tradiitonal ~100w power draw of intel I7s and the ~125-140 of phenoms. The 200+ that bulldozer was pulling was new territory.
  • Oxford Guy - Wednesday, March 31, 2021 - link

    Certain motherboard makers played loose with the VRMs. AsRock in particular was known for its 9000-series-certified boards frying. MSI was also bad. Only a few boards were suited to the 9000 series and any enthusiast would have skipped the 9000 series in favor of one of the lower-leakage chips, which could be overclocked to the same 4.7 GHz. 5 GHz with Piledriver was not stable, requiring too much voltage. ASUS tried to hide that by under-reporting the voltage used in its flagship board. 4.4 GHz was optimal, 4.5 was okay, and 4.7 was as far as one wanted to go for frequent use. That's with the lower-leakage 'E' parts.

    "The Stilt" said AMD would have sent the 9000 series to the crusher had it not come up with an after-the-fact lower standard for leakage. So, Hruska gets his take spectacularly wrong in his Rocket Lake article. The 9000 series was not aimed at 'the enthusiast faithful'. Those people knew better than to buy a 9000 series chip, even though there were a few astroturfers trying to get people to buy them — like one guy who claimed his was running at 5.1 GHz 24/7.

    It was aimed at people who could be tricked by the 5 Ghz number. It was the most cynical cash grab possible. Not only did AMD offer only 4 FPU cores (important for gaming) it offered a CPU that was priced into the stratosphere while having un-fixable single-core performance.

    Piledriver's fatal flaw was its abysmal single-thread performance, not its power consumption. It could have been okay enough with the lower-leakage standard (and a more strict socket standard as Zen 1 had). But, reportedly, the 32nm SOI wasn't very good for some time (Bulldozer and the first generation of Piledriver), so AMD let the AM3+ spec be pretty loose (although not as loose as FM).

    Overclocking Piledriver even to 5 GHz wasn't enough to give it decent single-thread performance.

    I do have to agree that the 9590 was the single worst consumer CPU product ever released. It even edges out the Pentium III that wasn't stable — since that one was actually pulled from the market. Not only was the 9590 100% cynical exploitation of consumer ignorance, it was really bad technologically. Figures that Hruska would praise it.

    (If, though, one lived in Iceland with a solar array backed by an iron-nickle battery complex, the 9590 would have been okay for playing Deserts of Kharak, provided one didn't buy it at its original price.)
  • blppt - Thursday, April 1, 2021 - link

    "Those people knew better than to buy a 9000 series chip, even though there were a few astroturfers trying to get people to buy them — like one guy who claimed his was running at 5.1 GHz 24/7."

    What is especially sad here is that even IF he managed to pump the 250-300W into that 9590 to run at 5.1 (all cores), it was probably still slower than a 4790K at stock speeds.
  • Oxford Guy - Saturday, April 3, 2021 - link

    In single core, certainly. However, 2011 is stamped onto the spreaders of Piledriver and it hit the market in 2012. The 4790K hit the market in Q2 2014.

    In 2014, the only FX to consider was the 8320E. Not only was it cheap (at least at MicroCenter), it could run in any AM3+ board without killing it — and could be overclocked better than a 9000 series with anything below nitrogen, due to its much superior leakage.

    The 8320E was the only FX worth anyone’s time. Paired with a UD3P board it could do 4.4 GHz readily and could manage 4.7 with a fast fan angled at the VRM sink. Total cost was very low for the CPU and board from MicroCenter, which is why I recommended that setup to the tightest budget people. But, the bad single core was a problem for frametime consistency.

    AMD should have been publicly tarred and feathered by the tech press for the 9590. All the light mockery wasn’t enough.
  • Spunjji - Friday, April 9, 2021 - link

    Broadly agreed, but I'd note that the 6300 was also reasonable if you were on a painfully low budget. I suggested it to a friend (his alternative was a Sandy Bridge i3) and it lasted him until a year back as his main gaming system. It's now moved on to another friend, who still uses it for games. Those chips have aged surprisingly well, all things considered, though it is probably holding his RX 470 back a little bit.
  • Oxford Guy - Wednesday, March 31, 2021 - link

    • The 9590 posted the highest results in the game Deserts of Kharak, in a dual 980 Ti setup at only 1080 or 1440. And, SLI setups showed competitive 4K scores for many games back then.

    • The overclocked 'The Stilt' said the 9000 series is not the chip to judge the design by because it has the worst leakage characteristics and would have been sent to the crusher had AMD not decided to create a lower standard after the fact. Instead, the chips that should be used to represent Piledriver are the 'E' series. They have the lowest leakage and can manage the same 4.7 GHz the 9590 uses with much more reasonable (although still non-competitive) demands. The 9000 series was really AMD's gift to Intel, by making the bad ancient Piledriver design look much worse.

    • AMD was a small cash-strapped company, thanks to Intel's monopoly abuses. When AMD was leading the x86 industry Intel kept it from getting the profit. So, Piledriver, although very bad in a number of ways, will never be as bad as Rocket Lake. The 9000 series is the only exception, though, since it was a purely cynical cash grab by AMD, using '5 GHz' to sucker people.
  • blppt - Thursday, April 1, 2021 - link

    "The 9590 posted the highest results in the game Deserts of Kharak, in a dual 980 Ti setup at only 1080 or 1440. And, SLI setups showed competitive 4K scores for many games back then."

    As I stated, in the (exceedingly rare) case where a game or app can saturate all 8 cores, when the 9590 was in its prime, it could be competitive.

    That almost never happened, especially in games. About the only 2 I can think of offhand that could do that in the 9590's prime was GTA5 and Company of Heroes 2. And even then, you were using 150+ more watts to get the same or slightly better performance than Intel's high-end quad cores. Along with the required AIO water cooling and required high-end mobo with a beastly VRM setup. As far as I know, only 3 pricey mobos were approved for the 9590, my CHVF-Z, one Gigabyte board, and an ASRock.

    9590 was one of the worst cpus ever. Probably the single worst (special edition) cpu. I had one for years.

    This rocket lake, while disappointing, hot, and power consuming, is consistently competitive in every game versus its direct competitors. The 9590 cannot come close to saying that.
  • Oxford Guy - Saturday, April 3, 2021 - link

    I cite Desert of Kharak because it’s the only game I’ve seen put the FX ahead of Intel at below 4K.

    Not only would the game need to be able to leverage 8 integer cores without needing more than 4 FPU cores, it would have to be able to saturate a narrow deep pipeline and not rely heavily on single thread IPC. It should also scale with clock and not need the best RAM and L3 performance. RTS is probably the best genre for the Piledriver design.
  • Gondalf - Tuesday, March 30, 2021 - link

    AMD FX-9590 had not AVX-512. Very high performance have a cost.
    Try to image Zen 3 with AVX-512, it could not be a champion in low power consumption at all.

    If you do not like high power draw, simply disable AVX-512.

Log in

Don't have an account? Sign up now