throbber
Power and Thermal Management in the Intel® CoreTM
`Duo Processor
`
`Alon Naveh, Mobility Group, Intel Corporation
`Efraim Rotem, Mobility Group, Intel Corporation
`Avi Mendelson, Mobility Group, Intel Corporation
`Simcha Gochman, Mobility Group, Intel Corporation
`Rajshree Chabukswar, Software Solutions Group, Intel Corporation
` Karthik Krishnan, Software Solutions Group, Intel Corporation
`Arun Kumar, Software Solutions Group, Intel Corporation
`
`Index words: Intel Core Duo, ACPI, power, thermal
`
`
`Comparing the Intel Core Duo processors with previous-
`generation Pentium® M processors [1] reveals that it uses
`newer process technology, runs faster, and uses the same-
`size caches, but the overall die size is increased in order to
`accommodate
`two cores. The use of new process
`technology, together with special design and layout
`optimizations, allows each core of the Intel Core Duo
`technology to control its dynamic power consumption. In
`this paper, therefore, we focus on the CMP-related aspects
`of this technology. Controlling the static power (leakage)
`was found to be a real challenge due to the fact that static
`power depends on the total area of the processor and on
`process technology. Since the overall area was increased
`and the new process was found to produce more leakage
`power, static and dynamic power management became
`one of the main issues when designing the system.
`
`ABSTRACT
`
`The Intel® Core™ Duo processor is the first mobile
`processor that uses Chip Multi-Processing (CMP) on-die
`technology to maximize performance and minimize power
`consumption. This paper provides an overview of the
`power control unit and its impact on the power-saving
`mechanisms. We describe the distribution of P-states and
`look at other techniques used to improve the system power
`consumption including the impact of multi-threading on
`power consumption. Then, we provide recommendations
`on optimizing for CMP and building power-friendly
`applications.
`
`INTRODUCTION
`
`Power and thermal management are becoming more
`challenging than ever before in all segments of computer-
`based systems. While in the server domain, the cost of
`electricity drives the need for low-power systems, in the
`mobility market, we are more concerned about battery life
`and thermal limitations. The increasing demand for
`computational power in conjunction with the relatively
`slow improvement in cooling technology caused power
`and thermal control methods to become primary parts of
`the architecture and the design process of the Intel Core
`Duo processor-based system.
`
`Intel Core Duo is the first general-purpose, multi-core on
`die Chip Multi-Processing (CMP) processor Intel has
`developed for the mobile market. The core was designed
`to achieve two main goals: (1) maximize the performance
`under the thermal limitation the platform allows; and (2)
`improve the battery life of the system relative to previous
`generations of processors.
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`109
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 1 of 15
`
`

`

`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`
`
`Figure 1: ACPI-based idle state management
`
`Optimizing the system for maximum performance at the
`minimum power consumption is usually done as a
`combination of software (operating system) and hardware
`elements. Most modern operating systems (OS) use the
`ACPI [2] standard for optimizing the system in these
`areas. The ACPI processor sleep state control assumes
`that the core can be in different power-saving states (also
`termed sleep states) marked as C0 to Cn. When the core is
`active, it always runs at C0, but when the core is idle, the
`OS tries to maintain a balance between the amount of
`power it can save and the overhead of entering and exiting
`to/from that sleep state. Thus C1 represents the power state
`that has the least power saving but can be switched on and
`off almost immediately, while extended deep-sleep (DC4)
`represents a power state where
`the static power
`consumption is negligible, but the time to enter into this
`state and respond to activity (back to C0) is quite long.
`Figure 1 shows the general principle of operation of how a
`traditional ACPI software layer controls the sleep states of
`the processor. When the OS scheduler detects there are no
`more tasks to run, it transitions the processor into the
`“selected” idle state, by executing a BIOS defined
`instruction sequence. The processor will remain in that
`
`idle state until a break event occurs and then return to the
`C0 state. Break events would typically be interrupts and
`similar indicators that new tasks need to be executed. The
`OS evaluates the CPU activity factor over a registry-based
`time window and, periodically, modifies the selected
`target C-state for the next evaluation window.
`
`The progressive increase in static power savings per state
`is achieved by employing more aggressive means as the
`C-state deepens. In C1 idle state, only processor-centric
`measures are employed: instruction execution is halted
`and the core clocks are gated. In C2 states and above,
`platform-level measures are also added to further increase
`the power savings. While in the C2 state, the processor is
`obligated not to access the bus without chipset consent.
`Thus, the front side bus can be placed in a lower power
`state, and the chipset can also initiate power-saving
`measures. In C3, the processor also disables its internal
`Phase Locked Loops. In C4, the processor also lowers its
`internal voltage level to the point where only content
`retention is possible, but no operations can be done. Deep-
`C4, a new Intel Core Duo power state, is described later in
`this paper.
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`110
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 2 of 15
`
`

`

`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`In order to control the dynamic power consumption, the
`ACPI standard
`tries
`to adjust
`the active power
`consumption to the “computational” needs of the software.
`The system controls the “active” state of the system (also
`termed P-states) by a combination of hardware and
`software
`interfaces:
`the hardware defines a set of
`“operational points” where each one defines a different
`frequency/voltage and therefore different levels of power
`consumption. The OS aims to keep the system at the
`lowest operational point yet still meet the performance
`requirements. In other words, if it anticipates that the
`software can efficiently use the maximum performance
`that the processor can provide, it puts it in the P0 state.
`When the OS predicts (based on history) that the lower
`(and slower) operational point can still meet
`the
`performance requirements, it switches to that operational
`point (marked as Pn) in order to save power.
`
`Power consumption produces heat that needs to be
`removed from the system. Naturally, each system has a
`total cooling capacity per its specific design, and each
`major component has a cooling
`limit or power
`consumption allowance. The system designers usually
`choose the maximum frequency the system can run at
`without overheating when running in a “normal” usage
`model; this is also called the system's Thermal Design
`Power (TDP). Thus, when an unexpected workload is
`used the thermal control logic aims to allow the system to
`provide maximum performance under
`the
`thermal
`constraints.
`
`The remainder of this paper is organized as follows: we
`first describe the Intel Core Duo implementation of power
`and thermal control; next, we provide experimental
`numbers that examine the efficiency of the power and
`thermal mechanisms; and finally we extend the discussion
`to software optimizations that can help save power.
`
`Figure 2: Internal power state in Intel Core Duo processor
`
`POWER AND THERMAL
`MANAGEMENT
`
`Power and thermal management of the Intel Core Duo
`processor raise a new challenge to the ACPI-based
`architecture since, from a software point of view, there is
`no difference between running under a CMP-based system
`and running under dual-core architecture, but from a
`hardware point of view, even though most of the logic is
`duplicated, major parts of the logic such as the power
`supply and the L2 cache are still shared. For example, in a
`Dual-Processor system, when the OS decides to reduce the
`frequency of a single core, the other core can still run at
`full speed. In the Intel Core Duo system, however,
`lowering the frequency to one core slows down the other
`core as well.
`
`C-state Architecture
`
`Since the OS views the Intel Core Duo processor as two
`independent execution units, and the platform views the
`whole processor as a single entity for all power-
`management related activities (C2 state and beyond), we
`chose to separate the core C-state control mechanism from
`that of the full CPU and platform C-state control.
`
`This was achieved by making the power and thermal
`control unit part of the core logic and not part of the
`chipset as before. Migration of the power and thermal
`management flow into the processor allows us to use a
`hardware coordination mechanism in which each core can
`request any C-state it wishes, thus allowing for individual
`core savings to be maximized. The CPU C-state is
`determined and entered based on the lowest common
`denominator of both cores’ requests, portraying a single
`CPU entity to the chipset power management hardware
`and flows. Thus, software can manage each core
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`111
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 3 of 15
`
`

`

`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`the actual power management
`independently, while
`adheres
`to
`the platform and CPU shared resource
`restrictions.
`
`As can be seen from Figure 2, the Intel Core Duo
`processor is partitioned into three domains. The cores,
`their respective Level-1 caches, and the local thermal
`management logic each operates as a power management
`domain. The shared resources–including the Level-2
`cache, bus interface, and interrupt controllers (APICs)–are
`yet another power management domain. All domains share
`a single power plane and a single-core PLL, thus
`operating at the same frequency and voltage levels.
`However, each of the domains has an independent clock
`distribution (spine). The core spines can be gated
`independently, allowing the most basic per core power
`savings (C1-Halt state). The shared-resource spine is
`gated only when both cores are idle and no shared
`operations (bus transactions, cache accesses) are taking
`place. If needed, the shared-resource clock can be kept
`active even when both cores’ clocks are halted, serving L2
`snoops and APIC message analysis.
`
`The coordination mechanism serves as a transparent layer
`between the individually controlled cores and the shared
`resource on die and on the platform. It determines the
`required CPU C-state, based on both cores’ individual
`requests, controls the state of the shared resources, such as
`the shared clock distribution, and also implements the C-
`state entry protocol with the chipset, emulating a legacy
`single-CPU mobile processor. When detecting that both
`core states are deeper
`than C1,
`the coordination
`mechanism issues the proper indication to the chipset,
`triggering the platform C2-4 entry sequence.
`
`When the platform detects a break event (such as an
`interrupt request), it negates the proper sideband signals
`(such as the STPCLK, SLP, DPSLP and DPRSTP). The
`coordination logic then analyzes the platform signaling,
`and initiates the proper C-state exit sequences for the
`shared resources, and if needed, the cores.
`
`Thus, the required goal of coordinating across the two
`cores is achieved in an efficient and transparent manner.
`Software operates as if it is managing two independent
`cores, while the platform and the shared resources are
`controlled as one by the coordination logic, reflecting a
`single platform Level C-state in a backwards-compatible
`manner.
`
`As part of the per core power savings, the independent
`cores’ L1 cache is also flushed during core C3 and C4
`states. Due to the ratio between L1 and L2 cache size, it is
`assumed that nearly all of the L1 is included in the L2.
`Therefore, the flushing should not incur a high overhead
`of a write-back into the system memory, and it will not
`
`
`
`incur a high warm-up penalty when restarting execution
`afterwards, since the data already reside nearby in the L2
`cache. By flushing the caches, the cores can be kept asleep
`even when the L2 cache is accessed heavily by the other
`core or by a system device, thus improving power savings
`even further.
`
`Dynamic Cache Sizing and Deep C4 State
`Now that the processor has been able to enter C4, we face
`the challenge of lowering the C4-state leakage power even
`further. Since leakage power is directly proportional to the
`operating voltage, the most efficient means to save
`leakage static power would be to lower the C4 operating
`point. Unfortunately, lowering the voltage impacts data
`retention, and the first cells to be affected are the small
`transistor data arrays such as in the L2 cache. Therefore,
`the first step in achieving a lower voltage idle state is to
`implement a mechanism that can dynamically shut down
`the L2 cache in preparation for the Deep C4 state.
`
`Cache Sizing
`When defining the L2 cache dynamic sizing algorithm, the
`following considerations need to be addressed:
`
`• L2 is a large array; therefore flushing it will incur
`some power and potentially C-state
`latencies,
`especially if done all at once or too frequently.
`
`• Many
`performance
`suffer
`applications will
`degradation if running for a long period of time with
`little or no L2 cache. However, it has been proven
`that the short interrupt handling tasks, occurring at
`periodic timer ticks, do not have much use for the L2
`cache and are not visibly impacted by running with
`the cache closed.
`
`To accommodate the above restrictions, the dynamic
`sizing mechanism
`is
`implemented as an adaptive
`algorithm, with various built-in filtering properties and
`heuristics.
`
`At the algorithms’ base is an assumption that during long
`periods of very
`low utilization and
`idle residency
`(mapping to the C4 state), shutting down the cache will
`not result in a perceivable performance impact. This
`condition is detected by a state machine, described in
`Figure 3. In order to start shrinking the cache, the Finite
`State Machine (FSM) checks that the CPU frequency,
`controlled by the OS, is below a programmable threshold.
`It also checks that the CPU on the whole has not stayed
`too long in C0–which may indicate a streaming task being
`executed (e.g., DVD playback). This is done by a pre-
`programmed count-down timer, reloaded once the whole
`CPU enters C2-4 (package) states. Finally, the FSM
`checks the second core’s idle state, requiring it to be in C4
`as well, before allowing a shrink operation to begin.
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`112
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 4 of 15
`
`

`

`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`Power control state machine
`
`
`
`
`
`Figure 3: Shrink/Expand heuristics
`
`The FSM freezes the shrink operation if a pending
`interrupt is detected, or if either core is in the active C0
`state.
`
`Cache expansion is requested once the activity indicators
`show that performance may be required. This is inferred
`in one of three ways: either the frequency is being
`increased over the programmable threshold, the CPU is
`staying in C0 for a period exceeding the pre-programmed
`timer, or one of the cores is entering a non C4 idle state
`(this is the OS’s way of signaling that the core was not
`idle enough).
`
`The actual cache flushing flow is performed by the
`microcode of the last core entering C4-state. In order to
`minimize the power impact and to filter too short C4
`periods, the microcode flushes only part of the L2 (1/8
`through 1/2 of
`the
`total cache size) during each
`consecutive C4 entry. The cache is flushed in chunks of
`lines (between 4 to 256 lines) checking for interrupts in
`between. Once a whole way is flushed, it is power-gated
`with sleep transistors, further reducing its leakage.
`
`Microcode typically automatically expands the cache to a
`minimum of two ways upon every DeepC4 exit. Once the
`CPU enters C4 again, microcode will shrink the cache
`back down to 0. At the initial expansion it is assumed that
`the CPU has just exited from DeepC4. As such the L2
`valid array may not be valid. Therefore, as part of the
`DeepC4 exit flow, the microcode also clears all of the L2
`valid bits, ensuring the cache is indeed perceived as empty
`by the snoop logic. The same initialization flow can be
`applied also to other sensitive arrays should testing detect
`them as unstable.
`
`DeepC4 (DC4) Entry
`After the L2 cache has been shrunk to 0 and the CPU
`enters C4, the CPU voltage may be further reduced.
`Moreover, since no data are cached, the data cache does
`not need to be awakened for snoops. This feature is
`performed by the chipset, during the DC4 state. Once this
`is detected, the chipset starts diverting snoopable traffic
`directly to memory. During the DC4 state, the chipset
`scans the incoming traffic for interrupts and APIC
`messages, and once they are detected, queues them
`separately, while initiating a break sequence for the
`processor. Once
`the processor
`is fully awake,
`the
`interrupts are delivered to the processor and the memory
`traffic is diverted back to the CPU for snooping.
`
`Handling P-states in a CMP Environment
`
`ACPI P-state (Performance) control algorithm’s goal is to
`optimize
`the
`runtime power consumption without
`significantly
`impacting performance. The algorithm
`dynamically adjusts the processor frequency such that it is
`just high enough to service the SW execution load.
`Operating point selection is done by the OS power
`management algorithms (OSPM) based on the CPU load
`observed over a window of time. Once the target point is
`set, the CPU is expected to modify its operating voltage
`and frequency to match the OSPM's request.
`
`Figure 4 shows one example of the relationship between
`different working points (P-state points) in the Intel Core
`Duo processor and their relative power consumption. As
`can be seen, the benefit of going to a lower working point
`can be divided into an “exponential” part and a “linear”
`part. The exponential part represents a range of operating
`points where both frequency and voltage can be scaled to
`meet the new working point, while the linear part
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`113
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 5 of 15
`
`

`

`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`represents a range of operating points where the system
`has already reached the minimum voltage allowed and so
`only the frequency can be scaled.
`
`100
`
`90
`
`80
`
`70
`
`60
`
`50
`
`40
`
`30
`
`20
`
`10
`
`0
`
`%Max Frequency
`
`P0 P1 P2 P3 P4 P5 P6 P7 P8
`
`Figure 4: P-state CPU frequencies
`
`
`
`When a multi-processor system was used, each core could
`be controlled independently by the OS and the individual
`processor state was set accordingly. The Intel Core Duo
`CMP implementation presents new challenges since some
`parts of the system, such as the PLLs and power planes,
`are shared and so both cores and the shared area need to
`operate at the same effective frequency/voltage point (i.e.,
`the same P-state).
`
`Initial SW versions that ran on the hyper-threading Intel
`NetBurst® microarchitecture-based CPUs utilized System
`Management Mode (SMM) code to coordinate the target
`P-state between the threads. This solution was targeted to
`optimize performance, which suited desktops and servers
`at which most hyper-threading CPUs were targeted.
`However, it incurred inherent overhead for every P-state
`request made by the OSPM, resulting in an average power
`impact prohibitive to the mobile focused Intel Core Duo
`architecture.
`
`Consequently, a HW coordination approach was adopted
`for the P-state architecture as well. As with the C-state
`mechanism, each core's OS Power Management
`component can request a P-state separately via the
`standard
`IA32_PERF_CONTROL MSR. The HW
`coordination logic in the shared region tracks these P-state
`requests from both cores and determines the required CPU
`level target operating point based on the current execution
`state of the CPU:
`
`• No
`the
`this case,
`In
`thermal control state:
`coordination logic will prefer performance over
`power, so as not to starve the SW threads. As a result,
`the higher operating point will be selected for the
`total CPU operating point, causing the whole CPU to
`the known advanced Intel SpeedStep®
`execute
`technology-based architecture
`transitions between
`operating points.
`
`• Thermal_Controlled_State: In this case, the HW has
`detected an overheating condition and is trying to
`drive the operating point down for cooling. Here, the
`coordination HW will select the lower of the two
`cores’ requests, allowing a quicker cooldown than if
`the maximum of both cores was used. Initial Intel
`Core Duo CPUs will use the same operation point for
`both cores; however, future implementations may
`choose to select a different operation point per core,
`thus taking advantage of this capability.
`
`Due to the hardware coordination nature of the P-state
`architecture, OSPM may have a hard time tracking the
`exact frequency each core has run at, since the actual
`runtime frequency may be higher than the OSPM
`requested, and may vary without the core's OSPM
`knowledge. This may cause some confusion to the OSPM
`when trying to use the core's activity factor (non C0 state
`%) and the expected frequency to determine the actual
`code load. Intel Core Duo technology augments the
`previous frequency visibility mechanism by supplying two
`64-bit counters, providing the OSPM information on the
`actual execution
`frequency of each core. ACNT
`(ActualCount) counts executed clocks (not Idle) at the
`currently
`coordinated CPU
`frequency. MCNT
`(MaxCount) counts the maximum number of non-idle
`clocks that the core could have run at, during the same
`period of time, had it run at the nominal frequency defined
`for the part. By dividing the ACNT by MCNT, the OSPM
`can determine the actual frequency each core has run at
`over a window of time, assisting the OSMP to assess the
`correct execution load on that core and then in turn
`determine the next operating point.
`
`Thermal Control in a CMP Environment
`
`The Intel Core Duo system was designed with power
`efficiency in mind and is aimed at different form factors,
`some of which are power and thermally constrained.
`Thermal management is a fundamental capability of all
`mobile platforms. Managing platform thermals enables us
`to achieve the maximum CPU and platform performance
`within thermal constraints. Thermal management features
`also improve ergonomics with cooler systems and lower
`fan acoustic noise.
`
`In order to better control the thermal conditions of the
`system, Intel Core Duo technology presents two new
`concepts: the use of digital sensors for high accuracy die
`temperature measurements and dual-core multiple-level
`thermal control.
`
`The general structure of the digital thermometer is
`described in Figure 5. It supports the legacy Intel®
`Centrino® mobile technology thermal sensor, PROCHOT
`and THERMTRIP, and the adding of a temperature
`reading capability to the fixed threshold sensor. A control
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`114
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 6 of 15
`
`

`

`In the above example, the die temperature is at 60oC and
`the thresholds are set to 50oC and 65oC. If the temperature
`rises above 65oC, a low-to-high interrupt is generated. The
`control software identifies the new temperature and
`initiates action, such as activating fans or initiating some
`passive cooling policy. The activation thresholds and
`policies are defined using BIOS and ACPI. New
`thresholds are loaded around the new temperature to
`further track new changes.
`
`In previous Intel Pentium M processor-based systems, a
`single analog thermal diode was used to measure die
`temperature. A thermal diode cannot be located at the
`hottest spot of the die and therefore some offset was
`applied to keep the CPU within specifications. For these
`systems it was sufficient, since the die had a single hot
`spot. In the Intel Core Duo system, there is more than a
`single hot spot that moves as a function of the combined
`workload of both cores. Figure 7 shows the differences
`between the usage of the traditional analog sensor and the
`new digital sensors.
`
`
`
`Figure 7: Analog vs. digital sensors in Intel Core Duo
`systems
`
`As we can see, the use of multiple sensing points provides
`high accuracy and close proximity to the hot spot at any
`time. An analog thermal diode is still available on the Intel
`Core Duo processor. An example of the difference
`between the diode and the DTS is presented in Figure 8.
`The information presented in this graph was taken from
`simulations and not from a real system, but was
`correletaed with data from real systems and found to be
`close enough.
`
`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`logic built around the thermal sensor performs a periodic
`scan and generates an output that represents the current
`temperature reading. The reading is loaded into an MSR,
`accessible to software. In order to improve temperature
`reading, multiple sense points monitor different hot spots
`on the die and report the maximum temperature of the die.
`An independent temperature reading from each core is
`available, with optional reporting of
`the maximum
`temperature of the entire die, e.g., the highest temperature
`of both cores.
`
`
`
`MSR (R/W)MSR (R/W)
`
`
`
`
`DigitalDigitalDigitalDigital
`
`
`
`SensorSensorSensorSensor
`
`
`
`Thermal shut down, THERMTRIPThermal shut down, THERMTRIP
`
`
`
`Thermal protection, PROCHOTThermal protection, PROCHOT
`
`
`
`77
`
`
`
`TEMPTEMP
`
`
`
`
`CompareCompareCompareCompare
`
`(Temp >(Temp >
`
`Threshold #1Threshold #1
`
`High Limit)High Limit)
`
`
`
`
`CompareCompareCompareCompare
`
`(Temp <(Temp <
`
`Threshold #2Threshold #2
`
`Low Limit)Low Limit)
`
`
`
`Enable 1Enable 1
`
`
`
`Enable 2Enable 2
`
`
`
`Out of SpecOut of Spec
`
`
`
`TEMP INT #1TEMP INT #1
`
`
`
`TEMP INT #2TEMP INT #2
`
`
`
`TEMP
`TEMP Threshold #1 Threshold #2
`TEMP
`TEMP Threshold #1 Threshold #2
`
`LOW
`LOW
`
`HIGH
`HIGH
`
`Figure 5: Digital thermometer block diagram
`
`Interrupt generation capability is provided in addition to
`the temperature reading. Two programmable thresholds
`are loaded by S/W and a thermal event is generated upon
`threshold crossing. This thermal event generates an
`interrupt to a single core or to both cores simultaneously
`through the APIC settings.
`
`The digital thermometer is intended to be used as an input
`to a software-based thermal control such as the ACPI. An
`interrupt threshold is defined to indicate the upper and
`lower temperature thresholds. An example of digital
`thermometer usage is illustrated in Figure 6.
`
`
`
`_TMP = 60
`
`-95-
`-90-
`-85-
`
`-80-
`-75-
`-70-
`-65-
`
`-60-
`-55-
`-50-
`-45-
`
`-40-
`-35-
`-30-
`-25-
`
`_CRT
`
`_AC0
`
`_AC1
`
`_PSV
`
`
`
`Figure 6: Digital thermometer and ACPI
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`115
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 7 of 15
`
`

`

`control behavior and further reduces power as needed.
`Continuous extreme conditions will eventually initiate an
`“Out Of Spec” thermal interrupt and register the warning
`in an internal status bit. The “Out Of Spec warning” signal
`is intended to initiate a managed system shutdown before
`a THERMTRIP shutdown occurs.
`
`Platform Power Optimization
`
`Intel Core Duo technology implemented power and
`thermal management features as described above to
`control the CPU power and thermals. Other components
`on the platform also contribute to the total platform power
`and work closely with the CPU. The CPU VR losses can
`get as high as 5W while running high workloads. A 3-
`phase VR was built to deliver high current operating at
`low efficiency while working at low utilization. Turning
`off phases or switching to asynchronous mode can save a
`significant amount of power. Efficiency, as a function of
`the number of phases, is described in Figure 9.
`
`
`
`ISL6247 Efficiency Comparision with Different # ChannelsISL6247 Efficiency Comparision with Different # Channels
`
`
`
`4-Channels4-Channels
`
`
`
`3-Channels3-Channels
`
`
`
`2-Channels2-Channels
`
`
`
`1-Channel1-Channel
`
`
`
`
`
`Power SavingsPower SavingsPower Savings
`
`
`Power SavingsPower SavingsPower Savings
`
`
`
`100.00%100.00%
`
`
`
`95.00%95.00%
`
`
`
`90.00%90.00%
`
`
`
`85.00%85.00%
`
`
`
`80.00%80.00%
`
`
`
`75.00%75.00%
`
`Efficiency
`Efficiency
`
`
`
`00
`
`
`
`1010
`
`
`
`2020
`
`
`
`3030
`
`
`
`4040
`
`
`
`5050
`
`
`
`6060
`
`
`
`7070
`
`
`
`8080
`
`
`
`Output CurrentOutput Current
`
`Figure 9: Efficiency as a function of number of phases
`
`It can be noted that improved power efficiency mode
`cannot handle high currents and there is a need for
`feedback on the current requirements. The Intel Core Duo
`processor keeps track of the power consumption and gives
`feedback to the VR using the PSI-2 and VID signals, as
`described in Figure 10.
`
`
`
`
`
`
`
`Intel Technology Journal, Volume 10, Issue 2, 2006
`
`
`
`Tjmax vs Diode TemperatureTjmax vs Diode Temperature
`
`
`
`Offset diodOffset diod
`
`
`
`1212
`
`
`
`1010
`
`
`
`88
`
`
`
`66
`
`
`
`44
`
`
`
`22
`
`
`
`00
`
`T offset ['C]
`T offset ['C]
`
`
`
`4545
`
`
`
`5555
`
`
`
`6565
`
`
`
`7575
`
`
`
`8585
`
`
`
`9595
`
`
`
`Tj [% of Tjm ax]Tj [% of Tjm ax]
`
`
`
`Figure 8: Diode to DTS temperature difference
`
`It can be seen that significant temperature gradients exist
`on the die. The use of a digital thermometer provides
`improved temperature readings, enables higher CPU
`performance within thermal limitations, and improves
`reliability.
`
`The Intel Core Duo system also implements hardware-
`based
`thermal
`control. Hardware-based
`thermal
`management is intended to handle abnormal thermal
`conditions and to protect the die from transient effects.
`Hardware-based thermal control ensures that the CPU will
`always operate within specified conditions. This improves
`reliability and allows higher performance with tighter
`control parameters.
`
`Legacy thermal control features implement two externally
`visible signals:
`
`• THERMTRIP: a fixed temperature sensor to detect
`catastrophic thermal conditions and to shut down the
`system if thermal runaway occurs.
`
`• PROCHOT: a fixed
`that
`threshold
`temperature
`provides the DVS with a self-control mechanism that
`drops frequency and voltage to a new working point
`(a more detailed description of this mechanism can be
`found in [1]).
`
`Intel Core Duo technology implements a new multiple-
`core thermal control algorithm. Both cores synchronize
`action requests and activity. A programmable policy can
`select DVS on both cores and linear control on each core
`either independently or as a locked operation. PROCHOT
`was also made available as an input, to allow thermal
`protection of platform components such as the voltage
`regulator (VR).
`
`is performed by
`A finer grain overheat detection
`monitoring the thermal activity. If an extreme thermal
`condition occurs (fan malfunction, operation within a bag,
`etc.) the processor self thermal management may not be
`sufficient and the temperature may not drop below the
`threshold point. The internal control algorithm tracks the
`
`Power and Thermal Management in the Intel® CoreTM Duo Processor
`
`116
`
`MyPAQ, Exhibit 2023
`IPR2022-00311
`Page 8 of 15
`
`

`

`calculation methodology in the following sections to
`compare ST and MT applications.
`
`
`
`P-State Average Pow er
`
`
`P 0, M T
`
`P 0, ST
`
`P 1, ST
`
`
`
`
`
`
`
`P 2, ST
`
`P 1, M T
`
`
`
`P 2, M T
`
`
`
`1
`
`
`261 521 781 1041 1301 1561 1821 2081 2341 2601 2861 3121
`
`T i mel i ne
`
`
`
`Figure 11: Energy calculation methodology
`
`Power Consumption of Different Power States
`
`In order to evaluate the benefit of using MT applications,
`we measure the power consumption of the system while in
`different working points when running an ST application
`vs. an MT application. Figure 12 shows the experimental
`data with both ST and MT code under different P states
`on an Intel Core Duo system. As can be seen, the power
`consumed by an MT code is less than twice the power
`consumed by an ST code in corresponding P states. This
`implies that the applications that are MT and demonstrate
`excellent scaling will not only reduce the total execution
`time, but also show greater savings in power consumed.
`
`P-C State Average Power
`
`MT (multithreaded)
`
`ST

This document is available on Docket Alarm but you must sign up to view it.


Or .

Accessing this document will incur an additional charge of $.

After purchase, you can access this document again without charge.

Accept $ Charge
throbber

Still Working On It

This document is taking longer than usual to download. This can happen if we need to contact the court directly to obtain the document and their servers are running slowly.

Give it another minute or two to complete, and then try the refresh button.

throbber

A few More Minutes ... Still Working

It can take up to 5 minutes for us to download a document if the court servers are running slowly.

Thank you for your continued patience.

This document could not be displayed.

We could not find this document within its docket. Please go back to the docket page and check the link. If that does not work, go back to the docket and refresh it to pull the newest information.

Your account does not support viewing this document.

You need a Paid Account to view this document. Click here to change your account type.

Your account does not support viewing this document.

Set your membership status to view this document.

With a Docket Alarm membership, you'll get a whole lot more, including:

  • Up-to-date information for this case.
  • Email alerts whenever there is an update.
  • Full text search for other cases.
  • Get email alerts whenever a new case matches your search.

Become a Member

One Moment Please

The filing “” is large (MB) and is being downloaded.

Please refresh this page in a few minutes to see if the filing has been downloaded. The filing will also be emailed to you when the download completes.

Your document is on its way!

If you do not receive the document in five minutes, contact support at support@docketalarm.com.

Sealed Document

We are unable to display this document, it may be under a court ordered seal.

If you have proper credentials to access the file, you may proceed directly to the court's system using your government issued username and password.


Access Government Site

We are redirecting you
to a mobile optimized page.





Document Unreadable or Corrupt

Refresh this Document
Go to the Docket

We are unable to display this document.

Refresh this Document
Go to the Docket