Mobile Pentium 4 Architecture Supporting Hyper-ThreadingTechnology presentation
Thread Rating:
  • 0 Vote(s) - 0 Average
  • 1
  • 2
  • 3
  • 4
  • 5
seminar presentation
Active In SP

Posts: 582
Joined: Apr 2010
12-05-2010, 09:19 PM

Mobile Pentium 4 Architecture Supporting Hyper-ThreadingTechnology

Presented By:
Hakan Burak Duygulu


The Intel Pentium 4 Processor Family (2000-2005)
based on Intel NetBurst® microarchitecture.
introduced Streaming SIMD Extensions 2 (SSE2)
introduced Streaming SIMD Extensions 3 (SSE3)
The Intel® Xeon Processor (2001-2005)
introduced support for Hyper-Threading Technology
The Intel® Pentium® M Processor (2003-2005)
designed for extending battery life
The Intel Pentium Processor Extreme Edition (2005)
64-bit addressing, 1024-Gbytes address space.
Intel® NetBurstTM microarchitecture.

Design Goals

to execute legacy IA-32 applications based on single-instruction, multiple-data (SIMD) technology at high throughput
to operate at high clock rates and to scale to higher performance and clock rates in the future
Intel® NetBurstTM microarchitecture.

Design Advantages

a deeply pipelined design, 20-stage pipeline, that allows for high clock rates (with different parts of the chip running at different clock rates).
a pipeline that optimizes for the common case of frequently executed instructions; the most frequently-executed instructions in common circumstances (such as a cache hit) are decoded efficiently and executed with short latencies
Employment of techniques to hide stall penalties; Among these are parallel execution, buffering, and speculation. The microarchitecture executes instructions dynamically and out-of-order, so the time it takes to execute each individual instruction is not always deterministic


The Intel NetBurst microarchitecture supports up to three levels of on-chip cache.
The first level cache (nearest to the execution core) contains separate caches for instructions and data. These include the first-level data cache and the trace cache (an advanced first-level instruction cache). All other caches are shared between instructions and data.
All caches use a pseudo-LRU (least recently used) replacement algorithm.

The Front End Pipeline

Consists of two parts:
Fetch/decode unit:
a hardware instruction fetcher that automatically prefetches instructions
a hardware mechanism that automatically fetches data and instructions into the unified second-level cache
Execution trace cache
The execution trace cache (TC) is the primary instruction cache in the Intel NetBurst microarchitecture. The TC stores decoded IA-32 instructions (µops).

The Front End Pipeline

Prefetches IA-32 instructions that are likely to be executed
Fetches instructions that have not already been prefetched
Decodes IA-32 instructions into micro-operations
Generates microcode for complex instructions and special-purpose code
Delivers decoded instructions from the execution trace cache
Predicts branches using highly advanced algorithm

The Front End Pipeline - Branch Prediction

Enables the processor to begin executing instructions long before the branch outcome is certain. Branch delay is the penalty that is incurred in the absence of correct prediction.
Branch prediction in the Intel NetBurst microarchitecture predicts all near branches (conditional calls, unconditional calls, returns and indirect branches). It does not predict far transfers (far calls and software interrupts).
Mechanisms have been implemented to aid in predicting branches accurately and to reduce the cost of taken branches:
the ability to dynamically predict the direction and target of branches based on an instructionâ„¢s linear address, using the branch target buffer (BTB)
if no dynamic prediction is available or if it is invalid, the ability to statically predict the outcome based on the offset of the target: a backward branch is predicted to be taken, a forward branch is predicted to be not taken
the ability to predict return addresses using the 16-entry return address stack
the ability to build a trace of instructions across predicted taken branches to avoid branch penalties.

The Static Predictor.

Once a branch instruction is decoded, the direction of the branch (forward or backward) is known. If there was no valid entry in the BTB for the branch, the static predictor makes a prediction based on the direction of the branch. The static prediction mechanism predicts backward conditional branches (those with negative displacement, such as loop-closing branches) as taken. Forward branches are predicted not taken.
To take advantage of the forward-not-taken and backward-taken static predictions, code should be arranged so that the likely target of the branch immediately follows forward branches

Branch Target Buffer.

Once branch history is available, the Pentium 4 processor can predict the branch outcome even before the branch instruction is decoded. The processor uses a branch history table and a branch target buffer (collectively called the BTB) to predict the direction and target of branches based on an instructionâ„¢s linear address. Once the branch is retired, the BTB is updated with the target address.

Out-Of-Order Execution Core

Ability to execute instructions out of order is a key factor in enabling parallelism. This feature enables the processor to reorder instructions so that if one µop is delayed, other µops may proceed around it. The processor employs several buffers to smooth the flow of µops.
The core is designed to facilitate parallel execution. It can dispatch up to six µops per cycle. Most pipelines can start executing a new µop every cycle, so several instructions can be in flight at a time for each pipeline. A number of arithmetic logical unit (ALU) instructions can start at two per cycle; many floating-point instructions can start once every two cycles.

Retirement Unit

The retirement unit receives the results of the executed µops from the out-of-order execution core and processes the results so that the architectural state updates according to the original program order.
When a µop completes and writes its result, it is retired. Up to three µops may be retired per cycle. The Reorder Buffer (ROB) is the unit in the processor which buffers completed µops, updates the architectural state in order, and manages the ordering of exceptions. The retirement section also keeps track of branches and sends updated branch target information to the branch target buffer(BTB). The BTB then purges pre-fetched traces that are no longer needed.
Hyper-Threading Technology
Enables software to take advantage of task-level, or thread-level parallelism by providing multiple logical processors within a physical processor package.
The two logical processors each have a complete set of architectural registers while sharing one single physical processor's resources. By maintaining the architecture state of two processors, an HT Technology capable processor looks like two processors to software, including operating system and application code.

Replicated Resources

Control registers (Architectural Registers AR)
8 general purpose registers (AR)
Machine state registers (AR)
Debug registers (AR)
Instruction pointers(IP)
Register renaming tables(RNT)
Return stack predictor (RSP)

Replicated Resources

ARâ„¢s are used by the operating system and application code to control program behavior and store data for computations.
IP and RNT are replicated for simultaneously track execution and state changes of the two logical processors.
The RSP is replicated to improve branch prediction of return instructions.

Partitioned Resources

Re-order Buffers(ROBâ„¢s)
Load/Store Buffers
Various queues, like the scheduling and µop queues

Partitioned Resources

operational fairness
permitting the ability to allow operations from one logical processor to bypass operations of the other logical processor that may have stalled.
For example: a cache miss, a branch misprediction, or instruction dependencies may prevent a logical processor from making forward progress for some number of cycles. The partitioning prevents the stalled logical processor from blocking forward progress.

Shared Resources

Caches: trace cache, L1, L2, L3
Execution Units
are fully shared to improve the dynamic
utilization of the resource.

Front End Pipeline

Execution trace cache access is arbitrated by the two logical processors every clock. If a cache line is fetched for one logical processor in one clock cycle, the next clock cycle a line would be fetched for the other logical processor provided that both logical processors are requesting access to the trace cache.
If one logical processor is stalled or is unable to use the execution trace cache, the other logical processor can use the full bandwidth of the trace cache.

Front End Pipeline

After fetching the instructions and building traces of µops, the µops are placed in a queue. This queue decouples the execution trace cache from the register rename pipeline stage. If both logical processors are active, the queue is partitioned so that both logical processors can make independent forward progress.

Execution Core

The core can dispatch up to six µops per cycle, provided the µops are ready to execute. Once the µops are placed in the queues waiting for execution, there is no distinction between instructions from the two logical processors.
After execution, instructions are placed in the re-order buffer. The re-order buffer decouples the execution stage from the retirement stage. The re-order buffer is partitioned such that each uses half the entries.


The retirement logic tracks when instructions from the two logical processors are ready to be retired. It retires the instruction in program order for each logical processor by alternating between the two logical processors. If one logical processor is not ready to retire any instructions, then all retirement bandwidth is dedicated to the other logical processor.
Once stores have retired, the processor needs to write the store data into the level-one data cache. Selection logic alternates between the two logical processors to commit store data to the cache.
Streaming SIMD Extensions 3 (SSE3)
Beginning with the Pentium II and Pentium Intel MMX technology processor families, four extensions have been introduced into the IA-32 architecture to permit IA-32 processors to perform single-instruction multiple-data (SIMD) operations.
These extensions include the MMX technology, SSE extensions, SSE2 extensions, and SSE3 extensions.
Streaming SIMD Extensions 3 (SSE3)
Streaming SIMD Extensions 3 (SSE3)

MMX„¢ Technology

MMX Technology introduced:
64-bit MMX registers
support for SIMD operations on packed byte, word, and doubleword integers
MMX instructions are useful for multimedia and communications software.
Streaming SIMD Extensions 3 (SSE3)

Streaming SIMD Extensions

Streaming SIMD extensions introduced:
128-bit XMM registers
data prefetch instructions
non-temporal store instructions and other cacheability and memory ordering instructions
extra 64-bit SIMD integer support
SSE instructions are useful for 3D geometry, 3D rendering, speech recognition, and video encoding and decoding.
Streaming SIMD Extensions 3 (SSE3)

Streaming SIMD Extensions 2

Streaming SIMD extensions 2 add the following:
support for SIMD arithmetic on 64-bit integer operands
instructions for converting between new and existing data types
extended support for data shuffling
extended support for cacheability and memory ordering operations
SSE2 instructions are useful for 3D graphics, video decoding/encoding, and encryption.
Streaming SIMD Extensions 3 (SSE3)

Streaming SIMD Extensions 3

Streaming SIMD extensions 3 add the following:
SIMD floating-point instructions for asymmetric and horizontal computation
a special-purpose 128-bit load instruction to avoid cache line splits
instructions to support thread synchronization
SSE3 instructions are useful for scientific, video and multi-threaded applications.
Enhanced Intel SpeedStep® technology
Enables real-time dynamic switching between multiple voltages and operating frequency points.
The processor features the Auto Halt, Stop Grant, Deep Sleep, and Deeper Sleep low power states.
Enhanced Intel SpeedStep® technology
The processor includes an address bus powerdown capability which removes power from the address and data pins when the FSB is not in use.

Deeply Pipelined, 20-stage pipeline , achieved Higher Clock Rate with NetBurst microarchitecture
Improved performance with very little additional die area with Hyper-Threading Technology
SSE3 offers 13 instructions that accelerate performance of Streaming SIMD Extensions technology

- IA-32 Intel® Architecture Software Developer's Manual, Volume 1: Basic Architecture
- IA-32 Intel® Architecture Optimization Reference Manual

download the ppt
Use Search at wisely To Get Information About Project Topic and Seminar ideas with report/source code along pdf and ppt presenaion

Important Note..!

If you are not satisfied with above reply ,..Please


So that we will collect data for you and will made reply to the request....OR try below "QUICK REPLY" box to add a reply to this page

Quick Reply
Type your reply to this message here.

Image Verification
Please enter the text contained within the image into the text box below it. This process is used to prevent automated spam bots.
Image Verification
(case insensitive)

Possibly Related Threads...
Thread Author Replies Views Last Post
  DETECTION OF LOST MOBILE USING SNIFFERS seminar class 64 34,003 12-04-2016, 03:24 PM
Last Post: mkaasees
  Report on Mobile TV study tips 1 477 11-04-2016, 10:14 AM
Last Post: mkaasees
  MOBILE TV PPT project girl 1 1,145 11-04-2016, 10:14 AM
Last Post: mkaasees
Last Post: mkaasees
  MOBILE TRAIN RADIO COMMUNICATION REPORT study tips 1 620 06-04-2016, 03:47 PM
Last Post: mkaasees
  6g mobile technology pdf jaseelati 0 355 22-01-2015, 02:00 PM
Last Post: jaseelati
Last Post: seminar project topic
  nanotechnology paper presentation computer science topics 4 27,682 07-03-2014, 04:41 PM
Last Post: seminar project topic
  wifi seminar or presentation report electronics seminars 28 31,180 10-02-2014, 01:52 PM
Last Post: JoshuapupE
  Electronics seminar or presentation lists2 computer science crazy 10 5,772 04-02-2014, 12:33 AM
Last Post: krupa gowda