0% found this document useful (0 votes)
43 views17 pages

Final Research

The document discusses the Intel Core i7 processor. It describes the instruction sets implemented including MMX, SSE, SSE2, SSE3, SSSE3, SSE4, and Hyper-Threading technology. It details the processor's memory support including triple channel memory and bandwidth. The document also covers clock speeds up to 5GHz, functionality as a quad core with hyperthreading, and advantages of hyperthreading for parallel workloads.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
43 views17 pages

Final Research

The document discusses the Intel Core i7 processor. It describes the instruction sets implemented including MMX, SSE, SSE2, SSE3, SSSE3, SSE4, and Hyper-Threading technology. It details the processor's memory support including triple channel memory and bandwidth. The document also covers clock speeds up to 5GHz, functionality as a quad core with hyperthreading, and advantages of hyperthreading for parallel workloads.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 17

MICROPROCESSOR

REPORT

Instructor: Lenard Ramirez


Reporters:
John Mark S. Santos
Maria Grace Abalos
INTEL CORE i7 PROCESSORS

INSTRUCTION SET

An instruction set is the list of all the instructions, and all their variations that a processor can
execute. A single instruction multiple data (SIMD) is one in which single instruction stream
has ability to process multiple data streams simultaneously these
machines are useful in applications such as general digital signal processing, image processing
and multimedia applications such as audio and video.

Originally, super computers known as array processors and vector processors provided SIMD
processing capabilities.

Intel’s core i7 implements the MMX, streaming SIMD extensions (SSE), streaming
SIMD extensions 2 (SSE2), and streaming SIMD
extensions 3 (SSE3), supplemental streaming SIMD extensions 3 (SSSE3), streaming SIMD
extensions 4 (SSE4) instruction sets that are capable of processing multiple data elements in a
single clock.
The multiple data elements are stored in the floating point registers. A multiple
instruction, multiple data (MIMD) machine is capable of is executing multiple instruction
streams, while working on a separate and independent data stream, the instruction set MMX is
64-bit instruction set. The instruction set SSE is 128-bit instruction set.

Intel Core i7

Produced : From 2008


Common manufacturer(s) : Intel
Max. CPU clock rate : 2.9 GHz to 4.2 GHz
Word size : 64 bit 8 octets 16 nibbles
QPI speeds: .8 GT/s to 8Gt/s
Min. feature size : 45 nm, 32 nm, or 22 nm
Instruction Set : x86, x86-64, MMX, SSE, SSE2,
SSE3, SSSE3, SSE4.1, SSE4.2
Number of Cores : 2, 4, 6, 8 or 10
Microarchitecture : Nahalem, Sandy Bridge,
Sandy Bridge-E, Ivy Bridge, Ivy Bridge-E,
Haswell, Haswell-E, Broadwell, Skylake
Core Names : Bloomfield, Lynnfield,
Clarksfield, Clarksfield XM
MEMORY

The i7 can support up to three memory channels, each consisting of a separate set of DIMMs,
and each of which can transfer in parallel. Using DDR3-1066 (DIMM PC8500), the i7 has a
peak memory bandwidth of just over 25 GB/sec. i7 uses 48-bit virtual addresses and 36-bit
physical addresses, yielding a maximum physical memory of 36 GB. Memory management is
handled with a two-level TLB.

After all, it is possible to configure a Core i7 processor with triple-channel memory


operating at 1066, 1333 and 1600MHz, or even faster. But what are the benefits of
using more aggressively clocked memory?
In the past, when testing the Core 2 family of processors, we found that spending more
money on high-speed DDR2 and DDR3 memory was a real waste of money, since the
processor was unable to take benefit from the added bandwidth.
Whereas previous Intel processors such as the Core 2, Pentium D, Pentium 4, and so
on, relied on the Front Side Bus (FSB) to access RAM, the new Core i7 processors do
not. The Intel FSB which was used on the Core 2 motherboards allowed for 4 transfers
per clock, which in marketing terms Intel referred to as their & quot;Quad-Pumped Front Side
Bus."
This meant that the Core 2 Duo E8500 processor which ran on a 333MHz FSB, could
perform 1333 million transfers per second. Now, because this FSB was 64-bit wide, the
end result was a bandwidth of just 10.6GB/s (1333x8). Here lies Intel’s problem; their
current 333MHz FSB supports a peak memory bandwidth of just 10.6GB/s, which is
being exceeded by DDR2-800 memory operating in dual-channel mode.
For some time we have known that DDR2-1066 does n't offer a substantial performance
gain over DDR2-800 memory for Core 2 users as DDR2-1066 memory operating in
dual-channel mode provides a peak bandwidth of 17GB/s, vastly in excess of the
10.6GB/s limitation of the 333MHz FSB.
Then along came the DDR3 memory standard, which in principle works much like
DDR2 memory, just at much greater frequencies. Yet again, DDR3 failed to have much
of an impact, even when compared to DDR2 memory operating at just 800MHz. Just
like faster DDR2-1066 memory, DDR3-1333 provides a peak bandwidth of 21.3GB/s,
while DDR3-1600 reaches 25.6GB/s, both of which exceed the 10.6GB/s limitation of
the 333MHz FSB, by miles.
This is the reason at large why DDR3 has failed to get off the ground. But with the new
Core i7 platform, DDR3 should finally get its chance to shine. As noted before, the Core

i7 features an on-die memory controller with support for triple-channel DDR3 memory.
In other words, three DDR3-1333 modules operating together will provide a peak
bandwidth of 32GB/s matching the L2 cache performance of a Core 2 Quad Q6600
processor.
Now if you are wondering how the FSB will cope with triple channel memory when it
could not even handle dual channel memory, the answer is simple, it’s not! Rather, by
using an Integrated Memory Controller (IMC), Intel was able to do away with the old
FSB.
So how do the Core i7 processors handle the additional memory bandwidth? Is it put to
good use or is it just wasted like on the Core 2 processors? This is what we plan to find
out, and while we will be testing a range of frequencies and timings, we will also look at
how single, dual, and triple-channel memory impacts the performance of the Core i7
processors.

SPEED

The speed range of i7 processors is 2.9GHz-4.2GHz.

In 2010, the fastest i7 had a clock rate of 3.3 GHz, which yields a peak instruction execution rate
of 13.2 billion instructions per second, or over 50 billion instructions per second for the four-
core design.

But now, Intel confirmed core i7-8086k has 6-core CPU, and will be the first Intel CPU to reach
5 GHz out of the box without overlocking.

Functionality

All i7 processors are Quad Core but they are equipped with Hyper Threading Technology
which enables each core to serve 2 threads, hence the OS sees it as an Octa Core Processor rather
than a Quad Core Processor.

An i7 is an ultra-fast CPU for professional users that need a shitload of performance for multiple
processes at once or advanced processes that make good use of multiple cores/threads (video
editors, 3D artists, professional twitch users).
Rule of thumb is that everything that doesn’t happen in an instant (like rendering videos or
exporting hundreds of files at once) gets faster with an i7 while everything that happens in an
instant (gaming, making a photo brighter) is about as fast with an i5 as it would be with an i7.

How does Hyper Threading work?


In a traditional multi-core CPU, each physical core has its own resources and each core consists
of a single thread which has independent access to all the resources.
Hyper-Threading involves 2 (or in rare cases more) threads sharing the same resources. The
scheduler can switch tasks and processes between these threads.

Advantages of Hyper Threading


 Hyper-Threading solves the problem of “starvation”. If a core or thread is free, the
scheduler can pass the data to it instead of the core remaining idle or waiting for some
other new data to flow through it.
 Much larger and parallel workloads can be done with greater efficiency. As there are
more threads to parallelize, applications that depend heavily on multiple threads can
boost their work significantly (not twice as fast though).
 If you are gaming and have some sort of important task running in the background, the
CPU will not struggle to provide adequate frames and run that task smoothly as it can
switch resources between threads.

The following are not much of disadvantages, rather they are more inconveniences.

 Hyper-Threading needs implementation from software level to take advantage of. Even
though more and more applications are being developed to take advantage of multiple
threads, applications that do not take advantage of any SMT (Simultaneous Multi-
Threading) technology or even multiple physical cores will run exactly the same
regardless. Performance of these applications are more dependent on the clock speed and
IPC of a CPU.

 Hyper-Threading can cause the CPU to create more heat. This is why i5s used to clock
much higher than i7s because they would not heat as much as they have fewer threads.

 Multiple threads share the same resources within a core. This is why performance does
not double. It is instead a very clever method to maximize efficiency and boost
performance where ever possible.

Hyper-Threading is old technology but one here to stay. As applications get more and more
demanding, and the increasing death rate of Moore’s law, the ability to parallelize workloads has
helped improve performance significantly. Being able to run partially parallel workloads helps
increase your productivity and gets your work done faster without stuttering.
The i7 supports the x86-64 instruction set architecture, a 64-bit extension of the 80x86
architecture. The i7 is an out-of-order execution processor that includes four cores.

Each core in an i7 can execute up to four 80x86 instructions per clock cycle, using a multiple
issue, dynamically scheduled and16-stage pipeline.
The core i7 processors series targets the gaming industry and for applications that demand
efficient performance

Core processor i7 is recommended for:

 multitasking, for running multiple programs at same time


 Multithreading applications
 Extreme 3D gaming
 Creating professional movies and editing graphical tasks
 More than enough for basic tasks such as word processing, internet browsing
and E-mail.

DURABILITY

Durability and toughness .Realizing great performance with Core(TM) i7 processor even under
the severe environments.

 High Performance
o  High performance CPU 
  Intel(R) Core(TM)i7-660UE Processor (1.33GHz)
o Improved dust and water resistance (conforming to IP65)
Conforming to IP54 when connecting to dust and water
resistant cable
o Guarantee a wide operating  temperature range   (-20 to 50 degrees C) 
when using wide temperature battery and HDD
o Long lasting battery 
Lasts up to 12 hours(when using long lasting battery)
o High brightness LCD
Max. 750cd/m²
Convertible Display
o Both Note Type and Tablet Type usage are available
 Long Term Supply / Maintenance 
o Maintain the system long

 Various Usage
o Field Maintenance Terminal
o Production Control Terminal
o Inventory Control
o Panel Computer
o Fish Market
o Police
o Disaster Center
o Vehicle Terminal

FEATURES OF INTEL COURE i7 PROCESSORS

1 – New Platform Architecture


2 – Higher-Performance Multiprocessor Systems with QPI
3 – CPU Performance Boost via Intel Turbo Boost Technology
4 – Improved Cache Latency with Smart L3 Cache
5 – Optimized Multithreaded Performance through Hyper-Threading
6 – Higher Data-Throughput via PCI Express 2.0 and DDR3 Memory Interface
7 – Improved Virtualization Performance
8 – Remote Management of Networked Systems with Intel Active Management Technology
(AMT)
1 – New Platform Architecture
As shown in Figure 1, the previous Intel microarchitectures for a single processor system
included three discrete components: a CPU; a Graphics and Memory Controller Hub (GMCH),
also known as the northbridge; and an I/O Controller Hub (ICH), also known as the southbridge.
The GMCH and ICH combined are referred to as the chipset.

In the older Penryn architecture, the front-side bus (FSB) was the interface for exchanging data
between the CPU and the northbridge. If the CPU had to read or write data into system memory
or over the PCI Express bus, then the data had to traverse over the external FSB. In the new
Nehalem microarchitecture, Intel moved the memory controller and PCI Express controller from
the northbridge onto the CPU die, reducing the number of external databus that the data had to
traverse. These changes help increase data-throughput and reduce the latency for memory and
PCI Express data transactions. These improvements make the Core i7 family of processors ideal
for test and measurement applications such as high-speed design validation and high-speed data
record and playback.

Figure 1. These block diagrams represent the higher-level architectural differences between the
previous generation of Intel microarchitectures and the new Nehalem microarchitecture for
single-processor systems.

2 – Higher-Performance Multiprocessor Systems with QPI


Not only was the memory controller moved to the CPU for Nehalem processors, Intel also
introduced a distributed shared memory architecture using Intel QuickPath Interconnect (QPI).
QPI is the new point-to-point interconnect for connecting a CPU to either a chipset or another
CPU. It provides up to 25.6 GB/s of total bidirectional data throughput per link.
Intel’s decision to move the memory controller in the CPU and introduce the new QPI databus
has had an impact for single-processor systems. However, this impact is much more significant
for multiprocessor systems. Figure 2 illustrates the typical block diagrams of multiprocessor
systems based on the previous generation and the Nehalem microarchitecture.

Figure 2:  These block diagrams represent the higher-level architectural differences between the
previous generation of Intel microarchitectures and the new Nehalem microarchitecture for
multiprocessor systems.

The Nehalem microarchitecture integrated the memory controller on the same die as the Core i7
processor and introduced the high-speed QPI databus. As shown in Figure 2, in a Nehalem-based
multiprocessor system each CPU has access to local memory but they also can access memory
that is local to other CPUs via QPI transactions. For example, one Core i7 processor can access
the memory region local to another processor through QPI either with one direct hop or through
multiple hops.

With these new features, the Core i7 processors lend themselves well to the creation of higher-
performance processing systems. For maximum performance gains in a multiprocessor system,
application software should be multithreaded and aware of this new architecture. Also, execution
threads should explicitly attempt to allocate memory for their operation within the memory space
local to the CPU on which they are executing.

By combining a multiprocessor computer with PXI-MXI-Express to a PXI system, processor


intensive applications can take advantage of the multiple CPUs. Examples of these types of
applications range from design simulation to hardware-in-the-loop (HIL).

3 – CPU Performance Boost via Intel Turbo Boost Technology


About five years ago, Intel and AMD introduced multicore CPUs. Since then a lot of
applications and development environments have been upgraded to take advantage of multiple
processing elements in a system. However, because of the software investment required to re-
architect applications, there are still a significant number of applications that are single threaded.
Before the advent of multicore CPUs, these applications saw performance gains by executing on
new CPUs that simply offered higher clock frequencies. With multicore CPUs, this trend was
broken as newer CPUs offered more discrete processing cores rather than higher clock
frequencies.

To provide a performance boost for lightly threaded applications and to also optimize the
processor power consumption, Intel introduced a new feature called Intel Turbo Boost. Intel
Turbo Boost is an innovative feature that automatically allows active processor cores to run
faster than the base operating frequency when certain conditions are met.

Intel Turbo Boost is activated when the OS requests the highest processor performance state. The
maximum frequency of the specific processing core on the Core i7 processor is dependent on the
number of active cores, and the amount of time the processor spends in the Turbo Boost state
depends on the workload and operating environment.

Figure 3. Intel Turbo Boost features offer processing performance gains for all applications
regardless of the number of execution threads created.

Figure 3 illustrates how the operating frequencies of the processing cores in the quad-core Core
i7 processor change to offer the best performance for a specific workload type. In an idle state,
all four cores operate at their base clock frequency. If an application that creates four discrete
execution threads is initiated, then all four processing cores start operating at the quad-core turbo
frequency. If the application creates only two execution threads, then two idle cores are put in a
low-power state and their power is diverted to the two active cores to allow them to run at an
even higher clock frequency. Similar behavior would apply in the case where the applications
generate only a single execution thread.
The Intel Core i7-820QM quad-core processor that is used in the NI PXIe-8133 embedded
controller has a base
Number Base Maximum clock frequency of 1.73
of Active Mode Frequenc Turbo Boos GHz. If the application is
Cores y t Frequency using only one CPU core,
Turbo Boost technology
automatically increases
Quad- the clock frequency of
4 1.73 GHz 2.0 GHz the active CPU core on
Core
the Intel Core i7-820QM
processor from 1.7
23 Dual-Core 1.73 GHz
GHz 2.8 GHz to
providing optimal
performance for all
Single- application types.
1 1.73 GHz 3.06 GHz
Core

Figure 4. This table showcases how Turbo Boost is able to increase the performance for a
variety of application

Theduration of time that the processor spends in a specific Turbo Boost state depends on how
soon it reaches thermal, power, and current thresholds. With adequate power supply and heat
dissipation solutions, a Core i7 processor can be made to operate in the Turbo Boost state for an
extended duration of time. In the case of the NI PXIe-8133 embedded controller, users can
manually control the number of active processor cores through the controller’s BIOS to fine tune
the operation of the Turbo Boost feature for optimizing performance for specific application
types.

For real-time applications, Intel Turbo Boost could be utilized, but to ensure best possible
execution determinism thorough testing should be done. When using the NI PXIe-8133
embedded controller, Intel Turbo Boost can be disabled through the BIOS for applications that
prefer to not use it.

4 – Improved Cache Latency with Smart L3 Cache


Cache is a block of high-speed memory for temporary data storage located on the same silicon
die as the CPU. If a single processing core, in a multicore CPU, requires specific data while
executing an instruction set, it first searches for the data in its local caches (L1 and L2). If the
data is not available, also known as a cache-miss, it then accesses the larger L3 cache. In an
exclusive L3 cache, if that attempt is unsuccessful, then the core performs cache snooping –
searches the local caches of other cores – to check whether they have data that it needs. If this
attempt also results in a cache-miss, it then accesses the slower system RAM for that
information. The latency of reading and writing from the cache is much lower than that from the
system RAM, therefore a smarter and larger cache greatly helps in improving processor
performance.

The Core i7 family of processors features an inclusive shared L3 cache that can be up to 12 MB
in size. Figure 4 shows the different types of caches and their layout for the Core i7-820QM
quad-core processor used in the NI PXIe-8133 embedded controller. The NI PXIe-8133
embedded controller features four cores, where each core has 32 kilobytes for instructions and 32
kilobytes for data of L1 cache, 256 kilobytes per core of L2 cache, along with 8 megabytes of
shared L3 cache. The L3 cache is shared across all cores and its inclusive nature helps increase
performance and reduces latency by reducing cache snooping traffic to the processor cores. An
inclusive shared L3 cache guarantees that if there is a cache-miss, then the data is outside the
processor and not available in the local caches of other cores, which eliminates unnecessary
cache snooping.

Figure 4. The inclusive shared L3 cache in the Core i7 processors offers better cache latency for
increased performance.

This feature provides improvement for the overall performance of the processor and is beneficial
for a variety of applications including test, measurement, and control.
5 – Optimized Multithreaded Performance through Hyper-Threading
Intel introduced Hyper-Threading Technology on its processors in 2002. Hyper-threading
exposes a single physical processing core as two logical cores to allow them to share resources
between execution threads and therefore increase the system efficiency (see Figure 5). Because
of the lack of OSs that could clearly differentiate between logical and physical processing cores,
Intel removed this feature when it introduced multicore CPUs. With the release of OSs such as
Windows Vista and Windows 7, which are fully aware of the differences between logical and
physical core, Intel brought back the hyper-threading feature in the Core i7 family of processors.

Hyper-Threading Technology benefits from larger caches and increased memory bandwidth of
the Core i7 processors, delivering greater throughput and responsiveness for multithreaded
applications.

Figure 5.  Hyper-threading allows simultaneous execution of two execution threads on the same
physical CPU core.

6 – Higher Data-Throughput via PCI Express 2.0 and DDR3 Memory Interface
To support the need of modern applications to move data at a faster rate, the Core i7 processors
offer increased throughput for the external databus and its memory channels.

The new processors feature the PCI Express 2.0 databus, which doubles the data throughput from
PCI Express 1.0 while maintaining full hardware and software compatibility with PCI Express
1.0. A x16 PCI Express 2.0 link has a maximum throughput of 8 GB/s/direction.

To allow data from the PCI Express 2.0 databus to be stored in system RAM, the Core i7
processors feature multiple DDR3 1333 MHz memory channels. A system with two channels of
DDR3 1333 MHz RAM had a theoretical memory bandwidth of 21.3 GB/s. This throughput
matches well with the theoretical maximum throughput of a x16 PCI Express 2.0 link. The NI
PXIe-8133 embedded controller uses both of these features to allow users to theoretical stream
data at 8 GB/s in a PXI Express system. 

Certain test and measurement applications – such as high-speed design validation and RF record
and playback – that require continuous acquisition or generation of data at extremely high rates
benefit greatly from these improvements.

7 – Improved Virtualization Performance


Virtualization is a technology that enables running multiple OSs side-by-side on the same
processing hardware. In the test, measurement, and control space, engineers and scientists have
used this technology to consolidate discrete computing nodes into a single system. With the
Nehalem mircoarchitecture, Intel has added new features such as hardware-assisted page-table
management and directed I/O in the Core i7 processors and its chipsets that allow software to
further improve their performance in virtualized environments.

These improvements coupled with increases in memory bandwidth and processing performance
allow engineers and scientists to build more capable and complex virtualized systems for test,
measurement, and control.

8 – Remote Management of Networked Systems with Intel Active Management Technology


(AMT)
AMT provides system administrators the ability to remotely monitor, maintain, and update
systems. Intel AMT is part of the Intel Management Engine, which is built into the chipset of a
Nehalem-based system. This feature allows administrators to boot systems from a remote media,
track hardware and software assets, and perform remote troubleshooting and recovery.

Engineers can use this feature for managing deployed automated test or control systems that need
high uptime. Test, measurement, and control applications are able to use AMT to perform remote
data collection and monitor application status. When an application or system failure occurs,
AMT enables the user to remotely diagnose the problem and access debug screens. This allows
for the problem to be resolved sooner and no longer requires interaction with the actual system.
When software updates are required, AMT allows for these to be done remotely, ensuring that
the system is updated as quickly as possible since downtime can be very costly. AMT is able to
provide many remote management benefits for PXI

COST OF INTEL CORE I7 PROCESSOR

Intel® Core™ i7-8705G Processor

 8 MB Cache
 4 Cores
 8 Threads
 4.10 GHz Max Turbo Frequency
 G - Includes discrete graphics on package
 8th Generation
 Php 9,895

Intel® Core™ i7-6700K Processor

 8 MB Cache
 4 Cores
 8 Threads
 4.10 GHz Max Turbo Frequency
 G - Includes discrete graphics on package
 8th Generation
 Php 17,500
QUESTIONS:
1. The year Intel Core i7 was produced.
2. Process that allows the two programs to be process at separate threads at the
same time.
3. The fastest i7 processor clock speed have by now.
4. is one in which single instruction stream has ability to process multiple data
streams simultaneously these machines are useful in applications such as
general digital signal processing, image processing and multimedia applications
such as audio and video.
5. What is (GMCH)?
6. What is (QPI)?
7. Also known as the northbridge
8. Also known as the southbridge
9. What is (FSB)?
10.What is (ICH)?

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy