MarketAdvanced Vector Extensions
Company Profile

Advanced Vector Extensions

Advanced Vector Extensions are SIMD extensions to the x86 instruction set architecture for microprocessors from Intel and Advanced Micro Devices (AMD). They were proposed by Intel in March 2008 and first supported by Intel with the Sandy Bridge microarchitecture shipping in Q1 2011 and later by AMD with the Bulldozer microarchitecture shipping in Q4 2011. AVX provides new features, new instructions, and a new coding scheme.

{{Anchor|AVX1}}Advanced Vector Extensions
AVX uses sixteen YMM registers to perform a single instruction on multiple pieces of data (see SIMD). Each YMM register can hold and do simultaneous operations (math) on: • eight 32-bit single-precision floating-point numbers or • four 64-bit double-precision floating-point numbers. The width of the SIMD registers is increased from 128 bits to 256 bits, and renamed from XMM0–XMM7 to YMM0–YMM7 (in x86-64 mode, from XMM0–XMM15 to YMM0–YMM15). The legacy SSE instructions can still be utilized via the VEX prefix to operate on the lower 128 bits of the YMM registers. AVX introduces a three-operand SIMD instruction format called VEX coding scheme, where the destination register is distinct from the two source operands. For example, an SSE instruction using the conventional two-operand form can now use a non-destructive three-operand form , preserving both source operands. Originally, AVX's three-operand format was limited to the instructions with SIMD operands (YMM), and did not include instructions with general purpose registers (e.g. EAX). It was later used for coding new instructions on general purpose registers in later extensions, such as BMI. VEX coding is also used for instructions operating on the k0-k7 mask registers that were introduced with AVX-512. The alignment requirement of SIMD memory operands is relaxed. Unlike their non-VEX coded counterparts, most VEX coded vector instructions no longer require their memory operands to be aligned to the vector size. Notably, the VMOVDQA instruction still requires its memory operand to be aligned. The new VEX coding scheme introduces a new set of code prefixes that extends the opcode space, allows instructions to have more than two operands, and allows SIMD vector registers to be longer than 128 bits. The VEX prefix can also be used on the legacy SSE instructions giving them a three-operand form, and making them interact more efficiently with AVX instructions without the need for VZEROUPPER and VZEROALL. The AVX instructions support both 128-bit and 256-bit SIMD. The 128-bit versions can be useful to improve old code without needing to widen the vectorization, and avoid the penalty of going from SSE to AVX; they are also faster on some early AMD implementations of AVX. This mode is sometimes known as AVX-128. Compared to SSE series, AVX further enhanced performance for digital media playback, web browsing, asymmetric encryption, and others. New instructions These AVX instructions are in addition to the ones that are 256-bit extensions of the legacy 128-bit SSE instructions; most are usable on both 128-bit and 256-bit operands. CPUs with AVX IntelSandy Bridge processors (Q1 2011) and newer, except models branded as Celeron and Pentium. • Pentium and Celeron branded processors starting with Tiger Lake (Q3 2020) and newer. • AMD: • Bulldozer processors (Q4 2011) and newer, including Jaguar and Puma. Issues regarding compatibility between future Intel and AMD processors are discussed under XOP instruction set. • VIA: • Nano QuadCore • Eden X4 • Zhaoxin: • WuDaoKou-based processors (KX-5000 and KH-20000) Compiler and assembler support Absoft supports with - flag. • The Free Pascal compiler supports AVX and AVX2 with the -CfAVX and -CfAVX2 switches from version 2.7.1. • RAD studio (v11.0 Alexandria) supports AVX2 and AVX-512. • The GNU Assembler (GAS) inline assembly functions support these instructions (accessible via GCC), as do Intel primitives and the Intel inline assembler (closely compatible to GAS, although more general in its handling of local references within inline code). GAS supports AVX starting with binutils version 2.19. • GCC starting with version 4.6 (although there was a 4.3 branch with certain support) and the Intel Compiler Suite starting with version 11.1 support AVX. • The Open64 compiler version 4.5.1 supports AVX with - flag. • PathScale supports via the - flag. • The Vector Pascal compiler supports AVX via the - flag. • The Visual Studio 2010/2012 compiler supports AVX via intrinsic and switch. • NASM starting with version 2.03 and newer. There were numerous bug fixes and updates related to AVX in version 2.04. • Other assemblers such as MASM VS2010 version, YASM, FASM and JWASM. Operating system support AVX adds new register-state through the 256-bit wide YMM register file, so explicit operating system support is required to properly save and restore AVX's expanded registers between context switches. The following operating system versions support AVX: • DragonFly BSD: support added in early 2013. • FreeBSD: support added in a patch submitted on January 21, 2012, which was included in the 9.1 stable release. • Linux: supported since kernel version 2.6.30, released on June 9, 2009. • macOS: support added in 10.6.8 (Snow Leopard) update released on June 23, 2011. In fact, macOS Ventura does not support x86 processors without the AVX2 instruction set. • OpenBSD: support added on March 21, 2015. • Solaris: supported in Solaris 10 Update 10 and Solaris 11. • Windows: supported since Windows 7 SP1 and Windows Server 2008 R2 SP1. • Windows Server 2008 R2 SP1 with Hyper-V requires a hotfix to support AMD AVX (Opteron 6200 and 4200 series) processors, KB2568088 • Windows XP and Windows Server 2003 do not support AVX in both kernel drivers and user applications. == Advanced Vector Extensions 2 ==
{{Anchor|AVX2}}Advanced Vector Extensions 2
Advanced Vector Extensions 2 (AVX2), also known as Haswell New Instructions, is an expansion of the AVX instruction set introduced in Intel's Haswell microarchitecture. AVX2 makes the following additions: • expansion of most vector integer SSE and AVX instructions to 256 bits • Gather support, enabling vector elements to be loaded from non-contiguous memory locations • DWORD- and QWORD-granularity any-to-any permutes • vector shifts. Sometimes three-operand fused multiply-accumulate (FMA3) extension is considered part of AVX2, as it was introduced by Intel in the same processor microarchitecture. This is a separate extension using its own CPUID flag, described on its own page and not below. New instructions CPUs with AVX2 IntelHaswell processors (Q2 2013) and newer, except models branded as Celeron and Pentium. • Celeron and Pentium branded processors starting with Tiger Lake (Q3 2020) and newer. • AMDExcavator processors (Q2 2015) and newer. • VIA: • Nano QuadCore • Eden X4 == AVX-512 ==
AVX-512
AVX-512 are 512-bit extensions to the 256-bit Advanced Vector Extensions SIMD instructions for x86 instruction set architecture proposed by Intel in July 2013. • AVX-512 Doubleword and Quadword Instructions (DQ) enhanced 32-bit and 64-bit integer operations • AVX-512 Vector Byte Manipulation Instructions (VBMI) adds vector byte permutation instructions which are not present in AVX-512BW. • AVX-512 Vector Neural Network Instructions Word variable precision (4VNNIW) vector instructions for deep learning. • AVX-512 Fused Multiply Accumulation Packed Single precision (4FMAPS) vector instructions for deep learning. • VPOPCNTDQ count of bits set to 1. • VPCLMULQDQ carry-less multiplication of quadwords. : Intel does not officially support AVX-512 family of instructions on the Alder Lake microprocessors. In early 2022, Intel began disabling in silicon (fusing off) AVX-512 in Alder Lake microprocessors to prevent customers from enabling AVX-512. In older Alder Lake family CPUs with some legacy combinations of BIOS and microcode revisions, it was possible to execute AVX-512 family instructions when disabling all the efficiency cores which do not contain the silicon for AVX-512. Compilers supporting AVX-512 Clang 3.9 and newer • GCC 4.9 and newer • ICC 15.0.1 and newer • Microsoft Visual Studio 2017 C++ Compiler Assemblers supporting AVX-512 FASMNASM 2.11 and newer == AVX-VNNI, AVX-IFMA ==
{{Anchor|AVX-VNNI|AVX-IFMA}}AVX-VNNI, AVX-IFMA
AVX-VNNI is a VEX-coded variant of the AVX512-VNNI instruction set extension. Similarly, AVX-IFMA is a VEX-coded variant of AVX512-IFMA. These extensions provide the same sets of operations as their AVX-512 counterparts, but are limited to 256-bit vectors and do not support any additional features of EVEX encoding, such as broadcasting, opmask registers or accessing more than 16 vector registers. These extensions allow for supporting VNNI and IFMA operations even when AVX-512 is not implemented in the processor. CPUs with AVX-VNNI IntelAlder Lake processors (Q4 2021) and newer. • AMDZen 5 processors (Q3 2024) and newer. CPUs with AVX-IFMA IntelSierra Forest E-core-only Xeon processors (Q2 2024) and newer. • Grand Ridge special-purpose processors and newer. • Meteor Lake mobile processors (Q4 2023) and newer. • Arrow Lake desktop processors (Q4 2024) and newer. • AMDZen 6 processors and newer. == AVX-NE-CONVERT ==
AVX-NE-CONVERT
AVX-NE-CONVERT introduces a set of instructions for converting between Bfloat16 (BF16), half-precision (FP16) and single-precision (FP32) floating-point numbers. The new instructions are VEX-coded, and therefore are limited to AVX2 vector registers and lack opmask support from AVX-512. Broadcast is only supported by means of two special instructions for loading numbers from memory, it is not supported as an instruction encoding feature. New instructions CPUs with AVX-NE-CONVERT IntelSierra Forest E-core-only Xeon processors (Q2 2024) and newer. • Grand Ridge special-purpose processors and newer. • Lunar Lake mobile processors (Q2 2024) and newer. • Arrow Lake desktop processors (Q4 2024) and newer. • Diamond Rapids Xeon processors and newer. • AMDZen 6 processors and newer. == AVX-VNNI-INT8, AVX-VNNI-INT16 ==
{{Anchor|AVX-VNNI-INT8|AVX-VNNI-INT16}}AVX-VNNI-INT8, AVX-VNNI-INT16
These instruction sets further extend the AVX-VNNI extension by adding support for more combinations of input data types for the VPDP* series of instructions. Where VPDPBUSD(S) instructions from AVX-VNNI take a vector of unsigned bytes as the first input operand and a vector of signed bytes as the second input operand, AVX-VNNI-INT8 adds variants of these instructions that support signed and unsigned byte inputs at any position. Similarly, where VPDPWSSD(S) from AVX-VNNI take two vectors of signed 16-bit words as input operands, AVX-VNNI-INT16 adds support for signed and unsigned 16-bit word inputs at any position. For the instructions accepting signed and unsigned inputs, there are distinct instructions for the two possible orders of inputs (signed/unsigned and unsigned/signed) because VEX encoding only supports the second input operand to be a memory operand. This allows any of the supported data types to be loaded from memory by the instruction. New instructions in AVX-VNNI-INT8 New instructions in AVX-VNNI-INT16 CPUs with AVX-VNNI-INT8 IntelSierra Forest E-core-only Xeon processors (Q2 2024) and newer. • Grand Ridge special-purpose processors and newer. • Lunar Lake mobile processors (Q2 2024) and newer. • Arrow Lake desktop processors (Q4 2024) and newer. • Diamond Rapids Xeon processors and newer. • AMDZen 6 processors and newer. CPUs with AVX-VNNI-INT16 IntelClearwater Forest E-core-only Xeon processors and newer. • Lunar Lake mobile processors (Q2 2024) and newer. • Arrow Lake desktop processors (Q4 2024) and newer. • Diamond Rapids Xeon processors and newer. == AVX10 ==
AVX10
AVX10, announced in July 2023, is a new, "converged" AVX instruction set. It addresses several issues of AVX-512; in particular, that it is split into too many parts (20 feature flags). The initial technical paper also made 512-bit vectors optional to support, but as of revision 3.0, vector length enumeration is removed and 512-bit vectors are mandatory. AVX10 presents a simplified CPUID interface to test for instruction support, consisting of the AVX10 version number (indicating the set of instructions supported, with later versions always being a superset of an earlier one). For example, AVX10.2 indicates that a CPU is capable of the second version of AVX10. AVX10.1 was first released in Intel Granite Rapids and Nova Lake. == Applications ==
Applications
• Suitable for floating-point-intensive calculations in multimedia, scientific and financial applications (AVX2 adds support for integer operations). • Increases parallelism and throughput in floating-point SIMD calculations. • Reduces register load due to the non-destructive instructions. • Improves Linux RAID software performance (requires AVX2, AVX is not sufficient) Software • Cryptography • BSAFE C toolkits uses AVX and AVX2 where appropriate to accelerate various cryptographic algorithms. • OpenSSL uses AVX- and AVX2-optimized cryptographic functions since version 1.0.2. Support for AVX-512 was added in version 3.0.0. Some of these optimizations are also present in various clones and forks, like LibreSSL. • Multimedia • Blender uses AVX, AVX2 and AVX-512 in the Cycles render engine. • Native Instruments' Massive X softsynth requires AVX. • dav1d AV1 decoder can use AVX2 and AVX-512 on supported CPUs. • SVT-AV1 AV1 encoder can use AVX2 and AVX-512 to accelerate video encoding. • Science, engineering an others • Esri ArcGIS Data Store uses AVX2 for graph storage. • Prime95/MPrime, the software used for GIMPS, started using the AVX instructions since version 27.1, AVX2 since 28.6 and AVX-512 since 29.1. • Einstein@Home uses AVX in some of their distributed applications that search for gravitational waves. • TensorFlow since version 1.6 and tensorflow above versions requires CPU supporting at least AVX. • EmEditor 19.0 and above uses AVX2 to speed up processing. • Microsoft Teams uses AVX2 instructions to create a blurred or custom background behind video chat participants, and for background noise suppression. • [https://github.com/simdjson/simdjson ], a JSON parsing library, uses AVX2 and AVX-512 to achieve improved decoding speed. • x86-simd-sort, a library with sorting algorithms for 16, 32 and 64-bit numeric data types, uses AVX2 and AVX-512. The library is used in NumPy and OpenJDK to accelerate sorting algorithms. • Tesseract OCR engine uses AVX, AVX2 and AVX-512 to accelerate character recognition. == Downclocking ==
Downclocking
Since AVX instructions are wider, they consume more power and generate more heat. Executing heavy AVX instructions at high CPU clock frequencies may affect CPU stability due to excessive voltage droop during load transients. Some Intel processors have provisions to reduce the Turbo Boost frequency limit when such instructions are being executed. This reduction happens even if the CPU hasn't reached its thermal and power consumption limits. On Skylake and its derivatives, the throttling is divided into three levels: • L0 (100%): The normal turbo boost limit. • L1 (~85%): The "AVX boost" limit. Soft-triggered by 256-bit "heavy" (floating-point unit: FP math and integer multiplication) instructions. Hard-triggered by "light" (all other) 512-bit instructions. • L2 (~60%): The "AVX-512 boost" limit. Soft-triggered by 512-bit heavy instructions. The frequency transition can be soft or hard. Hard transition means the frequency is reduced as soon as such an instruction is spotted; soft transition means that the frequency is reduced only after reaching a threshold number of matching instructions. The limit is per-thread. • L0 (100%): The normal turbo boost limit. • L1 (~97%): Triggered by any 512-bit instructions, but only when single-core boost is active; not triggered when multiple cores are loaded. Rocket Lake processors do not trigger frequency reduction upon executing any kind of vector instructions regardless of the vector size. On supported and unlocked variants of processors that down-clock, the clock ratio reduction offsets (typically called AVX and AVX-512 offsets) are adjustable and may be turned off entirely (set to 0x) via Intel's Overclocking / Tuning utility or in BIOS if supported there. == See also ==
tickerdossier.comtickerdossier.substack.com