Difference between revisions of "Power ISA/Vector Operations"

From RCS Wiki
Jump to navigation Jump to search
(Power ISA v3.1 adds MMA)
 
(4 intermediate revisions by one other user not shown)
Line 8: Line 8:
  
 
According to [[File:POWER9-Features-and-Specifications.pdf]] page 7, the Vector Scalar Unit (VSU)'s 128-bit hardware is dedicated per super-slice (2 threads).  This may indicate that trying to aggressively use 128-bit VSX instructions in two threads that use the same super-slice will be inefficient.  It is possible that clever usage of <code>taskset</code> may improve this situation.
 
According to [[File:POWER9-Features-and-Specifications.pdf]] page 7, the Vector Scalar Unit (VSU)'s 128-bit hardware is dedicated per super-slice (2 threads).  This may indicate that trying to aggressively use 128-bit VSX instructions in two threads that use the same super-slice will be inefficient.  It is possible that clever usage of <code>taskset</code> may improve this situation.
 +
 +
Power ISA v3.1 adds an optional VSX extension, MMA, targeted at matrix math applications.
  
 
== External Links ==
 
== External Links ==
Line 20: Line 22:
 
* Ruzhu Chen. [https://developer.ibm.com/learningpaths/port-your-app-to-lop/migrate-app-on-lop/ Linux on IBM Power Systems application porting and tuning guide]. Published September 27, 2020
 
* Ruzhu Chen. [https://developer.ibm.com/learningpaths/port-your-app-to-lop/migrate-app-on-lop/ Linux on IBM Power Systems application porting and tuning guide]. Published September 27, 2020
 
* OpenPOWER. [https://openpowerfoundation.org/compliance/vectorintrinsicprogrammingreference/ Vector Intrinsic Programming Reference Compliance Specification]. Published March 28, 2021
 
* OpenPOWER. [https://openpowerfoundation.org/compliance/vectorintrinsicprogrammingreference/ Vector Intrinsic Programming Reference Compliance Specification]. Published March 28, 2021
 +
* José E. Moreira. [https://arxiv.org/abs/2104.03142 A matrix math facility for Power ISA(TM) processors]. Published April 7, 2021
 +
* José E. Moreira. [https://www.redbooks.ibm.com/Redbooks.nsf/RedpieceAbstracts/redp5612.html Matrix-Multiply Assist Best Practices Guide]. Published April 15, 2021
 +
* Sridhar Venkat. [https://community.ibm.com/community/user/powerdeveloper/blogs/sridhar-venkat1/2022/08/29/power10-mma MMA in IBM Power10 processor]. Published August 29, 2022
  
== Github pages ==
+
== Github / Gitlab pages ==
  
* [https://github.com/ermig1979/Simd C++ image processing and machine learning library with using of SIMD]
+
* Eigen. [https://gitlab.com/libeigen/eigen A C++ template library for linear algebra: matrices, vectors, numerical solvers and related algorithms]
* [https://github.com/simd-everywhere/simde Implementations of SIMD instruction sets for systems which don't natively support them]
+
* Simd Library. [https://github.com/ermig1979/Simd C++ image processing and machine learning library with using of SIMD]
* [https://github.com/jfalcou/eve Expressive Vector Engine - SIMD in C++]
+
* SIMD Everywhere. [https://github.com/simd-everywhere/simde Implementations of SIMD instruction sets for systems which don't natively support them]
* [https://github.com/VectorChief/UniSIMD-assembler SIMD macro assembler unified for ARM, MIPS, PPC and x86]
+
* EVE - the Expressive Vector Engine. [https://github.com/jfalcou/eve SIMD in C++]
* [https://github.com/powturbo/Turbo-Base64 Turbo Base64 - Fastest Base64 SIMD:SSE/AVX2/AVX512/Neon/Altivec]
+
* UniSIMD Assembler. [https://github.com/VectorChief/UniSIMD-assembler SIMD macro assembler unified for ARM, MIPS, PPC and x86]
* [https://github.com/berenger-eu/inastemp Intrinsics as template - is a basic library to use vectorization easily in C++]
+
* Turbo Base64. [https://github.com/powturbo/Turbo-Base64 Fastest Base64 SIMD:SSE/AVX2/AVX512/Neon/Altivec]
* [https://github.com/p12tic/libsimdpp Portable header-only C++ low level SIMD library]
+
* Inastemp. [https://gitlab.inria.fr/bramas/inastemp Intrinsics as template - is a basic library to use vectorization easily in C++]
* [https://github.com/open-power-sdk/pveclib Power Vector Library]
+
* libsimdpp. [https://github.com/p12tic/libsimdpp Portable header-only C++ low level SIMD library]
* [https://github.com/shibatch/sleef SIMD Library for Evaluating Elementary Functions, vectorized libm and DFT]
+
* pveclib. [https://github.com/open-power-sdk/pveclib Power Vector Library]
* [https://github.com/VectorCamp/libfreevec SIMD optimized C library]
+
* SLEEF. [https://github.com/shibatch/sleef SIMD Library for Evaluating Elementary Functions, vectorized libm and DFT]
 +
* libjpeg-turbo. [https://github.com/libjpeg-turbo/libjpeg-turbo A JPEG image codec that uses SIMD instructions to accelerate baseline JPEG compression and decompression]
 +
* libfreevec. [https://github.com/VectorCamp/libfreevec SIMD optimized C library]

Latest revision as of 06:54, 8 November 2023

The Power Architecture ISA includes a specification of vector or SIMD operations. Prior to the Power ISA, i.e. PowerPC, some of these operations were available, but defined in an external standard, called Altivec by Freescale (Motorola spin-off), Vector Multimedia Extension (VMX) by IBM, and Velocity Engine by Apple.

The Vector operations are classified as Vector Facility and Vector Scalar Extension (VSX) in current versions of the Power ISA.

Power ISA v2.07 still refers to some instructions as VMX in its summary of changes since the previous version, but the rest of the document avoids mentioning VMX completely.

Power ISA v3.0 no longer mentions VMX at all.

According to File:POWER9-Features-and-Specifications.pdf page 7, the Vector Scalar Unit (VSU)'s 128-bit hardware is dedicated per super-slice (2 threads). This may indicate that trying to aggressively use 128-bit VSX instructions in two threads that use the same super-slice will be inefficient. It is possible that clever usage of taskset may improve this situation.

Power ISA v3.1 adds an optional VSX extension, MMA, targeted at matrix math applications.

External Links

Github / Gitlab pages