Ah that information is not obviously stated in the manual.
3.4. SIMD support
GLM provides some SIMD optimizations based on compiler intrinsics. These optimizations will be
automatically thanks to compiler arguments. For example, if a program is compiled with Visual Studio using
/arch:AVX, GLM will detect this argument and generate code using AVX instructions automatically when
available.
It’s possible to avoid the instruction set detection by forcing the use of a specific instruction set with one of
the fallowing define: GLM_FORCE_SSE2, GLM_FORCE_SSE3, GLM_FORCE_SSSE3, GLM_FORCE_SSE41, GLM_FORCE_SSE42,
GLM_FORCE_AVX, GLM_FORCE_AVX2 or GLM_FORCE_AVX512.
7.8. Is GLM fast?
Following the Pareto principle where 20% of the code consumes 80% of the execution time, GLM operates
perfectly on the 80% of the code that consumes 20% of the performances. Furthermore, thanks to the
lowp, mediump and highp qualifiers, GLM provides approximations which trade precision for performance.
Finally, GLM can automatically produce SIMD optimized code for functions of its implementation.
Here as you found in the advanced mathmatics section of the 9.2 api it states
GLM provides some SIMD optimizations based on compiler intrinsics. These optimizations will be automatically utilized based on the build environment. These optimizations are mainly available through the extensions GLM_GTX_simd_vec4: SIMD vec4 type and functions and GLM_GTX_simd_mat4: SIMD mat4 type and functions.
That information is a bit outdated at least for the simple fact that intrinsics are no longer separated out into SIMD types but are built into vec4,mat4,quat and that it is are no longer experimental in the API.
Aside from that is there any reason why someone still couldn't or shouldn't replace the entire vector/matrix/math implementations with GLM?
Dream Big Or Go Home.
Help Me Help You.