Skip to content

Conversation

@bosilca
Copy link
Member

@bosilca bosilca commented Jan 6, 2026

Check first for _Float128 then __float128 and then for _Quad type (Intel compiler) and pick the one that matches the Fortran REAL16 type. This commit adds support for pack, unpack as well as MPI_Op.

As of today, no hardware support (AVX* or SVE) exists for the 128 bits floating point, so the only valid MPI_Op for this type is the base op (provided in software by most recent compilers).

Fixes #13564

@bosilca bosilca force-pushed the topic/better_support_for_fortran_real16 branch from 9d83217 to 846dfb9 Compare January 6, 2026 05:00
Copy link
Contributor

@devreal devreal left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Minor nit-pick. Can we document in the code why we pick the order the way it is?

@bosilca bosilca force-pushed the topic/better_support_for_fortran_real16 branch from 846dfb9 to 0148404 Compare January 6, 2026 16:23
Check first for _Float128 then __float128 and then for _Quad type (Intel
compiler) and pick the one that matches the Fortran REAL16 type. This
commit adds support for pack, unpack as well as MPI_Op.

As of today, no hardware support (AVX* or SVE) exists for the 128 bits floating
point, so the only valid MPI_Op for this type is the base op (provided
in software by most recent compilers).

Signed-off-by: George Bosilca <gbosilca@nvidia.com>
@bosilca bosilca force-pushed the topic/better_support_for_fortran_real16 branch from 0148404 to 58cd5a4 Compare January 6, 2026 16:39
@bosilca
Copy link
Member Author

bosilca commented Jan 6, 2026

Minor nit-pick. Can we document in the code why we pick the order the way it is?

Which order ? During configure or in the header files ?

@bosilca bosilca self-assigned this Jan 6, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

quad precision (128 bit) reductions return bad results

2 participants