Opcode/ Instruction |
Op/ En |
64/32 bit Mode Support |
CPUID Feature Flag |
Description |
EVEX.128.NP.MAP5.W0 5B /r VCVTDQ2PH xmm1{k1}{z}, xmm2/ m128/m32bcst |
A |
V/V |
AVX512-FP16 AVX512VL |
Convert four packed signed doubleword integers from xmm2/m128/m32bcst to four packed FP16 values, and store the result in xmm1 subject to writemask k1. |
EVEX.256.NP.MAP5.W0 5B /r VCVTDQ2PH xmm1{k1}{z}, ymm2/ m256/m32bcst |
A |
V/V |
AVX512-FP16 AVX512VL |
Convert eight packed signed doubleword integers from ymm2/m256/m32bcst to eight packed FP16 values, and store the result in xmm1 subject to writemask k1. |
EVEX.512.NP.MAP5.W0 5B /r VCVTDQ2PH ymm1{k1}{z}, zmm2/ m512/m32bcst {er} |
A |
V/V |
AVX512-FP16 |
Convert sixteen packed signed doubleword integers from zmm2/m512/m32bcst to sixteen packed FP16 values, and store the result in ymm1 subject to writemask k1. |
Op/En |
Tuple |
Operand 1 |
Operand 2 |
Operand 3 |
Operand 4 |
A |
Full |
ModRM:reg (w) |
ModRM:r/m (r) |
N/A |
N/A |
This instruction converts four, eight, or sixteen packed signed doubleword integers in the source operand to four, eight, or sixteen packed FP16 values in the destination operand.
EVEX encoded versions: The source operand can be a ZMM/YMM/XMM register, a 512/256/128-bit memory loca- tion or a 512/256/128-bit vector broadcast from a 32-bit memory location. The destination operand is a YMM/XMM register conditionally updated with writemask k1.
EVEX.vvvv is reserved and must be 1111b, otherwise instructions will #UD.
If the result of the convert operation is overflow and MXCSR.OM=0 then a SIMD exception will be raised with OE=1, PE=1.
VL = 128, 256 or 512 KL := VL / 32 IF *SRC is a register* and (VL = 512) AND (EVEX.b = 1): SET_RM(EVEX.RC) ELSE: SET_RM(MXCSR.RC) FOR j := 0 TO KL-1: IF k1[j] OR *no writemask*: IF *SRC is memory* and EVEX.b = 1: tsrc := SRC.dword[0] ELSE tsrc := SRC.dword[j] DEST.fp16[j] := Convert_integer32_to_fp16(tsrc) ELSE IF *zeroing*: DEST.fp16[j] := 0 // else dest.fp16[j] remains unchanged DEST[MAXVL-1:VL/2] := 0
VCVTDQ2PH __m256h _mm512_cvt_roundepi32_ph (__m512i a, int rounding); VCVTDQ2PH __m256h _mm512_mask_cvt_roundepi32_ph (__m256h src, __mmask16 k, __m512i a, int rounding); VCVTDQ2PH __m256h _mm512_maskz_cvt_roundepi32_ph (__mmask16 k, __m512i a, int rounding); VCVTDQ2PH __m128h _mm_cvtepi32_ph (__m128i a); VCVTDQ2PH __m128h _mm_mask_cvtepi32_ph (__m128h src, __mmask8 k, __m128i a); VCVTDQ2PH __m128h _mm_maskz_cvtepi32_ph (__mmask8 k, __m128i a); VCVTDQ2PH __m128h _mm256_cvtepi32_ph (__m256i a); VCVTDQ2PH __m128h _mm256_mask_cvtepi32_ph (__m128h src, __mmask8 k, __m256i a); VCVTDQ2PH __m128h _mm256_maskz_cvtepi32_ph (__mmask8 k, __m256i a); VCVTDQ2PH __m256h _mm512_cvtepi32_ph (__m512i a); VCVTDQ2PH __m256h _mm512_mask_cvtepi32_ph (__m256h src, __mmask16 k, __m512i a); VCVTDQ2PH __m256h _mm512_maskz_cvtepi32_ph (__mmask16 k, __m512i a);
Overflow, Precision.
EVEX-encoded instructions, see Table 2-46, "Type E2 Class Exception Conditions."