Closed
Description
The reference and gcc (generated with https://gist.github.com/gnzlbg/3a27cca666f4a9f8a51d2eb008ce914c from the gcc docs):
-
fn add_a_b(i8x16, i8x16) -> i8x16
-
fn add_a_h(i16x8, i16x8) -> i16x8
-
fn add_a_w(i32x4, i32x4) -> i32x4
-
fn add_a_d(i64x2, i64x2) -> i64x2
-
fn adds_a_b(i8x16, i8x16) -> i8x16
-
fn adds_a_h(i16x8, i16x8) -> i16x8
-
fn adds_a_w(i32x4, i32x4) -> i32x4
-
fn adds_a_d(i64x2, i64x2) -> i64x2
-
fn adds_s_b(i8x16, i8x16) -> i8x16
-
fn adds_s_h(i16x8, i16x8) -> i16x8
-
fn adds_s_w(i32x4, i32x4) -> i32x4
-
fn adds_s_d(i64x2, i64x2) -> i64x2
-
fn adds_u_b(u8x16, u8x16) -> u8x16
-
fn adds_u_h(u16x8, u16x8) -> u16x8
-
fn adds_u_w(u32x4, u32x4) -> u32x4
-
fn adds_u_d(u64x2, u64x2) -> u64x2
-
fn addv_b(i8x16, i8x16) -> i8x16
-
fn addv_h(i16x8, i16x8) -> i16x8
-
fn addv_w(i32x4, i32x4) -> i32x4
-
fn addv_d(i64x2, i64x2) -> i64x2
-
fn addvi_b(i8x16, imm0_31) -> i8x16
-
fn addvi_h(i16x8, imm0_31) -> i16x8
-
fn addvi_w(i32x4, imm0_31) -> i32x4
-
fn addvi_d(i64x2, imm0_31) -> i64x2
-
fn and_v(u8x16, u8x16) -> u8x16
-
fn andi_b(u8x16, imm0_255) -> u8x16
-
fn asub_s_b(i8x16, i8x16) -> i8x16
-
fn asub_s_h(i16x8, i16x8) -> i16x8
-
fn asub_s_w(i32x4, i32x4) -> i32x4
-
fn asub_s_d(i64x2, i64x2) -> i64x2
-
fn asub_u_b(u8x16, u8x16) -> u8x16
-
fn asub_u_h(u16x8, u16x8) -> u16x8
-
fn asub_u_w(u32x4, u32x4) -> u32x4
-
fn asub_u_d(u64x2, u64x2) -> u64x2
-
fn ave_s_b(i8x16, i8x16) -> i8x16
-
fn ave_s_h(i16x8, i16x8) -> i16x8
-
fn ave_s_w(i32x4, i32x4) -> i32x4
-
fn ave_s_d(i64x2, i64x2) -> i64x2
-
fn ave_u_b(u8x16, u8x16) -> u8x16
-
fn ave_u_h(u16x8, u16x8) -> u16x8
-
fn ave_u_w(u32x4, u32x4) -> u32x4
-
fn ave_u_d(u64x2, u64x2) -> u64x2
-
fn aver_s_b(i8x16, i8x16) -> i8x16
-
fn aver_s_h(i16x8, i16x8) -> i16x8
-
fn aver_s_w(i32x4, i32x4) -> i32x4
-
fn aver_s_d(i64x2, i64x2) -> i64x2
-
fn aver_u_b(u8x16, u8x16) -> u8x16
-
fn aver_u_h(u16x8, u16x8) -> u16x8
-
fn aver_u_w(u32x4, u32x4) -> u32x4
-
fn aver_u_d(u64x2, u64x2) -> u64x2
-
fn bclr_b(u8x16, u8x16) -> u8x16
-
fn bclr_h(u16x8, u16x8) -> u16x8
-
fn bclr_w(u32x4, u32x4) -> u32x4
-
fn bclr_d(u64x2, u64x2) -> u64x2
-
fn bclri_b(u8x16, imm0_7) -> u8x16
-
fn bclri_h(u16x8, imm0_15) -> u16x8
-
fn bclri_w(u32x4, imm0_31) -> u32x4
-
fn bclri_d(u64x2, imm0_63) -> u64x2
-
fn binsl_b(u8x16, u8x16, u8x16) -> u8x16
-
fn binsl_h(u16x8, u16x8, u16x8) -> u16x8
-
fn binsl_w(u32x4, u32x4, u32x4) -> u32x4
-
fn binsl_d(u64x2, u64x2, u64x2) -> u64x2
-
fn binsli_b(u8x16, u8x16, imm0_7) -> u8x16
-
fn binsli_h(u16x8, u16x8, imm0_15) -> u16x8
-
fn binsli_w(u32x4, u32x4, imm0_31) -> u32x4
-
fn binsli_d(u64x2, u64x2, imm0_63) -> u64x2
-
fn binsr_b(u8x16, u8x16, u8x16) -> u8x16
-
fn binsr_h(u16x8, u16x8, u16x8) -> u16x8
-
fn binsr_w(u32x4, u32x4, u32x4) -> u32x4
-
fn binsr_d(u64x2, u64x2, u64x2) -> u64x2
-
fn binsri_b(u8x16, u8x16, imm0_7) -> u8x16
-
fn binsri_h(u16x8, u16x8, imm0_15) -> u16x8
-
fn binsri_w(u32x4, u32x4, imm0_31) -> u32x4
-
fn binsri_d(u64x2, u64x2, imm0_63) -> u64x2
-
fn bmnz_v(u8x16, u8x16, u8x16) -> u8x16
-
fn bmnzi_b(u8x16, u8x16, imm0_255) -> u8x16
-
fn bmz_v(u8x16, u8x16, u8x16) -> u8x16
-
fn bmzi_b(u8x16, u8x16, imm0_255) -> u8x16
-
fn bneg_b(u8x16, u8x16) -> u8x16
-
fn bneg_h(u16x8, u16x8) -> u16x8
-
fn bneg_w(u32x4, u32x4) -> u32x4
-
fn bneg_d(u64x2, u64x2) -> u64x2
-
fn bnegi_b(u8x16, imm0_7) -> u8x16
-
fn bnegi_h(u16x8, imm0_15) -> u16x8
-
fn bnegi_w(u32x4, imm0_31) -> u32x4
-
fn bnegi_d(u64x2, imm0_63) -> u64x2
-
fn bnz_b(u8x16) -> i32
-
fn bnz_h(u16x8) -> i32
-
fn bnz_w(u32x4) -> i32
-
fn bnz_d(u64x2) -> i32
-
fn bnz_v(u8x16) -> i32
-
fn bsel_v(u8x16, u8x16, u8x16) -> u8x16
-
fn bseli_b(u8x16, u8x16, imm0_255) -> u8x16
-
fn bset_b(u8x16, u8x16) -> u8x16
-
fn bset_h(u16x8, u16x8) -> u16x8
-
fn bset_w(u32x4, u32x4) -> u32x4
-
fn bset_d(u64x2, u64x2) -> u64x2
-
fn bseti_b(u8x16, imm0_7) -> u8x16
-
fn bseti_h(u16x8, imm0_15) -> u16x8
-
fn bseti_w(u32x4, imm0_31) -> u32x4
-
fn bseti_d(u64x2, imm0_63) -> u64x2
-
fn bz_b(u8x16) -> i32
-
fn bz_h(u16x8) -> i32
-
fn bz_w(u32x4) -> i32
-
fn bz_d(u64x2) -> i32
-
fn bz_v(u8x16) -> i32
-
fn ceq_b(i8x16, i8x16) -> i8x16
-
fn ceq_h(i16x8, i16x8) -> i16x8
-
fn ceq_w(i32x4, i32x4) -> i32x4
-
fn ceq_d(i64x2, i64x2) -> i64x2
-
fn ceqi_b(i8x16, imm_n16_15) -> i8x16
-
fn ceqi_h(i16x8, imm_n16_15) -> i16x8
-
fn ceqi_w(i32x4, imm_n16_15) -> i32x4
-
fn ceqi_d(i64x2, imm_n16_15) -> i64x2
-
fn cfcmsa(imm0_31) -> i32
-
fn cle_s_b(i8x16, i8x16) -> i8x16
-
fn cle_s_h(i16x8, i16x8) -> i16x8
-
fn cle_s_w(i32x4, i32x4) -> i32x4
-
fn cle_s_d(i64x2, i64x2) -> i64x2
-
fn cle_u_b(u8x16, u8x16) -> i8x16
-
fn cle_u_h(u16x8, u16x8) -> i16x8
-
fn cle_u_w(u32x4, u32x4) -> i32x4
-
fn cle_u_d(u64x2, u64x2) -> i64x2
-
fn clei_s_b(i8x16, imm_n16_15) -> i8x16
-
fn clei_s_h(i16x8, imm_n16_15) -> i16x8
-
fn clei_s_w(i32x4, imm_n16_15) -> i32x4
-
fn clei_s_d(i64x2, imm_n16_15) -> i64x2
-
fn clei_u_b(u8x16, imm0_31) -> i8x16
-
fn clei_u_h(u16x8, imm0_31) -> i16x8
-
fn clei_u_w(u32x4, imm0_31) -> i32x4
-
fn clei_u_d(u64x2, imm0_31) -> i64x2
-
fn clt_s_b(i8x16, i8x16) -> i8x16
-
fn clt_s_h(i16x8, i16x8) -> i16x8
-
fn clt_s_w(i32x4, i32x4) -> i32x4
-
fn clt_s_d(i64x2, i64x2) -> i64x2
-
fn clt_u_b(u8x16, u8x16) -> i8x16
-
fn clt_u_h(u16x8, u16x8) -> i16x8
-
fn clt_u_w(u32x4, u32x4) -> i32x4
-
fn clt_u_d(u64x2, u64x2) -> i64x2
-
fn clti_s_b(i8x16, imm_n16_15) -> i8x16
-
fn clti_s_h(i16x8, imm_n16_15) -> i16x8
-
fn clti_s_w(i32x4, imm_n16_15) -> i32x4
-
fn clti_s_d(i64x2, imm_n16_15) -> i64x2
-
fn clti_u_b(u8x16, imm0_31) -> i8x16
-
fn clti_u_h(u16x8, imm0_31) -> i16x8
-
fn clti_u_w(u32x4, imm0_31) -> i32x4
-
fn clti_u_d(u64x2, imm0_31) -> i64x2
-
fn copy_s_b(i8x16, imm0_15) -> i32
-
fn copy_s_h(i16x8, imm0_7) -> i32
-
fn copy_s_w(i32x4, imm0_3) -> i32
-
fn copy_s_d(i64x2, imm0_1) -> i64
-
fn copy_u_b(i8x16, imm0_15) -> u32
-
fn copy_u_h(i16x8, imm0_7) -> u32
-
fn copy_u_w(i32x4, imm0_3) -> u32
-
fn copy_u_d(i64x2, imm0_1) -> u64
-
fn ctcmsa(imm0_31, i32) -> ()
-
fn div_s_b(i8x16, i8x16) -> i8x16
-
fn div_s_h(i16x8, i16x8) -> i16x8
-
fn div_s_w(i32x4, i32x4) -> i32x4
-
fn div_s_d(i64x2, i64x2) -> i64x2
-
fn div_u_b(u8x16, u8x16) -> u8x16
-
fn div_u_h(u16x8, u16x8) -> u16x8
-
fn div_u_w(u32x4, u32x4) -> u32x4
-
fn div_u_d(u64x2, u64x2) -> u64x2
-
fn dotp_s_h(i8x16, i8x16) -> i16x8
-
fn dotp_s_w(i16x8, i16x8) -> i32x4
-
fn dotp_s_d(i32x4, i32x4) -> i64x2
-
fn dotp_u_h(u8x16, u8x16) -> u16x8
-
fn dotp_u_w(u16x8, u16x8) -> u32x4
-
fn dotp_u_d(u32x4, u32x4) -> u64x2
-
fn dpadd_s_h(i16x8, i8x16, i8x16) -> i16x8
-
fn dpadd_s_w(i32x4, i16x8, i16x8) -> i32x4
-
fn dpadd_s_d(i64x2, i32x4, i32x4) -> i64x2
-
fn dpadd_u_h(u16x8, u8x16, u8x16) -> u16x8
-
fn dpadd_u_w(u32x4, u16x8, u16x8) -> u32x4
-
fn dpadd_u_d(u64x2, u32x4, u32x4) -> u64x2
-
fn dpsub_s_h(i16x8, i8x16, i8x16) -> i16x8
-
fn dpsub_s_w(i32x4, i16x8, i16x8) -> i32x4
-
fn dpsub_s_d(i64x2, i32x4, i32x4) -> i64x2
-
fn dpsub_u_h(i16x8, u8x16, u8x16) -> i16x8
-
fn dpsub_u_w(i32x4, u16x8, u16x8) -> i32x4
-
fn dpsub_u_d(i64x2, u32x4, u32x4) -> i64x2
-
fn fadd_w(f32x4, f32x4) -> f32x4
-
fn fadd_d(f64x2, f64x2) -> f64x2
-
fn fcaf_w(f32x4, f32x4) -> i32x4
-
fn fcaf_d(f64x2, f64x2) -> i64x2
-
fn fceq_w(f32x4, f32x4) -> i32x4
-
fn fceq_d(f64x2, f64x2) -> i64x2
-
fn fclass_w(f32x4) -> i32x4
-
fn fclass_d(f64x2) -> i64x2
-
fn fcle_w(f32x4, f32x4) -> i32x4
-
fn fcle_d(f64x2, f64x2) -> i64x2
-
fn fclt_w(f32x4, f32x4) -> i32x4
-
fn fclt_d(f64x2, f64x2) -> i64x2
-
fn fcne_w(f32x4, f32x4) -> i32x4
-
fn fcne_d(f64x2, f64x2) -> i64x2
-
fn fcor_w(f32x4, f32x4) -> i32x4
-
fn fcor_d(f64x2, f64x2) -> i64x2
-
fn fcueq_w(f32x4, f32x4) -> i32x4
-
fn fcueq_d(f64x2, f64x2) -> i64x2
-
fn fcule_w(f32x4, f32x4) -> i32x4
-
fn fcule_d(f64x2, f64x2) -> i64x2
-
fn fcult_w(f32x4, f32x4) -> i32x4
-
fn fcult_d(f64x2, f64x2) -> i64x2
-
fn fcun_w(f32x4, f32x4) -> i32x4
-
fn fcun_d(f64x2, f64x2) -> i64x2
-
fn fcune_w(f32x4, f32x4) -> i32x4
-
fn fcune_d(f64x2, f64x2) -> i64x2
-
fn fdiv_w(f32x4, f32x4) -> f32x4
-
fn fdiv_d(f64x2, f64x2) -> f64x2
-
fn fexdo_h(f32x4, f32x4) -> i16x8
-
fn fexdo_w(f64x2, f64x2) -> f32x4
-
fn fexp2_w(f32x4, i32x4) -> f32x4
-
fn fexp2_d(f64x2, i64x2) -> f64x2
-
fn fexupl_w(i16x8) -> f32x4
-
fn fexupl_d(f32x4) -> f64x2
-
fn fexupr_w(i16x8) -> f32x4
-
fn fexupr_d(f32x4) -> f64x2
-
fn ffint_s_w(i32x4) -> f32x4
-
fn ffint_s_d(i64x2) -> f64x2
-
fn ffint_u_w(u32x4) -> f32x4
-
fn ffint_u_d(u64x2) -> f64x2
-
fn ffql_w(i16x8) -> f32x4
-
fn ffql_d(i32x4) -> f64x2
-
fn ffqr_w(i16x8) -> f32x4
-
fn ffqr_d(i32x4) -> f64x2
-
fn fill_b(i32) -> i8x16
-
fn fill_h(i32) -> i16x8
-
fn fill_w(i32) -> i32x4
-
fn fill_d(i64) -> i64x2
-
fn flog2_w(f32x4) -> f32x4
-
fn flog2_d(f64x2) -> f64x2
-
fn fmadd_w(f32x4, f32x4, f32x4) -> f32x4
-
fn fmadd_d(f64x2, f64x2, f64x2) -> f64x2
-
fn fmax_w(f32x4, f32x4) -> f32x4
-
fn fmax_d(f64x2, f64x2) -> f64x2
-
fn fmax_a_w(f32x4, f32x4) -> f32x4
-
fn fmax_a_d(f64x2, f64x2) -> f64x2
-
fn fmin_w(f32x4, f32x4) -> f32x4
-
fn fmin_d(f64x2, f64x2) -> f64x2
-
fn fmin_a_w(f32x4, f32x4) -> f32x4
-
fn fmin_a_d(f64x2, f64x2) -> f64x2
-
fn fmsub_w(f32x4, f32x4, f32x4) -> f32x4
-
fn fmsub_d(f64x2, f64x2, f64x2) -> f64x2
-
fn fmul_w(f32x4, f32x4) -> f32x4
-
fn fmul_d(f64x2, f64x2) -> f64x2
-
fn frint_w(f32x4) -> f32x4
-
fn frint_d(f64x2) -> f64x2
-
fn frcp_w(f32x4) -> f32x4
-
fn frcp_d(f64x2) -> f64x2
-
fn frsqrt_w(f32x4) -> f32x4
-
fn frsqrt_d(f64x2) -> f64x2
-
fn fsaf_w(f32x4, f32x4) -> i32x4
-
fn fsaf_d(f64x2, f64x2) -> i64x2
-
fn fseq_w(f32x4, f32x4) -> i32x4
-
fn fseq_d(f64x2, f64x2) -> i64x2
-
fn fsle_w(f32x4, f32x4) -> i32x4
-
fn fsle_d(f64x2, f64x2) -> i64x2
-
fn fslt_w(f32x4, f32x4) -> i32x4
-
fn fslt_d(f64x2, f64x2) -> i64x2
-
fn fsne_w(f32x4, f32x4) -> i32x4
-
fn fsne_d(f64x2, f64x2) -> i64x2
-
fn fsor_w(f32x4, f32x4) -> i32x4
-
fn fsor_d(f64x2, f64x2) -> i64x2
-
fn fsqrt_w(f32x4) -> f32x4
-
fn fsqrt_d(f64x2) -> f64x2
-
fn fsub_w(f32x4, f32x4) -> f32x4
-
fn fsub_d(f64x2, f64x2) -> f64x2
-
fn fsueq_w(f32x4, f32x4) -> i32x4
-
fn fsueq_d(f64x2, f64x2) -> i64x2
-
fn fsule_w(f32x4, f32x4) -> i32x4
-
fn fsule_d(f64x2, f64x2) -> i64x2
-
fn fsult_w(f32x4, f32x4) -> i32x4
-
fn fsult_d(f64x2, f64x2) -> i64x2
-
fn fsun_w(f32x4, f32x4) -> i32x4
-
fn fsun_d(f64x2, f64x2) -> i64x2
-
fn fsune_w(f32x4, f32x4) -> i32x4
-
fn fsune_d(f64x2, f64x2) -> i64x2
-
fn ftint_s_w(f32x4) -> i32x4
-
fn ftint_s_d(f64x2) -> i64x2
-
fn ftint_u_w(f32x4) -> u32x4
-
fn ftint_u_d(f64x2) -> u64x2
-
fn ftq_h(f32x4, f32x4) -> i16x8
-
fn ftq_w(f64x2, f64x2) -> i32x4
-
fn ftrunc_s_w(f32x4) -> i32x4
-
fn ftrunc_s_d(f64x2) -> i64x2
-
fn ftrunc_u_w(f32x4) -> u32x4
-
fn ftrunc_u_d(f64x2) -> u64x2
-
fn hadd_s_h(i8x16, i8x16) -> i16x8
-
fn hadd_s_w(i16x8, i16x8) -> i32x4
-
fn hadd_s_d(i32x4, i32x4) -> i64x2
-
fn hadd_u_h(u8x16, u8x16) -> u16x8
-
fn hadd_u_w(u16x8, u16x8) -> u32x4
-
fn hadd_u_d(u32x4, u32x4) -> u64x2
-
fn hsub_s_h(i8x16, i8x16) -> i16x8
-
fn hsub_s_w(i16x8, i16x8) -> i32x4
-
fn hsub_s_d(i32x4, i32x4) -> i64x2
-
fn hsub_u_h(u8x16, u8x16) -> i16x8
-
fn hsub_u_w(u16x8, u16x8) -> i32x4
-
fn hsub_u_d(u32x4, u32x4) -> i64x2
-
fn ilvev_b(i8x16, i8x16) -> i8x16
-
fn ilvev_h(i16x8, i16x8) -> i16x8
-
fn ilvev_w(i32x4, i32x4) -> i32x4
-
fn ilvev_d(i64x2, i64x2) -> i64x2
-
fn ilvl_b(i8x16, i8x16) -> i8x16
-
fn ilvl_h(i16x8, i16x8) -> i16x8
-
fn ilvl_w(i32x4, i32x4) -> i32x4
-
fn ilvl_d(i64x2, i64x2) -> i64x2
-
fn ilvod_b(i8x16, i8x16) -> i8x16
-
fn ilvod_h(i16x8, i16x8) -> i16x8
-
fn ilvod_w(i32x4, i32x4) -> i32x4
-
fn ilvod_d(i64x2, i64x2) -> i64x2
-
fn ilvr_b(i8x16, i8x16) -> i8x16
-
fn ilvr_h(i16x8, i16x8) -> i16x8
-
fn ilvr_w(i32x4, i32x4) -> i32x4
-
fn ilvr_d(i64x2, i64x2) -> i64x2
-
fn insert_b(i8x16, imm0_15, i32) -> i8x16
-
fn insert_h(i16x8, imm0_7, i32) -> i16x8
-
fn insert_w(i32x4, imm0_3, i32) -> i32x4
-
fn insert_d(i64x2, imm0_1, i64) -> i64x2
-
fn insve_b(i8x16, imm0_15, i8x16) -> i8x16
-
fn insve_h(i16x8, imm0_7, i16x8) -> i16x8
-
fn insve_w(i32x4, imm0_3, i32x4) -> i32x4
-
fn insve_d(i64x2, imm0_1, i64x2) -> i64x2
-
fn ld_b(*mut c_void, imm_n512_511) -> i8x16
-
fn ld_h(*mut c_void, imm_n1024_1022) -> i16x8
-
fn ld_w(*mut c_void, imm_n2048_2044) -> i32x4
-
fn ld_d(*mut c_void, imm_n4096_4088) -> i64x2
-
fn ldi_b(imm_n512_511) -> i8x16
-
fn ldi_h(imm_n512_511) -> i16x8
-
fn ldi_w(imm_n512_511) -> i32x4
-
fn ldi_d(imm_n512_511) -> i64x2
-
fn madd_q_h(i16x8, i16x8, i16x8) -> i16x8
-
fn madd_q_w(i32x4, i32x4, i32x4) -> i32x4
-
fn maddr_q_h(i16x8, i16x8, i16x8) -> i16x8
-
fn maddr_q_w(i32x4, i32x4, i32x4) -> i32x4
-
fn maddv_b(i8x16, i8x16, i8x16) -> i8x16
-
fn maddv_h(i16x8, i16x8, i16x8) -> i16x8
-
fn maddv_w(i32x4, i32x4, i32x4) -> i32x4
-
fn maddv_d(i64x2, i64x2, i64x2) -> i64x2
-
fn max_a_b(i8x16, i8x16) -> i8x16
-
fn max_a_h(i16x8, i16x8) -> i16x8
-
fn max_a_w(i32x4, i32x4) -> i32x4
-
fn max_a_d(i64x2, i64x2) -> i64x2
-
fn max_s_b(i8x16, i8x16) -> i8x16
-
fn max_s_h(i16x8, i16x8) -> i16x8
-
fn max_s_w(i32x4, i32x4) -> i32x4
-
fn max_s_d(i64x2, i64x2) -> i64x2
-
fn max_u_b(u8x16, u8x16) -> u8x16
-
fn max_u_h(u16x8, u16x8) -> u16x8
-
fn max_u_w(u32x4, u32x4) -> u32x4
-
fn max_u_d(u64x2, u64x2) -> u64x2
-
fn maxi_s_b(i8x16, imm_n16_15) -> i8x16
-
fn maxi_s_h(i16x8, imm_n16_15) -> i16x8
-
fn maxi_s_w(i32x4, imm_n16_15) -> i32x4
-
fn maxi_s_d(i64x2, imm_n16_15) -> i64x2
-
fn maxi_u_b(u8x16, imm0_31) -> u8x16
-
fn maxi_u_h(u16x8, imm0_31) -> u16x8
-
fn maxi_u_w(u32x4, imm0_31) -> u32x4
-
fn maxi_u_d(u64x2, imm0_31) -> u64x2
-
fn min_a_b(i8x16, i8x16) -> i8x16
-
fn min_a_h(i16x8, i16x8) -> i16x8
-
fn min_a_w(i32x4, i32x4) -> i32x4
-
fn min_a_d(i64x2, i64x2) -> i64x2
-
fn min_s_b(i8x16, i8x16) -> i8x16
-
fn min_s_h(i16x8, i16x8) -> i16x8
-
fn min_s_w(i32x4, i32x4) -> i32x4
-
fn min_s_d(i64x2, i64x2) -> i64x2
-
fn min_u_b(u8x16, u8x16) -> u8x16
-
fn min_u_h(u16x8, u16x8) -> u16x8
-
fn min_u_w(u32x4, u32x4) -> u32x4
-
fn min_u_d(u64x2, u64x2) -> u64x2
-
fn mini_s_b(i8x16, imm_n16_15) -> i8x16
-
fn mini_s_h(i16x8, imm_n16_15) -> i16x8
-
fn mini_s_w(i32x4, imm_n16_15) -> i32x4
-
fn mini_s_d(i64x2, imm_n16_15) -> i64x2
-
fn mini_u_b(u8x16, imm0_31) -> u8x16
-
fn mini_u_h(u16x8, imm0_31) -> u16x8
-
fn mini_u_w(u32x4, imm0_31) -> u32x4
-
fn mini_u_d(u64x2, imm0_31) -> u64x2
-
fn mod_s_b(i8x16, i8x16) -> i8x16
-
fn mod_s_h(i16x8, i16x8) -> i16x8
-
fn mod_s_w(i32x4, i32x4) -> i32x4
-
fn mod_s_d(i64x2, i64x2) -> i64x2
-
fn mod_u_b(u8x16, u8x16) -> u8x16
-
fn mod_u_h(u16x8, u16x8) -> u16x8
-
fn mod_u_w(u32x4, u32x4) -> u32x4
-
fn mod_u_d(u64x2, u64x2) -> u64x2
-
fn move_v(i8x16) -> i8x16
-
fn msub_q_h(i16x8, i16x8, i16x8) -> i16x8
-
fn msub_q_w(i32x4, i32x4, i32x4) -> i32x4
-
fn msubr_q_h(i16x8, i16x8, i16x8) -> i16x8
-
fn msubr_q_w(i32x4, i32x4, i32x4) -> i32x4
-
fn msubv_b(i8x16, i8x16, i8x16) -> i8x16
-
fn msubv_h(i16x8, i16x8, i16x8) -> i16x8
-
fn msubv_w(i32x4, i32x4, i32x4) -> i32x4
-
fn msubv_d(i64x2, i64x2, i64x2) -> i64x2
-
fn mul_q_h(i16x8, i16x8) -> i16x8
-
fn mul_q_w(i32x4, i32x4) -> i32x4
-
fn mulr_q_h(i16x8, i16x8) -> i16x8
-
fn mulr_q_w(i32x4, i32x4) -> i32x4
-
fn mulv_b(i8x16, i8x16) -> i8x16
-
fn mulv_h(i16x8, i16x8) -> i16x8
-
fn mulv_w(i32x4, i32x4) -> i32x4
-
fn mulv_d(i64x2, i64x2) -> i64x2
-
fn nloc_b(i8x16) -> i8x16
-
fn nloc_h(i16x8) -> i16x8
-
fn nloc_w(i32x4) -> i32x4
-
fn nloc_d(i64x2) -> i64x2
-
fn nlzc_b(i8x16) -> i8x16
-
fn nlzc_h(i16x8) -> i16x8
-
fn nlzc_w(i32x4) -> i32x4
-
fn nlzc_d(i64x2) -> i64x2
-
fn nor_v(u8x16, u8x16) -> u8x16
-
fn nori_b(u8x16, imm0_255) -> u8x16
-
fn or_v(u8x16, u8x16) -> u8x16
-
fn ori_b(u8x16, imm0_255) -> u8x16
-
fn pckev_b(i8x16, i8x16) -> i8x16
-
fn pckev_h(i16x8, i16x8) -> i16x8
-
fn pckev_w(i32x4, i32x4) -> i32x4
-
fn pckev_d(i64x2, i64x2) -> i64x2
-
fn pckod_b(i8x16, i8x16) -> i8x16
-
fn pckod_h(i16x8, i16x8) -> i16x8
-
fn pckod_w(i32x4, i32x4) -> i32x4
-
fn pckod_d(i64x2, i64x2) -> i64x2
-
fn pcnt_b(i8x16) -> i8x16
-
fn pcnt_h(i16x8) -> i16x8
-
fn pcnt_w(i32x4) -> i32x4
-
fn pcnt_d(i64x2) -> i64x2
-
fn sat_s_b(i8x16, imm0_7) -> i8x16
-
fn sat_s_h(i16x8, imm0_15) -> i16x8
-
fn sat_s_w(i32x4, imm0_31) -> i32x4
-
fn sat_s_d(i64x2, imm0_63) -> i64x2
-
fn sat_u_b(u8x16, imm0_7) -> u8x16
-
fn sat_u_h(u16x8, imm0_15) -> u16x8
-
fn sat_u_w(u32x4, imm0_31) -> u32x4
-
fn sat_u_d(u64x2, imm0_63) -> u64x2
-
fn shf_b(i8x16, imm0_255) -> i8x16
-
fn shf_h(i16x8, imm0_255) -> i16x8
-
fn shf_w(i32x4, imm0_255) -> i32x4
-
fn sld_b(i8x16, i8x16, i32) -> i8x16
-
fn sld_h(i16x8, i16x8, i32) -> i16x8
-
fn sld_w(i32x4, i32x4, i32) -> i32x4
-
fn sld_d(i64x2, i64x2, i32) -> i64x2
-
fn sldi_b(i8x16, i8x16, imm0_15) -> i8x16
-
fn sldi_h(i16x8, i16x8, imm0_7) -> i16x8
-
fn sldi_w(i32x4, i32x4, imm0_3) -> i32x4
-
fn sldi_d(i64x2, i64x2, imm0_1) -> i64x2
-
fn sll_b(i8x16, i8x16) -> i8x16
-
fn sll_h(i16x8, i16x8) -> i16x8
-
fn sll_w(i32x4, i32x4) -> i32x4
-
fn sll_d(i64x2, i64x2) -> i64x2
-
fn slli_b(i8x16, imm0_7) -> i8x16
-
fn slli_h(i16x8, imm0_15) -> i16x8
-
fn slli_w(i32x4, imm0_31) -> i32x4
-
fn slli_d(i64x2, imm0_63) -> i64x2
-
fn splat_b(i8x16, i32) -> i8x16
-
fn splat_h(i16x8, i32) -> i16x8
-
fn splat_w(i32x4, i32) -> i32x4
-
fn splat_d(i64x2, i32) -> i64x2
-
fn splati_b(i8x16, imm0_15) -> i8x16
-
fn splati_h(i16x8, imm0_7) -> i16x8
-
fn splati_w(i32x4, imm0_3) -> i32x4
-
fn splati_d(i64x2, imm0_1) -> i64x2
-
fn sra_b(i8x16, i8x16) -> i8x16
-
fn sra_h(i16x8, i16x8) -> i16x8
-
fn sra_w(i32x4, i32x4) -> i32x4
-
fn sra_d(i64x2, i64x2) -> i64x2
-
fn srai_b(i8x16, imm0_7) -> i8x16
-
fn srai_h(i16x8, imm0_15) -> i16x8
-
fn srai_w(i32x4, imm0_31) -> i32x4
-
fn srai_d(i64x2, imm0_63) -> i64x2
-
fn srar_b(i8x16, i8x16) -> i8x16
-
fn srar_h(i16x8, i16x8) -> i16x8
-
fn srar_w(i32x4, i32x4) -> i32x4
-
fn srar_d(i64x2, i64x2) -> i64x2
-
fn srari_b(i8x16, imm0_7) -> i8x16
-
fn srari_h(i16x8, imm0_15) -> i16x8
-
fn srari_w(i32x4, imm0_31) -> i32x4
-
fn srari_d(i64x2, imm0_63) -> i64x2
-
fn srl_b(i8x16, i8x16) -> i8x16
-
fn srl_h(i16x8, i16x8) -> i16x8
-
fn srl_w(i32x4, i32x4) -> i32x4
-
fn srl_d(i64x2, i64x2) -> i64x2
-
fn srli_b(i8x16, imm0_7) -> i8x16
-
fn srli_h(i16x8, imm0_15) -> i16x8
-
fn srli_w(i32x4, imm0_31) -> i32x4
-
fn srli_d(i64x2, imm0_63) -> i64x2
-
fn srlr_b(i8x16, i8x16) -> i8x16
-
fn srlr_h(i16x8, i16x8) -> i16x8
-
fn srlr_w(i32x4, i32x4) -> i32x4
-
fn srlr_d(i64x2, i64x2) -> i64x2
-
fn srlri_b(i8x16, imm0_7) -> i8x16
-
fn srlri_h(i16x8, imm0_15) -> i16x8
-
fn srlri_w(i32x4, imm0_31) -> i32x4
-
fn srlri_d(i64x2, imm0_63) -> i64x2
-
fn st_b(i8x16, *mut c_void, imm_n512_511) -> ()
-
fn st_h(i16x8, *mut c_void, imm_n1024_1022) -> ()
-
fn st_w(i32x4, *mut c_void, imm_n2048_2044) -> ()
-
fn st_d(i64x2, *mut c_void, imm_n4096_4088) -> ()
-
fn subs_s_b(i8x16, i8x16) -> i8x16
-
fn subs_s_h(i16x8, i16x8) -> i16x8
-
fn subs_s_w(i32x4, i32x4) -> i32x4
-
fn subs_s_d(i64x2, i64x2) -> i64x2
-
fn subs_u_b(u8x16, u8x16) -> u8x16
-
fn subs_u_h(u16x8, u16x8) -> u16x8
-
fn subs_u_w(u32x4, u32x4) -> u32x4
-
fn subs_u_d(u64x2, u64x2) -> u64x2
-
fn subsus_u_b(u8x16, i8x16) -> u8x16
-
fn subsus_u_h(u16x8, i16x8) -> u16x8
-
fn subsus_u_w(u32x4, i32x4) -> u32x4
-
fn subsus_u_d(u64x2, i64x2) -> u64x2
-
fn subsuu_s_b(u8x16, u8x16) -> i8x16
-
fn subsuu_s_h(u16x8, u16x8) -> i16x8
-
fn subsuu_s_w(u32x4, u32x4) -> i32x4
-
fn subsuu_s_d(u64x2, u64x2) -> i64x2
-
fn subv_b(i8x16, i8x16) -> i8x16
-
fn subv_h(i16x8, i16x8) -> i16x8
-
fn subv_w(i32x4, i32x4) -> i32x4
-
fn subv_d(i64x2, i64x2) -> i64x2
-
fn subvi_b(i8x16, imm0_31) -> i8x16
-
fn subvi_h(i16x8, imm0_31) -> i16x8
-
fn subvi_w(i32x4, imm0_31) -> i32x4
-
fn subvi_d(i64x2, imm0_31) -> i64x2
-
fn vshf_b(i8x16, i8x16, i8x16) -> i8x16
-
fn vshf_h(i16x8, i16x8, i16x8) -> i16x8
-
fn vshf_w(i32x4, i32x4, i32x4) -> i32x4
-
fn vshf_d(i64x2, i64x2, i64x2) -> i64x2
-
fn xor_v(u8x16, u8x16) -> u8x16
-
fn xori_b(u8x16, imm0_255) -> u8x16