1414#include "rv64v_settings.h"
1515
1616#ifndef MLK_RVV_WIDENING_MUL
17-
1817static inline vint16m1_t fq_redc (vint16m1_t rh , vint16m1_t rl , size_t vl )
1918{
2019 vint16m1_t t ;
@@ -27,8 +26,7 @@ static inline vint16m1_t fq_redc(vint16m1_t rh, vint16m1_t rl, size_t vl)
2726
2827 return t ;
2928}
30-
31- #endif
29+ #endif /* !MLK_RVV_WIDENING_MUL */
3230
3331/* Narrowing reduction */
3432
@@ -93,9 +91,9 @@ static inline vint16m1_t fq_mul_vv(vint16m1_t rx, vint16m1_t ry, size_t vl)
9391 rh = __riscv_vmulh_vv_i16m1 (rx , ry , vl ); /* h = (x * y) / R */
9492 rl = __riscv_vmul_vv_i16m1 (rx , ry , vl ); /* l = (x * y) % R */
9593 return fq_redc (rh , rl , vl );
96- #else
94+ #else /* !MLK_RVV_WIDENING_MUL */
9795 return fq_redc2 (__riscv_vwmul_vv_i32m2 (rx , ry , vl ), vl );
98- #endif
96+ #endif /* MLK_RVV_WIDENING_MUL */
9997}
10098
10199/* Montgomery multiply: vector-scalar */
@@ -108,9 +106,9 @@ static inline vint16m1_t fq_mul_vx(vint16m1_t rx, int16_t ry, size_t vl)
108106 rh = __riscv_vmulh_vx_i16m1 (rx , ry , vl ); /* h = (x * y) / R */
109107 rl = __riscv_vmul_vx_i16m1 (rx , ry , vl ); /* l = (x * y) % R */
110108 return fq_redc (rh , rl , vl );
111- #else
109+ #else /* !MLK_RVV_WIDENING_MUL */
112110 return fq_redc2 (__riscv_vwmul_vx_i32m2 (rx , ry , vl ), vl );
113- #endif
111+ #endif /* MLK_RVV_WIDENING_MUL */
114112}
115113
116114/* full normalization */
@@ -670,7 +668,7 @@ void mlk_rv64v_poly_basemul_mont_add_k4(int16_t *r, const int16_t *a,
670668 mlk_rv64v_poly_basemul_mont_add_k (r , a , b , 4 * MLKEM_N );
671669}
672670
673- #endif /* ( MLK_RVV_VLEN == 256) */
671+ #endif /* MLK_RVV_VLEN == 256 */
674672
675673/*************************************************
676674 * Name: poly_tomont
0 commit comments