|
@@ -260,7 +260,7 @@ vext_ldst_stride(void *vd, void *v0, target_ulong base, target_ulong stride,
|
|
uint32_t esz = 1 << log2_esz;
|
|
uint32_t esz = 1 << log2_esz;
|
|
uint32_t vma = vext_vma(desc);
|
|
uint32_t vma = vext_vma(desc);
|
|
|
|
|
|
- VSTART_CHECK_EARLY_EXIT(env);
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, env->vl);
|
|
|
|
|
|
for (i = env->vstart; i < env->vl; env->vstart = ++i) {
|
|
for (i = env->vstart; i < env->vl; env->vstart = ++i) {
|
|
k = 0;
|
|
k = 0;
|
|
@@ -383,10 +383,7 @@ vext_ldst_us(void *vd, target_ulong base, CPURISCVState *env, uint32_t desc,
|
|
uint32_t msize = nf * esz;
|
|
uint32_t msize = nf * esz;
|
|
int mmu_index = riscv_env_mmu_index(env, false);
|
|
int mmu_index = riscv_env_mmu_index(env, false);
|
|
|
|
|
|
- if (env->vstart >= evl) {
|
|
|
|
- env->vstart = 0;
|
|
|
|
- return;
|
|
|
|
- }
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, evl);
|
|
|
|
|
|
#if defined(CONFIG_USER_ONLY)
|
|
#if defined(CONFIG_USER_ONLY)
|
|
/*
|
|
/*
|
|
@@ -544,7 +541,7 @@ vext_ldst_index(void *vd, void *v0, target_ulong base,
|
|
uint32_t esz = 1 << log2_esz;
|
|
uint32_t esz = 1 << log2_esz;
|
|
uint32_t vma = vext_vma(desc);
|
|
uint32_t vma = vext_vma(desc);
|
|
|
|
|
|
- VSTART_CHECK_EARLY_EXIT(env);
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, env->vl);
|
|
|
|
|
|
/* load bytes from guest memory */
|
|
/* load bytes from guest memory */
|
|
for (i = env->vstart; i < env->vl; env->vstart = ++i) {
|
|
for (i = env->vstart; i < env->vl; env->vstart = ++i) {
|
|
@@ -638,7 +635,7 @@ vext_ldff(void *vd, void *v0, target_ulong base, CPURISCVState *env,
|
|
int flags;
|
|
int flags;
|
|
void *host;
|
|
void *host;
|
|
|
|
|
|
- VSTART_CHECK_EARLY_EXIT(env);
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, env->vl);
|
|
|
|
|
|
addr = base + ((env->vstart * nf) << log2_esz);
|
|
addr = base + ((env->vstart * nf) << log2_esz);
|
|
page_split = -(addr | TARGET_PAGE_MASK);
|
|
page_split = -(addr | TARGET_PAGE_MASK);
|
|
@@ -1116,7 +1113,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
@@ -1150,7 +1147,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -1187,7 +1184,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint32_t vta_all_1s = vext_vta_all_1s(desc); \
|
|
uint32_t vta_all_1s = vext_vta_all_1s(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
@@ -1227,7 +1224,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
|
|
uint32_t vta_all_1s = vext_vta_all_1s(desc); \
|
|
uint32_t vta_all_1s = vext_vta_all_1s(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -1325,7 +1322,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -1374,7 +1371,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -1438,7 +1435,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
@@ -1505,7 +1502,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -2054,7 +2051,7 @@ void HELPER(NAME)(void *vd, void *vs1, CPURISCVState *env, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
@@ -2080,7 +2077,7 @@ void HELPER(NAME)(void *vd, uint64_t s1, CPURISCVState *env, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
*((ETYPE *)vd + H(i)) = (ETYPE)s1; \
|
|
*((ETYPE *)vd + H(i)) = (ETYPE)s1; \
|
|
@@ -2105,7 +2102,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE *vt = (!vext_elem_mask(v0, i) ? vs2 : vs1); \
|
|
ETYPE *vt = (!vext_elem_mask(v0, i) ? vs2 : vs1); \
|
|
@@ -2131,7 +2128,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -2178,7 +2175,7 @@ vext_vv_rm_1(void *vd, void *v0, void *vs1, void *vs2,
|
|
uint32_t vl, uint32_t vm, int vxrm,
|
|
uint32_t vl, uint32_t vm, int vxrm,
|
|
opivv2_rm_fn *fn, uint32_t vma, uint32_t esz)
|
|
opivv2_rm_fn *fn, uint32_t vma, uint32_t esz)
|
|
{
|
|
{
|
|
- VSTART_CHECK_EARLY_EXIT(env);
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl);
|
|
|
|
|
|
for (uint32_t i = env->vstart; i < vl; i++) {
|
|
for (uint32_t i = env->vstart; i < vl; i++) {
|
|
if (!vm && !vext_elem_mask(v0, i)) {
|
|
if (!vm && !vext_elem_mask(v0, i)) {
|
|
@@ -2305,7 +2302,7 @@ vext_vx_rm_1(void *vd, void *v0, target_long s1, void *vs2,
|
|
uint32_t vl, uint32_t vm, int vxrm,
|
|
uint32_t vl, uint32_t vm, int vxrm,
|
|
opivx2_rm_fn *fn, uint32_t vma, uint32_t esz)
|
|
opivx2_rm_fn *fn, uint32_t vma, uint32_t esz)
|
|
{
|
|
{
|
|
- VSTART_CHECK_EARLY_EXIT(env);
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl);
|
|
|
|
|
|
for (uint32_t i = env->vstart; i < vl; i++) {
|
|
for (uint32_t i = env->vstart; i < vl; i++) {
|
|
if (!vm && !vext_elem_mask(v0, i)) {
|
|
if (!vm && !vext_elem_mask(v0, i)) {
|
|
@@ -3104,7 +3101,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -3149,7 +3146,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -3737,7 +3734,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
if (vl == 0) { \
|
|
if (vl == 0) { \
|
|
return; \
|
|
return; \
|
|
@@ -4260,7 +4257,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
ETYPE s1 = *((ETYPE *)vs1 + H(i)); \
|
|
@@ -4302,7 +4299,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -4497,7 +4494,7 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2, \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t vta = vext_vta(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
ETYPE s2 = *((ETYPE *)vs2 + H(i)); \
|
|
@@ -4827,7 +4824,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
int a, b; \
|
|
int a, b; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
a = vext_elem_mask(vs1, i); \
|
|
a = vext_elem_mask(vs1, i); \
|
|
@@ -5022,7 +5019,7 @@ void HELPER(NAME)(void *vd, void *v0, CPURISCVState *env, uint32_t desc) \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
int i; \
|
|
int i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -5059,7 +5056,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
target_ulong offset = s1, i_min, i; \
|
|
target_ulong offset = s1, i_min, i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
i_min = MAX(env->vstart, offset); \
|
|
i_min = MAX(env->vstart, offset); \
|
|
for (i = i_min; i < vl; i++) { \
|
|
for (i = i_min; i < vl; i++) { \
|
|
@@ -5094,7 +5091,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
target_ulong i_max, i_min, i; \
|
|
target_ulong i_max, i_min, i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
i_min = MIN(s1 < vlmax ? vlmax - s1 : 0, vl); \
|
|
i_min = MIN(s1 < vlmax ? vlmax - s1 : 0, vl); \
|
|
i_max = MAX(i_min, env->vstart); \
|
|
i_max = MAX(i_min, env->vstart); \
|
|
@@ -5138,7 +5135,7 @@ static void vslide1up_##BITWIDTH(void *vd, void *v0, uint64_t s1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -5189,7 +5186,7 @@ static void vslide1down_##BITWIDTH(void *vd, void *v0, uint64_t s1, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -5266,7 +5263,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2, \
|
|
uint64_t index; \
|
|
uint64_t index; \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -5311,7 +5308,7 @@ void HELPER(NAME)(void *vd, void *v0, target_ulong s1, void *vs2, \
|
|
uint64_t index = s1; \
|
|
uint64_t index = s1; \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
@@ -5407,7 +5404,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs2, \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t vma = vext_vma(desc); \
|
|
uint32_t i; \
|
|
uint32_t i; \
|
|
\
|
|
\
|
|
- VSTART_CHECK_EARLY_EXIT(env); \
|
|
|
|
|
|
+ VSTART_CHECK_EARLY_EXIT(env, vl); \
|
|
\
|
|
\
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
for (i = env->vstart; i < vl; i++) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|
|
if (!vm && !vext_elem_mask(v0, i)) { \
|