|
@@ -855,6 +855,7 @@ void tcg_optimize(TCGContext *s)
|
|
if ((arg_info(op->args[1])->mask & 0x80) != 0) {
|
|
if ((arg_info(op->args[1])->mask & 0x80) != 0) {
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
|
|
+ QEMU_FALLTHROUGH;
|
|
CASE_OP_32_64(ext8u):
|
|
CASE_OP_32_64(ext8u):
|
|
mask = 0xff;
|
|
mask = 0xff;
|
|
goto and_const;
|
|
goto and_const;
|
|
@@ -862,6 +863,7 @@ void tcg_optimize(TCGContext *s)
|
|
if ((arg_info(op->args[1])->mask & 0x8000) != 0) {
|
|
if ((arg_info(op->args[1])->mask & 0x8000) != 0) {
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
|
|
+ QEMU_FALLTHROUGH;
|
|
CASE_OP_32_64(ext16u):
|
|
CASE_OP_32_64(ext16u):
|
|
mask = 0xffff;
|
|
mask = 0xffff;
|
|
goto and_const;
|
|
goto and_const;
|
|
@@ -869,6 +871,7 @@ void tcg_optimize(TCGContext *s)
|
|
if ((arg_info(op->args[1])->mask & 0x80000000) != 0) {
|
|
if ((arg_info(op->args[1])->mask & 0x80000000) != 0) {
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
|
|
+ QEMU_FALLTHROUGH;
|
|
case INDEX_op_ext32u_i64:
|
|
case INDEX_op_ext32u_i64:
|
|
mask = 0xffffffffU;
|
|
mask = 0xffffffffU;
|
|
goto and_const;
|
|
goto and_const;
|
|
@@ -886,6 +889,7 @@ void tcg_optimize(TCGContext *s)
|
|
if ((arg_info(op->args[1])->mask & 0x80000000) != 0) {
|
|
if ((arg_info(op->args[1])->mask & 0x80000000) != 0) {
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
|
|
+ QEMU_FALLTHROUGH;
|
|
case INDEX_op_extu_i32_i64:
|
|
case INDEX_op_extu_i32_i64:
|
|
/* We do not compute affected as it is a size changing op. */
|
|
/* We do not compute affected as it is a size changing op. */
|
|
mask = (uint32_t)arg_info(op->args[1])->mask;
|
|
mask = (uint32_t)arg_info(op->args[1])->mask;
|