8000 Update · pytorch/pytorch@f6a621b · GitHub
[go: up one dir, main page]

Skip to content

Commit f6a621b

Browse files
committed
Update
[ghstack-poisoned]
1 parent 42374fb commit f6a621b

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

aten/src/ATen/native/mkldnn/xpu/detail/QMatmul.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -307,7 +307,7 @@ void quantized_matmul(
307307
args.insert({DNNL_ARG_ATTR_SCALES | DNNL_ARG_SRC, m1_sc_m});
308308
if (m1_need_zp) {
309309
m1_zp_m = dnnl_memory_from_host_scalar(
310-
static_cast<int32_t>(input_zero_point), m1_zp_tensor, engine);
310+
static_cast<int32_t>(input_zero_point), m1_zp_tensor, engine);
311311
args.insert({DNNL_ARG_ATTR_ZERO_POINTS | DNNL_ARG_SRC, m1_zp_m});
312312
}
313313

@@ -318,7 +318,7 @@ void quantized_matmul(
318318
args.insert({DNNL_ARG_ATTR_SCALES | DNNL_ARG_DST, dst_sc_m});
319319
if (dst_need_zp) {
320320
dst_zp_m = dnnl_memory_from_host_scalar(
321-
static_cast<int32_t>(output_zero_point), dst_zp_tensor, engine);
321+
static_cast<int32_t>(output_zero_point), dst_zp_tensor, engine);
322322
args.insert({DNNL_ARG_ATTR_ZERO_POINTS | DNNL_ARG_DST, dst_zp_m});
323323
}
324324

0 commit comments

Comments
 (0)
0