aclIntArray *padding = aclCreateIntArray(paddingVal, 1);
int64_t dilationVal[] = {1};
aclIntArray *dilation = aclCreateIntArray(dilationVal, 1);
- bool transposed = true;
- int64_t groups = 1;
int8_t cubeMathType = 0;
#ifdef ASCEND_310P
#endif
GGML_CANN_CALL_ACLNN_OP(ctx, Convolution, acl_input, acl_weight, nullptr, stride,
- padding, dilation, transposed, padding, groups, acl_dst, cubeMathType);
+ padding, dilation, true, padding, 1, acl_dst, cubeMathType);
ggml_cann_release_resources(ctx, acl_weight, acl_dst, stride, padding, dilation);
}
case GGML_OP_ARGMAX:
case GGML_OP_COS:
case GGML_OP_SIN:
- case GGML_OP_CONV_TRANSPOSE_1D:
case GGML_OP_LOG:
case GGML_OP_MEAN:
case GGML_OP_PAD_REFLECT_1D:
case GGML_OP_COUNT_EQUAL:
return true;
+ case GGML_OP_CONV_TRANSPOSE_1D:
+ // TODO: ((weightL - 1) * dilationW - padLeft)=1336 should not be larger than 255.
+ return (op->src[0]->ne[0] - 1) <= 255;
case GGML_OP_SCALE:
float bias;
memcpy(&bias, (const float *)(op->op_params) + 1, sizeof(float));