[PATCH v6 09/12] app/mldev: enable support for inference batches
Thomas Monjalon
thomas at monjalon.net
Thu Mar 16 18:47:33 CET 2023
11/03/2023 16:09, Srikanth Yalavarthi:
> @@ -528,8 +533,8 @@ ml_request_initialize(struct rte_mempool *mp, void *opaque, void *obj, unsigned
> req->niters = 0;
>
> /* quantize data */
> - rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id,
> - t->model[t->fid].info.batch_size, t->model[t->fid].input, req->input);
> + rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id, t->model[t->fid].nb_batches,
> + t->model[t->fid].input, req->input);
> }
>
> int
> @@ -547,7 +552,7 @@ ml_inference_iomem_setup(struct ml_test *test, struct ml_options *opt, uint16_t
> int ret;
>
> /* get input buffer size */
> - ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, t->model[fid].info.batch_size,
> + ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, t->model[fid].nb_batches,
> &t->model[fid].inp_qsize, &t->model[fid].inp_dsize);
> if (ret != 0) {
> ml_err("Failed to get input size, model : %s\n", opt->filelist[fid].model);
> @@ -555,9 +560,8 @@ ml_inference_iomem_setup(struct ml_test *test, struct ml_options *opt, uint16_t
> }
>
> /* get output buffer size */
> - ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id,
> - t->model[fid].info.batch_size, &t->model[fid].out_qsize,
> - &t->model[fid].out_dsize);
> + ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id, t->model[fid].nb_batches,
> + &t->model[fid].out_qsize, &t->model[fid].out_dsize);
> if (ret != 0) {
> ml_err("Failed to get input size, model : %s\n", opt->filelist[fid].model);
> return ret;
> @@ -702,7 +706,7 @@ ml_request_finish(struct rte_mempool *mp, void *opaque, void *obj, unsigned int
> return;
>
> t->nb_used++;
> - rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, t->model[req->fid].info.batch_size,
> + rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, t->model[req->fid].nb_batches,
> req->output, model->output);
These changes look unrelated with the topic of the patch.
You should probably fix it when adding those lines at first.
More information about the dev
mailing list