[PATCH v6 09/12] app/mldev: enable support for inference batches

Thomas Monjalon thomas at monjalon.net
Thu Mar 16 18:47:33 CET 2023


11/03/2023 16:09, Srikanth Yalavarthi:
> @@ -528,8 +533,8 @@ ml_request_initialize(struct rte_mempool *mp, void *opaque, void *obj, unsigned
>  	req->niters = 0;
>  
>  	/* quantize data */
> -	rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id,
> -			   t->model[t->fid].info.batch_size, t->model[t->fid].input, req->input);
> +	rte_ml_io_quantize(t->cmn.opt->dev_id, t->model[t->fid].id, t->model[t->fid].nb_batches,
> +			   t->model[t->fid].input, req->input);
>  }
>  
>  int
> @@ -547,7 +552,7 @@ ml_inference_iomem_setup(struct ml_test *test, struct ml_options *opt, uint16_t
>  	int ret;
>  
>  	/* get input buffer size */
> -	ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, t->model[fid].info.batch_size,
> +	ret = rte_ml_io_input_size_get(opt->dev_id, t->model[fid].id, t->model[fid].nb_batches,
>  				       &t->model[fid].inp_qsize, &t->model[fid].inp_dsize);
>  	if (ret != 0) {
>  		ml_err("Failed to get input size, model : %s\n", opt->filelist[fid].model);
> @@ -555,9 +560,8 @@ ml_inference_iomem_setup(struct ml_test *test, struct ml_options *opt, uint16_t
>  	}
>  
>  	/* get output buffer size */
> -	ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id,
> -					t->model[fid].info.batch_size, &t->model[fid].out_qsize,
> -					&t->model[fid].out_dsize);
> +	ret = rte_ml_io_output_size_get(opt->dev_id, t->model[fid].id, t->model[fid].nb_batches,
> +					&t->model[fid].out_qsize, &t->model[fid].out_dsize);
>  	if (ret != 0) {
>  		ml_err("Failed to get input size, model : %s\n", opt->filelist[fid].model);
>  		return ret;
> @@ -702,7 +706,7 @@ ml_request_finish(struct rte_mempool *mp, void *opaque, void *obj, unsigned int
>  		return;
>  
>  	t->nb_used++;
> -	rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, t->model[req->fid].info.batch_size,
> +	rte_ml_io_dequantize(t->cmn.opt->dev_id, model->id, t->model[req->fid].nb_batches,
>  			     req->output, model->output);

These changes look unrelated with the topic of the patch.
You should probably fix it when adding those lines at first.




More information about the dev mailing list