summaryrefslogtreecommitdiff
path: root/app/test-mldev/test_inference_interleave.c
blob: 9503dda042be9e913da25e59569850bf90f1ca02 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
/* SPDX-License-Identifier: BSD-3-Clause
 * Copyright (c) 2022 Marvell.
 */

#include <rte_common.h>
#include <rte_launch.h>

#include "ml_common.h"
#include "test_inference_common.h"
#include "test_stats.h"

static int
test_inference_interleave_driver(struct ml_test *test, struct ml_options *opt)
{
	struct test_inference *t;
	uint16_t fid = 0;
	int ret = 0;

	t = ml_test_priv(test);

	ret = ml_inference_mldev_setup(test, opt);
	if (ret != 0)
		return ret;

	ret = ml_inference_mem_setup(test, opt);
	if (ret != 0)
		return ret;

	/* load and start all models */
	for (fid = 0; fid < opt->nb_filelist; fid++) {
		ret = ml_model_load(test, opt, &t->model[fid], fid);
		if (ret != 0)
			goto error;

		ret = ml_model_start(test, opt, &t->model[fid], fid);
		if (ret != 0)
			goto error;

		ret = ml_inference_iomem_setup(test, opt, fid);
		if (ret != 0)
			goto error;
	}

	/* launch inference requests */
	ret = ml_inference_launch_cores(test, opt, 0, opt->nb_filelist - 1);
	if (ret != 0) {
		ml_err("failed to launch cores");
		goto error;
	}

	rte_eal_mp_wait_lcore();

	/* stop and unload all models */
	for (fid = 0; fid < opt->nb_filelist; fid++) {
		ret = ml_inference_result(test, opt, fid);
		if (ret != ML_TEST_SUCCESS)
			goto error;

		ml_inference_iomem_destroy(test, opt, fid);
	}

	for (fid = 0; fid < opt->nb_filelist; fid++)
		ml_stats_get(test, opt, RTE_ML_DEV_XSTATS_MODEL, fid);

	for (fid = 0; fid < opt->nb_filelist; fid++) {
		ret = ml_model_stop(test, opt, &t->model[fid], fid);
		if (ret != 0)
			goto error;

		ret = ml_model_unload(test, opt, &t->model[fid], fid);
		if (ret != 0)
			goto error;
	}

	ml_stats_get(test, opt, RTE_ML_DEV_XSTATS_DEVICE, -1);
	ml_throughput_get(test, opt);
	ml_inference_mem_destroy(test, opt);

	ret = ml_inference_mldev_destroy(test, opt);
	if (ret != 0)
		return ret;

	t->cmn.result = ML_TEST_SUCCESS;

	return 0;

error:
	ml_inference_mem_destroy(test, opt);
	for (fid = 0; fid < opt->nb_filelist; fid++) {
		ml_inference_iomem_destroy(test, opt, fid);
		ml_model_stop(test, opt, &t->model[fid], fid);
		ml_model_unload(test, opt, &t->model[fid], fid);
	}

	t->cmn.result = ML_TEST_FAILED;

	return ret;
}

static int
test_inference_interleave_result(struct ml_test *test, struct ml_options *opt)
{
	struct test_inference *t;

	RTE_SET_USED(opt);

	t = ml_test_priv(test);

	return t->cmn.result;
}

static const struct ml_test_ops inference_interleave = {
	.cap_check = test_inference_cap_check,
	.opt_check = test_inference_opt_check,
	.opt_dump = test_inference_opt_dump,
	.test_setup = test_inference_setup,
	.test_destroy = test_inference_destroy,
	.test_driver = test_inference_interleave_driver,
	.test_result = test_inference_interleave_result,
};

ML_TEST_REGISTER(inference_interleave);