16
16
#if defined(__AMX_INT8__)
17
17
18
18
// AMX buffer interface
19
- static const char * ggml_backend_amx_buffer_get_name (ggml_backend_buffer_t buffer) {
20
- return " AMX" ;
21
-
22
- GGML_UNUSED (buffer);
23
- }
24
-
25
19
static void ggml_backend_amx_buffer_free_buffer (ggml_backend_buffer_t buffer) {
26
20
free (buffer->context );
27
21
}
@@ -72,7 +66,6 @@ static void ggml_backend_amx_buffer_clear(ggml_backend_buffer_t buffer, uint8_t
72
66
}
73
67
74
68
static ggml_backend_buffer_i ggml_backend_amx_buffer_interface = {
75
- /* .get_name = */ ggml_backend_amx_buffer_get_name,
76
69
/* .free_buffer = */ ggml_backend_amx_buffer_free_buffer,
77
70
/* .get_base = */ ggml_backend_amx_buffer_get_base,
78
71
/* .init_tensor = */ NULL , // no initialization required
@@ -121,14 +114,14 @@ static bool ggml_backend_amx_buffer_type_is_host(ggml_backend_buffer_type_t buft
121
114
ggml_backend_buffer_type_t ggml_backend_amx_buffer_type () {
122
115
static struct ggml_backend_buffer_type ggml_backend_buffer_type_amx = {
123
116
/* .iface = */ {
124
- /* .get_name = */ ggml_backend_amx_buffer_type_get_name,
125
- /* .alloc_buffer = */ ggml_backend_amx_buffer_type_alloc_buffer,
126
- /* .get_alignment = */ ggml_backend_amx_buffer_type_get_alignment,
127
- /* .get_max_size = */ NULL , // defaults to SIZE_MAX
128
- /* .get_alloc_size = */ ggml_backend_amx_buffer_type_get_alloc_size,
129
- /* .is_host = */ ggml_backend_amx_buffer_type_is_host,
117
+ /* .get_name = */ ggml_backend_amx_buffer_type_get_name,
118
+ /* .alloc_buffer = */ ggml_backend_amx_buffer_type_alloc_buffer,
119
+ /* .get_alignment = */ ggml_backend_amx_buffer_type_get_alignment,
120
+ /* .get_max_size = */ NULL , // defaults to SIZE_MAX
121
+ /* .get_alloc_size = */ ggml_backend_amx_buffer_type_get_alloc_size,
122
+ /* .is_host = */ ggml_backend_amx_buffer_type_is_host,
130
123
},
131
- /* .device = */ NULL ,
124
+ /* .device = */ ggml_backend_reg_dev_get ( ggml_backend_amx_reg (), 0 ) ,
132
125
/* .context = */ NULL ,
133
126
};
134
127
@@ -149,12 +142,6 @@ static void ggml_backend_amx_free(ggml_backend_t backend) {
149
142
delete backend;
150
143
}
151
144
152
- static ggml_backend_buffer_type_t ggml_backend_amx_get_default_buffer_type (ggml_backend_t backend) {
153
- return ggml_backend_amx_buffer_type ();
154
-
155
- GGML_UNUSED (backend);
156
- }
157
-
158
145
static enum ggml_status ggml_backend_amx_graph_compute (ggml_backend_t backend, struct ggml_cgraph * cgraph) {
159
146
ggml_backend_amx_context * ctx = (ggml_backend_amx_context *)backend->context ;
160
147
@@ -187,7 +174,6 @@ static enum ggml_status ggml_backend_amx_graph_compute(ggml_backend_t backend, s
187
174
static struct ggml_backend_i ggml_backend_amx_i = {
188
175
/* .get_name = */ ggml_backend_amx_name,
189
176
/* .free = */ ggml_backend_amx_free,
190
- /* .get_default_buffer_type = */ ggml_backend_amx_get_default_buffer_type,
191
177
/* .set_tensor_async = */ NULL ,
192
178
/* .get_tensor_async = */ NULL ,
193
179
/* .cpy_tensor_async = */ NULL ,
@@ -197,9 +183,6 @@ static struct ggml_backend_i ggml_backend_amx_i = {
197
183
/* .graph_plan_update = */ NULL ,
198
184
/* .graph_plan_compute = */ NULL ,
199
185
/* .graph_compute = */ ggml_backend_amx_graph_compute,
200
- /* .supports_op = */ NULL ,
201
- /* .supports_buft = */ NULL ,
202
- /* .offload_op = */ NULL ,
203
186
/* .event_record = */ NULL ,
204
187
/* .event_wait = */ NULL ,
205
188
};
@@ -279,7 +262,7 @@ static void ggml_backend_amx_device_get_memory(ggml_backend_dev_t dev, size_t *
279
262
}
280
263
281
264
static enum ggml_backend_dev_type ggml_backend_amx_device_get_type (ggml_backend_dev_t dev) {
282
- return GGML_BACKEND_DEVICE_TYPE_CPU ;
265
+ return GGML_BACKEND_DEVICE_TYPE_ACCEL ;
283
266
284
267
GGML_UNUSED (dev);
285
268
}
0 commit comments