1 | /* |
2 | * This file is part of the MicroPython project, http://micropython.org/ |
3 | * |
4 | * The MIT License (MIT) |
5 | * |
6 | * Copyright (c) 2013, 2014 Damien P. George |
7 | * Copyright (c) 2014 Paul Sokolovsky |
8 | * |
9 | * Permission is hereby granted, free of charge, to any person obtaining a copy |
10 | * of this software and associated documentation files (the "Software"), to deal |
11 | * in the Software without restriction, including without limitation the rights |
12 | * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
13 | * copies of the Software, and to permit persons to whom the Software is |
14 | * furnished to do so, subject to the following conditions: |
15 | * |
16 | * The above copyright notice and this permission notice shall be included in |
17 | * all copies or substantial portions of the Software. |
18 | * |
19 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
20 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
21 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
22 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
23 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
24 | * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN |
25 | * THE SOFTWARE. |
26 | */ |
27 | |
28 | #include <string.h> |
29 | #include <assert.h> |
30 | |
31 | #include "py/objtuple.h" |
32 | #include "py/objfun.h" |
33 | #include "py/runtime.h" |
34 | #include "py/bc.h" |
35 | #include "py/stackctrl.h" |
36 | |
37 | #if MICROPY_DEBUG_VERBOSE // print debugging info |
38 | #define DEBUG_PRINT (1) |
39 | #else // don't print debugging info |
40 | #define DEBUG_PRINT (0) |
41 | #define DEBUG_printf(...) (void)0 |
42 | #endif |
43 | |
44 | // Note: the "name" entry in mp_obj_type_t for a function type must be |
45 | // MP_QSTR_function because it is used to determine if an object is of generic |
46 | // function type. |
47 | |
48 | /******************************************************************************/ |
49 | /* builtin functions */ |
50 | |
51 | STATIC mp_obj_t fun_builtin_0_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
52 | (void)args; |
53 | assert(mp_obj_is_type(self_in, &mp_type_fun_builtin_0)); |
54 | mp_obj_fun_builtin_fixed_t *self = MP_OBJ_TO_PTR(self_in); |
55 | mp_arg_check_num(n_args, n_kw, 0, 0, false); |
56 | return self->fun._0(); |
57 | } |
58 | |
59 | const mp_obj_type_t mp_type_fun_builtin_0 = { |
60 | { &mp_type_type }, |
61 | .flags = MP_TYPE_FLAG_BINDS_SELF | MP_TYPE_FLAG_BUILTIN_FUN, |
62 | .name = MP_QSTR_function, |
63 | .call = fun_builtin_0_call, |
64 | .unary_op = mp_generic_unary_op, |
65 | }; |
66 | |
67 | STATIC mp_obj_t fun_builtin_1_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
68 | assert(mp_obj_is_type(self_in, &mp_type_fun_builtin_1)); |
69 | mp_obj_fun_builtin_fixed_t *self = MP_OBJ_TO_PTR(self_in); |
70 | mp_arg_check_num(n_args, n_kw, 1, 1, false); |
71 | return self->fun._1(args[0]); |
72 | } |
73 | |
74 | const mp_obj_type_t mp_type_fun_builtin_1 = { |
75 | { &mp_type_type }, |
76 | .flags = MP_TYPE_FLAG_BINDS_SELF | MP_TYPE_FLAG_BUILTIN_FUN, |
77 | .name = MP_QSTR_function, |
78 | .call = fun_builtin_1_call, |
79 | .unary_op = mp_generic_unary_op, |
80 | }; |
81 | |
82 | STATIC mp_obj_t fun_builtin_2_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
83 | assert(mp_obj_is_type(self_in, &mp_type_fun_builtin_2)); |
84 | mp_obj_fun_builtin_fixed_t *self = MP_OBJ_TO_PTR(self_in); |
85 | mp_arg_check_num(n_args, n_kw, 2, 2, false); |
86 | return self->fun._2(args[0], args[1]); |
87 | } |
88 | |
89 | const mp_obj_type_t mp_type_fun_builtin_2 = { |
90 | { &mp_type_type }, |
91 | .flags = MP_TYPE_FLAG_BINDS_SELF | MP_TYPE_FLAG_BUILTIN_FUN, |
92 | .name = MP_QSTR_function, |
93 | .call = fun_builtin_2_call, |
94 | .unary_op = mp_generic_unary_op, |
95 | }; |
96 | |
97 | STATIC mp_obj_t fun_builtin_3_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
98 | assert(mp_obj_is_type(self_in, &mp_type_fun_builtin_3)); |
99 | mp_obj_fun_builtin_fixed_t *self = MP_OBJ_TO_PTR(self_in); |
100 | mp_arg_check_num(n_args, n_kw, 3, 3, false); |
101 | return self->fun._3(args[0], args[1], args[2]); |
102 | } |
103 | |
104 | const mp_obj_type_t mp_type_fun_builtin_3 = { |
105 | { &mp_type_type }, |
106 | .flags = MP_TYPE_FLAG_BINDS_SELF | MP_TYPE_FLAG_BUILTIN_FUN, |
107 | .name = MP_QSTR_function, |
108 | .call = fun_builtin_3_call, |
109 | .unary_op = mp_generic_unary_op, |
110 | }; |
111 | |
112 | STATIC mp_obj_t fun_builtin_var_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
113 | assert(mp_obj_is_type(self_in, &mp_type_fun_builtin_var)); |
114 | mp_obj_fun_builtin_var_t *self = MP_OBJ_TO_PTR(self_in); |
115 | |
116 | // check number of arguments |
117 | mp_arg_check_num_sig(n_args, n_kw, self->sig); |
118 | |
119 | if (self->sig & 1) { |
120 | // function allows keywords |
121 | |
122 | // we create a map directly from the given args array |
123 | mp_map_t kw_args; |
124 | mp_map_init_fixed_table(&kw_args, n_kw, args + n_args); |
125 | |
126 | return self->fun.kw(n_args, args, &kw_args); |
127 | |
128 | } else { |
129 | // function takes a variable number of arguments, but no keywords |
130 | |
131 | return self->fun.var(n_args, args); |
132 | } |
133 | } |
134 | |
135 | const mp_obj_type_t mp_type_fun_builtin_var = { |
136 | { &mp_type_type }, |
137 | .flags = MP_TYPE_FLAG_BINDS_SELF | MP_TYPE_FLAG_BUILTIN_FUN, |
138 | .name = MP_QSTR_function, |
139 | .call = fun_builtin_var_call, |
140 | .unary_op = mp_generic_unary_op, |
141 | }; |
142 | |
143 | /******************************************************************************/ |
144 | /* byte code functions */ |
145 | |
146 | qstr mp_obj_code_get_name(const byte *code_info) { |
147 | MP_BC_PRELUDE_SIZE_DECODE(code_info); |
148 | #if MICROPY_PERSISTENT_CODE |
149 | return code_info[0] | (code_info[1] << 8); |
150 | #else |
151 | return mp_decode_uint_value(code_info); |
152 | #endif |
153 | } |
154 | |
155 | #if MICROPY_EMIT_NATIVE |
156 | STATIC const mp_obj_type_t mp_type_fun_native; |
157 | #endif |
158 | |
159 | qstr mp_obj_fun_get_name(mp_const_obj_t fun_in) { |
160 | const mp_obj_fun_bc_t *fun = MP_OBJ_TO_PTR(fun_in); |
161 | #if MICROPY_EMIT_NATIVE |
162 | if (fun->base.type == &mp_type_fun_native || fun->base.type == &mp_type_native_gen_wrap) { |
163 | // TODO native functions don't have name stored |
164 | return MP_QSTR_; |
165 | } |
166 | #endif |
167 | |
168 | const byte *bc = fun->bytecode; |
169 | MP_BC_PRELUDE_SIG_DECODE(bc); |
170 | return mp_obj_code_get_name(bc); |
171 | } |
172 | |
173 | #if MICROPY_CPYTHON_COMPAT |
174 | STATIC void fun_bc_print(const mp_print_t *print, mp_obj_t o_in, mp_print_kind_t kind) { |
175 | (void)kind; |
176 | mp_obj_fun_bc_t *o = MP_OBJ_TO_PTR(o_in); |
177 | mp_printf(print, "<function %q at 0x%p>" , mp_obj_fun_get_name(o_in), o); |
178 | } |
179 | #endif |
180 | |
181 | #if DEBUG_PRINT |
182 | STATIC void dump_args(const mp_obj_t *a, size_t sz) { |
183 | DEBUG_printf("%p: " , a); |
184 | for (size_t i = 0; i < sz; i++) { |
185 | DEBUG_printf("%p " , a[i]); |
186 | } |
187 | DEBUG_printf("\n" ); |
188 | } |
189 | #else |
190 | #define dump_args(...) (void)0 |
191 | #endif |
192 | |
193 | // With this macro you can tune the maximum number of function state bytes |
194 | // that will be allocated on the stack. Any function that needs more |
195 | // than this will try to use the heap, with fallback to stack allocation. |
196 | #define VM_MAX_STATE_ON_STACK (sizeof(mp_uint_t) * 11) |
197 | |
198 | #define DECODE_CODESTATE_SIZE(bytecode, n_state_out_var, state_size_out_var) \ |
199 | { \ |
200 | const uint8_t *ip = bytecode; \ |
201 | size_t n_exc_stack, scope_flags, n_pos_args, n_kwonly_args, n_def_args; \ |
202 | MP_BC_PRELUDE_SIG_DECODE_INTO(ip, n_state_out_var, n_exc_stack, scope_flags, n_pos_args, n_kwonly_args, n_def_args); \ |
203 | \ |
204 | /* state size in bytes */ \ |
205 | state_size_out_var = n_state_out_var * sizeof(mp_obj_t) \ |
206 | + n_exc_stack * sizeof(mp_exc_stack_t); \ |
207 | } |
208 | |
209 | #define INIT_CODESTATE(code_state, _fun_bc, _n_state, n_args, n_kw, args) \ |
210 | code_state->fun_bc = _fun_bc; \ |
211 | code_state->ip = 0; \ |
212 | code_state->n_state = _n_state; \ |
213 | mp_setup_code_state(code_state, n_args, n_kw, args); \ |
214 | code_state->old_globals = mp_globals_get(); |
215 | |
216 | #if MICROPY_STACKLESS |
217 | mp_code_state_t *mp_obj_fun_bc_prepare_codestate(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
218 | MP_STACK_CHECK(); |
219 | mp_obj_fun_bc_t *self = MP_OBJ_TO_PTR(self_in); |
220 | |
221 | size_t n_state, state_size; |
222 | DECODE_CODESTATE_SIZE(self->bytecode, n_state, state_size); |
223 | |
224 | mp_code_state_t *code_state; |
225 | #if MICROPY_ENABLE_PYSTACK |
226 | code_state = mp_pystack_alloc(sizeof(mp_code_state_t) + state_size); |
227 | #else |
228 | // If we use m_new_obj_var(), then on no memory, MemoryError will be |
229 | // raised. But this is not correct exception for a function call, |
230 | // RuntimeError should be raised instead. So, we use m_new_obj_var_maybe(), |
231 | // return NULL, then vm.c takes the needed action (either raise |
232 | // RuntimeError or fallback to stack allocation). |
233 | code_state = m_new_obj_var_maybe(mp_code_state_t, byte, state_size); |
234 | if (!code_state) { |
235 | return NULL; |
236 | } |
237 | #endif |
238 | |
239 | INIT_CODESTATE(code_state, self, n_state, n_args, n_kw, args); |
240 | |
241 | // execute the byte code with the correct globals context |
242 | mp_globals_set(self->globals); |
243 | |
244 | return code_state; |
245 | } |
246 | #endif |
247 | |
248 | STATIC mp_obj_t fun_bc_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
249 | MP_STACK_CHECK(); |
250 | |
251 | DEBUG_printf("Input n_args: " UINT_FMT ", n_kw: " UINT_FMT "\n" , n_args, n_kw); |
252 | DEBUG_printf("Input pos args: " ); |
253 | dump_args(args, n_args); |
254 | DEBUG_printf("Input kw args: " ); |
255 | dump_args(args + n_args, n_kw * 2); |
256 | |
257 | mp_obj_fun_bc_t *self = MP_OBJ_TO_PTR(self_in); |
258 | |
259 | size_t n_state, state_size; |
260 | DECODE_CODESTATE_SIZE(self->bytecode, n_state, state_size); |
261 | |
262 | // allocate state for locals and stack |
263 | mp_code_state_t *code_state = NULL; |
264 | #if MICROPY_ENABLE_PYSTACK |
265 | code_state = mp_pystack_alloc(sizeof(mp_code_state_t) + state_size); |
266 | #else |
267 | if (state_size > VM_MAX_STATE_ON_STACK) { |
268 | code_state = m_new_obj_var_maybe(mp_code_state_t, byte, state_size); |
269 | #if MICROPY_DEBUG_VM_STACK_OVERFLOW |
270 | if (code_state != NULL) { |
271 | memset(code_state->state, 0, state_size); |
272 | } |
273 | #endif |
274 | } |
275 | if (code_state == NULL) { |
276 | code_state = alloca(sizeof(mp_code_state_t) + state_size); |
277 | #if MICROPY_DEBUG_VM_STACK_OVERFLOW |
278 | memset(code_state->state, 0, state_size); |
279 | #endif |
280 | state_size = 0; // indicate that we allocated using alloca |
281 | } |
282 | #endif |
283 | |
284 | INIT_CODESTATE(code_state, self, n_state, n_args, n_kw, args); |
285 | |
286 | // execute the byte code with the correct globals context |
287 | mp_globals_set(self->globals); |
288 | mp_vm_return_kind_t vm_return_kind = mp_execute_bytecode(code_state, MP_OBJ_NULL); |
289 | mp_globals_set(code_state->old_globals); |
290 | |
291 | #if MICROPY_DEBUG_VM_STACK_OVERFLOW |
292 | if (vm_return_kind == MP_VM_RETURN_NORMAL) { |
293 | if (code_state->sp < code_state->state) { |
294 | mp_printf(MICROPY_DEBUG_PRINTER, "VM stack underflow: " INT_FMT "\n" , code_state->sp - code_state->state); |
295 | assert(0); |
296 | } |
297 | } |
298 | const byte *bytecode_ptr = self->bytecode; |
299 | size_t n_state_unused, n_exc_stack_unused, scope_flags_unused; |
300 | size_t n_pos_args, n_kwonly_args, n_def_args_unused; |
301 | MP_BC_PRELUDE_SIG_DECODE_INTO(bytecode_ptr, n_state_unused, n_exc_stack_unused, |
302 | scope_flags_unused, n_pos_args, n_kwonly_args, n_def_args_unused); |
303 | // We can't check the case when an exception is returned in state[0] |
304 | // and there are no arguments, because in this case our detection slot may have |
305 | // been overwritten by the returned exception (which is allowed). |
306 | if (!(vm_return_kind == MP_VM_RETURN_EXCEPTION && n_pos_args + n_kwonly_args == 0)) { |
307 | // Just check to see that we have at least 1 null object left in the state. |
308 | bool overflow = true; |
309 | for (size_t i = 0; i < n_state - n_pos_args - n_kwonly_args; ++i) { |
310 | if (code_state->state[i] == MP_OBJ_NULL) { |
311 | overflow = false; |
312 | break; |
313 | } |
314 | } |
315 | if (overflow) { |
316 | mp_printf(MICROPY_DEBUG_PRINTER, "VM stack overflow state=%p n_state+1=" UINT_FMT "\n" , code_state->state, n_state); |
317 | assert(0); |
318 | } |
319 | } |
320 | #endif |
321 | |
322 | mp_obj_t result; |
323 | if (vm_return_kind == MP_VM_RETURN_NORMAL) { |
324 | // return value is in *sp |
325 | result = *code_state->sp; |
326 | } else { |
327 | // must be an exception because normal functions can't yield |
328 | assert(vm_return_kind == MP_VM_RETURN_EXCEPTION); |
329 | // returned exception is in state[0] |
330 | result = code_state->state[0]; |
331 | } |
332 | |
333 | #if MICROPY_ENABLE_PYSTACK |
334 | mp_pystack_free(code_state); |
335 | #else |
336 | // free the state if it was allocated on the heap |
337 | if (state_size != 0) { |
338 | m_del_var(mp_code_state_t, byte, state_size, code_state); |
339 | } |
340 | #endif |
341 | |
342 | if (vm_return_kind == MP_VM_RETURN_NORMAL) { |
343 | return result; |
344 | } else { // MP_VM_RETURN_EXCEPTION |
345 | nlr_raise(result); |
346 | } |
347 | } |
348 | |
349 | #if MICROPY_PY_FUNCTION_ATTRS |
350 | void mp_obj_fun_bc_attr(mp_obj_t self_in, qstr attr, mp_obj_t *dest) { |
351 | if (dest[0] != MP_OBJ_NULL) { |
352 | // not load attribute |
353 | return; |
354 | } |
355 | if (attr == MP_QSTR___name__) { |
356 | dest[0] = MP_OBJ_NEW_QSTR(mp_obj_fun_get_name(self_in)); |
357 | } |
358 | if (attr == MP_QSTR___globals__) { |
359 | mp_obj_fun_bc_t *self = MP_OBJ_TO_PTR(self_in); |
360 | dest[0] = MP_OBJ_FROM_PTR(self->globals); |
361 | } |
362 | } |
363 | #endif |
364 | |
365 | const mp_obj_type_t mp_type_fun_bc = { |
366 | { &mp_type_type }, |
367 | .flags = MP_TYPE_FLAG_BINDS_SELF, |
368 | .name = MP_QSTR_function, |
369 | #if MICROPY_CPYTHON_COMPAT |
370 | .print = fun_bc_print, |
371 | #endif |
372 | .call = fun_bc_call, |
373 | .unary_op = mp_generic_unary_op, |
374 | #if MICROPY_PY_FUNCTION_ATTRS |
375 | .attr = mp_obj_fun_bc_attr, |
376 | #endif |
377 | }; |
378 | |
379 | mp_obj_t mp_obj_new_fun_bc(mp_obj_t def_args_in, mp_obj_t def_kw_args, const byte *code, const mp_uint_t *const_table) { |
380 | size_t n_def_args = 0; |
381 | size_t = 0; |
382 | mp_obj_tuple_t *def_args = MP_OBJ_TO_PTR(def_args_in); |
383 | if (def_args_in != MP_OBJ_NULL) { |
384 | assert(mp_obj_is_type(def_args_in, &mp_type_tuple)); |
385 | n_def_args = def_args->len; |
386 | n_extra_args = def_args->len; |
387 | } |
388 | if (def_kw_args != MP_OBJ_NULL) { |
389 | n_extra_args += 1; |
390 | } |
391 | mp_obj_fun_bc_t *o = m_new_obj_var(mp_obj_fun_bc_t, mp_obj_t, n_extra_args); |
392 | o->base.type = &mp_type_fun_bc; |
393 | o->globals = mp_globals_get(); |
394 | o->bytecode = code; |
395 | o->const_table = const_table; |
396 | if (def_args != NULL) { |
397 | memcpy(o->extra_args, def_args->items, n_def_args * sizeof(mp_obj_t)); |
398 | } |
399 | if (def_kw_args != MP_OBJ_NULL) { |
400 | o->extra_args[n_def_args] = def_kw_args; |
401 | } |
402 | return MP_OBJ_FROM_PTR(o); |
403 | } |
404 | |
405 | /******************************************************************************/ |
406 | /* native functions */ |
407 | |
408 | #if MICROPY_EMIT_NATIVE |
409 | |
410 | STATIC mp_obj_t fun_native_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
411 | MP_STACK_CHECK(); |
412 | mp_obj_fun_bc_t *self = self_in; |
413 | mp_call_fun_t fun = MICROPY_MAKE_POINTER_CALLABLE((void *)self->bytecode); |
414 | return fun(self_in, n_args, n_kw, args); |
415 | } |
416 | |
417 | STATIC const mp_obj_type_t mp_type_fun_native = { |
418 | { &mp_type_type }, |
419 | .flags = MP_TYPE_FLAG_BINDS_SELF, |
420 | .name = MP_QSTR_function, |
421 | .call = fun_native_call, |
422 | .unary_op = mp_generic_unary_op, |
423 | }; |
424 | |
425 | mp_obj_t mp_obj_new_fun_native(mp_obj_t def_args_in, mp_obj_t def_kw_args, const void *fun_data, const mp_uint_t *const_table) { |
426 | mp_obj_fun_bc_t *o = mp_obj_new_fun_bc(def_args_in, def_kw_args, (const byte *)fun_data, const_table); |
427 | o->base.type = &mp_type_fun_native; |
428 | return o; |
429 | } |
430 | |
431 | #endif // MICROPY_EMIT_NATIVE |
432 | |
433 | /******************************************************************************/ |
434 | /* inline assembler functions */ |
435 | |
436 | #if MICROPY_EMIT_INLINE_ASM |
437 | |
438 | typedef struct _mp_obj_fun_asm_t { |
439 | mp_obj_base_t base; |
440 | size_t n_args; |
441 | const void *fun_data; // GC must be able to trace this pointer |
442 | mp_uint_t type_sig; |
443 | } mp_obj_fun_asm_t; |
444 | |
445 | typedef mp_uint_t (*inline_asm_fun_0_t)(void); |
446 | typedef mp_uint_t (*inline_asm_fun_1_t)(mp_uint_t); |
447 | typedef mp_uint_t (*inline_asm_fun_2_t)(mp_uint_t, mp_uint_t); |
448 | typedef mp_uint_t (*inline_asm_fun_3_t)(mp_uint_t, mp_uint_t, mp_uint_t); |
449 | typedef mp_uint_t (*inline_asm_fun_4_t)(mp_uint_t, mp_uint_t, mp_uint_t, mp_uint_t); |
450 | |
451 | // convert a MicroPython object to a sensible value for inline asm |
452 | STATIC mp_uint_t convert_obj_for_inline_asm(mp_obj_t obj) { |
453 | // TODO for byte_array, pass pointer to the array |
454 | if (mp_obj_is_small_int(obj)) { |
455 | return MP_OBJ_SMALL_INT_VALUE(obj); |
456 | } else if (obj == mp_const_none) { |
457 | return 0; |
458 | } else if (obj == mp_const_false) { |
459 | return 0; |
460 | } else if (obj == mp_const_true) { |
461 | return 1; |
462 | } else if (mp_obj_is_type(obj, &mp_type_int)) { |
463 | return mp_obj_int_get_truncated(obj); |
464 | } else if (mp_obj_is_str(obj)) { |
465 | // pointer to the string (it's probably constant though!) |
466 | size_t l; |
467 | return (mp_uint_t)mp_obj_str_get_data(obj, &l); |
468 | } else { |
469 | const mp_obj_type_t *type = mp_obj_get_type(obj); |
470 | #if MICROPY_PY_BUILTINS_FLOAT |
471 | if (type == &mp_type_float) { |
472 | // convert float to int (could also pass in float registers) |
473 | return (mp_int_t)mp_obj_float_get(obj); |
474 | } |
475 | #endif |
476 | if (type == &mp_type_tuple || type == &mp_type_list) { |
477 | // pointer to start of tuple (could pass length, but then could use len(x) for that) |
478 | size_t len; |
479 | mp_obj_t *items; |
480 | mp_obj_get_array(obj, &len, &items); |
481 | return (mp_uint_t)items; |
482 | } else { |
483 | mp_buffer_info_t bufinfo; |
484 | if (mp_get_buffer(obj, &bufinfo, MP_BUFFER_READ)) { |
485 | // supports the buffer protocol, return a pointer to the data |
486 | return (mp_uint_t)bufinfo.buf; |
487 | } else { |
488 | // just pass along a pointer to the object |
489 | return (mp_uint_t)obj; |
490 | } |
491 | } |
492 | } |
493 | } |
494 | |
495 | STATIC mp_obj_t fun_asm_call(mp_obj_t self_in, size_t n_args, size_t n_kw, const mp_obj_t *args) { |
496 | mp_obj_fun_asm_t *self = self_in; |
497 | |
498 | mp_arg_check_num(n_args, n_kw, self->n_args, self->n_args, false); |
499 | |
500 | const void *fun = MICROPY_MAKE_POINTER_CALLABLE(self->fun_data); |
501 | |
502 | mp_uint_t ret; |
503 | if (n_args == 0) { |
504 | ret = ((inline_asm_fun_0_t)fun)(); |
505 | } else if (n_args == 1) { |
506 | ret = ((inline_asm_fun_1_t)fun)(convert_obj_for_inline_asm(args[0])); |
507 | } else if (n_args == 2) { |
508 | ret = ((inline_asm_fun_2_t)fun)(convert_obj_for_inline_asm(args[0]), convert_obj_for_inline_asm(args[1])); |
509 | } else if (n_args == 3) { |
510 | ret = ((inline_asm_fun_3_t)fun)(convert_obj_for_inline_asm(args[0]), convert_obj_for_inline_asm(args[1]), convert_obj_for_inline_asm(args[2])); |
511 | } else { |
512 | // compiler allows at most 4 arguments |
513 | assert(n_args == 4); |
514 | ret = ((inline_asm_fun_4_t)fun)( |
515 | convert_obj_for_inline_asm(args[0]), |
516 | convert_obj_for_inline_asm(args[1]), |
517 | convert_obj_for_inline_asm(args[2]), |
518 | convert_obj_for_inline_asm(args[3]) |
519 | ); |
520 | } |
521 | |
522 | return mp_native_to_obj(ret, self->type_sig); |
523 | } |
524 | |
525 | STATIC const mp_obj_type_t mp_type_fun_asm = { |
526 | { &mp_type_type }, |
527 | .flags = MP_TYPE_FLAG_BINDS_SELF, |
528 | .name = MP_QSTR_function, |
529 | .call = fun_asm_call, |
530 | .unary_op = mp_generic_unary_op, |
531 | }; |
532 | |
533 | mp_obj_t mp_obj_new_fun_asm(size_t n_args, const void *fun_data, mp_uint_t type_sig) { |
534 | mp_obj_fun_asm_t *o = m_new_obj(mp_obj_fun_asm_t); |
535 | o->base.type = &mp_type_fun_asm; |
536 | o->n_args = n_args; |
537 | o->fun_data = fun_data; |
538 | o->type_sig = type_sig; |
539 | return o; |
540 | } |
541 | |
542 | #endif // MICROPY_EMIT_INLINE_ASM |
543 | |