hwrng: core - Don't use a stack buffer in add_early_randomness()
[linux/fpc-iii.git] / drivers / acpi / acpica / dsmethod.c
blob32e9ddc0cf2bbbf4a73afc640956fcc92715bbcd
1 /******************************************************************************
3 * Module Name: dsmethod - Parser/Interpreter interface - control method parsing
5 *****************************************************************************/
7 /*
8 * Copyright (C) 2000 - 2016, Intel Corp.
9 * All rights reserved.
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions, and the following disclaimer,
16 * without modification.
17 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
18 * substantially similar to the "NO WARRANTY" disclaimer below
19 * ("Disclaimer") and any redistribution must be conditioned upon
20 * including a substantially similar Disclaimer requirement for further
21 * binary redistribution.
22 * 3. Neither the names of the above-listed copyright holders nor the names
23 * of any contributors may be used to endorse or promote products derived
24 * from this software without specific prior written permission.
26 * Alternatively, this software may be distributed under the terms of the
27 * GNU General Public License ("GPL") version 2 as published by the Free
28 * Software Foundation.
30 * NO WARRANTY
31 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
32 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
33 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
34 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
35 * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
36 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
37 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
38 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
39 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
40 * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
41 * POSSIBILITY OF SUCH DAMAGES.
44 #include <acpi/acpi.h>
45 #include "accommon.h"
46 #include "acdispat.h"
47 #include "acinterp.h"
48 #include "acnamesp.h"
49 #include "acparser.h"
50 #include "amlcode.h"
51 #include "acdebug.h"
53 #define _COMPONENT ACPI_DISPATCHER
54 ACPI_MODULE_NAME("dsmethod")
56 /* Local prototypes */
57 static acpi_status
58 acpi_ds_detect_named_opcodes(struct acpi_walk_state *walk_state,
59 union acpi_parse_object **out_op);
61 static acpi_status
62 acpi_ds_create_method_mutex(union acpi_operand_object *method_desc);
64 /*******************************************************************************
66 * FUNCTION: acpi_ds_auto_serialize_method
68 * PARAMETERS: node - Namespace Node of the method
69 * obj_desc - Method object attached to node
71 * RETURN: Status
73 * DESCRIPTION: Parse a control method AML to scan for control methods that
74 * need serialization due to the creation of named objects.
76 * NOTE: It is a bit of overkill to mark all such methods serialized, since
77 * there is only a problem if the method actually blocks during execution.
78 * A blocking operation is, for example, a Sleep() operation, or any access
79 * to an operation region. However, it is probably not possible to easily
80 * detect whether a method will block or not, so we simply mark all suspicious
81 * methods as serialized.
83 * NOTE2: This code is essentially a generic routine for parsing a single
84 * control method.
86 ******************************************************************************/
88 acpi_status
89 acpi_ds_auto_serialize_method(struct acpi_namespace_node *node,
90 union acpi_operand_object *obj_desc)
92 acpi_status status;
93 union acpi_parse_object *op = NULL;
94 struct acpi_walk_state *walk_state;
96 ACPI_FUNCTION_TRACE_PTR(ds_auto_serialize_method, node);
98 ACPI_DEBUG_PRINT((ACPI_DB_PARSE,
99 "Method auto-serialization parse [%4.4s] %p\n",
100 acpi_ut_get_node_name(node), node));
102 acpi_ex_enter_interpreter();
104 /* Create/Init a root op for the method parse tree */
106 op = acpi_ps_alloc_op(AML_METHOD_OP, obj_desc->method.aml_start);
107 if (!op) {
108 status = AE_NO_MEMORY;
109 goto unlock;
112 acpi_ps_set_name(op, node->name.integer);
113 op->common.node = node;
115 /* Create and initialize a new walk state */
117 walk_state =
118 acpi_ds_create_walk_state(node->owner_id, NULL, NULL, NULL);
119 if (!walk_state) {
120 acpi_ps_free_op(op);
121 status = AE_NO_MEMORY;
122 goto unlock;
125 status = acpi_ds_init_aml_walk(walk_state, op, node,
126 obj_desc->method.aml_start,
127 obj_desc->method.aml_length, NULL, 0);
128 if (ACPI_FAILURE(status)) {
129 acpi_ds_delete_walk_state(walk_state);
130 acpi_ps_free_op(op);
131 return_ACPI_STATUS(status);
134 walk_state->descending_callback = acpi_ds_detect_named_opcodes;
136 /* Parse the method, scan for creation of named objects */
138 status = acpi_ps_parse_aml(walk_state);
140 acpi_ps_delete_parse_tree(op);
141 unlock:
142 acpi_ex_exit_interpreter();
143 return_ACPI_STATUS(status);
146 /*******************************************************************************
148 * FUNCTION: acpi_ds_detect_named_opcodes
150 * PARAMETERS: walk_state - Current state of the parse tree walk
151 * out_op - Unused, required for parser interface
153 * RETURN: Status
155 * DESCRIPTION: Descending callback used during the loading of ACPI tables.
156 * Currently used to detect methods that must be marked serialized
157 * in order to avoid problems with the creation of named objects.
159 ******************************************************************************/
161 static acpi_status
162 acpi_ds_detect_named_opcodes(struct acpi_walk_state *walk_state,
163 union acpi_parse_object **out_op)
166 ACPI_FUNCTION_NAME(acpi_ds_detect_named_opcodes);
168 /* We are only interested in opcodes that create a new name */
170 if (!
171 (walk_state->op_info->
172 flags & (AML_NAMED | AML_CREATE | AML_FIELD))) {
173 return (AE_OK);
177 * At this point, we know we have a Named object opcode.
178 * Mark the method as serialized. Later code will create a mutex for
179 * this method to enforce serialization.
181 * Note, ACPI_METHOD_IGNORE_SYNC_LEVEL flag means that we will ignore the
182 * Sync Level mechanism for this method, even though it is now serialized.
183 * Otherwise, there can be conflicts with existing ASL code that actually
184 * uses sync levels.
186 walk_state->method_desc->method.sync_level = 0;
187 walk_state->method_desc->method.info_flags |=
188 (ACPI_METHOD_SERIALIZED | ACPI_METHOD_IGNORE_SYNC_LEVEL);
190 ACPI_DEBUG_PRINT((ACPI_DB_INFO,
191 "Method serialized [%4.4s] %p - [%s] (%4.4X)\n",
192 walk_state->method_node->name.ascii,
193 walk_state->method_node, walk_state->op_info->name,
194 walk_state->opcode));
196 /* Abort the parse, no need to examine this method any further */
198 return (AE_CTRL_TERMINATE);
201 /*******************************************************************************
203 * FUNCTION: acpi_ds_method_error
205 * PARAMETERS: status - Execution status
206 * walk_state - Current state
208 * RETURN: Status
210 * DESCRIPTION: Called on method error. Invoke the global exception handler if
211 * present, dump the method data if the debugger is configured
213 * Note: Allows the exception handler to change the status code
215 ******************************************************************************/
217 acpi_status
218 acpi_ds_method_error(acpi_status status, struct acpi_walk_state *walk_state)
220 u32 aml_offset;
222 ACPI_FUNCTION_ENTRY();
224 /* Ignore AE_OK and control exception codes */
226 if (ACPI_SUCCESS(status) || (status & AE_CODE_CONTROL)) {
227 return (status);
230 /* Invoke the global exception handler */
232 if (acpi_gbl_exception_handler) {
234 /* Exit the interpreter, allow handler to execute methods */
236 acpi_ex_exit_interpreter();
239 * Handler can map the exception code to anything it wants, including
240 * AE_OK, in which case the executing method will not be aborted.
242 aml_offset = (u32)ACPI_PTR_DIFF(walk_state->aml,
243 walk_state->parser_state.
244 aml_start);
246 status = acpi_gbl_exception_handler(status,
247 walk_state->method_node ?
248 walk_state->method_node->
249 name.integer : 0,
250 walk_state->opcode,
251 aml_offset, NULL);
252 acpi_ex_enter_interpreter();
255 acpi_ds_clear_implicit_return(walk_state);
257 if (ACPI_FAILURE(status)) {
258 acpi_ds_dump_method_stack(status, walk_state, walk_state->op);
260 /* Display method locals/args if debugger is present */
262 #ifdef ACPI_DEBUGGER
263 acpi_db_dump_method_info(status, walk_state);
264 #endif
267 return (status);
270 /*******************************************************************************
272 * FUNCTION: acpi_ds_create_method_mutex
274 * PARAMETERS: obj_desc - The method object
276 * RETURN: Status
278 * DESCRIPTION: Create a mutex object for a serialized control method
280 ******************************************************************************/
282 static acpi_status
283 acpi_ds_create_method_mutex(union acpi_operand_object *method_desc)
285 union acpi_operand_object *mutex_desc;
286 acpi_status status;
288 ACPI_FUNCTION_TRACE(ds_create_method_mutex);
290 /* Create the new mutex object */
292 mutex_desc = acpi_ut_create_internal_object(ACPI_TYPE_MUTEX);
293 if (!mutex_desc) {
294 return_ACPI_STATUS(AE_NO_MEMORY);
297 /* Create the actual OS Mutex */
299 status = acpi_os_create_mutex(&mutex_desc->mutex.os_mutex);
300 if (ACPI_FAILURE(status)) {
301 acpi_ut_delete_object_desc(mutex_desc);
302 return_ACPI_STATUS(status);
305 mutex_desc->mutex.sync_level = method_desc->method.sync_level;
306 method_desc->method.mutex = mutex_desc;
307 return_ACPI_STATUS(AE_OK);
310 /*******************************************************************************
312 * FUNCTION: acpi_ds_begin_method_execution
314 * PARAMETERS: method_node - Node of the method
315 * obj_desc - The method object
316 * walk_state - current state, NULL if not yet executing
317 * a method.
319 * RETURN: Status
321 * DESCRIPTION: Prepare a method for execution. Parses the method if necessary,
322 * increments the thread count, and waits at the method semaphore
323 * for clearance to execute.
325 ******************************************************************************/
327 acpi_status
328 acpi_ds_begin_method_execution(struct acpi_namespace_node *method_node,
329 union acpi_operand_object *obj_desc,
330 struct acpi_walk_state *walk_state)
332 acpi_status status = AE_OK;
334 ACPI_FUNCTION_TRACE_PTR(ds_begin_method_execution, method_node);
336 if (!method_node) {
337 return_ACPI_STATUS(AE_NULL_ENTRY);
340 acpi_ex_start_trace_method(method_node, obj_desc, walk_state);
342 /* Prevent wraparound of thread count */
344 if (obj_desc->method.thread_count == ACPI_UINT8_MAX) {
345 ACPI_ERROR((AE_INFO,
346 "Method reached maximum reentrancy limit (255)"));
347 return_ACPI_STATUS(AE_AML_METHOD_LIMIT);
351 * If this method is serialized, we need to acquire the method mutex.
353 if (obj_desc->method.info_flags & ACPI_METHOD_SERIALIZED) {
355 * Create a mutex for the method if it is defined to be Serialized
356 * and a mutex has not already been created. We defer the mutex creation
357 * until a method is actually executed, to minimize the object count
359 if (!obj_desc->method.mutex) {
360 status = acpi_ds_create_method_mutex(obj_desc);
361 if (ACPI_FAILURE(status)) {
362 return_ACPI_STATUS(status);
367 * The current_sync_level (per-thread) must be less than or equal to
368 * the sync level of the method. This mechanism provides some
369 * deadlock prevention.
371 * If the method was auto-serialized, we just ignore the sync level
372 * mechanism, because auto-serialization of methods can interfere
373 * with ASL code that actually uses sync levels.
375 * Top-level method invocation has no walk state at this point
377 if (walk_state &&
378 (!(obj_desc->method.
379 info_flags & ACPI_METHOD_IGNORE_SYNC_LEVEL))
380 && (walk_state->thread->current_sync_level >
381 obj_desc->method.mutex->mutex.sync_level)) {
382 ACPI_ERROR((AE_INFO,
383 "Cannot acquire Mutex for method [%4.4s]"
384 ", current SyncLevel is too large (%u)",
385 acpi_ut_get_node_name(method_node),
386 walk_state->thread->current_sync_level));
388 return_ACPI_STATUS(AE_AML_MUTEX_ORDER);
392 * Obtain the method mutex if necessary. Do not acquire mutex for a
393 * recursive call.
395 if (!walk_state ||
396 !obj_desc->method.mutex->mutex.thread_id ||
397 (walk_state->thread->thread_id !=
398 obj_desc->method.mutex->mutex.thread_id)) {
400 * Acquire the method mutex. This releases the interpreter if we
401 * block (and reacquires it before it returns)
403 status =
404 acpi_ex_system_wait_mutex(obj_desc->method.mutex->
405 mutex.os_mutex,
406 ACPI_WAIT_FOREVER);
407 if (ACPI_FAILURE(status)) {
408 return_ACPI_STATUS(status);
411 /* Update the mutex and walk info and save the original sync_level */
413 if (walk_state) {
414 obj_desc->method.mutex->mutex.
415 original_sync_level =
416 walk_state->thread->current_sync_level;
418 obj_desc->method.mutex->mutex.thread_id =
419 walk_state->thread->thread_id;
422 * Update the current sync_level only if this is not an auto-
423 * serialized method. In the auto case, we have to ignore
424 * the sync level for the method mutex (created for the
425 * auto-serialization) because we have no idea of what the
426 * sync level should be. Therefore, just ignore it.
428 if (!(obj_desc->method.info_flags &
429 ACPI_METHOD_IGNORE_SYNC_LEVEL)) {
430 walk_state->thread->current_sync_level =
431 obj_desc->method.sync_level;
433 } else {
434 obj_desc->method.mutex->mutex.
435 original_sync_level =
436 obj_desc->method.mutex->mutex.sync_level;
438 obj_desc->method.mutex->mutex.thread_id =
439 acpi_os_get_thread_id();
443 /* Always increase acquisition depth */
445 obj_desc->method.mutex->mutex.acquisition_depth++;
449 * Allocate an Owner ID for this method, only if this is the first thread
450 * to begin concurrent execution. We only need one owner_id, even if the
451 * method is invoked recursively.
453 if (!obj_desc->method.owner_id) {
454 status = acpi_ut_allocate_owner_id(&obj_desc->method.owner_id);
455 if (ACPI_FAILURE(status)) {
456 goto cleanup;
461 * Increment the method parse tree thread count since it has been
462 * reentered one more time (even if it is the same thread)
464 obj_desc->method.thread_count++;
465 acpi_method_count++;
466 return_ACPI_STATUS(status);
468 cleanup:
469 /* On error, must release the method mutex (if present) */
471 if (obj_desc->method.mutex) {
472 acpi_os_release_mutex(obj_desc->method.mutex->mutex.os_mutex);
474 return_ACPI_STATUS(status);
477 /*******************************************************************************
479 * FUNCTION: acpi_ds_call_control_method
481 * PARAMETERS: thread - Info for this thread
482 * this_walk_state - Current walk state
483 * op - Current Op to be walked
485 * RETURN: Status
487 * DESCRIPTION: Transfer execution to a called control method
489 ******************************************************************************/
491 acpi_status
492 acpi_ds_call_control_method(struct acpi_thread_state *thread,
493 struct acpi_walk_state *this_walk_state,
494 union acpi_parse_object *op)
496 acpi_status status;
497 struct acpi_namespace_node *method_node;
498 struct acpi_walk_state *next_walk_state = NULL;
499 union acpi_operand_object *obj_desc;
500 struct acpi_evaluate_info *info;
501 u32 i;
503 ACPI_FUNCTION_TRACE_PTR(ds_call_control_method, this_walk_state);
505 ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH,
506 "Calling method %p, currentstate=%p\n",
507 this_walk_state->prev_op, this_walk_state));
510 * Get the namespace entry for the control method we are about to call
512 method_node = this_walk_state->method_call_node;
513 if (!method_node) {
514 return_ACPI_STATUS(AE_NULL_ENTRY);
517 obj_desc = acpi_ns_get_attached_object(method_node);
518 if (!obj_desc) {
519 return_ACPI_STATUS(AE_NULL_OBJECT);
522 /* Init for new method, possibly wait on method mutex */
524 status =
525 acpi_ds_begin_method_execution(method_node, obj_desc,
526 this_walk_state);
527 if (ACPI_FAILURE(status)) {
528 return_ACPI_STATUS(status);
531 /* Begin method parse/execution. Create a new walk state */
533 next_walk_state =
534 acpi_ds_create_walk_state(obj_desc->method.owner_id, NULL, obj_desc,
535 thread);
536 if (!next_walk_state) {
537 status = AE_NO_MEMORY;
538 goto cleanup;
542 * The resolved arguments were put on the previous walk state's operand
543 * stack. Operands on the previous walk state stack always
544 * start at index 0. Also, null terminate the list of arguments
546 this_walk_state->operands[this_walk_state->num_operands] = NULL;
549 * Allocate and initialize the evaluation information block
550 * TBD: this is somewhat inefficient, should change interface to
551 * ds_init_aml_walk. For now, keeps this struct off the CPU stack
553 info = ACPI_ALLOCATE_ZEROED(sizeof(struct acpi_evaluate_info));
554 if (!info) {
555 status = AE_NO_MEMORY;
556 goto cleanup;
559 info->parameters = &this_walk_state->operands[0];
561 status = acpi_ds_init_aml_walk(next_walk_state, NULL, method_node,
562 obj_desc->method.aml_start,
563 obj_desc->method.aml_length, info,
564 ACPI_IMODE_EXECUTE);
566 ACPI_FREE(info);
567 if (ACPI_FAILURE(status)) {
568 goto cleanup;
572 * Delete the operands on the previous walkstate operand stack
573 * (they were copied to new objects)
575 for (i = 0; i < obj_desc->method.param_count; i++) {
576 acpi_ut_remove_reference(this_walk_state->operands[i]);
577 this_walk_state->operands[i] = NULL;
580 /* Clear the operand stack */
582 this_walk_state->num_operands = 0;
584 ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH,
585 "**** Begin nested execution of [%4.4s] **** WalkState=%p\n",
586 method_node->name.ascii, next_walk_state));
588 /* Invoke an internal method if necessary */
590 if (obj_desc->method.info_flags & ACPI_METHOD_INTERNAL_ONLY) {
591 status =
592 obj_desc->method.dispatch.implementation(next_walk_state);
593 if (status == AE_OK) {
594 status = AE_CTRL_TERMINATE;
598 return_ACPI_STATUS(status);
600 cleanup:
602 /* On error, we must terminate the method properly */
604 acpi_ds_terminate_control_method(obj_desc, next_walk_state);
605 acpi_ds_delete_walk_state(next_walk_state);
607 return_ACPI_STATUS(status);
610 /*******************************************************************************
612 * FUNCTION: acpi_ds_restart_control_method
614 * PARAMETERS: walk_state - State for preempted method (caller)
615 * return_desc - Return value from the called method
617 * RETURN: Status
619 * DESCRIPTION: Restart a method that was preempted by another (nested) method
620 * invocation. Handle the return value (if any) from the callee.
622 ******************************************************************************/
624 acpi_status
625 acpi_ds_restart_control_method(struct acpi_walk_state *walk_state,
626 union acpi_operand_object *return_desc)
628 acpi_status status;
629 int same_as_implicit_return;
631 ACPI_FUNCTION_TRACE_PTR(ds_restart_control_method, walk_state);
633 ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH,
634 "****Restart [%4.4s] Op %p ReturnValueFromCallee %p\n",
635 acpi_ut_get_node_name(walk_state->method_node),
636 walk_state->method_call_op, return_desc));
638 ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH,
639 " ReturnFromThisMethodUsed?=%X ResStack %p Walk %p\n",
640 walk_state->return_used,
641 walk_state->results, walk_state));
643 /* Did the called method return a value? */
645 if (return_desc) {
647 /* Is the implicit return object the same as the return desc? */
649 same_as_implicit_return =
650 (walk_state->implicit_return_obj == return_desc);
652 /* Are we actually going to use the return value? */
654 if (walk_state->return_used) {
656 /* Save the return value from the previous method */
658 status = acpi_ds_result_push(return_desc, walk_state);
659 if (ACPI_FAILURE(status)) {
660 acpi_ut_remove_reference(return_desc);
661 return_ACPI_STATUS(status);
665 * Save as THIS method's return value in case it is returned
666 * immediately to yet another method
668 walk_state->return_desc = return_desc;
672 * The following code is the optional support for the so-called
673 * "implicit return". Some AML code assumes that the last value of the
674 * method is "implicitly" returned to the caller, in the absence of an
675 * explicit return value.
677 * Just save the last result of the method as the return value.
679 * NOTE: this is optional because the ASL language does not actually
680 * support this behavior.
682 else if (!acpi_ds_do_implicit_return
683 (return_desc, walk_state, FALSE)
684 || same_as_implicit_return) {
686 * Delete the return value if it will not be used by the
687 * calling method or remove one reference if the explicit return
688 * is the same as the implicit return value.
690 acpi_ut_remove_reference(return_desc);
694 return_ACPI_STATUS(AE_OK);
697 /*******************************************************************************
699 * FUNCTION: acpi_ds_terminate_control_method
701 * PARAMETERS: method_desc - Method object
702 * walk_state - State associated with the method
704 * RETURN: None
706 * DESCRIPTION: Terminate a control method. Delete everything that the method
707 * created, delete all locals and arguments, and delete the parse
708 * tree if requested.
710 * MUTEX: Interpreter is locked
712 ******************************************************************************/
714 void
715 acpi_ds_terminate_control_method(union acpi_operand_object *method_desc,
716 struct acpi_walk_state *walk_state)
719 ACPI_FUNCTION_TRACE_PTR(ds_terminate_control_method, walk_state);
721 /* method_desc is required, walk_state is optional */
723 if (!method_desc) {
724 return_VOID;
727 if (walk_state) {
729 /* Delete all arguments and locals */
731 acpi_ds_method_data_delete_all(walk_state);
734 * If method is serialized, release the mutex and restore the
735 * current sync level for this thread
737 if (method_desc->method.mutex) {
739 /* Acquisition Depth handles recursive calls */
741 method_desc->method.mutex->mutex.acquisition_depth--;
742 if (!method_desc->method.mutex->mutex.acquisition_depth) {
743 walk_state->thread->current_sync_level =
744 method_desc->method.mutex->mutex.
745 original_sync_level;
747 acpi_os_release_mutex(method_desc->method.
748 mutex->mutex.os_mutex);
749 method_desc->method.mutex->mutex.thread_id = 0;
754 * Delete any namespace objects created anywhere within the
755 * namespace by the execution of this method. Unless:
756 * 1) This method is a module-level executable code method, in which
757 * case we want make the objects permanent.
758 * 2) There are other threads executing the method, in which case we
759 * will wait until the last thread has completed.
761 if (!(method_desc->method.info_flags & ACPI_METHOD_MODULE_LEVEL)
762 && (method_desc->method.thread_count == 1)) {
764 /* Delete any direct children of (created by) this method */
766 (void)acpi_ex_exit_interpreter();
767 acpi_ns_delete_namespace_subtree(walk_state->
768 method_node);
769 (void)acpi_ex_enter_interpreter();
772 * Delete any objects that were created by this method
773 * elsewhere in the namespace (if any were created).
774 * Use of the ACPI_METHOD_MODIFIED_NAMESPACE optimizes the
775 * deletion such that we don't have to perform an entire
776 * namespace walk for every control method execution.
778 if (method_desc->method.
779 info_flags & ACPI_METHOD_MODIFIED_NAMESPACE) {
780 (void)acpi_ex_exit_interpreter();
781 acpi_ns_delete_namespace_by_owner(method_desc->
782 method.
783 owner_id);
784 (void)acpi_ex_enter_interpreter();
785 method_desc->method.info_flags &=
786 ~ACPI_METHOD_MODIFIED_NAMESPACE;
791 /* Decrement the thread count on the method */
793 if (method_desc->method.thread_count) {
794 method_desc->method.thread_count--;
795 } else {
796 ACPI_ERROR((AE_INFO, "Invalid zero thread count in method"));
799 /* Are there any other threads currently executing this method? */
801 if (method_desc->method.thread_count) {
803 * Additional threads. Do not release the owner_id in this case,
804 * we immediately reuse it for the next thread executing this method
806 ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH,
807 "*** Completed execution of one thread, %u threads remaining\n",
808 method_desc->method.thread_count));
809 } else {
810 /* This is the only executing thread for this method */
813 * Support to dynamically change a method from not_serialized to
814 * Serialized if it appears that the method is incorrectly written and
815 * does not support multiple thread execution. The best example of this
816 * is if such a method creates namespace objects and blocks. A second
817 * thread will fail with an AE_ALREADY_EXISTS exception.
819 * This code is here because we must wait until the last thread exits
820 * before marking the method as serialized.
822 if (method_desc->method.
823 info_flags & ACPI_METHOD_SERIALIZED_PENDING) {
824 if (walk_state) {
825 ACPI_INFO(("Marking method %4.4s as Serialized "
826 "because of AE_ALREADY_EXISTS error",
827 walk_state->method_node->name.
828 ascii));
832 * Method tried to create an object twice and was marked as
833 * "pending serialized". The probable cause is that the method
834 * cannot handle reentrancy.
836 * The method was created as not_serialized, but it tried to create
837 * a named object and then blocked, causing the second thread
838 * entrance to begin and then fail. Workaround this problem by
839 * marking the method permanently as Serialized when the last
840 * thread exits here.
842 method_desc->method.info_flags &=
843 ~ACPI_METHOD_SERIALIZED_PENDING;
845 method_desc->method.info_flags |=
846 (ACPI_METHOD_SERIALIZED |
847 ACPI_METHOD_IGNORE_SYNC_LEVEL);
848 method_desc->method.sync_level = 0;
851 /* No more threads, we can free the owner_id */
853 if (!
854 (method_desc->method.
855 info_flags & ACPI_METHOD_MODULE_LEVEL)) {
856 acpi_ut_release_owner_id(&method_desc->method.owner_id);
860 acpi_ex_stop_trace_method((struct acpi_namespace_node *)method_desc->
861 method.node, method_desc, walk_state);
863 return_VOID;