1 // Copyright 2013 The Chromium Authors. All rights reserved.
2 // Use of this source code is governed by a BSD-style license that can be
3 // found in the LICENSE file.
5 #include "base/message_loop/message_loop.h"
10 #include "base/compiler_specific.h"
11 #include "base/debug/alias.h"
12 #include "base/debug/trace_event.h"
13 #include "base/lazy_instance.h"
14 #include "base/logging.h"
15 #include "base/memory/scoped_ptr.h"
16 #include "base/message_loop/message_pump_default.h"
17 #include "base/metrics/histogram.h"
18 #include "base/metrics/statistics_recorder.h"
19 #include "base/run_loop.h"
20 #include "base/third_party/dynamic_annotations/dynamic_annotations.h"
21 #include "base/thread_task_runner_handle.h"
22 #include "base/threading/thread_local.h"
23 #include "base/time/time.h"
24 #include "base/tracked_objects.h"
26 #if defined(OS_MACOSX)
27 #include "base/message_loop/message_pump_mac.h"
29 #if defined(OS_POSIX) && !defined(OS_IOS)
30 #include "base/message_loop/message_pump_libevent.h"
32 #if defined(OS_ANDROID)
33 #include "base/message_loop/message_pump_android.h"
36 #if defined(TOOLKIT_GTK)
45 // A lazily created thread local storage for quick access to a thread's message
46 // loop, if one exists. This should be safe and free of static constructors.
47 LazyInstance
<base::ThreadLocalPointer
<MessageLoop
> >::Leaky lazy_tls_ptr
=
48 LAZY_INSTANCE_INITIALIZER
;
50 // Logical events for Histogram profiling. Run with -message-loop-histogrammer
51 // to get an accounting of messages and actions taken on each thread.
52 const int kTaskRunEvent
= 0x1;
53 const int kTimerEvent
= 0x2;
55 // Provide range of message IDs for use in histogramming and debug display.
56 const int kLeastNonZeroMessageId
= 1;
57 const int kMaxMessageId
= 1099;
58 const int kNumberOfDistinctMessagesDisplayed
= 1100;
60 // Provide a macro that takes an expression (such as a constant, or macro
61 // constant) and creates a pair to initalize an array of pairs. In this case,
62 // our pair consists of the expressions value, and the "stringized" version
63 // of the expression (i.e., the exrpression put in quotes). For example, if
67 // then the following:
68 // VALUE_TO_NUMBER_AND_NAME(FOO + BAR)
71 // We use the resulting array as an argument to our histogram, which reads the
72 // number as a bucket identifier, and proceeds to use the corresponding name
73 // in the pair (i.e., the quoted string) when printing out a histogram.
74 #define VALUE_TO_NUMBER_AND_NAME(name) {name, #name},
76 const LinearHistogram::DescriptionPair event_descriptions_
[] = {
77 // Provide some pretty print capability in our histogram for our internal
80 // A few events we handle (kindred to messages), and used to profile actions.
81 VALUE_TO_NUMBER_AND_NAME(kTaskRunEvent
)
82 VALUE_TO_NUMBER_AND_NAME(kTimerEvent
)
84 {-1, NULL
} // The list must be null terminated, per API to histogram.
87 bool enable_histogrammer_
= false;
89 MessageLoop::MessagePumpFactory
* message_pump_for_ui_factory_
= NULL
;
91 // Returns true if MessagePump::ScheduleWork() must be called one
92 // time for every task that is added to the MessageLoop incoming queue.
93 bool AlwaysNotifyPump(MessageLoop::Type type
) {
94 #if defined(OS_ANDROID)
95 return type
== MessageLoop::TYPE_UI
|| type
== MessageLoop::TYPE_JAVA
;
103 //------------------------------------------------------------------------------
107 // Upon a SEH exception in this thread, it restores the original unhandled
109 static int SEHFilter(LPTOP_LEVEL_EXCEPTION_FILTER old_filter
) {
110 ::SetUnhandledExceptionFilter(old_filter
);
111 return EXCEPTION_CONTINUE_SEARCH
;
114 // Retrieves a pointer to the current unhandled exception filter. There
115 // is no standalone getter method.
116 static LPTOP_LEVEL_EXCEPTION_FILTER
GetTopSEHFilter() {
117 LPTOP_LEVEL_EXCEPTION_FILTER top_filter
= NULL
;
118 top_filter
= ::SetUnhandledExceptionFilter(0);
119 ::SetUnhandledExceptionFilter(top_filter
);
123 #endif // defined(OS_WIN)
125 //------------------------------------------------------------------------------
127 MessageLoop::TaskObserver::TaskObserver() {
130 MessageLoop::TaskObserver::~TaskObserver() {
133 MessageLoop::DestructionObserver::~DestructionObserver() {
136 //------------------------------------------------------------------------------
138 MessageLoop::MessageLoop(Type type
)
140 exception_restoration_(false),
141 nestable_tasks_allowed_(true),
143 os_modal_loop_(false),
145 message_histogram_(NULL
),
147 DCHECK(!current()) << "should only have one message loop per thread";
148 lazy_tls_ptr
.Pointer()->Set(this);
150 incoming_task_queue_
= new internal::IncomingTaskQueue(this);
151 message_loop_proxy_
=
152 new internal::MessageLoopProxyImpl(incoming_task_queue_
);
153 thread_task_runner_handle_
.reset(
154 new ThreadTaskRunnerHandle(message_loop_proxy_
));
156 // TODO(rvargas): Get rid of the OS guards.
158 #define MESSAGE_PUMP_UI new MessagePumpForUI()
159 #define MESSAGE_PUMP_IO new MessagePumpForIO()
160 #elif defined(OS_IOS)
161 #define MESSAGE_PUMP_UI MessagePumpMac::Create()
162 #define MESSAGE_PUMP_IO new MessagePumpIOSForIO()
163 #elif defined(OS_MACOSX)
164 #define MESSAGE_PUMP_UI MessagePumpMac::Create()
165 #define MESSAGE_PUMP_IO new MessagePumpLibevent()
166 #elif defined(OS_NACL)
167 // Currently NaCl doesn't have a UI MessageLoop.
168 // TODO(abarth): Figure out if we need this.
169 #define MESSAGE_PUMP_UI NULL
170 // ipc_channel_nacl.cc uses a worker thread to do socket reads currently, and
171 // doesn't require extra support for watching file descriptors.
172 #define MESSAGE_PUMP_IO new MessagePumpDefault()
173 #elif defined(OS_POSIX) // POSIX but not MACOSX.
174 #define MESSAGE_PUMP_UI new MessagePumpForUI()
175 #define MESSAGE_PUMP_IO new MessagePumpLibevent()
177 #error Not implemented
180 if (type_
== TYPE_UI
) {
181 if (message_pump_for_ui_factory_
)
182 pump_
.reset(message_pump_for_ui_factory_());
184 pump_
.reset(MESSAGE_PUMP_UI
);
185 } else if (type_
== TYPE_IO
) {
186 pump_
.reset(MESSAGE_PUMP_IO
);
187 #if defined(TOOLKIT_GTK)
188 } else if (type_
== TYPE_GPU
) {
189 pump_
.reset(new MessagePumpX11());
191 #if defined(OS_ANDROID)
192 } else if (type_
== TYPE_JAVA
) {
193 pump_
.reset(MESSAGE_PUMP_UI
);
196 DCHECK_EQ(TYPE_DEFAULT
, type_
);
197 pump_
.reset(new MessagePumpDefault());
201 MessageLoop::~MessageLoop() {
202 DCHECK_EQ(this, current());
206 // Clean up any unprocessed tasks, but take care: deleting a task could
207 // result in the addition of more tasks (e.g., via DeleteSoon). We set a
208 // limit on the number of times we will allow a deleted task to generate more
209 // tasks. Normally, we should only pass through this loop once or twice. If
210 // we end up hitting the loop limit, then it is probably due to one task that
211 // is being stubborn. Inspect the queues to see who is left.
213 for (int i
= 0; i
< 100; ++i
) {
214 DeletePendingTasks();
216 // If we end up with empty queues, then break out of the loop.
217 did_work
= DeletePendingTasks();
223 // Let interested parties have one last shot at accessing this.
224 FOR_EACH_OBSERVER(DestructionObserver
, destruction_observers_
,
225 WillDestroyCurrentMessageLoop());
227 thread_task_runner_handle_
.reset();
229 // Tell the incoming queue that we are dying.
230 incoming_task_queue_
->WillDestroyCurrentMessageLoop();
231 incoming_task_queue_
= NULL
;
232 message_loop_proxy_
= NULL
;
234 // OK, now make it so that no one can find us.
235 lazy_tls_ptr
.Pointer()->Set(NULL
);
239 MessageLoop
* MessageLoop::current() {
240 // TODO(darin): sadly, we cannot enable this yet since people call us even
241 // when they have no intention of using us.
242 // DCHECK(loop) << "Ouch, did you forget to initialize me?";
243 return lazy_tls_ptr
.Pointer()->Get();
247 void MessageLoop::EnableHistogrammer(bool enable
) {
248 enable_histogrammer_
= enable
;
252 bool MessageLoop::InitMessagePumpForUIFactory(MessagePumpFactory
* factory
) {
253 if (message_pump_for_ui_factory_
)
256 message_pump_for_ui_factory_
= factory
;
260 void MessageLoop::AddDestructionObserver(
261 DestructionObserver
* destruction_observer
) {
262 DCHECK_EQ(this, current());
263 destruction_observers_
.AddObserver(destruction_observer
);
266 void MessageLoop::RemoveDestructionObserver(
267 DestructionObserver
* destruction_observer
) {
268 DCHECK_EQ(this, current());
269 destruction_observers_
.RemoveObserver(destruction_observer
);
272 void MessageLoop::PostTask(
273 const tracked_objects::Location
& from_here
,
274 const Closure
& task
) {
275 DCHECK(!task
.is_null()) << from_here
.ToString();
276 incoming_task_queue_
->AddToIncomingQueue(from_here
, task
, TimeDelta(), true);
279 bool MessageLoop::TryPostTask(
280 const tracked_objects::Location
& from_here
,
281 const Closure
& task
) {
282 DCHECK(!task
.is_null()) << from_here
.ToString();
283 return incoming_task_queue_
->TryAddToIncomingQueue(from_here
, task
);
286 void MessageLoop::PostDelayedTask(
287 const tracked_objects::Location
& from_here
,
290 DCHECK(!task
.is_null()) << from_here
.ToString();
291 incoming_task_queue_
->AddToIncomingQueue(from_here
, task
, delay
, true);
294 void MessageLoop::PostNonNestableTask(
295 const tracked_objects::Location
& from_here
,
296 const Closure
& task
) {
297 DCHECK(!task
.is_null()) << from_here
.ToString();
298 incoming_task_queue_
->AddToIncomingQueue(from_here
, task
, TimeDelta(), false);
301 void MessageLoop::PostNonNestableDelayedTask(
302 const tracked_objects::Location
& from_here
,
305 DCHECK(!task
.is_null()) << from_here
.ToString();
306 incoming_task_queue_
->AddToIncomingQueue(from_here
, task
, delay
, false);
309 void MessageLoop::Run() {
314 void MessageLoop::RunUntilIdle() {
316 run_loop
.RunUntilIdle();
319 void MessageLoop::QuitWhenIdle() {
320 DCHECK_EQ(this, current());
322 run_loop_
->quit_when_idle_received_
= true;
324 NOTREACHED() << "Must be inside Run to call Quit";
328 void MessageLoop::QuitNow() {
329 DCHECK_EQ(this, current());
333 NOTREACHED() << "Must be inside Run to call Quit";
337 bool MessageLoop::IsType(Type type
) const {
338 return type_
== type
;
341 static void QuitCurrentWhenIdle() {
342 MessageLoop::current()->QuitWhenIdle();
346 Closure
MessageLoop::QuitWhenIdleClosure() {
347 return Bind(&QuitCurrentWhenIdle
);
350 void MessageLoop::SetNestableTasksAllowed(bool allowed
) {
351 if (nestable_tasks_allowed_
!= allowed
) {
352 nestable_tasks_allowed_
= allowed
;
353 if (!nestable_tasks_allowed_
)
355 // Start the native pump if we are not already pumping.
356 pump_
->ScheduleWork();
360 bool MessageLoop::NestableTasksAllowed() const {
361 return nestable_tasks_allowed_
;
364 bool MessageLoop::IsNested() {
365 return run_loop_
->run_depth_
> 1;
368 void MessageLoop::AddTaskObserver(TaskObserver
* task_observer
) {
369 DCHECK_EQ(this, current());
370 task_observers_
.AddObserver(task_observer
);
373 void MessageLoop::RemoveTaskObserver(TaskObserver
* task_observer
) {
374 DCHECK_EQ(this, current());
375 task_observers_
.RemoveObserver(task_observer
);
378 bool MessageLoop::is_running() const {
379 DCHECK_EQ(this, current());
380 return run_loop_
!= NULL
;
383 bool MessageLoop::IsHighResolutionTimerEnabledForTesting() {
384 return incoming_task_queue_
->IsHighResolutionTimerEnabledForTesting();
387 bool MessageLoop::IsIdleForTesting() {
388 // We only check the imcoming queue|, since we don't want to lock the work
390 return incoming_task_queue_
->IsIdleForTesting();
393 void MessageLoop::LockWaitUnLockForTesting(WaitableEvent
* caller_wait
,
394 WaitableEvent
* caller_signal
) {
395 incoming_task_queue_
->LockWaitUnLockForTesting(caller_wait
, caller_signal
);
398 //------------------------------------------------------------------------------
400 // Runs the loop in two different SEH modes:
401 // enable_SEH_restoration_ = false : any unhandled exception goes to the last
402 // one that calls SetUnhandledExceptionFilter().
403 // enable_SEH_restoration_ = true : any unhandled exception goes to the filter
404 // that was existed before the loop was run.
405 void MessageLoop::RunHandler() {
407 if (exception_restoration_
) {
408 RunInternalInSEHFrame();
417 __declspec(noinline
) void MessageLoop::RunInternalInSEHFrame() {
418 LPTOP_LEVEL_EXCEPTION_FILTER current_filter
= GetTopSEHFilter();
421 } __except(SEHFilter(current_filter
)) {
427 void MessageLoop::RunInternal() {
428 DCHECK_EQ(this, current());
432 #if !defined(OS_MACOSX) && !defined(OS_ANDROID) && \
433 !defined(USE_GTK_MESSAGE_PUMP)
434 if (run_loop_
->dispatcher_
&& type() == TYPE_UI
) {
435 static_cast<MessagePumpForUI
*>(pump_
.get())->
436 RunWithDispatcher(this, run_loop_
->dispatcher_
);
444 bool MessageLoop::ProcessNextDelayedNonNestableTask() {
445 if (run_loop_
->run_depth_
!= 1)
448 if (deferred_non_nestable_work_queue_
.empty())
451 PendingTask pending_task
= deferred_non_nestable_work_queue_
.front();
452 deferred_non_nestable_work_queue_
.pop();
454 RunTask(pending_task
);
458 void MessageLoop::RunTask(const PendingTask
& pending_task
) {
459 tracked_objects::TrackedTime start_time
=
460 tracked_objects::ThreadData::NowForStartOfRun(pending_task
.birth_tally
);
462 TRACE_EVENT_FLOW_END1("task", "MessageLoop::PostTask",
463 TRACE_ID_MANGLE(GetTaskTraceID(pending_task
)),
465 (start_time
- pending_task
.EffectiveTimePosted()).InMilliseconds());
466 // When tracing memory for posted tasks it's more valuable to attribute the
467 // memory allocations to the source function than generically to "RunTask".
468 TRACE_EVENT_WITH_MEMORY_TAG2(
469 "task", "MessageLoop::RunTask",
470 pending_task
.posted_from
.function_name(), // Name for memory tracking.
471 "src_file", pending_task
.posted_from
.file_name(),
472 "src_func", pending_task
.posted_from
.function_name());
474 DCHECK(nestable_tasks_allowed_
);
475 // Execute the task and assume the worst: It is probably not reentrant.
476 nestable_tasks_allowed_
= false;
478 // Before running the task, store the program counter where it was posted
479 // and deliberately alias it to ensure it is on the stack if the task
480 // crashes. Be careful not to assume that the variable itself will have the
481 // expected value when displayed by the optimizer in an optimized build.
482 // Look at a memory dump of the stack.
483 const void* program_counter
=
484 pending_task
.posted_from
.program_counter();
485 debug::Alias(&program_counter
);
487 HistogramEvent(kTaskRunEvent
);
489 FOR_EACH_OBSERVER(TaskObserver
, task_observers_
,
490 WillProcessTask(pending_task
));
491 pending_task
.task
.Run();
492 FOR_EACH_OBSERVER(TaskObserver
, task_observers_
,
493 DidProcessTask(pending_task
));
495 tracked_objects::ThreadData::TallyRunOnNamedThreadIfTracking(pending_task
,
496 start_time
, tracked_objects::ThreadData::NowForEndOfRun());
498 nestable_tasks_allowed_
= true;
501 bool MessageLoop::DeferOrRunPendingTask(const PendingTask
& pending_task
) {
502 if (pending_task
.nestable
|| run_loop_
->run_depth_
== 1) {
503 RunTask(pending_task
);
504 // Show that we ran a task (Note: a new one might arrive as a
509 // We couldn't run the task now because we're in a nested message loop
510 // and the task isn't nestable.
511 deferred_non_nestable_work_queue_
.push(pending_task
);
515 void MessageLoop::AddToDelayedWorkQueue(const PendingTask
& pending_task
) {
516 // Move to the delayed work queue.
517 delayed_work_queue_
.push(pending_task
);
520 bool MessageLoop::DeletePendingTasks() {
521 bool did_work
= !work_queue_
.empty();
522 while (!work_queue_
.empty()) {
523 PendingTask pending_task
= work_queue_
.front();
525 if (!pending_task
.delayed_run_time
.is_null()) {
526 // We want to delete delayed tasks in the same order in which they would
527 // normally be deleted in case of any funny dependencies between delayed
529 AddToDelayedWorkQueue(pending_task
);
532 did_work
|= !deferred_non_nestable_work_queue_
.empty();
533 while (!deferred_non_nestable_work_queue_
.empty()) {
534 deferred_non_nestable_work_queue_
.pop();
536 did_work
|= !delayed_work_queue_
.empty();
538 // Historically, we always delete the task regardless of valgrind status. It's
539 // not completely clear why we want to leak them in the loops above. This
540 // code is replicating legacy behavior, and should not be considered
541 // absolutely "correct" behavior. See TODO above about deleting all tasks
543 while (!delayed_work_queue_
.empty()) {
544 delayed_work_queue_
.pop();
549 uint64
MessageLoop::GetTaskTraceID(const PendingTask
& task
) {
550 return (static_cast<uint64
>(task
.sequence_num
) << 32) |
551 ((static_cast<uint64
>(reinterpret_cast<intptr_t>(this)) << 32) >> 32);
554 void MessageLoop::ReloadWorkQueue() {
555 // We can improve performance of our loading tasks from the incoming queue to
556 // |*work_queue| by waiting until the last minute (|*work_queue| is empty) to
557 // load. That reduces the number of locks-per-task significantly when our
559 if (work_queue_
.empty())
560 incoming_task_queue_
->ReloadWorkQueue(&work_queue_
);
563 void MessageLoop::ScheduleWork(bool was_empty
) {
564 // The Android UI message loop needs to get notified each time
565 // a task is added to the incoming queue.
566 if (was_empty
|| AlwaysNotifyPump(type_
))
567 pump_
->ScheduleWork();
570 //------------------------------------------------------------------------------
571 // Method and data for histogramming events and actions taken by each instance
574 void MessageLoop::StartHistogrammer() {
575 #if !defined(OS_NACL) // NaCl build has no metrics code.
576 if (enable_histogrammer_
&& !message_histogram_
577 && StatisticsRecorder::IsActive()) {
578 DCHECK(!thread_name_
.empty());
579 message_histogram_
= LinearHistogram::FactoryGetWithRangeDescription(
580 "MsgLoop:" + thread_name_
,
581 kLeastNonZeroMessageId
, kMaxMessageId
,
582 kNumberOfDistinctMessagesDisplayed
,
583 message_histogram_
->kHexRangePrintingFlag
,
584 event_descriptions_
);
589 void MessageLoop::HistogramEvent(int event
) {
590 #if !defined(OS_NACL)
591 if (message_histogram_
)
592 message_histogram_
->Add(event
);
596 bool MessageLoop::DoWork() {
597 if (!nestable_tasks_allowed_
) {
598 // Task can't be executed right now.
604 if (work_queue_
.empty())
607 // Execute oldest task.
609 PendingTask pending_task
= work_queue_
.front();
611 if (!pending_task
.delayed_run_time
.is_null()) {
612 AddToDelayedWorkQueue(pending_task
);
613 // If we changed the topmost task, then it is time to reschedule.
614 if (delayed_work_queue_
.top().task
.Equals(pending_task
.task
))
615 pump_
->ScheduleDelayedWork(pending_task
.delayed_run_time
);
617 if (DeferOrRunPendingTask(pending_task
))
620 } while (!work_queue_
.empty());
627 bool MessageLoop::DoDelayedWork(TimeTicks
* next_delayed_work_time
) {
628 if (!nestable_tasks_allowed_
|| delayed_work_queue_
.empty()) {
629 recent_time_
= *next_delayed_work_time
= TimeTicks();
633 // When we "fall behind," there will be a lot of tasks in the delayed work
634 // queue that are ready to run. To increase efficiency when we fall behind,
635 // we will only call Time::Now() intermittently, and then process all tasks
636 // that are ready to run before calling it again. As a result, the more we
637 // fall behind (and have a lot of ready-to-run delayed tasks), the more
638 // efficient we'll be at handling the tasks.
640 TimeTicks next_run_time
= delayed_work_queue_
.top().delayed_run_time
;
641 if (next_run_time
> recent_time_
) {
642 recent_time_
= TimeTicks::Now(); // Get a better view of Now();
643 if (next_run_time
> recent_time_
) {
644 *next_delayed_work_time
= next_run_time
;
649 PendingTask pending_task
= delayed_work_queue_
.top();
650 delayed_work_queue_
.pop();
652 if (!delayed_work_queue_
.empty())
653 *next_delayed_work_time
= delayed_work_queue_
.top().delayed_run_time
;
655 return DeferOrRunPendingTask(pending_task
);
658 bool MessageLoop::DoIdleWork() {
659 if (ProcessNextDelayedNonNestableTask())
662 if (run_loop_
->quit_when_idle_received_
)
668 void MessageLoop::GetQueueingInformation(size_t* queue_size
,
669 TimeDelta
* queueing_delay
) {
670 *queue_size
= work_queue_
.size();
671 if (*queue_size
== 0) {
672 *queueing_delay
= TimeDelta();
676 const PendingTask
& next_to_run
= work_queue_
.front();
677 tracked_objects::Duration duration
=
678 tracked_objects::TrackedTime::Now() - next_to_run
.EffectiveTimePosted();
679 *queueing_delay
= TimeDelta::FromMilliseconds(duration
.InMilliseconds());
682 void MessageLoop::DeleteSoonInternal(const tracked_objects::Location
& from_here
,
683 void(*deleter
)(const void*),
684 const void* object
) {
685 PostNonNestableTask(from_here
, Bind(deleter
, object
));
688 void MessageLoop::ReleaseSoonInternal(
689 const tracked_objects::Location
& from_here
,
690 void(*releaser
)(const void*),
691 const void* object
) {
692 PostNonNestableTask(from_here
, Bind(releaser
, object
));
695 //------------------------------------------------------------------------------
698 #if defined(OS_ANDROID)
699 void MessageLoopForUI::Start() {
700 // No Histogram support for UI message loop as it is managed by Java side
701 static_cast<MessagePumpForUI
*>(pump_
.get())->Start(this);
706 void MessageLoopForUI::Attach() {
707 static_cast<MessagePumpUIApplication
*>(pump_
.get())->Attach(this);
711 #if !defined(OS_MACOSX) && !defined(OS_NACL) && !defined(OS_ANDROID)
712 void MessageLoopForUI::AddObserver(Observer
* observer
) {
713 pump_ui()->AddObserver(observer
);
716 void MessageLoopForUI::RemoveObserver(Observer
* observer
) {
717 pump_ui()->RemoveObserver(observer
);
720 #endif // !defined(OS_MACOSX) && !defined(OS_NACL) && !defined(OS_ANDROID)
722 //------------------------------------------------------------------------------
727 void MessageLoopForIO::RegisterIOHandler(HANDLE file
, IOHandler
* handler
) {
728 pump_io()->RegisterIOHandler(file
, handler
);
731 bool MessageLoopForIO::RegisterJobObject(HANDLE job
, IOHandler
* handler
) {
732 return pump_io()->RegisterJobObject(job
, handler
);
735 bool MessageLoopForIO::WaitForIOCompletion(DWORD timeout
, IOHandler
* filter
) {
736 return pump_io()->WaitForIOCompletion(timeout
, filter
);
739 #elif defined(OS_IOS)
741 bool MessageLoopForIO::WatchFileDescriptor(int fd
,
744 FileDescriptorWatcher
*controller
,
746 return pump_io()->WatchFileDescriptor(
754 #elif defined(OS_POSIX) && !defined(OS_NACL)
756 bool MessageLoopForIO::WatchFileDescriptor(int fd
,
759 FileDescriptorWatcher
*controller
,
761 return pump_libevent()->WatchFileDescriptor(