static ACE_THR_FUNC_RETURN event_loop (void *arg) { ACE_Reactor *reactor = static_cast<ACE_Reactor *> (arg); reactor->owner (ACE_OS::thr_self ()); reactor->run_reactor_event_loop (); return 0; }
static void * worker (void *args) { ACE_Reactor *reactor = (ACE_Reactor *) args; reactor->owner (ACE_Thread::self ()); ACE_Time_Value timeout (4); for (;;) { //ACE_DEBUG ((LM_DEBUG, "(%t) calling handle_events\n")); switch (reactor->handle_events (timeout)) { case -1: ACE_ERROR_RETURN ((LM_ERROR, "(%t) %p\n", "reactor"), 0); /* NOTREACHED */ case 0: ACE_ERROR_RETURN ((LM_ERROR, "(%t) timeout\n"), 0); /* NOTREACHED */ } // ACE_DEBUG ((LM_DEBUG, "(%t) done with handle_events\n")); } ACE_NOTREACHED(return 0); }
static void * worker (void *args) { ACE_Reactor *reactor = reinterpret_cast<ACE_Reactor *> (args); // Make this thread the owner of the Reactor's event loop. reactor->owner (ACE_Thread::self ()); // Use a timeout to inform the Reactor when to shutdown. ACE_Time_Value timeout (4); for (;;) switch (reactor->handle_events (timeout)) { case -1: ACE_ERROR_RETURN ((LM_ERROR, ACE_TEXT ("(%t) %p\n"), ACE_TEXT ("reactor")), 0); /* NOTREACHED */ case 0: ACE_DEBUG ((LM_DEBUG, ACE_TEXT ("(%t) Reactor shutdown\n"))); return 0; } ACE_NOTREACHED (return 0); }
static void * dispatch (void *arg) { // every thread must register the same stream to write to file if (out_stream) { ACE_LOG_MSG->set_flags (ACE_Log_Msg::OSTREAM); ACE_LOG_MSG->msg_ostream (out_stream); } ACE_DEBUG ((LM_DEBUG, ACE_TEXT (" (%t) Dispatcher Thread started!\n"))); ACE_Reactor *r = reinterpret_cast <ACE_Reactor *> (arg); int result; r->owner (ACE_OS::thr_self ()); while (1) { result = r->handle_events (); if (result <= 0) ACE_DEBUG ((LM_DEBUG, ACE_TEXT ("Dispatch: handle_events (): %d"), result)); } ACE_NOTREACHED (return 0); }
static void *run_reactor (void *pReactor) { ACE_Reactor *pLogReactor = (ACE_Reactor *)pReactor; pLogReactor->owner(ACE_Thread_Manager::instance ()->thr_self()); pLogReactor->run_reactor_event_loop (); ACE_DEBUG((LM_INFO, ACE_TEXT("(%P|%t) %M run_reactor exit[%N,%l]\n"))); return 0; }
//for ACE_TP_Reactor static ACE_THR_FUNC_RETURN event_loop (void *arg) { //ACE_Reactor *reactor = ACE_static_cast(ACE_Reactor *, arg); ACE_Reactor *reactor = static_cast< ACE_Reactor *> (arg); ACE_DEBUG((LM_INFO, "[%D] ace tp reactor event loop, tid=%d,pid=%d\n" , ACE_OS::thr_self () , ACE_OS::getpid() )); reactor->owner(ACE_OS::getpid()); reactor->run_reactor_event_loop (); return 0; }
int TAO_Leader_Follower::wait_for_event (TAO_LF_Event *event, TAO_Transport *transport, ACE_Time_Value *max_wait_time) { // Obtain the lock. ACE_GUARD_RETURN (TAO_SYNCH_MUTEX, ace_mon, this->lock (), -1); TAO::ORB_Countdown_Time countdown (max_wait_time); // Optimize the first iteration [no access to errno] int result = 1; // For some cases the transport may disappear like when waiting for // connection to be initiated or closed. So cache the id. // @@ NOTE: This is not completely safe either. We will be fine for // cases that don't access the id ie. when debug level is off but // with debugging level on we are on a sticky wicket. Hopefully none // of our users should run TAO with debugging enabled like they did // in PathFinder size_t t_id = 0; if (TAO_debug_level && transport != 0) { t_id = transport->id (); } { // Scope #1: All threads inside here are client threads // Calls this->set_client_thread () on construction and // this->reset_client_thread () on destruction. TAO_LF_Client_Thread_Helper client_thread_helper (*this); ACE_UNUSED_ARG (client_thread_helper); // The loop here is for when we get elected (client) leader and // then later relinquish the leader position and our event has // still not completed (and we haven't run out of time). // All the conditions below are basically the various ways the // leader loop below can end, other than the event being complete while (event->keep_waiting_i () && !(result == 0 && max_wait_time != 0 && *max_wait_time == ACE_Time_Value::zero) && result != -1) { // Scope #2: threads here alternate between being leader/followers // Check if there is a leader. Note that it cannot be us since we // gave up our leadership when we became a client. if (this->leader_available ()) { // Scope #3: threads here are followers // = Wait as a follower. // Grab a follower: TAO_LF_Follower_Auto_Ptr follower (*this); if (follower.get () == 0) return -1; if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" (follower), cond <%@>\n"), t_id, follower.get ())); // Bound the follower and the LF_Event, this is important to // get a signal when the event terminates TAO_LF_Event_Binder event_binder (event, follower.get ()); while (event->keep_waiting_i () && this->leader_available ()) { // Scope #4: this loop handles spurious wake-ups // Add ourselves to the list, do it everytime we wake up // from the CV loop. Because: // // - The leader thread could have elected us as the new // leader. // - Before we can assume the role another thread becomes // the leader // - But our condition variable could have been removed // already, if we don't add it again we will never wake // up. // // Notice that we can have spurious wake ups, in that case // adding the leader results in an error, that must be // ignored. // You may be thinking of not removing the condition // variable in the code that sends the signal, but // removing it here, that does not work either, in that // case the condition variable may be used twice: // // - Wake up because its reply arrived // - Wake up because it must become the leader // // but only the first one has any effect, so the leader is // lost. TAO_LF_Follower_Auto_Adder auto_adder (*this, follower); if (max_wait_time == 0) { if (follower->wait (max_wait_time) == -1) { if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event, ") ACE_TEXT (" (follower) [no timer, cond failed]\n"), t_id)); // @@ Michael: What is our error handling in this case? // We could be elected as leader and // no leader would come in? return -1; } } else { countdown.update (); ACE_Time_Value tv = ACE_OS::gettimeofday (); tv += *max_wait_time; if (follower->wait (&tv) == -1) { if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait, ") ACE_TEXT ("(follower) [has timer, follower failed]\n"), t_id )); // If we have timedout set the state in the // LF_Event. We call the non-locking, // no-signalling method on LF_Event. if (errno == ETIME) // We have timedout event->set_state (TAO_LF_Event::LFS_TIMEOUT); if (!event->successful_i ()) { // Remove follower can fail because either // 1) the condition was satisfied (i.e. reply // received or queue drained), or // 2) somebody elected us as leader, or // 3) the connection got closed. // // Therefore: // If remove_follower fails and the condition // was not satisfied, we know that we got // elected as a leader. // But we got a timeout, so we cannot become // the leader, therefore, we have to select a // new leader. // if (this->elect_new_leader () == -1 && TAO_debug_level > 0) { TAOLIB_ERROR ((LM_ERROR, ACE_TEXT("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event, ") ACE_TEXT("elect_new_leader failed\n"), t_id )); } } return -1; } } } // End Scope #4: loop to handle spurious wakeups countdown.update (); // @@ Michael: This is an old comment why we do not want to // remove the follower here. // We should not remove the follower here, we *must* remove it when // we signal it so the same condition is not signalled for // both wake up as a follower and as the next leader. if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" done (follower), successful %d\n"), t_id, event->successful_i ())); // Now somebody woke us up to become a leader or to handle our // input. We are already removed from the follower queue. if (event->successful_i ()) return 0; if (event->error_detected_i ()) return -1; // FALLTHROUGH // We only get here if we woke up but the reply is not // complete yet, time to assume the leader role.... // i.e. ACE_ASSERT (event->successful () == 0); } // End Scope #3: we are no longer a follower // One last attempt to avoid becoming a client-leader // If there is a new_leader_generator attached, give it a chance to // create a leader thread before becoming a leader. if (this->avoid_client_leader_ && this->no_leaders_available()) { if (TAO_debug_level >= 5) { TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event, ") ACE_TEXT ("Would become client leader, ") ACE_TEXT ("but generating new thread\n"), t_id)); } // Yield, providing the event thread some time to grab leadership ACE_GUARD_RETURN (ACE_Reverse_Lock<TAO_SYNCH_MUTEX>, rev_mon, this->reverse_lock (), -1); ACE_OS::thr_yield (); continue; } else { if (TAO_debug_level >= 5) { TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event, ") ACE_TEXT ("Becoming client leader.\n"), t_id)); } } // = Leader Code. // The only way to reach this point is if we must become the // leader, because there is no leader or we have to update to a // leader or we are doing nested upcalls in this case we do // increase the refcount on the leader in TAO_ORB_Core. // Calls this->set_client_leader_thread () on // construction and this->reset_client_leader_thread () // on destruction. Note that this may increase the refcount of // the leader. { // Scope #5: We are now the client-leader TAO_LF_Client_Leader_Thread_Helper client_leader_thread_helper (*this); ACE_UNUSED_ARG (client_leader_thread_helper); { // Scope #6: release the lock via a reverse lock ACE_GUARD_RETURN (ACE_Reverse_Lock<TAO_SYNCH_MUTEX>, rev_mon, this->reverse_lock (), -1); // Become owner of the reactor. ACE_Reactor *reactor = this->reactor_; reactor->owner (ACE_Thread::self ()); // Run the reactor event loop. if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" (leader) enter reactor event loop\n"), t_id)); // If we got our event, no need to run the event loop any // further. while (event->keep_waiting_i ()) { // Run the event loop. result = reactor->handle_events (max_wait_time); // Did we timeout? If so, stop running the loop. if (result == 0 && max_wait_time != 0 && *max_wait_time == ACE_Time_Value::zero) break; // Other errors? If so, stop running the loop. if (result == -1) break; // Has an event loop thread become available to take over? // Yes, we are checking this without the lock, however, if // we get a false reading we'll just circle around and // become leader again... if (this->event_loop_threads_waiting_) break; // Did we give up leadership? if (!this->is_client_leader_thread ()) break; // Otherwise, keep going... } if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" (leader) exit reactor event loop\n"), t_id)); } // End Scope #6: we should now hold the lock again // End artificial scope for auto_ptr like helpers calling: // this->reset_client_leader_thread (). } // End Scope #5: we are no longer a client-leader // We only get here if we were the client leader and either our // event completed or an event loop thread has become available to // become leader. // resume any deferred events before we switch to a new leader thread this->resume_events (); // Wake and yield to any event loop threads that may be waiting to // take leadership - otherwise we will just loop around and take // leadership again (because we hold the lock). if (this->event_loop_threads_waiting_ && !this->leader_available ()) { if (TAO_debug_level >= 5) TAOLIB_DEBUG ((LM_DEBUG, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" (client) waking and yielding to allow event thread leadership\n"), t_id)); // Wake up the next leader (in case not yet done) if (this->elect_new_leader () == -1) TAOLIB_ERROR_RETURN ((LM_ERROR, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" failed to elect new leader\n"), t_id), -1); // Yield, providing the event thread some time to grab leadership ACE_GUARD_RETURN (ACE_Reverse_Lock<TAO_SYNCH_MUTEX>, rev_mon, this->reverse_lock (), -1); ACE_OS::thr_yield (); } } // End Scope #2: we loop here if our event is incomplete // End artificial scope for auto_ptr like helpers calling: // this->reset_client_thread () // We should only get here when our event is complete or timed-out } // End Scope #1 // Wake up the next leader, we cannot do that in handle_input, // because the woken up thread would try to get into handle_events, // which is at the time in handle_input still occupied. But do it // before checking the error in <result>, even if there is an error // in our input we should continue running the loop in another // thread. if (this->elect_new_leader () == -1) TAOLIB_ERROR_RETURN ((LM_ERROR, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" failed to elect new leader\n"), t_id), -1); if (result == -1 && !this->reactor_->reactor_event_loop_done ()) TAOLIB_ERROR_RETURN ((LM_ERROR, ACE_TEXT ("TAO (%P|%t) - Leader_Follower[%d]::wait_for_event,") ACE_TEXT (" handle_events failed\n"), t_id), -1); // Return an error if there was a problem receiving the reply... if (max_wait_time != 0) { if (!event->successful_i () && *max_wait_time == ACE_Time_Value::zero) { result = -1; errno = ETIME; } else if (event->error_detected_i ()) { // If the time did not expire yet, but we get a failure, // e.g. the connections closed, we should still return an error. result = -1; } } else { /** * There should be no reason to reset the value of result * here. If there was an error in handle_events () that the * leader saw, I (Bala) believe it should be propagated to the * clients. * result = 0; */ if (event->error_detected_i ()) { result = -1; } } return result; }