PUBLIC void Ipc_gate::invoke(L4_obj_ref /*self*/, L4_fpage::Rights rights, Syscall_frame *f, Utcb *utcb) { Syscall_frame *ipc_f = f; //LOG_MSG_3VAL(current(), "gIPC", Mword(_thread), _id, f->obj_2_flags()); //printf("Invoke: Ipc_gate(%lx->%p)...\n", _id, _thread); Thread *ct = current_thread(); Thread *sender = 0; Thread *partner = 0; bool have_rcv = false; if (EXPECT_FALSE(!_thread)) { L4_error e = block(ct, f->timeout().snd, utcb); if (!e.ok()) { f->tag(commit_error(utcb, e)); return; } if (EXPECT_FALSE(!_thread)) { f->tag(commit_error(utcb, L4_error::Not_existent)); return; } } bool ipc = _thread->check_sys_ipc(f->ref().op(), &partner, &sender, &have_rcv); LOG_TRACE("IPC Gate invoke", "gate", current(), Log_ipc_gate_invoke, l->gate_dbg_id = dbg_id(); l->thread_dbg_id = _thread->dbg_id(); l->label = _id | cxx::int_value<L4_fpage::Rights>(rights); );
// result of ipc static unsigned formatter_ipc_res(Tb_entry *tb, const char *tidstr, unsigned tidlen, char *buf, int maxlen) { Tb_entry_ipc_res *e = static_cast<Tb_entry_ipc_res*>(tb); L4_error error; if (e->tag().has_error()) error = e->result(); else error = L4_error::None; const char *m = "answ"; //get_ipc_type(e); my_snprintf(buf, maxlen, " %s%-*s %s [%08lx] L=%lx err=%lx (%s) (%lx,%lx) ", e->is_np() ? "[np] " : "", tidlen, tidstr, m, e->tag().raw(), e->from(), error.raw(), error.str_error(), e->dword(0), e->dword(1)); return maxlen; }
PRIVATE inline L4_msg_tag Thread_object::sys_vcpu_resume(L4_msg_tag const &tag, Utcb *utcb) { if (this != current() || !(state() & Thread_vcpu_enabled)) return commit_result(-L4_err::EInval); Space *s = space(); Vcpu_state *vcpu = vcpu_state().access(true); L4_obj_ref user_task = vcpu->user_task; if (user_task.valid()) { L4_fpage::Rights task_rights = L4_fpage::Rights(0); Task *task = Kobject::dcast<Task*>(s->lookup_local(user_task.cap(), &task_rights)); if (EXPECT_FALSE(task && !(task_rights & L4_fpage::Rights::W()))) return commit_result(-L4_err::EPerm); if (task != vcpu_user_space()) vcpu_set_user_space(task); vcpu->user_task = L4_obj_ref(); } else if (user_task.op() == L4_obj_ref::Ipc_reply) vcpu_set_user_space(0); L4_snd_item_iter snd_items(utcb, tag.words()); int items = tag.items(); if (vcpu_user_space()) for (; items && snd_items.more(); --items) { if (EXPECT_FALSE(!snd_items.next())) break; Lock_guard<Lock> guard; if (!guard.check_and_lock(&static_cast<Task *>(vcpu_user_space())->existence_lock)) return commit_result(-L4_err::ENoent); cpu_lock.clear(); L4_snd_item_iter::Item const *const item = snd_items.get(); L4_fpage sfp(item->d); Reap_list rl; L4_error err = fpage_map(space(), sfp, vcpu_user_space(), L4_fpage::all_spaces(), item->b, &rl); rl.del(); cpu_lock.lock(); if (EXPECT_FALSE(!err.ok())) return commit_error(utcb, err); } if ((vcpu->_saved_state & Vcpu_state::F_irqs) && (vcpu->sticky_flags & Vcpu_state::Sf_irq_pending)) { assert_kdb(cpu_lock.test()); do_ipc(L4_msg_tag(), 0, 0, true, 0, L4_timeout_pair(L4_timeout::Zero, L4_timeout::Zero), &vcpu->_ipc_regs, L4_fpage::Rights::FULL()); vcpu = vcpu_state().access(true); if (EXPECT_TRUE(!vcpu->_ipc_regs.tag().has_error() || this->utcb().access(true)->error.error() == L4_error::R_timeout)) { vcpu->_ts.set_ipc_upcall(); Address sp; // tried to resume to user mode, so an IRQ enters from user mode if (vcpu->_saved_state & Vcpu_state::F_user_mode) sp = vcpu->_entry_sp; else sp = vcpu->_ts.sp(); arch_load_vcpu_kern_state(vcpu, true); LOG_TRACE("VCPU events", "vcpu", this, Vcpu_log, l->type = 4; l->state = vcpu->state; l->ip = vcpu->_entry_ip; l->sp = sp; l->space = static_cast<Task*>(_space.vcpu_aware())->dbg_id(); );