static void zynqmp_pwr_domain_suspend(const psci_power_state_t *target_state) { unsigned int cpu_id = plat_my_core_pos(); const struct pm_proc *proc = pm_get_proc(cpu_id); for (size_t i = 0; i <= PLAT_MAX_PWR_LVL; i++) VERBOSE("%s: target_state->pwr_domain_state[%lu]=%x\n", __func__, i, target_state->pwr_domain_state[i]); /* Send request to PMU to suspend this core */ pm_self_suspend(proc->node_id, MAX_LATENCY, 0, zynqmp_sec_entry); /* APU is to be turned off */ if (target_state->pwr_domain_state[1] > PLAT_MAX_RET_STATE) { /* Power down L2 cache */ pm_set_requirement(NODE_L2, 0, 0, REQ_ACK_NO); /* Send request for OCM retention state */ set_ocm_retention(); /* disable coherency */ plat_arm_interconnect_exit_coherency(); } }
static void zynqmp_pwr_domain_suspend(const psci_power_state_t *target_state) { unsigned int state; unsigned int cpu_id = plat_my_core_pos(); const struct pm_proc *proc = pm_get_proc(cpu_id); for (size_t i = 0; i <= PLAT_MAX_PWR_LVL; i++) VERBOSE("%s: target_state->pwr_domain_state[%lu]=%x\n", __func__, i, target_state->pwr_domain_state[i]); state = target_state->pwr_domain_state[1] > PLAT_MAX_RET_STATE ? PM_STATE_SUSPEND_TO_RAM : PM_STATE_CPU_IDLE; /* Send request to PMU to suspend this core */ pm_self_suspend(proc->node_id, MAX_LATENCY, state, zynqmp_sec_entry); /* APU is to be turned off */ if (target_state->pwr_domain_state[1] > PLAT_MAX_RET_STATE) { /* disable coherency */ plat_arm_interconnect_exit_coherency(); } }
static void zynqmp_pwr_domain_off(const psci_power_state_t *target_state) { unsigned int cpu_id = plat_my_core_pos(); const struct pm_proc *proc = pm_get_proc(cpu_id); for (size_t i = 0; i <= PLAT_MAX_PWR_LVL; i++) VERBOSE("%s: target_state->pwr_domain_state[%lu]=%x\n", __func__, i, target_state->pwr_domain_state[i]); /* Prevent interrupts from spuriously waking up this cpu */ gicv2_cpuif_disable(); /* * Send request to PMU to power down the appropriate APU CPU * core. * According to PSCI specification, CPU_off function does not * have resume address and CPU core can only be woken up * invoking CPU_on function, during which resume address will * be set. */ pm_self_suspend(proc->node_id, MAX_LATENCY, 0, 0); }
/** * pm_smc_handler() - SMC handler for PM-API calls coming from EL1/EL2. * @smc_fid - Function Identifier * @x1 - x4 - Arguments * @cookie - Unused * @handler - Pointer to caller's context structure * * @return - Unused * * Determines that smc_fid is valid and supported PM SMC Function ID from the * list of pm_api_ids, otherwise completes the request with * the unknown SMC Function ID * * The SMC calls for PM service are forwarded from SIP Service SMC handler * function with rt_svc_handle signature */ uint64_t pm_smc_handler(uint32_t smc_fid, uint64_t x1, uint64_t x2, uint64_t x3, uint64_t x4, void *cookie, void *handle, uint64_t flags) { enum pm_ret_status ret; uint32_t pm_arg[4]; /* Handle case where PM wasn't initialized properly */ if (pm_down) SMC_RET1(handle, SMC_UNK); pm_arg[0] = (uint32_t)x1; pm_arg[1] = (uint32_t)(x1 >> 32); pm_arg[2] = (uint32_t)x2; pm_arg[3] = (uint32_t)(x2 >> 32); switch (smc_fid & FUNCID_NUM_MASK) { /* PM API Functions */ case PM_SELF_SUSPEND: ret = pm_self_suspend(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_REQ_SUSPEND: ret = pm_req_suspend(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_REQ_WAKEUP: ret = pm_req_wakeup(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_FORCE_POWERDOWN: ret = pm_force_powerdown(pm_arg[0], pm_arg[1]); SMC_RET1(handle, (uint64_t)ret); case PM_ABORT_SUSPEND: ret = pm_abort_suspend(pm_arg[0]); SMC_RET1(handle, (uint64_t)ret); case PM_SET_WAKEUP_SOURCE: ret = pm_set_wakeup_source(pm_arg[0], pm_arg[1], pm_arg[2]); SMC_RET1(handle, (uint64_t)ret); case PM_SYSTEM_SHUTDOWN: ret = pm_system_shutdown(pm_arg[0]); SMC_RET1(handle, (uint64_t)ret); case PM_REQ_NODE: ret = pm_req_node(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_RELEASE_NODE: ret = pm_release_node(pm_arg[0]); SMC_RET1(handle, (uint64_t)ret); case PM_SET_REQUIREMENT: ret = pm_set_requirement(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_SET_MAX_LATENCY: ret = pm_set_max_latency(pm_arg[0], pm_arg[1]); SMC_RET1(handle, (uint64_t)ret); case PM_GET_API_VERSION: /* Check is PM API version already verified */ if (pm_ctx.api_version == PM_VERSION) SMC_RET1(handle, (uint64_t)PM_RET_SUCCESS | ((uint64_t)PM_VERSION << 32)); ret = pm_get_api_version(&pm_ctx.api_version); SMC_RET1(handle, (uint64_t)ret | ((uint64_t)pm_ctx.api_version << 32)); case PM_SET_CONFIGURATION: ret = pm_set_configuration(pm_arg[0]); SMC_RET1(handle, (uint64_t)ret); case PM_GET_NODE_STATUS: ret = pm_get_node_status(pm_arg[0]); SMC_RET1(handle, (uint64_t)ret); case PM_GET_OP_CHARACTERISTIC: { uint32_t result; ret = pm_get_op_characteristic(pm_arg[0], pm_arg[1], &result); SMC_RET1(handle, (uint64_t)ret | ((uint64_t)result << 32)); } case PM_REGISTER_NOTIFIER: ret = pm_register_notifier(pm_arg[0], pm_arg[1], pm_arg[2], pm_arg[3]); SMC_RET1(handle, (uint64_t)ret); case PM_RESET_ASSERT: ret = pm_reset_assert(pm_arg[0], pm_arg[1]); SMC_RET1(handle, (uint64_t)ret); case PM_RESET_GET_STATUS: { uint32_t reset_status; ret = pm_reset_get_status(pm_arg[0], &reset_status); SMC_RET1(handle, (uint64_t)ret | ((uint64_t)reset_status << 32)); } /* PM memory access functions */ case PM_MMIO_WRITE: ret = pm_mmio_write(pm_arg[0], pm_arg[1], pm_arg[2]); SMC_RET1(handle, (uint64_t)ret); case PM_MMIO_READ: { uint32_t value; ret = pm_mmio_read(pm_arg[0], &value); SMC_RET1(handle, (uint64_t)ret | ((uint64_t)value) << 32); } default: WARN("Unimplemented PM Service Call: 0x%x\n", smc_fid); SMC_RET1(handle, SMC_UNK); } }