static int pmc_core_mphy_pg_sts_show(struct seq_file *s, void *unused) { struct pmc_dev *pmcdev = s->private; const struct pmc_bit_map *map = pmcdev->map->mphy_sts; u32 mphy_core_reg_low, mphy_core_reg_high; u32 val_low, val_high; int index, err = 0; if (pmcdev->pmc_xram_read_bit) { seq_puts(s, "Access denied: please disable PMC_READ_DISABLE setting in BIOS."); return 0; } mphy_core_reg_low = (SPT_PMC_MPHY_CORE_STS_0 << 16); mphy_core_reg_high = (SPT_PMC_MPHY_CORE_STS_1 << 16); mutex_lock(&pmcdev->lock); if (pmc_core_send_msg(&mphy_core_reg_low) != 0) { err = -EBUSY; goto out_unlock; } msleep(10); val_low = pmc_core_reg_read(pmcdev, SPT_PMC_MFPMC_OFFSET); if (pmc_core_send_msg(&mphy_core_reg_high) != 0) { err = -EBUSY; goto out_unlock; } msleep(10); val_high = pmc_core_reg_read(pmcdev, SPT_PMC_MFPMC_OFFSET); for (index = 0; map[index].name && index < 8; index++) { seq_printf(s, "%-32s\tState: %s\n", map[index].name, map[index].bit_mask & val_low ? "Not power gated" : "Power gated"); } for (index = 8; map[index].name; index++) { seq_printf(s, "%-32s\tState: %s\n", map[index].name, map[index].bit_mask & val_high ? "Not power gated" : "Power gated"); } out_unlock: mutex_unlock(&pmcdev->lock); return err; }
static ssize_t pmc_core_ltr_ignore_write(struct file *file, const char __user *userbuf, size_t count, loff_t *ppos) { struct pmc_dev *pmcdev = &pmc; const struct pmc_reg_map *map = pmcdev->map; u32 val, buf_size, fd; int err = 0; buf_size = count < 64 ? count : 64; mutex_lock(&pmcdev->lock); if (kstrtou32_from_user(userbuf, buf_size, 10, &val)) { err = -EFAULT; goto out_unlock; } if (val > NUM_IP_IGN_ALLOWED) { err = -EINVAL; goto out_unlock; } fd = pmc_core_reg_read(pmcdev, map->ltr_ignore_offset); fd |= (1U << val); pmc_core_reg_write(pmcdev, map->ltr_ignore_offset, fd); out_unlock: mutex_unlock(&pmcdev->lock); return err == 0 ? count : err; }
static int pmc_core_pll_show(struct seq_file *s, void *unused) { struct pmc_dev *pmcdev = s->private; const struct pmc_bit_map *map = pmcdev->map->pll_sts; u32 mphy_common_reg, val; int index, err = 0; if (pmcdev->pmc_xram_read_bit) { seq_puts(s, "Access denied: please disable PMC_READ_DISABLE setting in BIOS."); return 0; } mphy_common_reg = (SPT_PMC_MPHY_COM_STS_0 << 16); mutex_lock(&pmcdev->lock); if (pmc_core_send_msg(&mphy_common_reg) != 0) { err = -EBUSY; goto out_unlock; } /* Observed PMC HW response latency for MTPMC-MFPMC is ~10 ms */ msleep(10); val = pmc_core_reg_read(pmcdev, SPT_PMC_MFPMC_OFFSET); for (index = 0; map[index].name ; index++) { seq_printf(s, "%-32s\tState: %s\n", map[index].name, map[index].bit_mask & val ? "Active" : "Idle"); } out_unlock: mutex_unlock(&pmcdev->lock); return err; }
/* This function should return link status, 0 means ready */ static int pmc_core_mtpmc_link_status(void) { struct pmc_dev *pmcdev = &pmc; u32 value; value = pmc_core_reg_read(pmcdev, SPT_PMC_PM_STS_OFFSET); return value & BIT(SPT_PMC_MSG_FULL_STS_BIT); }
static int pmc_core_check_read_lock_bit(void) { struct pmc_dev *pmcdev = &pmc; u32 value; value = pmc_core_reg_read(pmcdev, pmcdev->map->pm_cfg_offset); return value & BIT(pmcdev->map->pm_read_disable_bit); }
/* This function should return link status, 0 means ready */ static int pmc_core_mtpmc_link_status(void) { struct pmc_dev *pmcdev = &pmc; u32 value; value = pmc_core_reg_read(pmcdev, SPT_PMC_PM_STS_OFFSET); return test_bit(SPT_PMC_MSG_FULL_STS_BIT, (unsigned long *)&value); }
static int pmc_core_check_read_lock_bit(void) { struct pmc_dev *pmcdev = &pmc; u32 value; value = pmc_core_reg_read(pmcdev, SPT_PMC_PM_CFG_OFFSET); return test_bit(SPT_PMC_READ_DISABLE_BIT, (unsigned long *)&value); }
static int pmc_core_dev_state_get(void *data, u64 *val) { struct pmc_dev *pmcdev = data; u32 value; value = pmc_core_reg_read(pmcdev, SPT_PMC_SLP_S0_RES_COUNTER_OFFSET); *val = pmc_core_adjust_slp_s0_step(value); return 0; }
static int pmc_core_dev_state_get(void *data, u64 *val) { struct pmc_dev *pmcdev = data; const struct pmc_reg_map *map = pmcdev->map; u32 value; value = pmc_core_reg_read(pmcdev, map->slp_s0_offset); *val = pmc_core_adjust_slp_s0_step(value); return 0; }
/** * intel_pmc_slp_s0_counter_read() - Read SLP_S0 residency. * @data: Out param that contains current SLP_S0 count. * * This API currently supports Intel Skylake SoC and Sunrise * Point Platform Controller Hub. Future platform support * should be added for platforms that support low power modes * beyond Package C10 state. * * SLP_S0_RESIDENCY counter counts in 100 us granularity per * step hence function populates the multiplied value in out * parameter @data. * * Return: an error code or 0 on success. */ int intel_pmc_slp_s0_counter_read(u32 *data) { struct pmc_dev *pmcdev = &pmc; u32 value; if (!pmcdev->has_slp_s0_res) return -EACCES; value = pmc_core_reg_read(pmcdev, SPT_PMC_SLP_S0_RES_COUNTER_OFFSET); *data = pmc_core_adjust_slp_s0_step(value); return 0; }
/** * intel_pmc_slp_s0_counter_read() - Read SLP_S0 residency. * @data: Out param that contains current SLP_S0 count. * * This API currently supports Intel Skylake SoC and Sunrise * Point Platform Controller Hub. Future platform support * should be added for platforms that support low power modes * beyond Package C10 state. * * SLP_S0_RESIDENCY counter counts in 100 us granularity per * step hence function populates the multiplied value in out * parameter @data. * * Return: an error code or 0 on success. */ int intel_pmc_slp_s0_counter_read(u32 *data) { struct pmc_dev *pmcdev = &pmc; const struct pmc_reg_map *map = pmcdev->map; u32 value; if (!pmcdev->has_slp_s0_res) return -EACCES; value = pmc_core_reg_read(pmcdev, map->slp_s0_offset); *data = pmc_core_adjust_slp_s0_step(value); return 0; }