/* safety check for chipinfo */ static int brcmf_sdio_chip_cichk(struct brcmf_chip *ci) { u8 core_idx; /* check RAM core presence for ARM CM3 core */ core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_ARM_CM3); if (BRCMF_MAX_CORENUM != core_idx) { core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_INTERNAL_MEM); if (BRCMF_MAX_CORENUM == core_idx) { brcmf_err("RAM core not provided with ARM CM3 core\n"); return -ENODEV; } } /* check RAM base for ARM CR4 core */ core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_ARM_CR4); if (BRCMF_MAX_CORENUM != core_idx) { if (ci->rambase == 0) { brcmf_err("RAM base not provided with ARM CR4 core\n"); return -ENOMEM; } } return 0; }
static void brcmf_sdio_ai_coredisable(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u8 idx; u32 regdata; idx = brcmf_sdio_chip_getinfidx(ci, coreid); /* if core is already in reset, just return */ regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_RESET_CTL, NULL); if ((regdata & BCMA_RESET_CTL_RESET) != 0) return; brcmf_sdio_regwl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, 0, NULL); regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, NULL); udelay(10); brcmf_sdio_regwl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_RESET_CTL, BCMA_RESET_CTL_RESET, NULL); udelay(1); }
static void brcmf_sdio_ai_resetcore(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u16 coreid, u32 pre_resetbits, u32 in_resetbits, u32 post_resetbits) { u8 idx; u32 regdata; u32 wrapbase; idx = brcmf_sdio_chip_getinfidx(ci, coreid); if (idx == BRCMF_MAX_CORENUM) return; wrapbase = ci->c_inf[idx].wrapbase; /* must disable first to work for arbitrary current core state */ brcmf_sdio_ai_coredisable(sdiodev, ci, coreid, pre_resetbits, in_resetbits); while (brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_RESET_CTL, NULL) & BCMA_RESET_CTL_RESET) { brcmf_sdiod_regwl(sdiodev, wrapbase + BCMA_RESET_CTL, 0, NULL); usleep_range(40, 60); } brcmf_sdiod_regwl(sdiodev, wrapbase + BCMA_IOCTL, post_resetbits | BCMA_IOCTL_CLK, NULL); regdata = brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_IOCTL, NULL); }
static void brcmf_sdio_ai_resetcore(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u8 idx; u32 regdata; idx = brcmf_sdio_chip_getinfidx(ci, coreid); /* must disable first to work for arbitrary current core state */ brcmf_sdio_ai_coredisable(sdiodev, ci, coreid); /* now do initialization sequence */ brcmf_sdio_regwl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, BCMA_IOCTL_FGC | BCMA_IOCTL_CLK, NULL); regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, NULL); brcmf_sdio_regwl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_RESET_CTL, 0, NULL); udelay(1); brcmf_sdio_regwl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, BCMA_IOCTL_CLK, NULL); regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, NULL); udelay(1); }
static void brcmf_sdio_sb_resetcore(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u32 regdata; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); /* * Must do the disable sequence first to work for * arbitrary current core state. */ brcmf_sdio_sb_coredisable(sdiodev, ci, coreid); /* * Now do the initialization sequence. * set reset while enabling the clock and * forcing them on throughout the core */ brcmf_sdcard_reg_write(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4, SSB_TMSLOW_FGC | SSB_TMSLOW_CLOCK | SSB_TMSLOW_RESET); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4); udelay(1); /* clear any serror */ regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatehigh), 4); if (regdata & SSB_TMSHIGH_SERR) brcmf_sdcard_reg_write(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatehigh), 4, 0); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbimstate), 4); if (regdata & (SSB_IMSTATE_IBE | SSB_IMSTATE_TO)) brcmf_sdcard_reg_write(sdiodev, CORE_SB(ci->c_inf[idx].base, sbimstate), 4, regdata & ~(SSB_IMSTATE_IBE | SSB_IMSTATE_TO)); /* clear reset and allow it to propagate throughout the core */ brcmf_sdcard_reg_write(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4, SSB_TMSLOW_FGC | SSB_TMSLOW_CLOCK); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4); udelay(1); /* leave clock enabled */ brcmf_sdcard_reg_write(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4, SSB_TMSLOW_CLOCK); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4); udelay(1); }
static u32 brcmf_sdio_ai_corerev(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); return (ci->c_inf[idx].cib & CIB_REV_MASK) >> CIB_REV_SHIFT; }
bool brcmf_sdio_chip_exit_download(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u32 rstvec) { u8 arm_core_idx; arm_core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_ARM_CM3); if (BRCMF_MAX_CORENUM != arm_core_idx) return brcmf_sdio_chip_cm3_exitdl(sdiodev, ci); return brcmf_sdio_chip_cr4_exitdl(sdiodev, ci, rstvec); }
void brcmf_sdio_chip_enter_download(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci) { u8 arm_core_idx; arm_core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_ARM_CM3); if (BRCMF_MAX_CORENUM != arm_core_idx) { brcmf_sdio_chip_cm3_enterdl(sdiodev, ci); return; } brcmf_sdio_chip_cr4_enterdl(sdiodev, ci); }
static u32 brcmf_sdio_sb_corerev(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u32 regdata; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbidhigh), 4); return SBCOREREV(regdata); }
static bool brcmf_sdio_sb_iscoreup(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u32 regdata; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); regdata = brcmf_sdcard_reg_read(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 4); regdata &= (SSB_TMSLOW_RESET | SSB_TMSLOW_REJECT | SSB_IMSTATE_REJECT | SSB_TMSLOW_CLOCK); return (SSB_TMSLOW_CLOCK == regdata); }
static bool brcmf_sdio_sb_iscoreup(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u16 coreid) { u32 regdata; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); if (idx == BRCMF_MAX_CORENUM) return false; regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), NULL); regdata &= (SSB_TMSLOW_RESET | SSB_TMSLOW_REJECT | SSB_IMSTATE_REJECT | SSB_TMSLOW_CLOCK); return (SSB_TMSLOW_CLOCK == regdata); }
static inline void brcmf_sdio_chip_cr4_enterdl(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci) { u8 idx; u32 regdata; u32 wrapbase; idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_ARM_CR4); if (idx == BRCMF_MAX_CORENUM) return; wrapbase = ci->c_inf[idx].wrapbase; regdata = brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_IOCTL, NULL); regdata &= ARMCR4_BCMA_IOCTL_CPUHALT; ci->resetcore(sdiodev, ci, BCMA_CORE_ARM_CR4, regdata, ARMCR4_BCMA_IOCTL_CPUHALT, ARMCR4_BCMA_IOCTL_CPUHALT); ci->resetcore(sdiodev, ci, BCMA_CORE_80211, D11_BCMA_IOCTL_PHYRESET | D11_BCMA_IOCTL_PHYCLOCKEN, D11_BCMA_IOCTL_PHYCLOCKEN, D11_BCMA_IOCTL_PHYCLOCKEN); }
static bool brcmf_sdio_ai_iscoreup(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u32 regdata; u8 idx; bool ret; idx = brcmf_sdio_chip_getinfidx(ci, coreid); regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, NULL); ret = (regdata & (BCMA_IOCTL_FGC | BCMA_IOCTL_CLK)) == BCMA_IOCTL_CLK; regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_RESET_CTL, NULL); ret = ret && ((regdata & BCMA_RESET_CTL_RESET) == 0); return ret; }
static bool brcmf_sdio_chip_cm3_exitdl(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci) { u8 core_idx; u32 reg_addr; if (!ci->iscoreup(sdiodev, ci, BCMA_CORE_INTERNAL_MEM)) { brcmf_err("SOCRAM core is down after reset?\n"); return false; } /* clear all interrupts */ core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_SDIO_DEV); reg_addr = ci->c_inf[core_idx].base; reg_addr += offsetof(struct sdpcmd_regs, intstatus); brcmf_sdiod_regwl(sdiodev, reg_addr, 0xFFFFFFFF, NULL); ci->resetcore(sdiodev, ci, BCMA_CORE_ARM_CM3, 0, 0, 0); return true; }
static void brcmf_sdio_ai_coredisable(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u16 coreid, u32 pre_resetbits, u32 in_resetbits) { u8 idx; u32 regdata; u32 wrapbase; idx = brcmf_sdio_chip_getinfidx(ci, coreid); if (idx == BRCMF_MAX_CORENUM) return; wrapbase = ci->c_inf[idx].wrapbase; /* if core is already in reset, just return */ regdata = brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_RESET_CTL, NULL); if ((regdata & BCMA_RESET_CTL_RESET) != 0) return; /* configure reset */ brcmf_sdiod_regwl(sdiodev, wrapbase + BCMA_IOCTL, pre_resetbits | BCMA_IOCTL_FGC | BCMA_IOCTL_CLK, NULL); regdata = brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_IOCTL, NULL); /* put in reset */ brcmf_sdiod_regwl(sdiodev, wrapbase + BCMA_RESET_CTL, BCMA_RESET_CTL_RESET, NULL); usleep_range(10, 20); /* wait till reset is 1 */ SPINWAIT(brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_RESET_CTL, NULL) != BCMA_RESET_CTL_RESET, 300); /* post reset configure */ brcmf_sdiod_regwl(sdiodev, wrapbase + BCMA_IOCTL, pre_resetbits | BCMA_IOCTL_FGC | BCMA_IOCTL_CLK, NULL); regdata = brcmf_sdiod_regrl(sdiodev, wrapbase + BCMA_IOCTL, NULL); }
static bool brcmf_sdio_chip_cr4_exitdl(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u32 rstvec) { u8 core_idx; u32 reg_addr; /* clear all interrupts */ core_idx = brcmf_sdio_chip_getinfidx(ci, BCMA_CORE_SDIO_DEV); reg_addr = ci->c_inf[core_idx].base; reg_addr += offsetof(struct sdpcmd_regs, intstatus); brcmf_sdiod_regwl(sdiodev, reg_addr, 0xFFFFFFFF, NULL); /* Write reset vector to address 0 */ brcmf_sdiod_ramrw(sdiodev, true, 0, (void *)&rstvec, sizeof(rstvec)); /* restore ARM */ ci->resetcore(sdiodev, ci, BCMA_CORE_ARM_CR4, ARMCR4_BCMA_IOCTL_CPUHALT, 0, 0); return true; }
static void brcmf_sdio_sb_coredisable(struct brcmf_sdio_dev *sdiodev, struct chip_info *ci, u16 coreid) { u32 regdata, base; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); base = ci->c_inf[idx].base; regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatelow), NULL); if (regdata & SSB_TMSLOW_RESET) return; regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatelow), NULL); if ((regdata & SSB_TMSLOW_CLOCK) != 0) { /* * set target reject and spin until busy is clear * (preserve core-specific bits) */ regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatelow), NULL); brcmf_sdio_regwl(sdiodev, CORE_SB(base, sbtmstatelow), regdata | SSB_TMSLOW_REJECT, NULL); regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatelow), NULL); udelay(1); SPINWAIT((brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatehigh), NULL) & SSB_TMSHIGH_BUSY), 100000); regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatehigh), NULL); if (regdata & SSB_TMSHIGH_BUSY) brcmf_dbg(ERROR, "core state still busy\n"); regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbidlow), NULL); if (regdata & SSB_IDLOW_INITIATOR) { regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbimstate), NULL); regdata |= SSB_IMSTATE_REJECT; brcmf_sdio_regwl(sdiodev, CORE_SB(base, sbimstate), regdata, NULL); regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbimstate), NULL); udelay(1); SPINWAIT((brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbimstate), NULL) & SSB_IMSTATE_BUSY), 100000); } /* set reset and reject while enabling the clocks */ regdata = SSB_TMSLOW_FGC | SSB_TMSLOW_CLOCK | SSB_TMSLOW_REJECT | SSB_TMSLOW_RESET; brcmf_sdio_regwl(sdiodev, CORE_SB(base, sbtmstatelow), regdata, NULL); regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbtmstatelow), NULL); udelay(10); /* clear the initiator reject bit */ regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbidlow), NULL); if (regdata & SSB_IDLOW_INITIATOR) { regdata = brcmf_sdio_regrl(sdiodev, CORE_SB(base, sbimstate), NULL); regdata &= ~SSB_IMSTATE_REJECT; brcmf_sdio_regwl(sdiodev, CORE_SB(base, sbimstate), regdata, NULL); } } /* leave reset and reject asserted */ brcmf_sdio_regwl(sdiodev, CORE_SB(base, sbtmstatelow), (SSB_TMSLOW_REJECT | SSB_TMSLOW_RESET), NULL); udelay(1); }
static void brcmf_sdio_sb_resetcore(struct brcmf_sdio_dev *sdiodev, struct brcmf_chip *ci, u16 coreid, u32 pre_resetbits, u32 in_resetbits, u32 post_resetbits) { u32 regdata; u8 idx; idx = brcmf_sdio_chip_getinfidx(ci, coreid); if (idx == BRCMF_MAX_CORENUM) return; /* * Must do the disable sequence first to work for * arbitrary current core state. */ brcmf_sdio_sb_coredisable(sdiodev, ci, coreid, pre_resetbits, in_resetbits); /* * Now do the initialization sequence. * set reset while enabling the clock and * forcing them on throughout the core */ brcmf_sdiod_regwl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), SSB_TMSLOW_FGC | SSB_TMSLOW_CLOCK | SSB_TMSLOW_RESET, NULL); regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), NULL); udelay(1); /* clear any serror */ regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatehigh), NULL); if (regdata & SSB_TMSHIGH_SERR) brcmf_sdiod_regwl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatehigh), 0, NULL); regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbimstate), NULL); if (regdata & (SSB_IMSTATE_IBE | SSB_IMSTATE_TO)) brcmf_sdiod_regwl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbimstate), regdata & ~(SSB_IMSTATE_IBE | SSB_IMSTATE_TO), NULL); /* clear reset and allow it to propagate throughout the core */ brcmf_sdiod_regwl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), SSB_TMSLOW_FGC | SSB_TMSLOW_CLOCK, NULL); regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), NULL); udelay(1); /* leave clock enabled */ brcmf_sdiod_regwl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), SSB_TMSLOW_CLOCK, NULL); regdata = brcmf_sdiod_regrl(sdiodev, CORE_SB(ci->c_inf[idx].base, sbtmstatelow), NULL); udelay(1); }