10257 lines
270 KiB
C
Executable File
10257 lines
270 KiB
C
Executable File
/*
|
|
* Misc utility routines for accessing chip-specific features
|
|
* of the SiliconBackplane-based Broadcom chips.
|
|
*
|
|
* Copyright (C) 2022, Broadcom.
|
|
*
|
|
* Unless you and Broadcom execute a separate written software license
|
|
* agreement governing use of this software, this software is licensed to you
|
|
* under the terms of the GNU General Public License version 2 (the "GPL"),
|
|
* available at http://www.broadcom.com/licenses/GPLv2.php, with the
|
|
* following added to such license:
|
|
*
|
|
* As a special exception, the copyright holders of this software give you
|
|
* permission to link this software with independent modules, and to copy and
|
|
* distribute the resulting executable under terms of your choice, provided that
|
|
* you also meet, for each linked independent module, the terms and conditions of
|
|
* the license of that module. An independent module is a module which is not
|
|
* derived from this software. The special exception does not apply to any
|
|
* modifications of the software.
|
|
*
|
|
*
|
|
* <<Broadcom-WL-IPTag/Dual:>>
|
|
*/
|
|
|
|
#include <typedefs.h>
|
|
#include <bcmdefs.h>
|
|
#include <osl.h>
|
|
#include <bcmutils.h>
|
|
#include <siutils.h>
|
|
#include <bcmdevs.h>
|
|
#include <hndsoc.h>
|
|
#include <sbchipc.h>
|
|
#include <sbgci.h>
|
|
#ifndef BCMSDIO
|
|
#include <pcie_core.h>
|
|
#endif
|
|
#if !defined(BCMDONGLEHOST)
|
|
#include <pci_core.h>
|
|
#include <nicpci.h>
|
|
#include <bcmnvram.h>
|
|
#include <bcmsrom.h>
|
|
#include <hndtcam.h>
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
#ifdef BCMPCIEDEV
|
|
#include <pcieregsoffs.h>
|
|
#include <pciedev.h>
|
|
#endif /* BCMPCIEDEV */
|
|
#include <pcicfg.h>
|
|
#include <sbpcmcia.h>
|
|
#include <sbsysmem.h>
|
|
#include <sbsocram.h>
|
|
#if defined(BCMECICOEX) || !defined(BCMDONGLEHOST)
|
|
#include <bcmotp.h>
|
|
#endif /* BCMECICOEX || !BCMDONGLEHOST */
|
|
#ifdef BCMSDIO
|
|
#include <bcmsdh.h>
|
|
#include <sdio.h>
|
|
#include <sbsdio.h>
|
|
#include <sbhnddma.h>
|
|
#include <sbsdpcmdev.h>
|
|
#include <bcmsdpcm.h>
|
|
#endif /* BCMSDIO */
|
|
#include <hndpmu.h>
|
|
#ifdef BCMSPI
|
|
#include <spid.h>
|
|
#endif /* BCMSPI */
|
|
#if !defined(BCMDONGLEHOST) && !defined(BCM_BOOTLOADER) && defined(SR_ESSENTIALS)
|
|
#include <saverestore.h>
|
|
#endif
|
|
#include <dhd_config.h>
|
|
|
|
#ifdef BCM_SDRBL
|
|
#include <hndcpu.h>
|
|
#endif /* BCM_SDRBL */
|
|
#ifdef HNDGCI
|
|
#include <hndgci.h>
|
|
#endif /* HNDGCI */
|
|
#ifdef DONGLEBUILD
|
|
#include <hnd_gci.h>
|
|
#endif /* DONGLEBUILD */
|
|
#include <hndlhl.h>
|
|
#include <hndoobr.h>
|
|
#include <lpflags.h>
|
|
#ifdef BCM_SFLASH
|
|
#include <sflash.h>
|
|
#endif
|
|
#ifdef BCM_SH_SFLASH
|
|
#include <sh_sflash.h>
|
|
#endif
|
|
#ifdef BCMGCISHM
|
|
#include <hnd_gcishm.h>
|
|
#endif
|
|
#include "siutils_priv.h"
|
|
#include "sbhndarm.h"
|
|
#include <hndchipc.h>
|
|
|
|
#ifdef SECI_UART
|
|
/* Defines the set of GPIOs to be used for SECI UART if not specified in NVRAM */
|
|
/* For further details on each ppin functionality please refer to PINMUX table in
|
|
* Top level architecture of BCMXXXX Chip
|
|
*/
|
|
#define DEFAULT_SECI_UART_PINMUX 0x08090a0b
|
|
static bool force_seci_clk = 0;
|
|
#endif /* SECI_UART */
|
|
|
|
#define XTAL_FREQ_26000KHZ 26000
|
|
#define XTAL_FREQ_59970KHZ 59970
|
|
#define WCI2_UART_RX_BUF_SIZE 64
|
|
|
|
/**
|
|
* A set of PMU registers is clocked in the ILP domain, which has an implication on register write
|
|
* behavior: if such a register is written, it takes multiple ILP clocks for the PMU block to absorb
|
|
* the write. During that time the 'SlowWritePending' bit in the PMUStatus register is set.
|
|
*/
|
|
#define PMUREGS_ILP_SENSITIVE(regoff) \
|
|
((regoff) == OFFSETOF(pmuregs_t, pmutimer) || \
|
|
(regoff) == OFFSETOF(pmuregs_t, pmuwatchdog) || \
|
|
(regoff) == OFFSETOF(pmuregs_t, res_req_timer))
|
|
|
|
#define CHIPCREGS_ILP_SENSITIVE(regoff) \
|
|
((regoff) == OFFSETOF(chipcregs_t, pmutimer) || \
|
|
(regoff) == OFFSETOF(chipcregs_t, pmuwatchdog) || \
|
|
(regoff) == OFFSETOF(chipcregs_t, res_req_timer))
|
|
|
|
#define GCI_FEM_CTRL_WAR 0x11111111
|
|
|
|
#ifndef AXI_TO_VAL
|
|
#define AXI_TO_VAL 19
|
|
#endif /* AXI_TO_VAL */
|
|
|
|
#ifndef AXI_TO_VAL_25
|
|
/*
|
|
* Increase BP timeout for fast clock and short PCIe timeouts
|
|
* New timeout: 2 ** 25 cycles
|
|
*/
|
|
#define AXI_TO_VAL_25 25
|
|
#endif /* AXI_TO_VAL_25 */
|
|
|
|
#define si_srpwr_domain_mask(rval, mask) \
|
|
(((rval) >> SRPWR_STATUS_SHIFT) & (mask))
|
|
|
|
/* local prototypes */
|
|
#if !defined(BCMDONGLEHOST)
|
|
static void si_43012_lp_enable(si_t *sih);
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
static int32 si_alloc_wrapper(si_info_t *sii);
|
|
static si_info_t *si_doattach(si_info_t *sii, uint devid, osl_t *osh, volatile void *regs,
|
|
uint bustype, void *sdh, char **vars, uint *varsz);
|
|
static bool si_buscore_prep(si_info_t *sii, uint bustype, uint devid, void *sdh);
|
|
static bool si_buscore_setup(si_info_t *sii, chipcregs_t *cc, uint bustype, uint32 savewin,
|
|
uint *origidx, volatile const void *regs);
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
static void si_nvram_process(si_info_t *sii, char *pvars);
|
|
|
|
/* dev path concatenation util */
|
|
static char *si_devpathvar(const si_t *sih, char *var, int len, const char *name);
|
|
static char *si_pcie_devpathvar(const si_t *sih, char *var, int len, const char *name);
|
|
static bool _si_clkctl_cc(si_info_t *sii, uint mode);
|
|
static bool si_ispcie(const si_info_t *sii);
|
|
static uint sysmem_banksize(const si_info_t *sii, sysmemregs_t *r, uint8 idx);
|
|
static uint socram_banksize(const si_info_t *sii, sbsocramregs_t *r, uint8 idx, uint8 mtype);
|
|
static uint8 si_gci_get_chipctrlreg_ringidx_base8(uint32 pin, uint32 *regidx, uint32 *pos);
|
|
static void si_gci_gpio_chipcontrol(si_t *si, uint8 gpoi, uint8 opt);
|
|
static void si_gci_enable_gpioint(si_t *sih, bool enable);
|
|
#if defined(BCMECICOEX) || defined(SECI_UART)
|
|
static chipcregs_t * seci_set_core(si_t *sih, uint32 *origidx, bool *fast);
|
|
#endif
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
static void si_gci_get_chipctrlreg_ringidx_base4(uint32 pin, uint32 *regidx, uint32 *pos);
|
|
static bool si_pmu_is_ilp_sensitive(uint32 idx, uint regoff);
|
|
|
|
static void si_oob_war_BT_F1(si_t *sih);
|
|
|
|
#if defined(DONGLEBUILD)
|
|
#if !defined(NVSRCX)
|
|
static char * si_getkvars(void);
|
|
static int si_getkvarsz(void);
|
|
#endif
|
|
#endif /* DONGLEBUILD */
|
|
|
|
#if defined(BCMLTECOEX) && !defined(WLTEST)
|
|
static void si_wci2_rxfifo_intr_handler_process(si_t *sih, uint32 intstatus);
|
|
#endif /* BCMLTECOEX && !WLTEST */
|
|
|
|
/* global variable to indicate reservation/release of gpio's */
|
|
static uint32 si_gpioreservation = 0;
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* global variable to indicate GCI reset is done */
|
|
static bool gci_reset_done = FALSE;
|
|
#endif
|
|
/* global flag to prevent shared resources from being initialized multiple times in si_attach() */
|
|
static bool si_onetimeinit = FALSE;
|
|
|
|
#ifdef SR_DEBUG
|
|
static const uint32 si_power_island_test_array[] = {
|
|
0x0000, 0x0001, 0x0010, 0x0011,
|
|
0x0100, 0x0101, 0x0110, 0x0111,
|
|
0x1000, 0x1001, 0x1010, 0x1011,
|
|
0x1100, 0x1101, 0x1110, 0x1111
|
|
};
|
|
#endif /* SR_DEBUG */
|
|
|
|
/* 4360 pcie2 WAR */
|
|
int do_4360_pcie2_war = 0;
|
|
|
|
/* global kernel resource */
|
|
static si_info_t ksii;
|
|
static si_cores_info_t ksii_cores_info;
|
|
|
|
#ifndef BCMDONGLEHOST
|
|
static const char rstr_rmin[] = "rmin";
|
|
static const char rstr_rmax[] = "rmax";
|
|
|
|
static const char rstr_lhl_ps_mode[] = "lhl_ps_mode";
|
|
static const char rstr_ext_wakeup_dis[] = "ext_wakeup_dis";
|
|
#if defined(BCMSRTOPOFF) && !defined(BCMSRTOPOFF_DISABLED)
|
|
static const char rstr_srtopoff_enab[] = "srtopoff_enab";
|
|
#endif
|
|
#endif /* BCMDONGLEHOST */
|
|
|
|
static uint32 wd_msticks; /**< watchdog timer ticks normalized to ms */
|
|
|
|
#ifdef DONGLEBUILD
|
|
/**
|
|
* As si_kattach goes thru full srom initialisation same can be used
|
|
* for all subsequent calls
|
|
*/
|
|
#if !defined(NVSRCX)
|
|
static char *
|
|
si_getkvars(void)
|
|
{
|
|
return (ksii.vars);
|
|
}
|
|
|
|
static int
|
|
si_getkvarsz(void)
|
|
{
|
|
return (ksii.varsz);
|
|
}
|
|
#endif /* !defined(NVSRCX) */
|
|
#endif /* DONGLEBUILD */
|
|
|
|
/** Returns the backplane address of the chipcommon core for a particular chip */
|
|
uint32
|
|
si_enum_base(uint devid)
|
|
{
|
|
return SI_ENUM_BASE_DEFAULT;
|
|
}
|
|
|
|
/**
|
|
* Allocate an si handle. This function may be called multiple times.
|
|
*
|
|
* devid - pci device id (used to determine chip#)
|
|
* osh - opaque OS handle
|
|
* regs - virtual address of initial core registers
|
|
* bustype - pci/sb/sdio/etc
|
|
* vars - pointer to a to-be created pointer area for "environment" variables. Some callers of this
|
|
* function set 'vars' to NULL, making dereferencing of this parameter undesired.
|
|
* varsz - pointer to int to return the size of the vars
|
|
*/
|
|
si_t *
|
|
si_attach(uint devid, osl_t *osh, volatile void *regs,
|
|
uint bustype, void *sdh, char **vars, uint *varsz)
|
|
{
|
|
si_info_t *sii;
|
|
|
|
SI_MSG_DBG_REG(("%s: Enter\n", __FUNCTION__));
|
|
/* alloc si_info_t */
|
|
/* freed after ucode download for firmware builds */
|
|
if ((sii = MALLOCZ_NOPERSIST(osh, sizeof(si_info_t))) == NULL) {
|
|
SI_ERROR(("si_attach: malloc failed! malloced %d bytes\n", MALLOCED(osh)));
|
|
return (NULL);
|
|
}
|
|
|
|
#ifdef BCMDVFS
|
|
if (BCMDVFS_ENAB() && si_dvfs_info_init((si_t *)sii, osh) == NULL) {
|
|
SI_ERROR(("si_dvfs_info_init failed\n"));
|
|
return (NULL);
|
|
}
|
|
#endif /* BCMDVFS */
|
|
|
|
if (si_doattach(sii, devid, osh, regs, bustype, sdh, vars, varsz) == NULL) {
|
|
MFREE(osh, sii, sizeof(si_info_t));
|
|
return (NULL);
|
|
}
|
|
sii->vars = vars ? *vars : NULL;
|
|
sii->varsz = varsz ? *varsz : 0;
|
|
|
|
#if defined(BCM_SH_SFLASH) && !defined(BCM_SH_SFLASH_DISABLED)
|
|
sh_sflash_attach(osh, (si_t *)sii);
|
|
#endif
|
|
SI_MSG_DBG_REG(("%s: Exit\n", __FUNCTION__));
|
|
return (si_t *)sii;
|
|
}
|
|
|
|
/** generic kernel variant of si_attach(). Is not called for Linux WLAN NIC builds. */
|
|
si_t *
|
|
si_kattach(osl_t *osh)
|
|
{
|
|
static bool ksii_attached = FALSE;
|
|
si_cores_info_t *cores_info;
|
|
|
|
if (!ksii_attached) {
|
|
void *regs = NULL;
|
|
const uint device_id = BCM4710_DEVICE_ID; // pick an arbitrary default device_id
|
|
|
|
regs = REG_MAP(si_enum_base(device_id), SI_CORE_SIZE); // map physical to virtual
|
|
cores_info = (si_cores_info_t *)&ksii_cores_info;
|
|
ksii.cores_info = cores_info;
|
|
|
|
/* Use osh as the deciding factor if the memory management
|
|
* system has been initialized. Pass non-NULL vars & varsz only
|
|
* if memory management has been initialized. Otherwise MALLOC()
|
|
* will fail/crash.
|
|
*/
|
|
#if defined(BCMDONGLEHOST)
|
|
ASSERT(osh);
|
|
#endif
|
|
if (si_doattach(&ksii, device_id, osh, regs,
|
|
SI_BUS, NULL,
|
|
osh != SI_OSH ? &(ksii.vars) : NULL,
|
|
osh != SI_OSH ? &(ksii.varsz) : NULL) == NULL) {
|
|
SI_ERROR(("si_kattach: si_doattach failed\n"));
|
|
REG_UNMAP(regs);
|
|
return NULL;
|
|
}
|
|
REG_UNMAP(regs);
|
|
|
|
/* save ticks normalized to ms for si_watchdog_ms() */
|
|
if (PMUCTL_ENAB(&ksii.pub)) {
|
|
/* based on 32KHz ILP clock */
|
|
wd_msticks = 32;
|
|
} else {
|
|
#if !defined(BCMDONGLEHOST)
|
|
if (CCREV(ksii.pub.ccrev) < 18)
|
|
wd_msticks = si_clock(&ksii.pub) / 1000;
|
|
else
|
|
wd_msticks = si_alp_clock(&ksii.pub) / 1000;
|
|
#else
|
|
wd_msticks = ALP_CLOCK / 1000;
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
}
|
|
|
|
ksii_attached = TRUE;
|
|
SI_MSG(("si_kattach done. ccrev = %d, wd_msticks = %d\n",
|
|
CCREV(ksii.pub.ccrev), wd_msticks));
|
|
}
|
|
|
|
return &ksii.pub;
|
|
}
|
|
|
|
static bool
|
|
si_buscore_prep(si_info_t *sii, uint bustype, uint devid, void *sdh)
|
|
{
|
|
BCM_REFERENCE(sdh);
|
|
BCM_REFERENCE(devid);
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* kludge to enable the clock on the 4306 which lacks a slowclock */
|
|
if (BUSTYPE(bustype) == PCI_BUS && !si_ispcie(sii))
|
|
si_clkctl_xtal(&sii->pub, XTAL|PLL, ON);
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#if defined(BCMSDIO) && defined(BCMDONGLEHOST) && !defined(BCMSDIOLITE)
|
|
/* PR 39902, 43618, 44891, 41539 -- avoid backplane accesses that may
|
|
* cause SDIO clock requests before a stable ALP clock. Originally had
|
|
* this later (just before srom_var_init() below) to guarantee ALP for
|
|
* CIS read, but due to these PRs moving it here before backplane use.
|
|
*/
|
|
/* As it precedes any backplane access, can't check chipid; but may
|
|
* be able to qualify with devid if underlying SDIO allows. But should
|
|
* be ok for all our SDIO (4318 doesn't support clock and pullup regs,
|
|
* but the access attempts don't seem to hurt.) Might elimiante the
|
|
* the need for ALP for CIS at all if underlying SDIO uses CMD53...
|
|
*/
|
|
if (BUSTYPE(bustype) == SDIO_BUS) {
|
|
int err;
|
|
uint8 clkset;
|
|
|
|
/* Try forcing SDIO core to do ALPAvail request only */
|
|
clkset = SBSDIO_FORCE_HW_CLKREQ_OFF | SBSDIO_ALP_AVAIL_REQ;
|
|
bcmsdh_cfg_write(sdh, SDIO_FUNC_1, SBSDIO_FUNC1_CHIPCLKCSR, clkset, &err);
|
|
if (!err) {
|
|
uint8 clkval;
|
|
|
|
/* If register supported, wait for ALPAvail and then force ALP */
|
|
clkval = bcmsdh_cfg_read(sdh, SDIO_FUNC_1, SBSDIO_FUNC1_CHIPCLKCSR, NULL);
|
|
if ((clkval & ~SBSDIO_AVBITS) == clkset) {
|
|
SPINWAIT(((clkval = bcmsdh_cfg_read(sdh, SDIO_FUNC_1,
|
|
SBSDIO_FUNC1_CHIPCLKCSR, NULL)), !SBSDIO_ALPAV(clkval)),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
if (!SBSDIO_ALPAV(clkval)) {
|
|
SI_ERROR(("timeout on ALPAV wait, clkval 0x%02x\n",
|
|
clkval));
|
|
return FALSE;
|
|
}
|
|
clkset = SBSDIO_FORCE_HW_CLKREQ_OFF | SBSDIO_FORCE_ALP;
|
|
bcmsdh_cfg_write(sdh, SDIO_FUNC_1, SBSDIO_FUNC1_CHIPCLKCSR,
|
|
clkset, &err);
|
|
/* PR 40613: account for possible ALP delay */
|
|
OSL_DELAY(65);
|
|
}
|
|
}
|
|
|
|
/* Also, disable the extra SDIO pull-ups */
|
|
bcmsdh_cfg_write(sdh, SDIO_FUNC_1, SBSDIO_FUNC1_SDIOPULLUP, 0, NULL);
|
|
}
|
|
|
|
#ifdef BCMSPI
|
|
/* Avoid backplane accesses before wake-wlan (i.e. htavail) for spi.
|
|
* F1 read accesses may return correct data but with data-not-available dstatus bit set.
|
|
*/
|
|
if (BUSTYPE(bustype) == SPI_BUS) {
|
|
|
|
int err;
|
|
uint32 regdata;
|
|
/* wake up wlan function :WAKE_UP goes as HT_AVAIL request in hardware */
|
|
regdata = bcmsdh_cfg_read_word(sdh, SDIO_FUNC_0, SPID_CONFIG, NULL);
|
|
SI_MSG(("F0 REG0 rd = 0x%x\n", regdata));
|
|
regdata |= WAKE_UP;
|
|
|
|
bcmsdh_cfg_write_word(sdh, SDIO_FUNC_0, SPID_CONFIG, regdata, &err);
|
|
|
|
/* It takes time for wakeup to take effect. */
|
|
OSL_DELAY(100000);
|
|
}
|
|
#endif /* BCMSPI */
|
|
#endif /* BCMSDIO && BCMDONGLEHOST && !BCMSDIOLITE */
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* note: this function is used by dhd */
|
|
uint32
|
|
si_get_pmu_reg_addr(si_t *sih, uint32 offset)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
uint32 pmuaddr = INVALID_ADDR;
|
|
uint origidx = 0;
|
|
|
|
SI_MSG(("si_get_pmu_reg_addr: pmu access, offset: %x\n", offset));
|
|
if (!(sii->pub.cccaps & CC_CAP_PMU)) {
|
|
goto done;
|
|
}
|
|
if (AOB_ENAB(&sii->pub)) {
|
|
uint pmucoreidx;
|
|
pmuregs_t *pmu;
|
|
SI_MSG(("si_get_pmu_reg_addr: AOBENAB: %x\n", offset));
|
|
origidx = sii->curidx;
|
|
pmucoreidx = si_findcoreidx(&sii->pub, PMU_CORE_ID, 0);
|
|
pmu = si_setcoreidx(&sii->pub, pmucoreidx);
|
|
/* note: this function is used by dhd and possible 64 bit compilation needs
|
|
* a cast to (unsigned long) for avoiding a compilation error.
|
|
*/
|
|
pmuaddr = (uint32)(uintptr)((volatile uint8*)pmu + offset);
|
|
si_setcoreidx(sih, origidx);
|
|
} else
|
|
pmuaddr = SI_ENUM_BASE(sih) + offset;
|
|
|
|
done:
|
|
SI_MSG(("%s: addrRET: %x\n", __FUNCTION__, pmuaddr));
|
|
return pmuaddr;
|
|
}
|
|
|
|
static bool
|
|
si_buscore_setup(si_info_t *sii, chipcregs_t *cc, uint bustype, uint32 savewin,
|
|
uint *origidx, volatile const void *regs)
|
|
{
|
|
const si_cores_info_t *cores_info = sii->cores_info;
|
|
bool pci, pcie, pcie_gen2 = FALSE;
|
|
uint i;
|
|
uint pciidx, pcieidx, pcirev, pcierev;
|
|
|
|
#if defined(AXI_TIMEOUTS_NIC) || defined(AXI_TIMEOUTS)
|
|
/* first, enable backplane timeouts */
|
|
si_slave_wrapper_add(&sii->pub);
|
|
#endif
|
|
sii->curidx = 0;
|
|
|
|
cc = si_setcoreidx(&sii->pub, SI_CC_IDX);
|
|
ASSERT((uintptr)cc);
|
|
|
|
/* get chipcommon rev */
|
|
sii->pub.ccrev = (int)si_corerev(&sii->pub);
|
|
|
|
/* get chipcommon chipstatus */
|
|
if (CCREV(sii->pub.ccrev) >= 11)
|
|
sii->pub.chipst = R_REG(sii->osh, &cc->chipstatus);
|
|
|
|
/* get chipcommon capabilites */
|
|
sii->pub.cccaps = R_REG(sii->osh, &cc->capabilities);
|
|
/* get chipcommon extended capabilities */
|
|
|
|
if (CCREV(sii->pub.ccrev) >= 35) /* PR77565 */
|
|
sii->pub.cccaps_ext = R_REG(sii->osh, &cc->capabilities_ext);
|
|
|
|
/* get pmu rev and caps */
|
|
if (sii->pub.cccaps & CC_CAP_PMU) {
|
|
if (AOB_ENAB(&sii->pub)) {
|
|
uint pmucoreidx;
|
|
pmuregs_t *pmu;
|
|
|
|
pmucoreidx = si_findcoreidx(&sii->pub, PMU_CORE_ID, 0);
|
|
if (!GOODIDX(pmucoreidx, sii->numcores)) {
|
|
SI_ERROR(("si_buscore_setup: si_findcoreidx failed\n"));
|
|
return FALSE;
|
|
}
|
|
|
|
pmu = si_setcoreidx(&sii->pub, pmucoreidx);
|
|
sii->pub.pmucaps = R_REG(sii->osh, &pmu->pmucapabilities);
|
|
si_setcoreidx(&sii->pub, SI_CC_IDX);
|
|
|
|
sii->pub.gcirev = si_corereg(&sii->pub, GCI_CORE_IDX(&sii->pub),
|
|
GCI_OFFSETOF(&sii->pub, gci_corecaps0), 0, 0) & GCI_CAP0_REV_MASK;
|
|
|
|
if (GCIREV(sii->pub.gcirev) >= 9) {
|
|
#if !defined(BCMQT) && !defined(BCMFPGA)
|
|
sii->pub.lhlrev = si_corereg(&sii->pub, GCI_CORE_IDX(&sii->pub),
|
|
OFFSETOF(gciregs_t, lhl_core_capab_adr), 0, 0) &
|
|
LHL_CAP_REV_MASK;
|
|
#endif /* !defined(BCMQT && !defined(BCMFPGA */
|
|
} else {
|
|
sii->pub.lhlrev = NOREV;
|
|
}
|
|
} else
|
|
sii->pub.pmucaps = R_REG(sii->osh, &cc->pmucapabilities);
|
|
|
|
sii->pub.pmurev = sii->pub.pmucaps & PCAP_REV_MASK;
|
|
}
|
|
|
|
SI_MSG(("Chipc: rev %d, caps 0x%x, chipst 0x%x pmurev %d, pmucaps 0x%x\n",
|
|
CCREV(sii->pub.ccrev), sii->pub.cccaps, sii->pub.chipst, sii->pub.pmurev,
|
|
sii->pub.pmucaps));
|
|
|
|
/* figure out bus/orignal core idx */
|
|
/* note for PCI_BUS the buscoretype variable is setup in ai_scan() */
|
|
if (BUSTYPE(sii->pub.bustype) != PCI_BUS) {
|
|
sii->pub.buscoretype = NODEV_CORE_ID;
|
|
}
|
|
sii->pub.buscorerev = NOREV;
|
|
sii->pub.buscoreidx = BADIDX;
|
|
|
|
pci = pcie = FALSE;
|
|
pcirev = pcierev = NOREV;
|
|
pciidx = pcieidx = BADIDX;
|
|
|
|
/* This loop can be optimized */
|
|
for (i = 0; i < sii->numcores; i++) {
|
|
uint cid, crev;
|
|
|
|
si_setcoreidx(&sii->pub, i);
|
|
cid = si_coreid(&sii->pub);
|
|
crev = si_corerev(&sii->pub);
|
|
|
|
/* Display cores found */
|
|
if (CHIPTYPE(sii->pub.socitype) != SOCI_NCI) {
|
|
SI_VMSG(("CORE[%d]: id 0x%x rev %d base 0x%x size:%x regs 0x%p\n",
|
|
i, cid, crev, cores_info->coresba[i], cores_info->coresba_size[i],
|
|
OSL_OBFUSCATE_BUF(cores_info->regs[i])));
|
|
}
|
|
|
|
if (BUSTYPE(bustype) == SI_BUS) {
|
|
/* now look at the chipstatus register to figure the pacakge */
|
|
/* this shoudl be a general change to cover all teh chips */
|
|
/* this also shoudl validate the build where the dongle is built */
|
|
/* for SDIO but downloaded on PCIE dev */
|
|
#ifdef BCMPCIEDEV_ENABLED
|
|
if (cid == PCIE2_CORE_ID) {
|
|
pcieidx = i;
|
|
pcierev = crev;
|
|
pcie = TRUE;
|
|
pcie_gen2 = TRUE;
|
|
}
|
|
#endif
|
|
/* rest fill it up here */
|
|
|
|
} else if (BUSTYPE(bustype) == PCI_BUS) {
|
|
if (cid == PCI_CORE_ID) {
|
|
pciidx = i;
|
|
pcirev = crev;
|
|
pci = TRUE;
|
|
} else if ((cid == PCIE_CORE_ID) || (cid == PCIE2_CORE_ID)) {
|
|
pcieidx = i;
|
|
pcierev = crev;
|
|
pcie = TRUE;
|
|
if (cid == PCIE2_CORE_ID)
|
|
pcie_gen2 = TRUE;
|
|
}
|
|
}
|
|
#ifdef BCMSDIO
|
|
else if (((BUSTYPE(bustype) == SDIO_BUS) ||
|
|
(BUSTYPE(bustype) == SPI_BUS)) &&
|
|
(cid == SDIOD_CORE_ID)) {
|
|
sii->pub.buscorerev = (int16)crev;
|
|
sii->pub.buscoretype = (uint16)cid;
|
|
sii->pub.buscoreidx = (uint16)i;
|
|
}
|
|
#endif /* BCMSDIO */
|
|
|
|
/* find the core idx before entering this func. */
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
if (regs == sii->curmap) {
|
|
*origidx = i;
|
|
}
|
|
} else {
|
|
/* find the core idx before entering this func. */
|
|
if ((savewin && (savewin == cores_info->coresba[i])) ||
|
|
(regs == cores_info->regs[i])) {
|
|
*origidx = i;
|
|
}
|
|
}
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
if (pci && pcie) {
|
|
if (si_ispcie(sii))
|
|
pci = FALSE;
|
|
else
|
|
pcie = FALSE;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#if defined(PCIE_FULL_DONGLE)
|
|
if (pcie) {
|
|
if (pcie_gen2)
|
|
sii->pub.buscoretype = PCIE2_CORE_ID;
|
|
else
|
|
sii->pub.buscoretype = PCIE_CORE_ID;
|
|
sii->pub.buscorerev = (int16)pcierev;
|
|
sii->pub.buscoreidx = (uint16)pcieidx;
|
|
}
|
|
BCM_REFERENCE(pci);
|
|
BCM_REFERENCE(pcirev);
|
|
BCM_REFERENCE(pciidx);
|
|
#else
|
|
if (pci) {
|
|
sii->pub.buscoretype = PCI_CORE_ID;
|
|
sii->pub.buscorerev = (int16)pcirev;
|
|
sii->pub.buscoreidx = (uint16)pciidx;
|
|
} else if (pcie) {
|
|
if (pcie_gen2)
|
|
sii->pub.buscoretype = PCIE2_CORE_ID;
|
|
else
|
|
sii->pub.buscoretype = PCIE_CORE_ID;
|
|
sii->pub.buscorerev = (int16)pcierev;
|
|
sii->pub.buscoreidx = (uint16)pcieidx;
|
|
}
|
|
#endif /* defined(PCIE_FULL_DONGLE) */
|
|
|
|
SI_VMSG(("Buscore id/type/rev %d/0x%x/%d\n", sii->pub.buscoreidx, sii->pub.buscoretype,
|
|
sii->pub.buscorerev));
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* fixup necessary chip/core configurations */
|
|
if (BUSTYPE(sii->pub.bustype) == PCI_BUS) {
|
|
if (SI_FAST(sii)) {
|
|
if (!sii->pch &&
|
|
((sii->pch = (void *)(uintptr)pcicore_init(&sii->pub, sii->osh,
|
|
(volatile void *)PCIEREGS(sii))) == NULL))
|
|
return FALSE;
|
|
}
|
|
if (si_pci_fixcfg(&sii->pub)) {
|
|
SI_ERROR(("si_buscore_setup: si_pci_fixcfg failed\n"));
|
|
return FALSE;
|
|
}
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#if defined(BCMSDIO) && defined(BCMDONGLEHOST)
|
|
/* Make sure any on-chip ARM is off (in case strapping is wrong), or downloaded code was
|
|
* already running.
|
|
*/
|
|
if ((BUSTYPE(bustype) == SDIO_BUS) || (BUSTYPE(bustype) == SPI_BUS)) {
|
|
if (si_setcore(&sii->pub, ARM7S_CORE_ID, 0) ||
|
|
si_setcore(&sii->pub, ARMCM3_CORE_ID, 0))
|
|
si_core_disable(&sii->pub, 0);
|
|
}
|
|
#endif /* BCMSDIO && BCMDONGLEHOST */
|
|
|
|
/* return to the original core */
|
|
si_setcoreidx(&sii->pub, *origidx);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST) /* if not a DHD build */
|
|
|
|
static const char rstr_boardvendor[] = "boardvendor";
|
|
static const char rstr_boardtype[] = "boardtype";
|
|
#if defined(BCMPCIEDEV_SROM_FORMAT)
|
|
static const char rstr_subvid[] = "subvid";
|
|
#endif /* defined(BCMPCIEDEV_SROM_FORMAT) */
|
|
#ifdef BCMSDIO
|
|
static const char rstr_manfid[] = "manfid";
|
|
#endif
|
|
static const char rstr_prodid[] = "prodid";
|
|
static const char rstr_boardrev[] = "boardrev";
|
|
static const char rstr_boardflags[] = "boardflags";
|
|
static const char rstr_boardflags4[] = "boardflags4";
|
|
static const char rstr_xtalfreq[] = "xtalfreq";
|
|
static const char rstr_muxenab[] = "muxenab";
|
|
static const char rstr_gpiopulldown[] = "gpdn";
|
|
static const char rstr_devid[] = "devid";
|
|
static const char rstr_wl0id[] = "wl0id";
|
|
static const char rstr_devpathD[] = "devpath%d";
|
|
static const char rstr_D_S[] = "%d:%s";
|
|
static const char rstr_swdenab[] = "swdenable";
|
|
static const char rstr_spurconfig[] = "spurconfig";
|
|
static const char rstr_lpflags[] = "lpflags";
|
|
static const char rstr_armclk[] = "armclk";
|
|
static const char rstr_rfldo3p3_cap_war[] = "rfldo3p3_cap_war";
|
|
#if defined(SECI_UART)
|
|
static const char rstr_fuart_pup_rx_cts[] = "fuart_pup_rx_cts";
|
|
#endif /* defined(SECI_UART) */
|
|
|
|
static uint32
|
|
si_fixup_vid_overrides(si_info_t *sii, char *pvars, uint32 conf_vid)
|
|
{
|
|
BCM_REFERENCE(pvars);
|
|
|
|
if ((sii->pub.boardvendor != VENDOR_APPLE)) {
|
|
return conf_vid;
|
|
}
|
|
|
|
switch (sii->pub.boardtype)
|
|
{
|
|
/* Check for the SROM value */
|
|
case BCM94360X51P2:
|
|
case BCM94360X29C:
|
|
case BCM94360X29CP2:
|
|
case BCM94360X51:
|
|
case BCM943602X87:
|
|
case BCM943602X238D:
|
|
/* Take the PCIe configuration space subsystem ID */
|
|
sii->pub.boardtype = (conf_vid >> 16) & 0xffff;
|
|
break;
|
|
|
|
default:
|
|
/* Do nothing */
|
|
break;
|
|
}
|
|
|
|
return conf_vid;
|
|
}
|
|
|
|
static void
|
|
si_nvram_process(si_info_t *sii, char *pvars)
|
|
{
|
|
uint w = 0;
|
|
|
|
/* get boardtype and boardrev */
|
|
switch (BUSTYPE(sii->pub.bustype)) {
|
|
case PCI_BUS:
|
|
/* do a pci config read to get subsystem id and subvendor id */
|
|
w = OSL_PCI_READ_CONFIG(sii->osh, PCI_CFG_SVID, sizeof(uint32));
|
|
|
|
/* Let nvram variables override subsystem Vend/ID */
|
|
if ((sii->pub.boardvendor = (uint16)si_getdevpathintvar(&sii->pub,
|
|
rstr_boardvendor)) == 0) {
|
|
#ifdef BCMHOSTVARS
|
|
if ((w & 0xffff) == 0)
|
|
sii->pub.boardvendor = VENDOR_BROADCOM;
|
|
else
|
|
#endif /* BCMHOSTVARS */
|
|
sii->pub.boardvendor = w & 0xffff;
|
|
} else {
|
|
SI_ERROR(("Overriding boardvendor: 0x%x instead of 0x%x\n",
|
|
sii->pub.boardvendor, w & 0xffff));
|
|
}
|
|
|
|
if ((sii->pub.boardtype = (uint16)si_getdevpathintvar(&sii->pub, rstr_boardtype))
|
|
== 0) {
|
|
if ((sii->pub.boardtype = getintvar(pvars, rstr_boardtype)) == 0)
|
|
sii->pub.boardtype = (w >> 16) & 0xffff;
|
|
} else {
|
|
SI_ERROR(("Overriding boardtype: 0x%x instead of 0x%x\n",
|
|
sii->pub.boardtype, (w >> 16) & 0xffff));
|
|
}
|
|
|
|
/* Override high priority fixups */
|
|
si_fixup_vid_overrides(sii, pvars, w);
|
|
break;
|
|
|
|
#ifdef BCMSDIO
|
|
case SDIO_BUS:
|
|
sii->pub.boardvendor = getintvar(pvars, rstr_manfid);
|
|
sii->pub.boardtype = getintvar(pvars, rstr_prodid);
|
|
break;
|
|
|
|
case SPI_BUS:
|
|
sii->pub.boardvendor = VENDOR_BROADCOM;
|
|
sii->pub.boardtype = QT4710_BOARD;
|
|
break;
|
|
#endif
|
|
|
|
case SI_BUS:
|
|
#ifdef BCMPCIEDEV_SROM_FORMAT
|
|
if (BCMPCIEDEV_ENAB() && si_is_sprom_available(&sii->pub) && pvars &&
|
|
getvar(pvars, rstr_subvid)) {
|
|
sii->pub.boardvendor = getintvar(pvars, rstr_subvid);
|
|
} else
|
|
#endif
|
|
sii->pub.boardvendor = VENDOR_BROADCOM;
|
|
if (pvars == NULL || ((sii->pub.boardtype = getintvar(pvars, rstr_prodid)) == 0))
|
|
if ((sii->pub.boardtype = getintvar(pvars, rstr_boardtype)) == 0)
|
|
sii->pub.boardtype = 0xffff;
|
|
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_UBUS) {
|
|
/* do a pci config read to get subsystem id and subvendor id */
|
|
w = OSL_PCI_READ_CONFIG(sii->osh, PCI_CFG_SVID, sizeof(uint32));
|
|
sii->pub.boardvendor = w & 0xffff;
|
|
sii->pub.boardtype = (w >> 16) & 0xffff;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (sii->pub.boardtype == 0) {
|
|
SI_ERROR(("si_doattach: unknown board type\n"));
|
|
ASSERT(sii->pub.boardtype);
|
|
}
|
|
|
|
sii->pub.lpflags = getintvar(pvars, rstr_lpflags);
|
|
sii->pub.boardrev = getintvar(pvars, rstr_boardrev);
|
|
sii->pub.boardflags = getintvar(pvars, rstr_boardflags);
|
|
|
|
#ifdef BCM_SDRBL
|
|
sii->pub.boardflags2 |= ((!CHIP_HOSTIF_USB(&(sii->pub))) ? ((si_arm_sflags(&(sii->pub))
|
|
& SISF_SDRENABLE) ? BFL2_SDR_EN:0):
|
|
(((uint)getintvar(pvars, "boardflags2")) & BFL2_SDR_EN));
|
|
#endif /* BCM_SDRBL */
|
|
sii->pub.boardflags4 = getintvar(pvars, rstr_boardflags4);
|
|
|
|
}
|
|
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#if defined(CONFIG_XIP) && defined(BCMTCAM)
|
|
extern uint8 patch_pair;
|
|
#endif /* CONFIG_XIP && BCMTCAM */
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
typedef struct {
|
|
uint8 uart_tx;
|
|
uint32 uart_rx;
|
|
} si_mux_uartopt_t;
|
|
|
|
/* note: each index corr to MUXENAB43012_HOSTWAKE_MASK > shift - 1 */
|
|
static const uint8 mux43012_hostwakeopt[] = {
|
|
CC_PIN_GPIO_00
|
|
};
|
|
|
|
static const si_mux_uartopt_t mux_uartopt[] = {
|
|
{CC_PIN_GPIO_00, CC_PIN_GPIO_01},
|
|
{CC_PIN_GPIO_05, CC_PIN_GPIO_04},
|
|
{CC_PIN_GPIO_15, CC_PIN_GPIO_14},
|
|
};
|
|
|
|
/* note: each index corr to MUXENAB_DEF_HOSTWAKE mask >> shift - 1 */
|
|
static const uint8 mux_hostwakeopt[] = {
|
|
CC_PIN_GPIO_00,
|
|
};
|
|
|
|
#ifdef SECI_UART
|
|
#define NUM_SECI_UART_GPIOS 4
|
|
static bool fuart_pullup_rx_cts_enab = FALSE;
|
|
static bool fast_uart_init = FALSE;
|
|
static uint32 fast_uart_tx;
|
|
static uint32 fast_uart_functionsel;
|
|
static uint32 fast_uart_pup;
|
|
static uint32 fast_uart_rx;
|
|
static uint32 fast_uart_cts_in;
|
|
#endif /* SECI_UART */
|
|
|
|
void
|
|
si_swdenable(si_t *sih, uint32 swdflag)
|
|
{
|
|
/* FIXME Need a more generic test for SWD instead of check on specific chipid */
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
if (swdflag) {
|
|
/* Enable ARM debug clk, which is required for the ARM debug
|
|
* unit to operate
|
|
*/
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL5, (1 << ARMCR4_DBG_CLK_BIT),
|
|
(1 << ARMCR4_DBG_CLK_BIT));
|
|
/* Force HT clock in Chipcommon. The HT clock is required for backplane
|
|
* access via SWD
|
|
*/
|
|
si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, clk_ctl_st), CCS_FORCEHT,
|
|
CCS_FORCEHT);
|
|
/* Set TAP_SEL so that ARM is the first and the only TAP on the TAP chain.
|
|
* Must do a chip reset to clear this bit
|
|
*/
|
|
si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, jtagctrl),
|
|
JCTRL_TAPSEL_BIT, JCTRL_TAPSEL_BIT);
|
|
SI_MSG(("si_swdenable: set arm_dbgclk, ForceHTClock and tap_sel bit\n"));
|
|
}
|
|
break;
|
|
default:
|
|
/* swdenable specified for an unsupported chip */
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/** want to have this available all the time to switch mux for debugging */
|
|
void
|
|
si_muxenab(si_t *sih, uint32 w)
|
|
{
|
|
uint32 chipcontrol, pmu_chipcontrol;
|
|
|
|
pmu_chipcontrol = si_pmu_chipcontrol(sih, 1, 0, 0);
|
|
chipcontrol = si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, chipcontrol),
|
|
0, 0);
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4360_CHIP_ID:
|
|
case BCM43460_CHIP_ID:
|
|
case BCM4352_CHIP_ID:
|
|
case BCM43526_CHIP_ID:
|
|
CASE_BCM43602_CHIP:
|
|
if (w & MUXENAB_UART)
|
|
chipcontrol |= CCTRL4360_UART_MODE;
|
|
break;
|
|
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
/*
|
|
* 0x10 : use GPIO0 as host wake up pin
|
|
* 0x20 ~ 0xf0: Reserved
|
|
*/
|
|
if (w & MUXENAB43012_HOSTWAKE_MASK) {
|
|
uint8 hostwake = 0;
|
|
uint8 hostwake_ix = MUXENAB43012_GETIX(w, HOSTWAKE);
|
|
|
|
if (hostwake_ix >
|
|
sizeof(mux43012_hostwakeopt)/sizeof(mux43012_hostwakeopt[0]) - 1) {
|
|
SI_ERROR(("si_muxenab: wrong index %d for hostwake\n",
|
|
hostwake_ix));
|
|
break;
|
|
}
|
|
|
|
hostwake = mux43012_hostwakeopt[hostwake_ix];
|
|
si_gci_set_functionsel(sih, hostwake, CC_FNSEL_MISC1);
|
|
}
|
|
break;
|
|
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
if (w & MUXENAB_DEF_UART_MASK) {
|
|
uint32 uart_rx = 0, uart_tx = 0;
|
|
uint8 uartopt_idx = (w & MUXENAB_DEF_UART_MASK) - 1;
|
|
uint8 uartopt_size = sizeof(mux_uartopt)/sizeof(mux_uartopt[0]);
|
|
|
|
if (uartopt_idx < uartopt_size) {
|
|
uart_rx = mux_uartopt[uartopt_idx].uart_rx;
|
|
uart_tx = mux_uartopt[uartopt_idx].uart_tx;
|
|
#ifdef BOOTLOADER_CONSOLE_OUTPUT
|
|
uart_rx = 0;
|
|
uart_tx = 1;
|
|
#endif
|
|
if (CHIPREV(sih->chiprev) >= 3) {
|
|
si_gci_set_functionsel(sih, uart_rx, CC_FNSEL_GPIO1);
|
|
si_gci_set_functionsel(sih, uart_tx, CC_FNSEL_GPIO1);
|
|
} else {
|
|
si_gci_set_functionsel(sih, uart_rx, CC_FNSEL_GPIO0);
|
|
si_gci_set_functionsel(sih, uart_tx, CC_FNSEL_GPIO0);
|
|
}
|
|
} else {
|
|
SI_MSG(("si_muxenab: Invalid uart OTP setting\n"));
|
|
}
|
|
}
|
|
if (w & MUXENAB_DEF_HOSTWAKE_MASK) {
|
|
uint8 hostwake = 0;
|
|
/*
|
|
* SDIO
|
|
* 0x10 : use GPIO0 as host wake up pin
|
|
*/
|
|
uint8 hostwake_ix = MUXENAB_DEF_GETIX(w, HOSTWAKE);
|
|
|
|
if (hostwake_ix > (sizeof(mux_hostwakeopt) /
|
|
sizeof(mux_hostwakeopt[0]) - 1)) {
|
|
SI_ERROR(("si_muxenab: wrong index %d for hostwake\n",
|
|
hostwake_ix));
|
|
break;
|
|
}
|
|
|
|
hostwake = mux_hostwakeopt[hostwake_ix];
|
|
si_gci_set_functionsel(sih, hostwake, CC_FNSEL_GPIO0);
|
|
}
|
|
|
|
break;
|
|
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
/* TBD fill */
|
|
if (w & MUXENAB_HOST_WAKE) {
|
|
si_gci_set_functionsel(sih, CC_PIN_GPIO_00, CC_FNSEL_MISC1);
|
|
}
|
|
break;
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
/* TBD fill */
|
|
break;
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
/* TBD fill */
|
|
break;
|
|
default:
|
|
/* muxenab specified for an unsupported chip */
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
|
|
/* write both updated values to hw */
|
|
si_pmu_chipcontrol(sih, 1, ~0, pmu_chipcontrol);
|
|
si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, chipcontrol),
|
|
~0, chipcontrol);
|
|
}
|
|
|
|
/** ltecx GCI reg access */
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gci_direct)(si_t *sih, uint offset, uint32 mask, uint32 val)
|
|
{
|
|
/* gci direct reg access */
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), offset, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_gci_indirect(si_t *sih, uint regidx, uint offset, uint32 mask, uint32 val)
|
|
{
|
|
/* gci indirect reg access */
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, regidx);
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), offset, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_gci_input(si_t *sih, uint reg)
|
|
{
|
|
/* gci_input[] */
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_input[reg]), 0, 0);
|
|
}
|
|
|
|
uint32
|
|
si_gci_output(si_t *sih, uint reg, uint32 mask, uint32 val)
|
|
{
|
|
/* gci_output[] */
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_output[reg]), mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_gci_int_enable(si_t *sih, bool enable)
|
|
{
|
|
uint offs;
|
|
|
|
/* enable GCI interrupt */
|
|
offs = OFFSETOF(chipcregs_t, intmask);
|
|
return (si_corereg(sih, SI_CC_IDX, offs, CI_ECI, (enable ? CI_ECI : 0)));
|
|
}
|
|
|
|
void
|
|
si_gci_reset(si_t *sih)
|
|
{
|
|
int i;
|
|
|
|
if (gci_reset_done == FALSE) {
|
|
gci_reset_done = TRUE;
|
|
|
|
/* Set ForceRegClk and ForceSeciClk */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((1 << GCI_CCTL_FREGCLK_OFFSET)
|
|
|(1 << GCI_CCTL_FSECICLK_OFFSET)),
|
|
((1 << GCI_CCTL_FREGCLK_OFFSET)
|
|
|(1 << GCI_CCTL_FSECICLK_OFFSET)));
|
|
|
|
/* Some Delay */
|
|
for (i = 0; i < 2; i++) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl), 0, 0);
|
|
}
|
|
/* Reset SECI block */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((1 << GCI_CCTL_SECIRST_OFFSET)
|
|
|(1 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(1 << GCI_CCTL_RSTOCC_OFFSET)),
|
|
((1 << GCI_CCTL_SECIRST_OFFSET)
|
|
|(1 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(1 << GCI_CCTL_RSTOCC_OFFSET)));
|
|
/* Some Delay */
|
|
for (i = 0; i < 10; i++) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl), 0, 0);
|
|
}
|
|
/* Remove SECI Reset */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((1 << GCI_CCTL_SECIRST_OFFSET)
|
|
|(1 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(1 << GCI_CCTL_RSTOCC_OFFSET)),
|
|
((0 << GCI_CCTL_SECIRST_OFFSET)
|
|
|(0 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(0 << GCI_CCTL_RSTOCC_OFFSET)));
|
|
|
|
/* Some Delay */
|
|
for (i = 0; i < 2; i++) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl), 0, 0);
|
|
}
|
|
|
|
/* Clear ForceRegClk and ForceSeciClk */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((1 << GCI_CCTL_FREGCLK_OFFSET)
|
|
|(1 << GCI_CCTL_FSECICLK_OFFSET)),
|
|
((0 << GCI_CCTL_FREGCLK_OFFSET)
|
|
|(0 << GCI_CCTL_FSECICLK_OFFSET)));
|
|
}
|
|
/* clear events */
|
|
for (i = 0; i < 32; i++) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_event[i]), ALLONES_32, 0x00);
|
|
}
|
|
}
|
|
|
|
void
|
|
si_gci_gpio_chipcontrol_ex(si_t *sih, uint8 gci_gpio, uint8 opt)
|
|
{
|
|
si_gci_gpio_chipcontrol(sih, gci_gpio, opt);
|
|
}
|
|
|
|
static void
|
|
BCMPOSTTRAPFN(si_gci_gpio_chipcontrol)(si_t *sih, uint8 gci_gpio, uint8 opt)
|
|
{
|
|
uint32 ring_idx = 0, pos = 0;
|
|
|
|
si_gci_get_chipctrlreg_ringidx_base8(gci_gpio, &ring_idx, &pos);
|
|
SI_MSG(("si_gci_gpio_chipcontrol:rngidx is %d, pos is %d, opt is %d, mask is 0x%04x,"
|
|
" value is 0x%04x\n",
|
|
ring_idx, pos, opt, GCIMASK_8B(pos), GCIPOSVAL_8B(opt, pos)));
|
|
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, ring_idx);
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_gpioctl),
|
|
GCIMASK_8B(pos), GCIPOSVAL_8B(opt, pos));
|
|
}
|
|
|
|
static uint8
|
|
BCMPOSTTRAPFN(si_gci_gpio_reg)(si_t *sih, uint8 gci_gpio, uint8 mask, uint8 value,
|
|
uint32 reg_offset)
|
|
{
|
|
uint32 ring_idx = 0, pos = 0; /**< FunctionSel register idx and bits to use */
|
|
uint32 val_32;
|
|
|
|
si_gci_get_chipctrlreg_ringidx_base4(gci_gpio, &ring_idx, &pos);
|
|
SI_MSG(("si_gci_gpio_reg:rngidx is %d, pos is %d, val is %d, mask is 0x%04x,"
|
|
" value is 0x%04x\n",
|
|
ring_idx, pos, value, GCIMASK_4B(pos), GCIPOSVAL_4B(value, pos)));
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, ring_idx);
|
|
|
|
if (mask || value) {
|
|
/* set operation */
|
|
si_corereg(sih, GCI_CORE_IDX(sih),
|
|
reg_offset, GCIMASK_4B(pos), GCIPOSVAL_4B(value, pos));
|
|
}
|
|
val_32 = si_corereg(sih, GCI_CORE_IDX(sih), reg_offset, 0, 0);
|
|
|
|
value = (uint8)((val_32 >> pos) & 0xFF);
|
|
|
|
return value;
|
|
}
|
|
|
|
/**
|
|
* In order to route a ChipCommon originated GPIO towards a package pin, both CC and GCI cores have
|
|
* to be written to.
|
|
* @param[in] sih
|
|
* @param[in] gpio chip specific package pin number. See Toplevel Arch page, GCI chipcontrol reg
|
|
* section.
|
|
* @param[in] mask chip common gpio mask
|
|
* @param[in] val chip common gpio value
|
|
*/
|
|
void
|
|
BCMPOSTTRAPFN(si_gci_enable_gpio)(si_t *sih, uint8 gpio, uint32 mask, uint32 value)
|
|
{
|
|
uint32 ring_idx = 0, pos = 0;
|
|
|
|
si_gci_get_chipctrlreg_ringidx_base4(gpio, &ring_idx, &pos);
|
|
SI_MSG(("si_gci_enable_gpio:rngidx is %d, pos is %d, val is %d, mask is 0x%04x,"
|
|
" value is 0x%04x\n",
|
|
ring_idx, pos, value, GCIMASK_4B(pos), GCIPOSVAL_4B(value, pos)));
|
|
si_gci_set_functionsel(sih, gpio, CC_FNSEL_SAMEASPIN);
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, ring_idx);
|
|
|
|
si_gpiocontrol(sih, mask, 0, GPIO_HI_PRIORITY);
|
|
si_gpioouten(sih, mask, mask, GPIO_HI_PRIORITY);
|
|
si_gpioout(sih, mask, value, GPIO_HI_PRIORITY);
|
|
}
|
|
|
|
/*
|
|
* The above seems to be for gpio output only (forces gpioouten).
|
|
* This function is to configure GPIO as input, and accepts a mask of bits.
|
|
* Also: doesn't force the gpiocontrol (chipc) functionality, assumes it
|
|
* is the default, and rejects the request (BUSY => gpio in use) if it's
|
|
* already configured for a different function... but it will override the
|
|
* output enable.
|
|
*/
|
|
int
|
|
si_gpio_enable(si_t *sih, uint32 mask)
|
|
{
|
|
uint bit;
|
|
int fnsel = -1; /* Valid fnsel is a small positive number */
|
|
|
|
BCM_REFERENCE(bit);
|
|
BCM_REFERENCE(fnsel);
|
|
|
|
/* Bail if any bit is explicitly set for some other function */
|
|
if (si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, gpiocontrol), 0, 0) & mask) {
|
|
return BCME_BUSY;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* Some chips need to be explicitly set */
|
|
switch (CHIPID(sih->chip))
|
|
{
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
fnsel = CC_FNSEL_SAMEASPIN;
|
|
default:
|
|
;
|
|
}
|
|
|
|
if (fnsel != -1) {
|
|
for (bit = 0; mask; bit++) {
|
|
if (mask & (1 << bit)) {
|
|
si_gci_set_functionsel(sih, bit, (uint8)fnsel);
|
|
mask ^= (1 << bit);
|
|
}
|
|
}
|
|
}
|
|
#endif /* !BCMDONGLEHOST */
|
|
si_gpioouten(sih, mask, 0, GPIO_HI_PRIORITY);
|
|
|
|
return BCME_OK;
|
|
}
|
|
|
|
static const char rstr_host_wake_opt[] = "host_wake_opt";
|
|
uint8
|
|
si_gci_host_wake_gpio_init(si_t *sih)
|
|
{
|
|
uint8 host_wake_gpio = CC_GCI_GPIO_INVALID;
|
|
uint32 host_wake_opt;
|
|
|
|
/* parse the device wake opt from nvram */
|
|
/* decode what that means for specific chip */
|
|
if (getvar(NULL, rstr_host_wake_opt) == NULL)
|
|
return host_wake_gpio;
|
|
|
|
host_wake_opt = getintvar(NULL, rstr_host_wake_opt);
|
|
host_wake_gpio = host_wake_opt & 0xff;
|
|
si_gci_host_wake_gpio_enable(sih, host_wake_gpio, FALSE);
|
|
|
|
return host_wake_gpio;
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_gci_host_wake_gpio_enable)(si_t *sih, uint8 gpio, bool state)
|
|
{
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
si_gci_enable_gpio(sih, gpio, 1 << gpio,
|
|
state ? 1 << gpio : 0x00);
|
|
break;
|
|
default:
|
|
SI_ERROR(("host wake not supported for 0x%04x yet\n", CHIPID(sih->chip)));
|
|
break;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_gci_time_sync_gpio_enable(si_t *sih, uint8 gpio, bool state)
|
|
{
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
si_gci_enable_gpio(sih, gpio, 1 << gpio,
|
|
state ? 1 << gpio : 0x00);
|
|
break;
|
|
default:
|
|
SI_ERROR(("Time sync not supported for 0x%04x yet\n", CHIPID(sih->chip)));
|
|
break;
|
|
}
|
|
}
|
|
|
|
#define TIMESYNC_GPIO_NUM 12 /* Hardcoded for now. Will be removed later */
|
|
static const char rstr_time_sync_opt[] = "time_sync_opt";
|
|
uint8
|
|
si_gci_time_sync_gpio_init(si_t *sih)
|
|
{
|
|
uint8 time_sync_gpio = TIMESYNC_GPIO_NUM;
|
|
uint32 time_sync_opt;
|
|
|
|
/* parse the device wake opt from nvram */
|
|
/* decode what that means for specific chip */
|
|
if (getvar(NULL, rstr_time_sync_opt) == NULL) {
|
|
time_sync_opt = TIMESYNC_GPIO_NUM;
|
|
} else {
|
|
time_sync_opt = getintvar(NULL, rstr_time_sync_opt);
|
|
}
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
time_sync_gpio = time_sync_opt & 0xff;
|
|
si_gci_enable_gpio(sih, time_sync_gpio,
|
|
1 << time_sync_gpio, 0x00);
|
|
break;
|
|
default:
|
|
SI_ERROR(("time sync not supported for 0x%04x yet\n", CHIPID(sih->chip)));
|
|
break;
|
|
}
|
|
|
|
return time_sync_gpio;
|
|
}
|
|
|
|
uint8
|
|
BCMPOSTTRAPFN(si_gci_gpio_wakemask)(si_t *sih, uint8 gpio, uint8 mask, uint8 value)
|
|
{
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_wakemask),
|
|
GCI_WAKEMASK_GPIOWAKE, GCI_WAKEMASK_GPIOWAKE);
|
|
return (si_gci_gpio_reg(sih, gpio, mask, value, GCI_OFFSETOF(sih, gci_gpiowakemask)));
|
|
}
|
|
|
|
uint8
|
|
BCMPOSTTRAPFN(si_gci_gpio_intmask)(si_t *sih, uint8 gpio, uint8 mask, uint8 value)
|
|
{
|
|
return (si_gci_gpio_reg(sih, gpio, mask, value, GCI_OFFSETOF(sih, gci_gpiointmask)));
|
|
}
|
|
|
|
uint8
|
|
BCMPOSTTRAPFN(si_gci_gpio_status)(si_t *sih, uint8 gpio, uint8 mask, uint8 value)
|
|
{
|
|
return (si_gci_gpio_reg(sih, gpio, mask, value, GCI_OFFSETOF(sih, gci_gpiostatus)));
|
|
}
|
|
|
|
static void
|
|
si_gci_enable_gpioint(si_t *sih, bool enable)
|
|
{
|
|
if (enable)
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_intmask),
|
|
GCI_INTSTATUS_GPIOINT, GCI_INTSTATUS_GPIOINT);
|
|
else
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_intmask),
|
|
GCI_INTSTATUS_GPIOINT, 0);
|
|
}
|
|
|
|
/* assumes function select is performed separately */
|
|
void
|
|
si_enable_gpio_wake(si_t *sih, uint8 *wake_mask, uint8 *cur_status, uint8 gci_gpio,
|
|
uint32 pmu_cc2_mask, uint32 pmu_cc2_value)
|
|
{
|
|
si_gci_gpio_chipcontrol(sih, gci_gpio,
|
|
(1 << GCI_GPIO_CHIPCTRL_ENAB_IN_BIT));
|
|
|
|
si_gci_gpio_intmask(sih, gci_gpio, *wake_mask, *wake_mask);
|
|
si_gci_gpio_wakemask(sih, gci_gpio, *wake_mask, *wake_mask);
|
|
|
|
/* clear the existing status bits */
|
|
*cur_status = si_gci_gpio_status(sih, gci_gpio,
|
|
GCI_GPIO_STS_CLEAR, GCI_GPIO_STS_CLEAR);
|
|
|
|
/* top level gci int enable */
|
|
si_gci_enable_gpioint(sih, TRUE);
|
|
|
|
/* enable the pmu chip control bit to enable wake */
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL2, pmu_cc2_mask, pmu_cc2_value);
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_gci_config_wake_pin)(si_t *sih, uint8 gpio_n, uint8 wake_events, bool gci_gpio)
|
|
{
|
|
uint8 chipcontrol = 0;
|
|
uint32 pmu_chipcontrol2 = 0;
|
|
|
|
if (!gci_gpio)
|
|
chipcontrol = (1 << GCI_GPIO_CHIPCTRL_ENAB_EXT_GPIO_BIT);
|
|
|
|
chipcontrol |= (1 << GCI_GPIO_CHIPCTRL_PULLUP_BIT);
|
|
si_gci_gpio_chipcontrol(sih, gpio_n,
|
|
(chipcontrol | (1 << GCI_GPIO_CHIPCTRL_ENAB_IN_BIT)));
|
|
|
|
/* enable gci gpio int/wake events */
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, wake_events);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, wake_events);
|
|
|
|
/* clear the existing status bits */
|
|
si_gci_gpio_status(sih, gpio_n,
|
|
GCI_GPIO_STS_CLEAR, GCI_GPIO_STS_CLEAR);
|
|
|
|
/* Enable gci2wl_wake */
|
|
pmu_chipcontrol2 = si_pmu_chipcontrol(sih, PMU_CHIPCTL2, 0, 0);
|
|
pmu_chipcontrol2 |= si_pmu_wake_bit_offset(sih);
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL2, ~0, pmu_chipcontrol2);
|
|
|
|
/* enable gci int/wake events */
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_intmask),
|
|
GCI_INTSTATUS_GPIOINT, GCI_INTSTATUS_GPIOINT);
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_wakemask),
|
|
GCI_INTSTATUS_GPIOWAKE, GCI_INTSTATUS_GPIOWAKE);
|
|
}
|
|
|
|
void
|
|
si_gci_free_wake_pin(si_t *sih, uint8 gpio_n)
|
|
{
|
|
uint8 chipcontrol = 0;
|
|
uint8 wake_events;
|
|
|
|
si_gci_gpio_chipcontrol(sih, gpio_n, chipcontrol);
|
|
|
|
/* enable gci gpio int/wake events */
|
|
wake_events = si_gci_gpio_intmask(sih, gpio_n, 0, 0);
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, 0);
|
|
wake_events = si_gci_gpio_wakemask(sih, gpio_n, 0, 0);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, 0);
|
|
|
|
/* clear the existing status bits */
|
|
si_gci_gpio_status(sih, gpio_n,
|
|
GCI_GPIO_STS_CLEAR, GCI_GPIO_STS_CLEAR);
|
|
}
|
|
|
|
#if defined(BCMPCIEDEV)
|
|
static const char rstr_device_wake_opt[] = "device_wake_opt";
|
|
#else
|
|
static const char rstr_device_wake_opt[] = "sd_devwake";
|
|
#endif
|
|
#define DEVICE_WAKE_GPIO3 3
|
|
|
|
uint8
|
|
si_enable_perst_wake(si_t *sih, uint8 *perst_wake_mask, uint8 *perst_cur_status)
|
|
{
|
|
uint8 gci_perst = CC_GCI_GPIO_15;
|
|
switch (CHIPID(sih->chip)) {
|
|
default:
|
|
SI_ERROR(("device wake not supported for 0x%04x yet\n", CHIPID(sih->chip)));
|
|
break;
|
|
}
|
|
return gci_perst;
|
|
|
|
}
|
|
|
|
uint8
|
|
si_get_device_wake_opt(si_t *sih)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (getvar(NULL, rstr_device_wake_opt) == NULL)
|
|
return CC_GCI_GPIO_INVALID;
|
|
|
|
sii->device_wake_opt = (uint8)getintvar(NULL, rstr_device_wake_opt);
|
|
return sii->device_wake_opt;
|
|
}
|
|
|
|
uint8
|
|
si_enable_device_wake(si_t *sih, uint8 *wake_mask, uint8 *cur_status)
|
|
{
|
|
uint8 gci_gpio = CC_GCI_GPIO_INVALID; /* DEVICE_WAKE GCI GPIO */
|
|
uint32 device_wake_opt;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
device_wake_opt = sii->device_wake_opt;
|
|
|
|
if (device_wake_opt == CC_GCI_GPIO_INVALID) {
|
|
/* parse the device wake opt from nvram */
|
|
/* decode what that means for specific chip */
|
|
/* apply the right gci config */
|
|
/* enable the internal interrupts */
|
|
/* assume: caller already registered handler for that GCI int */
|
|
if (getvar(NULL, rstr_device_wake_opt) == NULL)
|
|
return gci_gpio;
|
|
|
|
device_wake_opt = getintvar(NULL, rstr_device_wake_opt);
|
|
}
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
/* device_wake op 1:
|
|
* gpio 1, func sel 4,
|
|
* gcigpioctrl: input pin, exra gpio
|
|
* since GCI_GPIO_CHIPCTRL_ENAB_EXT_GPIO_BIT is used, gci gpio is same as GPIO num
|
|
* GCI GPIO 1,wakemask/intmask: any edge, both positive negative
|
|
* enable the wake mask, intmask in GCI top level
|
|
* enable the chip common to get the G/ECI interrupt
|
|
* enable the PMU ctrl to wake the chip on wakemask set
|
|
*/
|
|
if (device_wake_opt == 1) {
|
|
gci_gpio = CC_GCI_GPIO_1;
|
|
*wake_mask = (1 << GCI_GPIO_STS_VALUE_BIT) |
|
|
(1 << GCI_GPIO_STS_POS_EDGE_BIT) |
|
|
(1 << GCI_GPIO_STS_NEG_EDGE_BIT);
|
|
si_gci_set_functionsel(sih, gci_gpio, CC_FNSEL_GCI0);
|
|
si_enable_gpio_wake(sih, wake_mask, cur_status, gci_gpio,
|
|
PMU_CC2_GCI2_WAKE | PMU_CC2_MASK_WL_DEV_WAKE,
|
|
PMU_CC2_GCI2_WAKE | PMU_CC2_MASK_WL_DEV_WAKE);
|
|
/* hack: add a pulldown to HOST_WAKE */
|
|
si_gci_gpio_chipcontrol(sih, 0,
|
|
(1 << GCI_GPIO_CHIPCTRL_PULLDN_BIT));
|
|
|
|
/* Enable wake on GciWake */
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_wakemask),
|
|
(GCI_INTSTATUS_GPIOWAKE | GCI_INTSTATUS_GPIOINT),
|
|
(GCI_INTSTATUS_GPIOWAKE | GCI_INTSTATUS_GPIOINT));
|
|
|
|
} else {
|
|
SI_ERROR(("0x%04x: don't know about device_wake_opt %d\n",
|
|
CHIPID(sih->chip), device_wake_opt));
|
|
}
|
|
break;
|
|
default:
|
|
SI_ERROR(("device wake not supported for 0x%04x yet\n", CHIPID(sih->chip)));
|
|
break;
|
|
}
|
|
return gci_gpio;
|
|
}
|
|
|
|
void
|
|
si_gci_gpioint_handler_unregister(si_t *sih, void *gci_i)
|
|
{
|
|
si_info_t *sii;
|
|
gci_gpio_item_t *p, *n;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
ASSERT(gci_i != NULL);
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
if (!(sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT)) {
|
|
SI_ERROR(("si_gci_gpioint_handler_unregister: not GCI capable\n"));
|
|
return;
|
|
}
|
|
ASSERT(sii->gci_gpio_head != NULL);
|
|
|
|
if ((void*)sii->gci_gpio_head == gci_i) {
|
|
sii->gci_gpio_head = sii->gci_gpio_head->next;
|
|
MFREE(sii->osh, gci_i, sizeof(gci_gpio_item_t));
|
|
return;
|
|
} else {
|
|
p = sii->gci_gpio_head;
|
|
n = p->next;
|
|
while (n) {
|
|
if ((void*)n == gci_i) {
|
|
p->next = n->next;
|
|
MFREE(sii->osh, gci_i, sizeof(gci_gpio_item_t));
|
|
return;
|
|
}
|
|
p = n;
|
|
n = n->next;
|
|
}
|
|
}
|
|
}
|
|
|
|
void*
|
|
si_gci_gpioint_handler_register(si_t *sih, uint8 gci_gpio, uint8 gpio_status,
|
|
gci_gpio_handler_t cb, void *arg)
|
|
{
|
|
si_info_t *sii;
|
|
gci_gpio_item_t *gci_i;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
ASSERT(cb != NULL);
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
if (!(sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT)) {
|
|
SI_ERROR(("si_gci_gpioint_handler_register: not GCI capable\n"));
|
|
return NULL;
|
|
}
|
|
|
|
SI_MSG(("si_gci_gpioint_handler_register: gci_gpio is %d\n", gci_gpio));
|
|
if (gci_gpio >= SI_GPIO_MAX) {
|
|
SI_ERROR(("isi_gci_gpioint_handler_register: Invalid GCI GPIO NUM %d\n", gci_gpio));
|
|
return NULL;
|
|
}
|
|
|
|
gci_i = MALLOC(sii->osh, (sizeof(gci_gpio_item_t)));
|
|
|
|
ASSERT(gci_i);
|
|
if (gci_i == NULL) {
|
|
SI_ERROR(("si_gci_gpioint_handler_register: GCI Item MALLOC failure\n"));
|
|
return NULL;
|
|
}
|
|
|
|
if (sii->gci_gpio_head)
|
|
gci_i->next = sii->gci_gpio_head;
|
|
else
|
|
gci_i->next = NULL;
|
|
|
|
sii->gci_gpio_head = gci_i;
|
|
|
|
gci_i->handler = cb;
|
|
gci_i->arg = arg;
|
|
gci_i->gci_gpio = gci_gpio;
|
|
gci_i->status = gpio_status;
|
|
|
|
return (void *)(gci_i);
|
|
}
|
|
|
|
static void
|
|
si_gci_gpioint_handler_process(si_t *sih)
|
|
{
|
|
si_info_t *sii;
|
|
uint32 gpio_status[2], status;
|
|
gci_gpio_item_t *gci_i;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
/* most probably there are going to be 1 or 2 GPIOs used this way, so do for each GPIO */
|
|
|
|
/* go through the GPIO handlers and call them back if their intstatus is set */
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, 0);
|
|
gpio_status[0] = si_corereg(sih, GCI_CORE_IDX(sih),
|
|
GCI_OFFSETOF(sih, gci_gpiostatus), 0, 0);
|
|
/* Only clear the status bits that have been read. Other bits (if present) should not
|
|
* get cleared, so that they can be handled later.
|
|
*/
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_gpiostatus), ~0, gpio_status[0]);
|
|
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, 1);
|
|
gpio_status[1] = si_corereg(sih, GCI_CORE_IDX(sih),
|
|
GCI_OFFSETOF(sih, gci_gpiostatus), 0, 0);
|
|
/* Only clear the status bits that have been read. Other bits (if present) should not
|
|
* get cleared, so that they can be handled later.
|
|
*/
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_gpiostatus), ~0, gpio_status[1]);
|
|
|
|
gci_i = sii->gci_gpio_head;
|
|
|
|
SI_MSG(("si_gci_gpioint_handler_process: status 0x%04x, 0x%04x\n",
|
|
gpio_status[0], gpio_status[1]));
|
|
|
|
while (gci_i) {
|
|
if (gci_i->gci_gpio < 8)
|
|
status = ((gpio_status[0] >> (gci_i->gci_gpio * 4)) & 0x0F);
|
|
else
|
|
status = ((gpio_status[1] >> ((gci_i->gci_gpio - 8) * 4)) & 0x0F);
|
|
/* should we mask these */
|
|
/* call back */
|
|
ASSERT(gci_i->handler);
|
|
if (gci_i->status & status)
|
|
gci_i->handler(status, gci_i->arg);
|
|
gci_i = gci_i->next;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_gci_handler_process(si_t *sih)
|
|
{
|
|
uint32 gci_intstatus;
|
|
|
|
/* check the intmask, wakemask in the interrupt routine and call the right ones */
|
|
/* for now call the gpio interrupt */
|
|
gci_intstatus = si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_intstat), 0, 0);
|
|
|
|
if (gci_intstatus & GCI_INTMASK_GPIOINT) {
|
|
SI_MSG(("si_gci_handler_process: gci_intstatus is 0x%04x\n", gci_intstatus));
|
|
si_gci_gpioint_handler_process(sih);
|
|
}
|
|
if ((gci_intstatus & ~(GCI_INTMASK_GPIOINT))) {
|
|
#ifdef HNDGCI
|
|
hndgci_handler_process(gci_intstatus, sih);
|
|
#endif /* HNDGCI */
|
|
}
|
|
#ifdef WLGCIMBHLR
|
|
if (gci_intstatus & GCI_INTSTATUS_EVENT) {
|
|
hnd_gci_mb_handler_process(gci_intstatus, sih);
|
|
}
|
|
#endif /* WLGCIMBHLR */
|
|
|
|
#if defined(BCMLTECOEX) && !defined(WLTEST)
|
|
if (gci_intstatus & GCI_INTMASK_SRFNE) {
|
|
si_wci2_rxfifo_intr_handler_process(sih, gci_intstatus);
|
|
}
|
|
#endif /* BCMLTECOEX && !WLTEST */
|
|
|
|
#ifdef BCMGCISHM
|
|
if (gci_intstatus & (GCI_INTSTATUS_EVENT | GCI_INTSTATUS_EVENTWAKE)) {
|
|
hnd_gcishm_handler_process(sih, gci_intstatus);
|
|
}
|
|
#endif /* BCMGCISHM */
|
|
}
|
|
|
|
void
|
|
si_gci_seci_init(si_t *sih)
|
|
{
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl), ALLONES_32,
|
|
(GCI_CCTL_SCS << GCI_CCTL_SCS_OFFSET) |
|
|
(GCI_MODE_SECI << GCI_CCTL_SMODE_OFFSET) |
|
|
(1 << GCI_CCTL_SECIEN_OFFSET));
|
|
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_chipctrl), ALLONES_32, 0x0080000); //0x200
|
|
|
|
si_gci_indirect(sih, 1, GCI_OFFSETOF(sih, gci_gpioctl), ALLONES_32, 0x00010280); //0x044
|
|
|
|
/* baudrate:4Mbps at 40MHz xtal, escseq:0xdb, high baudrate, enable seci_tx/rx */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv), //0x1e0
|
|
ALLONES_32, 0xF6);
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj), ALLONES_32, 0xFF); //0x1f8
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secifcr), ALLONES_32, 0x00); //0x1e4
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr), ALLONES_32, 0x08); //0x1ec
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secilcr), ALLONES_32, 0xA8); //0x1e8
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciuartescval), //0x1d0
|
|
ALLONES_32, 0xDB);
|
|
|
|
/* Atlas/GMAC3 configuration for SECI */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_miscctl), ALLONES_32, 0xFFFF); //0xc54
|
|
|
|
/* config GPIO pins 5/6 as SECI_IN/SECI_OUT */
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_seciin_ctrl), ALLONES_32, 0x161); //0x218
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_seciout_ctrl), ALLONES_32, 0x10051); //0x21c
|
|
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciout_txen_txbr), ALLONES_32, 0x01); //0x224
|
|
|
|
/* WLAN rx offset assignment */
|
|
/* WLCX: RX offset assignment from WLAN core to WLAN core (faked as BT TX) */
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_secif0rx_offset), ALLONES_32, 0x13121110); //0x1bc
|
|
si_gci_indirect(sih, 1,
|
|
GCI_OFFSETOF(sih, gci_secif0rx_offset), ALLONES_32, 0x17161514);
|
|
si_gci_indirect(sih, 2,
|
|
GCI_OFFSETOF(sih, gci_secif0rx_offset), ALLONES_32, 0x1b1a1918);
|
|
|
|
/* first 12 nibbles configured for format-0 */
|
|
/* note: we can only select 1st 12 nibbles of each IP for format_0 */
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_seciusef0tx_reg), //0x1b4
|
|
ALLONES_32, 0xFFF); // first 12 nibbles
|
|
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_secitx_datatag),
|
|
ALLONES_32, 0x0F0); // gci_secitx_datatag(nibbles 4 to 7 tagged)
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_secirx_datatag),
|
|
ALLONES_32, 0x0F0); // gci_secirx_datatag(nibbles 4 to 7 tagged)
|
|
|
|
/* TX offset assignment (wlan to bt) */
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_secif0tx_offset), 0xFFFFFFFF, 0x76543210); //0x1b8
|
|
si_gci_indirect(sih, 1,
|
|
GCI_OFFSETOF(sih, gci_secif0tx_offset), 0xFFFFFFFF, 0x0000ba98);
|
|
if (CHIPID(sih->chip) == BCM43602_CHIP_ID) {
|
|
/* Request BT side to update SECI information */
|
|
si_gci_direct(sih, OFFSETOF(chipcregs_t, gci_seciauxtx),
|
|
(SECI_AUX_TX_START | SECI_REFRESH_REQ),
|
|
(SECI_AUX_TX_START | SECI_REFRESH_REQ));
|
|
/* WLAN to update SECI information */
|
|
si_gci_direct(sih, OFFSETOF(chipcregs_t, gci_corectrl),
|
|
SECI_UPD_SECI, SECI_UPD_SECI);
|
|
}
|
|
|
|
// HW ECI bus directly driven from IP
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_control_0), ALLONES_32, 0x00000000);
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_control_1), ALLONES_32, 0x00000000);
|
|
}
|
|
|
|
#if defined(BCMLTECOEX) && !defined(WLTEST)
|
|
int
|
|
si_wci2_rxfifo_handler_register(si_t *sih, wci2_handler_t rx_cb, void *ctx)
|
|
{
|
|
si_info_t *sii;
|
|
wci2_rxfifo_info_t *wci2_info;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
ASSERT(rx_cb != NULL);
|
|
|
|
if (!(sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT)) {
|
|
SI_ERROR(("si_wci2_rxfifo_handler_register: not GCI capable\n"));
|
|
return BCME_ERROR;
|
|
}
|
|
|
|
if ((wci2_info = (wci2_rxfifo_info_t *)MALLOCZ(sii->osh,
|
|
sizeof(wci2_rxfifo_info_t))) == NULL) {
|
|
SI_ERROR(("si_wci2_rxfifo_handler_register: WCI2 RXFIFO INFO MALLOC failure\n"));
|
|
return BCME_NOMEM;
|
|
}
|
|
|
|
if ((wci2_info->rx_buf = (char *)MALLOCZ(sii->osh, WCI2_UART_RX_BUF_SIZE)) == NULL) {
|
|
MFREE(sii->osh, wci2_info, sizeof(wci2_rxfifo_info_t));
|
|
|
|
SI_ERROR(("si_wci2_rxfifo_handler_register: WCI2 RXFIFO INFO MALLOC failure\n"));
|
|
return BCME_NOMEM;
|
|
}
|
|
|
|
if ((wci2_info->cbs = (wci2_cbs_t *)MALLOCZ(sii->osh, sizeof(wci2_cbs_t))) == NULL) {
|
|
MFREE(sii->osh, wci2_info->rx_buf, WCI2_UART_RX_BUF_SIZE);
|
|
MFREE(sii->osh, wci2_info, sizeof(wci2_rxfifo_info_t));
|
|
|
|
SI_ERROR(("si_wci2_rxfifo_handler_register: WCI2 RXFIFO INFO MALLOC failure\n"));
|
|
return BCME_NOMEM;
|
|
}
|
|
|
|
sii->wci2_info = wci2_info;
|
|
|
|
/* init callback */
|
|
wci2_info->cbs->handler = rx_cb;
|
|
wci2_info->cbs->context = ctx;
|
|
|
|
return BCME_OK;
|
|
}
|
|
|
|
void
|
|
si_wci2_rxfifo_handler_unregister(si_t *sih)
|
|
{
|
|
|
|
si_info_t *sii;
|
|
wci2_rxfifo_info_t *wci2_info;
|
|
|
|
sii = SI_INFO(sih);
|
|
ASSERT(sii);
|
|
|
|
if (!(sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT)) {
|
|
SI_ERROR(("si_wci2_rxfifo_handler_unregister: not GCI capable\n"));
|
|
return;
|
|
}
|
|
|
|
wci2_info = sii->wci2_info;
|
|
|
|
if (wci2_info == NULL) {
|
|
return;
|
|
}
|
|
|
|
if (wci2_info->rx_buf != NULL) {
|
|
MFREE(sii->osh, wci2_info->rx_buf, WCI2_UART_RX_BUF_SIZE);
|
|
}
|
|
|
|
if (wci2_info->cbs != NULL) {
|
|
MFREE(sii->osh, wci2_info->cbs, sizeof(wci2_cbs_t));
|
|
}
|
|
|
|
MFREE(sii->osh, wci2_info, sizeof(wci2_rxfifo_info_t));
|
|
|
|
}
|
|
|
|
/* GCI WCI2 UART RXFIFO interrupt handler */
|
|
static void
|
|
si_wci2_rxfifo_intr_handler_process(si_t *sih, uint32 intstatus)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 udata;
|
|
char ubyte;
|
|
wci2_rxfifo_info_t *wci2_info;
|
|
bool call_cb = FALSE;
|
|
|
|
wci2_info = sii->wci2_info;
|
|
|
|
if (wci2_info == NULL) {
|
|
return;
|
|
}
|
|
|
|
if (intstatus & GCI_INTSTATUS_SRFOF) {
|
|
SI_ERROR(("*** rx fifo overflow *** \n"));
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_intstat),
|
|
GCI_INTSTATUS_SRFOF, GCI_INTSTATUS_SRFOF);
|
|
}
|
|
|
|
/* Check if RF FIFO has any data */
|
|
if (intstatus & GCI_INTMASK_SRFNE) {
|
|
|
|
/* Read seci uart data */
|
|
udata = si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciuartdata), 0, 0);
|
|
|
|
while (udata & SECI_UART_DATA_RF_NOT_EMPTY_BIT) {
|
|
|
|
ubyte = (char) udata;
|
|
if (wci2_info) {
|
|
wci2_info->rx_buf[wci2_info->rx_idx] = ubyte;
|
|
wci2_info->rx_idx++;
|
|
call_cb = TRUE;
|
|
/* if the buffer is full, break
|
|
* remaining will be processed in next callback
|
|
*/
|
|
if (wci2_info->rx_idx == WCI2_UART_RX_BUF_SIZE) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
udata = si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciuartdata), 0, 0);
|
|
}
|
|
|
|
/* if callback registered; call it */
|
|
if (call_cb && wci2_info && wci2_info->cbs) {
|
|
wci2_info->cbs->handler(wci2_info->cbs->context, wci2_info->rx_buf,
|
|
wci2_info->rx_idx);
|
|
bzero(wci2_info->rx_buf, WCI2_UART_RX_BUF_SIZE);
|
|
wci2_info->rx_idx = 0;
|
|
}
|
|
}
|
|
}
|
|
#endif /* BCMLTECOEX && !WLTEST */
|
|
|
|
#ifdef BCMLTECOEX
|
|
/* Program GCI GpioMask and GCI GpioControl Registers */
|
|
static void
|
|
si_config_gcigpio(si_t *sih, uint32 gci_pos, uint8 gcigpio,
|
|
uint8 gpioctl_mask, uint8 gpioctl_val)
|
|
{
|
|
uint32 indirect_idx =
|
|
GCI_REGIDX(gci_pos) | (gcigpio << GCI_GPIOIDX_OFFSET);
|
|
si_gci_indirect(sih, indirect_idx, GCI_OFFSETOF(sih, gci_gpiomask),
|
|
(1 << GCI_BITOFFSET(gci_pos)),
|
|
(1 << GCI_BITOFFSET(gci_pos)));
|
|
/* Write GPIO Configuration to GCI Registers */
|
|
si_gci_indirect(sih, gcigpio/4, GCI_OFFSETOF(sih, gci_gpioctl),
|
|
(gpioctl_mask << (gcigpio%4)*8), (gpioctl_val << (gcigpio%4)*8));
|
|
}
|
|
|
|
void
|
|
si_ercx_init(si_t *sih, uint32 ltecx_mux, uint32 ltecx_padnum,
|
|
uint32 ltecx_fnsel, uint32 ltecx_gcigpio)
|
|
{
|
|
uint8 fsync_padnum, lterx_padnum, ltetx_padnum, wlprio_padnum;
|
|
uint8 fsync_fnsel, lterx_fnsel, ltetx_fnsel, wlprio_fnsel;
|
|
uint8 fsync_gcigpio, lterx_gcigpio, ltetx_gcigpio, wlprio_gcigpio;
|
|
|
|
/* reset GCI block */
|
|
si_gci_reset(sih);
|
|
|
|
/* enable ERCX (pure gpio) mode, Keep SECI in Reset Mode Only */
|
|
/* Hopefully, keeping SECI in Reset Mode will draw lesser current */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((GCI_MODE_MASK << GCI_CCTL_SMODE_OFFSET)
|
|
|(1 << GCI_CCTL_SECIEN_OFFSET)
|
|
|(1 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIRST_OFFSET)),
|
|
((GCI_MODE_GPIO << GCI_CCTL_SMODE_OFFSET)
|
|
|(0 << GCI_CCTL_SECIEN_OFFSET)
|
|
|(1 << GCI_CCTL_RSTSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIRST_OFFSET)));
|
|
|
|
/* Extract Interface Configuration */
|
|
fsync_padnum = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_FSYNC_IDX);
|
|
lterx_padnum = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_LTERX_IDX);
|
|
ltetx_padnum = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_LTETX_IDX);
|
|
wlprio_padnum = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_WLPRIO_IDX);
|
|
|
|
fsync_fnsel = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_FSYNC_IDX);
|
|
lterx_fnsel = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_LTERX_IDX);
|
|
ltetx_fnsel = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_LTETX_IDX);
|
|
wlprio_fnsel = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_WLPRIO_IDX);
|
|
|
|
fsync_gcigpio = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_FSYNC_IDX);
|
|
lterx_gcigpio = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_LTERX_IDX);
|
|
ltetx_gcigpio = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_LTETX_IDX);
|
|
wlprio_gcigpio = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_WLPRIO_IDX);
|
|
|
|
/* Clear this Function Select for all GPIOs if programmed by default */
|
|
si_gci_clear_functionsel(sih, fsync_fnsel);
|
|
si_gci_clear_functionsel(sih, lterx_fnsel);
|
|
si_gci_clear_functionsel(sih, ltetx_fnsel);
|
|
si_gci_clear_functionsel(sih, wlprio_fnsel);
|
|
|
|
/* Program Function select for selected GPIOs */
|
|
si_gci_set_functionsel(sih, fsync_padnum, fsync_fnsel);
|
|
si_gci_set_functionsel(sih, lterx_padnum, lterx_fnsel);
|
|
si_gci_set_functionsel(sih, ltetx_padnum, ltetx_fnsel);
|
|
si_gci_set_functionsel(sih, wlprio_padnum, wlprio_fnsel);
|
|
|
|
/* NOTE: We are keeping Input PADs in Pull Down Mode to take care of the case
|
|
* when LTE Modem doesn't drive these lines for any reason.
|
|
* We should consider alternate ways to identify this situation and dynamically
|
|
* enable Pull Down PAD only when LTE Modem doesn't drive these lines.
|
|
*/
|
|
|
|
/* Configure Frame Sync as input */
|
|
si_config_gcigpio(sih, GCI_LTE_FRAMESYNC_POS, fsync_gcigpio, 0xFF,
|
|
((1 << GCI_GPIOCTL_INEN_OFFSET)|(1 << GCI_GPIOCTL_PDN_OFFSET)));
|
|
|
|
/* Configure LTE Rx as input */
|
|
si_config_gcigpio(sih, GCI_LTE_RX_POS, lterx_gcigpio, 0xFF,
|
|
((1 << GCI_GPIOCTL_INEN_OFFSET)|(1 << GCI_GPIOCTL_PDN_OFFSET)));
|
|
|
|
/* Configure LTE Tx as input */
|
|
si_config_gcigpio(sih, GCI_LTE_TX_POS, ltetx_gcigpio, 0xFF,
|
|
((1 << GCI_GPIOCTL_INEN_OFFSET)|(1 << GCI_GPIOCTL_PDN_OFFSET)));
|
|
|
|
/* Configure WLAN Prio as output. BT Need to configure its ISM Prio separately
|
|
* NOTE: LTE chip has to enable its internal pull-down whenever WL goes down
|
|
*/
|
|
si_config_gcigpio(sih, GCI_WLAN_PRIO_POS, wlprio_gcigpio, 0xFF,
|
|
(1 << GCI_GPIOCTL_OUTEN_OFFSET));
|
|
|
|
/* Enable inbandIntMask for FrmSync only, disable LTE_Rx and LTE_Tx
|
|
* Note: FrameSync, LTE Rx & LTE Tx happen to share the same REGIDX
|
|
* Hence a single Access is sufficient
|
|
*/
|
|
si_gci_indirect(sih, GCI_REGIDX(GCI_LTE_FRAMESYNC_POS),
|
|
GCI_OFFSETOF(sih, gci_inbandeventintmask),
|
|
((1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS))
|
|
|(1 << GCI_BITOFFSET(GCI_LTE_RX_POS))
|
|
|(1 << GCI_BITOFFSET(GCI_LTE_TX_POS))),
|
|
((1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS))
|
|
|(0 << GCI_BITOFFSET(GCI_LTE_RX_POS))
|
|
|(0 << GCI_BITOFFSET(GCI_LTE_TX_POS))));
|
|
|
|
/* Enable Inband interrupt polarity for LTE_FRMSYNC */
|
|
si_gci_indirect(sih, GCI_REGIDX(GCI_LTE_FRAMESYNC_POS),
|
|
GCI_OFFSETOF(sih, gci_intpolreg),
|
|
(1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS)),
|
|
(1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS)));
|
|
}
|
|
|
|
void
|
|
si_wci2_init(si_t *sih, uint8 baudrate, uint32 ltecx_mux, uint32 ltecx_padnum,
|
|
uint32 ltecx_fnsel, uint32 ltecx_gcigpio, uint32 xtalfreq)
|
|
{
|
|
/* BCMLTECOEXGCI_ENAB should be checked before calling si_wci2_init() */
|
|
uint8 baud = baudrate;
|
|
uint8 seciin, seciout, fnselin, fnselout, gcigpioin, gcigpioout;
|
|
|
|
/* Extract PAD GPIO number (1-byte) from "ltecx_padnum" for each LTECX pin */
|
|
seciin = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_WCI2IN_IDX);
|
|
seciout = LTECX_EXTRACT_PADNUM(ltecx_padnum, LTECX_NVRAM_WCI2OUT_IDX);
|
|
/* Extract FunctionSel (1-nibble) from "ltecx_fnsel" for each LTECX pin */
|
|
fnselin = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_WCI2IN_IDX);
|
|
fnselout = LTECX_EXTRACT_FNSEL(ltecx_fnsel, LTECX_NVRAM_WCI2OUT_IDX);
|
|
/* Extract GCI-GPIO number (1-nibble) from "ltecx_gcigpio" for each LTECX pin */
|
|
gcigpioin = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_WCI2IN_IDX);
|
|
gcigpioout = LTECX_EXTRACT_GCIGPIO(ltecx_gcigpio, LTECX_NVRAM_WCI2OUT_IDX);
|
|
|
|
/* reset GCI block */
|
|
si_gci_reset(sih);
|
|
|
|
/* NOTE: Writing Reserved bits of older GCI Revs is OK */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((GCI_CCTL_SCS_MASK << GCI_CCTL_SCS_OFFSET)
|
|
|(GCI_CCTL_LOWTOUT_MASK << GCI_CCTL_SILOWTOUT_OFFSET)
|
|
|(1 << GCI_CCTL_BRKONSLP_OFFSET)
|
|
|(1 << GCI_CCTL_US_OFFSET)
|
|
|(GCI_MODE_MASK << GCI_CCTL_SMODE_OFFSET)
|
|
|(1 << GCI_CCTL_FSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIEN_OFFSET)),
|
|
((GCI_CCTL_SCS_DEF << GCI_CCTL_SCS_OFFSET)
|
|
|(GCI_CCTL_LOWTOUT_30BIT << GCI_CCTL_SILOWTOUT_OFFSET)
|
|
|(0 << GCI_CCTL_BRKONSLP_OFFSET)
|
|
|(0 << GCI_CCTL_US_OFFSET)
|
|
|(GCI_MODE_BTSIG << GCI_CCTL_SMODE_OFFSET)
|
|
|(0 << GCI_CCTL_FSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIEN_OFFSET))); /* 19000024 */
|
|
|
|
/* Program Function select for selected GPIOs */
|
|
si_gci_set_functionsel(sih, seciin, fnselin);
|
|
si_gci_set_functionsel(sih, seciout, fnselout);
|
|
|
|
/* Enable inbandIntMask for FrmSync only; disable LTE_Rx and LTE_Tx
|
|
* Note: FrameSync, LTE Rx & LTE Tx happen to share the same REGIDX
|
|
* Hence a single Access is sufficient
|
|
*/
|
|
si_gci_indirect(sih,
|
|
GCI_REGIDX(GCI_LTE_FRAMESYNC_POS),
|
|
GCI_OFFSETOF(sih, gci_inbandeventintmask),
|
|
((1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS))
|
|
|(1 << GCI_BITOFFSET(GCI_LTE_RX_POS))
|
|
|(1 << GCI_BITOFFSET(GCI_LTE_TX_POS))),
|
|
((1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS))
|
|
|(0 << GCI_BITOFFSET(GCI_LTE_RX_POS))
|
|
|(0 << GCI_BITOFFSET(GCI_LTE_TX_POS))));
|
|
|
|
if (GCIREV(sih->gcirev) >= 1) {
|
|
/* Program inband interrupt polarity as posedge for FrameSync */
|
|
si_gci_indirect(sih, GCI_REGIDX(GCI_LTE_FRAMESYNC_POS),
|
|
GCI_OFFSETOF(sih, gci_intpolreg),
|
|
(1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS)),
|
|
(1 << GCI_BITOFFSET(GCI_LTE_FRAMESYNC_POS)));
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 4) {
|
|
/* Program SECI_IN Control Register */
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_seciin_ctrl), ALLONES_32,
|
|
((GCI_MODE_BTSIG << GCI_SECIIN_MODE_OFFSET)
|
|
|(gcigpioin << GCI_SECIIN_GCIGPIO_OFFSET)
|
|
|(GCI_LTE_IP_ID << GCI_SECIIN_RXID2IP_OFFSET)));
|
|
|
|
/* Program GPIO Control Register for SECI_IN GCI GPIO */
|
|
si_gci_indirect(sih, gcigpioin/4, GCI_OFFSETOF(sih, gci_gpioctl),
|
|
(0xFF << (gcigpioin%4)*8),
|
|
(((1 << GCI_GPIOCTL_INEN_OFFSET)
|
|
|(1 << GCI_GPIOCTL_PDN_OFFSET)) << (gcigpioin%4)*8));
|
|
|
|
/* Program SECI_OUT Control Register */
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_seciout_ctrl), ALLONES_32,
|
|
((GCI_MODE_BTSIG << GCI_SECIOUT_MODE_OFFSET)
|
|
|(gcigpioout << GCI_SECIOUT_GCIGPIO_OFFSET)
|
|
|((1 << GCI_LTECX_SECI_ID) << GCI_SECIOUT_SECIINRELATED_OFFSET)));
|
|
|
|
/* Program GPIO Control Register for SECI_OUT GCI GPIO */
|
|
si_gci_indirect(sih, gcigpioout/4, GCI_OFFSETOF(sih, gci_gpioctl),
|
|
(0xFF << (gcigpioout%4)*8),
|
|
(((1 << GCI_GPIOCTL_OUTEN_OFFSET)) << (gcigpioout%4)*8));
|
|
|
|
/* Program SECI_IN Aux FIFO enable for LTECX SECI_IN Port */
|
|
if (GCIREV(sih->gcirev) >= 16) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_seciin_auxfifo_en),
|
|
(((1 << GCI_LTECX_SECI_ID) << GCI_SECIAUX_RXENABLE_OFFSET)
|
|
|((1 << GCI_LTECX_SECI_ID) << GCI_SECIFIFO_RXENABLE_OFFSET)),
|
|
(((1 << GCI_LTECX_SECI_ID) << GCI_SECIAUX_RXENABLE_OFFSET)
|
|
|((1 << GCI_LTECX_SECI_ID) << GCI_SECIFIFO_RXENABLE_OFFSET)));
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciin_auxfifo_en),
|
|
(((1 << GCI_LTECX_SECI_ID) << GCI_SECIAUX_RXENABLE_OFFSET)
|
|
|((1 << GCI_LTECX_SECI_ID) << GCI_SECIFIFO_RXENABLE_OFFSET)),
|
|
(((1 << GCI_LTECX_SECI_ID) << GCI_SECIAUX_RXENABLE_OFFSET)
|
|
|((1 << GCI_LTECX_SECI_ID) << GCI_SECIFIFO_RXENABLE_OFFSET)));
|
|
}
|
|
/* Program SECI_OUT Tx Enable for LTECX SECI_OUT Port */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciout_txen_txbr), ALLONES_32,
|
|
((1 << GCI_LTECX_SECI_ID) << GCI_SECITX_ENABLE_OFFSET));
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 5) {
|
|
/* enable WlPrio/TxOn override from D11 */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_miscctl),
|
|
(1 << GCI_LTECX_TXCONF_EN_OFFSET | 1 << GCI_LTECX_PRISEL_EN_OFFSET),
|
|
(1 << GCI_LTECX_TXCONF_EN_OFFSET | 1 << GCI_LTECX_PRISEL_EN_OFFSET));
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_miscctl),
|
|
(1 << GCI_LTECX_TXCONF_EN_OFFSET | 1 << GCI_LTECX_PRISEL_EN_OFFSET),
|
|
0x0000);
|
|
}
|
|
/* baudrate: 1/2/3/4mbps, escseq:0xdb, high baudrate, enable seci_tx/rx */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secifcr), ALLONES_32, 0x00);
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secilcr),
|
|
ALLONES_32, 0x00);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secilcr), ALLONES_32, 0x00);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secilcr), ALLONES_32, 0x28);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_seciuartescval), ALLONES_32, 0xDB);
|
|
|
|
switch (baud) {
|
|
case 1:
|
|
/* baudrate:1mbps */
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xFE);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xFE);
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x80);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x80);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x81);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x23);
|
|
break;
|
|
|
|
case 2:
|
|
/* baudrate:2mbps */
|
|
if (xtalfreq == XTAL_FREQ_26000KHZ) {
|
|
/* 43430 A0 uses 26 MHz crystal.
|
|
* Baudrate settings for crystel freq 26 MHz
|
|
*/
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xFF);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xFF);
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secimcr), ALLONES_32, 0x80);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x80);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x0);
|
|
}
|
|
else {
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xFF);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xFF);
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secimcr), ALLONES_32, 0x80);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x80);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x81);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x11);
|
|
}
|
|
break;
|
|
|
|
case 4:
|
|
/* baudrate:4mbps */
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF7);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF7);
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x9);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x0);
|
|
break;
|
|
|
|
case 25:
|
|
/* baudrate:2.5mbps */
|
|
if (xtalfreq == XTAL_FREQ_26000KHZ) {
|
|
/* 43430 A0 uses 26 MHz crystal.
|
|
* Baudrate settings for crystel freq 26 MHz
|
|
*/
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xF6);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF6);
|
|
}
|
|
} else if (xtalfreq == XTAL_FREQ_59970KHZ) {
|
|
/* 4387 uses 60M MHz crystal.
|
|
* Baudrate settings for crystel freq/2 29.9 MHz
|
|
* set bauddiv to 0xF4 to achieve 2.5M for Xtal/2 @ 29.9MHz
|
|
* bauddiv = 256-Integer Part of (GCI clk freq/baudrate)
|
|
*/
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xF4);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF4);
|
|
}
|
|
} else {
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xF1);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF1);
|
|
}
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x9);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x0);
|
|
break;
|
|
|
|
case 3:
|
|
default:
|
|
/* baudrate:3mbps */
|
|
if (xtalfreq == XTAL_FREQ_26000KHZ) {
|
|
/* 43430 A0 uses 26 MHz crystal.
|
|
* Baudrate settings for crystel freq 26 MHz
|
|
*/
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xF7);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF7);
|
|
}
|
|
} else {
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID,
|
|
GCI_OFFSETOF(sih, gci_secibauddiv), ALLONES_32, 0xF4);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secibauddiv),
|
|
ALLONES_32, 0xF4);
|
|
}
|
|
}
|
|
if (GCIREV(sih->gcirev) >= 15) {
|
|
si_gci_indirect(sih, GCI_LTECX_SECI_ID, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else if (GCIREV(sih->gcirev) >= 4) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x8);
|
|
} else {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_secimcr),
|
|
ALLONES_32, 0x9);
|
|
}
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_baudadj),
|
|
ALLONES_32, 0x0);
|
|
break;
|
|
}
|
|
/* GCI Rev >= 1 */
|
|
if (GCIREV(sih->gcirev) >= 1) {
|
|
/* Route Rx-data through AUX register */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_rxfifo_common_ctrl),
|
|
GCI_RXFIFO_CTRL_AUX_EN, GCI_RXFIFO_CTRL_AUX_EN);
|
|
#if !defined(WLTEST)
|
|
/* Route RX Type 2 data through RX FIFO */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_rxfifo_common_ctrl),
|
|
GCI_RXFIFO_CTRL_FIFO_TYPE2_EN, GCI_RXFIFO_CTRL_FIFO_TYPE2_EN);
|
|
/* Enable Inband interrupt for RX FIFO status */
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_intmask),
|
|
(GCI_INTSTATUS_SRFNE | GCI_INTSTATUS_SRFOF),
|
|
(GCI_INTSTATUS_SRFNE | GCI_INTSTATUS_SRFOF));
|
|
#endif /* !WLTEST */
|
|
} else {
|
|
/* GPIO 3-7 as BT_SIG complaint */
|
|
/* config GPIO pins 3-7 as input */
|
|
si_gci_indirect(sih, 0,
|
|
GCI_OFFSETOF(sih, gci_gpioctl), 0x20000000, 0x20000010);
|
|
si_gci_indirect(sih, 1,
|
|
GCI_OFFSETOF(sih, gci_gpioctl), 0x20202020, 0x20202020);
|
|
/* gpio mapping: frmsync-gpio7, mws_rx-gpio6, mws_tx-gpio5,
|
|
* pat[0]-gpio4, pat[1]-gpio3
|
|
*/
|
|
si_gci_indirect(sih, 0x70010,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x00000001, 0x00000001);
|
|
si_gci_indirect(sih, 0x60010,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x00000002, 0x00000002);
|
|
si_gci_indirect(sih, 0x50010,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x00000004, 0x00000004);
|
|
si_gci_indirect(sih, 0x40010,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x02000000, 0x00000008);
|
|
si_gci_indirect(sih, 0x30010,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x04000000, 0x04000010);
|
|
/* gpio mapping: wlan_rx_prio-gpio5, wlan_tx_on-gpio4 */
|
|
si_gci_indirect(sih, 0x50000,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x00000010, 0x00000010);
|
|
si_gci_indirect(sih, 0x40000,
|
|
GCI_OFFSETOF(sih, gci_gpiomask), 0x00000020, 0x00000020);
|
|
/* enable gpio out on gpio4(wlanrxprio), gpio5(wlantxon) */
|
|
si_gci_direct(sih,
|
|
GCI_OFFSETOF(sih, gci_control_0), 0x00000030, 0x00000000);
|
|
}
|
|
}
|
|
#endif /* BCMLTECOEX */
|
|
|
|
/* This function is used in AIBSS mode by BTCX to enable strobing to BT */
|
|
bool
|
|
si_btcx_wci2_init(si_t *sih)
|
|
{
|
|
/* reset GCI block */
|
|
si_gci_reset(sih);
|
|
|
|
if (GCIREV(sih->gcirev) >= 1) {
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_corectrl),
|
|
((GCI_CCTL_SCS_MASK << GCI_CCTL_SCS_OFFSET)
|
|
|(GCI_CCTL_LOWTOUT_MASK << GCI_CCTL_SILOWTOUT_OFFSET)
|
|
|(1 << GCI_CCTL_BRKONSLP_OFFSET)
|
|
|(1 << GCI_CCTL_US_OFFSET)
|
|
|(GCI_MODE_MASK << GCI_CCTL_SMODE_OFFSET)
|
|
|(1 << GCI_CCTL_FSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIEN_OFFSET)),
|
|
((GCI_CCTL_SCS_DEF << GCI_CCTL_SCS_OFFSET)
|
|
|(GCI_CCTL_LOWTOUT_30BIT << GCI_CCTL_SILOWTOUT_OFFSET)
|
|
|(0 << GCI_CCTL_BRKONSLP_OFFSET)
|
|
|(0 << GCI_CCTL_US_OFFSET)
|
|
|(GCI_MODE_BTSIG << GCI_CCTL_SMODE_OFFSET)
|
|
|(0 << GCI_CCTL_FSL_OFFSET)
|
|
|(1 << GCI_CCTL_SECIEN_OFFSET))); /* 19000024 */
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
void
|
|
si_gci_uart_init(si_t *sih, osl_t *osh, uint8 seci_mode)
|
|
{
|
|
#ifdef HNDGCI
|
|
hndgci_init(sih, osh, HND_GCI_PLAIN_UART_MODE,
|
|
GCI_UART_BR_115200);
|
|
|
|
/* specify rx callback */
|
|
hndgci_uart_config_rx_complete(-1, -1, 0, NULL, NULL);
|
|
#else
|
|
BCM_REFERENCE(sih);
|
|
BCM_REFERENCE(osh);
|
|
BCM_REFERENCE(seci_mode);
|
|
#endif /* HNDGCI */
|
|
}
|
|
|
|
/* input: pin number
|
|
* output: chipcontrol reg(ring_index base) and
|
|
* bits to shift for pin first regbit.
|
|
* eg: gpio9 will give regidx: 2 and pos 16
|
|
*/
|
|
static uint8
|
|
BCMPOSTTRAPFN(si_gci_get_chipctrlreg_ringidx_base8)(uint32 pin, uint32 *regidx, uint32 *pos)
|
|
{
|
|
*regidx = (pin / 4);
|
|
*pos = (pin % 4)*8;
|
|
|
|
SI_MSG(("si_gci_get_chipctrlreg_ringidx_base8:%d:%d:%d\n", pin, *regidx, *pos));
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
/**
|
|
* A given GCI pin needs to be converted to a GCI FunctionSel register offset and the bit position
|
|
* in this register.
|
|
* @param[in] input pin number, see respective chip Toplevel Arch page, GCI chipstatus regs
|
|
* @param[out] regidx chipcontrol reg(ring_index base) and
|
|
* @param[out] pos bits to shift for pin first regbit
|
|
*
|
|
* eg: gpio9 will give regidx: 1 and pos 4
|
|
*/
|
|
static void
|
|
BCMPOSTTRAPFN(si_gci_get_chipctrlreg_ringidx_base4)(uint32 pin, uint32 *regidx, uint32 *pos)
|
|
{
|
|
*regidx = (pin / 8);
|
|
*pos = (pin % 8) * 4; // each pin occupies 4 FunctionSel register bits
|
|
|
|
SI_MSG(("si_gci_get_chipctrlreg_ringidx_base4:%d:%d:%d\n", pin, *regidx, *pos));
|
|
}
|
|
|
|
/** setup a given pin for fnsel function */
|
|
void
|
|
BCMPOSTTRAPFN(si_gci_set_functionsel)(si_t *sih, uint32 pin, uint8 fnsel)
|
|
{
|
|
uint32 reg = 0, pos = 0;
|
|
|
|
SI_MSG(("si_gci_set_functionsel:%d\n", pin));
|
|
|
|
si_gci_get_chipctrlreg_ringidx_base4(pin, ®, &pos);
|
|
si_gci_chipcontrol(sih, reg, GCIMASK_4B(pos), GCIPOSVAL_4B(fnsel, pos));
|
|
}
|
|
|
|
/* Returns a given pin's fnsel value */
|
|
uint32
|
|
si_gci_get_functionsel(si_t *sih, uint32 pin)
|
|
{
|
|
uint32 reg = 0, pos = 0, temp;
|
|
|
|
SI_MSG(("si_gci_get_functionsel: %d\n", pin));
|
|
|
|
si_gci_get_chipctrlreg_ringidx_base4(pin, ®, &pos);
|
|
temp = si_gci_chipstatus(sih, reg);
|
|
return GCIGETNBL(temp, pos);
|
|
}
|
|
|
|
/* Sets fnsel value to IND for all the GPIO pads that have fnsel set to given argument */
|
|
void
|
|
si_gci_clear_functionsel(si_t *sih, uint8 fnsel)
|
|
{
|
|
uint32 i;
|
|
SI_MSG(("si_gci_clear_functionsel: %d\n", fnsel));
|
|
for (i = 0; i <= CC_PIN_GPIO_LAST; i++) {
|
|
if (si_gci_get_functionsel(sih, i) == fnsel)
|
|
si_gci_set_functionsel(sih, i, CC_FNSEL_IND);
|
|
}
|
|
}
|
|
|
|
/** write 'val' to the gci chip control register indexed by 'reg' */
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gci_chipcontrol)(si_t *sih, uint reg, uint32 mask, uint32 val)
|
|
{
|
|
/* because NFLASH and GCI clashes in 0xC00 */
|
|
if ((CCREV(sih->ccrev) == 38) && ((sih->chipst & (1 << 4)) != 0)) {
|
|
/* CC NFLASH exist, prohibit to manipulate gci register */
|
|
ASSERT(0);
|
|
return ALLONES_32;
|
|
}
|
|
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, reg);
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_chipctrl), mask, val);
|
|
}
|
|
|
|
/* Read the gci chip status register indexed by 'reg' */
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gci_chipstatus)(si_t *sih, uint reg)
|
|
{
|
|
/* because NFLASH and GCI clashes in 0xC00 */
|
|
if ((CCREV(sih->ccrev) == 38) && ((sih->chipst & (1 << 4)) != 0)) {
|
|
/* CC NFLASH exist, prohibit to manipulate gci register */
|
|
ASSERT(0);
|
|
return ALLONES_32;
|
|
}
|
|
|
|
si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_indirect_addr), ~0, reg);
|
|
/* setting mask and value to '0' to use si_corereg for read only purpose */
|
|
return si_corereg(sih, GCI_CORE_IDX(sih), GCI_OFFSETOF(sih, gci_chipsts), 0, 0);
|
|
}
|
|
|
|
void
|
|
sflash_gpio_config(si_t *sih)
|
|
{
|
|
ASSERT(sih);
|
|
|
|
if (!si_is_sflash_available(sih)) {
|
|
return;
|
|
}
|
|
|
|
switch (CHIPID((sih)->chip)) {
|
|
case BCM4387_CHIP_GRPID:
|
|
/* config 4387 sflash gpio lines 12,15,18,19 */
|
|
si_gci_set_functionsel(sih, CC_PIN_GPIO_12, CC4387_FNSEL_SFLASH);
|
|
si_gci_set_functionsel(sih, CC_PIN_GPIO_15, CC4387_FNSEL_SFLASH);
|
|
si_gci_set_functionsel(sih, CC_PIN_GPIO_18, CC4387_FNSEL_SFLASH);
|
|
si_gci_set_functionsel(sih, CC_PIN_GPIO_19, CC4387_FNSEL_SFLASH);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
uint16
|
|
si_chipid(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
return (sii->chipnew) ? sii->chipnew : sih->chip;
|
|
}
|
|
|
|
/* CHIP_ID's being mapped here should not be used anywhere else in the code */
|
|
static void
|
|
si_chipid_fixup(si_t *sih)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT(sii->chipnew == 0);
|
|
switch (sih->chip) {
|
|
case BCM4377_CHIP_ID:
|
|
sii->chipnew = sih->chip; /* save it */
|
|
sii->pub.chip = BCM4369_CHIP_ID; /* chip class */
|
|
break;
|
|
case BCM4375_CHIP_ID:
|
|
sii->chipnew = sih->chip; /* save it */
|
|
sii->pub.chip = BCM4375_CHIP_ID; /* chip class */
|
|
break;
|
|
case BCM4362_CHIP_ID:
|
|
sii->chipnew = sih->chip; /* save it */
|
|
sii->pub.chip = BCM4362_CHIP_ID; /* chip class */
|
|
break;
|
|
case BCM4356_CHIP_ID:
|
|
case BCM4371_CHIP_ID:
|
|
sii->chipnew = sih->chip; /* save it */
|
|
sii->pub.chip = BCM4354_CHIP_ID; /* chip class */
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
#ifdef AXI_TIMEOUTS_NIC
|
|
uint32
|
|
BCMPOSTTRAPFN(si_clear_backplane_to_fast)(void *sih, void *addr)
|
|
{
|
|
si_t *_sih = DISCARD_QUAL(sih, si_t);
|
|
|
|
if (CHIPTYPE(_sih->socitype) == SOCI_AI) {
|
|
return ai_clear_backplane_to_fast(_sih, addr);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
const si_axi_error_info_t *
|
|
si_get_axi_errlog_info(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
return (const si_axi_error_info_t *)sih->err_info;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
si_reset_axi_errlog_info(const si_t *sih)
|
|
{
|
|
if (sih->err_info) {
|
|
sih->err_info->count = 0;
|
|
}
|
|
}
|
|
#endif /* AXI_TIMEOUTS_NIC */
|
|
|
|
/* TODO: Can we allocate only one instance? */
|
|
static int32
|
|
si_alloc_wrapper(si_info_t *sii)
|
|
{
|
|
if (sii->osh) {
|
|
sii->axi_wrapper = (axi_wrapper_t *)MALLOCZ(sii->osh,
|
|
(sizeof(axi_wrapper_t) * SI_MAX_AXI_WRAPPERS));
|
|
|
|
if (sii->axi_wrapper == NULL) {
|
|
return BCME_NOMEM;
|
|
}
|
|
} else {
|
|
sii->axi_wrapper = NULL;
|
|
return BCME_ERROR;
|
|
}
|
|
return BCME_OK;
|
|
}
|
|
|
|
static void
|
|
si_free_wrapper(si_info_t *sii)
|
|
{
|
|
if (sii->axi_wrapper) {
|
|
|
|
MFREE(sii->osh, sii->axi_wrapper, (sizeof(axi_wrapper_t) * SI_MAX_AXI_WRAPPERS));
|
|
}
|
|
}
|
|
|
|
static void *
|
|
si_alloc_coresinfo(si_info_t *sii, osl_t *osh, chipcregs_t *cc)
|
|
{
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
sii->nci_info = nci_init(&sii->pub, (void*)(uintptr)cc, sii->pub.bustype);
|
|
|
|
return sii->nci_info;
|
|
|
|
} else {
|
|
|
|
#ifdef _RTE_
|
|
sii->cores_info = (si_cores_info_t *)&ksii_cores_info;
|
|
#else
|
|
if (sii->cores_info == NULL) {
|
|
/* alloc si_cores_info_t */
|
|
if ((sii->cores_info = (si_cores_info_t *)MALLOCZ(osh,
|
|
sizeof(si_cores_info_t))) == NULL) {
|
|
SI_ERROR(("si_attach: malloc failed for cores_info! malloced"
|
|
" %d bytes\n", MALLOCED(osh)));
|
|
return (NULL);
|
|
}
|
|
} else {
|
|
ASSERT(sii->cores_info == &ksii_cores_info);
|
|
|
|
}
|
|
#endif /* _RTE_ */
|
|
return sii->cores_info;
|
|
}
|
|
|
|
}
|
|
|
|
static void
|
|
si_free_coresinfo(si_info_t *sii, osl_t *osh)
|
|
{
|
|
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
if (sii->nci_info) {
|
|
nci_uninit(sii->nci_info);
|
|
sii->nci_info = NULL;
|
|
}
|
|
} else {
|
|
if (sii->cores_info && (sii->cores_info != &ksii_cores_info)) {
|
|
MFREE(osh, sii->cores_info, sizeof(si_cores_info_t));
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Allocate an si handle. This function may be called multiple times. This function is called by
|
|
* both si_attach() and si_kattach().
|
|
*
|
|
* vars - pointer to a to-be created pointer area for "environment" variables. Some callers of this
|
|
* function set 'vars' to NULL.
|
|
*/
|
|
static si_info_t *
|
|
si_doattach(si_info_t *sii, uint devid, osl_t *osh, volatile void *regs,
|
|
uint bustype, void *sdh, char **vars, uint *varsz)
|
|
{
|
|
struct si_pub *sih = &sii->pub;
|
|
uint32 w, savewin;
|
|
chipcregs_t *cc;
|
|
char *pvars = NULL;
|
|
uint origidx;
|
|
#if defined(NVSRCX)
|
|
char *sromvars;
|
|
#endif
|
|
uint err_at = 0;
|
|
|
|
ASSERT(GOODREGS(regs));
|
|
|
|
savewin = 0;
|
|
|
|
sih->buscoreidx = BADIDX;
|
|
sii->device_removed = FALSE;
|
|
|
|
sii->curmap = regs;
|
|
sii->sdh = sdh;
|
|
sii->osh = osh;
|
|
sii->second_bar0win = ~0x0;
|
|
sih->enum_base = si_enum_base(devid);
|
|
|
|
#if defined(AXI_TIMEOUTS_NIC)
|
|
sih->err_info = MALLOCZ(osh, sizeof(si_axi_error_info_t));
|
|
if (sih->err_info == NULL) {
|
|
SI_ERROR(("si_doattach: %zu bytes MALLOC FAILED",
|
|
sizeof(si_axi_error_info_t)));
|
|
}
|
|
#endif /* AXI_TIMEOUTS_NIC */
|
|
|
|
#if defined(AXI_TIMEOUTS_NIC) && defined(__linux__)
|
|
osl_set_bpt_cb(osh, (void *)si_clear_backplane_to_fast, (void *)sih);
|
|
#endif /* AXI_TIMEOUTS_NIC && linux */
|
|
|
|
/* check to see if we are a si core mimic'ing a pci core */
|
|
if ((bustype == PCI_BUS) &&
|
|
(OSL_PCI_READ_CONFIG(sii->osh, PCI_SPROM_CONTROL, sizeof(uint32)) == 0xffffffff)) {
|
|
SI_ERROR(("si_doattach: incoming bus is PCI but it's a lie, switching to SI "
|
|
"devid:0x%x\n", devid));
|
|
bustype = SI_BUS;
|
|
}
|
|
|
|
/* find Chipcommon address */
|
|
if (bustype == PCI_BUS) {
|
|
savewin = OSL_PCI_READ_CONFIG(sii->osh, PCI_BAR0_WIN, sizeof(uint32));
|
|
/* PR 29857: init to core0 if bar0window is not programmed properly */
|
|
if (!GOODCOREADDR(savewin, SI_ENUM_BASE(sih)))
|
|
savewin = SI_ENUM_BASE(sih);
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_BAR0_WIN, 4, SI_ENUM_BASE(sih));
|
|
if (!regs) {
|
|
err_at = 1;
|
|
goto exit;
|
|
}
|
|
cc = (chipcregs_t *)regs;
|
|
#ifdef BCMSDIO
|
|
} else if ((bustype == SDIO_BUS) || (bustype == SPI_BUS)) {
|
|
cc = (chipcregs_t *)sii->curmap;
|
|
#endif
|
|
} else {
|
|
cc = (chipcregs_t *)REG_MAP(SI_ENUM_BASE(sih), SI_CORE_SIZE);
|
|
}
|
|
|
|
sih->bustype = (uint16)bustype;
|
|
#ifdef BCMBUSTYPE
|
|
if (bustype != BUSTYPE(bustype)) {
|
|
SI_ERROR(("si_doattach: bus type %d does not match configured bus type %d\n",
|
|
bustype, BUSTYPE(bustype)));
|
|
err_at = 2;
|
|
goto exit;
|
|
}
|
|
#endif
|
|
|
|
/* bus/core/clk setup for register access */
|
|
if (!si_buscore_prep(sii, bustype, devid, sdh)) {
|
|
SI_ERROR(("si_doattach: si_core_clk_prep failed %d\n", bustype));
|
|
err_at = 3;
|
|
goto exit;
|
|
}
|
|
|
|
/* ChipID recognition.
|
|
* We assume we can read chipid at offset 0 from the regs arg.
|
|
* If we add other chiptypes (or if we need to support old sdio hosts w/o chipcommon),
|
|
* some way of recognizing them needs to be added here.
|
|
*/
|
|
if (!cc) {
|
|
err_at = 3;
|
|
goto exit;
|
|
}
|
|
w = R_REG(osh, &cc->chipid);
|
|
#if defined(BCMDONGLEHOST)
|
|
/* plz refer to RB:13157 */
|
|
if ((w & 0xfffff) == 148277) w -= 65532;
|
|
#endif /* defined(BCMDONGLEHOST) */
|
|
sih->socitype = (w & CID_TYPE_MASK) >> CID_TYPE_SHIFT;
|
|
/* Might as wll fill in chip id rev & pkg */
|
|
sih->chip = w & CID_ID_MASK;
|
|
sih->chiprev = (w & CID_REV_MASK) >> CID_REV_SHIFT;
|
|
sih->chippkg = (w & CID_PKG_MASK) >> CID_PKG_SHIFT;
|
|
|
|
#if defined(BCMSDIO) && (defined(HW_OOB) || defined(FORCE_WOWLAN))
|
|
dhd_conf_set_hw_oob_intr(sdh, sih);
|
|
#endif
|
|
|
|
si_chipid_fixup(sih);
|
|
|
|
sih->issim = IS_SIM(sih->chippkg);
|
|
|
|
if (MULTIBP_CAP(sih)) {
|
|
sih->_multibp_enable = TRUE;
|
|
}
|
|
|
|
/* scan for cores */
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
|
|
if (si_alloc_coresinfo(sii, osh, cc) == NULL) {
|
|
err_at = 4;
|
|
goto exit;
|
|
}
|
|
ASSERT(sii->nci_info);
|
|
|
|
#ifndef BCM_BOOTLOADER
|
|
if ((si_alloc_wrapper(sii)) != BCME_OK) {
|
|
err_at = 5;
|
|
goto exit;
|
|
}
|
|
#endif /* BCM_BOOTLOADER */
|
|
|
|
#ifndef SOCI_NCI_BUS
|
|
/* If !SOCI_NCI_BUS, nci_scan(sih) is always 0. */
|
|
err_at = 6;
|
|
goto exit;
|
|
#else
|
|
if ((sii->numcores = nci_scan(sih)) == 0u) {
|
|
err_at = 6;
|
|
goto exit;
|
|
} else {
|
|
#ifndef BCM_BOOTLOADER
|
|
nci_dump_erom(sii->nci_info);
|
|
#endif /* BCM_BOOTLOADER */
|
|
}
|
|
#endif /* !SOCI_NCI_BUS */
|
|
} else {
|
|
|
|
if (si_alloc_coresinfo(sii, osh, cc) == NULL) {
|
|
err_at = 7;
|
|
goto exit;
|
|
}
|
|
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_SB) {
|
|
SI_MSG(("Found chip type SB (0x%08x)\n", w));
|
|
sb_scan(&sii->pub, regs, devid);
|
|
} else if ((CHIPTYPE(sii->pub.socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sii->pub.socitype) == SOCI_NAI) ||
|
|
(CHIPTYPE(sii->pub.socitype) == SOCI_DVTBUS)) {
|
|
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_AI)
|
|
SI_MSG(("Found chip type AI (0x%08x)\n", w));
|
|
else if (CHIPTYPE(sii->pub.socitype) == SOCI_NAI)
|
|
SI_MSG(("Found chip type NAI (0x%08x)\n", w));
|
|
else
|
|
SI_MSG(("Found chip type DVT (0x%08x)\n", w));
|
|
/* pass chipc address instead of original core base */
|
|
if ((si_alloc_wrapper(sii)) != BCME_OK) {
|
|
err_at = 8;
|
|
goto exit;
|
|
}
|
|
ai_scan(&sii->pub, (void *)(uintptr)cc, devid);
|
|
/* make sure the wrappers are properly accounted for */
|
|
if (sii->axi_num_wrappers == 0) {
|
|
SI_ERROR(("FATAL: Wrapper count 0\n"));
|
|
err_at = 16;
|
|
goto exit;
|
|
}
|
|
}
|
|
else if (CHIPTYPE(sii->pub.socitype) == SOCI_UBUS) {
|
|
SI_MSG(("Found chip type UBUS (0x%08x), chip id = 0x%4x\n", w, sih->chip));
|
|
/* pass chipc address instead of original core base */
|
|
ub_scan(&sii->pub, (void *)(uintptr)cc, devid);
|
|
} else {
|
|
SI_ERROR(("Found chip of unknown type (0x%08x)\n", w));
|
|
err_at = 9;
|
|
goto exit;
|
|
}
|
|
}
|
|
/* no cores found, bail out */
|
|
if (sii->numcores == 0) {
|
|
err_at = 10;
|
|
goto exit;
|
|
}
|
|
/* bus/core/clk setup */
|
|
origidx = SI_CC_IDX;
|
|
if (!si_buscore_setup(sii, cc, bustype, savewin, &origidx, regs)) {
|
|
err_at = 11;
|
|
goto exit;
|
|
}
|
|
|
|
/* JIRA: SWWLAN-98321: SPROM read showing wrong values */
|
|
/* Set the clkdiv2 divisor bits (2:0) to 0x4 if srom is present */
|
|
if (bustype == SI_BUS) {
|
|
uint32 clkdiv2, sromprsnt, capabilities, srom_supported;
|
|
capabilities = R_REG(osh, &cc->capabilities);
|
|
srom_supported = capabilities & SROM_SUPPORTED;
|
|
if (srom_supported) {
|
|
sromprsnt = R_REG(osh, &cc->sromcontrol);
|
|
sromprsnt = sromprsnt & SROM_PRSNT_MASK;
|
|
if (sromprsnt) {
|
|
/* SROM clock come from backplane clock/div2. Must <= 1Mhz */
|
|
clkdiv2 = (R_REG(osh, &cc->clkdiv2) & ~CLKD2_SROM);
|
|
clkdiv2 |= CLKD2_SROMDIV_192;
|
|
W_REG(osh, &cc->clkdiv2, clkdiv2);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (bustype == PCI_BUS) {
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* JIRA:SWWLAN-18243: SPROM access taking too long */
|
|
/* not required for 43602 */
|
|
if (((CHIPID(sih->chip) == BCM4360_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43460_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM4352_CHIP_ID)) &&
|
|
(CHIPREV(sih->chiprev) <= 2)) {
|
|
pcie_disable_TL_clk_gating(sii->pch);
|
|
pcie_set_L1_entry_time(sii->pch, 0x40);
|
|
}
|
|
#endif /* BCMDONGLEHOST */
|
|
|
|
}
|
|
#ifdef BCM_SDRBL
|
|
/* 4360 rom bootloader in PCIE case, if the SDR is enabled, But preotection is
|
|
* not turned on, then we want to hold arm in reset.
|
|
* Bottomline: In sdrenable case, we allow arm to boot only when protection is
|
|
* turned on.
|
|
*/
|
|
if (CHIP_HOSTIF_PCIE(&(sii->pub))) {
|
|
uint32 sflags = si_arm_sflags(&(sii->pub));
|
|
|
|
/* If SDR is enabled but protection is not turned on
|
|
* then we want to force arm to WFI.
|
|
*/
|
|
if ((sflags & (SISF_SDRENABLE | SISF_TCMPROT)) == SISF_SDRENABLE) {
|
|
disable_arm_irq();
|
|
while (1) {
|
|
hnd_cpu_wait(sih);
|
|
}
|
|
}
|
|
}
|
|
#endif /* BCM_SDRBL */
|
|
#ifdef SI_SPROM_PROBE
|
|
si_sprom_init(sih);
|
|
#endif /* SI_SPROM_PROBE */
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* Init nvram from flash if it exists */
|
|
if (nvram_init(&(sii->pub)) != BCME_OK) {
|
|
SI_ERROR(("si_doattach: nvram_init failed \n"));
|
|
goto exit;
|
|
}
|
|
|
|
/* Init nvram from sprom/otp if they exist */
|
|
GCC_DIAGNOSTIC_PUSH_SUPPRESS_CAST();
|
|
|
|
#ifdef DONGLEBUILD
|
|
#if !defined(NVSRCX)
|
|
/* Init nvram from sprom/otp if they exist and not inited */
|
|
if (si_getkvars()) {
|
|
*vars = si_getkvars();
|
|
*varsz = si_getkvarsz();
|
|
}
|
|
else
|
|
#endif
|
|
#endif /* DONGLEBUILD */
|
|
{
|
|
#if defined(NVSRCX)
|
|
sromvars = srom_get_sromvars();
|
|
if (sromvars == NULL) {
|
|
if (srom_var_init(&sii->pub, BUSTYPE(bustype), (void *)regs,
|
|
sii->osh, &sromvars, varsz)) {
|
|
err_at = 12;
|
|
goto exit;
|
|
}
|
|
}
|
|
#else
|
|
if (srom_var_init(&sii->pub, BUSTYPE(bustype), (void *)regs,
|
|
sii->osh, vars, varsz)) {
|
|
err_at = 13;
|
|
goto exit;
|
|
}
|
|
#endif /* NVSRCX */
|
|
}
|
|
GCC_DIAGNOSTIC_POP();
|
|
|
|
pvars = vars ? *vars : NULL;
|
|
|
|
si_nvram_process(sii, pvars);
|
|
|
|
/* xtalfreq is required for programming open loop calibration support changes */
|
|
sii->xtalfreq = getintvar(NULL, rstr_xtalfreq);
|
|
/* === NVRAM, clock is ready === */
|
|
#else
|
|
pvars = NULL;
|
|
BCM_REFERENCE(pvars);
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
#if defined(BCMSRTOPOFF) && !defined(BCMSRTOPOFF_DISABLED)
|
|
_srtopoff_enab = (bool)getintvar(NULL, rstr_srtopoff_enab);
|
|
#endif
|
|
|
|
if (HIB_EXT_WAKEUP_CAP(sih)) {
|
|
sii->lhl_ps_mode = (uint8)getintvar(NULL, rstr_lhl_ps_mode);
|
|
|
|
if (getintvar(NULL, rstr_ext_wakeup_dis)) {
|
|
sii->hib_ext_wakeup_enab = FALSE;
|
|
} else if (BCMSRTOPOFF_ENAB()) {
|
|
/* Has GPIO false wakeup issue on 4387, needs resolve */
|
|
sii->hib_ext_wakeup_enab = TRUE;
|
|
} else if (LHL_IS_PSMODE_1(sih)) {
|
|
sii->hib_ext_wakeup_enab = TRUE;
|
|
} else {
|
|
sii->hib_ext_wakeup_enab = FALSE;
|
|
}
|
|
}
|
|
|
|
sii->rfldo3p3_war = (bool)getintvar(NULL, rstr_rfldo3p3_cap_war);
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
if (!si_onetimeinit) {
|
|
#if !defined(BCMDONGLEHOST)
|
|
char *val;
|
|
|
|
(void) val;
|
|
|
|
/* Cache nvram override to min mask */
|
|
if ((val = getvar(NULL, rstr_rmin)) != NULL) {
|
|
sii->min_mask_valid = TRUE;
|
|
sii->nvram_min_mask = (uint32)bcm_strtoul(val, NULL, 0);
|
|
} else {
|
|
sii->min_mask_valid = FALSE;
|
|
}
|
|
/* Cache nvram override to max mask */
|
|
if ((val = getvar(NULL, rstr_rmax)) != NULL) {
|
|
sii->max_mask_valid = TRUE;
|
|
sii->nvram_max_mask = (uint32)bcm_strtoul(val, NULL, 0);
|
|
} else {
|
|
sii->max_mask_valid = FALSE;
|
|
}
|
|
|
|
#ifdef DONGLEBUILD
|
|
/* Handle armclk frequency setting from NVRAM file */
|
|
if (BCM4369_CHIP(sih->chip) || BCM4362_CHIP(sih->chip) ||
|
|
BCM4389_CHIP(sih->chip) ||
|
|
BCM4388_CHIP(sih->chip) || BCM4397_CHIP(sih->chip) || FALSE) {
|
|
if ((val = getvar(NULL, rstr_armclk)) != NULL) {
|
|
sii->armpllclkfreq = (uint32)bcm_strtoul(val, NULL, 0);
|
|
ASSERT(sii->armpllclkfreq > 0);
|
|
} else {
|
|
sii->armpllclkfreq = 0;
|
|
}
|
|
}
|
|
|
|
#endif /* DONGLEBUILD */
|
|
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
#if defined(CONFIG_XIP) && defined(BCMTCAM)
|
|
/* patch the ROM if there are any patch pairs from OTP/SPROM */
|
|
if (patch_pair) {
|
|
|
|
#if defined(__ARM_ARCH_7R__)
|
|
hnd_tcam_bootloader_load(si_setcore(sih, ARMCR4_CORE_ID, 0), pvars);
|
|
#elif defined(__ARM_ARCH_7A__)
|
|
hnd_tcam_bootloader_load(si_setcore(sih, SYSMEM_CORE_ID, 0), pvars);
|
|
#else
|
|
hnd_tcam_bootloader_load(si_setcore(sih, SOCRAM_CORE_ID, 0), pvars);
|
|
#endif
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
#endif /* CONFIG_XIP && BCMTCAM */
|
|
|
|
if (CCREV(sii->pub.ccrev) >= 20) {
|
|
uint32 gpiopullup = 0, gpiopulldown = 0;
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
ASSERT(cc != NULL);
|
|
|
|
#if !defined(BCMDONGLEHOST) /* if not a DHD build */
|
|
if (getvar(pvars, rstr_gpiopulldown) != NULL) {
|
|
uint32 value;
|
|
value = getintvar(pvars, rstr_gpiopulldown);
|
|
if (value != 0xFFFFFFFF) { /* non populated SROM fields are ffff */
|
|
gpiopulldown |= value;
|
|
}
|
|
}
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
W_REG(osh, &cc->gpiopullup, gpiopullup);
|
|
W_REG(osh, &cc->gpiopulldown, gpiopulldown);
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
#ifdef DONGLEBUILD
|
|
/* Ensure gci is initialized before PMU as PLL init needs to aquire gci semaphore */
|
|
hnd_gci_init(sih);
|
|
#endif /* DONGLEBUILD */
|
|
|
|
#if defined(BT_WLAN_REG_ON_WAR)
|
|
/*
|
|
* 4389B0/C0 - WLAN and BT turn on WAR - synchronize WLAN and BT firmware using GCI
|
|
* semaphore - THREAD_0_GCI_SEM_3_ID to ensure that simultaneous register accesses
|
|
* does not occur. The WLAN firmware will acquire the semaphore just to ensure that
|
|
* if BT firmware is already executing the WAR, then wait until it finishes.
|
|
* In BT firmware checking for WL_REG_ON status is sufficient to decide whether
|
|
* to apply the WAR or not (i.e, WLAN is turned ON/OFF).
|
|
*/
|
|
if ((hnd_gcisem_acquire(GCI_BT_WLAN_REG_ON_WAR_SEM, TRUE,
|
|
GCI_BT_WLAN_REG_ON_WAR_SEM_TIMEOUT) != BCME_OK)) {
|
|
err_at = 14;
|
|
hnd_gcisem_set_err(GCI_BT_WLAN_REG_ON_WAR_SEM);
|
|
goto exit;
|
|
}
|
|
|
|
/* WLAN/BT turn On WAR - Remove wlsc_btsc_prisel override after semaphore acquire
|
|
* BT sets the override at power up when WL_REG_ON is low - wlsc_btsc_prisel is in
|
|
* undefined state when wlan_reg_on is low
|
|
*/
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_23,
|
|
(CC_GCI_CHIPCTRL_23_WLSC_BTSC_PRISEL_FORCE_MASK |
|
|
CC_GCI_CHIPCTRL_23_WLSC_BTSC_PRISEL_VAL_MASK), 0u);
|
|
|
|
if ((hnd_gcisem_release(GCI_BT_WLAN_REG_ON_WAR_SEM) != BCME_OK)) {
|
|
hnd_gcisem_set_err(GCI_BT_WLAN_REG_ON_WAR_SEM);
|
|
err_at = 15;
|
|
goto exit;
|
|
}
|
|
#endif /* BT_WLAN_REG_ON_WAR */
|
|
|
|
/* Skip PMU initialization from the Dongle Host.
|
|
* Firmware will take care of it when it comes up.
|
|
*/
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* PMU specific initializations */
|
|
if (PMUCTL_ENAB(sih)) {
|
|
uint32 xtalfreq;
|
|
si_pmu_init(sih, sii->osh);
|
|
si_pmu_chip_init(sih, sii->osh);
|
|
xtalfreq = getintvar(pvars, rstr_xtalfreq);
|
|
|
|
/*
|
|
* workaround for chips that don't support external LPO, thus ALP clock
|
|
* can not be measured accurately:
|
|
*/
|
|
switch (CHIPID(sih->chip)) {
|
|
CASE_BCM43602_CHIP:
|
|
xtalfreq = 40000;
|
|
break;
|
|
case BCM4369_CHIP_GRPID:
|
|
if (xtalfreq == 0)
|
|
xtalfreq = 37400;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/* If xtalfreq var not available, try to measure it */
|
|
if (xtalfreq == 0)
|
|
xtalfreq = si_pmu_measure_alpclk(sih, sii->osh);
|
|
|
|
sii->xtalfreq = xtalfreq;
|
|
si_pmu_pll_init(sih, sii->osh, xtalfreq);
|
|
|
|
/* configure default spurmode */
|
|
sii->spurmode = getintvar(pvars, rstr_spurconfig) & 0xf;
|
|
|
|
#if defined(SAVERESTORE)
|
|
/* Only needs to be done once.
|
|
* Needs this before si_pmu_res_init() to use sr_isenab()
|
|
*/
|
|
if (SR_ENAB()) {
|
|
sr_save_restore_init(sih);
|
|
}
|
|
#endif
|
|
|
|
/* TODO: should move the per core srpwr out of
|
|
* si_doattach() to a function where it knows
|
|
* which core it should enable the power domain
|
|
* request for...
|
|
*/
|
|
if (SRPWR_CAP(sih) && !SRPWR_ENAB()) {
|
|
uint32 domain = SRPWR_DMN3_MACMAIN_MASK;
|
|
|
|
#if defined(WLRSDB) && !defined(WLRSDB_DISABLED)
|
|
domain |= SRPWR_DMN2_MACAUX_MASK;
|
|
#endif /* WLRSDB && !WLRSDB_DISABLED */
|
|
|
|
if (si_scan_core_present(sih)) {
|
|
domain |= SRPWR_DMN4_MACSCAN_MASK;
|
|
}
|
|
|
|
si_srpwr_request(sih, domain, domain);
|
|
}
|
|
|
|
si_pmu_res_init(sih, sii->osh);
|
|
si_pmu_swreg_init(sih, sii->osh);
|
|
#ifdef BCMGCISHM
|
|
hnd_gcishm_init(sih);
|
|
#endif
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
#ifdef _RTE_
|
|
si_onetimeinit = TRUE;
|
|
#endif
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
|
|
si_lowpwr_opt(sih);
|
|
|
|
if (PCIE(sii)) {
|
|
ASSERT(sii->pch != NULL);
|
|
pcicore_attach(sii->pch, pvars, SI_DOATTACH);
|
|
}
|
|
|
|
if ((CHIPID(sih->chip) == BCM43012_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43013_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43014_CHIP_ID) ||
|
|
(CCREV(sih->ccrev) >= 62)) {
|
|
/* Clear SFlash clock request */
|
|
CHIPC_REG(sih, clk_ctl_st, CCS_SFLASH_CLKREQ, 0);
|
|
}
|
|
|
|
#ifdef SECI_UART
|
|
/* Enable pull up on fast_uart_rx and fast_uart_cts_in
|
|
* when fast uart is disabled.
|
|
*/
|
|
if (getvar(pvars, rstr_fuart_pup_rx_cts) != NULL) {
|
|
w = getintvar(pvars, rstr_fuart_pup_rx_cts);
|
|
if (w)
|
|
fuart_pullup_rx_cts_enab = TRUE;
|
|
}
|
|
#endif
|
|
|
|
/* configure default pinmux enables for the chip */
|
|
if (getvar(pvars, rstr_muxenab) != NULL) {
|
|
w = getintvar(pvars, rstr_muxenab);
|
|
si_muxenab((si_t *)sii, w);
|
|
}
|
|
|
|
/* configure default swd enables for the chip */
|
|
if (getvar(pvars, rstr_swdenab) != NULL) {
|
|
w = getintvar(pvars, rstr_swdenab);
|
|
si_swdenable((si_t *)sii, w);
|
|
}
|
|
|
|
sii->device_wake_opt = CC_GCI_GPIO_INVALID;
|
|
#endif /* !BCMDONGLEHOST */
|
|
/* clear any previous epidiag-induced target abort */
|
|
ASSERT(!si_taclear(sih, FALSE));
|
|
|
|
#if defined(BCMPMU_STATS) && !defined(BCMPMU_STATS_DISABLED)
|
|
si_pmustatstimer_init(sih);
|
|
#endif /* BCMPMU_STATS */
|
|
|
|
#ifdef BOOTLOADER_CONSOLE_OUTPUT
|
|
/* Enable console prints */
|
|
si_muxenab(sii, 3);
|
|
#endif
|
|
|
|
if (((PCIECOREREV(sih->buscorerev) == 66) || (PCIECOREREV(sih->buscorerev) == 68)) &&
|
|
CST4378_CHIPMODE_BTOP(sih->chipst)) {
|
|
/*
|
|
* HW4378-413 :
|
|
* BT oob connections for pcie function 1 seen at oob_ain[5] instead of oob_ain[1]
|
|
*/
|
|
si_oob_war_BT_F1(sih);
|
|
}
|
|
|
|
return (sii);
|
|
|
|
exit:
|
|
#if !defined(BCMDONGLEHOST)
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
if (sii->pch)
|
|
pcicore_deinit(sii->pch);
|
|
sii->pch = NULL;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
if (err_at) {
|
|
SI_ERROR(("si_doattach Failed. Error at %d\n", err_at));
|
|
si_free_coresinfo(sii, osh);
|
|
si_free_wrapper(sii);
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/** may be called with core in reset */
|
|
void
|
|
si_detach(si_t *sih)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
|
|
uint idx;
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
struct si_pub *si_local = NULL;
|
|
bcopy(&sih, &si_local, sizeof(si_t*));
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
#ifdef BCM_SH_SFLASH
|
|
if (BCM_SH_SFLASH_ENAB()) {
|
|
sh_sflash_detach(sii->osh, sih);
|
|
}
|
|
#endif
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
if (sii->nci_info) {
|
|
nci_uninit(sii->nci_info);
|
|
sii->nci_info = NULL;
|
|
|
|
/* TODO: REG_UNMAP */
|
|
}
|
|
} else {
|
|
for (idx = 0; idx < SI_MAXCORES; idx++) {
|
|
if (cores_info->regs[idx]) {
|
|
REG_UNMAP(cores_info->regs[idx]);
|
|
cores_info->regs[idx] = NULL;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
srom_var_deinit(si_local);
|
|
nvram_exit(si_local); /* free up nvram buffers */
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
if (sii->pch)
|
|
pcicore_deinit(sii->pch);
|
|
sii->pch = NULL;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
si_free_coresinfo(sii, sii->osh);
|
|
|
|
#if defined(AXI_TIMEOUTS_NIC)
|
|
if (sih->err_info) {
|
|
MFREE(sii->osh, sih->err_info, sizeof(si_axi_error_info_t));
|
|
sii->pub.err_info = NULL;
|
|
}
|
|
#endif /* AXI_TIMEOUTS_NIC */
|
|
|
|
si_free_wrapper(sii);
|
|
|
|
#ifdef BCMDVFS
|
|
if (BCMDVFS_ENAB()) {
|
|
si_dvfs_info_deinit(sih, sii->osh);
|
|
}
|
|
#endif /* BCMDVFS */
|
|
|
|
if (sii != &ksii) {
|
|
MFREE(sii->osh, sii, sizeof(si_info_t));
|
|
}
|
|
}
|
|
|
|
void *
|
|
BCMPOSTTRAPFN(si_osh)(si_t *sih)
|
|
{
|
|
const si_info_t *sii;
|
|
|
|
sii = SI_INFO(sih);
|
|
return sii->osh;
|
|
}
|
|
|
|
void
|
|
si_setosh(si_t *sih, osl_t *osh)
|
|
{
|
|
si_info_t *sii;
|
|
|
|
sii = SI_INFO(sih);
|
|
if (sii->osh != NULL) {
|
|
SI_ERROR(("osh is already set....\n"));
|
|
ASSERT(!sii->osh);
|
|
}
|
|
sii->osh = osh;
|
|
}
|
|
|
|
/** register driver interrupt disabling and restoring callback functions */
|
|
void
|
|
si_register_intr_callback(si_t *sih, void *intrsoff_fn, void *intrsrestore_fn,
|
|
void *intrsenabled_fn, void *intr_arg)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
sii->intr_arg = intr_arg;
|
|
sii->intrsoff_fn = (si_intrsoff_t)intrsoff_fn;
|
|
sii->intrsrestore_fn = (si_intrsrestore_t)intrsrestore_fn;
|
|
sii->intrsenabled_fn = (si_intrsenabled_t)intrsenabled_fn;
|
|
/* save current core id. when this function called, the current core
|
|
* must be the core which provides driver functions(il, et, wl, etc.)
|
|
*/
|
|
sii->dev_coreid = si_coreid(sih);
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_deregister_intr_callback)(si_t *sih)
|
|
{
|
|
si_info_t *sii;
|
|
|
|
sii = SI_INFO(sih);
|
|
sii->intrsoff_fn = NULL;
|
|
sii->intrsrestore_fn = NULL;
|
|
sii->intrsenabled_fn = NULL;
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_intflag)(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_intflag(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return R_REG(sii->osh, ((uint32 *)(uintptr)
|
|
(sii->oob_router + OOB_STATUSA)));
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_intflag(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
uint
|
|
si_flag(si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_flag(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_flag(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_flag(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_flag(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
uint
|
|
si_flag_alt(const si_t *sih)
|
|
{
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_flag_alt(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_flag_alt(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_setint(const si_t *sih, int siflag)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_setint(sih, siflag);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_setint(sih, siflag);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_setint(sih, siflag);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_setint(sih, siflag);
|
|
else
|
|
ASSERT(0);
|
|
}
|
|
|
|
uint32
|
|
si_oobr_baseaddr(const si_t *sih, bool second)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return 0;
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return (second ? sii->oob_router1 : sii->oob_router);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_oobr_baseaddr(sih, second);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_coreid)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (const si_cores_info_t *)sii->cores_info;
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
return nci_coreid(sih, sii->curidx);
|
|
} else
|
|
{
|
|
return cores_info->coreid[sii->curidx];
|
|
}
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_coreidx)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii;
|
|
|
|
sii = SI_INFO(sih);
|
|
return sii->curidx;
|
|
}
|
|
|
|
uint
|
|
si_get_num_cores(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->numcores;
|
|
}
|
|
|
|
volatile void *
|
|
si_d11_switch_addrbase(si_t *sih, uint coreunit)
|
|
{
|
|
return si_setcore(sih, D11_CORE_ID, coreunit);
|
|
}
|
|
|
|
/** return the core-type instantiation # of the current core */
|
|
uint
|
|
si_coreunit(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (const si_cores_info_t *)sii->cores_info;
|
|
uint idx;
|
|
uint coreid;
|
|
uint coreunit;
|
|
uint i;
|
|
|
|
if (CHIPTYPE(sii->pub.socitype) == SOCI_NCI) {
|
|
return nci_coreunit(sih);
|
|
}
|
|
|
|
coreunit = 0;
|
|
|
|
idx = sii->curidx;
|
|
|
|
ASSERT(GOODREGS(sii->curmap));
|
|
coreid = si_coreid(sih);
|
|
|
|
/* count the cores of our type */
|
|
for (i = 0; i < idx; i++)
|
|
if (cores_info->coreid[i] == coreid)
|
|
coreunit++;
|
|
|
|
return (coreunit);
|
|
}
|
|
|
|
uint
|
|
si_corevendor(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_corevendor(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_corevendor(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_corevendor(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_corevendor(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
bool
|
|
si_backplane64(const si_t *sih)
|
|
{
|
|
return ((sih->cccaps & CC_CAP_BKPLN64) != 0);
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_corerev)(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_corerev(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_corerev(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_corerev(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_corerev(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
uint
|
|
si_corerev_minor(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
return ai_corerev_minor(sih);
|
|
}
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_corerev_minor(sih);
|
|
else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* return index of coreid or BADIDX if not found */
|
|
uint
|
|
BCMPOSTTRAPFN(si_findcoreidx)(const si_t *sih, uint coreid, uint coreunit)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (const si_cores_info_t *)sii->cores_info;
|
|
uint found;
|
|
uint i;
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_findcoreidx(sih, coreid, coreunit);
|
|
}
|
|
|
|
found = 0;
|
|
|
|
for (i = 0; i < sii->numcores; i++) {
|
|
if (cores_info->coreid[i] == coreid) {
|
|
if (found == coreunit)
|
|
return (i);
|
|
found++;
|
|
}
|
|
}
|
|
|
|
return (BADIDX);
|
|
}
|
|
|
|
bool
|
|
BCMPOSTTRAPFN(si_hwa_present)(const si_t *sih)
|
|
{
|
|
if (si_findcoreidx(sih, HWA_CORE_ID, 0) != BADIDX) {
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
bool
|
|
BCMPOSTTRAPFN(si_sysmem_present)(const si_t *sih)
|
|
{
|
|
if (si_findcoreidx(sih, SYSMEM_CORE_ID, 0) != BADIDX) {
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
/* return the coreid of the core at index */
|
|
uint
|
|
si_findcoreid(const si_t *sih, uint coreidx)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = sii->cores_info;
|
|
|
|
if (coreidx >= sii->numcores) {
|
|
return NODEV_CORE_ID;
|
|
}
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_coreid(sih, coreidx);
|
|
}
|
|
return cores_info->coreid[coreidx];
|
|
}
|
|
|
|
/** return total coreunit of coreid or zero if not found */
|
|
uint
|
|
BCMPOSTTRAPFN(si_numcoreunits)(const si_t *sih, uint coreid)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (si_cores_info_t *)sii->cores_info;
|
|
uint found = 0;
|
|
uint i;
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_numcoreunits(sih, coreid);
|
|
}
|
|
for (i = 0; i < sii->numcores; i++) {
|
|
if (cores_info->coreid[i] == coreid) {
|
|
found++;
|
|
}
|
|
}
|
|
|
|
return found;
|
|
}
|
|
|
|
/** return total D11 coreunits */
|
|
uint
|
|
BCMPOSTTRAPRAMFN(si_numd11coreunits)(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_numcoreunits(sih, D11_CORE_ID);
|
|
}
|
|
return si_numcoreunits(sih, D11_CORE_ID);
|
|
}
|
|
|
|
/** return list of found cores */
|
|
uint
|
|
si_corelist(const si_t *sih, uint coreid[])
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (const si_cores_info_t *)sii->cores_info;
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_corelist(sih, coreid);
|
|
}
|
|
(void)memcpy_s(coreid, (sii->numcores * sizeof(uint)), cores_info->coreid,
|
|
(sii->numcores * sizeof(uint)));
|
|
return (sii->numcores);
|
|
}
|
|
|
|
/** return current wrapper mapping */
|
|
void *
|
|
BCMPOSTTRAPFN(si_wrapperregs)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT(GOODREGS(sii->curwrap));
|
|
|
|
return (sii->curwrap);
|
|
}
|
|
|
|
/** return current register mapping */
|
|
volatile void *
|
|
BCMPOSTTRAPFN(si_coreregs)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT(GOODREGS(sii->curmap));
|
|
|
|
return (sii->curmap);
|
|
}
|
|
|
|
/**
|
|
* This function changes logical "focus" to the indicated core;
|
|
* must be called with interrupts off.
|
|
* Moreover, callers should keep interrupts off during switching out of and back to d11 core
|
|
*/
|
|
volatile void *
|
|
BCMPOSTTRAPFN(si_setcore)(si_t *sih, uint coreid, uint coreunit)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
uint idx;
|
|
|
|
idx = si_findcoreidx(sih, coreid, coreunit);
|
|
if (!GOODIDX(idx, sii->numcores)) {
|
|
return (NULL);
|
|
}
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_setcoreidx(sih, idx);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_setcoreidx(sih, idx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_setcoreidx(sih, idx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_setcoreidx(sih, idx);
|
|
else {
|
|
ASSERT(0);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
volatile void *
|
|
BCMPOSTTRAPFN(si_setcoreidx)(si_t *sih, uint coreidx)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_setcoreidx(sih, coreidx);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_setcoreidx(sih, coreidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_setcoreidx(sih, coreidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_setcoreidx(sih, coreidx);
|
|
else {
|
|
ASSERT(0);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
/** Turn off interrupt as required by sb_setcore, before switch core */
|
|
volatile void *
|
|
BCMPOSTTRAPFN(si_switch_core)(si_t *sih, uint coreid, uint *origidx, bcm_int_bitmask_t *intr_val)
|
|
{
|
|
volatile void *cc;
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (SI_FAST(sii)) {
|
|
/* Overloading the origidx variable to remember the coreid,
|
|
* this works because the core ids cannot be confused with
|
|
* core indices.
|
|
*/
|
|
*origidx = coreid;
|
|
if (coreid == CC_CORE_ID)
|
|
return (volatile void *)CCREGS_FAST(sii);
|
|
else if (coreid == BUSCORETYPE(sih->buscoretype))
|
|
return (volatile void *)PCIEREGS(sii);
|
|
}
|
|
INTR_OFF(sii, intr_val);
|
|
*origidx = sii->curidx;
|
|
cc = si_setcore(sih, coreid, 0);
|
|
ASSERT(cc != NULL);
|
|
|
|
return cc;
|
|
}
|
|
|
|
/* restore coreidx and restore interrupt */
|
|
void
|
|
BCMPOSTTRAPFN(si_restore_core)(si_t *sih, uint coreid, bcm_int_bitmask_t *intr_val)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (SI_FAST(sii) && ((coreid == CC_CORE_ID) || (coreid == BUSCORETYPE(sih->buscoretype))))
|
|
return;
|
|
|
|
si_setcoreidx(sih, coreid);
|
|
INTR_RESTORE(sii, intr_val);
|
|
}
|
|
|
|
/* Switch to particular core and get corerev */
|
|
#ifdef USE_NEW_COREREV_API
|
|
uint
|
|
BCMPOSTTRAPFN(si_corerev_ext)(si_t *sih, uint coreid, uint coreunit)
|
|
{
|
|
uint coreidx;
|
|
uint corerev;
|
|
|
|
coreidx = si_coreidx(sih);
|
|
(void)si_setcore(sih, coreid, coreunit);
|
|
|
|
corerev = si_corerev(sih);
|
|
|
|
si_setcoreidx(sih, coreidx);
|
|
return corerev;
|
|
}
|
|
#else
|
|
uint si_get_corerev(si_t *sih, uint core_id)
|
|
{
|
|
uint corerev, orig_coreid;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
si_switch_core(sih, core_id, &orig_coreid, &intr_val);
|
|
corerev = si_corerev(sih);
|
|
si_restore_core(sih, orig_coreid, &intr_val);
|
|
return corerev;
|
|
}
|
|
#endif /* !USE_NEW_COREREV_API */
|
|
|
|
int
|
|
si_numaddrspaces(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_numaddrspaces(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_numaddrspaces(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_numaddrspaces(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_numaddrspaces(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* Return the address of the nth address space in the current core
|
|
* Arguments:
|
|
* sih : Pointer to struct si_t
|
|
* spidx : slave port index
|
|
* baidx : base address index
|
|
*/
|
|
|
|
uint32
|
|
si_addrspace(const si_t *sih, uint spidx, uint baidx)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_addrspace(sih, baidx);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_addrspace(sih, spidx, baidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_addrspace(sih, baidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_addrspace(sih, spidx, baidx);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/* Return the size of the nth address space in the current core
|
|
* Arguments:
|
|
* sih : Pointer to struct si_t
|
|
* spidx : slave port index
|
|
* baidx : base address index
|
|
*/
|
|
uint32
|
|
si_addrspacesize(const si_t *sih, uint spidx, uint baidx)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_addrspacesize(sih, baidx);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_addrspacesize(sih, spidx, baidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_addrspacesize(sih, baidx);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_addrspacesize(sih, spidx, baidx);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_coreaddrspaceX(const si_t *sih, uint asidx, uint32 *addr, uint32 *size)
|
|
{
|
|
/* Only supported for SOCI_AI */
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_coreaddrspaceX(sih, asidx, addr, size);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_coreaddrspaceX(sih, asidx, addr, size);
|
|
else
|
|
*size = 0;
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_core_cflags)(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_core_cflags(sih, mask, val);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_core_cflags(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_core_cflags(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_core_cflags(sih, mask, val);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_core_cflags_wo(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_core_cflags_wo(sih, mask, val);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_core_cflags_wo(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_core_cflags_wo(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_core_cflags_wo(sih, mask, val);
|
|
else
|
|
ASSERT(0);
|
|
}
|
|
|
|
uint32
|
|
si_core_sflags(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_core_sflags(sih, mask, val);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_core_sflags(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_core_sflags(sih, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_core_sflags(sih, mask, val);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_commit(si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_commit(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
;
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
;
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
;
|
|
else {
|
|
ASSERT(0);
|
|
}
|
|
}
|
|
|
|
bool
|
|
BCMPOSTTRAPFN(si_iscoreup)(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_iscoreup(sih);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_iscoreup(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_iscoreup(sih);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_iscoreup(sih);
|
|
else {
|
|
ASSERT(0);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
/** Caller should make sure it is on the right core, before calling this routine */
|
|
uint
|
|
BCMPOSTTRAPFN(si_wrapperreg)(const si_t *sih, uint32 offset, uint32 mask, uint32 val)
|
|
{
|
|
/* only for AI back plane chips */
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return (ai_wrap_reg(sih, offset, mask, val));
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return (nci_get_wrap_reg(sih, offset, mask, val));
|
|
return 0;
|
|
}
|
|
/* si_backplane_access is used to read full backplane address from host for PCIE FD
|
|
* it uses secondary bar-0 window which lies at an offset of 16K from primary bar-0
|
|
* Provides support for read/write of 1/2/4 bytes of backplane address
|
|
* Can be used to read/write
|
|
* 1. core regs
|
|
* 2. Wrapper regs
|
|
* 3. memory
|
|
* 4. BT area
|
|
* For accessing any 32 bit backplane address, [31 : 12] of backplane should be given in "region"
|
|
* [11 : 0] should be the "regoff"
|
|
* for reading 4 bytes from reg 0x200 of d11 core use it like below
|
|
* : si_backplane_access(sih, 0x18001000, 0x200, 4, 0, TRUE)
|
|
*/
|
|
static int si_backplane_addr_sane(uint addr, uint size)
|
|
{
|
|
int bcmerror = BCME_OK;
|
|
|
|
/* For 2 byte access, address has to be 2 byte aligned */
|
|
if (size == 2) {
|
|
if (addr & 0x1) {
|
|
bcmerror = BCME_ERROR;
|
|
}
|
|
}
|
|
/* For 4 byte access, address has to be 4 byte aligned */
|
|
if (size == 4) {
|
|
if (addr & 0x3) {
|
|
bcmerror = BCME_ERROR;
|
|
}
|
|
}
|
|
|
|
return bcmerror;
|
|
}
|
|
|
|
void
|
|
si_invalidate_second_bar0win(si_t *sih)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
sii->second_bar0win = ~0x0;
|
|
}
|
|
|
|
int
|
|
si_backplane_access(si_t *sih, uint addr, uint size, uint *val, bool read)
|
|
{
|
|
volatile uint32 *r = NULL;
|
|
uint32 region = 0;
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
/* Valid only for pcie bus */
|
|
if (BUSTYPE(sih->bustype) != PCI_BUS) {
|
|
SI_ERROR(("Valid only for pcie bus \n"));
|
|
return BCME_ERROR;
|
|
}
|
|
|
|
/* Split adrr into region and address offset */
|
|
region = (addr & (0xFFFFF << 12));
|
|
addr = addr & 0xFFF;
|
|
|
|
/* check for address and size sanity */
|
|
if (si_backplane_addr_sane(addr, size) != BCME_OK)
|
|
return BCME_ERROR;
|
|
|
|
/* Update window if required */
|
|
if (sii->second_bar0win != region) {
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCIE2_BAR0_CORE2_WIN, 4, region);
|
|
sii->second_bar0win = region;
|
|
}
|
|
|
|
/* Estimate effective address
|
|
* sii->curmap : bar-0 virtual address
|
|
* PCI_SECOND_BAR0_OFFSET : secondar bar-0 offset
|
|
* regoff : actual reg offset
|
|
*/
|
|
r = (volatile uint32 *)((volatile char *)sii->curmap + PCI_SECOND_BAR0_OFFSET + addr);
|
|
|
|
SI_VMSG(("si curmap %p region %x regaddr %x effective addr %p READ %d\n",
|
|
(volatile char*)sii->curmap, region, addr, r, read));
|
|
|
|
switch (size) {
|
|
case sizeof(uint8) :
|
|
if (read)
|
|
*val = R_REG(sii->osh, (volatile uint8*)r);
|
|
else
|
|
W_REG(sii->osh, (volatile uint8*)r, *val);
|
|
break;
|
|
case sizeof(uint16) :
|
|
if (read)
|
|
*val = R_REG(sii->osh, (volatile uint16*)r);
|
|
else
|
|
W_REG(sii->osh, (volatile uint16*)r, *val);
|
|
break;
|
|
case sizeof(uint32) :
|
|
if (read)
|
|
*val = R_REG(sii->osh, (volatile uint32*)r);
|
|
else
|
|
W_REG(sii->osh, (volatile uint32*)r, *val);
|
|
break;
|
|
default :
|
|
SI_ERROR(("Invalid size %d \n", size));
|
|
return (BCME_ERROR);
|
|
break;
|
|
}
|
|
|
|
return (BCME_OK);
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_corereg)(si_t *sih, uint coreidx, uint regoff, uint mask, uint val)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_corereg(sih, coreidx, regoff, mask, val);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_corereg(sih, coreidx, regoff, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return ub_corereg(sih, coreidx, regoff, mask, val);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_corereg(sih, coreidx, regoff, mask, val);
|
|
else {
|
|
ASSERT(0);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_corereg_writeonly)(si_t *sih, uint coreidx, uint regoff, uint mask, uint val)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
return nci_corereg_writeonly(sih, coreidx, regoff, mask, val);
|
|
} else
|
|
{
|
|
return ai_corereg_writeonly(sih, coreidx, regoff, mask, val);
|
|
}
|
|
}
|
|
|
|
/** ILP sensitive register access needs special treatment to avoid backplane stalls */
|
|
bool
|
|
BCMPOSTTRAPFN(si_pmu_is_ilp_sensitive)(uint32 idx, uint regoff)
|
|
{
|
|
if (idx == SI_CC_IDX) {
|
|
if (CHIPCREGS_ILP_SENSITIVE(regoff))
|
|
return TRUE;
|
|
} else if (PMUREGS_ILP_SENSITIVE(regoff)) {
|
|
return TRUE;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
/** 'idx' should refer either to the chipcommon core or the PMU core */
|
|
uint
|
|
BCMPOSTTRAPFN(si_pmu_corereg)(si_t *sih, uint32 idx, uint regoff, uint mask, uint val)
|
|
{
|
|
int pmustatus_offset;
|
|
|
|
/* prevent backplane stall on double write to 'ILP domain' registers in the PMU */
|
|
if (mask != 0 && PMUREV(sih->pmurev) >= 22 &&
|
|
si_pmu_is_ilp_sensitive(idx, regoff)) {
|
|
pmustatus_offset = AOB_ENAB(sih) ? OFFSETOF(pmuregs_t, pmustatus) :
|
|
OFFSETOF(chipcregs_t, pmustatus);
|
|
|
|
while (si_corereg(sih, idx, pmustatus_offset, 0, 0) & PST_SLOW_WR_PENDING)
|
|
{};
|
|
}
|
|
|
|
return si_corereg(sih, idx, regoff, mask, val);
|
|
}
|
|
|
|
/*
|
|
* If there is no need for fiddling with interrupts or core switches (typically silicon
|
|
* back plane registers, pci registers and chipcommon registers), this function
|
|
* returns the register offset on this core to a mapped address. This address can
|
|
* be used for W_REG/R_REG directly.
|
|
*
|
|
* For accessing registers that would need a core switch, this function will return
|
|
* NULL.
|
|
*/
|
|
volatile uint32 *
|
|
BCMPOSTTRAPFN(si_corereg_addr)(si_t *sih, uint coreidx, uint regoff)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_corereg_addr(sih, coreidx, regoff);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return ai_corereg_addr(sih, coreidx, regoff);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
return nci_corereg_addr(sih, coreidx, regoff);
|
|
else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_core_disable(const si_t *sih, uint32 bits)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_core_disable(sih, bits);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_core_disable(sih, bits);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_core_disable(sih, bits);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_core_disable(sih, bits);
|
|
}
|
|
|
|
bool
|
|
si_core_reset(si_t *sih, uint32 bits, uint32 resetbits)
|
|
{
|
|
bool ret = TRUE;
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_core_reset(sih, bits, resetbits);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ret = ai_core_reset(sih, bits, resetbits);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_core_reset(sih, bits, resetbits);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_core_reset(sih, bits, resetbits);
|
|
return ret;
|
|
}
|
|
|
|
/** Run bist on current core. Caller needs to take care of core-specific bist hazards */
|
|
int
|
|
si_corebist(const si_t *sih)
|
|
{
|
|
uint32 cflags;
|
|
int result = 0;
|
|
|
|
/* Read core control flags */
|
|
cflags = si_core_cflags(sih, 0, 0);
|
|
|
|
/* Set bist & fgc */
|
|
si_core_cflags(sih, ~0, (SICF_BIST_EN | SICF_FGC));
|
|
|
|
/* Wait for bist done */
|
|
SPINWAIT(((si_core_sflags(sih, 0, 0) & SISF_BIST_DONE) == 0), 100000);
|
|
|
|
if (si_core_sflags(sih, 0, 0) & SISF_BIST_ERROR)
|
|
result = BCME_ERROR;
|
|
|
|
/* Reset core control flags */
|
|
si_core_cflags(sih, 0xffff, cflags);
|
|
|
|
return result;
|
|
}
|
|
|
|
uint
|
|
si_num_slaveports(const si_t *sih, uint coreid)
|
|
{
|
|
uint idx = si_findcoreidx(sih, coreid, 0);
|
|
uint num = 0;
|
|
|
|
if (idx != BADIDX) {
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
num = ai_num_slaveports(sih, idx);
|
|
}
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI) {
|
|
num = nci_num_slaveports(sih, idx);
|
|
}
|
|
}
|
|
return num;
|
|
}
|
|
|
|
/* TODO: Check if NCI has a slave port address */
|
|
uint32
|
|
si_get_slaveport_addr(si_t *sih, uint spidx, uint baidx, uint core_id, uint coreunit)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx = sii->curidx;
|
|
uint32 addr = 0x0;
|
|
|
|
if (!((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NCI)))
|
|
goto done;
|
|
|
|
si_setcore(sih, core_id, coreunit);
|
|
|
|
addr = si_addrspace(sih, spidx, baidx);
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
return addr;
|
|
}
|
|
|
|
/* TODO: Check if NCI has a d11 slave port address */
|
|
uint32
|
|
si_get_d11_slaveport_addr(si_t *sih, uint spidx, uint baidx, uint coreunit)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx = sii->curidx;
|
|
uint32 addr = 0x0;
|
|
|
|
if (!((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NCI)))
|
|
goto done;
|
|
|
|
si_setcore(sih, D11_CORE_ID, coreunit);
|
|
|
|
addr = si_addrspace(sih, spidx, baidx);
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
return addr;
|
|
}
|
|
|
|
static uint32
|
|
factor6(uint32 x)
|
|
{
|
|
switch (x) {
|
|
case CC_F6_2: return 2;
|
|
case CC_F6_3: return 3;
|
|
case CC_F6_4: return 4;
|
|
case CC_F6_5: return 5;
|
|
case CC_F6_6: return 6;
|
|
case CC_F6_7: return 7;
|
|
default: return 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Divide the clock by the divisor with protection for
|
|
* a zero divisor.
|
|
*/
|
|
static uint32
|
|
divide_clock(uint32 clock, uint32 div)
|
|
{
|
|
return div ? clock / div : 0;
|
|
}
|
|
|
|
/** calculate the speed the SI would run at given a set of clockcontrol values */
|
|
uint32
|
|
si_clock_rate(uint32 pll_type, uint32 n, uint32 m)
|
|
{
|
|
uint32 n1, n2, clock, m1, m2, m3, mc;
|
|
|
|
n1 = n & CN_N1_MASK;
|
|
n2 = (n & CN_N2_MASK) >> CN_N2_SHIFT;
|
|
|
|
if (pll_type == PLL_TYPE6) {
|
|
if (m & CC_T6_MMASK)
|
|
return CC_T6_M1;
|
|
else
|
|
return CC_T6_M0;
|
|
} else if ((pll_type == PLL_TYPE1) ||
|
|
(pll_type == PLL_TYPE3) ||
|
|
(pll_type == PLL_TYPE4) ||
|
|
(pll_type == PLL_TYPE7)) {
|
|
n1 = factor6(n1);
|
|
n2 += CC_F5_BIAS;
|
|
} else if (pll_type == PLL_TYPE2) {
|
|
n1 += CC_T2_BIAS;
|
|
n2 += CC_T2_BIAS;
|
|
ASSERT((n1 >= 2) && (n1 <= 7));
|
|
ASSERT((n2 >= 5) && (n2 <= 23));
|
|
} else if (pll_type == PLL_TYPE5) {
|
|
/* 5365 */
|
|
return (100000000);
|
|
} else
|
|
ASSERT(0);
|
|
/* PLL types 3 and 7 use BASE2 (25Mhz) */
|
|
if ((pll_type == PLL_TYPE3) ||
|
|
(pll_type == PLL_TYPE7)) {
|
|
clock = CC_CLOCK_BASE2 * n1 * n2;
|
|
} else
|
|
clock = CC_CLOCK_BASE1 * n1 * n2;
|
|
|
|
if (clock == 0)
|
|
return 0;
|
|
|
|
m1 = m & CC_M1_MASK;
|
|
m2 = (m & CC_M2_MASK) >> CC_M2_SHIFT;
|
|
m3 = (m & CC_M3_MASK) >> CC_M3_SHIFT;
|
|
mc = (m & CC_MC_MASK) >> CC_MC_SHIFT;
|
|
|
|
if ((pll_type == PLL_TYPE1) ||
|
|
(pll_type == PLL_TYPE3) ||
|
|
(pll_type == PLL_TYPE4) ||
|
|
(pll_type == PLL_TYPE7)) {
|
|
m1 = factor6(m1);
|
|
if ((pll_type == PLL_TYPE1) || (pll_type == PLL_TYPE3))
|
|
m2 += CC_F5_BIAS;
|
|
else
|
|
m2 = factor6(m2);
|
|
m3 = factor6(m3);
|
|
|
|
switch (mc) {
|
|
case CC_MC_BYPASS: return (clock);
|
|
case CC_MC_M1: return divide_clock(clock, m1);
|
|
case CC_MC_M1M2: return divide_clock(clock, m1 * m2);
|
|
case CC_MC_M1M2M3: return divide_clock(clock, m1 * m2 * m3);
|
|
case CC_MC_M1M3: return divide_clock(clock, m1 * m3);
|
|
default: return (0);
|
|
}
|
|
} else {
|
|
ASSERT(pll_type == PLL_TYPE2);
|
|
|
|
m1 += CC_T2_BIAS;
|
|
m2 += CC_T2M2_BIAS;
|
|
m3 += CC_T2_BIAS;
|
|
ASSERT((m1 >= 2) && (m1 <= 7));
|
|
ASSERT((m2 >= 3) && (m2 <= 10));
|
|
ASSERT((m3 >= 2) && (m3 <= 7));
|
|
|
|
if ((mc & CC_T2MC_M1BYP) == 0)
|
|
clock /= m1;
|
|
if ((mc & CC_T2MC_M2BYP) == 0)
|
|
clock /= m2;
|
|
if ((mc & CC_T2MC_M3BYP) == 0)
|
|
clock /= m3;
|
|
|
|
return (clock);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Some chips could have multiple host interfaces, however only one will be active.
|
|
* For a given chip. Depending pkgopt and cc_chipst return the active host interface.
|
|
*/
|
|
uint
|
|
si_chip_hostif(const si_t *sih)
|
|
{
|
|
uint hosti = 0;
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
hosti = CHIP_HOSTIF_SDIOMODE;
|
|
break;
|
|
CASE_BCM43602_CHIP:
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
break;
|
|
|
|
case BCM4360_CHIP_ID:
|
|
/* chippkg bit-0 == 0 is PCIE only pkgs
|
|
* chippkg bit-0 == 1 has both PCIE and USB cores enabled
|
|
*/
|
|
if ((sih->chippkg & 0x1) && (sih->chipst & CST4360_MODE_USB))
|
|
hosti = CHIP_HOSTIF_USBMODE;
|
|
else
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
|
|
break;
|
|
|
|
case BCM4369_CHIP_GRPID:
|
|
if (CST4369_CHIPMODE_SDIOD(sih->chipst))
|
|
hosti = CHIP_HOSTIF_SDIOMODE;
|
|
else if (CST4369_CHIPMODE_PCIE(sih->chipst))
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
break;
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
break;
|
|
case BCM4362_CHIP_GRPID:
|
|
if (CST4362_CHIPMODE_SDIOD(sih->chipst)) {
|
|
hosti = CHIP_HOSTIF_SDIOMODE;
|
|
} else if (CST4362_CHIPMODE_PCIE(sih->chipst)) {
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
}
|
|
break;
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
if (CST4381_CHIPMODE_PCIE(sih->chipst)) {
|
|
hosti = CHIP_HOSTIF_PCIEMODE;
|
|
} else if (CST4381_CHIPMODE_SDIO(sih->chipst)) {
|
|
hosti = CHIP_HOSTIF_SDIOMODE;
|
|
} else if (CST4381_CHIPMODE_USB(sih->chipst)) {
|
|
hosti = CHIP_HOSTIF_USBMODE;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return hosti;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
uint32
|
|
si_clock(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint32 n, m;
|
|
uint idx;
|
|
uint32 pll_type, rate;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
if (PMUCTL_ENAB(sih)) {
|
|
rate = si_pmu_si_clock(sih, sii->osh);
|
|
goto exit;
|
|
}
|
|
|
|
idx = sii->curidx;
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
ASSERT(cc != NULL);
|
|
|
|
n = R_REG(sii->osh, &cc->clockcontrol_n);
|
|
pll_type = sih->cccaps & CC_CAP_PLL_MASK;
|
|
if (pll_type == PLL_TYPE6)
|
|
m = R_REG(sii->osh, &cc->clockcontrol_m3);
|
|
else if (pll_type == PLL_TYPE3)
|
|
m = R_REG(sii->osh, &cc->clockcontrol_m2);
|
|
else
|
|
m = R_REG(sii->osh, &cc->clockcontrol_sb);
|
|
|
|
/* calculate rate */
|
|
rate = si_clock_rate(pll_type, n, m);
|
|
|
|
if (pll_type == PLL_TYPE3)
|
|
rate = rate / 2;
|
|
|
|
/* switch back to previous core */
|
|
si_setcoreidx(sih, idx);
|
|
exit:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
|
|
return rate;
|
|
}
|
|
|
|
/** returns value in [Hz] units */
|
|
uint32
|
|
si_alp_clock(si_t *sih)
|
|
{
|
|
if (PMUCTL_ENAB(sih)) {
|
|
return si_pmu_alp_clock(sih, si_osh(sih));
|
|
}
|
|
|
|
return ALP_CLOCK;
|
|
}
|
|
|
|
/** returns value in [Hz] units */
|
|
uint32
|
|
si_ilp_clock(si_t *sih)
|
|
{
|
|
if (PMUCTL_ENAB(sih))
|
|
return si_pmu_ilp_clock(sih, si_osh(sih));
|
|
|
|
return ILP_CLOCK;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
/** set chip watchdog reset timer to fire in 'ticks' */
|
|
void
|
|
si_watchdog(si_t *sih, uint ticks)
|
|
{
|
|
uint nb, maxt;
|
|
uint pmu_wdt = 1;
|
|
|
|
if ((CHIPID(sih->chip) == BCM4381_CHIP_GRPID) ||
|
|
(CHIPID(sih->chip) == BCM4382_CHIP_GRPID)) {
|
|
/* For chips in which PMU is shared, cc watchdog reset
|
|
* * should be done insted of PMU
|
|
* */
|
|
pmu_wdt = 0;
|
|
}
|
|
|
|
if (PMUCTL_ENAB(sih) && pmu_wdt) {
|
|
nb = (CCREV(sih->ccrev) < 26) ? 16 : ((CCREV(sih->ccrev) >= 37) ? 32 : 24);
|
|
/* The mips compiler uses the sllv instruction,
|
|
* so we specially handle the 32-bit case.
|
|
*/
|
|
if (nb == 32)
|
|
maxt = 0xffffffff;
|
|
else
|
|
maxt = ((1 << nb) - 1);
|
|
|
|
/* PR43821: PMU watchdog timer needs min. of 2 ticks */
|
|
if (ticks == 1)
|
|
ticks = 2;
|
|
else if (ticks > maxt)
|
|
ticks = maxt;
|
|
#ifndef DONGLEBUILD
|
|
if ((CHIPID(sih->chip) == BCM43012_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43013_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43014_CHIP_ID)) {
|
|
PMU_REG_NEW(sih, min_res_mask, ~0, DEFAULT_43012_MIN_RES_MASK);
|
|
PMU_REG_NEW(sih, watchdog_res_mask, ~0, DEFAULT_43012_MIN_RES_MASK);
|
|
PMU_REG_NEW(sih, pmustatus, PST_WDRESET, PST_WDRESET);
|
|
PMU_REG_NEW(sih, pmucontrol_ext, PCTL_EXT_FASTLPO_SWENAB, 0);
|
|
SPINWAIT((PMU_REG(sih, pmustatus, 0, 0) & PST_ILPFASTLPO),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
}
|
|
#endif /* DONGLEBUILD */
|
|
pmu_corereg(sih, SI_CC_IDX, pmuwatchdog, ~0, ticks);
|
|
} else {
|
|
#if !defined(BCMDONGLEHOST)
|
|
/* make sure we come up in fast clock mode; or if clearing, clear clock */
|
|
si_clkctl_cc(sih, ticks ? CLK_FAST : CLK_DYNAMIC);
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
maxt = (1 << 28) - 1;
|
|
if (ticks > maxt)
|
|
ticks = maxt;
|
|
|
|
if (CCREV(sih->ccrev) >= 65) {
|
|
si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, watchdog), ~0,
|
|
(ticks & WD_COUNTER_MASK) | WD_SSRESET_PCIE_F0_EN |
|
|
WD_SSRESET_PCIE_ALL_FN_EN);
|
|
} else {
|
|
si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, watchdog), ~0, ticks);
|
|
}
|
|
}
|
|
}
|
|
|
|
/** trigger watchdog reset after ms milliseconds */
|
|
void
|
|
si_watchdog_ms(si_t *sih, uint32 ms)
|
|
{
|
|
si_watchdog(sih, wd_msticks * ms);
|
|
}
|
|
|
|
uint32 si_watchdog_msticks(void)
|
|
{
|
|
return wd_msticks;
|
|
}
|
|
|
|
bool
|
|
si_taclear(si_t *sih, bool details)
|
|
{
|
|
#if defined(BCMDBG_ERR) || defined(BCMASSERT_SUPPORT) || defined(BCMDBG_DUMP)
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
return sb_taclear(sih, details);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NCI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
return FALSE;
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
return FALSE;
|
|
else {
|
|
ASSERT(0);
|
|
return FALSE;
|
|
}
|
|
#else
|
|
return FALSE;
|
|
#endif /* BCMDBG_ERR || BCMASSERT_SUPPORT || BCMDBG_DUMP */
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/**
|
|
* Map sb core id to pci device id.
|
|
*/
|
|
uint16
|
|
si_d11_devid(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint16 device;
|
|
|
|
(void) sii;
|
|
|
|
/* normal case: nvram variable with devpath->devid->wl0id */
|
|
if ((device = (uint16)si_getdevpathintvar(sih, rstr_devid)) != 0)
|
|
;
|
|
/* Get devid from OTP/SPROM depending on where the SROM is read */
|
|
else if ((device = (uint16)getintvar(sii->vars, rstr_devid)) != 0)
|
|
;
|
|
/* no longer support wl0id, but keep the code here for backward compatibility. */
|
|
else if ((device = (uint16)getintvar(sii->vars, rstr_wl0id)) != 0)
|
|
;
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
;
|
|
else {
|
|
/* ignore it */
|
|
device = 0xffff;
|
|
}
|
|
return device;
|
|
}
|
|
|
|
int
|
|
si_corepciid(si_t *sih, uint func, uint16 *pcivendor, uint16 *pcidevice,
|
|
uint8 *pciclass, uint8 *pcisubclass, uint8 *pciprogif,
|
|
uint8 *pciheader)
|
|
{
|
|
uint16 vendor = 0xffff, device = 0xffff;
|
|
uint8 class, subclass, progif = 0;
|
|
uint8 header = PCI_HEADER_NORMAL;
|
|
uint32 core = si_coreid(sih);
|
|
|
|
/* Verify whether the function exists for the core */
|
|
if (func >= (uint)((core == USB20H_CORE_ID) || (core == NS_USB20_CORE_ID) ? 2 : 1))
|
|
return BCME_ERROR;
|
|
|
|
/* Known vendor translations */
|
|
switch (si_corevendor(sih)) {
|
|
case SB_VEND_BCM:
|
|
case MFGID_BRCM:
|
|
vendor = VENDOR_BROADCOM;
|
|
break;
|
|
default:
|
|
return BCME_ERROR;
|
|
}
|
|
|
|
/* Determine class based on known core codes */
|
|
switch (core) {
|
|
case ENET_CORE_ID:
|
|
class = PCI_CLASS_NET;
|
|
subclass = PCI_NET_ETHER;
|
|
device = BCM47XX_ENET_ID;
|
|
break;
|
|
case GIGETH_CORE_ID:
|
|
class = PCI_CLASS_NET;
|
|
subclass = PCI_NET_ETHER;
|
|
device = BCM47XX_GIGETH_ID;
|
|
break;
|
|
case GMAC_CORE_ID:
|
|
class = PCI_CLASS_NET;
|
|
subclass = PCI_NET_ETHER;
|
|
device = BCM47XX_GMAC_ID;
|
|
break;
|
|
case SDRAM_CORE_ID:
|
|
case MEMC_CORE_ID:
|
|
case DMEMC_CORE_ID:
|
|
case SOCRAM_CORE_ID:
|
|
class = PCI_CLASS_MEMORY;
|
|
subclass = PCI_MEMORY_RAM;
|
|
device = (uint16)core;
|
|
break;
|
|
case PCI_CORE_ID:
|
|
case PCIE_CORE_ID:
|
|
case PCIE2_CORE_ID:
|
|
class = PCI_CLASS_BRIDGE;
|
|
subclass = PCI_BRIDGE_PCI;
|
|
device = (uint16)core;
|
|
header = PCI_HEADER_BRIDGE;
|
|
break;
|
|
case CODEC_CORE_ID:
|
|
class = PCI_CLASS_COMM;
|
|
subclass = PCI_COMM_MODEM;
|
|
device = BCM47XX_V90_ID;
|
|
break;
|
|
case I2S_CORE_ID:
|
|
class = PCI_CLASS_MMEDIA;
|
|
subclass = PCI_MMEDIA_AUDIO;
|
|
device = BCM47XX_AUDIO_ID;
|
|
break;
|
|
case USB_CORE_ID:
|
|
case USB11H_CORE_ID:
|
|
class = PCI_CLASS_SERIAL;
|
|
subclass = PCI_SERIAL_USB;
|
|
progif = 0x10; /* OHCI */
|
|
device = BCM47XX_USBH_ID;
|
|
break;
|
|
case USB20H_CORE_ID:
|
|
case NS_USB20_CORE_ID:
|
|
class = PCI_CLASS_SERIAL;
|
|
subclass = PCI_SERIAL_USB;
|
|
progif = func == 0 ? 0x10 : 0x20; /* OHCI/EHCI value defined in spec */
|
|
device = BCM47XX_USB20H_ID;
|
|
header = PCI_HEADER_MULTI; /* multifunction */
|
|
break;
|
|
case IPSEC_CORE_ID:
|
|
class = PCI_CLASS_CRYPT;
|
|
subclass = PCI_CRYPT_NETWORK;
|
|
device = BCM47XX_IPSEC_ID;
|
|
break;
|
|
case NS_USB30_CORE_ID:
|
|
class = PCI_CLASS_SERIAL;
|
|
subclass = PCI_SERIAL_USB;
|
|
progif = 0x30; /* XHCI */
|
|
device = BCM47XX_USB30H_ID;
|
|
break;
|
|
case ROBO_CORE_ID:
|
|
/* Don't use class NETWORK, so wl/et won't attempt to recognize it */
|
|
class = PCI_CLASS_COMM;
|
|
subclass = PCI_COMM_OTHER;
|
|
device = BCM47XX_ROBO_ID;
|
|
break;
|
|
case CC_CORE_ID:
|
|
class = PCI_CLASS_MEMORY;
|
|
subclass = PCI_MEMORY_FLASH;
|
|
device = (uint16)core;
|
|
break;
|
|
case SATAXOR_CORE_ID:
|
|
class = PCI_CLASS_XOR;
|
|
subclass = PCI_XOR_QDMA;
|
|
device = BCM47XX_SATAXOR_ID;
|
|
break;
|
|
case ATA100_CORE_ID:
|
|
class = PCI_CLASS_DASDI;
|
|
subclass = PCI_DASDI_IDE;
|
|
device = BCM47XX_ATA100_ID;
|
|
break;
|
|
case USB11D_CORE_ID:
|
|
class = PCI_CLASS_SERIAL;
|
|
subclass = PCI_SERIAL_USB;
|
|
device = BCM47XX_USBD_ID;
|
|
break;
|
|
case USB20D_CORE_ID:
|
|
class = PCI_CLASS_SERIAL;
|
|
subclass = PCI_SERIAL_USB;
|
|
device = BCM47XX_USB20D_ID;
|
|
break;
|
|
case D11_CORE_ID:
|
|
class = PCI_CLASS_NET;
|
|
subclass = PCI_NET_OTHER;
|
|
device = si_d11_devid(sih);
|
|
break;
|
|
|
|
default:
|
|
class = subclass = progif = 0xff;
|
|
device = (uint16)core;
|
|
break;
|
|
}
|
|
|
|
*pcivendor = vendor;
|
|
*pcidevice = device;
|
|
*pciclass = class;
|
|
*pcisubclass = subclass;
|
|
*pciprogif = progif;
|
|
*pciheader = header;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#if defined(BCMDBG) || defined(BCMDBG_DUMP) || defined(BCMDBG_PHYDUMP)
|
|
/** print interesting sbconfig registers */
|
|
void
|
|
si_dumpregs(si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
origidx = sii->curidx;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_dumpregs(sih, b);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_dumpregs(sih, b);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_dumpregs(sih, b);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_dumpregs(sih, b);
|
|
else
|
|
ASSERT(0);
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
#endif /* BCMDBG || BCMDBG_DUMP || BCMDBG_PHYDUMP */
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#ifdef BCMDBG
|
|
void
|
|
si_view(si_t *sih, bool verbose)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
sb_view(sih, verbose);
|
|
else if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_view(sih, verbose);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_UBUS)
|
|
ub_view(sih, verbose);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_view(sih, verbose);
|
|
else
|
|
ASSERT(0);
|
|
}
|
|
|
|
void
|
|
si_viewall(si_t *sih, bool verbose)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
uint curidx, i;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
curidx = sii->curidx;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_NAI))
|
|
ai_viewall(sih, verbose);
|
|
else if (CHIPTYPE(sih->socitype) == SOCI_NCI)
|
|
nci_viewall(sih, verbose);
|
|
else {
|
|
SI_ERROR(("si_viewall: num_cores %d\n", sii->numcores));
|
|
for (i = 0; i < sii->numcores; i++) {
|
|
si_setcoreidx(sih, i);
|
|
si_view(sih, verbose);
|
|
}
|
|
}
|
|
si_setcoreidx(sih, curidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
#endif /* BCMDBG */
|
|
|
|
/** return the slow clock source - LPO, XTAL, or PCI */
|
|
static uint
|
|
si_slowclk_src(si_info_t *sii)
|
|
{
|
|
chipcregs_t *cc;
|
|
|
|
ASSERT(SI_FAST(sii) || si_coreid(&sii->pub) == CC_CORE_ID);
|
|
|
|
if (CCREV(sii->pub.ccrev) < 6) {
|
|
if ((BUSTYPE(sii->pub.bustype) == PCI_BUS) &&
|
|
(OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_OUT, sizeof(uint32)) &
|
|
PCI_CFG_GPIO_SCS))
|
|
return (SCC_SS_PCI);
|
|
else
|
|
return (SCC_SS_XTAL);
|
|
} else if (CCREV(sii->pub.ccrev) < 10) {
|
|
cc = (chipcregs_t *)si_setcoreidx(&sii->pub, sii->curidx);
|
|
ASSERT(cc);
|
|
return (R_REG(sii->osh, &cc->slow_clk_ctl) & SCC_SS_MASK);
|
|
} else /* Insta-clock */
|
|
return (SCC_SS_XTAL);
|
|
}
|
|
|
|
/** return the ILP (slowclock) min or max frequency */
|
|
static uint
|
|
si_slowclk_freq(si_info_t *sii, bool max_freq, chipcregs_t *cc)
|
|
{
|
|
uint32 slowclk;
|
|
uint div;
|
|
|
|
ASSERT(SI_FAST(sii) || si_coreid(&sii->pub) == CC_CORE_ID);
|
|
|
|
/* shouldn't be here unless we've established the chip has dynamic clk control */
|
|
ASSERT(R_REG(sii->osh, &cc->capabilities) & CC_CAP_PWR_CTL);
|
|
|
|
slowclk = si_slowclk_src(sii);
|
|
if (CCREV(sii->pub.ccrev) < 6) {
|
|
if (slowclk == SCC_SS_PCI)
|
|
return (max_freq ? (PCIMAXFREQ / 64) : (PCIMINFREQ / 64));
|
|
else
|
|
return (max_freq ? (XTALMAXFREQ / 32) : (XTALMINFREQ / 32));
|
|
} else if (CCREV(sii->pub.ccrev) < 10) {
|
|
div = 4 *
|
|
(((R_REG(sii->osh, &cc->slow_clk_ctl) & SCC_CD_MASK) >> SCC_CD_SHIFT) + 1);
|
|
if (slowclk == SCC_SS_LPO)
|
|
return (max_freq ? LPOMAXFREQ : LPOMINFREQ);
|
|
else if (slowclk == SCC_SS_XTAL)
|
|
return (max_freq ? (XTALMAXFREQ / div) : (XTALMINFREQ / div));
|
|
else if (slowclk == SCC_SS_PCI)
|
|
return (max_freq ? (PCIMAXFREQ / div) : (PCIMINFREQ / div));
|
|
else
|
|
ASSERT(0);
|
|
} else {
|
|
/* Chipc rev 10 is InstaClock */
|
|
div = R_REG(sii->osh, &cc->system_clk_ctl) >> SYCC_CD_SHIFT;
|
|
div = 4 * (div + 1);
|
|
return (max_freq ? XTALMAXFREQ : (XTALMINFREQ / div));
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
static void
|
|
si_clkctl_setdelay(si_info_t *sii, void *chipcregs)
|
|
{
|
|
chipcregs_t *cc = (chipcregs_t *)chipcregs;
|
|
uint slowmaxfreq, pll_delay, slowclk;
|
|
uint pll_on_delay, fref_sel_delay;
|
|
|
|
pll_delay = PLL_DELAY;
|
|
|
|
/* If the slow clock is not sourced by the xtal then add the xtal_on_delay
|
|
* since the xtal will also be powered down by dynamic clk control logic.
|
|
*/
|
|
|
|
slowclk = si_slowclk_src(sii);
|
|
if (slowclk != SCC_SS_XTAL)
|
|
pll_delay += XTAL_ON_DELAY;
|
|
|
|
/* Starting with 4318 it is ILP that is used for the delays */
|
|
slowmaxfreq = si_slowclk_freq(sii, (CCREV(sii->pub.ccrev) >= 10) ? FALSE : TRUE, cc);
|
|
|
|
pll_on_delay = ((slowmaxfreq * pll_delay) + 999999) / 1000000;
|
|
fref_sel_delay = ((slowmaxfreq * FREF_DELAY) + 999999) / 1000000;
|
|
|
|
W_REG(sii->osh, &cc->pll_on_delay, pll_on_delay);
|
|
W_REG(sii->osh, &cc->fref_sel_delay, fref_sel_delay);
|
|
}
|
|
|
|
/** initialize power control delay registers */
|
|
void
|
|
si_clkctl_init(si_t *sih)
|
|
{
|
|
si_info_t *sii;
|
|
uint origidx = 0;
|
|
chipcregs_t *cc;
|
|
bool fast;
|
|
|
|
if (!CCCTL_ENAB(sih))
|
|
return;
|
|
|
|
sii = SI_INFO(sih);
|
|
fast = SI_FAST(sii);
|
|
if (!fast) {
|
|
origidx = sii->curidx;
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
return;
|
|
} else if ((cc = (chipcregs_t *)CCREGS_FAST(sii)) == NULL)
|
|
return;
|
|
ASSERT(cc != NULL);
|
|
|
|
/* set all Instaclk chip ILP to 1 MHz */
|
|
if (CCREV(sih->ccrev) >= 10)
|
|
SET_REG(sii->osh, &cc->system_clk_ctl, SYCC_CD_MASK,
|
|
(ILP_DIV_1MHZ << SYCC_CD_SHIFT));
|
|
|
|
si_clkctl_setdelay(sii, (void *)(uintptr)cc);
|
|
|
|
/* PR 110294 */
|
|
OSL_DELAY(20000);
|
|
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
/** return the value suitable for writing to the dot11 core FAST_PWRUP_DELAY register */
|
|
uint16
|
|
si_clkctl_fast_pwrup_delay(si_t *sih)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
uint origidx = 0;
|
|
chipcregs_t *cc;
|
|
uint slowminfreq;
|
|
uint16 fpdelay;
|
|
bcm_int_bitmask_t intr_val;
|
|
bool fast;
|
|
|
|
if (PMUCTL_ENAB(sih)) {
|
|
INTR_OFF(sii, &intr_val);
|
|
fpdelay = si_pmu_fast_pwrup_delay(sih, sii->osh);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
return fpdelay;
|
|
}
|
|
|
|
if (!CCCTL_ENAB(sih))
|
|
return 0;
|
|
|
|
fast = SI_FAST(sii);
|
|
fpdelay = 0;
|
|
if (!fast) {
|
|
origidx = sii->curidx;
|
|
INTR_OFF(sii, &intr_val);
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
goto done;
|
|
} else if ((cc = (chipcregs_t *)CCREGS_FAST(sii)) == NULL) {
|
|
goto done;
|
|
}
|
|
|
|
ASSERT(cc != NULL);
|
|
|
|
slowminfreq = si_slowclk_freq(sii, FALSE, cc);
|
|
if (slowminfreq > 0)
|
|
fpdelay = (((R_REG(sii->osh, &cc->pll_on_delay) + 2) * 1000000) +
|
|
(slowminfreq - 1)) / slowminfreq;
|
|
|
|
done:
|
|
if (!fast) {
|
|
si_setcoreidx(sih, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
return fpdelay;
|
|
}
|
|
|
|
/** turn primary xtal and/or pll off/on */
|
|
int
|
|
si_clkctl_xtal(si_t *sih, uint what, bool on)
|
|
{
|
|
si_info_t *sii;
|
|
uint32 in, out, outen;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
switch (BUSTYPE(sih->bustype)) {
|
|
|
|
#ifdef BCMSDIO
|
|
case SDIO_BUS:
|
|
return (-1);
|
|
#endif /* BCMSDIO */
|
|
|
|
case PCI_BUS:
|
|
/* pcie core doesn't have any mapping to control the xtal pu */
|
|
if (PCIE(sii) || PCIE_GEN2(sii))
|
|
return -1;
|
|
|
|
in = OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_IN, sizeof(uint32));
|
|
out = OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_OUT, sizeof(uint32));
|
|
outen = OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_OUTEN, sizeof(uint32));
|
|
|
|
/*
|
|
* Avoid glitching the clock if GPRS is already using it.
|
|
* We can't actually read the state of the PLLPD so we infer it
|
|
* by the value of XTAL_PU which *is* readable via gpioin.
|
|
*/
|
|
if (on && (in & PCI_CFG_GPIO_XTAL))
|
|
return (0);
|
|
|
|
if (what & XTAL)
|
|
outen |= PCI_CFG_GPIO_XTAL;
|
|
if (what & PLL)
|
|
outen |= PCI_CFG_GPIO_PLL;
|
|
|
|
if (on) {
|
|
/* turn primary xtal on */
|
|
if (what & XTAL) {
|
|
out |= PCI_CFG_GPIO_XTAL;
|
|
if (what & PLL)
|
|
out |= PCI_CFG_GPIO_PLL;
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_GPIO_OUT,
|
|
sizeof(uint32), out);
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_GPIO_OUTEN,
|
|
sizeof(uint32), outen);
|
|
OSL_DELAY(XTAL_ON_DELAY);
|
|
}
|
|
|
|
/* turn pll on */
|
|
if (what & PLL) {
|
|
out &= ~PCI_CFG_GPIO_PLL;
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_GPIO_OUT,
|
|
sizeof(uint32), out);
|
|
OSL_DELAY(2000);
|
|
}
|
|
} else {
|
|
if (what & XTAL)
|
|
out &= ~PCI_CFG_GPIO_XTAL;
|
|
if (what & PLL)
|
|
out |= PCI_CFG_GPIO_PLL;
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_GPIO_OUT, sizeof(uint32), out);
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_GPIO_OUTEN, sizeof(uint32),
|
|
outen);
|
|
}
|
|
return 0;
|
|
|
|
default:
|
|
return (-1);
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
/**
|
|
* clock control policy function throught chipcommon
|
|
*
|
|
* set dynamic clk control mode (forceslow, forcefast, dynamic)
|
|
* returns true if we are forcing fast clock
|
|
* this is a wrapper over the next internal function
|
|
* to allow flexible policy settings for outside caller
|
|
*/
|
|
bool
|
|
si_clkctl_cc(si_t *sih, uint mode)
|
|
{
|
|
si_info_t *sii;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
/* chipcommon cores prior to rev6 don't support dynamic clock control */
|
|
if (CCREV(sih->ccrev) < 6)
|
|
return FALSE;
|
|
|
|
return _si_clkctl_cc(sii, mode);
|
|
}
|
|
|
|
/* clk control mechanism through chipcommon, no policy checking */
|
|
static bool
|
|
_si_clkctl_cc(si_info_t *sii, uint mode)
|
|
{
|
|
uint origidx = 0;
|
|
chipcregs_t *cc;
|
|
uint32 scc;
|
|
bcm_int_bitmask_t intr_val;
|
|
bool fast = SI_FAST(sii);
|
|
|
|
/* chipcommon cores prior to rev6 don't support dynamic clock control */
|
|
if (CCREV(sii->pub.ccrev) < 6)
|
|
return (FALSE);
|
|
|
|
/* Chips with ccrev 10 are EOL and they don't have SYCC_HR which we use below */
|
|
ASSERT(CCREV(sii->pub.ccrev) != 10);
|
|
|
|
if (!fast) {
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = sii->curidx;
|
|
cc = (chipcregs_t *) si_setcore(&sii->pub, CC_CORE_ID, 0);
|
|
} else if ((cc = (chipcregs_t *) CCREGS_FAST(sii)) == NULL)
|
|
goto done;
|
|
ASSERT(cc != NULL);
|
|
|
|
if (!CCCTL_ENAB(&sii->pub) && (CCREV(sii->pub.ccrev) < 20))
|
|
goto done;
|
|
|
|
switch (mode) {
|
|
case CLK_FAST: /* FORCEHT, fast (pll) clock */
|
|
if (CCREV(sii->pub.ccrev) < 10) {
|
|
/* don't forget to force xtal back on before we clear SCC_DYN_XTAL.. */
|
|
si_clkctl_xtal(&sii->pub, XTAL, ON);
|
|
SET_REG(sii->osh, &cc->slow_clk_ctl, (SCC_XC | SCC_FS | SCC_IP), SCC_IP);
|
|
} else if (CCREV(sii->pub.ccrev) < 20) {
|
|
OR_REG(sii->osh, &cc->system_clk_ctl, SYCC_HR);
|
|
} else {
|
|
OR_REG(sii->osh, &cc->clk_ctl_st, CCS_FORCEHT);
|
|
}
|
|
|
|
/* wait for the PLL */
|
|
if (PMUCTL_ENAB(&sii->pub)) {
|
|
uint32 htavail = CCS_HTAVAIL;
|
|
SPINWAIT(((R_REG(sii->osh, &cc->clk_ctl_st) & htavail) == 0),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
ASSERT(R_REG(sii->osh, &cc->clk_ctl_st) & htavail);
|
|
} else {
|
|
OSL_DELAY(PLL_DELAY);
|
|
}
|
|
break;
|
|
|
|
case CLK_DYNAMIC: /* enable dynamic clock control */
|
|
if (CCREV(sii->pub.ccrev) < 10) {
|
|
scc = R_REG(sii->osh, &cc->slow_clk_ctl);
|
|
scc &= ~(SCC_FS | SCC_IP | SCC_XC);
|
|
if ((scc & SCC_SS_MASK) != SCC_SS_XTAL)
|
|
scc |= SCC_XC;
|
|
W_REG(sii->osh, &cc->slow_clk_ctl, scc);
|
|
|
|
/* for dynamic control, we have to release our xtal_pu "force on" */
|
|
if (scc & SCC_XC)
|
|
si_clkctl_xtal(&sii->pub, XTAL, OFF);
|
|
} else if (CCREV(sii->pub.ccrev) < 20) {
|
|
/* Instaclock */
|
|
AND_REG(sii->osh, &cc->system_clk_ctl, ~SYCC_HR);
|
|
} else {
|
|
AND_REG(sii->osh, &cc->clk_ctl_st, ~CCS_FORCEHT);
|
|
}
|
|
|
|
/* wait for the PLL */
|
|
if (PMUCTL_ENAB(&sii->pub)) {
|
|
uint32 htavail = CCS_HTAVAIL;
|
|
SPINWAIT(((R_REG(sii->osh, &cc->clk_ctl_st) & htavail) != 0),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
ASSERT(!(R_REG(sii->osh, &cc->clk_ctl_st) & htavail));
|
|
} else {
|
|
OSL_DELAY(PLL_DELAY);
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
ASSERT(0);
|
|
}
|
|
|
|
done:
|
|
if (!fast) {
|
|
si_setcoreidx(&sii->pub, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
return (mode == CLK_FAST);
|
|
}
|
|
|
|
/** Build device path. Support SI, PCI for now. */
|
|
int
|
|
BCMNMIATTACHFN(si_devpath)(const si_t *sih, char *path, int size)
|
|
{
|
|
int slen;
|
|
|
|
ASSERT(path != NULL);
|
|
ASSERT(size >= SI_DEVPATH_BUFSZ);
|
|
|
|
if (!path || size <= 0)
|
|
return -1;
|
|
|
|
switch (BUSTYPE(sih->bustype)) {
|
|
case SI_BUS:
|
|
slen = snprintf(path, (size_t)size, "sb/%u/", si_coreidx(sih));
|
|
break;
|
|
case PCI_BUS:
|
|
ASSERT((SI_INFO(sih))->osh != NULL);
|
|
slen = snprintf(path, (size_t)size, "pci/%u/%u/",
|
|
OSL_PCI_BUS((SI_INFO(sih))->osh),
|
|
OSL_PCI_SLOT((SI_INFO(sih))->osh));
|
|
break;
|
|
#ifdef BCMSDIO
|
|
case SDIO_BUS:
|
|
SI_ERROR(("si_devpath: device 0 assumed\n"));
|
|
slen = snprintf(path, (size_t)size, "sd/%u/", si_coreidx(sih));
|
|
break;
|
|
#endif
|
|
default:
|
|
slen = -1;
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
|
|
if (slen < 0 || slen >= size) {
|
|
path[0] = '\0';
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
BCMNMIATTACHFN(si_devpath_pcie)(const si_t *sih, char *path, int size)
|
|
{
|
|
ASSERT(path != NULL);
|
|
ASSERT(size >= SI_DEVPATH_BUFSZ);
|
|
|
|
if (!path || size <= 0)
|
|
return -1;
|
|
|
|
ASSERT((SI_INFO(sih))->osh != NULL);
|
|
snprintf(path, (size_t)size, "pcie/%u/%u/",
|
|
OSL_PCIE_DOMAIN((SI_INFO(sih))->osh),
|
|
OSL_PCIE_BUS((SI_INFO(sih))->osh));
|
|
|
|
return 0;
|
|
}
|
|
|
|
char *
|
|
si_coded_devpathvar(const si_t *sih, char *varname, int var_len, const char *name)
|
|
{
|
|
char pathname[SI_DEVPATH_BUFSZ + 32];
|
|
char devpath[SI_DEVPATH_BUFSZ + 32];
|
|
char devpath_pcie[SI_DEVPATH_BUFSZ + 32];
|
|
char *p;
|
|
int idx;
|
|
int len1;
|
|
int len2;
|
|
int len3 = 0;
|
|
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
snprintf(devpath_pcie, SI_DEVPATH_BUFSZ, "pcie/%u/%u",
|
|
OSL_PCIE_DOMAIN((SI_INFO(sih))->osh),
|
|
OSL_PCIE_BUS((SI_INFO(sih))->osh));
|
|
len3 = strlen(devpath_pcie);
|
|
}
|
|
|
|
/* try to get compact devpath if it exist */
|
|
if (si_devpath(sih, devpath, SI_DEVPATH_BUFSZ) == 0) {
|
|
/* devpath now is 'zzz/zz/', adjust length to */
|
|
/* eliminate ending '/' (if present) */
|
|
len1 = strlen(devpath);
|
|
if (devpath[len1 - 1] == '/')
|
|
len1--;
|
|
|
|
for (idx = 0; idx < SI_MAXCORES; idx++) {
|
|
snprintf(pathname, SI_DEVPATH_BUFSZ, rstr_devpathD, idx);
|
|
if ((p = getvar(NULL, pathname)) == NULL)
|
|
continue;
|
|
|
|
/* eliminate ending '/' (if present) */
|
|
len2 = strlen(p);
|
|
if (p[len2 - 1] == '/')
|
|
len2--;
|
|
|
|
/* check that both lengths match and if so compare */
|
|
/* the strings (minus trailing '/'s if present */
|
|
if ((len1 == len2) && (memcmp(p, devpath, len1) == 0)) {
|
|
snprintf(varname, var_len, rstr_D_S, idx, name);
|
|
return varname;
|
|
}
|
|
|
|
/* try the new PCIe devpath format if it exists */
|
|
if (len3 && (len3 == len2) && (memcmp(p, devpath_pcie, len3) == 0)) {
|
|
snprintf(varname, var_len, rstr_D_S, idx, name);
|
|
return varname;
|
|
}
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/** Get a variable, but only if it has a devpath prefix */
|
|
char *
|
|
si_getdevpathvar(const si_t *sih, const char *name)
|
|
{
|
|
char varname[SI_DEVPATH_BUFSZ + 32];
|
|
char *val;
|
|
|
|
si_devpathvar(sih, varname, sizeof(varname), name);
|
|
|
|
if ((val = getvar(NULL, varname)) != NULL)
|
|
return val;
|
|
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
si_pcie_devpathvar(sih, varname, sizeof(varname), name);
|
|
if ((val = getvar(NULL, varname)) != NULL)
|
|
return val;
|
|
}
|
|
|
|
/* try to get compact devpath if it exist */
|
|
if (si_coded_devpathvar(sih, varname, sizeof(varname), name) == NULL)
|
|
return NULL;
|
|
|
|
return (getvar(NULL, varname));
|
|
}
|
|
|
|
/** Get a variable, but only if it has a devpath prefix */
|
|
int
|
|
si_getdevpathintvar(const si_t *sih, const char *name)
|
|
{
|
|
#if defined(BCMBUSTYPE) && (BCMBUSTYPE == SI_BUS)
|
|
BCM_REFERENCE(sih);
|
|
return (getintvar(NULL, name));
|
|
#else
|
|
char varname[SI_DEVPATH_BUFSZ + 32];
|
|
int val;
|
|
|
|
si_devpathvar(sih, varname, sizeof(varname), name);
|
|
|
|
if ((val = getintvar(NULL, varname)) != 0)
|
|
return val;
|
|
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
si_pcie_devpathvar(sih, varname, sizeof(varname), name);
|
|
if ((val = getintvar(NULL, varname)) != 0)
|
|
return val;
|
|
}
|
|
|
|
/* try to get compact devpath if it exist */
|
|
if (si_coded_devpathvar(sih, varname, sizeof(varname), name) == NULL)
|
|
return 0;
|
|
|
|
return (getintvar(NULL, varname));
|
|
#endif /* BCMBUSTYPE && BCMBUSTYPE == SI_BUS */
|
|
}
|
|
|
|
/**
|
|
* Concatenate the dev path with a varname into the given 'var' buffer
|
|
* and return the 'var' pointer.
|
|
* Nothing is done to the arguments if len == 0 or var is NULL, var is still returned.
|
|
* On overflow, the first char will be set to '\0'.
|
|
*/
|
|
static char *
|
|
si_devpathvar(const si_t *sih, char *var, int len, const char *name)
|
|
{
|
|
uint path_len;
|
|
|
|
if (!var || len <= 0)
|
|
return var;
|
|
|
|
if (si_devpath(sih, var, len) == 0) {
|
|
path_len = strlen(var);
|
|
|
|
if (strlen(name) + 1 > (uint)(len - path_len))
|
|
var[0] = '\0';
|
|
else
|
|
strlcpy(var + path_len, name, len - path_len);
|
|
}
|
|
|
|
return var;
|
|
}
|
|
|
|
static char *
|
|
si_pcie_devpathvar(const si_t *sih, char *var, int len, const char *name)
|
|
{
|
|
uint path_len;
|
|
|
|
if (!var || len <= 0)
|
|
return var;
|
|
|
|
if (si_devpath_pcie(sih, var, len) == 0) {
|
|
path_len = strlen(var);
|
|
|
|
if (strlen(name) + 1 > (uint)(len - path_len))
|
|
var[0] = '\0';
|
|
else
|
|
strlcpy(var + path_len, name, len - path_len);
|
|
}
|
|
|
|
return var;
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_ccreg)(si_t *sih, uint32 offset, uint32 mask, uint32 val)
|
|
{
|
|
si_info_t *sii;
|
|
uint32 reg_val = 0;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
/* abort for invalid offset */
|
|
if (offset > sizeof(chipcregs_t))
|
|
return 0;
|
|
|
|
reg_val = si_corereg(&sii->pub, SI_CC_IDX, offset, mask, val);
|
|
|
|
return reg_val;
|
|
}
|
|
|
|
void
|
|
sih_write_sraon(si_t *sih, int offset, int len, const uint32* data)
|
|
{
|
|
chipcregs_t *cc;
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
W_REG(si_osh(sih), &cc->sr_memrw_addr, offset);
|
|
while (len > 0) {
|
|
W_REG(si_osh(sih), &cc->sr_memrw_data, *data);
|
|
data++;
|
|
len -= sizeof(uint32);
|
|
}
|
|
}
|
|
|
|
#ifdef SR_DEBUG
|
|
void
|
|
si_dump_pmu(si_t *sih, void *arg)
|
|
{
|
|
uint i;
|
|
uint32 pmu_chip_ctl_reg;
|
|
uint32 pmu_chip_reg_reg;
|
|
uint32 pmu_chip_pll_reg;
|
|
uint32 pmu_chip_res_reg;
|
|
pmu_reg_t *pmu_var = (pmu_reg_t*)arg;
|
|
pmu_var->pmu_control = si_ccreg(sih, PMU_CTL, 0, 0);
|
|
pmu_var->pmu_capabilities = si_ccreg(sih, PMU_CAP, 0, 0);
|
|
pmu_var->pmu_status = si_ccreg(sih, PMU_ST, 0, 0);
|
|
pmu_var->res_state = si_ccreg(sih, PMU_RES_STATE, 0, 0);
|
|
pmu_var->res_pending = si_ccreg(sih, PMU_RES_PENDING, 0, 0);
|
|
pmu_var->pmu_timer1 = si_ccreg(sih, PMU_TIMER, 0, 0);
|
|
pmu_var->min_res_mask = si_ccreg(sih, MINRESMASKREG, 0, 0);
|
|
pmu_var->max_res_mask = si_ccreg(sih, MAXRESMASKREG, 0, 0);
|
|
pmu_chip_ctl_reg = (pmu_var->pmu_capabilities & 0xf8000000);
|
|
pmu_chip_ctl_reg = pmu_chip_ctl_reg >> 27;
|
|
for (i = 0; i < pmu_chip_ctl_reg; i++) {
|
|
pmu_var->pmu_chipcontrol1[i] = si_pmu_chipcontrol(sih, i, 0, 0);
|
|
}
|
|
pmu_chip_reg_reg = (pmu_var->pmu_capabilities & 0x07c00000);
|
|
pmu_chip_reg_reg = pmu_chip_reg_reg >> 22;
|
|
for (i = 0; i < pmu_chip_reg_reg; i++) {
|
|
pmu_var->pmu_regcontrol[i] = si_pmu_vreg_control(sih, i, 0, 0);
|
|
}
|
|
pmu_chip_pll_reg = (pmu_var->pmu_capabilities & 0x003e0000);
|
|
pmu_chip_pll_reg = pmu_chip_pll_reg >> 17;
|
|
for (i = 0; i < pmu_chip_pll_reg; i++) {
|
|
pmu_var->pmu_pllcontrol[i] = si_pmu_pllcontrol(sih, i, 0, 0);
|
|
}
|
|
pmu_chip_res_reg = (pmu_var->pmu_capabilities & 0x00001f00);
|
|
pmu_chip_res_reg = pmu_chip_res_reg >> 8;
|
|
for (i = 0; i < pmu_chip_res_reg; i++) {
|
|
si_corereg(sih, SI_CC_IDX, RSRCTABLEADDR, ~0, i);
|
|
pmu_var->pmu_rsrc_up_down_timer[i] = si_corereg(sih, SI_CC_IDX,
|
|
RSRCUPDWNTIME, 0, 0);
|
|
}
|
|
pmu_chip_res_reg = (pmu_var->pmu_capabilities & 0x00001f00);
|
|
pmu_chip_res_reg = pmu_chip_res_reg >> 8;
|
|
for (i = 0; i < pmu_chip_res_reg; i++) {
|
|
si_corereg(sih, SI_CC_IDX, RSRCTABLEADDR, ~0, i);
|
|
pmu_var->rsrc_dep_mask[i] = si_corereg(sih, SI_CC_IDX, PMU_RES_DEP_MASK, 0, 0);
|
|
}
|
|
}
|
|
|
|
void
|
|
si_pmu_keep_on(const si_t *sih, int32 int_val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
uint32 res_dep_mask;
|
|
uint32 min_res_mask;
|
|
uint32 max_res_mask;
|
|
|
|
/* Corresponding Resource Dependancy Mask */
|
|
W_REG(sii->osh, &cc->res_table_sel, int_val);
|
|
res_dep_mask = R_REG(sii->osh, &cc->res_dep_mask);
|
|
/* Local change of minimum resource mask */
|
|
min_res_mask = res_dep_mask | 1 << int_val;
|
|
/* Corresponding change of Maximum Resource Mask */
|
|
max_res_mask = R_REG(sii->osh, &cc->max_res_mask);
|
|
max_res_mask = max_res_mask | min_res_mask;
|
|
W_REG(sii->osh, &cc->max_res_mask, max_res_mask);
|
|
/* Corresponding change of Minimum Resource Mask */
|
|
W_REG(sii->osh, &cc->min_res_mask, min_res_mask);
|
|
}
|
|
|
|
uint32
|
|
si_pmu_keep_on_get(const si_t *sih)
|
|
{
|
|
uint i;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
uint32 res_dep_mask;
|
|
uint32 min_res_mask;
|
|
|
|
/* Read min res mask */
|
|
min_res_mask = R_REG(sii->osh, &cc->min_res_mask);
|
|
/* Get corresponding Resource Dependancy Mask */
|
|
for (i = 0; i < PMU_RES; i++) {
|
|
W_REG(sii->osh, &cc->res_table_sel, i);
|
|
res_dep_mask = R_REG(sii->osh, &cc->res_dep_mask);
|
|
res_dep_mask = res_dep_mask | 1 << i;
|
|
/* Compare with the present min res mask */
|
|
if (res_dep_mask == min_res_mask) {
|
|
return i;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
uint32
|
|
si_power_island_set(si_t *sih, uint32 int_val)
|
|
{
|
|
uint32 i = 0x0;
|
|
uint32 j;
|
|
uint32 k;
|
|
int cnt = 0;
|
|
for (k = 0; k < ARRAYSIZE(si_power_island_test_array); k++) {
|
|
if (int_val == si_power_island_test_array[k]) {
|
|
cnt = cnt + 1;
|
|
}
|
|
}
|
|
if (cnt > 0) {
|
|
if (int_val & SUBCORE_POWER_ON) {
|
|
i = i | 0x1;
|
|
}
|
|
if (int_val & PHY_POWER_ON) {
|
|
i = i | 0x2;
|
|
}
|
|
if (int_val & VDDM_POWER_ON) {
|
|
i = i | 0x4;
|
|
}
|
|
if (int_val & MEMLPLDO_POWER_ON) {
|
|
i = i | 0x8;
|
|
}
|
|
j = (i << 18) & 0x003c0000;
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0x003c0000, j);
|
|
} else {
|
|
return 0;
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
uint32
|
|
si_power_island_get(si_t *sih)
|
|
{
|
|
uint32 sc_on = 0x0;
|
|
uint32 phy_on = 0x0;
|
|
uint32 vddm_on = 0x0;
|
|
uint32 memlpldo_on = 0x0;
|
|
uint32 res;
|
|
uint32 reg_val;
|
|
reg_val = si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0, 0);
|
|
if (reg_val & SUBCORE_POWER_ON_CHK) {
|
|
sc_on = SUBCORE_POWER_ON;
|
|
}
|
|
if (reg_val & PHY_POWER_ON_CHK) {
|
|
phy_on = PHY_POWER_ON;
|
|
}
|
|
if (reg_val & VDDM_POWER_ON_CHK) {
|
|
vddm_on = VDDM_POWER_ON;
|
|
}
|
|
if (reg_val & MEMLPLDO_POWER_ON_CHK) {
|
|
memlpldo_on = MEMLPLDO_POWER_ON;
|
|
}
|
|
res = (sc_on | phy_on | vddm_on | memlpldo_on);
|
|
return res;
|
|
}
|
|
#endif /* SR_DEBUG */
|
|
|
|
uint32
|
|
si_pciereg(const si_t *sih, uint32 offset, uint32 mask, uint32 val, uint type)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii)) {
|
|
SI_ERROR(("si_pciereg: Not a PCIE device\n"));
|
|
return 0;
|
|
}
|
|
|
|
return pcicore_pciereg(sii->pch, offset, mask, val, type);
|
|
}
|
|
|
|
uint32
|
|
si_pcieserdesreg(const si_t *sih, uint32 mdioslave, uint32 offset, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii)) {
|
|
SI_ERROR(("si_pcieserdesreg: Not a PCIE device\n"));
|
|
return 0;
|
|
}
|
|
|
|
return pcicore_pcieserdesreg(sii->pch, mdioslave, offset, mask, val);
|
|
|
|
}
|
|
|
|
/** return TRUE if PCIE capability exists in the pci config space */
|
|
static bool
|
|
si_ispcie(const si_info_t *sii)
|
|
{
|
|
uint8 cap_ptr;
|
|
|
|
if (BUSTYPE(sii->pub.bustype) != PCI_BUS)
|
|
return FALSE;
|
|
|
|
cap_ptr = pcicore_find_pci_capability(sii->osh, PCI_CAP_PCIECAP_ID, NULL, NULL);
|
|
if (!cap_ptr)
|
|
return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* Wake-on-wireless-LAN (WOWL) support functions */
|
|
/** Enable PME generation and disable clkreq */
|
|
void
|
|
si_pci_pmeen(const si_t *sih)
|
|
{
|
|
pcicore_pmeen(SI_INFO(sih)->pch);
|
|
}
|
|
|
|
/** Return TRUE if PME status is set */
|
|
bool
|
|
si_pci_pmestat(const si_t *sih)
|
|
{
|
|
return pcicore_pmestat(SI_INFO(sih)->pch);
|
|
}
|
|
|
|
/** Disable PME generation, clear the PME status bit if set */
|
|
void
|
|
si_pci_pmeclr(const si_t *sih)
|
|
{
|
|
pcicore_pmeclr(SI_INFO(sih)->pch);
|
|
}
|
|
|
|
void
|
|
si_pci_pmestatclr(const si_t *sih)
|
|
{
|
|
pcicore_pmestatclr(SI_INFO(sih)->pch);
|
|
}
|
|
|
|
#ifdef BCMSDIO
|
|
/** initialize the sdio core */
|
|
void
|
|
si_sdio_init(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (BUSCORETYPE(sih->buscoretype) == SDIOD_CORE_ID) {
|
|
uint idx;
|
|
sdpcmd_regs_t *sdpregs;
|
|
|
|
/* get the current core index */
|
|
/* could do stuff like tcpflag in pci, but why? */
|
|
idx = sii->curidx;
|
|
ASSERT(idx == si_findcoreidx(sih, D11_CORE_ID, 0));
|
|
|
|
/* switch to sdio core */
|
|
/* could use buscoreidx? */
|
|
sdpregs = (sdpcmd_regs_t *)si_setcore(sih, SDIOD_CORE_ID, 0);
|
|
ASSERT(sdpregs);
|
|
|
|
SI_MSG(("si_sdio_init: For SDIO Corerev %d, enable ints from core %d "
|
|
"through SD core %d (%p)\n",
|
|
sih->buscorerev, idx, sii->curidx, OSL_OBFUSCATE_BUF(sdpregs)));
|
|
|
|
/* enable backplane error and core interrupts */
|
|
W_REG(sii->osh, &sdpregs->hostintmask, I_SBINT);
|
|
W_REG(sii->osh, &sdpregs->sbintmask, (I_SB_SERR | I_SB_RESPERR | (1 << idx)));
|
|
|
|
/* switch back to previous core */
|
|
si_setcoreidx(sih, idx);
|
|
}
|
|
|
|
/* enable interrupts */
|
|
bcmsdh_intr_enable(sii->sdh);
|
|
|
|
/* What else */
|
|
}
|
|
#endif /* BCMSDIO */
|
|
|
|
/**
|
|
* Disable pcie_war_ovr for some platforms (sigh!)
|
|
* This is for boards that have BFL2_PCIEWAR_OVR set
|
|
* but are in systems that still want the benefits of ASPM
|
|
* Note that this should be done AFTER si_doattach
|
|
*/
|
|
void
|
|
si_pcie_war_ovr_update(const si_t *sih, uint8 aspm)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii))
|
|
return;
|
|
|
|
pcie_war_ovr_aspm_update(sii->pch, aspm);
|
|
}
|
|
|
|
void
|
|
si_pcie_power_save_enable(const si_t *sih, bool enable)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii))
|
|
return;
|
|
|
|
pcie_power_save_enable(sii->pch, enable);
|
|
}
|
|
|
|
void
|
|
si_pcie_set_maxpayload_size(const si_t *sih, uint16 size)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return;
|
|
|
|
pcie_set_maxpayload_size(sii->pch, size);
|
|
}
|
|
|
|
uint16
|
|
si_pcie_get_maxpayload_size(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return (0);
|
|
|
|
return pcie_get_maxpayload_size(sii->pch);
|
|
}
|
|
|
|
void
|
|
si_pcie_set_request_size(const si_t *sih, uint16 size)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return;
|
|
|
|
pcie_set_request_size(sii->pch, size);
|
|
}
|
|
|
|
uint16
|
|
si_pcie_get_request_size(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii))
|
|
return (0);
|
|
|
|
return pcie_get_request_size(sii->pch);
|
|
}
|
|
|
|
uint16
|
|
si_pcie_get_ssid(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii))
|
|
return (0);
|
|
|
|
return pcie_get_ssid(sii->pch);
|
|
}
|
|
|
|
uint32
|
|
si_pcie_get_bar0(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return (0);
|
|
|
|
return pcie_get_bar0(sii->pch);
|
|
}
|
|
|
|
int
|
|
si_pcie_configspace_cache(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return BCME_UNSUPPORTED;
|
|
|
|
return pcie_configspace_cache(sii->pch);
|
|
}
|
|
|
|
int
|
|
si_pcie_configspace_restore(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return BCME_UNSUPPORTED;
|
|
|
|
return pcie_configspace_restore(sii->pch);
|
|
}
|
|
|
|
int
|
|
si_pcie_configspace_get(const si_t *sih, uint8 *buf, uint size)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii) || size > PCI_CONFIG_SPACE_SIZE)
|
|
return -1;
|
|
|
|
return pcie_configspace_get(sii->pch, buf, size);
|
|
}
|
|
|
|
void
|
|
si_pcie_hw_L1SS_war(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
/* SWWLAN-41753: WAR intermittent issue with D3Cold and L1.2 exit,
|
|
* need to update PMU rsrc dependency
|
|
*/
|
|
if (PCIE_GEN2(sii))
|
|
pcie_hw_L1SS_war(sii->pch);
|
|
}
|
|
|
|
void
|
|
si_pci_up(const si_t *sih)
|
|
{
|
|
const si_info_t *sii;
|
|
|
|
/* if not pci bus, we're done */
|
|
if (BUSTYPE(sih->bustype) != PCI_BUS)
|
|
return;
|
|
|
|
sii = SI_INFO(sih);
|
|
|
|
if (PCIE(sii)) {
|
|
pcicore_up(sii->pch, SI_PCIUP);
|
|
}
|
|
}
|
|
|
|
/** Unconfigure and/or apply various WARs when system is going to sleep mode */
|
|
void
|
|
si_pci_sleep(const si_t *sih)
|
|
{
|
|
/* 4360 pcie2 WAR */
|
|
do_4360_pcie2_war = 0;
|
|
|
|
pcicore_sleep(SI_INFO(sih)->pch);
|
|
}
|
|
|
|
/** Unconfigure and/or apply various WARs when the wireless interface is going down */
|
|
void
|
|
si_pci_down(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
BCM_REFERENCE(sii);
|
|
|
|
/* if not pci bus, we're done */
|
|
if (BUSTYPE(sih->bustype) != PCI_BUS)
|
|
return;
|
|
|
|
pcicore_down(sii->pch, SI_PCIDOWN);
|
|
}
|
|
|
|
/**
|
|
* Configure the pci core for pci client (NIC) action
|
|
* coremask is the bitvec of cores by index to be enabled.
|
|
*/
|
|
void
|
|
si_pci_setup(si_t *sih, uint coremask)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
sbpciregs_t *pciregs = NULL;
|
|
uint32 siflag = 0, w;
|
|
uint idx = 0;
|
|
|
|
if (BUSTYPE(sii->pub.bustype) != PCI_BUS)
|
|
return;
|
|
|
|
ASSERT(PCI(sii) || PCIE(sii));
|
|
ASSERT(sii->pub.buscoreidx != BADIDX);
|
|
|
|
if (PCI(sii)) {
|
|
/* get current core index */
|
|
idx = sii->curidx;
|
|
|
|
/* we interrupt on this backplane flag number */
|
|
siflag = si_flag(sih);
|
|
|
|
/* switch over to pci core */
|
|
pciregs = (sbpciregs_t *)si_setcoreidx(sih, sii->pub.buscoreidx);
|
|
}
|
|
|
|
/*
|
|
* Enable sb->pci interrupts. Assume
|
|
* PCI rev 2.3 support was added in pci core rev 6 and things changed..
|
|
*/
|
|
if (PCIE(sii) || (PCI(sii) && ((sii->pub.buscorerev) >= 6))) {
|
|
/* pci config write to set this core bit in PCIIntMask */
|
|
w = OSL_PCI_READ_CONFIG(sii->osh, PCI_INT_MASK, sizeof(uint32));
|
|
w |= (coremask << PCI_SBIM_SHIFT);
|
|
#ifdef USER_MODE
|
|
/* User mode operate with interrupt disabled */
|
|
w &= !(coremask << PCI_SBIM_SHIFT);
|
|
#endif
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, PCI_INT_MASK, sizeof(uint32), w);
|
|
} else {
|
|
/* set sbintvec bit for our flag number */
|
|
si_setint(sih, siflag);
|
|
}
|
|
|
|
/*
|
|
* enable prefetch and bursts for dma big window
|
|
* enable read multiple for dma big window corerev >= 11
|
|
* PR 9962/4708: Set initiator timeouts. corerev < 5
|
|
*/
|
|
if (PCI(sii)) {
|
|
OR_REG(sii->osh, &pciregs->sbtopci2, (SBTOPCI_PREF | SBTOPCI_BURST));
|
|
if (sii->pub.buscorerev >= 11) {
|
|
OR_REG(sii->osh, &pciregs->sbtopci2, SBTOPCI_RC_READMULTI);
|
|
/* PR50531: On some Laptops, the 4321 CB shows bad
|
|
* UDP performance on one direction
|
|
*/
|
|
w = R_REG(sii->osh, &pciregs->clkrun);
|
|
W_REG(sii->osh, &pciregs->clkrun, (w | PCI_CLKRUN_DSBL));
|
|
w = R_REG(sii->osh, &pciregs->clkrun);
|
|
}
|
|
|
|
/* switch back to previous core */
|
|
si_setcoreidx(sih, idx);
|
|
}
|
|
}
|
|
|
|
/* In NIC mode is there any better way to find out what ARM core is there? */
|
|
static uint
|
|
si_get_armcoreidx(si_t *sih)
|
|
{
|
|
uint saveidx = si_coreidx(sih);
|
|
uint coreidx = BADIDX;
|
|
|
|
if (si_setcore(sih, ARMCR4_CORE_ID, 0) != NULL ||
|
|
si_setcore(sih, ARMCA7_CORE_ID, 0) != NULL) {
|
|
coreidx = si_coreidx(sih);
|
|
}
|
|
|
|
si_setcoreidx(sih, saveidx);
|
|
|
|
return coreidx;
|
|
}
|
|
|
|
/**
|
|
* Configure the pcie core for pcie client (NIC) action
|
|
* coreidx is the index of the core to be enabled.
|
|
*/
|
|
int
|
|
si_pcie_setup(si_t *sih, uint coreidx)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
int main_intr, alt_intr;
|
|
uint pciepidx;
|
|
uint32 w;
|
|
osl_t *osh = si_osh(sih);
|
|
uint saveidx = si_coreidx(sih);
|
|
volatile void *oobrregs;
|
|
uint armcidx, armpidx;
|
|
int ret = BCME_OK;
|
|
|
|
/* try the new hnd oobr first */
|
|
if ((oobrregs = si_setcore(sih, HND_OOBR_CORE_ID, 0)) == NULL) {
|
|
goto exit;
|
|
}
|
|
|
|
ASSERT(BUSTYPE(sih->bustype) == PCI_BUS);
|
|
ASSERT(BUSTYPE(sih->buscoretype) == PCIE2_CORE_ID);
|
|
|
|
/* ==== Enable sb->pci interrupts ==== */
|
|
|
|
/* 1) query the pcie interrupt port index and
|
|
* re-route the main interrupt to pcie (from ARM) if necessary
|
|
*/
|
|
main_intr = hnd_oobr_get_intr_config(sih, coreidx,
|
|
HND_CORE_MAIN_INTR, sih->buscoreidx, &pciepidx);
|
|
if (main_intr < 0) {
|
|
/* query failure means the main interrupt is not routed
|
|
* to the pcie core... re-route!
|
|
*/
|
|
armcidx = si_get_armcoreidx(sih);
|
|
if (!GOODIDX(armcidx, sii->numcores)) {
|
|
SI_MSG(("si_pcie_setup: arm core not found\n"));
|
|
ret = BCME_NOTFOUND;
|
|
goto exit;
|
|
}
|
|
|
|
/* query main and alt interrupt info */
|
|
main_intr = hnd_oobr_get_intr_config(sih, coreidx,
|
|
HND_CORE_MAIN_INTR, armcidx, &armpidx);
|
|
alt_intr = hnd_oobr_get_intr_config(sih, coreidx,
|
|
HND_CORE_ALT_INTR, sih->buscoreidx, &pciepidx);
|
|
if ((ret = main_intr) < 0 || (ret = alt_intr) < 0) {
|
|
SI_MSG(("si_pcie_setup: coreidx %u main (=%d) or "
|
|
"alt (=%d) interrupt query failed\n",
|
|
coreidx, main_intr, alt_intr));
|
|
goto exit;
|
|
}
|
|
|
|
/* swap main and alt interrupts at pcie input interrupts */
|
|
hnd_oobr_set_intr_src(sih, sih->buscoreidx, pciepidx, main_intr);
|
|
/* TODO: route the alternate interrupt to arm */
|
|
/* hnd_oobr_set_intr_src(sih, armcidx, armppidx, alt_intr); */
|
|
BCM_REFERENCE(armpidx);
|
|
|
|
/* query main interrupt info again.
|
|
* is it really necessary?
|
|
* it can't fail as we just set it up...
|
|
*/
|
|
main_intr = hnd_oobr_get_intr_config(sih, coreidx,
|
|
HND_CORE_MAIN_INTR, sih->buscoreidx, &pciepidx);
|
|
ASSERT(main_intr >= 0);
|
|
}
|
|
/* hnd_oobr_dump(sih); */
|
|
|
|
/* 2) pcie config write to set this core bit in PCIIntMask */
|
|
w = OSL_PCI_READ_CONFIG(osh, PCI_INT_MASK, sizeof(w));
|
|
w |= ((1 << pciepidx) << PCI_SBIM_SHIFT);
|
|
OSL_PCI_WRITE_CONFIG(osh, PCI_INT_MASK, sizeof(w), w);
|
|
|
|
/* ==== other setups ==== */
|
|
|
|
/* reset the return value */
|
|
ret = BCME_OK;
|
|
exit:
|
|
/* return to the original core */
|
|
si_setcoreidx(sih, saveidx);
|
|
if (ret != BCME_OK) {
|
|
return ret;
|
|
}
|
|
|
|
/* fall back to the old way... */
|
|
if (oobrregs == NULL) {
|
|
uint coremask = (1 << coreidx);
|
|
si_pci_setup(sih, coremask);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
uint8
|
|
si_pcieclkreq(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return 0;
|
|
|
|
return pcie_clkreq(sii->pch, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_pcielcreg(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return 0;
|
|
|
|
return pcie_lcreg(sii->pch, mask, val);
|
|
}
|
|
|
|
uint8
|
|
si_pcieltrenable(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!(PCIE(sii)))
|
|
return 0;
|
|
|
|
return pcie_ltrenable(sii->pch, mask, val);
|
|
}
|
|
|
|
uint8
|
|
si_pcieobffenable(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!(PCIE(sii)))
|
|
return 0;
|
|
|
|
return pcie_obffenable(sii->pch, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_pcieltr_reg(const si_t *sih, uint32 reg, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!(PCIE(sii)))
|
|
return 0;
|
|
|
|
return pcie_ltr_reg(sii->pch, reg, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_pcieltrspacing_reg(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!(PCIE(sii)))
|
|
return 0;
|
|
|
|
return pcieltrspacing_reg(sii->pch, mask, val);
|
|
}
|
|
|
|
uint32
|
|
si_pcieltrhysteresiscnt_reg(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!(PCIE(sii)))
|
|
return 0;
|
|
|
|
return pcieltrhysteresiscnt_reg(sii->pch, mask, val);
|
|
}
|
|
|
|
void
|
|
si_pcie_set_error_injection(const si_t *sih, uint32 mode)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE(sii))
|
|
return;
|
|
|
|
pcie_set_error_injection(sii->pch, mode);
|
|
}
|
|
|
|
void
|
|
si_pcie_set_L1substate(const si_t *sih, uint32 substate)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pcie_set_L1substate(sii->pch, substate);
|
|
}
|
|
#ifndef BCM_BOOTLOADER
|
|
uint32
|
|
si_pcie_get_L1substate(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
return pcie_get_L1substate(sii->pch);
|
|
|
|
return 0;
|
|
}
|
|
#endif /* BCM_BOOTLOADER */
|
|
/** indirect way to read pcie config regs */
|
|
uint
|
|
si_pcie_readreg(void *sih, uint addrtype, uint offset)
|
|
{
|
|
return pcie_readreg(sih, (sbpcieregs_t *)PCIEREGS(((si_info_t *)sih)),
|
|
addrtype, offset);
|
|
}
|
|
|
|
/* indirect way to write pcie config regs */
|
|
uint
|
|
si_pcie_writereg(void *sih, uint addrtype, uint offset, uint val)
|
|
{
|
|
return pcie_writereg(sih, (sbpcieregs_t *)PCIEREGS(((si_info_t *)sih)),
|
|
addrtype, offset, val);
|
|
}
|
|
|
|
/**
|
|
* PCI(e) core requires additional software initialization in an SROMless system. In such a system,
|
|
* the PCIe core will assume POR defaults, which are mostly ok, with the exception of the mapping of
|
|
* two address subwindows within the BAR0 window.
|
|
* Note: the current core may be changed upon return.
|
|
*/
|
|
int
|
|
si_pci_fixcfg(si_t *sih)
|
|
{
|
|
#ifndef DONGLEBUILD
|
|
|
|
uint origidx, pciidx;
|
|
sbpciregs_t *pciregs = NULL;
|
|
sbpcieregs_t *pcieregs = NULL;
|
|
uint16 val16;
|
|
volatile uint16 *reg16 = NULL;
|
|
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT(BUSTYPE(sii->pub.bustype) == PCI_BUS);
|
|
|
|
/* Fixup PI in SROM shadow area to enable the correct PCI core access */
|
|
origidx = si_coreidx(&sii->pub);
|
|
|
|
/* check 'pi' is correct and fix it if not. */
|
|
if (BUSCORETYPE(sii->pub.buscoretype) == PCIE2_CORE_ID) {
|
|
pcieregs = (sbpcieregs_t *)si_setcore(&sii->pub, PCIE2_CORE_ID, 0);
|
|
ASSERT(pcieregs != NULL);
|
|
reg16 = &pcieregs->sprom[SRSH_PI_OFFSET];
|
|
} else if (BUSCORETYPE(sii->pub.buscoretype) == PCIE_CORE_ID) {
|
|
pcieregs = (sbpcieregs_t *)si_setcore(&sii->pub, PCIE_CORE_ID, 0);
|
|
ASSERT(pcieregs != NULL);
|
|
reg16 = &pcieregs->sprom[SRSH_PI_OFFSET];
|
|
} else if (BUSCORETYPE(sii->pub.buscoretype) == PCI_CORE_ID) {
|
|
pciregs = (sbpciregs_t *)si_setcore(&sii->pub, PCI_CORE_ID, 0);
|
|
ASSERT(pciregs != NULL);
|
|
reg16 = &pciregs->sprom[SRSH_PI_OFFSET];
|
|
}
|
|
pciidx = si_coreidx(&sii->pub);
|
|
|
|
if (!reg16) return -1;
|
|
|
|
val16 = R_REG(sii->osh, reg16);
|
|
if (((val16 & SRSH_PI_MASK) >> SRSH_PI_SHIFT) != (uint16)pciidx) {
|
|
/* write bitfield used to translate 3rd and 7th 4K chunk in the Bar0 space. */
|
|
val16 = (uint16)(pciidx << SRSH_PI_SHIFT) | (val16 & ~SRSH_PI_MASK);
|
|
W_REG(sii->osh, reg16, val16);
|
|
}
|
|
|
|
/* restore the original index */
|
|
si_setcoreidx(&sii->pub, origidx);
|
|
|
|
pcicore_hwup(sii->pch);
|
|
#endif /* DONGLEBUILD */
|
|
return 0;
|
|
} /* si_pci_fixcfg */
|
|
|
|
#if defined(BCMDBG) || defined(BCMDBG_DUMP) || defined(WLTEST)
|
|
int
|
|
si_dump_pcieinfo(const si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii) && !PCIE_GEN2(sii))
|
|
return BCME_ERROR;
|
|
|
|
return pcicore_dump_pcieinfo(sii->pch, b);
|
|
}
|
|
|
|
void
|
|
si_dump_pmuregs(si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
uint i;
|
|
uint32 pmu_cap;
|
|
uint32 pmu_chip_reg;
|
|
|
|
bcm_bprintf(b, "===pmu(rev %d)===\n", sih->pmurev);
|
|
if (!(sih->pmurev == 0x11 || (sih->pmurev >= 0x15 && sih->pmurev <= 0x19))) {
|
|
bcm_bprintf(b, "PMU dump not supported\n");
|
|
return;
|
|
}
|
|
pmu_cap = si_ccreg(sih, PMU_CAP, 0, 0);
|
|
bcm_bprintf(b, "pmu_control 0x%x\n", si_ccreg(sih, PMU_CTL, 0, 0));
|
|
bcm_bprintf(b, "pmu_capabilities 0x%x\n", pmu_cap);
|
|
bcm_bprintf(b, "pmu_status 0x%x\n", si_ccreg(sih, PMU_ST, 0, 0));
|
|
bcm_bprintf(b, "res_state 0x%x\n", si_ccreg(sih, PMU_RES_STATE, 0, 0));
|
|
bcm_bprintf(b, "res_pending 0x%x\n", si_ccreg(sih, PMU_RES_PENDING, 0, 0));
|
|
bcm_bprintf(b, "pmu_timer1 %d\n", si_ccreg(sih, PMU_TIMER, 0, 0));
|
|
bcm_bprintf(b, "min_res_mask 0x%x\n", si_ccreg(sih, MINRESMASKREG, 0, 0));
|
|
bcm_bprintf(b, "max_res_mask 0x%x\n", si_ccreg(sih, MAXRESMASKREG, 0, 0));
|
|
|
|
pmu_chip_reg = (pmu_cap & 0xf8000000);
|
|
pmu_chip_reg = pmu_chip_reg >> 27;
|
|
bcm_bprintf(b, "si_pmu_chipcontrol: ");
|
|
for (i = 0; i < pmu_chip_reg; i++) {
|
|
bcm_bprintf(b, "[%d]=0x%x ", i, si_pmu_chipcontrol(sih, i, 0, 0));
|
|
}
|
|
|
|
pmu_chip_reg = (pmu_cap & 0x07c00000);
|
|
pmu_chip_reg = pmu_chip_reg >> 22;
|
|
bcm_bprintf(b, "\nsi_pmu_vregcontrol: ");
|
|
for (i = 0; i < pmu_chip_reg; i++) {
|
|
bcm_bprintf(b, "[%d]=0x%x ", i, si_pmu_vreg_control(sih, i, 0, 0));
|
|
}
|
|
pmu_chip_reg = (pmu_cap & 0x003e0000);
|
|
pmu_chip_reg = pmu_chip_reg >> 17;
|
|
bcm_bprintf(b, "\nsi_pmu_pllcontrol: ");
|
|
for (i = 0; i < pmu_chip_reg; i++) {
|
|
bcm_bprintf(b, "[%d]=0x%x ", i, si_pmu_pllcontrol(sih, i, 0, 0));
|
|
}
|
|
pmu_chip_reg = (pmu_cap & 0x0001e000);
|
|
pmu_chip_reg = pmu_chip_reg >> 13;
|
|
bcm_bprintf(b, "\nsi_pmu_res u/d timer: ");
|
|
for (i = 0; i < pmu_chip_reg; i++) {
|
|
si_corereg(sih, SI_CC_IDX, RSRCTABLEADDR, ~0, i);
|
|
bcm_bprintf(b, "[%d]=0x%x ", i, si_corereg(sih, SI_CC_IDX, RSRCUPDWNTIME, 0, 0));
|
|
}
|
|
pmu_chip_reg = (pmu_cap & 0x00001f00);
|
|
pmu_chip_reg = pmu_chip_reg >> 8;
|
|
bcm_bprintf(b, "\nsi_pmu_res dep_mask: ");
|
|
for (i = 0; i < pmu_chip_reg; i++) {
|
|
si_corereg(sih, SI_CC_IDX, RSRCTABLEADDR, ~0, i);
|
|
bcm_bprintf(b, "[%d]=0x%x ", i, si_corereg(sih, SI_CC_IDX, PMU_RES_DEP_MASK, 0, 0));
|
|
}
|
|
bcm_bprintf(b, "\n");
|
|
}
|
|
|
|
int
|
|
si_dump_pcieregs(const si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (!PCIE_GEN1(sii) && !PCIE_GEN2(sii))
|
|
return BCME_ERROR;
|
|
|
|
return pcicore_dump_pcieregs(sii->pch, b);
|
|
}
|
|
|
|
#endif /* BCMDBG || BCMDBG_DUMP || WLTEST */
|
|
|
|
#if defined(BCMDBG) || defined(BCMDBG_DUMP) || defined(BCMDBG_PHYDUMP)
|
|
void
|
|
si_dump(const si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const si_cores_info_t *cores_info = (const si_cores_info_t *)sii->cores_info;
|
|
uint i;
|
|
|
|
bcm_bprintf(b, "si %p chip 0x%x chiprev 0x%x boardtype 0x%x boardvendor 0x%x bus %d\n",
|
|
OSL_OBFUSCATE_BUF(sii), sih->chip, sih->chiprev,
|
|
sih->boardtype, sih->boardvendor, sih->bustype);
|
|
bcm_bprintf(b, "osh %p curmap %p\n",
|
|
OSL_OBFUSCATE_BUF(sii->osh), OSL_OBFUSCATE_BUF(sii->curmap));
|
|
|
|
if (CHIPTYPE(sih->socitype) == SOCI_SB)
|
|
bcm_bprintf(b, "sonicsrev %d ", sih->socirev);
|
|
bcm_bprintf(b, "ccrev %d buscoretype 0x%x buscorerev %d curidx %d\n",
|
|
CCREV(sih->ccrev), sih->buscoretype, sih->buscorerev, sii->curidx);
|
|
|
|
#ifdef BCMDBG
|
|
if ((BUSTYPE(sih->bustype) == PCI_BUS) && (sii->pch))
|
|
pcicore_dump(sii->pch, b);
|
|
#endif
|
|
|
|
bcm_bprintf(b, "cores: ");
|
|
for (i = 0; i < sii->numcores; i++)
|
|
bcm_bprintf(b, "0x%x ", cores_info->coreid[i]);
|
|
bcm_bprintf(b, "\n");
|
|
}
|
|
|
|
void
|
|
si_ccreg_dump(si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
uint i;
|
|
bcm_int_bitmask_t intr_val;
|
|
chipcregs_t *cc;
|
|
|
|
/* only support corerev 22 for now */
|
|
if (CCREV(sih->ccrev) != 23)
|
|
return;
|
|
|
|
origidx = sii->curidx;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
ASSERT(cc);
|
|
|
|
bcm_bprintf(b, "\n===cc(rev %d) registers(offset val)===\n", CCREV(sih->ccrev));
|
|
for (i = 0; i <= 0xc4; i += 4) {
|
|
if (i == 0x4c) {
|
|
bcm_bprintf(b, "\n");
|
|
continue;
|
|
}
|
|
bcm_bprintf(b, "0x%x\t0x%x\n", i, *(uint32 *)((uintptr)cc + i));
|
|
}
|
|
|
|
bcm_bprintf(b, "\n");
|
|
|
|
for (i = 0x1e0; i <= 0x1e4; i += 4) {
|
|
bcm_bprintf(b, "0x%x\t0x%x\n", i, *(uint32 *)((uintptr)cc + i));
|
|
}
|
|
bcm_bprintf(b, "\n");
|
|
|
|
if (sih->cccaps & CC_CAP_PMU) {
|
|
for (i = 0x600; i <= 0x660; i += 4) {
|
|
bcm_bprintf(b, "0x%x\t0x%x\n", i, *(uint32 *)((uintptr)cc + i));
|
|
}
|
|
}
|
|
bcm_bprintf(b, "\n");
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
/** dump dynamic clock control related registers */
|
|
void
|
|
si_clkctl_dump(si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
if (!(sih->cccaps & CC_CAP_PWR_CTL))
|
|
return;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = sii->curidx;
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
goto done;
|
|
|
|
bcm_bprintf(b, "pll_on_delay 0x%x fref_sel_delay 0x%x ",
|
|
cc->pll_on_delay, cc->fref_sel_delay);
|
|
if ((CCREV(sih->ccrev) >= 6) && (CCREV(sih->ccrev) < 10))
|
|
bcm_bprintf(b, "slow_clk_ctl 0x%x ", cc->slow_clk_ctl);
|
|
if (CCREV(sih->ccrev) >= 10) {
|
|
bcm_bprintf(b, "system_clk_ctl 0x%x ", cc->system_clk_ctl);
|
|
bcm_bprintf(b, "clkstatestretch 0x%x ", cc->clkstatestretch);
|
|
}
|
|
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS)
|
|
bcm_bprintf(b, "gpioout 0x%x gpioouten 0x%x ",
|
|
OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_OUT, sizeof(uint32)),
|
|
OSL_PCI_READ_CONFIG(sii->osh, PCI_GPIO_OUTEN, sizeof(uint32)));
|
|
|
|
if (sih->cccaps & CC_CAP_PMU) {
|
|
/* dump some PMU register ? */
|
|
}
|
|
bcm_bprintf(b, "\n");
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
int
|
|
si_gpiodump(si_t *sih, struct bcmstrbuf *b)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
chipcregs_t *cc;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
|
|
origidx = si_coreidx(sih);
|
|
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
ASSERT(cc);
|
|
|
|
bcm_bprintf(b, "GPIOregs\t");
|
|
|
|
bcm_bprintf(b, "gpioin 0x%x ", R_REG(sii->osh, &cc->gpioin));
|
|
bcm_bprintf(b, "gpioout 0x%x ", R_REG(sii->osh, &cc->gpioout));
|
|
bcm_bprintf(b, "gpioouten 0x%x ", R_REG(sii->osh, &cc->gpioouten));
|
|
bcm_bprintf(b, "gpiocontrol 0x%x ", R_REG(sii->osh, &cc->gpiocontrol));
|
|
bcm_bprintf(b, "gpiointpolarity 0x%x ", R_REG(sii->osh, &cc->gpiointpolarity));
|
|
bcm_bprintf(b, "gpiointmask 0x%x ", R_REG(sii->osh, &cc->gpiointmask));
|
|
|
|
bcm_bprintf(b, "\n");
|
|
|
|
/* restore the original index */
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
INTR_RESTORE(sii, &intr_val);
|
|
return 0;
|
|
|
|
}
|
|
#endif /* BCMDBG || BCMDBG_DUMP || BCMDBG_PHYDUMP */
|
|
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
/** change logical "focus" to the gpio core for optimized access */
|
|
volatile void *
|
|
si_gpiosetcore(si_t *sih)
|
|
{
|
|
return (si_setcoreidx(sih, SI_CC_IDX));
|
|
}
|
|
|
|
/**
|
|
* mask & set gpiocontrol bits.
|
|
* If a gpiocontrol bit is set to 0, chipcommon controls the corresponding GPIO pin.
|
|
* If a gpiocontrol bit is set to 1, the GPIO pin is no longer a GPIO and becomes dedicated
|
|
* to some chip-specific purpose.
|
|
*/
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gpiocontrol)(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
|
|
regoff = 0;
|
|
|
|
/* gpios could be shared on router platforms
|
|
* ignore reservation if it's high priority (e.g., test apps)
|
|
*/
|
|
if ((priority != GPIO_HI_PRIORITY) &&
|
|
(BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpiocontrol);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
/** mask&set gpio output enable bits */
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gpioouten)(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
|
|
regoff = 0;
|
|
|
|
/* gpios could be shared on router platforms
|
|
* ignore reservation if it's high priority (e.g., test apps)
|
|
*/
|
|
if ((priority != GPIO_HI_PRIORITY) &&
|
|
(BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpioouten);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
/** mask&set gpio output bits */
|
|
uint32
|
|
BCMPOSTTRAPFN(si_gpioout)(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
|
|
regoff = 0;
|
|
|
|
/* gpios could be shared on router platforms
|
|
* ignore reservation if it's high priority (e.g., test apps)
|
|
*/
|
|
if ((priority != GPIO_HI_PRIORITY) &&
|
|
(BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpioout);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
/** reserve one gpio */
|
|
uint32
|
|
si_gpioreserve(const si_t *sih, uint32 gpio_bitmask, uint8 priority)
|
|
{
|
|
/* only cores on SI_BUS share GPIO's and only applcation users need to
|
|
* reserve/release GPIO
|
|
*/
|
|
if ((BUSTYPE(sih->bustype) != SI_BUS) || (!priority)) {
|
|
ASSERT((BUSTYPE(sih->bustype) == SI_BUS) && (priority));
|
|
return 0xffffffff;
|
|
}
|
|
/* make sure only one bit is set */
|
|
if ((!gpio_bitmask) || ((gpio_bitmask) & (gpio_bitmask - 1))) {
|
|
ASSERT((gpio_bitmask) && !((gpio_bitmask) & (gpio_bitmask - 1)));
|
|
return 0xffffffff;
|
|
}
|
|
|
|
/* already reserved */
|
|
if (si_gpioreservation & gpio_bitmask)
|
|
return 0xffffffff;
|
|
/* set reservation */
|
|
si_gpioreservation |= gpio_bitmask;
|
|
|
|
return si_gpioreservation;
|
|
}
|
|
|
|
/**
|
|
* release one gpio.
|
|
*
|
|
* releasing the gpio doesn't change the current value on the GPIO last write value
|
|
* persists till someone overwrites it.
|
|
*/
|
|
uint32
|
|
si_gpiorelease(const si_t *sih, uint32 gpio_bitmask, uint8 priority)
|
|
{
|
|
/* only cores on SI_BUS share GPIO's and only applcation users need to
|
|
* reserve/release GPIO
|
|
*/
|
|
if ((BUSTYPE(sih->bustype) != SI_BUS) || (!priority)) {
|
|
ASSERT((BUSTYPE(sih->bustype) == SI_BUS) && (priority));
|
|
return 0xffffffff;
|
|
}
|
|
/* make sure only one bit is set */
|
|
if ((!gpio_bitmask) || ((gpio_bitmask) & (gpio_bitmask - 1))) {
|
|
ASSERT((gpio_bitmask) && !((gpio_bitmask) & (gpio_bitmask - 1)));
|
|
return 0xffffffff;
|
|
}
|
|
|
|
/* already released */
|
|
if (!(si_gpioreservation & gpio_bitmask))
|
|
return 0xffffffff;
|
|
|
|
/* clear reservation */
|
|
si_gpioreservation &= ~gpio_bitmask;
|
|
|
|
return si_gpioreservation;
|
|
}
|
|
|
|
/* return the current gpioin register value */
|
|
uint32
|
|
si_gpioin(si_t *sih)
|
|
{
|
|
uint regoff;
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpioin);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, 0, 0));
|
|
}
|
|
|
|
/* mask&set gpio interrupt polarity bits */
|
|
uint32
|
|
si_gpiointpolarity(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
|
|
/* gpios could be shared on router platforms */
|
|
if ((BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpiointpolarity);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
/* mask&set gpio interrupt mask bits */
|
|
uint32
|
|
si_gpiointmask(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
|
|
/* gpios could be shared on router platforms */
|
|
if ((BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
|
|
regoff = OFFSETOF(chipcregs_t, gpiointmask);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
uint32
|
|
si_gpioeventintmask(si_t *sih, uint32 mask, uint32 val, uint8 priority)
|
|
{
|
|
uint regoff;
|
|
/* gpios could be shared on router platforms */
|
|
if ((BUSTYPE(sih->bustype) == SI_BUS) && (val || mask)) {
|
|
mask = priority ? (si_gpioreservation & mask) :
|
|
((si_gpioreservation | mask) & ~(si_gpioreservation));
|
|
val &= mask;
|
|
}
|
|
regoff = OFFSETOF(chipcregs_t, gpioeventintmask);
|
|
return (si_corereg(sih, SI_CC_IDX, regoff, mask, val));
|
|
}
|
|
|
|
uint32
|
|
si_gpiopull(si_t *sih, bool updown, uint32 mask, uint32 val)
|
|
{
|
|
uint offs;
|
|
|
|
if (CCREV(sih->ccrev) < 20)
|
|
return 0xffffffff;
|
|
|
|
offs = (updown ? OFFSETOF(chipcregs_t, gpiopulldown) : OFFSETOF(chipcregs_t, gpiopullup));
|
|
return (si_corereg(sih, SI_CC_IDX, offs, mask, val));
|
|
}
|
|
|
|
uint32
|
|
si_gpioevent(si_t *sih, uint regtype, uint32 mask, uint32 val)
|
|
{
|
|
uint offs;
|
|
|
|
if (CCREV(sih->ccrev) < 11)
|
|
return 0xffffffff;
|
|
|
|
if (regtype == GPIO_REGEVT)
|
|
offs = OFFSETOF(chipcregs_t, gpioevent);
|
|
else if (regtype == GPIO_REGEVT_INTMSK)
|
|
offs = OFFSETOF(chipcregs_t, gpioeventintmask);
|
|
else if (regtype == GPIO_REGEVT_INTPOL)
|
|
offs = OFFSETOF(chipcregs_t, gpioeventintpolarity);
|
|
else
|
|
return 0xffffffff;
|
|
|
|
return (si_corereg(sih, SI_CC_IDX, offs, mask, val));
|
|
}
|
|
|
|
uint32
|
|
si_gpio_int_enable(si_t *sih, bool enable)
|
|
{
|
|
uint offs;
|
|
|
|
if (CCREV(sih->ccrev) < 11)
|
|
return 0xffffffff;
|
|
|
|
offs = OFFSETOF(chipcregs_t, intmask);
|
|
return (si_corereg(sih, SI_CC_IDX, offs, CI_GPIO, (enable ? CI_GPIO : 0)));
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
void
|
|
si_gci_shif_config_wake_pin(si_t *sih, uint8 gpio_n, uint8 wake_events,
|
|
bool gci_gpio)
|
|
{
|
|
uint8 chipcontrol = 0;
|
|
uint32 gci_wakset;
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4376_CHIP_GRPID :
|
|
case BCM4378_CHIP_GRPID :
|
|
{
|
|
if (!gci_gpio) {
|
|
chipcontrol = (1 << GCI_GPIO_CHIPCTRL_ENAB_EXT_GPIO_BIT);
|
|
}
|
|
chipcontrol |= (1 << GCI_GPIO_CHIPCTRL_PULLUP_BIT);
|
|
chipcontrol |= (1 << GCI_GPIO_CHIPCTRL_INVERT_BIT);
|
|
si_gci_gpio_chipcontrol(sih, gpio_n,
|
|
(chipcontrol | (1 << GCI_GPIO_CHIPCTRL_ENAB_IN_BIT)));
|
|
|
|
/* enable gci gpio int/wake events */
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, wake_events);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, wake_events);
|
|
|
|
/* clear the existing status bits */
|
|
si_gci_gpio_status(sih, gpio_n,
|
|
GCI_GPIO_STS_CLEAR, GCI_GPIO_STS_CLEAR);
|
|
|
|
/* Enable gci2wl_wake for 4378 */
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL2,
|
|
CC2_4378_GCI2WAKE_MASK, CC2_4378_GCI2WAKE_MASK);
|
|
|
|
/* enable gci int/wake events */
|
|
gci_wakset = (GCI_INTSTATUS_GPIOWAKE) | (GCI_INTSTATUS_GPIOINT);
|
|
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_intmask),
|
|
gci_wakset, gci_wakset);
|
|
/* Enable wake on GciWake */
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_wakemask),
|
|
gci_wakset, gci_wakset);
|
|
break;
|
|
}
|
|
case BCM4385_CHIP_GRPID :
|
|
case BCM4387_CHIP_GRPID :
|
|
case BCM4389_CHIP_GRPID :
|
|
{
|
|
if (!gci_gpio) {
|
|
chipcontrol = (1 << GCI_GPIO_CHIPCTRL_ENAB_EXT_GPIO_BIT);
|
|
}
|
|
chipcontrol |= (1 << GCI_GPIO_CHIPCTRL_PULLUP_BIT);
|
|
chipcontrol |= (1 << GCI_GPIO_CHIPCTRL_INVERT_BIT);
|
|
si_gci_gpio_chipcontrol(sih, gpio_n,
|
|
(chipcontrol | (1 << GCI_GPIO_CHIPCTRL_ENAB_IN_BIT)));
|
|
|
|
/* enable gci gpio int/wake events */
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, wake_events);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, wake_events);
|
|
|
|
/* clear the existing status bits */
|
|
si_gci_gpio_status(sih, gpio_n,
|
|
GCI_GPIO_STS_CLEAR, GCI_GPIO_STS_CLEAR);
|
|
|
|
/* Enable gci2wl_wake for 4387 */
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL2,
|
|
CC2_4387_GCI2WAKE_MASK, CC2_4387_GCI2WAKE_MASK);
|
|
|
|
/* enable gci int/wake events */
|
|
gci_wakset = (GCI_INTSTATUS_GPIOWAKE) | (GCI_INTSTATUS_GPIOINT);
|
|
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_intmask),
|
|
gci_wakset, gci_wakset);
|
|
/* Enable wake on GciWake */
|
|
si_gci_indirect(sih, 0, GCI_OFFSETOF(sih, gci_wakemask),
|
|
gci_wakset, gci_wakset);
|
|
break;
|
|
}
|
|
default:;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_shif_int_enable(si_t *sih, uint8 gpio_n, uint8 wake_events, bool enable)
|
|
{
|
|
if (enable) {
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, wake_events);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, wake_events);
|
|
} else {
|
|
si_gci_gpio_intmask(sih, gpio_n, wake_events, 0);
|
|
si_gci_gpio_wakemask(sih, gpio_n, wake_events, 0);
|
|
}
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#define SI_BANKINFO_DELAY_USEC 10u
|
|
|
|
/** Return the size of the specified SYSMEM bank */
|
|
static uint
|
|
sysmem_banksize(const si_info_t *sii, sysmemregs_t *regs, uint8 idx)
|
|
{
|
|
uint banksize, bankinfo;
|
|
uint bankidx = idx;
|
|
|
|
W_REG(sii->osh, ®s->bankidx, bankidx);
|
|
/* Add some delay before reading back bankinfo */
|
|
OSL_DELAY(SI_BANKINFO_DELAY_USEC);
|
|
bankinfo = R_REG(sii->osh, ®s->bankinfo);
|
|
banksize = SYSMEM_BANKINFO_SZBASE * ((bankinfo & SYSMEM_BANKINFO_SZMASK) + 1);
|
|
|
|
SI_PRINT(("%s: bankidx:%d bankinfo:0x%x banksize:%d(0x%x)\n",
|
|
__FUNCTION__, bankidx, bankinfo, banksize, banksize));
|
|
|
|
return banksize;
|
|
}
|
|
|
|
#define SI_CORE_RESET_DELAY_USEC 10u
|
|
#define SI_CORE_RESET_RETRY 5u
|
|
|
|
/** Return the RAM size of the SYSMEM core */
|
|
uint32
|
|
si_sysmem_size(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
sysmemregs_t *regs;
|
|
bool wasup;
|
|
uint32 coreinfo;
|
|
uint memsize = 0;
|
|
uint8 i;
|
|
uint nb, nrb;
|
|
uint32 retry = SI_CORE_RESET_RETRY;
|
|
|
|
SI_MSG_DBG_REG(("%s: Enter\n", __FUNCTION__));
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Switch to SYSMEM core */
|
|
if (!(regs = si_setcore(sih, SYSMEM_CORE_ID, 0))) {
|
|
SI_ERROR(("%s: si_setcore SYSMEM_CORE_ID failed\n", __FUNCTION__));
|
|
goto done;
|
|
}
|
|
|
|
while (retry--) {
|
|
/* retry if core is not up */
|
|
if (!(wasup = si_iscoreup(sih))) {
|
|
SI_PRINT(("%s: core not up, do si_core_reset, retry:%d\n",
|
|
__FUNCTION__, retry));
|
|
si_core_reset(sih, 0, 0);
|
|
/* Add some delay after reset */
|
|
OSL_DELAY(SI_CORE_RESET_DELAY_USEC);
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (retry == 0) {
|
|
SI_ERROR(("%s: failed\n", __FUNCTION__));
|
|
goto restore_core;
|
|
}
|
|
|
|
coreinfo = R_REG(sii->osh, ®s->coreinfo);
|
|
if (coreinfo == (uint32)-1) {
|
|
SI_ERROR(("%s: invalid coreinfo:0x%x\n", __FUNCTION__, coreinfo));
|
|
goto restore_core;
|
|
}
|
|
|
|
/* Number of ROM banks, SW need to skip the ROM banks. */
|
|
if (si_corerev(sih) < 12) {
|
|
nrb = (coreinfo & SYSMEM_SRCI_ROMNB_MASK) >> SYSMEM_SRCI_ROMNB_SHIFT;
|
|
nb = (coreinfo & SYSMEM_SRCI_SRNB_MASK) >> SYSMEM_SRCI_SRNB_SHIFT;
|
|
} else {
|
|
nrb = (coreinfo & SYSMEM_SRCI_NEW_ROMNB_MASK) >> SYSMEM_SRCI_NEW_ROMNB_SHIFT;
|
|
nb = (coreinfo & SYSMEM_SRCI_NEW_SRNB_MASK) >> SYSMEM_SRCI_NEW_SRNB_SHIFT;
|
|
}
|
|
|
|
SI_PRINT(("%s: coreinfo:0x%x num_rom_banks:%d num_ram_baks:%d\n",
|
|
__FUNCTION__, coreinfo, nrb, nb));
|
|
|
|
for (i = 0; i < nb; i++)
|
|
memsize += sysmem_banksize(sii, regs, i + nrb);
|
|
|
|
restore_core:
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
|
|
SI_MSG_DBG_REG(("%s: Exit memsize=%d\n", __FUNCTION__, memsize));
|
|
return memsize;
|
|
}
|
|
|
|
/** Return the size of the specified SOCRAM bank */
|
|
static uint
|
|
socram_banksize(const si_info_t *sii, sbsocramregs_t *regs, uint8 idx, uint8 mem_type)
|
|
{
|
|
uint banksize, bankinfo;
|
|
uint bankidx = idx | (mem_type << SOCRAM_BANKIDX_MEMTYPE_SHIFT);
|
|
|
|
ASSERT(mem_type <= SOCRAM_MEMTYPE_DEVRAM);
|
|
|
|
W_REG(sii->osh, ®s->bankidx, bankidx);
|
|
bankinfo = R_REG(sii->osh, ®s->bankinfo);
|
|
banksize = SOCRAM_BANKINFO_SZBASE * ((bankinfo & SOCRAM_BANKINFO_SZMASK) + 1);
|
|
return banksize;
|
|
}
|
|
|
|
void si_socram_set_bankpda(si_t *sih, uint32 bankidx, uint32 bankpda)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
sbsocramregs_t *regs;
|
|
bool wasup;
|
|
uint corerev;
|
|
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Switch to SOCRAM core */
|
|
if (!(regs = si_setcore(sih, SOCRAM_CORE_ID, 0)))
|
|
goto done;
|
|
|
|
if (!(wasup = si_iscoreup(sih)))
|
|
si_core_reset(sih, 0, 0);
|
|
|
|
corerev = si_corerev(sih);
|
|
if (corerev >= 16) {
|
|
W_REG(sii->osh, ®s->bankidx, bankidx);
|
|
W_REG(sii->osh, ®s->bankpda, bankpda);
|
|
}
|
|
|
|
/* Return to previous state and core */
|
|
if (!wasup)
|
|
si_core_disable(sih, 0);
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
/** Return the RAM size of the SOCRAM core */
|
|
uint32
|
|
si_socram_size(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
sbsocramregs_t *regs;
|
|
bool wasup;
|
|
uint corerev;
|
|
uint32 coreinfo;
|
|
uint memsize = 0;
|
|
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Switch to SOCRAM core */
|
|
if (!(regs = si_setcore(sih, SOCRAM_CORE_ID, 0)))
|
|
goto done;
|
|
|
|
/* Get info for determining size */
|
|
if (!(wasup = si_iscoreup(sih)))
|
|
si_core_reset(sih, 0, 0);
|
|
corerev = si_corerev(sih);
|
|
coreinfo = R_REG(sii->osh, ®s->coreinfo);
|
|
|
|
/* Calculate size from coreinfo based on rev */
|
|
if (corerev == 0)
|
|
memsize = 1 << (16 + (coreinfo & SRCI_MS0_MASK));
|
|
else if (corerev < 3) {
|
|
memsize = 1 << (SR_BSZ_BASE + (coreinfo & SRCI_SRBSZ_MASK));
|
|
memsize *= (coreinfo & SRCI_SRNB_MASK) >> SRCI_SRNB_SHIFT;
|
|
} else if ((corerev <= 7) || (corerev == 12)) {
|
|
uint nb = (coreinfo & SRCI_SRNB_MASK) >> SRCI_SRNB_SHIFT;
|
|
uint bsz = (coreinfo & SRCI_SRBSZ_MASK);
|
|
uint lss = (coreinfo & SRCI_LSS_MASK) >> SRCI_LSS_SHIFT;
|
|
if (lss != 0)
|
|
nb --;
|
|
memsize = nb * (1 << (bsz + SR_BSZ_BASE));
|
|
if (lss != 0)
|
|
memsize += (1 << ((lss - 1) + SR_BSZ_BASE));
|
|
} else {
|
|
uint8 i;
|
|
uint nb;
|
|
/* length of SRAM Banks increased for corerev greater than 23 */
|
|
if (corerev >= 23) {
|
|
nb = (coreinfo & (SRCI_SRNB_MASK | SRCI_SRNB_MASK_EXT)) >> SRCI_SRNB_SHIFT;
|
|
} else {
|
|
nb = (coreinfo & SRCI_SRNB_MASK) >> SRCI_SRNB_SHIFT;
|
|
}
|
|
for (i = 0; i < nb; i++)
|
|
memsize += socram_banksize(sii, regs, i, SOCRAM_MEMTYPE_RAM);
|
|
}
|
|
|
|
/* Return to previous state and core */
|
|
if (!wasup)
|
|
si_core_disable(sih, 0);
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
|
|
return memsize;
|
|
}
|
|
|
|
/* Return true if bus MPU is present */
|
|
bool
|
|
si_is_bus_mpu_present(si_t *sih)
|
|
{
|
|
uint origidx, newidx = NODEV_CORE_ID;
|
|
sysmemregs_t *sysmemregs = NULL;
|
|
cr4regs_t *cr4regs;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint ret = 0;
|
|
bool wasup;
|
|
|
|
origidx = si_coreidx(sih);
|
|
|
|
cr4regs = si_setcore(sih, ARMCR4_CORE_ID, 0);
|
|
if (cr4regs) {
|
|
/* ARMCR4 */
|
|
newidx = ARMCR4_CORE_ID;
|
|
} else {
|
|
sysmemregs = si_setcore(sih, SYSMEM_CORE_ID, 0);
|
|
if (sysmemregs) {
|
|
/* ARMCA7 */
|
|
newidx = SYSMEM_CORE_ID;
|
|
}
|
|
}
|
|
|
|
if (newidx != NODEV_CORE_ID) {
|
|
if (!(wasup = si_iscoreup(sih))) {
|
|
si_core_reset(sih, 0, 0);
|
|
}
|
|
if (newidx == ARMCR4_CORE_ID) {
|
|
/* ARMCR4 */
|
|
ret = R_REG(sii->osh, &cr4regs->corecapabilities) & CAP_MPU_MASK;
|
|
} else {
|
|
/* ARMCA7 */
|
|
ret = R_REG(sii->osh, &sysmemregs->mpucapabilities) &
|
|
ACC_MPU_REGION_CNT_MASK;
|
|
}
|
|
if (!wasup) {
|
|
si_core_disable(sih, 0);
|
|
}
|
|
}
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
return ret ? TRUE : FALSE;
|
|
}
|
|
|
|
#if defined(BCMDONGLEHOST)
|
|
|
|
/** Return the TCM-RAM size of the ARMCR4 core. */
|
|
uint32
|
|
si_tcm_size(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
volatile uint8 *regs;
|
|
bool wasup;
|
|
uint32 corecap;
|
|
uint memsize = 0;
|
|
uint banku_size = 0;
|
|
uint32 nab = 0;
|
|
uint32 nbb = 0;
|
|
uint32 totb = 0;
|
|
uint32 bxinfo = 0;
|
|
uint32 idx = 0;
|
|
volatile uint32 *arm_cap_reg;
|
|
volatile uint32 *arm_bidx;
|
|
volatile uint32 *arm_binfo;
|
|
bool ret = TRUE;
|
|
|
|
SI_MSG_DBG_REG(("%s: Enter\n", __FUNCTION__));
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Switch to CR4 core */
|
|
if (!(regs = si_setcore(sih, ARMCR4_CORE_ID, 0)))
|
|
goto done;
|
|
|
|
/* Get info for determining size. If in reset, come out of reset,
|
|
* but remain in halt
|
|
*/
|
|
if (!(wasup = si_iscoreup(sih))) {
|
|
ret = si_core_reset(sih, SICF_CPUHALT, SICF_CPUHALT);
|
|
if (!ret) {
|
|
goto done;
|
|
}
|
|
} else
|
|
SI_ERROR(("si_iscoreup!!\n"));
|
|
|
|
arm_cap_reg = (volatile uint32 *)(regs + SI_CR4_CAP);
|
|
corecap = R_REG(sii->osh, arm_cap_reg);
|
|
|
|
nab = (corecap & ARMCR4_TCBANB_MASK) >> ARMCR4_TCBANB_SHIFT;
|
|
nbb = (corecap & ARMCR4_TCBBNB_MASK) >> ARMCR4_TCBBNB_SHIFT;
|
|
totb = nab + nbb;
|
|
|
|
arm_bidx = (volatile uint32 *)(regs + SI_CR4_BANKIDX);
|
|
arm_binfo = (volatile uint32 *)(regs + SI_CR4_BANKINFO);
|
|
for (idx = 0; idx < totb; idx++) {
|
|
W_REG(sii->osh, arm_bidx, idx);
|
|
|
|
bxinfo = R_REG(sii->osh, arm_binfo);
|
|
if (bxinfo & ARMCR4_BUNITSZ_MASK) {
|
|
banku_size = ARMCR4_BSZ_1K;
|
|
} else {
|
|
banku_size = ARMCR4_BSZ_8K;
|
|
}
|
|
memsize += ((bxinfo & ARMCR4_BSZ_MASK) + 1) * banku_size;
|
|
}
|
|
|
|
/* Return to previous state and core */
|
|
if (!wasup)
|
|
si_core_disable(sih, 0);
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
SI_MSG_DBG_REG(("%s: Exit memsize=%d\n", __FUNCTION__, memsize));
|
|
return memsize;
|
|
}
|
|
|
|
bool
|
|
si_has_flops(si_t *sih)
|
|
{
|
|
uint origidx, cr4_rev;
|
|
|
|
/* Find out CR4 core revision */
|
|
origidx = si_coreidx(sih);
|
|
if (si_setcore(sih, ARMCR4_CORE_ID, 0)) {
|
|
cr4_rev = si_corerev(sih);
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
if (cr4_rev == 1 || cr4_rev >= 3)
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
#endif /* BCMDONGLEHOST */
|
|
|
|
uint32
|
|
si_socram_srmem_size(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
sbsocramregs_t *regs;
|
|
bool wasup;
|
|
uint corerev;
|
|
uint32 coreinfo;
|
|
uint memsize = 0;
|
|
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Switch to SOCRAM core */
|
|
if (!(regs = si_setcore(sih, SOCRAM_CORE_ID, 0)))
|
|
goto done;
|
|
|
|
/* Get info for determining size */
|
|
if (!(wasup = si_iscoreup(sih)))
|
|
si_core_reset(sih, 0, 0);
|
|
corerev = si_corerev(sih);
|
|
coreinfo = R_REG(sii->osh, ®s->coreinfo);
|
|
|
|
/* Calculate size from coreinfo based on rev */
|
|
if (corerev >= 16) {
|
|
uint8 i;
|
|
uint nb = (coreinfo & SRCI_SRNB_MASK) >> SRCI_SRNB_SHIFT;
|
|
for (i = 0; i < nb; i++) {
|
|
W_REG(sii->osh, ®s->bankidx, i);
|
|
if (R_REG(sii->osh, ®s->bankinfo) & SOCRAM_BANKINFO_RETNTRAM_MASK)
|
|
memsize += socram_banksize(sii, regs, i, SOCRAM_MEMTYPE_RAM);
|
|
}
|
|
}
|
|
|
|
/* Return to previous state and core */
|
|
if (!wasup)
|
|
si_core_disable(sih, 0);
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
done:
|
|
INTR_RESTORE(sii, &intr_val);
|
|
|
|
return memsize;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
static bool
|
|
BCMPOSTTRAPFN(si_seci_uart)(const si_t *sih)
|
|
{
|
|
return (sih->cccaps_ext & CC_CAP_EXT_SECI_PUART_PRESENT);
|
|
}
|
|
|
|
/** seci clock enable/disable */
|
|
static void
|
|
BCMPOSTTRAPFN(si_seci_clkreq)(si_t *sih, bool enable)
|
|
{
|
|
uint32 clk_ctl_st;
|
|
uint32 offset;
|
|
uint32 val;
|
|
pmuregs_t *pmu;
|
|
uint32 origidx = 0;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
#ifdef SECI_UART
|
|
bool fast;
|
|
chipcregs_t *cc = seci_set_core(sih, &origidx, &fast);
|
|
ASSERT(cc);
|
|
#endif /* SECI_UART */
|
|
if (!si_seci(sih) && !si_seci_uart(sih))
|
|
return;
|
|
offset = OFFSETOF(chipcregs_t, clk_ctl_st);
|
|
clk_ctl_st = si_corereg(sih, 0, offset, 0, 0);
|
|
|
|
if (enable && !(clk_ctl_st & CLKCTL_STS_SECI_CLK_REQ)) {
|
|
val = CLKCTL_STS_SECI_CLK_REQ | CLKCTL_STS_HT_AVAIL_REQ;
|
|
#ifdef SECI_UART
|
|
/* Restore the fast UART function select when enabling */
|
|
if (fast_uart_init) {
|
|
si_gci_set_functionsel(sih, fast_uart_tx, fast_uart_functionsel);
|
|
if (fuart_pullup_rx_cts_enab) {
|
|
si_gci_set_functionsel(sih, fast_uart_rx, fast_uart_functionsel);
|
|
si_gci_set_functionsel(sih, fast_uart_cts_in,
|
|
fast_uart_functionsel);
|
|
}
|
|
}
|
|
#endif /* SECI_UART */
|
|
} else if (!enable && (clk_ctl_st & CLKCTL_STS_SECI_CLK_REQ)) {
|
|
val = 0;
|
|
#ifdef SECI_UART
|
|
if (force_seci_clk) {
|
|
return;
|
|
}
|
|
#endif /* SECI_UART */
|
|
} else {
|
|
return;
|
|
}
|
|
#ifdef SECI_UART
|
|
/* park the fast UART as PULL UP when disabling the clocks to avoid sending
|
|
* breaks to the host
|
|
*/
|
|
if (!enable && fast_uart_init) {
|
|
si_gci_set_functionsel(sih, fast_uart_tx, fast_uart_pup);
|
|
if (fuart_pullup_rx_cts_enab) {
|
|
W_REG(sii->osh, &cc->SECI_status, SECI_STAT_BI);
|
|
si_gci_set_functionsel(sih, fast_uart_rx, fast_uart_pup);
|
|
si_gci_set_functionsel(sih, fast_uart_cts_in, fast_uart_pup);
|
|
SPINWAIT(!(R_REG(sii->osh, &cc->SECI_status) & SECI_STAT_BI), 1000);
|
|
}
|
|
}
|
|
#endif /* SECI_UART */
|
|
|
|
/* Setting/clearing bit 4 along with bit 8 of 0x1e0 block. the core requests that
|
|
* the PMU set the device state such that the HT clock will be available on short notice.
|
|
*/
|
|
si_corereg(sih, SI_CC_IDX, offset,
|
|
CLKCTL_STS_SECI_CLK_REQ | CLKCTL_STS_HT_AVAIL_REQ, val);
|
|
|
|
if (!enable)
|
|
return;
|
|
#ifndef SECI_UART
|
|
/* Remember original core before switch to chipc/pmu */
|
|
origidx = si_coreidx(sih);
|
|
#endif
|
|
|
|
if (AOB_ENAB(sih)) {
|
|
pmu = si_setcore(sih, PMU_CORE_ID, 0);
|
|
} else {
|
|
pmu = si_setcoreidx(sih, SI_CC_IDX);
|
|
}
|
|
ASSERT(pmu != NULL);
|
|
(void)si_pmu_wait_for_steady_state(sih, sii->osh, pmu);
|
|
/* Return to original core */
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
SPINWAIT(!(si_corereg(sih, 0, offset, 0, 0) & CLKCTL_STS_SECI_CLK_AVAIL),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
|
|
clk_ctl_st = si_corereg(sih, 0, offset, 0, 0);
|
|
if (enable) {
|
|
if (!(clk_ctl_st & CLKCTL_STS_SECI_CLK_AVAIL)) {
|
|
SI_ERROR(("SECI clock is not available\n"));
|
|
ASSERT(0);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
#if defined(BCMECICOEX) || defined(SECI_UART)
|
|
static chipcregs_t *
|
|
BCMPOSTTRAPFN(seci_set_core)(si_t *sih, uint32 *origidx, bool *fast)
|
|
{
|
|
chipcregs_t *cc;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
*fast = SI_FAST(sii);
|
|
|
|
if (!*fast) {
|
|
*origidx = sii->curidx;
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
} else {
|
|
*origidx = 0;
|
|
cc = (chipcregs_t *)CCREGS_FAST(sii);
|
|
}
|
|
return cc;
|
|
}
|
|
|
|
static chipcregs_t *
|
|
BCMPOSTTRAPFN(si_seci_access_preamble)(si_t *sih, const si_info_t *sii, uint32 *origidx, bool *fast)
|
|
{
|
|
chipcregs_t *cc = seci_set_core(sih, origidx, fast);
|
|
|
|
if (cc) {
|
|
if (((R_REG(sii->osh, &cc->clk_ctl_st) & CCS_SECICLKREQ) != CCS_SECICLKREQ)) {
|
|
/* enable SECI clock */
|
|
si_seci_clkreq(sih, TRUE);
|
|
}
|
|
}
|
|
return cc;
|
|
}
|
|
#endif /* BCMECICOEX||SECI_UART */
|
|
#ifdef SECI_UART
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_seci_access)(si_t *sih, uint32 val, int access)
|
|
{
|
|
uint32 origidx;
|
|
bool fast;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
bcm_int_bitmask_t intr_val;
|
|
uint32 offset, retval = 1;
|
|
|
|
if (!si_seci_uart(sih))
|
|
return 0;
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
if (!(cc = si_seci_access_preamble(sih, sii, &origidx, &fast)))
|
|
goto exit;
|
|
|
|
switch (access) {
|
|
case SECI_ACCESS_STATUSMASK_SET:
|
|
offset = OFFSETOF(chipcregs_t, SECI_statusmask);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, ALLONES_32, val);
|
|
break;
|
|
case SECI_ACCESS_STATUSMASK_GET:
|
|
offset = OFFSETOF(chipcregs_t, SECI_statusmask);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, 0, 0);
|
|
break;
|
|
case SECI_ACCESS_INTRS:
|
|
offset = OFFSETOF(chipcregs_t, SECI_status);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset,
|
|
ALLONES_32, ALLONES_32);
|
|
break;
|
|
case SECI_ACCESS_UART_CTS:
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_msr);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, 0, 0);
|
|
retval = retval & SECI_UART_MSR_CTS_STATE;
|
|
break;
|
|
case SECI_ACCESS_UART_RTS:
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_mcr);
|
|
if (val) {
|
|
/* clear forced flow control; enable auto rts */
|
|
retval = si_corereg(sih, SI_CC_IDX, offset,
|
|
SECI_UART_MCR_PRTS | SECI_UART_MCR_AUTO_RTS,
|
|
SECI_UART_MCR_AUTO_RTS);
|
|
} else {
|
|
/* set forced flow control; clear auto rts */
|
|
retval = si_corereg(sih, SI_CC_IDX, offset,
|
|
SECI_UART_MCR_PRTS | SECI_UART_MCR_AUTO_RTS,
|
|
SECI_UART_MCR_PRTS);
|
|
}
|
|
break;
|
|
case SECI_ACCESS_UART_RXEMPTY:
|
|
offset = OFFSETOF(chipcregs_t, SECI_status);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, 0, 0);
|
|
retval = (retval & SECI_STAT_SRFE) == SECI_STAT_SRFE;
|
|
break;
|
|
case SECI_ACCESS_UART_GETC:
|
|
/* assumes caller checked for nonempty rx FIFO */
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_data);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, 0, 0) & 0xff;
|
|
break;
|
|
case SECI_ACCESS_UART_TXFULL:
|
|
offset = OFFSETOF(chipcregs_t, SECI_status);
|
|
retval = si_corereg(sih, SI_CC_IDX, offset, 0, 0);
|
|
retval = (retval & SECI_STAT_STFF) == SECI_STAT_STFF;
|
|
break;
|
|
case SECI_ACCESS_UART_PUTC:
|
|
/* This register must not do a RMW otherwise it will affect the RX FIFO */
|
|
W_REG(sii->osh, &cc->seci_uart_data, (uint32)(val & 0xff));
|
|
retval = 0;
|
|
break;
|
|
default:
|
|
ASSERT(0);
|
|
}
|
|
|
|
exit:
|
|
/* restore previous core */
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
INTR_RESTORE(sii, &intr_val);
|
|
|
|
return retval;
|
|
}
|
|
|
|
void si_seci_clk_force(si_t *sih, bool val)
|
|
{
|
|
force_seci_clk = val;
|
|
if (force_seci_clk) {
|
|
si_seci_clkreq(sih, TRUE);
|
|
} else {
|
|
si_seci_down(sih);
|
|
}
|
|
}
|
|
|
|
bool si_seci_clk_force_status(si_t *sih)
|
|
{
|
|
return force_seci_clk;
|
|
}
|
|
#endif /* SECI_UART */
|
|
|
|
/** SECI Init routine, pass in seci_mode */
|
|
volatile void *
|
|
si_seci_init(si_t *sih, uint8 seci_mode)
|
|
{
|
|
uint32 origidx = 0;
|
|
uint32 offset;
|
|
const si_info_t *sii;
|
|
volatile void *ptr;
|
|
chipcregs_t *cc;
|
|
bool fast;
|
|
uint32 seci_conf;
|
|
|
|
if (sih->ccrev < 35)
|
|
return NULL;
|
|
|
|
#ifdef SECI_UART
|
|
if (seci_mode == SECI_MODE_UART) {
|
|
if (!si_seci_uart(sih))
|
|
return NULL;
|
|
}
|
|
else {
|
|
#endif /* SECI_UART */
|
|
if (!si_seci(sih))
|
|
return NULL;
|
|
#ifdef SECI_UART
|
|
}
|
|
#endif /* SECI_UART */
|
|
|
|
if (seci_mode > SECI_MODE_MASK)
|
|
return NULL;
|
|
|
|
sii = SI_INFO(sih);
|
|
fast = SI_FAST(sii);
|
|
if (!fast) {
|
|
origidx = sii->curidx;
|
|
if ((ptr = si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
return NULL;
|
|
} else if ((ptr = CCREGS_FAST(sii)) == NULL)
|
|
return NULL;
|
|
cc = (chipcregs_t *)ptr;
|
|
ASSERT(cc);
|
|
|
|
/* enable SECI clock */
|
|
if (seci_mode != SECI_MODE_LEGACY_3WIRE_WLAN)
|
|
si_seci_clkreq(sih, TRUE);
|
|
|
|
/* put the SECI in reset */
|
|
seci_conf = R_REG(sii->osh, &cc->SECI_config);
|
|
seci_conf &= ~SECI_ENAB_SECI_ECI;
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
seci_conf = SECI_RESET;
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
|
|
/* set force-low, and set EN_SECI for all non-legacy modes */
|
|
seci_conf |= SECI_ENAB_SECIOUT_DIS;
|
|
if ((seci_mode == SECI_MODE_UART) || (seci_mode == SECI_MODE_SECI) ||
|
|
(seci_mode == SECI_MODE_HALF_SECI))
|
|
{
|
|
seci_conf |= SECI_ENAB_SECI_ECI;
|
|
}
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
|
|
if (seci_mode != SECI_MODE_LEGACY_3WIRE_WLAN) {
|
|
/* take seci out of reset */
|
|
seci_conf = R_REG(sii->osh, &cc->SECI_config);
|
|
seci_conf &= ~(SECI_RESET);
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
}
|
|
/* set UART/SECI baud rate */
|
|
/* hard-coded at 4MBaud for now */
|
|
if ((seci_mode == SECI_MODE_UART) || (seci_mode == SECI_MODE_SECI) ||
|
|
(seci_mode == SECI_MODE_HALF_SECI)) {
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_bauddiv);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFF, 0xFF); /* 4MBaud */
|
|
if ((CHIPID(sih->chip) == BCM4360_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43460_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43526_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM4352_CHIP_ID)) {
|
|
/* MAC clk is 160MHz */
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_bauddiv);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFF, 0xFE);
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_baudadj);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFF, 0x44);
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_mcr);
|
|
si_corereg(sih, SI_CC_IDX, offset,
|
|
0xFF, SECI_UART_MCR_BAUD_ADJ_EN); /* 0x81 */
|
|
}
|
|
#ifdef SECI_UART
|
|
else if (CCREV(sih->ccrev) >= 62) {
|
|
/* rx FIFO level at which an interrupt is generated */
|
|
offset = OFFSETOF(chipcregs_t, eci.ge35.eci_uartfifolevel);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xff, 0x01);
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_mcr);
|
|
si_corereg(sih, SI_CC_IDX, offset, SECI_UART_MCR_AUTO_RTS,
|
|
SECI_UART_MCR_AUTO_RTS);
|
|
}
|
|
#endif /* SECI_UART */
|
|
else {
|
|
/* 4336 MAC clk is 80MHz */
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_baudadj);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFF, 0x22);
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_mcr);
|
|
si_corereg(sih, SI_CC_IDX, offset,
|
|
0xFF, SECI_UART_MCR_BAUD_ADJ_EN); /* 0x80 */
|
|
}
|
|
|
|
/* LCR/MCR settings */
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_lcr);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFF,
|
|
(SECI_UART_LCR_RX_EN | SECI_UART_LCR_TXO_EN)); /* 0x28 */
|
|
offset = OFFSETOF(chipcregs_t, seci_uart_mcr);
|
|
si_corereg(sih, SI_CC_IDX, offset,
|
|
SECI_UART_MCR_TX_EN, SECI_UART_MCR_TX_EN); /* 0x01 */
|
|
|
|
#ifndef SECI_UART
|
|
/* Give control of ECI output regs to MAC core */
|
|
offset = OFFSETOF(chipcregs_t, eci.ge35.eci_controllo);
|
|
si_corereg(sih, SI_CC_IDX, offset, ALLONES_32, ECI_MACCTRLLO_BITS);
|
|
offset = OFFSETOF(chipcregs_t, eci.ge35.eci_controlhi);
|
|
si_corereg(sih, SI_CC_IDX, offset, 0xFFFF, ECI_MACCTRLHI_BITS);
|
|
#endif /* SECI_UART */
|
|
}
|
|
|
|
/* set the seci mode in seci conf register */
|
|
seci_conf = R_REG(sii->osh, &cc->SECI_config);
|
|
seci_conf &= ~(SECI_MODE_MASK << SECI_MODE_SHIFT);
|
|
seci_conf |= (seci_mode << SECI_MODE_SHIFT);
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
|
|
/* Clear force-low bit */
|
|
seci_conf = R_REG(sii->osh, &cc->SECI_config);
|
|
seci_conf &= ~SECI_ENAB_SECIOUT_DIS;
|
|
W_REG(sii->osh, &cc->SECI_config, seci_conf);
|
|
|
|
/* restore previous core */
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
return ptr;
|
|
}
|
|
|
|
#ifdef BCMECICOEX
|
|
#define NOTIFY_BT_FM_DISABLE(sih, val) \
|
|
si_eci_notify_bt((sih), ECI_OUT_FM_DISABLE_MASK(CCREV(sih->ccrev)), \
|
|
((val) << ECI_OUT_FM_DISABLE_SHIFT(CCREV(sih->ccrev))), FALSE)
|
|
|
|
/** Query OTP to see if FM is disabled */
|
|
static int
|
|
si_query_FMDisabled_from_OTP(si_t *sih, uint16 *FMDisabled)
|
|
{
|
|
int error = BCME_OK;
|
|
uint bitoff = 0;
|
|
bool wasup;
|
|
void *oh;
|
|
uint32 min_res_mask = 0;
|
|
|
|
/* If there is a bit for this chip, check it */
|
|
if (bitoff) {
|
|
if (!(wasup = si_is_otp_powered(sih))) {
|
|
si_otp_power(sih, TRUE, &min_res_mask);
|
|
}
|
|
|
|
if ((oh = otp_init(sih)) != NULL)
|
|
*FMDisabled = !otp_read_bit(oh, OTP4325_FM_DISABLED_OFFSET);
|
|
else
|
|
error = BCME_NOTFOUND;
|
|
|
|
if (!wasup) {
|
|
si_otp_power(sih, FALSE, &min_res_mask);
|
|
}
|
|
}
|
|
|
|
return error;
|
|
}
|
|
|
|
bool
|
|
si_eci(const si_t *sih)
|
|
{
|
|
return (!!(sih->cccaps & CC_CAP_ECI));
|
|
}
|
|
|
|
bool
|
|
BCMPOSTTRAPFN(si_seci)(const si_t *sih)
|
|
{
|
|
return (sih->cccaps_ext & CC_CAP_EXT_SECI_PRESENT);
|
|
}
|
|
|
|
bool
|
|
si_gci(const si_t *sih)
|
|
{
|
|
return (sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT);
|
|
}
|
|
|
|
bool
|
|
si_sraon(const si_t *sih)
|
|
{
|
|
return (sih->cccaps_ext & CC_CAP_SR_AON_PRESENT);
|
|
}
|
|
|
|
/** ECI Init routine */
|
|
int
|
|
si_eci_init(si_t *sih)
|
|
{
|
|
uint32 origidx = 0;
|
|
const si_info_t *sii;
|
|
chipcregs_t *cc;
|
|
bool fast;
|
|
uint16 FMDisabled = FALSE;
|
|
|
|
/* check for ECI capability */
|
|
if (!(sih->cccaps & CC_CAP_ECI))
|
|
return BCME_ERROR;
|
|
|
|
sii = SI_INFO(sih);
|
|
fast = SI_FAST(sii);
|
|
if (!fast) {
|
|
origidx = sii->curidx;
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
return BCME_ERROR;
|
|
} else if ((cc = (chipcregs_t *)CCREGS_FAST(sii)) == NULL)
|
|
return BCME_ERROR;
|
|
ASSERT(cc);
|
|
|
|
/* disable level based interrupts */
|
|
if (CCREV(sih->ccrev) < 35) {
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_intmaskhi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_intmaskmi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_intmasklo, 0x0);
|
|
} else {
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_intmaskhi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_intmasklo, 0x0);
|
|
}
|
|
|
|
/* Assign eci_output bits between 'wl' and dot11mac */
|
|
if (CCREV(sih->ccrev) < 35) {
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_control, ECI_MACCTRL_BITS);
|
|
} else {
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_controllo, ECI_MACCTRLLO_BITS);
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_controlhi, ECI_MACCTRLHI_BITS);
|
|
}
|
|
|
|
/* enable only edge based interrupts
|
|
* only toggle on bit 62 triggers an interrupt
|
|
*/
|
|
if (CCREV(sih->ccrev) < 35) {
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_eventmaskhi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_eventmaskmi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.lt35.eci_eventmasklo, 0x0);
|
|
} else {
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_eventmaskhi, 0x0);
|
|
W_REG(sii->osh, &cc->eci.ge35.eci_eventmasklo, 0x0);
|
|
}
|
|
|
|
/* restore previous core */
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
/* if FM disabled in OTP, let BT know */
|
|
if (!si_query_FMDisabled_from_OTP(sih, &FMDisabled)) {
|
|
if (FMDisabled) {
|
|
NOTIFY_BT_FM_DISABLE(sih, 1);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/** Write values to BT on eci_output. */
|
|
void
|
|
si_eci_notify_bt(si_t *sih, uint32 mask, uint32 val, bool is_interrupt)
|
|
{
|
|
uint32 offset;
|
|
|
|
if ((sih->cccaps & CC_CAP_ECI) ||
|
|
(si_seci(sih)))
|
|
{
|
|
/* ECI or SECI mode */
|
|
/* Clear interrupt bit by default */
|
|
if (is_interrupt) {
|
|
si_corereg(sih, SI_CC_IDX,
|
|
(CCREV(sih->ccrev) < 35 ?
|
|
OFFSETOF(chipcregs_t, eci.lt35.eci_output) :
|
|
OFFSETOF(chipcregs_t, eci.ge35.eci_outputlo)),
|
|
(1 << 30), 0);
|
|
}
|
|
|
|
if (CCREV(sih->ccrev) >= 35) {
|
|
if ((mask & 0xFFFF0000) == ECI48_OUT_MASKMAGIC_HIWORD) {
|
|
offset = OFFSETOF(chipcregs_t, eci.ge35.eci_outputhi);
|
|
mask = mask & ~0xFFFF0000;
|
|
} else {
|
|
offset = OFFSETOF(chipcregs_t, eci.ge35.eci_outputlo);
|
|
mask = mask | (1<<30);
|
|
val = val & ~(1 << 30);
|
|
}
|
|
} else {
|
|
offset = OFFSETOF(chipcregs_t, eci.lt35.eci_output);
|
|
val = val & ~(1 << 30);
|
|
}
|
|
|
|
si_corereg(sih, SI_CC_IDX, offset, mask, val);
|
|
|
|
/* Set interrupt bit if needed */
|
|
if (is_interrupt) {
|
|
si_corereg(sih, SI_CC_IDX,
|
|
(CCREV(sih->ccrev) < 35 ?
|
|
OFFSETOF(chipcregs_t, eci.lt35.eci_output) :
|
|
OFFSETOF(chipcregs_t, eci.ge35.eci_outputlo)),
|
|
(1 << 30), (1 << 30));
|
|
}
|
|
} else if (sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT) {
|
|
/* GCI Mode */
|
|
if ((mask & 0xFFFF0000) == ECI48_OUT_MASKMAGIC_HIWORD) {
|
|
mask = mask & ~0xFFFF0000;
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_output[1]), mask, val);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
BCMPOSTTRAPFN(seci_restore_coreidx)(si_t *sih, uint32 origidx, bool fast)
|
|
{
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
return;
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_seci_down)(si_t *sih)
|
|
{
|
|
uint32 origidx;
|
|
bool fast;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
const chipcregs_t *cc;
|
|
uint32 offset;
|
|
|
|
if (!si_seci(sih) && !si_seci_uart(sih))
|
|
return;
|
|
/* Don't proceed if request is already made to bring down the clock */
|
|
offset = OFFSETOF(chipcregs_t, clk_ctl_st);
|
|
if (!(si_corereg(sih, 0, offset, 0, 0) & CLKCTL_STS_SECI_CLK_REQ))
|
|
return;
|
|
if (!(cc = si_seci_access_preamble(sih, sii, &origidx, &fast)))
|
|
goto exit;
|
|
|
|
exit:
|
|
/* bring down the clock if up */
|
|
si_seci_clkreq(sih, FALSE);
|
|
|
|
/* restore previous core */
|
|
seci_restore_coreidx(sih, origidx, fast);
|
|
}
|
|
|
|
void
|
|
si_seci_upd(si_t *sih, bool enable)
|
|
{
|
|
uint32 origidx = 0;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
bool fast;
|
|
uint32 regval, seci_ctrl;
|
|
bcm_int_bitmask_t intr_val;
|
|
|
|
if (!si_seci(sih))
|
|
return;
|
|
|
|
fast = SI_FAST(sii);
|
|
INTR_OFF(sii, &intr_val);
|
|
if (!fast) {
|
|
origidx = sii->curidx;
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL)
|
|
goto exit;
|
|
} else if ((cc = (chipcregs_t *)CCREGS_FAST(sii)) == NULL)
|
|
goto exit;
|
|
|
|
ASSERT(cc);
|
|
|
|
/* Select SECI based on enable input */
|
|
if ((CHIPID(sih->chip) == BCM4352_CHIP_ID) || (CHIPID(sih->chip) == BCM4360_CHIP_ID)) {
|
|
regval = R_REG(sii->osh, &cc->chipcontrol);
|
|
|
|
seci_ctrl = CCTRL4360_SECI_ON_GPIO01;
|
|
|
|
if (enable) {
|
|
regval |= seci_ctrl;
|
|
} else {
|
|
regval &= ~seci_ctrl;
|
|
}
|
|
W_REG(sii->osh, &cc->chipcontrol, regval);
|
|
|
|
if (enable) {
|
|
/* Send ECI update to BT */
|
|
regval = R_REG(sii->osh, &cc->SECI_config);
|
|
regval |= SECI_UPD_SECI;
|
|
W_REG(sii->osh, &cc->SECI_config, regval);
|
|
SPINWAIT((R_REG(sii->osh, &cc->SECI_config) & SECI_UPD_SECI), 1000);
|
|
/* Request ECI update from BT */
|
|
W_REG(sii->osh, &cc->seci_uart_data, SECI_SLIP_ESC_CHAR);
|
|
W_REG(sii->osh, &cc->seci_uart_data, SECI_REFRESH_REQ);
|
|
}
|
|
}
|
|
|
|
exit:
|
|
/* restore previous core */
|
|
if (!fast)
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
void *
|
|
si_gci_init(si_t *sih)
|
|
{
|
|
#ifdef HNDGCI
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
#endif /* HNDGCI */
|
|
|
|
if (sih->cccaps_ext & CC_CAP_EXT_GCI_PRESENT)
|
|
{
|
|
si_gci_reset(sih);
|
|
|
|
if (sih->boardflags4 & BFL4_BTCOEX_OVER_SECI) {
|
|
si_gci_seci_init(sih);
|
|
}
|
|
|
|
/* Set GCI Control bits 40 - 47 to be SW Controlled. These bits
|
|
contain WL channel info and are sent to BT.
|
|
*/
|
|
si_gci_direct(sih, GCI_OFFSETOF(sih, gci_control_1),
|
|
GCI_WL_CHN_INFO_MASK, GCI_WL_CHN_INFO_MASK);
|
|
}
|
|
#ifdef HNDGCI
|
|
hndgci_init(sih, sii->osh, HND_GCI_PLAIN_UART_MODE,
|
|
GCI_UART_BR_115200);
|
|
#endif /* HNDGCI */
|
|
|
|
return (NULL);
|
|
}
|
|
#endif /* BCMECICOEX */
|
|
#endif /* !(BCMDONGLEHOST) */
|
|
|
|
/**
|
|
* For boards that use GPIO(8) is used for Bluetooth Coex TX_WLAN pin,
|
|
* when GPIOControl for Pin 8 is with ChipCommon core,
|
|
* if UART_TX_1 (bit 5: Chipc capabilities) strapping option is set, then
|
|
* GPIO pin 8 is driven by Uart0MCR:2 rather than GPIOOut:8. To drive this pin
|
|
* low, one has to set Uart0MCR:2 to 1. This is required when the BTC is disabled,
|
|
* or the driver goes down. Refer to PR35488.
|
|
*/
|
|
void
|
|
si_btcgpiowar(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
chipcregs_t *cc;
|
|
|
|
/* Make sure that there is ChipCommon core present &&
|
|
* UART_TX is strapped to 1
|
|
*/
|
|
if (!(sih->cccaps & CC_CAP_UARTGPIO))
|
|
return;
|
|
|
|
/* si_corereg cannot be used as we have to guarantee 8-bit read/writes */
|
|
INTR_OFF(sii, &intr_val);
|
|
|
|
origidx = si_coreidx(sih);
|
|
|
|
cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0);
|
|
ASSERT(cc != NULL);
|
|
|
|
W_REG(sii->osh, &cc->uart0mcr, R_REG(sii->osh, &cc->uart0mcr) | 0x04);
|
|
|
|
/* restore the original index */
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
void
|
|
si_chipcontrl_restore(si_t *sih, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL) {
|
|
SI_ERROR(("si_chipcontrl_restore: Failed to find CORE ID!\n"));
|
|
return;
|
|
}
|
|
W_REG(sii->osh, &cc->chipcontrol, val);
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
uint32
|
|
si_chipcontrl_read(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
uint32 val;
|
|
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL) {
|
|
SI_ERROR(("si_chipcontrl_read: Failed to find CORE ID!\n"));
|
|
return -1;
|
|
}
|
|
val = R_REG(sii->osh, &cc->chipcontrol);
|
|
si_setcoreidx(sih, origidx);
|
|
return val;
|
|
}
|
|
|
|
/** switch muxed pins, on: SROM, off: FEMCTRL. Called for a family of ac chips, not just 4360. */
|
|
void
|
|
si_chipcontrl_srom4360(si_t *sih, bool on)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
uint32 val;
|
|
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL) {
|
|
SI_ERROR(("si_chipcontrl_srom4360: Failed to find CORE ID!\n"));
|
|
return;
|
|
}
|
|
val = R_REG(sii->osh, &cc->chipcontrol);
|
|
|
|
if (on) {
|
|
val &= ~(CCTRL4360_SECI_MODE |
|
|
CCTRL4360_BTSWCTRL_MODE |
|
|
CCTRL4360_EXTRA_FEMCTRL_MODE |
|
|
CCTRL4360_BT_LGCY_MODE |
|
|
CCTRL4360_CORE2FEMCTRL4_ON);
|
|
|
|
W_REG(sii->osh, &cc->chipcontrol, val);
|
|
} else {
|
|
/* huh, nothing here? */
|
|
}
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
/**
|
|
* The SROM clock is derived from the backplane clock. For chips having a fast
|
|
* backplane clock that requires a higher-than-POR-default clock divisor ratio for the SROM clock.
|
|
*/
|
|
void
|
|
si_srom_clk_set(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
uint32 val;
|
|
uint32 divisor = 1;
|
|
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL) {
|
|
SI_ERROR(("si_srom_clk_set: Failed to find CORE ID!\n"));
|
|
return;
|
|
}
|
|
|
|
val = R_REG(sii->osh, &cc->clkdiv2);
|
|
ASSERT(0);
|
|
|
|
W_REG(sii->osh, &cc->clkdiv2, ((val & ~CLKD2_SROM) | divisor));
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
void
|
|
si_pmu_avb_clk_set(si_t *sih, osl_t *osh, bool set_flag)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM43460_CHIP_ID:
|
|
case BCM4360_CHIP_ID:
|
|
si_pmu_avbtimer_enable(sih, osh, set_flag);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
void
|
|
si_btc_enable_chipcontrol(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
|
|
if ((cc = (chipcregs_t *)si_setcore(sih, CC_CORE_ID, 0)) == NULL) {
|
|
SI_ERROR(("si_btc_enable_chipcontrol: Failed to find CORE ID!\n"));
|
|
return;
|
|
}
|
|
|
|
/* BT fix */
|
|
W_REG(sii->osh, &cc->chipcontrol,
|
|
R_REG(sii->osh, &cc->chipcontrol) | CC_BTCOEX_EN_MASK);
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
/** cache device removed state */
|
|
void si_set_device_removed(si_t *sih, bool status)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
sii->device_removed = status;
|
|
}
|
|
|
|
/** check if the device is removed */
|
|
bool
|
|
si_deviceremoved(const si_t *sih)
|
|
{
|
|
uint32 w;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (sii->device_removed) {
|
|
return TRUE;
|
|
}
|
|
|
|
switch (BUSTYPE(sih->bustype)) {
|
|
case PCI_BUS:
|
|
ASSERT(SI_INFO(sih)->osh != NULL);
|
|
w = OSL_PCI_READ_CONFIG(SI_INFO(sih)->osh, PCI_CFG_VID, sizeof(uint32));
|
|
if ((w & 0xFFFF) != VENDOR_BROADCOM)
|
|
return TRUE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
bool
|
|
si_is_warmboot(void)
|
|
{
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
bool
|
|
si_is_sprom_available(si_t *sih)
|
|
{
|
|
if (CCREV(sih->ccrev) >= 31) {
|
|
const si_info_t *sii;
|
|
uint origidx;
|
|
chipcregs_t *cc;
|
|
uint32 sromctrl;
|
|
|
|
if ((sih->cccaps & CC_CAP_SROM) == 0)
|
|
return FALSE;
|
|
|
|
sii = SI_INFO(sih);
|
|
origidx = sii->curidx;
|
|
cc = si_setcoreidx(sih, SI_CC_IDX);
|
|
ASSERT(cc);
|
|
sromctrl = R_REG(sii->osh, &cc->sromcontrol);
|
|
si_setcoreidx(sih, origidx);
|
|
return (sromctrl & SRC_PRESENT);
|
|
}
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
if (CHIPREV(sih->chiprev) == 0) {
|
|
/* WAR for 4369a0: HW4369-1729. no sprom, default to otp always. */
|
|
return 0;
|
|
} else {
|
|
return (sih->chipst & CST4369_SPROM_PRESENT) != 0;
|
|
}
|
|
break;
|
|
CASE_BCM43602_CHIP:
|
|
return (sih->chipst & CST43602_SPROM_PRESENT) != 0;
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
return FALSE;
|
|
case BCM4362_CHIP_GRPID:
|
|
return (sih->chipst & CST4362_SPROM_PRESENT) != 0;
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
return (sih->chipst & CST4378_SPROM_PRESENT) != 0;
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
return (sih->chipst & CST4387_SPROM_PRESENT) != 0;
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
/* 4389 supports only OTP */
|
|
return FALSE;
|
|
default:
|
|
return TRUE;
|
|
}
|
|
}
|
|
|
|
bool
|
|
si_is_sflash_available(const si_t *sih)
|
|
{
|
|
return (sih->chipst & CST_SFLASH_PRESENT) != 0;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
bool
|
|
si_is_otp_disabled(const si_t *sih)
|
|
{
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4360_CHIP_ID:
|
|
case BCM43526_CHIP_ID:
|
|
case BCM43460_CHIP_ID:
|
|
case BCM4352_CHIP_ID:
|
|
case BCM43602_CHIP_ID:
|
|
/* 4360 OTP is always powered and enabled */
|
|
return FALSE;
|
|
/* These chips always have their OTP on */
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
default:
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
bool
|
|
si_is_otp_powered(si_t *sih)
|
|
{
|
|
if (PMUCTL_ENAB(sih))
|
|
return si_pmu_is_otp_powered(sih, si_osh(sih));
|
|
return TRUE;
|
|
}
|
|
|
|
void
|
|
si_otp_power(si_t *sih, bool on, uint32* min_res_mask)
|
|
{
|
|
if (PMUCTL_ENAB(sih))
|
|
si_pmu_otp_power(sih, si_osh(sih), on, min_res_mask);
|
|
OSL_DELAY(1000);
|
|
}
|
|
|
|
/* Return BCME_NOTFOUND if the card doesn't have CIS format nvram */
|
|
int
|
|
si_cis_source(const si_t *sih)
|
|
{
|
|
/* Most PCI chips use SROM format instead of CIS */
|
|
if (BUSTYPE(sih->bustype) == PCI_BUS) {
|
|
return BCME_NOTFOUND;
|
|
}
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4360_CHIP_ID:
|
|
case BCM43460_CHIP_ID:
|
|
case BCM4352_CHIP_ID:
|
|
case BCM43526_CHIP_ID: {
|
|
if ((sih->chipst & CST4360_OTP_ENABLED))
|
|
return CIS_OTP;
|
|
return CIS_DEFAULT;
|
|
}
|
|
CASE_BCM43602_CHIP:
|
|
if (sih->chipst & CST43602_SPROM_PRESENT) {
|
|
/* Don't support CIS formatted SROM, use 'real' SROM format instead */
|
|
return BCME_NOTFOUND;
|
|
}
|
|
return CIS_OTP;
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
return CIS_OTP;
|
|
case BCM4369_CHIP_GRPID:
|
|
if (CHIPREV(sih->chiprev) == 0) {
|
|
/* WAR for 4369a0: HW4369-1729 */
|
|
return CIS_OTP;
|
|
} else if (sih->chipst & CST4369_SPROM_PRESENT) {
|
|
return CIS_SROM;
|
|
}
|
|
return CIS_OTP;
|
|
case BCM4362_CHIP_GRPID:
|
|
return ((sih->chipst & CST4362_SPROM_PRESENT)? CIS_SROM : CIS_OTP);
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
if (sih->chipst & CST4378_SPROM_PRESENT)
|
|
return CIS_SROM;
|
|
return CIS_OTP;
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
if (sih->chipst & CST4387_SPROM_PRESENT)
|
|
return CIS_SROM;
|
|
return CIS_OTP;
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
/* 4389 supports only OTP */
|
|
return CIS_OTP;
|
|
default:
|
|
return CIS_DEFAULT;
|
|
}
|
|
}
|
|
|
|
uint16 si_fabid(si_t *sih)
|
|
{
|
|
uint32 data;
|
|
uint16 fabid = 0;
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
CASE_BCM43602_CHIP:
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
data = si_corereg(sih, SI_CC_IDX, OFFSETOF(chipcregs_t, fabid), 0, 0);
|
|
fabid = data & 0xf;
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return fabid;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
uint32 si_get_sromctl(si_t *sih)
|
|
{
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
uint32 sromctl;
|
|
osl_t *osh = si_osh(sih);
|
|
|
|
cc = si_setcoreidx(sih, SI_CC_IDX);
|
|
ASSERT((uintptr)cc);
|
|
|
|
sromctl = R_REG(osh, &cc->sromcontrol);
|
|
|
|
/* return to the original core */
|
|
si_setcoreidx(sih, origidx);
|
|
return sromctl;
|
|
}
|
|
|
|
int si_set_sromctl(si_t *sih, uint32 value)
|
|
{
|
|
chipcregs_t *cc;
|
|
uint origidx = si_coreidx(sih);
|
|
osl_t *osh = si_osh(sih);
|
|
int ret = BCME_OK;
|
|
|
|
cc = si_setcoreidx(sih, SI_CC_IDX);
|
|
ASSERT((uintptr)cc);
|
|
|
|
/* get chipcommon rev */
|
|
if (si_corerev(sih) >= 32) {
|
|
/* SpromCtrl is only accessible if CoreCapabilities.SpromSupported and
|
|
* SpromPresent is 1.
|
|
*/
|
|
if ((R_REG(osh, &cc->capabilities) & CC_CAP_SROM) != 0 &&
|
|
(R_REG(osh, &cc->sromcontrol) & SRC_PRESENT)) {
|
|
W_REG(osh, &cc->sromcontrol, value);
|
|
} else {
|
|
ret = BCME_NODEVICE;
|
|
}
|
|
} else {
|
|
ret = BCME_UNSUPPORTED;
|
|
}
|
|
|
|
/* return to the original core */
|
|
si_setcoreidx(sih, origidx);
|
|
|
|
return ret;
|
|
}
|
|
|
|
uint
|
|
BCMPOSTTRAPFN(si_core_wrapperreg)(si_t *sih, uint32 coreidx, uint32 offset, uint32 mask, uint32 val)
|
|
{
|
|
uint origidx;
|
|
bcm_int_bitmask_t intr_val;
|
|
uint ret_val;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
origidx = si_coreidx(sih);
|
|
|
|
INTR_OFF(sii, &intr_val);
|
|
/* Validate the core idx */
|
|
si_setcoreidx(sih, coreidx);
|
|
|
|
ret_val = si_wrapperreg(sih, offset, mask, val);
|
|
|
|
/* return to the original core */
|
|
si_setcoreidx(sih, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
return ret_val;
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
static void
|
|
si_pmu_sr_upd(si_t *sih)
|
|
{
|
|
#if defined(SAVERESTORE)
|
|
if (SR_ENAB()) {
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
/* min_mask is updated after SR code is downloaded to txfifo */
|
|
if (PMUCTL_ENAB(sih))
|
|
si_pmu_res_minmax_update(sih, sii->osh);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* To make sure that, res mask is minimal to save power and also, to indicate
|
|
* specifically to host about the SR logic.
|
|
*/
|
|
void
|
|
si_update_masks(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
CASE_BCM43602_CHIP:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
/* Assumes SR engine has been enabled */
|
|
if (PMUCTL_ENAB(sih))
|
|
si_pmu_res_minmax_update(sih, sii->osh);
|
|
break;
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID:
|
|
/* min_mask is updated after SR code is downloaded to txfifo */
|
|
si_pmu_sr_upd(sih);
|
|
PMU_REG(sih, mac_res_req_timer, ~0x0, PMU43012_MAC_RES_REQ_TIMER);
|
|
PMU_REG(sih, mac_res_req_mask, ~0x0, PMU43012_MAC_RES_REQ_MASK);
|
|
break;
|
|
|
|
default:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void
|
|
si_force_islanding(si_t *sih, bool enable)
|
|
{
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM43012_CHIP_ID:
|
|
case BCM43013_CHIP_ID:
|
|
case BCM43014_CHIP_ID: {
|
|
if (enable) {
|
|
/* Turn on the islands */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0x00000053, 0x0);
|
|
#ifdef USE_MEMLPLDO
|
|
/* Force vddm pwrsw always on */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0x000003, 0x000003);
|
|
#endif
|
|
#ifdef BCMQT
|
|
/* Turn off the islands */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0x000050, 0x000050);
|
|
#endif
|
|
} else {
|
|
/* Turn off the islands */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, 0x000050, 0x000050);
|
|
}
|
|
}
|
|
break;
|
|
|
|
default:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
/* cleanup the timer from the host when ARM is been halted
|
|
* without a chance for ARM cleanup its resources
|
|
* If left not cleanup, Intr from a software timer can still
|
|
* request HT clk when ARM is halted.
|
|
*/
|
|
uint32
|
|
si_pmu_res_req_timer_clr(si_t *sih)
|
|
{
|
|
uint32 mask;
|
|
|
|
mask = PRRT_REQ_ACTIVE | PRRT_INTEN | PRRT_HT_REQ;
|
|
mask <<= 14;
|
|
/* clear mask bits */
|
|
pmu_corereg(sih, SI_CC_IDX, res_req_timer, mask, 0);
|
|
/* readback to ensure write completes */
|
|
return pmu_corereg(sih, SI_CC_IDX, res_req_timer, 0, 0);
|
|
}
|
|
|
|
/** turn on/off rfldo */
|
|
void
|
|
si_pmu_rfldo(si_t *sih, bool on)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4360_CHIP_ID:
|
|
case BCM4352_CHIP_ID:
|
|
case BCM43526_CHIP_ID: {
|
|
CASE_BCM43602_CHIP:
|
|
si_pmu_vreg_control(sih, PMU_VREG_0, RCTRL4360_RFLDO_PWR_DOWN,
|
|
on ? 0 : RCTRL4360_RFLDO_PWR_DOWN);
|
|
break;
|
|
}
|
|
default:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/* Caller of this function should make sure is on PCIE core
|
|
* Used in pciedev.c.
|
|
*/
|
|
void
|
|
si_pcie_disable_oobselltr(const si_t *sih)
|
|
{
|
|
ASSERT(si_coreid(sih) == PCIE2_CORE_ID);
|
|
if (PCIECOREREV(sih->buscorerev) >= 23)
|
|
si_wrapperreg(sih, AI_OOBSELIND74, ~0, 0);
|
|
else
|
|
si_wrapperreg(sih, AI_OOBSELIND30, ~0, 0);
|
|
}
|
|
|
|
void
|
|
si_pcie_ltr_war(const si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pcie_ltr_war(sii->pch, si_pcieltrenable(sih, 0, 0));
|
|
#endif /* !defined(BCMDONGLEHOST */
|
|
}
|
|
|
|
void
|
|
si_pcie_hw_LTR_war(const si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pcie_hw_LTR_war(sii->pch);
|
|
#endif /* !defined(BCMDONGLEHOST */
|
|
}
|
|
|
|
void
|
|
si_pciedev_reg_pm_clk_period(const si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pciedev_reg_pm_clk_period(sii->pch);
|
|
#endif /* !defined(BCMDONGLEHOST */
|
|
}
|
|
|
|
void
|
|
si_pciedev_crwlpciegen2(const si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pciedev_crwlpciegen2(sii->pch);
|
|
#endif /* !defined(BCMDONGLEHOST */
|
|
}
|
|
|
|
void
|
|
si_pcie_prep_D3(const si_t *sih, bool enter_D3)
|
|
{
|
|
#if !defined(BCMDONGLEHOST)
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
if (PCIE_GEN2(sii))
|
|
pciedev_prep_D3(sii->pch, enter_D3);
|
|
#endif /* !defined(BCMDONGLEHOST */
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
uint
|
|
BCMPOSTTRAPFN(si_corereg_ifup)(si_t *sih, uint core_id, uint regoff, uint mask, uint val)
|
|
{
|
|
bool isup;
|
|
volatile void *regs;
|
|
uint origidx, ret_val, coreidx;
|
|
|
|
/* Remember original core before switch to chipc */
|
|
origidx = si_coreidx(sih);
|
|
regs = si_setcore(sih, core_id, 0);
|
|
BCM_REFERENCE(regs);
|
|
ASSERT(regs != NULL);
|
|
|
|
coreidx = si_coreidx(sih);
|
|
|
|
isup = si_iscoreup(sih);
|
|
if (isup == TRUE) {
|
|
ret_val = si_corereg(sih, coreidx, regoff, mask, val);
|
|
} else {
|
|
ret_val = 0;
|
|
}
|
|
|
|
/* Return to original core */
|
|
si_setcoreidx(sih, origidx);
|
|
return ret_val;
|
|
}
|
|
|
|
/* 43012 specific low power settings.
|
|
* See http://confluence.broadcom.com/display/WLAN/BCM43012+Low+Power+Settings.
|
|
* See 47xxtcl/43012.tcl proc lp_enable.
|
|
*/
|
|
void si_43012_lp_enable(si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
bcm_int_bitmask_t intr_val;
|
|
uint origidx;
|
|
int count;
|
|
gciregs_t *gciregs;
|
|
|
|
/* Block ints and save current core */
|
|
INTR_OFF(sii, &intr_val);
|
|
origidx = si_coreidx(sih);
|
|
|
|
/* Enable radiodig clk gating */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG5, PMUCCTL05_43012_RADIO_DIG_CLK_GATING_EN,
|
|
PMUCCTL05_43012_RADIO_DIG_CLK_GATING_EN);
|
|
|
|
/* Disable SPM clock */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG5, PMUCCTL05_43012_DISABLE_SPM_CLK,
|
|
PMUCCTL05_43012_DISABLE_SPM_CLK);
|
|
|
|
/* Enable access of radiodig registers using async apb interface */
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG6, PMUCCTL06_43012_GCI2RDIG_USE_ASYNCAPB,
|
|
PMUCCTL06_43012_GCI2RDIG_USE_ASYNCAPB);
|
|
|
|
/* Remove SFLASH clock request (which is default on for boot-from-flash support) */
|
|
CHIPC_REG(sih, clk_ctl_st, CCS_SFLASH_CLKREQ | CCS_HQCLKREQ, CCS_HQCLKREQ);
|
|
|
|
/* Switch to GCI core */
|
|
if (!(gciregs = si_setcore(sih, GCI_CORE_ID, 0))) {
|
|
goto done;
|
|
}
|
|
|
|
/* GCIForceRegClk Off */
|
|
if (!(sih->lpflags & LPFLAGS_SI_GCI_FORCE_REGCLK_DISABLE)) {
|
|
si_gci_direct(sih, GET_GCI_OFFSET(sih, gci_corectrl),
|
|
GCI_CORECTRL_FORCEREGCLK_MASK, 0);
|
|
}
|
|
|
|
/* Disable the sflash pad */
|
|
if (!(sih->lpflags & LPFLAGS_SI_SFLASH_DISABLE)) {
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03,
|
|
CC_GCI_03_LPFLAGS_SFLASH_MASK, CC_GCI_03_LPFLAGS_SFLASH_VAL);
|
|
}
|
|
|
|
/* Input disable all LHL I/O pins */
|
|
for (count = 0; count < GPIO_CTRL_REG_COUNT; count++) {
|
|
OR_REG(sii->osh, &gciregs->gpio_ctrl_iocfg_p_adr[count],
|
|
GPIO_CTRL_REG_DISABLE_INTERRUPT);
|
|
}
|
|
|
|
/* Power down BT LDO3p3 */
|
|
if (!(sih->lpflags & LPFLAGS_SI_BTLDO3P3_DISABLE)) {
|
|
si_pmu_chipcontrol(sih, CHIPCTRLREG2, PMUCCTL02_43012_BTLDO3P3_PU_FORCE_OFF,
|
|
PMUCCTL02_43012_BTLDO3P3_PU_FORCE_OFF);
|
|
}
|
|
|
|
done:
|
|
si_setcoreidx(sih, origidx);
|
|
INTR_RESTORE(sii, &intr_val);
|
|
}
|
|
|
|
/** this function is called from the BMAC during (re) initialisation */
|
|
void
|
|
si_lowpwr_opt(si_t *sih)
|
|
{
|
|
uint mask, val;
|
|
|
|
/* 43602 chip (all revision) related changes */
|
|
if (BCM43602_CHIP(sih->chip)) {
|
|
uint hosti = si_chip_hostif(sih);
|
|
uint origidx = si_coreidx(sih);
|
|
volatile void *regs;
|
|
|
|
regs = si_setcore(sih, CC_CORE_ID, 0);
|
|
BCM_REFERENCE(regs);
|
|
ASSERT(regs != NULL);
|
|
|
|
/* disable usb app clk */
|
|
/* Can be done any time. If it is not USB, then do it. In case */
|
|
/* of USB, do not write it */
|
|
if (hosti != CHIP_HOSTIF_USBMODE && !BCM43602_CHIP(sih->chip)) {
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL5, (1 << USBAPP_CLK_BIT), 0);
|
|
}
|
|
/* disable pcie clks */
|
|
if (hosti != CHIP_HOSTIF_PCIEMODE) {
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL5, (1 << PCIE_CLK_BIT), 0);
|
|
}
|
|
|
|
/* disable armcr4 debug clk */
|
|
/* Can be done anytime as long as driver is functional. */
|
|
/* In TCL, dhalt commands needs to change to undo this */
|
|
switch (CHIPID(sih->chip)) {
|
|
CASE_BCM43602_CHIP:
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL3,
|
|
PMU43602_CC3_ARMCR4_DBG_CLK, 0);
|
|
break;
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
{
|
|
uint32 tapsel = si_corereg(sih, SI_CC_IDX,
|
|
OFFSETOF(chipcregs_t, jtagctrl), 0, 0)
|
|
& JCTRL_TAPSEL_BIT;
|
|
/* SWD: if tap sel bit set, */
|
|
/* enable armcr4 debug clock */
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL5,
|
|
(1 << ARMCR4_DBG_CLK_BIT),
|
|
tapsel?(1 << ARMCR4_DBG_CLK_BIT):0);
|
|
}
|
|
break;
|
|
default:
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL5,
|
|
(1 << ARMCR4_DBG_CLK_BIT), 0);
|
|
break;
|
|
}
|
|
|
|
/* Power down unused BBPLL ch-6(pcie_tl_clk) and ch-5(sample-sync-clk), */
|
|
/* valid in all modes, ch-5 needs to be reenabled for sample-capture */
|
|
/* this needs to be done in the pmu init path, at the beginning. Should not be */
|
|
/* a pcie driver. Enable the sample-sync-clk in the sample capture function */
|
|
if (BCM43602_CHIP(sih->chip)) {
|
|
/* configure open loop PLL parameters, open loop is used during S/R */
|
|
val = (3 << PMU1_PLL0_PC1_M1DIV_SHIFT) | (6 << PMU1_PLL0_PC1_M2DIV_SHIFT) |
|
|
(6 << PMU1_PLL0_PC1_M3DIV_SHIFT) | (8 << PMU1_PLL0_PC1_M4DIV_SHIFT);
|
|
si_pmu_pllcontrol(sih, PMU1_PLL0_PLLCTL4, ~0, val);
|
|
si_pmu_pllupd(sih);
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL2,
|
|
PMU43602_CC2_PCIE_CLKREQ_L_WAKE_EN | PMU43602_CC2_PMU_WAKE_ALP_AVAIL_EN,
|
|
PMU43602_CC2_PCIE_CLKREQ_L_WAKE_EN | PMU43602_CC2_PMU_WAKE_ALP_AVAIL_EN);
|
|
}
|
|
|
|
/* Return to original core */
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
if ((CHIPID(sih->chip) == BCM43012_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43013_CHIP_ID) ||
|
|
(CHIPID(sih->chip) == BCM43014_CHIP_ID)) {
|
|
/* Enable memory standby based on lpflags */
|
|
if (sih->lpflags & LPFLAGS_SI_GLOBAL_DISABLE) {
|
|
SI_MSG(("si_lowpwr_opt: Disable lower power configuration!\n"));
|
|
goto exit;
|
|
}
|
|
|
|
SI_MSG(("si_lowpwr_opt: Enable lower power configuration!\n"));
|
|
|
|
/* Enable mem clk gating */
|
|
mask = (0x1 << MEM_CLK_GATE_BIT);
|
|
val = (0x1 << MEM_CLK_GATE_BIT);
|
|
|
|
si_corereg_ifup(sih, SDIOD_CORE_ID, SI_PWR_CTL_ST, mask, val);
|
|
si_corereg_ifup(sih, SOCRAM_CORE_ID, SI_PWR_CTL_ST, mask, val);
|
|
|
|
si_43012_lp_enable(sih);
|
|
}
|
|
exit:
|
|
return;
|
|
}
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#if defined(AXI_TIMEOUTS) || defined(AXI_TIMEOUTS_NIC)
|
|
uint32
|
|
BCMPOSTTRAPFN(si_clear_backplane_to_per_core)(si_t *sih, uint coreid, uint coreunit, void * wrap)
|
|
{
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS)) {
|
|
return ai_clear_backplane_to_per_core(sih, coreid, coreunit, wrap);
|
|
}
|
|
return AXI_WRAP_STS_NONE;
|
|
}
|
|
#endif /* AXI_TIMEOUTS || AXI_TIMEOUTS_NIC */
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_clear_backplane_to)(si_t *sih)
|
|
{
|
|
if ((CHIPTYPE(sih->socitype) == SOCI_AI) ||
|
|
(CHIPTYPE(sih->socitype) == SOCI_DVTBUS)) {
|
|
return ai_clear_backplane_to(sih);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
si_update_backplane_timeouts(const si_t *sih, bool enable, uint32 timeout_exp,
|
|
uint32 cid)
|
|
{
|
|
#if defined(AXI_TIMEOUTS) || defined(AXI_TIMEOUTS_NIC)
|
|
/* Enable only for AXI */
|
|
if (CHIPTYPE(sih->socitype) != SOCI_AI) {
|
|
return;
|
|
}
|
|
|
|
ai_update_backplane_timeouts(sih, enable, timeout_exp, cid);
|
|
#endif /* AXI_TIMEOUTS || AXI_TIMEOUTS_NIC */
|
|
}
|
|
|
|
/*
|
|
* This routine adds the AXI timeouts for
|
|
* chipcommon, pcie and ARM slave wrappers
|
|
*/
|
|
void
|
|
si_slave_wrapper_add(si_t *sih)
|
|
{
|
|
#if defined(AXI_TIMEOUTS) || defined(AXI_TIMEOUTS_NIC)
|
|
uint32 axi_to = 0;
|
|
|
|
/* Enable only for AXI */
|
|
if ((CHIPTYPE(sih->socitype) != SOCI_AI) &&
|
|
(CHIPTYPE(sih->socitype) != SOCI_DVTBUS)) {
|
|
return;
|
|
}
|
|
|
|
axi_to = AXI_TO_VAL;
|
|
|
|
/* All required slave wrappers are added in ai_scan */
|
|
ai_update_backplane_timeouts(sih, TRUE, axi_to, 0);
|
|
|
|
#ifdef DISABLE_PCIE2_AXI_TIMEOUT
|
|
ai_update_backplane_timeouts(sih, FALSE, 0, PCIE_CORE_ID);
|
|
ai_update_backplane_timeouts(sih, FALSE, 0, PCIE2_CORE_ID);
|
|
#endif
|
|
|
|
#endif /* AXI_TIMEOUTS || AXI_TIMEOUTS_NIC */
|
|
|
|
}
|
|
|
|
#ifndef BCMDONGLEHOST
|
|
/* read from pcie space using back plane indirect access */
|
|
/* Set Below mask for reading 1, 2, 4 bytes in single read */
|
|
/* #define SI_BPIND_1BYTE 0x1 */
|
|
/* #define SI_BPIND_2BYTE 0x3 */
|
|
/* #define SI_BPIND_4BYTE 0xF */
|
|
int
|
|
BCMPOSTTRAPFN(si_bpind_access)(si_t *sih, uint32 addr_high, uint32 addr_low,
|
|
int32 * data, bool read, uint32 us_timeout)
|
|
{
|
|
|
|
uint32 status = 0;
|
|
uint8 mask = SI_BPIND_4BYTE;
|
|
int ret_val = BCME_OK;
|
|
|
|
/* Program Address low and high fields */
|
|
si_ccreg(sih, OFFSETOF(chipcregs_t, bp_addrlow), ~0, addr_low);
|
|
si_ccreg(sih, OFFSETOF(chipcregs_t, bp_addrhigh), ~0, addr_high);
|
|
|
|
if (read) {
|
|
/* Start the read */
|
|
si_ccreg(sih, OFFSETOF(chipcregs_t, bp_indaccess), ~0,
|
|
CC_BP_IND_ACCESS_START_MASK | mask);
|
|
} else {
|
|
/* Write the data and force the trigger */
|
|
si_ccreg(sih, OFFSETOF(chipcregs_t, bp_data), ~0, *data);
|
|
si_ccreg(sih, OFFSETOF(chipcregs_t, bp_indaccess), ~0,
|
|
CC_BP_IND_ACCESS_START_MASK |
|
|
CC_BP_IND_ACCESS_RDWR_MASK | mask);
|
|
|
|
}
|
|
|
|
/* Wait for status to be cleared */
|
|
SPINWAIT(((status = si_ccreg(sih, OFFSETOF(chipcregs_t, bp_indaccess), 0, 0)) &
|
|
CC_BP_IND_ACCESS_START_MASK), us_timeout);
|
|
|
|
if (status & (CC_BP_IND_ACCESS_START_MASK | CC_BP_IND_ACCESS_ERROR_MASK)) {
|
|
ret_val = BCME_ERROR;
|
|
SI_ERROR(("Action Failed for address 0x%08x:0x%08x \t status: 0x%x\n",
|
|
addr_high, addr_low, status));
|
|
/* For ATE, Stop execution here, to catch BPind timeout */
|
|
#ifdef ATE_BUILD
|
|
hnd_die();
|
|
#endif /* ATE_BUILD */
|
|
} else {
|
|
/* read data */
|
|
if (read)
|
|
*data = si_ccreg(sih, OFFSETOF(chipcregs_t, bp_data), 0, 0);
|
|
}
|
|
|
|
return ret_val;
|
|
}
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
void
|
|
si_pll_sr_reinit(si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST) && !defined(DONGLEBUILD)
|
|
osl_t *osh = si_osh(sih);
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 data;
|
|
|
|
/* disable PLL open loop operation */
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM43602_CHIP_ID:
|
|
/* read back the pll openloop state */
|
|
data = si_pmu_pllcontrol(sih, PMU1_PLL0_PLLCTL8, 0, 0);
|
|
/* check current pll mode */
|
|
if ((data & PMU1_PLLCTL8_OPENLOOP_MASK) == 0) {
|
|
/* no POR; don't required pll and saverestore init */
|
|
return;
|
|
}
|
|
si_pmu_pll_init(sih, osh, sii->xtalfreq);
|
|
si_pmu_pllcontrol(sih, PMU1_PLL0_PLLCTL8, PMU1_PLLCTL8_OPENLOOP_MASK, 0);
|
|
si_pmu_pllupd(sih);
|
|
/* allow PLL to settle after config PLL for closeloop operation */
|
|
OSL_DELAY(100);
|
|
break;
|
|
default:
|
|
/* any unsupported chip bail */
|
|
return;
|
|
}
|
|
si_pmu_init(sih, osh);
|
|
si_pmu_chip_init(sih, osh);
|
|
#if defined(BCMPMU_STATS)
|
|
if (PMU_STATS_ENAB()) {
|
|
si_pmustatstimer_init(sih);
|
|
}
|
|
#endif /* BCMPMU_STATS */
|
|
#if defined(SR_ESSENTIALS)
|
|
/* Module can be power down during D3 state, thus
|
|
* needs this before si_pmu_res_init() to use sr_isenab()
|
|
* Full dongle may not need to reinit saverestore
|
|
*/
|
|
if (SR_ESSENTIALS_ENAB()) {
|
|
sr_save_restore_init(sih);
|
|
}
|
|
#endif /* SR_ESSENTIALS */
|
|
si_pmu_res_init(sih, sii->osh);
|
|
si_pmu_swreg_init(sih, osh);
|
|
si_lowpwr_opt(sih);
|
|
#endif /* !BCMDONGLEHOST && !DONGLEBUILD */
|
|
}
|
|
|
|
void
|
|
si_pll_closeloop(si_t *sih)
|
|
{
|
|
#if !defined(BCMDONGLEHOST) && !defined(DONGLEBUILD) || defined(SAVERESTORE)
|
|
uint32 data;
|
|
|
|
BCM_REFERENCE(data);
|
|
|
|
/* disable PLL open loop operation */
|
|
switch (CHIPID(sih->chip)) {
|
|
#if !defined(BCMDONGLEHOST) && !defined(DONGLEBUILD)
|
|
/* Don't apply those changes to FULL DONGLE mode since the
|
|
* behaviour was not verified
|
|
*/
|
|
case BCM43602_CHIP_ID:
|
|
/* read back the pll openloop state */
|
|
data = si_pmu_pllcontrol(sih, PMU1_PLL0_PLLCTL8, 0, 0);
|
|
/* current mode is openloop (possible POR) */
|
|
if ((data & PMU1_PLLCTL8_OPENLOOP_MASK) != 0) {
|
|
si_pmu_pllcontrol(sih, PMU1_PLL0_PLLCTL8,
|
|
PMU1_PLLCTL8_OPENLOOP_MASK, 0);
|
|
si_pmu_pllupd(sih);
|
|
/* allow PLL to settle after config PLL for closeloop operation */
|
|
OSL_DELAY(100);
|
|
}
|
|
break;
|
|
#endif /* !BCMDONGLEHOST && !DONGLEBUILD */
|
|
case BCM4369_CHIP_GRPID:
|
|
case BCM4362_CHIP_GRPID:
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL1,
|
|
PMU_CC1_ENABLE_CLOSED_LOOP_MASK, PMU_CC1_ENABLE_CLOSED_LOOP);
|
|
break;
|
|
default:
|
|
/* any unsupported chip bail */
|
|
return;
|
|
}
|
|
#endif /* !BCMDONGLEHOST && !DONGLEBUILD || SAVERESTORE */
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
void
|
|
BCMPOSTTRAPFN(si_introff)(const si_t *sih, bcm_int_bitmask_t *intr_val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
INTR_OFF(sii, intr_val);
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_intrrestore)(const si_t *sih, bcm_int_bitmask_t *intr_val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
INTR_RESTORE(sii, intr_val);
|
|
}
|
|
|
|
bool
|
|
si_get_nvram_rfldo3p3_war(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->rfldo3p3_war;
|
|
}
|
|
|
|
void
|
|
si_nvram_res_masks(const si_t *sih, uint32 *min_mask, uint32 *max_mask)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
/* Apply nvram override to min mask */
|
|
if (sii->min_mask_valid == TRUE) {
|
|
SI_MSG(("Applying rmin=%d to min_mask\n", sii->nvram_min_mask));
|
|
*min_mask = sii->nvram_min_mask;
|
|
}
|
|
/* Apply nvram override to max mask */
|
|
if (sii->max_mask_valid == TRUE) {
|
|
SI_MSG(("Applying rmax=%d to max_mask\n", sii->nvram_max_mask));
|
|
*max_mask = sii->nvram_max_mask;
|
|
}
|
|
}
|
|
|
|
uint8
|
|
si_getspurmode(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->spurmode;
|
|
}
|
|
|
|
uint32
|
|
si_xtalfreq(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->xtalfreq;
|
|
}
|
|
|
|
uint32
|
|
si_get_openloop_dco_code(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->openloop_dco_code;
|
|
}
|
|
|
|
void
|
|
si_set_openloop_dco_code(si_t *sih, uint32 _openloop_dco_code)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
sii->openloop_dco_code = _openloop_dco_code;
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_get_armpllclkfreq)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 armpllclkfreq = ARMPLL_FREQ_400MHZ;
|
|
BCM_REFERENCE(sii);
|
|
|
|
#ifdef DONGLEBUILD
|
|
uint32 armpllclk_max;
|
|
|
|
#if defined(__ARM_ARCH_7R__)
|
|
armpllclk_max = ARMPLL_FREQ_400MHZ;
|
|
#elif defined(__ARM_ARCH_7A__)
|
|
armpllclk_max = ARMPLL_FREQ_1000MHZ;
|
|
#else
|
|
#error "Unknown CPU architecture for armpllclkfreq!"
|
|
#endif
|
|
|
|
armpllclkfreq = (sii->armpllclkfreq) ? sii->armpllclkfreq : armpllclk_max;
|
|
|
|
SI_MSG(("armpllclkfreq = %d\n", armpllclkfreq));
|
|
#endif /* DONGLEBUILD */
|
|
|
|
return armpllclkfreq;
|
|
}
|
|
|
|
uint8
|
|
BCMPOSTTRAPFN(si_get_ccidiv)(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint8 ccidiv = 0xFF;
|
|
BCM_REFERENCE(sii);
|
|
|
|
#ifdef DONGLEBUILD
|
|
ccidiv = (sii->ccidiv) ? sii->ccidiv : CCIDIV_3_TO_1;
|
|
#endif /* DONGLEBUILD */
|
|
|
|
return ccidiv;
|
|
}
|
|
#ifdef DONGLEBUILD
|
|
uint32
|
|
si_wrapper_dump_buf_size(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI)
|
|
return ai_wrapper_dump_buf_size(sih);
|
|
return 0;
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_wrapper_dump_binary)(const si_t *sih, uchar *p)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI)
|
|
return ai_wrapper_dump_binary(sih, p);
|
|
return 0;
|
|
}
|
|
|
|
#if defined(ETD) && !defined(ETD_DISABLED)
|
|
uint32
|
|
BCMPOSTTRAPFN(si_wrapper_dump_last_timeout)(const si_t *sih, uint32 *error, uint32 *core,
|
|
uint32 *ba, uchar *p)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI)
|
|
return ai_wrapper_dump_last_timeout(sih, error, core, ba, p);
|
|
return 0;
|
|
}
|
|
#endif /* ETD && !ETD_DISABLED */
|
|
#endif /* DONGLEBUILD */
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_findcoreidx_by_axiid)(const si_t *sih, uint32 axiid)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI)
|
|
return ai_findcoreidx_by_axiid(sih, axiid);
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
BCMPOSTTRAPFN(si_wrapper_get_last_error)(const si_t *sih, uint32 *error_status, uint32 *core,
|
|
uint32 *lo, uint32 *hi, uint32 *id)
|
|
{
|
|
#if defined(AXI_TIMEOUTS_NIC) || defined(AXI_TIMEOUTS)
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI)
|
|
ai_wrapper_get_last_error(sih, error_status, core, lo, hi, id);
|
|
#endif /* (AXI_TIMEOUTS_NIC) || (AXI_TIMEOUTS) */
|
|
return;
|
|
}
|
|
|
|
uint32
|
|
si_get_axi_timeout_reg(const si_t *sih)
|
|
{
|
|
#if defined(AXI_TIMEOUTS_NIC) || defined(AXI_TIMEOUTS)
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
return ai_get_axi_timeout_reg();
|
|
}
|
|
#endif /* (AXI_TIMEOUTS_NIC) || (AXI_TIMEOUTS) */
|
|
return 0;
|
|
}
|
|
|
|
#if defined(BCMSRPWR) && !defined(BCMSRPWR_DISABLED)
|
|
bool _bcmsrpwr = TRUE;
|
|
#else
|
|
bool _bcmsrpwr = FALSE;
|
|
#endif
|
|
|
|
#define PWRREQ_OFFSET(sih) OFFSETOF(chipcregs_t, powerctl)
|
|
|
|
static void
|
|
BCMPOSTTRAPFN(si_corereg_pciefast_write)(const si_t *sih, uint regoff, uint val)
|
|
{
|
|
volatile uint32 *r = NULL;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT((BUSTYPE(sih->bustype) == PCI_BUS));
|
|
|
|
r = (volatile uint32 *)((volatile char *)sii->curmap +
|
|
PCI_16KB0_PCIREGS_OFFSET + regoff);
|
|
|
|
W_REG(sii->osh, r, val);
|
|
}
|
|
|
|
static uint
|
|
BCMPOSTTRAPFN(si_corereg_pciefast_read)(const si_t *sih, uint regoff)
|
|
{
|
|
volatile uint32 *r = NULL;
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
ASSERT((BUSTYPE(sih->bustype) == PCI_BUS));
|
|
|
|
r = (volatile uint32 *)((volatile char *)sii->curmap +
|
|
PCI_16KB0_PCIREGS_OFFSET + regoff);
|
|
|
|
return R_REG(sii->osh, r);
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_srpwr_request)(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 r, offset = (BUSTYPE(sih->bustype) == SI_BUS) ?
|
|
OFFSETOF(chipcregs_t, powerctl) : PWRREQ_OFFSET(sih);
|
|
uint32 mask2 = mask;
|
|
uint32 val2 = val;
|
|
volatile uint32 *fast_srpwr_addr = (volatile uint32 *)((uintptr)SI_ENUM_BASE(sih)
|
|
+ (uintptr)offset);
|
|
|
|
if (mask || val) {
|
|
mask <<= SRPWR_REQON_SHIFT;
|
|
val <<= SRPWR_REQON_SHIFT;
|
|
|
|
/* Return if requested power request is already set */
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
|
|
if ((r & mask) == val) {
|
|
return r;
|
|
}
|
|
|
|
r = (r & ~mask) | val;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
W_REG(sii->osh, fast_srpwr_addr, r);
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
si_corereg_pciefast_write(sih, offset, r);
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
|
|
if (val2) {
|
|
if ((r & (mask2 << SRPWR_STATUS_SHIFT)) ==
|
|
(val2 << SRPWR_STATUS_SHIFT)) {
|
|
return r;
|
|
}
|
|
si_srpwr_stat_spinwait(sih, mask2, val2);
|
|
}
|
|
} else {
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
#ifdef CORE_PWRUP_WAR
|
|
uint32
|
|
BCMPOSTTRAPFN(si_srpwr_request_on_rev80)(si_t *sih, uint32 mask, uint32 val, uint32 ucode_awake)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 r, offset = OFFSETOF(chipcregs_t, powerctl); /* Same 0x1e8 per core */
|
|
uint cidx = (BUSTYPE(sih->bustype) == SI_BUS) ? SI_CC_IDX : sih->buscoreidx;
|
|
uint32 mask2 = mask;
|
|
uint32 val2 = val;
|
|
volatile uint32 *fast_srpwr_addr = (volatile uint32 *)((uintptr)SI_ENUM_BASE(sih)
|
|
+ (uintptr)offset);
|
|
if (mask || val) {
|
|
mask <<= SRPWR_REQON_SHIFT;
|
|
val <<= SRPWR_REQON_SHIFT;
|
|
|
|
/* Return if requested power request is already set */
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
r = si_corereg(sih, cidx, offset, 0, 0);
|
|
}
|
|
|
|
if ((r & mask) == val) {
|
|
W_REG(sii->osh, fast_srpwr_addr, r);
|
|
return r;
|
|
}
|
|
|
|
r = (r & ~mask) | val;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
W_REG(sii->osh, fast_srpwr_addr, r);
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
r = si_corereg(sih, cidx, offset, ~0, r);
|
|
}
|
|
|
|
if (val2) {
|
|
|
|
/*
|
|
* When ucode is not requested to be awake by FW,
|
|
* the power status may indicate ON due to FW or
|
|
* ucode's earlier power down request is not
|
|
* honored yet. In such case, FW will find the
|
|
* power status high at this stage, but as it is in
|
|
* transition (from ON to OFF), it may go down any
|
|
* time and lead to AXI slave error. Hence we need
|
|
* a fixed delay to cross any such transition state.
|
|
*/
|
|
if (ucode_awake == 0) {
|
|
hnd_delay(SRPWR_UP_DOWN_DELAY);
|
|
}
|
|
|
|
if ((r & (mask2 << SRPWR_STATUS_SHIFT)) ==
|
|
(val2 << SRPWR_STATUS_SHIFT)) {
|
|
return r;
|
|
}
|
|
si_srpwr_stat_spinwait(sih, mask2, val2);
|
|
}
|
|
} else {
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = R_REG(sii->osh, fast_srpwr_addr);
|
|
} else {
|
|
r = si_corereg(sih, cidx, offset, 0, 0);
|
|
}
|
|
SPINWAIT(((R_REG(sii->osh, fast_srpwr_addr) &
|
|
(mask2 << SRPWR_REQON_SHIFT)) != 0),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
#endif /* CORE_PWRUP_WAR */
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_srpwr_stat_spinwait)(const si_t *sih, uint32 mask, uint32 val)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 r, offset = (BUSTYPE(sih->bustype) == SI_BUS) ?
|
|
OFFSETOF(chipcregs_t, powerctl) : PWRREQ_OFFSET(sih);
|
|
volatile uint32 *fast_srpwr_addr = (volatile uint32 *)((uintptr)SI_ENUM_BASE(sih)
|
|
+ (uintptr)offset);
|
|
|
|
ASSERT(mask);
|
|
ASSERT(val);
|
|
|
|
/* spinwait on pwrstatus */
|
|
mask <<= SRPWR_STATUS_SHIFT;
|
|
val <<= SRPWR_STATUS_SHIFT;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
SPINWAIT(((R_REG(sii->osh, fast_srpwr_addr) & mask) != val),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
r = R_REG(sii->osh, fast_srpwr_addr) & mask;
|
|
ASSERT(r == val);
|
|
} else {
|
|
SPINWAIT(((si_corereg_pciefast_read(sih, offset) & mask) != val),
|
|
PMU_MAX_TRANSITION_DLY);
|
|
r = si_corereg_pciefast_read(sih, offset) & mask;
|
|
ASSERT(r == val);
|
|
}
|
|
|
|
r = (r >> SRPWR_STATUS_SHIFT) & SRPWR_DMN_ALL_MASK(sih);
|
|
|
|
return r;
|
|
}
|
|
|
|
uint32
|
|
si_srpwr_stat(si_t *sih)
|
|
{
|
|
uint32 r, offset = (BUSTYPE(sih->bustype) == SI_BUS) ?
|
|
OFFSETOF(chipcregs_t, powerctl) : PWRREQ_OFFSET(sih);
|
|
uint cidx = (BUSTYPE(sih->bustype) == SI_BUS) ? SI_CC_IDX : sih->buscoreidx;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = si_corereg(sih, cidx, offset, 0, 0);
|
|
} else {
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
|
|
r = (r >> SRPWR_STATUS_SHIFT) & SRPWR_DMN_ALL_MASK(sih);
|
|
|
|
return r;
|
|
}
|
|
|
|
uint32
|
|
si_srpwr_domain(si_t *sih)
|
|
{
|
|
uint32 r, offset = (BUSTYPE(sih->bustype) == SI_BUS) ?
|
|
OFFSETOF(chipcregs_t, powerctl) : PWRREQ_OFFSET(sih);
|
|
uint cidx = (BUSTYPE(sih->bustype) == SI_BUS) ? SI_CC_IDX : sih->buscoreidx;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = si_corereg(sih, cidx, offset, 0, 0);
|
|
} else {
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
|
|
r = (r >> SRPWR_DMN_ID_SHIFT) & SRPWR_DMN_ID_MASK;
|
|
|
|
return r;
|
|
}
|
|
|
|
uint8
|
|
si_srpwr_domain_wl(si_t *sih)
|
|
{
|
|
return SRPWR_DMN1_ARMBPSD;
|
|
}
|
|
|
|
bool
|
|
si_srpwr_cap(si_t *sih)
|
|
{
|
|
/* If domain ID is non-zero, chip supports power domain control */
|
|
return si_srpwr_domain(sih) != 0 ? TRUE : FALSE;
|
|
}
|
|
|
|
uint32
|
|
BCMPOSTTRAPFN(si_srpwr_domain_all_mask)(const si_t *sih)
|
|
{
|
|
uint32 mask = SRPWR_DMN0_PCIE_MASK |
|
|
SRPWR_DMN1_ARMBPSD_MASK |
|
|
SRPWR_DMN2_MACAUX_MASK |
|
|
SRPWR_DMN3_MACMAIN_MASK;
|
|
|
|
if (si_scan_core_present(sih)) {
|
|
mask |= SRPWR_DMN4_MACSCAN_MASK;
|
|
}
|
|
|
|
return mask;
|
|
}
|
|
|
|
uint32
|
|
si_srpwr_bt_status(si_t *sih)
|
|
{
|
|
uint32 r;
|
|
uint32 offset = (BUSTYPE(sih->bustype) == SI_BUS) ?
|
|
OFFSETOF(chipcregs_t, powerctl) : PWRREQ_OFFSET(sih);
|
|
uint32 cidx = (BUSTYPE(sih->bustype) == SI_BUS) ? SI_CC_IDX : sih->buscoreidx;
|
|
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
r = si_corereg(sih, cidx, offset, 0, 0);
|
|
} else {
|
|
r = si_corereg_pciefast_read(sih, offset);
|
|
}
|
|
|
|
r = (r >> SRPWR_BT_STATUS_SHIFT) & SRPWR_BT_STATUS_MASK;
|
|
|
|
return r;
|
|
}
|
|
/* Utility API to read/write the raw registers with absolute address.
|
|
* This function can be invoked from either FW or host driver.
|
|
*/
|
|
uint32
|
|
si_raw_reg(const si_t *sih, uint32 reg, uint32 val, uint32 wrire_req)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
uint32 address_space = reg & ~0xFFF;
|
|
volatile uint32 * addr = (void*)(uintptr)(reg);
|
|
uint32 prev_value = 0;
|
|
uint32 cfg_reg = 0;
|
|
|
|
if (sii == NULL) {
|
|
return 0;
|
|
}
|
|
|
|
/* No need to translate the absolute address on SI bus */
|
|
if (BUSTYPE(sih->bustype) == SI_BUS) {
|
|
goto skip_cfg;
|
|
}
|
|
|
|
/* This API supports only the PCI host interface */
|
|
if (BUSTYPE(sih->bustype) != PCI_BUS) {
|
|
return ID32_INVALID;
|
|
}
|
|
|
|
if (PCIE_GEN2(sii)) {
|
|
/* Use BAR0 Secondary window is PCIe Gen2.
|
|
* Set the secondary BAR0 Window to current register of interest
|
|
*/
|
|
addr = (volatile uint32*)(((volatile uint8*)sii->curmap) +
|
|
PCI_SEC_BAR0_WIN_OFFSET + (reg & 0xfff));
|
|
cfg_reg = PCIE2_BAR0_CORE2_WIN;
|
|
|
|
} else {
|
|
/* PCIe Gen1 do not have secondary BAR0 window.
|
|
* reuse the BAR0 WIN2
|
|
*/
|
|
addr = (volatile uint32*)(((volatile uint8*)sii->curmap) +
|
|
PCI_BAR0_WIN2_OFFSET + (reg & 0xfff));
|
|
cfg_reg = PCI_BAR0_WIN2;
|
|
}
|
|
|
|
prev_value = OSL_PCI_READ_CONFIG(sii->osh, cfg_reg, 4);
|
|
|
|
if (prev_value != address_space) {
|
|
OSL_PCI_WRITE_CONFIG(sii->osh, cfg_reg,
|
|
sizeof(uint32), address_space);
|
|
} else {
|
|
prev_value = 0;
|
|
}
|
|
|
|
skip_cfg:
|
|
if (wrire_req) {
|
|
W_REG(sii->osh, addr, val);
|
|
} else {
|
|
val = R_REG(sii->osh, addr);
|
|
}
|
|
|
|
if (prev_value) {
|
|
/* Restore BAR0 WIN2 for PCIE GEN1 devices */
|
|
OSL_PCI_WRITE_CONFIG(sii->osh,
|
|
cfg_reg, sizeof(uint32), prev_value);
|
|
}
|
|
|
|
return val;
|
|
}
|
|
|
|
#ifdef DONGLEBUILD
|
|
/* if the logs could be gathered, host could be notified with to take logs or not */
|
|
bool
|
|
BCMPOSTTRAPFN(si_check_enable_backplane_log)(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
return ai_check_enable_backplane_log(sih);
|
|
}
|
|
return TRUE;
|
|
}
|
|
#endif /* DONGLEBUILD */
|
|
|
|
uint8
|
|
si_lhl_ps_mode(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->lhl_ps_mode;
|
|
}
|
|
|
|
uint8
|
|
si_hib_ext_wakeup_isenab(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
return sii->hib_ext_wakeup_enab;
|
|
}
|
|
|
|
static void
|
|
si_oob_war_BT_F1(si_t *sih)
|
|
{
|
|
uint origidx = si_coreidx(sih);
|
|
volatile void *regs;
|
|
|
|
regs = si_setcore(sih, AXI2AHB_BRIDGE_ID, 0);
|
|
ASSERT(regs);
|
|
BCM_REFERENCE(regs);
|
|
|
|
si_wrapperreg(sih, AI_OOBSELINA30, 0xF00, 0x300);
|
|
|
|
si_setcoreidx(sih, origidx);
|
|
}
|
|
|
|
#ifndef BCMDONGLEHOST
|
|
|
|
#define RF_SW_CTRL_ELNABYP_ANT_MASK 0x000CC330
|
|
|
|
/* These are the outputs to the rfem which go out via the CLB */
|
|
#define RF_SW_CTRL_ELNABYP_2G0_MASK 0x00000010
|
|
#define RF_SW_CTRL_ELNABYP_5G0_MASK 0x00000020
|
|
#define RF_SW_CTRL_ELNABYP_2G1_MASK 0x00004000
|
|
#define RF_SW_CTRL_ELNABYP_5G1_MASK 0x00008000
|
|
|
|
/* Feedback values go into the phy from CLB output
|
|
* The polarity of the feedback is opposite to the elnabyp signal going out to the rfem
|
|
*/
|
|
#define RF_SW_CTRL_ELNABYP_2G0_MASK_FB 0x00000100
|
|
#define RF_SW_CTRL_ELNABYP_5G0_MASK_FB 0x00000200
|
|
#define RF_SW_CTRL_ELNABYP_2G1_MASK_FB 0x00040000
|
|
#define RF_SW_CTRL_ELNABYP_5G1_MASK_FB 0x00080000
|
|
|
|
/* The elnabyp override values for each rfem */
|
|
#define ELNABYP_IOVAR_2G0_VALUE_MASK 0x01
|
|
#define ELNABYP_IOVAR_5G0_VALUE_MASK 0x02
|
|
#define ELNABYP_IOVAR_2G1_VALUE_MASK 0x04
|
|
#define ELNABYP_IOVAR_5G1_VALUE_MASK 0x08
|
|
|
|
/* The elnabyp override enables for each rfem
|
|
* The values are 'don't care' if the corresponding enables are 0
|
|
*/
|
|
#define ELNABYP_IOVAR_2G0_ENABLE_MASK 0x10
|
|
#define ELNABYP_IOVAR_5G0_ENABLE_MASK 0x20
|
|
#define ELNABYP_IOVAR_2G1_ENABLE_MASK 0x40
|
|
#define ELNABYP_IOVAR_5G1_ENABLE_MASK 0x80
|
|
|
|
#define ANTENNA_0_ENABLE 0x00000044
|
|
#define ANTENNA_1_ENABLE 0x20000000
|
|
#define RFFE_CTRL_START 0x80000000
|
|
#define RFFE_CTRL_READ 0x40000000
|
|
#define RFFE_CTRL_RFEM_SEL 0x08000000
|
|
#define RFFE_MISC_EN_PHYCYCLES 0x00000002
|
|
|
|
void
|
|
si_rffe_rfem_init(si_t *sih)
|
|
{
|
|
ASSERT(GCI_OFFSETOF(sih, gci_chipctrl) == OFFSETOF(gciregs_t, gci_chipctrl));
|
|
/* Enable RFFE clock
|
|
* GCI Chip Control reg 15 - Bits 29 & 30 (Global 509 & 510)
|
|
*/
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_15, ALLONES_32, 0x60000000);
|
|
/* SDATA0/1 rf_sw_ctrl pull down
|
|
* GCI chip control reg 23 - Bits 29 & 30 (Global 765 & 766)
|
|
*/
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_23, 0x3 << 29, 0x3 << 29);
|
|
/* RFFE Clk Ctrl Reg */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_clk_ctrl), ALLONES_32, 0x101);
|
|
|
|
/* Disable override control of RFFE controller and enable phy control */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_change_detect_ovr_wlmc), ALLONES_32, 0);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_change_detect_ovr_wlac), ALLONES_32, 0);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_change_detect_ovr_wlsc), ALLONES_32, 0);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_change_detect_ovr_btmc), ALLONES_32, 0);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_change_detect_ovr_btsc), ALLONES_32, 0);
|
|
|
|
/* reg address = 0x16, deviceID of rffe dev1 = 0xE, deviceID of dev0 = 0xC,
|
|
* last_mux_ctrl = 0, disable_preemption = 0 (1 for 4387b0), tssi_mask = 3, tssi_en = 0,
|
|
* rffe_disable_line1 = 0, enable rffe_en_phyaccess = 1,
|
|
* disable BRCM proprietary reg0 wr = 0
|
|
*/
|
|
if (sih->ccrev == 68) {
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_misc_ctrl), ALLONES_32, 0x0016EC72);
|
|
} else {
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_misc_ctrl), ALLONES_32, 0x0016EC32);
|
|
}
|
|
|
|
/* Enable Dual RFFE Master: rffe_single_master = 0
|
|
* Use Master0 SW interface only : rffe_dis_sw_intf_m1 = 1
|
|
*/
|
|
if (sih->ccrev >= 71) {
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_clk_ctrl),
|
|
1u << 20u | 1u << 26u, 1u << 26u);
|
|
}
|
|
|
|
/* Enable antenna access for both cores */
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, ALLONES_32, ANTENNA_0_ENABLE);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, ALLONES_32, ANTENNA_1_ENABLE);
|
|
}
|
|
|
|
void
|
|
si_rffe_set_debug_mode(si_t *sih, bool enable)
|
|
{
|
|
uint32 misc_ctrl_set = 0;
|
|
/* Enable/Disable rffe_en_phyaccess bit */
|
|
if (!enable) {
|
|
misc_ctrl_set = RFFE_MISC_EN_PHYCYCLES;
|
|
}
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_misc_ctrl), RFFE_MISC_EN_PHYCYCLES,
|
|
misc_ctrl_set);
|
|
|
|
sih->rffe_debug_mode = enable;
|
|
}
|
|
|
|
bool
|
|
si_rffe_get_debug_mode(si_t *sih)
|
|
{
|
|
return sih->rffe_debug_mode;
|
|
}
|
|
|
|
int8
|
|
si_rffe_get_elnabyp_mode(si_t *sih)
|
|
{
|
|
return sih->rffe_elnabyp_mode;
|
|
}
|
|
|
|
int
|
|
si_rffe_set_elnabyp_mode(si_t *sih, uint8 mode)
|
|
{
|
|
int ret = BCME_OK;
|
|
uint32 elnabyp_ovr_val = 0;
|
|
uint32 elnabyp_ovr_en = 0;
|
|
|
|
if ((mode & ELNABYP_IOVAR_2G0_VALUE_MASK) && (mode & ELNABYP_IOVAR_2G0_ENABLE_MASK)) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_2G0_MASK;
|
|
} else if (mode & ELNABYP_IOVAR_2G0_ENABLE_MASK) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_2G0_MASK_FB;
|
|
}
|
|
if ((mode & ELNABYP_IOVAR_5G0_VALUE_MASK) && (mode & ELNABYP_IOVAR_5G0_ENABLE_MASK)) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_5G0_MASK;
|
|
} else if (mode & ELNABYP_IOVAR_5G0_ENABLE_MASK) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_5G0_MASK_FB;
|
|
}
|
|
if ((mode & ELNABYP_IOVAR_2G1_VALUE_MASK) && (mode & ELNABYP_IOVAR_2G1_ENABLE_MASK)) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_2G1_MASK;
|
|
} else if (mode & ELNABYP_IOVAR_2G1_ENABLE_MASK) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_2G1_MASK_FB;
|
|
}
|
|
if ((mode & ELNABYP_IOVAR_5G1_VALUE_MASK) && (mode & ELNABYP_IOVAR_5G1_ENABLE_MASK)) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_5G1_MASK;
|
|
} else if (mode & ELNABYP_IOVAR_5G1_ENABLE_MASK) {
|
|
elnabyp_ovr_val |= RF_SW_CTRL_ELNABYP_5G1_MASK_FB;
|
|
}
|
|
|
|
if (mode & ELNABYP_IOVAR_2G0_ENABLE_MASK) {
|
|
elnabyp_ovr_en |= (RF_SW_CTRL_ELNABYP_2G0_MASK | RF_SW_CTRL_ELNABYP_2G0_MASK_FB);
|
|
}
|
|
if (mode & ELNABYP_IOVAR_5G0_ENABLE_MASK) {
|
|
elnabyp_ovr_en |= (RF_SW_CTRL_ELNABYP_5G0_MASK | RF_SW_CTRL_ELNABYP_5G0_MASK_FB);
|
|
}
|
|
if (mode & ELNABYP_IOVAR_2G1_ENABLE_MASK) {
|
|
elnabyp_ovr_en |= (RF_SW_CTRL_ELNABYP_2G1_MASK | RF_SW_CTRL_ELNABYP_2G1_MASK_FB);
|
|
}
|
|
if (mode & ELNABYP_IOVAR_5G1_ENABLE_MASK) {
|
|
elnabyp_ovr_en |= (RF_SW_CTRL_ELNABYP_5G1_MASK | RF_SW_CTRL_ELNABYP_5G1_MASK_FB);
|
|
}
|
|
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_14, RF_SW_CTRL_ELNABYP_ANT_MASK,
|
|
elnabyp_ovr_val);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_15, RF_SW_CTRL_ELNABYP_ANT_MASK,
|
|
elnabyp_ovr_en);
|
|
|
|
sih->rffe_elnabyp_mode = mode;
|
|
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
BCMPOSTTRAPFN(si_rffe_rfem_read)(si_t *sih, uint8 dev_id, uint8 antenna, uint16 reg_addr,
|
|
uint32 *val)
|
|
{
|
|
int ret = BCME_OK;
|
|
uint32 gci_rffe_ctrl_val, antenna_0_enable, antenna_1_enable;
|
|
uint32 gci_rffe_ctrl = si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0);
|
|
uint32 gci_chipcontrol_03 = si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, 0, 0);
|
|
uint32 gci_chipcontrol_02 = si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, 0, 0);
|
|
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), ALLONES_32, 0);
|
|
|
|
switch (antenna) {
|
|
case 1:
|
|
gci_rffe_ctrl_val = RFFE_CTRL_START | RFFE_CTRL_READ;
|
|
antenna_0_enable = ANTENNA_0_ENABLE;
|
|
antenna_1_enable = 0;
|
|
break;
|
|
case 2:
|
|
gci_rffe_ctrl_val = RFFE_CTRL_START | RFFE_CTRL_READ | RFFE_CTRL_RFEM_SEL;
|
|
antenna_0_enable = 0;
|
|
antenna_1_enable = ANTENNA_1_ENABLE;
|
|
break;
|
|
default:
|
|
ret = BCME_BADOPTION;
|
|
}
|
|
|
|
if (ret == BCME_OK) {
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_config), ALLONES_32,
|
|
((uint16) dev_id) << 8);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_rfem_addr), ALLONES_32, reg_addr);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, ANTENNA_0_ENABLE, antenna_0_enable);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, ANTENNA_1_ENABLE, antenna_1_enable);
|
|
/* Initiate read */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl),
|
|
RFFE_CTRL_START | RFFE_CTRL_READ | RFFE_CTRL_RFEM_SEL, gci_rffe_ctrl_val);
|
|
/* Wait for status */
|
|
SPINWAIT(si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0) &
|
|
RFFE_CTRL_START, 100);
|
|
if (si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0) &
|
|
RFFE_CTRL_START) {
|
|
OSL_SYS_HALT();
|
|
ret = BCME_NOTREADY;
|
|
} else {
|
|
*val = si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_rfem_data0), 0, 0);
|
|
/* Clear read and rfem_sel flags */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl),
|
|
RFFE_CTRL_READ | RFFE_CTRL_RFEM_SEL, 0);
|
|
}
|
|
}
|
|
|
|
/* Restore the values */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), ALLONES_32, gci_rffe_ctrl);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, ALLONES_32, gci_chipcontrol_03);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, ALLONES_32, gci_chipcontrol_02);
|
|
return ret;
|
|
}
|
|
|
|
int
|
|
BCMPOSTTRAPFN(si_rffe_rfem_write)(si_t *sih, uint8 dev_id, uint8 antenna, uint16 reg_addr,
|
|
uint32 data)
|
|
{
|
|
int ret = BCME_OK;
|
|
uint32 antenna_0_enable, antenna_1_enable;
|
|
uint32 gci_rffe_ctrl = si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0);
|
|
uint32 gci_chipcontrol_03 = si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, 0, 0);
|
|
uint32 gci_chipcontrol_02 = si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, 0, 0);
|
|
uint8 repeat = (sih->ccrev == 69) ? 2 : 1; /* WAR for 4387c0 */
|
|
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), ALLONES_32, 0);
|
|
|
|
switch (antenna) {
|
|
case 1:
|
|
antenna_0_enable = ANTENNA_0_ENABLE;
|
|
antenna_1_enable = 0;
|
|
break;
|
|
case 2:
|
|
antenna_0_enable = 0;
|
|
antenna_1_enable = ANTENNA_1_ENABLE;
|
|
break;
|
|
case 3:
|
|
antenna_0_enable = ANTENNA_0_ENABLE;
|
|
antenna_1_enable = ANTENNA_1_ENABLE;
|
|
break;
|
|
default:
|
|
ret = BCME_BADOPTION;
|
|
}
|
|
|
|
if (ret == BCME_OK) {
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_config), ALLONES_32,
|
|
((uint16) dev_id) << 8);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_rfem_addr), ALLONES_32, reg_addr);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, ANTENNA_0_ENABLE, antenna_0_enable);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, ANTENNA_1_ENABLE, antenna_1_enable);
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_rfem_data0), ALLONES_32, data);
|
|
while (repeat--) {
|
|
/* Initiate write */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), RFFE_CTRL_START |
|
|
RFFE_CTRL_READ | RFFE_CTRL_RFEM_SEL, RFFE_CTRL_START);
|
|
/* Wait for status */
|
|
SPINWAIT(si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0) &
|
|
RFFE_CTRL_START, 100);
|
|
if (si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), 0, 0) &
|
|
RFFE_CTRL_START) {
|
|
OSL_SYS_HALT();
|
|
ret = BCME_NOTREADY;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Restore the values */
|
|
si_gci_direct(sih, OFFSETOF(gciregs_t, gci_rffe_ctrl), ALLONES_32, gci_rffe_ctrl);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_03, ALLONES_32, gci_chipcontrol_03);
|
|
si_gci_chipcontrol(sih, CC_GCI_CHIPCTRL_02, ALLONES_32, gci_chipcontrol_02);
|
|
return ret;
|
|
}
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
#if defined(BCMSDIODEV_ENABLED) && defined(ATE_BUILD)
|
|
bool
|
|
si_chipcap_sdio_ate_only(const si_t *sih)
|
|
{
|
|
bool ate_build = FALSE;
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4369_CHIP_GRPID:
|
|
if (CST4369_CHIPMODE_SDIOD(sih->chipst) &&
|
|
CST4369_CHIPMODE_PCIE(sih->chipst)) {
|
|
ate_build = TRUE;
|
|
}
|
|
break;
|
|
case BCM4376_CHIP_GRPID:
|
|
case BCM4378_CHIP_GRPID:
|
|
case BCM4381_CHIP_GRPID:
|
|
case BCM4383_CHIP_GRPID:
|
|
case BCM4385_CHIP_GRPID:
|
|
case BCM4387_CHIP_GRPID:
|
|
case BCM4388_CHIP_GRPID:
|
|
case BCM4389_CHIP_GRPID:
|
|
case BCM4397_CHIP_GRPID:
|
|
ate_build = TRUE;
|
|
break;
|
|
case BCM4362_CHIP_GRPID:
|
|
if (CST4362_CHIPMODE_SDIOD(sih->chipst) &&
|
|
CST4362_CHIPMODE_PCIE(sih->chipst)) {
|
|
ate_build = TRUE;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
return ate_build;
|
|
}
|
|
#endif /* BCMSDIODEV_ENABLED && ATE_BUILD */
|
|
|
|
#ifdef UART_TRAP_DBG
|
|
void
|
|
si_dump_APB_Bridge_registers(const si_t *sih)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
ai_dump_APB_Bridge_registers(sih);
|
|
}
|
|
}
|
|
#endif /* UART_TRAP_DBG */
|
|
|
|
void
|
|
si_force_clocks(const si_t *sih, uint clock_state)
|
|
{
|
|
if (CHIPTYPE(sih->socitype) == SOCI_AI) {
|
|
ai_force_clocks(sih, clock_state);
|
|
}
|
|
}
|
|
|
|
/* Indicates to the siutils how the PICe BAR0 is mappend,
|
|
* used for siutils to arrange BAR0 window management,
|
|
* for PCI NIC driver.
|
|
*
|
|
* Here is the current scheme, which are all using BAR0:
|
|
*
|
|
* id enum wrapper
|
|
* ==== ========= =========
|
|
* 0 0000-0FFF 1000-1FFF
|
|
* 1 4000-4FFF 5000-5FFF
|
|
* 2 9000-9FFF A000-AFFF
|
|
* >= 3 not supported
|
|
*/
|
|
void
|
|
si_set_slice_id(si_t *sih, uint8 slice)
|
|
{
|
|
si_info_t *sii = SI_INFO(sih);
|
|
|
|
sii->slice = slice;
|
|
}
|
|
|
|
uint8
|
|
si_get_slice_id(const si_t *sih)
|
|
{
|
|
const si_info_t *sii = SI_INFO(sih);
|
|
|
|
return sii->slice;
|
|
}
|
|
|
|
bool
|
|
BCMPOSTTRAPRAMFN(si_scan_core_present)(const si_t *sih)
|
|
{
|
|
return (si_numcoreunits(sih, D11_CORE_ID) > 2);
|
|
}
|
|
|
|
#if !defined(BCMDONGLEHOST)
|
|
bool
|
|
si_btc_bt_status_in_reset(si_t *sih)
|
|
{
|
|
uint32 chipst = 0;
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4387_CHIP_GRPID:
|
|
chipst = si_corereg(sih, SI_CC_IDX,
|
|
OFFSETOF(chipcregs_t, chipstatus), 0, 0);
|
|
/* 1 =bt in reset 0 = bt out of reset */
|
|
return (chipst & (1 << BT_IN_RESET_BIT_SHIFT)) ? TRUE : FALSE;
|
|
break;
|
|
default:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
bool
|
|
si_btc_bt_status_in_pds(si_t *sih)
|
|
{
|
|
return !((si_gci_chipstatus(sih, GCI_CHIPSTATUS_04) >>
|
|
BT_IN_PDS_BIT_SHIFT) & 0x1);
|
|
}
|
|
|
|
int
|
|
si_btc_bt_pds_wakeup_force(si_t *sih, bool force)
|
|
{
|
|
if (force) {
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL0,
|
|
PMU_CC0_4387_BT_PU_WAKE_MASK, PMU_CC0_4387_BT_PU_WAKE_MASK);
|
|
SPINWAIT((si_btc_bt_status_in_pds(sih) == TRUE), PMU_MAX_TRANSITION_DLY);
|
|
if (si_btc_bt_status_in_pds(sih) == TRUE) {
|
|
SI_ERROR(("si_btc_bt_pds_wakeup_force"
|
|
" ERROR : BT still in PDS after pds_wakeup_force \n"));
|
|
return BCME_ERROR;
|
|
} else {
|
|
return BCME_OK;
|
|
}
|
|
} else {
|
|
si_pmu_chipcontrol(sih, PMU_CHIPCTL0, PMU_CC0_4387_BT_PU_WAKE_MASK, 0);
|
|
return BCME_OK;
|
|
}
|
|
}
|
|
|
|
#endif /* !defined(BCMDONGLEHOST) */
|
|
|
|
#ifndef BCMDONGLEHOST
|
|
/* query d11 core type */
|
|
uint
|
|
si_core_d11_type(si_t *sih, uint coreunit)
|
|
{
|
|
#ifdef WL_SCAN_CORE_SIM
|
|
/* use the core unit WL_SCAN_CORE_SIM as the scan core */
|
|
return (coreunit == WL_SCAN_CORE_SIM) ?
|
|
D11_CORE_TYPE_SCAN : D11_CORE_TYPE_NORM;
|
|
#else
|
|
uint coreidx;
|
|
volatile void *regs;
|
|
uint coretype;
|
|
|
|
coreidx = si_coreidx(sih);
|
|
regs = si_setcore(sih, D11_CORE_ID, coreunit);
|
|
ASSERT(regs != NULL);
|
|
BCM_REFERENCE(regs);
|
|
|
|
coretype = (si_core_sflags(sih, 0, 0) & SISF_CORE_BITS_SCAN) != 0 ?
|
|
D11_CORE_TYPE_SCAN : D11_CORE_TYPE_NORM;
|
|
|
|
si_setcoreidx(sih, coreidx);
|
|
return coretype;
|
|
#endif /* WL_SCAN_CORE_SIM */
|
|
}
|
|
|
|
/* decide if this core is allowed by the package option or not... */
|
|
bool
|
|
si_pkgopt_d11_allowed(si_t *sih, uint coreunit)
|
|
{
|
|
uint coreidx;
|
|
volatile void *regs;
|
|
bool allowed;
|
|
|
|
coreidx = si_coreidx(sih);
|
|
regs = si_setcore(sih, D11_CORE_ID, coreunit);
|
|
ASSERT(regs != NULL);
|
|
BCM_REFERENCE(regs);
|
|
|
|
allowed = ((si_core_sflags(sih, 0, 0) & SISF_CORE_BITS_SCAN) == 0 ||
|
|
(si_gci_chipstatus(sih, GCI_CHIPSTATUS_09) & GCI_CST9_SCAN_DIS) == 0);
|
|
|
|
si_setcoreidx(sih, coreidx);
|
|
return allowed;
|
|
}
|
|
|
|
void
|
|
si_configure_pwrthrottle_gpio(si_t *sih, uint8 pwrthrottle_gpio_in)
|
|
{
|
|
uint32 board_gpio = 0;
|
|
if (CHIPID(sih->chip) == BCM4369_CHIP_ID || CHIPID(sih->chip) == BCM4377_CHIP_ID) {
|
|
si_gci_set_functionsel(sih, pwrthrottle_gpio_in, 1);
|
|
}
|
|
board_gpio = 1 << pwrthrottle_gpio_in;
|
|
si_gpiocontrol(sih, board_gpio, 0, GPIO_DRV_PRIORITY);
|
|
si_gpioouten(sih, board_gpio, 0, GPIO_DRV_PRIORITY);
|
|
}
|
|
|
|
void
|
|
si_configure_onbody_gpio(si_t *sih, uint8 onbody_gpio_in)
|
|
{
|
|
uint32 board_gpio = 0;
|
|
if (CHIPID(sih->chip) == BCM4369_CHIP_ID || CHIPID(sih->chip) == BCM4377_CHIP_ID) {
|
|
si_gci_set_functionsel(sih, onbody_gpio_in, 1);
|
|
}
|
|
board_gpio = 1 << onbody_gpio_in;
|
|
si_gpiocontrol(sih, board_gpio, 0, GPIO_DRV_PRIORITY);
|
|
si_gpioouten(sih, board_gpio, 0, GPIO_DRV_PRIORITY);
|
|
}
|
|
|
|
#endif /* !BCMDONGLEHOST */
|
|
|
|
void
|
|
si_jtag_udr_pwrsw_main_toggle(si_t *sih, bool on)
|
|
{
|
|
#ifdef DONGLEBUILD
|
|
int val = on ? 0 : 1;
|
|
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4387_CHIP_GRPID:
|
|
jtag_setbit_128(sih, 8, 99, val);
|
|
jtag_setbit_128(sih, 8, 101, val);
|
|
jtag_setbit_128(sih, 8, 105, val);
|
|
break;
|
|
default:
|
|
SI_ERROR(("si_jtag_udr_pwrsw_main_toggle: add support for this chip!\n"));
|
|
OSL_SYS_HALT();
|
|
break;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/* return the backplane address where the sssr dumps are stored per D11 core */
|
|
uint32
|
|
si_d11_core_sssr_addr(si_t *sih, uint unit, uint32 *sssr_size)
|
|
{
|
|
uint32 sssr_dmp_src = 0;
|
|
*sssr_size = 0;
|
|
/* ideally these addresses should be grok'ed from EROM map */
|
|
switch (CHIPID(sih->chip)) {
|
|
case BCM4387_CHIP_GRPID:
|
|
if (unit == 0) {
|
|
sssr_dmp_src = BCM4387_SSSR_DUMP_AXI_MAIN;
|
|
*sssr_size = (uint32)BCM4387_SSSR_DUMP_MAIN_SIZE;
|
|
} else if (unit == 1) {
|
|
sssr_dmp_src = BCM4387_SSSR_DUMP_AXI_AUX;
|
|
*sssr_size = (uint32)BCM4387_SSSR_DUMP_AUX_SIZE;
|
|
} else if (unit == 2) {
|
|
sssr_dmp_src = BCM4387_SSSR_DUMP_AXI_SCAN;
|
|
*sssr_size = (uint32)BCM4387_SSSR_DUMP_SCAN_SIZE;
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
return (sssr_dmp_src);
|
|
}
|
|
|
|
#ifdef USE_LHL_TIMER
|
|
/* Get current HIB time API */
|
|
uint32
|
|
si_cur_hib_time(si_t *sih)
|
|
{
|
|
uint32 hib_time;
|
|
|
|
hib_time = LHL_REG(sih, lhl_hibtim_adr, 0, 0);
|
|
|
|
/* there is no HW sync on the read path for LPO regs,
|
|
* so SW should read twice and if values are same,
|
|
* then use the value, else read again and use the
|
|
* latest value
|
|
*/
|
|
if (hib_time != LHL_REG(sih, lhl_hibtim_adr, 0, 0)) {
|
|
hib_time = LHL_REG(sih, lhl_hibtim_adr, 0, 0);
|
|
}
|
|
|
|
return (hib_time);
|
|
}
|
|
#endif /* USE_LHL_TIMER */
|