linux/drivers/bcma/host_pci.c
<<
>>
Prefs
   1/*
   2 * Broadcom specific AMBA
   3 * PCI Host
   4 *
   5 * Licensed under the GNU/GPL. See COPYING for details.
   6 */
   7
   8#include "bcma_private.h"
   9#include <linux/slab.h>
  10#include <linux/bcma/bcma.h>
  11#include <linux/pci.h>
  12#include <linux/module.h>
  13
  14static void bcma_host_pci_switch_core(struct bcma_device *core)
  15{
  16        int win2 = core->bus->host_is_pcie2 ?
  17                BCMA_PCIE2_BAR0_WIN2 : BCMA_PCI_BAR0_WIN2;
  18
  19        pci_write_config_dword(core->bus->host_pci, BCMA_PCI_BAR0_WIN,
  20                               core->addr);
  21        pci_write_config_dword(core->bus->host_pci, win2, core->wrap);
  22        core->bus->mapped_core = core;
  23        bcma_debug(core->bus, "Switched to core: 0x%X\n", core->id.id);
  24}
  25
  26/* Provides access to the requested core. Returns base offset that has to be
  27 * used. It makes use of fixed windows when possible. */
  28static u16 bcma_host_pci_provide_access_to_core(struct bcma_device *core)
  29{
  30        switch (core->id.id) {
  31        case BCMA_CORE_CHIPCOMMON:
  32                return 3 * BCMA_CORE_SIZE;
  33        case BCMA_CORE_PCIE:
  34                return 2 * BCMA_CORE_SIZE;
  35        }
  36
  37        if (core->bus->mapped_core != core)
  38                bcma_host_pci_switch_core(core);
  39        return 0;
  40}
  41
  42static u8 bcma_host_pci_read8(struct bcma_device *core, u16 offset)
  43{
  44        offset += bcma_host_pci_provide_access_to_core(core);
  45        return ioread8(core->bus->mmio + offset);
  46}
  47
  48static u16 bcma_host_pci_read16(struct bcma_device *core, u16 offset)
  49{
  50        offset += bcma_host_pci_provide_access_to_core(core);
  51        return ioread16(core->bus->mmio + offset);
  52}
  53
  54static u32 bcma_host_pci_read32(struct bcma_device *core, u16 offset)
  55{
  56        offset += bcma_host_pci_provide_access_to_core(core);
  57        return ioread32(core->bus->mmio + offset);
  58}
  59
  60static void bcma_host_pci_write8(struct bcma_device *core, u16 offset,
  61                                 u8 value)
  62{
  63        offset += bcma_host_pci_provide_access_to_core(core);
  64        iowrite8(value, core->bus->mmio + offset);
  65}
  66
  67static void bcma_host_pci_write16(struct bcma_device *core, u16 offset,
  68                                 u16 value)
  69{
  70        offset += bcma_host_pci_provide_access_to_core(core);
  71        iowrite16(value, core->bus->mmio + offset);
  72}
  73
  74static void bcma_host_pci_write32(struct bcma_device *core, u16 offset,
  75                                 u32 value)
  76{
  77        offset += bcma_host_pci_provide_access_to_core(core);
  78        iowrite32(value, core->bus->mmio + offset);
  79}
  80
  81#ifdef CONFIG_BCMA_BLOCKIO
  82static void bcma_host_pci_block_read(struct bcma_device *core, void *buffer,
  83                                     size_t count, u16 offset, u8 reg_width)
  84{
  85        void __iomem *addr = core->bus->mmio + offset;
  86        if (core->bus->mapped_core != core)
  87                bcma_host_pci_switch_core(core);
  88        switch (reg_width) {
  89        case sizeof(u8):
  90                ioread8_rep(addr, buffer, count);
  91                break;
  92        case sizeof(u16):
  93                WARN_ON(count & 1);
  94                ioread16_rep(addr, buffer, count >> 1);
  95                break;
  96        case sizeof(u32):
  97                WARN_ON(count & 3);
  98                ioread32_rep(addr, buffer, count >> 2);
  99                break;
 100        default:
 101                WARN_ON(1);
 102        }
 103}
 104
 105static void bcma_host_pci_block_write(struct bcma_device *core,
 106                                      const void *buffer, size_t count,
 107                                      u16 offset, u8 reg_width)
 108{
 109        void __iomem *addr = core->bus->mmio + offset;
 110        if (core->bus->mapped_core != core)
 111                bcma_host_pci_switch_core(core);
 112        switch (reg_width) {
 113        case sizeof(u8):
 114                iowrite8_rep(addr, buffer, count);
 115                break;
 116        case sizeof(u16):
 117                WARN_ON(count & 1);
 118                iowrite16_rep(addr, buffer, count >> 1);
 119                break;
 120        case sizeof(u32):
 121                WARN_ON(count & 3);
 122                iowrite32_rep(addr, buffer, count >> 2);
 123                break;
 124        default:
 125                WARN_ON(1);
 126        }
 127}
 128#endif
 129
 130static u32 bcma_host_pci_aread32(struct bcma_device *core, u16 offset)
 131{
 132        if (core->bus->mapped_core != core)
 133                bcma_host_pci_switch_core(core);
 134        return ioread32(core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset);
 135}
 136
 137static void bcma_host_pci_awrite32(struct bcma_device *core, u16 offset,
 138                                  u32 value)
 139{
 140        if (core->bus->mapped_core != core)
 141                bcma_host_pci_switch_core(core);
 142        iowrite32(value, core->bus->mmio + (1 * BCMA_CORE_SIZE) + offset);
 143}
 144
 145static const struct bcma_host_ops bcma_host_pci_ops = {
 146        .read8          = bcma_host_pci_read8,
 147        .read16         = bcma_host_pci_read16,
 148        .read32         = bcma_host_pci_read32,
 149        .write8         = bcma_host_pci_write8,
 150        .write16        = bcma_host_pci_write16,
 151        .write32        = bcma_host_pci_write32,
 152#ifdef CONFIG_BCMA_BLOCKIO
 153        .block_read     = bcma_host_pci_block_read,
 154        .block_write    = bcma_host_pci_block_write,
 155#endif
 156        .aread32        = bcma_host_pci_aread32,
 157        .awrite32       = bcma_host_pci_awrite32,
 158};
 159
 160static int bcma_host_pci_probe(struct pci_dev *dev,
 161                               const struct pci_device_id *id)
 162{
 163        struct bcma_bus *bus;
 164        int err = -ENOMEM;
 165        const char *name;
 166        u32 val;
 167
 168        /* Alloc */
 169        bus = kzalloc(sizeof(*bus), GFP_KERNEL);
 170        if (!bus)
 171                goto out;
 172
 173        /* Basic PCI configuration */
 174        err = pci_enable_device(dev);
 175        if (err)
 176                goto err_kfree_bus;
 177
 178        name = dev_name(&dev->dev);
 179        if (dev->driver && dev->driver->name)
 180                name = dev->driver->name;
 181        err = pci_request_regions(dev, name);
 182        if (err)
 183                goto err_pci_disable;
 184        pci_set_master(dev);
 185
 186        /* Disable the RETRY_TIMEOUT register (0x41) to keep
 187         * PCI Tx retries from interfering with C3 CPU state */
 188        pci_read_config_dword(dev, 0x40, &val);
 189        if ((val & 0x0000ff00) != 0)
 190                pci_write_config_dword(dev, 0x40, val & 0xffff00ff);
 191
 192        /* SSB needed additional powering up, do we have any AMBA PCI cards? */
 193        if (!pci_is_pcie(dev)) {
 194                bcma_err(bus, "PCI card detected, they are not supported.\n");
 195                err = -ENXIO;
 196                goto err_pci_release_regions;
 197        }
 198
 199        /* Map MMIO */
 200        err = -ENOMEM;
 201        bus->mmio = pci_iomap(dev, 0, ~0UL);
 202        if (!bus->mmio)
 203                goto err_pci_release_regions;
 204
 205        /* Host specific */
 206        bus->host_pci = dev;
 207        bus->hosttype = BCMA_HOSTTYPE_PCI;
 208        bus->ops = &bcma_host_pci_ops;
 209
 210        bus->boardinfo.vendor = bus->host_pci->subsystem_vendor;
 211        bus->boardinfo.type = bus->host_pci->subsystem_device;
 212
 213        /* Initialize struct, detect chip */
 214        bcma_init_bus(bus);
 215
 216        /* Scan bus to find out generation of PCIe core */
 217        err = bcma_bus_scan(bus);
 218        if (err)
 219                goto err_pci_unmap_mmio;
 220
 221        if (bcma_find_core(bus, BCMA_CORE_PCIE2))
 222                bus->host_is_pcie2 = true;
 223
 224        /* Register */
 225        err = bcma_bus_register(bus);
 226        if (err)
 227                goto err_unregister_cores;
 228
 229        pci_set_drvdata(dev, bus);
 230
 231out:
 232        return err;
 233
 234err_unregister_cores:
 235        bcma_unregister_cores(bus);
 236err_pci_unmap_mmio:
 237        pci_iounmap(dev, bus->mmio);
 238err_pci_release_regions:
 239        pci_release_regions(dev);
 240err_pci_disable:
 241        pci_disable_device(dev);
 242err_kfree_bus:
 243        kfree(bus);
 244        return err;
 245}
 246
 247static void bcma_host_pci_remove(struct pci_dev *dev)
 248{
 249        struct bcma_bus *bus = pci_get_drvdata(dev);
 250
 251        bcma_bus_unregister(bus);
 252        pci_iounmap(dev, bus->mmio);
 253        pci_release_regions(dev);
 254        pci_disable_device(dev);
 255        kfree(bus);
 256}
 257
 258#ifdef CONFIG_PM_SLEEP
 259static int bcma_host_pci_suspend(struct device *dev)
 260{
 261        struct pci_dev *pdev = to_pci_dev(dev);
 262        struct bcma_bus *bus = pci_get_drvdata(pdev);
 263
 264        bus->mapped_core = NULL;
 265
 266        return bcma_bus_suspend(bus);
 267}
 268
 269static int bcma_host_pci_resume(struct device *dev)
 270{
 271        struct pci_dev *pdev = to_pci_dev(dev);
 272        struct bcma_bus *bus = pci_get_drvdata(pdev);
 273
 274        return bcma_bus_resume(bus);
 275}
 276
 277static SIMPLE_DEV_PM_OPS(bcma_pm_ops, bcma_host_pci_suspend,
 278                         bcma_host_pci_resume);
 279#define BCMA_PM_OPS     (&bcma_pm_ops)
 280
 281#else /* CONFIG_PM_SLEEP */
 282
 283#define BCMA_PM_OPS     NULL
 284
 285#endif /* CONFIG_PM_SLEEP */
 286
 287static const struct pci_device_id bcma_pci_bridge_tbl[] = {
 288        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x0576) },
 289        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4313) },
 290        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43224) },  /* 0xa8d8 */
 291        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4331) },
 292        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4353) },
 293        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4357) },
 294        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4358) },
 295        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4359) },
 296        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4360) },
 297        { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_DELL, 0x0016) },
 298        { PCI_DEVICE_SUB(PCI_VENDOR_ID_BROADCOM, 0x4365, PCI_VENDOR_ID_FOXCONN, 0xe092) },
 299        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a0) },
 300        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43a9) },
 301        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43aa) },
 302        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x43b1) },
 303        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 0x4727) },
 304        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43227) },  /* 0xa8db, BCM43217 (sic!) */
 305        { PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, 43228) },  /* 0xa8dc */
 306        { 0, },
 307};
 308MODULE_DEVICE_TABLE(pci, bcma_pci_bridge_tbl);
 309
 310static struct pci_driver bcma_pci_bridge_driver = {
 311        .name = "bcma-pci-bridge",
 312        .id_table = bcma_pci_bridge_tbl,
 313        .probe = bcma_host_pci_probe,
 314        .remove = bcma_host_pci_remove,
 315        .driver.pm = BCMA_PM_OPS,
 316};
 317
 318int __init bcma_host_pci_init(void)
 319{
 320        return pci_register_driver(&bcma_pci_bridge_driver);
 321}
 322
 323void __exit bcma_host_pci_exit(void)
 324{
 325        pci_unregister_driver(&bcma_pci_bridge_driver);
 326}
 327
 328/**************************************************
 329 * Runtime ops for drivers.
 330 **************************************************/
 331
 332/* See also pcicore_up */
 333void bcma_host_pci_up(struct bcma_bus *bus)
 334{
 335        if (bus->hosttype != BCMA_HOSTTYPE_PCI)
 336                return;
 337
 338        if (bus->host_is_pcie2)
 339                bcma_core_pcie2_up(&bus->drv_pcie2);
 340        else
 341                bcma_core_pci_up(&bus->drv_pci[0]);
 342}
 343EXPORT_SYMBOL_GPL(bcma_host_pci_up);
 344
 345/* See also pcicore_down */
 346void bcma_host_pci_down(struct bcma_bus *bus)
 347{
 348        if (bus->hosttype != BCMA_HOSTTYPE_PCI)
 349                return;
 350
 351        if (!bus->host_is_pcie2)
 352                bcma_core_pci_down(&bus->drv_pci[0]);
 353}
 354EXPORT_SYMBOL_GPL(bcma_host_pci_down);
 355
 356/* See also si_pci_setup */
 357int bcma_host_pci_irq_ctl(struct bcma_bus *bus, struct bcma_device *core,
 358                          bool enable)
 359{
 360        struct pci_dev *pdev;
 361        u32 coremask, tmp;
 362        int err = 0;
 363
 364        if (bus->hosttype != BCMA_HOSTTYPE_PCI) {
 365                /* This bcma device is not on a PCI host-bus. So the IRQs are
 366                 * not routed through the PCI core.
 367                 * So we must not enable routing through the PCI core. */
 368                goto out;
 369        }
 370
 371        pdev = bus->host_pci;
 372
 373        err = pci_read_config_dword(pdev, BCMA_PCI_IRQMASK, &tmp);
 374        if (err)
 375                goto out;
 376
 377        coremask = BIT(core->core_index) << 8;
 378        if (enable)
 379                tmp |= coremask;
 380        else
 381                tmp &= ~coremask;
 382
 383        err = pci_write_config_dword(pdev, BCMA_PCI_IRQMASK, tmp);
 384
 385out:
 386        return err;
 387}
 388EXPORT_SYMBOL_GPL(bcma_host_pci_irq_ctl);
 389
lxr.linux.no kindly hosted by Redpill Linpro AS, provider of Linux consulting and operations services since 1995.