9dc9fd9484
This patch adds a bunch of fixes 1. Reduce sg table size to 64 (SG_MX) instead of default SG_ALL 2. clear task lists on phy down events 3. release all tasks on port deformation 4. release current task for device gone notification 5. Add sata abort handing 6. Add 10ms delay to each port reset (currently done serially and with interrupts disabled) [jejb: whitespace fixes and clean ups plus added description added dummy 94xx_clear_srs_irq function just to prevent the mismatch in the mvs_dispatch structure killing 94xx cards] Signed-off-by: Srinivas <satyasrinivasp@hcl.in> Cc: Andy Yan <ayan@marvell.com> Cc: qswang@marvell.com Cc: jfeng@marvell.com Signed-off-by: James Bottomley <James.Bottomley@suse.de>
714 lines
18 KiB
C
714 lines
18 KiB
C
/*
|
|
* Marvell 88SE64xx/88SE94xx pci init
|
|
*
|
|
* Copyright 2007 Red Hat, Inc.
|
|
* Copyright 2008 Marvell. <kewei@marvell.com>
|
|
*
|
|
* This file is licensed under GPLv2.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License as
|
|
* published by the Free Software Foundation; version 2 of the
|
|
* License.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
|
|
* USA
|
|
*/
|
|
|
|
|
|
#include "mv_sas.h"
|
|
|
|
static struct scsi_transport_template *mvs_stt;
|
|
static const struct mvs_chip_info mvs_chips[] = {
|
|
[chip_6320] = { 1, 2, 0x400, 17, 16, 9, &mvs_64xx_dispatch, },
|
|
[chip_6440] = { 1, 4, 0x400, 17, 16, 9, &mvs_64xx_dispatch, },
|
|
[chip_6485] = { 1, 8, 0x800, 33, 32, 10, &mvs_64xx_dispatch, },
|
|
[chip_9180] = { 2, 4, 0x800, 17, 64, 9, &mvs_94xx_dispatch, },
|
|
[chip_9480] = { 2, 4, 0x800, 17, 64, 9, &mvs_94xx_dispatch, },
|
|
[chip_1300] = { 1, 4, 0x400, 17, 16, 9, &mvs_64xx_dispatch, },
|
|
[chip_1320] = { 2, 4, 0x800, 17, 64, 9, &mvs_94xx_dispatch, },
|
|
};
|
|
|
|
#define SOC_SAS_NUM 2
|
|
#define SG_MX 64
|
|
|
|
static struct scsi_host_template mvs_sht = {
|
|
.module = THIS_MODULE,
|
|
.name = DRV_NAME,
|
|
.queuecommand = sas_queuecommand,
|
|
.target_alloc = sas_target_alloc,
|
|
.slave_configure = mvs_slave_configure,
|
|
.slave_destroy = sas_slave_destroy,
|
|
.scan_finished = mvs_scan_finished,
|
|
.scan_start = mvs_scan_start,
|
|
.change_queue_depth = sas_change_queue_depth,
|
|
.change_queue_type = sas_change_queue_type,
|
|
.bios_param = sas_bios_param,
|
|
.can_queue = 1,
|
|
.cmd_per_lun = 1,
|
|
.this_id = -1,
|
|
.sg_tablesize = SG_MX,
|
|
.max_sectors = SCSI_DEFAULT_MAX_SECTORS,
|
|
.use_clustering = ENABLE_CLUSTERING,
|
|
.eh_device_reset_handler = sas_eh_device_reset_handler,
|
|
.eh_bus_reset_handler = sas_eh_bus_reset_handler,
|
|
.slave_alloc = mvs_slave_alloc,
|
|
.target_destroy = sas_target_destroy,
|
|
.ioctl = sas_ioctl,
|
|
};
|
|
|
|
static struct sas_domain_function_template mvs_transport_ops = {
|
|
.lldd_dev_found = mvs_dev_found,
|
|
.lldd_dev_gone = mvs_dev_gone,
|
|
.lldd_execute_task = mvs_queue_command,
|
|
.lldd_control_phy = mvs_phy_control,
|
|
|
|
.lldd_abort_task = mvs_abort_task,
|
|
.lldd_abort_task_set = mvs_abort_task_set,
|
|
.lldd_clear_aca = mvs_clear_aca,
|
|
.lldd_clear_task_set = mvs_clear_task_set,
|
|
.lldd_I_T_nexus_reset = mvs_I_T_nexus_reset,
|
|
.lldd_lu_reset = mvs_lu_reset,
|
|
.lldd_query_task = mvs_query_task,
|
|
.lldd_port_formed = mvs_port_formed,
|
|
.lldd_port_deformed = mvs_port_deformed,
|
|
|
|
};
|
|
|
|
static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[phy_id];
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
phy->mvi = mvi;
|
|
init_timer(&phy->timer);
|
|
sas_phy->enabled = (phy_id < mvi->chip->n_phy) ? 1 : 0;
|
|
sas_phy->class = SAS;
|
|
sas_phy->iproto = SAS_PROTOCOL_ALL;
|
|
sas_phy->tproto = 0;
|
|
sas_phy->type = PHY_TYPE_PHYSICAL;
|
|
sas_phy->role = PHY_ROLE_INITIATOR;
|
|
sas_phy->oob_mode = OOB_NOT_CONNECTED;
|
|
sas_phy->linkrate = SAS_LINK_RATE_UNKNOWN;
|
|
|
|
sas_phy->id = phy_id;
|
|
sas_phy->sas_addr = &mvi->sas_addr[0];
|
|
sas_phy->frame_rcvd = &phy->frame_rcvd[0];
|
|
sas_phy->ha = (struct sas_ha_struct *)mvi->shost->hostdata;
|
|
sas_phy->lldd_phy = phy;
|
|
}
|
|
|
|
static void mvs_free(struct mvs_info *mvi)
|
|
{
|
|
int i;
|
|
struct mvs_wq *mwq;
|
|
int slot_nr;
|
|
|
|
if (!mvi)
|
|
return;
|
|
|
|
if (mvi->flags & MVF_FLAG_SOC)
|
|
slot_nr = MVS_SOC_SLOTS;
|
|
else
|
|
slot_nr = MVS_SLOTS;
|
|
|
|
for (i = 0; i < mvi->tags_num; i++) {
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
if (slot->buf)
|
|
dma_free_coherent(mvi->dev, MVS_SLOT_BUF_SZ,
|
|
slot->buf, slot->buf_dma);
|
|
}
|
|
|
|
if (mvi->tx)
|
|
dma_free_coherent(mvi->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
mvi->tx, mvi->tx_dma);
|
|
if (mvi->rx_fis)
|
|
dma_free_coherent(mvi->dev, MVS_RX_FISL_SZ,
|
|
mvi->rx_fis, mvi->rx_fis_dma);
|
|
if (mvi->rx)
|
|
dma_free_coherent(mvi->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
mvi->rx, mvi->rx_dma);
|
|
if (mvi->slot)
|
|
dma_free_coherent(mvi->dev,
|
|
sizeof(*mvi->slot) * slot_nr,
|
|
mvi->slot, mvi->slot_dma);
|
|
#ifndef DISABLE_HOTPLUG_DMA_FIX
|
|
if (mvi->bulk_buffer)
|
|
dma_free_coherent(mvi->dev, TRASH_BUCKET_SIZE,
|
|
mvi->bulk_buffer, mvi->bulk_buffer_dma);
|
|
#endif
|
|
|
|
MVS_CHIP_DISP->chip_iounmap(mvi);
|
|
if (mvi->shost)
|
|
scsi_host_put(mvi->shost);
|
|
list_for_each_entry(mwq, &mvi->wq_list, entry)
|
|
cancel_delayed_work(&mwq->work_q);
|
|
kfree(mvi);
|
|
}
|
|
|
|
#ifdef MVS_USE_TASKLET
|
|
struct tasklet_struct mv_tasklet;
|
|
static void mvs_tasklet(unsigned long opaque)
|
|
{
|
|
unsigned long flags;
|
|
u32 stat;
|
|
u16 core_nr, i = 0;
|
|
|
|
struct mvs_info *mvi;
|
|
struct sas_ha_struct *sha = (struct sas_ha_struct *)opaque;
|
|
|
|
core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
if (unlikely(!mvi))
|
|
BUG_ON(1);
|
|
|
|
for (i = 0; i < core_nr; i++) {
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
stat = MVS_CHIP_DISP->isr_status(mvi, mvi->irq);
|
|
if (stat)
|
|
MVS_CHIP_DISP->isr(mvi, mvi->irq, stat);
|
|
}
|
|
|
|
}
|
|
#endif
|
|
|
|
static irqreturn_t mvs_interrupt(int irq, void *opaque)
|
|
{
|
|
u32 core_nr, i = 0;
|
|
u32 stat;
|
|
struct mvs_info *mvi;
|
|
struct sas_ha_struct *sha = opaque;
|
|
|
|
core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
if (unlikely(!mvi))
|
|
return IRQ_NONE;
|
|
|
|
stat = MVS_CHIP_DISP->isr_status(mvi, irq);
|
|
if (!stat)
|
|
return IRQ_NONE;
|
|
|
|
#ifdef MVS_USE_TASKLET
|
|
tasklet_schedule(&mv_tasklet);
|
|
#else
|
|
for (i = 0; i < core_nr; i++) {
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
MVS_CHIP_DISP->isr(mvi, irq, stat);
|
|
}
|
|
#endif
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int __devinit mvs_alloc(struct mvs_info *mvi, struct Scsi_Host *shost)
|
|
{
|
|
int i = 0, slot_nr;
|
|
|
|
if (mvi->flags & MVF_FLAG_SOC)
|
|
slot_nr = MVS_SOC_SLOTS;
|
|
else
|
|
slot_nr = MVS_SLOTS;
|
|
|
|
spin_lock_init(&mvi->lock);
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
mvs_phy_init(mvi, i);
|
|
mvi->port[i].wide_port_phymap = 0;
|
|
mvi->port[i].port_attached = 0;
|
|
INIT_LIST_HEAD(&mvi->port[i].list);
|
|
}
|
|
for (i = 0; i < MVS_MAX_DEVICES; i++) {
|
|
mvi->devices[i].taskfileset = MVS_ID_NOT_MAPPED;
|
|
mvi->devices[i].dev_type = NO_DEVICE;
|
|
mvi->devices[i].device_id = i;
|
|
mvi->devices[i].dev_status = MVS_DEV_NORMAL;
|
|
init_timer(&mvi->devices[i].timer);
|
|
}
|
|
|
|
/*
|
|
* alloc and init our DMA areas
|
|
*/
|
|
mvi->tx = dma_alloc_coherent(mvi->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
&mvi->tx_dma, GFP_KERNEL);
|
|
if (!mvi->tx)
|
|
goto err_out;
|
|
memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ);
|
|
mvi->rx_fis = dma_alloc_coherent(mvi->dev, MVS_RX_FISL_SZ,
|
|
&mvi->rx_fis_dma, GFP_KERNEL);
|
|
if (!mvi->rx_fis)
|
|
goto err_out;
|
|
memset(mvi->rx_fis, 0, MVS_RX_FISL_SZ);
|
|
|
|
mvi->rx = dma_alloc_coherent(mvi->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
&mvi->rx_dma, GFP_KERNEL);
|
|
if (!mvi->rx)
|
|
goto err_out;
|
|
memset(mvi->rx, 0, sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1));
|
|
mvi->rx[0] = cpu_to_le32(0xfff);
|
|
mvi->rx_cons = 0xfff;
|
|
|
|
mvi->slot = dma_alloc_coherent(mvi->dev,
|
|
sizeof(*mvi->slot) * slot_nr,
|
|
&mvi->slot_dma, GFP_KERNEL);
|
|
if (!mvi->slot)
|
|
goto err_out;
|
|
memset(mvi->slot, 0, sizeof(*mvi->slot) * slot_nr);
|
|
|
|
#ifndef DISABLE_HOTPLUG_DMA_FIX
|
|
mvi->bulk_buffer = dma_alloc_coherent(mvi->dev,
|
|
TRASH_BUCKET_SIZE,
|
|
&mvi->bulk_buffer_dma, GFP_KERNEL);
|
|
if (!mvi->bulk_buffer)
|
|
goto err_out;
|
|
#endif
|
|
for (i = 0; i < slot_nr; i++) {
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
|
|
slot->buf = dma_alloc_coherent(mvi->dev, MVS_SLOT_BUF_SZ,
|
|
&slot->buf_dma, GFP_KERNEL);
|
|
if (!slot->buf) {
|
|
printk(KERN_DEBUG"failed to allocate slot->buf.\n");
|
|
goto err_out;
|
|
}
|
|
memset(slot->buf, 0, MVS_SLOT_BUF_SZ);
|
|
++mvi->tags_num;
|
|
}
|
|
/* Initialize tags */
|
|
mvs_tag_init(mvi);
|
|
return 0;
|
|
err_out:
|
|
return 1;
|
|
}
|
|
|
|
|
|
int mvs_ioremap(struct mvs_info *mvi, int bar, int bar_ex)
|
|
{
|
|
unsigned long res_start, res_len, res_flag, res_flag_ex = 0;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
if (bar_ex != -1) {
|
|
/*
|
|
* ioremap main and peripheral registers
|
|
*/
|
|
res_start = pci_resource_start(pdev, bar_ex);
|
|
res_len = pci_resource_len(pdev, bar_ex);
|
|
if (!res_start || !res_len)
|
|
goto err_out;
|
|
|
|
res_flag_ex = pci_resource_flags(pdev, bar_ex);
|
|
if (res_flag_ex & IORESOURCE_MEM) {
|
|
if (res_flag_ex & IORESOURCE_CACHEABLE)
|
|
mvi->regs_ex = ioremap(res_start, res_len);
|
|
else
|
|
mvi->regs_ex = ioremap_nocache(res_start,
|
|
res_len);
|
|
} else
|
|
mvi->regs_ex = (void *)res_start;
|
|
if (!mvi->regs_ex)
|
|
goto err_out;
|
|
}
|
|
|
|
res_start = pci_resource_start(pdev, bar);
|
|
res_len = pci_resource_len(pdev, bar);
|
|
if (!res_start || !res_len)
|
|
goto err_out;
|
|
|
|
res_flag = pci_resource_flags(pdev, bar);
|
|
if (res_flag & IORESOURCE_CACHEABLE)
|
|
mvi->regs = ioremap(res_start, res_len);
|
|
else
|
|
mvi->regs = ioremap_nocache(res_start, res_len);
|
|
|
|
if (!mvi->regs) {
|
|
if (mvi->regs_ex && (res_flag_ex & IORESOURCE_MEM))
|
|
iounmap(mvi->regs_ex);
|
|
mvi->regs_ex = NULL;
|
|
goto err_out;
|
|
}
|
|
|
|
return 0;
|
|
err_out:
|
|
return -1;
|
|
}
|
|
|
|
void mvs_iounmap(void __iomem *regs)
|
|
{
|
|
iounmap(regs);
|
|
}
|
|
|
|
static struct mvs_info *__devinit mvs_pci_alloc(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent,
|
|
struct Scsi_Host *shost, unsigned int id)
|
|
{
|
|
struct mvs_info *mvi;
|
|
struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
mvi = kzalloc(sizeof(*mvi) + MVS_SLOTS * sizeof(struct mvs_slot_info),
|
|
GFP_KERNEL);
|
|
if (!mvi)
|
|
return NULL;
|
|
|
|
mvi->pdev = pdev;
|
|
mvi->dev = &pdev->dev;
|
|
mvi->chip_id = ent->driver_data;
|
|
mvi->chip = &mvs_chips[mvi->chip_id];
|
|
INIT_LIST_HEAD(&mvi->wq_list);
|
|
mvi->irq = pdev->irq;
|
|
|
|
((struct mvs_prv_info *)sha->lldd_ha)->mvi[id] = mvi;
|
|
((struct mvs_prv_info *)sha->lldd_ha)->n_phy = mvi->chip->n_phy;
|
|
|
|
mvi->id = id;
|
|
mvi->sas = sha;
|
|
mvi->shost = shost;
|
|
#ifdef MVS_USE_TASKLET
|
|
tasklet_init(&mv_tasklet, mvs_tasklet, (unsigned long)sha);
|
|
#endif
|
|
|
|
if (MVS_CHIP_DISP->chip_ioremap(mvi))
|
|
goto err_out;
|
|
if (!mvs_alloc(mvi, shost))
|
|
return mvi;
|
|
err_out:
|
|
mvs_free(mvi);
|
|
return NULL;
|
|
}
|
|
|
|
/* move to PCI layer or libata core? */
|
|
static int pci_go_64(struct pci_dev *pdev)
|
|
{
|
|
int rc;
|
|
|
|
if (!pci_set_dma_mask(pdev, DMA_BIT_MASK(64))) {
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64));
|
|
if (rc) {
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"64-bit DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
}
|
|
} else {
|
|
rc = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"32-bit DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"32-bit consistent DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int __devinit mvs_prep_sas_ha_init(struct Scsi_Host *shost,
|
|
const struct mvs_chip_info *chip_info)
|
|
{
|
|
int phy_nr, port_nr; unsigned short core_nr;
|
|
struct asd_sas_phy **arr_phy;
|
|
struct asd_sas_port **arr_port;
|
|
struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
|
|
core_nr = chip_info->n_host;
|
|
phy_nr = core_nr * chip_info->n_phy;
|
|
port_nr = phy_nr;
|
|
|
|
memset(sha, 0x00, sizeof(struct sas_ha_struct));
|
|
arr_phy = kcalloc(phy_nr, sizeof(void *), GFP_KERNEL);
|
|
arr_port = kcalloc(port_nr, sizeof(void *), GFP_KERNEL);
|
|
if (!arr_phy || !arr_port)
|
|
goto exit_free;
|
|
|
|
sha->sas_phy = arr_phy;
|
|
sha->sas_port = arr_port;
|
|
sha->core.shost = shost;
|
|
|
|
sha->lldd_ha = kzalloc(sizeof(struct mvs_prv_info), GFP_KERNEL);
|
|
if (!sha->lldd_ha)
|
|
goto exit_free;
|
|
|
|
((struct mvs_prv_info *)sha->lldd_ha)->n_host = core_nr;
|
|
|
|
shost->transportt = mvs_stt;
|
|
shost->max_id = 128;
|
|
shost->max_lun = ~0;
|
|
shost->max_channel = 1;
|
|
shost->max_cmd_len = 16;
|
|
|
|
return 0;
|
|
exit_free:
|
|
kfree(arr_phy);
|
|
kfree(arr_port);
|
|
return -1;
|
|
|
|
}
|
|
|
|
static void __devinit mvs_post_sas_ha_init(struct Scsi_Host *shost,
|
|
const struct mvs_chip_info *chip_info)
|
|
{
|
|
int can_queue, i = 0, j = 0;
|
|
struct mvs_info *mvi = NULL;
|
|
struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost);
|
|
unsigned short nr_core = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
|
|
for (j = 0; j < nr_core; j++) {
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[j];
|
|
for (i = 0; i < chip_info->n_phy; i++) {
|
|
sha->sas_phy[j * chip_info->n_phy + i] =
|
|
&mvi->phy[i].sas_phy;
|
|
sha->sas_port[j * chip_info->n_phy + i] =
|
|
&mvi->port[i].sas_port;
|
|
}
|
|
}
|
|
|
|
sha->sas_ha_name = DRV_NAME;
|
|
sha->dev = mvi->dev;
|
|
sha->lldd_module = THIS_MODULE;
|
|
sha->sas_addr = &mvi->sas_addr[0];
|
|
|
|
sha->num_phys = nr_core * chip_info->n_phy;
|
|
|
|
sha->lldd_max_execute_num = 1;
|
|
|
|
if (mvi->flags & MVF_FLAG_SOC)
|
|
can_queue = MVS_SOC_CAN_QUEUE;
|
|
else
|
|
can_queue = MVS_CAN_QUEUE;
|
|
|
|
sha->lldd_queue_size = can_queue;
|
|
shost->can_queue = can_queue;
|
|
mvi->shost->cmd_per_lun = MVS_SLOTS/sha->num_phys;
|
|
sha->core.shost = mvi->shost;
|
|
}
|
|
|
|
static void mvs_init_sas_add(struct mvs_info *mvi)
|
|
{
|
|
u8 i;
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
mvi->phy[i].dev_sas_addr = 0x5005043011ab0000ULL;
|
|
mvi->phy[i].dev_sas_addr =
|
|
cpu_to_be64((u64)(*(u64 *)&mvi->phy[i].dev_sas_addr));
|
|
}
|
|
|
|
memcpy(mvi->sas_addr, &mvi->phy[0].dev_sas_addr, SAS_ADDR_SIZE);
|
|
}
|
|
|
|
static int __devinit mvs_pci_init(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent)
|
|
{
|
|
unsigned int rc, nhost = 0;
|
|
struct mvs_info *mvi;
|
|
irq_handler_t irq_handler = mvs_interrupt;
|
|
struct Scsi_Host *shost = NULL;
|
|
const struct mvs_chip_info *chip;
|
|
|
|
dev_printk(KERN_INFO, &pdev->dev,
|
|
"mvsas: driver version %s\n", DRV_VERSION);
|
|
rc = pci_enable_device(pdev);
|
|
if (rc)
|
|
goto err_out_enable;
|
|
|
|
pci_set_master(pdev);
|
|
|
|
rc = pci_request_regions(pdev, DRV_NAME);
|
|
if (rc)
|
|
goto err_out_disable;
|
|
|
|
rc = pci_go_64(pdev);
|
|
if (rc)
|
|
goto err_out_regions;
|
|
|
|
shost = scsi_host_alloc(&mvs_sht, sizeof(void *));
|
|
if (!shost) {
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
}
|
|
|
|
chip = &mvs_chips[ent->driver_data];
|
|
SHOST_TO_SAS_HA(shost) =
|
|
kcalloc(1, sizeof(struct sas_ha_struct), GFP_KERNEL);
|
|
if (!SHOST_TO_SAS_HA(shost)) {
|
|
kfree(shost);
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
}
|
|
|
|
rc = mvs_prep_sas_ha_init(shost, chip);
|
|
if (rc) {
|
|
kfree(shost);
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
}
|
|
|
|
pci_set_drvdata(pdev, SHOST_TO_SAS_HA(shost));
|
|
|
|
do {
|
|
mvi = mvs_pci_alloc(pdev, ent, shost, nhost);
|
|
if (!mvi) {
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
}
|
|
|
|
mvs_init_sas_add(mvi);
|
|
|
|
mvi->instance = nhost;
|
|
rc = MVS_CHIP_DISP->chip_init(mvi);
|
|
if (rc) {
|
|
mvs_free(mvi);
|
|
goto err_out_regions;
|
|
}
|
|
nhost++;
|
|
} while (nhost < chip->n_host);
|
|
#ifdef MVS_USE_TASKLET
|
|
tasklet_init(&mv_tasklet, mvs_tasklet,
|
|
(unsigned long)SHOST_TO_SAS_HA(shost));
|
|
#endif
|
|
|
|
mvs_post_sas_ha_init(shost, chip);
|
|
|
|
rc = scsi_add_host(shost, &pdev->dev);
|
|
if (rc)
|
|
goto err_out_shost;
|
|
|
|
rc = sas_register_ha(SHOST_TO_SAS_HA(shost));
|
|
if (rc)
|
|
goto err_out_shost;
|
|
rc = request_irq(pdev->irq, irq_handler, IRQF_SHARED,
|
|
DRV_NAME, SHOST_TO_SAS_HA(shost));
|
|
if (rc)
|
|
goto err_not_sas;
|
|
|
|
MVS_CHIP_DISP->interrupt_enable(mvi);
|
|
|
|
scsi_scan_host(mvi->shost);
|
|
|
|
return 0;
|
|
|
|
err_not_sas:
|
|
sas_unregister_ha(SHOST_TO_SAS_HA(shost));
|
|
err_out_shost:
|
|
scsi_remove_host(mvi->shost);
|
|
err_out_regions:
|
|
pci_release_regions(pdev);
|
|
err_out_disable:
|
|
pci_disable_device(pdev);
|
|
err_out_enable:
|
|
return rc;
|
|
}
|
|
|
|
static void __devexit mvs_pci_remove(struct pci_dev *pdev)
|
|
{
|
|
unsigned short core_nr, i = 0;
|
|
struct sas_ha_struct *sha = pci_get_drvdata(pdev);
|
|
struct mvs_info *mvi = NULL;
|
|
|
|
core_nr = ((struct mvs_prv_info *)sha->lldd_ha)->n_host;
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[0];
|
|
|
|
#ifdef MVS_USE_TASKLET
|
|
tasklet_kill(&mv_tasklet);
|
|
#endif
|
|
|
|
pci_set_drvdata(pdev, NULL);
|
|
sas_unregister_ha(sha);
|
|
sas_remove_host(mvi->shost);
|
|
scsi_remove_host(mvi->shost);
|
|
|
|
MVS_CHIP_DISP->interrupt_disable(mvi);
|
|
free_irq(mvi->irq, sha);
|
|
for (i = 0; i < core_nr; i++) {
|
|
mvi = ((struct mvs_prv_info *)sha->lldd_ha)->mvi[i];
|
|
mvs_free(mvi);
|
|
}
|
|
kfree(sha->sas_phy);
|
|
kfree(sha->sas_port);
|
|
kfree(sha);
|
|
pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
return;
|
|
}
|
|
|
|
static struct pci_device_id __devinitdata mvs_pci_table[] = {
|
|
{ PCI_VDEVICE(MARVELL, 0x6320), chip_6320 },
|
|
{ PCI_VDEVICE(MARVELL, 0x6340), chip_6440 },
|
|
{
|
|
.vendor = PCI_VENDOR_ID_MARVELL,
|
|
.device = 0x6440,
|
|
.subvendor = PCI_ANY_ID,
|
|
.subdevice = 0x6480,
|
|
.class = 0,
|
|
.class_mask = 0,
|
|
.driver_data = chip_6485,
|
|
},
|
|
{ PCI_VDEVICE(MARVELL, 0x6440), chip_6440 },
|
|
{ PCI_VDEVICE(MARVELL, 0x6485), chip_6485 },
|
|
{ PCI_VDEVICE(MARVELL, 0x9480), chip_9480 },
|
|
{ PCI_VDEVICE(MARVELL, 0x9180), chip_9180 },
|
|
{ PCI_VDEVICE(ARECA, PCI_DEVICE_ID_ARECA_1300), chip_1300 },
|
|
{ PCI_VDEVICE(ARECA, PCI_DEVICE_ID_ARECA_1320), chip_1320 },
|
|
{ PCI_VDEVICE(ADAPTEC2, 0x0450), chip_6440 },
|
|
|
|
{ } /* terminate list */
|
|
};
|
|
|
|
static struct pci_driver mvs_pci_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = mvs_pci_table,
|
|
.probe = mvs_pci_init,
|
|
.remove = __devexit_p(mvs_pci_remove),
|
|
};
|
|
|
|
/* task handler */
|
|
struct task_struct *mvs_th;
|
|
static int __init mvs_init(void)
|
|
{
|
|
int rc;
|
|
mvs_stt = sas_domain_attach_transport(&mvs_transport_ops);
|
|
if (!mvs_stt)
|
|
return -ENOMEM;
|
|
|
|
rc = pci_register_driver(&mvs_pci_driver);
|
|
|
|
if (rc)
|
|
goto err_out;
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
sas_release_transport(mvs_stt);
|
|
return rc;
|
|
}
|
|
|
|
static void __exit mvs_exit(void)
|
|
{
|
|
pci_unregister_driver(&mvs_pci_driver);
|
|
sas_release_transport(mvs_stt);
|
|
}
|
|
|
|
module_init(mvs_init);
|
|
module_exit(mvs_exit);
|
|
|
|
MODULE_AUTHOR("Jeff Garzik <jgarzik@pobox.com>");
|
|
MODULE_DESCRIPTION("Marvell 88SE6440 SAS/SATA controller driver");
|
|
MODULE_VERSION(DRV_VERSION);
|
|
MODULE_LICENSE("GPL");
|
|
#ifdef CONFIG_PCI
|
|
MODULE_DEVICE_TABLE(pci, mvs_pci_table);
|
|
#endif
|