Lines Matching defs:dev

255 static int hpsa_ioctl(struct scsi_device *dev, unsigned int cmd,
263 static int hpsa_compat_ioctl(struct scsi_device *dev, unsigned int cmd,
332 struct hpsa_scsi_dev_t *dev,
392 dev_warn(&h->pdev->dev,
397 dev_warn(&h->pdev->dev,
401 dev_warn(&h->pdev->dev,
409 dev_warn(&h->pdev->dev,
414 dev_warn(&h->pdev->dev,
419 dev_warn(&h->pdev->dev,
433 dev_warn(&h->pdev->dev, HPSA "device busy");
438 static ssize_t host_show_lockup_detected(struct device *dev,
443 struct Scsi_Host *shost = class_to_shost(dev);
451 static ssize_t host_store_hp_ssd_smart_path_status(struct device *dev,
457 struct Scsi_Host *shost = class_to_shost(dev);
469 dev_warn(&h->pdev->dev,
475 static ssize_t host_store_raid_offload_debug(struct device *dev,
481 struct Scsi_Host *shost = class_to_shost(dev);
495 dev_warn(&h->pdev->dev, "hpsa: Set raid_offload_debug level = %d\n",
500 static ssize_t host_store_rescan(struct device *dev,
505 struct Scsi_Host *shost = class_to_shost(dev);
517 static ssize_t host_show_firmware_revision(struct device *dev,
521 struct Scsi_Host *shost = class_to_shost(dev);
532 static ssize_t host_show_commands_outstanding(struct device *dev,
535 struct Scsi_Host *shost = class_to_shost(dev);
542 static ssize_t host_show_transport_mode(struct device *dev,
546 struct Scsi_Host *shost = class_to_shost(dev);
554 static ssize_t host_show_hp_ssd_smart_path_status(struct device *dev,
558 struct Scsi_Host *shost = class_to_shost(dev);
638 static ssize_t host_show_resettable(struct device *dev,
642 struct Scsi_Host *shost = class_to_shost(dev);
671 static ssize_t raid_level_show(struct device *dev,
681 sdev = to_scsi_device(dev);
705 static ssize_t lunid_show(struct device *dev,
714 sdev = to_scsi_device(dev);
727 static ssize_t unique_id_show(struct device *dev,
736 sdev = to_scsi_device(dev);
755 static ssize_t sas_address_show(struct device *dev,
764 sdev = to_scsi_device(dev);
778 static ssize_t host_show_hp_ssd_smart_path_enabled(struct device *dev,
787 sdev = to_scsi_device(dev);
806 static ssize_t path_info_show(struct device *dev,
821 sdev = to_scsi_device(dev);
891 static ssize_t host_show_ctlr_num(struct device *dev,
895 struct Scsi_Host *shost = class_to_shost(dev);
901 static ssize_t host_show_legacy_board(struct device *dev,
905 struct Scsi_Host *shost = class_to_shost(dev);
1214 if (h->dev[i]->bus == bus && h->dev[i]->target != -1)
1215 __set_bit(h->dev[i]->target, lun_taken);
1229 struct hpsa_scsi_dev_t *dev, char *description)
1237 switch (dev->devtype) {
1246 if (dev->external)
1248 else if (!is_logical_dev_addr_mode(dev->scsi3addr))
1253 dev->raid_level > RAID_UNKNOWN ? "?" :
1254 raid_label[dev->raid_level]);
1270 dev_printk(level, &h->pdev->dev,
1272 h->scsi_host->host_no, dev->bus, dev->target, dev->lun,
1274 scsi_device_type(dev->devtype),
1275 dev->vendor,
1276 dev->model,
1278 dev->offload_config ? '+' : '-',
1279 dev->offload_to_be_enabled ? '+' : '-',
1280 dev->expose_device);
1283 /* Add an entry into h->dev[] array. */
1295 dev_err(&h->pdev->dev, "too many devices, some will be "
1327 sd = h->dev[i];
1340 dev_warn(&h->pdev->dev, "physical device with no LUN=0,"
1348 h->dev[n] = device;
1360 * Update an entry in h->dev[] array.
1369 h->dev[entry]->raid_level = new_entry->raid_level;
1374 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
1384 * h->dev[entry]->offload_enabled is currently 0.
1386 h->dev[entry]->raid_map = new_entry->raid_map;
1387 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
1390 h->dev[entry]->ioaccel_handle = new_entry->ioaccel_handle;
1393 h->dev[entry]->hba_ioaccel_enabled = new_entry->hba_ioaccel_enabled;
1394 h->dev[entry]->offload_config = new_entry->offload_config;
1395 h->dev[entry]->offload_to_mirror = new_entry->offload_to_mirror;
1396 h->dev[entry]->queue_depth = new_entry->queue_depth;
1400 * ioaccel on until we can update h->dev[entry]->phys_disk[], but we
1403 h->dev[entry]->offload_to_be_enabled = new_entry->offload_to_be_enabled;
1409 h->dev[entry]->offload_enabled = 0;
1411 hpsa_show_dev_msg(KERN_INFO, h, h->dev[entry], "updated");
1414 /* Replace an entry from h->dev[] array. */
1422 removed[*nremoved] = h->dev[entry];
1430 new_entry->target = h->dev[entry]->target;
1431 new_entry->lun = h->dev[entry]->lun;
1434 h->dev[entry] = new_entry;
1441 /* Remove an entry from h->dev[] array. */
1451 sd = h->dev[entry];
1452 removed[*nremoved] = h->dev[entry];
1456 h->dev[i] = h->dev[i+1];
1475 * h->dev[] to match the mid layer's view.
1482 if (h->dev[i] == added) {
1484 h->dev[j] = h->dev[j+1];
1621 dev_info(&h->pdev->dev,
1629 dev_info(&h->pdev->dev,
1635 dev_info(&h->pdev->dev,
1641 dev_info(&h->pdev->dev,
1647 dev_info(&h->pdev->dev,
1653 dev_info(&h->pdev->dev,
1659 dev_info(&h->pdev->dev,
1665 dev_info(&h->pdev->dev,
1671 dev_info(&h->pdev->dev,
1677 dev_info(&h->pdev->dev,
1683 dev_info(&h->pdev->dev,
1689 dev_info(&h->pdev->dev,
1702 struct hpsa_scsi_dev_t *dev[], int ndevices,
1728 if (dev[j] == NULL)
1730 if (dev[j]->devtype != TYPE_DISK &&
1731 dev[j]->devtype != TYPE_ZBC)
1733 if (is_logical_device(dev[j]))
1735 if (dev[j]->ioaccel_handle != dd[i].ioaccel_handle)
1738 logical_drive->phys_disk[i] = dev[j];
1753 dev_warn(&h->pdev->dev,
1777 struct hpsa_scsi_dev_t *dev[], int ndevices)
1782 if (dev[i] == NULL)
1784 if (dev[i]->devtype != TYPE_DISK &&
1785 dev[i]->devtype != TYPE_ZBC)
1787 if (!is_logical_device(dev[i]))
1809 if (!dev[i]->offload_enabled && dev[i]->offload_to_be_enabled)
1810 hpsa_figure_phys_disk_ptrs(h, dev, ndevices, dev[i]);
1831 struct hpsa_scsi_dev_t *dev)
1840 if (refcount > 1 && hpsa_cmd_dev_match(h, c, dev,
1841 dev->scsi3addr)) {
1877 dev_warn(&h->pdev->dev,
1925 * reality and updates h->dev[] to reflect that reality.
1949 dev_warn(&h->pdev->dev, "out of memory in "
1956 /* find any devices in h->dev[] that are not in
1957 * sd[] and remove them from h->dev[], and for any
1967 csd = h->dev[i];
1988 * listed in h->dev[], adding them if they aren't found
2006 device_change = hpsa_scsi_find_entry(sd[i], h->dev,
2016 dev_warn(&h->pdev->dev,
2021 hpsa_update_log_drive_phys_drive_ptrs(h, h->dev, h->ndevices);
2024 * Now that h->dev[]->phys_disk[] is coherent, we can enable
2032 if (h->dev[i] == NULL)
2034 h->dev[i]->offload_enabled = h->dev[i]->offload_to_be_enabled;
2041 * so don't touch h->dev[]
2078 dev_warn(&h->pdev->dev,
2080 /* now we have to remove it from h->dev,
2103 sd = h->dev[i];
2277 temp64 = dma_map_single(&h->pdev->dev, chain_block, chain_size,
2279 if (dma_mapping_error(&h->pdev->dev, temp64)) {
2298 dma_unmap_single(&h->pdev->dev, temp64, chain_size, DMA_TO_DEVICE);
2314 temp64 = dma_map_single(&h->pdev->dev, chain_block, chain_len,
2316 if (dma_mapping_error(&h->pdev->dev, temp64)) {
2334 dma_unmap_single(&h->pdev->dev, le64_to_cpu(chain_sg->Addr),
2347 struct hpsa_scsi_dev_t *dev)
2420 if (dev->physical_device && dev->expose_device) {
2422 dev->removed = 1;
2424 dev_warn(&h->pdev->dev,
2452 if (dev->in_reset)
2461 struct hpsa_scsi_dev_t *dev = c->device;
2470 if (dev) {
2471 atomic_dec(&dev->commands_outstanding);
2472 if (dev->in_reset &&
2473 atomic_read(&dev->commands_outstanding) <= 0)
2501 struct hpsa_scsi_dev_t *dev)
2517 if (is_logical_device(dev) &&
2522 hpsa_turn_off_ioaccel_for_device(dev);
2525 if (dev->in_reset) {
2533 if (handle_ioaccel_mode2_error(h, c, cmd, c2, dev))
2560 dev_warn(&h->pdev->dev, "Unknown TMF status: 0x%02x\n",
2572 struct hpsa_scsi_dev_t *dev;
2589 dev = cmd->device->hostdata;
2590 if (!dev) {
2608 if (dev->was_removed) {
2614 if (dev->physical_device && dev->expose_device &&
2615 dev->removed) {
2635 return process_ioaccel2_completion(h, cp, cmd, dev);
2658 if (is_logical_device(dev)) {
2660 dev->offload_enabled = 0;
2692 dev->removed = 1;
2703 dev_warn(&h->pdev->dev, "cp %p has status 0x%x "
2710 dev_warn(&h->pdev->dev, "cp %p SCSI status was 0. "
2732 dev_warn(&h->pdev->dev,
2749 dev_warn(&h->pdev->dev, "CDB %16phN : protocol error\n",
2754 dev_warn(&h->pdev->dev, "CDB %16phN : hardware error\n",
2759 dev_warn(&h->pdev->dev, "CDB %16phN : connection lost\n",
2767 dev_warn(&h->pdev->dev, "CDB %16phN : abort failed\n",
2772 dev_warn(&h->pdev->dev, "CDB %16phN : unsolicited abort\n",
2777 dev_warn(&h->pdev->dev, "CDB %16phN timed out\n",
2782 dev_warn(&h->pdev->dev, "Command unabortable\n");
2793 dev_warn(&h->pdev->dev,
2798 dev_warn(&h->pdev->dev, "cp %p returned unknown status %x\n",
2811 dma_unmap_single(&pdev->dev, le64_to_cpu(c->SG[i].Addr),
2830 addr64 = dma_map_single(&pdev->dev, buf, buflen, data_direction);
2831 if (dma_mapping_error(&pdev->dev, addr64)) {
2861 dev_warn(&h->pdev->dev, "Command timed out.\n");
2924 dev_warn(&h->pdev->dev, "%s: LUN:%8phN CDB:%16phN\n",
2932 struct device *d = &cp->h->pdev->dev;
3083 static int hpsa_send_reset(struct ctlr_info *h, struct hpsa_scsi_dev_t *dev,
3091 c->device = dev;
3094 (void) fill_cmd(c, reset_type, h, NULL, 0, 0, dev->scsi3addr, TYPE_MSG);
3097 dev_warn(&h->pdev->dev, "Failed to send reset command\n");
3113 struct hpsa_scsi_dev_t *dev,
3133 if (c->phys_disk == dev) {
3137 /* Possible RAID mode -- check each phys dev. */
3141 for (i = 0; i < dev->nphysical_disks && !match; i++) {
3144 * match = dev->phys_disk[i]->ioaccel_handle
3146 match = dev->phys_disk[i] == c->phys_disk;
3152 for (i = 0; i < dev->nphysical_disks && !match; i++) {
3153 match = dev->phys_disk[i]->ioaccel_handle ==
3163 dev_err(&h->pdev->dev, "unexpected cmd_type: %d\n",
3171 static int hpsa_do_reset(struct ctlr_info *h, struct hpsa_scsi_dev_t *dev,
3178 dev_warn(&h->pdev->dev, "concurrent reset wait interrupted.\n");
3182 rc = hpsa_send_reset(h, dev, reset_type, reply_queue);
3185 atomic_dec(&dev->commands_outstanding);
3187 atomic_read(&dev->commands_outstanding) <= 0 ||
3192 dev_warn(&h->pdev->dev,
3198 rc = wait_for_device_to_become_ready(h, dev->scsi3addr, 0);
3247 dev_info(&h->pdev->dev, "structure_size = %u\n",
3249 dev_info(&h->pdev->dev, "volume_blk_size = %u\n",
3251 dev_info(&h->pdev->dev, "volume_blk_cnt = 0x%llx\n",
3253 dev_info(&h->pdev->dev, "physicalBlockShift = %u\n",
3255 dev_info(&h->pdev->dev, "parity_rotation_shift = %u\n",
3257 dev_info(&h->pdev->dev, "strip_size = %u\n",
3259 dev_info(&h->pdev->dev, "disk_starting_blk = 0x%llx\n",
3261 dev_info(&h->pdev->dev, "disk_blk_cnt = 0x%llx\n",
3263 dev_info(&h->pdev->dev, "data_disks_per_row = %u\n",
3265 dev_info(&h->pdev->dev, "metadata_disks_per_row = %u\n",
3267 dev_info(&h->pdev->dev, "row_cnt = %u\n",
3269 dev_info(&h->pdev->dev, "layout_map_count = %u\n",
3271 dev_info(&h->pdev->dev, "flags = 0x%x\n",
3273 dev_info(&h->pdev->dev, "encryption = %s\n",
3276 dev_info(&h->pdev->dev, "dekindex = %u\n",
3280 dev_info(&h->pdev->dev, "Map%u:\n", map);
3283 dev_info(&h->pdev->dev, " Row%u:\n", row);
3287 dev_info(&h->pdev->dev,
3294 dev_info(&h->pdev->dev,
3321 dev_warn(&h->pdev->dev, "hpsa_get_raid_map fill_cmd failed\n");
3340 dev_warn(&h->pdev->dev, "RAID map size is too large!\n");
3490 dev_warn(&h->pdev->dev, "%s: id_phys failed %d bdi[0x%x]\n",
3549 dev_err(&h->pdev->dev, "report physical LUNs failed.\n");
3567 struct hpsa_scsi_dev_t *dev)
3590 dev->sas_address = sa;
3608 dev_info(&h->pdev->dev,
3776 dev_err(&h->pdev->dev,
3962 dev_err(&h->pdev->dev,
3983 dev_err(&h->pdev->dev,
4009 dev_info(&h->pdev->dev,
4017 dev_err(&h->pdev->dev,
4120 dev_err(&h->pdev->dev, "report physical LUNs failed.\n");
4125 dev_warn(&h->pdev->dev, "maximum physical LUNs (%d) exceeded. %d LUNs ignored.\n",
4130 dev_err(&h->pdev->dev, "report logical LUNs failed.\n");
4136 dev_warn(&h->pdev->dev,
4143 dev_warn(&h->pdev->dev,
4181 struct hpsa_scsi_dev_t *dev,
4193 dev->ioaccel_handle = rle->ioaccel_handle;
4194 if ((rle->device_flags & 0x08) && dev->ioaccel_handle)
4195 dev->hba_ioaccel_enabled = 1;
4204 dev->queue_depth =
4208 dev->queue_depth = DRIVE_QUEUE_DEPTH; /* conservative */
4250 dev_warn(&h->pdev->dev, "%s: id_ctlr buffer is NULL.\n",
4367 dev_err(&h->pdev->dev, "out of memory\n");
4381 dev_warn(&h->pdev->dev,
4397 dev_warn(&h->pdev->dev, "maximum devices (%d) exceeded."
4451 dev_warn(&h->pdev->dev,
4541 dev_warn(&h->pdev->dev,
4649 dev_warn(&h->pdev->dev,
4801 dev_err(&h->pdev->dev, "unknown data direction: %d\n",
4832 struct hpsa_scsi_dev_t *dev = cmd->device->hostdata;
4834 if (!dev)
4837 c->phys_disk = dev;
4839 if (dev->in_reset)
4842 return hpsa_scsi_ioaccel_queue_command(h, c, dev->ioaccel_handle,
4843 cmd->cmnd, cmd->cmd_len, dev->scsi3addr, dev);
4853 struct hpsa_scsi_dev_t *dev = cmd->device->hostdata;
4854 struct raid_map_data *map = &dev->raid_map;
4890 dev_err(&h->pdev->dev,
4998 dev_err(&h->pdev->dev, "unknown data direction: %d\n",
5106 struct hpsa_scsi_dev_t *dev = cmd->device->hostdata;
5107 struct raid_map_data *map = &dev->raid_map;
5137 if (!dev)
5140 if (dev->in_reset)
5209 if (is_write && dev->raid_level != 0)
5256 switch (dev->raid_level) {
5266 hpsa_turn_off_ioaccel_for_device(dev);
5269 if (dev->offload_to_mirror)
5271 dev->offload_to_mirror = !dev->offload_to_mirror;
5279 hpsa_turn_off_ioaccel_for_device(dev);
5283 offload_to_mirror = dev->offload_to_mirror;
5291 dev->offload_to_mirror = offload_to_mirror;
5292 /* Avoid direct use of dev->offload_to_mirror within this
5294 * increment it beyond the range of dev->layout_map_count -1.
5307 hpsa_turn_off_ioaccel_for_device(dev);
5396 c->phys_disk = dev->phys_disk[map_index];
5447 dev->scsi3addr,
5448 dev->phys_disk[map_index]);
5458 struct hpsa_scsi_dev_t *dev)
5464 memcpy(&c->Header.LUN.LunAddrBytes[0], &dev->scsi3addr[0], 8);
5505 dev_err(&h->pdev->dev, "unknown data direction: %d\n",
5516 if (dev->in_reset) {
5521 c->device = dev;
5577 struct hpsa_scsi_dev_t *dev = cmd->device->hostdata;
5580 if (!dev)
5583 if (dev->in_reset)
5591 if (dev->offload_enabled) {
5595 c->device = dev;
5601 } else if (dev->hba_ioaccel_enabled) {
5605 c->device = dev;
5618 struct hpsa_scsi_dev_t *dev;
5622 dev = cmd->device->hostdata;
5623 if (!dev) {
5628 if (dev->in_reset) {
5666 if (hpsa_ciss_submit(c->h, c, cmd, dev)) {
5684 struct hpsa_scsi_dev_t *dev;
5693 dev = cmd->device->hostdata;
5694 if (!dev) {
5700 if (dev->removed) {
5712 if (dev->in_reset)
5745 return hpsa_ciss_submit(h, c, cmd, dev);
5855 dev_err(&h->pdev->dev, "scsi_host_alloc failed\n");
5882 rv = scsi_add_host(h->scsi_host, &h->pdev->dev);
5884 dev_err(&h->pdev->dev, "scsi_add_host failed\n");
5973 dev_warn(&h->pdev->dev,
6013 dev_warn(&h->pdev->dev, "giving up on device.\n");
6015 dev_warn(&h->pdev->dev, "device is ready.\n");
6029 struct hpsa_scsi_dev_t *dev = NULL;
6048 dev = scsicmd->device->hostdata;
6049 if (!dev) {
6050 dev_err(&h->pdev->dev, "%s: device lookup failed\n", __func__);
6055 if (dev->devtype == TYPE_ENCLOSURE) {
6065 hpsa_show_dev_msg(KERN_WARNING, h, dev, msg);
6075 hpsa_show_dev_msg(KERN_WARNING, h, dev, msg);
6081 if (is_hba_lunid(dev->scsi3addr)) {
6086 if (is_logical_dev_addr_mode(dev->scsi3addr))
6093 hpsa_show_dev_msg(KERN_WARNING, h, dev, msg);
6098 dev->in_reset = true; /* block any new cmds from OS for this device */
6100 if (atomic_read(&dev->commands_outstanding) > 0)
6107 rc = hpsa_do_reset(h, dev, reset_type, DEFAULT_REPLY_QUEUE);
6116 hpsa_show_dev_msg(KERN_WARNING, h, dev, msg);
6121 if (dev)
6122 dev->in_reset = false;
6141 dev_err(&h->pdev->dev, "Bad block tag: %d not in [%d..%d]\n",
6157 dev_warn(&h->pdev->dev,
6269 static int hpsa_ioctl32_passthru(struct scsi_device *dev, unsigned int cmd,
6272 struct ctlr_info *h = sdev_to_hba(dev);
6300 static int hpsa_ioctl32_big_passthru(struct scsi_device *dev,
6303 struct ctlr_info *h = sdev_to_hba(dev);
6331 static int hpsa_compat_ioctl(struct scsi_device *dev, unsigned int cmd,
6350 return hpsa_ioctl(dev, cmd, arg);
6353 return hpsa_ioctl32_passthru(dev, cmd, arg);
6355 return hpsa_ioctl32_big_passthru(dev, cmd, arg);
6387 dev_info(&h->pdev->dev, "driver version string '%s' "
6452 temp64 = dma_map_single(&h->pdev->dev, buff,
6454 if (dma_mapping_error(&h->pdev->dev, (dma_addr_t) temp64)) {
6559 temp64 = dma_map_single(&h->pdev->dev, buff[i],
6561 if (dma_mapping_error(&h->pdev->dev,
6627 static int hpsa_ioctl(struct scsi_device *dev, unsigned int cmd,
6630 struct ctlr_info *h = sdev_to_hba(dev);
6859 dev_warn(&h->pdev->dev, "unknown command 0x%c\n", cmd);
6895 dev_warn(&h->pdev->dev, "unknown message type %d\n",
6900 dev_warn(&h->pdev->dev, "unknown command type %d\n", cmd_type);
6955 dev_warn(&h->pdev->dev, "bad tag 0x%08x ignored.\n", raw_tag);
6998 dev_info(&h->pdev->dev, "Received interrupt while interrupts disabled "
7113 err = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32));
7119 cmd = dma_alloc_coherent(&pdev->dev, cmd_sz, &paddr64, GFP_KERNEL);
7163 dev_err(&pdev->dev, "controller message %02x:%02x timed out\n",
7168 dma_free_coherent(&pdev->dev, cmd_sz, cmd, paddr64);
7171 dev_err(&pdev->dev, "controller message %02x:%02x failed\n",
7176 dev_info(&pdev->dev, "controller message %02x:%02x succeeded\n",
7192 dev_info(&pdev->dev, "using doorbell to reset controller\n");
7213 dev_info(&pdev->dev, "using PCI PM to reset controller\n");
7318 dev_warn(&pdev->dev, "Controller not resettable\n");
7363 dev_warn(&pdev->dev,
7383 dev_warn(&pdev->dev,
7392 dev_warn(&pdev->dev, "Unable to successfully reset "
7396 dev_info(&pdev->dev, "board ready after hard reset.\n");
7412 static void print_cfg_table(struct device *dev, struct CfgTable __iomem *tb)
7418 dev_info(dev, "Controller Configuration information\n");
7419 dev_info(dev, "------------------------------------\n");
7423 dev_info(dev, " Signature = %s\n", temp_name);
7424 dev_info(dev, " Spec Number = %d\n", readl(&(tb->SpecValence)));
7425 dev_info(dev, " Transport methods supported = 0x%x\n",
7427 dev_info(dev, " Transport methods active = 0x%x\n",
7429 dev_info(dev, " Requested transport Method = 0x%x\n",
7431 dev_info(dev, " Coalesce Interrupt Delay = 0x%x\n",
7433 dev_info(dev, " Coalesce Interrupt Count = 0x%x\n",
7435 dev_info(dev, " Max outstanding commands = %d\n",
7437 dev_info(dev, " Bus Types = 0x%x\n", readl(&(tb->BusTypes)));
7441 dev_info(dev, " Server Name = %s\n", temp_name);
7442 dev_info(dev, " Heartbeat Counter = 0x%x\n\n\n",
7470 dev_warn(&pdev->dev,
7558 dev_warn(&pdev->dev,
7566 dev_warn(&pdev->dev, "unrecognized board ID: 0x%08x\n", *board_id);
7581 dev_dbg(&pdev->dev, "memory BAR = %lx\n",
7585 dev_warn(&pdev->dev, "no memory BAR found\n");
7610 dev_warn(&pdev->dev, "board not ready, timed out.\n");
7623 dev_warn(&pdev->dev, "cannot find cfg_base_addr_index\n");
7659 dev_err(&h->pdev->dev, "Failed mapping cfgtable\n");
7671 dev_err(&h->pdev->dev, "Failed mapping transfer table\n");
7690 dev_warn(&h->pdev->dev,
7736 dev_warn(&h->pdev->dev, "Physical aborts not supported\n");
7738 dev_warn(&h->pdev->dev, "Logical aborts not supported\n");
7740 dev_warn(&h->pdev->dev, "HP SSD Smart Path aborts not supported\n");
7746 dev_err(&h->pdev->dev, "not a valid CISS config table\n");
7842 print_cfg_table(&h->pdev->dev, h->cfgtable);
7848 dev_err(&h->pdev->dev, "failed to enter simple mode\n");
7884 dev_err(&h->pdev->dev, "failed to enable PCI device\n");
7891 dev_err(&h->pdev->dev,
7911 dev_err(&h->pdev->dev, "failed to remap PCI mem\n");
7976 * the pci interface. The dev->enable_cnt is equal zero
7981 dev_warn(&pdev->dev, "Failed to enable PCI device\n");
7988 dev_warn(&pdev->dev, "failed to enable device.\n");
8014 dev_info(&pdev->dev, "Waiting for controller to respond to no-op\n");
8019 dev_warn(&pdev->dev, "no-op failed%s\n",
8034 dma_free_coherent(&h->pdev->dev,
8042 dma_free_coherent(&h->pdev->dev,
8054 h->cmd_pool = dma_alloc_coherent(&h->pdev->dev,
8057 h->errinfo_pool = dma_alloc_coherent(&h->pdev->dev,
8063 dev_err(&h->pdev->dev, "out of memory in %s", __func__);
8126 dev_err(&h->pdev->dev,
8157 dev_err(&h->pdev->dev, "failed to get irq %d for %s\n",
8170 dev_info(&h->pdev->dev, "Waiting for board to soft reset.\n");
8173 dev_warn(&h->pdev->dev, "Soft reset had no effect.\n");
8177 dev_info(&h->pdev->dev, "Board reset, awaiting READY status.\n");
8180 dev_warn(&h->pdev->dev, "Board failed to become ready "
8195 dma_free_coherent(&h->pdev->dev,
8251 dev_warn(&h->pdev->dev,
8277 dev_warn(&h->pdev->dev,
8284 dev_warn(&h->pdev->dev, "Controller lockup detected: 0x%08x after %d\n",
8287 dev_warn(&h->pdev->dev, "Telling controller to do a CHKPT\n");
8361 device = h->dev[i];
8434 dev_warn(&h->pdev->dev,
8514 dev_warn(&h->pdev->dev,
8519 dev_info(&h->pdev->dev,
8599 dev_info(&h->pdev->dev,
8635 dev_err(&h->pdev->dev, "failed to create %s workqueue\n", name);
8675 dev_warn(&pdev->dev, "Board ID not found\n");
8701 dev_err(&pdev->dev, "Failed to allocate controller head\n");
8718 dev_err(&h->pdev->dev, "Failed to allocate lockup detector\n");
8739 rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(64));
8741 rc = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
8743 dev_err(&pdev->dev, "no suitable DMA available\n");
8814 dev_warn(&h->pdev->dev,
8835 dev_info(&h->pdev->dev, "Board READY.\n");
8836 dev_info(&h->pdev->dev,
8844 dev_info(&h->pdev->dev,
8873 dev_info(&h->pdev->dev,
8956 dev_warn(&h->pdev->dev,
9017 dev_err(&h->pdev->dev,
9050 kfree(h->dev[i]);
9051 h->dev[i] = NULL;
9061 dev_err(&pdev->dev, "unable to remove device\n");
9116 __attribute__((unused)) struct device *dev)
9122 (__attribute__((unused)) struct device *dev)
9265 dev_err(&h->pdev->dev,
9271 dev_err(&h->pdev->dev,
9345 dev_err(&h->pdev->dev,
9356 dma_free_coherent(&h->pdev->dev,
9382 dma_alloc_coherent(&h->pdev->dev,
9409 dma_free_coherent(&h->pdev->dev,
9435 dma_alloc_coherent(&h->pdev->dev,
9504 h->reply_queue[i].head = dma_alloc_coherent(&h->pdev->dev,
9737 device = h->dev[i];