1
2
3
4
5
6#include "qla_def.h"
7#include "qla_target.h"
8
9#include <linux/kthread.h>
10#include <linux/vmalloc.h>
11#include <linux/slab.h>
12#include <linux/delay.h>
13
14static int qla24xx_vport_disable(struct fc_vport *, bool);
15
16
17
18static ssize_t
19qla2x00_sysfs_read_fw_dump(struct file *filp, struct kobject *kobj,
20 struct bin_attribute *bin_attr,
21 char *buf, loff_t off, size_t count)
22{
23 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
24 struct device, kobj)));
25 struct qla_hw_data *ha = vha->hw;
26 int rval = 0;
27
28 if (!(ha->fw_dump_reading || ha->mctp_dump_reading ||
29 ha->mpi_fw_dump_reading))
30 return 0;
31
32 mutex_lock(&ha->optrom_mutex);
33 if (IS_P3P_TYPE(ha)) {
34 if (off < ha->md_template_size) {
35 rval = memory_read_from_buffer(buf, count,
36 &off, ha->md_tmplt_hdr, ha->md_template_size);
37 } else {
38 off -= ha->md_template_size;
39 rval = memory_read_from_buffer(buf, count,
40 &off, ha->md_dump, ha->md_dump_size);
41 }
42 } else if (ha->mctp_dumped && ha->mctp_dump_reading) {
43 rval = memory_read_from_buffer(buf, count, &off, ha->mctp_dump,
44 MCTP_DUMP_SIZE);
45 } else if (ha->mpi_fw_dumped && ha->mpi_fw_dump_reading) {
46 rval = memory_read_from_buffer(buf, count, &off,
47 ha->mpi_fw_dump,
48 ha->mpi_fw_dump_len);
49 } else if (ha->fw_dump_reading) {
50 rval = memory_read_from_buffer(buf, count, &off, ha->fw_dump,
51 ha->fw_dump_len);
52 } else {
53 rval = 0;
54 }
55 mutex_unlock(&ha->optrom_mutex);
56 return rval;
57}
58
59static ssize_t
60qla2x00_sysfs_write_fw_dump(struct file *filp, struct kobject *kobj,
61 struct bin_attribute *bin_attr,
62 char *buf, loff_t off, size_t count)
63{
64 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
65 struct device, kobj)));
66 struct qla_hw_data *ha = vha->hw;
67 int reading;
68
69 if (off != 0)
70 return (0);
71
72 reading = simple_strtol(buf, NULL, 10);
73 switch (reading) {
74 case 0:
75 if (!ha->fw_dump_reading)
76 break;
77
78 ql_log(ql_log_info, vha, 0x705d,
79 "Firmware dump cleared on (%ld).\n", vha->host_no);
80
81 if (IS_P3P_TYPE(ha)) {
82 qla82xx_md_free(vha);
83 qla82xx_md_prep(vha);
84 }
85 ha->fw_dump_reading = 0;
86 ha->fw_dumped = false;
87 break;
88 case 1:
89 if (ha->fw_dumped && !ha->fw_dump_reading) {
90 ha->fw_dump_reading = 1;
91
92 ql_log(ql_log_info, vha, 0x705e,
93 "Raw firmware dump ready for read on (%ld).\n",
94 vha->host_no);
95 }
96 break;
97 case 2:
98 qla2x00_alloc_fw_dump(vha);
99 break;
100 case 3:
101 if (IS_QLA82XX(ha)) {
102 qla82xx_idc_lock(ha);
103 qla82xx_set_reset_owner(vha);
104 qla82xx_idc_unlock(ha);
105 } else if (IS_QLA8044(ha)) {
106 qla8044_idc_lock(ha);
107 qla82xx_set_reset_owner(vha);
108 qla8044_idc_unlock(ha);
109 } else {
110 qla2x00_system_error(vha);
111 }
112 break;
113 case 4:
114 if (IS_P3P_TYPE(ha)) {
115 if (ha->md_tmplt_hdr)
116 ql_dbg(ql_dbg_user, vha, 0x705b,
117 "MiniDump supported with this firmware.\n");
118 else
119 ql_dbg(ql_dbg_user, vha, 0x709d,
120 "MiniDump not supported with this firmware.\n");
121 }
122 break;
123 case 5:
124 if (IS_P3P_TYPE(ha))
125 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
126 break;
127 case 6:
128 if (!ha->mctp_dump_reading)
129 break;
130 ql_log(ql_log_info, vha, 0x70c1,
131 "MCTP dump cleared on (%ld).\n", vha->host_no);
132 ha->mctp_dump_reading = 0;
133 ha->mctp_dumped = 0;
134 break;
135 case 7:
136 if (ha->mctp_dumped && !ha->mctp_dump_reading) {
137 ha->mctp_dump_reading = 1;
138 ql_log(ql_log_info, vha, 0x70c2,
139 "Raw mctp dump ready for read on (%ld).\n",
140 vha->host_no);
141 }
142 break;
143 case 8:
144 if (!ha->mpi_fw_dump_reading)
145 break;
146 ql_log(ql_log_info, vha, 0x70e7,
147 "MPI firmware dump cleared on (%ld).\n", vha->host_no);
148 ha->mpi_fw_dump_reading = 0;
149 ha->mpi_fw_dumped = 0;
150 break;
151 case 9:
152 if (ha->mpi_fw_dumped && !ha->mpi_fw_dump_reading) {
153 ha->mpi_fw_dump_reading = 1;
154 ql_log(ql_log_info, vha, 0x70e8,
155 "Raw MPI firmware dump ready for read on (%ld).\n",
156 vha->host_no);
157 }
158 break;
159 case 10:
160 if (IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
161 ql_log(ql_log_info, vha, 0x70e9,
162 "Issuing MPI firmware dump on host#%ld.\n",
163 vha->host_no);
164 ha->isp_ops->mpi_fw_dump(vha, 0);
165 }
166 break;
167 }
168 return count;
169}
170
171static struct bin_attribute sysfs_fw_dump_attr = {
172 .attr = {
173 .name = "fw_dump",
174 .mode = S_IRUSR | S_IWUSR,
175 },
176 .size = 0,
177 .read = qla2x00_sysfs_read_fw_dump,
178 .write = qla2x00_sysfs_write_fw_dump,
179};
180
181static ssize_t
182qla2x00_sysfs_read_nvram(struct file *filp, struct kobject *kobj,
183 struct bin_attribute *bin_attr,
184 char *buf, loff_t off, size_t count)
185{
186 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
187 struct device, kobj)));
188 struct qla_hw_data *ha = vha->hw;
189 uint32_t faddr;
190 struct active_regions active_regions = { };
191
192 if (!capable(CAP_SYS_ADMIN))
193 return 0;
194
195 mutex_lock(&ha->optrom_mutex);
196 if (qla2x00_chip_is_down(vha)) {
197 mutex_unlock(&ha->optrom_mutex);
198 return -EAGAIN;
199 }
200
201 if (!IS_NOCACHE_VPD_TYPE(ha)) {
202 mutex_unlock(&ha->optrom_mutex);
203 goto skip;
204 }
205
206 faddr = ha->flt_region_nvram;
207 if (IS_QLA28XX(ha)) {
208 qla28xx_get_aux_images(vha, &active_regions);
209 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
210 faddr = ha->flt_region_nvram_sec;
211 }
212 ha->isp_ops->read_optrom(vha, ha->nvram, faddr << 2, ha->nvram_size);
213
214 mutex_unlock(&ha->optrom_mutex);
215
216skip:
217 return memory_read_from_buffer(buf, count, &off, ha->nvram,
218 ha->nvram_size);
219}
220
221static ssize_t
222qla2x00_sysfs_write_nvram(struct file *filp, struct kobject *kobj,
223 struct bin_attribute *bin_attr,
224 char *buf, loff_t off, size_t count)
225{
226 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
227 struct device, kobj)));
228 struct qla_hw_data *ha = vha->hw;
229 uint16_t cnt;
230
231 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->nvram_size ||
232 !ha->isp_ops->write_nvram)
233 return -EINVAL;
234
235
236 if (IS_FWI2_CAPABLE(ha)) {
237 __le32 *iter = (__force __le32 *)buf;
238 uint32_t chksum;
239
240 chksum = 0;
241 for (cnt = 0; cnt < ((count >> 2) - 1); cnt++, iter++)
242 chksum += le32_to_cpu(*iter);
243 chksum = ~chksum + 1;
244 *iter = cpu_to_le32(chksum);
245 } else {
246 uint8_t *iter;
247 uint8_t chksum;
248
249 iter = (uint8_t *)buf;
250 chksum = 0;
251 for (cnt = 0; cnt < count - 1; cnt++)
252 chksum += *iter++;
253 chksum = ~chksum + 1;
254 *iter = chksum;
255 }
256
257 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
258 ql_log(ql_log_warn, vha, 0x705f,
259 "HBA not online, failing NVRAM update.\n");
260 return -EAGAIN;
261 }
262
263 mutex_lock(&ha->optrom_mutex);
264 if (qla2x00_chip_is_down(vha)) {
265 mutex_unlock(&ha->optrom_mutex);
266 return -EAGAIN;
267 }
268
269
270 ha->isp_ops->write_nvram(vha, buf, ha->nvram_base, count);
271 ha->isp_ops->read_nvram(vha, ha->nvram, ha->nvram_base,
272 count);
273 mutex_unlock(&ha->optrom_mutex);
274
275 ql_dbg(ql_dbg_user, vha, 0x7060,
276 "Setting ISP_ABORT_NEEDED\n");
277
278 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
279 qla2xxx_wake_dpc(vha);
280 qla2x00_wait_for_chip_reset(vha);
281
282 return count;
283}
284
285static struct bin_attribute sysfs_nvram_attr = {
286 .attr = {
287 .name = "nvram",
288 .mode = S_IRUSR | S_IWUSR,
289 },
290 .size = 512,
291 .read = qla2x00_sysfs_read_nvram,
292 .write = qla2x00_sysfs_write_nvram,
293};
294
295static ssize_t
296qla2x00_sysfs_read_optrom(struct file *filp, struct kobject *kobj,
297 struct bin_attribute *bin_attr,
298 char *buf, loff_t off, size_t count)
299{
300 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
301 struct device, kobj)));
302 struct qla_hw_data *ha = vha->hw;
303 ssize_t rval = 0;
304
305 mutex_lock(&ha->optrom_mutex);
306
307 if (ha->optrom_state != QLA_SREADING)
308 goto out;
309
310 rval = memory_read_from_buffer(buf, count, &off, ha->optrom_buffer,
311 ha->optrom_region_size);
312
313out:
314 mutex_unlock(&ha->optrom_mutex);
315
316 return rval;
317}
318
319static ssize_t
320qla2x00_sysfs_write_optrom(struct file *filp, struct kobject *kobj,
321 struct bin_attribute *bin_attr,
322 char *buf, loff_t off, size_t count)
323{
324 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
325 struct device, kobj)));
326 struct qla_hw_data *ha = vha->hw;
327
328 mutex_lock(&ha->optrom_mutex);
329
330 if (ha->optrom_state != QLA_SWRITING) {
331 mutex_unlock(&ha->optrom_mutex);
332 return -EINVAL;
333 }
334 if (off > ha->optrom_region_size) {
335 mutex_unlock(&ha->optrom_mutex);
336 return -ERANGE;
337 }
338 if (off + count > ha->optrom_region_size)
339 count = ha->optrom_region_size - off;
340
341 memcpy(&ha->optrom_buffer[off], buf, count);
342 mutex_unlock(&ha->optrom_mutex);
343
344 return count;
345}
346
347static struct bin_attribute sysfs_optrom_attr = {
348 .attr = {
349 .name = "optrom",
350 .mode = S_IRUSR | S_IWUSR,
351 },
352 .size = 0,
353 .read = qla2x00_sysfs_read_optrom,
354 .write = qla2x00_sysfs_write_optrom,
355};
356
357static ssize_t
358qla2x00_sysfs_write_optrom_ctl(struct file *filp, struct kobject *kobj,
359 struct bin_attribute *bin_attr,
360 char *buf, loff_t off, size_t count)
361{
362 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
363 struct device, kobj)));
364 struct qla_hw_data *ha = vha->hw;
365 uint32_t start = 0;
366 uint32_t size = ha->optrom_size;
367 int val, valid;
368 ssize_t rval = count;
369
370 if (off)
371 return -EINVAL;
372
373 if (unlikely(pci_channel_offline(ha->pdev)))
374 return -EAGAIN;
375
376 if (sscanf(buf, "%d:%x:%x", &val, &start, &size) < 1)
377 return -EINVAL;
378 if (start > ha->optrom_size)
379 return -EINVAL;
380 if (size > ha->optrom_size - start)
381 size = ha->optrom_size - start;
382
383 mutex_lock(&ha->optrom_mutex);
384 if (qla2x00_chip_is_down(vha)) {
385 mutex_unlock(&ha->optrom_mutex);
386 return -EAGAIN;
387 }
388 switch (val) {
389 case 0:
390 if (ha->optrom_state != QLA_SREADING &&
391 ha->optrom_state != QLA_SWRITING) {
392 rval = -EINVAL;
393 goto out;
394 }
395 ha->optrom_state = QLA_SWAITING;
396
397 ql_dbg(ql_dbg_user, vha, 0x7061,
398 "Freeing flash region allocation -- 0x%x bytes.\n",
399 ha->optrom_region_size);
400
401 vfree(ha->optrom_buffer);
402 ha->optrom_buffer = NULL;
403 break;
404 case 1:
405 if (ha->optrom_state != QLA_SWAITING) {
406 rval = -EINVAL;
407 goto out;
408 }
409
410 ha->optrom_region_start = start;
411 ha->optrom_region_size = size;
412
413 ha->optrom_state = QLA_SREADING;
414 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
415 if (ha->optrom_buffer == NULL) {
416 ql_log(ql_log_warn, vha, 0x7062,
417 "Unable to allocate memory for optrom retrieval "
418 "(%x).\n", ha->optrom_region_size);
419
420 ha->optrom_state = QLA_SWAITING;
421 rval = -ENOMEM;
422 goto out;
423 }
424
425 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
426 ql_log(ql_log_warn, vha, 0x7063,
427 "HBA not online, failing NVRAM update.\n");
428 rval = -EAGAIN;
429 goto out;
430 }
431
432 ql_dbg(ql_dbg_user, vha, 0x7064,
433 "Reading flash region -- 0x%x/0x%x.\n",
434 ha->optrom_region_start, ha->optrom_region_size);
435
436 ha->isp_ops->read_optrom(vha, ha->optrom_buffer,
437 ha->optrom_region_start, ha->optrom_region_size);
438 break;
439 case 2:
440 if (ha->optrom_state != QLA_SWAITING) {
441 rval = -EINVAL;
442 goto out;
443 }
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469 valid = 0;
470 if (ha->optrom_size == OPTROM_SIZE_2300 && start == 0)
471 valid = 1;
472 else if (IS_QLA24XX_TYPE(ha) || IS_QLA25XX(ha))
473 valid = 1;
474 if (!valid) {
475 ql_log(ql_log_warn, vha, 0x7065,
476 "Invalid start region 0x%x/0x%x.\n", start, size);
477 rval = -EINVAL;
478 goto out;
479 }
480
481 ha->optrom_region_start = start;
482 ha->optrom_region_size = size;
483
484 ha->optrom_state = QLA_SWRITING;
485 ha->optrom_buffer = vzalloc(ha->optrom_region_size);
486 if (ha->optrom_buffer == NULL) {
487 ql_log(ql_log_warn, vha, 0x7066,
488 "Unable to allocate memory for optrom update "
489 "(%x)\n", ha->optrom_region_size);
490
491 ha->optrom_state = QLA_SWAITING;
492 rval = -ENOMEM;
493 goto out;
494 }
495
496 ql_dbg(ql_dbg_user, vha, 0x7067,
497 "Staging flash region write -- 0x%x/0x%x.\n",
498 ha->optrom_region_start, ha->optrom_region_size);
499
500 break;
501 case 3:
502 if (ha->optrom_state != QLA_SWRITING) {
503 rval = -EINVAL;
504 goto out;
505 }
506
507 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
508 ql_log(ql_log_warn, vha, 0x7068,
509 "HBA not online, failing flash update.\n");
510 rval = -EAGAIN;
511 goto out;
512 }
513
514 ql_dbg(ql_dbg_user, vha, 0x7069,
515 "Writing flash region -- 0x%x/0x%x.\n",
516 ha->optrom_region_start, ha->optrom_region_size);
517
518 rval = ha->isp_ops->write_optrom(vha, ha->optrom_buffer,
519 ha->optrom_region_start, ha->optrom_region_size);
520 if (rval)
521 rval = -EIO;
522 break;
523 default:
524 rval = -EINVAL;
525 }
526
527out:
528 mutex_unlock(&ha->optrom_mutex);
529 return rval;
530}
531
532static struct bin_attribute sysfs_optrom_ctl_attr = {
533 .attr = {
534 .name = "optrom_ctl",
535 .mode = S_IWUSR,
536 },
537 .size = 0,
538 .write = qla2x00_sysfs_write_optrom_ctl,
539};
540
541static ssize_t
542qla2x00_sysfs_read_vpd(struct file *filp, struct kobject *kobj,
543 struct bin_attribute *bin_attr,
544 char *buf, loff_t off, size_t count)
545{
546 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
547 struct device, kobj)));
548 struct qla_hw_data *ha = vha->hw;
549 uint32_t faddr;
550 struct active_regions active_regions = { };
551
552 if (unlikely(pci_channel_offline(ha->pdev)))
553 return -EAGAIN;
554
555 if (!capable(CAP_SYS_ADMIN))
556 return -EINVAL;
557
558 if (IS_NOCACHE_VPD_TYPE(ha))
559 goto skip;
560
561 faddr = ha->flt_region_vpd << 2;
562
563 if (IS_QLA28XX(ha)) {
564 qla28xx_get_aux_images(vha, &active_regions);
565 if (active_regions.aux.vpd_nvram == QLA27XX_SECONDARY_IMAGE)
566 faddr = ha->flt_region_vpd_sec << 2;
567
568 ql_dbg(ql_dbg_init, vha, 0x7070,
569 "Loading %s nvram image.\n",
570 active_regions.aux.vpd_nvram == QLA27XX_PRIMARY_IMAGE ?
571 "primary" : "secondary");
572 }
573
574 mutex_lock(&ha->optrom_mutex);
575 if (qla2x00_chip_is_down(vha)) {
576 mutex_unlock(&ha->optrom_mutex);
577 return -EAGAIN;
578 }
579
580 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
581 mutex_unlock(&ha->optrom_mutex);
582
583 ha->isp_ops->read_optrom(vha, ha->vpd, faddr, ha->vpd_size);
584skip:
585 return memory_read_from_buffer(buf, count, &off, ha->vpd, ha->vpd_size);
586}
587
588static ssize_t
589qla2x00_sysfs_write_vpd(struct file *filp, struct kobject *kobj,
590 struct bin_attribute *bin_attr,
591 char *buf, loff_t off, size_t count)
592{
593 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
594 struct device, kobj)));
595 struct qla_hw_data *ha = vha->hw;
596 uint8_t *tmp_data;
597
598 if (unlikely(pci_channel_offline(ha->pdev)))
599 return 0;
600
601 if (qla2x00_chip_is_down(vha))
602 return 0;
603
604 if (!capable(CAP_SYS_ADMIN) || off != 0 || count != ha->vpd_size ||
605 !ha->isp_ops->write_nvram)
606 return 0;
607
608 if (qla2x00_wait_for_hba_online(vha) != QLA_SUCCESS) {
609 ql_log(ql_log_warn, vha, 0x706a,
610 "HBA not online, failing VPD update.\n");
611 return -EAGAIN;
612 }
613
614 mutex_lock(&ha->optrom_mutex);
615 if (qla2x00_chip_is_down(vha)) {
616 mutex_unlock(&ha->optrom_mutex);
617 return -EAGAIN;
618 }
619
620
621 ha->isp_ops->write_nvram(vha, buf, ha->vpd_base, count);
622 ha->isp_ops->read_nvram(vha, ha->vpd, ha->vpd_base, count);
623
624
625 if (!IS_FWI2_CAPABLE(ha)) {
626 mutex_unlock(&ha->optrom_mutex);
627 return -EINVAL;
628 }
629
630 tmp_data = vmalloc(256);
631 if (!tmp_data) {
632 mutex_unlock(&ha->optrom_mutex);
633 ql_log(ql_log_warn, vha, 0x706b,
634 "Unable to allocate memory for VPD information update.\n");
635 return -ENOMEM;
636 }
637 ha->isp_ops->get_flash_version(vha, tmp_data);
638 vfree(tmp_data);
639
640 mutex_unlock(&ha->optrom_mutex);
641
642 return count;
643}
644
645static struct bin_attribute sysfs_vpd_attr = {
646 .attr = {
647 .name = "vpd",
648 .mode = S_IRUSR | S_IWUSR,
649 },
650 .size = 0,
651 .read = qla2x00_sysfs_read_vpd,
652 .write = qla2x00_sysfs_write_vpd,
653};
654
655static ssize_t
656qla2x00_sysfs_read_sfp(struct file *filp, struct kobject *kobj,
657 struct bin_attribute *bin_attr,
658 char *buf, loff_t off, size_t count)
659{
660 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
661 struct device, kobj)));
662 int rval;
663
664 if (!capable(CAP_SYS_ADMIN) || off != 0 || count < SFP_DEV_SIZE)
665 return 0;
666
667 mutex_lock(&vha->hw->optrom_mutex);
668 if (qla2x00_chip_is_down(vha)) {
669 mutex_unlock(&vha->hw->optrom_mutex);
670 return 0;
671 }
672
673 rval = qla2x00_read_sfp_dev(vha, buf, count);
674 mutex_unlock(&vha->hw->optrom_mutex);
675
676 if (rval)
677 return -EIO;
678
679 return count;
680}
681
682static struct bin_attribute sysfs_sfp_attr = {
683 .attr = {
684 .name = "sfp",
685 .mode = S_IRUSR | S_IWUSR,
686 },
687 .size = SFP_DEV_SIZE,
688 .read = qla2x00_sysfs_read_sfp,
689};
690
691static ssize_t
692qla2x00_sysfs_write_reset(struct file *filp, struct kobject *kobj,
693 struct bin_attribute *bin_attr,
694 char *buf, loff_t off, size_t count)
695{
696 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
697 struct device, kobj)));
698 struct qla_hw_data *ha = vha->hw;
699 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
700 int type;
701 uint32_t idc_control;
702 uint8_t *tmp_data = NULL;
703
704 if (off != 0)
705 return -EINVAL;
706
707 type = simple_strtol(buf, NULL, 10);
708 switch (type) {
709 case 0x2025c:
710 ql_log(ql_log_info, vha, 0x706e,
711 "Issuing ISP reset.\n");
712
713 if (vha->hw->flags.port_isolated) {
714 ql_log(ql_log_info, vha, 0x706e,
715 "Port is isolated, returning.\n");
716 return -EINVAL;
717 }
718
719 scsi_block_requests(vha->host);
720 if (IS_QLA82XX(ha)) {
721 ha->flags.isp82xx_no_md_cap = 1;
722 qla82xx_idc_lock(ha);
723 qla82xx_set_reset_owner(vha);
724 qla82xx_idc_unlock(ha);
725 } else if (IS_QLA8044(ha)) {
726 qla8044_idc_lock(ha);
727 idc_control = qla8044_rd_reg(ha,
728 QLA8044_IDC_DRV_CTRL);
729 qla8044_wr_reg(ha, QLA8044_IDC_DRV_CTRL,
730 (idc_control | GRACEFUL_RESET_BIT1));
731 qla82xx_set_reset_owner(vha);
732 qla8044_idc_unlock(ha);
733 } else {
734 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
735 qla2xxx_wake_dpc(vha);
736 }
737 qla2x00_wait_for_chip_reset(vha);
738 scsi_unblock_requests(vha->host);
739 break;
740 case 0x2025d:
741 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
742 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
743 return -EPERM;
744
745 ql_log(ql_log_info, vha, 0x706f,
746 "Issuing MPI reset.\n");
747
748 if (IS_QLA83XX(ha) || IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
749 uint32_t idc_control;
750
751 qla83xx_idc_lock(vha, 0);
752 __qla83xx_get_idc_control(vha, &idc_control);
753 idc_control |= QLA83XX_IDC_GRACEFUL_RESET;
754 __qla83xx_set_idc_control(vha, idc_control);
755 qla83xx_wr_reg(vha, QLA83XX_IDC_DEV_STATE,
756 QLA8XXX_DEV_NEED_RESET);
757 qla83xx_idc_audit(vha, IDC_AUDIT_TIMESTAMP);
758 qla83xx_idc_unlock(vha, 0);
759 break;
760 } else {
761
762 WARN_ON_ONCE(qla2x00_wait_for_hba_online(vha) !=
763 QLA_SUCCESS);
764
765
766 scsi_block_requests(vha->host);
767 if (qla81xx_restart_mpi_firmware(vha) != QLA_SUCCESS)
768 ql_log(ql_log_warn, vha, 0x7070,
769 "MPI reset failed.\n");
770 scsi_unblock_requests(vha->host);
771 break;
772 }
773 break;
774 case 0x2025e:
775 if (!IS_P3P_TYPE(ha) || vha != base_vha) {
776 ql_log(ql_log_info, vha, 0x7071,
777 "FCoE ctx reset not supported.\n");
778 return -EPERM;
779 }
780
781 ql_log(ql_log_info, vha, 0x7072,
782 "Issuing FCoE ctx reset.\n");
783 set_bit(FCOE_CTX_RESET_NEEDED, &vha->dpc_flags);
784 qla2xxx_wake_dpc(vha);
785 qla2x00_wait_for_fcoe_ctx_reset(vha);
786 break;
787 case 0x2025f:
788 if (!IS_QLA8031(ha))
789 return -EPERM;
790 ql_log(ql_log_info, vha, 0x70bc,
791 "Disabling Reset by IDC control\n");
792 qla83xx_idc_lock(vha, 0);
793 __qla83xx_get_idc_control(vha, &idc_control);
794 idc_control |= QLA83XX_IDC_RESET_DISABLED;
795 __qla83xx_set_idc_control(vha, idc_control);
796 qla83xx_idc_unlock(vha, 0);
797 break;
798 case 0x20260:
799 if (!IS_QLA8031(ha))
800 return -EPERM;
801 ql_log(ql_log_info, vha, 0x70bd,
802 "Enabling Reset by IDC control\n");
803 qla83xx_idc_lock(vha, 0);
804 __qla83xx_get_idc_control(vha, &idc_control);
805 idc_control &= ~QLA83XX_IDC_RESET_DISABLED;
806 __qla83xx_set_idc_control(vha, idc_control);
807 qla83xx_idc_unlock(vha, 0);
808 break;
809 case 0x20261:
810 ql_dbg(ql_dbg_user, vha, 0x70e0,
811 "Updating cache versions without reset ");
812
813 tmp_data = vmalloc(256);
814 if (!tmp_data) {
815 ql_log(ql_log_warn, vha, 0x70e1,
816 "Unable to allocate memory for VPD information update.\n");
817 return -ENOMEM;
818 }
819 ha->isp_ops->get_flash_version(vha, tmp_data);
820 vfree(tmp_data);
821 break;
822 }
823 return count;
824}
825
826static struct bin_attribute sysfs_reset_attr = {
827 .attr = {
828 .name = "reset",
829 .mode = S_IWUSR,
830 },
831 .size = 0,
832 .write = qla2x00_sysfs_write_reset,
833};
834
835static ssize_t
836qla2x00_issue_logo(struct file *filp, struct kobject *kobj,
837 struct bin_attribute *bin_attr,
838 char *buf, loff_t off, size_t count)
839{
840 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
841 struct device, kobj)));
842 int type;
843 port_id_t did;
844
845 if (!capable(CAP_SYS_ADMIN))
846 return 0;
847
848 if (unlikely(pci_channel_offline(vha->hw->pdev)))
849 return 0;
850
851 if (qla2x00_chip_is_down(vha))
852 return 0;
853
854 type = simple_strtol(buf, NULL, 10);
855
856 did.b.domain = (type & 0x00ff0000) >> 16;
857 did.b.area = (type & 0x0000ff00) >> 8;
858 did.b.al_pa = (type & 0x000000ff);
859
860 ql_log(ql_log_info, vha, 0xd04d, "portid=%02x%02x%02x done\n",
861 did.b.domain, did.b.area, did.b.al_pa);
862
863 ql_log(ql_log_info, vha, 0x70e4, "%s: %d\n", __func__, type);
864
865 qla24xx_els_dcmd_iocb(vha, ELS_DCMD_LOGO, did);
866 return count;
867}
868
869static struct bin_attribute sysfs_issue_logo_attr = {
870 .attr = {
871 .name = "issue_logo",
872 .mode = S_IWUSR,
873 },
874 .size = 0,
875 .write = qla2x00_issue_logo,
876};
877
878static ssize_t
879qla2x00_sysfs_read_xgmac_stats(struct file *filp, struct kobject *kobj,
880 struct bin_attribute *bin_attr,
881 char *buf, loff_t off, size_t count)
882{
883 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
884 struct device, kobj)));
885 struct qla_hw_data *ha = vha->hw;
886 int rval;
887 uint16_t actual_size;
888
889 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > XGMAC_DATA_SIZE)
890 return 0;
891
892 if (unlikely(pci_channel_offline(ha->pdev)))
893 return 0;
894 mutex_lock(&vha->hw->optrom_mutex);
895 if (qla2x00_chip_is_down(vha)) {
896 mutex_unlock(&vha->hw->optrom_mutex);
897 return 0;
898 }
899
900 if (ha->xgmac_data)
901 goto do_read;
902
903 ha->xgmac_data = dma_alloc_coherent(&ha->pdev->dev, XGMAC_DATA_SIZE,
904 &ha->xgmac_data_dma, GFP_KERNEL);
905 if (!ha->xgmac_data) {
906 mutex_unlock(&vha->hw->optrom_mutex);
907 ql_log(ql_log_warn, vha, 0x7076,
908 "Unable to allocate memory for XGMAC read-data.\n");
909 return 0;
910 }
911
912do_read:
913 actual_size = 0;
914 memset(ha->xgmac_data, 0, XGMAC_DATA_SIZE);
915
916 rval = qla2x00_get_xgmac_stats(vha, ha->xgmac_data_dma,
917 XGMAC_DATA_SIZE, &actual_size);
918
919 mutex_unlock(&vha->hw->optrom_mutex);
920 if (rval != QLA_SUCCESS) {
921 ql_log(ql_log_warn, vha, 0x7077,
922 "Unable to read XGMAC data (%x).\n", rval);
923 count = 0;
924 }
925
926 count = actual_size > count ? count : actual_size;
927 memcpy(buf, ha->xgmac_data, count);
928
929 return count;
930}
931
932static struct bin_attribute sysfs_xgmac_stats_attr = {
933 .attr = {
934 .name = "xgmac_stats",
935 .mode = S_IRUSR,
936 },
937 .size = 0,
938 .read = qla2x00_sysfs_read_xgmac_stats,
939};
940
941static ssize_t
942qla2x00_sysfs_read_dcbx_tlv(struct file *filp, struct kobject *kobj,
943 struct bin_attribute *bin_attr,
944 char *buf, loff_t off, size_t count)
945{
946 struct scsi_qla_host *vha = shost_priv(dev_to_shost(container_of(kobj,
947 struct device, kobj)));
948 struct qla_hw_data *ha = vha->hw;
949 int rval;
950
951 if (!capable(CAP_SYS_ADMIN) || off != 0 || count > DCBX_TLV_DATA_SIZE)
952 return 0;
953
954 if (ha->dcbx_tlv)
955 goto do_read;
956 mutex_lock(&vha->hw->optrom_mutex);
957 if (qla2x00_chip_is_down(vha)) {
958 mutex_unlock(&vha->hw->optrom_mutex);
959 return 0;
960 }
961
962 ha->dcbx_tlv = dma_alloc_coherent(&ha->pdev->dev, DCBX_TLV_DATA_SIZE,
963 &ha->dcbx_tlv_dma, GFP_KERNEL);
964 if (!ha->dcbx_tlv) {
965 mutex_unlock(&vha->hw->optrom_mutex);
966 ql_log(ql_log_warn, vha, 0x7078,
967 "Unable to allocate memory for DCBX TLV read-data.\n");
968 return -ENOMEM;
969 }
970
971do_read:
972 memset(ha->dcbx_tlv, 0, DCBX_TLV_DATA_SIZE);
973
974 rval = qla2x00_get_dcbx_params(vha, ha->dcbx_tlv_dma,
975 DCBX_TLV_DATA_SIZE);
976
977 mutex_unlock(&vha->hw->optrom_mutex);
978
979 if (rval != QLA_SUCCESS) {
980 ql_log(ql_log_warn, vha, 0x7079,
981 "Unable to read DCBX TLV (%x).\n", rval);
982 return -EIO;
983 }
984
985 memcpy(buf, ha->dcbx_tlv, count);
986
987 return count;
988}
989
990static struct bin_attribute sysfs_dcbx_tlv_attr = {
991 .attr = {
992 .name = "dcbx_tlv",
993 .mode = S_IRUSR,
994 },
995 .size = 0,
996 .read = qla2x00_sysfs_read_dcbx_tlv,
997};
998
999static struct sysfs_entry {
1000 char *name;
1001 struct bin_attribute *attr;
1002 int type;
1003} bin_file_entries[] = {
1004 { "fw_dump", &sysfs_fw_dump_attr, },
1005 { "nvram", &sysfs_nvram_attr, },
1006 { "optrom", &sysfs_optrom_attr, },
1007 { "optrom_ctl", &sysfs_optrom_ctl_attr, },
1008 { "vpd", &sysfs_vpd_attr, 1 },
1009 { "sfp", &sysfs_sfp_attr, 1 },
1010 { "reset", &sysfs_reset_attr, },
1011 { "issue_logo", &sysfs_issue_logo_attr, },
1012 { "xgmac_stats", &sysfs_xgmac_stats_attr, 3 },
1013 { "dcbx_tlv", &sysfs_dcbx_tlv_attr, 3 },
1014 { NULL },
1015};
1016
1017void
1018qla2x00_alloc_sysfs_attr(scsi_qla_host_t *vha)
1019{
1020 struct Scsi_Host *host = vha->host;
1021 struct sysfs_entry *iter;
1022 int ret;
1023
1024 for (iter = bin_file_entries; iter->name; iter++) {
1025 if (iter->type && !IS_FWI2_CAPABLE(vha->hw))
1026 continue;
1027 if (iter->type == 2 && !IS_QLA25XX(vha->hw))
1028 continue;
1029 if (iter->type == 3 && !(IS_CNA_CAPABLE(vha->hw)))
1030 continue;
1031
1032 ret = sysfs_create_bin_file(&host->shost_gendev.kobj,
1033 iter->attr);
1034 if (ret)
1035 ql_log(ql_log_warn, vha, 0x00f3,
1036 "Unable to create sysfs %s binary attribute (%d).\n",
1037 iter->name, ret);
1038 else
1039 ql_dbg(ql_dbg_init, vha, 0x00f4,
1040 "Successfully created sysfs %s binary attribute.\n",
1041 iter->name);
1042 }
1043}
1044
1045void
1046qla2x00_free_sysfs_attr(scsi_qla_host_t *vha, bool stop_beacon)
1047{
1048 struct Scsi_Host *host = vha->host;
1049 struct sysfs_entry *iter;
1050 struct qla_hw_data *ha = vha->hw;
1051
1052 for (iter = bin_file_entries; iter->name; iter++) {
1053 if (iter->type && !IS_FWI2_CAPABLE(ha))
1054 continue;
1055 if (iter->type == 2 && !IS_QLA25XX(ha))
1056 continue;
1057 if (iter->type == 3 && !(IS_CNA_CAPABLE(ha)))
1058 continue;
1059 if (iter->type == 0x27 &&
1060 (!IS_QLA27XX(ha) || !IS_QLA28XX(ha)))
1061 continue;
1062
1063 sysfs_remove_bin_file(&host->shost_gendev.kobj,
1064 iter->attr);
1065 }
1066
1067 if (stop_beacon && ha->beacon_blink_led == 1)
1068 ha->isp_ops->beacon_off(vha);
1069}
1070
1071
1072
1073static ssize_t
1074qla2x00_driver_version_show(struct device *dev,
1075 struct device_attribute *attr, char *buf)
1076{
1077 return scnprintf(buf, PAGE_SIZE, "%s\n", qla2x00_version_str);
1078}
1079
1080static ssize_t
1081qla2x00_fw_version_show(struct device *dev,
1082 struct device_attribute *attr, char *buf)
1083{
1084 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1085 struct qla_hw_data *ha = vha->hw;
1086 char fw_str[128];
1087
1088 return scnprintf(buf, PAGE_SIZE, "%s\n",
1089 ha->isp_ops->fw_version_str(vha, fw_str, sizeof(fw_str)));
1090}
1091
1092static ssize_t
1093qla2x00_serial_num_show(struct device *dev, struct device_attribute *attr,
1094 char *buf)
1095{
1096 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1097 struct qla_hw_data *ha = vha->hw;
1098 uint32_t sn;
1099
1100 if (IS_QLAFX00(vha->hw)) {
1101 return scnprintf(buf, PAGE_SIZE, "%s\n",
1102 vha->hw->mr.serial_num);
1103 } else if (IS_FWI2_CAPABLE(ha)) {
1104 qla2xxx_get_vpd_field(vha, "SN", buf, PAGE_SIZE - 1);
1105 return strlen(strcat(buf, "\n"));
1106 }
1107
1108 sn = ((ha->serial0 & 0x1f) << 16) | (ha->serial2 << 8) | ha->serial1;
1109 return scnprintf(buf, PAGE_SIZE, "%c%05d\n", 'A' + sn / 100000,
1110 sn % 100000);
1111}
1112
1113static ssize_t
1114qla2x00_isp_name_show(struct device *dev, struct device_attribute *attr,
1115 char *buf)
1116{
1117 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1118
1119 return scnprintf(buf, PAGE_SIZE, "ISP%04X\n", vha->hw->pdev->device);
1120}
1121
1122static ssize_t
1123qla2x00_isp_id_show(struct device *dev, struct device_attribute *attr,
1124 char *buf)
1125{
1126 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1127 struct qla_hw_data *ha = vha->hw;
1128
1129 if (IS_QLAFX00(vha->hw))
1130 return scnprintf(buf, PAGE_SIZE, "%s\n",
1131 vha->hw->mr.hw_version);
1132
1133 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
1134 ha->product_id[0], ha->product_id[1], ha->product_id[2],
1135 ha->product_id[3]);
1136}
1137
1138static ssize_t
1139qla2x00_model_name_show(struct device *dev, struct device_attribute *attr,
1140 char *buf)
1141{
1142 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1143
1144 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_number);
1145}
1146
1147static ssize_t
1148qla2x00_model_desc_show(struct device *dev, struct device_attribute *attr,
1149 char *buf)
1150{
1151 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1152
1153 return scnprintf(buf, PAGE_SIZE, "%s\n", vha->hw->model_desc);
1154}
1155
1156static ssize_t
1157qla2x00_pci_info_show(struct device *dev, struct device_attribute *attr,
1158 char *buf)
1159{
1160 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1161 char pci_info[30];
1162
1163 return scnprintf(buf, PAGE_SIZE, "%s\n",
1164 vha->hw->isp_ops->pci_info_str(vha, pci_info,
1165 sizeof(pci_info)));
1166}
1167
1168static ssize_t
1169qla2x00_link_state_show(struct device *dev, struct device_attribute *attr,
1170 char *buf)
1171{
1172 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1173 struct qla_hw_data *ha = vha->hw;
1174 int len = 0;
1175
1176 if (atomic_read(&vha->loop_state) == LOOP_DOWN ||
1177 atomic_read(&vha->loop_state) == LOOP_DEAD ||
1178 vha->device_flags & DFLG_NO_CABLE)
1179 len = scnprintf(buf, PAGE_SIZE, "Link Down\n");
1180 else if (atomic_read(&vha->loop_state) != LOOP_READY ||
1181 qla2x00_chip_is_down(vha))
1182 len = scnprintf(buf, PAGE_SIZE, "Unknown Link State\n");
1183 else {
1184 len = scnprintf(buf, PAGE_SIZE, "Link Up - ");
1185
1186 switch (ha->current_topology) {
1187 case ISP_CFG_NL:
1188 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1189 break;
1190 case ISP_CFG_FL:
1191 len += scnprintf(buf + len, PAGE_SIZE-len, "FL_Port\n");
1192 break;
1193 case ISP_CFG_N:
1194 len += scnprintf(buf + len, PAGE_SIZE-len,
1195 "N_Port to N_Port\n");
1196 break;
1197 case ISP_CFG_F:
1198 len += scnprintf(buf + len, PAGE_SIZE-len, "F_Port\n");
1199 break;
1200 default:
1201 len += scnprintf(buf + len, PAGE_SIZE-len, "Loop\n");
1202 break;
1203 }
1204 }
1205 return len;
1206}
1207
1208static ssize_t
1209qla2x00_zio_show(struct device *dev, struct device_attribute *attr,
1210 char *buf)
1211{
1212 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1213 int len = 0;
1214
1215 switch (vha->hw->zio_mode) {
1216 case QLA_ZIO_MODE_6:
1217 len += scnprintf(buf + len, PAGE_SIZE-len, "Mode 6\n");
1218 break;
1219 case QLA_ZIO_DISABLED:
1220 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1221 break;
1222 }
1223 return len;
1224}
1225
1226static ssize_t
1227qla2x00_zio_store(struct device *dev, struct device_attribute *attr,
1228 const char *buf, size_t count)
1229{
1230 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1231 struct qla_hw_data *ha = vha->hw;
1232 int val = 0;
1233 uint16_t zio_mode;
1234
1235 if (!IS_ZIO_SUPPORTED(ha))
1236 return -ENOTSUPP;
1237
1238 if (sscanf(buf, "%d", &val) != 1)
1239 return -EINVAL;
1240
1241 if (val)
1242 zio_mode = QLA_ZIO_MODE_6;
1243 else
1244 zio_mode = QLA_ZIO_DISABLED;
1245
1246
1247 if (zio_mode != QLA_ZIO_DISABLED || ha->zio_mode != QLA_ZIO_DISABLED) {
1248 ha->zio_mode = zio_mode;
1249 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
1250 }
1251 return strlen(buf);
1252}
1253
1254static ssize_t
1255qla2x00_zio_timer_show(struct device *dev, struct device_attribute *attr,
1256 char *buf)
1257{
1258 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1259
1260 return scnprintf(buf, PAGE_SIZE, "%d us\n", vha->hw->zio_timer * 100);
1261}
1262
1263static ssize_t
1264qla2x00_zio_timer_store(struct device *dev, struct device_attribute *attr,
1265 const char *buf, size_t count)
1266{
1267 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1268 int val = 0;
1269 uint16_t zio_timer;
1270
1271 if (sscanf(buf, "%d", &val) != 1)
1272 return -EINVAL;
1273 if (val > 25500 || val < 100)
1274 return -ERANGE;
1275
1276 zio_timer = (uint16_t)(val / 100);
1277 vha->hw->zio_timer = zio_timer;
1278
1279 return strlen(buf);
1280}
1281
1282static ssize_t
1283qla_zio_threshold_show(struct device *dev, struct device_attribute *attr,
1284 char *buf)
1285{
1286 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1287
1288 return scnprintf(buf, PAGE_SIZE, "%d exchanges\n",
1289 vha->hw->last_zio_threshold);
1290}
1291
1292static ssize_t
1293qla_zio_threshold_store(struct device *dev, struct device_attribute *attr,
1294 const char *buf, size_t count)
1295{
1296 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1297 int val = 0;
1298
1299 if (vha->hw->zio_mode != QLA_ZIO_MODE_6)
1300 return -EINVAL;
1301 if (sscanf(buf, "%d", &val) != 1)
1302 return -EINVAL;
1303 if (val < 0 || val > 256)
1304 return -ERANGE;
1305
1306 atomic_set(&vha->hw->zio_threshold, val);
1307 return strlen(buf);
1308}
1309
1310static ssize_t
1311qla2x00_beacon_show(struct device *dev, struct device_attribute *attr,
1312 char *buf)
1313{
1314 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1315 int len = 0;
1316
1317 if (vha->hw->beacon_blink_led)
1318 len += scnprintf(buf + len, PAGE_SIZE-len, "Enabled\n");
1319 else
1320 len += scnprintf(buf + len, PAGE_SIZE-len, "Disabled\n");
1321 return len;
1322}
1323
1324static ssize_t
1325qla2x00_beacon_store(struct device *dev, struct device_attribute *attr,
1326 const char *buf, size_t count)
1327{
1328 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1329 struct qla_hw_data *ha = vha->hw;
1330 int val = 0;
1331 int rval;
1332
1333 if (IS_QLA2100(ha) || IS_QLA2200(ha))
1334 return -EPERM;
1335
1336 if (sscanf(buf, "%d", &val) != 1)
1337 return -EINVAL;
1338
1339 mutex_lock(&vha->hw->optrom_mutex);
1340 if (qla2x00_chip_is_down(vha)) {
1341 mutex_unlock(&vha->hw->optrom_mutex);
1342 ql_log(ql_log_warn, vha, 0x707a,
1343 "Abort ISP active -- ignoring beacon request.\n");
1344 return -EBUSY;
1345 }
1346
1347 if (val)
1348 rval = ha->isp_ops->beacon_on(vha);
1349 else
1350 rval = ha->isp_ops->beacon_off(vha);
1351
1352 if (rval != QLA_SUCCESS)
1353 count = 0;
1354
1355 mutex_unlock(&vha->hw->optrom_mutex);
1356
1357 return count;
1358}
1359
1360static ssize_t
1361qla2x00_beacon_config_show(struct device *dev, struct device_attribute *attr,
1362 char *buf)
1363{
1364 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1365 struct qla_hw_data *ha = vha->hw;
1366 uint16_t led[3] = { 0 };
1367
1368 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1369 return -EPERM;
1370
1371 if (ql26xx_led_config(vha, 0, led))
1372 return scnprintf(buf, PAGE_SIZE, "\n");
1373
1374 return scnprintf(buf, PAGE_SIZE, "%#04hx %#04hx %#04hx\n",
1375 led[0], led[1], led[2]);
1376}
1377
1378static ssize_t
1379qla2x00_beacon_config_store(struct device *dev, struct device_attribute *attr,
1380 const char *buf, size_t count)
1381{
1382 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1383 struct qla_hw_data *ha = vha->hw;
1384 uint16_t options = BIT_0;
1385 uint16_t led[3] = { 0 };
1386 uint16_t word[4];
1387 int n;
1388
1389 if (!IS_QLA2031(ha) && !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1390 return -EPERM;
1391
1392 n = sscanf(buf, "%hx %hx %hx %hx", word+0, word+1, word+2, word+3);
1393 if (n == 4) {
1394 if (word[0] == 3) {
1395 options |= BIT_3|BIT_2|BIT_1;
1396 led[0] = word[1];
1397 led[1] = word[2];
1398 led[2] = word[3];
1399 goto write;
1400 }
1401 return -EINVAL;
1402 }
1403
1404 if (n == 2) {
1405
1406 if (word[0] == 0) {
1407 options |= BIT_2;
1408 led[0] = word[1];
1409 goto write;
1410 }
1411 if (word[0] == 1) {
1412 options |= BIT_3;
1413 led[1] = word[1];
1414 goto write;
1415 }
1416 if (word[0] == 2) {
1417 options |= BIT_1;
1418 led[2] = word[1];
1419 goto write;
1420 }
1421 return -EINVAL;
1422 }
1423
1424 return -EINVAL;
1425
1426write:
1427 if (ql26xx_led_config(vha, options, led))
1428 return -EFAULT;
1429
1430 return count;
1431}
1432
1433static ssize_t
1434qla2x00_optrom_bios_version_show(struct device *dev,
1435 struct device_attribute *attr, char *buf)
1436{
1437 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1438 struct qla_hw_data *ha = vha->hw;
1439
1440 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->bios_revision[1],
1441 ha->bios_revision[0]);
1442}
1443
1444static ssize_t
1445qla2x00_optrom_efi_version_show(struct device *dev,
1446 struct device_attribute *attr, char *buf)
1447{
1448 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1449 struct qla_hw_data *ha = vha->hw;
1450
1451 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->efi_revision[1],
1452 ha->efi_revision[0]);
1453}
1454
1455static ssize_t
1456qla2x00_optrom_fcode_version_show(struct device *dev,
1457 struct device_attribute *attr, char *buf)
1458{
1459 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1460 struct qla_hw_data *ha = vha->hw;
1461
1462 return scnprintf(buf, PAGE_SIZE, "%d.%02d\n", ha->fcode_revision[1],
1463 ha->fcode_revision[0]);
1464}
1465
1466static ssize_t
1467qla2x00_optrom_fw_version_show(struct device *dev,
1468 struct device_attribute *attr, char *buf)
1469{
1470 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1471 struct qla_hw_data *ha = vha->hw;
1472
1473 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d %d\n",
1474 ha->fw_revision[0], ha->fw_revision[1], ha->fw_revision[2],
1475 ha->fw_revision[3]);
1476}
1477
1478static ssize_t
1479qla2x00_optrom_gold_fw_version_show(struct device *dev,
1480 struct device_attribute *attr, char *buf)
1481{
1482 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1483 struct qla_hw_data *ha = vha->hw;
1484
1485 if (!IS_QLA81XX(ha) && !IS_QLA83XX(ha) &&
1486 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1487 return scnprintf(buf, PAGE_SIZE, "\n");
1488
1489 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%d)\n",
1490 ha->gold_fw_version[0], ha->gold_fw_version[1],
1491 ha->gold_fw_version[2], ha->gold_fw_version[3]);
1492}
1493
1494static ssize_t
1495qla2x00_total_isp_aborts_show(struct device *dev,
1496 struct device_attribute *attr, char *buf)
1497{
1498 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1499
1500 return scnprintf(buf, PAGE_SIZE, "%d\n",
1501 vha->qla_stats.total_isp_aborts);
1502}
1503
1504static ssize_t
1505qla24xx_84xx_fw_version_show(struct device *dev,
1506 struct device_attribute *attr, char *buf)
1507{
1508 int rval = QLA_SUCCESS;
1509 uint16_t status[2] = { 0 };
1510 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1511 struct qla_hw_data *ha = vha->hw;
1512
1513 if (!IS_QLA84XX(ha))
1514 return scnprintf(buf, PAGE_SIZE, "\n");
1515
1516 if (!ha->cs84xx->op_fw_version) {
1517 rval = qla84xx_verify_chip(vha, status);
1518
1519 if (!rval && !status[0])
1520 return scnprintf(buf, PAGE_SIZE, "%u\n",
1521 (uint32_t)ha->cs84xx->op_fw_version);
1522 }
1523
1524 return scnprintf(buf, PAGE_SIZE, "\n");
1525}
1526
1527static ssize_t
1528qla2x00_serdes_version_show(struct device *dev, struct device_attribute *attr,
1529 char *buf)
1530{
1531 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1532 struct qla_hw_data *ha = vha->hw;
1533
1534 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1535 return scnprintf(buf, PAGE_SIZE, "\n");
1536
1537 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1538 ha->serdes_version[0], ha->serdes_version[1],
1539 ha->serdes_version[2]);
1540}
1541
1542static ssize_t
1543qla2x00_mpi_version_show(struct device *dev, struct device_attribute *attr,
1544 char *buf)
1545{
1546 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1547 struct qla_hw_data *ha = vha->hw;
1548
1549 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha) && !IS_QLA8044(ha) &&
1550 !IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1551 return scnprintf(buf, PAGE_SIZE, "\n");
1552
1553 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d (%x)\n",
1554 ha->mpi_version[0], ha->mpi_version[1], ha->mpi_version[2],
1555 ha->mpi_capabilities);
1556}
1557
1558static ssize_t
1559qla2x00_phy_version_show(struct device *dev, struct device_attribute *attr,
1560 char *buf)
1561{
1562 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1563 struct qla_hw_data *ha = vha->hw;
1564
1565 if (!IS_QLA81XX(ha) && !IS_QLA8031(ha))
1566 return scnprintf(buf, PAGE_SIZE, "\n");
1567
1568 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1569 ha->phy_version[0], ha->phy_version[1], ha->phy_version[2]);
1570}
1571
1572static ssize_t
1573qla2x00_flash_block_size_show(struct device *dev,
1574 struct device_attribute *attr, char *buf)
1575{
1576 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1577 struct qla_hw_data *ha = vha->hw;
1578
1579 return scnprintf(buf, PAGE_SIZE, "0x%x\n", ha->fdt_block_size);
1580}
1581
1582static ssize_t
1583qla2x00_vlan_id_show(struct device *dev, struct device_attribute *attr,
1584 char *buf)
1585{
1586 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1587
1588 if (!IS_CNA_CAPABLE(vha->hw))
1589 return scnprintf(buf, PAGE_SIZE, "\n");
1590
1591 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->fcoe_vlan_id);
1592}
1593
1594static ssize_t
1595qla2x00_vn_port_mac_address_show(struct device *dev,
1596 struct device_attribute *attr, char *buf)
1597{
1598 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1599
1600 if (!IS_CNA_CAPABLE(vha->hw))
1601 return scnprintf(buf, PAGE_SIZE, "\n");
1602
1603 return scnprintf(buf, PAGE_SIZE, "%pMR\n", vha->fcoe_vn_port_mac);
1604}
1605
1606static ssize_t
1607qla2x00_fabric_param_show(struct device *dev, struct device_attribute *attr,
1608 char *buf)
1609{
1610 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1611
1612 return scnprintf(buf, PAGE_SIZE, "%d\n", vha->hw->switch_cap);
1613}
1614
1615static ssize_t
1616qla2x00_thermal_temp_show(struct device *dev,
1617 struct device_attribute *attr, char *buf)
1618{
1619 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1620 uint16_t temp = 0;
1621 int rc;
1622
1623 mutex_lock(&vha->hw->optrom_mutex);
1624 if (qla2x00_chip_is_down(vha)) {
1625 mutex_unlock(&vha->hw->optrom_mutex);
1626 ql_log(ql_log_warn, vha, 0x70dc, "ISP reset active.\n");
1627 goto done;
1628 }
1629
1630 if (vha->hw->flags.eeh_busy) {
1631 mutex_unlock(&vha->hw->optrom_mutex);
1632 ql_log(ql_log_warn, vha, 0x70dd, "PCI EEH busy.\n");
1633 goto done;
1634 }
1635
1636 rc = qla2x00_get_thermal_temp(vha, &temp);
1637 mutex_unlock(&vha->hw->optrom_mutex);
1638 if (rc == QLA_SUCCESS)
1639 return scnprintf(buf, PAGE_SIZE, "%d\n", temp);
1640
1641done:
1642 return scnprintf(buf, PAGE_SIZE, "\n");
1643}
1644
1645static ssize_t
1646qla2x00_fw_state_show(struct device *dev, struct device_attribute *attr,
1647 char *buf)
1648{
1649 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1650 int rval = QLA_FUNCTION_FAILED;
1651 uint16_t state[6];
1652 uint32_t pstate;
1653
1654 if (IS_QLAFX00(vha->hw)) {
1655 pstate = qlafx00_fw_state_show(dev, attr, buf);
1656 return scnprintf(buf, PAGE_SIZE, "0x%x\n", pstate);
1657 }
1658
1659 mutex_lock(&vha->hw->optrom_mutex);
1660 if (qla2x00_chip_is_down(vha)) {
1661 mutex_unlock(&vha->hw->optrom_mutex);
1662 ql_log(ql_log_warn, vha, 0x707c,
1663 "ISP reset active.\n");
1664 goto out;
1665 } else if (vha->hw->flags.eeh_busy) {
1666 mutex_unlock(&vha->hw->optrom_mutex);
1667 goto out;
1668 }
1669
1670 rval = qla2x00_get_firmware_state(vha, state);
1671 mutex_unlock(&vha->hw->optrom_mutex);
1672out:
1673 if (rval != QLA_SUCCESS) {
1674 memset(state, -1, sizeof(state));
1675 rval = qla2x00_get_firmware_state(vha, state);
1676 }
1677
1678 return scnprintf(buf, PAGE_SIZE, "0x%x 0x%x 0x%x 0x%x 0x%x 0x%x\n",
1679 state[0], state[1], state[2], state[3], state[4], state[5]);
1680}
1681
1682static ssize_t
1683qla2x00_diag_requests_show(struct device *dev,
1684 struct device_attribute *attr, char *buf)
1685{
1686 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1687
1688 if (!IS_BIDI_CAPABLE(vha->hw))
1689 return scnprintf(buf, PAGE_SIZE, "\n");
1690
1691 return scnprintf(buf, PAGE_SIZE, "%llu\n", vha->bidi_stats.io_count);
1692}
1693
1694static ssize_t
1695qla2x00_diag_megabytes_show(struct device *dev,
1696 struct device_attribute *attr, char *buf)
1697{
1698 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1699
1700 if (!IS_BIDI_CAPABLE(vha->hw))
1701 return scnprintf(buf, PAGE_SIZE, "\n");
1702
1703 return scnprintf(buf, PAGE_SIZE, "%llu\n",
1704 vha->bidi_stats.transfer_bytes >> 20);
1705}
1706
1707static ssize_t
1708qla2x00_fw_dump_size_show(struct device *dev, struct device_attribute *attr,
1709 char *buf)
1710{
1711 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1712 struct qla_hw_data *ha = vha->hw;
1713 uint32_t size;
1714
1715 if (!ha->fw_dumped)
1716 size = 0;
1717 else if (IS_P3P_TYPE(ha))
1718 size = ha->md_template_size + ha->md_dump_size;
1719 else
1720 size = ha->fw_dump_len;
1721
1722 return scnprintf(buf, PAGE_SIZE, "%d\n", size);
1723}
1724
1725static ssize_t
1726qla2x00_allow_cna_fw_dump_show(struct device *dev,
1727 struct device_attribute *attr, char *buf)
1728{
1729 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1730
1731 if (!IS_P3P_TYPE(vha->hw))
1732 return scnprintf(buf, PAGE_SIZE, "\n");
1733 else
1734 return scnprintf(buf, PAGE_SIZE, "%s\n",
1735 vha->hw->allow_cna_fw_dump ? "true" : "false");
1736}
1737
1738static ssize_t
1739qla2x00_allow_cna_fw_dump_store(struct device *dev,
1740 struct device_attribute *attr, const char *buf, size_t count)
1741{
1742 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1743 int val = 0;
1744
1745 if (!IS_P3P_TYPE(vha->hw))
1746 return -EINVAL;
1747
1748 if (sscanf(buf, "%d", &val) != 1)
1749 return -EINVAL;
1750
1751 vha->hw->allow_cna_fw_dump = val != 0;
1752
1753 return strlen(buf);
1754}
1755
1756static ssize_t
1757qla2x00_pep_version_show(struct device *dev, struct device_attribute *attr,
1758 char *buf)
1759{
1760 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1761 struct qla_hw_data *ha = vha->hw;
1762
1763 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1764 return scnprintf(buf, PAGE_SIZE, "\n");
1765
1766 return scnprintf(buf, PAGE_SIZE, "%d.%02d.%02d\n",
1767 ha->pep_version[0], ha->pep_version[1], ha->pep_version[2]);
1768}
1769
1770static ssize_t
1771qla2x00_min_supported_speed_show(struct device *dev,
1772 struct device_attribute *attr, char *buf)
1773{
1774 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1775 struct qla_hw_data *ha = vha->hw;
1776
1777 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1778 return scnprintf(buf, PAGE_SIZE, "\n");
1779
1780 return scnprintf(buf, PAGE_SIZE, "%s\n",
1781 ha->min_supported_speed == 6 ? "64Gps" :
1782 ha->min_supported_speed == 5 ? "32Gps" :
1783 ha->min_supported_speed == 4 ? "16Gps" :
1784 ha->min_supported_speed == 3 ? "8Gps" :
1785 ha->min_supported_speed == 2 ? "4Gps" :
1786 ha->min_supported_speed != 0 ? "unknown" : "");
1787}
1788
1789static ssize_t
1790qla2x00_max_supported_speed_show(struct device *dev,
1791 struct device_attribute *attr, char *buf)
1792{
1793 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1794 struct qla_hw_data *ha = vha->hw;
1795
1796 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
1797 return scnprintf(buf, PAGE_SIZE, "\n");
1798
1799 return scnprintf(buf, PAGE_SIZE, "%s\n",
1800 ha->max_supported_speed == 2 ? "64Gps" :
1801 ha->max_supported_speed == 1 ? "32Gps" :
1802 ha->max_supported_speed == 0 ? "16Gps" : "unknown");
1803}
1804
1805static ssize_t
1806qla2x00_port_speed_store(struct device *dev, struct device_attribute *attr,
1807 const char *buf, size_t count)
1808{
1809 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1810 ulong type, speed;
1811 int oldspeed, rval;
1812 int mode = QLA_SET_DATA_RATE_LR;
1813 struct qla_hw_data *ha = vha->hw;
1814
1815 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha)) {
1816 ql_log(ql_log_warn, vha, 0x70d8,
1817 "Speed setting not supported \n");
1818 return -EINVAL;
1819 }
1820
1821 rval = kstrtol(buf, 10, &type);
1822 if (rval)
1823 return rval;
1824 speed = type;
1825 if (type == 40 || type == 80 || type == 160 ||
1826 type == 320) {
1827 ql_dbg(ql_dbg_user, vha, 0x70d9,
1828 "Setting will be affected after a loss of sync\n");
1829 type = type/10;
1830 mode = QLA_SET_DATA_RATE_NOLR;
1831 }
1832
1833 oldspeed = ha->set_data_rate;
1834
1835 switch (type) {
1836 case 0:
1837 ha->set_data_rate = PORT_SPEED_AUTO;
1838 break;
1839 case 4:
1840 ha->set_data_rate = PORT_SPEED_4GB;
1841 break;
1842 case 8:
1843 ha->set_data_rate = PORT_SPEED_8GB;
1844 break;
1845 case 16:
1846 ha->set_data_rate = PORT_SPEED_16GB;
1847 break;
1848 case 32:
1849 ha->set_data_rate = PORT_SPEED_32GB;
1850 break;
1851 default:
1852 ql_log(ql_log_warn, vha, 0x1199,
1853 "Unrecognized speed setting:%lx. Setting Autoneg\n",
1854 speed);
1855 ha->set_data_rate = PORT_SPEED_AUTO;
1856 }
1857
1858 if (qla2x00_chip_is_down(vha) || (oldspeed == ha->set_data_rate))
1859 return -EINVAL;
1860
1861 ql_log(ql_log_info, vha, 0x70da,
1862 "Setting speed to %lx Gbps \n", type);
1863
1864 rval = qla2x00_set_data_rate(vha, mode);
1865 if (rval != QLA_SUCCESS)
1866 return -EIO;
1867
1868 return strlen(buf);
1869}
1870
1871static ssize_t
1872qla2x00_port_speed_show(struct device *dev, struct device_attribute *attr,
1873 char *buf)
1874{
1875 struct scsi_qla_host *vha = shost_priv(dev_to_shost(dev));
1876 struct qla_hw_data *ha = vha->hw;
1877 ssize_t rval;
1878 char *spd[7] = {"0", "0", "0", "4", "8", "16", "32"};
1879
1880 rval = qla2x00_get_data_rate(vha);
1881 if (rval != QLA_SUCCESS) {
1882 ql_log(ql_log_warn, vha, 0x70db,
1883 "Unable to get port speed rval:%zd\n", rval);
1884 return -EINVAL;
1885 }
1886
1887 return scnprintf(buf, PAGE_SIZE, "%s\n", spd[ha->link_data_rate]);
1888}
1889
1890
1891
1892static ssize_t
1893qlini_mode_show(struct device *dev, struct device_attribute *attr, char *buf)
1894{
1895 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
1896 int len = 0;
1897
1898 len += scnprintf(buf + len, PAGE_SIZE-len,
1899 "Supported options: enabled | disabled | dual | exclusive\n");
1900
1901
1902 len += scnprintf(buf + len, PAGE_SIZE-len, "Current selection: ");
1903
1904 switch (vha->qlini_mode) {
1905 case QLA2XXX_INI_MODE_EXCLUSIVE:
1906 len += scnprintf(buf + len, PAGE_SIZE-len,
1907 QLA2XXX_INI_MODE_STR_EXCLUSIVE);
1908 break;
1909 case QLA2XXX_INI_MODE_DISABLED:
1910 len += scnprintf(buf + len, PAGE_SIZE-len,
1911 QLA2XXX_INI_MODE_STR_DISABLED);
1912 break;
1913 case QLA2XXX_INI_MODE_ENABLED:
1914 len += scnprintf(buf + len, PAGE_SIZE-len,
1915 QLA2XXX_INI_MODE_STR_ENABLED);
1916 break;
1917 case QLA2XXX_INI_MODE_DUAL:
1918 len += scnprintf(buf + len, PAGE_SIZE-len,
1919 QLA2XXX_INI_MODE_STR_DUAL);
1920 break;
1921 }
1922 len += scnprintf(buf + len, PAGE_SIZE-len, "\n");
1923
1924 return len;
1925}
1926
1927static char *mode_to_str[] = {
1928 "exclusive",
1929 "disabled",
1930 "enabled",
1931 "dual",
1932};
1933
1934#define NEED_EXCH_OFFLOAD(_exchg) ((_exchg) > FW_DEF_EXCHANGES_CNT)
1935static void qla_set_ini_mode(scsi_qla_host_t *vha, int op)
1936{
1937 enum {
1938 NO_ACTION,
1939 MODE_CHANGE_ACCEPT,
1940 MODE_CHANGE_NO_ACTION,
1941 TARGET_STILL_ACTIVE,
1942 };
1943 int action = NO_ACTION;
1944 int set_mode = 0;
1945 u8 eo_toggle = 0;
1946
1947 switch (vha->qlini_mode) {
1948 case QLA2XXX_INI_MODE_DISABLED:
1949 switch (op) {
1950 case QLA2XXX_INI_MODE_DISABLED:
1951 if (qla_tgt_mode_enabled(vha)) {
1952 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
1953 vha->hw->flags.exchoffld_enabled)
1954 eo_toggle = 1;
1955 if (((vha->ql2xexchoffld !=
1956 vha->u_ql2xexchoffld) &&
1957 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
1958 eo_toggle) {
1959
1960
1961
1962
1963
1964 action = MODE_CHANGE_ACCEPT;
1965 } else {
1966 action = MODE_CHANGE_NO_ACTION;
1967 }
1968 } else {
1969 action = MODE_CHANGE_NO_ACTION;
1970 }
1971 break;
1972 case QLA2XXX_INI_MODE_EXCLUSIVE:
1973 if (qla_tgt_mode_enabled(vha)) {
1974 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
1975 vha->hw->flags.exchoffld_enabled)
1976 eo_toggle = 1;
1977 if (((vha->ql2xexchoffld !=
1978 vha->u_ql2xexchoffld) &&
1979 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
1980 eo_toggle) {
1981
1982
1983
1984
1985
1986 action = MODE_CHANGE_ACCEPT;
1987 } else {
1988 action = MODE_CHANGE_NO_ACTION;
1989 }
1990 } else {
1991 action = MODE_CHANGE_ACCEPT;
1992 }
1993 break;
1994 case QLA2XXX_INI_MODE_DUAL:
1995 action = MODE_CHANGE_ACCEPT;
1996
1997 if (qla_tgt_mode_enabled(vha)) {
1998 set_mode = 1;
1999 action = MODE_CHANGE_ACCEPT;
2000 } else {
2001 action = MODE_CHANGE_NO_ACTION;
2002 }
2003 break;
2004
2005 case QLA2XXX_INI_MODE_ENABLED:
2006 if (qla_tgt_mode_enabled(vha))
2007 action = TARGET_STILL_ACTIVE;
2008 else {
2009 action = MODE_CHANGE_ACCEPT;
2010 set_mode = 1;
2011 }
2012 break;
2013 }
2014 break;
2015
2016 case QLA2XXX_INI_MODE_EXCLUSIVE:
2017 switch (op) {
2018 case QLA2XXX_INI_MODE_EXCLUSIVE:
2019 if (qla_tgt_mode_enabled(vha)) {
2020 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2021 vha->hw->flags.exchoffld_enabled)
2022 eo_toggle = 1;
2023 if (((vha->ql2xexchoffld !=
2024 vha->u_ql2xexchoffld) &&
2025 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2026 eo_toggle)
2027
2028
2029
2030
2031
2032 action = MODE_CHANGE_ACCEPT;
2033 else
2034 action = NO_ACTION;
2035 } else
2036 action = NO_ACTION;
2037
2038 break;
2039
2040 case QLA2XXX_INI_MODE_DISABLED:
2041 if (qla_tgt_mode_enabled(vha)) {
2042 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld) !=
2043 vha->hw->flags.exchoffld_enabled)
2044 eo_toggle = 1;
2045 if (((vha->ql2xexchoffld !=
2046 vha->u_ql2xexchoffld) &&
2047 NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld)) ||
2048 eo_toggle)
2049 action = MODE_CHANGE_ACCEPT;
2050 else
2051 action = MODE_CHANGE_NO_ACTION;
2052 } else
2053 action = MODE_CHANGE_NO_ACTION;
2054 break;
2055
2056 case QLA2XXX_INI_MODE_DUAL:
2057 if (qla_tgt_mode_enabled(vha)) {
2058 action = MODE_CHANGE_ACCEPT;
2059 set_mode = 1;
2060 } else
2061 action = MODE_CHANGE_ACCEPT;
2062 break;
2063
2064 case QLA2XXX_INI_MODE_ENABLED:
2065 if (qla_tgt_mode_enabled(vha))
2066 action = TARGET_STILL_ACTIVE;
2067 else {
2068 if (vha->hw->flags.fw_started)
2069 action = MODE_CHANGE_NO_ACTION;
2070 else
2071 action = MODE_CHANGE_ACCEPT;
2072 }
2073 break;
2074 }
2075 break;
2076
2077 case QLA2XXX_INI_MODE_ENABLED:
2078 switch (op) {
2079 case QLA2XXX_INI_MODE_ENABLED:
2080 if (NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg) !=
2081 vha->hw->flags.exchoffld_enabled)
2082 eo_toggle = 1;
2083 if (((vha->ql2xiniexchg != vha->u_ql2xiniexchg) &&
2084 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg)) ||
2085 eo_toggle)
2086 action = MODE_CHANGE_ACCEPT;
2087 else
2088 action = NO_ACTION;
2089 break;
2090 case QLA2XXX_INI_MODE_DUAL:
2091 case QLA2XXX_INI_MODE_DISABLED:
2092 action = MODE_CHANGE_ACCEPT;
2093 break;
2094 default:
2095 action = MODE_CHANGE_NO_ACTION;
2096 break;
2097 }
2098 break;
2099
2100 case QLA2XXX_INI_MODE_DUAL:
2101 switch (op) {
2102 case QLA2XXX_INI_MODE_DUAL:
2103 if (qla_tgt_mode_enabled(vha) ||
2104 qla_dual_mode_enabled(vha)) {
2105 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2106 vha->u_ql2xiniexchg) !=
2107 vha->hw->flags.exchoffld_enabled)
2108 eo_toggle = 1;
2109
2110 if ((((vha->ql2xexchoffld +
2111 vha->ql2xiniexchg) !=
2112 (vha->u_ql2xiniexchg +
2113 vha->u_ql2xexchoffld)) &&
2114 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2115 vha->u_ql2xexchoffld)) || eo_toggle)
2116 action = MODE_CHANGE_ACCEPT;
2117 else
2118 action = NO_ACTION;
2119 } else {
2120 if (NEED_EXCH_OFFLOAD(vha->u_ql2xexchoffld +
2121 vha->u_ql2xiniexchg) !=
2122 vha->hw->flags.exchoffld_enabled)
2123 eo_toggle = 1;
2124
2125 if ((((vha->ql2xexchoffld + vha->ql2xiniexchg)
2126 != (vha->u_ql2xiniexchg +
2127 vha->u_ql2xexchoffld)) &&
2128 NEED_EXCH_OFFLOAD(vha->u_ql2xiniexchg +
2129 vha->u_ql2xexchoffld)) || eo_toggle)
2130 action = MODE_CHANGE_NO_ACTION;
2131 else
2132 action = NO_ACTION;
2133 }
2134 break;
2135
2136 case QLA2XXX_INI_MODE_DISABLED:
2137 if (qla_tgt_mode_enabled(vha) ||
2138 qla_dual_mode_enabled(vha)) {
2139
2140 set_mode = 1;
2141 action = MODE_CHANGE_ACCEPT;
2142 } else {
2143 action = MODE_CHANGE_NO_ACTION;
2144 }
2145 break;
2146
2147 case QLA2XXX_INI_MODE_EXCLUSIVE:
2148 if (qla_tgt_mode_enabled(vha) ||
2149 qla_dual_mode_enabled(vha)) {
2150 set_mode = 1;
2151 action = MODE_CHANGE_ACCEPT;
2152 } else {
2153 action = MODE_CHANGE_ACCEPT;
2154 }
2155 break;
2156
2157 case QLA2XXX_INI_MODE_ENABLED:
2158 if (qla_tgt_mode_enabled(vha) ||
2159 qla_dual_mode_enabled(vha)) {
2160 action = TARGET_STILL_ACTIVE;
2161 } else {
2162 action = MODE_CHANGE_ACCEPT;
2163 }
2164 }
2165 break;
2166 }
2167
2168 switch (action) {
2169 case MODE_CHANGE_ACCEPT:
2170 ql_log(ql_log_warn, vha, 0xffff,
2171 "Mode change accepted. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2172 mode_to_str[vha->qlini_mode], mode_to_str[op],
2173 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2174 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2175
2176 vha->qlini_mode = op;
2177 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2178 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2179 if (set_mode)
2180 qlt_set_mode(vha);
2181 vha->flags.online = 1;
2182 set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags);
2183 break;
2184
2185 case MODE_CHANGE_NO_ACTION:
2186 ql_log(ql_log_warn, vha, 0xffff,
2187 "Mode is set. No action taken. From %s to %s, Tgt exchg %d|%d. ini exchg %d|%d\n",
2188 mode_to_str[vha->qlini_mode], mode_to_str[op],
2189 vha->ql2xexchoffld, vha->u_ql2xexchoffld,
2190 vha->ql2xiniexchg, vha->u_ql2xiniexchg);
2191 vha->qlini_mode = op;
2192 vha->ql2xexchoffld = vha->u_ql2xexchoffld;
2193 vha->ql2xiniexchg = vha->u_ql2xiniexchg;
2194 break;
2195
2196 case TARGET_STILL_ACTIVE:
2197 ql_log(ql_log_warn, vha, 0xffff,
2198 "Target Mode is active. Unable to change Mode.\n");
2199 break;
2200
2201 case NO_ACTION:
2202 default:
2203 ql_log(ql_log_warn, vha, 0xffff,
2204 "Mode unchange. No action taken. %d|%d pct %d|%d.\n",
2205 vha->qlini_mode, op,
2206 vha->ql2xexchoffld, vha->u_ql2xexchoffld);
2207 break;
2208 }
2209}
2210
2211static ssize_t
2212qlini_mode_store(struct device *dev, struct device_attribute *attr,
2213 const char *buf, size_t count)
2214{
2215 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2216 int ini;
2217
2218 if (!buf)
2219 return -EINVAL;
2220
2221 if (strncasecmp(QLA2XXX_INI_MODE_STR_EXCLUSIVE, buf,
2222 strlen(QLA2XXX_INI_MODE_STR_EXCLUSIVE)) == 0)
2223 ini = QLA2XXX_INI_MODE_EXCLUSIVE;
2224 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DISABLED, buf,
2225 strlen(QLA2XXX_INI_MODE_STR_DISABLED)) == 0)
2226 ini = QLA2XXX_INI_MODE_DISABLED;
2227 else if (strncasecmp(QLA2XXX_INI_MODE_STR_ENABLED, buf,
2228 strlen(QLA2XXX_INI_MODE_STR_ENABLED)) == 0)
2229 ini = QLA2XXX_INI_MODE_ENABLED;
2230 else if (strncasecmp(QLA2XXX_INI_MODE_STR_DUAL, buf,
2231 strlen(QLA2XXX_INI_MODE_STR_DUAL)) == 0)
2232 ini = QLA2XXX_INI_MODE_DUAL;
2233 else
2234 return -EINVAL;
2235
2236 qla_set_ini_mode(vha, ini);
2237 return strlen(buf);
2238}
2239
2240static ssize_t
2241ql2xexchoffld_show(struct device *dev, struct device_attribute *attr,
2242 char *buf)
2243{
2244 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2245 int len = 0;
2246
2247 len += scnprintf(buf + len, PAGE_SIZE-len,
2248 "target exchange: new %d : current: %d\n\n",
2249 vha->u_ql2xexchoffld, vha->ql2xexchoffld);
2250
2251 len += scnprintf(buf + len, PAGE_SIZE-len,
2252 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2253 vha->host_no);
2254
2255 return len;
2256}
2257
2258static ssize_t
2259ql2xexchoffld_store(struct device *dev, struct device_attribute *attr,
2260 const char *buf, size_t count)
2261{
2262 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2263 int val = 0;
2264
2265 if (sscanf(buf, "%d", &val) != 1)
2266 return -EINVAL;
2267
2268 if (val > FW_MAX_EXCHANGES_CNT)
2269 val = FW_MAX_EXCHANGES_CNT;
2270 else if (val < 0)
2271 val = 0;
2272
2273 vha->u_ql2xexchoffld = val;
2274 return strlen(buf);
2275}
2276
2277static ssize_t
2278ql2xiniexchg_show(struct device *dev, struct device_attribute *attr,
2279 char *buf)
2280{
2281 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2282 int len = 0;
2283
2284 len += scnprintf(buf + len, PAGE_SIZE-len,
2285 "target exchange: new %d : current: %d\n\n",
2286 vha->u_ql2xiniexchg, vha->ql2xiniexchg);
2287
2288 len += scnprintf(buf + len, PAGE_SIZE-len,
2289 "Please (re)set operating mode via \"/sys/class/scsi_host/host%ld/qlini_mode\" to load new setting.\n",
2290 vha->host_no);
2291
2292 return len;
2293}
2294
2295static ssize_t
2296ql2xiniexchg_store(struct device *dev, struct device_attribute *attr,
2297 const char *buf, size_t count)
2298{
2299 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2300 int val = 0;
2301
2302 if (sscanf(buf, "%d", &val) != 1)
2303 return -EINVAL;
2304
2305 if (val > FW_MAX_EXCHANGES_CNT)
2306 val = FW_MAX_EXCHANGES_CNT;
2307 else if (val < 0)
2308 val = 0;
2309
2310 vha->u_ql2xiniexchg = val;
2311 return strlen(buf);
2312}
2313
2314static ssize_t
2315qla2x00_dif_bundle_statistics_show(struct device *dev,
2316 struct device_attribute *attr, char *buf)
2317{
2318 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2319 struct qla_hw_data *ha = vha->hw;
2320
2321 return scnprintf(buf, PAGE_SIZE,
2322 "cross=%llu read=%llu write=%llu kalloc=%llu dma_alloc=%llu unusable=%u\n",
2323 ha->dif_bundle_crossed_pages, ha->dif_bundle_reads,
2324 ha->dif_bundle_writes, ha->dif_bundle_kallocs,
2325 ha->dif_bundle_dma_allocs, ha->pool.unusable.count);
2326}
2327
2328static ssize_t
2329qla2x00_fw_attr_show(struct device *dev,
2330 struct device_attribute *attr, char *buf)
2331{
2332 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2333 struct qla_hw_data *ha = vha->hw;
2334
2335 if (!IS_QLA27XX(ha) && !IS_QLA28XX(ha))
2336 return scnprintf(buf, PAGE_SIZE, "\n");
2337
2338 return scnprintf(buf, PAGE_SIZE, "%llx\n",
2339 (uint64_t)ha->fw_attributes_ext[1] << 48 |
2340 (uint64_t)ha->fw_attributes_ext[0] << 32 |
2341 (uint64_t)ha->fw_attributes_h << 16 |
2342 (uint64_t)ha->fw_attributes);
2343}
2344
2345static ssize_t
2346qla2x00_port_no_show(struct device *dev, struct device_attribute *attr,
2347 char *buf)
2348{
2349 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2350
2351 return scnprintf(buf, PAGE_SIZE, "%u\n", vha->hw->port_no);
2352}
2353
2354static ssize_t
2355qla2x00_dport_diagnostics_show(struct device *dev,
2356 struct device_attribute *attr, char *buf)
2357{
2358 scsi_qla_host_t *vha = shost_priv(class_to_shost(dev));
2359
2360 if (!IS_QLA83XX(vha->hw) && !IS_QLA27XX(vha->hw) &&
2361 !IS_QLA28XX(vha->hw))
2362 return scnprintf(buf, PAGE_SIZE, "\n");
2363
2364 if (!*vha->dport_data)
2365 return scnprintf(buf, PAGE_SIZE, "\n");
2366
2367 return scnprintf(buf, PAGE_SIZE, "%04x %04x %04x %04x\n",
2368 vha->dport_data[0], vha->dport_data[1],
2369 vha->dport_data[2], vha->dport_data[3]);
2370}
2371static DEVICE_ATTR(dport_diagnostics, 0444,
2372 qla2x00_dport_diagnostics_show, NULL);
2373
2374static DEVICE_ATTR(driver_version, S_IRUGO, qla2x00_driver_version_show, NULL);
2375static DEVICE_ATTR(fw_version, S_IRUGO, qla2x00_fw_version_show, NULL);
2376static DEVICE_ATTR(serial_num, S_IRUGO, qla2x00_serial_num_show, NULL);
2377static DEVICE_ATTR(isp_name, S_IRUGO, qla2x00_isp_name_show, NULL);
2378static DEVICE_ATTR(isp_id, S_IRUGO, qla2x00_isp_id_show, NULL);
2379static DEVICE_ATTR(model_name, S_IRUGO, qla2x00_model_name_show, NULL);
2380static DEVICE_ATTR(model_desc, S_IRUGO, qla2x00_model_desc_show, NULL);
2381static DEVICE_ATTR(pci_info, S_IRUGO, qla2x00_pci_info_show, NULL);
2382static DEVICE_ATTR(link_state, S_IRUGO, qla2x00_link_state_show, NULL);
2383static DEVICE_ATTR(zio, S_IRUGO | S_IWUSR, qla2x00_zio_show, qla2x00_zio_store);
2384static DEVICE_ATTR(zio_timer, S_IRUGO | S_IWUSR, qla2x00_zio_timer_show,
2385 qla2x00_zio_timer_store);
2386static DEVICE_ATTR(beacon, S_IRUGO | S_IWUSR, qla2x00_beacon_show,
2387 qla2x00_beacon_store);
2388static DEVICE_ATTR(beacon_config, 0644, qla2x00_beacon_config_show,
2389 qla2x00_beacon_config_store);
2390static DEVICE_ATTR(optrom_bios_version, S_IRUGO,
2391 qla2x00_optrom_bios_version_show, NULL);
2392static DEVICE_ATTR(optrom_efi_version, S_IRUGO,
2393 qla2x00_optrom_efi_version_show, NULL);
2394static DEVICE_ATTR(optrom_fcode_version, S_IRUGO,
2395 qla2x00_optrom_fcode_version_show, NULL);
2396static DEVICE_ATTR(optrom_fw_version, S_IRUGO, qla2x00_optrom_fw_version_show,
2397 NULL);
2398static DEVICE_ATTR(optrom_gold_fw_version, S_IRUGO,
2399 qla2x00_optrom_gold_fw_version_show, NULL);
2400static DEVICE_ATTR(84xx_fw_version, S_IRUGO, qla24xx_84xx_fw_version_show,
2401 NULL);
2402static DEVICE_ATTR(total_isp_aborts, S_IRUGO, qla2x00_total_isp_aborts_show,
2403 NULL);
2404static DEVICE_ATTR(serdes_version, 0444, qla2x00_serdes_version_show, NULL);
2405static DEVICE_ATTR(mpi_version, S_IRUGO, qla2x00_mpi_version_show, NULL);
2406static DEVICE_ATTR(phy_version, S_IRUGO, qla2x00_phy_version_show, NULL);
2407static DEVICE_ATTR(flash_block_size, S_IRUGO, qla2x00_flash_block_size_show,
2408 NULL);
2409static DEVICE_ATTR(vlan_id, S_IRUGO, qla2x00_vlan_id_show, NULL);
2410static DEVICE_ATTR(vn_port_mac_address, S_IRUGO,
2411 qla2x00_vn_port_mac_address_show, NULL);
2412static DEVICE_ATTR(fabric_param, S_IRUGO, qla2x00_fabric_param_show, NULL);
2413static DEVICE_ATTR(fw_state, S_IRUGO, qla2x00_fw_state_show, NULL);
2414static DEVICE_ATTR(thermal_temp, S_IRUGO, qla2x00_thermal_temp_show, NULL);
2415static DEVICE_ATTR(diag_requests, S_IRUGO, qla2x00_diag_requests_show, NULL);
2416static DEVICE_ATTR(diag_megabytes, S_IRUGO, qla2x00_diag_megabytes_show, NULL);
2417static DEVICE_ATTR(fw_dump_size, S_IRUGO, qla2x00_fw_dump_size_show, NULL);
2418static DEVICE_ATTR(allow_cna_fw_dump, S_IRUGO | S_IWUSR,
2419 qla2x00_allow_cna_fw_dump_show,
2420 qla2x00_allow_cna_fw_dump_store);
2421static DEVICE_ATTR(pep_version, S_IRUGO, qla2x00_pep_version_show, NULL);
2422static DEVICE_ATTR(min_supported_speed, 0444,
2423 qla2x00_min_supported_speed_show, NULL);
2424static DEVICE_ATTR(max_supported_speed, 0444,
2425 qla2x00_max_supported_speed_show, NULL);
2426static DEVICE_ATTR(zio_threshold, 0644,
2427 qla_zio_threshold_show,
2428 qla_zio_threshold_store);
2429static DEVICE_ATTR_RW(qlini_mode);
2430static DEVICE_ATTR_RW(ql2xexchoffld);
2431static DEVICE_ATTR_RW(ql2xiniexchg);
2432static DEVICE_ATTR(dif_bundle_statistics, 0444,
2433 qla2x00_dif_bundle_statistics_show, NULL);
2434static DEVICE_ATTR(port_speed, 0644, qla2x00_port_speed_show,
2435 qla2x00_port_speed_store);
2436static DEVICE_ATTR(port_no, 0444, qla2x00_port_no_show, NULL);
2437static DEVICE_ATTR(fw_attr, 0444, qla2x00_fw_attr_show, NULL);
2438
2439
2440struct device_attribute *qla2x00_host_attrs[] = {
2441 &dev_attr_driver_version,
2442 &dev_attr_fw_version,
2443 &dev_attr_serial_num,
2444 &dev_attr_isp_name,
2445 &dev_attr_isp_id,
2446 &dev_attr_model_name,
2447 &dev_attr_model_desc,
2448 &dev_attr_pci_info,
2449 &dev_attr_link_state,
2450 &dev_attr_zio,
2451 &dev_attr_zio_timer,
2452 &dev_attr_beacon,
2453 &dev_attr_beacon_config,
2454 &dev_attr_optrom_bios_version,
2455 &dev_attr_optrom_efi_version,
2456 &dev_attr_optrom_fcode_version,
2457 &dev_attr_optrom_fw_version,
2458 &dev_attr_84xx_fw_version,
2459 &dev_attr_total_isp_aborts,
2460 &dev_attr_serdes_version,
2461 &dev_attr_mpi_version,
2462 &dev_attr_phy_version,
2463 &dev_attr_flash_block_size,
2464 &dev_attr_vlan_id,
2465 &dev_attr_vn_port_mac_address,
2466 &dev_attr_fabric_param,
2467 &dev_attr_fw_state,
2468 &dev_attr_optrom_gold_fw_version,
2469 &dev_attr_thermal_temp,
2470 &dev_attr_diag_requests,
2471 &dev_attr_diag_megabytes,
2472 &dev_attr_fw_dump_size,
2473 &dev_attr_allow_cna_fw_dump,
2474 &dev_attr_pep_version,
2475 &dev_attr_min_supported_speed,
2476 &dev_attr_max_supported_speed,
2477 &dev_attr_zio_threshold,
2478 &dev_attr_dif_bundle_statistics,
2479 &dev_attr_port_speed,
2480 &dev_attr_port_no,
2481 &dev_attr_fw_attr,
2482 &dev_attr_dport_diagnostics,
2483 NULL,
2484 NULL,
2485 NULL,
2486 NULL,
2487};
2488
2489void qla_insert_tgt_attrs(void)
2490{
2491 struct device_attribute **attr;
2492
2493
2494 for (attr = &qla2x00_host_attrs[0]; *attr; ++attr)
2495 continue;
2496
2497 *attr = &dev_attr_qlini_mode;
2498 attr++;
2499 *attr = &dev_attr_ql2xiniexchg;
2500 attr++;
2501 *attr = &dev_attr_ql2xexchoffld;
2502}
2503
2504
2505
2506static void
2507qla2x00_get_host_port_id(struct Scsi_Host *shost)
2508{
2509 scsi_qla_host_t *vha = shost_priv(shost);
2510
2511 fc_host_port_id(shost) = vha->d_id.b.domain << 16 |
2512 vha->d_id.b.area << 8 | vha->d_id.b.al_pa;
2513}
2514
2515static void
2516qla2x00_get_host_speed(struct Scsi_Host *shost)
2517{
2518 scsi_qla_host_t *vha = shost_priv(shost);
2519 u32 speed;
2520
2521 if (IS_QLAFX00(vha->hw)) {
2522 qlafx00_get_host_speed(shost);
2523 return;
2524 }
2525
2526 switch (vha->hw->link_data_rate) {
2527 case PORT_SPEED_1GB:
2528 speed = FC_PORTSPEED_1GBIT;
2529 break;
2530 case PORT_SPEED_2GB:
2531 speed = FC_PORTSPEED_2GBIT;
2532 break;
2533 case PORT_SPEED_4GB:
2534 speed = FC_PORTSPEED_4GBIT;
2535 break;
2536 case PORT_SPEED_8GB:
2537 speed = FC_PORTSPEED_8GBIT;
2538 break;
2539 case PORT_SPEED_10GB:
2540 speed = FC_PORTSPEED_10GBIT;
2541 break;
2542 case PORT_SPEED_16GB:
2543 speed = FC_PORTSPEED_16GBIT;
2544 break;
2545 case PORT_SPEED_32GB:
2546 speed = FC_PORTSPEED_32GBIT;
2547 break;
2548 case PORT_SPEED_64GB:
2549 speed = FC_PORTSPEED_64GBIT;
2550 break;
2551 default:
2552 speed = FC_PORTSPEED_UNKNOWN;
2553 break;
2554 }
2555
2556 fc_host_speed(shost) = speed;
2557}
2558
2559static void
2560qla2x00_get_host_port_type(struct Scsi_Host *shost)
2561{
2562 scsi_qla_host_t *vha = shost_priv(shost);
2563 uint32_t port_type;
2564
2565 if (vha->vp_idx) {
2566 fc_host_port_type(shost) = FC_PORTTYPE_NPIV;
2567 return;
2568 }
2569 switch (vha->hw->current_topology) {
2570 case ISP_CFG_NL:
2571 port_type = FC_PORTTYPE_LPORT;
2572 break;
2573 case ISP_CFG_FL:
2574 port_type = FC_PORTTYPE_NLPORT;
2575 break;
2576 case ISP_CFG_N:
2577 port_type = FC_PORTTYPE_PTP;
2578 break;
2579 case ISP_CFG_F:
2580 port_type = FC_PORTTYPE_NPORT;
2581 break;
2582 default:
2583 port_type = FC_PORTTYPE_UNKNOWN;
2584 break;
2585 }
2586
2587 fc_host_port_type(shost) = port_type;
2588}
2589
2590static void
2591qla2x00_get_starget_node_name(struct scsi_target *starget)
2592{
2593 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2594 scsi_qla_host_t *vha = shost_priv(host);
2595 fc_port_t *fcport;
2596 u64 node_name = 0;
2597
2598 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2599 if (fcport->rport &&
2600 starget->id == fcport->rport->scsi_target_id) {
2601 node_name = wwn_to_u64(fcport->node_name);
2602 break;
2603 }
2604 }
2605
2606 fc_starget_node_name(starget) = node_name;
2607}
2608
2609static void
2610qla2x00_get_starget_port_name(struct scsi_target *starget)
2611{
2612 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2613 scsi_qla_host_t *vha = shost_priv(host);
2614 fc_port_t *fcport;
2615 u64 port_name = 0;
2616
2617 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2618 if (fcport->rport &&
2619 starget->id == fcport->rport->scsi_target_id) {
2620 port_name = wwn_to_u64(fcport->port_name);
2621 break;
2622 }
2623 }
2624
2625 fc_starget_port_name(starget) = port_name;
2626}
2627
2628static void
2629qla2x00_get_starget_port_id(struct scsi_target *starget)
2630{
2631 struct Scsi_Host *host = dev_to_shost(starget->dev.parent);
2632 scsi_qla_host_t *vha = shost_priv(host);
2633 fc_port_t *fcport;
2634 uint32_t port_id = ~0U;
2635
2636 list_for_each_entry(fcport, &vha->vp_fcports, list) {
2637 if (fcport->rport &&
2638 starget->id == fcport->rport->scsi_target_id) {
2639 port_id = fcport->d_id.b.domain << 16 |
2640 fcport->d_id.b.area << 8 | fcport->d_id.b.al_pa;
2641 break;
2642 }
2643 }
2644
2645 fc_starget_port_id(starget) = port_id;
2646}
2647
2648static inline void
2649qla2x00_set_rport_loss_tmo(struct fc_rport *rport, uint32_t timeout)
2650{
2651 rport->dev_loss_tmo = timeout ? timeout : 1;
2652}
2653
2654static void
2655qla2x00_dev_loss_tmo_callbk(struct fc_rport *rport)
2656{
2657 struct Scsi_Host *host = rport_to_shost(rport);
2658 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2659 unsigned long flags;
2660
2661 if (!fcport)
2662 return;
2663
2664
2665
2666 qla2x00_set_fcport_state(fcport, FCS_DEVICE_DEAD);
2667
2668
2669
2670
2671
2672 spin_lock_irqsave(host->host_lock, flags);
2673 fcport->rport = fcport->drport = NULL;
2674 *((fc_port_t **)rport->dd_data) = NULL;
2675 spin_unlock_irqrestore(host->host_lock, flags);
2676
2677 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2678 return;
2679
2680 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2681 qla2x00_abort_all_cmds(fcport->vha, DID_NO_CONNECT << 16);
2682 return;
2683 }
2684}
2685
2686static void
2687qla2x00_terminate_rport_io(struct fc_rport *rport)
2688{
2689 fc_port_t *fcport = *(fc_port_t **)rport->dd_data;
2690
2691 if (!fcport)
2692 return;
2693
2694 if (test_bit(UNLOADING, &fcport->vha->dpc_flags))
2695 return;
2696
2697 if (test_bit(ABORT_ISP_ACTIVE, &fcport->vha->dpc_flags))
2698 return;
2699
2700 if (unlikely(pci_channel_offline(fcport->vha->hw->pdev))) {
2701 qla2x00_abort_all_cmds(fcport->vha, DID_NO_CONNECT << 16);
2702 return;
2703 }
2704
2705
2706
2707
2708 if (fcport->loop_id != FC_NO_LOOP_ID) {
2709 if (IS_FWI2_CAPABLE(fcport->vha->hw))
2710 fcport->vha->hw->isp_ops->fabric_logout(fcport->vha,
2711 fcport->loop_id, fcport->d_id.b.domain,
2712 fcport->d_id.b.area, fcport->d_id.b.al_pa);
2713 else
2714 qla2x00_port_logout(fcport->vha, fcport);
2715 }
2716}
2717
2718static int
2719qla2x00_issue_lip(struct Scsi_Host *shost)
2720{
2721 scsi_qla_host_t *vha = shost_priv(shost);
2722
2723 if (IS_QLAFX00(vha->hw))
2724 return 0;
2725
2726 if (vha->hw->flags.port_isolated)
2727 return 0;
2728
2729 qla2x00_loop_reset(vha);
2730 return 0;
2731}
2732
2733static struct fc_host_statistics *
2734qla2x00_get_fc_host_stats(struct Scsi_Host *shost)
2735{
2736 scsi_qla_host_t *vha = shost_priv(shost);
2737 struct qla_hw_data *ha = vha->hw;
2738 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2739 int rval;
2740 struct link_statistics *stats;
2741 dma_addr_t stats_dma;
2742 struct fc_host_statistics *p = &vha->fc_host_stat;
2743 struct qla_qpair *qpair;
2744 int i;
2745 u64 ib = 0, ob = 0, ir = 0, or = 0;
2746
2747 memset(p, -1, sizeof(*p));
2748
2749 if (IS_QLAFX00(vha->hw))
2750 goto done;
2751
2752 if (test_bit(UNLOADING, &vha->dpc_flags))
2753 goto done;
2754
2755 if (unlikely(pci_channel_offline(ha->pdev)))
2756 goto done;
2757
2758 if (qla2x00_chip_is_down(vha))
2759 goto done;
2760
2761 stats = dma_alloc_coherent(&ha->pdev->dev, sizeof(*stats), &stats_dma,
2762 GFP_KERNEL);
2763 if (!stats) {
2764 ql_log(ql_log_warn, vha, 0x707d,
2765 "Failed to allocate memory for stats.\n");
2766 goto done;
2767 }
2768
2769 rval = QLA_FUNCTION_FAILED;
2770 if (IS_FWI2_CAPABLE(ha)) {
2771 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, 0);
2772 } else if (atomic_read(&base_vha->loop_state) == LOOP_READY &&
2773 !ha->dpc_active) {
2774
2775 rval = qla2x00_get_link_status(base_vha, base_vha->loop_id,
2776 stats, stats_dma);
2777 }
2778
2779 if (rval != QLA_SUCCESS)
2780 goto done_free;
2781
2782
2783 for (i = 0; i < vha->hw->max_qpairs; i++) {
2784 qpair = vha->hw->queue_pair_map[i];
2785 if (!qpair)
2786 continue;
2787 ir += qpair->counters.input_requests;
2788 or += qpair->counters.output_requests;
2789 ib += qpair->counters.input_bytes;
2790 ob += qpair->counters.output_bytes;
2791 }
2792 ir += ha->base_qpair->counters.input_requests;
2793 or += ha->base_qpair->counters.output_requests;
2794 ib += ha->base_qpair->counters.input_bytes;
2795 ob += ha->base_qpair->counters.output_bytes;
2796
2797 ir += vha->qla_stats.input_requests;
2798 or += vha->qla_stats.output_requests;
2799 ib += vha->qla_stats.input_bytes;
2800 ob += vha->qla_stats.output_bytes;
2801
2802
2803 p->link_failure_count = le32_to_cpu(stats->link_fail_cnt);
2804 p->loss_of_sync_count = le32_to_cpu(stats->loss_sync_cnt);
2805 p->loss_of_signal_count = le32_to_cpu(stats->loss_sig_cnt);
2806 p->prim_seq_protocol_err_count = le32_to_cpu(stats->prim_seq_err_cnt);
2807 p->invalid_tx_word_count = le32_to_cpu(stats->inval_xmit_word_cnt);
2808 p->invalid_crc_count = le32_to_cpu(stats->inval_crc_cnt);
2809 if (IS_FWI2_CAPABLE(ha)) {
2810 p->lip_count = le32_to_cpu(stats->lip_cnt);
2811 p->tx_frames = le32_to_cpu(stats->tx_frames);
2812 p->rx_frames = le32_to_cpu(stats->rx_frames);
2813 p->dumped_frames = le32_to_cpu(stats->discarded_frames);
2814 p->nos_count = le32_to_cpu(stats->nos_rcvd);
2815 p->error_frames =
2816 le32_to_cpu(stats->dropped_frames) +
2817 le32_to_cpu(stats->discarded_frames);
2818 if (IS_QLA83XX(ha) || IS_QLA27XX(ha) || IS_QLA28XX(ha)) {
2819 p->rx_words = le64_to_cpu(stats->fpm_recv_word_cnt);
2820 p->tx_words = le64_to_cpu(stats->fpm_xmit_word_cnt);
2821 } else {
2822 p->rx_words = ib >> 2;
2823 p->tx_words = ob >> 2;
2824 }
2825 }
2826
2827 p->fcp_control_requests = vha->qla_stats.control_requests;
2828 p->fcp_input_requests = ir;
2829 p->fcp_output_requests = or;
2830 p->fcp_input_megabytes = ib >> 20;
2831 p->fcp_output_megabytes = ob >> 20;
2832 p->seconds_since_last_reset =
2833 get_jiffies_64() - vha->qla_stats.jiffies_at_last_reset;
2834 do_div(p->seconds_since_last_reset, HZ);
2835
2836done_free:
2837 dma_free_coherent(&ha->pdev->dev, sizeof(struct link_statistics),
2838 stats, stats_dma);
2839done:
2840 return p;
2841}
2842
2843static void
2844qla2x00_reset_host_stats(struct Scsi_Host *shost)
2845{
2846 scsi_qla_host_t *vha = shost_priv(shost);
2847 struct qla_hw_data *ha = vha->hw;
2848 struct scsi_qla_host *base_vha = pci_get_drvdata(ha->pdev);
2849 struct link_statistics *stats;
2850 dma_addr_t stats_dma;
2851 int i;
2852 struct qla_qpair *qpair;
2853
2854 memset(&vha->qla_stats, 0, sizeof(vha->qla_stats));
2855 memset(&vha->fc_host_stat, 0, sizeof(vha->fc_host_stat));
2856 for (i = 0; i < vha->hw->max_qpairs; i++) {
2857 qpair = vha->hw->queue_pair_map[i];
2858 if (!qpair)
2859 continue;
2860 memset(&qpair->counters, 0, sizeof(qpair->counters));
2861 }
2862 memset(&ha->base_qpair->counters, 0, sizeof(qpair->counters));
2863
2864 vha->qla_stats.jiffies_at_last_reset = get_jiffies_64();
2865
2866 if (IS_FWI2_CAPABLE(ha)) {
2867 int rval;
2868
2869 stats = dma_alloc_coherent(&ha->pdev->dev,
2870 sizeof(*stats), &stats_dma, GFP_KERNEL);
2871 if (!stats) {
2872 ql_log(ql_log_warn, vha, 0x70d7,
2873 "Failed to allocate memory for stats.\n");
2874 return;
2875 }
2876
2877
2878 rval = qla24xx_get_isp_stats(base_vha, stats, stats_dma, BIT_0);
2879 if (rval != QLA_SUCCESS)
2880 ql_log(ql_log_warn, vha, 0x70de,
2881 "Resetting ISP statistics failed: rval = %d\n",
2882 rval);
2883
2884 dma_free_coherent(&ha->pdev->dev, sizeof(*stats),
2885 stats, stats_dma);
2886 }
2887}
2888
2889static void
2890qla2x00_get_host_symbolic_name(struct Scsi_Host *shost)
2891{
2892 scsi_qla_host_t *vha = shost_priv(shost);
2893
2894 qla2x00_get_sym_node_name(vha, fc_host_symbolic_name(shost),
2895 sizeof(fc_host_symbolic_name(shost)));
2896}
2897
2898static void
2899qla2x00_set_host_system_hostname(struct Scsi_Host *shost)
2900{
2901 scsi_qla_host_t *vha = shost_priv(shost);
2902
2903 set_bit(REGISTER_FDMI_NEEDED, &vha->dpc_flags);
2904}
2905
2906static void
2907qla2x00_get_host_fabric_name(struct Scsi_Host *shost)
2908{
2909 scsi_qla_host_t *vha = shost_priv(shost);
2910 static const uint8_t node_name[WWN_SIZE] = {
2911 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF
2912 };
2913 u64 fabric_name = wwn_to_u64(node_name);
2914
2915 if (vha->device_flags & SWITCH_FOUND)
2916 fabric_name = wwn_to_u64(vha->fabric_node_name);
2917
2918 fc_host_fabric_name(shost) = fabric_name;
2919}
2920
2921static void
2922qla2x00_get_host_port_state(struct Scsi_Host *shost)
2923{
2924 scsi_qla_host_t *vha = shost_priv(shost);
2925 struct scsi_qla_host *base_vha = pci_get_drvdata(vha->hw->pdev);
2926
2927 if (!base_vha->flags.online) {
2928 fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE;
2929 return;
2930 }
2931
2932 switch (atomic_read(&base_vha->loop_state)) {
2933 case LOOP_UPDATE:
2934 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
2935 break;
2936 case LOOP_DOWN:
2937 if (test_bit(LOOP_RESYNC_NEEDED, &base_vha->dpc_flags))
2938 fc_host_port_state(shost) = FC_PORTSTATE_DIAGNOSTICS;
2939 else
2940 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
2941 break;
2942 case LOOP_DEAD:
2943 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
2944 break;
2945 case LOOP_READY:
2946 fc_host_port_state(shost) = FC_PORTSTATE_ONLINE;
2947 break;
2948 default:
2949 fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN;
2950 break;
2951 }
2952}
2953
2954static int
2955qla24xx_vport_create(struct fc_vport *fc_vport, bool disable)
2956{
2957 int ret = 0;
2958 uint8_t qos = 0;
2959 scsi_qla_host_t *base_vha = shost_priv(fc_vport->shost);
2960 scsi_qla_host_t *vha = NULL;
2961 struct qla_hw_data *ha = base_vha->hw;
2962 int cnt;
2963 struct req_que *req = ha->req_q_map[0];
2964 struct qla_qpair *qpair;
2965
2966 ret = qla24xx_vport_create_req_sanity_check(fc_vport);
2967 if (ret) {
2968 ql_log(ql_log_warn, vha, 0x707e,
2969 "Vport sanity check failed, status %x\n", ret);
2970 return (ret);
2971 }
2972
2973 vha = qla24xx_create_vhost(fc_vport);
2974 if (vha == NULL) {
2975 ql_log(ql_log_warn, vha, 0x707f, "Vport create host failed.\n");
2976 return FC_VPORT_FAILED;
2977 }
2978 if (disable) {
2979 atomic_set(&vha->vp_state, VP_OFFLINE);
2980 fc_vport_set_state(fc_vport, FC_VPORT_DISABLED);
2981 } else
2982 atomic_set(&vha->vp_state, VP_FAILED);
2983
2984
2985 ql_log(ql_log_info, vha, 0x7080,
2986 "VP entry id %d assigned.\n", vha->vp_idx);
2987
2988
2989 atomic_set(&vha->loop_state, LOOP_DOWN);
2990 vha->vp_err_state = VP_ERR_PORTDWN;
2991 vha->vp_prev_err_state = VP_ERR_UNKWN;
2992
2993 if (atomic_read(&base_vha->loop_state) == LOOP_DOWN ||
2994 atomic_read(&base_vha->loop_state) == LOOP_DEAD) {
2995
2996 ql_dbg(ql_dbg_user, vha, 0x7081,
2997 "Vport loop state is not UP.\n");
2998 atomic_set(&vha->loop_state, LOOP_DEAD);
2999 if (!disable)
3000 fc_vport_set_state(fc_vport, FC_VPORT_LINKDOWN);
3001 }
3002
3003 if (IS_T10_PI_CAPABLE(ha) && ql2xenabledif) {
3004 if (ha->fw_attributes & BIT_4) {
3005 int prot = 0, guard;
3006
3007 vha->flags.difdix_supported = 1;
3008 ql_dbg(ql_dbg_user, vha, 0x7082,
3009 "Registered for DIF/DIX type 1 and 3 protection.\n");
3010 if (ql2xenabledif == 1)
3011 prot = SHOST_DIX_TYPE0_PROTECTION;
3012 scsi_host_set_prot(vha->host,
3013 prot | SHOST_DIF_TYPE1_PROTECTION
3014 | SHOST_DIF_TYPE2_PROTECTION
3015 | SHOST_DIF_TYPE3_PROTECTION
3016 | SHOST_DIX_TYPE1_PROTECTION
3017 | SHOST_DIX_TYPE2_PROTECTION
3018 | SHOST_DIX_TYPE3_PROTECTION);
3019
3020 guard = SHOST_DIX_GUARD_CRC;
3021
3022 if (IS_PI_IPGUARD_CAPABLE(ha) &&
3023 (ql2xenabledif > 1 || IS_PI_DIFB_DIX0_CAPABLE(ha)))
3024 guard |= SHOST_DIX_GUARD_IP;
3025
3026 scsi_host_set_guard(vha->host, guard);
3027 } else
3028 vha->flags.difdix_supported = 0;
3029 }
3030
3031 if (scsi_add_host_with_dma(vha->host, &fc_vport->dev,
3032 &ha->pdev->dev)) {
3033 ql_dbg(ql_dbg_user, vha, 0x7083,
3034 "scsi_add_host failure for VP[%d].\n", vha->vp_idx);
3035 goto vport_create_failed_2;
3036 }
3037
3038
3039 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3040 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3041 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3042 fc_host_supported_classes(vha->host) =
3043 fc_host_supported_classes(base_vha->host);
3044 fc_host_supported_speeds(vha->host) =
3045 fc_host_supported_speeds(base_vha->host);
3046
3047 qlt_vport_create(vha, ha);
3048 qla24xx_vport_disable(fc_vport, disable);
3049
3050 if (!ql2xmqsupport || !ha->npiv_info)
3051 goto vport_queue;
3052
3053
3054 for (cnt = 0; cnt < ha->nvram_npiv_size; cnt++) {
3055 if (memcmp(ha->npiv_info[cnt].port_name, vha->port_name, 8) == 0
3056 && memcmp(ha->npiv_info[cnt].node_name, vha->node_name,
3057 8) == 0) {
3058 qos = ha->npiv_info[cnt].q_qos;
3059 break;
3060 }
3061 }
3062
3063 if (qos) {
3064 qpair = qla2xxx_create_qpair(vha, qos, vha->vp_idx, true);
3065 if (!qpair)
3066 ql_log(ql_log_warn, vha, 0x7084,
3067 "Can't create qpair for VP[%d]\n",
3068 vha->vp_idx);
3069 else {
3070 ql_dbg(ql_dbg_multiq, vha, 0xc001,
3071 "Queue pair: %d Qos: %d) created for VP[%d]\n",
3072 qpair->id, qos, vha->vp_idx);
3073 ql_dbg(ql_dbg_user, vha, 0x7085,
3074 "Queue Pair: %d Qos: %d) created for VP[%d]\n",
3075 qpair->id, qos, vha->vp_idx);
3076 req = qpair->req;
3077 vha->qpair = qpair;
3078 }
3079 }
3080
3081vport_queue:
3082 vha->req = req;
3083 return 0;
3084
3085vport_create_failed_2:
3086 qla24xx_disable_vp(vha);
3087 qla24xx_deallocate_vp_id(vha);
3088 scsi_host_put(vha->host);
3089 return FC_VPORT_FAILED;
3090}
3091
3092static int
3093qla24xx_vport_delete(struct fc_vport *fc_vport)
3094{
3095 scsi_qla_host_t *vha = fc_vport->dd_data;
3096 struct qla_hw_data *ha = vha->hw;
3097 uint16_t id = vha->vp_idx;
3098
3099 set_bit(VPORT_DELETE, &vha->dpc_flags);
3100
3101 while (test_bit(LOOP_RESYNC_ACTIVE, &vha->dpc_flags) ||
3102 test_bit(FCPORT_UPDATE_NEEDED, &vha->dpc_flags))
3103 msleep(1000);
3104
3105
3106 qla24xx_disable_vp(vha);
3107 qla2x00_wait_for_sess_deletion(vha);
3108
3109 qla_nvme_delete(vha);
3110 vha->flags.delete_progress = 1;
3111
3112 qlt_remove_target(ha, vha);
3113
3114 fc_remove_host(vha->host);
3115
3116 scsi_remove_host(vha->host);
3117
3118
3119 qla24xx_deallocate_vp_id(vha);
3120
3121 if (vha->timer_active) {
3122 qla2x00_vp_stop_timer(vha);
3123 ql_dbg(ql_dbg_user, vha, 0x7086,
3124 "Timer for the VP[%d] has stopped\n", vha->vp_idx);
3125 }
3126
3127 qla2x00_free_fcports(vha);
3128
3129 mutex_lock(&ha->vport_lock);
3130 ha->cur_vport_count--;
3131 clear_bit(vha->vp_idx, ha->vp_idx_map);
3132 mutex_unlock(&ha->vport_lock);
3133
3134 dma_free_coherent(&ha->pdev->dev, vha->gnl.size, vha->gnl.l,
3135 vha->gnl.ldma);
3136
3137 vha->gnl.l = NULL;
3138
3139 vfree(vha->scan.l);
3140
3141 if (vha->qpair && vha->qpair->vp_idx == vha->vp_idx) {
3142 if (qla2xxx_delete_qpair(vha, vha->qpair) != QLA_SUCCESS)
3143 ql_log(ql_log_warn, vha, 0x7087,
3144 "Queue Pair delete failed.\n");
3145 }
3146
3147 ql_log(ql_log_info, vha, 0x7088, "VP[%d] deleted.\n", id);
3148 scsi_host_put(vha->host);
3149 return 0;
3150}
3151
3152static int
3153qla24xx_vport_disable(struct fc_vport *fc_vport, bool disable)
3154{
3155 scsi_qla_host_t *vha = fc_vport->dd_data;
3156
3157 if (disable)
3158 qla24xx_disable_vp(vha);
3159 else
3160 qla24xx_enable_vp(vha);
3161
3162 return 0;
3163}
3164
3165struct fc_function_template qla2xxx_transport_functions = {
3166
3167 .show_host_node_name = 1,
3168 .show_host_port_name = 1,
3169 .show_host_supported_classes = 1,
3170 .show_host_supported_speeds = 1,
3171
3172 .get_host_port_id = qla2x00_get_host_port_id,
3173 .show_host_port_id = 1,
3174 .get_host_speed = qla2x00_get_host_speed,
3175 .show_host_speed = 1,
3176 .get_host_port_type = qla2x00_get_host_port_type,
3177 .show_host_port_type = 1,
3178 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3179 .show_host_symbolic_name = 1,
3180 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3181 .show_host_system_hostname = 1,
3182 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3183 .show_host_fabric_name = 1,
3184 .get_host_port_state = qla2x00_get_host_port_state,
3185 .show_host_port_state = 1,
3186
3187 .dd_fcrport_size = sizeof(struct fc_port *),
3188 .show_rport_supported_classes = 1,
3189
3190 .get_starget_node_name = qla2x00_get_starget_node_name,
3191 .show_starget_node_name = 1,
3192 .get_starget_port_name = qla2x00_get_starget_port_name,
3193 .show_starget_port_name = 1,
3194 .get_starget_port_id = qla2x00_get_starget_port_id,
3195 .show_starget_port_id = 1,
3196
3197 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3198 .show_rport_dev_loss_tmo = 1,
3199
3200 .issue_fc_host_lip = qla2x00_issue_lip,
3201 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3202 .terminate_rport_io = qla2x00_terminate_rport_io,
3203 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3204 .reset_fc_host_stats = qla2x00_reset_host_stats,
3205
3206 .vport_create = qla24xx_vport_create,
3207 .vport_disable = qla24xx_vport_disable,
3208 .vport_delete = qla24xx_vport_delete,
3209 .bsg_request = qla24xx_bsg_request,
3210 .bsg_timeout = qla24xx_bsg_timeout,
3211};
3212
3213struct fc_function_template qla2xxx_transport_vport_functions = {
3214
3215 .show_host_node_name = 1,
3216 .show_host_port_name = 1,
3217 .show_host_supported_classes = 1,
3218
3219 .get_host_port_id = qla2x00_get_host_port_id,
3220 .show_host_port_id = 1,
3221 .get_host_speed = qla2x00_get_host_speed,
3222 .show_host_speed = 1,
3223 .get_host_port_type = qla2x00_get_host_port_type,
3224 .show_host_port_type = 1,
3225 .get_host_symbolic_name = qla2x00_get_host_symbolic_name,
3226 .show_host_symbolic_name = 1,
3227 .set_host_system_hostname = qla2x00_set_host_system_hostname,
3228 .show_host_system_hostname = 1,
3229 .get_host_fabric_name = qla2x00_get_host_fabric_name,
3230 .show_host_fabric_name = 1,
3231 .get_host_port_state = qla2x00_get_host_port_state,
3232 .show_host_port_state = 1,
3233
3234 .dd_fcrport_size = sizeof(struct fc_port *),
3235 .show_rport_supported_classes = 1,
3236
3237 .get_starget_node_name = qla2x00_get_starget_node_name,
3238 .show_starget_node_name = 1,
3239 .get_starget_port_name = qla2x00_get_starget_port_name,
3240 .show_starget_port_name = 1,
3241 .get_starget_port_id = qla2x00_get_starget_port_id,
3242 .show_starget_port_id = 1,
3243
3244 .set_rport_dev_loss_tmo = qla2x00_set_rport_loss_tmo,
3245 .show_rport_dev_loss_tmo = 1,
3246
3247 .issue_fc_host_lip = qla2x00_issue_lip,
3248 .dev_loss_tmo_callbk = qla2x00_dev_loss_tmo_callbk,
3249 .terminate_rport_io = qla2x00_terminate_rport_io,
3250 .get_fc_host_stats = qla2x00_get_fc_host_stats,
3251 .reset_fc_host_stats = qla2x00_reset_host_stats,
3252
3253 .bsg_request = qla24xx_bsg_request,
3254 .bsg_timeout = qla24xx_bsg_timeout,
3255};
3256
3257void
3258qla2x00_init_host_attr(scsi_qla_host_t *vha)
3259{
3260 struct qla_hw_data *ha = vha->hw;
3261 u32 speeds = FC_PORTSPEED_UNKNOWN;
3262
3263 fc_host_dev_loss_tmo(vha->host) = ha->port_down_retry_count;
3264 fc_host_node_name(vha->host) = wwn_to_u64(vha->node_name);
3265 fc_host_port_name(vha->host) = wwn_to_u64(vha->port_name);
3266 fc_host_supported_classes(vha->host) = ha->base_qpair->enable_class_2 ?
3267 (FC_COS_CLASS2|FC_COS_CLASS3) : FC_COS_CLASS3;
3268 fc_host_max_npiv_vports(vha->host) = ha->max_npiv_vports;
3269 fc_host_npiv_vports_inuse(vha->host) = ha->cur_vport_count;
3270
3271 speeds = qla25xx_fdmi_port_speed_capability(ha);
3272
3273 fc_host_supported_speeds(vha->host) = speeds;
3274}
3275