]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * Adaptec AAC series RAID controller driver | |
3 | * (c) Copyright 2001 Red Hat Inc. | |
4 | * | |
5 | * based on the old aacraid driver that is.. | |
6 | * Adaptec aacraid device driver for Linux. | |
7 | * | |
8 | * Copyright (c) 2000-2010 Adaptec, Inc. | |
9 | * 2010-2015 PMC-Sierra, Inc. (aacraid@pmc-sierra.com) | |
10 | * 2016-2017 Microsemi Corp. (aacraid@microsemi.com) | |
11 | * | |
12 | * This program is free software; you can redistribute it and/or modify | |
13 | * it under the terms of the GNU General Public License as published by | |
14 | * the Free Software Foundation; either version 2, or (at your option) | |
15 | * any later version. | |
16 | * | |
17 | * This program is distributed in the hope that it will be useful, | |
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
20 | * GNU General Public License for more details. | |
21 | * | |
22 | * You should have received a copy of the GNU General Public License | |
23 | * along with this program; see the file COPYING. If not, write to | |
24 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | |
25 | * | |
26 | * Module Name: | |
27 | * linit.c | |
28 | * | |
29 | * Abstract: Linux Driver entry module for Adaptec RAID Array Controller | |
30 | */ | |
31 | ||
32 | ||
33 | #include <linux/compat.h> | |
34 | #include <linux/blkdev.h> | |
35 | #include <linux/completion.h> | |
36 | #include <linux/init.h> | |
37 | #include <linux/interrupt.h> | |
38 | #include <linux/kernel.h> | |
39 | #include <linux/module.h> | |
40 | #include <linux/moduleparam.h> | |
41 | #include <linux/pci.h> | |
42 | #include <linux/aer.h> | |
43 | #include <linux/pci-aspm.h> | |
44 | #include <linux/slab.h> | |
45 | #include <linux/mutex.h> | |
46 | #include <linux/spinlock.h> | |
47 | #include <linux/syscalls.h> | |
48 | #include <linux/delay.h> | |
49 | #include <linux/kthread.h> | |
50 | ||
51 | #include <scsi/scsi.h> | |
52 | #include <scsi/scsi_cmnd.h> | |
53 | #include <scsi/scsi_device.h> | |
54 | #include <scsi/scsi_host.h> | |
55 | #include <scsi/scsi_tcq.h> | |
56 | #include <scsi/scsicam.h> | |
57 | #include <scsi/scsi_eh.h> | |
58 | ||
59 | #include "aacraid.h" | |
60 | ||
61 | #define AAC_DRIVER_VERSION "1.2.1" | |
62 | #ifndef AAC_DRIVER_BRANCH | |
63 | #define AAC_DRIVER_BRANCH "" | |
64 | #endif | |
65 | #define AAC_DRIVERNAME "aacraid" | |
66 | ||
67 | #ifdef AAC_DRIVER_BUILD | |
68 | #define _str(x) #x | |
69 | #define str(x) _str(x) | |
70 | #define AAC_DRIVER_FULL_VERSION AAC_DRIVER_VERSION "[" str(AAC_DRIVER_BUILD) "]" AAC_DRIVER_BRANCH | |
71 | #else | |
72 | #define AAC_DRIVER_FULL_VERSION AAC_DRIVER_VERSION AAC_DRIVER_BRANCH | |
73 | #endif | |
74 | ||
75 | MODULE_AUTHOR("Red Hat Inc and Adaptec"); | |
76 | MODULE_DESCRIPTION("Dell PERC2, 2/Si, 3/Si, 3/Di, " | |
77 | "Adaptec Advanced Raid Products, " | |
78 | "HP NetRAID-4M, IBM ServeRAID & ICP SCSI driver"); | |
79 | MODULE_LICENSE("GPL"); | |
80 | MODULE_VERSION(AAC_DRIVER_FULL_VERSION); | |
81 | ||
82 | static DEFINE_MUTEX(aac_mutex); | |
83 | static LIST_HEAD(aac_devices); | |
84 | static int aac_cfg_major = AAC_CHARDEV_UNREGISTERED; | |
85 | char aac_driver_version[] = AAC_DRIVER_FULL_VERSION; | |
86 | ||
87 | /* | |
88 | * Because of the way Linux names scsi devices, the order in this table has | |
89 | * become important. Check for on-board Raid first, add-in cards second. | |
90 | * | |
91 | * Note: The last field is used to index into aac_drivers below. | |
92 | */ | |
93 | static const struct pci_device_id aac_pci_tbl[] = { | |
94 | { 0x1028, 0x0001, 0x1028, 0x0001, 0, 0, 0 }, /* PERC 2/Si (Iguana/PERC2Si) */ | |
95 | { 0x1028, 0x0002, 0x1028, 0x0002, 0, 0, 1 }, /* PERC 3/Di (Opal/PERC3Di) */ | |
96 | { 0x1028, 0x0003, 0x1028, 0x0003, 0, 0, 2 }, /* PERC 3/Si (SlimFast/PERC3Si */ | |
97 | { 0x1028, 0x0004, 0x1028, 0x00d0, 0, 0, 3 }, /* PERC 3/Di (Iguana FlipChip/PERC3DiF */ | |
98 | { 0x1028, 0x0002, 0x1028, 0x00d1, 0, 0, 4 }, /* PERC 3/Di (Viper/PERC3DiV) */ | |
99 | { 0x1028, 0x0002, 0x1028, 0x00d9, 0, 0, 5 }, /* PERC 3/Di (Lexus/PERC3DiL) */ | |
100 | { 0x1028, 0x000a, 0x1028, 0x0106, 0, 0, 6 }, /* PERC 3/Di (Jaguar/PERC3DiJ) */ | |
101 | { 0x1028, 0x000a, 0x1028, 0x011b, 0, 0, 7 }, /* PERC 3/Di (Dagger/PERC3DiD) */ | |
102 | { 0x1028, 0x000a, 0x1028, 0x0121, 0, 0, 8 }, /* PERC 3/Di (Boxster/PERC3DiB) */ | |
103 | { 0x9005, 0x0283, 0x9005, 0x0283, 0, 0, 9 }, /* catapult */ | |
104 | { 0x9005, 0x0284, 0x9005, 0x0284, 0, 0, 10 }, /* tomcat */ | |
105 | { 0x9005, 0x0285, 0x9005, 0x0286, 0, 0, 11 }, /* Adaptec 2120S (Crusader) */ | |
106 | { 0x9005, 0x0285, 0x9005, 0x0285, 0, 0, 12 }, /* Adaptec 2200S (Vulcan) */ | |
107 | { 0x9005, 0x0285, 0x9005, 0x0287, 0, 0, 13 }, /* Adaptec 2200S (Vulcan-2m) */ | |
108 | { 0x9005, 0x0285, 0x17aa, 0x0286, 0, 0, 14 }, /* Legend S220 (Legend Crusader) */ | |
109 | { 0x9005, 0x0285, 0x17aa, 0x0287, 0, 0, 15 }, /* Legend S230 (Legend Vulcan) */ | |
110 | ||
111 | { 0x9005, 0x0285, 0x9005, 0x0288, 0, 0, 16 }, /* Adaptec 3230S (Harrier) */ | |
112 | { 0x9005, 0x0285, 0x9005, 0x0289, 0, 0, 17 }, /* Adaptec 3240S (Tornado) */ | |
113 | { 0x9005, 0x0285, 0x9005, 0x028a, 0, 0, 18 }, /* ASR-2020ZCR SCSI PCI-X ZCR (Skyhawk) */ | |
114 | { 0x9005, 0x0285, 0x9005, 0x028b, 0, 0, 19 }, /* ASR-2025ZCR SCSI SO-DIMM PCI-X ZCR (Terminator) */ | |
115 | { 0x9005, 0x0286, 0x9005, 0x028c, 0, 0, 20 }, /* ASR-2230S + ASR-2230SLP PCI-X (Lancer) */ | |
116 | { 0x9005, 0x0286, 0x9005, 0x028d, 0, 0, 21 }, /* ASR-2130S (Lancer) */ | |
117 | { 0x9005, 0x0286, 0x9005, 0x029b, 0, 0, 22 }, /* AAR-2820SA (Intruder) */ | |
118 | { 0x9005, 0x0286, 0x9005, 0x029c, 0, 0, 23 }, /* AAR-2620SA (Intruder) */ | |
119 | { 0x9005, 0x0286, 0x9005, 0x029d, 0, 0, 24 }, /* AAR-2420SA (Intruder) */ | |
120 | { 0x9005, 0x0286, 0x9005, 0x029e, 0, 0, 25 }, /* ICP9024RO (Lancer) */ | |
121 | { 0x9005, 0x0286, 0x9005, 0x029f, 0, 0, 26 }, /* ICP9014RO (Lancer) */ | |
122 | { 0x9005, 0x0286, 0x9005, 0x02a0, 0, 0, 27 }, /* ICP9047MA (Lancer) */ | |
123 | { 0x9005, 0x0286, 0x9005, 0x02a1, 0, 0, 28 }, /* ICP9087MA (Lancer) */ | |
124 | { 0x9005, 0x0286, 0x9005, 0x02a3, 0, 0, 29 }, /* ICP5445AU (Hurricane44) */ | |
125 | { 0x9005, 0x0285, 0x9005, 0x02a4, 0, 0, 30 }, /* ICP9085LI (Marauder-X) */ | |
126 | { 0x9005, 0x0285, 0x9005, 0x02a5, 0, 0, 31 }, /* ICP5085BR (Marauder-E) */ | |
127 | { 0x9005, 0x0286, 0x9005, 0x02a6, 0, 0, 32 }, /* ICP9067MA (Intruder-6) */ | |
128 | { 0x9005, 0x0287, 0x9005, 0x0800, 0, 0, 33 }, /* Themisto Jupiter Platform */ | |
129 | { 0x9005, 0x0200, 0x9005, 0x0200, 0, 0, 33 }, /* Themisto Jupiter Platform */ | |
130 | { 0x9005, 0x0286, 0x9005, 0x0800, 0, 0, 34 }, /* Callisto Jupiter Platform */ | |
131 | { 0x9005, 0x0285, 0x9005, 0x028e, 0, 0, 35 }, /* ASR-2020SA SATA PCI-X ZCR (Skyhawk) */ | |
132 | { 0x9005, 0x0285, 0x9005, 0x028f, 0, 0, 36 }, /* ASR-2025SA SATA SO-DIMM PCI-X ZCR (Terminator) */ | |
133 | { 0x9005, 0x0285, 0x9005, 0x0290, 0, 0, 37 }, /* AAR-2410SA PCI SATA 4ch (Jaguar II) */ | |
134 | { 0x9005, 0x0285, 0x1028, 0x0291, 0, 0, 38 }, /* CERC SATA RAID 2 PCI SATA 6ch (DellCorsair) */ | |
135 | { 0x9005, 0x0285, 0x9005, 0x0292, 0, 0, 39 }, /* AAR-2810SA PCI SATA 8ch (Corsair-8) */ | |
136 | { 0x9005, 0x0285, 0x9005, 0x0293, 0, 0, 40 }, /* AAR-21610SA PCI SATA 16ch (Corsair-16) */ | |
137 | { 0x9005, 0x0285, 0x9005, 0x0294, 0, 0, 41 }, /* ESD SO-DIMM PCI-X SATA ZCR (Prowler) */ | |
138 | { 0x9005, 0x0285, 0x103C, 0x3227, 0, 0, 42 }, /* AAR-2610SA PCI SATA 6ch */ | |
139 | { 0x9005, 0x0285, 0x9005, 0x0296, 0, 0, 43 }, /* ASR-2240S (SabreExpress) */ | |
140 | { 0x9005, 0x0285, 0x9005, 0x0297, 0, 0, 44 }, /* ASR-4005 */ | |
141 | { 0x9005, 0x0285, 0x1014, 0x02F2, 0, 0, 45 }, /* IBM 8i (AvonPark) */ | |
142 | { 0x9005, 0x0285, 0x1014, 0x0312, 0, 0, 45 }, /* IBM 8i (AvonPark Lite) */ | |
143 | { 0x9005, 0x0286, 0x1014, 0x9580, 0, 0, 46 }, /* IBM 8k/8k-l8 (Aurora) */ | |
144 | { 0x9005, 0x0286, 0x1014, 0x9540, 0, 0, 47 }, /* IBM 8k/8k-l4 (Aurora Lite) */ | |
145 | { 0x9005, 0x0285, 0x9005, 0x0298, 0, 0, 48 }, /* ASR-4000 (BlackBird) */ | |
146 | { 0x9005, 0x0285, 0x9005, 0x0299, 0, 0, 49 }, /* ASR-4800SAS (Marauder-X) */ | |
147 | { 0x9005, 0x0285, 0x9005, 0x029a, 0, 0, 50 }, /* ASR-4805SAS (Marauder-E) */ | |
148 | { 0x9005, 0x0286, 0x9005, 0x02a2, 0, 0, 51 }, /* ASR-3800 (Hurricane44) */ | |
149 | ||
150 | { 0x9005, 0x0285, 0x1028, 0x0287, 0, 0, 52 }, /* Perc 320/DC*/ | |
151 | { 0x1011, 0x0046, 0x9005, 0x0365, 0, 0, 53 }, /* Adaptec 5400S (Mustang)*/ | |
152 | { 0x1011, 0x0046, 0x9005, 0x0364, 0, 0, 54 }, /* Adaptec 5400S (Mustang)*/ | |
153 | { 0x1011, 0x0046, 0x9005, 0x1364, 0, 0, 55 }, /* Dell PERC2/QC */ | |
154 | { 0x1011, 0x0046, 0x103c, 0x10c2, 0, 0, 56 }, /* HP NetRAID-4M */ | |
155 | ||
156 | { 0x9005, 0x0285, 0x1028, PCI_ANY_ID, 0, 0, 57 }, /* Dell Catchall */ | |
157 | { 0x9005, 0x0285, 0x17aa, PCI_ANY_ID, 0, 0, 58 }, /* Legend Catchall */ | |
158 | { 0x9005, 0x0285, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 59 }, /* Adaptec Catch All */ | |
159 | { 0x9005, 0x0286, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 60 }, /* Adaptec Rocket Catch All */ | |
160 | { 0x9005, 0x0288, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 61 }, /* Adaptec NEMER/ARK Catch All */ | |
161 | { 0x9005, 0x028b, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 62 }, /* Adaptec PMC Series 6 (Tupelo) */ | |
162 | { 0x9005, 0x028c, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 63 }, /* Adaptec PMC Series 7 (Denali) */ | |
163 | { 0x9005, 0x028d, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 64 }, /* Adaptec PMC Series 8 */ | |
164 | { 0,} | |
165 | }; | |
166 | MODULE_DEVICE_TABLE(pci, aac_pci_tbl); | |
167 | ||
168 | /* | |
169 | * dmb - For now we add the number of channels to this structure. | |
170 | * In the future we should add a fib that reports the number of channels | |
171 | * for the card. At that time we can remove the channels from here | |
172 | */ | |
173 | static struct aac_driver_ident aac_drivers[] = { | |
174 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 2/Si (Iguana/PERC2Si) */ | |
175 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Opal/PERC3Di) */ | |
176 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Si (SlimFast/PERC3Si */ | |
177 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Iguana FlipChip/PERC3DiF */ | |
178 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Viper/PERC3DiV) */ | |
179 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Lexus/PERC3DiL) */ | |
180 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 1, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Jaguar/PERC3DiJ) */ | |
181 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Dagger/PERC3DiD) */ | |
182 | { aac_rx_init, "percraid", "DELL ", "PERCRAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* PERC 3/Di (Boxster/PERC3DiB) */ | |
183 | { aac_rx_init, "aacraid", "ADAPTEC ", "catapult ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* catapult */ | |
184 | { aac_rx_init, "aacraid", "ADAPTEC ", "tomcat ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* tomcat */ | |
185 | { aac_rx_init, "aacraid", "ADAPTEC ", "Adaptec 2120S ", 1, AAC_QUIRK_31BIT | AAC_QUIRK_34SG }, /* Adaptec 2120S (Crusader) */ | |
186 | { aac_rx_init, "aacraid", "ADAPTEC ", "Adaptec 2200S ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG }, /* Adaptec 2200S (Vulcan) */ | |
187 | { aac_rx_init, "aacraid", "ADAPTEC ", "Adaptec 2200S ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* Adaptec 2200S (Vulcan-2m) */ | |
188 | { aac_rx_init, "aacraid", "Legend ", "Legend S220 ", 1, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* Legend S220 (Legend Crusader) */ | |
189 | { aac_rx_init, "aacraid", "Legend ", "Legend S230 ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* Legend S230 (Legend Vulcan) */ | |
190 | ||
191 | { aac_rx_init, "aacraid", "ADAPTEC ", "Adaptec 3230S ", 2 }, /* Adaptec 3230S (Harrier) */ | |
192 | { aac_rx_init, "aacraid", "ADAPTEC ", "Adaptec 3240S ", 2 }, /* Adaptec 3240S (Tornado) */ | |
193 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2020ZCR ", 2 }, /* ASR-2020ZCR SCSI PCI-X ZCR (Skyhawk) */ | |
194 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2025ZCR ", 2 }, /* ASR-2025ZCR SCSI SO-DIMM PCI-X ZCR (Terminator) */ | |
195 | { aac_rkt_init, "aacraid", "ADAPTEC ", "ASR-2230S PCI-X ", 2 }, /* ASR-2230S + ASR-2230SLP PCI-X (Lancer) */ | |
196 | { aac_rkt_init, "aacraid", "ADAPTEC ", "ASR-2130S PCI-X ", 1 }, /* ASR-2130S (Lancer) */ | |
197 | { aac_rkt_init, "aacraid", "ADAPTEC ", "AAR-2820SA ", 1 }, /* AAR-2820SA (Intruder) */ | |
198 | { aac_rkt_init, "aacraid", "ADAPTEC ", "AAR-2620SA ", 1 }, /* AAR-2620SA (Intruder) */ | |
199 | { aac_rkt_init, "aacraid", "ADAPTEC ", "AAR-2420SA ", 1 }, /* AAR-2420SA (Intruder) */ | |
200 | { aac_rkt_init, "aacraid", "ICP ", "ICP9024RO ", 2 }, /* ICP9024RO (Lancer) */ | |
201 | { aac_rkt_init, "aacraid", "ICP ", "ICP9014RO ", 1 }, /* ICP9014RO (Lancer) */ | |
202 | { aac_rkt_init, "aacraid", "ICP ", "ICP9047MA ", 1 }, /* ICP9047MA (Lancer) */ | |
203 | { aac_rkt_init, "aacraid", "ICP ", "ICP9087MA ", 1 }, /* ICP9087MA (Lancer) */ | |
204 | { aac_rkt_init, "aacraid", "ICP ", "ICP5445AU ", 1 }, /* ICP5445AU (Hurricane44) */ | |
205 | { aac_rx_init, "aacraid", "ICP ", "ICP9085LI ", 1 }, /* ICP9085LI (Marauder-X) */ | |
206 | { aac_rx_init, "aacraid", "ICP ", "ICP5085BR ", 1 }, /* ICP5085BR (Marauder-E) */ | |
207 | { aac_rkt_init, "aacraid", "ICP ", "ICP9067MA ", 1 }, /* ICP9067MA (Intruder-6) */ | |
208 | { NULL , "aacraid", "ADAPTEC ", "Themisto ", 0, AAC_QUIRK_SLAVE }, /* Jupiter Platform */ | |
209 | { aac_rkt_init, "aacraid", "ADAPTEC ", "Callisto ", 2, AAC_QUIRK_MASTER }, /* Jupiter Platform */ | |
210 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2020SA ", 1 }, /* ASR-2020SA SATA PCI-X ZCR (Skyhawk) */ | |
211 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2025SA ", 1 }, /* ASR-2025SA SATA SO-DIMM PCI-X ZCR (Terminator) */ | |
212 | { aac_rx_init, "aacraid", "ADAPTEC ", "AAR-2410SA SATA ", 1, AAC_QUIRK_17SG }, /* AAR-2410SA PCI SATA 4ch (Jaguar II) */ | |
213 | { aac_rx_init, "aacraid", "DELL ", "CERC SR2 ", 1, AAC_QUIRK_17SG }, /* CERC SATA RAID 2 PCI SATA 6ch (DellCorsair) */ | |
214 | { aac_rx_init, "aacraid", "ADAPTEC ", "AAR-2810SA SATA ", 1, AAC_QUIRK_17SG }, /* AAR-2810SA PCI SATA 8ch (Corsair-8) */ | |
215 | { aac_rx_init, "aacraid", "ADAPTEC ", "AAR-21610SA SATA", 1, AAC_QUIRK_17SG }, /* AAR-21610SA PCI SATA 16ch (Corsair-16) */ | |
216 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2026ZCR ", 1 }, /* ESD SO-DIMM PCI-X SATA ZCR (Prowler) */ | |
217 | { aac_rx_init, "aacraid", "ADAPTEC ", "AAR-2610SA ", 1 }, /* SATA 6Ch (Bearcat) */ | |
218 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-2240S ", 1 }, /* ASR-2240S (SabreExpress) */ | |
219 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-4005 ", 1 }, /* ASR-4005 */ | |
220 | { aac_rx_init, "ServeRAID","IBM ", "ServeRAID 8i ", 1 }, /* IBM 8i (AvonPark) */ | |
221 | { aac_rkt_init, "ServeRAID","IBM ", "ServeRAID 8k-l8 ", 1 }, /* IBM 8k/8k-l8 (Aurora) */ | |
222 | { aac_rkt_init, "ServeRAID","IBM ", "ServeRAID 8k-l4 ", 1 }, /* IBM 8k/8k-l4 (Aurora Lite) */ | |
223 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-4000 ", 1 }, /* ASR-4000 (BlackBird & AvonPark) */ | |
224 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-4800SAS ", 1 }, /* ASR-4800SAS (Marauder-X) */ | |
225 | { aac_rx_init, "aacraid", "ADAPTEC ", "ASR-4805SAS ", 1 }, /* ASR-4805SAS (Marauder-E) */ | |
226 | { aac_rkt_init, "aacraid", "ADAPTEC ", "ASR-3800 ", 1 }, /* ASR-3800 (Hurricane44) */ | |
227 | ||
228 | { aac_rx_init, "percraid", "DELL ", "PERC 320/DC ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG }, /* Perc 320/DC*/ | |
229 | { aac_sa_init, "aacraid", "ADAPTEC ", "Adaptec 5400S ", 4, AAC_QUIRK_34SG }, /* Adaptec 5400S (Mustang)*/ | |
230 | { aac_sa_init, "aacraid", "ADAPTEC ", "AAC-364 ", 4, AAC_QUIRK_34SG }, /* Adaptec 5400S (Mustang)*/ | |
231 | { aac_sa_init, "percraid", "DELL ", "PERCRAID ", 4, AAC_QUIRK_34SG }, /* Dell PERC2/QC */ | |
232 | { aac_sa_init, "hpnraid", "HP ", "NetRAID ", 4, AAC_QUIRK_34SG }, /* HP NetRAID-4M */ | |
233 | ||
234 | { aac_rx_init, "aacraid", "DELL ", "RAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* Dell Catchall */ | |
235 | { aac_rx_init, "aacraid", "Legend ", "RAID ", 2, AAC_QUIRK_31BIT | AAC_QUIRK_34SG | AAC_QUIRK_SCSI_32 }, /* Legend Catchall */ | |
236 | { aac_rx_init, "aacraid", "ADAPTEC ", "RAID ", 2 }, /* Adaptec Catch All */ | |
237 | { aac_rkt_init, "aacraid", "ADAPTEC ", "RAID ", 2 }, /* Adaptec Rocket Catch All */ | |
238 | { aac_nark_init, "aacraid", "ADAPTEC ", "RAID ", 2 }, /* Adaptec NEMER/ARK Catch All */ | |
239 | { aac_src_init, "aacraid", "ADAPTEC ", "RAID ", 2, AAC_QUIRK_SRC }, /* Adaptec PMC Series 6 (Tupelo) */ | |
240 | { aac_srcv_init, "aacraid", "ADAPTEC ", "RAID ", 2, AAC_QUIRK_SRC }, /* Adaptec PMC Series 7 (Denali) */ | |
241 | { aac_srcv_init, "aacraid", "ADAPTEC ", "RAID ", 2, AAC_QUIRK_SRC }, /* Adaptec PMC Series 8 */ | |
242 | }; | |
243 | ||
244 | /** | |
245 | * aac_queuecommand - queue a SCSI command | |
246 | * @cmd: SCSI command to queue | |
247 | * @done: Function to call on command completion | |
248 | * | |
249 | * Queues a command for execution by the associated Host Adapter. | |
250 | * | |
251 | * TODO: unify with aac_scsi_cmd(). | |
252 | */ | |
253 | ||
254 | static int aac_queuecommand(struct Scsi_Host *shost, | |
255 | struct scsi_cmnd *cmd) | |
256 | { | |
257 | int r = 0; | |
258 | cmd->SCp.phase = AAC_OWNER_LOWLEVEL; | |
259 | r = (aac_scsi_cmd(cmd) ? FAILED : 0); | |
260 | return r; | |
261 | } | |
262 | ||
263 | /** | |
264 | * aac_info - Returns the host adapter name | |
265 | * @shost: Scsi host to report on | |
266 | * | |
267 | * Returns a static string describing the device in question | |
268 | */ | |
269 | ||
270 | static const char *aac_info(struct Scsi_Host *shost) | |
271 | { | |
272 | struct aac_dev *dev = (struct aac_dev *)shost->hostdata; | |
273 | return aac_drivers[dev->cardtype].name; | |
274 | } | |
275 | ||
276 | /** | |
277 | * aac_get_driver_ident | |
278 | * @devtype: index into lookup table | |
279 | * | |
280 | * Returns a pointer to the entry in the driver lookup table. | |
281 | */ | |
282 | ||
283 | struct aac_driver_ident* aac_get_driver_ident(int devtype) | |
284 | { | |
285 | return &aac_drivers[devtype]; | |
286 | } | |
287 | ||
288 | /** | |
289 | * aac_biosparm - return BIOS parameters for disk | |
290 | * @sdev: The scsi device corresponding to the disk | |
291 | * @bdev: the block device corresponding to the disk | |
292 | * @capacity: the sector capacity of the disk | |
293 | * @geom: geometry block to fill in | |
294 | * | |
295 | * Return the Heads/Sectors/Cylinders BIOS Disk Parameters for Disk. | |
296 | * The default disk geometry is 64 heads, 32 sectors, and the appropriate | |
297 | * number of cylinders so as not to exceed drive capacity. In order for | |
298 | * disks equal to or larger than 1 GB to be addressable by the BIOS | |
299 | * without exceeding the BIOS limitation of 1024 cylinders, Extended | |
300 | * Translation should be enabled. With Extended Translation enabled, | |
301 | * drives between 1 GB inclusive and 2 GB exclusive are given a disk | |
302 | * geometry of 128 heads and 32 sectors, and drives above 2 GB inclusive | |
303 | * are given a disk geometry of 255 heads and 63 sectors. However, if | |
304 | * the BIOS detects that the Extended Translation setting does not match | |
305 | * the geometry in the partition table, then the translation inferred | |
306 | * from the partition table will be used by the BIOS, and a warning may | |
307 | * be displayed. | |
308 | */ | |
309 | ||
310 | static int aac_biosparm(struct scsi_device *sdev, struct block_device *bdev, | |
311 | sector_t capacity, int *geom) | |
312 | { | |
313 | struct diskparm *param = (struct diskparm *)geom; | |
314 | unsigned char *buf; | |
315 | ||
316 | dprintk((KERN_DEBUG "aac_biosparm.\n")); | |
317 | ||
318 | /* | |
319 | * Assuming extended translation is enabled - #REVISIT# | |
320 | */ | |
321 | if (capacity >= 2 * 1024 * 1024) { /* 1 GB in 512 byte sectors */ | |
322 | if(capacity >= 4 * 1024 * 1024) { /* 2 GB in 512 byte sectors */ | |
323 | param->heads = 255; | |
324 | param->sectors = 63; | |
325 | } else { | |
326 | param->heads = 128; | |
327 | param->sectors = 32; | |
328 | } | |
329 | } else { | |
330 | param->heads = 64; | |
331 | param->sectors = 32; | |
332 | } | |
333 | ||
334 | param->cylinders = cap_to_cyls(capacity, param->heads * param->sectors); | |
335 | ||
336 | /* | |
337 | * Read the first 1024 bytes from the disk device, if the boot | |
338 | * sector partition table is valid, search for a partition table | |
339 | * entry whose end_head matches one of the standard geometry | |
340 | * translations ( 64/32, 128/32, 255/63 ). | |
341 | */ | |
342 | buf = scsi_bios_ptable(bdev); | |
343 | if (!buf) | |
344 | return 0; | |
345 | if(*(__le16 *)(buf + 0x40) == cpu_to_le16(0xaa55)) { | |
346 | struct partition *first = (struct partition * )buf; | |
347 | struct partition *entry = first; | |
348 | int saved_cylinders = param->cylinders; | |
349 | int num; | |
350 | unsigned char end_head, end_sec; | |
351 | ||
352 | for(num = 0; num < 4; num++) { | |
353 | end_head = entry->end_head; | |
354 | end_sec = entry->end_sector & 0x3f; | |
355 | ||
356 | if(end_head == 63) { | |
357 | param->heads = 64; | |
358 | param->sectors = 32; | |
359 | break; | |
360 | } else if(end_head == 127) { | |
361 | param->heads = 128; | |
362 | param->sectors = 32; | |
363 | break; | |
364 | } else if(end_head == 254) { | |
365 | param->heads = 255; | |
366 | param->sectors = 63; | |
367 | break; | |
368 | } | |
369 | entry++; | |
370 | } | |
371 | ||
372 | if (num == 4) { | |
373 | end_head = first->end_head; | |
374 | end_sec = first->end_sector & 0x3f; | |
375 | } | |
376 | ||
377 | param->cylinders = cap_to_cyls(capacity, param->heads * param->sectors); | |
378 | if (num < 4 && end_sec == param->sectors) { | |
379 | if (param->cylinders != saved_cylinders) | |
380 | dprintk((KERN_DEBUG "Adopting geometry: heads=%d, sectors=%d from partition table %d.\n", | |
381 | param->heads, param->sectors, num)); | |
382 | } else if (end_head > 0 || end_sec > 0) { | |
383 | dprintk((KERN_DEBUG "Strange geometry: heads=%d, sectors=%d in partition table %d.\n", | |
384 | end_head + 1, end_sec, num)); | |
385 | dprintk((KERN_DEBUG "Using geometry: heads=%d, sectors=%d.\n", | |
386 | param->heads, param->sectors)); | |
387 | } | |
388 | } | |
389 | kfree(buf); | |
390 | return 0; | |
391 | } | |
392 | ||
393 | /** | |
394 | * aac_slave_configure - compute queue depths | |
395 | * @sdev: SCSI device we are considering | |
396 | * | |
397 | * Selects queue depths for each target device based on the host adapter's | |
398 | * total capacity and the queue depth supported by the target device. | |
399 | * A queue depth of one automatically disables tagged queueing. | |
400 | */ | |
401 | ||
402 | static int aac_slave_configure(struct scsi_device *sdev) | |
403 | { | |
404 | struct aac_dev *aac = (struct aac_dev *)sdev->host->hostdata; | |
405 | int chn, tid; | |
406 | unsigned int depth = 0; | |
407 | unsigned int set_timeout = 0; | |
408 | ||
409 | chn = aac_logical_to_phys(sdev_channel(sdev)); | |
410 | tid = sdev_id(sdev); | |
411 | if (chn < AAC_MAX_BUSES && tid < AAC_MAX_TARGETS && | |
412 | aac->hba_map[chn][tid].devtype == AAC_DEVTYPE_NATIVE_RAW) { | |
413 | depth = aac->hba_map[chn][tid].qd_limit; | |
414 | set_timeout = 1; | |
415 | goto common_config; | |
416 | } | |
417 | ||
418 | ||
419 | if (aac->jbod && (sdev->type == TYPE_DISK)) | |
420 | sdev->removable = 1; | |
421 | ||
422 | if (sdev->type == TYPE_DISK | |
423 | && sdev_channel(sdev) != CONTAINER_CHANNEL | |
424 | && (!aac->jbod || sdev->inq_periph_qual) | |
425 | && (!aac->raid_scsi_mode || (sdev_channel(sdev) != 2))) { | |
426 | ||
427 | if (expose_physicals == 0) | |
428 | return -ENXIO; | |
429 | ||
430 | if (expose_physicals < 0) | |
431 | sdev->no_uld_attach = 1; | |
432 | } | |
433 | ||
434 | if (sdev->tagged_supported | |
435 | && sdev->type == TYPE_DISK | |
436 | && (!aac->raid_scsi_mode || (sdev_channel(sdev) != 2)) | |
437 | && !sdev->no_uld_attach) { | |
438 | ||
439 | struct scsi_device * dev; | |
440 | struct Scsi_Host *host = sdev->host; | |
441 | unsigned num_lsu = 0; | |
442 | unsigned num_one = 0; | |
443 | unsigned cid; | |
444 | ||
445 | set_timeout = 1; | |
446 | ||
447 | for (cid = 0; cid < aac->maximum_num_containers; ++cid) | |
448 | if (aac->fsa_dev[cid].valid) | |
449 | ++num_lsu; | |
450 | ||
451 | __shost_for_each_device(dev, host) { | |
452 | if (dev->tagged_supported | |
453 | && dev->type == TYPE_DISK | |
454 | && (!aac->raid_scsi_mode || (sdev_channel(sdev) != 2)) | |
455 | && !dev->no_uld_attach) { | |
456 | if ((sdev_channel(dev) != CONTAINER_CHANNEL) | |
457 | || !aac->fsa_dev[sdev_id(dev)].valid) { | |
458 | ++num_lsu; | |
459 | } | |
460 | } else { | |
461 | ++num_one; | |
462 | } | |
463 | } | |
464 | ||
465 | if (num_lsu == 0) | |
466 | ++num_lsu; | |
467 | ||
468 | depth = (host->can_queue - num_one) / num_lsu; | |
469 | ||
470 | if (sdev_channel(sdev) != NATIVE_CHANNEL) | |
471 | goto common_config; | |
472 | ||
473 | /* | |
474 | * Check if SATA drive | |
475 | */ | |
476 | if (strncmp(sdev->vendor, "ATA", 3) == 0) | |
477 | depth = 32; | |
478 | else | |
479 | depth = 64; | |
480 | } | |
481 | ||
482 | common_config: | |
483 | /* | |
484 | * Firmware has an individual device recovery time typically | |
485 | * of 35 seconds, give us a margin. | |
486 | */ | |
487 | if (set_timeout && sdev->request_queue->rq_timeout < (45 * HZ)) | |
488 | blk_queue_rq_timeout(sdev->request_queue, 45*HZ); | |
489 | ||
490 | if (depth > 256) | |
491 | depth = 256; | |
492 | else if (depth < 1) | |
493 | depth = 1; | |
494 | ||
495 | scsi_change_queue_depth(sdev, depth); | |
496 | ||
497 | sdev->tagged_supported = 1; | |
498 | ||
499 | return 0; | |
500 | } | |
501 | ||
502 | /** | |
503 | * aac_change_queue_depth - alter queue depths | |
504 | * @sdev: SCSI device we are considering | |
505 | * @depth: desired queue depth | |
506 | * | |
507 | * Alters queue depths for target device based on the host adapter's | |
508 | * total capacity and the queue depth supported by the target device. | |
509 | */ | |
510 | ||
511 | static int aac_change_queue_depth(struct scsi_device *sdev, int depth) | |
512 | { | |
513 | struct aac_dev *aac = (struct aac_dev *)(sdev->host->hostdata); | |
514 | int chn, tid, is_native_device = 0; | |
515 | ||
516 | chn = aac_logical_to_phys(sdev_channel(sdev)); | |
517 | tid = sdev_id(sdev); | |
518 | if (chn < AAC_MAX_BUSES && tid < AAC_MAX_TARGETS && | |
519 | aac->hba_map[chn][tid].devtype == AAC_DEVTYPE_NATIVE_RAW) | |
520 | is_native_device = 1; | |
521 | ||
522 | if (sdev->tagged_supported && (sdev->type == TYPE_DISK) && | |
523 | (sdev_channel(sdev) == CONTAINER_CHANNEL)) { | |
524 | struct scsi_device * dev; | |
525 | struct Scsi_Host *host = sdev->host; | |
526 | unsigned num = 0; | |
527 | ||
528 | __shost_for_each_device(dev, host) { | |
529 | if (dev->tagged_supported && (dev->type == TYPE_DISK) && | |
530 | (sdev_channel(dev) == CONTAINER_CHANNEL)) | |
531 | ++num; | |
532 | ++num; | |
533 | } | |
534 | if (num >= host->can_queue) | |
535 | num = host->can_queue - 1; | |
536 | if (depth > (host->can_queue - num)) | |
537 | depth = host->can_queue - num; | |
538 | if (depth > 256) | |
539 | depth = 256; | |
540 | else if (depth < 2) | |
541 | depth = 2; | |
542 | return scsi_change_queue_depth(sdev, depth); | |
543 | } else if (is_native_device) { | |
544 | scsi_change_queue_depth(sdev, aac->hba_map[chn][tid].qd_limit); | |
545 | } else { | |
546 | scsi_change_queue_depth(sdev, 1); | |
547 | } | |
548 | return sdev->queue_depth; | |
549 | } | |
550 | ||
551 | static ssize_t aac_show_raid_level(struct device *dev, struct device_attribute *attr, char *buf) | |
552 | { | |
553 | struct scsi_device *sdev = to_scsi_device(dev); | |
554 | struct aac_dev *aac = (struct aac_dev *)(sdev->host->hostdata); | |
555 | if (sdev_channel(sdev) != CONTAINER_CHANNEL) | |
556 | return snprintf(buf, PAGE_SIZE, sdev->no_uld_attach | |
557 | ? "Hidden\n" : | |
558 | ((aac->jbod && (sdev->type == TYPE_DISK)) ? "JBOD\n" : "")); | |
559 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
560 | get_container_type(aac->fsa_dev[sdev_id(sdev)].type)); | |
561 | } | |
562 | ||
563 | static struct device_attribute aac_raid_level_attr = { | |
564 | .attr = { | |
565 | .name = "level", | |
566 | .mode = S_IRUGO, | |
567 | }, | |
568 | .show = aac_show_raid_level | |
569 | }; | |
570 | ||
571 | static ssize_t aac_show_unique_id(struct device *dev, | |
572 | struct device_attribute *attr, char *buf) | |
573 | { | |
574 | struct scsi_device *sdev = to_scsi_device(dev); | |
575 | struct aac_dev *aac = (struct aac_dev *)(sdev->host->hostdata); | |
576 | unsigned char sn[16]; | |
577 | ||
578 | memset(sn, 0, sizeof(sn)); | |
579 | ||
580 | if (sdev_channel(sdev) == CONTAINER_CHANNEL) | |
581 | memcpy(sn, aac->fsa_dev[sdev_id(sdev)].identifier, sizeof(sn)); | |
582 | ||
583 | return snprintf(buf, 16 * 2 + 2, | |
584 | "%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X%02X\n", | |
585 | sn[0], sn[1], sn[2], sn[3], | |
586 | sn[4], sn[5], sn[6], sn[7], | |
587 | sn[8], sn[9], sn[10], sn[11], | |
588 | sn[12], sn[13], sn[14], sn[15]); | |
589 | } | |
590 | ||
591 | static struct device_attribute aac_unique_id_attr = { | |
592 | .attr = { | |
593 | .name = "unique_id", | |
594 | .mode = 0444, | |
595 | }, | |
596 | .show = aac_show_unique_id | |
597 | }; | |
598 | ||
599 | ||
600 | ||
601 | static struct device_attribute *aac_dev_attrs[] = { | |
602 | &aac_raid_level_attr, | |
603 | &aac_unique_id_attr, | |
604 | NULL, | |
605 | }; | |
606 | ||
607 | static int aac_ioctl(struct scsi_device *sdev, int cmd, void __user * arg) | |
608 | { | |
609 | struct aac_dev *dev = (struct aac_dev *)sdev->host->hostdata; | |
610 | if (!capable(CAP_SYS_RAWIO)) | |
611 | return -EPERM; | |
612 | return aac_do_ioctl(dev, cmd, arg); | |
613 | } | |
614 | ||
615 | static int aac_eh_abort(struct scsi_cmnd* cmd) | |
616 | { | |
617 | struct scsi_device * dev = cmd->device; | |
618 | struct Scsi_Host * host = dev->host; | |
619 | struct aac_dev * aac = (struct aac_dev *)host->hostdata; | |
620 | int count, found; | |
621 | u32 bus, cid; | |
622 | int ret = FAILED; | |
623 | ||
624 | bus = aac_logical_to_phys(scmd_channel(cmd)); | |
625 | cid = scmd_id(cmd); | |
626 | if (aac->hba_map[bus][cid].devtype == AAC_DEVTYPE_NATIVE_RAW) { | |
627 | struct fib *fib; | |
628 | struct aac_hba_tm_req *tmf; | |
629 | int status; | |
630 | u64 address; | |
631 | __le32 managed_request_id; | |
632 | ||
633 | pr_err("%s: Host adapter abort request (%d,%d,%d,%d)\n", | |
634 | AAC_DRIVERNAME, | |
635 | host->host_no, sdev_channel(dev), sdev_id(dev), (int)dev->lun); | |
636 | ||
637 | found = 0; | |
638 | for (count = 0; count < (host->can_queue + AAC_NUM_MGT_FIB); ++count) { | |
639 | fib = &aac->fibs[count]; | |
640 | if (*(u8 *)fib->hw_fib_va != 0 && | |
641 | (fib->flags & FIB_CONTEXT_FLAG_NATIVE_HBA) && | |
642 | (fib->callback_data == cmd)) { | |
643 | found = 1; | |
644 | managed_request_id = ((struct aac_hba_cmd_req *) | |
645 | fib->hw_fib_va)->request_id; | |
646 | break; | |
647 | } | |
648 | } | |
649 | if (!found) | |
650 | return ret; | |
651 | ||
652 | /* start a HBA_TMF_ABORT_TASK TMF request */ | |
653 | fib = aac_fib_alloc(aac); | |
654 | if (!fib) | |
655 | return ret; | |
656 | ||
657 | tmf = (struct aac_hba_tm_req *)fib->hw_fib_va; | |
658 | memset(tmf, 0, sizeof(*tmf)); | |
659 | tmf->tmf = HBA_TMF_ABORT_TASK; | |
660 | tmf->it_nexus = aac->hba_map[bus][cid].rmw_nexus; | |
661 | tmf->lun[1] = cmd->device->lun; | |
662 | ||
663 | address = (u64)fib->hw_error_pa; | |
664 | tmf->error_ptr_hi = cpu_to_le32((u32)(address >> 32)); | |
665 | tmf->error_ptr_lo = cpu_to_le32((u32)(address & 0xffffffff)); | |
666 | tmf->error_length = cpu_to_le32(FW_ERROR_BUFFER_SIZE); | |
667 | ||
668 | fib->hbacmd_size = sizeof(*tmf); | |
669 | cmd->SCp.sent_command = 0; | |
670 | ||
671 | status = aac_hba_send(HBA_IU_TYPE_SCSI_TM_REQ, fib, | |
672 | (fib_callback) aac_hba_callback, | |
673 | (void *) cmd); | |
674 | ||
675 | /* Wait up to 2 minutes for completion */ | |
676 | for (count = 0; count < 120; ++count) { | |
677 | if (cmd->SCp.sent_command) { | |
678 | ret = SUCCESS; | |
679 | break; | |
680 | } | |
681 | msleep(1000); | |
682 | } | |
683 | ||
684 | if (ret != SUCCESS) | |
685 | pr_err("%s: Host adapter abort request timed out\n", | |
686 | AAC_DRIVERNAME); | |
687 | } else { | |
688 | pr_err( | |
689 | "%s: Host adapter abort request.\n" | |
690 | "%s: Outstanding commands on (%d,%d,%d,%d):\n", | |
691 | AAC_DRIVERNAME, AAC_DRIVERNAME, | |
692 | host->host_no, sdev_channel(dev), sdev_id(dev), | |
693 | (int)dev->lun); | |
694 | switch (cmd->cmnd[0]) { | |
695 | case SERVICE_ACTION_IN_16: | |
696 | if (!(aac->raw_io_interface) || | |
697 | !(aac->raw_io_64) || | |
698 | ((cmd->cmnd[1] & 0x1f) != SAI_READ_CAPACITY_16)) | |
699 | break; | |
700 | case INQUIRY: | |
701 | case READ_CAPACITY: | |
702 | /* | |
703 | * Mark associated FIB to not complete, | |
704 | * eh handler does this | |
705 | */ | |
706 | for (count = 0; | |
707 | count < (host->can_queue + AAC_NUM_MGT_FIB); | |
708 | ++count) { | |
709 | struct fib *fib = &aac->fibs[count]; | |
710 | ||
711 | if (fib->hw_fib_va->header.XferState && | |
712 | (fib->flags & FIB_CONTEXT_FLAG) && | |
713 | (fib->callback_data == cmd)) { | |
714 | fib->flags |= | |
715 | FIB_CONTEXT_FLAG_TIMED_OUT; | |
716 | cmd->SCp.phase = | |
717 | AAC_OWNER_ERROR_HANDLER; | |
718 | ret = SUCCESS; | |
719 | } | |
720 | } | |
721 | break; | |
722 | case TEST_UNIT_READY: | |
723 | /* | |
724 | * Mark associated FIB to not complete, | |
725 | * eh handler does this | |
726 | */ | |
727 | for (count = 0; | |
728 | count < (host->can_queue + AAC_NUM_MGT_FIB); | |
729 | ++count) { | |
730 | struct scsi_cmnd *command; | |
731 | struct fib *fib = &aac->fibs[count]; | |
732 | ||
733 | command = fib->callback_data; | |
734 | ||
735 | if ((fib->hw_fib_va->header.XferState & | |
736 | cpu_to_le32 | |
737 | (Async | NoResponseExpected)) && | |
738 | (fib->flags & FIB_CONTEXT_FLAG) && | |
739 | ((command)) && | |
740 | (command->device == cmd->device)) { | |
741 | fib->flags |= | |
742 | FIB_CONTEXT_FLAG_TIMED_OUT; | |
743 | command->SCp.phase = | |
744 | AAC_OWNER_ERROR_HANDLER; | |
745 | if (command == cmd) | |
746 | ret = SUCCESS; | |
747 | } | |
748 | } | |
749 | break; | |
750 | } | |
751 | } | |
752 | return ret; | |
753 | } | |
754 | ||
755 | /* | |
756 | * aac_eh_reset - Reset command handling | |
757 | * @scsi_cmd: SCSI command block causing the reset | |
758 | * | |
759 | */ | |
760 | static int aac_eh_reset(struct scsi_cmnd* cmd) | |
761 | { | |
762 | struct scsi_device * dev = cmd->device; | |
763 | struct Scsi_Host * host = dev->host; | |
764 | struct aac_dev * aac = (struct aac_dev *)host->hostdata; | |
765 | int count; | |
766 | u32 bus, cid; | |
767 | int ret = FAILED; | |
768 | ||
769 | bus = aac_logical_to_phys(scmd_channel(cmd)); | |
770 | cid = scmd_id(cmd); | |
771 | if (bus < AAC_MAX_BUSES && cid < AAC_MAX_TARGETS && | |
772 | aac->hba_map[bus][cid].devtype == AAC_DEVTYPE_NATIVE_RAW) { | |
773 | struct fib *fib; | |
774 | int status; | |
775 | u64 address; | |
776 | u8 command; | |
777 | ||
778 | pr_err("%s: Host adapter reset request. SCSI hang ?\n", | |
779 | AAC_DRIVERNAME); | |
780 | ||
781 | fib = aac_fib_alloc(aac); | |
782 | if (!fib) | |
783 | return ret; | |
784 | ||
785 | ||
786 | if (aac->hba_map[bus][cid].reset_state == 0) { | |
787 | struct aac_hba_tm_req *tmf; | |
788 | ||
789 | /* start a HBA_TMF_LUN_RESET TMF request */ | |
790 | tmf = (struct aac_hba_tm_req *)fib->hw_fib_va; | |
791 | memset(tmf, 0, sizeof(*tmf)); | |
792 | tmf->tmf = HBA_TMF_LUN_RESET; | |
793 | tmf->it_nexus = aac->hba_map[bus][cid].rmw_nexus; | |
794 | tmf->lun[1] = cmd->device->lun; | |
795 | ||
796 | address = (u64)fib->hw_error_pa; | |
797 | tmf->error_ptr_hi = cpu_to_le32 | |
798 | ((u32)(address >> 32)); | |
799 | tmf->error_ptr_lo = cpu_to_le32 | |
800 | ((u32)(address & 0xffffffff)); | |
801 | tmf->error_length = cpu_to_le32(FW_ERROR_BUFFER_SIZE); | |
802 | fib->hbacmd_size = sizeof(*tmf); | |
803 | ||
804 | command = HBA_IU_TYPE_SCSI_TM_REQ; | |
805 | aac->hba_map[bus][cid].reset_state++; | |
806 | } else if (aac->hba_map[bus][cid].reset_state >= 1) { | |
807 | struct aac_hba_reset_req *rst; | |
808 | ||
809 | /* already tried, start a hard reset now */ | |
810 | rst = (struct aac_hba_reset_req *)fib->hw_fib_va; | |
811 | memset(rst, 0, sizeof(*rst)); | |
812 | /* reset_type is already zero... */ | |
813 | rst->it_nexus = aac->hba_map[bus][cid].rmw_nexus; | |
814 | ||
815 | address = (u64)fib->hw_error_pa; | |
816 | rst->error_ptr_hi = cpu_to_le32((u32)(address >> 32)); | |
817 | rst->error_ptr_lo = cpu_to_le32 | |
818 | ((u32)(address & 0xffffffff)); | |
819 | rst->error_length = cpu_to_le32(FW_ERROR_BUFFER_SIZE); | |
820 | fib->hbacmd_size = sizeof(*rst); | |
821 | ||
822 | command = HBA_IU_TYPE_SATA_REQ; | |
823 | aac->hba_map[bus][cid].reset_state = 0; | |
824 | } | |
825 | cmd->SCp.sent_command = 0; | |
826 | ||
827 | status = aac_hba_send(command, fib, | |
828 | (fib_callback) aac_hba_callback, | |
829 | (void *) cmd); | |
830 | ||
831 | /* Wait up to 2 minutes for completion */ | |
832 | for (count = 0; count < 120; ++count) { | |
833 | if (cmd->SCp.sent_command) { | |
834 | ret = SUCCESS; | |
835 | break; | |
836 | } | |
837 | msleep(1000); | |
838 | } | |
839 | ||
840 | if (ret != SUCCESS) | |
841 | pr_err("%s: Host adapter reset request timed out\n", | |
842 | AAC_DRIVERNAME); | |
843 | } else { | |
844 | struct scsi_cmnd *command; | |
845 | unsigned long flags; | |
846 | ||
847 | /* Mark the assoc. FIB to not complete, eh handler does this */ | |
848 | for (count = 0; | |
849 | count < (host->can_queue + AAC_NUM_MGT_FIB); | |
850 | ++count) { | |
851 | struct fib *fib = &aac->fibs[count]; | |
852 | ||
853 | if (fib->hw_fib_va->header.XferState && | |
854 | (fib->flags & FIB_CONTEXT_FLAG) && | |
855 | (fib->callback_data == cmd)) { | |
856 | fib->flags |= FIB_CONTEXT_FLAG_TIMED_OUT; | |
857 | cmd->SCp.phase = AAC_OWNER_ERROR_HANDLER; | |
858 | } | |
859 | } | |
860 | ||
861 | pr_err("%s: Host adapter reset request. SCSI hang ?\n", | |
862 | AAC_DRIVERNAME); | |
863 | ||
864 | count = aac_check_health(aac); | |
865 | if (count) | |
866 | return count; | |
867 | /* | |
868 | * Wait for all commands to complete to this specific | |
869 | * target (block maximum 60 seconds). | |
870 | */ | |
871 | for (count = 60; count; --count) { | |
872 | int active = aac->in_reset; | |
873 | ||
874 | if (active == 0) | |
875 | __shost_for_each_device(dev, host) { | |
876 | spin_lock_irqsave(&dev->list_lock, flags); | |
877 | list_for_each_entry(command, &dev->cmd_list, | |
878 | list) { | |
879 | if ((command != cmd) && | |
880 | (command->SCp.phase == | |
881 | AAC_OWNER_FIRMWARE)) { | |
882 | active++; | |
883 | break; | |
884 | } | |
885 | } | |
886 | spin_unlock_irqrestore(&dev->list_lock, flags); | |
887 | if (active) | |
888 | break; | |
889 | ||
890 | } | |
891 | /* | |
892 | * We can exit If all the commands are complete | |
893 | */ | |
894 | if (active == 0) | |
895 | return SUCCESS; | |
896 | ssleep(1); | |
897 | } | |
898 | pr_err("%s: SCSI bus appears hung\n", AAC_DRIVERNAME); | |
899 | ||
900 | /* | |
901 | * This adapter needs a blind reset, only do so for | |
902 | * Adapters that support a register, instead of a commanded, | |
903 | * reset. | |
904 | */ | |
905 | if (((aac->supplement_adapter_info.supported_options2 & | |
906 | AAC_OPTION_MU_RESET) || | |
907 | (aac->supplement_adapter_info.supported_options2 & | |
908 | AAC_OPTION_DOORBELL_RESET)) && | |
909 | aac_check_reset && | |
910 | ((aac_check_reset != 1) || | |
911 | !(aac->supplement_adapter_info.supported_options2 & | |
912 | AAC_OPTION_IGNORE_RESET))) { | |
913 | /* Bypass wait for command quiesce */ | |
914 | aac_reset_adapter(aac, 2, IOP_HWSOFT_RESET); | |
915 | } | |
916 | ret = SUCCESS; | |
917 | } | |
918 | /* | |
919 | * Cause an immediate retry of the command with a ten second delay | |
920 | * after successful tur | |
921 | */ | |
922 | return ret; | |
923 | } | |
924 | ||
925 | /** | |
926 | * aac_cfg_open - open a configuration file | |
927 | * @inode: inode being opened | |
928 | * @file: file handle attached | |
929 | * | |
930 | * Called when the configuration device is opened. Does the needed | |
931 | * set up on the handle and then returns | |
932 | * | |
933 | * Bugs: This needs extending to check a given adapter is present | |
934 | * so we can support hot plugging, and to ref count adapters. | |
935 | */ | |
936 | ||
937 | static int aac_cfg_open(struct inode *inode, struct file *file) | |
938 | { | |
939 | struct aac_dev *aac; | |
940 | unsigned minor_number = iminor(inode); | |
941 | int err = -ENODEV; | |
942 | ||
943 | mutex_lock(&aac_mutex); /* BKL pushdown: nothing else protects this list */ | |
944 | list_for_each_entry(aac, &aac_devices, entry) { | |
945 | if (aac->id == minor_number) { | |
946 | file->private_data = aac; | |
947 | err = 0; | |
948 | break; | |
949 | } | |
950 | } | |
951 | mutex_unlock(&aac_mutex); | |
952 | ||
953 | return err; | |
954 | } | |
955 | ||
956 | /** | |
957 | * aac_cfg_ioctl - AAC configuration request | |
958 | * @inode: inode of device | |
959 | * @file: file handle | |
960 | * @cmd: ioctl command code | |
961 | * @arg: argument | |
962 | * | |
963 | * Handles a configuration ioctl. Currently this involves wrapping it | |
964 | * up and feeding it into the nasty windowsalike glue layer. | |
965 | * | |
966 | * Bugs: Needs locking against parallel ioctls lower down | |
967 | * Bugs: Needs to handle hot plugging | |
968 | */ | |
969 | ||
970 | static long aac_cfg_ioctl(struct file *file, | |
971 | unsigned int cmd, unsigned long arg) | |
972 | { | |
973 | struct aac_dev *aac = (struct aac_dev *)file->private_data; | |
974 | ||
975 | if (!capable(CAP_SYS_RAWIO)) | |
976 | return -EPERM; | |
977 | ||
978 | return aac_do_ioctl(aac, cmd, (void __user *)arg); | |
979 | } | |
980 | ||
981 | #ifdef CONFIG_COMPAT | |
982 | static long aac_compat_do_ioctl(struct aac_dev *dev, unsigned cmd, unsigned long arg) | |
983 | { | |
984 | long ret; | |
985 | switch (cmd) { | |
986 | case FSACTL_MINIPORT_REV_CHECK: | |
987 | case FSACTL_SENDFIB: | |
988 | case FSACTL_OPEN_GET_ADAPTER_FIB: | |
989 | case FSACTL_CLOSE_GET_ADAPTER_FIB: | |
990 | case FSACTL_SEND_RAW_SRB: | |
991 | case FSACTL_GET_PCI_INFO: | |
992 | case FSACTL_QUERY_DISK: | |
993 | case FSACTL_DELETE_DISK: | |
994 | case FSACTL_FORCE_DELETE_DISK: | |
995 | case FSACTL_GET_CONTAINERS: | |
996 | case FSACTL_SEND_LARGE_FIB: | |
997 | ret = aac_do_ioctl(dev, cmd, (void __user *)arg); | |
998 | break; | |
999 | ||
1000 | case FSACTL_GET_NEXT_ADAPTER_FIB: { | |
1001 | struct fib_ioctl __user *f; | |
1002 | ||
1003 | f = compat_alloc_user_space(sizeof(*f)); | |
1004 | ret = 0; | |
1005 | if (clear_user(f, sizeof(*f))) | |
1006 | ret = -EFAULT; | |
1007 | if (copy_in_user(f, (void __user *)arg, sizeof(struct fib_ioctl) - sizeof(u32))) | |
1008 | ret = -EFAULT; | |
1009 | if (!ret) | |
1010 | ret = aac_do_ioctl(dev, cmd, f); | |
1011 | break; | |
1012 | } | |
1013 | ||
1014 | default: | |
1015 | ret = -ENOIOCTLCMD; | |
1016 | break; | |
1017 | } | |
1018 | return ret; | |
1019 | } | |
1020 | ||
1021 | static int aac_compat_ioctl(struct scsi_device *sdev, int cmd, void __user *arg) | |
1022 | { | |
1023 | struct aac_dev *dev = (struct aac_dev *)sdev->host->hostdata; | |
1024 | if (!capable(CAP_SYS_RAWIO)) | |
1025 | return -EPERM; | |
1026 | return aac_compat_do_ioctl(dev, cmd, (unsigned long)arg); | |
1027 | } | |
1028 | ||
1029 | static long aac_compat_cfg_ioctl(struct file *file, unsigned cmd, unsigned long arg) | |
1030 | { | |
1031 | if (!capable(CAP_SYS_RAWIO)) | |
1032 | return -EPERM; | |
1033 | return aac_compat_do_ioctl(file->private_data, cmd, arg); | |
1034 | } | |
1035 | #endif | |
1036 | ||
1037 | static ssize_t aac_show_model(struct device *device, | |
1038 | struct device_attribute *attr, char *buf) | |
1039 | { | |
1040 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1041 | int len; | |
1042 | ||
1043 | if (dev->supplement_adapter_info.adapter_type_text[0]) { | |
1044 | char *cp = dev->supplement_adapter_info.adapter_type_text; | |
1045 | while (*cp && *cp != ' ') | |
1046 | ++cp; | |
1047 | while (*cp == ' ') | |
1048 | ++cp; | |
1049 | len = snprintf(buf, PAGE_SIZE, "%s\n", cp); | |
1050 | } else | |
1051 | len = snprintf(buf, PAGE_SIZE, "%s\n", | |
1052 | aac_drivers[dev->cardtype].model); | |
1053 | return len; | |
1054 | } | |
1055 | ||
1056 | static ssize_t aac_show_vendor(struct device *device, | |
1057 | struct device_attribute *attr, char *buf) | |
1058 | { | |
1059 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1060 | struct aac_supplement_adapter_info *sup_adap_info; | |
1061 | int len; | |
1062 | ||
1063 | sup_adap_info = &dev->supplement_adapter_info; | |
1064 | if (sup_adap_info->adapter_type_text[0]) { | |
1065 | char *cp = sup_adap_info->adapter_type_text; | |
1066 | while (*cp && *cp != ' ') | |
1067 | ++cp; | |
1068 | len = snprintf(buf, PAGE_SIZE, "%.*s\n", | |
1069 | (int)(cp - (char *)sup_adap_info->adapter_type_text), | |
1070 | sup_adap_info->adapter_type_text); | |
1071 | } else | |
1072 | len = snprintf(buf, PAGE_SIZE, "%s\n", | |
1073 | aac_drivers[dev->cardtype].vname); | |
1074 | return len; | |
1075 | } | |
1076 | ||
1077 | static ssize_t aac_show_flags(struct device *cdev, | |
1078 | struct device_attribute *attr, char *buf) | |
1079 | { | |
1080 | int len = 0; | |
1081 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(cdev)->hostdata; | |
1082 | ||
1083 | if (nblank(dprintk(x))) | |
1084 | len = snprintf(buf, PAGE_SIZE, "dprintk\n"); | |
1085 | #ifdef AAC_DETAILED_STATUS_INFO | |
1086 | len += snprintf(buf + len, PAGE_SIZE - len, | |
1087 | "AAC_DETAILED_STATUS_INFO\n"); | |
1088 | #endif | |
1089 | if (dev->raw_io_interface && dev->raw_io_64) | |
1090 | len += snprintf(buf + len, PAGE_SIZE - len, | |
1091 | "SAI_READ_CAPACITY_16\n"); | |
1092 | if (dev->jbod) | |
1093 | len += snprintf(buf + len, PAGE_SIZE - len, "SUPPORTED_JBOD\n"); | |
1094 | if (dev->supplement_adapter_info.supported_options2 & | |
1095 | AAC_OPTION_POWER_MANAGEMENT) | |
1096 | len += snprintf(buf + len, PAGE_SIZE - len, | |
1097 | "SUPPORTED_POWER_MANAGEMENT\n"); | |
1098 | if (dev->msi) | |
1099 | len += snprintf(buf + len, PAGE_SIZE - len, "PCI_HAS_MSI\n"); | |
1100 | return len; | |
1101 | } | |
1102 | ||
1103 | static ssize_t aac_show_kernel_version(struct device *device, | |
1104 | struct device_attribute *attr, | |
1105 | char *buf) | |
1106 | { | |
1107 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1108 | int len, tmp; | |
1109 | ||
1110 | tmp = le32_to_cpu(dev->adapter_info.kernelrev); | |
1111 | len = snprintf(buf, PAGE_SIZE, "%d.%d-%d[%d]\n", | |
1112 | tmp >> 24, (tmp >> 16) & 0xff, tmp & 0xff, | |
1113 | le32_to_cpu(dev->adapter_info.kernelbuild)); | |
1114 | return len; | |
1115 | } | |
1116 | ||
1117 | static ssize_t aac_show_monitor_version(struct device *device, | |
1118 | struct device_attribute *attr, | |
1119 | char *buf) | |
1120 | { | |
1121 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1122 | int len, tmp; | |
1123 | ||
1124 | tmp = le32_to_cpu(dev->adapter_info.monitorrev); | |
1125 | len = snprintf(buf, PAGE_SIZE, "%d.%d-%d[%d]\n", | |
1126 | tmp >> 24, (tmp >> 16) & 0xff, tmp & 0xff, | |
1127 | le32_to_cpu(dev->adapter_info.monitorbuild)); | |
1128 | return len; | |
1129 | } | |
1130 | ||
1131 | static ssize_t aac_show_bios_version(struct device *device, | |
1132 | struct device_attribute *attr, | |
1133 | char *buf) | |
1134 | { | |
1135 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1136 | int len, tmp; | |
1137 | ||
1138 | tmp = le32_to_cpu(dev->adapter_info.biosrev); | |
1139 | len = snprintf(buf, PAGE_SIZE, "%d.%d-%d[%d]\n", | |
1140 | tmp >> 24, (tmp >> 16) & 0xff, tmp & 0xff, | |
1141 | le32_to_cpu(dev->adapter_info.biosbuild)); | |
1142 | return len; | |
1143 | } | |
1144 | ||
1145 | static ssize_t aac_show_driver_version(struct device *device, | |
1146 | struct device_attribute *attr, | |
1147 | char *buf) | |
1148 | { | |
1149 | return snprintf(buf, PAGE_SIZE, "%s\n", aac_driver_version); | |
1150 | } | |
1151 | ||
1152 | static ssize_t aac_show_serial_number(struct device *device, | |
1153 | struct device_attribute *attr, char *buf) | |
1154 | { | |
1155 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1156 | int len = 0; | |
1157 | ||
1158 | if (le32_to_cpu(dev->adapter_info.serial[0]) != 0xBAD0) | |
1159 | len = snprintf(buf, 16, "%06X\n", | |
1160 | le32_to_cpu(dev->adapter_info.serial[0])); | |
1161 | if (len && | |
1162 | !memcmp(&dev->supplement_adapter_info.mfg_pcba_serial_no[ | |
1163 | sizeof(dev->supplement_adapter_info.mfg_pcba_serial_no)-len], | |
1164 | buf, len-1)) | |
1165 | len = snprintf(buf, 16, "%.*s\n", | |
1166 | (int)sizeof(dev->supplement_adapter_info.mfg_pcba_serial_no), | |
1167 | dev->supplement_adapter_info.mfg_pcba_serial_no); | |
1168 | ||
1169 | return min(len, 16); | |
1170 | } | |
1171 | ||
1172 | static ssize_t aac_show_max_channel(struct device *device, | |
1173 | struct device_attribute *attr, char *buf) | |
1174 | { | |
1175 | return snprintf(buf, PAGE_SIZE, "%d\n", | |
1176 | class_to_shost(device)->max_channel); | |
1177 | } | |
1178 | ||
1179 | static ssize_t aac_show_max_id(struct device *device, | |
1180 | struct device_attribute *attr, char *buf) | |
1181 | { | |
1182 | return snprintf(buf, PAGE_SIZE, "%d\n", | |
1183 | class_to_shost(device)->max_id); | |
1184 | } | |
1185 | ||
1186 | static ssize_t aac_store_reset_adapter(struct device *device, | |
1187 | struct device_attribute *attr, | |
1188 | const char *buf, size_t count) | |
1189 | { | |
1190 | int retval = -EACCES; | |
1191 | int bled = 0; | |
1192 | struct aac_dev *aac; | |
1193 | ||
1194 | ||
1195 | if (!capable(CAP_SYS_ADMIN)) | |
1196 | return retval; | |
1197 | ||
1198 | aac = (struct aac_dev *)class_to_shost(device)->hostdata; | |
1199 | bled = buf[0] == '!' ? 1:0; | |
1200 | retval = aac_reset_adapter(aac, bled, IOP_HWSOFT_RESET); | |
1201 | if (retval >= 0) | |
1202 | retval = count; | |
1203 | return retval; | |
1204 | } | |
1205 | ||
1206 | static ssize_t aac_show_reset_adapter(struct device *device, | |
1207 | struct device_attribute *attr, | |
1208 | char *buf) | |
1209 | { | |
1210 | struct aac_dev *dev = (struct aac_dev*)class_to_shost(device)->hostdata; | |
1211 | int len, tmp; | |
1212 | ||
1213 | tmp = aac_adapter_check_health(dev); | |
1214 | if ((tmp == 0) && dev->in_reset) | |
1215 | tmp = -EBUSY; | |
1216 | len = snprintf(buf, PAGE_SIZE, "0x%x\n", tmp); | |
1217 | return len; | |
1218 | } | |
1219 | ||
1220 | static struct device_attribute aac_model = { | |
1221 | .attr = { | |
1222 | .name = "model", | |
1223 | .mode = S_IRUGO, | |
1224 | }, | |
1225 | .show = aac_show_model, | |
1226 | }; | |
1227 | static struct device_attribute aac_vendor = { | |
1228 | .attr = { | |
1229 | .name = "vendor", | |
1230 | .mode = S_IRUGO, | |
1231 | }, | |
1232 | .show = aac_show_vendor, | |
1233 | }; | |
1234 | static struct device_attribute aac_flags = { | |
1235 | .attr = { | |
1236 | .name = "flags", | |
1237 | .mode = S_IRUGO, | |
1238 | }, | |
1239 | .show = aac_show_flags, | |
1240 | }; | |
1241 | static struct device_attribute aac_kernel_version = { | |
1242 | .attr = { | |
1243 | .name = "hba_kernel_version", | |
1244 | .mode = S_IRUGO, | |
1245 | }, | |
1246 | .show = aac_show_kernel_version, | |
1247 | }; | |
1248 | static struct device_attribute aac_monitor_version = { | |
1249 | .attr = { | |
1250 | .name = "hba_monitor_version", | |
1251 | .mode = S_IRUGO, | |
1252 | }, | |
1253 | .show = aac_show_monitor_version, | |
1254 | }; | |
1255 | static struct device_attribute aac_bios_version = { | |
1256 | .attr = { | |
1257 | .name = "hba_bios_version", | |
1258 | .mode = S_IRUGO, | |
1259 | }, | |
1260 | .show = aac_show_bios_version, | |
1261 | }; | |
1262 | static struct device_attribute aac_lld_version = { | |
1263 | .attr = { | |
1264 | .name = "driver_version", | |
1265 | .mode = 0444, | |
1266 | }, | |
1267 | .show = aac_show_driver_version, | |
1268 | }; | |
1269 | static struct device_attribute aac_serial_number = { | |
1270 | .attr = { | |
1271 | .name = "serial_number", | |
1272 | .mode = S_IRUGO, | |
1273 | }, | |
1274 | .show = aac_show_serial_number, | |
1275 | }; | |
1276 | static struct device_attribute aac_max_channel = { | |
1277 | .attr = { | |
1278 | .name = "max_channel", | |
1279 | .mode = S_IRUGO, | |
1280 | }, | |
1281 | .show = aac_show_max_channel, | |
1282 | }; | |
1283 | static struct device_attribute aac_max_id = { | |
1284 | .attr = { | |
1285 | .name = "max_id", | |
1286 | .mode = S_IRUGO, | |
1287 | }, | |
1288 | .show = aac_show_max_id, | |
1289 | }; | |
1290 | static struct device_attribute aac_reset = { | |
1291 | .attr = { | |
1292 | .name = "reset_host", | |
1293 | .mode = S_IWUSR|S_IRUGO, | |
1294 | }, | |
1295 | .store = aac_store_reset_adapter, | |
1296 | .show = aac_show_reset_adapter, | |
1297 | }; | |
1298 | ||
1299 | static struct device_attribute *aac_attrs[] = { | |
1300 | &aac_model, | |
1301 | &aac_vendor, | |
1302 | &aac_flags, | |
1303 | &aac_kernel_version, | |
1304 | &aac_monitor_version, | |
1305 | &aac_bios_version, | |
1306 | &aac_lld_version, | |
1307 | &aac_serial_number, | |
1308 | &aac_max_channel, | |
1309 | &aac_max_id, | |
1310 | &aac_reset, | |
1311 | NULL | |
1312 | }; | |
1313 | ||
1314 | ssize_t aac_get_serial_number(struct device *device, char *buf) | |
1315 | { | |
1316 | return aac_show_serial_number(device, &aac_serial_number, buf); | |
1317 | } | |
1318 | ||
1319 | static const struct file_operations aac_cfg_fops = { | |
1320 | .owner = THIS_MODULE, | |
1321 | .unlocked_ioctl = aac_cfg_ioctl, | |
1322 | #ifdef CONFIG_COMPAT | |
1323 | .compat_ioctl = aac_compat_cfg_ioctl, | |
1324 | #endif | |
1325 | .open = aac_cfg_open, | |
1326 | .llseek = noop_llseek, | |
1327 | }; | |
1328 | ||
1329 | static struct scsi_host_template aac_driver_template = { | |
1330 | .module = THIS_MODULE, | |
1331 | .name = "AAC", | |
1332 | .proc_name = AAC_DRIVERNAME, | |
1333 | .info = aac_info, | |
1334 | .ioctl = aac_ioctl, | |
1335 | #ifdef CONFIG_COMPAT | |
1336 | .compat_ioctl = aac_compat_ioctl, | |
1337 | #endif | |
1338 | .queuecommand = aac_queuecommand, | |
1339 | .bios_param = aac_biosparm, | |
1340 | .shost_attrs = aac_attrs, | |
1341 | .slave_configure = aac_slave_configure, | |
1342 | .change_queue_depth = aac_change_queue_depth, | |
1343 | .sdev_attrs = aac_dev_attrs, | |
1344 | .eh_abort_handler = aac_eh_abort, | |
1345 | .eh_host_reset_handler = aac_eh_reset, | |
1346 | .can_queue = AAC_NUM_IO_FIB, | |
1347 | .this_id = MAXIMUM_NUM_CONTAINERS, | |
1348 | .sg_tablesize = 16, | |
1349 | .max_sectors = 128, | |
1350 | #if (AAC_NUM_IO_FIB > 256) | |
1351 | .cmd_per_lun = 256, | |
1352 | #else | |
1353 | .cmd_per_lun = AAC_NUM_IO_FIB, | |
1354 | #endif | |
1355 | .use_clustering = ENABLE_CLUSTERING, | |
1356 | .emulated = 1, | |
1357 | .no_write_same = 1, | |
1358 | }; | |
1359 | ||
1360 | static void __aac_shutdown(struct aac_dev * aac) | |
1361 | { | |
1362 | int i; | |
1363 | ||
1364 | aac->adapter_shutdown = 1; | |
1365 | aac_send_shutdown(aac); | |
1366 | ||
1367 | if (aac->aif_thread) { | |
1368 | int i; | |
1369 | /* Clear out events first */ | |
1370 | for (i = 0; i < (aac->scsi_host_ptr->can_queue + AAC_NUM_MGT_FIB); i++) { | |
1371 | struct fib *fib = &aac->fibs[i]; | |
1372 | if (!(fib->hw_fib_va->header.XferState & cpu_to_le32(NoResponseExpected | Async)) && | |
1373 | (fib->hw_fib_va->header.XferState & cpu_to_le32(ResponseExpected))) | |
1374 | up(&fib->event_wait); | |
1375 | } | |
1376 | kthread_stop(aac->thread); | |
1377 | } | |
1378 | aac_adapter_disable_int(aac); | |
1379 | if (aac->pdev->device == PMC_DEVICE_S6 || | |
1380 | aac->pdev->device == PMC_DEVICE_S7 || | |
1381 | aac->pdev->device == PMC_DEVICE_S8 || | |
1382 | aac->pdev->device == PMC_DEVICE_S9) { | |
1383 | if (aac->max_msix > 1) { | |
1384 | for (i = 0; i < aac->max_msix; i++) { | |
1385 | free_irq(pci_irq_vector(aac->pdev, i), | |
1386 | &(aac->aac_msix[i])); | |
1387 | } | |
1388 | } else { | |
1389 | free_irq(aac->pdev->irq, | |
1390 | &(aac->aac_msix[0])); | |
1391 | } | |
1392 | } else { | |
1393 | free_irq(aac->pdev->irq, aac); | |
1394 | } | |
1395 | if (aac->msi) | |
1396 | pci_disable_msi(aac->pdev); | |
1397 | else if (aac->max_msix > 1) | |
1398 | pci_disable_msix(aac->pdev); | |
1399 | } | |
1400 | static void aac_init_char(void) | |
1401 | { | |
1402 | aac_cfg_major = register_chrdev(0, "aac", &aac_cfg_fops); | |
1403 | if (aac_cfg_major < 0) { | |
1404 | pr_err("aacraid: unable to register \"aac\" device.\n"); | |
1405 | } | |
1406 | } | |
1407 | ||
1408 | static int aac_probe_one(struct pci_dev *pdev, const struct pci_device_id *id) | |
1409 | { | |
1410 | unsigned index = id->driver_data; | |
1411 | struct Scsi_Host *shost; | |
1412 | struct aac_dev *aac; | |
1413 | struct list_head *insert = &aac_devices; | |
1414 | int error = -ENODEV; | |
1415 | int unique_id = 0; | |
1416 | u64 dmamask; | |
1417 | int mask_bits = 0; | |
1418 | extern int aac_sync_mode; | |
1419 | ||
1420 | /* | |
1421 | * Only series 7 needs freset. | |
1422 | */ | |
1423 | if (pdev->device == PMC_DEVICE_S7) | |
1424 | pdev->needs_freset = 1; | |
1425 | ||
1426 | list_for_each_entry(aac, &aac_devices, entry) { | |
1427 | if (aac->id > unique_id) | |
1428 | break; | |
1429 | insert = &aac->entry; | |
1430 | unique_id++; | |
1431 | } | |
1432 | ||
1433 | pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S | PCIE_LINK_STATE_L1 | | |
1434 | PCIE_LINK_STATE_CLKPM); | |
1435 | ||
1436 | error = pci_enable_device(pdev); | |
1437 | if (error) | |
1438 | goto out; | |
1439 | error = -ENODEV; | |
1440 | ||
1441 | if (!(aac_drivers[index].quirks & AAC_QUIRK_SRC)) { | |
1442 | error = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); | |
1443 | if (error) { | |
1444 | dev_err(&pdev->dev, "PCI 32 BIT dma mask set failed"); | |
1445 | goto out_disable_pdev; | |
1446 | } | |
1447 | } | |
1448 | ||
1449 | /* | |
1450 | * If the quirk31 bit is set, the adapter needs adapter | |
1451 | * to driver communication memory to be allocated below 2gig | |
1452 | */ | |
1453 | if (aac_drivers[index].quirks & AAC_QUIRK_31BIT) { | |
1454 | dmamask = DMA_BIT_MASK(31); | |
1455 | mask_bits = 31; | |
1456 | } else { | |
1457 | dmamask = DMA_BIT_MASK(32); | |
1458 | mask_bits = 32; | |
1459 | } | |
1460 | ||
1461 | error = pci_set_consistent_dma_mask(pdev, dmamask); | |
1462 | if (error) { | |
1463 | dev_err(&pdev->dev, "PCI %d B consistent dma mask set failed\n" | |
1464 | , mask_bits); | |
1465 | goto out_disable_pdev; | |
1466 | } | |
1467 | ||
1468 | pci_set_master(pdev); | |
1469 | ||
1470 | shost = scsi_host_alloc(&aac_driver_template, sizeof(struct aac_dev)); | |
1471 | if (!shost) | |
1472 | goto out_disable_pdev; | |
1473 | ||
1474 | shost->irq = pdev->irq; | |
1475 | shost->unique_id = unique_id; | |
1476 | shost->max_cmd_len = 16; | |
1477 | shost->use_cmd_list = 1; | |
1478 | ||
1479 | if (aac_cfg_major == AAC_CHARDEV_NEEDS_REINIT) | |
1480 | aac_init_char(); | |
1481 | ||
1482 | aac = (struct aac_dev *)shost->hostdata; | |
1483 | aac->base_start = pci_resource_start(pdev, 0); | |
1484 | aac->scsi_host_ptr = shost; | |
1485 | aac->pdev = pdev; | |
1486 | aac->name = aac_driver_template.name; | |
1487 | aac->id = shost->unique_id; | |
1488 | aac->cardtype = index; | |
1489 | INIT_LIST_HEAD(&aac->entry); | |
1490 | ||
1491 | aac->fibs = kzalloc(sizeof(struct fib) * (shost->can_queue + AAC_NUM_MGT_FIB), GFP_KERNEL); | |
1492 | if (!aac->fibs) | |
1493 | goto out_free_host; | |
1494 | spin_lock_init(&aac->fib_lock); | |
1495 | ||
1496 | mutex_init(&aac->ioctl_mutex); | |
1497 | /* | |
1498 | * Map in the registers from the adapter. | |
1499 | */ | |
1500 | aac->base_size = AAC_MIN_FOOTPRINT_SIZE; | |
1501 | if ((*aac_drivers[index].init)(aac)) | |
1502 | goto out_unmap; | |
1503 | ||
1504 | if (aac->sync_mode) { | |
1505 | if (aac_sync_mode) | |
1506 | printk(KERN_INFO "%s%d: Sync. mode enforced " | |
1507 | "by driver parameter. This will cause " | |
1508 | "a significant performance decrease!\n", | |
1509 | aac->name, | |
1510 | aac->id); | |
1511 | else | |
1512 | printk(KERN_INFO "%s%d: Async. mode not supported " | |
1513 | "by current driver, sync. mode enforced." | |
1514 | "\nPlease update driver to get full performance.\n", | |
1515 | aac->name, | |
1516 | aac->id); | |
1517 | } | |
1518 | ||
1519 | /* | |
1520 | * Start any kernel threads needed | |
1521 | */ | |
1522 | aac->thread = kthread_run(aac_command_thread, aac, AAC_DRIVERNAME); | |
1523 | if (IS_ERR(aac->thread)) { | |
1524 | printk(KERN_ERR "aacraid: Unable to create command thread.\n"); | |
1525 | error = PTR_ERR(aac->thread); | |
1526 | aac->thread = NULL; | |
1527 | goto out_deinit; | |
1528 | } | |
1529 | ||
1530 | aac->maximum_num_channels = aac_drivers[index].channels; | |
1531 | error = aac_get_adapter_info(aac); | |
1532 | if (error < 0) | |
1533 | goto out_deinit; | |
1534 | ||
1535 | /* | |
1536 | * Lets override negotiations and drop the maximum SG limit to 34 | |
1537 | */ | |
1538 | if ((aac_drivers[index].quirks & AAC_QUIRK_34SG) && | |
1539 | (shost->sg_tablesize > 34)) { | |
1540 | shost->sg_tablesize = 34; | |
1541 | shost->max_sectors = (shost->sg_tablesize * 8) + 112; | |
1542 | } | |
1543 | ||
1544 | if ((aac_drivers[index].quirks & AAC_QUIRK_17SG) && | |
1545 | (shost->sg_tablesize > 17)) { | |
1546 | shost->sg_tablesize = 17; | |
1547 | shost->max_sectors = (shost->sg_tablesize * 8) + 112; | |
1548 | } | |
1549 | ||
1550 | error = pci_set_dma_max_seg_size(pdev, | |
1551 | (aac->adapter_info.options & AAC_OPT_NEW_COMM) ? | |
1552 | (shost->max_sectors << 9) : 65536); | |
1553 | if (error) | |
1554 | goto out_deinit; | |
1555 | ||
1556 | /* | |
1557 | * Firmware printf works only with older firmware. | |
1558 | */ | |
1559 | if (aac_drivers[index].quirks & AAC_QUIRK_34SG) | |
1560 | aac->printf_enabled = 1; | |
1561 | else | |
1562 | aac->printf_enabled = 0; | |
1563 | ||
1564 | /* | |
1565 | * max channel will be the physical channels plus 1 virtual channel | |
1566 | * all containers are on the virtual channel 0 (CONTAINER_CHANNEL) | |
1567 | * physical channels are address by their actual physical number+1 | |
1568 | */ | |
1569 | if (aac->nondasd_support || expose_physicals || aac->jbod) | |
1570 | shost->max_channel = aac->maximum_num_channels; | |
1571 | else | |
1572 | shost->max_channel = 0; | |
1573 | ||
1574 | aac_get_config_status(aac, 0); | |
1575 | aac_get_containers(aac); | |
1576 | list_add(&aac->entry, insert); | |
1577 | ||
1578 | shost->max_id = aac->maximum_num_containers; | |
1579 | if (shost->max_id < aac->maximum_num_physicals) | |
1580 | shost->max_id = aac->maximum_num_physicals; | |
1581 | if (shost->max_id < MAXIMUM_NUM_CONTAINERS) | |
1582 | shost->max_id = MAXIMUM_NUM_CONTAINERS; | |
1583 | else | |
1584 | shost->this_id = shost->max_id; | |
1585 | ||
1586 | if (!aac->sa_firmware && aac_drivers[index].quirks & AAC_QUIRK_SRC) | |
1587 | aac_intr_normal(aac, 0, 2, 0, NULL); | |
1588 | ||
1589 | /* | |
1590 | * dmb - we may need to move the setting of these parms somewhere else once | |
1591 | * we get a fib that can report the actual numbers | |
1592 | */ | |
1593 | shost->max_lun = AAC_MAX_LUN; | |
1594 | ||
1595 | pci_set_drvdata(pdev, shost); | |
1596 | ||
1597 | error = scsi_add_host(shost, &pdev->dev); | |
1598 | if (error) | |
1599 | goto out_deinit; | |
1600 | scsi_scan_host(shost); | |
1601 | ||
1602 | pci_enable_pcie_error_reporting(pdev); | |
1603 | pci_save_state(pdev); | |
1604 | ||
1605 | return 0; | |
1606 | ||
1607 | out_deinit: | |
1608 | __aac_shutdown(aac); | |
1609 | out_unmap: | |
1610 | aac_fib_map_free(aac); | |
1611 | if (aac->comm_addr) | |
1612 | dma_free_coherent(&aac->pdev->dev, aac->comm_size, | |
1613 | aac->comm_addr, aac->comm_phys); | |
1614 | kfree(aac->queues); | |
1615 | aac_adapter_ioremap(aac, 0); | |
1616 | kfree(aac->fibs); | |
1617 | kfree(aac->fsa_dev); | |
1618 | out_free_host: | |
1619 | scsi_host_put(shost); | |
1620 | out_disable_pdev: | |
1621 | pci_disable_device(pdev); | |
1622 | out: | |
1623 | return error; | |
1624 | } | |
1625 | ||
1626 | static void aac_release_resources(struct aac_dev *aac) | |
1627 | { | |
1628 | aac_adapter_disable_int(aac); | |
1629 | aac_free_irq(aac); | |
1630 | } | |
1631 | ||
1632 | static int aac_acquire_resources(struct aac_dev *dev) | |
1633 | { | |
1634 | unsigned long status; | |
1635 | /* | |
1636 | * First clear out all interrupts. Then enable the one's that we | |
1637 | * can handle. | |
1638 | */ | |
1639 | while (!((status = src_readl(dev, MUnit.OMR)) & KERNEL_UP_AND_RUNNING) | |
1640 | || status == 0xffffffff) | |
1641 | msleep(20); | |
1642 | ||
1643 | aac_adapter_disable_int(dev); | |
1644 | aac_adapter_enable_int(dev); | |
1645 | ||
1646 | ||
1647 | if ((dev->pdev->device == PMC_DEVICE_S7 || | |
1648 | dev->pdev->device == PMC_DEVICE_S8 || | |
1649 | dev->pdev->device == PMC_DEVICE_S9)) | |
1650 | aac_define_int_mode(dev); | |
1651 | ||
1652 | if (dev->msi_enabled) | |
1653 | aac_src_access_devreg(dev, AAC_ENABLE_MSIX); | |
1654 | ||
1655 | if (aac_acquire_irq(dev)) | |
1656 | goto error_iounmap; | |
1657 | ||
1658 | aac_adapter_enable_int(dev); | |
1659 | ||
1660 | /*max msix may change after EEH | |
1661 | * Re-assign vectors to fibs | |
1662 | */ | |
1663 | aac_fib_vector_assign(dev); | |
1664 | ||
1665 | if (!dev->sync_mode) { | |
1666 | /* After EEH recovery or suspend resume, max_msix count | |
1667 | * may change, therefore updating in init as well. | |
1668 | */ | |
1669 | dev->init->r7.no_of_msix_vectors = cpu_to_le32(dev->max_msix); | |
1670 | aac_adapter_start(dev); | |
1671 | } | |
1672 | return 0; | |
1673 | ||
1674 | error_iounmap: | |
1675 | return -1; | |
1676 | ||
1677 | } | |
1678 | ||
1679 | #if (defined(CONFIG_PM)) | |
1680 | static int aac_suspend(struct pci_dev *pdev, pm_message_t state) | |
1681 | { | |
1682 | ||
1683 | struct Scsi_Host *shost = pci_get_drvdata(pdev); | |
1684 | struct aac_dev *aac = (struct aac_dev *)shost->hostdata; | |
1685 | ||
1686 | scsi_block_requests(shost); | |
1687 | aac_send_shutdown(aac); | |
1688 | ||
1689 | aac_release_resources(aac); | |
1690 | ||
1691 | pci_set_drvdata(pdev, shost); | |
1692 | pci_save_state(pdev); | |
1693 | pci_disable_device(pdev); | |
1694 | pci_set_power_state(pdev, pci_choose_state(pdev, state)); | |
1695 | ||
1696 | return 0; | |
1697 | } | |
1698 | ||
1699 | static int aac_resume(struct pci_dev *pdev) | |
1700 | { | |
1701 | struct Scsi_Host *shost = pci_get_drvdata(pdev); | |
1702 | struct aac_dev *aac = (struct aac_dev *)shost->hostdata; | |
1703 | int r; | |
1704 | ||
1705 | pci_set_power_state(pdev, PCI_D0); | |
1706 | pci_enable_wake(pdev, PCI_D0, 0); | |
1707 | pci_restore_state(pdev); | |
1708 | r = pci_enable_device(pdev); | |
1709 | ||
1710 | if (r) | |
1711 | goto fail_device; | |
1712 | ||
1713 | pci_set_master(pdev); | |
1714 | if (aac_acquire_resources(aac)) | |
1715 | goto fail_device; | |
1716 | /* | |
1717 | * reset this flag to unblock ioctl() as it was set at | |
1718 | * aac_send_shutdown() to block ioctls from upperlayer | |
1719 | */ | |
1720 | aac->adapter_shutdown = 0; | |
1721 | scsi_unblock_requests(shost); | |
1722 | ||
1723 | return 0; | |
1724 | ||
1725 | fail_device: | |
1726 | printk(KERN_INFO "%s%d: resume failed.\n", aac->name, aac->id); | |
1727 | scsi_host_put(shost); | |
1728 | pci_disable_device(pdev); | |
1729 | return -ENODEV; | |
1730 | } | |
1731 | #endif | |
1732 | ||
1733 | static void aac_shutdown(struct pci_dev *dev) | |
1734 | { | |
1735 | struct Scsi_Host *shost = pci_get_drvdata(dev); | |
1736 | scsi_block_requests(shost); | |
1737 | __aac_shutdown((struct aac_dev *)shost->hostdata); | |
1738 | } | |
1739 | ||
1740 | static void aac_remove_one(struct pci_dev *pdev) | |
1741 | { | |
1742 | struct Scsi_Host *shost = pci_get_drvdata(pdev); | |
1743 | struct aac_dev *aac = (struct aac_dev *)shost->hostdata; | |
1744 | ||
1745 | scsi_remove_host(shost); | |
1746 | ||
1747 | __aac_shutdown(aac); | |
1748 | aac_fib_map_free(aac); | |
1749 | dma_free_coherent(&aac->pdev->dev, aac->comm_size, aac->comm_addr, | |
1750 | aac->comm_phys); | |
1751 | kfree(aac->queues); | |
1752 | ||
1753 | aac_adapter_ioremap(aac, 0); | |
1754 | ||
1755 | kfree(aac->fibs); | |
1756 | kfree(aac->fsa_dev); | |
1757 | ||
1758 | list_del(&aac->entry); | |
1759 | scsi_host_put(shost); | |
1760 | pci_disable_device(pdev); | |
1761 | if (list_empty(&aac_devices)) { | |
1762 | unregister_chrdev(aac_cfg_major, "aac"); | |
1763 | aac_cfg_major = AAC_CHARDEV_NEEDS_REINIT; | |
1764 | } | |
1765 | } | |
1766 | ||
1767 | static void aac_flush_ios(struct aac_dev *aac) | |
1768 | { | |
1769 | int i; | |
1770 | struct scsi_cmnd *cmd; | |
1771 | ||
1772 | for (i = 0; i < aac->scsi_host_ptr->can_queue; i++) { | |
1773 | cmd = (struct scsi_cmnd *)aac->fibs[i].callback_data; | |
1774 | if (cmd && (cmd->SCp.phase == AAC_OWNER_FIRMWARE)) { | |
1775 | scsi_dma_unmap(cmd); | |
1776 | ||
1777 | if (aac->handle_pci_error) | |
1778 | cmd->result = DID_NO_CONNECT << 16; | |
1779 | else | |
1780 | cmd->result = DID_RESET << 16; | |
1781 | ||
1782 | cmd->scsi_done(cmd); | |
1783 | } | |
1784 | } | |
1785 | } | |
1786 | ||
1787 | static pci_ers_result_t aac_pci_error_detected(struct pci_dev *pdev, | |
1788 | enum pci_channel_state error) | |
1789 | { | |
1790 | struct Scsi_Host *shost = pci_get_drvdata(pdev); | |
1791 | struct aac_dev *aac = shost_priv(shost); | |
1792 | ||
1793 | dev_err(&pdev->dev, "aacraid: PCI error detected %x\n", error); | |
1794 | ||
1795 | switch (error) { | |
1796 | case pci_channel_io_normal: | |
1797 | return PCI_ERS_RESULT_CAN_RECOVER; | |
1798 | case pci_channel_io_frozen: | |
1799 | aac->handle_pci_error = 1; | |
1800 | ||
1801 | scsi_block_requests(aac->scsi_host_ptr); | |
1802 | aac_flush_ios(aac); | |
1803 | aac_release_resources(aac); | |
1804 | ||
1805 | pci_disable_pcie_error_reporting(pdev); | |
1806 | aac_adapter_ioremap(aac, 0); | |
1807 | ||
1808 | return PCI_ERS_RESULT_NEED_RESET; | |
1809 | case pci_channel_io_perm_failure: | |
1810 | aac->handle_pci_error = 1; | |
1811 | ||
1812 | aac_flush_ios(aac); | |
1813 | return PCI_ERS_RESULT_DISCONNECT; | |
1814 | } | |
1815 | ||
1816 | return PCI_ERS_RESULT_NEED_RESET; | |
1817 | } | |
1818 | ||
1819 | static pci_ers_result_t aac_pci_mmio_enabled(struct pci_dev *pdev) | |
1820 | { | |
1821 | dev_err(&pdev->dev, "aacraid: PCI error - mmio enabled\n"); | |
1822 | return PCI_ERS_RESULT_NEED_RESET; | |
1823 | } | |
1824 | ||
1825 | static pci_ers_result_t aac_pci_slot_reset(struct pci_dev *pdev) | |
1826 | { | |
1827 | dev_err(&pdev->dev, "aacraid: PCI error - slot reset\n"); | |
1828 | pci_restore_state(pdev); | |
1829 | if (pci_enable_device(pdev)) { | |
1830 | dev_warn(&pdev->dev, | |
1831 | "aacraid: failed to enable slave\n"); | |
1832 | goto fail_device; | |
1833 | } | |
1834 | ||
1835 | pci_set_master(pdev); | |
1836 | ||
1837 | if (pci_enable_device_mem(pdev)) { | |
1838 | dev_err(&pdev->dev, "pci_enable_device_mem failed\n"); | |
1839 | goto fail_device; | |
1840 | } | |
1841 | ||
1842 | return PCI_ERS_RESULT_RECOVERED; | |
1843 | ||
1844 | fail_device: | |
1845 | dev_err(&pdev->dev, "aacraid: PCI error - slot reset failed\n"); | |
1846 | return PCI_ERS_RESULT_DISCONNECT; | |
1847 | } | |
1848 | ||
1849 | ||
1850 | static void aac_pci_resume(struct pci_dev *pdev) | |
1851 | { | |
1852 | struct Scsi_Host *shost = pci_get_drvdata(pdev); | |
1853 | struct scsi_device *sdev = NULL; | |
1854 | struct aac_dev *aac = (struct aac_dev *)shost_priv(shost); | |
1855 | ||
1856 | pci_cleanup_aer_uncorrect_error_status(pdev); | |
1857 | ||
1858 | if (aac_adapter_ioremap(aac, aac->base_size)) { | |
1859 | ||
1860 | dev_err(&pdev->dev, "aacraid: ioremap failed\n"); | |
1861 | /* remap failed, go back ... */ | |
1862 | aac->comm_interface = AAC_COMM_PRODUCER; | |
1863 | if (aac_adapter_ioremap(aac, AAC_MIN_FOOTPRINT_SIZE)) { | |
1864 | dev_warn(&pdev->dev, | |
1865 | "aacraid: unable to map adapter.\n"); | |
1866 | ||
1867 | return; | |
1868 | } | |
1869 | } | |
1870 | ||
1871 | msleep(10000); | |
1872 | ||
1873 | aac_acquire_resources(aac); | |
1874 | ||
1875 | /* | |
1876 | * reset this flag to unblock ioctl() as it was set | |
1877 | * at aac_send_shutdown() to block ioctls from upperlayer | |
1878 | */ | |
1879 | aac->adapter_shutdown = 0; | |
1880 | aac->handle_pci_error = 0; | |
1881 | ||
1882 | shost_for_each_device(sdev, shost) | |
1883 | if (sdev->sdev_state == SDEV_OFFLINE) | |
1884 | sdev->sdev_state = SDEV_RUNNING; | |
1885 | scsi_unblock_requests(aac->scsi_host_ptr); | |
1886 | scsi_scan_host(aac->scsi_host_ptr); | |
1887 | pci_save_state(pdev); | |
1888 | ||
1889 | dev_err(&pdev->dev, "aacraid: PCI error - resume\n"); | |
1890 | } | |
1891 | ||
1892 | static struct pci_error_handlers aac_pci_err_handler = { | |
1893 | .error_detected = aac_pci_error_detected, | |
1894 | .mmio_enabled = aac_pci_mmio_enabled, | |
1895 | .slot_reset = aac_pci_slot_reset, | |
1896 | .resume = aac_pci_resume, | |
1897 | }; | |
1898 | ||
1899 | static struct pci_driver aac_pci_driver = { | |
1900 | .name = AAC_DRIVERNAME, | |
1901 | .id_table = aac_pci_tbl, | |
1902 | .probe = aac_probe_one, | |
1903 | .remove = aac_remove_one, | |
1904 | #if (defined(CONFIG_PM)) | |
1905 | .suspend = aac_suspend, | |
1906 | .resume = aac_resume, | |
1907 | #endif | |
1908 | .shutdown = aac_shutdown, | |
1909 | .err_handler = &aac_pci_err_handler, | |
1910 | }; | |
1911 | ||
1912 | static int __init aac_init(void) | |
1913 | { | |
1914 | int error; | |
1915 | ||
1916 | printk(KERN_INFO "Adaptec %s driver %s\n", | |
1917 | AAC_DRIVERNAME, aac_driver_version); | |
1918 | ||
1919 | error = pci_register_driver(&aac_pci_driver); | |
1920 | if (error < 0) | |
1921 | return error; | |
1922 | ||
1923 | aac_init_char(); | |
1924 | ||
1925 | ||
1926 | return 0; | |
1927 | } | |
1928 | ||
1929 | static void __exit aac_exit(void) | |
1930 | { | |
1931 | if (aac_cfg_major > -1) | |
1932 | unregister_chrdev(aac_cfg_major, "aac"); | |
1933 | pci_unregister_driver(&aac_pci_driver); | |
1934 | } | |
1935 | ||
1936 | module_init(aac_init); | |
1937 | module_exit(aac_exit); |