1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Support PCI/PCIe on PowerNV platforms
4 *
5 * Copyright 2011 Benjamin Herrenschmidt, IBM Corp.
6 */
7
8 #include <linux/kernel.h>
9 #include <linux/pci.h>
10 #include <linux/delay.h>
11 #include <linux/string.h>
12 #include <linux/init.h>
13 #include <linux/irq.h>
14 #include <linux/io.h>
15 #include <linux/msi.h>
16 #include <linux/iommu.h>
17 #include <linux/sched/mm.h>
18
19 #include <asm/sections.h>
20 #include <asm/io.h>
21 #include <asm/pci-bridge.h>
22 #include <asm/machdep.h>
23 #include <asm/msi_bitmap.h>
24 #include <asm/ppc-pci.h>
25 #include <asm/pnv-pci.h>
26 #include <asm/opal.h>
27 #include <asm/iommu.h>
28 #include <asm/tce.h>
29 #include <asm/firmware.h>
30 #include <asm/eeh_event.h>
31 #include <asm/eeh.h>
32
33 #include "powernv.h"
34 #include "pci.h"
35
36 static DEFINE_MUTEX(tunnel_mutex);
37
pnv_pci_get_slot_id(struct device_node * np,uint64_t * id)38 int pnv_pci_get_slot_id(struct device_node *np, uint64_t *id)
39 {
40 struct device_node *node = np;
41 u32 bdfn;
42 u64 phbid;
43 int ret;
44
45 ret = of_property_read_u32(np, "reg", &bdfn);
46 if (ret)
47 return -ENXIO;
48
49 bdfn = ((bdfn & 0x00ffff00) >> 8);
50 for (node = np; node; node = of_get_parent(node)) {
51 if (!PCI_DN(node)) {
52 of_node_put(node);
53 break;
54 }
55
56 if (!of_device_is_compatible(node, "ibm,ioda2-phb") &&
57 !of_device_is_compatible(node, "ibm,ioda3-phb") &&
58 !of_device_is_compatible(node, "ibm,ioda2-npu2-opencapi-phb")) {
59 of_node_put(node);
60 continue;
61 }
62
63 ret = of_property_read_u64(node, "ibm,opal-phbid", &phbid);
64 if (ret) {
65 of_node_put(node);
66 return -ENXIO;
67 }
68
69 if (of_device_is_compatible(node, "ibm,ioda2-npu2-opencapi-phb"))
70 *id = PCI_PHB_SLOT_ID(phbid);
71 else
72 *id = PCI_SLOT_ID(phbid, bdfn);
73 return 0;
74 }
75
76 return -ENODEV;
77 }
78 EXPORT_SYMBOL_GPL(pnv_pci_get_slot_id);
79
pnv_pci_get_device_tree(uint32_t phandle,void * buf,uint64_t len)80 int pnv_pci_get_device_tree(uint32_t phandle, void *buf, uint64_t len)
81 {
82 int64_t rc;
83
84 if (!opal_check_token(OPAL_GET_DEVICE_TREE))
85 return -ENXIO;
86
87 rc = opal_get_device_tree(phandle, (uint64_t)buf, len);
88 if (rc < OPAL_SUCCESS)
89 return -EIO;
90
91 return rc;
92 }
93 EXPORT_SYMBOL_GPL(pnv_pci_get_device_tree);
94
pnv_pci_get_presence_state(uint64_t id,uint8_t * state)95 int pnv_pci_get_presence_state(uint64_t id, uint8_t *state)
96 {
97 int64_t rc;
98
99 if (!opal_check_token(OPAL_PCI_GET_PRESENCE_STATE))
100 return -ENXIO;
101
102 rc = opal_pci_get_presence_state(id, (uint64_t)state);
103 if (rc != OPAL_SUCCESS)
104 return -EIO;
105
106 return 0;
107 }
108 EXPORT_SYMBOL_GPL(pnv_pci_get_presence_state);
109
pnv_pci_get_power_state(uint64_t id,uint8_t * state)110 int pnv_pci_get_power_state(uint64_t id, uint8_t *state)
111 {
112 int64_t rc;
113
114 if (!opal_check_token(OPAL_PCI_GET_POWER_STATE))
115 return -ENXIO;
116
117 rc = opal_pci_get_power_state(id, (uint64_t)state);
118 if (rc != OPAL_SUCCESS)
119 return -EIO;
120
121 return 0;
122 }
123 EXPORT_SYMBOL_GPL(pnv_pci_get_power_state);
124
pnv_pci_set_power_state(uint64_t id,uint8_t state,struct opal_msg * msg)125 int pnv_pci_set_power_state(uint64_t id, uint8_t state, struct opal_msg *msg)
126 {
127 struct opal_msg m;
128 int token, ret;
129 int64_t rc;
130
131 if (!opal_check_token(OPAL_PCI_SET_POWER_STATE))
132 return -ENXIO;
133
134 token = opal_async_get_token_interruptible();
135 if (unlikely(token < 0))
136 return token;
137
138 rc = opal_pci_set_power_state(token, id, (uint64_t)&state);
139 if (rc == OPAL_SUCCESS) {
140 ret = 0;
141 goto exit;
142 } else if (rc != OPAL_ASYNC_COMPLETION) {
143 ret = -EIO;
144 goto exit;
145 }
146
147 ret = opal_async_wait_response(token, &m);
148 if (ret < 0)
149 goto exit;
150
151 if (msg) {
152 ret = 1;
153 memcpy(msg, &m, sizeof(m));
154 }
155
156 exit:
157 opal_async_release_token(token);
158 return ret;
159 }
160 EXPORT_SYMBOL_GPL(pnv_pci_set_power_state);
161
162 /* Nicely print the contents of the PE State Tables (PEST). */
pnv_pci_dump_pest(__be64 pestA[],__be64 pestB[],int pest_size)163 static void pnv_pci_dump_pest(__be64 pestA[], __be64 pestB[], int pest_size)
164 {
165 __be64 prevA = ULONG_MAX, prevB = ULONG_MAX;
166 bool dup = false;
167 int i;
168
169 for (i = 0; i < pest_size; i++) {
170 __be64 peA = be64_to_cpu(pestA[i]);
171 __be64 peB = be64_to_cpu(pestB[i]);
172
173 if (peA != prevA || peB != prevB) {
174 if (dup) {
175 pr_info("PE[..%03x] A/B: as above\n", i-1);
176 dup = false;
177 }
178 prevA = peA;
179 prevB = peB;
180 if (peA & PNV_IODA_STOPPED_STATE ||
181 peB & PNV_IODA_STOPPED_STATE)
182 pr_info("PE[%03x] A/B: %016llx %016llx\n",
183 i, peA, peB);
184 } else if (!dup && (peA & PNV_IODA_STOPPED_STATE ||
185 peB & PNV_IODA_STOPPED_STATE)) {
186 dup = true;
187 }
188 }
189 }
190
pnv_pci_dump_p7ioc_diag_data(struct pci_controller * hose,struct OpalIoPhbErrorCommon * common)191 static void pnv_pci_dump_p7ioc_diag_data(struct pci_controller *hose,
192 struct OpalIoPhbErrorCommon *common)
193 {
194 struct OpalIoP7IOCPhbErrorData *data;
195
196 data = (struct OpalIoP7IOCPhbErrorData *)common;
197 pr_info("P7IOC PHB#%x Diag-data (Version: %d)\n",
198 hose->global_number, be32_to_cpu(common->version));
199
200 if (data->brdgCtl)
201 pr_info("brdgCtl: %08x\n",
202 be32_to_cpu(data->brdgCtl));
203 if (data->portStatusReg || data->rootCmplxStatus ||
204 data->busAgentStatus)
205 pr_info("UtlSts: %08x %08x %08x\n",
206 be32_to_cpu(data->portStatusReg),
207 be32_to_cpu(data->rootCmplxStatus),
208 be32_to_cpu(data->busAgentStatus));
209 if (data->deviceStatus || data->slotStatus ||
210 data->linkStatus || data->devCmdStatus ||
211 data->devSecStatus)
212 pr_info("RootSts: %08x %08x %08x %08x %08x\n",
213 be32_to_cpu(data->deviceStatus),
214 be32_to_cpu(data->slotStatus),
215 be32_to_cpu(data->linkStatus),
216 be32_to_cpu(data->devCmdStatus),
217 be32_to_cpu(data->devSecStatus));
218 if (data->rootErrorStatus || data->uncorrErrorStatus ||
219 data->corrErrorStatus)
220 pr_info("RootErrSts: %08x %08x %08x\n",
221 be32_to_cpu(data->rootErrorStatus),
222 be32_to_cpu(data->uncorrErrorStatus),
223 be32_to_cpu(data->corrErrorStatus));
224 if (data->tlpHdr1 || data->tlpHdr2 ||
225 data->tlpHdr3 || data->tlpHdr4)
226 pr_info("RootErrLog: %08x %08x %08x %08x\n",
227 be32_to_cpu(data->tlpHdr1),
228 be32_to_cpu(data->tlpHdr2),
229 be32_to_cpu(data->tlpHdr3),
230 be32_to_cpu(data->tlpHdr4));
231 if (data->sourceId || data->errorClass ||
232 data->correlator)
233 pr_info("RootErrLog1: %08x %016llx %016llx\n",
234 be32_to_cpu(data->sourceId),
235 be64_to_cpu(data->errorClass),
236 be64_to_cpu(data->correlator));
237 if (data->p7iocPlssr || data->p7iocCsr)
238 pr_info("PhbSts: %016llx %016llx\n",
239 be64_to_cpu(data->p7iocPlssr),
240 be64_to_cpu(data->p7iocCsr));
241 if (data->lemFir)
242 pr_info("Lem: %016llx %016llx %016llx\n",
243 be64_to_cpu(data->lemFir),
244 be64_to_cpu(data->lemErrorMask),
245 be64_to_cpu(data->lemWOF));
246 if (data->phbErrorStatus)
247 pr_info("PhbErr: %016llx %016llx %016llx %016llx\n",
248 be64_to_cpu(data->phbErrorStatus),
249 be64_to_cpu(data->phbFirstErrorStatus),
250 be64_to_cpu(data->phbErrorLog0),
251 be64_to_cpu(data->phbErrorLog1));
252 if (data->mmioErrorStatus)
253 pr_info("OutErr: %016llx %016llx %016llx %016llx\n",
254 be64_to_cpu(data->mmioErrorStatus),
255 be64_to_cpu(data->mmioFirstErrorStatus),
256 be64_to_cpu(data->mmioErrorLog0),
257 be64_to_cpu(data->mmioErrorLog1));
258 if (data->dma0ErrorStatus)
259 pr_info("InAErr: %016llx %016llx %016llx %016llx\n",
260 be64_to_cpu(data->dma0ErrorStatus),
261 be64_to_cpu(data->dma0FirstErrorStatus),
262 be64_to_cpu(data->dma0ErrorLog0),
263 be64_to_cpu(data->dma0ErrorLog1));
264 if (data->dma1ErrorStatus)
265 pr_info("InBErr: %016llx %016llx %016llx %016llx\n",
266 be64_to_cpu(data->dma1ErrorStatus),
267 be64_to_cpu(data->dma1FirstErrorStatus),
268 be64_to_cpu(data->dma1ErrorLog0),
269 be64_to_cpu(data->dma1ErrorLog1));
270
271 pnv_pci_dump_pest(data->pestA, data->pestB, OPAL_P7IOC_NUM_PEST_REGS);
272 }
273
pnv_pci_dump_phb3_diag_data(struct pci_controller * hose,struct OpalIoPhbErrorCommon * common)274 static void pnv_pci_dump_phb3_diag_data(struct pci_controller *hose,
275 struct OpalIoPhbErrorCommon *common)
276 {
277 struct OpalIoPhb3ErrorData *data;
278
279 data = (struct OpalIoPhb3ErrorData*)common;
280 pr_info("PHB3 PHB#%x Diag-data (Version: %d)\n",
281 hose->global_number, be32_to_cpu(common->version));
282 if (data->brdgCtl)
283 pr_info("brdgCtl: %08x\n",
284 be32_to_cpu(data->brdgCtl));
285 if (data->portStatusReg || data->rootCmplxStatus ||
286 data->busAgentStatus)
287 pr_info("UtlSts: %08x %08x %08x\n",
288 be32_to_cpu(data->portStatusReg),
289 be32_to_cpu(data->rootCmplxStatus),
290 be32_to_cpu(data->busAgentStatus));
291 if (data->deviceStatus || data->slotStatus ||
292 data->linkStatus || data->devCmdStatus ||
293 data->devSecStatus)
294 pr_info("RootSts: %08x %08x %08x %08x %08x\n",
295 be32_to_cpu(data->deviceStatus),
296 be32_to_cpu(data->slotStatus),
297 be32_to_cpu(data->linkStatus),
298 be32_to_cpu(data->devCmdStatus),
299 be32_to_cpu(data->devSecStatus));
300 if (data->rootErrorStatus || data->uncorrErrorStatus ||
301 data->corrErrorStatus)
302 pr_info("RootErrSts: %08x %08x %08x\n",
303 be32_to_cpu(data->rootErrorStatus),
304 be32_to_cpu(data->uncorrErrorStatus),
305 be32_to_cpu(data->corrErrorStatus));
306 if (data->tlpHdr1 || data->tlpHdr2 ||
307 data->tlpHdr3 || data->tlpHdr4)
308 pr_info("RootErrLog: %08x %08x %08x %08x\n",
309 be32_to_cpu(data->tlpHdr1),
310 be32_to_cpu(data->tlpHdr2),
311 be32_to_cpu(data->tlpHdr3),
312 be32_to_cpu(data->tlpHdr4));
313 if (data->sourceId || data->errorClass ||
314 data->correlator)
315 pr_info("RootErrLog1: %08x %016llx %016llx\n",
316 be32_to_cpu(data->sourceId),
317 be64_to_cpu(data->errorClass),
318 be64_to_cpu(data->correlator));
319 if (data->nFir)
320 pr_info("nFir: %016llx %016llx %016llx\n",
321 be64_to_cpu(data->nFir),
322 be64_to_cpu(data->nFirMask),
323 be64_to_cpu(data->nFirWOF));
324 if (data->phbPlssr || data->phbCsr)
325 pr_info("PhbSts: %016llx %016llx\n",
326 be64_to_cpu(data->phbPlssr),
327 be64_to_cpu(data->phbCsr));
328 if (data->lemFir)
329 pr_info("Lem: %016llx %016llx %016llx\n",
330 be64_to_cpu(data->lemFir),
331 be64_to_cpu(data->lemErrorMask),
332 be64_to_cpu(data->lemWOF));
333 if (data->phbErrorStatus)
334 pr_info("PhbErr: %016llx %016llx %016llx %016llx\n",
335 be64_to_cpu(data->phbErrorStatus),
336 be64_to_cpu(data->phbFirstErrorStatus),
337 be64_to_cpu(data->phbErrorLog0),
338 be64_to_cpu(data->phbErrorLog1));
339 if (data->mmioErrorStatus)
340 pr_info("OutErr: %016llx %016llx %016llx %016llx\n",
341 be64_to_cpu(data->mmioErrorStatus),
342 be64_to_cpu(data->mmioFirstErrorStatus),
343 be64_to_cpu(data->mmioErrorLog0),
344 be64_to_cpu(data->mmioErrorLog1));
345 if (data->dma0ErrorStatus)
346 pr_info("InAErr: %016llx %016llx %016llx %016llx\n",
347 be64_to_cpu(data->dma0ErrorStatus),
348 be64_to_cpu(data->dma0FirstErrorStatus),
349 be64_to_cpu(data->dma0ErrorLog0),
350 be64_to_cpu(data->dma0ErrorLog1));
351 if (data->dma1ErrorStatus)
352 pr_info("InBErr: %016llx %016llx %016llx %016llx\n",
353 be64_to_cpu(data->dma1ErrorStatus),
354 be64_to_cpu(data->dma1FirstErrorStatus),
355 be64_to_cpu(data->dma1ErrorLog0),
356 be64_to_cpu(data->dma1ErrorLog1));
357
358 pnv_pci_dump_pest(data->pestA, data->pestB, OPAL_PHB3_NUM_PEST_REGS);
359 }
360
pnv_pci_dump_phb4_diag_data(struct pci_controller * hose,struct OpalIoPhbErrorCommon * common)361 static void pnv_pci_dump_phb4_diag_data(struct pci_controller *hose,
362 struct OpalIoPhbErrorCommon *common)
363 {
364 struct OpalIoPhb4ErrorData *data;
365
366 data = (struct OpalIoPhb4ErrorData*)common;
367 pr_info("PHB4 PHB#%d Diag-data (Version: %d)\n",
368 hose->global_number, be32_to_cpu(common->version));
369 if (data->brdgCtl)
370 pr_info("brdgCtl: %08x\n",
371 be32_to_cpu(data->brdgCtl));
372 if (data->deviceStatus || data->slotStatus ||
373 data->linkStatus || data->devCmdStatus ||
374 data->devSecStatus)
375 pr_info("RootSts: %08x %08x %08x %08x %08x\n",
376 be32_to_cpu(data->deviceStatus),
377 be32_to_cpu(data->slotStatus),
378 be32_to_cpu(data->linkStatus),
379 be32_to_cpu(data->devCmdStatus),
380 be32_to_cpu(data->devSecStatus));
381 if (data->rootErrorStatus || data->uncorrErrorStatus ||
382 data->corrErrorStatus)
383 pr_info("RootErrSts: %08x %08x %08x\n",
384 be32_to_cpu(data->rootErrorStatus),
385 be32_to_cpu(data->uncorrErrorStatus),
386 be32_to_cpu(data->corrErrorStatus));
387 if (data->tlpHdr1 || data->tlpHdr2 ||
388 data->tlpHdr3 || data->tlpHdr4)
389 pr_info("RootErrLog: %08x %08x %08x %08x\n",
390 be32_to_cpu(data->tlpHdr1),
391 be32_to_cpu(data->tlpHdr2),
392 be32_to_cpu(data->tlpHdr3),
393 be32_to_cpu(data->tlpHdr4));
394 if (data->sourceId)
395 pr_info("sourceId: %08x\n", be32_to_cpu(data->sourceId));
396 if (data->nFir)
397 pr_info("nFir: %016llx %016llx %016llx\n",
398 be64_to_cpu(data->nFir),
399 be64_to_cpu(data->nFirMask),
400 be64_to_cpu(data->nFirWOF));
401 if (data->phbPlssr || data->phbCsr)
402 pr_info("PhbSts: %016llx %016llx\n",
403 be64_to_cpu(data->phbPlssr),
404 be64_to_cpu(data->phbCsr));
405 if (data->lemFir)
406 pr_info("Lem: %016llx %016llx %016llx\n",
407 be64_to_cpu(data->lemFir),
408 be64_to_cpu(data->lemErrorMask),
409 be64_to_cpu(data->lemWOF));
410 if (data->phbErrorStatus)
411 pr_info("PhbErr: %016llx %016llx %016llx %016llx\n",
412 be64_to_cpu(data->phbErrorStatus),
413 be64_to_cpu(data->phbFirstErrorStatus),
414 be64_to_cpu(data->phbErrorLog0),
415 be64_to_cpu(data->phbErrorLog1));
416 if (data->phbTxeErrorStatus)
417 pr_info("PhbTxeErr: %016llx %016llx %016llx %016llx\n",
418 be64_to_cpu(data->phbTxeErrorStatus),
419 be64_to_cpu(data->phbTxeFirstErrorStatus),
420 be64_to_cpu(data->phbTxeErrorLog0),
421 be64_to_cpu(data->phbTxeErrorLog1));
422 if (data->phbRxeArbErrorStatus)
423 pr_info("RxeArbErr: %016llx %016llx %016llx %016llx\n",
424 be64_to_cpu(data->phbRxeArbErrorStatus),
425 be64_to_cpu(data->phbRxeArbFirstErrorStatus),
426 be64_to_cpu(data->phbRxeArbErrorLog0),
427 be64_to_cpu(data->phbRxeArbErrorLog1));
428 if (data->phbRxeMrgErrorStatus)
429 pr_info("RxeMrgErr: %016llx %016llx %016llx %016llx\n",
430 be64_to_cpu(data->phbRxeMrgErrorStatus),
431 be64_to_cpu(data->phbRxeMrgFirstErrorStatus),
432 be64_to_cpu(data->phbRxeMrgErrorLog0),
433 be64_to_cpu(data->phbRxeMrgErrorLog1));
434 if (data->phbRxeTceErrorStatus)
435 pr_info("RxeTceErr: %016llx %016llx %016llx %016llx\n",
436 be64_to_cpu(data->phbRxeTceErrorStatus),
437 be64_to_cpu(data->phbRxeTceFirstErrorStatus),
438 be64_to_cpu(data->phbRxeTceErrorLog0),
439 be64_to_cpu(data->phbRxeTceErrorLog1));
440
441 if (data->phbPblErrorStatus)
442 pr_info("PblErr: %016llx %016llx %016llx %016llx\n",
443 be64_to_cpu(data->phbPblErrorStatus),
444 be64_to_cpu(data->phbPblFirstErrorStatus),
445 be64_to_cpu(data->phbPblErrorLog0),
446 be64_to_cpu(data->phbPblErrorLog1));
447 if (data->phbPcieDlpErrorStatus)
448 pr_info("PcieDlp: %016llx %016llx %016llx\n",
449 be64_to_cpu(data->phbPcieDlpErrorLog1),
450 be64_to_cpu(data->phbPcieDlpErrorLog2),
451 be64_to_cpu(data->phbPcieDlpErrorStatus));
452 if (data->phbRegbErrorStatus)
453 pr_info("RegbErr: %016llx %016llx %016llx %016llx\n",
454 be64_to_cpu(data->phbRegbErrorStatus),
455 be64_to_cpu(data->phbRegbFirstErrorStatus),
456 be64_to_cpu(data->phbRegbErrorLog0),
457 be64_to_cpu(data->phbRegbErrorLog1));
458
459
460 pnv_pci_dump_pest(data->pestA, data->pestB, OPAL_PHB4_NUM_PEST_REGS);
461 }
462
pnv_pci_dump_phb_diag_data(struct pci_controller * hose,unsigned char * log_buff)463 void pnv_pci_dump_phb_diag_data(struct pci_controller *hose,
464 unsigned char *log_buff)
465 {
466 struct OpalIoPhbErrorCommon *common;
467
468 if (!hose || !log_buff)
469 return;
470
471 common = (struct OpalIoPhbErrorCommon *)log_buff;
472 switch (be32_to_cpu(common->ioType)) {
473 case OPAL_PHB_ERROR_DATA_TYPE_P7IOC:
474 pnv_pci_dump_p7ioc_diag_data(hose, common);
475 break;
476 case OPAL_PHB_ERROR_DATA_TYPE_PHB3:
477 pnv_pci_dump_phb3_diag_data(hose, common);
478 break;
479 case OPAL_PHB_ERROR_DATA_TYPE_PHB4:
480 pnv_pci_dump_phb4_diag_data(hose, common);
481 break;
482 default:
483 pr_warn("%s: Unrecognized ioType %d\n",
484 __func__, be32_to_cpu(common->ioType));
485 }
486 }
487
pnv_pci_handle_eeh_config(struct pnv_phb * phb,u32 pe_no)488 static void pnv_pci_handle_eeh_config(struct pnv_phb *phb, u32 pe_no)
489 {
490 unsigned long flags, rc;
491 int has_diag, ret = 0;
492
493 spin_lock_irqsave(&phb->lock, flags);
494
495 /* Fetch PHB diag-data */
496 rc = opal_pci_get_phb_diag_data2(phb->opal_id, phb->diag_data,
497 phb->diag_data_size);
498 has_diag = (rc == OPAL_SUCCESS);
499
500 /* If PHB supports compound PE, to handle it */
501 if (phb->unfreeze_pe) {
502 ret = phb->unfreeze_pe(phb,
503 pe_no,
504 OPAL_EEH_ACTION_CLEAR_FREEZE_ALL);
505 } else {
506 rc = opal_pci_eeh_freeze_clear(phb->opal_id,
507 pe_no,
508 OPAL_EEH_ACTION_CLEAR_FREEZE_ALL);
509 if (rc) {
510 pr_warn("%s: Failure %ld clearing frozen "
511 "PHB#%x-PE#%x\n",
512 __func__, rc, phb->hose->global_number,
513 pe_no);
514 ret = -EIO;
515 }
516 }
517
518 /*
519 * For now, let's only display the diag buffer when we fail to clear
520 * the EEH status. We'll do more sensible things later when we have
521 * proper EEH support. We need to make sure we don't pollute ourselves
522 * with the normal errors generated when probing empty slots
523 */
524 if (has_diag && ret)
525 pnv_pci_dump_phb_diag_data(phb->hose, phb->diag_data);
526
527 spin_unlock_irqrestore(&phb->lock, flags);
528 }
529
pnv_pci_config_check_eeh(struct pci_dn * pdn)530 static void pnv_pci_config_check_eeh(struct pci_dn *pdn)
531 {
532 struct pnv_phb *phb = pdn->phb->private_data;
533 u8 fstate = 0;
534 __be16 pcierr = 0;
535 unsigned int pe_no;
536 s64 rc;
537
538 /*
539 * Get the PE#. During the PCI probe stage, we might not
540 * setup that yet. So all ER errors should be mapped to
541 * reserved PE.
542 */
543 pe_no = pdn->pe_number;
544 if (pe_no == IODA_INVALID_PE) {
545 pe_no = phb->ioda.reserved_pe_idx;
546 }
547
548 /*
549 * Fetch frozen state. If the PHB support compound PE,
550 * we need handle that case.
551 */
552 if (phb->get_pe_state) {
553 fstate = phb->get_pe_state(phb, pe_no);
554 } else {
555 rc = opal_pci_eeh_freeze_status(phb->opal_id,
556 pe_no,
557 &fstate,
558 &pcierr,
559 NULL);
560 if (rc) {
561 pr_warn("%s: Failure %lld getting PHB#%x-PE#%x state\n",
562 __func__, rc, phb->hose->global_number, pe_no);
563 return;
564 }
565 }
566
567 pr_devel(" -> EEH check, bdfn=%04x PE#%x fstate=%x\n",
568 (pdn->busno << 8) | (pdn->devfn), pe_no, fstate);
569
570 /* Clear the frozen state if applicable */
571 if (fstate == OPAL_EEH_STOPPED_MMIO_FREEZE ||
572 fstate == OPAL_EEH_STOPPED_DMA_FREEZE ||
573 fstate == OPAL_EEH_STOPPED_MMIO_DMA_FREEZE) {
574 /*
575 * If PHB supports compound PE, freeze it for
576 * consistency.
577 */
578 if (phb->freeze_pe)
579 phb->freeze_pe(phb, pe_no);
580
581 pnv_pci_handle_eeh_config(phb, pe_no);
582 }
583 }
584
pnv_pci_cfg_read(struct pci_dn * pdn,int where,int size,u32 * val)585 int pnv_pci_cfg_read(struct pci_dn *pdn,
586 int where, int size, u32 *val)
587 {
588 struct pnv_phb *phb = pdn->phb->private_data;
589 u32 bdfn = (pdn->busno << 8) | pdn->devfn;
590 s64 rc;
591
592 switch (size) {
593 case 1: {
594 u8 v8;
595 rc = opal_pci_config_read_byte(phb->opal_id, bdfn, where, &v8);
596 *val = (rc == OPAL_SUCCESS) ? v8 : 0xff;
597 break;
598 }
599 case 2: {
600 __be16 v16;
601 rc = opal_pci_config_read_half_word(phb->opal_id, bdfn, where,
602 &v16);
603 *val = (rc == OPAL_SUCCESS) ? be16_to_cpu(v16) : 0xffff;
604 break;
605 }
606 case 4: {
607 __be32 v32;
608 rc = opal_pci_config_read_word(phb->opal_id, bdfn, where, &v32);
609 *val = (rc == OPAL_SUCCESS) ? be32_to_cpu(v32) : 0xffffffff;
610 break;
611 }
612 default:
613 return PCIBIOS_FUNC_NOT_SUPPORTED;
614 }
615
616 pr_devel("%s: bus: %x devfn: %x +%x/%x -> %08x\n",
617 __func__, pdn->busno, pdn->devfn, where, size, *val);
618 return PCIBIOS_SUCCESSFUL;
619 }
620
pnv_pci_cfg_write(struct pci_dn * pdn,int where,int size,u32 val)621 int pnv_pci_cfg_write(struct pci_dn *pdn,
622 int where, int size, u32 val)
623 {
624 struct pnv_phb *phb = pdn->phb->private_data;
625 u32 bdfn = (pdn->busno << 8) | pdn->devfn;
626
627 pr_devel("%s: bus: %x devfn: %x +%x/%x -> %08x\n",
628 __func__, pdn->busno, pdn->devfn, where, size, val);
629 switch (size) {
630 case 1:
631 opal_pci_config_write_byte(phb->opal_id, bdfn, where, val);
632 break;
633 case 2:
634 opal_pci_config_write_half_word(phb->opal_id, bdfn, where, val);
635 break;
636 case 4:
637 opal_pci_config_write_word(phb->opal_id, bdfn, where, val);
638 break;
639 default:
640 return PCIBIOS_FUNC_NOT_SUPPORTED;
641 }
642
643 return PCIBIOS_SUCCESSFUL;
644 }
645
646 #ifdef CONFIG_EEH
pnv_pci_cfg_check(struct pci_dn * pdn)647 static bool pnv_pci_cfg_check(struct pci_dn *pdn)
648 {
649 struct eeh_dev *edev = NULL;
650 struct pnv_phb *phb = pdn->phb->private_data;
651
652 /* EEH not enabled ? */
653 if (!(phb->flags & PNV_PHB_FLAG_EEH))
654 return true;
655
656 /* PE reset or device removed ? */
657 edev = pdn->edev;
658 if (edev) {
659 if (edev->pe &&
660 (edev->pe->state & EEH_PE_CFG_BLOCKED))
661 return false;
662
663 if (edev->mode & EEH_DEV_REMOVED)
664 return false;
665 }
666
667 return true;
668 }
669 #else
pnv_pci_cfg_check(struct pci_dn * pdn)670 static inline pnv_pci_cfg_check(struct pci_dn *pdn)
671 {
672 return true;
673 }
674 #endif /* CONFIG_EEH */
675
pnv_pci_read_config(struct pci_bus * bus,unsigned int devfn,int where,int size,u32 * val)676 static int pnv_pci_read_config(struct pci_bus *bus,
677 unsigned int devfn,
678 int where, int size, u32 *val)
679 {
680 struct pci_dn *pdn;
681 struct pnv_phb *phb;
682 int ret;
683
684 *val = 0xFFFFFFFF;
685 pdn = pci_get_pdn_by_devfn(bus, devfn);
686 if (!pdn)
687 return PCIBIOS_DEVICE_NOT_FOUND;
688
689 if (!pnv_pci_cfg_check(pdn))
690 return PCIBIOS_DEVICE_NOT_FOUND;
691
692 ret = pnv_pci_cfg_read(pdn, where, size, val);
693 phb = pdn->phb->private_data;
694 if (phb->flags & PNV_PHB_FLAG_EEH && pdn->edev) {
695 if (*val == EEH_IO_ERROR_VALUE(size) &&
696 eeh_dev_check_failure(pdn->edev))
697 return PCIBIOS_DEVICE_NOT_FOUND;
698 } else {
699 pnv_pci_config_check_eeh(pdn);
700 }
701
702 return ret;
703 }
704
pnv_pci_write_config(struct pci_bus * bus,unsigned int devfn,int where,int size,u32 val)705 static int pnv_pci_write_config(struct pci_bus *bus,
706 unsigned int devfn,
707 int where, int size, u32 val)
708 {
709 struct pci_dn *pdn;
710 struct pnv_phb *phb;
711 int ret;
712
713 pdn = pci_get_pdn_by_devfn(bus, devfn);
714 if (!pdn)
715 return PCIBIOS_DEVICE_NOT_FOUND;
716
717 if (!pnv_pci_cfg_check(pdn))
718 return PCIBIOS_DEVICE_NOT_FOUND;
719
720 ret = pnv_pci_cfg_write(pdn, where, size, val);
721 phb = pdn->phb->private_data;
722 if (!(phb->flags & PNV_PHB_FLAG_EEH))
723 pnv_pci_config_check_eeh(pdn);
724
725 return ret;
726 }
727
728 struct pci_ops pnv_pci_ops = {
729 .read = pnv_pci_read_config,
730 .write = pnv_pci_write_config,
731 };
732
pnv_pci_table_alloc(int nid)733 struct iommu_table *pnv_pci_table_alloc(int nid)
734 {
735 struct iommu_table *tbl;
736
737 tbl = kzalloc_node(sizeof(struct iommu_table), GFP_KERNEL, nid);
738 if (!tbl)
739 return NULL;
740
741 INIT_LIST_HEAD_RCU(&tbl->it_group_list);
742 kref_init(&tbl->it_kref);
743
744 return tbl;
745 }
746
pnv_pci_get_phb_node(struct pci_dev * dev)747 struct device_node *pnv_pci_get_phb_node(struct pci_dev *dev)
748 {
749 struct pci_controller *hose = pci_bus_to_host(dev->bus);
750
751 return of_node_get(hose->dn);
752 }
753 EXPORT_SYMBOL(pnv_pci_get_phb_node);
754
pnv_pci_set_tunnel_bar(struct pci_dev * dev,u64 addr,int enable)755 int pnv_pci_set_tunnel_bar(struct pci_dev *dev, u64 addr, int enable)
756 {
757 struct pnv_phb *phb = pci_bus_to_pnvhb(dev->bus);
758 u64 tunnel_bar;
759 __be64 val;
760 int rc;
761
762 if (!opal_check_token(OPAL_PCI_GET_PBCQ_TUNNEL_BAR))
763 return -ENXIO;
764 if (!opal_check_token(OPAL_PCI_SET_PBCQ_TUNNEL_BAR))
765 return -ENXIO;
766
767 mutex_lock(&tunnel_mutex);
768 rc = opal_pci_get_pbcq_tunnel_bar(phb->opal_id, &val);
769 if (rc != OPAL_SUCCESS) {
770 rc = -EIO;
771 goto out;
772 }
773 tunnel_bar = be64_to_cpu(val);
774 if (enable) {
775 /*
776 * Only one device per PHB can use atomics.
777 * Our policy is first-come, first-served.
778 */
779 if (tunnel_bar) {
780 if (tunnel_bar != addr)
781 rc = -EBUSY;
782 else
783 rc = 0; /* Setting same address twice is ok */
784 goto out;
785 }
786 } else {
787 /*
788 * The device that owns atomics and wants to release
789 * them must pass the same address with enable == 0.
790 */
791 if (tunnel_bar != addr) {
792 rc = -EPERM;
793 goto out;
794 }
795 addr = 0x0ULL;
796 }
797 rc = opal_pci_set_pbcq_tunnel_bar(phb->opal_id, addr);
798 rc = opal_error_code(rc);
799 out:
800 mutex_unlock(&tunnel_mutex);
801 return rc;
802 }
803 EXPORT_SYMBOL_GPL(pnv_pci_set_tunnel_bar);
804
pnv_pci_shutdown(void)805 void pnv_pci_shutdown(void)
806 {
807 struct pci_controller *hose;
808
809 list_for_each_entry(hose, &hose_list, list_node)
810 if (hose->controller_ops.shutdown)
811 hose->controller_ops.shutdown(hose);
812 }
813
814 /* Fixup wrong class code in p7ioc and p8 root complex */
pnv_p7ioc_rc_quirk(struct pci_dev * dev)815 static void pnv_p7ioc_rc_quirk(struct pci_dev *dev)
816 {
817 dev->class = PCI_CLASS_BRIDGE_PCI_NORMAL;
818 }
819 DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_IBM, 0x3b9, pnv_p7ioc_rc_quirk);
820
pnv_pci_init(void)821 void __init pnv_pci_init(void)
822 {
823 struct device_node *np;
824
825 pci_add_flags(PCI_CAN_SKIP_ISA_ALIGN);
826
827 /* If we don't have OPAL, eg. in sim, just skip PCI probe */
828 if (!firmware_has_feature(FW_FEATURE_OPAL))
829 return;
830
831 #ifdef CONFIG_PCIEPORTBUS
832 /*
833 * On PowerNV PCIe devices are (currently) managed in cooperation
834 * with firmware. This isn't *strictly* required, but there's enough
835 * assumptions baked into both firmware and the platform code that
836 * it's unwise to allow the portbus services to be used.
837 *
838 * We need to fix this eventually, but for now set this flag to disable
839 * the portbus driver. The AER service isn't required since that AER
840 * events are handled via EEH. The pciehp hotplug driver can't work
841 * without kernel changes (and portbus binding breaks pnv_php). The
842 * other services also require some thinking about how we're going
843 * to integrate them.
844 */
845 pcie_ports_disabled = true;
846 #endif
847
848 /* Look for IODA IO-Hubs. */
849 for_each_compatible_node(np, NULL, "ibm,ioda-hub") {
850 pnv_pci_init_ioda_hub(np);
851 }
852
853 /* Look for ioda2 built-in PHB3's */
854 for_each_compatible_node(np, NULL, "ibm,ioda2-phb")
855 pnv_pci_init_ioda2_phb(np);
856
857 /* Look for ioda3 built-in PHB4's, we treat them as IODA2 */
858 for_each_compatible_node(np, NULL, "ibm,ioda3-phb")
859 pnv_pci_init_ioda2_phb(np);
860
861 /* Look for NPU2 OpenCAPI PHBs */
862 for_each_compatible_node(np, NULL, "ibm,ioda2-npu2-opencapi-phb")
863 pnv_pci_init_npu2_opencapi_phb(np);
864
865 /* Configure IOMMU DMA hooks */
866 set_pci_dma_ops(&dma_iommu_ops);
867 }
868
pnv_tce_iommu_bus_notifier(struct notifier_block * nb,unsigned long action,void * data)869 static int pnv_tce_iommu_bus_notifier(struct notifier_block *nb,
870 unsigned long action, void *data)
871 {
872 struct device *dev = data;
873
874 switch (action) {
875 case BUS_NOTIFY_DEL_DEVICE:
876 iommu_del_device(dev);
877 return 0;
878 default:
879 return 0;
880 }
881 }
882
883 static struct notifier_block pnv_tce_iommu_bus_nb = {
884 .notifier_call = pnv_tce_iommu_bus_notifier,
885 };
886
pnv_tce_iommu_bus_notifier_init(void)887 static int __init pnv_tce_iommu_bus_notifier_init(void)
888 {
889 bus_register_notifier(&pci_bus_type, &pnv_tce_iommu_bus_nb);
890 return 0;
891 }
892 machine_subsys_initcall_sync(powernv, pnv_tce_iommu_bus_notifier_init);
893