blob: b595a94c4d16bb93b5dc61c7880534c1ef5a528e [file] [log] [blame]
Christoph Hellwig8c165672019-04-30 14:42:39 -04001// SPDX-License-Identifier: GPL-2.0
Christoph Hellwig973c4e32016-09-14 16:18:56 +02002/*
3 * Copyright (c) 2016 Christoph Hellwig.
Christoph Hellwig973c4e32016-09-14 16:18:56 +02004 */
Stephen Rothwell8ec2ef22016-09-19 15:50:16 +10005#include <linux/kobject.h>
6#include <linux/blkdev.h>
Christoph Hellwig973c4e32016-09-14 16:18:56 +02007#include <linux/blk-mq.h>
8#include <linux/blk-mq-pci.h>
9#include <linux/pci.h>
10#include <linux/module.h>
11
Minwoo Im0da73d02018-07-02 23:46:43 +090012#include "blk-mq.h"
13
Christoph Hellwig973c4e32016-09-14 16:18:56 +020014/**
15 * blk_mq_pci_map_queues - provide a default queue mapping for PCI device
Bart Van Assche0542cd52019-05-30 17:00:49 -070016 * @qmap: CPU to hardware queue map.
Christoph Hellwig973c4e32016-09-14 16:18:56 +020017 * @pdev: PCI device associated with @set.
Keith Buschf23f5bec2018-03-27 09:39:06 -060018 * @offset: Offset to use for the pci irq vector
Christoph Hellwig973c4e32016-09-14 16:18:56 +020019 *
20 * This function assumes the PCI device @pdev has at least as many available
Sagi Grimberg018c2592017-03-29 20:04:36 +030021 * interrupt vectors as @set has queues. It will then query the vector
Christoph Hellwig973c4e32016-09-14 16:18:56 +020022 * corresponding to each queue for it's affinity mask and built queue mapping
23 * that maps a queue to the CPUs that have irq affinity for the corresponding
24 * vector.
25 */
Jens Axboeed76e322018-10-29 13:06:14 -060026int blk_mq_pci_map_queues(struct blk_mq_queue_map *qmap, struct pci_dev *pdev,
Keith Buschf23f5bec2018-03-27 09:39:06 -060027 int offset)
Christoph Hellwig973c4e32016-09-14 16:18:56 +020028{
29 const struct cpumask *mask;
30 unsigned int queue, cpu;
31
Jens Axboeed76e322018-10-29 13:06:14 -060032 for (queue = 0; queue < qmap->nr_queues; queue++) {
Keith Buschf23f5bec2018-03-27 09:39:06 -060033 mask = pci_irq_get_affinity(pdev, queue + offset);
Christoph Hellwig973c4e32016-09-14 16:18:56 +020034 if (!mask)
Christoph Hellwigc0053902017-08-17 12:24:47 +020035 goto fallback;
Christoph Hellwig973c4e32016-09-14 16:18:56 +020036
37 for_each_cpu(cpu, mask)
Jens Axboe843477d2018-10-24 13:16:11 -060038 qmap->mq_map[cpu] = qmap->queue_offset + queue;
Christoph Hellwig973c4e32016-09-14 16:18:56 +020039 }
40
41 return 0;
Christoph Hellwigc0053902017-08-17 12:24:47 +020042
43fallback:
Jens Axboeed76e322018-10-29 13:06:14 -060044 WARN_ON_ONCE(qmap->nr_queues > 1);
45 blk_mq_clear_mq_map(qmap);
Christoph Hellwigc0053902017-08-17 12:24:47 +020046 return 0;
Christoph Hellwig973c4e32016-09-14 16:18:56 +020047}
48EXPORT_SYMBOL_GPL(blk_mq_pci_map_queues);