This source file includes following definitions.
- pf2vf_read_mbox
- pf2vf_write_mbox
- pf2vf_send_response
- pf2vf_resp_handler
- nitrox_pf2vf_mbox_handler
- nitrox_mbox_init
- nitrox_mbox_cleanup
1
2 #include <linux/workqueue.h>
3
4 #include "nitrox_csr.h"
5 #include "nitrox_hal.h"
6 #include "nitrox_dev.h"
7
8 #define RING_TO_VFNO(_x, _y) ((_x) / (_y))
9
10
11
12
13 enum mbx_msg_type {
14 MBX_MSG_TYPE_NOP,
15 MBX_MSG_TYPE_REQ,
16 MBX_MSG_TYPE_ACK,
17 MBX_MSG_TYPE_NACK,
18 };
19
20
21
22
23 enum mbx_msg_opcode {
24 MSG_OP_VF_MODE = 1,
25 MSG_OP_VF_UP,
26 MSG_OP_VF_DOWN,
27 MSG_OP_CHIPID_VFID,
28 };
29
30 struct pf2vf_work {
31 struct nitrox_vfdev *vfdev;
32 struct nitrox_device *ndev;
33 struct work_struct pf2vf_resp;
34 };
35
36 static inline u64 pf2vf_read_mbox(struct nitrox_device *ndev, int ring)
37 {
38 u64 reg_addr;
39
40 reg_addr = NPS_PKT_MBOX_VF_PF_PFDATAX(ring);
41 return nitrox_read_csr(ndev, reg_addr);
42 }
43
44 static inline void pf2vf_write_mbox(struct nitrox_device *ndev, u64 value,
45 int ring)
46 {
47 u64 reg_addr;
48
49 reg_addr = NPS_PKT_MBOX_PF_VF_PFDATAX(ring);
50 nitrox_write_csr(ndev, reg_addr, value);
51 }
52
53 static void pf2vf_send_response(struct nitrox_device *ndev,
54 struct nitrox_vfdev *vfdev)
55 {
56 union mbox_msg msg;
57
58 msg.value = vfdev->msg.value;
59
60 switch (vfdev->msg.opcode) {
61 case MSG_OP_VF_MODE:
62 msg.data = ndev->mode;
63 break;
64 case MSG_OP_VF_UP:
65 vfdev->nr_queues = vfdev->msg.data;
66 atomic_set(&vfdev->state, __NDEV_READY);
67 break;
68 case MSG_OP_CHIPID_VFID:
69 msg.id.chipid = ndev->idx;
70 msg.id.vfid = vfdev->vfno;
71 break;
72 case MSG_OP_VF_DOWN:
73 vfdev->nr_queues = 0;
74 atomic_set(&vfdev->state, __NDEV_NOT_READY);
75 break;
76 default:
77 msg.type = MBX_MSG_TYPE_NOP;
78 break;
79 }
80
81 if (msg.type == MBX_MSG_TYPE_NOP)
82 return;
83
84
85 msg.type = MBX_MSG_TYPE_ACK;
86 pf2vf_write_mbox(ndev, msg.value, vfdev->ring);
87
88 vfdev->msg.value = 0;
89 atomic64_inc(&vfdev->mbx_resp);
90 }
91
92 static void pf2vf_resp_handler(struct work_struct *work)
93 {
94 struct pf2vf_work *pf2vf_resp = container_of(work, struct pf2vf_work,
95 pf2vf_resp);
96 struct nitrox_vfdev *vfdev = pf2vf_resp->vfdev;
97 struct nitrox_device *ndev = pf2vf_resp->ndev;
98
99 switch (vfdev->msg.type) {
100 case MBX_MSG_TYPE_REQ:
101
102 pf2vf_send_response(ndev, vfdev);
103 break;
104 case MBX_MSG_TYPE_ACK:
105 case MBX_MSG_TYPE_NACK:
106 break;
107 };
108
109 kfree(pf2vf_resp);
110 }
111
112 void nitrox_pf2vf_mbox_handler(struct nitrox_device *ndev)
113 {
114 struct nitrox_vfdev *vfdev;
115 struct pf2vf_work *pfwork;
116 u64 value, reg_addr;
117 u32 i;
118 int vfno;
119
120
121 reg_addr = NPS_PKT_MBOX_INT_LO;
122 value = nitrox_read_csr(ndev, reg_addr);
123 for_each_set_bit(i, (const unsigned long *)&value, BITS_PER_LONG) {
124
125 vfno = RING_TO_VFNO(i, ndev->iov.max_vf_queues);
126 vfdev = ndev->iov.vfdev + vfno;
127 vfdev->ring = i;
128
129 vfdev->msg.value = pf2vf_read_mbox(ndev, vfdev->ring);
130 pfwork = kzalloc(sizeof(*pfwork), GFP_ATOMIC);
131 if (!pfwork)
132 continue;
133
134 pfwork->vfdev = vfdev;
135 pfwork->ndev = ndev;
136 INIT_WORK(&pfwork->pf2vf_resp, pf2vf_resp_handler);
137 queue_work(ndev->iov.pf2vf_wq, &pfwork->pf2vf_resp);
138
139 nitrox_write_csr(ndev, reg_addr, BIT_ULL(i));
140 }
141
142
143 reg_addr = NPS_PKT_MBOX_INT_HI;
144 value = nitrox_read_csr(ndev, reg_addr);
145 for_each_set_bit(i, (const unsigned long *)&value, BITS_PER_LONG) {
146
147 vfno = RING_TO_VFNO(i + 64, ndev->iov.max_vf_queues);
148 vfdev = ndev->iov.vfdev + vfno;
149 vfdev->ring = (i + 64);
150
151 vfdev->msg.value = pf2vf_read_mbox(ndev, vfdev->ring);
152
153 pfwork = kzalloc(sizeof(*pfwork), GFP_ATOMIC);
154 if (!pfwork)
155 continue;
156
157 pfwork->vfdev = vfdev;
158 pfwork->ndev = ndev;
159 INIT_WORK(&pfwork->pf2vf_resp, pf2vf_resp_handler);
160 queue_work(ndev->iov.pf2vf_wq, &pfwork->pf2vf_resp);
161
162 nitrox_write_csr(ndev, reg_addr, BIT_ULL(i));
163 }
164 }
165
166 int nitrox_mbox_init(struct nitrox_device *ndev)
167 {
168 struct nitrox_vfdev *vfdev;
169 int i;
170
171 ndev->iov.vfdev = kcalloc(ndev->iov.num_vfs,
172 sizeof(struct nitrox_vfdev), GFP_KERNEL);
173 if (!ndev->iov.vfdev)
174 return -ENOMEM;
175
176 for (i = 0; i < ndev->iov.num_vfs; i++) {
177 vfdev = ndev->iov.vfdev + i;
178 vfdev->vfno = i;
179 }
180
181
182 ndev->iov.pf2vf_wq = alloc_workqueue("nitrox_pf2vf", 0, 0);
183 if (!ndev->iov.pf2vf_wq) {
184 kfree(ndev->iov.vfdev);
185 return -ENOMEM;
186 }
187
188 enable_pf2vf_mbox_interrupts(ndev);
189
190 return 0;
191 }
192
193 void nitrox_mbox_cleanup(struct nitrox_device *ndev)
194 {
195
196 disable_pf2vf_mbox_interrupts(ndev);
197
198 if (ndev->iov.pf2vf_wq)
199 destroy_workqueue(ndev->iov.pf2vf_wq);
200
201 kfree(ndev->iov.vfdev);
202 ndev->iov.pf2vf_wq = NULL;
203 ndev->iov.vfdev = NULL;
204 }