1 // SPDX-License-Identifier: GPL-2.0+
2 /* Microchip Sparx5 Switch driver
3  *
4  * Copyright (c) 2021 Microchip Technology Inc. and its subsidiaries.
5  */
6 
7 #include "sparx5_main_regs.h"
8 #include "sparx5_main.h"
9 
sparx5_vlant_set_mask(struct sparx5 * sparx5,u16 vid)10 static int sparx5_vlant_set_mask(struct sparx5 *sparx5, u16 vid)
11 {
12 	u32 mask[3];
13 
14 	/* Divide up mask in 32 bit words */
15 	bitmap_to_arr32(mask, sparx5->vlan_mask[vid], SPX5_PORTS);
16 
17 	/* Output mask to respective registers */
18 	spx5_wr(mask[0], sparx5, ANA_L3_VLAN_MASK_CFG(vid));
19 	spx5_wr(mask[1], sparx5, ANA_L3_VLAN_MASK_CFG1(vid));
20 	spx5_wr(mask[2], sparx5, ANA_L3_VLAN_MASK_CFG2(vid));
21 
22 	return 0;
23 }
24 
sparx5_vlan_init(struct sparx5 * sparx5)25 void sparx5_vlan_init(struct sparx5 *sparx5)
26 {
27 	u16 vid;
28 
29 	spx5_rmw(ANA_L3_VLAN_CTRL_VLAN_ENA_SET(1),
30 		 ANA_L3_VLAN_CTRL_VLAN_ENA,
31 		 sparx5,
32 		 ANA_L3_VLAN_CTRL);
33 
34 	/* Map VLAN = FID */
35 	for (vid = NULL_VID; vid < VLAN_N_VID; vid++)
36 		spx5_rmw(ANA_L3_VLAN_CFG_VLAN_FID_SET(vid),
37 			 ANA_L3_VLAN_CFG_VLAN_FID,
38 			 sparx5,
39 			 ANA_L3_VLAN_CFG(vid));
40 }
41 
sparx5_vlan_port_setup(struct sparx5 * sparx5,int portno)42 void sparx5_vlan_port_setup(struct sparx5 *sparx5, int portno)
43 {
44 	struct sparx5_port *port = sparx5->ports[portno];
45 
46 	/* Configure PVID */
47 	spx5_rmw(ANA_CL_VLAN_CTRL_VLAN_AWARE_ENA_SET(0) |
48 		 ANA_CL_VLAN_CTRL_PORT_VID_SET(port->pvid),
49 		 ANA_CL_VLAN_CTRL_VLAN_AWARE_ENA |
50 		 ANA_CL_VLAN_CTRL_PORT_VID,
51 		 sparx5,
52 		 ANA_CL_VLAN_CTRL(port->portno));
53 }
54 
sparx5_vlan_vid_add(struct sparx5_port * port,u16 vid,bool pvid,bool untagged)55 int sparx5_vlan_vid_add(struct sparx5_port *port, u16 vid, bool pvid,
56 			bool untagged)
57 {
58 	struct sparx5 *sparx5 = port->sparx5;
59 	int ret;
60 
61 	/* Untagged egress vlan classification */
62 	if (untagged && port->vid != vid) {
63 		if (port->vid) {
64 			netdev_err(port->ndev,
65 				   "Port already has a native VLAN: %d\n",
66 				   port->vid);
67 			return -EBUSY;
68 		}
69 		port->vid = vid;
70 	}
71 
72 	/* Make the port a member of the VLAN */
73 	set_bit(port->portno, sparx5->vlan_mask[vid]);
74 	ret = sparx5_vlant_set_mask(sparx5, vid);
75 	if (ret)
76 		return ret;
77 
78 	/* Default ingress vlan classification */
79 	if (pvid)
80 		port->pvid = vid;
81 
82 	sparx5_vlan_port_apply(sparx5, port);
83 
84 	return 0;
85 }
86 
sparx5_vlan_vid_del(struct sparx5_port * port,u16 vid)87 int sparx5_vlan_vid_del(struct sparx5_port *port, u16 vid)
88 {
89 	struct sparx5 *sparx5 = port->sparx5;
90 	int ret;
91 
92 	/* 8021q removes VID 0 on module unload for all interfaces
93 	 * with VLAN filtering feature. We need to keep it to receive
94 	 * untagged traffic.
95 	 */
96 	if (vid == 0)
97 		return 0;
98 
99 	/* Stop the port from being a member of the vlan */
100 	clear_bit(port->portno, sparx5->vlan_mask[vid]);
101 	ret = sparx5_vlant_set_mask(sparx5, vid);
102 	if (ret)
103 		return ret;
104 
105 	/* Ingress */
106 	if (port->pvid == vid)
107 		port->pvid = 0;
108 
109 	/* Egress */
110 	if (port->vid == vid)
111 		port->vid = 0;
112 
113 	sparx5_vlan_port_apply(sparx5, port);
114 
115 	return 0;
116 }
117 
sparx5_pgid_update_mask(struct sparx5_port * port,int pgid,bool enable)118 void sparx5_pgid_update_mask(struct sparx5_port *port, int pgid, bool enable)
119 {
120 	struct sparx5 *sparx5 = port->sparx5;
121 	u32 val, mask;
122 
123 	/* mask is spread across 3 registers x 32 bit */
124 	if (port->portno < 32) {
125 		mask = BIT(port->portno);
126 		val = enable ? mask : 0;
127 		spx5_rmw(val, mask, sparx5, ANA_AC_PGID_CFG(pgid));
128 	} else if (port->portno < 64) {
129 		mask = BIT(port->portno - 32);
130 		val = enable ? mask : 0;
131 		spx5_rmw(val, mask, sparx5, ANA_AC_PGID_CFG1(pgid));
132 	} else if (port->portno < SPX5_PORTS) {
133 		mask = BIT(port->portno - 64);
134 		val = enable ? mask : 0;
135 		spx5_rmw(val, mask, sparx5, ANA_AC_PGID_CFG2(pgid));
136 	} else {
137 		netdev_err(port->ndev, "Invalid port no: %d\n", port->portno);
138 	}
139 }
140 
sparx5_pgid_clear(struct sparx5 * spx5,int pgid)141 void sparx5_pgid_clear(struct sparx5 *spx5, int pgid)
142 {
143 	spx5_wr(0, spx5, ANA_AC_PGID_CFG(pgid));
144 	spx5_wr(0, spx5, ANA_AC_PGID_CFG1(pgid));
145 	spx5_wr(0, spx5, ANA_AC_PGID_CFG2(pgid));
146 }
147 
sparx5_pgid_read_mask(struct sparx5 * spx5,int pgid,u32 portmask[3])148 void sparx5_pgid_read_mask(struct sparx5 *spx5, int pgid, u32 portmask[3])
149 {
150 	portmask[0] = spx5_rd(spx5, ANA_AC_PGID_CFG(pgid));
151 	portmask[1] = spx5_rd(spx5, ANA_AC_PGID_CFG1(pgid));
152 	portmask[2] = spx5_rd(spx5, ANA_AC_PGID_CFG2(pgid));
153 }
154 
sparx5_update_fwd(struct sparx5 * sparx5)155 void sparx5_update_fwd(struct sparx5 *sparx5)
156 {
157 	DECLARE_BITMAP(workmask, SPX5_PORTS);
158 	u32 mask[3];
159 	int port;
160 
161 	/* Divide up fwd mask in 32 bit words */
162 	bitmap_to_arr32(mask, sparx5->bridge_fwd_mask, SPX5_PORTS);
163 
164 	/* Update flood masks */
165 	for (port = PGID_UC_FLOOD; port <= PGID_BCAST; port++) {
166 		spx5_wr(mask[0], sparx5, ANA_AC_PGID_CFG(port));
167 		spx5_wr(mask[1], sparx5, ANA_AC_PGID_CFG1(port));
168 		spx5_wr(mask[2], sparx5, ANA_AC_PGID_CFG2(port));
169 	}
170 
171 	/* Update SRC masks */
172 	for (port = 0; port < SPX5_PORTS; port++) {
173 		if (test_bit(port, sparx5->bridge_fwd_mask)) {
174 			/* Allow to send to all bridged but self */
175 			bitmap_copy(workmask, sparx5->bridge_fwd_mask, SPX5_PORTS);
176 			clear_bit(port, workmask);
177 			bitmap_to_arr32(mask, workmask, SPX5_PORTS);
178 			spx5_wr(mask[0], sparx5, ANA_AC_SRC_CFG(port));
179 			spx5_wr(mask[1], sparx5, ANA_AC_SRC_CFG1(port));
180 			spx5_wr(mask[2], sparx5, ANA_AC_SRC_CFG2(port));
181 		} else {
182 			spx5_wr(0, sparx5, ANA_AC_SRC_CFG(port));
183 			spx5_wr(0, sparx5, ANA_AC_SRC_CFG1(port));
184 			spx5_wr(0, sparx5, ANA_AC_SRC_CFG2(port));
185 		}
186 	}
187 
188 	/* Learning enabled only for bridged ports */
189 	bitmap_and(workmask, sparx5->bridge_fwd_mask,
190 		   sparx5->bridge_lrn_mask, SPX5_PORTS);
191 	bitmap_to_arr32(mask, workmask, SPX5_PORTS);
192 
193 	/* Apply learning mask */
194 	spx5_wr(mask[0], sparx5, ANA_L2_AUTO_LRN_CFG);
195 	spx5_wr(mask[1], sparx5, ANA_L2_AUTO_LRN_CFG1);
196 	spx5_wr(mask[2], sparx5, ANA_L2_AUTO_LRN_CFG2);
197 }
198 
sparx5_vlan_port_apply(struct sparx5 * sparx5,struct sparx5_port * port)199 void sparx5_vlan_port_apply(struct sparx5 *sparx5,
200 			    struct sparx5_port *port)
201 
202 {
203 	u32 val;
204 
205 	/* Configure PVID, vlan aware */
206 	val = ANA_CL_VLAN_CTRL_VLAN_AWARE_ENA_SET(port->vlan_aware) |
207 		ANA_CL_VLAN_CTRL_VLAN_POP_CNT_SET(port->vlan_aware) |
208 		ANA_CL_VLAN_CTRL_PORT_VID_SET(port->pvid);
209 	spx5_wr(val, sparx5, ANA_CL_VLAN_CTRL(port->portno));
210 
211 	val = 0;
212 	if (port->vlan_aware && !port->pvid)
213 		/* If port is vlan-aware and tagged, drop untagged and
214 		 * priority tagged frames.
215 		 */
216 		val = ANA_CL_VLAN_FILTER_CTRL_TAG_REQUIRED_ENA_SET(1) |
217 			ANA_CL_VLAN_FILTER_CTRL_PRIO_CTAG_DIS_SET(1) |
218 			ANA_CL_VLAN_FILTER_CTRL_PRIO_STAG_DIS_SET(1);
219 	spx5_wr(val, sparx5,
220 		ANA_CL_VLAN_FILTER_CTRL(port->portno, 0));
221 
222 	/* Egress configuration (REW_TAG_CFG): VLAN tag type to 8021Q */
223 	val = REW_TAG_CTRL_TAG_TPID_CFG_SET(0);
224 	if (port->vlan_aware) {
225 		if (port->vid)
226 			/* Tag all frames except when VID == DEFAULT_VLAN */
227 			val |= REW_TAG_CTRL_TAG_CFG_SET(1);
228 		else
229 			val |= REW_TAG_CTRL_TAG_CFG_SET(3);
230 	}
231 	spx5_wr(val, sparx5, REW_TAG_CTRL(port->portno));
232 
233 	/* Egress VID */
234 	spx5_rmw(REW_PORT_VLAN_CFG_PORT_VID_SET(port->vid),
235 		 REW_PORT_VLAN_CFG_PORT_VID,
236 		 sparx5,
237 		 REW_PORT_VLAN_CFG(port->portno));
238 }
239