1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Greybus CPort control protocol.
4 *
5 * Copyright 2015 Google Inc.
6 * Copyright 2015 Linaro Ltd.
7 */
8
9 #include <linux/kernel.h>
10 #include <linux/module.h>
11 #include <linux/slab.h>
12 #include <linux/greybus.h>
13
14 /* Highest control-protocol version supported */
15 #define GB_CONTROL_VERSION_MAJOR 0
16 #define GB_CONTROL_VERSION_MINOR 1
17
gb_control_get_version(struct gb_control * control)18 static int gb_control_get_version(struct gb_control *control)
19 {
20 struct gb_interface *intf = control->connection->intf;
21 struct gb_control_version_request request;
22 struct gb_control_version_response response;
23 int ret;
24
25 request.major = GB_CONTROL_VERSION_MAJOR;
26 request.minor = GB_CONTROL_VERSION_MINOR;
27
28 ret = gb_operation_sync(control->connection,
29 GB_CONTROL_TYPE_VERSION,
30 &request, sizeof(request), &response,
31 sizeof(response));
32 if (ret) {
33 dev_err(&intf->dev,
34 "failed to get control-protocol version: %d\n",
35 ret);
36 return ret;
37 }
38
39 if (response.major > request.major) {
40 dev_err(&intf->dev,
41 "unsupported major control-protocol version (%u > %u)\n",
42 response.major, request.major);
43 return -ENOTSUPP;
44 }
45
46 control->protocol_major = response.major;
47 control->protocol_minor = response.minor;
48
49 dev_dbg(&intf->dev, "%s - %u.%u\n", __func__, response.major,
50 response.minor);
51
52 return 0;
53 }
54
gb_control_get_bundle_version(struct gb_control * control,struct gb_bundle * bundle)55 static int gb_control_get_bundle_version(struct gb_control *control,
56 struct gb_bundle *bundle)
57 {
58 struct gb_interface *intf = control->connection->intf;
59 struct gb_control_bundle_version_request request;
60 struct gb_control_bundle_version_response response;
61 int ret;
62
63 request.bundle_id = bundle->id;
64
65 ret = gb_operation_sync(control->connection,
66 GB_CONTROL_TYPE_BUNDLE_VERSION,
67 &request, sizeof(request),
68 &response, sizeof(response));
69 if (ret) {
70 dev_err(&intf->dev,
71 "failed to get bundle %u class version: %d\n",
72 bundle->id, ret);
73 return ret;
74 }
75
76 bundle->class_major = response.major;
77 bundle->class_minor = response.minor;
78
79 dev_dbg(&intf->dev, "%s - %u: %u.%u\n", __func__, bundle->id,
80 response.major, response.minor);
81
82 return 0;
83 }
84
gb_control_get_bundle_versions(struct gb_control * control)85 int gb_control_get_bundle_versions(struct gb_control *control)
86 {
87 struct gb_interface *intf = control->connection->intf;
88 struct gb_bundle *bundle;
89 int ret;
90
91 if (!control->has_bundle_version)
92 return 0;
93
94 list_for_each_entry(bundle, &intf->bundles, links) {
95 ret = gb_control_get_bundle_version(control, bundle);
96 if (ret)
97 return ret;
98 }
99
100 return 0;
101 }
102
103 /* Get Manifest's size from the interface */
gb_control_get_manifest_size_operation(struct gb_interface * intf)104 int gb_control_get_manifest_size_operation(struct gb_interface *intf)
105 {
106 struct gb_control_get_manifest_size_response response;
107 struct gb_connection *connection = intf->control->connection;
108 int ret;
109
110 ret = gb_operation_sync(connection, GB_CONTROL_TYPE_GET_MANIFEST_SIZE,
111 NULL, 0, &response, sizeof(response));
112 if (ret) {
113 dev_err(&connection->intf->dev,
114 "failed to get manifest size: %d\n", ret);
115 return ret;
116 }
117
118 return le16_to_cpu(response.size);
119 }
120
121 /* Reads Manifest from the interface */
gb_control_get_manifest_operation(struct gb_interface * intf,void * manifest,size_t size)122 int gb_control_get_manifest_operation(struct gb_interface *intf, void *manifest,
123 size_t size)
124 {
125 struct gb_connection *connection = intf->control->connection;
126
127 return gb_operation_sync(connection, GB_CONTROL_TYPE_GET_MANIFEST,
128 NULL, 0, manifest, size);
129 }
130
gb_control_connected_operation(struct gb_control * control,u16 cport_id)131 int gb_control_connected_operation(struct gb_control *control, u16 cport_id)
132 {
133 struct gb_control_connected_request request;
134
135 request.cport_id = cpu_to_le16(cport_id);
136 return gb_operation_sync(control->connection, GB_CONTROL_TYPE_CONNECTED,
137 &request, sizeof(request), NULL, 0);
138 }
139
gb_control_disconnected_operation(struct gb_control * control,u16 cport_id)140 int gb_control_disconnected_operation(struct gb_control *control, u16 cport_id)
141 {
142 struct gb_control_disconnected_request request;
143
144 request.cport_id = cpu_to_le16(cport_id);
145 return gb_operation_sync(control->connection,
146 GB_CONTROL_TYPE_DISCONNECTED, &request,
147 sizeof(request), NULL, 0);
148 }
149
gb_control_disconnecting_operation(struct gb_control * control,u16 cport_id)150 int gb_control_disconnecting_operation(struct gb_control *control,
151 u16 cport_id)
152 {
153 struct gb_control_disconnecting_request *request;
154 struct gb_operation *operation;
155 int ret;
156
157 operation = gb_operation_create_core(control->connection,
158 GB_CONTROL_TYPE_DISCONNECTING,
159 sizeof(*request), 0, 0,
160 GFP_KERNEL);
161 if (!operation)
162 return -ENOMEM;
163
164 request = operation->request->payload;
165 request->cport_id = cpu_to_le16(cport_id);
166
167 ret = gb_operation_request_send_sync(operation);
168 if (ret) {
169 dev_err(&control->dev, "failed to send disconnecting: %d\n",
170 ret);
171 }
172
173 gb_operation_put(operation);
174
175 return ret;
176 }
177
gb_control_mode_switch_operation(struct gb_control * control)178 int gb_control_mode_switch_operation(struct gb_control *control)
179 {
180 struct gb_operation *operation;
181 int ret;
182
183 operation = gb_operation_create_core(control->connection,
184 GB_CONTROL_TYPE_MODE_SWITCH,
185 0, 0,
186 GB_OPERATION_FLAG_UNIDIRECTIONAL,
187 GFP_KERNEL);
188 if (!operation)
189 return -ENOMEM;
190
191 ret = gb_operation_request_send_sync(operation);
192 if (ret)
193 dev_err(&control->dev, "failed to send mode switch: %d\n", ret);
194
195 gb_operation_put(operation);
196
197 return ret;
198 }
199
gb_control_bundle_pm_status_map(u8 status)200 static int gb_control_bundle_pm_status_map(u8 status)
201 {
202 switch (status) {
203 case GB_CONTROL_BUNDLE_PM_INVAL:
204 return -EINVAL;
205 case GB_CONTROL_BUNDLE_PM_BUSY:
206 return -EBUSY;
207 case GB_CONTROL_BUNDLE_PM_NA:
208 return -ENOMSG;
209 case GB_CONTROL_BUNDLE_PM_FAIL:
210 default:
211 return -EREMOTEIO;
212 }
213 }
214
gb_control_bundle_suspend(struct gb_control * control,u8 bundle_id)215 int gb_control_bundle_suspend(struct gb_control *control, u8 bundle_id)
216 {
217 struct gb_control_bundle_pm_request request;
218 struct gb_control_bundle_pm_response response;
219 int ret;
220
221 request.bundle_id = bundle_id;
222 ret = gb_operation_sync(control->connection,
223 GB_CONTROL_TYPE_BUNDLE_SUSPEND, &request,
224 sizeof(request), &response, sizeof(response));
225 if (ret) {
226 dev_err(&control->dev, "failed to send bundle %u suspend: %d\n",
227 bundle_id, ret);
228 return ret;
229 }
230
231 if (response.status != GB_CONTROL_BUNDLE_PM_OK) {
232 dev_err(&control->dev, "failed to suspend bundle %u: %d\n",
233 bundle_id, response.status);
234 return gb_control_bundle_pm_status_map(response.status);
235 }
236
237 return 0;
238 }
239
gb_control_bundle_resume(struct gb_control * control,u8 bundle_id)240 int gb_control_bundle_resume(struct gb_control *control, u8 bundle_id)
241 {
242 struct gb_control_bundle_pm_request request;
243 struct gb_control_bundle_pm_response response;
244 int ret;
245
246 request.bundle_id = bundle_id;
247 ret = gb_operation_sync(control->connection,
248 GB_CONTROL_TYPE_BUNDLE_RESUME, &request,
249 sizeof(request), &response, sizeof(response));
250 if (ret) {
251 dev_err(&control->dev, "failed to send bundle %u resume: %d\n",
252 bundle_id, ret);
253 return ret;
254 }
255
256 if (response.status != GB_CONTROL_BUNDLE_PM_OK) {
257 dev_err(&control->dev, "failed to resume bundle %u: %d\n",
258 bundle_id, response.status);
259 return gb_control_bundle_pm_status_map(response.status);
260 }
261
262 return 0;
263 }
264
gb_control_bundle_deactivate(struct gb_control * control,u8 bundle_id)265 int gb_control_bundle_deactivate(struct gb_control *control, u8 bundle_id)
266 {
267 struct gb_control_bundle_pm_request request;
268 struct gb_control_bundle_pm_response response;
269 int ret;
270
271 request.bundle_id = bundle_id;
272 ret = gb_operation_sync(control->connection,
273 GB_CONTROL_TYPE_BUNDLE_DEACTIVATE, &request,
274 sizeof(request), &response, sizeof(response));
275 if (ret) {
276 dev_err(&control->dev,
277 "failed to send bundle %u deactivate: %d\n", bundle_id,
278 ret);
279 return ret;
280 }
281
282 if (response.status != GB_CONTROL_BUNDLE_PM_OK) {
283 dev_err(&control->dev, "failed to deactivate bundle %u: %d\n",
284 bundle_id, response.status);
285 return gb_control_bundle_pm_status_map(response.status);
286 }
287
288 return 0;
289 }
290
gb_control_bundle_activate(struct gb_control * control,u8 bundle_id)291 int gb_control_bundle_activate(struct gb_control *control, u8 bundle_id)
292 {
293 struct gb_control_bundle_pm_request request;
294 struct gb_control_bundle_pm_response response;
295 int ret;
296
297 if (!control->has_bundle_activate)
298 return 0;
299
300 request.bundle_id = bundle_id;
301 ret = gb_operation_sync(control->connection,
302 GB_CONTROL_TYPE_BUNDLE_ACTIVATE, &request,
303 sizeof(request), &response, sizeof(response));
304 if (ret) {
305 dev_err(&control->dev,
306 "failed to send bundle %u activate: %d\n", bundle_id,
307 ret);
308 return ret;
309 }
310
311 if (response.status != GB_CONTROL_BUNDLE_PM_OK) {
312 dev_err(&control->dev, "failed to activate bundle %u: %d\n",
313 bundle_id, response.status);
314 return gb_control_bundle_pm_status_map(response.status);
315 }
316
317 return 0;
318 }
319
gb_control_interface_pm_status_map(u8 status)320 static int gb_control_interface_pm_status_map(u8 status)
321 {
322 switch (status) {
323 case GB_CONTROL_INTF_PM_BUSY:
324 return -EBUSY;
325 case GB_CONTROL_INTF_PM_NA:
326 return -ENOMSG;
327 default:
328 return -EREMOTEIO;
329 }
330 }
331
gb_control_interface_suspend_prepare(struct gb_control * control)332 int gb_control_interface_suspend_prepare(struct gb_control *control)
333 {
334 struct gb_control_intf_pm_response response;
335 int ret;
336
337 ret = gb_operation_sync(control->connection,
338 GB_CONTROL_TYPE_INTF_SUSPEND_PREPARE, NULL, 0,
339 &response, sizeof(response));
340 if (ret) {
341 dev_err(&control->dev,
342 "failed to send interface suspend prepare: %d\n", ret);
343 return ret;
344 }
345
346 if (response.status != GB_CONTROL_INTF_PM_OK) {
347 dev_err(&control->dev, "interface error while preparing suspend: %d\n",
348 response.status);
349 return gb_control_interface_pm_status_map(response.status);
350 }
351
352 return 0;
353 }
354
gb_control_interface_deactivate_prepare(struct gb_control * control)355 int gb_control_interface_deactivate_prepare(struct gb_control *control)
356 {
357 struct gb_control_intf_pm_response response;
358 int ret;
359
360 ret = gb_operation_sync(control->connection,
361 GB_CONTROL_TYPE_INTF_DEACTIVATE_PREPARE, NULL,
362 0, &response, sizeof(response));
363 if (ret) {
364 dev_err(&control->dev, "failed to send interface deactivate prepare: %d\n",
365 ret);
366 return ret;
367 }
368
369 if (response.status != GB_CONTROL_INTF_PM_OK) {
370 dev_err(&control->dev, "interface error while preparing deactivate: %d\n",
371 response.status);
372 return gb_control_interface_pm_status_map(response.status);
373 }
374
375 return 0;
376 }
377
gb_control_interface_hibernate_abort(struct gb_control * control)378 int gb_control_interface_hibernate_abort(struct gb_control *control)
379 {
380 struct gb_control_intf_pm_response response;
381 int ret;
382
383 ret = gb_operation_sync(control->connection,
384 GB_CONTROL_TYPE_INTF_HIBERNATE_ABORT, NULL, 0,
385 &response, sizeof(response));
386 if (ret) {
387 dev_err(&control->dev,
388 "failed to send interface aborting hibernate: %d\n",
389 ret);
390 return ret;
391 }
392
393 if (response.status != GB_CONTROL_INTF_PM_OK) {
394 dev_err(&control->dev, "interface error while aborting hibernate: %d\n",
395 response.status);
396 return gb_control_interface_pm_status_map(response.status);
397 }
398
399 return 0;
400 }
401
vendor_string_show(struct device * dev,struct device_attribute * attr,char * buf)402 static ssize_t vendor_string_show(struct device *dev,
403 struct device_attribute *attr, char *buf)
404 {
405 struct gb_control *control = to_gb_control(dev);
406
407 return scnprintf(buf, PAGE_SIZE, "%s\n", control->vendor_string);
408 }
409 static DEVICE_ATTR_RO(vendor_string);
410
product_string_show(struct device * dev,struct device_attribute * attr,char * buf)411 static ssize_t product_string_show(struct device *dev,
412 struct device_attribute *attr, char *buf)
413 {
414 struct gb_control *control = to_gb_control(dev);
415
416 return scnprintf(buf, PAGE_SIZE, "%s\n", control->product_string);
417 }
418 static DEVICE_ATTR_RO(product_string);
419
420 static struct attribute *control_attrs[] = {
421 &dev_attr_vendor_string.attr,
422 &dev_attr_product_string.attr,
423 NULL,
424 };
425 ATTRIBUTE_GROUPS(control);
426
gb_control_release(struct device * dev)427 static void gb_control_release(struct device *dev)
428 {
429 struct gb_control *control = to_gb_control(dev);
430
431 gb_connection_destroy(control->connection);
432
433 kfree(control->vendor_string);
434 kfree(control->product_string);
435
436 kfree(control);
437 }
438
439 struct device_type greybus_control_type = {
440 .name = "greybus_control",
441 .release = gb_control_release,
442 };
443
gb_control_create(struct gb_interface * intf)444 struct gb_control *gb_control_create(struct gb_interface *intf)
445 {
446 struct gb_connection *connection;
447 struct gb_control *control;
448
449 control = kzalloc(sizeof(*control), GFP_KERNEL);
450 if (!control)
451 return ERR_PTR(-ENOMEM);
452
453 control->intf = intf;
454
455 connection = gb_connection_create_control(intf);
456 if (IS_ERR(connection)) {
457 dev_err(&intf->dev,
458 "failed to create control connection: %ld\n",
459 PTR_ERR(connection));
460 kfree(control);
461 return ERR_CAST(connection);
462 }
463
464 control->connection = connection;
465
466 control->dev.parent = &intf->dev;
467 control->dev.bus = &greybus_bus_type;
468 control->dev.type = &greybus_control_type;
469 control->dev.groups = control_groups;
470 control->dev.dma_mask = intf->dev.dma_mask;
471 device_initialize(&control->dev);
472 dev_set_name(&control->dev, "%s.ctrl", dev_name(&intf->dev));
473
474 gb_connection_set_data(control->connection, control);
475
476 return control;
477 }
478
gb_control_enable(struct gb_control * control)479 int gb_control_enable(struct gb_control *control)
480 {
481 int ret;
482
483 dev_dbg(&control->connection->intf->dev, "%s\n", __func__);
484
485 ret = gb_connection_enable_tx(control->connection);
486 if (ret) {
487 dev_err(&control->connection->intf->dev,
488 "failed to enable control connection: %d\n",
489 ret);
490 return ret;
491 }
492
493 ret = gb_control_get_version(control);
494 if (ret)
495 goto err_disable_connection;
496
497 if (control->protocol_major > 0 || control->protocol_minor > 1)
498 control->has_bundle_version = true;
499
500 /* FIXME: use protocol version instead */
501 if (!(control->intf->quirks & GB_INTERFACE_QUIRK_NO_BUNDLE_ACTIVATE))
502 control->has_bundle_activate = true;
503
504 return 0;
505
506 err_disable_connection:
507 gb_connection_disable(control->connection);
508
509 return ret;
510 }
511
gb_control_disable(struct gb_control * control)512 void gb_control_disable(struct gb_control *control)
513 {
514 dev_dbg(&control->connection->intf->dev, "%s\n", __func__);
515
516 if (control->intf->disconnected)
517 gb_connection_disable_forced(control->connection);
518 else
519 gb_connection_disable(control->connection);
520 }
521
gb_control_suspend(struct gb_control * control)522 int gb_control_suspend(struct gb_control *control)
523 {
524 gb_connection_disable(control->connection);
525
526 return 0;
527 }
528
gb_control_resume(struct gb_control * control)529 int gb_control_resume(struct gb_control *control)
530 {
531 int ret;
532
533 ret = gb_connection_enable_tx(control->connection);
534 if (ret) {
535 dev_err(&control->connection->intf->dev,
536 "failed to enable control connection: %d\n", ret);
537 return ret;
538 }
539
540 return 0;
541 }
542
gb_control_add(struct gb_control * control)543 int gb_control_add(struct gb_control *control)
544 {
545 int ret;
546
547 ret = device_add(&control->dev);
548 if (ret) {
549 dev_err(&control->dev,
550 "failed to register control device: %d\n",
551 ret);
552 return ret;
553 }
554
555 return 0;
556 }
557
gb_control_del(struct gb_control * control)558 void gb_control_del(struct gb_control *control)
559 {
560 if (device_is_registered(&control->dev))
561 device_del(&control->dev);
562 }
563
gb_control_get(struct gb_control * control)564 struct gb_control *gb_control_get(struct gb_control *control)
565 {
566 get_device(&control->dev);
567
568 return control;
569 }
570
gb_control_put(struct gb_control * control)571 void gb_control_put(struct gb_control *control)
572 {
573 put_device(&control->dev);
574 }
575
gb_control_mode_switch_prepare(struct gb_control * control)576 void gb_control_mode_switch_prepare(struct gb_control *control)
577 {
578 gb_connection_mode_switch_prepare(control->connection);
579 }
580
gb_control_mode_switch_complete(struct gb_control * control)581 void gb_control_mode_switch_complete(struct gb_control *control)
582 {
583 gb_connection_mode_switch_complete(control->connection);
584 }
585