1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Copyright (C) 2006-2010 Red Hat, Inc. All rights reserved.
4 */
5
6 #include <linux/miscdevice.h>
7 #include <linux/init.h>
8 #include <linux/wait.h>
9 #include <linux/file.h>
10 #include <linux/fs.h>
11 #include <linux/poll.h>
12 #include <linux/signal.h>
13 #include <linux/spinlock.h>
14 #include <linux/dlm.h>
15 #include <linux/dlm_device.h>
16 #include <linux/slab.h>
17 #include <linux/sched/signal.h>
18
19 #include <trace/events/dlm.h>
20
21 #include "dlm_internal.h"
22 #include "lockspace.h"
23 #include "lock.h"
24 #include "lvb_table.h"
25 #include "user.h"
26 #include "ast.h"
27 #include "config.h"
28 #include "memory.h"
29
30 static const char name_prefix[] = "dlm";
31 static const struct file_operations device_fops;
32 static atomic_t dlm_monitor_opened;
33 static int dlm_monitor_unused = 1;
34
35 #ifdef CONFIG_COMPAT
36
37 struct dlm_lock_params32 {
38 __u8 mode;
39 __u8 namelen;
40 __u16 unused;
41 __u32 flags;
42 __u32 lkid;
43 __u32 parent;
44 __u64 xid;
45 __u64 timeout;
46 __u32 castparam;
47 __u32 castaddr;
48 __u32 bastparam;
49 __u32 bastaddr;
50 __u32 lksb;
51 char lvb[DLM_USER_LVB_LEN];
52 char name[];
53 };
54
55 struct dlm_write_request32 {
56 __u32 version[3];
57 __u8 cmd;
58 __u8 is64bit;
59 __u8 unused[2];
60
61 union {
62 struct dlm_lock_params32 lock;
63 struct dlm_lspace_params lspace;
64 struct dlm_purge_params purge;
65 } i;
66 };
67
68 struct dlm_lksb32 {
69 __u32 sb_status;
70 __u32 sb_lkid;
71 __u8 sb_flags;
72 __u32 sb_lvbptr;
73 };
74
75 struct dlm_lock_result32 {
76 __u32 version[3];
77 __u32 length;
78 __u32 user_astaddr;
79 __u32 user_astparam;
80 __u32 user_lksb;
81 struct dlm_lksb32 lksb;
82 __u8 bast_mode;
83 __u8 unused[3];
84 /* Offsets may be zero if no data is present */
85 __u32 lvb_offset;
86 };
87
compat_input(struct dlm_write_request * kb,struct dlm_write_request32 * kb32,int namelen)88 static void compat_input(struct dlm_write_request *kb,
89 struct dlm_write_request32 *kb32,
90 int namelen)
91 {
92 kb->version[0] = kb32->version[0];
93 kb->version[1] = kb32->version[1];
94 kb->version[2] = kb32->version[2];
95
96 kb->cmd = kb32->cmd;
97 kb->is64bit = kb32->is64bit;
98 if (kb->cmd == DLM_USER_CREATE_LOCKSPACE ||
99 kb->cmd == DLM_USER_REMOVE_LOCKSPACE) {
100 kb->i.lspace.flags = kb32->i.lspace.flags;
101 kb->i.lspace.minor = kb32->i.lspace.minor;
102 memcpy(kb->i.lspace.name, kb32->i.lspace.name, namelen);
103 } else if (kb->cmd == DLM_USER_PURGE) {
104 kb->i.purge.nodeid = kb32->i.purge.nodeid;
105 kb->i.purge.pid = kb32->i.purge.pid;
106 } else {
107 kb->i.lock.mode = kb32->i.lock.mode;
108 kb->i.lock.namelen = kb32->i.lock.namelen;
109 kb->i.lock.flags = kb32->i.lock.flags;
110 kb->i.lock.lkid = kb32->i.lock.lkid;
111 kb->i.lock.parent = kb32->i.lock.parent;
112 kb->i.lock.xid = kb32->i.lock.xid;
113 kb->i.lock.timeout = kb32->i.lock.timeout;
114 kb->i.lock.castparam = (__user void *)(long)kb32->i.lock.castparam;
115 kb->i.lock.castaddr = (__user void *)(long)kb32->i.lock.castaddr;
116 kb->i.lock.bastparam = (__user void *)(long)kb32->i.lock.bastparam;
117 kb->i.lock.bastaddr = (__user void *)(long)kb32->i.lock.bastaddr;
118 kb->i.lock.lksb = (__user void *)(long)kb32->i.lock.lksb;
119 memcpy(kb->i.lock.lvb, kb32->i.lock.lvb, DLM_USER_LVB_LEN);
120 memcpy(kb->i.lock.name, kb32->i.lock.name, namelen);
121 }
122 }
123
compat_output(struct dlm_lock_result * res,struct dlm_lock_result32 * res32)124 static void compat_output(struct dlm_lock_result *res,
125 struct dlm_lock_result32 *res32)
126 {
127 memset(res32, 0, sizeof(*res32));
128
129 res32->version[0] = res->version[0];
130 res32->version[1] = res->version[1];
131 res32->version[2] = res->version[2];
132
133 res32->user_astaddr = (__u32)(__force long)res->user_astaddr;
134 res32->user_astparam = (__u32)(__force long)res->user_astparam;
135 res32->user_lksb = (__u32)(__force long)res->user_lksb;
136 res32->bast_mode = res->bast_mode;
137
138 res32->lvb_offset = res->lvb_offset;
139 res32->length = res->length;
140
141 res32->lksb.sb_status = res->lksb.sb_status;
142 res32->lksb.sb_flags = res->lksb.sb_flags;
143 res32->lksb.sb_lkid = res->lksb.sb_lkid;
144 res32->lksb.sb_lvbptr = (__u32)(long)res->lksb.sb_lvbptr;
145 }
146 #endif
147
148 /* should held proc->asts_spin lock */
dlm_purge_lkb_callbacks(struct dlm_lkb * lkb)149 void dlm_purge_lkb_callbacks(struct dlm_lkb *lkb)
150 {
151 struct dlm_callback *cb, *safe;
152
153 list_for_each_entry_safe(cb, safe, &lkb->lkb_callbacks, list) {
154 list_del(&cb->list);
155 kref_put(&cb->ref, dlm_release_callback);
156 }
157
158 clear_bit(DLM_IFL_CB_PENDING_BIT, &lkb->lkb_iflags);
159
160 /* invalidate */
161 dlm_callback_set_last_ptr(&lkb->lkb_last_cast, NULL);
162 dlm_callback_set_last_ptr(&lkb->lkb_last_cb, NULL);
163 lkb->lkb_last_bast_mode = -1;
164 }
165
166 /* Figure out if this lock is at the end of its life and no longer
167 available for the application to use. The lkb still exists until
168 the final ast is read. A lock becomes EOL in three situations:
169 1. a noqueue request fails with EAGAIN
170 2. an unlock completes with EUNLOCK
171 3. a cancel of a waiting request completes with ECANCEL/EDEADLK
172 An EOL lock needs to be removed from the process's list of locks.
173 And we can't allow any new operation on an EOL lock. This is
174 not related to the lifetime of the lkb struct which is managed
175 entirely by refcount. */
176
lkb_is_endoflife(int mode,int status)177 static int lkb_is_endoflife(int mode, int status)
178 {
179 switch (status) {
180 case -DLM_EUNLOCK:
181 return 1;
182 case -DLM_ECANCEL:
183 case -ETIMEDOUT:
184 case -EDEADLK:
185 case -EAGAIN:
186 if (mode == DLM_LOCK_IV)
187 return 1;
188 break;
189 }
190 return 0;
191 }
192
193 /* we could possibly check if the cancel of an orphan has resulted in the lkb
194 being removed and then remove that lkb from the orphans list and free it */
195
dlm_user_add_ast(struct dlm_lkb * lkb,uint32_t flags,int mode,int status,uint32_t sbflags)196 void dlm_user_add_ast(struct dlm_lkb *lkb, uint32_t flags, int mode,
197 int status, uint32_t sbflags)
198 {
199 struct dlm_ls *ls;
200 struct dlm_user_args *ua;
201 struct dlm_user_proc *proc;
202 int rv;
203
204 if (test_bit(DLM_DFL_ORPHAN_BIT, &lkb->lkb_dflags) ||
205 test_bit(DLM_IFL_DEAD_BIT, &lkb->lkb_iflags))
206 return;
207
208 ls = lkb->lkb_resource->res_ls;
209 spin_lock(&ls->ls_clear_proc_locks);
210
211 /* If ORPHAN/DEAD flag is set, it means the process is dead so an ast
212 can't be delivered. For ORPHAN's, dlm_clear_proc_locks() freed
213 lkb->ua so we can't try to use it. This second check is necessary
214 for cases where a completion ast is received for an operation that
215 began before clear_proc_locks did its cancel/unlock. */
216
217 if (test_bit(DLM_DFL_ORPHAN_BIT, &lkb->lkb_dflags) ||
218 test_bit(DLM_IFL_DEAD_BIT, &lkb->lkb_iflags))
219 goto out;
220
221 DLM_ASSERT(lkb->lkb_ua, dlm_print_lkb(lkb););
222 ua = lkb->lkb_ua;
223 proc = ua->proc;
224
225 if ((flags & DLM_CB_BAST) && ua->bastaddr == NULL)
226 goto out;
227
228 if ((flags & DLM_CB_CAST) && lkb_is_endoflife(mode, status))
229 set_bit(DLM_IFL_ENDOFLIFE_BIT, &lkb->lkb_iflags);
230
231 spin_lock(&proc->asts_spin);
232
233 rv = dlm_enqueue_lkb_callback(lkb, flags, mode, status, sbflags);
234 switch (rv) {
235 case DLM_ENQUEUE_CALLBACK_FAILURE:
236 spin_unlock(&proc->asts_spin);
237 WARN_ON_ONCE(1);
238 goto out;
239 case DLM_ENQUEUE_CALLBACK_NEED_SCHED:
240 kref_get(&lkb->lkb_ref);
241 list_add_tail(&lkb->lkb_cb_list, &proc->asts);
242 wake_up_interruptible(&proc->wait);
243 break;
244 case DLM_ENQUEUE_CALLBACK_SUCCESS:
245 break;
246 default:
247 WARN_ON_ONCE(1);
248 break;
249 }
250 spin_unlock(&proc->asts_spin);
251
252 if (test_bit(DLM_IFL_ENDOFLIFE_BIT, &lkb->lkb_iflags)) {
253 /* N.B. spin_lock locks_spin, not asts_spin */
254 spin_lock(&proc->locks_spin);
255 if (!list_empty(&lkb->lkb_ownqueue)) {
256 list_del_init(&lkb->lkb_ownqueue);
257 dlm_put_lkb(lkb);
258 }
259 spin_unlock(&proc->locks_spin);
260 }
261 out:
262 spin_unlock(&ls->ls_clear_proc_locks);
263 }
264
device_user_lock(struct dlm_user_proc * proc,struct dlm_lock_params * params)265 static int device_user_lock(struct dlm_user_proc *proc,
266 struct dlm_lock_params *params)
267 {
268 struct dlm_ls *ls;
269 struct dlm_user_args *ua;
270 uint32_t lkid;
271 int error = -ENOMEM;
272
273 ls = dlm_find_lockspace_local(proc->lockspace);
274 if (!ls)
275 return -ENOENT;
276
277 if (!params->castaddr || !params->lksb) {
278 error = -EINVAL;
279 goto out;
280 }
281
282 ua = kzalloc(sizeof(struct dlm_user_args), GFP_NOFS);
283 if (!ua)
284 goto out;
285 ua->proc = proc;
286 ua->user_lksb = params->lksb;
287 ua->castparam = params->castparam;
288 ua->castaddr = params->castaddr;
289 ua->bastparam = params->bastparam;
290 ua->bastaddr = params->bastaddr;
291 ua->xid = params->xid;
292
293 if (params->flags & DLM_LKF_CONVERT) {
294 error = dlm_user_convert(ls, ua,
295 params->mode, params->flags,
296 params->lkid, params->lvb);
297 } else if (params->flags & DLM_LKF_ORPHAN) {
298 error = dlm_user_adopt_orphan(ls, ua,
299 params->mode, params->flags,
300 params->name, params->namelen,
301 &lkid);
302 if (!error)
303 error = lkid;
304 } else {
305 error = dlm_user_request(ls, ua,
306 params->mode, params->flags,
307 params->name, params->namelen);
308 if (!error)
309 error = ua->lksb.sb_lkid;
310 }
311 out:
312 dlm_put_lockspace(ls);
313 return error;
314 }
315
device_user_unlock(struct dlm_user_proc * proc,struct dlm_lock_params * params)316 static int device_user_unlock(struct dlm_user_proc *proc,
317 struct dlm_lock_params *params)
318 {
319 struct dlm_ls *ls;
320 struct dlm_user_args *ua;
321 int error = -ENOMEM;
322
323 ls = dlm_find_lockspace_local(proc->lockspace);
324 if (!ls)
325 return -ENOENT;
326
327 ua = kzalloc(sizeof(struct dlm_user_args), GFP_NOFS);
328 if (!ua)
329 goto out;
330 ua->proc = proc;
331 ua->user_lksb = params->lksb;
332 ua->castparam = params->castparam;
333 ua->castaddr = params->castaddr;
334
335 if (params->flags & DLM_LKF_CANCEL)
336 error = dlm_user_cancel(ls, ua, params->flags, params->lkid);
337 else
338 error = dlm_user_unlock(ls, ua, params->flags, params->lkid,
339 params->lvb);
340 out:
341 dlm_put_lockspace(ls);
342 return error;
343 }
344
device_user_deadlock(struct dlm_user_proc * proc,struct dlm_lock_params * params)345 static int device_user_deadlock(struct dlm_user_proc *proc,
346 struct dlm_lock_params *params)
347 {
348 struct dlm_ls *ls;
349 int error;
350
351 ls = dlm_find_lockspace_local(proc->lockspace);
352 if (!ls)
353 return -ENOENT;
354
355 error = dlm_user_deadlock(ls, params->flags, params->lkid);
356
357 dlm_put_lockspace(ls);
358 return error;
359 }
360
dlm_device_register(struct dlm_ls * ls,char * name)361 static int dlm_device_register(struct dlm_ls *ls, char *name)
362 {
363 int error, len;
364
365 /* The device is already registered. This happens when the
366 lockspace is created multiple times from userspace. */
367 if (ls->ls_device.name)
368 return 0;
369
370 error = -ENOMEM;
371 len = strlen(name) + strlen(name_prefix) + 2;
372 ls->ls_device.name = kzalloc(len, GFP_NOFS);
373 if (!ls->ls_device.name)
374 goto fail;
375
376 snprintf((char *)ls->ls_device.name, len, "%s_%s", name_prefix,
377 name);
378 ls->ls_device.fops = &device_fops;
379 ls->ls_device.minor = MISC_DYNAMIC_MINOR;
380
381 error = misc_register(&ls->ls_device);
382 if (error) {
383 kfree(ls->ls_device.name);
384 /* this has to be set to NULL
385 * to avoid a double-free in dlm_device_deregister
386 */
387 ls->ls_device.name = NULL;
388 }
389 fail:
390 return error;
391 }
392
dlm_device_deregister(struct dlm_ls * ls)393 int dlm_device_deregister(struct dlm_ls *ls)
394 {
395 /* The device is not registered. This happens when the lockspace
396 was never used from userspace, or when device_create_lockspace()
397 calls dlm_release_lockspace() after the register fails. */
398 if (!ls->ls_device.name)
399 return 0;
400
401 misc_deregister(&ls->ls_device);
402 kfree(ls->ls_device.name);
403 return 0;
404 }
405
device_user_purge(struct dlm_user_proc * proc,struct dlm_purge_params * params)406 static int device_user_purge(struct dlm_user_proc *proc,
407 struct dlm_purge_params *params)
408 {
409 struct dlm_ls *ls;
410 int error;
411
412 ls = dlm_find_lockspace_local(proc->lockspace);
413 if (!ls)
414 return -ENOENT;
415
416 error = dlm_user_purge(ls, proc, params->nodeid, params->pid);
417
418 dlm_put_lockspace(ls);
419 return error;
420 }
421
device_create_lockspace(struct dlm_lspace_params * params)422 static int device_create_lockspace(struct dlm_lspace_params *params)
423 {
424 dlm_lockspace_t *lockspace;
425 struct dlm_ls *ls;
426 int error;
427
428 if (!capable(CAP_SYS_ADMIN))
429 return -EPERM;
430
431 error = dlm_new_user_lockspace(params->name, dlm_config.ci_cluster_name,
432 params->flags, DLM_USER_LVB_LEN, NULL,
433 NULL, NULL, &lockspace);
434 if (error)
435 return error;
436
437 ls = dlm_find_lockspace_local(lockspace);
438 if (!ls)
439 return -ENOENT;
440
441 error = dlm_device_register(ls, params->name);
442 dlm_put_lockspace(ls);
443
444 if (error)
445 dlm_release_lockspace(lockspace, 0);
446 else
447 error = ls->ls_device.minor;
448
449 return error;
450 }
451
device_remove_lockspace(struct dlm_lspace_params * params)452 static int device_remove_lockspace(struct dlm_lspace_params *params)
453 {
454 dlm_lockspace_t *lockspace;
455 struct dlm_ls *ls;
456 int error, force = 0;
457
458 if (!capable(CAP_SYS_ADMIN))
459 return -EPERM;
460
461 ls = dlm_find_lockspace_device(params->minor);
462 if (!ls)
463 return -ENOENT;
464
465 if (params->flags & DLM_USER_LSFLG_FORCEFREE)
466 force = 2;
467
468 lockspace = ls->ls_local_handle;
469 dlm_put_lockspace(ls);
470
471 /* The final dlm_release_lockspace waits for references to go to
472 zero, so all processes will need to close their device for the
473 ls before the release will proceed. release also calls the
474 device_deregister above. Converting a positive return value
475 from release to zero means that userspace won't know when its
476 release was the final one, but it shouldn't need to know. */
477
478 error = dlm_release_lockspace(lockspace, force);
479 if (error > 0)
480 error = 0;
481 return error;
482 }
483
484 /* Check the user's version matches ours */
check_version(struct dlm_write_request * req)485 static int check_version(struct dlm_write_request *req)
486 {
487 if (req->version[0] != DLM_DEVICE_VERSION_MAJOR ||
488 (req->version[0] == DLM_DEVICE_VERSION_MAJOR &&
489 req->version[1] > DLM_DEVICE_VERSION_MINOR)) {
490
491 printk(KERN_DEBUG "dlm: process %s (%d) version mismatch "
492 "user (%d.%d.%d) kernel (%d.%d.%d)\n",
493 current->comm,
494 task_pid_nr(current),
495 req->version[0],
496 req->version[1],
497 req->version[2],
498 DLM_DEVICE_VERSION_MAJOR,
499 DLM_DEVICE_VERSION_MINOR,
500 DLM_DEVICE_VERSION_PATCH);
501 return -EINVAL;
502 }
503 return 0;
504 }
505
506 /*
507 * device_write
508 *
509 * device_user_lock
510 * dlm_user_request -> request_lock
511 * dlm_user_convert -> convert_lock
512 *
513 * device_user_unlock
514 * dlm_user_unlock -> unlock_lock
515 * dlm_user_cancel -> cancel_lock
516 *
517 * device_create_lockspace
518 * dlm_new_lockspace
519 *
520 * device_remove_lockspace
521 * dlm_release_lockspace
522 */
523
524 /* a write to a lockspace device is a lock or unlock request, a write
525 to the control device is to create/remove a lockspace */
526
device_write(struct file * file,const char __user * buf,size_t count,loff_t * ppos)527 static ssize_t device_write(struct file *file, const char __user *buf,
528 size_t count, loff_t *ppos)
529 {
530 struct dlm_user_proc *proc = file->private_data;
531 struct dlm_write_request *kbuf;
532 int error;
533
534 #ifdef CONFIG_COMPAT
535 if (count < sizeof(struct dlm_write_request32))
536 #else
537 if (count < sizeof(struct dlm_write_request))
538 #endif
539 return -EINVAL;
540
541 /*
542 * can't compare against COMPAT/dlm_write_request32 because
543 * we don't yet know if is64bit is zero
544 */
545 if (count > sizeof(struct dlm_write_request) + DLM_RESNAME_MAXLEN)
546 return -EINVAL;
547
548 kbuf = memdup_user_nul(buf, count);
549 if (IS_ERR(kbuf))
550 return PTR_ERR(kbuf);
551
552 if (check_version(kbuf)) {
553 error = -EBADE;
554 goto out_free;
555 }
556
557 #ifdef CONFIG_COMPAT
558 if (!kbuf->is64bit) {
559 struct dlm_write_request32 *k32buf;
560 int namelen = 0;
561
562 if (count > sizeof(struct dlm_write_request32))
563 namelen = count - sizeof(struct dlm_write_request32);
564
565 k32buf = (struct dlm_write_request32 *)kbuf;
566
567 /* add 1 after namelen so that the name string is terminated */
568 kbuf = kzalloc(sizeof(struct dlm_write_request) + namelen + 1,
569 GFP_NOFS);
570 if (!kbuf) {
571 kfree(k32buf);
572 return -ENOMEM;
573 }
574
575 if (proc)
576 set_bit(DLM_PROC_FLAGS_COMPAT, &proc->flags);
577
578 compat_input(kbuf, k32buf, namelen);
579 kfree(k32buf);
580 }
581 #endif
582
583 /* do we really need this? can a write happen after a close? */
584 if ((kbuf->cmd == DLM_USER_LOCK || kbuf->cmd == DLM_USER_UNLOCK) &&
585 (proc && test_bit(DLM_PROC_FLAGS_CLOSING, &proc->flags))) {
586 error = -EINVAL;
587 goto out_free;
588 }
589
590 error = -EINVAL;
591
592 switch (kbuf->cmd)
593 {
594 case DLM_USER_LOCK:
595 if (!proc) {
596 log_print("no locking on control device");
597 goto out_free;
598 }
599 error = device_user_lock(proc, &kbuf->i.lock);
600 break;
601
602 case DLM_USER_UNLOCK:
603 if (!proc) {
604 log_print("no locking on control device");
605 goto out_free;
606 }
607 error = device_user_unlock(proc, &kbuf->i.lock);
608 break;
609
610 case DLM_USER_DEADLOCK:
611 if (!proc) {
612 log_print("no locking on control device");
613 goto out_free;
614 }
615 error = device_user_deadlock(proc, &kbuf->i.lock);
616 break;
617
618 case DLM_USER_CREATE_LOCKSPACE:
619 if (proc) {
620 log_print("create/remove only on control device");
621 goto out_free;
622 }
623 error = device_create_lockspace(&kbuf->i.lspace);
624 break;
625
626 case DLM_USER_REMOVE_LOCKSPACE:
627 if (proc) {
628 log_print("create/remove only on control device");
629 goto out_free;
630 }
631 error = device_remove_lockspace(&kbuf->i.lspace);
632 break;
633
634 case DLM_USER_PURGE:
635 if (!proc) {
636 log_print("no locking on control device");
637 goto out_free;
638 }
639 error = device_user_purge(proc, &kbuf->i.purge);
640 break;
641
642 default:
643 log_print("Unknown command passed to DLM device : %d\n",
644 kbuf->cmd);
645 }
646
647 out_free:
648 kfree(kbuf);
649 return error;
650 }
651
652 /* Every process that opens the lockspace device has its own "proc" structure
653 hanging off the open file that's used to keep track of locks owned by the
654 process and asts that need to be delivered to the process. */
655
device_open(struct inode * inode,struct file * file)656 static int device_open(struct inode *inode, struct file *file)
657 {
658 struct dlm_user_proc *proc;
659 struct dlm_ls *ls;
660
661 ls = dlm_find_lockspace_device(iminor(inode));
662 if (!ls)
663 return -ENOENT;
664
665 proc = kzalloc(sizeof(struct dlm_user_proc), GFP_NOFS);
666 if (!proc) {
667 dlm_put_lockspace(ls);
668 return -ENOMEM;
669 }
670
671 proc->lockspace = ls->ls_local_handle;
672 INIT_LIST_HEAD(&proc->asts);
673 INIT_LIST_HEAD(&proc->locks);
674 INIT_LIST_HEAD(&proc->unlocking);
675 spin_lock_init(&proc->asts_spin);
676 spin_lock_init(&proc->locks_spin);
677 init_waitqueue_head(&proc->wait);
678 file->private_data = proc;
679
680 return 0;
681 }
682
device_close(struct inode * inode,struct file * file)683 static int device_close(struct inode *inode, struct file *file)
684 {
685 struct dlm_user_proc *proc = file->private_data;
686 struct dlm_ls *ls;
687
688 ls = dlm_find_lockspace_local(proc->lockspace);
689 if (!ls)
690 return -ENOENT;
691
692 set_bit(DLM_PROC_FLAGS_CLOSING, &proc->flags);
693
694 dlm_clear_proc_locks(ls, proc);
695
696 /* at this point no more lkb's should exist for this lockspace,
697 so there's no chance of dlm_user_add_ast() being called and
698 looking for lkb->ua->proc */
699
700 kfree(proc);
701 file->private_data = NULL;
702
703 dlm_put_lockspace(ls);
704 dlm_put_lockspace(ls); /* for the find in device_open() */
705
706 /* FIXME: AUTOFREE: if this ls is no longer used do
707 device_remove_lockspace() */
708
709 return 0;
710 }
711
copy_result_to_user(struct dlm_user_args * ua,int compat,uint32_t flags,int mode,int copy_lvb,char __user * buf,size_t count)712 static int copy_result_to_user(struct dlm_user_args *ua, int compat,
713 uint32_t flags, int mode, int copy_lvb,
714 char __user *buf, size_t count)
715 {
716 #ifdef CONFIG_COMPAT
717 struct dlm_lock_result32 result32;
718 #endif
719 struct dlm_lock_result result;
720 void *resultptr;
721 int error=0;
722 int len;
723 int struct_len;
724
725 memset(&result, 0, sizeof(struct dlm_lock_result));
726 result.version[0] = DLM_DEVICE_VERSION_MAJOR;
727 result.version[1] = DLM_DEVICE_VERSION_MINOR;
728 result.version[2] = DLM_DEVICE_VERSION_PATCH;
729 memcpy(&result.lksb, &ua->lksb, offsetof(struct dlm_lksb, sb_lvbptr));
730 result.user_lksb = ua->user_lksb;
731
732 /* FIXME: dlm1 provides for the user's bastparam/addr to not be updated
733 in a conversion unless the conversion is successful. See code
734 in dlm_user_convert() for updating ua from ua_tmp. OpenVMS, though,
735 notes that a new blocking AST address and parameter are set even if
736 the conversion fails, so maybe we should just do that. */
737
738 if (flags & DLM_CB_BAST) {
739 result.user_astaddr = ua->bastaddr;
740 result.user_astparam = ua->bastparam;
741 result.bast_mode = mode;
742 } else {
743 result.user_astaddr = ua->castaddr;
744 result.user_astparam = ua->castparam;
745 }
746
747 #ifdef CONFIG_COMPAT
748 if (compat)
749 len = sizeof(struct dlm_lock_result32);
750 else
751 #endif
752 len = sizeof(struct dlm_lock_result);
753 struct_len = len;
754
755 /* copy lvb to userspace if there is one, it's been updated, and
756 the user buffer has space for it */
757
758 if (copy_lvb && ua->lksb.sb_lvbptr && count >= len + DLM_USER_LVB_LEN) {
759 if (copy_to_user(buf+len, ua->lksb.sb_lvbptr,
760 DLM_USER_LVB_LEN)) {
761 error = -EFAULT;
762 goto out;
763 }
764
765 result.lvb_offset = len;
766 len += DLM_USER_LVB_LEN;
767 }
768
769 result.length = len;
770 resultptr = &result;
771 #ifdef CONFIG_COMPAT
772 if (compat) {
773 compat_output(&result, &result32);
774 resultptr = &result32;
775 }
776 #endif
777
778 if (copy_to_user(buf, resultptr, struct_len))
779 error = -EFAULT;
780 else
781 error = len;
782 out:
783 return error;
784 }
785
copy_version_to_user(char __user * buf,size_t count)786 static int copy_version_to_user(char __user *buf, size_t count)
787 {
788 struct dlm_device_version ver;
789
790 memset(&ver, 0, sizeof(struct dlm_device_version));
791 ver.version[0] = DLM_DEVICE_VERSION_MAJOR;
792 ver.version[1] = DLM_DEVICE_VERSION_MINOR;
793 ver.version[2] = DLM_DEVICE_VERSION_PATCH;
794
795 if (copy_to_user(buf, &ver, sizeof(struct dlm_device_version)))
796 return -EFAULT;
797 return sizeof(struct dlm_device_version);
798 }
799
800 /* a read returns a single ast described in a struct dlm_lock_result */
801
device_read(struct file * file,char __user * buf,size_t count,loff_t * ppos)802 static ssize_t device_read(struct file *file, char __user *buf, size_t count,
803 loff_t *ppos)
804 {
805 struct dlm_user_proc *proc = file->private_data;
806 struct dlm_lkb *lkb;
807 DECLARE_WAITQUEUE(wait, current);
808 struct dlm_callback *cb;
809 int rv, copy_lvb = 0;
810 int old_mode, new_mode;
811
812 if (count == sizeof(struct dlm_device_version)) {
813 rv = copy_version_to_user(buf, count);
814 return rv;
815 }
816
817 if (!proc) {
818 log_print("non-version read from control device %zu", count);
819 return -EINVAL;
820 }
821
822 #ifdef CONFIG_COMPAT
823 if (count < sizeof(struct dlm_lock_result32))
824 #else
825 if (count < sizeof(struct dlm_lock_result))
826 #endif
827 return -EINVAL;
828
829 try_another:
830
831 /* do we really need this? can a read happen after a close? */
832 if (test_bit(DLM_PROC_FLAGS_CLOSING, &proc->flags))
833 return -EINVAL;
834
835 spin_lock(&proc->asts_spin);
836 if (list_empty(&proc->asts)) {
837 if (file->f_flags & O_NONBLOCK) {
838 spin_unlock(&proc->asts_spin);
839 return -EAGAIN;
840 }
841
842 add_wait_queue(&proc->wait, &wait);
843
844 repeat:
845 set_current_state(TASK_INTERRUPTIBLE);
846 if (list_empty(&proc->asts) && !signal_pending(current)) {
847 spin_unlock(&proc->asts_spin);
848 schedule();
849 spin_lock(&proc->asts_spin);
850 goto repeat;
851 }
852 set_current_state(TASK_RUNNING);
853 remove_wait_queue(&proc->wait, &wait);
854
855 if (signal_pending(current)) {
856 spin_unlock(&proc->asts_spin);
857 return -ERESTARTSYS;
858 }
859 }
860
861 /* if we empty lkb_callbacks, we don't want to unlock the spinlock
862 without removing lkb_cb_list; so empty lkb_cb_list is always
863 consistent with empty lkb_callbacks */
864
865 lkb = list_first_entry(&proc->asts, struct dlm_lkb, lkb_cb_list);
866
867 /* rem_lkb_callback sets a new lkb_last_cast */
868 old_mode = lkb->lkb_last_cast->mode;
869
870 rv = dlm_dequeue_lkb_callback(lkb, &cb);
871 switch (rv) {
872 case DLM_DEQUEUE_CALLBACK_EMPTY:
873 /* this shouldn't happen; lkb should have been removed from
874 * list when last item was dequeued
875 */
876 log_print("dlm_rem_lkb_callback empty %x", lkb->lkb_id);
877 list_del_init(&lkb->lkb_cb_list);
878 spin_unlock(&proc->asts_spin);
879 /* removes ref for proc->asts, may cause lkb to be freed */
880 dlm_put_lkb(lkb);
881 WARN_ON_ONCE(1);
882 goto try_another;
883 case DLM_DEQUEUE_CALLBACK_LAST:
884 list_del_init(&lkb->lkb_cb_list);
885 clear_bit(DLM_IFL_CB_PENDING_BIT, &lkb->lkb_iflags);
886 break;
887 case DLM_DEQUEUE_CALLBACK_SUCCESS:
888 break;
889 default:
890 WARN_ON_ONCE(1);
891 break;
892 }
893 spin_unlock(&proc->asts_spin);
894
895 if (cb->flags & DLM_CB_BAST) {
896 trace_dlm_bast(lkb->lkb_resource->res_ls, lkb, cb->mode);
897 } else if (cb->flags & DLM_CB_CAST) {
898 new_mode = cb->mode;
899
900 if (!cb->sb_status && lkb->lkb_lksb->sb_lvbptr &&
901 dlm_lvb_operations[old_mode + 1][new_mode + 1])
902 copy_lvb = 1;
903
904 lkb->lkb_lksb->sb_status = cb->sb_status;
905 lkb->lkb_lksb->sb_flags = cb->sb_flags;
906 trace_dlm_ast(lkb->lkb_resource->res_ls, lkb);
907 }
908
909 rv = copy_result_to_user(lkb->lkb_ua,
910 test_bit(DLM_PROC_FLAGS_COMPAT, &proc->flags),
911 cb->flags, cb->mode, copy_lvb, buf, count);
912
913 kref_put(&cb->ref, dlm_release_callback);
914
915 /* removes ref for proc->asts, may cause lkb to be freed */
916 if (rv == DLM_DEQUEUE_CALLBACK_LAST)
917 dlm_put_lkb(lkb);
918
919 return rv;
920 }
921
device_poll(struct file * file,poll_table * wait)922 static __poll_t device_poll(struct file *file, poll_table *wait)
923 {
924 struct dlm_user_proc *proc = file->private_data;
925
926 poll_wait(file, &proc->wait, wait);
927
928 spin_lock(&proc->asts_spin);
929 if (!list_empty(&proc->asts)) {
930 spin_unlock(&proc->asts_spin);
931 return EPOLLIN | EPOLLRDNORM;
932 }
933 spin_unlock(&proc->asts_spin);
934 return 0;
935 }
936
dlm_user_daemon_available(void)937 int dlm_user_daemon_available(void)
938 {
939 /* dlm_controld hasn't started (or, has started, but not
940 properly populated configfs) */
941
942 if (!dlm_our_nodeid())
943 return 0;
944
945 /* This is to deal with versions of dlm_controld that don't
946 know about the monitor device. We assume that if the
947 dlm_controld was started (above), but the monitor device
948 was never opened, that it's an old version. dlm_controld
949 should open the monitor device before populating configfs. */
950
951 if (dlm_monitor_unused)
952 return 1;
953
954 return atomic_read(&dlm_monitor_opened) ? 1 : 0;
955 }
956
ctl_device_open(struct inode * inode,struct file * file)957 static int ctl_device_open(struct inode *inode, struct file *file)
958 {
959 file->private_data = NULL;
960 return 0;
961 }
962
ctl_device_close(struct inode * inode,struct file * file)963 static int ctl_device_close(struct inode *inode, struct file *file)
964 {
965 return 0;
966 }
967
monitor_device_open(struct inode * inode,struct file * file)968 static int monitor_device_open(struct inode *inode, struct file *file)
969 {
970 atomic_inc(&dlm_monitor_opened);
971 dlm_monitor_unused = 0;
972 return 0;
973 }
974
monitor_device_close(struct inode * inode,struct file * file)975 static int monitor_device_close(struct inode *inode, struct file *file)
976 {
977 if (atomic_dec_and_test(&dlm_monitor_opened))
978 dlm_stop_lockspaces();
979 return 0;
980 }
981
982 static const struct file_operations device_fops = {
983 .open = device_open,
984 .release = device_close,
985 .read = device_read,
986 .write = device_write,
987 .poll = device_poll,
988 .owner = THIS_MODULE,
989 .llseek = noop_llseek,
990 };
991
992 static const struct file_operations ctl_device_fops = {
993 .open = ctl_device_open,
994 .release = ctl_device_close,
995 .read = device_read,
996 .write = device_write,
997 .owner = THIS_MODULE,
998 .llseek = noop_llseek,
999 };
1000
1001 static struct miscdevice ctl_device = {
1002 .name = "dlm-control",
1003 .fops = &ctl_device_fops,
1004 .minor = MISC_DYNAMIC_MINOR,
1005 };
1006
1007 static const struct file_operations monitor_device_fops = {
1008 .open = monitor_device_open,
1009 .release = monitor_device_close,
1010 .owner = THIS_MODULE,
1011 .llseek = noop_llseek,
1012 };
1013
1014 static struct miscdevice monitor_device = {
1015 .name = "dlm-monitor",
1016 .fops = &monitor_device_fops,
1017 .minor = MISC_DYNAMIC_MINOR,
1018 };
1019
dlm_user_init(void)1020 int __init dlm_user_init(void)
1021 {
1022 int error;
1023
1024 atomic_set(&dlm_monitor_opened, 0);
1025
1026 error = misc_register(&ctl_device);
1027 if (error) {
1028 log_print("misc_register failed for control device");
1029 goto out;
1030 }
1031
1032 error = misc_register(&monitor_device);
1033 if (error) {
1034 log_print("misc_register failed for monitor device");
1035 misc_deregister(&ctl_device);
1036 }
1037 out:
1038 return error;
1039 }
1040
dlm_user_exit(void)1041 void dlm_user_exit(void)
1042 {
1043 misc_deregister(&ctl_device);
1044 misc_deregister(&monitor_device);
1045 }
1046
1047