1 /**************************************************************************
2 *
3 * Copyright (c) 2007-2009 VMware, Inc., Palo Alto, CA., USA
4 * All Rights Reserved.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the
8 * "Software"), to deal in the Software without restriction, including
9 * without limitation the rights to use, copy, modify, merge, publish,
10 * distribute, sub license, and/or sell copies of the Software, and to
11 * permit persons to whom the Software is furnished to do so, subject to
12 * the following conditions:
13 *
14 * The above copyright notice and this permission notice (including the
15 * next paragraph) shall be included in all copies or substantial portions
16 * of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
22 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24 * USE OR OTHER DEALINGS IN THE SOFTWARE.
25 *
26 **************************************************************************/
27 /*
28 * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com>
29 */
30
31 #include "ttm/ttm_lock.h"
32 #include "ttm/ttm_module.h"
33 #include <linux/atomic.h>
34 #include <linux/errno.h>
35 #include <linux/wait.h>
36 #include <linux/sched.h>
37 #include <linux/module.h>
38
39 #define TTM_WRITE_LOCK_PENDING (1 << 0)
40 #define TTM_VT_LOCK_PENDING (1 << 1)
41 #define TTM_SUSPEND_LOCK_PENDING (1 << 2)
42 #define TTM_VT_LOCK (1 << 3)
43 #define TTM_SUSPEND_LOCK (1 << 4)
44
ttm_lock_init(struct ttm_lock * lock)45 void ttm_lock_init(struct ttm_lock *lock)
46 {
47 spin_lock_init(&lock->lock);
48 init_waitqueue_head(&lock->queue);
49 lock->rw = 0;
50 lock->flags = 0;
51 lock->kill_takers = false;
52 lock->signal = SIGKILL;
53 }
54 EXPORT_SYMBOL(ttm_lock_init);
55
ttm_read_unlock(struct ttm_lock * lock)56 void ttm_read_unlock(struct ttm_lock *lock)
57 {
58 spin_lock(&lock->lock);
59 if (--lock->rw == 0)
60 wake_up_all(&lock->queue);
61 spin_unlock(&lock->lock);
62 }
63 EXPORT_SYMBOL(ttm_read_unlock);
64
__ttm_read_lock(struct ttm_lock * lock)65 static bool __ttm_read_lock(struct ttm_lock *lock)
66 {
67 bool locked = false;
68
69 spin_lock(&lock->lock);
70 if (unlikely(lock->kill_takers)) {
71 send_sig(lock->signal, current, 0);
72 spin_unlock(&lock->lock);
73 return false;
74 }
75 if (lock->rw >= 0 && lock->flags == 0) {
76 ++lock->rw;
77 locked = true;
78 }
79 spin_unlock(&lock->lock);
80 return locked;
81 }
82
ttm_read_lock(struct ttm_lock * lock,bool interruptible)83 int ttm_read_lock(struct ttm_lock *lock, bool interruptible)
84 {
85 int ret = 0;
86
87 if (interruptible)
88 ret = wait_event_interruptible(lock->queue,
89 __ttm_read_lock(lock));
90 else
91 wait_event(lock->queue, __ttm_read_lock(lock));
92 return ret;
93 }
94 EXPORT_SYMBOL(ttm_read_lock);
95
__ttm_read_trylock(struct ttm_lock * lock,bool * locked)96 static bool __ttm_read_trylock(struct ttm_lock *lock, bool *locked)
97 {
98 bool block = true;
99
100 *locked = false;
101
102 spin_lock(&lock->lock);
103 if (unlikely(lock->kill_takers)) {
104 send_sig(lock->signal, current, 0);
105 spin_unlock(&lock->lock);
106 return false;
107 }
108 if (lock->rw >= 0 && lock->flags == 0) {
109 ++lock->rw;
110 block = false;
111 *locked = true;
112 } else if (lock->flags == 0) {
113 block = false;
114 }
115 spin_unlock(&lock->lock);
116
117 return !block;
118 }
119
ttm_read_trylock(struct ttm_lock * lock,bool interruptible)120 int ttm_read_trylock(struct ttm_lock *lock, bool interruptible)
121 {
122 int ret = 0;
123 bool locked;
124
125 if (interruptible)
126 ret = wait_event_interruptible
127 (lock->queue, __ttm_read_trylock(lock, &locked));
128 else
129 wait_event(lock->queue, __ttm_read_trylock(lock, &locked));
130
131 if (unlikely(ret != 0)) {
132 BUG_ON(locked);
133 return ret;
134 }
135
136 return (locked) ? 0 : -EBUSY;
137 }
138
ttm_write_unlock(struct ttm_lock * lock)139 void ttm_write_unlock(struct ttm_lock *lock)
140 {
141 spin_lock(&lock->lock);
142 lock->rw = 0;
143 wake_up_all(&lock->queue);
144 spin_unlock(&lock->lock);
145 }
146 EXPORT_SYMBOL(ttm_write_unlock);
147
__ttm_write_lock(struct ttm_lock * lock)148 static bool __ttm_write_lock(struct ttm_lock *lock)
149 {
150 bool locked = false;
151
152 spin_lock(&lock->lock);
153 if (unlikely(lock->kill_takers)) {
154 send_sig(lock->signal, current, 0);
155 spin_unlock(&lock->lock);
156 return false;
157 }
158 if (lock->rw == 0 && ((lock->flags & ~TTM_WRITE_LOCK_PENDING) == 0)) {
159 lock->rw = -1;
160 lock->flags &= ~TTM_WRITE_LOCK_PENDING;
161 locked = true;
162 } else {
163 lock->flags |= TTM_WRITE_LOCK_PENDING;
164 }
165 spin_unlock(&lock->lock);
166 return locked;
167 }
168
ttm_write_lock(struct ttm_lock * lock,bool interruptible)169 int ttm_write_lock(struct ttm_lock *lock, bool interruptible)
170 {
171 int ret = 0;
172
173 if (interruptible) {
174 ret = wait_event_interruptible(lock->queue,
175 __ttm_write_lock(lock));
176 if (unlikely(ret != 0)) {
177 spin_lock(&lock->lock);
178 lock->flags &= ~TTM_WRITE_LOCK_PENDING;
179 wake_up_all(&lock->queue);
180 spin_unlock(&lock->lock);
181 }
182 } else
183 wait_event(lock->queue, __ttm_read_lock(lock));
184
185 return ret;
186 }
187 EXPORT_SYMBOL(ttm_write_lock);
188
ttm_write_lock_downgrade(struct ttm_lock * lock)189 void ttm_write_lock_downgrade(struct ttm_lock *lock)
190 {
191 spin_lock(&lock->lock);
192 lock->rw = 1;
193 wake_up_all(&lock->queue);
194 spin_unlock(&lock->lock);
195 }
196
__ttm_vt_unlock(struct ttm_lock * lock)197 static int __ttm_vt_unlock(struct ttm_lock *lock)
198 {
199 int ret = 0;
200
201 spin_lock(&lock->lock);
202 if (unlikely(!(lock->flags & TTM_VT_LOCK)))
203 ret = -EINVAL;
204 lock->flags &= ~TTM_VT_LOCK;
205 wake_up_all(&lock->queue);
206 spin_unlock(&lock->lock);
207
208 return ret;
209 }
210
ttm_vt_lock_remove(struct ttm_base_object ** p_base)211 static void ttm_vt_lock_remove(struct ttm_base_object **p_base)
212 {
213 struct ttm_base_object *base = *p_base;
214 struct ttm_lock *lock = container_of(base, struct ttm_lock, base);
215 int ret;
216
217 *p_base = NULL;
218 ret = __ttm_vt_unlock(lock);
219 BUG_ON(ret != 0);
220 }
221
__ttm_vt_lock(struct ttm_lock * lock)222 static bool __ttm_vt_lock(struct ttm_lock *lock)
223 {
224 bool locked = false;
225
226 spin_lock(&lock->lock);
227 if (lock->rw == 0) {
228 lock->flags &= ~TTM_VT_LOCK_PENDING;
229 lock->flags |= TTM_VT_LOCK;
230 locked = true;
231 } else {
232 lock->flags |= TTM_VT_LOCK_PENDING;
233 }
234 spin_unlock(&lock->lock);
235 return locked;
236 }
237
ttm_vt_lock(struct ttm_lock * lock,bool interruptible,struct ttm_object_file * tfile)238 int ttm_vt_lock(struct ttm_lock *lock,
239 bool interruptible,
240 struct ttm_object_file *tfile)
241 {
242 int ret = 0;
243
244 if (interruptible) {
245 ret = wait_event_interruptible(lock->queue,
246 __ttm_vt_lock(lock));
247 if (unlikely(ret != 0)) {
248 spin_lock(&lock->lock);
249 lock->flags &= ~TTM_VT_LOCK_PENDING;
250 wake_up_all(&lock->queue);
251 spin_unlock(&lock->lock);
252 return ret;
253 }
254 } else
255 wait_event(lock->queue, __ttm_vt_lock(lock));
256
257 /*
258 * Add a base-object, the destructor of which will
259 * make sure the lock is released if the client dies
260 * while holding it.
261 */
262
263 ret = ttm_base_object_init(tfile, &lock->base, false,
264 ttm_lock_type, &ttm_vt_lock_remove, NULL);
265 if (ret)
266 (void)__ttm_vt_unlock(lock);
267 else
268 lock->vt_holder = tfile;
269
270 return ret;
271 }
272 EXPORT_SYMBOL(ttm_vt_lock);
273
ttm_vt_unlock(struct ttm_lock * lock)274 int ttm_vt_unlock(struct ttm_lock *lock)
275 {
276 return ttm_ref_object_base_unref(lock->vt_holder,
277 lock->base.hash.key, TTM_REF_USAGE);
278 }
279 EXPORT_SYMBOL(ttm_vt_unlock);
280
ttm_suspend_unlock(struct ttm_lock * lock)281 void ttm_suspend_unlock(struct ttm_lock *lock)
282 {
283 spin_lock(&lock->lock);
284 lock->flags &= ~TTM_SUSPEND_LOCK;
285 wake_up_all(&lock->queue);
286 spin_unlock(&lock->lock);
287 }
288 EXPORT_SYMBOL(ttm_suspend_unlock);
289
__ttm_suspend_lock(struct ttm_lock * lock)290 static bool __ttm_suspend_lock(struct ttm_lock *lock)
291 {
292 bool locked = false;
293
294 spin_lock(&lock->lock);
295 if (lock->rw == 0) {
296 lock->flags &= ~TTM_SUSPEND_LOCK_PENDING;
297 lock->flags |= TTM_SUSPEND_LOCK;
298 locked = true;
299 } else {
300 lock->flags |= TTM_SUSPEND_LOCK_PENDING;
301 }
302 spin_unlock(&lock->lock);
303 return locked;
304 }
305
ttm_suspend_lock(struct ttm_lock * lock)306 void ttm_suspend_lock(struct ttm_lock *lock)
307 {
308 wait_event(lock->queue, __ttm_suspend_lock(lock));
309 }
310 EXPORT_SYMBOL(ttm_suspend_lock);
311