1 /*
2 * fs/cifs/misc.c
3 *
4 * Copyright (C) International Business Machines Corp., 2002,2008
5 * Author(s): Steve French (sfrench@us.ibm.com)
6 *
7 * This library is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU Lesser General Public License as published
9 * by the Free Software Foundation; either version 2.1 of the License, or
10 * (at your option) any later version.
11 *
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
15 * the GNU Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public License
18 * along with this library; if not, write to the Free Software
19 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20 */
21
22 #include <linux/slab.h>
23 #include <linux/ctype.h>
24 #include <linux/mempool.h>
25 #include "cifspdu.h"
26 #include "cifsglob.h"
27 #include "cifsproto.h"
28 #include "cifs_debug.h"
29 #include "smberr.h"
30 #include "nterr.h"
31 #include "cifs_unicode.h"
32
33 extern mempool_t *cifs_sm_req_poolp;
34 extern mempool_t *cifs_req_poolp;
35
36 /* The xid serves as a useful identifier for each incoming vfs request,
37 in a similar way to the mid which is useful to track each sent smb,
38 and CurrentXid can also provide a running counter (although it
39 will eventually wrap past zero) of the total vfs operations handled
40 since the cifs fs was mounted */
41
42 unsigned int
_GetXid(void)43 _GetXid(void)
44 {
45 unsigned int xid;
46
47 spin_lock(&GlobalMid_Lock);
48 GlobalTotalActiveXid++;
49
50 /* keep high water mark for number of simultaneous ops in filesystem */
51 if (GlobalTotalActiveXid > GlobalMaxActiveXid)
52 GlobalMaxActiveXid = GlobalTotalActiveXid;
53 if (GlobalTotalActiveXid > 65000)
54 cFYI(1, "warning: more than 65000 requests active");
55 xid = GlobalCurrentXid++;
56 spin_unlock(&GlobalMid_Lock);
57 return xid;
58 }
59
60 void
_FreeXid(unsigned int xid)61 _FreeXid(unsigned int xid)
62 {
63 spin_lock(&GlobalMid_Lock);
64 /* if (GlobalTotalActiveXid == 0)
65 BUG(); */
66 GlobalTotalActiveXid--;
67 spin_unlock(&GlobalMid_Lock);
68 }
69
70 struct cifsSesInfo *
sesInfoAlloc(void)71 sesInfoAlloc(void)
72 {
73 struct cifsSesInfo *ret_buf;
74
75 ret_buf = kzalloc(sizeof(struct cifsSesInfo), GFP_KERNEL);
76 if (ret_buf) {
77 atomic_inc(&sesInfoAllocCount);
78 ret_buf->status = CifsNew;
79 ++ret_buf->ses_count;
80 INIT_LIST_HEAD(&ret_buf->smb_ses_list);
81 INIT_LIST_HEAD(&ret_buf->tcon_list);
82 mutex_init(&ret_buf->session_mutex);
83 }
84 return ret_buf;
85 }
86
87 void
sesInfoFree(struct cifsSesInfo * buf_to_free)88 sesInfoFree(struct cifsSesInfo *buf_to_free)
89 {
90 if (buf_to_free == NULL) {
91 cFYI(1, "Null buffer passed to sesInfoFree");
92 return;
93 }
94
95 atomic_dec(&sesInfoAllocCount);
96 kfree(buf_to_free->serverOS);
97 kfree(buf_to_free->serverDomain);
98 kfree(buf_to_free->serverNOS);
99 if (buf_to_free->password) {
100 memset(buf_to_free->password, 0, strlen(buf_to_free->password));
101 kfree(buf_to_free->password);
102 }
103 kfree(buf_to_free->user_name);
104 kfree(buf_to_free->domainName);
105 kfree(buf_to_free);
106 }
107
108 struct cifsTconInfo *
tconInfoAlloc(void)109 tconInfoAlloc(void)
110 {
111 struct cifsTconInfo *ret_buf;
112 ret_buf = kzalloc(sizeof(struct cifsTconInfo), GFP_KERNEL);
113 if (ret_buf) {
114 atomic_inc(&tconInfoAllocCount);
115 ret_buf->tidStatus = CifsNew;
116 ++ret_buf->tc_count;
117 INIT_LIST_HEAD(&ret_buf->openFileList);
118 INIT_LIST_HEAD(&ret_buf->tcon_list);
119 #ifdef CONFIG_CIFS_STATS
120 spin_lock_init(&ret_buf->stat_lock);
121 #endif
122 }
123 return ret_buf;
124 }
125
126 void
tconInfoFree(struct cifsTconInfo * buf_to_free)127 tconInfoFree(struct cifsTconInfo *buf_to_free)
128 {
129 if (buf_to_free == NULL) {
130 cFYI(1, "Null buffer passed to tconInfoFree");
131 return;
132 }
133 atomic_dec(&tconInfoAllocCount);
134 kfree(buf_to_free->nativeFileSystem);
135 if (buf_to_free->password) {
136 memset(buf_to_free->password, 0, strlen(buf_to_free->password));
137 kfree(buf_to_free->password);
138 }
139 kfree(buf_to_free);
140 }
141
142 struct smb_hdr *
cifs_buf_get(void)143 cifs_buf_get(void)
144 {
145 struct smb_hdr *ret_buf = NULL;
146
147 /* We could use negotiated size instead of max_msgsize -
148 but it may be more efficient to always alloc same size
149 albeit slightly larger than necessary and maxbuffersize
150 defaults to this and can not be bigger */
151 ret_buf = mempool_alloc(cifs_req_poolp, GFP_NOFS);
152
153 /* clear the first few header bytes */
154 /* for most paths, more is cleared in header_assemble */
155 if (ret_buf) {
156 memset(ret_buf, 0, sizeof(struct smb_hdr) + 3);
157 atomic_inc(&bufAllocCount);
158 #ifdef CONFIG_CIFS_STATS2
159 atomic_inc(&totBufAllocCount);
160 #endif /* CONFIG_CIFS_STATS2 */
161 }
162
163 return ret_buf;
164 }
165
166 void
cifs_buf_release(void * buf_to_free)167 cifs_buf_release(void *buf_to_free)
168 {
169 if (buf_to_free == NULL) {
170 /* cFYI(1, "Null buffer passed to cifs_buf_release");*/
171 return;
172 }
173 mempool_free(buf_to_free, cifs_req_poolp);
174
175 atomic_dec(&bufAllocCount);
176 return;
177 }
178
179 struct smb_hdr *
cifs_small_buf_get(void)180 cifs_small_buf_get(void)
181 {
182 struct smb_hdr *ret_buf = NULL;
183
184 /* We could use negotiated size instead of max_msgsize -
185 but it may be more efficient to always alloc same size
186 albeit slightly larger than necessary and maxbuffersize
187 defaults to this and can not be bigger */
188 ret_buf = mempool_alloc(cifs_sm_req_poolp, GFP_NOFS);
189 if (ret_buf) {
190 /* No need to clear memory here, cleared in header assemble */
191 /* memset(ret_buf, 0, sizeof(struct smb_hdr) + 27);*/
192 atomic_inc(&smBufAllocCount);
193 #ifdef CONFIG_CIFS_STATS2
194 atomic_inc(&totSmBufAllocCount);
195 #endif /* CONFIG_CIFS_STATS2 */
196
197 }
198 return ret_buf;
199 }
200
201 void
cifs_small_buf_release(void * buf_to_free)202 cifs_small_buf_release(void *buf_to_free)
203 {
204
205 if (buf_to_free == NULL) {
206 cFYI(1, "Null buffer passed to cifs_small_buf_release");
207 return;
208 }
209 mempool_free(buf_to_free, cifs_sm_req_poolp);
210
211 atomic_dec(&smBufAllocCount);
212 return;
213 }
214
215 /*
216 Find a free multiplex id (SMB mid). Otherwise there could be
217 mid collisions which might cause problems, demultiplexing the
218 wrong response to this request. Multiplex ids could collide if
219 one of a series requests takes much longer than the others, or
220 if a very large number of long lived requests (byte range
221 locks or FindNotify requests) are pending. No more than
222 64K-1 requests can be outstanding at one time. If no
223 mids are available, return zero. A future optimization
224 could make the combination of mids and uid the key we use
225 to demultiplex on (rather than mid alone).
226 In addition to the above check, the cifs demultiplex
227 code already used the command code as a secondary
228 check of the frame and if signing is negotiated the
229 response would be discarded if the mid were the same
230 but the signature was wrong. Since the mid is not put in the
231 pending queue until later (when it is about to be dispatched)
232 we do have to limit the number of outstanding requests
233 to somewhat less than 64K-1 although it is hard to imagine
234 so many threads being in the vfs at one time.
235 */
GetNextMid(struct TCP_Server_Info * server)236 __u16 GetNextMid(struct TCP_Server_Info *server)
237 {
238 __u16 mid = 0;
239 __u16 last_mid;
240 bool collision;
241
242 spin_lock(&GlobalMid_Lock);
243 last_mid = server->CurrentMid; /* we do not want to loop forever */
244 server->CurrentMid++;
245 /* This nested loop looks more expensive than it is.
246 In practice the list of pending requests is short,
247 fewer than 50, and the mids are likely to be unique
248 on the first pass through the loop unless some request
249 takes longer than the 64 thousand requests before it
250 (and it would also have to have been a request that
251 did not time out) */
252 while (server->CurrentMid != last_mid) {
253 struct mid_q_entry *mid_entry;
254 unsigned int num_mids;
255
256 collision = false;
257 if (server->CurrentMid == 0)
258 server->CurrentMid++;
259
260 num_mids = 0;
261 list_for_each_entry(mid_entry, &server->pending_mid_q, qhead) {
262 ++num_mids;
263 if (mid_entry->mid == server->CurrentMid &&
264 mid_entry->midState == MID_REQUEST_SUBMITTED) {
265 /* This mid is in use, try a different one */
266 collision = true;
267 break;
268 }
269 }
270
271 /*
272 * if we have more than 32k mids in the list, then something
273 * is very wrong. Possibly a local user is trying to DoS the
274 * box by issuing long-running calls and SIGKILL'ing them. If
275 * we get to 2^16 mids then we're in big trouble as this
276 * function could loop forever.
277 *
278 * Go ahead and assign out the mid in this situation, but force
279 * an eventual reconnect to clean out the pending_mid_q.
280 */
281 if (num_mids > 32768)
282 server->tcpStatus = CifsNeedReconnect;
283
284 if (!collision) {
285 mid = server->CurrentMid;
286 break;
287 }
288 server->CurrentMid++;
289 }
290 spin_unlock(&GlobalMid_Lock);
291 return mid;
292 }
293
294 /* NB: MID can not be set if treeCon not passed in, in that
295 case it is responsbility of caller to set the mid */
296 void
header_assemble(struct smb_hdr * buffer,char smb_command,const struct cifsTconInfo * treeCon,int word_count)297 header_assemble(struct smb_hdr *buffer, char smb_command /* command */ ,
298 const struct cifsTconInfo *treeCon, int word_count
299 /* length of fixed section (word count) in two byte units */)
300 {
301 struct list_head *temp_item;
302 struct cifsSesInfo *ses;
303 char *temp = (char *) buffer;
304
305 memset(temp, 0, 256); /* bigger than MAX_CIFS_HDR_SIZE */
306
307 buffer->smb_buf_length =
308 (2 * word_count) + sizeof(struct smb_hdr) -
309 4 /* RFC 1001 length field does not count */ +
310 2 /* for bcc field itself */ ;
311 /* Note that this is the only network field that has to be converted
312 to big endian and it is done just before we send it */
313
314 buffer->Protocol[0] = 0xFF;
315 buffer->Protocol[1] = 'S';
316 buffer->Protocol[2] = 'M';
317 buffer->Protocol[3] = 'B';
318 buffer->Command = smb_command;
319 buffer->Flags = 0x00; /* case sensitive */
320 buffer->Flags2 = SMBFLG2_KNOWS_LONG_NAMES;
321 buffer->Pid = cpu_to_le16((__u16)current->tgid);
322 buffer->PidHigh = cpu_to_le16((__u16)(current->tgid >> 16));
323 if (treeCon) {
324 buffer->Tid = treeCon->tid;
325 if (treeCon->ses) {
326 if (treeCon->ses->capabilities & CAP_UNICODE)
327 buffer->Flags2 |= SMBFLG2_UNICODE;
328 if (treeCon->ses->capabilities & CAP_STATUS32)
329 buffer->Flags2 |= SMBFLG2_ERR_STATUS;
330
331 /* Uid is not converted */
332 buffer->Uid = treeCon->ses->Suid;
333 buffer->Mid = GetNextMid(treeCon->ses->server);
334 if (multiuser_mount != 0) {
335 /* For the multiuser case, there are few obvious technically */
336 /* possible mechanisms to match the local linux user (uid) */
337 /* to a valid remote smb user (smb_uid): */
338 /* 1) Query Winbind (or other local pam/nss daemon */
339 /* for userid/password/logon_domain or credential */
340 /* 2) Query Winbind for uid to sid to username mapping */
341 /* and see if we have a matching password for existing*/
342 /* session for that user perhas getting password by */
343 /* adding a new pam_cifs module that stores passwords */
344 /* so that the cifs vfs can get at that for all logged*/
345 /* on users */
346 /* 3) (Which is the mechanism we have chosen) */
347 /* Search through sessions to the same server for a */
348 /* a match on the uid that was passed in on mount */
349 /* with the current processes uid (or euid?) and use */
350 /* that smb uid. If no existing smb session for */
351 /* that uid found, use the default smb session ie */
352 /* the smb session for the volume mounted which is */
353 /* the same as would be used if the multiuser mount */
354 /* flag were disabled. */
355
356 /* BB Add support for establishing new tCon and SMB Session */
357 /* with userid/password pairs found on the smb session */
358 /* for other target tcp/ip addresses BB */
359 if (current_fsuid() != treeCon->ses->linux_uid) {
360 cFYI(1, "Multiuser mode and UID "
361 "did not match tcon uid");
362 spin_lock(&cifs_tcp_ses_lock);
363 list_for_each(temp_item, &treeCon->ses->server->smb_ses_list) {
364 ses = list_entry(temp_item, struct cifsSesInfo, smb_ses_list);
365 if (ses->linux_uid == current_fsuid()) {
366 if (ses->server == treeCon->ses->server) {
367 cFYI(1, "found matching uid substitute right smb_uid");
368 buffer->Uid = ses->Suid;
369 break;
370 } else {
371 /* BB eventually call cifs_setup_session here */
372 cFYI(1, "local UID found but no smb sess with this server exists");
373 }
374 }
375 }
376 spin_unlock(&cifs_tcp_ses_lock);
377 }
378 }
379 }
380 if (treeCon->Flags & SMB_SHARE_IS_IN_DFS)
381 buffer->Flags2 |= SMBFLG2_DFS;
382 if (treeCon->nocase)
383 buffer->Flags |= SMBFLG_CASELESS;
384 if ((treeCon->ses) && (treeCon->ses->server))
385 if (treeCon->ses->server->secMode &
386 (SECMODE_SIGN_REQUIRED | SECMODE_SIGN_ENABLED))
387 buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE;
388 }
389
390 /* endian conversion of flags is now done just before sending */
391 buffer->WordCount = (char) word_count;
392 return;
393 }
394
395 static int
check_smb_hdr(struct smb_hdr * smb,__u16 mid)396 check_smb_hdr(struct smb_hdr *smb, __u16 mid)
397 {
398 /* does it have the right SMB "signature" ? */
399 if (*(__le32 *) smb->Protocol != cpu_to_le32(0x424d53ff)) {
400 cERROR(1, "Bad protocol string signature header 0x%x",
401 *(unsigned int *)smb->Protocol);
402 return 1;
403 }
404
405 /* Make sure that message ids match */
406 if (mid != smb->Mid) {
407 cERROR(1, "Mids do not match. received=%u expected=%u",
408 smb->Mid, mid);
409 return 1;
410 }
411
412 /* if it's a response then accept */
413 if (smb->Flags & SMBFLG_RESPONSE)
414 return 0;
415
416 /* only one valid case where server sends us request */
417 if (smb->Command == SMB_COM_LOCKING_ANDX)
418 return 0;
419
420 cERROR(1, "Server sent request, not response. mid=%u", smb->Mid);
421 return 1;
422 }
423
424 int
checkSMB(struct smb_hdr * smb,__u16 mid,unsigned int length)425 checkSMB(struct smb_hdr *smb, __u16 mid, unsigned int length)
426 {
427 __u32 len = smb->smb_buf_length;
428 __u32 clc_len; /* calculated length */
429 cFYI(0, "checkSMB Length: 0x%x, smb_buf_length: 0x%x", length, len);
430
431 if (length < 2 + sizeof(struct smb_hdr)) {
432 if ((length >= sizeof(struct smb_hdr) - 1)
433 && (smb->Status.CifsError != 0)) {
434 smb->WordCount = 0;
435 /* some error cases do not return wct and bcc */
436 return 0;
437 } else if ((length == sizeof(struct smb_hdr) + 1) &&
438 (smb->WordCount == 0)) {
439 char *tmp = (char *)smb;
440 /* Need to work around a bug in two servers here */
441 /* First, check if the part of bcc they sent was zero */
442 if (tmp[sizeof(struct smb_hdr)] == 0) {
443 /* some servers return only half of bcc
444 * on simple responses (wct, bcc both zero)
445 * in particular have seen this on
446 * ulogoffX and FindClose. This leaves
447 * one byte of bcc potentially unitialized
448 */
449 /* zero rest of bcc */
450 tmp[sizeof(struct smb_hdr)+1] = 0;
451 return 0;
452 }
453 cERROR(1, "rcvd invalid byte count (bcc)");
454 } else {
455 cERROR(1, "Length less than smb header size");
456 }
457 return 1;
458 }
459 if (len > CIFSMaxBufSize + MAX_CIFS_HDR_SIZE - 4) {
460 cERROR(1, "smb length greater than MaxBufSize, mid=%d",
461 smb->Mid);
462 return 1;
463 }
464
465 if (check_smb_hdr(smb, mid))
466 return 1;
467 clc_len = smbCalcSize_LE(smb);
468
469 if (4 + len != length) {
470 cERROR(1, "Length read does not match RFC1001 length %d",
471 len);
472 return 1;
473 }
474
475 if (4 + len != clc_len) {
476 /* check if bcc wrapped around for large read responses */
477 if ((len > 64 * 1024) && (len > clc_len)) {
478 /* check if lengths match mod 64K */
479 if (((4 + len) & 0xFFFF) == (clc_len & 0xFFFF))
480 return 0; /* bcc wrapped */
481 }
482 cFYI(1, "Calculated size %u vs length %u mismatch for mid=%u",
483 clc_len, 4 + len, smb->Mid);
484
485 if (4 + len < clc_len) {
486 cERROR(1, "RFC1001 size %u smaller than SMB for mid=%u",
487 len, smb->Mid);
488 return 1;
489 } else if (len > clc_len + 512) {
490 /*
491 * Some servers (Windows XP in particular) send more
492 * data than the lengths in the SMB packet would
493 * indicate on certain calls (byte range locks and
494 * trans2 find first calls in particular). While the
495 * client can handle such a frame by ignoring the
496 * trailing data, we choose limit the amount of extra
497 * data to 512 bytes.
498 */
499 cERROR(1, "RFC1001 size %u more than 512 bytes larger "
500 "than SMB for mid=%u", len, smb->Mid);
501 return 1;
502 }
503 }
504 return 0;
505 }
506
507 bool
is_valid_oplock_break(struct smb_hdr * buf,struct TCP_Server_Info * srv)508 is_valid_oplock_break(struct smb_hdr *buf, struct TCP_Server_Info *srv)
509 {
510 struct smb_com_lock_req *pSMB = (struct smb_com_lock_req *)buf;
511 struct list_head *tmp, *tmp1, *tmp2;
512 struct cifsSesInfo *ses;
513 struct cifsTconInfo *tcon;
514 struct cifsInodeInfo *pCifsInode;
515 struct cifsFileInfo *netfile;
516
517 cFYI(1, "Checking for oplock break or dnotify response");
518 if ((pSMB->hdr.Command == SMB_COM_NT_TRANSACT) &&
519 (pSMB->hdr.Flags & SMBFLG_RESPONSE)) {
520 struct smb_com_transaction_change_notify_rsp *pSMBr =
521 (struct smb_com_transaction_change_notify_rsp *)buf;
522 struct file_notify_information *pnotify;
523 __u32 data_offset = 0;
524 if (get_bcc_le(buf) > sizeof(struct file_notify_information)) {
525 data_offset = le32_to_cpu(pSMBr->DataOffset);
526
527 pnotify = (struct file_notify_information *)
528 ((char *)&pSMBr->hdr.Protocol + data_offset);
529 cFYI(1, "dnotify on %s Action: 0x%x",
530 pnotify->FileName, pnotify->Action);
531 /* cifs_dump_mem("Rcvd notify Data: ",buf,
532 sizeof(struct smb_hdr)+60); */
533 return true;
534 }
535 if (pSMBr->hdr.Status.CifsError) {
536 cFYI(1, "notify err 0x%d",
537 pSMBr->hdr.Status.CifsError);
538 return true;
539 }
540 return false;
541 }
542 if (pSMB->hdr.Command != SMB_COM_LOCKING_ANDX)
543 return false;
544 if (pSMB->hdr.Flags & SMBFLG_RESPONSE) {
545 /* no sense logging error on invalid handle on oplock
546 break - harmless race between close request and oplock
547 break response is expected from time to time writing out
548 large dirty files cached on the client */
549 if ((NT_STATUS_INVALID_HANDLE) ==
550 le32_to_cpu(pSMB->hdr.Status.CifsError)) {
551 cFYI(1, "invalid handle on oplock break");
552 return true;
553 } else if (ERRbadfid ==
554 le16_to_cpu(pSMB->hdr.Status.DosError.Error)) {
555 return true;
556 } else {
557 return false; /* on valid oplock brk we get "request" */
558 }
559 }
560 if (pSMB->hdr.WordCount != 8)
561 return false;
562
563 cFYI(1, "oplock type 0x%d level 0x%d",
564 pSMB->LockType, pSMB->OplockLevel);
565 if (!(pSMB->LockType & LOCKING_ANDX_OPLOCK_RELEASE))
566 return false;
567
568 /* look up tcon based on tid & uid */
569 spin_lock(&cifs_tcp_ses_lock);
570 list_for_each(tmp, &srv->smb_ses_list) {
571 ses = list_entry(tmp, struct cifsSesInfo, smb_ses_list);
572 list_for_each(tmp1, &ses->tcon_list) {
573 tcon = list_entry(tmp1, struct cifsTconInfo, tcon_list);
574 if (tcon->tid != buf->Tid)
575 continue;
576
577 cifs_stats_inc(&tcon->num_oplock_brks);
578 spin_lock(&cifs_file_list_lock);
579 list_for_each(tmp2, &tcon->openFileList) {
580 netfile = list_entry(tmp2, struct cifsFileInfo,
581 tlist);
582 if (pSMB->Fid != netfile->netfid)
583 continue;
584
585 cFYI(1, "file id match, oplock break");
586 pCifsInode = CIFS_I(netfile->dentry->d_inode);
587
588 cifs_set_oplock_level(pCifsInode,
589 pSMB->OplockLevel ? OPLOCK_READ : 0);
590 /*
591 * cifs_oplock_break_put() can't be called
592 * from here. Get reference after queueing
593 * succeeded. cifs_oplock_break() will
594 * synchronize using cifs_file_list_lock.
595 */
596 if (queue_work(system_nrt_wq,
597 &netfile->oplock_break))
598 cifs_oplock_break_get(netfile);
599 netfile->oplock_break_cancelled = false;
600
601 spin_unlock(&cifs_file_list_lock);
602 spin_unlock(&cifs_tcp_ses_lock);
603 return true;
604 }
605 spin_unlock(&cifs_file_list_lock);
606 spin_unlock(&cifs_tcp_ses_lock);
607 cFYI(1, "No matching file for oplock break");
608 return true;
609 }
610 }
611 spin_unlock(&cifs_tcp_ses_lock);
612 cFYI(1, "Can not process oplock break for non-existent connection");
613 return true;
614 }
615
616 void
dump_smb(struct smb_hdr * smb_buf,int smb_buf_length)617 dump_smb(struct smb_hdr *smb_buf, int smb_buf_length)
618 {
619 int i, j;
620 char debug_line[17];
621 unsigned char *buffer;
622
623 if (traceSMB == 0)
624 return;
625
626 buffer = (unsigned char *) smb_buf;
627 for (i = 0, j = 0; i < smb_buf_length; i++, j++) {
628 if (i % 8 == 0) {
629 /* have reached the beginning of line */
630 printk(KERN_DEBUG "| ");
631 j = 0;
632 }
633 printk("%0#4x ", buffer[i]);
634 debug_line[2 * j] = ' ';
635 if (isprint(buffer[i]))
636 debug_line[1 + (2 * j)] = buffer[i];
637 else
638 debug_line[1 + (2 * j)] = '_';
639
640 if (i % 8 == 7) {
641 /* reached end of line, time to print ascii */
642 debug_line[16] = 0;
643 printk(" | %s\n", debug_line);
644 }
645 }
646 for (; j < 8; j++) {
647 printk(" ");
648 debug_line[2 * j] = ' ';
649 debug_line[1 + (2 * j)] = ' ';
650 }
651 printk(" | %s\n", debug_line);
652 return;
653 }
654
655 void
cifs_autodisable_serverino(struct cifs_sb_info * cifs_sb)656 cifs_autodisable_serverino(struct cifs_sb_info *cifs_sb)
657 {
658 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) {
659 cifs_sb->mnt_cifs_flags &= ~CIFS_MOUNT_SERVER_INUM;
660 cERROR(1, "Autodisabling the use of server inode numbers on "
661 "%s. This server doesn't seem to support them "
662 "properly. Hardlinks will not be recognized on this "
663 "mount. Consider mounting with the \"noserverino\" "
664 "option to silence this message.",
665 cifs_sb_master_tcon(cifs_sb)->treeName);
666 }
667 }
668
cifs_set_oplock_level(struct cifsInodeInfo * cinode,__u32 oplock)669 void cifs_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock)
670 {
671 oplock &= 0xF;
672
673 if (oplock == OPLOCK_EXCLUSIVE) {
674 cinode->clientCanCacheAll = true;
675 cinode->clientCanCacheRead = true;
676 cFYI(1, "Exclusive Oplock granted on inode %p",
677 &cinode->vfs_inode);
678 } else if (oplock == OPLOCK_READ) {
679 cinode->clientCanCacheAll = false;
680 cinode->clientCanCacheRead = true;
681 cFYI(1, "Level II Oplock granted on inode %p",
682 &cinode->vfs_inode);
683 } else {
684 cinode->clientCanCacheAll = false;
685 cinode->clientCanCacheRead = false;
686 }
687 }
688