Merge branch 'master' into next
[deliverable/linux.git] / fs / cifs / misc.c
1 /*
2 * fs/cifs/misc.c
3 *
4 * Copyright (C) International Business Machines Corp., 2002,2008
5 * Author(s): Steve French (sfrench@us.ibm.com)
6 *
7 * This library is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU Lesser General Public License as published
9 * by the Free Software Foundation; either version 2.1 of the License, or
10 * (at your option) any later version.
11 *
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
15 * the GNU Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public License
18 * along with this library; if not, write to the Free Software
19 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
20 */
21
22 #include <linux/slab.h>
23 #include <linux/ctype.h>
24 #include <linux/mempool.h>
25 #include "cifspdu.h"
26 #include "cifsglob.h"
27 #include "cifsproto.h"
28 #include "cifs_debug.h"
29 #include "smberr.h"
30 #include "nterr.h"
31 #include "cifs_unicode.h"
32
33 extern mempool_t *cifs_sm_req_poolp;
34 extern mempool_t *cifs_req_poolp;
35 extern struct task_struct *oplockThread;
36
37 /* The xid serves as a useful identifier for each incoming vfs request,
38 in a similar way to the mid which is useful to track each sent smb,
39 and CurrentXid can also provide a running counter (although it
40 will eventually wrap past zero) of the total vfs operations handled
41 since the cifs fs was mounted */
42
43 unsigned int
44 _GetXid(void)
45 {
46 unsigned int xid;
47
48 spin_lock(&GlobalMid_Lock);
49 GlobalTotalActiveXid++;
50
51 /* keep high water mark for number of simultaneous ops in filesystem */
52 if (GlobalTotalActiveXid > GlobalMaxActiveXid)
53 GlobalMaxActiveXid = GlobalTotalActiveXid;
54 if (GlobalTotalActiveXid > 65000)
55 cFYI(1, ("warning: more than 65000 requests active"));
56 xid = GlobalCurrentXid++;
57 spin_unlock(&GlobalMid_Lock);
58 return xid;
59 }
60
61 void
62 _FreeXid(unsigned int xid)
63 {
64 spin_lock(&GlobalMid_Lock);
65 /* if (GlobalTotalActiveXid == 0)
66 BUG(); */
67 GlobalTotalActiveXid--;
68 spin_unlock(&GlobalMid_Lock);
69 }
70
71 struct cifsSesInfo *
72 sesInfoAlloc(void)
73 {
74 struct cifsSesInfo *ret_buf;
75
76 ret_buf = kzalloc(sizeof(struct cifsSesInfo), GFP_KERNEL);
77 if (ret_buf) {
78 atomic_inc(&sesInfoAllocCount);
79 ret_buf->status = CifsNew;
80 ++ret_buf->ses_count;
81 INIT_LIST_HEAD(&ret_buf->smb_ses_list);
82 INIT_LIST_HEAD(&ret_buf->tcon_list);
83 init_MUTEX(&ret_buf->sesSem);
84 }
85 return ret_buf;
86 }
87
88 void
89 sesInfoFree(struct cifsSesInfo *buf_to_free)
90 {
91 if (buf_to_free == NULL) {
92 cFYI(1, ("Null buffer passed to sesInfoFree"));
93 return;
94 }
95
96 atomic_dec(&sesInfoAllocCount);
97 kfree(buf_to_free->serverOS);
98 kfree(buf_to_free->serverDomain);
99 kfree(buf_to_free->serverNOS);
100 kfree(buf_to_free->password);
101 kfree(buf_to_free->domainName);
102 kfree(buf_to_free);
103 }
104
105 struct cifsTconInfo *
106 tconInfoAlloc(void)
107 {
108 struct cifsTconInfo *ret_buf;
109 ret_buf = kzalloc(sizeof(struct cifsTconInfo), GFP_KERNEL);
110 if (ret_buf) {
111 atomic_inc(&tconInfoAllocCount);
112 ret_buf->tidStatus = CifsNew;
113 ++ret_buf->tc_count;
114 INIT_LIST_HEAD(&ret_buf->openFileList);
115 INIT_LIST_HEAD(&ret_buf->tcon_list);
116 #ifdef CONFIG_CIFS_STATS
117 spin_lock_init(&ret_buf->stat_lock);
118 #endif
119 }
120 return ret_buf;
121 }
122
123 void
124 tconInfoFree(struct cifsTconInfo *buf_to_free)
125 {
126 if (buf_to_free == NULL) {
127 cFYI(1, ("Null buffer passed to tconInfoFree"));
128 return;
129 }
130 atomic_dec(&tconInfoAllocCount);
131 kfree(buf_to_free->nativeFileSystem);
132 kfree(buf_to_free);
133 }
134
135 struct smb_hdr *
136 cifs_buf_get(void)
137 {
138 struct smb_hdr *ret_buf = NULL;
139
140 /* We could use negotiated size instead of max_msgsize -
141 but it may be more efficient to always alloc same size
142 albeit slightly larger than necessary and maxbuffersize
143 defaults to this and can not be bigger */
144 ret_buf = mempool_alloc(cifs_req_poolp, GFP_NOFS);
145
146 /* clear the first few header bytes */
147 /* for most paths, more is cleared in header_assemble */
148 if (ret_buf) {
149 memset(ret_buf, 0, sizeof(struct smb_hdr) + 3);
150 atomic_inc(&bufAllocCount);
151 #ifdef CONFIG_CIFS_STATS2
152 atomic_inc(&totBufAllocCount);
153 #endif /* CONFIG_CIFS_STATS2 */
154 }
155
156 return ret_buf;
157 }
158
159 void
160 cifs_buf_release(void *buf_to_free)
161 {
162 if (buf_to_free == NULL) {
163 /* cFYI(1, ("Null buffer passed to cifs_buf_release"));*/
164 return;
165 }
166 mempool_free(buf_to_free, cifs_req_poolp);
167
168 atomic_dec(&bufAllocCount);
169 return;
170 }
171
172 struct smb_hdr *
173 cifs_small_buf_get(void)
174 {
175 struct smb_hdr *ret_buf = NULL;
176
177 /* We could use negotiated size instead of max_msgsize -
178 but it may be more efficient to always alloc same size
179 albeit slightly larger than necessary and maxbuffersize
180 defaults to this and can not be bigger */
181 ret_buf = mempool_alloc(cifs_sm_req_poolp, GFP_NOFS);
182 if (ret_buf) {
183 /* No need to clear memory here, cleared in header assemble */
184 /* memset(ret_buf, 0, sizeof(struct smb_hdr) + 27);*/
185 atomic_inc(&smBufAllocCount);
186 #ifdef CONFIG_CIFS_STATS2
187 atomic_inc(&totSmBufAllocCount);
188 #endif /* CONFIG_CIFS_STATS2 */
189
190 }
191 return ret_buf;
192 }
193
194 void
195 cifs_small_buf_release(void *buf_to_free)
196 {
197
198 if (buf_to_free == NULL) {
199 cFYI(1, ("Null buffer passed to cifs_small_buf_release"));
200 return;
201 }
202 mempool_free(buf_to_free, cifs_sm_req_poolp);
203
204 atomic_dec(&smBufAllocCount);
205 return;
206 }
207
208 /*
209 Find a free multiplex id (SMB mid). Otherwise there could be
210 mid collisions which might cause problems, demultiplexing the
211 wrong response to this request. Multiplex ids could collide if
212 one of a series requests takes much longer than the others, or
213 if a very large number of long lived requests (byte range
214 locks or FindNotify requests) are pending. No more than
215 64K-1 requests can be outstanding at one time. If no
216 mids are available, return zero. A future optimization
217 could make the combination of mids and uid the key we use
218 to demultiplex on (rather than mid alone).
219 In addition to the above check, the cifs demultiplex
220 code already used the command code as a secondary
221 check of the frame and if signing is negotiated the
222 response would be discarded if the mid were the same
223 but the signature was wrong. Since the mid is not put in the
224 pending queue until later (when it is about to be dispatched)
225 we do have to limit the number of outstanding requests
226 to somewhat less than 64K-1 although it is hard to imagine
227 so many threads being in the vfs at one time.
228 */
229 __u16 GetNextMid(struct TCP_Server_Info *server)
230 {
231 __u16 mid = 0;
232 __u16 last_mid;
233 int collision;
234
235 if (server == NULL)
236 return mid;
237
238 spin_lock(&GlobalMid_Lock);
239 last_mid = server->CurrentMid; /* we do not want to loop forever */
240 server->CurrentMid++;
241 /* This nested loop looks more expensive than it is.
242 In practice the list of pending requests is short,
243 fewer than 50, and the mids are likely to be unique
244 on the first pass through the loop unless some request
245 takes longer than the 64 thousand requests before it
246 (and it would also have to have been a request that
247 did not time out) */
248 while (server->CurrentMid != last_mid) {
249 struct list_head *tmp;
250 struct mid_q_entry *mid_entry;
251
252 collision = 0;
253 if (server->CurrentMid == 0)
254 server->CurrentMid++;
255
256 list_for_each(tmp, &server->pending_mid_q) {
257 mid_entry = list_entry(tmp, struct mid_q_entry, qhead);
258
259 if ((mid_entry->mid == server->CurrentMid) &&
260 (mid_entry->midState == MID_REQUEST_SUBMITTED)) {
261 /* This mid is in use, try a different one */
262 collision = 1;
263 break;
264 }
265 }
266 if (collision == 0) {
267 mid = server->CurrentMid;
268 break;
269 }
270 server->CurrentMid++;
271 }
272 spin_unlock(&GlobalMid_Lock);
273 return mid;
274 }
275
276 /* NB: MID can not be set if treeCon not passed in, in that
277 case it is responsbility of caller to set the mid */
278 void
279 header_assemble(struct smb_hdr *buffer, char smb_command /* command */ ,
280 const struct cifsTconInfo *treeCon, int word_count
281 /* length of fixed section (word count) in two byte units */)
282 {
283 struct list_head *temp_item;
284 struct cifsSesInfo *ses;
285 char *temp = (char *) buffer;
286
287 memset(temp, 0, 256); /* bigger than MAX_CIFS_HDR_SIZE */
288
289 buffer->smb_buf_length =
290 (2 * word_count) + sizeof(struct smb_hdr) -
291 4 /* RFC 1001 length field does not count */ +
292 2 /* for bcc field itself */ ;
293 /* Note that this is the only network field that has to be converted
294 to big endian and it is done just before we send it */
295
296 buffer->Protocol[0] = 0xFF;
297 buffer->Protocol[1] = 'S';
298 buffer->Protocol[2] = 'M';
299 buffer->Protocol[3] = 'B';
300 buffer->Command = smb_command;
301 buffer->Flags = 0x00; /* case sensitive */
302 buffer->Flags2 = SMBFLG2_KNOWS_LONG_NAMES;
303 buffer->Pid = cpu_to_le16((__u16)current->tgid);
304 buffer->PidHigh = cpu_to_le16((__u16)(current->tgid >> 16));
305 if (treeCon) {
306 buffer->Tid = treeCon->tid;
307 if (treeCon->ses) {
308 if (treeCon->ses->capabilities & CAP_UNICODE)
309 buffer->Flags2 |= SMBFLG2_UNICODE;
310 if (treeCon->ses->capabilities & CAP_STATUS32)
311 buffer->Flags2 |= SMBFLG2_ERR_STATUS;
312
313 /* Uid is not converted */
314 buffer->Uid = treeCon->ses->Suid;
315 buffer->Mid = GetNextMid(treeCon->ses->server);
316 if (multiuser_mount != 0) {
317 /* For the multiuser case, there are few obvious technically */
318 /* possible mechanisms to match the local linux user (uid) */
319 /* to a valid remote smb user (smb_uid): */
320 /* 1) Query Winbind (or other local pam/nss daemon */
321 /* for userid/password/logon_domain or credential */
322 /* 2) Query Winbind for uid to sid to username mapping */
323 /* and see if we have a matching password for existing*/
324 /* session for that user perhas getting password by */
325 /* adding a new pam_cifs module that stores passwords */
326 /* so that the cifs vfs can get at that for all logged*/
327 /* on users */
328 /* 3) (Which is the mechanism we have chosen) */
329 /* Search through sessions to the same server for a */
330 /* a match on the uid that was passed in on mount */
331 /* with the current processes uid (or euid?) and use */
332 /* that smb uid. If no existing smb session for */
333 /* that uid found, use the default smb session ie */
334 /* the smb session for the volume mounted which is */
335 /* the same as would be used if the multiuser mount */
336 /* flag were disabled. */
337
338 /* BB Add support for establishing new tCon and SMB Session */
339 /* with userid/password pairs found on the smb session */
340 /* for other target tcp/ip addresses BB */
341 if (current_fsuid() != treeCon->ses->linux_uid) {
342 cFYI(1, ("Multiuser mode and UID "
343 "did not match tcon uid"));
344 read_lock(&cifs_tcp_ses_lock);
345 list_for_each(temp_item, &treeCon->ses->server->smb_ses_list) {
346 ses = list_entry(temp_item, struct cifsSesInfo, smb_ses_list);
347 if (ses->linux_uid == current_fsuid()) {
348 if (ses->server == treeCon->ses->server) {
349 cFYI(1, ("found matching uid substitute right smb_uid"));
350 buffer->Uid = ses->Suid;
351 break;
352 } else {
353 /* BB eventually call cifs_setup_session here */
354 cFYI(1, ("local UID found but no smb sess with this server exists"));
355 }
356 }
357 }
358 read_unlock(&cifs_tcp_ses_lock);
359 }
360 }
361 }
362 if (treeCon->Flags & SMB_SHARE_IS_IN_DFS)
363 buffer->Flags2 |= SMBFLG2_DFS;
364 if (treeCon->nocase)
365 buffer->Flags |= SMBFLG_CASELESS;
366 if ((treeCon->ses) && (treeCon->ses->server))
367 if (treeCon->ses->server->secMode &
368 (SECMODE_SIGN_REQUIRED | SECMODE_SIGN_ENABLED))
369 buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE;
370 }
371
372 /* endian conversion of flags is now done just before sending */
373 buffer->WordCount = (char) word_count;
374 return;
375 }
376
377 static int
378 checkSMBhdr(struct smb_hdr *smb, __u16 mid)
379 {
380 /* Make sure that this really is an SMB, that it is a response,
381 and that the message ids match */
382 if ((*(__le32 *) smb->Protocol == cpu_to_le32(0x424d53ff)) &&
383 (mid == smb->Mid)) {
384 if (smb->Flags & SMBFLG_RESPONSE)
385 return 0;
386 else {
387 /* only one valid case where server sends us request */
388 if (smb->Command == SMB_COM_LOCKING_ANDX)
389 return 0;
390 else
391 cERROR(1, ("Received Request not response"));
392 }
393 } else { /* bad signature or mid */
394 if (*(__le32 *) smb->Protocol != cpu_to_le32(0x424d53ff))
395 cERROR(1,
396 ("Bad protocol string signature header %x",
397 *(unsigned int *) smb->Protocol));
398 if (mid != smb->Mid)
399 cERROR(1, ("Mids do not match"));
400 }
401 cERROR(1, ("bad smb detected. The Mid=%d", smb->Mid));
402 return 1;
403 }
404
405 int
406 checkSMB(struct smb_hdr *smb, __u16 mid, unsigned int length)
407 {
408 __u32 len = smb->smb_buf_length;
409 __u32 clc_len; /* calculated length */
410 cFYI(0, ("checkSMB Length: 0x%x, smb_buf_length: 0x%x", length, len));
411
412 if (length < 2 + sizeof(struct smb_hdr)) {
413 if ((length >= sizeof(struct smb_hdr) - 1)
414 && (smb->Status.CifsError != 0)) {
415 smb->WordCount = 0;
416 /* some error cases do not return wct and bcc */
417 return 0;
418 } else if ((length == sizeof(struct smb_hdr) + 1) &&
419 (smb->WordCount == 0)) {
420 char *tmp = (char *)smb;
421 /* Need to work around a bug in two servers here */
422 /* First, check if the part of bcc they sent was zero */
423 if (tmp[sizeof(struct smb_hdr)] == 0) {
424 /* some servers return only half of bcc
425 * on simple responses (wct, bcc both zero)
426 * in particular have seen this on
427 * ulogoffX and FindClose. This leaves
428 * one byte of bcc potentially unitialized
429 */
430 /* zero rest of bcc */
431 tmp[sizeof(struct smb_hdr)+1] = 0;
432 return 0;
433 }
434 cERROR(1, ("rcvd invalid byte count (bcc)"));
435 } else {
436 cERROR(1, ("Length less than smb header size"));
437 }
438 return 1;
439 }
440 if (len > CIFSMaxBufSize + MAX_CIFS_HDR_SIZE - 4) {
441 cERROR(1, ("smb length greater than MaxBufSize, mid=%d",
442 smb->Mid));
443 return 1;
444 }
445
446 if (checkSMBhdr(smb, mid))
447 return 1;
448 clc_len = smbCalcSize_LE(smb);
449
450 if (4 + len != length) {
451 cERROR(1, ("Length read does not match RFC1001 length %d",
452 len));
453 return 1;
454 }
455
456 if (4 + len != clc_len) {
457 /* check if bcc wrapped around for large read responses */
458 if ((len > 64 * 1024) && (len > clc_len)) {
459 /* check if lengths match mod 64K */
460 if (((4 + len) & 0xFFFF) == (clc_len & 0xFFFF))
461 return 0; /* bcc wrapped */
462 }
463 cFYI(1, ("Calculated size %d vs length %d mismatch for mid %d",
464 clc_len, 4 + len, smb->Mid));
465 /* Windows XP can return a few bytes too much, presumably
466 an illegal pad, at the end of byte range lock responses
467 so we allow for that three byte pad, as long as actual
468 received length is as long or longer than calculated length */
469 /* We have now had to extend this more, since there is a
470 case in which it needs to be bigger still to handle a
471 malformed response to transact2 findfirst from WinXP when
472 access denied is returned and thus bcc and wct are zero
473 but server says length is 0x21 bytes too long as if the server
474 forget to reset the smb rfc1001 length when it reset the
475 wct and bcc to minimum size and drop the t2 parms and data */
476 if ((4+len > clc_len) && (len <= clc_len + 512))
477 return 0;
478 else {
479 cERROR(1, ("RFC1001 size %d bigger than SMB for Mid=%d",
480 len, smb->Mid));
481 return 1;
482 }
483 }
484 return 0;
485 }
486
487 bool
488 is_valid_oplock_break(struct smb_hdr *buf, struct TCP_Server_Info *srv)
489 {
490 struct smb_com_lock_req *pSMB = (struct smb_com_lock_req *)buf;
491 struct list_head *tmp, *tmp1, *tmp2;
492 struct cifsSesInfo *ses;
493 struct cifsTconInfo *tcon;
494 struct cifsInodeInfo *pCifsInode;
495 struct cifsFileInfo *netfile;
496
497 cFYI(1, ("Checking for oplock break or dnotify response"));
498 if ((pSMB->hdr.Command == SMB_COM_NT_TRANSACT) &&
499 (pSMB->hdr.Flags & SMBFLG_RESPONSE)) {
500 struct smb_com_transaction_change_notify_rsp *pSMBr =
501 (struct smb_com_transaction_change_notify_rsp *)buf;
502 struct file_notify_information *pnotify;
503 __u32 data_offset = 0;
504 if (pSMBr->ByteCount > sizeof(struct file_notify_information)) {
505 data_offset = le32_to_cpu(pSMBr->DataOffset);
506
507 pnotify = (struct file_notify_information *)
508 ((char *)&pSMBr->hdr.Protocol + data_offset);
509 cFYI(1, ("dnotify on %s Action: 0x%x",
510 pnotify->FileName, pnotify->Action));
511 /* cifs_dump_mem("Rcvd notify Data: ",buf,
512 sizeof(struct smb_hdr)+60); */
513 return true;
514 }
515 if (pSMBr->hdr.Status.CifsError) {
516 cFYI(1, ("notify err 0x%d",
517 pSMBr->hdr.Status.CifsError));
518 return true;
519 }
520 return false;
521 }
522 if (pSMB->hdr.Command != SMB_COM_LOCKING_ANDX)
523 return false;
524 if (pSMB->hdr.Flags & SMBFLG_RESPONSE) {
525 /* no sense logging error on invalid handle on oplock
526 break - harmless race between close request and oplock
527 break response is expected from time to time writing out
528 large dirty files cached on the client */
529 if ((NT_STATUS_INVALID_HANDLE) ==
530 le32_to_cpu(pSMB->hdr.Status.CifsError)) {
531 cFYI(1, ("invalid handle on oplock break"));
532 return true;
533 } else if (ERRbadfid ==
534 le16_to_cpu(pSMB->hdr.Status.DosError.Error)) {
535 return true;
536 } else {
537 return false; /* on valid oplock brk we get "request" */
538 }
539 }
540 if (pSMB->hdr.WordCount != 8)
541 return false;
542
543 cFYI(1, ("oplock type 0x%d level 0x%d",
544 pSMB->LockType, pSMB->OplockLevel));
545 if (!(pSMB->LockType & LOCKING_ANDX_OPLOCK_RELEASE))
546 return false;
547
548 /* look up tcon based on tid & uid */
549 read_lock(&cifs_tcp_ses_lock);
550 list_for_each(tmp, &srv->smb_ses_list) {
551 ses = list_entry(tmp, struct cifsSesInfo, smb_ses_list);
552 list_for_each(tmp1, &ses->tcon_list) {
553 tcon = list_entry(tmp1, struct cifsTconInfo, tcon_list);
554 if (tcon->tid != buf->Tid)
555 continue;
556
557 cifs_stats_inc(&tcon->num_oplock_brks);
558 list_for_each(tmp2, &tcon->openFileList) {
559 netfile = list_entry(tmp2, struct cifsFileInfo,
560 tlist);
561 if (pSMB->Fid != netfile->netfid)
562 continue;
563
564 read_unlock(&cifs_tcp_ses_lock);
565 cFYI(1, ("file id match, oplock break"));
566 pCifsInode = CIFS_I(netfile->pInode);
567 pCifsInode->clientCanCacheAll = false;
568 if (pSMB->OplockLevel == 0)
569 pCifsInode->clientCanCacheRead = false;
570 pCifsInode->oplockPending = true;
571 AllocOplockQEntry(netfile->pInode,
572 netfile->netfid, tcon);
573 cFYI(1, ("about to wake up oplock thread"));
574 if (oplockThread)
575 wake_up_process(oplockThread);
576
577 return true;
578 }
579 read_unlock(&cifs_tcp_ses_lock);
580 cFYI(1, ("No matching file for oplock break"));
581 return true;
582 }
583 }
584 read_unlock(&cifs_tcp_ses_lock);
585 cFYI(1, ("Can not process oplock break for non-existent connection"));
586 return true;
587 }
588
589 void
590 dump_smb(struct smb_hdr *smb_buf, int smb_buf_length)
591 {
592 int i, j;
593 char debug_line[17];
594 unsigned char *buffer;
595
596 if (traceSMB == 0)
597 return;
598
599 buffer = (unsigned char *) smb_buf;
600 for (i = 0, j = 0; i < smb_buf_length; i++, j++) {
601 if (i % 8 == 0) {
602 /* have reached the beginning of line */
603 printk(KERN_DEBUG "| ");
604 j = 0;
605 }
606 printk("%0#4x ", buffer[i]);
607 debug_line[2 * j] = ' ';
608 if (isprint(buffer[i]))
609 debug_line[1 + (2 * j)] = buffer[i];
610 else
611 debug_line[1 + (2 * j)] = '_';
612
613 if (i % 8 == 7) {
614 /* reached end of line, time to print ascii */
615 debug_line[16] = 0;
616 printk(" | %s\n", debug_line);
617 }
618 }
619 for (; j < 8; j++) {
620 printk(" ");
621 debug_line[2 * j] = ' ';
622 debug_line[1 + (2 * j)] = ' ';
623 }
624 printk(" | %s\n", debug_line);
625 return;
626 }
627
628 /* Windows maps these to the user defined 16 bit Unicode range since they are
629 reserved symbols (along with \ and /), otherwise illegal to store
630 in filenames in NTFS */
631 #define UNI_ASTERIK (__u16) ('*' + 0xF000)
632 #define UNI_QUESTION (__u16) ('?' + 0xF000)
633 #define UNI_COLON (__u16) (':' + 0xF000)
634 #define UNI_GRTRTHAN (__u16) ('>' + 0xF000)
635 #define UNI_LESSTHAN (__u16) ('<' + 0xF000)
636 #define UNI_PIPE (__u16) ('|' + 0xF000)
637 #define UNI_SLASH (__u16) ('\\' + 0xF000)
638
639 /* Convert 16 bit Unicode pathname from wire format to string in current code
640 page. Conversion may involve remapping up the seven characters that are
641 only legal in POSIX-like OS (if they are present in the string). Path
642 names are little endian 16 bit Unicode on the wire */
643 int
644 cifs_convertUCSpath(char *target, const __le16 *source, int maxlen,
645 const struct nls_table *cp)
646 {
647 int i, j, len;
648 __u16 src_char;
649
650 for (i = 0, j = 0; i < maxlen; i++) {
651 src_char = le16_to_cpu(source[i]);
652 switch (src_char) {
653 case 0:
654 goto cUCS_out; /* BB check this BB */
655 case UNI_COLON:
656 target[j] = ':';
657 break;
658 case UNI_ASTERIK:
659 target[j] = '*';
660 break;
661 case UNI_QUESTION:
662 target[j] = '?';
663 break;
664 /* BB We can not handle remapping slash until
665 all the calls to build_path_from_dentry
666 are modified, as they use slash as separator BB */
667 /* case UNI_SLASH:
668 target[j] = '\\';
669 break;*/
670 case UNI_PIPE:
671 target[j] = '|';
672 break;
673 case UNI_GRTRTHAN:
674 target[j] = '>';
675 break;
676 case UNI_LESSTHAN:
677 target[j] = '<';
678 break;
679 default:
680 len = cp->uni2char(src_char, &target[j],
681 NLS_MAX_CHARSET_SIZE);
682 if (len > 0) {
683 j += len;
684 continue;
685 } else {
686 target[j] = '?';
687 }
688 }
689 j++;
690 /* make sure we do not overrun callers allocated temp buffer */
691 if (j >= (2 * NAME_MAX))
692 break;
693 }
694 cUCS_out:
695 target[j] = 0;
696 return j;
697 }
698
699 /* Convert 16 bit Unicode pathname to wire format from string in current code
700 page. Conversion may involve remapping up the seven characters that are
701 only legal in POSIX-like OS (if they are present in the string). Path
702 names are little endian 16 bit Unicode on the wire */
703 int
704 cifsConvertToUCS(__le16 *target, const char *source, int maxlen,
705 const struct nls_table *cp, int mapChars)
706 {
707 int i, j, charlen;
708 int len_remaining = maxlen;
709 char src_char;
710 __u16 temp;
711
712 if (!mapChars)
713 return cifs_strtoUCS(target, source, PATH_MAX, cp);
714
715 for (i = 0, j = 0; i < maxlen; j++) {
716 src_char = source[i];
717 switch (src_char) {
718 case 0:
719 target[j] = 0;
720 goto ctoUCS_out;
721 case ':':
722 target[j] = cpu_to_le16(UNI_COLON);
723 break;
724 case '*':
725 target[j] = cpu_to_le16(UNI_ASTERIK);
726 break;
727 case '?':
728 target[j] = cpu_to_le16(UNI_QUESTION);
729 break;
730 case '<':
731 target[j] = cpu_to_le16(UNI_LESSTHAN);
732 break;
733 case '>':
734 target[j] = cpu_to_le16(UNI_GRTRTHAN);
735 break;
736 case '|':
737 target[j] = cpu_to_le16(UNI_PIPE);
738 break;
739 /* BB We can not handle remapping slash until
740 all the calls to build_path_from_dentry
741 are modified, as they use slash as separator BB */
742 /* case '\\':
743 target[j] = cpu_to_le16(UNI_SLASH);
744 break;*/
745 default:
746 charlen = cp->char2uni(source+i,
747 len_remaining, &temp);
748 /* if no match, use question mark, which
749 at least in some cases servers as wild card */
750 if (charlen < 1) {
751 target[j] = cpu_to_le16(0x003f);
752 charlen = 1;
753 } else
754 target[j] = cpu_to_le16(temp);
755 len_remaining -= charlen;
756 /* character may take more than one byte in the
757 the source string, but will take exactly two
758 bytes in the target string */
759 i += charlen;
760 continue;
761 }
762 i++; /* move to next char in source string */
763 len_remaining--;
764 }
765
766 ctoUCS_out:
767 return i;
768 }
This page took 0.054223 seconds and 5 git commands to generate.