blob: 41d9d0725f0f3a47d432aa754803a4596c0381ac [file] [log] [blame]
Pavel Shilovskyec2e4522011-12-27 16:12:43 +04001/*
2 * fs/cifs/smb2pdu.c
3 *
Steve Frenche4aa25e2012-10-01 12:26:22 -05004 * Copyright (C) International Business Machines Corp., 2009, 2012
Pavel Shilovskyec2e4522011-12-27 16:12:43 +04005 * Etersoft, 2012
6 * Author(s): Steve French (sfrench@us.ibm.com)
7 * Pavel Shilovsky (pshilovsky@samba.org) 2012
8 *
9 * Contains the routines for constructing the SMB2 PDUs themselves
10 *
11 * This library is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU Lesser General Public License as published
13 * by the Free Software Foundation; either version 2.1 of the License, or
14 * (at your option) any later version.
15 *
16 * This library is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
19 * the GNU Lesser General Public License for more details.
20 *
21 * You should have received a copy of the GNU Lesser General Public License
22 * along with this library; if not, write to the Free Software
23 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
24 */
25
26 /* SMB2 PDU handling routines here - except for leftovers (eg session setup) */
27 /* Note that there are handle based routines which must be */
28 /* treated slightly differently for reconnection purposes since we never */
29 /* want to reuse a stale file handle and only the caller knows the file info */
30
31#include <linux/fs.h>
32#include <linux/kernel.h>
33#include <linux/vfs.h>
Pavel Shilovsky09a47072012-09-18 16:20:29 -070034#include <linux/task_io_accounting_ops.h>
Pavel Shilovskyec2e4522011-12-27 16:12:43 +040035#include <linux/uaccess.h>
Pavel Shilovsky33319142012-09-18 16:20:29 -070036#include <linux/pagemap.h>
Pavel Shilovskyec2e4522011-12-27 16:12:43 +040037#include <linux/xattr.h>
38#include "smb2pdu.h"
39#include "cifsglob.h"
40#include "cifsacl.h"
41#include "cifsproto.h"
42#include "smb2proto.h"
43#include "cifs_unicode.h"
44#include "cifs_debug.h"
45#include "ntlmssp.h"
46#include "smb2status.h"
Pavel Shilovsky09a47072012-09-18 16:20:29 -070047#include "smb2glob.h"
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -070048#include "cifspdu.h"
Pavel Shilovskyec2e4522011-12-27 16:12:43 +040049
50/*
51 * The following table defines the expected "StructureSize" of SMB2 requests
52 * in order by SMB2 command. This is similar to "wct" in SMB/CIFS requests.
53 *
54 * Note that commands are defined in smb2pdu.h in le16 but the array below is
55 * indexed by command in host byte order.
56 */
57static const int smb2_req_struct_sizes[NUMBER_OF_SMB2_COMMANDS] = {
58 /* SMB2_NEGOTIATE */ 36,
59 /* SMB2_SESSION_SETUP */ 25,
60 /* SMB2_LOGOFF */ 4,
61 /* SMB2_TREE_CONNECT */ 9,
62 /* SMB2_TREE_DISCONNECT */ 4,
63 /* SMB2_CREATE */ 57,
64 /* SMB2_CLOSE */ 24,
65 /* SMB2_FLUSH */ 24,
66 /* SMB2_READ */ 49,
67 /* SMB2_WRITE */ 49,
68 /* SMB2_LOCK */ 48,
69 /* SMB2_IOCTL */ 57,
70 /* SMB2_CANCEL */ 4,
71 /* SMB2_ECHO */ 4,
72 /* SMB2_QUERY_DIRECTORY */ 33,
73 /* SMB2_CHANGE_NOTIFY */ 32,
74 /* SMB2_QUERY_INFO */ 41,
75 /* SMB2_SET_INFO */ 33,
76 /* SMB2_OPLOCK_BREAK */ 24 /* BB this is 36 for LEASE_BREAK variant */
77};
78
79
80static void
81smb2_hdr_assemble(struct smb2_hdr *hdr, __le16 smb2_cmd /* command */ ,
82 const struct cifs_tcon *tcon)
83{
84 struct smb2_pdu *pdu = (struct smb2_pdu *)hdr;
85 char *temp = (char *)hdr;
86 /* lookup word count ie StructureSize from table */
87 __u16 parmsize = smb2_req_struct_sizes[le16_to_cpu(smb2_cmd)];
88
89 /*
90 * smaller than SMALL_BUFFER_SIZE but bigger than fixed area of
91 * largest operations (Create)
92 */
93 memset(temp, 0, 256);
94
95 /* Note this is only network field converted to big endian */
96 hdr->smb2_buf_length = cpu_to_be32(parmsize + sizeof(struct smb2_hdr)
97 - 4 /* RFC 1001 length field itself not counted */);
98
99 hdr->ProtocolId[0] = 0xFE;
100 hdr->ProtocolId[1] = 'S';
101 hdr->ProtocolId[2] = 'M';
102 hdr->ProtocolId[3] = 'B';
103 hdr->StructureSize = cpu_to_le16(64);
104 hdr->Command = smb2_cmd;
105 hdr->CreditRequest = cpu_to_le16(2); /* BB make this dynamic */
106 hdr->ProcessId = cpu_to_le32((__u16)current->tgid);
107
108 if (!tcon)
109 goto out;
110
111 hdr->TreeId = tcon->tid;
112 /* Uid is not converted */
113 if (tcon->ses)
114 hdr->SessionId = tcon->ses->Suid;
115 /* BB check following DFS flags BB */
116 /* BB do we have to add check for SHI1005_FLAGS_DFS_ROOT too? */
Pavel Shilovskyfaaf9462011-12-27 16:04:00 +0400117 if (tcon->share_flags & SHI1005_FLAGS_DFS)
118 hdr->Flags |= SMB2_FLAGS_DFS_OPERATIONS;
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400119 /* BB how does SMB2 do case sensitive? */
120 /* if (tcon->nocase)
121 hdr->Flags |= SMBFLG_CASELESS; */
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -0700122 if (tcon->ses && tcon->ses->server &&
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400123 (tcon->ses->server->sec_mode & SECMODE_SIGN_REQUIRED))
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -0700124 hdr->Flags |= SMB2_FLAGS_SIGNED;
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400125out:
126 pdu->StructureSize2 = cpu_to_le16(parmsize);
127 return;
128}
129
130static int
131smb2_reconnect(__le16 smb2_command, struct cifs_tcon *tcon)
132{
133 int rc = 0;
Pavel Shilovskyaa24d1e2011-12-27 16:23:34 +0400134 struct nls_table *nls_codepage;
135 struct cifs_ses *ses;
136 struct TCP_Server_Info *server;
137
138 /*
139 * SMB2s NegProt, SessSetup, Logoff do not have tcon yet so
140 * check for tcp and smb session status done differently
141 * for those three - in the calling routine.
142 */
143 if (tcon == NULL)
144 return rc;
145
146 if (smb2_command == SMB2_TREE_CONNECT)
147 return rc;
148
149 if (tcon->tidStatus == CifsExiting) {
150 /*
151 * only tree disconnect, open, and write,
152 * (and ulogoff which does not have tcon)
153 * are allowed as we start force umount.
154 */
155 if ((smb2_command != SMB2_WRITE) &&
156 (smb2_command != SMB2_CREATE) &&
157 (smb2_command != SMB2_TREE_DISCONNECT)) {
158 cFYI(1, "can not send cmd %d while umounting",
159 smb2_command);
160 return -ENODEV;
161 }
162 }
163 if ((!tcon->ses) || (tcon->ses->status == CifsExiting) ||
164 (!tcon->ses->server))
165 return -EIO;
166
167 ses = tcon->ses;
168 server = ses->server;
169
170 /*
171 * Give demultiplex thread up to 10 seconds to reconnect, should be
172 * greater than cifs socket timeout which is 7 seconds
173 */
174 while (server->tcpStatus == CifsNeedReconnect) {
175 /*
176 * Return to caller for TREE_DISCONNECT and LOGOFF and CLOSE
177 * here since they are implicitly done when session drops.
178 */
179 switch (smb2_command) {
180 /*
181 * BB Should we keep oplock break and add flush to exceptions?
182 */
183 case SMB2_TREE_DISCONNECT:
184 case SMB2_CANCEL:
185 case SMB2_CLOSE:
186 case SMB2_OPLOCK_BREAK:
187 return -EAGAIN;
188 }
189
190 wait_event_interruptible_timeout(server->response_q,
191 (server->tcpStatus != CifsNeedReconnect), 10 * HZ);
192
193 /* are we still trying to reconnect? */
194 if (server->tcpStatus != CifsNeedReconnect)
195 break;
196
197 /*
198 * on "soft" mounts we wait once. Hard mounts keep
199 * retrying until process is killed or server comes
200 * back on-line
201 */
202 if (!tcon->retry) {
203 cFYI(1, "gave up waiting on reconnect in smb_init");
204 return -EHOSTDOWN;
205 }
206 }
207
208 if (!tcon->ses->need_reconnect && !tcon->need_reconnect)
209 return rc;
210
211 nls_codepage = load_nls_default();
212
213 /*
214 * need to prevent multiple threads trying to simultaneously reconnect
215 * the same SMB session
216 */
217 mutex_lock(&tcon->ses->session_mutex);
218 rc = cifs_negotiate_protocol(0, tcon->ses);
219 if (!rc && tcon->ses->need_reconnect)
220 rc = cifs_setup_session(0, tcon->ses, nls_codepage);
221
222 if (rc || !tcon->need_reconnect) {
223 mutex_unlock(&tcon->ses->session_mutex);
224 goto out;
225 }
226
227 cifs_mark_open_files_invalid(tcon);
228 rc = SMB2_tcon(0, tcon->ses, tcon->treeName, tcon, nls_codepage);
229 mutex_unlock(&tcon->ses->session_mutex);
230 cFYI(1, "reconnect tcon rc = %d", rc);
231 if (rc)
232 goto out;
233 atomic_inc(&tconInfoReconnectCount);
234 /*
235 * BB FIXME add code to check if wsize needs update due to negotiated
236 * smb buffer size shrinking.
237 */
238out:
239 /*
240 * Check if handle based operation so we know whether we can continue
241 * or not without returning to caller to reset file handle.
242 */
243 /*
244 * BB Is flush done by server on drop of tcp session? Should we special
245 * case it and skip above?
246 */
247 switch (smb2_command) {
248 case SMB2_FLUSH:
249 case SMB2_READ:
250 case SMB2_WRITE:
251 case SMB2_LOCK:
252 case SMB2_IOCTL:
253 case SMB2_QUERY_DIRECTORY:
254 case SMB2_CHANGE_NOTIFY:
255 case SMB2_QUERY_INFO:
256 case SMB2_SET_INFO:
257 return -EAGAIN;
258 }
259 unload_nls(nls_codepage);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400260 return rc;
261}
262
263/*
264 * Allocate and return pointer to an SMB request hdr, and set basic
265 * SMB information in the SMB header. If the return code is zero, this
266 * function must have filled in request_buf pointer.
267 */
268static int
269small_smb2_init(__le16 smb2_command, struct cifs_tcon *tcon,
270 void **request_buf)
271{
272 int rc = 0;
273
274 rc = smb2_reconnect(smb2_command, tcon);
275 if (rc)
276 return rc;
277
278 /* BB eventually switch this to SMB2 specific small buf size */
279 *request_buf = cifs_small_buf_get();
280 if (*request_buf == NULL) {
281 /* BB should we add a retry in here if not a writepage? */
282 return -ENOMEM;
283 }
284
285 smb2_hdr_assemble((struct smb2_hdr *) *request_buf, smb2_command, tcon);
286
287 if (tcon != NULL) {
288#ifdef CONFIG_CIFS_STATS2
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400289 uint16_t com_code = le16_to_cpu(smb2_command);
290 cifs_stats_inc(&tcon->stats.smb2_stats.smb2_com_sent[com_code]);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400291#endif
292 cifs_stats_inc(&tcon->num_smbs_sent);
293 }
294
295 return rc;
296}
297
298static void
299free_rsp_buf(int resp_buftype, void *rsp)
300{
301 if (resp_buftype == CIFS_SMALL_BUFFER)
302 cifs_small_buf_release(rsp);
303 else if (resp_buftype == CIFS_LARGE_BUFFER)
304 cifs_buf_release(rsp);
305}
306
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400307
308/*
309 *
310 * SMB2 Worker functions follow:
311 *
312 * The general structure of the worker functions is:
313 * 1) Call smb2_init (assembles SMB2 header)
314 * 2) Initialize SMB2 command specific fields in fixed length area of SMB
315 * 3) Call smb_sendrcv2 (sends request on socket and waits for response)
316 * 4) Decode SMB2 command specific fields in the fixed length area
317 * 5) Decode variable length data area (if any for this SMB2 command type)
318 * 6) Call free smb buffer
319 * 7) return
320 *
321 */
322
323int
324SMB2_negotiate(const unsigned int xid, struct cifs_ses *ses)
325{
326 struct smb2_negotiate_req *req;
327 struct smb2_negotiate_rsp *rsp;
328 struct kvec iov[1];
329 int rc = 0;
330 int resp_buftype;
331 struct TCP_Server_Info *server;
332 unsigned int sec_flags;
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400333 u16 temp = 0;
334 int blob_offset, blob_length;
335 char *security_blob;
336 int flags = CIFS_NEG_OP;
337
338 cFYI(1, "Negotiate protocol");
339
340 if (ses->server)
341 server = ses->server;
342 else {
343 rc = -EIO;
344 return rc;
345 }
346
347 rc = small_smb2_init(SMB2_NEGOTIATE, NULL, (void **) &req);
348 if (rc)
349 return rc;
350
351 /* if any of auth flags (ie not sign or seal) are overriden use them */
352 if (ses->overrideSecFlg & (~(CIFSSEC_MUST_SIGN | CIFSSEC_MUST_SEAL)))
353 sec_flags = ses->overrideSecFlg; /* BB FIXME fix sign flags?*/
354 else /* if override flags set only sign/seal OR them with global auth */
355 sec_flags = global_secflags | ses->overrideSecFlg;
356
357 cFYI(1, "sec_flags 0x%x", sec_flags);
358
359 req->hdr.SessionId = 0;
360
Steve Frenche4aa25e2012-10-01 12:26:22 -0500361 req->Dialects[0] = cpu_to_le16(ses->server->vals->protocol_id);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400362
Steve Frenche4aa25e2012-10-01 12:26:22 -0500363 req->DialectCount = cpu_to_le16(1); /* One vers= at a time for now */
364 inc_rfc1001_len(req, 2);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400365
366 /* only one of SMB2 signing flags may be set in SMB2 request */
367 if ((sec_flags & CIFSSEC_MUST_SIGN) == CIFSSEC_MUST_SIGN)
368 temp = SMB2_NEGOTIATE_SIGNING_REQUIRED;
369 else if (sec_flags & CIFSSEC_MAY_SIGN) /* MAY_SIGN is a single flag */
370 temp = SMB2_NEGOTIATE_SIGNING_ENABLED;
371
372 req->SecurityMode = cpu_to_le16(temp);
373
Steve Frenche4aa25e2012-10-01 12:26:22 -0500374 req->Capabilities = cpu_to_le32(ses->server->vals->req_capabilities);
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400375
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700376 memcpy(req->ClientGUID, cifs_client_guid, SMB2_CLIENT_GUID_SIZE);
377
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400378 iov[0].iov_base = (char *)req;
379 /* 4 for rfc1002 length field */
380 iov[0].iov_len = get_rfc1002_length(req) + 4;
381
382 rc = SendReceive2(xid, ses, iov, 1, &resp_buftype, flags);
383
384 rsp = (struct smb2_negotiate_rsp *)iov[0].iov_base;
385 /*
386 * No tcon so can't do
387 * cifs_stats_inc(&tcon->stats.smb2_stats.smb2_com_fail[SMB2...]);
388 */
389 if (rc != 0)
390 goto neg_exit;
391
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400392 cFYI(1, "mode 0x%x", rsp->SecurityMode);
393
Steve Frenche4aa25e2012-10-01 12:26:22 -0500394 /* BB we may eventually want to match the negotiated vs. requested
395 dialect, even though we are only requesting one at a time */
396 if (rsp->DialectRevision == cpu_to_le16(SMB20_PROT_ID))
397 cFYI(1, "negotiated smb2.0 dialect");
398 else if (rsp->DialectRevision == cpu_to_le16(SMB21_PROT_ID))
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400399 cFYI(1, "negotiated smb2.1 dialect");
Steve Frenche4aa25e2012-10-01 12:26:22 -0500400 else if (rsp->DialectRevision == cpu_to_le16(SMB30_PROT_ID))
401 cFYI(1, "negotiated smb3.0 dialect");
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400402 else {
403 cERROR(1, "Illegal dialect returned by server %d",
404 le16_to_cpu(rsp->DialectRevision));
405 rc = -EIO;
406 goto neg_exit;
407 }
408 server->dialect = le16_to_cpu(rsp->DialectRevision);
409
410 server->maxBuf = le32_to_cpu(rsp->MaxTransactSize);
411 server->max_read = le32_to_cpu(rsp->MaxReadSize);
412 server->max_write = le32_to_cpu(rsp->MaxWriteSize);
413 /* BB Do we need to validate the SecurityMode? */
414 server->sec_mode = le16_to_cpu(rsp->SecurityMode);
415 server->capabilities = le32_to_cpu(rsp->Capabilities);
Pavel Shilovsky29e20f92012-07-13 13:58:14 +0400416 /* Internal types */
417 server->capabilities |= SMB2_NT_FIND | SMB2_LARGE_FILES;
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400418
419 security_blob = smb2_get_data_area_len(&blob_offset, &blob_length,
420 &rsp->hdr);
421 if (blob_length == 0) {
422 cERROR(1, "missing security blob on negprot");
423 rc = -EIO;
424 goto neg_exit;
425 }
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -0700426
427 cFYI(1, "sec_flags 0x%x", sec_flags);
Steve French52c0f4a2012-12-04 16:56:37 -0600428 if ((sec_flags & CIFSSEC_MUST_SIGN) == CIFSSEC_MUST_SIGN) {
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -0700429 cFYI(1, "Signing required");
430 if (!(server->sec_mode & (SMB2_NEGOTIATE_SIGNING_REQUIRED |
431 SMB2_NEGOTIATE_SIGNING_ENABLED))) {
432 cERROR(1, "signing required but server lacks support");
433 rc = -EOPNOTSUPP;
434 goto neg_exit;
435 }
436 server->sec_mode |= SECMODE_SIGN_REQUIRED;
437 } else if (sec_flags & CIFSSEC_MAY_SIGN) {
438 cFYI(1, "Signing optional");
439 if (server->sec_mode & SMB2_NEGOTIATE_SIGNING_REQUIRED) {
440 cFYI(1, "Server requires signing");
441 server->sec_mode |= SECMODE_SIGN_REQUIRED;
442 } else {
443 server->sec_mode &=
444 ~(SECMODE_SIGN_ENABLED | SECMODE_SIGN_REQUIRED);
445 }
446 } else {
447 cFYI(1, "Signing disabled");
448 if (server->sec_mode & SMB2_NEGOTIATE_SIGNING_REQUIRED) {
449 cERROR(1, "Server requires packet signing to be enabled"
450 " in /proc/fs/cifs/SecurityFlags.");
451 rc = -EOPNOTSUPP;
452 goto neg_exit;
453 }
454 server->sec_mode &=
455 ~(SECMODE_SIGN_ENABLED | SECMODE_SIGN_REQUIRED);
456 }
457
Pavel Shilovskyec2e4522011-12-27 16:12:43 +0400458#ifdef CONFIG_SMB2_ASN1 /* BB REMOVEME when updated asn1.c ready */
459 rc = decode_neg_token_init(security_blob, blob_length,
460 &server->sec_type);
461 if (rc == 1)
462 rc = 0;
463 else if (rc == 0) {
464 rc = -EIO;
465 goto neg_exit;
466 }
467#endif
468
469neg_exit:
470 free_rsp_buf(resp_buftype, rsp);
471 return rc;
472}
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400473
474int
475SMB2_sess_setup(const unsigned int xid, struct cifs_ses *ses,
476 const struct nls_table *nls_cp)
477{
478 struct smb2_sess_setup_req *req;
479 struct smb2_sess_setup_rsp *rsp = NULL;
480 struct kvec iov[2];
481 int rc = 0;
482 int resp_buftype;
483 __le32 phase = NtLmNegotiate; /* NTLMSSP, if needed, is multistage */
484 struct TCP_Server_Info *server;
485 unsigned int sec_flags;
486 u8 temp = 0;
487 u16 blob_length = 0;
488 char *security_blob;
489 char *ntlmssp_blob = NULL;
490 bool use_spnego = false; /* else use raw ntlmssp */
491
492 cFYI(1, "Session Setup");
493
494 if (ses->server)
495 server = ses->server;
496 else {
497 rc = -EIO;
498 return rc;
499 }
500
501 /*
502 * If memory allocation is successful, caller of this function
503 * frees it.
504 */
505 ses->ntlmssp = kmalloc(sizeof(struct ntlmssp_auth), GFP_KERNEL);
506 if (!ses->ntlmssp)
507 return -ENOMEM;
508
509 ses->server->secType = RawNTLMSSP;
510
511ssetup_ntlmssp_authenticate:
512 if (phase == NtLmChallenge)
513 phase = NtLmAuthenticate; /* if ntlmssp, now final phase */
514
515 rc = small_smb2_init(SMB2_SESSION_SETUP, NULL, (void **) &req);
516 if (rc)
517 return rc;
518
519 /* if any of auth flags (ie not sign or seal) are overriden use them */
520 if (ses->overrideSecFlg & (~(CIFSSEC_MUST_SIGN | CIFSSEC_MUST_SEAL)))
521 sec_flags = ses->overrideSecFlg; /* BB FIXME fix sign flags?*/
522 else /* if override flags set only sign/seal OR them with global auth */
523 sec_flags = global_secflags | ses->overrideSecFlg;
524
525 cFYI(1, "sec_flags 0x%x", sec_flags);
526
527 req->hdr.SessionId = 0; /* First session, not a reauthenticate */
528 req->VcNumber = 0; /* MBZ */
529 /* to enable echos and oplocks */
530 req->hdr.CreditRequest = cpu_to_le16(3);
531
532 /* only one of SMB2 signing flags may be set in SMB2 request */
533 if ((sec_flags & CIFSSEC_MUST_SIGN) == CIFSSEC_MUST_SIGN)
534 temp = SMB2_NEGOTIATE_SIGNING_REQUIRED;
535 else if (ses->server->sec_mode & SMB2_NEGOTIATE_SIGNING_REQUIRED)
536 temp = SMB2_NEGOTIATE_SIGNING_REQUIRED;
537 else if (sec_flags & CIFSSEC_MAY_SIGN) /* MAY_SIGN is a single flag */
538 temp = SMB2_NEGOTIATE_SIGNING_ENABLED;
539
540 req->SecurityMode = temp;
541 req->Capabilities = 0;
542 req->Channel = 0; /* MBZ */
543
544 iov[0].iov_base = (char *)req;
545 /* 4 for rfc1002 length field and 1 for pad */
546 iov[0].iov_len = get_rfc1002_length(req) + 4 - 1;
547 if (phase == NtLmNegotiate) {
548 ntlmssp_blob = kmalloc(sizeof(struct _NEGOTIATE_MESSAGE),
549 GFP_KERNEL);
550 if (ntlmssp_blob == NULL) {
551 rc = -ENOMEM;
552 goto ssetup_exit;
553 }
554 build_ntlmssp_negotiate_blob(ntlmssp_blob, ses);
555 if (use_spnego) {
556 /* blob_length = build_spnego_ntlmssp_blob(
557 &security_blob,
558 sizeof(struct _NEGOTIATE_MESSAGE),
559 ntlmssp_blob); */
560 /* BB eventually need to add this */
561 cERROR(1, "spnego not supported for SMB2 yet");
562 rc = -EOPNOTSUPP;
563 kfree(ntlmssp_blob);
564 goto ssetup_exit;
565 } else {
566 blob_length = sizeof(struct _NEGOTIATE_MESSAGE);
567 /* with raw NTLMSSP we don't encapsulate in SPNEGO */
568 security_blob = ntlmssp_blob;
569 }
570 } else if (phase == NtLmAuthenticate) {
571 req->hdr.SessionId = ses->Suid;
572 ntlmssp_blob = kzalloc(sizeof(struct _NEGOTIATE_MESSAGE) + 500,
573 GFP_KERNEL);
574 if (ntlmssp_blob == NULL) {
575 cERROR(1, "failed to malloc ntlmssp blob");
576 rc = -ENOMEM;
577 goto ssetup_exit;
578 }
579 rc = build_ntlmssp_auth_blob(ntlmssp_blob, &blob_length, ses,
580 nls_cp);
581 if (rc) {
582 cFYI(1, "build_ntlmssp_auth_blob failed %d", rc);
583 goto ssetup_exit; /* BB double check error handling */
584 }
585 if (use_spnego) {
586 /* blob_length = build_spnego_ntlmssp_blob(
587 &security_blob,
588 blob_length,
589 ntlmssp_blob); */
590 cERROR(1, "spnego not supported for SMB2 yet");
591 rc = -EOPNOTSUPP;
592 kfree(ntlmssp_blob);
593 goto ssetup_exit;
594 } else {
595 security_blob = ntlmssp_blob;
596 }
597 } else {
598 cERROR(1, "illegal ntlmssp phase");
599 rc = -EIO;
600 goto ssetup_exit;
601 }
602
603 /* Testing shows that buffer offset must be at location of Buffer[0] */
604 req->SecurityBufferOffset =
605 cpu_to_le16(sizeof(struct smb2_sess_setup_req) -
606 1 /* pad */ - 4 /* rfc1001 len */);
607 req->SecurityBufferLength = cpu_to_le16(blob_length);
608 iov[1].iov_base = security_blob;
609 iov[1].iov_len = blob_length;
610
611 inc_rfc1001_len(req, blob_length - 1 /* pad */);
612
613 /* BB add code to build os and lm fields */
614
Steve French6d8b59d2012-12-08 22:36:29 -0600615 rc = SendReceive2(xid, ses, iov, 2, &resp_buftype,
616 CIFS_LOG_ERROR | CIFS_NEG_OP);
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400617
618 kfree(security_blob);
619 rsp = (struct smb2_sess_setup_rsp *)iov[0].iov_base;
Pavel Shilovsky4ca3a992012-09-25 11:00:09 +0400620 if (resp_buftype != CIFS_NO_BUFFER &&
621 rsp->hdr.Status == STATUS_MORE_PROCESSING_REQUIRED) {
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400622 if (phase != NtLmNegotiate) {
623 cERROR(1, "Unexpected more processing error");
624 goto ssetup_exit;
625 }
626 if (offsetof(struct smb2_sess_setup_rsp, Buffer) - 4 !=
Pavel Shilovsky4ca3a992012-09-25 11:00:09 +0400627 le16_to_cpu(rsp->SecurityBufferOffset)) {
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400628 cERROR(1, "Invalid security buffer offset %d",
629 le16_to_cpu(rsp->SecurityBufferOffset));
630 rc = -EIO;
631 goto ssetup_exit;
632 }
633
634 /* NTLMSSP Negotiate sent now processing challenge (response) */
635 phase = NtLmChallenge; /* process ntlmssp challenge */
636 rc = 0; /* MORE_PROCESSING is not an error here but expected */
637 ses->Suid = rsp->hdr.SessionId;
638 rc = decode_ntlmssp_challenge(rsp->Buffer,
639 le16_to_cpu(rsp->SecurityBufferLength), ses);
640 }
641
642 /*
643 * BB eventually add code for SPNEGO decoding of NtlmChallenge blob,
644 * but at least the raw NTLMSSP case works.
645 */
646 /*
647 * No tcon so can't do
648 * cifs_stats_inc(&tcon->stats.smb2_stats.smb2_com_fail[SMB2...]);
649 */
650 if (rc != 0)
651 goto ssetup_exit;
652
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400653 ses->session_flags = le16_to_cpu(rsp->SessionFlags);
654ssetup_exit:
655 free_rsp_buf(resp_buftype, rsp);
656
657 /* if ntlmssp, and negotiate succeeded, proceed to authenticate phase */
658 if ((phase == NtLmChallenge) && (rc == 0))
659 goto ssetup_ntlmssp_authenticate;
660 return rc;
661}
662
663int
664SMB2_logoff(const unsigned int xid, struct cifs_ses *ses)
665{
666 struct smb2_logoff_req *req; /* response is also trivial struct */
667 int rc = 0;
668 struct TCP_Server_Info *server;
669
670 cFYI(1, "disconnect session %p", ses);
671
672 if (ses && (ses->server))
673 server = ses->server;
674 else
675 return -EIO;
676
677 rc = small_smb2_init(SMB2_LOGOFF, NULL, (void **) &req);
678 if (rc)
679 return rc;
680
681 /* since no tcon, smb2_init can not do this, so do here */
682 req->hdr.SessionId = ses->Suid;
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -0700683 if (server->sec_mode & SECMODE_SIGN_REQUIRED)
684 req->hdr.Flags |= SMB2_FLAGS_SIGNED;
Pavel Shilovsky5478f9b2011-12-27 16:22:00 +0400685
686 rc = SendReceiveNoRsp(xid, ses, (char *) &req->hdr, 0);
687 /*
688 * No tcon so can't do
689 * cifs_stats_inc(&tcon->stats.smb2_stats.smb2_com_fail[SMB2...]);
690 */
691 return rc;
692}
Pavel Shilovskyfaaf9462011-12-27 16:04:00 +0400693
694static inline void cifs_stats_fail_inc(struct cifs_tcon *tcon, uint16_t code)
695{
Pavel Shilovskyd60622e2012-05-28 15:19:39 +0400696 cifs_stats_inc(&tcon->stats.smb2_stats.smb2_com_failed[code]);
Pavel Shilovskyfaaf9462011-12-27 16:04:00 +0400697}
698
699#define MAX_SHARENAME_LENGTH (255 /* server */ + 80 /* share */ + 1 /* NULL */)
700
701int
702SMB2_tcon(const unsigned int xid, struct cifs_ses *ses, const char *tree,
703 struct cifs_tcon *tcon, const struct nls_table *cp)
704{
705 struct smb2_tree_connect_req *req;
706 struct smb2_tree_connect_rsp *rsp = NULL;
707 struct kvec iov[2];
708 int rc = 0;
709 int resp_buftype;
710 int unc_path_len;
711 struct TCP_Server_Info *server;
712 __le16 *unc_path = NULL;
713
714 cFYI(1, "TCON");
715
716 if ((ses->server) && tree)
717 server = ses->server;
718 else
719 return -EIO;
720
721 if (tcon && tcon->bad_network_name)
722 return -ENOENT;
723
724 unc_path = kmalloc(MAX_SHARENAME_LENGTH * 2, GFP_KERNEL);
725 if (unc_path == NULL)
726 return -ENOMEM;
727
728 unc_path_len = cifs_strtoUTF16(unc_path, tree, strlen(tree), cp) + 1;
729 unc_path_len *= 2;
730 if (unc_path_len < 2) {
731 kfree(unc_path);
732 return -EINVAL;
733 }
734
735 rc = small_smb2_init(SMB2_TREE_CONNECT, tcon, (void **) &req);
736 if (rc) {
737 kfree(unc_path);
738 return rc;
739 }
740
741 if (tcon == NULL) {
742 /* since no tcon, smb2_init can not do this, so do here */
743 req->hdr.SessionId = ses->Suid;
744 /* if (ses->server->sec_mode & SECMODE_SIGN_REQUIRED)
745 req->hdr.Flags |= SMB2_FLAGS_SIGNED; */
746 }
747
748 iov[0].iov_base = (char *)req;
749 /* 4 for rfc1002 length field and 1 for pad */
750 iov[0].iov_len = get_rfc1002_length(req) + 4 - 1;
751
752 /* Testing shows that buffer offset must be at location of Buffer[0] */
753 req->PathOffset = cpu_to_le16(sizeof(struct smb2_tree_connect_req)
754 - 1 /* pad */ - 4 /* do not count rfc1001 len field */);
755 req->PathLength = cpu_to_le16(unc_path_len - 2);
756 iov[1].iov_base = unc_path;
757 iov[1].iov_len = unc_path_len;
758
759 inc_rfc1001_len(req, unc_path_len - 1 /* pad */);
760
761 rc = SendReceive2(xid, ses, iov, 2, &resp_buftype, 0);
762 rsp = (struct smb2_tree_connect_rsp *)iov[0].iov_base;
763
764 if (rc != 0) {
765 if (tcon) {
766 cifs_stats_fail_inc(tcon, SMB2_TREE_CONNECT_HE);
767 tcon->need_reconnect = true;
768 }
769 goto tcon_error_exit;
770 }
771
Pavel Shilovskyfaaf9462011-12-27 16:04:00 +0400772 if (tcon == NULL) {
773 ses->ipc_tid = rsp->hdr.TreeId;
774 goto tcon_exit;
775 }
776
777 if (rsp->ShareType & SMB2_SHARE_TYPE_DISK)
778 cFYI(1, "connection to disk share");
779 else if (rsp->ShareType & SMB2_SHARE_TYPE_PIPE) {
780 tcon->ipc = true;
781 cFYI(1, "connection to pipe share");
782 } else if (rsp->ShareType & SMB2_SHARE_TYPE_PRINT) {
783 tcon->print = true;
784 cFYI(1, "connection to printer");
785 } else {
786 cERROR(1, "unknown share type %d", rsp->ShareType);
787 rc = -EOPNOTSUPP;
788 goto tcon_error_exit;
789 }
790
791 tcon->share_flags = le32_to_cpu(rsp->ShareFlags);
792 tcon->maximal_access = le32_to_cpu(rsp->MaximalAccess);
793 tcon->tidStatus = CifsGood;
794 tcon->need_reconnect = false;
795 tcon->tid = rsp->hdr.TreeId;
796 strncpy(tcon->treeName, tree, MAX_TREE_SIZE);
797
798 if ((rsp->Capabilities & SMB2_SHARE_CAP_DFS) &&
799 ((tcon->share_flags & SHI1005_FLAGS_DFS) == 0))
800 cERROR(1, "DFS capability contradicts DFS flag");
801
802tcon_exit:
803 free_rsp_buf(resp_buftype, rsp);
804 kfree(unc_path);
805 return rc;
806
807tcon_error_exit:
808 if (rsp->hdr.Status == STATUS_BAD_NETWORK_NAME) {
809 cERROR(1, "BAD_NETWORK_NAME: %s", tree);
810 tcon->bad_network_name = true;
811 }
812 goto tcon_exit;
813}
814
815int
816SMB2_tdis(const unsigned int xid, struct cifs_tcon *tcon)
817{
818 struct smb2_tree_disconnect_req *req; /* response is trivial */
819 int rc = 0;
820 struct TCP_Server_Info *server;
821 struct cifs_ses *ses = tcon->ses;
822
823 cFYI(1, "Tree Disconnect");
824
825 if (ses && (ses->server))
826 server = ses->server;
827 else
828 return -EIO;
829
830 if ((tcon->need_reconnect) || (tcon->ses->need_reconnect))
831 return 0;
832
833 rc = small_smb2_init(SMB2_TREE_DISCONNECT, tcon, (void **) &req);
834 if (rc)
835 return rc;
836
837 rc = SendReceiveNoRsp(xid, ses, (char *)&req->hdr, 0);
838 if (rc)
839 cifs_stats_fail_inc(tcon, SMB2_TREE_DISCONNECT_HE);
840
841 return rc;
842}
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400843
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700844static struct create_lease *
845create_lease_buf(u8 *lease_key, u8 oplock)
846{
847 struct create_lease *buf;
848
849 buf = kmalloc(sizeof(struct create_lease), GFP_KERNEL);
850 if (!buf)
851 return NULL;
852
853 memset(buf, 0, sizeof(struct create_lease));
854
855 buf->lcontext.LeaseKeyLow = cpu_to_le64(*((u64 *)lease_key));
856 buf->lcontext.LeaseKeyHigh = cpu_to_le64(*((u64 *)(lease_key + 8)));
857 if (oplock == SMB2_OPLOCK_LEVEL_EXCLUSIVE)
858 buf->lcontext.LeaseState = SMB2_LEASE_WRITE_CACHING |
859 SMB2_LEASE_READ_CACHING;
860 else if (oplock == SMB2_OPLOCK_LEVEL_II)
861 buf->lcontext.LeaseState = SMB2_LEASE_READ_CACHING;
862 else if (oplock == SMB2_OPLOCK_LEVEL_BATCH)
863 buf->lcontext.LeaseState = SMB2_LEASE_HANDLE_CACHING |
864 SMB2_LEASE_READ_CACHING |
865 SMB2_LEASE_WRITE_CACHING;
866
867 buf->ccontext.DataOffset = cpu_to_le16(offsetof
868 (struct create_lease, lcontext));
869 buf->ccontext.DataLength = cpu_to_le32(sizeof(struct lease_context));
870 buf->ccontext.NameOffset = cpu_to_le16(offsetof
871 (struct create_lease, Name));
872 buf->ccontext.NameLength = cpu_to_le16(4);
873 buf->Name[0] = 'R';
874 buf->Name[1] = 'q';
875 buf->Name[2] = 'L';
876 buf->Name[3] = 's';
877 return buf;
878}
879
880static __u8
881parse_lease_state(struct smb2_create_rsp *rsp)
882{
883 char *data_offset;
884 struct create_lease *lc;
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700885 bool found = false;
886
887 data_offset = (char *)rsp;
888 data_offset += 4 + le32_to_cpu(rsp->CreateContextsOffset);
889 lc = (struct create_lease *)data_offset;
890 do {
891 char *name = le16_to_cpu(lc->ccontext.NameOffset) + (char *)lc;
892 if (le16_to_cpu(lc->ccontext.NameLength) != 4 ||
893 strncmp(name, "RqLs", 4)) {
894 lc = (struct create_lease *)((char *)lc
895 + le32_to_cpu(lc->ccontext.Next));
896 continue;
897 }
898 if (lc->lcontext.LeaseFlags & SMB2_LEASE_FLAG_BREAK_IN_PROGRESS)
899 return SMB2_OPLOCK_LEVEL_NOCHANGE;
900 found = true;
901 break;
902 } while (le32_to_cpu(lc->ccontext.Next) != 0);
903
904 if (!found)
Pavel Shilovsky0822f512012-09-19 06:22:45 -0700905 return 0;
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700906
Pavel Shilovsky0822f512012-09-19 06:22:45 -0700907 return smb2_map_lease_to_oplock(lc->lcontext.LeaseState);
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700908}
909
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400910int
911SMB2_open(const unsigned int xid, struct cifs_tcon *tcon, __le16 *path,
912 u64 *persistent_fid, u64 *volatile_fid, __u32 desired_access,
Pavel Shilovskyf0df7372012-09-18 16:20:26 -0700913 __u32 create_disposition, __u32 file_attributes, __u32 create_options,
Pavel Shilovsky2e44b282012-09-18 16:20:33 -0700914 __u8 *oplock, struct smb2_file_all_info *buf)
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400915{
916 struct smb2_create_req *req;
917 struct smb2_create_rsp *rsp;
918 struct TCP_Server_Info *server;
919 struct cifs_ses *ses = tcon->ses;
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700920 struct kvec iov[3];
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400921 int resp_buftype;
922 int uni_path_len;
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700923 __le16 *copy_path = NULL;
924 int copy_size;
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400925 int rc = 0;
926 int num_iovecs = 2;
927
928 cFYI(1, "create/open");
929
930 if (ses && (ses->server))
931 server = ses->server;
932 else
933 return -EIO;
934
935 rc = small_smb2_init(SMB2_CREATE, tcon, (void **) &req);
936 if (rc)
937 return rc;
938
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400939 req->ImpersonationLevel = IL_IMPERSONATION;
940 req->DesiredAccess = cpu_to_le32(desired_access);
941 /* File attributes ignored on open (used in create though) */
942 req->FileAttributes = cpu_to_le32(file_attributes);
943 req->ShareAccess = FILE_SHARE_ALL_LE;
944 req->CreateDisposition = cpu_to_le32(create_disposition);
945 req->CreateOptions = cpu_to_le32(create_options);
946 uni_path_len = (2 * UniStrnlen((wchar_t *)path, PATH_MAX)) + 2;
947 req->NameOffset = cpu_to_le16(sizeof(struct smb2_create_req)
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700948 - 8 /* pad */ - 4 /* do not count rfc1001 len field */);
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400949
950 iov[0].iov_base = (char *)req;
951 /* 4 for rfc1002 length field */
952 iov[0].iov_len = get_rfc1002_length(req) + 4;
953
954 /* MUST set path len (NameLength) to 0 opening root of share */
955 if (uni_path_len >= 4) {
956 req->NameLength = cpu_to_le16(uni_path_len - 2);
957 /* -1 since last byte is buf[0] which is sent below (path) */
958 iov[0].iov_len--;
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700959 if (uni_path_len % 8 != 0) {
960 copy_size = uni_path_len / 8 * 8;
961 if (copy_size < uni_path_len)
962 copy_size += 8;
963
964 copy_path = kzalloc(copy_size, GFP_KERNEL);
965 if (!copy_path)
966 return -ENOMEM;
967 memcpy((char *)copy_path, (const char *)path,
968 uni_path_len);
969 uni_path_len = copy_size;
970 path = copy_path;
971 }
972
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400973 iov[1].iov_len = uni_path_len;
974 iov[1].iov_base = path;
975 /*
976 * -1 since last byte is buf[0] which was counted in
977 * smb2_buf_len.
978 */
979 inc_rfc1001_len(req, uni_path_len - 1);
980 } else {
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700981 iov[0].iov_len += 7;
982 req->hdr.smb2_buf_length = cpu_to_be32(be32_to_cpu(
983 req->hdr.smb2_buf_length) + 8 - 1);
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +0400984 num_iovecs = 1;
985 req->NameLength = 0;
986 }
987
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -0700988 if (!server->oplocks)
989 *oplock = SMB2_OPLOCK_LEVEL_NONE;
990
991 if (!(tcon->ses->server->capabilities & SMB2_GLOBAL_CAP_LEASING) ||
992 *oplock == SMB2_OPLOCK_LEVEL_NONE)
993 req->RequestedOplockLevel = *oplock;
994 else {
995 iov[num_iovecs].iov_base = create_lease_buf(oplock+1, *oplock);
996 if (iov[num_iovecs].iov_base == NULL) {
997 cifs_small_buf_release(req);
998 kfree(copy_path);
999 return -ENOMEM;
1000 }
1001 iov[num_iovecs].iov_len = sizeof(struct create_lease);
1002 req->RequestedOplockLevel = SMB2_OPLOCK_LEVEL_LEASE;
1003 req->CreateContextsOffset = cpu_to_le32(
1004 sizeof(struct smb2_create_req) - 4 - 8 +
1005 iov[num_iovecs-1].iov_len);
1006 req->CreateContextsLength = cpu_to_le32(
1007 sizeof(struct create_lease));
1008 inc_rfc1001_len(&req->hdr, sizeof(struct create_lease));
1009 num_iovecs++;
1010 }
1011
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001012 rc = SendReceive2(xid, ses, iov, num_iovecs, &resp_buftype, 0);
1013 rsp = (struct smb2_create_rsp *)iov[0].iov_base;
1014
1015 if (rc != 0) {
1016 cifs_stats_fail_inc(tcon, SMB2_CREATE_HE);
1017 goto creat_exit;
1018 }
1019
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001020 *persistent_fid = rsp->PersistentFileId;
1021 *volatile_fid = rsp->VolatileFileId;
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001022
1023 if (buf) {
1024 memcpy(buf, &rsp->CreationTime, 32);
1025 buf->AllocationSize = rsp->AllocationSize;
1026 buf->EndOfFile = rsp->EndofFile;
1027 buf->Attributes = rsp->FileAttributes;
1028 buf->NumberOfLinks = cpu_to_le32(1);
1029 buf->DeletePending = 0;
1030 }
Pavel Shilovsky2e44b282012-09-18 16:20:33 -07001031
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -07001032 if (rsp->OplockLevel == SMB2_OPLOCK_LEVEL_LEASE)
1033 *oplock = parse_lease_state(rsp);
1034 else
1035 *oplock = rsp->OplockLevel;
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001036creat_exit:
Pavel Shilovskyb8c32db2012-09-19 06:22:44 -07001037 kfree(copy_path);
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001038 free_rsp_buf(resp_buftype, rsp);
1039 return rc;
1040}
1041
1042int
1043SMB2_close(const unsigned int xid, struct cifs_tcon *tcon,
1044 u64 persistent_fid, u64 volatile_fid)
1045{
1046 struct smb2_close_req *req;
1047 struct smb2_close_rsp *rsp;
1048 struct TCP_Server_Info *server;
1049 struct cifs_ses *ses = tcon->ses;
1050 struct kvec iov[1];
1051 int resp_buftype;
1052 int rc = 0;
1053
1054 cFYI(1, "Close");
1055
1056 if (ses && (ses->server))
1057 server = ses->server;
1058 else
1059 return -EIO;
1060
1061 rc = small_smb2_init(SMB2_CLOSE, tcon, (void **) &req);
1062 if (rc)
1063 return rc;
1064
1065 req->PersistentFileId = persistent_fid;
1066 req->VolatileFileId = volatile_fid;
1067
1068 iov[0].iov_base = (char *)req;
1069 /* 4 for rfc1002 length field */
1070 iov[0].iov_len = get_rfc1002_length(req) + 4;
1071
1072 rc = SendReceive2(xid, ses, iov, 1, &resp_buftype, 0);
1073 rsp = (struct smb2_close_rsp *)iov[0].iov_base;
1074
1075 if (rc != 0) {
1076 if (tcon)
1077 cifs_stats_fail_inc(tcon, SMB2_CLOSE_HE);
1078 goto close_exit;
1079 }
1080
Pavel Shilovsky2503a0d2011-12-26 22:58:46 +04001081 /* BB FIXME - decode close response, update inode for caching */
1082
1083close_exit:
1084 free_rsp_buf(resp_buftype, rsp);
1085 return rc;
1086}
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001087
1088static int
1089validate_buf(unsigned int offset, unsigned int buffer_length,
1090 struct smb2_hdr *hdr, unsigned int min_buf_size)
1091
1092{
1093 unsigned int smb_len = be32_to_cpu(hdr->smb2_buf_length);
1094 char *end_of_smb = smb_len + 4 /* RFC1001 length field */ + (char *)hdr;
1095 char *begin_of_buf = 4 /* RFC1001 len field */ + offset + (char *)hdr;
1096 char *end_of_buf = begin_of_buf + buffer_length;
1097
1098
1099 if (buffer_length < min_buf_size) {
1100 cERROR(1, "buffer length %d smaller than minimum size %d",
1101 buffer_length, min_buf_size);
1102 return -EINVAL;
1103 }
1104
1105 /* check if beyond RFC1001 maximum length */
1106 if ((smb_len > 0x7FFFFF) || (buffer_length > 0x7FFFFF)) {
1107 cERROR(1, "buffer length %d or smb length %d too large",
1108 buffer_length, smb_len);
1109 return -EINVAL;
1110 }
1111
1112 if ((begin_of_buf > end_of_smb) || (end_of_buf > end_of_smb)) {
1113 cERROR(1, "illegal server response, bad offset to data");
1114 return -EINVAL;
1115 }
1116
1117 return 0;
1118}
1119
1120/*
1121 * If SMB buffer fields are valid, copy into temporary buffer to hold result.
1122 * Caller must free buffer.
1123 */
1124static int
1125validate_and_copy_buf(unsigned int offset, unsigned int buffer_length,
1126 struct smb2_hdr *hdr, unsigned int minbufsize,
1127 char *data)
1128
1129{
1130 char *begin_of_buf = 4 /* RFC1001 len field */ + offset + (char *)hdr;
1131 int rc;
1132
1133 if (!data)
1134 return -EINVAL;
1135
1136 rc = validate_buf(offset, buffer_length, hdr, minbufsize);
1137 if (rc)
1138 return rc;
1139
1140 memcpy(data, begin_of_buf, buffer_length);
1141
1142 return 0;
1143}
1144
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001145static int
1146query_info(const unsigned int xid, struct cifs_tcon *tcon,
1147 u64 persistent_fid, u64 volatile_fid, u8 info_class,
1148 size_t output_len, size_t min_len, void *data)
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001149{
1150 struct smb2_query_info_req *req;
1151 struct smb2_query_info_rsp *rsp = NULL;
1152 struct kvec iov[2];
1153 int rc = 0;
1154 int resp_buftype;
1155 struct TCP_Server_Info *server;
1156 struct cifs_ses *ses = tcon->ses;
1157
1158 cFYI(1, "Query Info");
1159
1160 if (ses && (ses->server))
1161 server = ses->server;
1162 else
1163 return -EIO;
1164
1165 rc = small_smb2_init(SMB2_QUERY_INFO, tcon, (void **) &req);
1166 if (rc)
1167 return rc;
1168
1169 req->InfoType = SMB2_O_INFO_FILE;
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001170 req->FileInfoClass = info_class;
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001171 req->PersistentFileId = persistent_fid;
1172 req->VolatileFileId = volatile_fid;
1173 /* 4 for rfc1002 length field and 1 for Buffer */
1174 req->InputBufferOffset =
1175 cpu_to_le16(sizeof(struct smb2_query_info_req) - 1 - 4);
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001176 req->OutputBufferLength = cpu_to_le32(output_len);
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001177
1178 iov[0].iov_base = (char *)req;
1179 /* 4 for rfc1002 length field */
1180 iov[0].iov_len = get_rfc1002_length(req) + 4;
1181
1182 rc = SendReceive2(xid, ses, iov, 1, &resp_buftype, 0);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001183 rsp = (struct smb2_query_info_rsp *)iov[0].iov_base;
1184
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001185 if (rc) {
1186 cifs_stats_fail_inc(tcon, SMB2_QUERY_INFO_HE);
1187 goto qinf_exit;
1188 }
1189
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001190 rc = validate_and_copy_buf(le16_to_cpu(rsp->OutputBufferOffset),
1191 le32_to_cpu(rsp->OutputBufferLength),
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001192 &rsp->hdr, min_len, data);
Pavel Shilovskybe4cb9e2011-12-29 17:06:33 +04001193
1194qinf_exit:
1195 free_rsp_buf(resp_buftype, rsp);
1196 return rc;
1197}
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04001198
Pavel Shilovskyf0df7372012-09-18 16:20:26 -07001199int
1200SMB2_query_info(const unsigned int xid, struct cifs_tcon *tcon,
1201 u64 persistent_fid, u64 volatile_fid,
1202 struct smb2_file_all_info *data)
1203{
1204 return query_info(xid, tcon, persistent_fid, volatile_fid,
1205 FILE_ALL_INFORMATION,
1206 sizeof(struct smb2_file_all_info) + MAX_NAME * 2,
1207 sizeof(struct smb2_file_all_info), data);
1208}
1209
1210int
1211SMB2_get_srv_num(const unsigned int xid, struct cifs_tcon *tcon,
1212 u64 persistent_fid, u64 volatile_fid, __le64 *uniqueid)
1213{
1214 return query_info(xid, tcon, persistent_fid, volatile_fid,
1215 FILE_INTERNAL_INFORMATION,
1216 sizeof(struct smb2_file_internal_info),
1217 sizeof(struct smb2_file_internal_info), uniqueid);
1218}
1219
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04001220/*
1221 * This is a no-op for now. We're not really interested in the reply, but
1222 * rather in the fact that the server sent one and that server->lstrp
1223 * gets updated.
1224 *
1225 * FIXME: maybe we should consider checking that the reply matches request?
1226 */
1227static void
1228smb2_echo_callback(struct mid_q_entry *mid)
1229{
1230 struct TCP_Server_Info *server = mid->callback_data;
1231 struct smb2_echo_rsp *smb2 = (struct smb2_echo_rsp *)mid->resp_buf;
1232 unsigned int credits_received = 1;
1233
1234 if (mid->mid_state == MID_RESPONSE_RECEIVED)
1235 credits_received = le16_to_cpu(smb2->hdr.CreditRequest);
1236
1237 DeleteMidQEntry(mid);
1238 add_credits(server, credits_received, CIFS_ECHO_OP);
1239}
1240
1241int
1242SMB2_echo(struct TCP_Server_Info *server)
1243{
1244 struct smb2_echo_req *req;
1245 int rc = 0;
1246 struct kvec iov;
Jeff Laytonfec344e2012-09-18 16:20:35 -07001247 struct smb_rqst rqst = { .rq_iov = &iov,
1248 .rq_nvec = 1 };
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04001249
1250 cFYI(1, "In echo request");
1251
1252 rc = small_smb2_init(SMB2_ECHO, NULL, (void **)&req);
1253 if (rc)
1254 return rc;
1255
1256 req->hdr.CreditRequest = cpu_to_le16(1);
1257
1258 iov.iov_base = (char *)req;
1259 /* 4 for rfc1002 length field */
1260 iov.iov_len = get_rfc1002_length(req) + 4;
1261
Jeff Laytonfec344e2012-09-18 16:20:35 -07001262 rc = cifs_call_async(server, &rqst, NULL, smb2_echo_callback, server,
Pavel Shilovsky9094fad2012-07-12 18:30:44 +04001263 CIFS_ECHO_OP);
1264 if (rc)
1265 cFYI(1, "Echo request failed: %d", rc);
1266
1267 cifs_small_buf_release(req);
1268 return rc;
1269}
Pavel Shilovsky7a5cfb12012-09-18 16:20:28 -07001270
1271int
1272SMB2_flush(const unsigned int xid, struct cifs_tcon *tcon, u64 persistent_fid,
1273 u64 volatile_fid)
1274{
1275 struct smb2_flush_req *req;
1276 struct TCP_Server_Info *server;
1277 struct cifs_ses *ses = tcon->ses;
1278 struct kvec iov[1];
1279 int resp_buftype;
1280 int rc = 0;
1281
1282 cFYI(1, "Flush");
1283
1284 if (ses && (ses->server))
1285 server = ses->server;
1286 else
1287 return -EIO;
1288
1289 rc = small_smb2_init(SMB2_FLUSH, tcon, (void **) &req);
1290 if (rc)
1291 return rc;
1292
1293 req->PersistentFileId = persistent_fid;
1294 req->VolatileFileId = volatile_fid;
1295
1296 iov[0].iov_base = (char *)req;
1297 /* 4 for rfc1002 length field */
1298 iov[0].iov_len = get_rfc1002_length(req) + 4;
1299
1300 rc = SendReceive2(xid, ses, iov, 1, &resp_buftype, 0);
1301
1302 if ((rc != 0) && tcon)
1303 cifs_stats_fail_inc(tcon, SMB2_FLUSH_HE);
1304
1305 free_rsp_buf(resp_buftype, iov[0].iov_base);
1306 return rc;
1307}
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001308
1309/*
1310 * To form a chain of read requests, any read requests after the first should
1311 * have the end_of_chain boolean set to true.
1312 */
1313static int
1314smb2_new_read_req(struct kvec *iov, struct cifs_io_parms *io_parms,
1315 unsigned int remaining_bytes, int request_type)
1316{
1317 int rc = -EACCES;
1318 struct smb2_read_req *req = NULL;
1319
1320 rc = small_smb2_init(SMB2_READ, io_parms->tcon, (void **) &req);
1321 if (rc)
1322 return rc;
1323 if (io_parms->tcon->ses->server == NULL)
1324 return -ECONNABORTED;
1325
1326 req->hdr.ProcessId = cpu_to_le32(io_parms->pid);
1327
1328 req->PersistentFileId = io_parms->persistent_fid;
1329 req->VolatileFileId = io_parms->volatile_fid;
1330 req->ReadChannelInfoOffset = 0; /* reserved */
1331 req->ReadChannelInfoLength = 0; /* reserved */
1332 req->Channel = 0; /* reserved */
1333 req->MinimumCount = 0;
1334 req->Length = cpu_to_le32(io_parms->length);
1335 req->Offset = cpu_to_le64(io_parms->offset);
1336
1337 if (request_type & CHAINED_REQUEST) {
1338 if (!(request_type & END_OF_CHAIN)) {
1339 /* 4 for rfc1002 length field */
1340 req->hdr.NextCommand =
1341 cpu_to_le32(get_rfc1002_length(req) + 4);
1342 } else /* END_OF_CHAIN */
1343 req->hdr.NextCommand = 0;
1344 if (request_type & RELATED_REQUEST) {
1345 req->hdr.Flags |= SMB2_FLAGS_RELATED_OPERATIONS;
1346 /*
1347 * Related requests use info from previous read request
1348 * in chain.
1349 */
1350 req->hdr.SessionId = 0xFFFFFFFF;
1351 req->hdr.TreeId = 0xFFFFFFFF;
1352 req->PersistentFileId = 0xFFFFFFFF;
1353 req->VolatileFileId = 0xFFFFFFFF;
1354 }
1355 }
1356 if (remaining_bytes > io_parms->length)
1357 req->RemainingBytes = cpu_to_le32(remaining_bytes);
1358 else
1359 req->RemainingBytes = 0;
1360
1361 iov[0].iov_base = (char *)req;
1362 /* 4 for rfc1002 length field */
1363 iov[0].iov_len = get_rfc1002_length(req) + 4;
1364 return rc;
1365}
1366
1367static void
1368smb2_readv_callback(struct mid_q_entry *mid)
1369{
1370 struct cifs_readdata *rdata = mid->callback_data;
1371 struct cifs_tcon *tcon = tlink_tcon(rdata->cfile->tlink);
1372 struct TCP_Server_Info *server = tcon->ses->server;
Jeff Layton58195752012-09-19 06:22:34 -07001373 struct smb2_hdr *buf = (struct smb2_hdr *)rdata->iov.iov_base;
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001374 unsigned int credits_received = 1;
Jeff Layton58195752012-09-19 06:22:34 -07001375 struct smb_rqst rqst = { .rq_iov = &rdata->iov,
Jeff Layton8321fec2012-09-19 06:22:32 -07001376 .rq_nvec = 1,
1377 .rq_pages = rdata->pages,
1378 .rq_npages = rdata->nr_pages,
1379 .rq_pagesz = rdata->pagesz,
1380 .rq_tailsz = rdata->tailsz };
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001381
1382 cFYI(1, "%s: mid=%llu state=%d result=%d bytes=%u", __func__,
1383 mid->mid, mid->mid_state, rdata->result, rdata->bytes);
1384
1385 switch (mid->mid_state) {
1386 case MID_RESPONSE_RECEIVED:
1387 credits_received = le16_to_cpu(buf->CreditRequest);
1388 /* result already set, check signature */
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -07001389 if (server->sec_mode &
1390 (SECMODE_SIGN_REQUIRED | SECMODE_SIGN_ENABLED)) {
1391 int rc;
1392
Jeff Layton0b688cf2012-09-18 16:20:34 -07001393 rc = smb2_verify_signature(&rqst, server);
Pavel Shilovsky3c1bf7e2012-09-18 16:20:30 -07001394 if (rc)
1395 cERROR(1, "SMB signature verification returned "
1396 "error = %d", rc);
1397 }
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001398 /* FIXME: should this be counted toward the initiating task? */
1399 task_io_account_read(rdata->bytes);
1400 cifs_stats_bytes_read(tcon, rdata->bytes);
1401 break;
1402 case MID_REQUEST_SUBMITTED:
1403 case MID_RETRY_NEEDED:
1404 rdata->result = -EAGAIN;
1405 break;
1406 default:
1407 if (rdata->result != -ENODATA)
1408 rdata->result = -EIO;
1409 }
1410
1411 if (rdata->result)
1412 cifs_stats_fail_inc(tcon, SMB2_READ_HE);
1413
1414 queue_work(cifsiod_wq, &rdata->work);
1415 DeleteMidQEntry(mid);
1416 add_credits(server, credits_received, 0);
1417}
1418
1419/* smb2_async_readv - send an async write, and set up mid to handle result */
1420int
1421smb2_async_readv(struct cifs_readdata *rdata)
1422{
1423 int rc;
1424 struct smb2_hdr *buf;
1425 struct cifs_io_parms io_parms;
Jeff Layton58195752012-09-19 06:22:34 -07001426 struct smb_rqst rqst = { .rq_iov = &rdata->iov,
Jeff Laytonfec344e2012-09-18 16:20:35 -07001427 .rq_nvec = 1 };
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001428
1429 cFYI(1, "%s: offset=%llu bytes=%u", __func__,
1430 rdata->offset, rdata->bytes);
1431
1432 io_parms.tcon = tlink_tcon(rdata->cfile->tlink);
1433 io_parms.offset = rdata->offset;
1434 io_parms.length = rdata->bytes;
1435 io_parms.persistent_fid = rdata->cfile->fid.persistent_fid;
1436 io_parms.volatile_fid = rdata->cfile->fid.volatile_fid;
1437 io_parms.pid = rdata->pid;
Jeff Layton58195752012-09-19 06:22:34 -07001438 rc = smb2_new_read_req(&rdata->iov, &io_parms, 0, 0);
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001439 if (rc)
1440 return rc;
1441
Jeff Layton58195752012-09-19 06:22:34 -07001442 buf = (struct smb2_hdr *)rdata->iov.iov_base;
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001443 /* 4 for rfc1002 length field */
Jeff Layton58195752012-09-19 06:22:34 -07001444 rdata->iov.iov_len = get_rfc1002_length(rdata->iov.iov_base) + 4;
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001445
1446 kref_get(&rdata->refcount);
Jeff Laytonfec344e2012-09-18 16:20:35 -07001447 rc = cifs_call_async(io_parms.tcon->ses->server, &rqst,
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001448 cifs_readv_receive, smb2_readv_callback,
1449 rdata, 0);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001450 if (rc) {
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001451 kref_put(&rdata->refcount, cifs_readdata_release);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001452 cifs_stats_fail_inc(io_parms.tcon, SMB2_READ_HE);
1453 }
Pavel Shilovsky09a47072012-09-18 16:20:29 -07001454
1455 cifs_small_buf_release(buf);
1456 return rc;
1457}
Pavel Shilovsky33319142012-09-18 16:20:29 -07001458
Pavel Shilovskyd8e05032012-09-18 16:20:30 -07001459int
1460SMB2_read(const unsigned int xid, struct cifs_io_parms *io_parms,
1461 unsigned int *nbytes, char **buf, int *buf_type)
1462{
1463 int resp_buftype, rc = -EACCES;
1464 struct smb2_read_rsp *rsp = NULL;
1465 struct kvec iov[1];
1466
1467 *nbytes = 0;
1468 rc = smb2_new_read_req(iov, io_parms, 0, 0);
1469 if (rc)
1470 return rc;
1471
1472 rc = SendReceive2(xid, io_parms->tcon->ses, iov, 1,
1473 &resp_buftype, CIFS_LOG_ERROR);
1474
1475 rsp = (struct smb2_read_rsp *)iov[0].iov_base;
1476
1477 if (rsp->hdr.Status == STATUS_END_OF_FILE) {
1478 free_rsp_buf(resp_buftype, iov[0].iov_base);
1479 return 0;
1480 }
1481
1482 if (rc) {
1483 cifs_stats_fail_inc(io_parms->tcon, SMB2_READ_HE);
1484 cERROR(1, "Send error in read = %d", rc);
1485 } else {
1486 *nbytes = le32_to_cpu(rsp->DataLength);
1487 if ((*nbytes > CIFS_MAX_MSGSIZE) ||
1488 (*nbytes > io_parms->length)) {
1489 cFYI(1, "bad length %d for count %d", *nbytes,
1490 io_parms->length);
1491 rc = -EIO;
1492 *nbytes = 0;
1493 }
1494 }
1495
1496 if (*buf) {
1497 memcpy(*buf, (char *)rsp->hdr.ProtocolId + rsp->DataOffset,
1498 *nbytes);
1499 free_rsp_buf(resp_buftype, iov[0].iov_base);
1500 } else if (resp_buftype != CIFS_NO_BUFFER) {
1501 *buf = iov[0].iov_base;
1502 if (resp_buftype == CIFS_SMALL_BUFFER)
1503 *buf_type = CIFS_SMALL_BUFFER;
1504 else if (resp_buftype == CIFS_LARGE_BUFFER)
1505 *buf_type = CIFS_LARGE_BUFFER;
1506 }
1507 return rc;
1508}
1509
Pavel Shilovsky33319142012-09-18 16:20:29 -07001510/*
1511 * Check the mid_state and signature on received buffer (if any), and queue the
1512 * workqueue completion task.
1513 */
1514static void
1515smb2_writev_callback(struct mid_q_entry *mid)
1516{
1517 struct cifs_writedata *wdata = mid->callback_data;
1518 struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink);
1519 unsigned int written;
1520 struct smb2_write_rsp *rsp = (struct smb2_write_rsp *)mid->resp_buf;
1521 unsigned int credits_received = 1;
1522
1523 switch (mid->mid_state) {
1524 case MID_RESPONSE_RECEIVED:
1525 credits_received = le16_to_cpu(rsp->hdr.CreditRequest);
1526 wdata->result = smb2_check_receive(mid, tcon->ses->server, 0);
1527 if (wdata->result != 0)
1528 break;
1529
1530 written = le32_to_cpu(rsp->DataLength);
1531 /*
1532 * Mask off high 16 bits when bytes written as returned
1533 * by the server is greater than bytes requested by the
1534 * client. OS/2 servers are known to set incorrect
1535 * CountHigh values.
1536 */
1537 if (written > wdata->bytes)
1538 written &= 0xFFFF;
1539
1540 if (written < wdata->bytes)
1541 wdata->result = -ENOSPC;
1542 else
1543 wdata->bytes = written;
1544 break;
1545 case MID_REQUEST_SUBMITTED:
1546 case MID_RETRY_NEEDED:
1547 wdata->result = -EAGAIN;
1548 break;
1549 default:
1550 wdata->result = -EIO;
1551 break;
1552 }
1553
1554 if (wdata->result)
1555 cifs_stats_fail_inc(tcon, SMB2_WRITE_HE);
1556
1557 queue_work(cifsiod_wq, &wdata->work);
1558 DeleteMidQEntry(mid);
1559 add_credits(tcon->ses->server, credits_received, 0);
1560}
1561
1562/* smb2_async_writev - send an async write, and set up mid to handle result */
1563int
1564smb2_async_writev(struct cifs_writedata *wdata)
1565{
Jeff Laytoneddb0792012-09-18 16:20:35 -07001566 int rc = -EACCES;
Pavel Shilovsky33319142012-09-18 16:20:29 -07001567 struct smb2_write_req *req = NULL;
1568 struct cifs_tcon *tcon = tlink_tcon(wdata->cfile->tlink);
Jeff Laytoneddb0792012-09-18 16:20:35 -07001569 struct kvec iov;
Jeff Laytonfec344e2012-09-18 16:20:35 -07001570 struct smb_rqst rqst;
Pavel Shilovsky33319142012-09-18 16:20:29 -07001571
1572 rc = small_smb2_init(SMB2_WRITE, tcon, (void **) &req);
1573 if (rc)
1574 goto async_writev_out;
1575
Pavel Shilovsky33319142012-09-18 16:20:29 -07001576 req->hdr.ProcessId = cpu_to_le32(wdata->cfile->pid);
1577
1578 req->PersistentFileId = wdata->cfile->fid.persistent_fid;
1579 req->VolatileFileId = wdata->cfile->fid.volatile_fid;
1580 req->WriteChannelInfoOffset = 0;
1581 req->WriteChannelInfoLength = 0;
1582 req->Channel = 0;
1583 req->Offset = cpu_to_le64(wdata->offset);
1584 /* 4 for rfc1002 length field */
1585 req->DataOffset = cpu_to_le16(
1586 offsetof(struct smb2_write_req, Buffer) - 4);
1587 req->RemainingBytes = 0;
1588
1589 /* 4 for rfc1002 length field and 1 for Buffer */
Jeff Laytoneddb0792012-09-18 16:20:35 -07001590 iov.iov_len = get_rfc1002_length(req) + 4 - 1;
1591 iov.iov_base = req;
Pavel Shilovsky33319142012-09-18 16:20:29 -07001592
Jeff Laytoneddb0792012-09-18 16:20:35 -07001593 rqst.rq_iov = &iov;
1594 rqst.rq_nvec = 1;
1595 rqst.rq_pages = wdata->pages;
1596 rqst.rq_npages = wdata->nr_pages;
1597 rqst.rq_pagesz = wdata->pagesz;
1598 rqst.rq_tailsz = wdata->tailsz;
Pavel Shilovsky33319142012-09-18 16:20:29 -07001599
1600 cFYI(1, "async write at %llu %u bytes", wdata->offset, wdata->bytes);
1601
1602 req->Length = cpu_to_le32(wdata->bytes);
1603
1604 inc_rfc1001_len(&req->hdr, wdata->bytes - 1 /* Buffer */);
1605
1606 kref_get(&wdata->refcount);
Jeff Laytonfec344e2012-09-18 16:20:35 -07001607 rc = cifs_call_async(tcon->ses->server, &rqst, NULL,
1608 smb2_writev_callback, wdata, 0);
Pavel Shilovsky33319142012-09-18 16:20:29 -07001609
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001610 if (rc) {
Pavel Shilovsky33319142012-09-18 16:20:29 -07001611 kref_put(&wdata->refcount, cifs_writedata_release);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001612 cifs_stats_fail_inc(tcon, SMB2_WRITE_HE);
1613 }
Pavel Shilovsky33319142012-09-18 16:20:29 -07001614
Pavel Shilovsky33319142012-09-18 16:20:29 -07001615async_writev_out:
1616 cifs_small_buf_release(req);
Pavel Shilovsky33319142012-09-18 16:20:29 -07001617 return rc;
1618}
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001619
1620/*
1621 * SMB2_write function gets iov pointer to kvec array with n_vec as a length.
1622 * The length field from io_parms must be at least 1 and indicates a number of
1623 * elements with data to write that begins with position 1 in iov array. All
1624 * data length is specified by count.
1625 */
1626int
1627SMB2_write(const unsigned int xid, struct cifs_io_parms *io_parms,
1628 unsigned int *nbytes, struct kvec *iov, int n_vec)
1629{
1630 int rc = 0;
1631 struct smb2_write_req *req = NULL;
1632 struct smb2_write_rsp *rsp = NULL;
1633 int resp_buftype;
1634 *nbytes = 0;
1635
1636 if (n_vec < 1)
1637 return rc;
1638
1639 rc = small_smb2_init(SMB2_WRITE, io_parms->tcon, (void **) &req);
1640 if (rc)
1641 return rc;
1642
1643 if (io_parms->tcon->ses->server == NULL)
1644 return -ECONNABORTED;
1645
1646 req->hdr.ProcessId = cpu_to_le32(io_parms->pid);
1647
1648 req->PersistentFileId = io_parms->persistent_fid;
1649 req->VolatileFileId = io_parms->volatile_fid;
1650 req->WriteChannelInfoOffset = 0;
1651 req->WriteChannelInfoLength = 0;
1652 req->Channel = 0;
1653 req->Length = cpu_to_le32(io_parms->length);
1654 req->Offset = cpu_to_le64(io_parms->offset);
1655 /* 4 for rfc1002 length field */
1656 req->DataOffset = cpu_to_le16(
1657 offsetof(struct smb2_write_req, Buffer) - 4);
1658 req->RemainingBytes = 0;
1659
1660 iov[0].iov_base = (char *)req;
1661 /* 4 for rfc1002 length field and 1 for Buffer */
1662 iov[0].iov_len = get_rfc1002_length(req) + 4 - 1;
1663
1664 /* length of entire message including data to be written */
1665 inc_rfc1001_len(req, io_parms->length - 1 /* Buffer */);
1666
1667 rc = SendReceive2(xid, io_parms->tcon->ses, iov, n_vec + 1,
1668 &resp_buftype, 0);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001669 rsp = (struct smb2_write_rsp *)iov[0].iov_base;
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001670
1671 if (rc) {
1672 cifs_stats_fail_inc(io_parms->tcon, SMB2_WRITE_HE);
1673 cERROR(1, "Send error in write = %d", rc);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001674 } else
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001675 *nbytes = le32_to_cpu(rsp->DataLength);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001676
1677 free_rsp_buf(resp_buftype, rsp);
Pavel Shilovsky009d3442012-09-18 16:20:30 -07001678 return rc;
1679}
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001680
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07001681static unsigned int
1682num_entries(char *bufstart, char *end_of_buf, char **lastentry, size_t size)
1683{
1684 int len;
1685 unsigned int entrycount = 0;
1686 unsigned int next_offset = 0;
1687 FILE_DIRECTORY_INFO *entryptr;
1688
1689 if (bufstart == NULL)
1690 return 0;
1691
1692 entryptr = (FILE_DIRECTORY_INFO *)bufstart;
1693
1694 while (1) {
1695 entryptr = (FILE_DIRECTORY_INFO *)
1696 ((char *)entryptr + next_offset);
1697
1698 if ((char *)entryptr + size > end_of_buf) {
1699 cERROR(1, "malformed search entry would overflow");
1700 break;
1701 }
1702
1703 len = le32_to_cpu(entryptr->FileNameLength);
1704 if ((char *)entryptr + len + size > end_of_buf) {
1705 cERROR(1, "directory entry name would overflow frame "
1706 "end of buf %p", end_of_buf);
1707 break;
1708 }
1709
1710 *lastentry = (char *)entryptr;
1711 entrycount++;
1712
1713 next_offset = le32_to_cpu(entryptr->NextEntryOffset);
1714 if (!next_offset)
1715 break;
1716 }
1717
1718 return entrycount;
1719}
1720
1721/*
1722 * Readdir/FindFirst
1723 */
1724int
1725SMB2_query_directory(const unsigned int xid, struct cifs_tcon *tcon,
1726 u64 persistent_fid, u64 volatile_fid, int index,
1727 struct cifs_search_info *srch_inf)
1728{
1729 struct smb2_query_directory_req *req;
1730 struct smb2_query_directory_rsp *rsp = NULL;
1731 struct kvec iov[2];
1732 int rc = 0;
1733 int len;
1734 int resp_buftype;
1735 unsigned char *bufptr;
1736 struct TCP_Server_Info *server;
1737 struct cifs_ses *ses = tcon->ses;
1738 __le16 asteriks = cpu_to_le16('*');
1739 char *end_of_smb;
1740 unsigned int output_size = CIFSMaxBufSize;
1741 size_t info_buf_size;
1742
1743 if (ses && (ses->server))
1744 server = ses->server;
1745 else
1746 return -EIO;
1747
1748 rc = small_smb2_init(SMB2_QUERY_DIRECTORY, tcon, (void **) &req);
1749 if (rc)
1750 return rc;
1751
1752 switch (srch_inf->info_level) {
1753 case SMB_FIND_FILE_DIRECTORY_INFO:
1754 req->FileInformationClass = FILE_DIRECTORY_INFORMATION;
1755 info_buf_size = sizeof(FILE_DIRECTORY_INFO) - 1;
1756 break;
1757 case SMB_FIND_FILE_ID_FULL_DIR_INFO:
1758 req->FileInformationClass = FILEID_FULL_DIRECTORY_INFORMATION;
1759 info_buf_size = sizeof(SEARCH_ID_FULL_DIR_INFO) - 1;
1760 break;
1761 default:
1762 cERROR(1, "info level %u isn't supported",
1763 srch_inf->info_level);
1764 rc = -EINVAL;
1765 goto qdir_exit;
1766 }
1767
1768 req->FileIndex = cpu_to_le32(index);
1769 req->PersistentFileId = persistent_fid;
1770 req->VolatileFileId = volatile_fid;
1771
1772 len = 0x2;
1773 bufptr = req->Buffer;
1774 memcpy(bufptr, &asteriks, len);
1775
1776 req->FileNameOffset =
1777 cpu_to_le16(sizeof(struct smb2_query_directory_req) - 1 - 4);
1778 req->FileNameLength = cpu_to_le16(len);
1779 /*
1780 * BB could be 30 bytes or so longer if we used SMB2 specific
1781 * buffer lengths, but this is safe and close enough.
1782 */
1783 output_size = min_t(unsigned int, output_size, server->maxBuf);
1784 output_size = min_t(unsigned int, output_size, 2 << 15);
1785 req->OutputBufferLength = cpu_to_le32(output_size);
1786
1787 iov[0].iov_base = (char *)req;
1788 /* 4 for RFC1001 length and 1 for Buffer */
1789 iov[0].iov_len = get_rfc1002_length(req) + 4 - 1;
1790
1791 iov[1].iov_base = (char *)(req->Buffer);
1792 iov[1].iov_len = len;
1793
1794 inc_rfc1001_len(req, len - 1 /* Buffer */);
1795
1796 rc = SendReceive2(xid, ses, iov, 2, &resp_buftype, 0);
Pavel Shilovskye5d04882012-09-19 16:03:26 +04001797 rsp = (struct smb2_query_directory_rsp *)iov[0].iov_base;
1798
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07001799 if (rc) {
1800 cifs_stats_fail_inc(tcon, SMB2_QUERY_DIRECTORY_HE);
1801 goto qdir_exit;
1802 }
Pavel Shilovskyd324f08d2012-09-18 16:20:33 -07001803
1804 rc = validate_buf(le16_to_cpu(rsp->OutputBufferOffset),
1805 le32_to_cpu(rsp->OutputBufferLength), &rsp->hdr,
1806 info_buf_size);
1807 if (rc)
1808 goto qdir_exit;
1809
1810 srch_inf->unicode = true;
1811
1812 if (srch_inf->ntwrk_buf_start) {
1813 if (srch_inf->smallBuf)
1814 cifs_small_buf_release(srch_inf->ntwrk_buf_start);
1815 else
1816 cifs_buf_release(srch_inf->ntwrk_buf_start);
1817 }
1818 srch_inf->ntwrk_buf_start = (char *)rsp;
1819 srch_inf->srch_entries_start = srch_inf->last_entry = 4 /* rfclen */ +
1820 (char *)&rsp->hdr + le16_to_cpu(rsp->OutputBufferOffset);
1821 /* 4 for rfc1002 length field */
1822 end_of_smb = get_rfc1002_length(rsp) + 4 + (char *)&rsp->hdr;
1823 srch_inf->entries_in_buffer =
1824 num_entries(srch_inf->srch_entries_start, end_of_smb,
1825 &srch_inf->last_entry, info_buf_size);
1826 srch_inf->index_of_last_entry += srch_inf->entries_in_buffer;
1827 cFYI(1, "num entries %d last_index %lld srch start %p srch end %p",
1828 srch_inf->entries_in_buffer, srch_inf->index_of_last_entry,
1829 srch_inf->srch_entries_start, srch_inf->last_entry);
1830 if (resp_buftype == CIFS_LARGE_BUFFER)
1831 srch_inf->smallBuf = false;
1832 else if (resp_buftype == CIFS_SMALL_BUFFER)
1833 srch_inf->smallBuf = true;
1834 else
1835 cERROR(1, "illegal search buffer type");
1836
1837 if (rsp->hdr.Status == STATUS_NO_MORE_FILES)
1838 srch_inf->endOfSearch = 1;
1839 else
1840 srch_inf->endOfSearch = 0;
1841
1842 return rc;
1843
1844qdir_exit:
1845 free_rsp_buf(resp_buftype, rsp);
1846 return rc;
1847}
1848
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001849static int
1850send_set_info(const unsigned int xid, struct cifs_tcon *tcon,
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07001851 u64 persistent_fid, u64 volatile_fid, u32 pid, int info_class,
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001852 unsigned int num, void **data, unsigned int *size)
1853{
1854 struct smb2_set_info_req *req;
1855 struct smb2_set_info_rsp *rsp = NULL;
1856 struct kvec *iov;
1857 int rc = 0;
1858 int resp_buftype;
1859 unsigned int i;
1860 struct TCP_Server_Info *server;
1861 struct cifs_ses *ses = tcon->ses;
1862
1863 if (ses && (ses->server))
1864 server = ses->server;
1865 else
1866 return -EIO;
1867
1868 if (!num)
1869 return -EINVAL;
1870
1871 iov = kmalloc(sizeof(struct kvec) * num, GFP_KERNEL);
1872 if (!iov)
1873 return -ENOMEM;
1874
1875 rc = small_smb2_init(SMB2_SET_INFO, tcon, (void **) &req);
1876 if (rc) {
1877 kfree(iov);
1878 return rc;
1879 }
1880
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07001881 req->hdr.ProcessId = cpu_to_le32(pid);
1882
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001883 req->InfoType = SMB2_O_INFO_FILE;
1884 req->FileInfoClass = info_class;
1885 req->PersistentFileId = persistent_fid;
1886 req->VolatileFileId = volatile_fid;
1887
1888 /* 4 for RFC1001 length and 1 for Buffer */
1889 req->BufferOffset =
1890 cpu_to_le16(sizeof(struct smb2_set_info_req) - 1 - 4);
1891 req->BufferLength = cpu_to_le32(*size);
1892
1893 inc_rfc1001_len(req, *size - 1 /* Buffer */);
1894
1895 memcpy(req->Buffer, *data, *size);
1896
1897 iov[0].iov_base = (char *)req;
1898 /* 4 for RFC1001 length */
1899 iov[0].iov_len = get_rfc1002_length(req) + 4;
1900
1901 for (i = 1; i < num; i++) {
1902 inc_rfc1001_len(req, size[i]);
1903 le32_add_cpu(&req->BufferLength, size[i]);
1904 iov[i].iov_base = (char *)data[i];
1905 iov[i].iov_len = size[i];
1906 }
1907
1908 rc = SendReceive2(xid, ses, iov, num, &resp_buftype, 0);
1909 rsp = (struct smb2_set_info_rsp *)iov[0].iov_base;
1910
1911 if (rc != 0) {
1912 cifs_stats_fail_inc(tcon, SMB2_SET_INFO_HE);
1913 goto out;
1914 }
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001915out:
1916 free_rsp_buf(resp_buftype, rsp);
1917 kfree(iov);
1918 return rc;
1919}
1920
1921int
1922SMB2_rename(const unsigned int xid, struct cifs_tcon *tcon,
1923 u64 persistent_fid, u64 volatile_fid, __le16 *target_file)
1924{
1925 struct smb2_file_rename_info info;
1926 void **data;
1927 unsigned int size[2];
1928 int rc;
1929 int len = (2 * UniStrnlen((wchar_t *)target_file, PATH_MAX));
1930
1931 data = kmalloc(sizeof(void *) * 2, GFP_KERNEL);
1932 if (!data)
1933 return -ENOMEM;
1934
1935 info.ReplaceIfExists = 1; /* 1 = replace existing target with new */
1936 /* 0 = fail if target already exists */
1937 info.RootDirectory = 0; /* MBZ for network ops (why does spec say?) */
1938 info.FileNameLength = cpu_to_le32(len);
1939
1940 data[0] = &info;
1941 size[0] = sizeof(struct smb2_file_rename_info);
1942
1943 data[1] = target_file;
1944 size[1] = len + 2 /* null */;
1945
1946 rc = send_set_info(xid, tcon, persistent_fid, volatile_fid,
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07001947 current->tgid, FILE_RENAME_INFORMATION, 2, data,
1948 size);
Pavel Shilovsky35143eb2012-09-18 16:20:31 -07001949 kfree(data);
1950 return rc;
1951}
Pavel Shilovsky568798c2012-09-18 16:20:31 -07001952
1953int
1954SMB2_set_hardlink(const unsigned int xid, struct cifs_tcon *tcon,
1955 u64 persistent_fid, u64 volatile_fid, __le16 *target_file)
1956{
1957 struct smb2_file_link_info info;
1958 void **data;
1959 unsigned int size[2];
1960 int rc;
1961 int len = (2 * UniStrnlen((wchar_t *)target_file, PATH_MAX));
1962
1963 data = kmalloc(sizeof(void *) * 2, GFP_KERNEL);
1964 if (!data)
1965 return -ENOMEM;
1966
1967 info.ReplaceIfExists = 0; /* 1 = replace existing link with new */
1968 /* 0 = fail if link already exists */
1969 info.RootDirectory = 0; /* MBZ for network ops (why does spec say?) */
1970 info.FileNameLength = cpu_to_le32(len);
1971
1972 data[0] = &info;
1973 size[0] = sizeof(struct smb2_file_link_info);
1974
1975 data[1] = target_file;
1976 size[1] = len + 2 /* null */;
1977
1978 rc = send_set_info(xid, tcon, persistent_fid, volatile_fid,
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07001979 current->tgid, FILE_LINK_INFORMATION, 2, data, size);
Pavel Shilovsky568798c2012-09-18 16:20:31 -07001980 kfree(data);
1981 return rc;
1982}
Pavel Shilovskyc839ff22012-09-18 16:20:32 -07001983
1984int
1985SMB2_set_eof(const unsigned int xid, struct cifs_tcon *tcon, u64 persistent_fid,
1986 u64 volatile_fid, u32 pid, __le64 *eof)
1987{
1988 struct smb2_file_eof_info info;
1989 void *data;
1990 unsigned int size;
1991
1992 info.EndOfFile = *eof;
1993
1994 data = &info;
1995 size = sizeof(struct smb2_file_eof_info);
1996
1997 return send_set_info(xid, tcon, persistent_fid, volatile_fid, pid,
1998 FILE_END_OF_FILE_INFORMATION, 1, &data, &size);
1999}
Pavel Shilovsky1feeaac2012-09-18 16:20:32 -07002000
2001int
2002SMB2_set_info(const unsigned int xid, struct cifs_tcon *tcon,
2003 u64 persistent_fid, u64 volatile_fid, FILE_BASIC_INFO *buf)
2004{
2005 unsigned int size;
2006 size = sizeof(FILE_BASIC_INFO);
2007 return send_set_info(xid, tcon, persistent_fid, volatile_fid,
2008 current->tgid, FILE_BASIC_INFORMATION, 1,
2009 (void **)&buf, &size);
2010}
Pavel Shilovsky983c88a2012-09-18 16:20:33 -07002011
2012int
2013SMB2_oplock_break(const unsigned int xid, struct cifs_tcon *tcon,
2014 const u64 persistent_fid, const u64 volatile_fid,
2015 __u8 oplock_level)
2016{
2017 int rc;
2018 struct smb2_oplock_break *req = NULL;
2019
2020 cFYI(1, "SMB2_oplock_break");
2021 rc = small_smb2_init(SMB2_OPLOCK_BREAK, tcon, (void **) &req);
2022
2023 if (rc)
2024 return rc;
2025
2026 req->VolatileFid = volatile_fid;
2027 req->PersistentFid = persistent_fid;
2028 req->OplockLevel = oplock_level;
2029 req->hdr.CreditRequest = cpu_to_le16(1);
2030
2031 rc = SendReceiveNoRsp(xid, tcon->ses, (char *) req, CIFS_OBREAK_OP);
2032 /* SMB2 buffer freed by function above */
2033
2034 if (rc) {
2035 cifs_stats_fail_inc(tcon, SMB2_OPLOCK_BREAK_HE);
2036 cFYI(1, "Send error in Oplock Break = %d", rc);
2037 }
2038
2039 return rc;
2040}
Pavel Shilovsky6fc05c22012-09-18 16:20:34 -07002041
2042static void
2043copy_fs_info_to_kstatfs(struct smb2_fs_full_size_info *pfs_inf,
2044 struct kstatfs *kst)
2045{
2046 kst->f_bsize = le32_to_cpu(pfs_inf->BytesPerSector) *
2047 le32_to_cpu(pfs_inf->SectorsPerAllocationUnit);
2048 kst->f_blocks = le64_to_cpu(pfs_inf->TotalAllocationUnits);
2049 kst->f_bfree = le64_to_cpu(pfs_inf->ActualAvailableAllocationUnits);
2050 kst->f_bavail = le64_to_cpu(pfs_inf->CallerAvailableAllocationUnits);
2051 return;
2052}
2053
2054static int
2055build_qfs_info_req(struct kvec *iov, struct cifs_tcon *tcon, int level,
2056 int outbuf_len, u64 persistent_fid, u64 volatile_fid)
2057{
2058 int rc;
2059 struct smb2_query_info_req *req;
2060
2061 cFYI(1, "Query FSInfo level %d", level);
2062
2063 if ((tcon->ses == NULL) || (tcon->ses->server == NULL))
2064 return -EIO;
2065
2066 rc = small_smb2_init(SMB2_QUERY_INFO, tcon, (void **) &req);
2067 if (rc)
2068 return rc;
2069
2070 req->InfoType = SMB2_O_INFO_FILESYSTEM;
2071 req->FileInfoClass = level;
2072 req->PersistentFileId = persistent_fid;
2073 req->VolatileFileId = volatile_fid;
2074 /* 4 for rfc1002 length field and 1 for pad */
2075 req->InputBufferOffset =
2076 cpu_to_le16(sizeof(struct smb2_query_info_req) - 1 - 4);
2077 req->OutputBufferLength = cpu_to_le32(
2078 outbuf_len + sizeof(struct smb2_query_info_rsp) - 1 - 4);
2079
2080 iov->iov_base = (char *)req;
2081 /* 4 for rfc1002 length field */
2082 iov->iov_len = get_rfc1002_length(req) + 4;
2083 return 0;
2084}
2085
2086int
2087SMB2_QFS_info(const unsigned int xid, struct cifs_tcon *tcon,
2088 u64 persistent_fid, u64 volatile_fid, struct kstatfs *fsdata)
2089{
2090 struct smb2_query_info_rsp *rsp = NULL;
2091 struct kvec iov;
2092 int rc = 0;
2093 int resp_buftype;
2094 struct cifs_ses *ses = tcon->ses;
2095 struct smb2_fs_full_size_info *info = NULL;
2096
2097 rc = build_qfs_info_req(&iov, tcon, FS_FULL_SIZE_INFORMATION,
2098 sizeof(struct smb2_fs_full_size_info),
2099 persistent_fid, volatile_fid);
2100 if (rc)
2101 return rc;
2102
2103 rc = SendReceive2(xid, ses, &iov, 1, &resp_buftype, 0);
2104 if (rc) {
2105 cifs_stats_fail_inc(tcon, SMB2_QUERY_INFO_HE);
2106 goto qinf_exit;
2107 }
2108 rsp = (struct smb2_query_info_rsp *)iov.iov_base;
2109
2110 info = (struct smb2_fs_full_size_info *)(4 /* RFC1001 len */ +
2111 le16_to_cpu(rsp->OutputBufferOffset) + (char *)&rsp->hdr);
2112 rc = validate_buf(le16_to_cpu(rsp->OutputBufferOffset),
2113 le32_to_cpu(rsp->OutputBufferLength), &rsp->hdr,
2114 sizeof(struct smb2_fs_full_size_info));
2115 if (!rc)
2116 copy_fs_info_to_kstatfs(info, fsdata);
2117
2118qinf_exit:
2119 free_rsp_buf(resp_buftype, iov.iov_base);
2120 return rc;
2121}
Pavel Shilovskyf7ba7fe2012-09-19 06:22:43 -07002122
2123int
2124smb2_lockv(const unsigned int xid, struct cifs_tcon *tcon,
2125 const __u64 persist_fid, const __u64 volatile_fid, const __u32 pid,
2126 const __u32 num_lock, struct smb2_lock_element *buf)
2127{
2128 int rc = 0;
2129 struct smb2_lock_req *req = NULL;
2130 struct kvec iov[2];
2131 int resp_buf_type;
2132 unsigned int count;
2133
2134 cFYI(1, "smb2_lockv num lock %d", num_lock);
2135
2136 rc = small_smb2_init(SMB2_LOCK, tcon, (void **) &req);
2137 if (rc)
2138 return rc;
2139
2140 req->hdr.ProcessId = cpu_to_le32(pid);
2141 req->LockCount = cpu_to_le16(num_lock);
2142
2143 req->PersistentFileId = persist_fid;
2144 req->VolatileFileId = volatile_fid;
2145
2146 count = num_lock * sizeof(struct smb2_lock_element);
2147 inc_rfc1001_len(req, count - sizeof(struct smb2_lock_element));
2148
2149 iov[0].iov_base = (char *)req;
2150 /* 4 for rfc1002 length field and count for all locks */
2151 iov[0].iov_len = get_rfc1002_length(req) + 4 - count;
2152 iov[1].iov_base = (char *)buf;
2153 iov[1].iov_len = count;
2154
2155 cifs_stats_inc(&tcon->stats.cifs_stats.num_locks);
2156 rc = SendReceive2(xid, tcon->ses, iov, 2, &resp_buf_type, CIFS_NO_RESP);
2157 if (rc) {
2158 cFYI(1, "Send error in smb2_lockv = %d", rc);
2159 cifs_stats_fail_inc(tcon, SMB2_LOCK_HE);
2160 }
2161
2162 return rc;
2163}
2164
2165int
2166SMB2_lock(const unsigned int xid, struct cifs_tcon *tcon,
2167 const __u64 persist_fid, const __u64 volatile_fid, const __u32 pid,
2168 const __u64 length, const __u64 offset, const __u32 lock_flags,
2169 const bool wait)
2170{
2171 struct smb2_lock_element lock;
2172
2173 lock.Offset = cpu_to_le64(offset);
2174 lock.Length = cpu_to_le64(length);
2175 lock.Flags = cpu_to_le32(lock_flags);
2176 if (!wait && lock_flags != SMB2_LOCKFLAG_UNLOCK)
2177 lock.Flags |= cpu_to_le32(SMB2_LOCKFLAG_FAIL_IMMEDIATELY);
2178
2179 return smb2_lockv(xid, tcon, persist_fid, volatile_fid, pid, 1, &lock);
2180}
Pavel Shilovsky0822f512012-09-19 06:22:45 -07002181
2182int
2183SMB2_lease_break(const unsigned int xid, struct cifs_tcon *tcon,
2184 __u8 *lease_key, const __le32 lease_state)
2185{
2186 int rc;
2187 struct smb2_lease_ack *req = NULL;
2188
2189 cFYI(1, "SMB2_lease_break");
2190 rc = small_smb2_init(SMB2_OPLOCK_BREAK, tcon, (void **) &req);
2191
2192 if (rc)
2193 return rc;
2194
2195 req->hdr.CreditRequest = cpu_to_le16(1);
2196 req->StructureSize = cpu_to_le16(36);
2197 inc_rfc1001_len(req, 12);
2198
2199 memcpy(req->LeaseKey, lease_key, 16);
2200 req->LeaseState = lease_state;
2201
2202 rc = SendReceiveNoRsp(xid, tcon->ses, (char *) req, CIFS_OBREAK_OP);
2203 /* SMB2 buffer freed by function above */
2204
2205 if (rc) {
2206 cifs_stats_fail_inc(tcon, SMB2_OPLOCK_BREAK_HE);
2207 cFYI(1, "Send error in Lease Break = %d", rc);
2208 }
2209
2210 return rc;
2211}