2021-06-18 05:31:49 +00:00
|
|
|
// SPDX-License-Identifier: LGPL-2.1
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
*
|
2008-02-07 23:25:02 +00:00
|
|
|
* Copyright (C) International Business Machines Corp., 2002,2008
|
2005-04-16 22:20:36 +00:00
|
|
|
* Author(s): Steve French (sfrench@us.ibm.com)
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/ctype.h>
|
|
|
|
#include <linux/mempool.h>
|
2017-04-25 18:52:29 +00:00
|
|
|
#include <linux/vmalloc.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "cifspdu.h"
|
|
|
|
#include "cifsglob.h"
|
|
|
|
#include "cifsproto.h"
|
|
|
|
#include "cifs_debug.h"
|
|
|
|
#include "smberr.h"
|
|
|
|
#include "nterr.h"
|
2005-04-29 05:41:06 +00:00
|
|
|
#include "cifs_unicode.h"
|
2012-01-12 18:40:50 +00:00
|
|
|
#include "smb2pdu.h"
|
2020-02-20 22:49:34 +00:00
|
|
|
#include "cifsfs.h"
|
2020-05-19 18:38:28 +00:00
|
|
|
#ifdef CONFIG_CIFS_DFS_UPCALL
|
|
|
|
#include "dns_resolve.h"
|
2023-02-28 22:01:54 +00:00
|
|
|
#include "dfs_cache.h"
|
2023-03-14 23:32:54 +00:00
|
|
|
#include "dfs.h"
|
2020-05-19 18:38:28 +00:00
|
|
|
#endif
|
2020-12-12 19:40:50 +00:00
|
|
|
#include "fs_context.h"
|
2022-08-12 00:04:29 +00:00
|
|
|
#include "cached_dir.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
extern mempool_t *cifs_sm_req_poolp;
|
|
|
|
extern mempool_t *cifs_req_poolp;
|
|
|
|
|
2007-07-10 01:16:18 +00:00
|
|
|
/* The xid serves as a useful identifier for each incoming vfs request,
|
|
|
|
in a similar way to the mid which is useful to track each sent smb,
|
|
|
|
and CurrentXid can also provide a running counter (although it
|
|
|
|
will eventually wrap past zero) of the total vfs operations handled
|
2005-04-16 22:20:36 +00:00
|
|
|
since the cifs fs was mounted */
|
|
|
|
|
|
|
|
unsigned int
|
2012-06-20 07:21:16 +00:00
|
|
|
_get_xid(void)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
unsigned int xid;
|
|
|
|
|
|
|
|
spin_lock(&GlobalMid_Lock);
|
|
|
|
GlobalTotalActiveXid++;
|
2007-07-13 00:33:32 +00:00
|
|
|
|
|
|
|
/* keep high water mark for number of simultaneous ops in filesystem */
|
2005-04-16 22:20:36 +00:00
|
|
|
if (GlobalTotalActiveXid > GlobalMaxActiveXid)
|
2007-07-13 00:33:32 +00:00
|
|
|
GlobalMaxActiveXid = GlobalTotalActiveXid;
|
2007-07-07 19:25:05 +00:00
|
|
|
if (GlobalTotalActiveXid > 65000)
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "warning: more than 65000 requests active\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
xid = GlobalCurrentXid++;
|
|
|
|
spin_unlock(&GlobalMid_Lock);
|
|
|
|
return xid;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2012-06-20 07:21:16 +00:00
|
|
|
_free_xid(unsigned int xid)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
spin_lock(&GlobalMid_Lock);
|
2007-07-07 19:25:05 +00:00
|
|
|
/* if (GlobalTotalActiveXid == 0)
|
2005-04-16 22:20:36 +00:00
|
|
|
BUG(); */
|
|
|
|
GlobalTotalActiveXid--;
|
|
|
|
spin_unlock(&GlobalMid_Lock);
|
|
|
|
}
|
|
|
|
|
2011-05-27 04:34:02 +00:00
|
|
|
struct cifs_ses *
|
2005-04-16 22:20:36 +00:00
|
|
|
sesInfoAlloc(void)
|
|
|
|
{
|
2011-05-27 04:34:02 +00:00
|
|
|
struct cifs_ses *ret_buf;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2011-05-27 04:34:02 +00:00
|
|
|
ret_buf = kzalloc(sizeof(struct cifs_ses), GFP_KERNEL);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (ret_buf) {
|
|
|
|
atomic_inc(&sesInfoAllocCount);
|
2022-07-27 19:49:56 +00:00
|
|
|
spin_lock_init(&ret_buf->ses_lock);
|
2022-04-07 13:15:49 +00:00
|
|
|
ret_buf->ses_status = SES_NEW;
|
2008-11-14 18:53:46 +00:00
|
|
|
++ret_buf->ses_count;
|
|
|
|
INIT_LIST_HEAD(&ret_buf->smb_ses_list);
|
2008-11-15 16:12:47 +00:00
|
|
|
INIT_LIST_HEAD(&ret_buf->tcon_list);
|
2010-02-25 05:36:46 +00:00
|
|
|
mutex_init(&ret_buf->session_mutex);
|
2018-06-14 13:43:18 +00:00
|
|
|
spin_lock_init(&ret_buf->iface_lock);
|
2022-01-01 12:50:21 +00:00
|
|
|
INIT_LIST_HEAD(&ret_buf->iface_list);
|
2021-07-19 10:54:46 +00:00
|
|
|
spin_lock_init(&ret_buf->chan_lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
return ret_buf;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2011-05-27 04:34:02 +00:00
|
|
|
sesInfoFree(struct cifs_ses *buf_to_free)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2022-01-01 12:50:21 +00:00
|
|
|
struct cifs_server_iface *iface = NULL, *niface = NULL;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (buf_to_free == NULL) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Null buffer passed to sesInfoFree\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2023-07-24 02:10:57 +00:00
|
|
|
unload_nls(buf_to_free->local_nls);
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_dec(&sesInfoAllocCount);
|
2005-11-07 09:01:34 +00:00
|
|
|
kfree(buf_to_free->serverOS);
|
|
|
|
kfree(buf_to_free->serverDomain);
|
|
|
|
kfree(buf_to_free->serverNOS);
|
2020-08-07 06:18:13 +00:00
|
|
|
kfree_sensitive(buf_to_free->password);
|
2011-02-25 07:11:56 +00:00
|
|
|
kfree(buf_to_free->user_name);
|
2006-05-31 22:40:51 +00:00
|
|
|
kfree(buf_to_free->domainName);
|
2020-08-07 06:18:13 +00:00
|
|
|
kfree_sensitive(buf_to_free->auth_key.response);
|
2022-01-01 12:50:21 +00:00
|
|
|
spin_lock(&buf_to_free->iface_lock);
|
|
|
|
list_for_each_entry_safe(iface, niface, &buf_to_free->iface_list,
|
|
|
|
iface_head)
|
|
|
|
kref_put(&iface->refcount, release_iface);
|
|
|
|
spin_unlock(&buf_to_free->iface_lock);
|
2020-08-07 06:18:13 +00:00
|
|
|
kfree_sensitive(buf_to_free);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2011-05-27 04:34:02 +00:00
|
|
|
struct cifs_tcon *
|
2023-09-19 16:35:53 +00:00
|
|
|
tcon_info_alloc(bool dir_leases_enabled)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2011-05-27 04:34:02 +00:00
|
|
|
struct cifs_tcon *ret_buf;
|
2018-12-21 05:50:48 +00:00
|
|
|
|
|
|
|
ret_buf = kzalloc(sizeof(*ret_buf), GFP_KERNEL);
|
|
|
|
if (!ret_buf)
|
|
|
|
return NULL;
|
2023-09-19 16:35:53 +00:00
|
|
|
|
|
|
|
if (dir_leases_enabled == true) {
|
|
|
|
ret_buf->cfids = init_cached_dirs();
|
|
|
|
if (!ret_buf->cfids) {
|
|
|
|
kfree(ret_buf);
|
|
|
|
return NULL;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2023-09-19 16:35:53 +00:00
|
|
|
/* else ret_buf->cfids is already set to NULL above */
|
2018-12-21 05:50:48 +00:00
|
|
|
|
|
|
|
atomic_inc(&tconInfoAllocCount);
|
smb3: cleanup and clarify status of tree connections
Currently the way the tid (tree connection) status is tracked
is confusing. The same enum is used for structs cifs_tcon
and cifs_ses and TCP_Server_info, but each of these three has
different states that they transition among. The current
code also unnecessarily uses camelCase.
Convert from use of statusEnum to a new tid_status_enum for
tree connections. The valid states for a tid are:
TID_NEW = 0,
TID_GOOD,
TID_EXITING,
TID_NEED_RECON,
TID_NEED_TCON,
TID_IN_TCON,
TID_NEED_FILES_INVALIDATE, /* unused, considering removing in future */
TID_IN_FILES_INVALIDATE
It also removes CifsNeedTcon, CifsInTcon, CifsNeedFilesInvalidate and
CifsInFilesInvalidate from the statusEnum used for session and
TCP_Server_Info since they are not relevant for those.
A follow on patch will fix the places where we use the
tcon->need_reconnect flag to be more consistent with the tid->status.
Also fixes a bug that was:
Reported-by: kernel test robot <lkp@intel.com>
Reviewed-by: Shyam Prasad N <sprasad@microsoft.com>
Reviewed-by: Ronnie Sahlberg <lsahlber@redhat.com>
Signed-off-by: Steve French <stfrench@microsoft.com>
2022-03-27 21:07:30 +00:00
|
|
|
ret_buf->status = TID_NEW;
|
2018-12-21 05:50:48 +00:00
|
|
|
++ret_buf->tc_count;
|
2022-07-27 19:49:56 +00:00
|
|
|
spin_lock_init(&ret_buf->tc_lock);
|
2018-12-21 05:50:48 +00:00
|
|
|
INIT_LIST_HEAD(&ret_buf->openFileList);
|
|
|
|
INIT_LIST_HEAD(&ret_buf->tcon_list);
|
|
|
|
spin_lock_init(&ret_buf->open_file_lock);
|
|
|
|
spin_lock_init(&ret_buf->stat_lock);
|
|
|
|
atomic_set(&ret_buf->num_local_opens, 0);
|
|
|
|
atomic_set(&ret_buf->num_remote_opens, 0);
|
2024-01-17 22:15:18 +00:00
|
|
|
ret_buf->stats_from_time = ktime_get_real_seconds();
|
2023-03-14 23:32:54 +00:00
|
|
|
#ifdef CONFIG_CIFS_DFS_UPCALL
|
|
|
|
INIT_LIST_HEAD(&ret_buf->dfs_ses_list);
|
|
|
|
#endif
|
2018-12-21 05:50:48 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return ret_buf;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2022-08-12 00:04:29 +00:00
|
|
|
tconInfoFree(struct cifs_tcon *tcon)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2022-08-12 00:04:29 +00:00
|
|
|
if (tcon == NULL) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Null buffer passed to tconInfoFree\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
2022-08-31 02:49:42 +00:00
|
|
|
free_cached_dirs(tcon->cfids);
|
2005-04-16 22:20:36 +00:00
|
|
|
atomic_dec(&tconInfoAllocCount);
|
2022-08-12 00:04:29 +00:00
|
|
|
kfree(tcon->nativeFileSystem);
|
|
|
|
kfree_sensitive(tcon->password);
|
2023-03-14 23:32:54 +00:00
|
|
|
#ifdef CONFIG_CIFS_DFS_UPCALL
|
|
|
|
dfs_put_root_smb_sessions(&tcon->dfs_ses_list);
|
|
|
|
#endif
|
2023-06-26 19:04:17 +00:00
|
|
|
kfree(tcon->origin_fullpath);
|
2022-08-12 00:04:29 +00:00
|
|
|
kfree(tcon);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
struct smb_hdr *
|
|
|
|
cifs_buf_get(void)
|
|
|
|
{
|
|
|
|
struct smb_hdr *ret_buf = NULL;
|
2012-01-12 18:40:50 +00:00
|
|
|
/*
|
|
|
|
* SMB2 header is bigger than CIFS one - no problems to clean some
|
|
|
|
* more bytes for CIFS.
|
|
|
|
*/
|
2021-11-04 23:39:01 +00:00
|
|
|
size_t buf_size = sizeof(struct smb2_hdr);
|
2017-07-08 23:48:15 +00:00
|
|
|
|
2012-01-12 18:40:50 +00:00
|
|
|
/*
|
|
|
|
* We could use negotiated size instead of max_msgsize -
|
|
|
|
* but it may be more efficient to always alloc same size
|
|
|
|
* albeit slightly larger than necessary and maxbuffersize
|
|
|
|
* defaults to this and can not be bigger.
|
|
|
|
*/
|
2008-09-15 10:22:54 +00:00
|
|
|
ret_buf = mempool_alloc(cifs_req_poolp, GFP_NOFS);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* clear the first few header bytes */
|
|
|
|
/* for most paths, more is cleared in header_assemble */
|
2017-04-10 02:08:53 +00:00
|
|
|
memset(ret_buf, 0, buf_size + 3);
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_inc(&buf_alloc_count);
|
2005-12-03 21:58:57 +00:00
|
|
|
#ifdef CONFIG_CIFS_STATS2
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_inc(&total_buf_alloc_count);
|
2005-12-03 21:58:57 +00:00
|
|
|
#endif /* CONFIG_CIFS_STATS2 */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
return ret_buf;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_buf_release(void *buf_to_free)
|
|
|
|
{
|
|
|
|
if (buf_to_free == NULL) {
|
2013-05-05 03:12:25 +00:00
|
|
|
/* cifs_dbg(FYI, "Null buffer passed to cifs_buf_release\n");*/
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
2007-07-10 01:16:18 +00:00
|
|
|
mempool_free(buf_to_free, cifs_req_poolp);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_dec(&buf_alloc_count);
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct smb_hdr *
|
|
|
|
cifs_small_buf_get(void)
|
|
|
|
{
|
|
|
|
struct smb_hdr *ret_buf = NULL;
|
|
|
|
|
2007-07-10 01:16:18 +00:00
|
|
|
/* We could use negotiated size instead of max_msgsize -
|
|
|
|
but it may be more efficient to always alloc same size
|
|
|
|
albeit slightly larger than necessary and maxbuffersize
|
2005-04-16 22:20:36 +00:00
|
|
|
defaults to this and can not be bigger */
|
2008-09-15 10:22:54 +00:00
|
|
|
ret_buf = mempool_alloc(cifs_sm_req_poolp, GFP_NOFS);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* No need to clear memory here, cleared in header assemble */
|
|
|
|
/* memset(ret_buf, 0, sizeof(struct smb_hdr) + 27);*/
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_inc(&small_buf_alloc_count);
|
2005-12-03 21:58:57 +00:00
|
|
|
#ifdef CONFIG_CIFS_STATS2
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_inc(&total_small_buf_alloc_count);
|
2005-12-03 21:58:57 +00:00
|
|
|
#endif /* CONFIG_CIFS_STATS2 */
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return ret_buf;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_small_buf_release(void *buf_to_free)
|
|
|
|
{
|
|
|
|
|
|
|
|
if (buf_to_free == NULL) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Null buffer passed to cifs_small_buf_release\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
2007-07-10 01:16:18 +00:00
|
|
|
mempool_free(buf_to_free, cifs_sm_req_poolp);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2022-07-16 04:45:45 +00:00
|
|
|
atomic_dec(&small_buf_alloc_count);
|
2005-04-16 22:20:36 +00:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2014-06-16 14:35:24 +00:00
|
|
|
void
|
|
|
|
free_rsp_buf(int resp_buftype, void *rsp)
|
|
|
|
{
|
|
|
|
if (resp_buftype == CIFS_SMALL_BUFFER)
|
|
|
|
cifs_small_buf_release(rsp);
|
|
|
|
else if (resp_buftype == CIFS_LARGE_BUFFER)
|
|
|
|
cifs_buf_release(rsp);
|
|
|
|
}
|
|
|
|
|
2005-08-17 19:38:22 +00:00
|
|
|
/* NB: MID can not be set if treeCon not passed in, in that
|
|
|
|
case it is responsbility of caller to set the mid */
|
2005-04-16 22:20:36 +00:00
|
|
|
void
|
|
|
|
header_assemble(struct smb_hdr *buffer, char smb_command /* command */ ,
|
2011-05-27 04:34:02 +00:00
|
|
|
const struct cifs_tcon *treeCon, int word_count
|
2005-04-16 22:20:36 +00:00
|
|
|
/* length of fixed section (word count) in two byte units */)
|
|
|
|
{
|
|
|
|
char *temp = (char *) buffer;
|
|
|
|
|
2007-07-10 01:16:18 +00:00
|
|
|
memset(temp, 0, 256); /* bigger than MAX_CIFS_HDR_SIZE */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2011-04-29 05:40:20 +00:00
|
|
|
buffer->smb_buf_length = cpu_to_be32(
|
2007-10-25 21:17:17 +00:00
|
|
|
(2 * word_count) + sizeof(struct smb_hdr) -
|
2005-04-16 22:20:36 +00:00
|
|
|
4 /* RFC 1001 length field does not count */ +
|
2011-04-29 05:40:20 +00:00
|
|
|
2 /* for bcc field itself */) ;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
buffer->Protocol[0] = 0xFF;
|
|
|
|
buffer->Protocol[1] = 'S';
|
|
|
|
buffer->Protocol[2] = 'M';
|
|
|
|
buffer->Protocol[3] = 'B';
|
|
|
|
buffer->Command = smb_command;
|
|
|
|
buffer->Flags = 0x00; /* case sensitive */
|
|
|
|
buffer->Flags2 = SMBFLG2_KNOWS_LONG_NAMES;
|
|
|
|
buffer->Pid = cpu_to_le16((__u16)current->tgid);
|
|
|
|
buffer->PidHigh = cpu_to_le16((__u16)(current->tgid >> 16));
|
|
|
|
if (treeCon) {
|
|
|
|
buffer->Tid = treeCon->tid;
|
|
|
|
if (treeCon->ses) {
|
|
|
|
if (treeCon->ses->capabilities & CAP_UNICODE)
|
|
|
|
buffer->Flags2 |= SMBFLG2_UNICODE;
|
2008-02-07 23:25:02 +00:00
|
|
|
if (treeCon->ses->capabilities & CAP_STATUS32)
|
2005-04-16 22:20:36 +00:00
|
|
|
buffer->Flags2 |= SMBFLG2_ERR_STATUS;
|
2008-02-07 23:25:02 +00:00
|
|
|
|
2005-08-17 19:38:22 +00:00
|
|
|
/* Uid is not converted */
|
|
|
|
buffer->Uid = treeCon->ses->Suid;
|
2021-09-23 23:52:40 +00:00
|
|
|
if (treeCon->ses->server)
|
|
|
|
buffer->Mid = get_next_mid(treeCon->ses->server);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
if (treeCon->Flags & SMB_SHARE_IS_IN_DFS)
|
|
|
|
buffer->Flags2 |= SMBFLG2_DFS;
|
2005-08-19 18:04:29 +00:00
|
|
|
if (treeCon->nocase)
|
|
|
|
buffer->Flags |= SMBFLG_CASELESS;
|
2007-07-07 19:25:05 +00:00
|
|
|
if ((treeCon->ses) && (treeCon->ses->server))
|
2013-05-26 11:01:00 +00:00
|
|
|
if (treeCon->ses->server->sign)
|
2005-04-16 22:20:36 +00:00
|
|
|
buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* endian conversion of flags is now done just before sending */
|
|
|
|
buffer->WordCount = (char) word_count;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2006-09-28 19:43:08 +00:00
|
|
|
static int
|
2013-10-16 15:09:49 +00:00
|
|
|
check_smb_hdr(struct smb_hdr *smb)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2011-01-28 20:05:42 +00:00
|
|
|
/* does it have the right SMB "signature" ? */
|
|
|
|
if (*(__le32 *) smb->Protocol != cpu_to_le32(0x424d53ff)) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "Bad protocol string signature header 0x%x\n",
|
|
|
|
*(unsigned int *)smb->Protocol);
|
2011-01-28 20:05:42 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* if it's a response then accept */
|
|
|
|
if (smb->Flags & SMBFLG_RESPONSE)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* only one valid case where server sends us request */
|
|
|
|
if (smb->Command == SMB_COM_LOCKING_ANDX)
|
|
|
|
return 0;
|
|
|
|
|
2013-11-02 17:50:34 +00:00
|
|
|
cifs_dbg(VFS, "Server sent request, not response. mid=%u\n",
|
|
|
|
get_mid(smb));
|
2005-04-16 22:20:36 +00:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
2015-12-18 19:05:30 +00:00
|
|
|
checkSMB(char *buf, unsigned int total_read, struct TCP_Server_Info *server)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2012-03-23 18:28:02 +00:00
|
|
|
struct smb_hdr *smb = (struct smb_hdr *)buf;
|
2011-10-11 10:41:32 +00:00
|
|
|
__u32 rfclen = be32_to_cpu(smb->smb_buf_length);
|
2005-10-10 18:48:26 +00:00
|
|
|
__u32 clc_len; /* calculated length */
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "checkSMB Length: 0x%x, smb_buf_length: 0x%x\n",
|
|
|
|
total_read, rfclen);
|
2006-10-12 17:49:24 +00:00
|
|
|
|
2011-10-11 10:41:32 +00:00
|
|
|
/* is this frame too small to even get to a BCC? */
|
|
|
|
if (total_read < 2 + sizeof(struct smb_hdr)) {
|
|
|
|
if ((total_read >= sizeof(struct smb_hdr) - 1)
|
2005-04-16 22:20:36 +00:00
|
|
|
&& (smb->Status.CifsError != 0)) {
|
2011-10-11 10:41:32 +00:00
|
|
|
/* it's an error return */
|
2006-10-12 17:49:24 +00:00
|
|
|
smb->WordCount = 0;
|
|
|
|
/* some error cases do not return wct and bcc */
|
|
|
|
return 0;
|
2011-10-11 10:41:32 +00:00
|
|
|
} else if ((total_read == sizeof(struct smb_hdr) + 1) &&
|
2006-10-12 17:49:24 +00:00
|
|
|
(smb->WordCount == 0)) {
|
2007-07-10 01:16:18 +00:00
|
|
|
char *tmp = (char *)smb;
|
2006-10-12 17:49:24 +00:00
|
|
|
/* Need to work around a bug in two servers here */
|
|
|
|
/* First, check if the part of bcc they sent was zero */
|
|
|
|
if (tmp[sizeof(struct smb_hdr)] == 0) {
|
|
|
|
/* some servers return only half of bcc
|
|
|
|
* on simple responses (wct, bcc both zero)
|
|
|
|
* in particular have seen this on
|
|
|
|
* ulogoffX and FindClose. This leaves
|
|
|
|
* one byte of bcc potentially unitialized
|
|
|
|
*/
|
|
|
|
/* zero rest of bcc */
|
|
|
|
tmp[sizeof(struct smb_hdr)+1] = 0;
|
2006-03-02 00:07:08 +00:00
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "rcvd invalid byte count (bcc)\n");
|
2006-10-12 17:49:24 +00:00
|
|
|
} else {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "Length less than smb header size\n");
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2011-10-11 10:41:32 +00:00
|
|
|
return -EIO;
|
2023-12-15 22:59:14 +00:00
|
|
|
} else if (total_read < sizeof(*smb) + 2 * smb->WordCount) {
|
|
|
|
cifs_dbg(VFS, "%s: can't read BCC due to invalid WordCount(%u)\n",
|
|
|
|
__func__, smb->WordCount);
|
|
|
|
return -EIO;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2011-10-11 10:41:32 +00:00
|
|
|
/* otherwise, there is enough to get to the BCC */
|
2013-10-16 15:09:49 +00:00
|
|
|
if (check_smb_hdr(smb))
|
2011-10-11 10:41:32 +00:00
|
|
|
return -EIO;
|
2022-08-17 17:14:02 +00:00
|
|
|
clc_len = smbCalcSize(smb);
|
2006-02-24 06:15:11 +00:00
|
|
|
|
2011-10-11 10:41:32 +00:00
|
|
|
if (4 + rfclen != total_read) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "Length read does not match RFC1001 length %d\n",
|
|
|
|
rfclen);
|
2011-10-11 10:41:32 +00:00
|
|
|
return -EIO;
|
2006-02-24 06:15:11 +00:00
|
|
|
}
|
|
|
|
|
2011-10-11 10:41:32 +00:00
|
|
|
if (4 + rfclen != clc_len) {
|
2013-11-02 17:50:34 +00:00
|
|
|
__u16 mid = get_mid(smb);
|
2006-02-24 06:15:11 +00:00
|
|
|
/* check if bcc wrapped around for large read responses */
|
2011-10-11 10:41:32 +00:00
|
|
|
if ((rfclen > 64 * 1024) && (rfclen > clc_len)) {
|
2006-02-24 06:15:11 +00:00
|
|
|
/* check if lengths match mod 64K */
|
2011-10-11 10:41:32 +00:00
|
|
|
if (((4 + rfclen) & 0xFFFF) == (clc_len & 0xFFFF))
|
2007-07-10 01:16:18 +00:00
|
|
|
return 0; /* bcc wrapped */
|
2006-02-24 06:15:11 +00:00
|
|
|
}
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Calculated size %u vs length %u mismatch for mid=%u\n",
|
2013-11-02 17:50:34 +00:00
|
|
|
clc_len, 4 + rfclen, mid);
|
2011-01-31 14:14:17 +00:00
|
|
|
|
2011-10-11 10:41:32 +00:00
|
|
|
if (4 + rfclen < clc_len) {
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "RFC1001 size %u smaller than SMB for mid=%u\n",
|
2013-11-02 17:50:34 +00:00
|
|
|
rfclen, mid);
|
2011-10-11 10:41:32 +00:00
|
|
|
return -EIO;
|
|
|
|
} else if (rfclen > clc_len + 512) {
|
2011-01-31 14:14:17 +00:00
|
|
|
/*
|
|
|
|
* Some servers (Windows XP in particular) send more
|
|
|
|
* data than the lengths in the SMB packet would
|
|
|
|
* indicate on certain calls (byte range locks and
|
|
|
|
* trans2 find first calls in particular). While the
|
|
|
|
* client can handle such a frame by ignoring the
|
|
|
|
* trailing data, we choose limit the amount of extra
|
|
|
|
* data to 512 bytes.
|
|
|
|
*/
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(VFS, "RFC1001 size %u more than 512 bytes larger than SMB for mid=%u\n",
|
2013-11-02 17:50:34 +00:00
|
|
|
rfclen, mid);
|
2011-10-11 10:41:32 +00:00
|
|
|
return -EIO;
|
2006-03-02 00:07:08 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2005-09-22 05:05:57 +00:00
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-04-29 00:06:05 +00:00
|
|
|
|
|
|
|
bool
|
2012-03-23 18:28:02 +00:00
|
|
|
is_valid_oplock_break(char *buffer, struct TCP_Server_Info *srv)
|
2007-07-10 01:16:18 +00:00
|
|
|
{
|
2012-03-23 18:28:02 +00:00
|
|
|
struct smb_hdr *buf = (struct smb_hdr *)buffer;
|
2007-07-10 01:16:18 +00:00
|
|
|
struct smb_com_lock_req *pSMB = (struct smb_com_lock_req *)buf;
|
2022-10-28 09:52:26 +00:00
|
|
|
struct TCP_Server_Info *pserver;
|
2011-05-27 04:34:02 +00:00
|
|
|
struct cifs_ses *ses;
|
|
|
|
struct cifs_tcon *tcon;
|
2008-11-15 16:12:47 +00:00
|
|
|
struct cifsInodeInfo *pCifsInode;
|
2005-04-16 22:20:36 +00:00
|
|
|
struct cifsFileInfo *netfile;
|
|
|
|
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Checking for oplock break or dnotify response\n");
|
2007-07-07 19:25:05 +00:00
|
|
|
if ((pSMB->hdr.Command == SMB_COM_NT_TRANSACT) &&
|
2005-04-16 22:20:36 +00:00
|
|
|
(pSMB->hdr.Flags & SMBFLG_RESPONSE)) {
|
2007-07-10 01:16:18 +00:00
|
|
|
struct smb_com_transaction_change_notify_rsp *pSMBr =
|
2005-04-16 22:20:36 +00:00
|
|
|
(struct smb_com_transaction_change_notify_rsp *)buf;
|
2007-07-10 01:16:18 +00:00
|
|
|
struct file_notify_information *pnotify;
|
2005-04-16 22:20:36 +00:00
|
|
|
__u32 data_offset = 0;
|
2018-09-06 09:47:01 +00:00
|
|
|
size_t len = srv->total_read - sizeof(pSMBr->hdr.smb_buf_length);
|
|
|
|
|
2011-05-04 12:05:26 +00:00
|
|
|
if (get_bcc(buf) > sizeof(struct file_notify_information)) {
|
2005-04-16 22:20:36 +00:00
|
|
|
data_offset = le32_to_cpu(pSMBr->DataOffset);
|
|
|
|
|
2018-09-06 09:47:01 +00:00
|
|
|
if (data_offset >
|
|
|
|
len - sizeof(struct file_notify_information)) {
|
2020-04-15 05:42:53 +00:00
|
|
|
cifs_dbg(FYI, "Invalid data_offset %u\n",
|
2018-09-06 09:47:01 +00:00
|
|
|
data_offset);
|
|
|
|
return true;
|
|
|
|
}
|
2006-05-31 22:40:51 +00:00
|
|
|
pnotify = (struct file_notify_information *)
|
|
|
|
((char *)&pSMBr->hdr.Protocol + data_offset);
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "dnotify on %s Action: 0x%x\n",
|
2010-04-21 03:50:45 +00:00
|
|
|
pnotify->FileName, pnotify->Action);
|
2007-07-10 01:16:18 +00:00
|
|
|
/* cifs_dump_mem("Rcvd notify Data: ",buf,
|
2006-05-31 22:40:51 +00:00
|
|
|
sizeof(struct smb_hdr)+60); */
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-07-07 19:25:05 +00:00
|
|
|
if (pSMBr->hdr.Status.CifsError) {
|
2014-08-03 02:16:48 +00:00
|
|
|
cifs_dbg(FYI, "notify err 0x%x\n",
|
2013-05-05 03:12:25 +00:00
|
|
|
pSMBr->hdr.Status.CifsError);
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-04-29 00:06:05 +00:00
|
|
|
return false;
|
2007-07-10 01:16:18 +00:00
|
|
|
}
|
2007-07-07 19:25:05 +00:00
|
|
|
if (pSMB->hdr.Command != SMB_COM_LOCKING_ANDX)
|
2008-04-29 00:06:05 +00:00
|
|
|
return false;
|
2007-07-07 19:25:05 +00:00
|
|
|
if (pSMB->hdr.Flags & SMBFLG_RESPONSE) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/* no sense logging error on invalid handle on oplock
|
|
|
|
break - harmless race between close request and oplock
|
|
|
|
break response is expected from time to time writing out
|
|
|
|
large dirty files cached on the client */
|
2007-07-10 01:16:18 +00:00
|
|
|
if ((NT_STATUS_INVALID_HANDLE) ==
|
|
|
|
le32_to_cpu(pSMB->hdr.Status.CifsError)) {
|
2020-04-15 05:42:53 +00:00
|
|
|
cifs_dbg(FYI, "Invalid handle on oplock break\n");
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2007-07-10 01:16:18 +00:00
|
|
|
} else if (ERRbadfid ==
|
2005-04-16 22:20:36 +00:00
|
|
|
le16_to_cpu(pSMB->hdr.Status.DosError.Error)) {
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
} else {
|
2008-04-29 00:06:05 +00:00
|
|
|
return false; /* on valid oplock brk we get "request" */
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
2007-07-07 19:25:05 +00:00
|
|
|
if (pSMB->hdr.WordCount != 8)
|
2008-04-29 00:06:05 +00:00
|
|
|
return false;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2014-08-03 02:16:48 +00:00
|
|
|
cifs_dbg(FYI, "oplock type 0x%x level 0x%x\n",
|
2010-04-21 03:50:45 +00:00
|
|
|
pSMB->LockType, pSMB->OplockLevel);
|
2007-07-07 19:25:05 +00:00
|
|
|
if (!(pSMB->LockType & LOCKING_ANDX_OPLOCK_RELEASE))
|
2008-04-29 00:06:05 +00:00
|
|
|
return false;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2022-10-28 09:52:26 +00:00
|
|
|
/* If server is a channel, select the primary channel */
|
2023-08-25 04:44:05 +00:00
|
|
|
pserver = SERVER_IS_CHAN(srv) ? srv->primary_server : srv;
|
2022-10-28 09:52:26 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/* look up tcon based on tid & uid */
|
2010-10-18 17:59:37 +00:00
|
|
|
spin_lock(&cifs_tcp_ses_lock);
|
2022-10-28 09:52:26 +00:00
|
|
|
list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) {
|
2022-07-22 17:02:59 +00:00
|
|
|
list_for_each_entry(tcon, &ses->tcon_list, tcon_list) {
|
2008-11-15 16:12:47 +00:00
|
|
|
if (tcon->tid != buf->Tid)
|
|
|
|
continue;
|
|
|
|
|
2012-05-28 10:16:31 +00:00
|
|
|
cifs_stats_inc(&tcon->stats.cifs_stats.num_oplock_brks);
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_lock(&tcon->open_file_lock);
|
2022-07-22 17:02:59 +00:00
|
|
|
list_for_each_entry(netfile, &tcon->openFileList, tlist) {
|
2012-09-18 23:20:26 +00:00
|
|
|
if (pSMB->Fid != netfile->fid.netfid)
|
2008-11-15 16:12:47 +00:00
|
|
|
continue;
|
|
|
|
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "file id match, oplock break\n");
|
2015-03-17 22:25:59 +00:00
|
|
|
pCifsInode = CIFS_I(d_inode(netfile->dentry));
|
2010-07-20 20:09:02 +00:00
|
|
|
|
2014-03-11 16:11:47 +00:00
|
|
|
set_bit(CIFS_INODE_PENDING_OPLOCK_BREAK,
|
|
|
|
&pCifsInode->flags);
|
|
|
|
|
2019-10-29 23:51:19 +00:00
|
|
|
netfile->oplock_epoch = 0;
|
|
|
|
netfile->oplock_level = pSMB->OplockLevel;
|
2010-07-20 20:09:02 +00:00
|
|
|
netfile->oplock_break_cancelled = false;
|
2019-10-29 23:51:19 +00:00
|
|
|
cifs_queue_oplock_break(netfile);
|
2010-07-20 20:09:02 +00:00
|
|
|
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_unlock(&tcon->open_file_lock);
|
2010-10-18 17:59:37 +00:00
|
|
|
spin_unlock(&cifs_tcp_ses_lock);
|
2008-11-15 16:12:47 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_unlock(&tcon->open_file_lock);
|
2010-10-18 17:59:37 +00:00
|
|
|
spin_unlock(&cifs_tcp_ses_lock);
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "No matching file for oplock break\n");
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
2010-10-18 17:59:37 +00:00
|
|
|
spin_unlock(&cifs_tcp_ses_lock);
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Can not process oplock break for non-existent connection\n");
|
2008-04-29 00:06:05 +00:00
|
|
|
return true;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2012-03-23 18:28:02 +00:00
|
|
|
dump_smb(void *buf, int smb_buf_length)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
if (traceSMB == 0)
|
|
|
|
return;
|
|
|
|
|
2014-08-27 13:49:43 +00:00
|
|
|
print_hex_dump(KERN_DEBUG, "", DUMP_PREFIX_NONE, 8, 2, buf,
|
|
|
|
smb_buf_length, true);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2005-04-29 05:41:05 +00:00
|
|
|
|
2009-11-06 19:18:29 +00:00
|
|
|
void
|
|
|
|
cifs_autodisable_serverino(struct cifs_sb_info *cifs_sb)
|
|
|
|
{
|
|
|
|
if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) {
|
2018-11-16 15:13:25 +00:00
|
|
|
struct cifs_tcon *tcon = NULL;
|
|
|
|
|
|
|
|
if (cifs_sb->master_tlink)
|
|
|
|
tcon = cifs_sb_master_tcon(cifs_sb);
|
|
|
|
|
2009-11-16 06:33:16 +00:00
|
|
|
cifs_sb->mnt_cifs_flags &= ~CIFS_MOUNT_SERVER_INUM;
|
2019-06-18 19:16:02 +00:00
|
|
|
cifs_sb->mnt_cifs_serverino_autodisabled = true;
|
2020-04-15 05:42:53 +00:00
|
|
|
cifs_dbg(VFS, "Autodisabling the use of server inode numbers on %s\n",
|
2022-09-21 19:05:53 +00:00
|
|
|
tcon ? tcon->tree_name : "new server");
|
2020-04-15 05:42:53 +00:00
|
|
|
cifs_dbg(VFS, "The server doesn't seem to support them properly or the files might be on different servers (DFS)\n");
|
2018-11-16 15:13:25 +00:00
|
|
|
cifs_dbg(VFS, "Hardlinks will not be recognized on this mount. Consider mounting with the \"noserverino\" option to silence this message.\n");
|
|
|
|
|
2009-11-06 19:18:29 +00:00
|
|
|
}
|
|
|
|
}
|
2010-11-02 09:00:42 +00:00
|
|
|
|
2010-11-03 07:58:57 +00:00
|
|
|
void cifs_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock)
|
2010-11-02 09:00:42 +00:00
|
|
|
{
|
2010-11-03 07:58:57 +00:00
|
|
|
oplock &= 0xF;
|
2010-11-02 09:00:42 +00:00
|
|
|
|
2010-11-03 07:58:57 +00:00
|
|
|
if (oplock == OPLOCK_EXCLUSIVE) {
|
2013-09-05 09:01:06 +00:00
|
|
|
cinode->oplock = CIFS_CACHE_WRITE_FLG | CIFS_CACHE_READ_FLG;
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Exclusive Oplock granted on inode %p\n",
|
netfs: Fix gcc-12 warning by embedding vfs inode in netfs_i_context
While randstruct was satisfied with using an open-coded "void *" offset
cast for the netfs_i_context <-> inode casting, __builtin_object_size() as
used by FORTIFY_SOURCE was not as easily fooled. This was causing the
following complaint[1] from gcc v12:
In file included from include/linux/string.h:253,
from include/linux/ceph/ceph_debug.h:7,
from fs/ceph/inode.c:2:
In function 'fortify_memset_chk',
inlined from 'netfs_i_context_init' at include/linux/netfs.h:326:2,
inlined from 'ceph_alloc_inode' at fs/ceph/inode.c:463:2:
include/linux/fortify-string.h:242:25: warning: call to '__write_overflow_field' declared with attribute warning: detected write beyond size of field (1st parameter); maybe use struct_group()? [-Wattribute-warning]
242 | __write_overflow_field(p_size_field, size);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Fix this by embedding a struct inode into struct netfs_i_context (which
should perhaps be renamed to struct netfs_inode). The struct inode
vfs_inode fields are then removed from the 9p, afs, ceph and cifs inode
structs and vfs_inode is then simply changed to "netfs.inode" in those
filesystems.
Further, rename netfs_i_context to netfs_inode, get rid of the
netfs_inode() function that converted a netfs_i_context pointer to an
inode pointer (that can now be done with &ctx->inode) and rename the
netfs_i_context() function to netfs_inode() (which is now a wrapper
around container_of()).
Most of the changes were done with:
perl -p -i -e 's/vfs_inode/netfs.inode/'g \
`git grep -l 'vfs_inode' -- fs/{9p,afs,ceph,cifs}/*.[ch]`
Kees suggested doing it with a pair structure[2] and a special
declarator to insert that into the network filesystem's inode
wrapper[3], but I think it's cleaner to embed it - and then it doesn't
matter if struct randomisation reorders things.
Dave Chinner suggested using a filesystem-specific VFS_I() function in
each filesystem to convert that filesystem's own inode wrapper struct
into the VFS inode struct[4].
Version #2:
- Fix a couple of missed name changes due to a disabled cifs option.
- Rename nfs_i_context to nfs_inode
- Use "netfs" instead of "nic" as the member name in per-fs inode wrapper
structs.
[ This also undoes commit 507160f46c55 ("netfs: gcc-12: temporarily
disable '-Wattribute-warning' for now") that is no longer needed ]
Fixes: bc899ee1c898 ("netfs: Add a netfs inode context")
Reported-by: Jeff Layton <jlayton@kernel.org>
Signed-off-by: David Howells <dhowells@redhat.com>
Reviewed-by: Jeff Layton <jlayton@kernel.org>
Reviewed-by: Kees Cook <keescook@chromium.org>
Reviewed-by: Xiubo Li <xiubli@redhat.com>
cc: Jonathan Corbet <corbet@lwn.net>
cc: Eric Van Hensbergen <ericvh@gmail.com>
cc: Latchesar Ionkov <lucho@ionkov.net>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: Marc Dionne <marc.dionne@auristor.com>
cc: Ilya Dryomov <idryomov@gmail.com>
cc: Steve French <smfrench@gmail.com>
cc: William Kucharski <william.kucharski@oracle.com>
cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
cc: Dave Chinner <david@fromorbit.com>
cc: linux-doc@vger.kernel.org
cc: v9fs-developer@lists.sourceforge.net
cc: linux-afs@lists.infradead.org
cc: ceph-devel@vger.kernel.org
cc: linux-cifs@vger.kernel.org
cc: samba-technical@lists.samba.org
cc: linux-fsdevel@vger.kernel.org
cc: linux-hardening@vger.kernel.org
Link: https://lore.kernel.org/r/d2ad3a3d7bdd794c6efb562d2f2b655fb67756b9.camel@kernel.org/ [1]
Link: https://lore.kernel.org/r/20220517210230.864239-1-keescook@chromium.org/ [2]
Link: https://lore.kernel.org/r/20220518202212.2322058-1-keescook@chromium.org/ [3]
Link: https://lore.kernel.org/r/20220524101205.GI2306852@dread.disaster.area/ [4]
Link: https://lore.kernel.org/r/165296786831.3591209.12111293034669289733.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165305805651.4094995.7763502506786714216.stgit@warthog.procyon.org.uk # v2
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2022-06-09 20:46:04 +00:00
|
|
|
&cinode->netfs.inode);
|
2010-11-03 07:58:57 +00:00
|
|
|
} else if (oplock == OPLOCK_READ) {
|
2013-09-05 09:01:06 +00:00
|
|
|
cinode->oplock = CIFS_CACHE_READ_FLG;
|
2013-05-05 03:12:25 +00:00
|
|
|
cifs_dbg(FYI, "Level II Oplock granted on inode %p\n",
|
netfs: Fix gcc-12 warning by embedding vfs inode in netfs_i_context
While randstruct was satisfied with using an open-coded "void *" offset
cast for the netfs_i_context <-> inode casting, __builtin_object_size() as
used by FORTIFY_SOURCE was not as easily fooled. This was causing the
following complaint[1] from gcc v12:
In file included from include/linux/string.h:253,
from include/linux/ceph/ceph_debug.h:7,
from fs/ceph/inode.c:2:
In function 'fortify_memset_chk',
inlined from 'netfs_i_context_init' at include/linux/netfs.h:326:2,
inlined from 'ceph_alloc_inode' at fs/ceph/inode.c:463:2:
include/linux/fortify-string.h:242:25: warning: call to '__write_overflow_field' declared with attribute warning: detected write beyond size of field (1st parameter); maybe use struct_group()? [-Wattribute-warning]
242 | __write_overflow_field(p_size_field, size);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Fix this by embedding a struct inode into struct netfs_i_context (which
should perhaps be renamed to struct netfs_inode). The struct inode
vfs_inode fields are then removed from the 9p, afs, ceph and cifs inode
structs and vfs_inode is then simply changed to "netfs.inode" in those
filesystems.
Further, rename netfs_i_context to netfs_inode, get rid of the
netfs_inode() function that converted a netfs_i_context pointer to an
inode pointer (that can now be done with &ctx->inode) and rename the
netfs_i_context() function to netfs_inode() (which is now a wrapper
around container_of()).
Most of the changes were done with:
perl -p -i -e 's/vfs_inode/netfs.inode/'g \
`git grep -l 'vfs_inode' -- fs/{9p,afs,ceph,cifs}/*.[ch]`
Kees suggested doing it with a pair structure[2] and a special
declarator to insert that into the network filesystem's inode
wrapper[3], but I think it's cleaner to embed it - and then it doesn't
matter if struct randomisation reorders things.
Dave Chinner suggested using a filesystem-specific VFS_I() function in
each filesystem to convert that filesystem's own inode wrapper struct
into the VFS inode struct[4].
Version #2:
- Fix a couple of missed name changes due to a disabled cifs option.
- Rename nfs_i_context to nfs_inode
- Use "netfs" instead of "nic" as the member name in per-fs inode wrapper
structs.
[ This also undoes commit 507160f46c55 ("netfs: gcc-12: temporarily
disable '-Wattribute-warning' for now") that is no longer needed ]
Fixes: bc899ee1c898 ("netfs: Add a netfs inode context")
Reported-by: Jeff Layton <jlayton@kernel.org>
Signed-off-by: David Howells <dhowells@redhat.com>
Reviewed-by: Jeff Layton <jlayton@kernel.org>
Reviewed-by: Kees Cook <keescook@chromium.org>
Reviewed-by: Xiubo Li <xiubli@redhat.com>
cc: Jonathan Corbet <corbet@lwn.net>
cc: Eric Van Hensbergen <ericvh@gmail.com>
cc: Latchesar Ionkov <lucho@ionkov.net>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: Marc Dionne <marc.dionne@auristor.com>
cc: Ilya Dryomov <idryomov@gmail.com>
cc: Steve French <smfrench@gmail.com>
cc: William Kucharski <william.kucharski@oracle.com>
cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
cc: Dave Chinner <david@fromorbit.com>
cc: linux-doc@vger.kernel.org
cc: v9fs-developer@lists.sourceforge.net
cc: linux-afs@lists.infradead.org
cc: ceph-devel@vger.kernel.org
cc: linux-cifs@vger.kernel.org
cc: samba-technical@lists.samba.org
cc: linux-fsdevel@vger.kernel.org
cc: linux-hardening@vger.kernel.org
Link: https://lore.kernel.org/r/d2ad3a3d7bdd794c6efb562d2f2b655fb67756b9.camel@kernel.org/ [1]
Link: https://lore.kernel.org/r/20220517210230.864239-1-keescook@chromium.org/ [2]
Link: https://lore.kernel.org/r/20220518202212.2322058-1-keescook@chromium.org/ [3]
Link: https://lore.kernel.org/r/20220524101205.GI2306852@dread.disaster.area/ [4]
Link: https://lore.kernel.org/r/165296786831.3591209.12111293034669289733.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165305805651.4094995.7763502506786714216.stgit@warthog.procyon.org.uk # v2
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2022-06-09 20:46:04 +00:00
|
|
|
&cinode->netfs.inode);
|
2013-09-05 09:01:06 +00:00
|
|
|
} else
|
|
|
|
cinode->oplock = 0;
|
2010-11-02 09:00:42 +00:00
|
|
|
}
|
2011-09-26 14:56:44 +00:00
|
|
|
|
2014-03-11 16:11:47 +00:00
|
|
|
/*
|
|
|
|
* We wait for oplock breaks to be processed before we attempt to perform
|
|
|
|
* writes.
|
|
|
|
*/
|
|
|
|
int cifs_get_writer(struct cifsInodeInfo *cinode)
|
|
|
|
{
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
start:
|
|
|
|
rc = wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_OPLOCK_BREAK,
|
sched: Remove proliferation of wait_on_bit() action functions
The current "wait_on_bit" interface requires an 'action'
function to be provided which does the actual waiting.
There are over 20 such functions, many of them identical.
Most cases can be satisfied by one of just two functions, one
which uses io_schedule() and one which just uses schedule().
So:
Rename wait_on_bit and wait_on_bit_lock to
wait_on_bit_action and wait_on_bit_lock_action
to make it explicit that they need an action function.
Introduce new wait_on_bit{,_lock} and wait_on_bit{,_lock}_io
which are *not* given an action function but implicitly use
a standard one.
The decision to error-out if a signal is pending is now made
based on the 'mode' argument rather than being encoded in the action
function.
All instances of the old wait_on_bit and wait_on_bit_lock which
can use the new version have been changed accordingly and their
action functions have been discarded.
wait_on_bit{_lock} does not return any specific error code in the
event of a signal so the caller must check for non-zero and
interpolate their own error code as appropriate.
The wait_on_bit() call in __fscache_wait_on_invalidate() was
ambiguous as it specified TASK_UNINTERRUPTIBLE but used
fscache_wait_bit_interruptible as an action function.
David Howells confirms this should be uniformly
"uninterruptible"
The main remaining user of wait_on_bit{,_lock}_action is NFS
which needs to use a freezer-aware schedule() call.
A comment in fs/gfs2/glock.c notes that having multiple 'action'
functions is useful as they display differently in the 'wchan'
field of 'ps'. (and /proc/$PID/wchan).
As the new bit_wait{,_io} functions are tagged "__sched", they
will not show up at all, but something higher in the stack. So
the distinction will still be visible, only with different
function names (gds2_glock_wait versus gfs2_glock_dq_wait in the
gfs2/glock.c case).
Since first version of this patch (against 3.15) two new action
functions appeared, on in NFS and one in CIFS. CIFS also now
uses an action function that makes the same freezer aware
schedule call as NFS.
Signed-off-by: NeilBrown <neilb@suse.de>
Acked-by: David Howells <dhowells@redhat.com> (fscache, keys)
Acked-by: Steven Whitehouse <swhiteho@redhat.com> (gfs2)
Acked-by: Peter Zijlstra <peterz@infradead.org>
Cc: Oleg Nesterov <oleg@redhat.com>
Cc: Steve French <sfrench@samba.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Link: http://lkml.kernel.org/r/20140707051603.28027.72349.stgit@notabene.brown
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2014-07-07 05:16:04 +00:00
|
|
|
TASK_KILLABLE);
|
2014-03-11 16:11:47 +00:00
|
|
|
if (rc)
|
|
|
|
return rc;
|
|
|
|
|
|
|
|
spin_lock(&cinode->writers_lock);
|
|
|
|
if (!cinode->writers)
|
|
|
|
set_bit(CIFS_INODE_PENDING_WRITERS, &cinode->flags);
|
|
|
|
cinode->writers++;
|
|
|
|
/* Check to see if we have started servicing an oplock break */
|
|
|
|
if (test_bit(CIFS_INODE_PENDING_OPLOCK_BREAK, &cinode->flags)) {
|
|
|
|
cinode->writers--;
|
|
|
|
if (cinode->writers == 0) {
|
|
|
|
clear_bit(CIFS_INODE_PENDING_WRITERS, &cinode->flags);
|
|
|
|
wake_up_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS);
|
|
|
|
}
|
|
|
|
spin_unlock(&cinode->writers_lock);
|
|
|
|
goto start;
|
|
|
|
}
|
|
|
|
spin_unlock(&cinode->writers_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void cifs_put_writer(struct cifsInodeInfo *cinode)
|
|
|
|
{
|
|
|
|
spin_lock(&cinode->writers_lock);
|
|
|
|
cinode->writers--;
|
|
|
|
if (cinode->writers == 0) {
|
|
|
|
clear_bit(CIFS_INODE_PENDING_WRITERS, &cinode->flags);
|
|
|
|
wake_up_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS);
|
|
|
|
}
|
|
|
|
spin_unlock(&cinode->writers_lock);
|
|
|
|
}
|
|
|
|
|
2019-03-29 09:49:12 +00:00
|
|
|
/**
|
|
|
|
* cifs_queue_oplock_break - queue the oplock break handler for cfile
|
2021-09-20 12:14:15 +00:00
|
|
|
* @cfile: The file to break the oplock on
|
2019-03-29 09:49:12 +00:00
|
|
|
*
|
|
|
|
* This function is called from the demultiplex thread when it
|
|
|
|
* receives an oplock break for @cfile.
|
|
|
|
*
|
|
|
|
* Assumes the tcon->open_file_lock is held.
|
|
|
|
* Assumes cfile->file_info_lock is NOT held.
|
|
|
|
*/
|
|
|
|
void cifs_queue_oplock_break(struct cifsFileInfo *cfile)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Bump the handle refcount now while we hold the
|
|
|
|
* open_file_lock to enforce the validity of it for the oplock
|
|
|
|
* break handler. The matching put is done at the end of the
|
|
|
|
* handler.
|
|
|
|
*/
|
|
|
|
cifsFileInfo_get(cfile);
|
|
|
|
|
|
|
|
queue_work(cifsoplockd_wq, &cfile->oplock_break);
|
|
|
|
}
|
|
|
|
|
2014-03-11 16:11:47 +00:00
|
|
|
void cifs_done_oplock_break(struct cifsInodeInfo *cinode)
|
|
|
|
{
|
|
|
|
clear_bit(CIFS_INODE_PENDING_OPLOCK_BREAK, &cinode->flags);
|
|
|
|
wake_up_bit(&cinode->flags, CIFS_INODE_PENDING_OPLOCK_BREAK);
|
|
|
|
}
|
|
|
|
|
2011-09-26 14:56:44 +00:00
|
|
|
bool
|
|
|
|
backup_cred(struct cifs_sb_info *cifs_sb)
|
|
|
|
{
|
|
|
|
if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_CIFS_BACKUPUID) {
|
2020-12-12 19:40:50 +00:00
|
|
|
if (uid_eq(cifs_sb->ctx->backupuid, current_fsuid()))
|
2011-09-26 14:56:44 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_CIFS_BACKUPGID) {
|
2020-12-12 19:40:50 +00:00
|
|
|
if (in_group_p(cifs_sb->ctx->backupgid))
|
2011-09-26 14:56:44 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
2012-09-19 13:22:45 +00:00
|
|
|
|
|
|
|
void
|
|
|
|
cifs_del_pending_open(struct cifs_pending_open *open)
|
|
|
|
{
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_lock(&tlink_tcon(open->tlink)->open_file_lock);
|
2012-09-19 13:22:45 +00:00
|
|
|
list_del(&open->olist);
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_unlock(&tlink_tcon(open->tlink)->open_file_lock);
|
2012-09-19 13:22:45 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_add_pending_open_locked(struct cifs_fid *fid, struct tcon_link *tlink,
|
|
|
|
struct cifs_pending_open *open)
|
|
|
|
{
|
|
|
|
memcpy(open->lease_key, fid->lease_key, SMB2_LEASE_KEY_SIZE);
|
|
|
|
open->oplock = CIFS_OPLOCK_NO_CHANGE;
|
|
|
|
open->tlink = tlink;
|
|
|
|
fid->pending_open = open;
|
|
|
|
list_add_tail(&open->olist, &tlink_tcon(tlink)->pending_opens);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_add_pending_open(struct cifs_fid *fid, struct tcon_link *tlink,
|
|
|
|
struct cifs_pending_open *open)
|
|
|
|
{
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_lock(&tlink_tcon(tlink)->open_file_lock);
|
2012-09-19 13:22:45 +00:00
|
|
|
cifs_add_pending_open_locked(fid, tlink, open);
|
2016-09-22 23:58:16 +00:00
|
|
|
spin_unlock(&tlink_tcon(open->tlink)->open_file_lock);
|
2012-09-19 13:22:45 +00:00
|
|
|
}
|
2017-02-13 15:03:47 +00:00
|
|
|
|
2021-05-05 10:56:47 +00:00
|
|
|
/*
|
|
|
|
* Critical section which runs after acquiring deferred_lock.
|
2021-05-20 16:45:01 +00:00
|
|
|
* As there is no reference count on cifs_deferred_close, pdclose
|
|
|
|
* should not be used outside deferred_lock.
|
2021-05-05 10:56:47 +00:00
|
|
|
*/
|
2021-04-13 05:26:42 +00:00
|
|
|
bool
|
|
|
|
cifs_is_deferred_close(struct cifsFileInfo *cfile, struct cifs_deferred_close **pdclose)
|
|
|
|
{
|
|
|
|
struct cifs_deferred_close *dclose;
|
|
|
|
|
|
|
|
list_for_each_entry(dclose, &CIFS_I(d_inode(cfile->dentry))->deferred_closes, dlist) {
|
|
|
|
if ((dclose->netfid == cfile->fid.netfid) &&
|
|
|
|
(dclose->persistent_fid == cfile->fid.persistent_fid) &&
|
|
|
|
(dclose->volatile_fid == cfile->fid.volatile_fid)) {
|
|
|
|
*pdclose = dclose;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2021-05-05 10:56:47 +00:00
|
|
|
/*
|
|
|
|
* Critical section which runs after acquiring deferred_lock.
|
|
|
|
*/
|
2021-04-13 05:26:42 +00:00
|
|
|
void
|
|
|
|
cifs_add_deferred_close(struct cifsFileInfo *cfile, struct cifs_deferred_close *dclose)
|
|
|
|
{
|
|
|
|
bool is_deferred = false;
|
|
|
|
struct cifs_deferred_close *pdclose;
|
|
|
|
|
|
|
|
is_deferred = cifs_is_deferred_close(cfile, &pdclose);
|
|
|
|
if (is_deferred) {
|
|
|
|
kfree(dclose);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
dclose->tlink = cfile->tlink;
|
|
|
|
dclose->netfid = cfile->fid.netfid;
|
|
|
|
dclose->persistent_fid = cfile->fid.persistent_fid;
|
|
|
|
dclose->volatile_fid = cfile->fid.volatile_fid;
|
|
|
|
list_add_tail(&dclose->dlist, &CIFS_I(d_inode(cfile->dentry))->deferred_closes);
|
|
|
|
}
|
|
|
|
|
2021-05-05 10:56:47 +00:00
|
|
|
/*
|
|
|
|
* Critical section which runs after acquiring deferred_lock.
|
|
|
|
*/
|
2021-04-13 05:26:42 +00:00
|
|
|
void
|
|
|
|
cifs_del_deferred_close(struct cifsFileInfo *cfile)
|
|
|
|
{
|
|
|
|
bool is_deferred = false;
|
|
|
|
struct cifs_deferred_close *dclose;
|
|
|
|
|
|
|
|
is_deferred = cifs_is_deferred_close(cfile, &dclose);
|
|
|
|
if (!is_deferred)
|
|
|
|
return;
|
|
|
|
list_del(&dclose->dlist);
|
|
|
|
kfree(dclose);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_close_deferred_file(struct cifsInodeInfo *cifs_inode)
|
|
|
|
{
|
|
|
|
struct cifsFileInfo *cfile = NULL;
|
2021-08-09 09:32:46 +00:00
|
|
|
struct file_list *tmp_list, *tmp_next_list;
|
|
|
|
struct list_head file_head;
|
2021-07-29 07:45:29 +00:00
|
|
|
|
|
|
|
if (cifs_inode == NULL)
|
|
|
|
return;
|
2021-04-13 05:26:42 +00:00
|
|
|
|
2021-08-09 09:32:46 +00:00
|
|
|
INIT_LIST_HEAD(&file_head);
|
|
|
|
spin_lock(&cifs_inode->open_file_lock);
|
2021-04-13 05:26:42 +00:00
|
|
|
list_for_each_entry(cfile, &cifs_inode->openFileList, flist) {
|
2021-07-29 07:45:29 +00:00
|
|
|
if (delayed_work_pending(&cfile->deferred)) {
|
2021-08-09 09:32:46 +00:00
|
|
|
if (cancel_delayed_work(&cfile->deferred)) {
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_lock(&cifs_inode->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
cifs_del_deferred_close(cfile);
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_unlock(&cifs_inode->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
|
2021-08-09 09:32:46 +00:00
|
|
|
tmp_list = kmalloc(sizeof(struct file_list), GFP_ATOMIC);
|
|
|
|
if (tmp_list == NULL)
|
2021-09-17 17:29:42 +00:00
|
|
|
break;
|
2021-08-09 09:32:46 +00:00
|
|
|
tmp_list->cfile = cfile;
|
|
|
|
list_add_tail(&tmp_list->list, &file_head);
|
|
|
|
}
|
2021-07-29 07:45:29 +00:00
|
|
|
}
|
2021-04-13 05:26:42 +00:00
|
|
|
}
|
2021-08-09 09:32:46 +00:00
|
|
|
spin_unlock(&cifs_inode->open_file_lock);
|
|
|
|
|
|
|
|
list_for_each_entry_safe(tmp_list, tmp_next_list, &file_head, list) {
|
2023-04-26 14:05:16 +00:00
|
|
|
_cifsFileInfo_put(tmp_list->cfile, false, false);
|
2021-08-09 09:32:46 +00:00
|
|
|
list_del(&tmp_list->list);
|
|
|
|
kfree(tmp_list);
|
|
|
|
}
|
2021-04-13 05:26:42 +00:00
|
|
|
}
|
|
|
|
|
2021-04-19 19:02:03 +00:00
|
|
|
void
|
|
|
|
cifs_close_all_deferred_files(struct cifs_tcon *tcon)
|
|
|
|
{
|
|
|
|
struct cifsFileInfo *cfile;
|
2021-08-09 09:32:46 +00:00
|
|
|
struct file_list *tmp_list, *tmp_next_list;
|
|
|
|
struct list_head file_head;
|
2021-04-19 19:02:03 +00:00
|
|
|
|
2021-08-09 09:32:46 +00:00
|
|
|
INIT_LIST_HEAD(&file_head);
|
2021-04-19 19:02:03 +00:00
|
|
|
spin_lock(&tcon->open_file_lock);
|
2022-07-22 17:02:59 +00:00
|
|
|
list_for_each_entry(cfile, &tcon->openFileList, tlist) {
|
2021-05-20 16:45:01 +00:00
|
|
|
if (delayed_work_pending(&cfile->deferred)) {
|
2021-08-09 09:32:46 +00:00
|
|
|
if (cancel_delayed_work(&cfile->deferred)) {
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_lock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
cifs_del_deferred_close(cfile);
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_unlock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
|
2021-08-09 09:32:46 +00:00
|
|
|
tmp_list = kmalloc(sizeof(struct file_list), GFP_ATOMIC);
|
|
|
|
if (tmp_list == NULL)
|
2021-09-17 17:29:42 +00:00
|
|
|
break;
|
2021-08-09 09:32:46 +00:00
|
|
|
tmp_list->cfile = cfile;
|
|
|
|
list_add_tail(&tmp_list->list, &file_head);
|
|
|
|
}
|
2021-05-20 16:45:01 +00:00
|
|
|
}
|
2021-04-19 19:02:03 +00:00
|
|
|
}
|
|
|
|
spin_unlock(&tcon->open_file_lock);
|
2021-08-09 09:32:46 +00:00
|
|
|
|
|
|
|
list_for_each_entry_safe(tmp_list, tmp_next_list, &file_head, list) {
|
|
|
|
_cifsFileInfo_put(tmp_list->cfile, true, false);
|
|
|
|
list_del(&tmp_list->list);
|
|
|
|
kfree(tmp_list);
|
|
|
|
}
|
2021-04-19 19:02:03 +00:00
|
|
|
}
|
2021-09-17 18:14:26 +00:00
|
|
|
void
|
|
|
|
cifs_close_deferred_file_under_dentry(struct cifs_tcon *tcon, const char *path)
|
|
|
|
{
|
|
|
|
struct cifsFileInfo *cfile;
|
|
|
|
struct file_list *tmp_list, *tmp_next_list;
|
|
|
|
struct list_head file_head;
|
|
|
|
void *page;
|
|
|
|
const char *full_path;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&file_head);
|
|
|
|
page = alloc_dentry_path();
|
|
|
|
spin_lock(&tcon->open_file_lock);
|
2022-07-22 17:02:59 +00:00
|
|
|
list_for_each_entry(cfile, &tcon->openFileList, tlist) {
|
2021-09-17 18:14:26 +00:00
|
|
|
full_path = build_path_from_dentry(cfile->dentry, page);
|
|
|
|
if (strstr(full_path, path)) {
|
|
|
|
if (delayed_work_pending(&cfile->deferred)) {
|
|
|
|
if (cancel_delayed_work(&cfile->deferred)) {
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_lock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
cifs_del_deferred_close(cfile);
|
2023-04-20 13:54:33 +00:00
|
|
|
spin_unlock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
|
2022-08-18 13:50:44 +00:00
|
|
|
|
2021-09-17 18:14:26 +00:00
|
|
|
tmp_list = kmalloc(sizeof(struct file_list), GFP_ATOMIC);
|
|
|
|
if (tmp_list == NULL)
|
|
|
|
break;
|
|
|
|
tmp_list->cfile = cfile;
|
|
|
|
list_add_tail(&tmp_list->list, &file_head);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&tcon->open_file_lock);
|
|
|
|
|
|
|
|
list_for_each_entry_safe(tmp_list, tmp_next_list, &file_head, list) {
|
|
|
|
_cifsFileInfo_put(tmp_list->cfile, true, false);
|
|
|
|
list_del(&tmp_list->list);
|
|
|
|
kfree(tmp_list);
|
|
|
|
}
|
|
|
|
free_dentry_path(page);
|
|
|
|
}
|
2021-04-19 19:02:03 +00:00
|
|
|
|
2022-09-06 03:24:35 +00:00
|
|
|
/* parses DFS referral V3 structure
|
2017-02-13 15:03:47 +00:00
|
|
|
* caller is responsible for freeing target_nodes
|
|
|
|
* returns:
|
|
|
|
* - on success - 0
|
|
|
|
* - on failure - errno
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
parse_dfs_referrals(struct get_dfs_referral_rsp *rsp, u32 rsp_size,
|
|
|
|
unsigned int *num_of_nodes,
|
|
|
|
struct dfs_info3_param **target_nodes,
|
|
|
|
const struct nls_table *nls_codepage, int remap,
|
|
|
|
const char *searchName, bool is_unicode)
|
|
|
|
{
|
|
|
|
int i, rc = 0;
|
|
|
|
char *data_end;
|
|
|
|
struct dfs_referral_level_3 *ref;
|
|
|
|
|
|
|
|
*num_of_nodes = le16_to_cpu(rsp->NumberOfReferrals);
|
|
|
|
|
|
|
|
if (*num_of_nodes < 1) {
|
|
|
|
cifs_dbg(VFS, "num_referrals: must be at least > 0, but we get num_referrals = %d\n",
|
|
|
|
*num_of_nodes);
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
ref = (struct dfs_referral_level_3 *) &(rsp->referrals);
|
|
|
|
if (ref->VersionNumber != cpu_to_le16(3)) {
|
|
|
|
cifs_dbg(VFS, "Referrals of V%d version are not supported, should be V3\n",
|
|
|
|
le16_to_cpu(ref->VersionNumber));
|
|
|
|
rc = -EINVAL;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* get the upper boundary of the resp buffer */
|
|
|
|
data_end = (char *)rsp + rsp_size;
|
|
|
|
|
|
|
|
cifs_dbg(FYI, "num_referrals: %d dfs flags: 0x%x ...\n",
|
|
|
|
*num_of_nodes, le32_to_cpu(rsp->DFSFlags));
|
|
|
|
|
|
|
|
*target_nodes = kcalloc(*num_of_nodes, sizeof(struct dfs_info3_param),
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (*target_nodes == NULL) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* collect necessary data from referrals */
|
|
|
|
for (i = 0; i < *num_of_nodes; i++) {
|
|
|
|
char *temp;
|
|
|
|
int max_len;
|
|
|
|
struct dfs_info3_param *node = (*target_nodes)+i;
|
|
|
|
|
|
|
|
node->flags = le32_to_cpu(rsp->DFSFlags);
|
|
|
|
if (is_unicode) {
|
|
|
|
__le16 *tmp = kmalloc(strlen(searchName)*2 + 2,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (tmp == NULL) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
cifsConvertToUTF16((__le16 *) tmp, searchName,
|
|
|
|
PATH_MAX, nls_codepage, remap);
|
|
|
|
node->path_consumed = cifs_utf16_bytes(tmp,
|
|
|
|
le16_to_cpu(rsp->PathConsumed),
|
|
|
|
nls_codepage);
|
|
|
|
kfree(tmp);
|
|
|
|
} else
|
|
|
|
node->path_consumed = le16_to_cpu(rsp->PathConsumed);
|
|
|
|
|
|
|
|
node->server_type = le16_to_cpu(ref->ServerType);
|
|
|
|
node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags);
|
|
|
|
|
|
|
|
/* copy DfsPath */
|
|
|
|
temp = (char *)ref + le16_to_cpu(ref->DfsPathOffset);
|
|
|
|
max_len = data_end - temp;
|
|
|
|
node->path_name = cifs_strndup_from_utf16(temp, max_len,
|
|
|
|
is_unicode, nls_codepage);
|
|
|
|
if (!node->path_name) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* copy link target UNC */
|
|
|
|
temp = (char *)ref + le16_to_cpu(ref->NetworkAddressOffset);
|
|
|
|
max_len = data_end - temp;
|
|
|
|
node->node_name = cifs_strndup_from_utf16(temp, max_len,
|
|
|
|
is_unicode, nls_codepage);
|
|
|
|
if (!node->node_name) {
|
|
|
|
rc = -ENOMEM;
|
|
|
|
goto parse_DFS_referrals_exit;
|
|
|
|
}
|
|
|
|
|
2018-11-14 17:38:51 +00:00
|
|
|
node->ttl = le32_to_cpu(ref->TimeToLive);
|
|
|
|
|
2017-02-13 15:03:47 +00:00
|
|
|
ref++;
|
|
|
|
}
|
|
|
|
|
|
|
|
parse_DFS_referrals_exit:
|
|
|
|
if (rc) {
|
|
|
|
free_dfs_info_array(*target_nodes, *num_of_nodes);
|
|
|
|
*target_nodes = NULL;
|
|
|
|
*num_of_nodes = 0;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
2017-04-25 18:52:29 +00:00
|
|
|
|
|
|
|
struct cifs_aio_ctx *
|
|
|
|
cifs_aio_ctx_alloc(void)
|
|
|
|
{
|
|
|
|
struct cifs_aio_ctx *ctx;
|
|
|
|
|
2019-04-10 19:37:47 +00:00
|
|
|
/*
|
|
|
|
* Must use kzalloc to initialize ctx->bv to NULL and ctx->direct_io
|
|
|
|
* to false so that we know when we have to unreference pages within
|
|
|
|
* cifs_aio_ctx_release()
|
|
|
|
*/
|
2017-04-25 18:52:29 +00:00
|
|
|
ctx = kzalloc(sizeof(struct cifs_aio_ctx), GFP_KERNEL);
|
|
|
|
if (!ctx)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&ctx->list);
|
|
|
|
mutex_init(&ctx->aio_mutex);
|
|
|
|
init_completion(&ctx->done);
|
|
|
|
kref_init(&ctx->refcount);
|
|
|
|
return ctx;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
cifs_aio_ctx_release(struct kref *refcount)
|
|
|
|
{
|
|
|
|
struct cifs_aio_ctx *ctx = container_of(refcount,
|
|
|
|
struct cifs_aio_ctx, refcount);
|
|
|
|
|
|
|
|
cifsFileInfo_put(ctx->cfile);
|
2019-04-10 19:37:47 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* ctx->bv is only set if setup_aio_ctx_iter() was call successfuly
|
cifs: Change the I/O paths to use an iterator rather than a page list
Currently, the cifs I/O paths hand lists of pages from the VM interface
routines at the top all the way through the intervening layers to the
socket interface at the bottom.
This is a problem, however, for interfacing with netfslib which passes an
iterator through to the ->issue_read() method (and will pass an iterator
through to the ->issue_write() method in future). Netfslib takes over
bounce buffering for direct I/O, async I/O and encrypted content, so cifs
doesn't need to do that. Netfslib also converts IOVEC-type iterators into
BVEC-type iterators if necessary.
Further, cifs needs foliating - and folios may come in a variety of sizes,
so a page list pointing to an array of heterogeneous pages may cause
problems in places such as where crypto is done.
Change the cifs I/O paths to hand iov_iter iterators all the way through
instead.
Notes:
(1) Some old routines are #if'd out to be removed in a follow up patch so
as to avoid confusing diff, thereby making the diff output easier to
follow. I've removed functions that don't overlap with anything
added.
(2) struct smb_rqst loses rq_pages, rq_offset, rq_npages, rq_pagesz and
rq_tailsz which describe the pages forming the buffer; instead there's
an rq_iter describing the source buffer and an rq_buffer which is used
to hold the buffer for encryption.
(3) struct cifs_readdata and cifs_writedata are similarly modified to
smb_rqst. The ->read_into_pages() and ->copy_into_pages() are then
replaced with passing the iterator directly to the socket.
The iterators are stored in these structs so that they are persistent
and don't get deallocated when the function returns (unlike if they
were stack variables).
(4) Buffered writeback is overhauled, borrowing the code from the afs
filesystem to gather up contiguous runs of folios. The XARRAY-type
iterator is then used to refer directly to the pagecache and can be
passed to the socket to transmit data directly from there.
This includes:
cifs_extend_writeback()
cifs_write_back_from_locked_folio()
cifs_writepages_region()
cifs_writepages()
(5) Pages are converted to folios.
(6) Direct I/O uses netfs_extract_user_iter() to create a BVEC-type
iterator from an IOBUF/UBUF-type source iterator.
(7) smb2_get_aead_req() uses netfs_extract_iter_to_sg() to extract page
fragments from the iterator into the scatterlists that the crypto
layer prefers.
(8) smb2_init_transform_rq() attached pages to smb_rqst::rq_buffer, an
xarray, to use as a bounce buffer for encryption. An XARRAY-type
iterator can then be used to pass the bounce buffer to lower layers.
Signed-off-by: David Howells <dhowells@redhat.com>
cc: Steve French <sfrench@samba.org>
cc: Shyam Prasad N <nspmangalore@gmail.com>
cc: Rohith Surabattula <rohiths.msft@gmail.com>
cc: Paulo Alcantara <pc@cjr.nz>
cc: Jeff Layton <jlayton@kernel.org>
cc: linux-cifs@vger.kernel.org
Link: https://lore.kernel.org/r/164311907995.2806745.400147335497304099.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/164928620163.457102.11602306234438271112.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165211420279.3154751.15923591172438186144.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165348880385.2106726.3220789453472800240.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165364827111.3334034.934805882842932881.stgit@warthog.procyon.org.uk/ # v3
Link: https://lore.kernel.org/r/166126396180.708021.271013668175370826.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/166697259595.61150.5982032408321852414.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/166732031756.3186319.12528413619888902872.stgit@warthog.procyon.org.uk/ # rfc
Signed-off-by: Steve French <stfrench@microsoft.com>
2022-01-24 21:13:24 +00:00
|
|
|
* which means that iov_iter_extract_pages() was a success and thus
|
|
|
|
* that we may have references or pins on pages that we need to
|
|
|
|
* release.
|
2019-04-10 19:37:47 +00:00
|
|
|
*/
|
|
|
|
if (ctx->bv) {
|
cifs: Change the I/O paths to use an iterator rather than a page list
Currently, the cifs I/O paths hand lists of pages from the VM interface
routines at the top all the way through the intervening layers to the
socket interface at the bottom.
This is a problem, however, for interfacing with netfslib which passes an
iterator through to the ->issue_read() method (and will pass an iterator
through to the ->issue_write() method in future). Netfslib takes over
bounce buffering for direct I/O, async I/O and encrypted content, so cifs
doesn't need to do that. Netfslib also converts IOVEC-type iterators into
BVEC-type iterators if necessary.
Further, cifs needs foliating - and folios may come in a variety of sizes,
so a page list pointing to an array of heterogeneous pages may cause
problems in places such as where crypto is done.
Change the cifs I/O paths to hand iov_iter iterators all the way through
instead.
Notes:
(1) Some old routines are #if'd out to be removed in a follow up patch so
as to avoid confusing diff, thereby making the diff output easier to
follow. I've removed functions that don't overlap with anything
added.
(2) struct smb_rqst loses rq_pages, rq_offset, rq_npages, rq_pagesz and
rq_tailsz which describe the pages forming the buffer; instead there's
an rq_iter describing the source buffer and an rq_buffer which is used
to hold the buffer for encryption.
(3) struct cifs_readdata and cifs_writedata are similarly modified to
smb_rqst. The ->read_into_pages() and ->copy_into_pages() are then
replaced with passing the iterator directly to the socket.
The iterators are stored in these structs so that they are persistent
and don't get deallocated when the function returns (unlike if they
were stack variables).
(4) Buffered writeback is overhauled, borrowing the code from the afs
filesystem to gather up contiguous runs of folios. The XARRAY-type
iterator is then used to refer directly to the pagecache and can be
passed to the socket to transmit data directly from there.
This includes:
cifs_extend_writeback()
cifs_write_back_from_locked_folio()
cifs_writepages_region()
cifs_writepages()
(5) Pages are converted to folios.
(6) Direct I/O uses netfs_extract_user_iter() to create a BVEC-type
iterator from an IOBUF/UBUF-type source iterator.
(7) smb2_get_aead_req() uses netfs_extract_iter_to_sg() to extract page
fragments from the iterator into the scatterlists that the crypto
layer prefers.
(8) smb2_init_transform_rq() attached pages to smb_rqst::rq_buffer, an
xarray, to use as a bounce buffer for encryption. An XARRAY-type
iterator can then be used to pass the bounce buffer to lower layers.
Signed-off-by: David Howells <dhowells@redhat.com>
cc: Steve French <sfrench@samba.org>
cc: Shyam Prasad N <nspmangalore@gmail.com>
cc: Rohith Surabattula <rohiths.msft@gmail.com>
cc: Paulo Alcantara <pc@cjr.nz>
cc: Jeff Layton <jlayton@kernel.org>
cc: linux-cifs@vger.kernel.org
Link: https://lore.kernel.org/r/164311907995.2806745.400147335497304099.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/164928620163.457102.11602306234438271112.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165211420279.3154751.15923591172438186144.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165348880385.2106726.3220789453472800240.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165364827111.3334034.934805882842932881.stgit@warthog.procyon.org.uk/ # v3
Link: https://lore.kernel.org/r/166126396180.708021.271013668175370826.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/166697259595.61150.5982032408321852414.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/166732031756.3186319.12528413619888902872.stgit@warthog.procyon.org.uk/ # rfc
Signed-off-by: Steve French <stfrench@microsoft.com>
2022-01-24 21:13:24 +00:00
|
|
|
if (ctx->should_dirty || ctx->bv_need_unpin) {
|
|
|
|
unsigned int i;
|
2019-04-10 19:37:47 +00:00
|
|
|
|
cifs: Change the I/O paths to use an iterator rather than a page list
Currently, the cifs I/O paths hand lists of pages from the VM interface
routines at the top all the way through the intervening layers to the
socket interface at the bottom.
This is a problem, however, for interfacing with netfslib which passes an
iterator through to the ->issue_read() method (and will pass an iterator
through to the ->issue_write() method in future). Netfslib takes over
bounce buffering for direct I/O, async I/O and encrypted content, so cifs
doesn't need to do that. Netfslib also converts IOVEC-type iterators into
BVEC-type iterators if necessary.
Further, cifs needs foliating - and folios may come in a variety of sizes,
so a page list pointing to an array of heterogeneous pages may cause
problems in places such as where crypto is done.
Change the cifs I/O paths to hand iov_iter iterators all the way through
instead.
Notes:
(1) Some old routines are #if'd out to be removed in a follow up patch so
as to avoid confusing diff, thereby making the diff output easier to
follow. I've removed functions that don't overlap with anything
added.
(2) struct smb_rqst loses rq_pages, rq_offset, rq_npages, rq_pagesz and
rq_tailsz which describe the pages forming the buffer; instead there's
an rq_iter describing the source buffer and an rq_buffer which is used
to hold the buffer for encryption.
(3) struct cifs_readdata and cifs_writedata are similarly modified to
smb_rqst. The ->read_into_pages() and ->copy_into_pages() are then
replaced with passing the iterator directly to the socket.
The iterators are stored in these structs so that they are persistent
and don't get deallocated when the function returns (unlike if they
were stack variables).
(4) Buffered writeback is overhauled, borrowing the code from the afs
filesystem to gather up contiguous runs of folios. The XARRAY-type
iterator is then used to refer directly to the pagecache and can be
passed to the socket to transmit data directly from there.
This includes:
cifs_extend_writeback()
cifs_write_back_from_locked_folio()
cifs_writepages_region()
cifs_writepages()
(5) Pages are converted to folios.
(6) Direct I/O uses netfs_extract_user_iter() to create a BVEC-type
iterator from an IOBUF/UBUF-type source iterator.
(7) smb2_get_aead_req() uses netfs_extract_iter_to_sg() to extract page
fragments from the iterator into the scatterlists that the crypto
layer prefers.
(8) smb2_init_transform_rq() attached pages to smb_rqst::rq_buffer, an
xarray, to use as a bounce buffer for encryption. An XARRAY-type
iterator can then be used to pass the bounce buffer to lower layers.
Signed-off-by: David Howells <dhowells@redhat.com>
cc: Steve French <sfrench@samba.org>
cc: Shyam Prasad N <nspmangalore@gmail.com>
cc: Rohith Surabattula <rohiths.msft@gmail.com>
cc: Paulo Alcantara <pc@cjr.nz>
cc: Jeff Layton <jlayton@kernel.org>
cc: linux-cifs@vger.kernel.org
Link: https://lore.kernel.org/r/164311907995.2806745.400147335497304099.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/164928620163.457102.11602306234438271112.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165211420279.3154751.15923591172438186144.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165348880385.2106726.3220789453472800240.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165364827111.3334034.934805882842932881.stgit@warthog.procyon.org.uk/ # v3
Link: https://lore.kernel.org/r/166126396180.708021.271013668175370826.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/166697259595.61150.5982032408321852414.stgit@warthog.procyon.org.uk/ # rfc
Link: https://lore.kernel.org/r/166732031756.3186319.12528413619888902872.stgit@warthog.procyon.org.uk/ # rfc
Signed-off-by: Steve French <stfrench@microsoft.com>
2022-01-24 21:13:24 +00:00
|
|
|
for (i = 0; i < ctx->nr_pinned_pages; i++) {
|
|
|
|
struct page *page = ctx->bv[i].bv_page;
|
|
|
|
|
|
|
|
if (ctx->should_dirty)
|
|
|
|
set_page_dirty(page);
|
|
|
|
if (ctx->bv_need_unpin)
|
|
|
|
unpin_user_page(page);
|
|
|
|
}
|
2019-04-10 19:37:47 +00:00
|
|
|
}
|
|
|
|
kvfree(ctx->bv);
|
|
|
|
}
|
|
|
|
|
2017-04-25 18:52:29 +00:00
|
|
|
kfree(ctx);
|
|
|
|
}
|
|
|
|
|
2018-02-16 18:19:27 +00:00
|
|
|
/**
|
|
|
|
* cifs_alloc_hash - allocate hash and hash context together
|
2021-09-20 12:14:15 +00:00
|
|
|
* @name: The name of the crypto hash algo
|
2022-09-29 20:36:50 +00:00
|
|
|
* @sdesc: SHASH descriptor where to put the pointer to the hash TFM
|
2018-02-16 18:19:27 +00:00
|
|
|
*
|
|
|
|
* The caller has to make sure @sdesc is initialized to either NULL or
|
2022-09-29 20:36:50 +00:00
|
|
|
* a valid context. It can be freed via cifs_free_hash().
|
2018-02-16 18:19:27 +00:00
|
|
|
*/
|
|
|
|
int
|
2022-09-29 20:36:50 +00:00
|
|
|
cifs_alloc_hash(const char *name, struct shash_desc **sdesc)
|
2018-02-16 18:19:27 +00:00
|
|
|
{
|
|
|
|
int rc = 0;
|
2022-09-29 20:36:50 +00:00
|
|
|
struct crypto_shash *alg = NULL;
|
2018-02-16 18:19:27 +00:00
|
|
|
|
2022-09-29 20:36:50 +00:00
|
|
|
if (*sdesc)
|
2018-02-16 18:19:27 +00:00
|
|
|
return 0;
|
|
|
|
|
2022-09-29 20:36:50 +00:00
|
|
|
alg = crypto_alloc_shash(name, 0, 0);
|
|
|
|
if (IS_ERR(alg)) {
|
|
|
|
cifs_dbg(VFS, "Could not allocate shash TFM '%s'\n", name);
|
|
|
|
rc = PTR_ERR(alg);
|
2018-02-16 18:19:27 +00:00
|
|
|
*sdesc = NULL;
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2022-09-29 20:36:50 +00:00
|
|
|
*sdesc = kmalloc(sizeof(struct shash_desc) + crypto_shash_descsize(alg), GFP_KERNEL);
|
2018-02-16 18:19:27 +00:00
|
|
|
if (*sdesc == NULL) {
|
2022-09-29 20:36:50 +00:00
|
|
|
cifs_dbg(VFS, "no memory left to allocate shash TFM '%s'\n", name);
|
|
|
|
crypto_free_shash(alg);
|
2018-02-16 18:19:27 +00:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2022-09-29 20:36:50 +00:00
|
|
|
(*sdesc)->tfm = alg;
|
2018-02-16 18:19:27 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* cifs_free_hash - free hash and hash context together
|
2022-09-29 20:36:50 +00:00
|
|
|
* @sdesc: Where to find the pointer to the hash TFM
|
2018-02-16 18:19:27 +00:00
|
|
|
*
|
2022-09-29 20:36:50 +00:00
|
|
|
* Freeing a NULL descriptor is safe.
|
2018-02-16 18:19:27 +00:00
|
|
|
*/
|
|
|
|
void
|
2022-09-29 20:36:50 +00:00
|
|
|
cifs_free_hash(struct shash_desc **sdesc)
|
2018-02-16 18:19:27 +00:00
|
|
|
{
|
2022-09-29 20:36:50 +00:00
|
|
|
if (unlikely(!sdesc) || !*sdesc)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if ((*sdesc)->tfm) {
|
|
|
|
crypto_free_shash((*sdesc)->tfm);
|
|
|
|
(*sdesc)->tfm = NULL;
|
|
|
|
}
|
|
|
|
|
2022-09-20 18:10:35 +00:00
|
|
|
kfree_sensitive(*sdesc);
|
2018-02-16 18:19:27 +00:00
|
|
|
*sdesc = NULL;
|
|
|
|
}
|
2018-05-30 19:47:58 +00:00
|
|
|
|
2018-11-14 19:20:31 +00:00
|
|
|
void extract_unc_hostname(const char *unc, const char **h, size_t *len)
|
|
|
|
{
|
|
|
|
const char *end;
|
|
|
|
|
|
|
|
/* skip initial slashes */
|
|
|
|
while (*unc && (*unc == '\\' || *unc == '/'))
|
|
|
|
unc++;
|
|
|
|
|
|
|
|
end = unc;
|
|
|
|
|
|
|
|
while (*end && !(*end == '\\' || *end == '/'))
|
|
|
|
end++;
|
|
|
|
|
|
|
|
*h = unc;
|
|
|
|
*len = end - unc;
|
|
|
|
}
|
2019-08-26 23:30:14 +00:00
|
|
|
|
|
|
|
/**
|
|
|
|
* copy_path_name - copy src path to dst, possibly truncating
|
2021-09-20 12:14:15 +00:00
|
|
|
* @dst: The destination buffer
|
|
|
|
* @src: The source name
|
2019-08-26 23:30:14 +00:00
|
|
|
*
|
|
|
|
* returns number of bytes written (including trailing nul)
|
|
|
|
*/
|
|
|
|
int copy_path_name(char *dst, const char *src)
|
|
|
|
{
|
|
|
|
int name_len;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* PATH_MAX includes nul, so if strlen(src) >= PATH_MAX it
|
|
|
|
* will truncate and strlen(dst) will be PATH_MAX-1
|
|
|
|
*/
|
|
|
|
name_len = strscpy(dst, src, PATH_MAX);
|
|
|
|
if (WARN_ON_ONCE(name_len < 0))
|
|
|
|
name_len = PATH_MAX-1;
|
|
|
|
|
|
|
|
/* we count the trailing nul */
|
|
|
|
name_len++;
|
|
|
|
return name_len;
|
|
|
|
}
|
2020-02-20 22:49:34 +00:00
|
|
|
|
|
|
|
struct super_cb_data {
|
2020-04-20 22:43:04 +00:00
|
|
|
void *data;
|
2020-02-20 22:49:34 +00:00
|
|
|
struct super_block *sb;
|
|
|
|
};
|
|
|
|
|
2023-06-26 19:04:17 +00:00
|
|
|
static void tcon_super_cb(struct super_block *sb, void *arg)
|
2020-02-20 22:49:34 +00:00
|
|
|
{
|
2020-04-20 22:43:04 +00:00
|
|
|
struct super_cb_data *sd = arg;
|
2020-02-20 22:49:34 +00:00
|
|
|
struct cifs_sb_info *cifs_sb;
|
2023-06-26 19:04:17 +00:00
|
|
|
struct cifs_tcon *t1 = sd->data, *t2;
|
2020-02-20 22:49:34 +00:00
|
|
|
|
2020-04-20 22:43:04 +00:00
|
|
|
if (sd->sb)
|
2020-02-20 22:49:34 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
cifs_sb = CIFS_SB(sb);
|
2023-06-26 19:04:17 +00:00
|
|
|
t2 = cifs_sb_master_tcon(cifs_sb);
|
|
|
|
|
|
|
|
spin_lock(&t2->tc_lock);
|
|
|
|
if (t1->ses == t2->ses &&
|
|
|
|
t1->ses->server == t2->ses->server &&
|
|
|
|
t2->origin_fullpath &&
|
|
|
|
dfs_src_pathname_equal(t2->origin_fullpath, t1->origin_fullpath))
|
2020-04-20 22:43:04 +00:00
|
|
|
sd->sb = sb;
|
2023-06-26 19:04:17 +00:00
|
|
|
spin_unlock(&t2->tc_lock);
|
2020-02-20 22:49:34 +00:00
|
|
|
}
|
|
|
|
|
2020-04-20 22:43:04 +00:00
|
|
|
static struct super_block *__cifs_get_super(void (*f)(struct super_block *, void *),
|
|
|
|
void *data)
|
2020-02-20 22:49:34 +00:00
|
|
|
{
|
2020-04-20 22:43:04 +00:00
|
|
|
struct super_cb_data sd = {
|
|
|
|
.data = data,
|
2020-02-20 22:49:34 +00:00
|
|
|
.sb = NULL,
|
|
|
|
};
|
2022-06-05 22:54:26 +00:00
|
|
|
struct file_system_type **fs_type = (struct file_system_type *[]) {
|
|
|
|
&cifs_fs_type, &smb3_fs_type, NULL,
|
|
|
|
};
|
2020-02-20 22:49:34 +00:00
|
|
|
|
2022-06-05 22:54:26 +00:00
|
|
|
for (; *fs_type; fs_type++) {
|
|
|
|
iterate_supers_type(*fs_type, f, &sd);
|
|
|
|
if (sd.sb) {
|
|
|
|
/*
|
|
|
|
* Grab an active reference in order to prevent automounts (DFS links)
|
|
|
|
* of expiring and then freeing up our cifs superblock pointer while
|
|
|
|
* we're doing failover.
|
|
|
|
*/
|
|
|
|
cifs_sb_active(sd.sb);
|
|
|
|
return sd.sb;
|
|
|
|
}
|
|
|
|
}
|
2023-06-26 19:04:17 +00:00
|
|
|
pr_warn_once("%s: could not find dfs superblock\n", __func__);
|
2022-06-05 22:54:26 +00:00
|
|
|
return ERR_PTR(-EINVAL);
|
2020-02-20 22:49:34 +00:00
|
|
|
}
|
|
|
|
|
2020-04-20 22:43:04 +00:00
|
|
|
static void __cifs_put_super(struct super_block *sb)
|
2020-02-20 22:49:34 +00:00
|
|
|
{
|
|
|
|
if (!IS_ERR_OR_NULL(sb))
|
|
|
|
cifs_sb_deactive(sb);
|
|
|
|
}
|
|
|
|
|
2023-06-26 19:04:17 +00:00
|
|
|
struct super_block *cifs_get_dfs_tcon_super(struct cifs_tcon *tcon)
|
2020-04-20 22:43:04 +00:00
|
|
|
{
|
2023-06-26 19:04:17 +00:00
|
|
|
spin_lock(&tcon->tc_lock);
|
|
|
|
if (!tcon->origin_fullpath) {
|
|
|
|
spin_unlock(&tcon->tc_lock);
|
|
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
}
|
|
|
|
spin_unlock(&tcon->tc_lock);
|
|
|
|
return __cifs_get_super(tcon_super_cb, tcon);
|
2020-04-20 22:43:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
void cifs_put_tcp_super(struct super_block *sb)
|
|
|
|
{
|
|
|
|
__cifs_put_super(sb);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_CIFS_DFS_UPCALL
|
2020-05-19 18:38:28 +00:00
|
|
|
int match_target_ip(struct TCP_Server_Info *server,
|
|
|
|
const char *share, size_t share_len,
|
|
|
|
bool *result)
|
|
|
|
{
|
|
|
|
int rc;
|
2022-10-04 21:41:36 +00:00
|
|
|
char *target;
|
|
|
|
struct sockaddr_storage ss;
|
2020-05-19 18:38:28 +00:00
|
|
|
|
|
|
|
*result = false;
|
|
|
|
|
|
|
|
target = kzalloc(share_len + 3, GFP_KERNEL);
|
2022-10-04 21:41:36 +00:00
|
|
|
if (!target)
|
|
|
|
return -ENOMEM;
|
2020-05-19 18:38:28 +00:00
|
|
|
|
|
|
|
scnprintf(target, share_len + 3, "\\\\%.*s", (int)share_len, share);
|
|
|
|
|
|
|
|
cifs_dbg(FYI, "%s: target name: %s\n", __func__, target + 2);
|
|
|
|
|
2022-10-04 21:41:36 +00:00
|
|
|
rc = dns_resolve_server_name_to_ip(target, (struct sockaddr *)&ss, NULL);
|
|
|
|
kfree(target);
|
2020-05-19 18:38:28 +00:00
|
|
|
|
2022-10-04 21:41:36 +00:00
|
|
|
if (rc < 0)
|
|
|
|
return rc;
|
2020-05-19 18:38:28 +00:00
|
|
|
|
2022-12-29 21:43:46 +00:00
|
|
|
spin_lock(&server->srv_lock);
|
2022-10-04 21:41:36 +00:00
|
|
|
*result = cifs_match_ipaddr((struct sockaddr *)&server->dstaddr, (struct sockaddr *)&ss);
|
2022-12-29 21:43:46 +00:00
|
|
|
spin_unlock(&server->srv_lock);
|
2020-05-19 18:38:28 +00:00
|
|
|
cifs_dbg(FYI, "%s: ip addresses match: %u\n", __func__, *result);
|
2022-10-04 21:41:36 +00:00
|
|
|
return 0;
|
2020-05-19 18:38:28 +00:00
|
|
|
}
|
|
|
|
|
2021-11-03 16:53:29 +00:00
|
|
|
int cifs_update_super_prepath(struct cifs_sb_info *cifs_sb, char *prefix)
|
2020-04-20 22:43:04 +00:00
|
|
|
{
|
2023-06-28 00:24:47 +00:00
|
|
|
int rc;
|
|
|
|
|
2020-02-20 22:49:34 +00:00
|
|
|
kfree(cifs_sb->prepath);
|
2023-06-28 00:24:47 +00:00
|
|
|
cifs_sb->prepath = NULL;
|
2020-02-20 22:49:34 +00:00
|
|
|
|
2020-07-21 12:36:42 +00:00
|
|
|
if (prefix && *prefix) {
|
2023-04-05 13:16:48 +00:00
|
|
|
cifs_sb->prepath = cifs_sanitize_prepath(prefix, GFP_ATOMIC);
|
2023-06-28 00:24:47 +00:00
|
|
|
if (IS_ERR(cifs_sb->prepath)) {
|
|
|
|
rc = PTR_ERR(cifs_sb->prepath);
|
|
|
|
cifs_sb->prepath = NULL;
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
if (cifs_sb->prepath)
|
|
|
|
convert_delimiter(cifs_sb->prepath, CIFS_DIR_SEP(cifs_sb));
|
|
|
|
}
|
2020-02-20 22:49:34 +00:00
|
|
|
|
|
|
|
cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH;
|
2021-11-03 16:53:29 +00:00
|
|
|
return 0;
|
2020-02-20 22:49:34 +00:00
|
|
|
}
|
2023-02-28 22:01:54 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Handle weird Windows SMB server behaviour. It responds with
|
|
|
|
* STATUS_OBJECT_NAME_INVALID code to SMB2 QUERY_INFO request for
|
|
|
|
* "\<server>\<dfsname>\<linkpath>" DFS reference, where <dfsname> contains
|
|
|
|
* non-ASCII unicode symbols.
|
|
|
|
*/
|
|
|
|
int cifs_inval_name_dfs_link_error(const unsigned int xid,
|
|
|
|
struct cifs_tcon *tcon,
|
|
|
|
struct cifs_sb_info *cifs_sb,
|
|
|
|
const char *full_path,
|
|
|
|
bool *islink)
|
|
|
|
{
|
|
|
|
struct cifs_ses *ses = tcon->ses;
|
|
|
|
size_t len;
|
|
|
|
char *path;
|
|
|
|
char *ref_path;
|
|
|
|
|
|
|
|
*islink = false;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Fast path - skip check when @full_path doesn't have a prefix path to
|
|
|
|
* look up or tcon is not DFS.
|
|
|
|
*/
|
|
|
|
if (strlen(full_path) < 2 || !cifs_sb ||
|
|
|
|
(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_DFS) ||
|
2023-06-26 19:04:17 +00:00
|
|
|
!is_tcon_dfs(tcon))
|
2023-02-28 22:01:54 +00:00
|
|
|
return 0;
|
|
|
|
|
2023-06-26 19:04:17 +00:00
|
|
|
spin_lock(&tcon->tc_lock);
|
|
|
|
if (!tcon->origin_fullpath) {
|
|
|
|
spin_unlock(&tcon->tc_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
spin_unlock(&tcon->tc_lock);
|
|
|
|
|
2023-02-28 22:01:54 +00:00
|
|
|
/*
|
|
|
|
* Slow path - tcon is DFS and @full_path has prefix path, so attempt
|
|
|
|
* to get a referral to figure out whether it is an DFS link.
|
|
|
|
*/
|
|
|
|
len = strnlen(tcon->tree_name, MAX_TREE_SIZE + 1) + strlen(full_path) + 1;
|
|
|
|
path = kmalloc(len, GFP_KERNEL);
|
|
|
|
if (!path)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
scnprintf(path, len, "%s%s", tcon->tree_name, full_path);
|
|
|
|
ref_path = dfs_cache_canonical_path(path + 1, cifs_sb->local_nls,
|
|
|
|
cifs_remap(cifs_sb));
|
|
|
|
kfree(path);
|
|
|
|
|
|
|
|
if (IS_ERR(ref_path)) {
|
|
|
|
if (PTR_ERR(ref_path) != -EINVAL)
|
|
|
|
return PTR_ERR(ref_path);
|
|
|
|
} else {
|
|
|
|
struct dfs_info3_param *refs = NULL;
|
|
|
|
int num_refs = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* XXX: we are not using dfs_cache_find() here because we might
|
2023-06-26 19:04:17 +00:00
|
|
|
* end up filling all the DFS cache and thus potentially
|
2023-02-28 22:01:54 +00:00
|
|
|
* removing cached DFS targets that the client would eventually
|
|
|
|
* need during failover.
|
|
|
|
*/
|
2023-03-14 23:32:56 +00:00
|
|
|
ses = CIFS_DFS_ROOT_SES(ses);
|
2023-02-28 22:01:54 +00:00
|
|
|
if (ses->server->ops->get_dfs_refer &&
|
|
|
|
!ses->server->ops->get_dfs_refer(xid, ses, ref_path, &refs,
|
|
|
|
&num_refs, cifs_sb->local_nls,
|
|
|
|
cifs_remap(cifs_sb)))
|
|
|
|
*islink = refs[0].server_type == DFS_TYPE_LINK;
|
|
|
|
free_dfs_info_array(refs, num_refs);
|
|
|
|
kfree(ref_path);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2021-11-03 16:53:29 +00:00
|
|
|
#endif
|
2023-02-28 22:01:55 +00:00
|
|
|
|
|
|
|
int cifs_wait_for_server_reconnect(struct TCP_Server_Info *server, bool retry)
|
|
|
|
{
|
|
|
|
int timeout = 10;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
spin_lock(&server->srv_lock);
|
|
|
|
if (server->tcpStatus != CifsNeedReconnect) {
|
|
|
|
spin_unlock(&server->srv_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
timeout *= server->nr_targets;
|
|
|
|
spin_unlock(&server->srv_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Give demultiplex thread up to 10 seconds to each target available for
|
|
|
|
* reconnect -- should be greater than cifs socket timeout which is 7
|
|
|
|
* seconds.
|
|
|
|
*
|
|
|
|
* On "soft" mounts we wait once. Hard mounts keep retrying until
|
|
|
|
* process is killed or server comes back on-line.
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
rc = wait_event_interruptible_timeout(server->response_q,
|
|
|
|
(server->tcpStatus != CifsNeedReconnect),
|
|
|
|
timeout * HZ);
|
|
|
|
if (rc < 0) {
|
|
|
|
cifs_dbg(FYI, "%s: aborting reconnect due to received signal\n",
|
|
|
|
__func__);
|
|
|
|
return -ERESTARTSYS;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* are we still trying to reconnect? */
|
|
|
|
spin_lock(&server->srv_lock);
|
|
|
|
if (server->tcpStatus != CifsNeedReconnect) {
|
|
|
|
spin_unlock(&server->srv_lock);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
spin_unlock(&server->srv_lock);
|
|
|
|
} while (retry);
|
|
|
|
|
|
|
|
cifs_dbg(FYI, "%s: gave up waiting on reconnect\n", __func__);
|
|
|
|
return -EHOSTDOWN;
|
|
|
|
}
|