| /******************************************************************************* |
| * This file contains main functions related to the iSCSI Target Core Driver. |
| * |
| * \u00a9 Copyright 2007-2011 RisingTide Systems LLC. |
| * |
| * Licensed to the Linux Foundation under the General Public License (GPL) version 2. |
| * |
| * Author: Nicholas A. Bellinger <nab@linux-iscsi.org> |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation; either version 2 of the License, or |
| * (at your option) any later version. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| ******************************************************************************/ |
| |
| #include <linux/string.h> |
| #include <linux/kthread.h> |
| #include <linux/crypto.h> |
| #include <linux/completion.h> |
| #include <linux/module.h> |
| #include <asm/unaligned.h> |
| #include <scsi/scsi_device.h> |
| #include <scsi/iscsi_proto.h> |
| #include <target/target_core_base.h> |
| #include <target/target_core_tmr.h> |
| #include <target/target_core_transport.h> |
| |
| #include "iscsi_target_core.h" |
| #include "iscsi_target_parameters.h" |
| #include "iscsi_target_seq_pdu_list.h" |
| #include "iscsi_target_tq.h" |
| #include "iscsi_target_configfs.h" |
| #include "iscsi_target_datain_values.h" |
| #include "iscsi_target_erl0.h" |
| #include "iscsi_target_erl1.h" |
| #include "iscsi_target_erl2.h" |
| #include "iscsi_target_login.h" |
| #include "iscsi_target_tmr.h" |
| #include "iscsi_target_tpg.h" |
| #include "iscsi_target_util.h" |
| #include "iscsi_target.h" |
| #include "iscsi_target_device.h" |
| #include "iscsi_target_stat.h" |
| |
| static LIST_HEAD(g_tiqn_list); |
| static LIST_HEAD(g_np_list); |
| static DEFINE_SPINLOCK(tiqn_lock); |
| static DEFINE_SPINLOCK(np_lock); |
| |
| static struct idr tiqn_idr; |
| struct idr sess_idr; |
| struct mutex auth_id_lock; |
| spinlock_t sess_idr_lock; |
| |
| struct iscsit_global *iscsit_global; |
| |
| struct kmem_cache *lio_cmd_cache; |
| struct kmem_cache *lio_qr_cache; |
| struct kmem_cache *lio_dr_cache; |
| struct kmem_cache *lio_ooo_cache; |
| struct kmem_cache *lio_r2t_cache; |
| |
| static int iscsit_handle_immediate_data(struct iscsi_cmd *, |
| unsigned char *buf, u32); |
| static int iscsit_logout_post_handler(struct iscsi_cmd *, struct iscsi_conn *); |
| |
| struct iscsi_tiqn *iscsit_get_tiqn_for_login(unsigned char *buf) |
| { |
| struct iscsi_tiqn *tiqn = NULL; |
| |
| spin_lock(&tiqn_lock); |
| list_for_each_entry(tiqn, &g_tiqn_list, tiqn_list) { |
| if (!strcmp(tiqn->tiqn, buf)) { |
| |
| spin_lock(&tiqn->tiqn_state_lock); |
| if (tiqn->tiqn_state == TIQN_STATE_ACTIVE) { |
| tiqn->tiqn_access_count++; |
| spin_unlock(&tiqn->tiqn_state_lock); |
| spin_unlock(&tiqn_lock); |
| return tiqn; |
| } |
| spin_unlock(&tiqn->tiqn_state_lock); |
| } |
| } |
| spin_unlock(&tiqn_lock); |
| |
| return NULL; |
| } |
| |
| static int iscsit_set_tiqn_shutdown(struct iscsi_tiqn *tiqn) |
| { |
| spin_lock(&tiqn->tiqn_state_lock); |
| if (tiqn->tiqn_state == TIQN_STATE_ACTIVE) { |
| tiqn->tiqn_state = TIQN_STATE_SHUTDOWN; |
| spin_unlock(&tiqn->tiqn_state_lock); |
| return 0; |
| } |
| spin_unlock(&tiqn->tiqn_state_lock); |
| |
| return -1; |
| } |
| |
| void iscsit_put_tiqn_for_login(struct iscsi_tiqn *tiqn) |
| { |
| spin_lock(&tiqn->tiqn_state_lock); |
| tiqn->tiqn_access_count--; |
| spin_unlock(&tiqn->tiqn_state_lock); |
| } |
| |
| /* |
| * Note that IQN formatting is expected to be done in userspace, and |
| * no explict IQN format checks are done here. |
| */ |
| struct iscsi_tiqn *iscsit_add_tiqn(unsigned char *buf) |
| { |
| struct iscsi_tiqn *tiqn = NULL; |
| int ret; |
| |
| if (strlen(buf) >= ISCSI_IQN_LEN) { |
| pr_err("Target IQN exceeds %d bytes\n", |
| ISCSI_IQN_LEN); |
| return ERR_PTR(-EINVAL); |
| } |
| |
| tiqn = kzalloc(sizeof(struct iscsi_tiqn), GFP_KERNEL); |
| if (!tiqn) { |
| pr_err("Unable to allocate struct iscsi_tiqn\n"); |
| return ERR_PTR(-ENOMEM); |
| } |
| |
| sprintf(tiqn->tiqn, "%s", buf); |
| INIT_LIST_HEAD(&tiqn->tiqn_list); |
| INIT_LIST_HEAD(&tiqn->tiqn_tpg_list); |
| spin_lock_init(&tiqn->tiqn_state_lock); |
| spin_lock_init(&tiqn->tiqn_tpg_lock); |
| spin_lock_init(&tiqn->sess_err_stats.lock); |
| spin_lock_init(&tiqn->login_stats.lock); |
| spin_lock_init(&tiqn->logout_stats.lock); |
| |
| if (!idr_pre_get(&tiqn_idr, GFP_KERNEL)) { |
| pr_err("idr_pre_get() for tiqn_idr failed\n"); |
| kfree(tiqn); |
| return ERR_PTR(-ENOMEM); |
| } |
| tiqn->tiqn_state = TIQN_STATE_ACTIVE; |
| |
| spin_lock(&tiqn_lock); |
| ret = idr_get_new(&tiqn_idr, NULL, &tiqn->tiqn_index); |
| if (ret < 0) { |
| pr_err("idr_get_new() failed for tiqn->tiqn_index\n"); |
| spin_unlock(&tiqn_lock); |
| kfree(tiqn); |
| return ERR_PTR(ret); |
| } |
| list_add_tail(&tiqn->tiqn_list, &g_tiqn_list); |
| spin_unlock(&tiqn_lock); |
| |
| pr_debug("CORE[0] - Added iSCSI Target IQN: %s\n", tiqn->tiqn); |
| |
| return tiqn; |
| |
| } |
| |
| static void iscsit_wait_for_tiqn(struct iscsi_tiqn *tiqn) |
| { |
| /* |
| * Wait for accesses to said struct iscsi_tiqn to end. |
| */ |
| spin_lock(&tiqn->tiqn_state_lock); |
| while (tiqn->tiqn_access_count != 0) { |
| spin_unlock(&tiqn->tiqn_state_lock); |
| msleep(10); |
| spin_lock(&tiqn->tiqn_state_lock); |
| } |
| spin_unlock(&tiqn->tiqn_state_lock); |
| } |
| |
| void iscsit_del_tiqn(struct iscsi_tiqn *tiqn) |
| { |
| /* |
| * iscsit_set_tiqn_shutdown sets tiqn->tiqn_state = TIQN_STATE_SHUTDOWN |
| * while holding tiqn->tiqn_state_lock. This means that all subsequent |
| * attempts to access this struct iscsi_tiqn will fail from both transport |
| * fabric and control code paths. |
| */ |
| if (iscsit_set_tiqn_shutdown(tiqn) < 0) { |
| pr_err("iscsit_set_tiqn_shutdown() failed\n"); |
| return; |
| } |
| |
| iscsit_wait_for_tiqn(tiqn); |
| |
| spin_lock(&tiqn_lock); |
| list_del(&tiqn->tiqn_list); |
| idr_remove(&tiqn_idr, tiqn->tiqn_index); |
| spin_unlock(&tiqn_lock); |
| |
| pr_debug("CORE[0] - Deleted iSCSI Target IQN: %s\n", |
| tiqn->tiqn); |
| kfree(tiqn); |
| } |
| |
| int iscsit_access_np(struct iscsi_np *np, struct iscsi_portal_group *tpg) |
| { |
| int ret; |
| /* |
| * Determine if the network portal is accepting storage traffic. |
| */ |
| spin_lock_bh(&np->np_thread_lock); |
| if (np->np_thread_state != ISCSI_NP_THREAD_ACTIVE) { |
| spin_unlock_bh(&np->np_thread_lock); |
| return -1; |
| } |
| if (np->np_login_tpg) { |
| pr_err("np->np_login_tpg() is not NULL!\n"); |
| spin_unlock_bh(&np->np_thread_lock); |
| return -1; |
| } |
| spin_unlock_bh(&np->np_thread_lock); |
| /* |
| * Determine if the portal group is accepting storage traffic. |
| */ |
| spin_lock_bh(&tpg->tpg_state_lock); |
| if (tpg->tpg_state != TPG_STATE_ACTIVE) { |
| spin_unlock_bh(&tpg->tpg_state_lock); |
| return -1; |
| } |
| spin_unlock_bh(&tpg->tpg_state_lock); |
| |
| /* |
| * Here we serialize access across the TIQN+TPG Tuple. |
| */ |
| ret = mutex_lock_interruptible(&tpg->np_login_lock); |
| if ((ret != 0) || signal_pending(current)) |
| return -1; |
| |
| spin_lock_bh(&np->np_thread_lock); |
| np->np_login_tpg = tpg; |
| spin_unlock_bh(&np->np_thread_lock); |
| |
| return 0; |
| } |
| |
| int iscsit_deaccess_np(struct iscsi_np *np, struct iscsi_portal_group *tpg) |
| { |
| struct iscsi_tiqn *tiqn = tpg->tpg_tiqn; |
| |
| spin_lock_bh(&np->np_thread_lock); |
| np->np_login_tpg = NULL; |
| spin_unlock_bh(&np->np_thread_lock); |
| |
| mutex_unlock(&tpg->np_login_lock); |
| |
| if (tiqn) |
| iscsit_put_tiqn_for_login(tiqn); |
| |
| return 0; |
| } |
| |
| static struct iscsi_np *iscsit_get_np( |
| struct __kernel_sockaddr_storage *sockaddr, |
| int network_transport) |
| { |
| struct sockaddr_in *sock_in, *sock_in_e; |
| struct sockaddr_in6 *sock_in6, *sock_in6_e; |
| struct iscsi_np *np; |
| int ip_match = 0; |
| u16 port; |
| |
| spin_lock_bh(&np_lock); |
| list_for_each_entry(np, &g_np_list, np_list) { |
| spin_lock(&np->np_thread_lock); |
| if (np->np_thread_state != ISCSI_NP_THREAD_ACTIVE) { |
| spin_unlock(&np->np_thread_lock); |
| continue; |
| } |
| |
| if (sockaddr->ss_family == AF_INET6) { |
| sock_in6 = (struct sockaddr_in6 *)sockaddr; |
| sock_in6_e = (struct sockaddr_in6 *)&np->np_sockaddr; |
| |
| if (!memcmp((void *)&sock_in6->sin6_addr.in6_u, |
| (void *)&sock_in6_e->sin6_addr.in6_u, |
| sizeof(struct in6_addr))) |
| ip_match = 1; |
| |
| port = ntohs(sock_in6->sin6_port); |
| } else { |
| sock_in = (struct sockaddr_in *)sockaddr; |
| sock_in_e = (struct sockaddr_in *)&np->np_sockaddr; |
| |
| if (sock_in->sin_addr.s_addr == |
| sock_in_e->sin_addr.s_addr) |
| ip_match = 1; |
| |
| port = ntohs(sock_in->sin_port); |
| } |
| |
| if ((ip_match == 1) && (np->np_port == port) && |
| (np->np_network_transport == network_transport)) { |
| /* |
| * Increment the np_exports reference count now to |
| * prevent iscsit_del_np() below from being called |
| * while iscsi_tpg_add_network_portal() is called. |
| */ |
| np->np_exports++; |
| spin_unlock(&np->np_thread_lock); |
| spin_unlock_bh(&np_lock); |
| return np; |
| } |
| spin_unlock(&np->np_thread_lock); |
| } |
| spin_unlock_bh(&np_lock); |
| |
| return NULL; |
| } |
| |
| struct iscsi_np *iscsit_add_np( |
| struct __kernel_sockaddr_storage *sockaddr, |
| char *ip_str, |
| int network_transport) |
| { |
| struct sockaddr_in *sock_in; |
| struct sockaddr_in6 *sock_in6; |
| struct iscsi_np *np; |
| int ret; |
| /* |
| * Locate the existing struct iscsi_np if already active.. |
| */ |
| np = iscsit_get_np(sockaddr, network_transport); |
| if (np) |
| return np; |
| |
| np = kzalloc(sizeof(struct iscsi_np), GFP_KERNEL); |
| if (!np) { |
| pr_err("Unable to allocate memory for struct iscsi_np\n"); |
| return ERR_PTR(-ENOMEM); |
| } |
| |
| np->np_flags |= NPF_IP_NETWORK; |
| if (sockaddr->ss_family == AF_INET6) { |
| sock_in6 = (struct sockaddr_in6 *)sockaddr; |
| snprintf(np->np_ip, IPV6_ADDRESS_SPACE, "%s", ip_str); |
| np->np_port = ntohs(sock_in6->sin6_port); |
| } else { |
| sock_in = (struct sockaddr_in *)sockaddr; |
| sprintf(np->np_ip, "%s", ip_str); |
| np->np_port = ntohs(sock_in->sin_port); |
| } |
| |
| np->np_network_transport = network_transport; |
| spin_lock_init(&np->np_thread_lock); |
| init_completion(&np->np_restart_comp); |
| INIT_LIST_HEAD(&np->np_list); |
| |
| ret = iscsi_target_setup_login_socket(np, sockaddr); |
| if (ret != 0) { |
| kfree(np); |
| return ERR_PTR(ret); |
| } |
| |
| np->np_thread = kthread_run(iscsi_target_login_thread, np, "iscsi_np"); |
| if (IS_ERR(np->np_thread)) { |
| pr_err("Unable to create kthread: iscsi_np\n"); |
| ret = PTR_ERR(np->np_thread); |
| kfree(np); |
| return ERR_PTR(ret); |
| } |
| /* |
| * Increment the np_exports reference count now to prevent |
| * iscsit_del_np() below from being run while a new call to |
| * iscsi_tpg_add_network_portal() for a matching iscsi_np is |
| * active. We don't need to hold np->np_thread_lock at this |
| * point because iscsi_np has not been added to g_np_list yet. |
| */ |
| np->np_exports = 1; |
| |
| spin_lock_bh(&np_lock); |
| list_add_tail(&np->np_list, &g_np_list); |
| spin_unlock_bh(&np_lock); |
| |
| pr_debug("CORE[0] - Added Network Portal: %s:%hu on %s\n", |
| np->np_ip, np->np_port, (np->np_network_transport == ISCSI_TCP) ? |
| "TCP" : "SCTP"); |
| |
| return np; |
| } |
| |
| int iscsit_reset_np_thread( |
| struct iscsi_np *np, |
| struct iscsi_tpg_np *tpg_np, |
| struct iscsi_portal_group *tpg) |
| { |
| spin_lock_bh(&np->np_thread_lock); |
| if (tpg && tpg_np) { |
| /* |
| * The reset operation need only be performed when the |
| * passed struct iscsi_portal_group has a login in progress |
| * to one of the network portals. |
| */ |
| if (tpg_np->tpg_np->np_login_tpg != tpg) { |
| spin_unlock_bh(&np->np_thread_lock); |
| return 0; |
| } |
| } |
| if (np->np_thread_state == ISCSI_NP_THREAD_INACTIVE) { |
| spin_unlock_bh(&np->np_thread_lock); |
| return 0; |
| } |
| np->np_thread_state = ISCSI_NP_THREAD_RESET; |
| |
| if (np->np_thread) { |
| spin_unlock_bh(&np->np_thread_lock); |
| send_sig(SIGINT, np->np_thread, 1); |
| wait_for_completion(&np->np_restart_comp); |
| spin_lock_bh(&np->np_thread_lock); |
| } |
| spin_unlock_bh(&np->np_thread_lock); |
| |
| return 0; |
| } |
| |
| int iscsit_del_np_comm(struct iscsi_np *np) |
| { |
| if (!np->np_socket) |
| return 0; |
| |
| /* |
| * Some network transports allocate their own struct sock->file, |
| * see if we need to free any additional allocated resources. |
| */ |
| if (np->np_flags & NPF_SCTP_STRUCT_FILE) { |
| kfree(np->np_socket->file); |
| np->np_socket->file = NULL; |
| } |
| |
| sock_release(np->np_socket); |
| return 0; |
| } |
| |
| int iscsit_del_np(struct iscsi_np *np) |
| { |
| spin_lock_bh(&np->np_thread_lock); |
| np->np_exports--; |
| if (np->np_exports) { |
| spin_unlock_bh(&np->np_thread_lock); |
| return 0; |
| } |
| np->np_thread_state = ISCSI_NP_THREAD_SHUTDOWN; |
| spin_unlock_bh(&np->np_thread_lock); |
| |
| if (np->np_thread) { |
| /* |
| * We need to send the signal to wakeup Linux/Net |
| * which may be sleeping in sock_accept().. |
| */ |
| send_sig(SIGINT, np->np_thread, 1); |
| kthread_stop(np->np_thread); |
| } |
| iscsit_del_np_comm(np); |
| |
| spin_lock_bh(&np_lock); |
| list_del(&np->np_list); |
| spin_unlock_bh(&np_lock); |
| |
| pr_debug("CORE[0] - Removed Network Portal: %s:%hu on %s\n", |
| np->np_ip, np->np_port, (np->np_network_transport == ISCSI_TCP) ? |
| "TCP" : "SCTP"); |
| |
| kfree(np); |
| return 0; |
| } |
| |
| static int __init iscsi_target_init_module(void) |
| { |
| int ret = 0; |
| |
| pr_debug("iSCSI-Target "ISCSIT_VERSION"\n"); |
| |
| iscsit_global = kzalloc(sizeof(struct iscsit_global), GFP_KERNEL); |
| if (!iscsit_global) { |
| pr_err("Unable to allocate memory for iscsit_global\n"); |
| return -1; |
| } |
| mutex_init(&auth_id_lock); |
| spin_lock_init(&sess_idr_lock); |
| idr_init(&tiqn_idr); |
| idr_init(&sess_idr); |
| |
| ret = iscsi_target_register_configfs(); |
| if (ret < 0) |
| goto out; |
| |
| ret = iscsi_thread_set_init(); |
| if (ret < 0) |
| goto configfs_out; |
| |
| if (iscsi_allocate_thread_sets(TARGET_THREAD_SET_COUNT) != |
| TARGET_THREAD_SET_COUNT) { |
| pr_err("iscsi_allocate_thread_sets() returned" |
| " unexpected value!\n"); |
| goto ts_out1; |
| } |
| |
| lio_cmd_cache = kmem_cache_create("lio_cmd_cache", |
| sizeof(struct iscsi_cmd), __alignof__(struct iscsi_cmd), |
| 0, NULL); |
| if (!lio_cmd_cache) { |
| pr_err("Unable to kmem_cache_create() for" |
| " lio_cmd_cache\n"); |
| goto ts_out2; |
| } |
| |
| lio_qr_cache = kmem_cache_create("lio_qr_cache", |
| sizeof(struct iscsi_queue_req), |
| __alignof__(struct iscsi_queue_req), 0, NULL); |
| if (!lio_qr_cache) { |
| pr_err("nable to kmem_cache_create() for" |
| " lio_qr_cache\n"); |
| goto cmd_out; |
| } |
| |
| lio_dr_cache = kmem_cache_create("lio_dr_cache", |
| sizeof(struct iscsi_datain_req), |
| __alignof__(struct iscsi_datain_req), 0, NULL); |
| if (!lio_dr_cache) { |
| pr_err("Unable to kmem_cache_create() for" |
| " lio_dr_cache\n"); |
| goto qr_out; |
| } |
| |
| lio_ooo_cache = kmem_cache_create("lio_ooo_cache", |
| sizeof(struct iscsi_ooo_cmdsn), |
| __alignof__(struct iscsi_ooo_cmdsn), 0, NULL); |
| if (!lio_ooo_cache) { |
| pr_err("Unable to kmem_cache_create() for" |
| " lio_ooo_cache\n"); |
| goto dr_out; |
| } |
| |
| lio_r2t_cache = kmem_cache_create("lio_r2t_cache", |
| sizeof(struct iscsi_r2t), __alignof__(struct iscsi_r2t), |
| 0, NULL); |
| if (!lio_r2t_cache) { |
| pr_err("Unable to kmem_cache_create() for" |
| " lio_r2t_cache\n"); |
| goto ooo_out; |
| } |
| |
| if (iscsit_load_discovery_tpg() < 0) |
| goto r2t_out; |
| |
| return ret; |
| r2t_out: |
| kmem_cache_destroy(lio_r2t_cache); |
| ooo_out: |
| kmem_cache_destroy(lio_ooo_cache); |
| dr_out: |
| kmem_cache_destroy(lio_dr_cache); |
| qr_out: |
| kmem_cache_destroy(lio_qr_cache); |
| cmd_out: |
| kmem_cache_destroy(lio_cmd_cache); |
| ts_out2: |
| iscsi_deallocate_thread_sets(); |
| ts_out1: |
| iscsi_thread_set_free(); |
| configfs_out: |
| iscsi_target_deregister_configfs(); |
| out: |
| kfree(iscsit_global); |
| return -ENOMEM; |
| } |
| |
| static void __exit iscsi_target_cleanup_module(void) |
| { |
| iscsi_deallocate_thread_sets(); |
| iscsi_thread_set_free(); |
| iscsit_release_discovery_tpg(); |
| kmem_cache_destroy(lio_cmd_cache); |
| kmem_cache_destroy(lio_qr_cache); |
| kmem_cache_destroy(lio_dr_cache); |
| kmem_cache_destroy(lio_ooo_cache); |
| kmem_cache_destroy(lio_r2t_cache); |
| |
| iscsi_target_deregister_configfs(); |
| |
| kfree(iscsit_global); |
| } |
| |
| int iscsit_add_reject( |
| u8 reason, |
| int fail_conn, |
| unsigned char *buf, |
| struct iscsi_conn *conn) |
| { |
| struct iscsi_cmd *cmd; |
| struct iscsi_reject *hdr; |
| int ret; |
| |
| cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); |
| if (!cmd) |
| return -1; |
| |
| cmd->iscsi_opcode = ISCSI_OP_REJECT; |
| if (fail_conn) |
| cmd->cmd_flags |= ICF_REJECT_FAIL_CONN; |
| |
| hdr = (struct iscsi_reject *) cmd->pdu; |
| hdr->reason = reason; |
| |
| cmd->buf_ptr = kmemdup(buf, ISCSI_HDR_LEN, GFP_KERNEL); |
| if (!cmd->buf_ptr) { |
| pr_err("Unable to allocate memory for cmd->buf_ptr\n"); |
| iscsit_release_cmd(cmd); |
| return -1; |
| } |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| cmd->i_state = ISTATE_SEND_REJECT; |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| |
| ret = wait_for_completion_interruptible(&cmd->reject_comp); |
| if (ret != 0) |
| return -1; |
| |
| return (!fail_conn) ? 0 : -1; |
| } |
| |
| int iscsit_add_reject_from_cmd( |
| u8 reason, |
| int fail_conn, |
| int add_to_conn, |
| unsigned char *buf, |
| struct iscsi_cmd *cmd) |
| { |
| struct iscsi_conn *conn; |
| struct iscsi_reject *hdr; |
| int ret; |
| |
| if (!cmd->conn) { |
| pr_err("cmd->conn is NULL for ITT: 0x%08x\n", |
| cmd->init_task_tag); |
| return -1; |
| } |
| conn = cmd->conn; |
| |
| cmd->iscsi_opcode = ISCSI_OP_REJECT; |
| if (fail_conn) |
| cmd->cmd_flags |= ICF_REJECT_FAIL_CONN; |
| |
| hdr = (struct iscsi_reject *) cmd->pdu; |
| hdr->reason = reason; |
| |
| cmd->buf_ptr = kmemdup(buf, ISCSI_HDR_LEN, GFP_KERNEL); |
| if (!cmd->buf_ptr) { |
| pr_err("Unable to allocate memory for cmd->buf_ptr\n"); |
| iscsit_release_cmd(cmd); |
| return -1; |
| } |
| |
| if (add_to_conn) { |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| } |
| |
| cmd->i_state = ISTATE_SEND_REJECT; |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| |
| ret = wait_for_completion_interruptible(&cmd->reject_comp); |
| if (ret != 0) |
| return -1; |
| |
| return (!fail_conn) ? 0 : -1; |
| } |
| |
| /* |
| * Map some portion of the allocated scatterlist to an iovec, suitable for |
| * kernel sockets to copy data in/out. This handles both pages and slab-allocated |
| * buffers, since we have been tricky and mapped t_mem_sg to the buffer in |
| * either case (see iscsit_alloc_buffs) |
| */ |
| static int iscsit_map_iovec( |
| struct iscsi_cmd *cmd, |
| struct kvec *iov, |
| u32 data_offset, |
| u32 data_length) |
| { |
| u32 i = 0; |
| struct scatterlist *sg; |
| unsigned int page_off; |
| |
| /* |
| * We have a private mapping of the allocated pages in t_mem_sg. |
| * At this point, we also know each contains a page. |
| */ |
| sg = &cmd->t_mem_sg[data_offset / PAGE_SIZE]; |
| page_off = (data_offset % PAGE_SIZE); |
| |
| cmd->first_data_sg = sg; |
| cmd->first_data_sg_off = page_off; |
| |
| while (data_length) { |
| u32 cur_len = min_t(u32, data_length, sg->length - page_off); |
| |
| iov[i].iov_base = kmap(sg_page(sg)) + sg->offset + page_off; |
| iov[i].iov_len = cur_len; |
| |
| data_length -= cur_len; |
| page_off = 0; |
| sg = sg_next(sg); |
| i++; |
| } |
| |
| cmd->kmapped_nents = i; |
| |
| return i; |
| } |
| |
| static void iscsit_unmap_iovec(struct iscsi_cmd *cmd) |
| { |
| u32 i; |
| struct scatterlist *sg; |
| |
| sg = cmd->first_data_sg; |
| |
| for (i = 0; i < cmd->kmapped_nents; i++) |
| kunmap(sg_page(&sg[i])); |
| } |
| |
| static void iscsit_ack_from_expstatsn(struct iscsi_conn *conn, u32 exp_statsn) |
| { |
| struct iscsi_cmd *cmd; |
| |
| conn->exp_statsn = exp_statsn; |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_for_each_entry(cmd, &conn->conn_cmd_list, i_list) { |
| spin_lock(&cmd->istate_lock); |
| if ((cmd->i_state == ISTATE_SENT_STATUS) && |
| (cmd->stat_sn < exp_statsn)) { |
| cmd->i_state = ISTATE_REMOVE; |
| spin_unlock(&cmd->istate_lock); |
| iscsit_add_cmd_to_immediate_queue(cmd, conn, |
| cmd->i_state); |
| continue; |
| } |
| spin_unlock(&cmd->istate_lock); |
| } |
| spin_unlock_bh(&conn->cmd_lock); |
| } |
| |
| static int iscsit_allocate_iovecs(struct iscsi_cmd *cmd) |
| { |
| u32 iov_count = (cmd->se_cmd.t_data_nents == 0) ? 1 : |
| cmd->se_cmd.t_data_nents; |
| |
| iov_count += ISCSI_IOV_DATA_BUFFER; |
| |
| cmd->iov_data = kzalloc(iov_count * sizeof(struct kvec), GFP_KERNEL); |
| if (!cmd->iov_data) { |
| pr_err("Unable to allocate cmd->iov_data\n"); |
| return -ENOMEM; |
| } |
| |
| cmd->orig_iov_data_count = iov_count; |
| return 0; |
| } |
| |
| static int iscsit_alloc_buffs(struct iscsi_cmd *cmd) |
| { |
| struct scatterlist *sgl; |
| u32 length = cmd->se_cmd.data_length; |
| int nents = DIV_ROUND_UP(length, PAGE_SIZE); |
| int i = 0, ret; |
| /* |
| * If no SCSI payload is present, allocate the default iovecs used for |
| * iSCSI PDU Header |
| */ |
| if (!length) |
| return iscsit_allocate_iovecs(cmd); |
| |
| sgl = kzalloc(sizeof(*sgl) * nents, GFP_KERNEL); |
| if (!sgl) |
| return -ENOMEM; |
| |
| sg_init_table(sgl, nents); |
| |
| while (length) { |
| int buf_size = min_t(int, length, PAGE_SIZE); |
| struct page *page; |
| |
| page = alloc_page(GFP_KERNEL | __GFP_ZERO); |
| if (!page) |
| goto page_alloc_failed; |
| |
| sg_set_page(&sgl[i], page, buf_size, 0); |
| |
| length -= buf_size; |
| i++; |
| } |
| |
| cmd->t_mem_sg = sgl; |
| cmd->t_mem_sg_nents = nents; |
| |
| /* BIDI ops not supported */ |
| |
| /* Tell the core about our preallocated memory */ |
| transport_generic_map_mem_to_cmd(&cmd->se_cmd, sgl, nents, NULL, 0); |
| /* |
| * Allocate iovecs for SCSI payload after transport_generic_map_mem_to_cmd |
| * so that cmd->se_cmd.t_tasks_se_num has been set. |
| */ |
| ret = iscsit_allocate_iovecs(cmd); |
| if (ret < 0) |
| goto page_alloc_failed; |
| |
| return 0; |
| |
| page_alloc_failed: |
| while (i >= 0) { |
| __free_page(sg_page(&sgl[i])); |
| i--; |
| } |
| kfree(cmd->t_mem_sg); |
| cmd->t_mem_sg = NULL; |
| return -ENOMEM; |
| } |
| |
| static int iscsit_handle_scsi_cmd( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| int data_direction, cmdsn_ret = 0, immed_ret, ret, transport_ret; |
| int dump_immediate_data = 0, send_check_condition = 0, payload_length; |
| struct iscsi_cmd *cmd = NULL; |
| struct iscsi_scsi_req *hdr; |
| |
| spin_lock_bh(&conn->sess->session_stats_lock); |
| conn->sess->cmd_pdus++; |
| if (conn->sess->se_sess->se_node_acl) { |
| spin_lock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| conn->sess->se_sess->se_node_acl->num_cmds++; |
| spin_unlock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| } |
| spin_unlock_bh(&conn->sess->session_stats_lock); |
| |
| hdr = (struct iscsi_scsi_req *) buf; |
| payload_length = ntoh24(hdr->dlength); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->data_length = be32_to_cpu(hdr->data_length); |
| hdr->cmdsn = be32_to_cpu(hdr->cmdsn); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| |
| /* FIXME; Add checks for AdditionalHeaderSegment */ |
| |
| if (!(hdr->flags & ISCSI_FLAG_CMD_WRITE) && |
| !(hdr->flags & ISCSI_FLAG_CMD_FINAL)) { |
| pr_err("ISCSI_FLAG_CMD_WRITE & ISCSI_FLAG_CMD_FINAL" |
| " not set. Bad iSCSI Initiator.\n"); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| |
| if (((hdr->flags & ISCSI_FLAG_CMD_READ) || |
| (hdr->flags & ISCSI_FLAG_CMD_WRITE)) && !hdr->data_length) { |
| /* |
| * Vmware ESX v3.0 uses a modified Cisco Initiator (v3.4.2) |
| * that adds support for RESERVE/RELEASE. There is a bug |
| * add with this new functionality that sets R/W bits when |
| * neither CDB carries any READ or WRITE datapayloads. |
| */ |
| if ((hdr->cdb[0] == 0x16) || (hdr->cdb[0] == 0x17)) { |
| hdr->flags &= ~ISCSI_FLAG_CMD_READ; |
| hdr->flags &= ~ISCSI_FLAG_CMD_WRITE; |
| goto done; |
| } |
| |
| pr_err("ISCSI_FLAG_CMD_READ or ISCSI_FLAG_CMD_WRITE" |
| " set when Expected Data Transfer Length is 0 for" |
| " CDB: 0x%02x. Bad iSCSI Initiator.\n", hdr->cdb[0]); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| done: |
| |
| if (!(hdr->flags & ISCSI_FLAG_CMD_READ) && |
| !(hdr->flags & ISCSI_FLAG_CMD_WRITE) && (hdr->data_length != 0)) { |
| pr_err("ISCSI_FLAG_CMD_READ and/or ISCSI_FLAG_CMD_WRITE" |
| " MUST be set if Expected Data Transfer Length is not 0." |
| " Bad iSCSI Initiator\n"); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| |
| if ((hdr->flags & ISCSI_FLAG_CMD_READ) && |
| (hdr->flags & ISCSI_FLAG_CMD_WRITE)) { |
| pr_err("Bidirectional operations not supported!\n"); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| |
| if (hdr->opcode & ISCSI_OP_IMMEDIATE) { |
| pr_err("Illegally set Immediate Bit in iSCSI Initiator" |
| " Scsi Command PDU.\n"); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| |
| if (payload_length && !conn->sess->sess_ops->ImmediateData) { |
| pr_err("ImmediateData=No but DataSegmentLength=%u," |
| " protocol error.\n", payload_length); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| if ((hdr->data_length == payload_length) && |
| (!(hdr->flags & ISCSI_FLAG_CMD_FINAL))) { |
| pr_err("Expected Data Transfer Length and Length of" |
| " Immediate Data are the same, but ISCSI_FLAG_CMD_FINAL" |
| " bit is not set protocol error\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| if (payload_length > hdr->data_length) { |
| pr_err("DataSegmentLength: %u is greater than" |
| " EDTL: %u, protocol error.\n", payload_length, |
| hdr->data_length); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| if (payload_length > conn->conn_ops->MaxRecvDataSegmentLength) { |
| pr_err("DataSegmentLength: %u is greater than" |
| " MaxRecvDataSegmentLength: %u, protocol error.\n", |
| payload_length, conn->conn_ops->MaxRecvDataSegmentLength); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| if (payload_length > conn->sess->sess_ops->FirstBurstLength) { |
| pr_err("DataSegmentLength: %u is greater than" |
| " FirstBurstLength: %u, protocol error.\n", |
| payload_length, conn->sess->sess_ops->FirstBurstLength); |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_INVALID, 1, |
| buf, conn); |
| } |
| |
| data_direction = (hdr->flags & ISCSI_FLAG_CMD_WRITE) ? DMA_TO_DEVICE : |
| (hdr->flags & ISCSI_FLAG_CMD_READ) ? DMA_FROM_DEVICE : |
| DMA_NONE; |
| |
| cmd = iscsit_allocate_se_cmd(conn, hdr->data_length, data_direction, |
| (hdr->flags & ISCSI_FLAG_CMD_ATTR_MASK)); |
| if (!cmd) |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_NO_RESOURCES, 1, |
| buf, conn); |
| |
| pr_debug("Got SCSI Command, ITT: 0x%08x, CmdSN: 0x%08x," |
| " ExpXferLen: %u, Length: %u, CID: %hu\n", hdr->itt, |
| hdr->cmdsn, hdr->data_length, payload_length, conn->cid); |
| |
| cmd->iscsi_opcode = ISCSI_OP_SCSI_CMD; |
| cmd->i_state = ISTATE_NEW_CMD; |
| cmd->immediate_cmd = ((hdr->opcode & ISCSI_OP_IMMEDIATE) ? 1 : 0); |
| cmd->immediate_data = (payload_length) ? 1 : 0; |
| cmd->unsolicited_data = ((!(hdr->flags & ISCSI_FLAG_CMD_FINAL) && |
| (hdr->flags & ISCSI_FLAG_CMD_WRITE)) ? 1 : 0); |
| if (cmd->unsolicited_data) |
| cmd->cmd_flags |= ICF_NON_IMMEDIATE_UNSOLICITED_DATA; |
| |
| conn->sess->init_task_tag = cmd->init_task_tag = hdr->itt; |
| if (hdr->flags & ISCSI_FLAG_CMD_READ) { |
| spin_lock_bh(&conn->sess->ttt_lock); |
| cmd->targ_xfer_tag = conn->sess->targ_xfer_tag++; |
| if (cmd->targ_xfer_tag == 0xFFFFFFFF) |
| cmd->targ_xfer_tag = conn->sess->targ_xfer_tag++; |
| spin_unlock_bh(&conn->sess->ttt_lock); |
| } else if (hdr->flags & ISCSI_FLAG_CMD_WRITE) |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| cmd->cmd_sn = hdr->cmdsn; |
| cmd->exp_stat_sn = hdr->exp_statsn; |
| cmd->first_burst_len = payload_length; |
| |
| if (cmd->data_direction == DMA_FROM_DEVICE) { |
| struct iscsi_datain_req *dr; |
| |
| dr = iscsit_allocate_datain_req(); |
| if (!dr) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, 1, buf, cmd); |
| |
| iscsit_attach_datain_req(cmd, dr); |
| } |
| |
| /* |
| * The CDB is going to an se_device_t. |
| */ |
| ret = iscsit_get_lun_for_cmd(cmd, hdr->cdb, |
| get_unaligned_le64(&hdr->lun)); |
| if (ret < 0) { |
| if (cmd->se_cmd.scsi_sense_reason == TCM_NON_EXISTENT_LUN) { |
| pr_debug("Responding to non-acl'ed," |
| " non-existent or non-exported iSCSI LUN:" |
| " 0x%016Lx\n", get_unaligned_le64(&hdr->lun)); |
| } |
| send_check_condition = 1; |
| goto attach_cmd; |
| } |
| /* |
| * The Initiator Node has access to the LUN (the addressing method |
| * is handled inside of iscsit_get_lun_for_cmd()). Now it's time to |
| * allocate 1->N transport tasks (depending on sector count and |
| * maximum request size the physical HBA(s) can handle. |
| */ |
| transport_ret = transport_generic_allocate_tasks(&cmd->se_cmd, hdr->cdb); |
| if (transport_ret == -ENOMEM) { |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, 1, buf, cmd); |
| } else if (transport_ret == -EINVAL) { |
| /* |
| * Unsupported SAM Opcode. CHECK_CONDITION will be sent |
| * in iscsit_execute_cmd() during the CmdSN OOO Execution |
| * Mechinism. |
| */ |
| send_check_condition = 1; |
| } else { |
| cmd->data_length = cmd->se_cmd.data_length; |
| |
| if (iscsit_decide_list_to_build(cmd, payload_length) < 0) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, 1, buf, cmd); |
| } |
| |
| attach_cmd: |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| /* |
| * Check if we need to delay processing because of ALUA |
| * Active/NonOptimized primary access state.. |
| */ |
| core_alua_check_nonop_delay(&cmd->se_cmd); |
| /* |
| * Allocate and setup SGL used with transport_generic_map_mem_to_cmd(). |
| * also call iscsit_allocate_iovecs() |
| */ |
| ret = iscsit_alloc_buffs(cmd); |
| if (ret < 0) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, 1, buf, cmd); |
| /* |
| * Check the CmdSN against ExpCmdSN/MaxCmdSN here if |
| * the Immediate Bit is not set, and no Immediate |
| * Data is attached. |
| * |
| * A PDU/CmdSN carrying Immediate Data can only |
| * be processed after the DataCRC has passed. |
| * If the DataCRC fails, the CmdSN MUST NOT |
| * be acknowledged. (See below) |
| */ |
| if (!cmd->immediate_data) { |
| cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| if (cmdsn_ret == CMDSN_LOWER_THAN_EXP) |
| return 0; |
| else if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| } |
| |
| iscsit_ack_from_expstatsn(conn, hdr->exp_statsn); |
| |
| /* |
| * If no Immediate Data is attached, it's OK to return now. |
| */ |
| if (!cmd->immediate_data) { |
| if (send_check_condition) |
| return 0; |
| |
| if (cmd->unsolicited_data) { |
| iscsit_set_dataout_sequence_values(cmd); |
| |
| spin_lock_bh(&cmd->dataout_timeout_lock); |
| iscsit_start_dataout_timer(cmd, cmd->conn); |
| spin_unlock_bh(&cmd->dataout_timeout_lock); |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * Early CHECK_CONDITIONs never make it to the transport processing |
| * thread. They are processed in CmdSN order by |
| * iscsit_check_received_cmdsn() below. |
| */ |
| if (send_check_condition) { |
| immed_ret = IMMEDIATE_DATA_NORMAL_OPERATION; |
| dump_immediate_data = 1; |
| goto after_immediate_data; |
| } |
| /* |
| * Call directly into transport_generic_new_cmd() to perform |
| * the backend memory allocation. |
| */ |
| ret = transport_generic_new_cmd(&cmd->se_cmd); |
| if (ret < 0) { |
| immed_ret = IMMEDIATE_DATA_NORMAL_OPERATION; |
| dump_immediate_data = 1; |
| goto after_immediate_data; |
| } |
| |
| immed_ret = iscsit_handle_immediate_data(cmd, buf, payload_length); |
| after_immediate_data: |
| if (immed_ret == IMMEDIATE_DATA_NORMAL_OPERATION) { |
| /* |
| * A PDU/CmdSN carrying Immediate Data passed |
| * DataCRC, check against ExpCmdSN/MaxCmdSN if |
| * Immediate Bit is not set. |
| */ |
| cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| /* |
| * Special case for Unsupported SAM WRITE Opcodes |
| * and ImmediateData=Yes. |
| */ |
| if (dump_immediate_data) { |
| if (iscsit_dump_data_payload(conn, payload_length, 1) < 0) |
| return -1; |
| } else if (cmd->unsolicited_data) { |
| iscsit_set_dataout_sequence_values(cmd); |
| |
| spin_lock_bh(&cmd->dataout_timeout_lock); |
| iscsit_start_dataout_timer(cmd, cmd->conn); |
| spin_unlock_bh(&cmd->dataout_timeout_lock); |
| } |
| |
| if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| |
| } else if (immed_ret == IMMEDIATE_DATA_ERL1_CRC_FAILURE) { |
| /* |
| * Immediate Data failed DataCRC and ERL>=1, |
| * silently drop this PDU and let the initiator |
| * plug the CmdSN gap. |
| * |
| * FIXME: Send Unsolicited NOPIN with reserved |
| * TTT here to help the initiator figure out |
| * the missing CmdSN, although they should be |
| * intelligent enough to determine the missing |
| * CmdSN and issue a retry to plug the sequence. |
| */ |
| cmd->i_state = ISTATE_REMOVE; |
| iscsit_add_cmd_to_immediate_queue(cmd, conn, cmd->i_state); |
| } else /* immed_ret == IMMEDIATE_DATA_CANNOT_RECOVER */ |
| return -1; |
| |
| return 0; |
| } |
| |
| static u32 iscsit_do_crypto_hash_sg( |
| struct hash_desc *hash, |
| struct iscsi_cmd *cmd, |
| u32 data_offset, |
| u32 data_length, |
| u32 padding, |
| u8 *pad_bytes) |
| { |
| u32 data_crc; |
| u32 i; |
| struct scatterlist *sg; |
| unsigned int page_off; |
| |
| crypto_hash_init(hash); |
| |
| sg = cmd->first_data_sg; |
| page_off = cmd->first_data_sg_off; |
| |
| i = 0; |
| while (data_length) { |
| u32 cur_len = min_t(u32, data_length, (sg[i].length - page_off)); |
| |
| crypto_hash_update(hash, &sg[i], cur_len); |
| |
| data_length -= cur_len; |
| page_off = 0; |
| i++; |
| } |
| |
| if (padding) { |
| struct scatterlist pad_sg; |
| |
| sg_init_one(&pad_sg, pad_bytes, padding); |
| crypto_hash_update(hash, &pad_sg, padding); |
| } |
| crypto_hash_final(hash, (u8 *) &data_crc); |
| |
| return data_crc; |
| } |
| |
| static void iscsit_do_crypto_hash_buf( |
| struct hash_desc *hash, |
| unsigned char *buf, |
| u32 payload_length, |
| u32 padding, |
| u8 *pad_bytes, |
| u8 *data_crc) |
| { |
| struct scatterlist sg; |
| |
| crypto_hash_init(hash); |
| |
| sg_init_one(&sg, (u8 *)buf, payload_length); |
| crypto_hash_update(hash, &sg, payload_length); |
| |
| if (padding) { |
| sg_init_one(&sg, pad_bytes, padding); |
| crypto_hash_update(hash, &sg, padding); |
| } |
| crypto_hash_final(hash, data_crc); |
| } |
| |
| static int iscsit_handle_data_out(struct iscsi_conn *conn, unsigned char *buf) |
| { |
| int iov_ret, ooo_cmdsn = 0, ret; |
| u8 data_crc_failed = 0; |
| u32 checksum, iov_count = 0, padding = 0, rx_got = 0; |
| u32 rx_size = 0, payload_length; |
| struct iscsi_cmd *cmd = NULL; |
| struct se_cmd *se_cmd; |
| struct iscsi_data *hdr; |
| struct kvec *iov; |
| unsigned long flags; |
| |
| hdr = (struct iscsi_data *) buf; |
| payload_length = ntoh24(hdr->dlength); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->ttt = be32_to_cpu(hdr->ttt); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| hdr->datasn = be32_to_cpu(hdr->datasn); |
| hdr->offset = be32_to_cpu(hdr->offset); |
| |
| if (!payload_length) { |
| pr_err("DataOUT payload is ZERO, protocol error.\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| /* iSCSI write */ |
| spin_lock_bh(&conn->sess->session_stats_lock); |
| conn->sess->rx_data_octets += payload_length; |
| if (conn->sess->se_sess->se_node_acl) { |
| spin_lock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| conn->sess->se_sess->se_node_acl->write_bytes += payload_length; |
| spin_unlock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| } |
| spin_unlock_bh(&conn->sess->session_stats_lock); |
| |
| if (payload_length > conn->conn_ops->MaxRecvDataSegmentLength) { |
| pr_err("DataSegmentLength: %u is greater than" |
| " MaxRecvDataSegmentLength: %u\n", payload_length, |
| conn->conn_ops->MaxRecvDataSegmentLength); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| cmd = iscsit_find_cmd_from_itt_or_dump(conn, hdr->itt, |
| payload_length); |
| if (!cmd) |
| return 0; |
| |
| pr_debug("Got DataOut ITT: 0x%08x, TTT: 0x%08x," |
| " DataSN: 0x%08x, Offset: %u, Length: %u, CID: %hu\n", |
| hdr->itt, hdr->ttt, hdr->datasn, hdr->offset, |
| payload_length, conn->cid); |
| |
| if (cmd->cmd_flags & ICF_GOT_LAST_DATAOUT) { |
| pr_err("Command ITT: 0x%08x received DataOUT after" |
| " last DataOUT received, dumping payload\n", |
| cmd->init_task_tag); |
| return iscsit_dump_data_payload(conn, payload_length, 1); |
| } |
| |
| if (cmd->data_direction != DMA_TO_DEVICE) { |
| pr_err("Command ITT: 0x%08x received DataOUT for a" |
| " NON-WRITE command.\n", cmd->init_task_tag); |
| return iscsit_add_reject_from_cmd(ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| } |
| se_cmd = &cmd->se_cmd; |
| iscsit_mod_dataout_timer(cmd); |
| |
| if ((hdr->offset + payload_length) > cmd->data_length) { |
| pr_err("DataOut Offset: %u, Length %u greater than" |
| " iSCSI Command EDTL %u, protocol error.\n", |
| hdr->offset, payload_length, cmd->data_length); |
| return iscsit_add_reject_from_cmd(ISCSI_REASON_BOOKMARK_INVALID, |
| 1, 0, buf, cmd); |
| } |
| |
| if (cmd->unsolicited_data) { |
| int dump_unsolicited_data = 0; |
| |
| if (conn->sess->sess_ops->InitialR2T) { |
| pr_err("Received unexpected unsolicited data" |
| " while InitialR2T=Yes, protocol error.\n"); |
| transport_send_check_condition_and_sense(&cmd->se_cmd, |
| TCM_UNEXPECTED_UNSOLICITED_DATA, 0); |
| return -1; |
| } |
| /* |
| * Special case for dealing with Unsolicited DataOUT |
| * and Unsupported SAM WRITE Opcodes and SE resource allocation |
| * failures; |
| */ |
| |
| /* Something's amiss if we're not in WRITE_PENDING state... */ |
| spin_lock_irqsave(&se_cmd->t_state_lock, flags); |
| WARN_ON(se_cmd->t_state != TRANSPORT_WRITE_PENDING); |
| spin_unlock_irqrestore(&se_cmd->t_state_lock, flags); |
| |
| spin_lock_irqsave(&se_cmd->t_state_lock, flags); |
| if (!(se_cmd->se_cmd_flags & SCF_SUPPORTED_SAM_OPCODE) || |
| (se_cmd->se_cmd_flags & SCF_SCSI_CDB_EXCEPTION)) |
| dump_unsolicited_data = 1; |
| spin_unlock_irqrestore(&se_cmd->t_state_lock, flags); |
| |
| if (dump_unsolicited_data) { |
| /* |
| * Check if a delayed TASK_ABORTED status needs to |
| * be sent now if the ISCSI_FLAG_CMD_FINAL has been |
| * received with the unsolicitied data out. |
| */ |
| if (hdr->flags & ISCSI_FLAG_CMD_FINAL) |
| iscsit_stop_dataout_timer(cmd); |
| |
| transport_check_aborted_status(se_cmd, |
| (hdr->flags & ISCSI_FLAG_CMD_FINAL)); |
| return iscsit_dump_data_payload(conn, payload_length, 1); |
| } |
| } else { |
| /* |
| * For the normal solicited data path: |
| * |
| * Check for a delayed TASK_ABORTED status and dump any |
| * incoming data out payload if one exists. Also, when the |
| * ISCSI_FLAG_CMD_FINAL is set to denote the end of the current |
| * data out sequence, we decrement outstanding_r2ts. Once |
| * outstanding_r2ts reaches zero, go ahead and send the delayed |
| * TASK_ABORTED status. |
| */ |
| if (atomic_read(&se_cmd->t_transport_aborted) != 0) { |
| if (hdr->flags & ISCSI_FLAG_CMD_FINAL) |
| if (--cmd->outstanding_r2ts < 1) { |
| iscsit_stop_dataout_timer(cmd); |
| transport_check_aborted_status( |
| se_cmd, 1); |
| } |
| |
| return iscsit_dump_data_payload(conn, payload_length, 1); |
| } |
| } |
| /* |
| * Preform DataSN, DataSequenceInOrder, DataPDUInOrder, and |
| * within-command recovery checks before receiving the payload. |
| */ |
| ret = iscsit_check_pre_dataout(cmd, buf); |
| if (ret == DATAOUT_WITHIN_COMMAND_RECOVERY) |
| return 0; |
| else if (ret == DATAOUT_CANNOT_RECOVER) |
| return -1; |
| |
| rx_size += payload_length; |
| iov = &cmd->iov_data[0]; |
| |
| iov_ret = iscsit_map_iovec(cmd, iov, hdr->offset, payload_length); |
| if (iov_ret < 0) |
| return -1; |
| |
| iov_count += iov_ret; |
| |
| padding = ((-payload_length) & 3); |
| if (padding != 0) { |
| iov[iov_count].iov_base = cmd->pad_bytes; |
| iov[iov_count++].iov_len = padding; |
| rx_size += padding; |
| pr_debug("Receiving %u padding bytes.\n", padding); |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iov[iov_count].iov_base = &checksum; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| rx_size += ISCSI_CRC_LEN; |
| } |
| |
| rx_got = rx_data(conn, &cmd->iov_data[0], iov_count, rx_size); |
| |
| iscsit_unmap_iovec(cmd); |
| |
| if (rx_got != rx_size) |
| return -1; |
| |
| if (conn->conn_ops->DataDigest) { |
| u32 data_crc; |
| |
| data_crc = iscsit_do_crypto_hash_sg(&conn->conn_rx_hash, cmd, |
| hdr->offset, payload_length, padding, |
| cmd->pad_bytes); |
| |
| if (checksum != data_crc) { |
| pr_err("ITT: 0x%08x, Offset: %u, Length: %u," |
| " DataSN: 0x%08x, CRC32C DataDigest 0x%08x" |
| " does not match computed 0x%08x\n", |
| hdr->itt, hdr->offset, payload_length, |
| hdr->datasn, checksum, data_crc); |
| data_crc_failed = 1; |
| } else { |
| pr_debug("Got CRC32C DataDigest 0x%08x for" |
| " %u bytes of Data Out\n", checksum, |
| payload_length); |
| } |
| } |
| /* |
| * Increment post receive data and CRC values or perform |
| * within-command recovery. |
| */ |
| ret = iscsit_check_post_dataout(cmd, buf, data_crc_failed); |
| if ((ret == DATAOUT_NORMAL) || (ret == DATAOUT_WITHIN_COMMAND_RECOVERY)) |
| return 0; |
| else if (ret == DATAOUT_SEND_R2T) { |
| iscsit_set_dataout_sequence_values(cmd); |
| iscsit_build_r2ts_for_cmd(cmd, conn, 0); |
| } else if (ret == DATAOUT_SEND_TO_TRANSPORT) { |
| /* |
| * Handle extra special case for out of order |
| * Unsolicited Data Out. |
| */ |
| spin_lock_bh(&cmd->istate_lock); |
| ooo_cmdsn = (cmd->cmd_flags & ICF_OOO_CMDSN); |
| cmd->cmd_flags |= ICF_GOT_LAST_DATAOUT; |
| cmd->i_state = ISTATE_RECEIVED_LAST_DATAOUT; |
| spin_unlock_bh(&cmd->istate_lock); |
| |
| iscsit_stop_dataout_timer(cmd); |
| return (!ooo_cmdsn) ? transport_generic_handle_data( |
| &cmd->se_cmd) : 0; |
| } else /* DATAOUT_CANNOT_RECOVER */ |
| return -1; |
| |
| return 0; |
| } |
| |
| static int iscsit_handle_nop_out( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| unsigned char *ping_data = NULL; |
| int cmdsn_ret, niov = 0, ret = 0, rx_got, rx_size; |
| u32 checksum, data_crc, padding = 0, payload_length; |
| u64 lun; |
| struct iscsi_cmd *cmd = NULL; |
| struct kvec *iov = NULL; |
| struct iscsi_nopout *hdr; |
| |
| hdr = (struct iscsi_nopout *) buf; |
| payload_length = ntoh24(hdr->dlength); |
| lun = get_unaligned_le64(&hdr->lun); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->ttt = be32_to_cpu(hdr->ttt); |
| hdr->cmdsn = be32_to_cpu(hdr->cmdsn); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| |
| if ((hdr->itt == 0xFFFFFFFF) && !(hdr->opcode & ISCSI_OP_IMMEDIATE)) { |
| pr_err("NOPOUT ITT is reserved, but Immediate Bit is" |
| " not set, protocol error.\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| if (payload_length > conn->conn_ops->MaxRecvDataSegmentLength) { |
| pr_err("NOPOUT Ping Data DataSegmentLength: %u is" |
| " greater than MaxRecvDataSegmentLength: %u, protocol" |
| " error.\n", payload_length, |
| conn->conn_ops->MaxRecvDataSegmentLength); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| pr_debug("Got NOPOUT Ping %s ITT: 0x%08x, TTT: 0x%09x," |
| " CmdSN: 0x%08x, ExpStatSN: 0x%08x, Length: %u\n", |
| (hdr->itt == 0xFFFFFFFF) ? "Response" : "Request", |
| hdr->itt, hdr->ttt, hdr->cmdsn, hdr->exp_statsn, |
| payload_length); |
| /* |
| * This is not a response to a Unsolicited NopIN, which means |
| * it can either be a NOPOUT ping request (with a valid ITT), |
| * or a NOPOUT not requesting a NOPIN (with a reserved ITT). |
| * Either way, make sure we allocate an struct iscsi_cmd, as both |
| * can contain ping data. |
| */ |
| if (hdr->ttt == 0xFFFFFFFF) { |
| cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); |
| if (!cmd) |
| return iscsit_add_reject( |
| ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, buf, conn); |
| |
| cmd->iscsi_opcode = ISCSI_OP_NOOP_OUT; |
| cmd->i_state = ISTATE_SEND_NOPIN; |
| cmd->immediate_cmd = ((hdr->opcode & ISCSI_OP_IMMEDIATE) ? |
| 1 : 0); |
| conn->sess->init_task_tag = cmd->init_task_tag = hdr->itt; |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| cmd->cmd_sn = hdr->cmdsn; |
| cmd->exp_stat_sn = hdr->exp_statsn; |
| cmd->data_direction = DMA_NONE; |
| } |
| |
| if (payload_length && (hdr->ttt == 0xFFFFFFFF)) { |
| rx_size = payload_length; |
| ping_data = kzalloc(payload_length + 1, GFP_KERNEL); |
| if (!ping_data) { |
| pr_err("Unable to allocate memory for" |
| " NOPOUT ping data.\n"); |
| ret = -1; |
| goto out; |
| } |
| |
| iov = &cmd->iov_misc[0]; |
| iov[niov].iov_base = ping_data; |
| iov[niov++].iov_len = payload_length; |
| |
| padding = ((-payload_length) & 3); |
| if (padding != 0) { |
| pr_debug("Receiving %u additional bytes" |
| " for padding.\n", padding); |
| iov[niov].iov_base = &cmd->pad_bytes; |
| iov[niov++].iov_len = padding; |
| rx_size += padding; |
| } |
| if (conn->conn_ops->DataDigest) { |
| iov[niov].iov_base = &checksum; |
| iov[niov++].iov_len = ISCSI_CRC_LEN; |
| rx_size += ISCSI_CRC_LEN; |
| } |
| |
| rx_got = rx_data(conn, &cmd->iov_misc[0], niov, rx_size); |
| if (rx_got != rx_size) { |
| ret = -1; |
| goto out; |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_rx_hash, |
| ping_data, payload_length, |
| padding, cmd->pad_bytes, |
| (u8 *)&data_crc); |
| |
| if (checksum != data_crc) { |
| pr_err("Ping data CRC32C DataDigest" |
| " 0x%08x does not match computed 0x%08x\n", |
| checksum, data_crc); |
| if (!conn->sess->sess_ops->ErrorRecoveryLevel) { |
| pr_err("Unable to recover from" |
| " NOPOUT Ping DataCRC failure while in" |
| " ERL=0.\n"); |
| ret = -1; |
| goto out; |
| } else { |
| /* |
| * Silently drop this PDU and let the |
| * initiator plug the CmdSN gap. |
| */ |
| pr_debug("Dropping NOPOUT" |
| " Command CmdSN: 0x%08x due to" |
| " DataCRC error.\n", hdr->cmdsn); |
| ret = 0; |
| goto out; |
| } |
| } else { |
| pr_debug("Got CRC32C DataDigest" |
| " 0x%08x for %u bytes of ping data.\n", |
| checksum, payload_length); |
| } |
| } |
| |
| ping_data[payload_length] = '\0'; |
| /* |
| * Attach ping data to struct iscsi_cmd->buf_ptr. |
| */ |
| cmd->buf_ptr = (void *)ping_data; |
| cmd->buf_ptr_size = payload_length; |
| |
| pr_debug("Got %u bytes of NOPOUT ping" |
| " data.\n", payload_length); |
| pr_debug("Ping Data: \"%s\"\n", ping_data); |
| } |
| |
| if (hdr->itt != 0xFFFFFFFF) { |
| if (!cmd) { |
| pr_err("Checking CmdSN for NOPOUT," |
| " but cmd is NULL!\n"); |
| return -1; |
| } |
| /* |
| * Initiator is expecting a NopIN ping reply, |
| */ |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| iscsit_ack_from_expstatsn(conn, hdr->exp_statsn); |
| |
| if (hdr->opcode & ISCSI_OP_IMMEDIATE) { |
| iscsit_add_cmd_to_response_queue(cmd, conn, |
| cmd->i_state); |
| return 0; |
| } |
| |
| cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| if (cmdsn_ret == CMDSN_LOWER_THAN_EXP) { |
| ret = 0; |
| goto ping_out; |
| } |
| if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| |
| return 0; |
| } |
| |
| if (hdr->ttt != 0xFFFFFFFF) { |
| /* |
| * This was a response to a unsolicited NOPIN ping. |
| */ |
| cmd = iscsit_find_cmd_from_ttt(conn, hdr->ttt); |
| if (!cmd) |
| return -1; |
| |
| iscsit_stop_nopin_response_timer(conn); |
| |
| cmd->i_state = ISTATE_REMOVE; |
| iscsit_add_cmd_to_immediate_queue(cmd, conn, cmd->i_state); |
| iscsit_start_nopin_timer(conn); |
| } else { |
| /* |
| * Initiator is not expecting a NOPIN is response. |
| * Just ignore for now. |
| * |
| * iSCSI v19-91 10.18 |
| * "A NOP-OUT may also be used to confirm a changed |
| * ExpStatSN if another PDU will not be available |
| * for a long time." |
| */ |
| ret = 0; |
| goto out; |
| } |
| |
| return 0; |
| out: |
| if (cmd) |
| iscsit_release_cmd(cmd); |
| ping_out: |
| kfree(ping_data); |
| return ret; |
| } |
| |
| static int iscsit_handle_task_mgt_cmd( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| struct iscsi_cmd *cmd; |
| struct se_tmr_req *se_tmr; |
| struct iscsi_tmr_req *tmr_req; |
| struct iscsi_tm *hdr; |
| u32 payload_length; |
| int out_of_order_cmdsn = 0; |
| int ret; |
| u8 function; |
| |
| hdr = (struct iscsi_tm *) buf; |
| payload_length = ntoh24(hdr->dlength); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->rtt = be32_to_cpu(hdr->rtt); |
| hdr->cmdsn = be32_to_cpu(hdr->cmdsn); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| hdr->refcmdsn = be32_to_cpu(hdr->refcmdsn); |
| hdr->exp_datasn = be32_to_cpu(hdr->exp_datasn); |
| hdr->flags &= ~ISCSI_FLAG_CMD_FINAL; |
| function = hdr->flags; |
| |
| pr_debug("Got Task Management Request ITT: 0x%08x, CmdSN:" |
| " 0x%08x, Function: 0x%02x, RefTaskTag: 0x%08x, RefCmdSN:" |
| " 0x%08x, CID: %hu\n", hdr->itt, hdr->cmdsn, function, |
| hdr->rtt, hdr->refcmdsn, conn->cid); |
| |
| if ((function != ISCSI_TM_FUNC_ABORT_TASK) && |
| ((function != ISCSI_TM_FUNC_TASK_REASSIGN) && |
| (hdr->rtt != ISCSI_RESERVED_TAG))) { |
| pr_err("RefTaskTag should be set to 0xFFFFFFFF.\n"); |
| hdr->rtt = ISCSI_RESERVED_TAG; |
| } |
| |
| if ((function == ISCSI_TM_FUNC_TASK_REASSIGN) && |
| !(hdr->opcode & ISCSI_OP_IMMEDIATE)) { |
| pr_err("Task Management Request TASK_REASSIGN not" |
| " issued as immediate command, bad iSCSI Initiator" |
| "implementation\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| if ((function != ISCSI_TM_FUNC_ABORT_TASK) && |
| (hdr->refcmdsn != ISCSI_RESERVED_TAG)) |
| hdr->refcmdsn = ISCSI_RESERVED_TAG; |
| |
| cmd = iscsit_allocate_se_cmd_for_tmr(conn, function); |
| if (!cmd) |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, buf, conn); |
| |
| cmd->iscsi_opcode = ISCSI_OP_SCSI_TMFUNC; |
| cmd->i_state = ISTATE_SEND_TASKMGTRSP; |
| cmd->immediate_cmd = ((hdr->opcode & ISCSI_OP_IMMEDIATE) ? 1 : 0); |
| cmd->init_task_tag = hdr->itt; |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| cmd->cmd_sn = hdr->cmdsn; |
| cmd->exp_stat_sn = hdr->exp_statsn; |
| se_tmr = cmd->se_cmd.se_tmr_req; |
| tmr_req = cmd->tmr_req; |
| /* |
| * Locate the struct se_lun for all TMRs not related to ERL=2 TASK_REASSIGN |
| */ |
| if (function != ISCSI_TM_FUNC_TASK_REASSIGN) { |
| ret = iscsit_get_lun_for_tmr(cmd, |
| get_unaligned_le64(&hdr->lun)); |
| if (ret < 0) { |
| cmd->se_cmd.se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION; |
| se_tmr->response = ISCSI_TMF_RSP_NO_LUN; |
| goto attach; |
| } |
| } |
| |
| switch (function) { |
| case ISCSI_TM_FUNC_ABORT_TASK: |
| se_tmr->response = iscsit_tmr_abort_task(cmd, buf); |
| if (se_tmr->response != ISCSI_TMF_RSP_COMPLETE) { |
| cmd->se_cmd.se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION; |
| goto attach; |
| } |
| break; |
| case ISCSI_TM_FUNC_ABORT_TASK_SET: |
| case ISCSI_TM_FUNC_CLEAR_ACA: |
| case ISCSI_TM_FUNC_CLEAR_TASK_SET: |
| case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET: |
| break; |
| case ISCSI_TM_FUNC_TARGET_WARM_RESET: |
| if (iscsit_tmr_task_warm_reset(conn, tmr_req, buf) < 0) { |
| cmd->se_cmd.se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION; |
| se_tmr->response = ISCSI_TMF_RSP_AUTH_FAILED; |
| goto attach; |
| } |
| break; |
| case ISCSI_TM_FUNC_TARGET_COLD_RESET: |
| if (iscsit_tmr_task_cold_reset(conn, tmr_req, buf) < 0) { |
| cmd->se_cmd.se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION; |
| se_tmr->response = ISCSI_TMF_RSP_AUTH_FAILED; |
| goto attach; |
| } |
| break; |
| case ISCSI_TM_FUNC_TASK_REASSIGN: |
| se_tmr->response = iscsit_tmr_task_reassign(cmd, buf); |
| /* |
| * Perform sanity checks on the ExpDataSN only if the |
| * TASK_REASSIGN was successful. |
| */ |
| if (se_tmr->response != ISCSI_TMF_RSP_COMPLETE) |
| break; |
| |
| if (iscsit_check_task_reassign_expdatasn(tmr_req, conn) < 0) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_BOOKMARK_INVALID, 1, 1, |
| buf, cmd); |
| break; |
| default: |
| pr_err("Unknown TMR function: 0x%02x, protocol" |
| " error.\n", function); |
| cmd->se_cmd.se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION; |
| se_tmr->response = ISCSI_TMF_RSP_NOT_SUPPORTED; |
| goto attach; |
| } |
| |
| if ((function != ISCSI_TM_FUNC_TASK_REASSIGN) && |
| (se_tmr->response == ISCSI_TMF_RSP_COMPLETE)) |
| se_tmr->call_transport = 1; |
| attach: |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| if (!(hdr->opcode & ISCSI_OP_IMMEDIATE)) { |
| int cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| if (cmdsn_ret == CMDSN_HIGHER_THAN_EXP) |
| out_of_order_cmdsn = 1; |
| else if (cmdsn_ret == CMDSN_LOWER_THAN_EXP) |
| return 0; |
| else if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| } |
| iscsit_ack_from_expstatsn(conn, hdr->exp_statsn); |
| |
| if (out_of_order_cmdsn || !(hdr->opcode & ISCSI_OP_IMMEDIATE)) |
| return 0; |
| /* |
| * Found the referenced task, send to transport for processing. |
| */ |
| if (se_tmr->call_transport) |
| return transport_generic_handle_tmr(&cmd->se_cmd); |
| |
| /* |
| * Could not find the referenced LUN, task, or Task Management |
| * command not authorized or supported. Change state and |
| * let the tx_thread send the response. |
| * |
| * For connection recovery, this is also the default action for |
| * TMR TASK_REASSIGN. |
| */ |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| return 0; |
| } |
| |
| /* #warning FIXME: Support Text Command parameters besides SendTargets */ |
| static int iscsit_handle_text_cmd( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| char *text_ptr, *text_in; |
| int cmdsn_ret, niov = 0, rx_got, rx_size; |
| u32 checksum = 0, data_crc = 0, payload_length; |
| u32 padding = 0, pad_bytes = 0, text_length = 0; |
| struct iscsi_cmd *cmd; |
| struct kvec iov[3]; |
| struct iscsi_text *hdr; |
| |
| hdr = (struct iscsi_text *) buf; |
| payload_length = ntoh24(hdr->dlength); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->ttt = be32_to_cpu(hdr->ttt); |
| hdr->cmdsn = be32_to_cpu(hdr->cmdsn); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| |
| if (payload_length > conn->conn_ops->MaxRecvDataSegmentLength) { |
| pr_err("Unable to accept text parameter length: %u" |
| "greater than MaxRecvDataSegmentLength %u.\n", |
| payload_length, conn->conn_ops->MaxRecvDataSegmentLength); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| pr_debug("Got Text Request: ITT: 0x%08x, CmdSN: 0x%08x," |
| " ExpStatSN: 0x%08x, Length: %u\n", hdr->itt, hdr->cmdsn, |
| hdr->exp_statsn, payload_length); |
| |
| rx_size = text_length = payload_length; |
| if (text_length) { |
| text_in = kzalloc(text_length, GFP_KERNEL); |
| if (!text_in) { |
| pr_err("Unable to allocate memory for" |
| " incoming text parameters\n"); |
| return -1; |
| } |
| |
| memset(iov, 0, 3 * sizeof(struct kvec)); |
| iov[niov].iov_base = text_in; |
| iov[niov++].iov_len = text_length; |
| |
| padding = ((-payload_length) & 3); |
| if (padding != 0) { |
| iov[niov].iov_base = &pad_bytes; |
| iov[niov++].iov_len = padding; |
| rx_size += padding; |
| pr_debug("Receiving %u additional bytes" |
| " for padding.\n", padding); |
| } |
| if (conn->conn_ops->DataDigest) { |
| iov[niov].iov_base = &checksum; |
| iov[niov++].iov_len = ISCSI_CRC_LEN; |
| rx_size += ISCSI_CRC_LEN; |
| } |
| |
| rx_got = rx_data(conn, &iov[0], niov, rx_size); |
| if (rx_got != rx_size) { |
| kfree(text_in); |
| return -1; |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_rx_hash, |
| text_in, text_length, |
| padding, (u8 *)&pad_bytes, |
| (u8 *)&data_crc); |
| |
| if (checksum != data_crc) { |
| pr_err("Text data CRC32C DataDigest" |
| " 0x%08x does not match computed" |
| " 0x%08x\n", checksum, data_crc); |
| if (!conn->sess->sess_ops->ErrorRecoveryLevel) { |
| pr_err("Unable to recover from" |
| " Text Data digest failure while in" |
| " ERL=0.\n"); |
| kfree(text_in); |
| return -1; |
| } else { |
| /* |
| * Silently drop this PDU and let the |
| * initiator plug the CmdSN gap. |
| */ |
| pr_debug("Dropping Text" |
| " Command CmdSN: 0x%08x due to" |
| " DataCRC error.\n", hdr->cmdsn); |
| kfree(text_in); |
| return 0; |
| } |
| } else { |
| pr_debug("Got CRC32C DataDigest" |
| " 0x%08x for %u bytes of text data.\n", |
| checksum, text_length); |
| } |
| } |
| text_in[text_length - 1] = '\0'; |
| pr_debug("Successfully read %d bytes of text" |
| " data.\n", text_length); |
| |
| if (strncmp("SendTargets", text_in, 11) != 0) { |
| pr_err("Received Text Data that is not" |
| " SendTargets, cannot continue.\n"); |
| kfree(text_in); |
| return -1; |
| } |
| text_ptr = strchr(text_in, '='); |
| if (!text_ptr) { |
| pr_err("No \"=\" separator found in Text Data," |
| " cannot continue.\n"); |
| kfree(text_in); |
| return -1; |
| } |
| if (strncmp("=All", text_ptr, 4) != 0) { |
| pr_err("Unable to locate All value for" |
| " SendTargets key, cannot continue.\n"); |
| kfree(text_in); |
| return -1; |
| } |
| /*#warning Support SendTargets=(iSCSI Target Name/Nothing) values. */ |
| kfree(text_in); |
| } |
| |
| cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); |
| if (!cmd) |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_NO_RESOURCES, |
| 1, buf, conn); |
| |
| cmd->iscsi_opcode = ISCSI_OP_TEXT; |
| cmd->i_state = ISTATE_SEND_TEXTRSP; |
| cmd->immediate_cmd = ((hdr->opcode & ISCSI_OP_IMMEDIATE) ? 1 : 0); |
| conn->sess->init_task_tag = cmd->init_task_tag = hdr->itt; |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| cmd->cmd_sn = hdr->cmdsn; |
| cmd->exp_stat_sn = hdr->exp_statsn; |
| cmd->data_direction = DMA_NONE; |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| iscsit_ack_from_expstatsn(conn, hdr->exp_statsn); |
| |
| if (!(hdr->opcode & ISCSI_OP_IMMEDIATE)) { |
| cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| |
| return 0; |
| } |
| |
| return iscsit_execute_cmd(cmd, 0); |
| } |
| |
| int iscsit_logout_closesession(struct iscsi_cmd *cmd, struct iscsi_conn *conn) |
| { |
| struct iscsi_conn *conn_p; |
| struct iscsi_session *sess = conn->sess; |
| |
| pr_debug("Received logout request CLOSESESSION on CID: %hu" |
| " for SID: %u.\n", conn->cid, conn->sess->sid); |
| |
| atomic_set(&sess->session_logout, 1); |
| atomic_set(&conn->conn_logout_remove, 1); |
| conn->conn_logout_reason = ISCSI_LOGOUT_REASON_CLOSE_SESSION; |
| |
| iscsit_inc_conn_usage_count(conn); |
| iscsit_inc_session_usage_count(sess); |
| |
| spin_lock_bh(&sess->conn_lock); |
| list_for_each_entry(conn_p, &sess->sess_conn_list, conn_list) { |
| if (conn_p->conn_state != TARG_CONN_STATE_LOGGED_IN) |
| continue; |
| |
| pr_debug("Moving to TARG_CONN_STATE_IN_LOGOUT.\n"); |
| conn_p->conn_state = TARG_CONN_STATE_IN_LOGOUT; |
| } |
| spin_unlock_bh(&sess->conn_lock); |
| |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| |
| return 0; |
| } |
| |
| int iscsit_logout_closeconnection(struct iscsi_cmd *cmd, struct iscsi_conn *conn) |
| { |
| struct iscsi_conn *l_conn; |
| struct iscsi_session *sess = conn->sess; |
| |
| pr_debug("Received logout request CLOSECONNECTION for CID:" |
| " %hu on CID: %hu.\n", cmd->logout_cid, conn->cid); |
| |
| /* |
| * A Logout Request with a CLOSECONNECTION reason code for a CID |
| * can arrive on a connection with a differing CID. |
| */ |
| if (conn->cid == cmd->logout_cid) { |
| spin_lock_bh(&conn->state_lock); |
| pr_debug("Moving to TARG_CONN_STATE_IN_LOGOUT.\n"); |
| conn->conn_state = TARG_CONN_STATE_IN_LOGOUT; |
| |
| atomic_set(&conn->conn_logout_remove, 1); |
| conn->conn_logout_reason = ISCSI_LOGOUT_REASON_CLOSE_CONNECTION; |
| iscsit_inc_conn_usage_count(conn); |
| |
| spin_unlock_bh(&conn->state_lock); |
| } else { |
| /* |
| * Handle all different cid CLOSECONNECTION requests in |
| * iscsit_logout_post_handler_diffcid() as to give enough |
| * time for any non immediate command's CmdSN to be |
| * acknowledged on the connection in question. |
| * |
| * Here we simply make sure the CID is still around. |
| */ |
| l_conn = iscsit_get_conn_from_cid(sess, |
| cmd->logout_cid); |
| if (!l_conn) { |
| cmd->logout_response = ISCSI_LOGOUT_CID_NOT_FOUND; |
| iscsit_add_cmd_to_response_queue(cmd, conn, |
| cmd->i_state); |
| return 0; |
| } |
| |
| iscsit_dec_conn_usage_count(l_conn); |
| } |
| |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| |
| return 0; |
| } |
| |
| int iscsit_logout_removeconnforrecovery(struct iscsi_cmd *cmd, struct iscsi_conn *conn) |
| { |
| struct iscsi_session *sess = conn->sess; |
| |
| pr_debug("Received explicit REMOVECONNFORRECOVERY logout for" |
| " CID: %hu on CID: %hu.\n", cmd->logout_cid, conn->cid); |
| |
| if (sess->sess_ops->ErrorRecoveryLevel != 2) { |
| pr_err("Received Logout Request REMOVECONNFORRECOVERY" |
| " while ERL!=2.\n"); |
| cmd->logout_response = ISCSI_LOGOUT_RECOVERY_UNSUPPORTED; |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| return 0; |
| } |
| |
| if (conn->cid == cmd->logout_cid) { |
| pr_err("Received Logout Request REMOVECONNFORRECOVERY" |
| " with CID: %hu on CID: %hu, implementation error.\n", |
| cmd->logout_cid, conn->cid); |
| cmd->logout_response = ISCSI_LOGOUT_CLEANUP_FAILED; |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| return 0; |
| } |
| |
| iscsit_add_cmd_to_response_queue(cmd, conn, cmd->i_state); |
| |
| return 0; |
| } |
| |
| static int iscsit_handle_logout_cmd( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| int cmdsn_ret, logout_remove = 0; |
| u8 reason_code = 0; |
| struct iscsi_cmd *cmd; |
| struct iscsi_logout *hdr; |
| struct iscsi_tiqn *tiqn = iscsit_snmp_get_tiqn(conn); |
| |
| hdr = (struct iscsi_logout *) buf; |
| reason_code = (hdr->flags & 0x7f); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->cid = be16_to_cpu(hdr->cid); |
| hdr->cmdsn = be32_to_cpu(hdr->cmdsn); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| |
| if (tiqn) { |
| spin_lock(&tiqn->logout_stats.lock); |
| if (reason_code == ISCSI_LOGOUT_REASON_CLOSE_SESSION) |
| tiqn->logout_stats.normal_logouts++; |
| else |
| tiqn->logout_stats.abnormal_logouts++; |
| spin_unlock(&tiqn->logout_stats.lock); |
| } |
| |
| pr_debug("Got Logout Request ITT: 0x%08x CmdSN: 0x%08x" |
| " ExpStatSN: 0x%08x Reason: 0x%02x CID: %hu on CID: %hu\n", |
| hdr->itt, hdr->cmdsn, hdr->exp_statsn, reason_code, |
| hdr->cid, conn->cid); |
| |
| if (conn->conn_state != TARG_CONN_STATE_LOGGED_IN) { |
| pr_err("Received logout request on connection that" |
| " is not in logged in state, ignoring request.\n"); |
| return 0; |
| } |
| |
| cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); |
| if (!cmd) |
| return iscsit_add_reject(ISCSI_REASON_BOOKMARK_NO_RESOURCES, 1, |
| buf, conn); |
| |
| cmd->iscsi_opcode = ISCSI_OP_LOGOUT; |
| cmd->i_state = ISTATE_SEND_LOGOUTRSP; |
| cmd->immediate_cmd = ((hdr->opcode & ISCSI_OP_IMMEDIATE) ? 1 : 0); |
| conn->sess->init_task_tag = cmd->init_task_tag = hdr->itt; |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| cmd->cmd_sn = hdr->cmdsn; |
| cmd->exp_stat_sn = hdr->exp_statsn; |
| cmd->logout_cid = hdr->cid; |
| cmd->logout_reason = reason_code; |
| cmd->data_direction = DMA_NONE; |
| |
| /* |
| * We need to sleep in these cases (by returning 1) until the Logout |
| * Response gets sent in the tx thread. |
| */ |
| if ((reason_code == ISCSI_LOGOUT_REASON_CLOSE_SESSION) || |
| ((reason_code == ISCSI_LOGOUT_REASON_CLOSE_CONNECTION) && |
| (hdr->cid == conn->cid))) |
| logout_remove = 1; |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn->conn_cmd_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| if (reason_code != ISCSI_LOGOUT_REASON_RECOVERY) |
| iscsit_ack_from_expstatsn(conn, hdr->exp_statsn); |
| |
| /* |
| * Immediate commands are executed, well, immediately. |
| * Non-Immediate Logout Commands are executed in CmdSN order. |
| */ |
| if (hdr->opcode & ISCSI_OP_IMMEDIATE) { |
| int ret = iscsit_execute_cmd(cmd, 0); |
| |
| if (ret < 0) |
| return ret; |
| } else { |
| cmdsn_ret = iscsit_sequence_cmd(conn, cmd, hdr->cmdsn); |
| if (cmdsn_ret == CMDSN_LOWER_THAN_EXP) { |
| logout_remove = 0; |
| } else if (cmdsn_ret == CMDSN_ERROR_CANNOT_RECOVER) { |
| return iscsit_add_reject_from_cmd( |
| ISCSI_REASON_PROTOCOL_ERROR, |
| 1, 0, buf, cmd); |
| } |
| } |
| |
| return logout_remove; |
| } |
| |
| static int iscsit_handle_snack( |
| struct iscsi_conn *conn, |
| unsigned char *buf) |
| { |
| u32 unpacked_lun; |
| u64 lun; |
| struct iscsi_snack *hdr; |
| |
| hdr = (struct iscsi_snack *) buf; |
| hdr->flags &= ~ISCSI_FLAG_CMD_FINAL; |
| lun = get_unaligned_le64(&hdr->lun); |
| unpacked_lun = scsilun_to_int((struct scsi_lun *)&lun); |
| hdr->itt = be32_to_cpu(hdr->itt); |
| hdr->ttt = be32_to_cpu(hdr->ttt); |
| hdr->exp_statsn = be32_to_cpu(hdr->exp_statsn); |
| hdr->begrun = be32_to_cpu(hdr->begrun); |
| hdr->runlength = be32_to_cpu(hdr->runlength); |
| |
| pr_debug("Got ISCSI_INIT_SNACK, ITT: 0x%08x, ExpStatSN:" |
| " 0x%08x, Type: 0x%02x, BegRun: 0x%08x, RunLength: 0x%08x," |
| " CID: %hu\n", hdr->itt, hdr->exp_statsn, hdr->flags, |
| hdr->begrun, hdr->runlength, conn->cid); |
| |
| if (!conn->sess->sess_ops->ErrorRecoveryLevel) { |
| pr_err("Initiator sent SNACK request while in" |
| " ErrorRecoveryLevel=0.\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| /* |
| * SNACK_DATA and SNACK_R2T are both 0, so check which function to |
| * call from inside iscsi_send_recovery_datain_or_r2t(). |
| */ |
| switch (hdr->flags & ISCSI_FLAG_SNACK_TYPE_MASK) { |
| case 0: |
| return iscsit_handle_recovery_datain_or_r2t(conn, buf, |
| hdr->itt, hdr->ttt, hdr->begrun, hdr->runlength); |
| case ISCSI_FLAG_SNACK_TYPE_STATUS: |
| return iscsit_handle_status_snack(conn, hdr->itt, hdr->ttt, |
| hdr->begrun, hdr->runlength); |
| case ISCSI_FLAG_SNACK_TYPE_DATA_ACK: |
| return iscsit_handle_data_ack(conn, hdr->ttt, hdr->begrun, |
| hdr->runlength); |
| case ISCSI_FLAG_SNACK_TYPE_RDATA: |
| /* FIXME: Support R-Data SNACK */ |
| pr_err("R-Data SNACK Not Supported.\n"); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| default: |
| pr_err("Unknown SNACK type 0x%02x, protocol" |
| " error.\n", hdr->flags & 0x0f); |
| return iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buf, conn); |
| } |
| |
| return 0; |
| } |
| |
| static void iscsit_rx_thread_wait_for_tcp(struct iscsi_conn *conn) |
| { |
| if ((conn->sock->sk->sk_shutdown & SEND_SHUTDOWN) || |
| (conn->sock->sk->sk_shutdown & RCV_SHUTDOWN)) { |
| wait_for_completion_interruptible_timeout( |
| &conn->rx_half_close_comp, |
| ISCSI_RX_THREAD_TCP_TIMEOUT * HZ); |
| } |
| } |
| |
| static int iscsit_handle_immediate_data( |
| struct iscsi_cmd *cmd, |
| unsigned char *buf, |
| u32 length) |
| { |
| int iov_ret, rx_got = 0, rx_size = 0; |
| u32 checksum, iov_count = 0, padding = 0; |
| struct iscsi_conn *conn = cmd->conn; |
| struct kvec *iov; |
| |
| iov_ret = iscsit_map_iovec(cmd, cmd->iov_data, cmd->write_data_done, length); |
| if (iov_ret < 0) |
| return IMMEDIATE_DATA_CANNOT_RECOVER; |
| |
| rx_size = length; |
| iov_count = iov_ret; |
| iov = &cmd->iov_data[0]; |
| |
| padding = ((-length) & 3); |
| if (padding != 0) { |
| iov[iov_count].iov_base = cmd->pad_bytes; |
| iov[iov_count++].iov_len = padding; |
| rx_size += padding; |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iov[iov_count].iov_base = &checksum; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| rx_size += ISCSI_CRC_LEN; |
| } |
| |
| rx_got = rx_data(conn, &cmd->iov_data[0], iov_count, rx_size); |
| |
| iscsit_unmap_iovec(cmd); |
| |
| if (rx_got != rx_size) { |
| iscsit_rx_thread_wait_for_tcp(conn); |
| return IMMEDIATE_DATA_CANNOT_RECOVER; |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| u32 data_crc; |
| |
| data_crc = iscsit_do_crypto_hash_sg(&conn->conn_rx_hash, cmd, |
| cmd->write_data_done, length, padding, |
| cmd->pad_bytes); |
| |
| if (checksum != data_crc) { |
| pr_err("ImmediateData CRC32C DataDigest 0x%08x" |
| " does not match computed 0x%08x\n", checksum, |
| data_crc); |
| |
| if (!conn->sess->sess_ops->ErrorRecoveryLevel) { |
| pr_err("Unable to recover from" |
| " Immediate Data digest failure while" |
| " in ERL=0.\n"); |
| iscsit_add_reject_from_cmd( |
| ISCSI_REASON_DATA_DIGEST_ERROR, |
| 1, 0, buf, cmd); |
| return IMMEDIATE_DATA_CANNOT_RECOVER; |
| } else { |
| iscsit_add_reject_from_cmd( |
| ISCSI_REASON_DATA_DIGEST_ERROR, |
| 0, 0, buf, cmd); |
| return IMMEDIATE_DATA_ERL1_CRC_FAILURE; |
| } |
| } else { |
| pr_debug("Got CRC32C DataDigest 0x%08x for" |
| " %u bytes of Immediate Data\n", checksum, |
| length); |
| } |
| } |
| |
| cmd->write_data_done += length; |
| |
| if (cmd->write_data_done == cmd->data_length) { |
| spin_lock_bh(&cmd->istate_lock); |
| cmd->cmd_flags |= ICF_GOT_LAST_DATAOUT; |
| cmd->i_state = ISTATE_RECEIVED_LAST_DATAOUT; |
| spin_unlock_bh(&cmd->istate_lock); |
| } |
| |
| return IMMEDIATE_DATA_NORMAL_OPERATION; |
| } |
| |
| /* |
| * Called with sess->conn_lock held. |
| */ |
| /* #warning iscsi_build_conn_drop_async_message() only sends out on connections |
| with active network interface */ |
| static void iscsit_build_conn_drop_async_message(struct iscsi_conn *conn) |
| { |
| struct iscsi_cmd *cmd; |
| struct iscsi_conn *conn_p; |
| |
| /* |
| * Only send a Asynchronous Message on connections whos network |
| * interface is still functional. |
| */ |
| list_for_each_entry(conn_p, &conn->sess->sess_conn_list, conn_list) { |
| if (conn_p->conn_state == TARG_CONN_STATE_LOGGED_IN) { |
| iscsit_inc_conn_usage_count(conn_p); |
| break; |
| } |
| } |
| |
| if (!conn_p) |
| return; |
| |
| cmd = iscsit_allocate_cmd(conn_p, GFP_KERNEL); |
| if (!cmd) { |
| iscsit_dec_conn_usage_count(conn_p); |
| return; |
| } |
| |
| cmd->logout_cid = conn->cid; |
| cmd->iscsi_opcode = ISCSI_OP_ASYNC_EVENT; |
| cmd->i_state = ISTATE_SEND_ASYNCMSG; |
| |
| spin_lock_bh(&conn_p->cmd_lock); |
| list_add_tail(&cmd->i_list, &conn_p->conn_cmd_list); |
| spin_unlock_bh(&conn_p->cmd_lock); |
| |
| iscsit_add_cmd_to_response_queue(cmd, conn_p, cmd->i_state); |
| iscsit_dec_conn_usage_count(conn_p); |
| } |
| |
| static int iscsit_send_conn_drop_async_message( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| struct iscsi_async *hdr; |
| |
| cmd->tx_size = ISCSI_HDR_LEN; |
| cmd->iscsi_opcode = ISCSI_OP_ASYNC_EVENT; |
| |
| hdr = (struct iscsi_async *) cmd->pdu; |
| hdr->opcode = ISCSI_OP_ASYNC_EVENT; |
| hdr->flags = ISCSI_FLAG_CMD_FINAL; |
| cmd->init_task_tag = 0xFFFFFFFF; |
| cmd->targ_xfer_tag = 0xFFFFFFFF; |
| put_unaligned_be64(0xFFFFFFFFFFFFFFFFULL, &hdr->rsvd4[0]); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| hdr->async_event = ISCSI_ASYNC_MSG_DROPPING_CONNECTION; |
| hdr->param1 = cpu_to_be16(cmd->logout_cid); |
| hdr->param2 = cpu_to_be16(conn->sess->sess_ops->DefaultTime2Wait); |
| hdr->param3 = cpu_to_be16(conn->sess->sess_ops->DefaultTime2Retain); |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| cmd->tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32C HeaderDigest to" |
| " Async Message 0x%08x\n", *header_digest); |
| } |
| |
| cmd->iov_misc[0].iov_base = cmd->pdu; |
| cmd->iov_misc[0].iov_len = cmd->tx_size; |
| cmd->iov_misc_count = 1; |
| |
| pr_debug("Sending Connection Dropped Async Message StatSN:" |
| " 0x%08x, for CID: %hu on CID: %hu\n", cmd->stat_sn, |
| cmd->logout_cid, conn->cid); |
| return 0; |
| } |
| |
| static int iscsit_send_data_in( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn, |
| int *eodr) |
| { |
| int iov_ret = 0, set_statsn = 0; |
| u32 iov_count = 0, tx_size = 0; |
| struct iscsi_datain datain; |
| struct iscsi_datain_req *dr; |
| struct iscsi_data_rsp *hdr; |
| struct kvec *iov; |
| |
| memset(&datain, 0, sizeof(struct iscsi_datain)); |
| dr = iscsit_get_datain_values(cmd, &datain); |
| if (!dr) { |
| pr_err("iscsit_get_datain_values failed for ITT: 0x%08x\n", |
| cmd->init_task_tag); |
| return -1; |
| } |
| |
| /* |
| * Be paranoid and double check the logic for now. |
| */ |
| if ((datain.offset + datain.length) > cmd->data_length) { |
| pr_err("Command ITT: 0x%08x, datain.offset: %u and" |
| " datain.length: %u exceeds cmd->data_length: %u\n", |
| cmd->init_task_tag, datain.offset, datain.length, |
| cmd->data_length); |
| return -1; |
| } |
| |
| spin_lock_bh(&conn->sess->session_stats_lock); |
| conn->sess->tx_data_octets += datain.length; |
| if (conn->sess->se_sess->se_node_acl) { |
| spin_lock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| conn->sess->se_sess->se_node_acl->read_bytes += datain.length; |
| spin_unlock(&conn->sess->se_sess->se_node_acl->stats_lock); |
| } |
| spin_unlock_bh(&conn->sess->session_stats_lock); |
| /* |
| * Special case for successfully execution w/ both DATAIN |
| * and Sense Data. |
| */ |
| if ((datain.flags & ISCSI_FLAG_DATA_STATUS) && |
| (cmd->se_cmd.se_cmd_flags & SCF_TRANSPORT_TASK_SENSE)) |
| datain.flags &= ~ISCSI_FLAG_DATA_STATUS; |
| else { |
| if ((dr->dr_complete == DATAIN_COMPLETE_NORMAL) || |
| (dr->dr_complete == DATAIN_COMPLETE_CONNECTION_RECOVERY)) { |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| cmd->stat_sn = conn->stat_sn++; |
| set_statsn = 1; |
| } else if (dr->dr_complete == |
| DATAIN_COMPLETE_WITHIN_COMMAND_RECOVERY) |
| set_statsn = 1; |
| } |
| |
| hdr = (struct iscsi_data_rsp *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_SCSI_DATA_IN; |
| hdr->flags = datain.flags; |
| if (hdr->flags & ISCSI_FLAG_DATA_STATUS) { |
| if (cmd->se_cmd.se_cmd_flags & SCF_OVERFLOW_BIT) { |
| hdr->flags |= ISCSI_FLAG_DATA_OVERFLOW; |
| hdr->residual_count = cpu_to_be32(cmd->se_cmd.residual_count); |
| } else if (cmd->se_cmd.se_cmd_flags & SCF_UNDERFLOW_BIT) { |
| hdr->flags |= ISCSI_FLAG_DATA_UNDERFLOW; |
| hdr->residual_count = cpu_to_be32(cmd->se_cmd.residual_count); |
| } |
| } |
| hton24(hdr->dlength, datain.length); |
| if (hdr->flags & ISCSI_FLAG_DATA_ACK) |
| int_to_scsilun(cmd->se_cmd.orig_fe_lun, |
| (struct scsi_lun *)&hdr->lun); |
| else |
| put_unaligned_le64(0xFFFFFFFFFFFFFFFFULL, &hdr->lun); |
| |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| hdr->ttt = (hdr->flags & ISCSI_FLAG_DATA_ACK) ? |
| cpu_to_be32(cmd->targ_xfer_tag) : |
| 0xFFFFFFFF; |
| hdr->statsn = (set_statsn) ? cpu_to_be32(cmd->stat_sn) : |
| 0xFFFFFFFF; |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| hdr->datasn = cpu_to_be32(datain.data_sn); |
| hdr->offset = cpu_to_be32(datain.offset); |
| |
| iov = &cmd->iov_data[0]; |
| iov[iov_count].iov_base = cmd->pdu; |
| iov[iov_count++].iov_len = ISCSI_HDR_LEN; |
| tx_size += ISCSI_HDR_LEN; |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| |
| pr_debug("Attaching CRC32 HeaderDigest" |
| " for DataIN PDU 0x%08x\n", *header_digest); |
| } |
| |
| iov_ret = iscsit_map_iovec(cmd, &cmd->iov_data[1], datain.offset, datain.length); |
| if (iov_ret < 0) |
| return -1; |
| |
| iov_count += iov_ret; |
| tx_size += datain.length; |
| |
| cmd->padding = ((-datain.length) & 3); |
| if (cmd->padding) { |
| iov[iov_count].iov_base = cmd->pad_bytes; |
| iov[iov_count++].iov_len = cmd->padding; |
| tx_size += cmd->padding; |
| |
| pr_debug("Attaching %u padding bytes\n", |
| cmd->padding); |
| } |
| if (conn->conn_ops->DataDigest) { |
| cmd->data_crc = iscsit_do_crypto_hash_sg(&conn->conn_tx_hash, cmd, |
| datain.offset, datain.length, cmd->padding, cmd->pad_bytes); |
| |
| iov[iov_count].iov_base = &cmd->data_crc; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| |
| pr_debug("Attached CRC32C DataDigest %d bytes, crc" |
| " 0x%08x\n", datain.length+cmd->padding, cmd->data_crc); |
| } |
| |
| cmd->iov_data_count = iov_count; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Built DataIN ITT: 0x%08x, StatSN: 0x%08x," |
| " DataSN: 0x%08x, Offset: %u, Length: %u, CID: %hu\n", |
| cmd->init_task_tag, ntohl(hdr->statsn), ntohl(hdr->datasn), |
| ntohl(hdr->offset), datain.length, conn->cid); |
| |
| if (dr->dr_complete) { |
| *eodr = (cmd->se_cmd.se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) ? |
| 2 : 1; |
| iscsit_free_datain_req(cmd, dr); |
| } |
| |
| return 0; |
| } |
| |
| static int iscsit_send_logout_response( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| int niov = 0, tx_size; |
| struct iscsi_conn *logout_conn = NULL; |
| struct iscsi_conn_recovery *cr = NULL; |
| struct iscsi_session *sess = conn->sess; |
| struct kvec *iov; |
| struct iscsi_logout_rsp *hdr; |
| /* |
| * The actual shutting down of Sessions and/or Connections |
| * for CLOSESESSION and CLOSECONNECTION Logout Requests |
| * is done in scsi_logout_post_handler(). |
| */ |
| switch (cmd->logout_reason) { |
| case ISCSI_LOGOUT_REASON_CLOSE_SESSION: |
| pr_debug("iSCSI session logout successful, setting" |
| " logout response to ISCSI_LOGOUT_SUCCESS.\n"); |
| cmd->logout_response = ISCSI_LOGOUT_SUCCESS; |
| break; |
| case ISCSI_LOGOUT_REASON_CLOSE_CONNECTION: |
| if (cmd->logout_response == ISCSI_LOGOUT_CID_NOT_FOUND) |
| break; |
| /* |
| * For CLOSECONNECTION logout requests carrying |
| * a matching logout CID -> local CID, the reference |
| * for the local CID will have been incremented in |
| * iscsi_logout_closeconnection(). |
| * |
| * For CLOSECONNECTION logout requests carrying |
| * a different CID than the connection it arrived |
| * on, the connection responding to cmd->logout_cid |
| * is stopped in iscsit_logout_post_handler_diffcid(). |
| */ |
| |
| pr_debug("iSCSI CID: %hu logout on CID: %hu" |
| " successful.\n", cmd->logout_cid, conn->cid); |
| cmd->logout_response = ISCSI_LOGOUT_SUCCESS; |
| break; |
| case ISCSI_LOGOUT_REASON_RECOVERY: |
| if ((cmd->logout_response == ISCSI_LOGOUT_RECOVERY_UNSUPPORTED) || |
| (cmd->logout_response == ISCSI_LOGOUT_CLEANUP_FAILED)) |
| break; |
| /* |
| * If the connection is still active from our point of view |
| * force connection recovery to occur. |
| */ |
| logout_conn = iscsit_get_conn_from_cid_rcfr(sess, |
| cmd->logout_cid); |
| if ((logout_conn)) { |
| iscsit_connection_reinstatement_rcfr(logout_conn); |
| iscsit_dec_conn_usage_count(logout_conn); |
| } |
| |
| cr = iscsit_get_inactive_connection_recovery_entry( |
| conn->sess, cmd->logout_cid); |
| if (!cr) { |
| pr_err("Unable to locate CID: %hu for" |
| " REMOVECONNFORRECOVERY Logout Request.\n", |
| cmd->logout_cid); |
| cmd->logout_response = ISCSI_LOGOUT_CID_NOT_FOUND; |
| break; |
| } |
| |
| iscsit_discard_cr_cmds_by_expstatsn(cr, cmd->exp_stat_sn); |
| |
| pr_debug("iSCSI REMOVECONNFORRECOVERY logout" |
| " for recovery for CID: %hu on CID: %hu successful.\n", |
| cmd->logout_cid, conn->cid); |
| cmd->logout_response = ISCSI_LOGOUT_SUCCESS; |
| break; |
| default: |
| pr_err("Unknown cmd->logout_reason: 0x%02x\n", |
| cmd->logout_reason); |
| return -1; |
| } |
| |
| tx_size = ISCSI_HDR_LEN; |
| hdr = (struct iscsi_logout_rsp *)cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_LOGOUT_RSP; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| hdr->response = cmd->logout_response; |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| iov = &cmd->iov_misc[0]; |
| iov[niov].iov_base = cmd->pdu; |
| iov[niov++].iov_len = ISCSI_HDR_LEN; |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32C HeaderDigest to" |
| " Logout Response 0x%08x\n", *header_digest); |
| } |
| cmd->iov_misc_count = niov; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Sending Logout Response ITT: 0x%08x StatSN:" |
| " 0x%08x Response: 0x%02x CID: %hu on CID: %hu\n", |
| cmd->init_task_tag, cmd->stat_sn, hdr->response, |
| cmd->logout_cid, conn->cid); |
| |
| return 0; |
| } |
| |
| /* |
| * Unsolicited NOPIN, either requesting a response or not. |
| */ |
| static int iscsit_send_unsolicited_nopin( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn, |
| int want_response) |
| { |
| int tx_size = ISCSI_HDR_LEN; |
| struct iscsi_nopin *hdr; |
| |
| hdr = (struct iscsi_nopin *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_NOOP_IN; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| hdr->ttt = cpu_to_be32(cmd->targ_xfer_tag); |
| cmd->stat_sn = conn->stat_sn; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32C HeaderDigest to" |
| " NopIN 0x%08x\n", *header_digest); |
| } |
| |
| cmd->iov_misc[0].iov_base = cmd->pdu; |
| cmd->iov_misc[0].iov_len = tx_size; |
| cmd->iov_misc_count = 1; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Sending Unsolicited NOPIN TTT: 0x%08x StatSN:" |
| " 0x%08x CID: %hu\n", hdr->ttt, cmd->stat_sn, conn->cid); |
| |
| return 0; |
| } |
| |
| static int iscsit_send_nopin_response( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| int niov = 0, tx_size; |
| u32 padding = 0; |
| struct kvec *iov; |
| struct iscsi_nopin *hdr; |
| |
| tx_size = ISCSI_HDR_LEN; |
| hdr = (struct iscsi_nopin *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_NOOP_IN; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| hton24(hdr->dlength, cmd->buf_ptr_size); |
| put_unaligned_le64(0xFFFFFFFFFFFFFFFFULL, &hdr->lun); |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| hdr->ttt = cpu_to_be32(cmd->targ_xfer_tag); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| iov = &cmd->iov_misc[0]; |
| iov[niov].iov_base = cmd->pdu; |
| iov[niov++].iov_len = ISCSI_HDR_LEN; |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32C HeaderDigest" |
| " to NopIn 0x%08x\n", *header_digest); |
| } |
| |
| /* |
| * NOPOUT Ping Data is attached to struct iscsi_cmd->buf_ptr. |
| * NOPOUT DataSegmentLength is at struct iscsi_cmd->buf_ptr_size. |
| */ |
| if (cmd->buf_ptr_size) { |
| iov[niov].iov_base = cmd->buf_ptr; |
| iov[niov++].iov_len = cmd->buf_ptr_size; |
| tx_size += cmd->buf_ptr_size; |
| |
| pr_debug("Echoing back %u bytes of ping" |
| " data.\n", cmd->buf_ptr_size); |
| |
| padding = ((-cmd->buf_ptr_size) & 3); |
| if (padding != 0) { |
| iov[niov].iov_base = &cmd->pad_bytes; |
| iov[niov++].iov_len = padding; |
| tx_size += padding; |
| pr_debug("Attaching %u additional" |
| " padding bytes.\n", padding); |
| } |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| cmd->buf_ptr, cmd->buf_ptr_size, |
| padding, (u8 *)&cmd->pad_bytes, |
| (u8 *)&cmd->data_crc); |
| |
| iov[niov].iov_base = &cmd->data_crc; |
| iov[niov++].iov_len = ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attached DataDigest for %u" |
| " bytes of ping data, CRC 0x%08x\n", |
| cmd->buf_ptr_size, cmd->data_crc); |
| } |
| } |
| |
| cmd->iov_misc_count = niov; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Sending NOPIN Response ITT: 0x%08x, TTT:" |
| " 0x%08x, StatSN: 0x%08x, Length %u\n", cmd->init_task_tag, |
| cmd->targ_xfer_tag, cmd->stat_sn, cmd->buf_ptr_size); |
| |
| return 0; |
| } |
| |
| int iscsit_send_r2t( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| int tx_size = 0; |
| struct iscsi_r2t *r2t; |
| struct iscsi_r2t_rsp *hdr; |
| |
| r2t = iscsit_get_r2t_from_list(cmd); |
| if (!r2t) |
| return -1; |
| |
| hdr = (struct iscsi_r2t_rsp *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_R2T; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| int_to_scsilun(cmd->se_cmd.orig_fe_lun, |
| (struct scsi_lun *)&hdr->lun); |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| spin_lock_bh(&conn->sess->ttt_lock); |
| r2t->targ_xfer_tag = conn->sess->targ_xfer_tag++; |
| if (r2t->targ_xfer_tag == 0xFFFFFFFF) |
| r2t->targ_xfer_tag = conn->sess->targ_xfer_tag++; |
| spin_unlock_bh(&conn->sess->ttt_lock); |
| hdr->ttt = cpu_to_be32(r2t->targ_xfer_tag); |
| hdr->statsn = cpu_to_be32(conn->stat_sn); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| hdr->r2tsn = cpu_to_be32(r2t->r2t_sn); |
| hdr->data_offset = cpu_to_be32(r2t->offset); |
| hdr->data_length = cpu_to_be32(r2t->xfer_len); |
| |
| cmd->iov_misc[0].iov_base = cmd->pdu; |
| cmd->iov_misc[0].iov_len = ISCSI_HDR_LEN; |
| tx_size += ISCSI_HDR_LEN; |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| cmd->iov_misc[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 HeaderDigest for R2T" |
| " PDU 0x%08x\n", *header_digest); |
| } |
| |
| pr_debug("Built %sR2T, ITT: 0x%08x, TTT: 0x%08x, StatSN:" |
| " 0x%08x, R2TSN: 0x%08x, Offset: %u, DDTL: %u, CID: %hu\n", |
| (!r2t->recovery_r2t) ? "" : "Recovery ", cmd->init_task_tag, |
| r2t->targ_xfer_tag, ntohl(hdr->statsn), r2t->r2t_sn, |
| r2t->offset, r2t->xfer_len, conn->cid); |
| |
| cmd->iov_misc_count = 1; |
| cmd->tx_size = tx_size; |
| |
| spin_lock_bh(&cmd->r2t_lock); |
| r2t->sent_r2t = 1; |
| spin_unlock_bh(&cmd->r2t_lock); |
| |
| return 0; |
| } |
| |
| /* |
| * type 0: Normal Operation. |
| * type 1: Called from Storage Transport. |
| * type 2: Called from iscsi_task_reassign_complete_write() for |
| * connection recovery. |
| */ |
| int iscsit_build_r2ts_for_cmd( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn, |
| int type) |
| { |
| int first_r2t = 1; |
| u32 offset = 0, xfer_len = 0; |
| |
| spin_lock_bh(&cmd->r2t_lock); |
| if (cmd->cmd_flags & ICF_SENT_LAST_R2T) { |
| spin_unlock_bh(&cmd->r2t_lock); |
| return 0; |
| } |
| |
| if (conn->sess->sess_ops->DataSequenceInOrder && (type != 2)) |
| if (cmd->r2t_offset < cmd->write_data_done) |
| cmd->r2t_offset = cmd->write_data_done; |
| |
| while (cmd->outstanding_r2ts < conn->sess->sess_ops->MaxOutstandingR2T) { |
| if (conn->sess->sess_ops->DataSequenceInOrder) { |
| offset = cmd->r2t_offset; |
| |
| if (first_r2t && (type == 2)) { |
| xfer_len = ((offset + |
| (conn->sess->sess_ops->MaxBurstLength - |
| cmd->next_burst_len) > |
| cmd->data_length) ? |
| (cmd->data_length - offset) : |
| (conn->sess->sess_ops->MaxBurstLength - |
| cmd->next_burst_len)); |
| } else { |
| xfer_len = ((offset + |
| conn->sess->sess_ops->MaxBurstLength) > |
| cmd->data_length) ? |
| (cmd->data_length - offset) : |
| conn->sess->sess_ops->MaxBurstLength; |
| } |
| cmd->r2t_offset += xfer_len; |
| |
| if (cmd->r2t_offset == cmd->data_length) |
| cmd->cmd_flags |= ICF_SENT_LAST_R2T; |
| } else { |
| struct iscsi_seq *seq; |
| |
| seq = iscsit_get_seq_holder_for_r2t(cmd); |
| if (!seq) { |
| spin_unlock_bh(&cmd->r2t_lock); |
| return -1; |
| } |
| |
| offset = seq->offset; |
| xfer_len = seq->xfer_len; |
| |
| if (cmd->seq_send_order == cmd->seq_count) |
| cmd->cmd_flags |= ICF_SENT_LAST_R2T; |
| } |
| cmd->outstanding_r2ts++; |
| first_r2t = 0; |
| |
| if (iscsit_add_r2t_to_list(cmd, offset, xfer_len, 0, 0) < 0) { |
| spin_unlock_bh(&cmd->r2t_lock); |
| return -1; |
| } |
| |
| if (cmd->cmd_flags & ICF_SENT_LAST_R2T) |
| break; |
| } |
| spin_unlock_bh(&cmd->r2t_lock); |
| |
| return 0; |
| } |
| |
| static int iscsit_send_status( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| u8 iov_count = 0, recovery; |
| u32 padding = 0, tx_size = 0; |
| struct iscsi_scsi_rsp *hdr; |
| struct kvec *iov; |
| |
| recovery = (cmd->i_state != ISTATE_SEND_STATUS); |
| if (!recovery) |
| cmd->stat_sn = conn->stat_sn++; |
| |
| spin_lock_bh(&conn->sess->session_stats_lock); |
| conn->sess->rsp_pdus++; |
| spin_unlock_bh(&conn->sess->session_stats_lock); |
| |
| hdr = (struct iscsi_scsi_rsp *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_SCSI_CMD_RSP; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| if (cmd->se_cmd.se_cmd_flags & SCF_OVERFLOW_BIT) { |
| hdr->flags |= ISCSI_FLAG_CMD_OVERFLOW; |
| hdr->residual_count = cpu_to_be32(cmd->se_cmd.residual_count); |
| } else if (cmd->se_cmd.se_cmd_flags & SCF_UNDERFLOW_BIT) { |
| hdr->flags |= ISCSI_FLAG_CMD_UNDERFLOW; |
| hdr->residual_count = cpu_to_be32(cmd->se_cmd.residual_count); |
| } |
| hdr->response = cmd->iscsi_response; |
| hdr->cmd_status = cmd->se_cmd.scsi_status; |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| iov = &cmd->iov_misc[0]; |
| iov[iov_count].iov_base = cmd->pdu; |
| iov[iov_count++].iov_len = ISCSI_HDR_LEN; |
| tx_size += ISCSI_HDR_LEN; |
| |
| /* |
| * Attach SENSE DATA payload to iSCSI Response PDU |
| */ |
| if (cmd->se_cmd.sense_buffer && |
| ((cmd->se_cmd.se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) || |
| (cmd->se_cmd.se_cmd_flags & SCF_EMULATED_TASK_SENSE))) { |
| padding = -(cmd->se_cmd.scsi_sense_length) & 3; |
| hton24(hdr->dlength, cmd->se_cmd.scsi_sense_length); |
| iov[iov_count].iov_base = cmd->se_cmd.sense_buffer; |
| iov[iov_count++].iov_len = |
| (cmd->se_cmd.scsi_sense_length + padding); |
| tx_size += cmd->se_cmd.scsi_sense_length; |
| |
| if (padding) { |
| memset(cmd->se_cmd.sense_buffer + |
| cmd->se_cmd.scsi_sense_length, 0, padding); |
| tx_size += padding; |
| pr_debug("Adding %u bytes of padding to" |
| " SENSE.\n", padding); |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| cmd->se_cmd.sense_buffer, |
| (cmd->se_cmd.scsi_sense_length + padding), |
| 0, NULL, (u8 *)&cmd->data_crc); |
| |
| iov[iov_count].iov_base = &cmd->data_crc; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| |
| pr_debug("Attaching CRC32 DataDigest for" |
| " SENSE, %u bytes CRC 0x%08x\n", |
| (cmd->se_cmd.scsi_sense_length + padding), |
| cmd->data_crc); |
| } |
| |
| pr_debug("Attaching SENSE DATA: %u bytes to iSCSI" |
| " Response PDU\n", |
| cmd->se_cmd.scsi_sense_length); |
| } |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 HeaderDigest for Response" |
| " PDU 0x%08x\n", *header_digest); |
| } |
| |
| cmd->iov_misc_count = iov_count; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Built %sSCSI Response, ITT: 0x%08x, StatSN: 0x%08x," |
| " Response: 0x%02x, SAM Status: 0x%02x, CID: %hu\n", |
| (!recovery) ? "" : "Recovery ", cmd->init_task_tag, |
| cmd->stat_sn, 0x00, cmd->se_cmd.scsi_status, conn->cid); |
| |
| return 0; |
| } |
| |
| static u8 iscsit_convert_tcm_tmr_rsp(struct se_tmr_req *se_tmr) |
| { |
| switch (se_tmr->response) { |
| case TMR_FUNCTION_COMPLETE: |
| return ISCSI_TMF_RSP_COMPLETE; |
| case TMR_TASK_DOES_NOT_EXIST: |
| return ISCSI_TMF_RSP_NO_TASK; |
| case TMR_LUN_DOES_NOT_EXIST: |
| return ISCSI_TMF_RSP_NO_LUN; |
| case TMR_TASK_MGMT_FUNCTION_NOT_SUPPORTED: |
| return ISCSI_TMF_RSP_NOT_SUPPORTED; |
| case TMR_FUNCTION_AUTHORIZATION_FAILED: |
| return ISCSI_TMF_RSP_AUTH_FAILED; |
| case TMR_FUNCTION_REJECTED: |
| default: |
| return ISCSI_TMF_RSP_REJECTED; |
| } |
| } |
| |
| static int iscsit_send_task_mgt_rsp( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| struct se_tmr_req *se_tmr = cmd->se_cmd.se_tmr_req; |
| struct iscsi_tm_rsp *hdr; |
| u32 tx_size = 0; |
| |
| hdr = (struct iscsi_tm_rsp *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_SCSI_TMFUNC_RSP; |
| hdr->flags = ISCSI_FLAG_CMD_FINAL; |
| hdr->response = iscsit_convert_tcm_tmr_rsp(se_tmr); |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| cmd->iov_misc[0].iov_base = cmd->pdu; |
| cmd->iov_misc[0].iov_len = ISCSI_HDR_LEN; |
| tx_size += ISCSI_HDR_LEN; |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| cmd->iov_misc[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 HeaderDigest for Task" |
| " Mgmt Response PDU 0x%08x\n", *header_digest); |
| } |
| |
| cmd->iov_misc_count = 1; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Built Task Management Response ITT: 0x%08x," |
| " StatSN: 0x%08x, Response: 0x%02x, CID: %hu\n", |
| cmd->init_task_tag, cmd->stat_sn, hdr->response, conn->cid); |
| |
| return 0; |
| } |
| |
| static int iscsit_build_sendtargets_response(struct iscsi_cmd *cmd) |
| { |
| char *payload = NULL; |
| struct iscsi_conn *conn = cmd->conn; |
| struct iscsi_portal_group *tpg; |
| struct iscsi_tiqn *tiqn; |
| struct iscsi_tpg_np *tpg_np; |
| int buffer_len, end_of_buf = 0, len = 0, payload_len = 0; |
| unsigned char buf[256]; |
| |
| buffer_len = (conn->conn_ops->MaxRecvDataSegmentLength > 32768) ? |
| 32768 : conn->conn_ops->MaxRecvDataSegmentLength; |
| |
| memset(buf, 0, 256); |
| |
| payload = kzalloc(buffer_len, GFP_KERNEL); |
| if (!payload) { |
| pr_err("Unable to allocate memory for sendtargets" |
| " response.\n"); |
| return -ENOMEM; |
| } |
| |
| spin_lock(&tiqn_lock); |
| list_for_each_entry(tiqn, &g_tiqn_list, tiqn_list) { |
| len = sprintf(buf, "TargetName=%s", tiqn->tiqn); |
| len += 1; |
| |
| if ((len + payload_len) > buffer_len) { |
| spin_unlock(&tiqn->tiqn_tpg_lock); |
| end_of_buf = 1; |
| goto eob; |
| } |
| memcpy((void *)payload + payload_len, buf, len); |
| payload_len += len; |
| |
| spin_lock(&tiqn->tiqn_tpg_lock); |
| list_for_each_entry(tpg, &tiqn->tiqn_tpg_list, tpg_list) { |
| |
| spin_lock(&tpg->tpg_state_lock); |
| if ((tpg->tpg_state == TPG_STATE_FREE) || |
| (tpg->tpg_state == TPG_STATE_INACTIVE)) { |
| spin_unlock(&tpg->tpg_state_lock); |
| continue; |
| } |
| spin_unlock(&tpg->tpg_state_lock); |
| |
| spin_lock(&tpg->tpg_np_lock); |
| list_for_each_entry(tpg_np, &tpg->tpg_gnp_list, |
| tpg_np_list) { |
| len = sprintf(buf, "TargetAddress=" |
| "%s%s%s:%hu,%hu", |
| (tpg_np->tpg_np->np_sockaddr.ss_family == AF_INET6) ? |
| "[" : "", tpg_np->tpg_np->np_ip, |
| (tpg_np->tpg_np->np_sockaddr.ss_family == AF_INET6) ? |
| "]" : "", tpg_np->tpg_np->np_port, |
| tpg->tpgt); |
| len += 1; |
| |
| if ((len + payload_len) > buffer_len) { |
| spin_unlock(&tpg->tpg_np_lock); |
| spin_unlock(&tiqn->tiqn_tpg_lock); |
| end_of_buf = 1; |
| goto eob; |
| } |
| memcpy((void *)payload + payload_len, buf, len); |
| payload_len += len; |
| } |
| spin_unlock(&tpg->tpg_np_lock); |
| } |
| spin_unlock(&tiqn->tiqn_tpg_lock); |
| eob: |
| if (end_of_buf) |
| break; |
| } |
| spin_unlock(&tiqn_lock); |
| |
| cmd->buf_ptr = payload; |
| |
| return payload_len; |
| } |
| |
| /* |
| * FIXME: Add support for F_BIT and C_BIT when the length is longer than |
| * MaxRecvDataSegmentLength. |
| */ |
| static int iscsit_send_text_rsp( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| struct iscsi_text_rsp *hdr; |
| struct kvec *iov; |
| u32 padding = 0, tx_size = 0; |
| int text_length, iov_count = 0; |
| |
| text_length = iscsit_build_sendtargets_response(cmd); |
| if (text_length < 0) |
| return text_length; |
| |
| padding = ((-text_length) & 3); |
| if (padding != 0) { |
| memset(cmd->buf_ptr + text_length, 0, padding); |
| pr_debug("Attaching %u additional bytes for" |
| " padding.\n", padding); |
| } |
| |
| hdr = (struct iscsi_text_rsp *) cmd->pdu; |
| memset(hdr, 0, ISCSI_HDR_LEN); |
| hdr->opcode = ISCSI_OP_TEXT_RSP; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| hton24(hdr->dlength, text_length); |
| hdr->itt = cpu_to_be32(cmd->init_task_tag); |
| hdr->ttt = cpu_to_be32(cmd->targ_xfer_tag); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| |
| iscsit_increment_maxcmdsn(cmd, conn->sess); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| iov = &cmd->iov_misc[0]; |
| |
| iov[iov_count].iov_base = cmd->pdu; |
| iov[iov_count++].iov_len = ISCSI_HDR_LEN; |
| iov[iov_count].iov_base = cmd->buf_ptr; |
| iov[iov_count++].iov_len = text_length + padding; |
| |
| tx_size += (ISCSI_HDR_LEN + text_length + padding); |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 HeaderDigest for" |
| " Text Response PDU 0x%08x\n", *header_digest); |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| cmd->buf_ptr, (text_length + padding), |
| 0, NULL, (u8 *)&cmd->data_crc); |
| |
| iov[iov_count].iov_base = &cmd->data_crc; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| |
| pr_debug("Attaching DataDigest for %u bytes of text" |
| " data, CRC 0x%08x\n", (text_length + padding), |
| cmd->data_crc); |
| } |
| |
| cmd->iov_misc_count = iov_count; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Built Text Response: ITT: 0x%08x, StatSN: 0x%08x," |
| " Length: %u, CID: %hu\n", cmd->init_task_tag, cmd->stat_sn, |
| text_length, conn->cid); |
| return 0; |
| } |
| |
| static int iscsit_send_reject( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| u32 iov_count = 0, tx_size = 0; |
| struct iscsi_reject *hdr; |
| struct kvec *iov; |
| |
| hdr = (struct iscsi_reject *) cmd->pdu; |
| hdr->opcode = ISCSI_OP_REJECT; |
| hdr->flags |= ISCSI_FLAG_CMD_FINAL; |
| hton24(hdr->dlength, ISCSI_HDR_LEN); |
| cmd->stat_sn = conn->stat_sn++; |
| hdr->statsn = cpu_to_be32(cmd->stat_sn); |
| hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn); |
| hdr->max_cmdsn = cpu_to_be32(conn->sess->max_cmd_sn); |
| |
| iov = &cmd->iov_misc[0]; |
| |
| iov[iov_count].iov_base = cmd->pdu; |
| iov[iov_count++].iov_len = ISCSI_HDR_LEN; |
| iov[iov_count].iov_base = cmd->buf_ptr; |
| iov[iov_count++].iov_len = ISCSI_HDR_LEN; |
| |
| tx_size = (ISCSI_HDR_LEN + ISCSI_HDR_LEN); |
| |
| if (conn->conn_ops->HeaderDigest) { |
| u32 *header_digest = (u32 *)&cmd->pdu[ISCSI_HDR_LEN]; |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)hdr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)header_digest); |
| |
| iov[0].iov_len += ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 HeaderDigest for" |
| " REJECT PDU 0x%08x\n", *header_digest); |
| } |
| |
| if (conn->conn_ops->DataDigest) { |
| iscsit_do_crypto_hash_buf(&conn->conn_tx_hash, |
| (unsigned char *)cmd->buf_ptr, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)&cmd->data_crc); |
| |
| iov[iov_count].iov_base = &cmd->data_crc; |
| iov[iov_count++].iov_len = ISCSI_CRC_LEN; |
| tx_size += ISCSI_CRC_LEN; |
| pr_debug("Attaching CRC32 DataDigest for REJECT" |
| " PDU 0x%08x\n", cmd->data_crc); |
| } |
| |
| cmd->iov_misc_count = iov_count; |
| cmd->tx_size = tx_size; |
| |
| pr_debug("Built Reject PDU StatSN: 0x%08x, Reason: 0x%02x," |
| " CID: %hu\n", ntohl(hdr->statsn), hdr->reason, conn->cid); |
| |
| return 0; |
| } |
| |
| static void iscsit_tx_thread_wait_for_tcp(struct iscsi_conn *conn) |
| { |
| if ((conn->sock->sk->sk_shutdown & SEND_SHUTDOWN) || |
| (conn->sock->sk->sk_shutdown & RCV_SHUTDOWN)) { |
| wait_for_completion_interruptible_timeout( |
| &conn->tx_half_close_comp, |
| ISCSI_TX_THREAD_TCP_TIMEOUT * HZ); |
| } |
| } |
| |
| #ifdef CONFIG_SMP |
| |
| void iscsit_thread_get_cpumask(struct iscsi_conn *conn) |
| { |
| struct iscsi_thread_set *ts = conn->thread_set; |
| int ord, cpu; |
| /* |
| * thread_id is assigned from iscsit_global->ts_bitmap from |
| * within iscsi_thread_set.c:iscsi_allocate_thread_sets() |
| * |
| * Here we use thread_id to determine which CPU that this |
| * iSCSI connection's iscsi_thread_set will be scheduled to |
| * execute upon. |
| */ |
| ord = ts->thread_id % cpumask_weight(cpu_online_mask); |
| #if 0 |
| pr_debug(">>>>>>>>>>>>>>>>>>>> Generated ord: %d from" |
| " thread_id: %d\n", ord, ts->thread_id); |
| #endif |
| for_each_online_cpu(cpu) { |
| if (ord-- == 0) { |
| cpumask_set_cpu(cpu, conn->conn_cpumask); |
| return; |
| } |
| } |
| /* |
| * This should never be reached.. |
| */ |
| dump_stack(); |
| cpumask_setall(conn->conn_cpumask); |
| } |
| |
| static inline void iscsit_thread_check_cpumask( |
| struct iscsi_conn *conn, |
| struct task_struct *p, |
| int mode) |
| { |
| char buf[128]; |
| /* |
| * mode == 1 signals iscsi_target_tx_thread() usage. |
| * mode == 0 signals iscsi_target_rx_thread() usage. |
| */ |
| if (mode == 1) { |
| if (!conn->conn_tx_reset_cpumask) |
| return; |
| conn->conn_tx_reset_cpumask = 0; |
| } else { |
| if (!conn->conn_rx_reset_cpumask) |
| return; |
| conn->conn_rx_reset_cpumask = 0; |
| } |
| /* |
| * Update the CPU mask for this single kthread so that |
| * both TX and RX kthreads are scheduled to run on the |
| * same CPU. |
| */ |
| memset(buf, 0, 128); |
| cpumask_scnprintf(buf, 128, conn->conn_cpumask); |
| #if 0 |
| pr_debug(">>>>>>>>>>>>>> Calling set_cpus_allowed_ptr():" |
| " %s for %s\n", buf, p->comm); |
| #endif |
| set_cpus_allowed_ptr(p, conn->conn_cpumask); |
| } |
| |
| #else |
| |
| void iscsit_thread_get_cpumask(struct iscsi_conn *conn) |
| { |
| return; |
| } |
| |
| #define iscsit_thread_check_cpumask(X, Y, Z) ({}) |
| #endif /* CONFIG_SMP */ |
| |
| int iscsi_target_tx_thread(void *arg) |
| { |
| u8 state; |
| int eodr = 0; |
| int ret = 0; |
| int sent_status = 0; |
| int use_misc = 0; |
| int map_sg = 0; |
| struct iscsi_cmd *cmd = NULL; |
| struct iscsi_conn *conn; |
| struct iscsi_queue_req *qr = NULL; |
| struct se_cmd *se_cmd; |
| struct iscsi_thread_set *ts = (struct iscsi_thread_set *)arg; |
| /* |
| * Allow ourselves to be interrupted by SIGINT so that a |
| * connection recovery / failure event can be triggered externally. |
| */ |
| allow_signal(SIGINT); |
| |
| restart: |
| conn = iscsi_tx_thread_pre_handler(ts); |
| if (!conn) |
| goto out; |
| |
| eodr = map_sg = ret = sent_status = use_misc = 0; |
| |
| while (!kthread_should_stop()) { |
| /* |
| * Ensure that both TX and RX per connection kthreads |
| * are scheduled to run on the same CPU. |
| */ |
| iscsit_thread_check_cpumask(conn, current, 1); |
| |
| schedule_timeout_interruptible(MAX_SCHEDULE_TIMEOUT); |
| |
| if ((ts->status == ISCSI_THREAD_SET_RESET) || |
| signal_pending(current)) |
| goto transport_err; |
| |
| get_immediate: |
| qr = iscsit_get_cmd_from_immediate_queue(conn); |
| if (qr) { |
| atomic_set(&conn->check_immediate_queue, 0); |
| cmd = qr->cmd; |
| state = qr->state; |
| kmem_cache_free(lio_qr_cache, qr); |
| |
| spin_lock_bh(&cmd->istate_lock); |
| switch (state) { |
| case ISTATE_SEND_R2T: |
| spin_unlock_bh(&cmd->istate_lock); |
| ret = iscsit_send_r2t(cmd, conn); |
| break; |
| case ISTATE_REMOVE: |
| spin_unlock_bh(&cmd->istate_lock); |
| |
| if (cmd->data_direction == DMA_TO_DEVICE) |
| iscsit_stop_dataout_timer(cmd); |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_del(&cmd->i_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| iscsit_free_cmd(cmd); |
| goto get_immediate; |
| case ISTATE_SEND_NOPIN_WANT_RESPONSE: |
| spin_unlock_bh(&cmd->istate_lock); |
| iscsit_mod_nopin_response_timer(conn); |
| ret = iscsit_send_unsolicited_nopin(cmd, |
| conn, 1); |
| break; |
| case ISTATE_SEND_NOPIN_NO_RESPONSE: |
| spin_unlock_bh(&cmd->istate_lock); |
| ret = iscsit_send_unsolicited_nopin(cmd, |
| conn, 0); |
| break; |
| default: |
| pr_err("Unknown Opcode: 0x%02x ITT:" |
| " 0x%08x, i_state: %d on CID: %hu\n", |
| cmd->iscsi_opcode, cmd->init_task_tag, state, |
| conn->cid); |
| spin_unlock_bh(&cmd->istate_lock); |
| goto transport_err; |
| } |
| if (ret < 0) { |
| conn->tx_immediate_queue = 0; |
| goto transport_err; |
| } |
| |
| if (iscsit_send_tx_data(cmd, conn, 1) < 0) { |
| conn->tx_immediate_queue = 0; |
| iscsit_tx_thread_wait_for_tcp(conn); |
| goto transport_err; |
| } |
| |
| spin_lock_bh(&cmd->istate_lock); |
| switch (state) { |
| case ISTATE_SEND_R2T: |
| spin_unlock_bh(&cmd->istate_lock); |
| spin_lock_bh(&cmd->dataout_timeout_lock); |
| iscsit_start_dataout_timer(cmd, conn); |
| spin_unlock_bh(&cmd->dataout_timeout_lock); |
| break; |
| case ISTATE_SEND_NOPIN_WANT_RESPONSE: |
| cmd->i_state = ISTATE_SENT_NOPIN_WANT_RESPONSE; |
| spin_unlock_bh(&cmd->istate_lock); |
| break; |
| case ISTATE_SEND_NOPIN_NO_RESPONSE: |
| cmd->i_state = ISTATE_SENT_STATUS; |
| spin_unlock_bh(&cmd->istate_lock); |
| break; |
| default: |
| pr_err("Unknown Opcode: 0x%02x ITT:" |
| " 0x%08x, i_state: %d on CID: %hu\n", |
| cmd->iscsi_opcode, cmd->init_task_tag, |
| state, conn->cid); |
| spin_unlock_bh(&cmd->istate_lock); |
| goto transport_err; |
| } |
| goto get_immediate; |
| } else |
| conn->tx_immediate_queue = 0; |
| |
| get_response: |
| qr = iscsit_get_cmd_from_response_queue(conn); |
| if (qr) { |
| cmd = qr->cmd; |
| state = qr->state; |
| kmem_cache_free(lio_qr_cache, qr); |
| |
| spin_lock_bh(&cmd->istate_lock); |
| check_rsp_state: |
| switch (state) { |
| case ISTATE_SEND_DATAIN: |
| spin_unlock_bh(&cmd->istate_lock); |
| ret = iscsit_send_data_in(cmd, conn, |
| &eodr); |
| map_sg = 1; |
| break; |
| case ISTATE_SEND_STATUS: |
| case ISTATE_SEND_STATUS_RECOVERY: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_status(cmd, conn); |
| break; |
| case ISTATE_SEND_LOGOUTRSP: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_logout_response(cmd, conn); |
| break; |
| case ISTATE_SEND_ASYNCMSG: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_conn_drop_async_message( |
| cmd, conn); |
| break; |
| case ISTATE_SEND_NOPIN: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_nopin_response(cmd, conn); |
| break; |
| case ISTATE_SEND_REJECT: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_reject(cmd, conn); |
| break; |
| case ISTATE_SEND_TASKMGTRSP: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_task_mgt_rsp(cmd, conn); |
| if (ret != 0) |
| break; |
| ret = iscsit_tmr_post_handler(cmd, conn); |
| if (ret != 0) |
| iscsit_fall_back_to_erl0(conn->sess); |
| break; |
| case ISTATE_SEND_TEXTRSP: |
| spin_unlock_bh(&cmd->istate_lock); |
| use_misc = 1; |
| ret = iscsit_send_text_rsp(cmd, conn); |
| break; |
| default: |
| pr_err("Unknown Opcode: 0x%02x ITT:" |
| " 0x%08x, i_state: %d on CID: %hu\n", |
| cmd->iscsi_opcode, cmd->init_task_tag, |
| state, conn->cid); |
| spin_unlock_bh(&cmd->istate_lock); |
| goto transport_err; |
| } |
| if (ret < 0) { |
| conn->tx_response_queue = 0; |
| goto transport_err; |
| } |
| |
| se_cmd = &cmd->se_cmd; |
| |
| if (map_sg && !conn->conn_ops->IFMarker) { |
| if (iscsit_fe_sendpage_sg(cmd, conn) < 0) { |
| conn->tx_response_queue = 0; |
| iscsit_tx_thread_wait_for_tcp(conn); |
| iscsit_unmap_iovec(cmd); |
| goto transport_err; |
| } |
| } else { |
| if (iscsit_send_tx_data(cmd, conn, use_misc) < 0) { |
| conn->tx_response_queue = 0; |
| iscsit_tx_thread_wait_for_tcp(conn); |
| iscsit_unmap_iovec(cmd); |
| goto transport_err; |
| } |
| } |
| map_sg = 0; |
| iscsit_unmap_iovec(cmd); |
| |
| spin_lock_bh(&cmd->istate_lock); |
| switch (state) { |
| case ISTATE_SEND_DATAIN: |
| if (!eodr) |
| goto check_rsp_state; |
| |
| if (eodr == 1) { |
| cmd->i_state = ISTATE_SENT_LAST_DATAIN; |
| sent_status = 1; |
| eodr = use_misc = 0; |
| } else if (eodr == 2) { |
| cmd->i_state = state = |
| ISTATE_SEND_STATUS; |
| sent_status = 0; |
| eodr = use_misc = 0; |
| goto check_rsp_state; |
| } |
| break; |
| case ISTATE_SEND_STATUS: |
| use_misc = 0; |
| sent_status = 1; |
| break; |
| case ISTATE_SEND_ASYNCMSG: |
| case ISTATE_SEND_NOPIN: |
| case ISTATE_SEND_STATUS_RECOVERY: |
| case ISTATE_SEND_TEXTRSP: |
| use_misc = 0; |
| sent_status = 1; |
| break; |
| case ISTATE_SEND_REJECT: |
| use_misc = 0; |
| if (cmd->cmd_flags & ICF_REJECT_FAIL_CONN) { |
| cmd->cmd_flags &= ~ICF_REJECT_FAIL_CONN; |
| spin_unlock_bh(&cmd->istate_lock); |
| complete(&cmd->reject_comp); |
| goto transport_err; |
| } |
| complete(&cmd->reject_comp); |
| break; |
| case ISTATE_SEND_TASKMGTRSP: |
| use_misc = 0; |
| sent_status = 1; |
| break; |
| case ISTATE_SEND_LOGOUTRSP: |
| spin_unlock_bh(&cmd->istate_lock); |
| if (!iscsit_logout_post_handler(cmd, conn)) |
| goto restart; |
| spin_lock_bh(&cmd->istate_lock); |
| use_misc = 0; |
| sent_status = 1; |
| break; |
| default: |
| pr_err("Unknown Opcode: 0x%02x ITT:" |
| " 0x%08x, i_state: %d on CID: %hu\n", |
| cmd->iscsi_opcode, cmd->init_task_tag, |
| cmd->i_state, conn->cid); |
| spin_unlock_bh(&cmd->istate_lock); |
| goto transport_err; |
| } |
| |
| if (sent_status) { |
| cmd->i_state = ISTATE_SENT_STATUS; |
| sent_status = 0; |
| } |
| spin_unlock_bh(&cmd->istate_lock); |
| |
| if (atomic_read(&conn->check_immediate_queue)) |
| goto get_immediate; |
| |
| goto get_response; |
| } else |
| conn->tx_response_queue = 0; |
| } |
| |
| transport_err: |
| iscsit_take_action_for_connection_exit(conn); |
| goto restart; |
| out: |
| return 0; |
| } |
| |
| int iscsi_target_rx_thread(void *arg) |
| { |
| int ret; |
| u8 buffer[ISCSI_HDR_LEN], opcode; |
| u32 checksum = 0, digest = 0; |
| struct iscsi_conn *conn = NULL; |
| struct iscsi_thread_set *ts = (struct iscsi_thread_set *)arg; |
| struct kvec iov; |
| /* |
| * Allow ourselves to be interrupted by SIGINT so that a |
| * connection recovery / failure event can be triggered externally. |
| */ |
| allow_signal(SIGINT); |
| |
| restart: |
| conn = iscsi_rx_thread_pre_handler(ts); |
| if (!conn) |
| goto out; |
| |
| while (!kthread_should_stop()) { |
| /* |
| * Ensure that both TX and RX per connection kthreads |
| * are scheduled to run on the same CPU. |
| */ |
| iscsit_thread_check_cpumask(conn, current, 0); |
| |
| memset(buffer, 0, ISCSI_HDR_LEN); |
| memset(&iov, 0, sizeof(struct kvec)); |
| |
| iov.iov_base = buffer; |
| iov.iov_len = ISCSI_HDR_LEN; |
| |
| ret = rx_data(conn, &iov, 1, ISCSI_HDR_LEN); |
| if (ret != ISCSI_HDR_LEN) { |
| iscsit_rx_thread_wait_for_tcp(conn); |
| goto transport_err; |
| } |
| |
| /* |
| * Set conn->bad_hdr for use with REJECT PDUs. |
| */ |
| memcpy(&conn->bad_hdr, &buffer, ISCSI_HDR_LEN); |
| |
| if (conn->conn_ops->HeaderDigest) { |
| iov.iov_base = &digest; |
| iov.iov_len = ISCSI_CRC_LEN; |
| |
| ret = rx_data(conn, &iov, 1, ISCSI_CRC_LEN); |
| if (ret != ISCSI_CRC_LEN) { |
| iscsit_rx_thread_wait_for_tcp(conn); |
| goto transport_err; |
| } |
| |
| iscsit_do_crypto_hash_buf(&conn->conn_rx_hash, |
| buffer, ISCSI_HDR_LEN, |
| 0, NULL, (u8 *)&checksum); |
| |
| if (digest != checksum) { |
| pr_err("HeaderDigest CRC32C failed," |
| " received 0x%08x, computed 0x%08x\n", |
| digest, checksum); |
| /* |
| * Set the PDU to 0xff so it will intentionally |
| * hit default in the switch below. |
| */ |
| memset(buffer, 0xff, ISCSI_HDR_LEN); |
| spin_lock_bh(&conn->sess->session_stats_lock); |
| conn->sess->conn_digest_errors++; |
| spin_unlock_bh(&conn->sess->session_stats_lock); |
| } else { |
| pr_debug("Got HeaderDigest CRC32C" |
| " 0x%08x\n", checksum); |
| } |
| } |
| |
| if (conn->conn_state == TARG_CONN_STATE_IN_LOGOUT) |
| goto transport_err; |
| |
| opcode = buffer[0] & ISCSI_OPCODE_MASK; |
| |
| if (conn->sess->sess_ops->SessionType && |
| ((!(opcode & ISCSI_OP_TEXT)) || |
| (!(opcode & ISCSI_OP_LOGOUT)))) { |
| pr_err("Received illegal iSCSI Opcode: 0x%02x" |
| " while in Discovery Session, rejecting.\n", opcode); |
| iscsit_add_reject(ISCSI_REASON_PROTOCOL_ERROR, 1, |
| buffer, conn); |
| goto transport_err; |
| } |
| |
| switch (opcode) { |
| case ISCSI_OP_SCSI_CMD: |
| if (iscsit_handle_scsi_cmd(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_SCSI_DATA_OUT: |
| if (iscsit_handle_data_out(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_NOOP_OUT: |
| if (iscsit_handle_nop_out(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_SCSI_TMFUNC: |
| if (iscsit_handle_task_mgt_cmd(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_TEXT: |
| if (iscsit_handle_text_cmd(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_LOGOUT: |
| ret = iscsit_handle_logout_cmd(conn, buffer); |
| if (ret > 0) { |
| wait_for_completion_timeout(&conn->conn_logout_comp, |
| SECONDS_FOR_LOGOUT_COMP * HZ); |
| goto transport_err; |
| } else if (ret < 0) |
| goto transport_err; |
| break; |
| case ISCSI_OP_SNACK: |
| if (iscsit_handle_snack(conn, buffer) < 0) |
| goto transport_err; |
| break; |
| default: |
| pr_err("Got unknown iSCSI OpCode: 0x%02x\n", |
| opcode); |
| if (!conn->sess->sess_ops->ErrorRecoveryLevel) { |
| pr_err("Cannot recover from unknown" |
| " opcode while ERL=0, closing iSCSI connection" |
| ".\n"); |
| goto transport_err; |
| } |
| if (!conn->conn_ops->OFMarker) { |
| pr_err("Unable to recover from unknown" |
| " opcode while OFMarker=No, closing iSCSI" |
| " connection.\n"); |
| goto transport_err; |
| } |
| if (iscsit_recover_from_unknown_opcode(conn) < 0) { |
| pr_err("Unable to recover from unknown" |
| " opcode, closing iSCSI connection.\n"); |
| goto transport_err; |
| } |
| break; |
| } |
| } |
| |
| transport_err: |
| if (!signal_pending(current)) |
| atomic_set(&conn->transport_failed, 1); |
| iscsit_take_action_for_connection_exit(conn); |
| goto restart; |
| out: |
| return 0; |
| } |
| |
| static void iscsit_release_commands_from_conn(struct iscsi_conn *conn) |
| { |
| struct iscsi_cmd *cmd = NULL, *cmd_tmp = NULL; |
| struct iscsi_session *sess = conn->sess; |
| /* |
| * We expect this function to only ever be called from either RX or TX |
| * thread context via iscsit_close_connection() once the other context |
| * has been reset -> returned sleeping pre-handler state. |
| */ |
| spin_lock_bh(&conn->cmd_lock); |
| list_for_each_entry_safe(cmd, cmd_tmp, &conn->conn_cmd_list, i_list) { |
| |
| list_del(&cmd->i_list); |
| spin_unlock_bh(&conn->cmd_lock); |
| |
| iscsit_increment_maxcmdsn(cmd, sess); |
| |
| iscsit_free_cmd(cmd); |
| |
| spin_lock_bh(&conn->cmd_lock); |
| } |
| spin_unlock_bh(&conn->cmd_lock); |
| } |
| |
| static void iscsit_stop_timers_for_cmds( |
| struct iscsi_conn *conn) |
| { |
| struct iscsi_cmd *cmd; |
| |
| spin_lock_bh(&conn->cmd_lock); |
| list_for_each_entry(cmd, &conn->conn_cmd_list, i_list) { |
| if (cmd->data_direction == DMA_TO_DEVICE) |
| iscsit_stop_dataout_timer(cmd); |
| } |
| spin_unlock_bh(&conn->cmd_lock); |
| } |
| |
| int iscsit_close_connection( |
| struct iscsi_conn *conn) |
| { |
| int conn_logout = (conn->conn_state == TARG_CONN_STATE_IN_LOGOUT); |
| struct iscsi_session *sess = conn->sess; |
| |
| pr_debug("Closing iSCSI connection CID %hu on SID:" |
| " %u\n", conn->cid, sess->sid); |
| /* |
| * Always up conn_logout_comp just in case the RX Thread is sleeping |
| * and the logout response never got sent because the connection |
| * failed. |
| */ |
| complete(&conn->conn_logout_comp); |
| |
| iscsi_release_thread_set(conn); |
| |
| iscsit_stop_timers_for_cmds(conn); |
| iscsit_stop_nopin_response_timer(conn); |
| iscsit_stop_nopin_timer(conn); |
| iscsit_free_queue_reqs_for_conn(conn); |
| |
| /* |
| * During Connection recovery drop unacknowledged out of order |
| * commands for this connection, and prepare the other commands |
| * for realligence. |
| * |
| * During normal operation clear the out of order commands (but |
| * do not free the struct iscsi_ooo_cmdsn's) and release all |
| * struct iscsi_cmds. |
| */ |
| if (atomic_read(&conn->connection_recovery)) { |
| iscsit_discard_unacknowledged_ooo_cmdsns_for_conn(conn); |
| iscsit_prepare_cmds_for_realligance(conn); |
| } else { |
| iscsit_clear_ooo_cmdsns_for_conn(conn); |
| iscsit_release_commands_from_conn(conn); |
| } |
| |
| /* |
| * Handle decrementing session or connection usage count if |
| * a logout response was not able to be sent because the |
| * connection failed. Fall back to Session Recovery here. |
| */ |
| if (atomic_read(&conn->conn_logout_remove)) { |
| if (conn->conn_logout_reason == ISCSI_LOGOUT_REASON_CLOSE_SESSION) { |
| iscsit_dec_conn_usage_count(conn); |
| iscsit_dec_session_usage_count(sess); |
| } |
| if (conn->conn_logout_reason == ISCSI_LOGOUT_REASON_CLOSE_CONNECTION) |
| iscsit_dec_conn_usage_count(conn); |
| |
| atomic_set(&conn->conn_logout_remove, 0); |
| atomic_set(&sess->session_reinstatement, 0); |
| atomic_set(&sess->session_fall_back_to_erl0, 1); |
| } |
| |
| spin_lock_bh(&sess->conn_lock); |
| list_del(&conn->conn_list); |
| |
| /* |
| * Attempt to let the Initiator know this connection failed by |
| * sending an Connection Dropped Async Message on another |
| * active connection. |
| */ |
| if (atomic_read(&conn->connection_recovery)) |
| iscsit_build_conn_drop_async_message(conn); |
| |
| spin_unlock_bh(&sess->conn_lock); |
| |
| /* |
| * If connection reinstatement is being performed on this connection, |
| * up the connection reinstatement semaphore that is being blocked on |
| * in iscsit_cause_connection_reinstatement(). |
| */ |
| spin_lock_bh(&conn->state_lock); |
| if (atomic_read(&conn->sleep_on_conn_wait_comp)) { |
| spin_unlock_bh(&conn->state_lock); |
| complete(&conn->conn_wait_comp); |
| wait_for_completion(&conn->conn_post_wait_comp); |
| spin_lock_bh(&conn->state_lock); |
| } |
| |
| /* |
| * If connection reinstatement is being performed on this connection |
| * by receiving a REMOVECONNFORRECOVERY logout request, up the |
| * connection wait rcfr semaphore that is being blocked on |
| * an iscsit_connection_reinstatement_rcfr(). |
| */ |
| if (atomic_read(&conn->connection_wait_rcfr)) { |
| spin_unlock_bh(&conn->state_lock); |
| complete(&conn->conn_wait_rcfr_comp); |
| wait_for_completion(&conn->conn_post_wait_comp); |
| spin_lock_bh(&conn->state_lock); |
| } |
| atomic_set(&conn->connection_reinstatement, 1); |
| spin_unlock_bh(&conn->state_lock); |
| |
| /* |
| * If any other processes are accessing this connection pointer we |
| * must wait until they have completed. |
| */ |
| iscsit_check_conn_usage_count(conn); |
| |
| if (conn->conn_rx_hash.tfm) |
| crypto_free_hash(conn->conn_rx_hash.tfm); |
| if (conn->conn_tx_hash.tfm) |
| crypto_free_hash(conn->conn_tx_hash.tfm); |
| |
| if (conn->conn_cpumask) |
| free_cpumask_var(conn->conn_cpumask); |
| |
| kfree(conn->conn_ops); |
| conn->conn_ops = NULL; |
| |
| if (conn->sock) { |
| if (conn->conn_flags & CONNFLAG_SCTP_STRUCT_FILE) { |
| kfree(conn->sock->file); |
| conn->sock->file = NULL; |
| } |
| sock_release(conn->sock); |
| } |
| conn->thread_set = NULL; |
| |
| pr_debug("Moving to TARG_CONN_STATE_FREE.\n"); |
| conn->conn_state = TARG_CONN_STATE_FREE; |
| kfree(conn); |
| |
| spin_lock_bh(&sess->conn_lock); |
| atomic_dec(&sess->nconn); |
| pr_debug("Decremented iSCSI connection count to %hu from node:" |
| " %s\n", atomic_read(&sess->nconn), |
| sess->sess_ops->InitiatorName); |
| /* |
| * Make sure that if one connection fails in an non ERL=2 iSCSI |
| * Session that they all fail. |
| */ |
| if ((sess->sess_ops->ErrorRecoveryLevel != 2) && !conn_logout && |
| !atomic_read(&sess->session_logout)) |
| atomic_set(&sess->session_fall_back_to_erl0, 1); |
| |
| /* |
| * If this was not the last connection in the session, and we are |
| * performing session reinstatement or falling back to ERL=0, call |
| * iscsit_stop_session() without sleeping to shutdown the other |
| * active connections. |
| */ |
| if (atomic_read(&sess->nconn)) { |
| if (!atomic_read(&sess->session_reinstatement) && |
| !atomic_read(&sess->session_fall_back_to_erl0)) { |
| spin_unlock_bh(&sess->conn_lock); |
| return 0; |
| } |
| if (!atomic_read(&sess->session_stop_active)) { |
| atomic_set(&sess->session_stop_active, 1); |
| spin_unlock_bh(&sess->conn_lock); |
| iscsit_stop_session(sess, 0, 0); |
| return 0; |
| } |
| spin_unlock_bh(&sess->conn_lock); |
| return 0; |
| } |
| |
| /* |
| * If this was the last connection in the session and one of the |
| * following is occurring: |
| * |
| * Session Reinstatement is not being performed, and are falling back |
| * to ERL=0 call iscsit_close_session(). |
| * |
| * Session Logout was requested. iscsit_close_session() will be called |
| * elsewhere. |
| * |
| * Session Continuation is not being performed, start the Time2Retain |
| * handler and check if sleep_on_sess_wait_sem is active. |
| */ |
| if (!atomic_read(&sess->session_reinstatement) && |
| atomic_read(&sess->session_fall_back_to_erl0)) { |
| spin_unlock_bh(&sess->conn_lock); |
| iscsit_close_session(sess); |
| |
| return 0; |
| } else if (atomic_read(&sess->session_logout)) { |
| pr_debug("Moving to TARG_SESS_STATE_FREE.\n"); |
| sess->session_state = TARG_SESS_STATE_FREE; |
| spin_unlock_bh(&sess->conn_lock); |
| |
| if (atomic_read(&sess->sleep_on_sess_wait_comp)) |
| complete(&sess->session_wait_comp); |
| |
| return 0; |
| } else { |
| pr_debug("Moving to TARG_SESS_STATE_FAILED.\n"); |
| sess->session_state = TARG_SESS_STATE_FAILED; |
| |
| if (!atomic_read(&sess->session_continuation)) { |
| spin_unlock_bh(&sess->conn_lock); |
| iscsit_start_time2retain_handler(sess); |
| } else |
| spin_unlock_bh(&sess->conn_lock); |
| |
| if (atomic_read(&sess->sleep_on_sess_wait_comp)) |
| complete(&sess->session_wait_comp); |
| |
| return 0; |
| } |
| spin_unlock_bh(&sess->conn_lock); |
| |
| return 0; |
| } |
| |
| int iscsit_close_session(struct iscsi_session *sess) |
| { |
| struct iscsi_portal_group *tpg = ISCSI_TPG_S(sess); |
| struct se_portal_group *se_tpg = &tpg->tpg_se_tpg; |
| |
| if (atomic_read(&sess->nconn)) { |
| pr_err("%d connection(s) still exist for iSCSI session" |
| " to %s\n", atomic_read(&sess->nconn), |
| sess->sess_ops->InitiatorName); |
| BUG(); |
| } |
| |
| spin_lock_bh(&se_tpg->session_lock); |
| atomic_set(&sess->session_logout, 1); |
| atomic_set(&sess->session_reinstatement, 1); |
| iscsit_stop_time2retain_timer(sess); |
| spin_unlock_bh(&se_tpg->session_lock); |
| |
| /* |
| * transport_deregister_session_configfs() will clear the |
| * struct se_node_acl->nacl_sess pointer now as a iscsi_np process context |
| * can be setting it again with __transport_register_session() in |
| * iscsi_post_login_handler() again after the iscsit_stop_session() |
| * completes in iscsi_np context. |
| */ |
| transport_deregister_session_configfs(sess->se_sess); |
| |
| /* |
| * If any other processes are accessing this session pointer we must |
| * wait until they have completed. If we are in an interrupt (the |
| * time2retain handler) and contain and active session usage count we |
| * restart the timer and exit. |
| */ |
| if (!in_interrupt()) { |
| if (iscsit_check_session_usage_count(sess) == 1) |
| iscsit_stop_session(sess, 1, 1); |
| } else { |
| if (iscsit_check_session_usage_count(sess) == 2) { |
| atomic_set(&sess->session_logout, 0); |
| iscsit_start_time2retain_handler(sess); |
| return 0; |
| } |
| } |
| |
| transport_deregister_session(sess->se_sess); |
| |
| if (sess->sess_ops->ErrorRecoveryLevel == 2) |
| iscsit_free_connection_recovery_entires(sess); |
| |
| iscsit_free_all_ooo_cmdsns(sess); |
| |
| spin_lock_bh(&se_tpg->session_lock); |
| pr_debug("Moving to TARG_SESS_STATE_FREE.\n"); |
| sess->session_state = TARG_SESS_STATE_FREE; |
| pr_debug("Released iSCSI session from node: %s\n", |
| sess->sess_ops->InitiatorName); |
| tpg->nsessions--; |
| if (tpg->tpg_tiqn) |
| tpg->tpg_tiqn->tiqn_nsessions--; |
| |
| pr_debug("Decremented number of active iSCSI Sessions on" |
| " iSCSI TPG: %hu to %u\n", tpg->tpgt, tpg->nsessions); |
| |
| spin_lock(&sess_idr_lock); |
| idr_remove(&sess_idr, sess->session_index); |
| spin_unlock(&sess_idr_lock); |
| |
| kfree(sess->sess_ops); |
| sess->sess_ops = NULL; |
| spin_unlock_bh(&se_tpg->session_lock); |
| |
| kfree(sess); |
| return 0; |
| } |
| |
| static void iscsit_logout_post_handler_closesession( |
| struct iscsi_conn *conn) |
| { |
| struct iscsi_session *sess = conn->sess; |
| |
| iscsi_set_thread_clear(conn, ISCSI_CLEAR_TX_THREAD); |
| iscsi_set_thread_set_signal(conn, ISCSI_SIGNAL_TX_THREAD); |
| |
| atomic_set(&conn->conn_logout_remove, 0); |
| complete(&conn->conn_logout_comp); |
| |
| iscsit_dec_conn_usage_count(conn); |
| iscsit_stop_session(sess, 1, 1); |
| iscsit_dec_session_usage_count(sess); |
| iscsit_close_session(sess); |
| } |
| |
| static void iscsit_logout_post_handler_samecid( |
| struct iscsi_conn *conn) |
| { |
| iscsi_set_thread_clear(conn, ISCSI_CLEAR_TX_THREAD); |
| iscsi_set_thread_set_signal(conn, ISCSI_SIGNAL_TX_THREAD); |
| |
| atomic_set(&conn->conn_logout_remove, 0); |
| complete(&conn->conn_logout_comp); |
| |
| iscsit_cause_connection_reinstatement(conn, 1); |
| iscsit_dec_conn_usage_count(conn); |
| } |
| |
| static void iscsit_logout_post_handler_diffcid( |
| struct iscsi_conn *conn, |
| u16 cid) |
| { |
| struct iscsi_conn *l_conn; |
| struct iscsi_session *sess = conn->sess; |
| |
| if (!sess) |
| return; |
| |
| spin_lock_bh(&sess->conn_lock); |
| list_for_each_entry(l_conn, &sess->sess_conn_list, conn_list) { |
| if (l_conn->cid == cid) { |
| iscsit_inc_conn_usage_count(l_conn); |
| break; |
| } |
| } |
| spin_unlock_bh(&sess->conn_lock); |
| |
| if (!l_conn) |
| return; |
| |
| if (l_conn->sock) |
| l_conn->sock->ops->shutdown(l_conn->sock, RCV_SHUTDOWN); |
| |
| spin_lock_bh(&l_conn->state_lock); |
| pr_debug("Moving to TARG_CONN_STATE_IN_LOGOUT.\n"); |
| l_conn->conn_state = TARG_CONN_STATE_IN_LOGOUT; |
| spin_unlock_bh(&l_conn->state_lock); |
| |
| iscsit_cause_connection_reinstatement(l_conn, 1); |
| iscsit_dec_conn_usage_count(l_conn); |
| } |
| |
| /* |
| * Return of 0 causes the TX thread to restart. |
| */ |
| static int iscsit_logout_post_handler( |
| struct iscsi_cmd *cmd, |
| struct iscsi_conn *conn) |
| { |
| int ret = 0; |
| |
| switch (cmd->logout_reason) { |
| case ISCSI_LOGOUT_REASON_CLOSE_SESSION: |
| switch (cmd->logout_response) { |
| case ISCSI_LOGOUT_SUCCESS: |
| case ISCSI_LOGOUT_CLEANUP_FAILED: |
| default: |
| iscsit_logout_post_handler_closesession(conn); |
| break; |
| } |
| ret = 0; |
| break; |
| case ISCSI_LOGOUT_REASON_CLOSE_CONNECTION: |
| if (conn->cid == cmd->logout_cid) { |
| switch (cmd->logout_response) { |
| case ISCSI_LOGOUT_SUCCESS: |
| case ISCSI_LOGOUT_CLEANUP_FAILED: |
| default: |
| iscsit_logout_post_handler_samecid(conn); |
| break; |
| } |
| ret = 0; |
| } else { |
| switch (cmd->logout_response) { |
| case ISCSI_LOGOUT_SUCCESS: |
| iscsit_logout_post_handler_diffcid(conn, |
| cmd->logout_cid); |
| break; |
| case ISCSI_LOGOUT_CID_NOT_FOUND: |
| case ISCSI_LOGOUT_CLEANUP_FAILED: |
| default: |
| break; |
| } |
| ret = 1; |
| } |
| break; |
| case ISCSI_LOGOUT_REASON_RECOVERY: |
| switch (cmd->logout_response) { |
| case ISCSI_LOGOUT_SUCCESS: |
| case ISCSI_LOGOUT_CID_NOT_FOUND: |
| case ISCSI_LOGOUT_RECOVERY_UNSUPPORTED: |
| case ISCSI_LOGOUT_CLEANUP_FAILED: |
| default: |
| break; |
| } |
| ret = 1; |
| break; |
| default: |
| break; |
| |
| } |
| return ret; |
| } |
| |
| void iscsit_fail_session(struct iscsi_session *sess) |
| { |
| struct iscsi_conn *conn; |
| |
| spin_lock_bh(&sess->conn_lock); |
| list_for_each_entry(conn, &sess->sess_conn_list, conn_list) { |
| pr_debug("Moving to TARG_CONN_STATE_CLEANUP_WAIT.\n"); |
| conn->conn_state = TARG_CONN_STATE_CLEANUP_WAIT; |
| } |
| spin_unlock_bh(&sess->conn_lock); |
| |
| pr_debug("Moving to TARG_SESS_STATE_FAILED.\n"); |
| sess->session_state = TARG_SESS_STATE_FAILED; |
| } |
| |
| int iscsit_free_session(struct iscsi_session *sess) |
| { |
| u16 conn_count = atomic_read(&sess->nconn); |
| struct iscsi_conn *conn, *conn_tmp = NULL; |
| int is_last; |
| |
| spin_lock_bh(&sess->conn_lock); |
| atomic_set(&sess->sleep_on_sess_wait_comp, 1); |
| |
| list_for_each_entry_safe(conn, conn_tmp, &sess->sess_conn_list, |
| conn_list) { |
| if (conn_count == 0) |
| break; |
| |
| if (list_is_last(&conn->conn_list, &sess->sess_conn_list)) { |
| is_last = 1; |
| } else { |
| iscsit_inc_conn_usage_count(conn_tmp); |
| is_last = 0; |
| } |
| iscsit_inc_conn_usage_count(conn); |
| |
| spin_unlock_bh(&sess->conn_lock); |
| iscsit_cause_connection_reinstatement(conn, 1); |
| spin_lock_bh(&sess->conn_lock); |
| |
| iscsit_dec_conn_usage_count(conn); |
| if (is_last == 0) |
| iscsit_dec_conn_usage_count(conn_tmp); |
| |
| conn_count--; |
| } |
| |
| if (atomic_read(&sess->nconn)) { |
| spin_unlock_bh(&sess->conn_lock); |
| wait_for_completion(&sess->session_wait_comp); |
| } else |
| spin_unlock_bh(&sess->conn_lock); |
| |
| iscsit_close_session(sess); |
| return 0; |
| } |
| |
| void iscsit_stop_session( |
| struct iscsi_session *sess, |
| int session_sleep, |
| int connection_sleep) |
| { |
| u16 conn_count = atomic_read(&sess->nconn); |
| struct iscsi_conn *conn, *conn_tmp = NULL; |
| int is_last; |
| |
| spin_lock_bh(&sess->conn_lock); |
| if (session_sleep) |
| atomic_set(&sess->sleep_on_sess_wait_comp, 1); |
| |
| if (connection_sleep) { |
| list_for_each_entry_safe(conn, conn_tmp, &sess->sess_conn_list, |
| conn_list) { |
| if (conn_count == 0) |
| break; |
| |
| if (list_is_last(&conn->conn_list, &sess->sess_conn_list)) { |
| is_last = 1; |
| } else { |
| iscsit_inc_conn_usage_count(conn_tmp); |
| is_last = 0; |
| } |
| iscsit_inc_conn_usage_count(conn); |
| |
| spin_unlock_bh(&sess->conn_lock); |
| iscsit_cause_connection_reinstatement(conn, 1); |
| spin_lock_bh(&sess->conn_lock); |
| |
| iscsit_dec_conn_usage_count(conn); |
| if (is_last == 0) |
| iscsit_dec_conn_usage_count(conn_tmp); |
| conn_count--; |
| } |
| } else { |
| list_for_each_entry(conn, &sess->sess_conn_list, conn_list) |
| iscsit_cause_connection_reinstatement(conn, 0); |
| } |
| |
| if (session_sleep && atomic_read(&sess->nconn)) { |
| spin_unlock_bh(&sess->conn_lock); |
| wait_for_completion(&sess->session_wait_comp); |
| } else |
| spin_unlock_bh(&sess->conn_lock); |
| } |
| |
| int iscsit_release_sessions_for_tpg(struct iscsi_portal_group *tpg, int force) |
| { |
| struct iscsi_session *sess; |
| struct se_portal_group *se_tpg = &tpg->tpg_se_tpg; |
| struct se_session *se_sess, *se_sess_tmp; |
| int session_count = 0; |
| |
| spin_lock_bh(&se_tpg->session_lock); |
| if (tpg->nsessions && !force) { |
| spin_unlock_bh(&se_tpg->session_lock); |
| return -1; |
| } |
| |
| list_for_each_entry_safe(se_sess, se_sess_tmp, &se_tpg->tpg_sess_list, |
| sess_list) { |
| sess = (struct iscsi_session *)se_sess->fabric_sess_ptr; |
| |
| spin_lock(&sess->conn_lock); |
| if (atomic_read(&sess->session_fall_back_to_erl0) || |
| atomic_read(&sess->session_logout) || |
| (sess->time2retain_timer_flags & ISCSI_TF_EXPIRED)) { |
| spin_unlock(&sess->conn_lock); |
| continue; |
| } |
| atomic_set(&sess->session_reinstatement, 1); |
| spin_unlock(&sess->conn_lock); |
| spin_unlock_bh(&se_tpg->session_lock); |
| |
| iscsit_free_session(sess); |
| spin_lock_bh(&se_tpg->session_lock); |
| |
| session_count++; |
| } |
| spin_unlock_bh(&se_tpg->session_lock); |
| |
| pr_debug("Released %d iSCSI Session(s) from Target Portal" |
| " Group: %hu\n", session_count, tpg->tpgt); |
| return 0; |
| } |
| |
| MODULE_DESCRIPTION("iSCSI-Target Driver for mainline target infrastructure"); |
| MODULE_VERSION("4.1.x"); |
| MODULE_AUTHOR("nab@Linux-iSCSI.org"); |
| MODULE_LICENSE("GPL"); |
| |
| module_init(iscsi_target_init_module); |
| module_exit(iscsi_target_cleanup_module); |