| /* connection.c: Rx connection routines |
| * |
| * Copyright (C) 2002 Red Hat, Inc. All Rights Reserved. |
| * Written by David Howells (dhowells@redhat.com) |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License |
| * as published by the Free Software Foundation; either version |
| * 2 of the License, or (at your option) any later version. |
| */ |
| |
| #include <linux/sched.h> |
| #include <linux/slab.h> |
| #include <linux/module.h> |
| #include <rxrpc/rxrpc.h> |
| #include <rxrpc/transport.h> |
| #include <rxrpc/peer.h> |
| #include <rxrpc/connection.h> |
| #include <rxrpc/call.h> |
| #include <rxrpc/message.h> |
| #include <linux/udp.h> |
| #include <linux/ip.h> |
| #include <net/sock.h> |
| #include <asm/uaccess.h> |
| #include "internal.h" |
| |
| __RXACCT_DECL(atomic_t rxrpc_connection_count); |
| |
| LIST_HEAD(rxrpc_conns); |
| DECLARE_RWSEM(rxrpc_conns_sem); |
| unsigned long rxrpc_conn_timeout = 60 * 60; |
| |
| static void rxrpc_conn_do_timeout(struct rxrpc_connection *conn); |
| |
| static void __rxrpc_conn_timeout(rxrpc_timer_t *timer) |
| { |
| struct rxrpc_connection *conn = |
| list_entry(timer, struct rxrpc_connection, timeout); |
| |
| _debug("Rx CONN TIMEOUT [%p{u=%d}]", conn, atomic_read(&conn->usage)); |
| |
| rxrpc_conn_do_timeout(conn); |
| } |
| |
| static const struct rxrpc_timer_ops rxrpc_conn_timer_ops = { |
| .timed_out = __rxrpc_conn_timeout, |
| }; |
| |
| /*****************************************************************************/ |
| /* |
| * create a new connection record |
| */ |
| static inline int __rxrpc_create_connection(struct rxrpc_peer *peer, |
| struct rxrpc_connection **_conn) |
| { |
| struct rxrpc_connection *conn; |
| |
| _enter("%p",peer); |
| |
| /* allocate and initialise a connection record */ |
| conn = kzalloc(sizeof(struct rxrpc_connection), GFP_KERNEL); |
| if (!conn) { |
| _leave(" = -ENOMEM"); |
| return -ENOMEM; |
| } |
| |
| atomic_set(&conn->usage, 1); |
| |
| INIT_LIST_HEAD(&conn->link); |
| INIT_LIST_HEAD(&conn->id_link); |
| init_waitqueue_head(&conn->chanwait); |
| spin_lock_init(&conn->lock); |
| rxrpc_timer_init(&conn->timeout, &rxrpc_conn_timer_ops); |
| |
| do_gettimeofday(&conn->atime); |
| conn->mtu_size = 1024; |
| conn->peer = peer; |
| conn->trans = peer->trans; |
| |
| __RXACCT(atomic_inc(&rxrpc_connection_count)); |
| *_conn = conn; |
| _leave(" = 0 (%p)", conn); |
| |
| return 0; |
| } /* end __rxrpc_create_connection() */ |
| |
| /*****************************************************************************/ |
| /* |
| * create a new connection record for outgoing connections |
| */ |
| int rxrpc_create_connection(struct rxrpc_transport *trans, |
| __be16 port, |
| __be32 addr, |
| uint16_t service_id, |
| void *security, |
| struct rxrpc_connection **_conn) |
| { |
| struct rxrpc_connection *candidate, *conn; |
| struct rxrpc_peer *peer; |
| struct list_head *_p; |
| __be32 connid; |
| int ret; |
| |
| _enter("%p{%hu},%u,%hu", trans, trans->port, ntohs(port), service_id); |
| |
| /* get a peer record */ |
| ret = rxrpc_peer_lookup(trans, addr, &peer); |
| if (ret < 0) { |
| _leave(" = %d", ret); |
| return ret; |
| } |
| |
| /* allocate and initialise a connection record */ |
| ret = __rxrpc_create_connection(peer, &candidate); |
| if (ret < 0) { |
| rxrpc_put_peer(peer); |
| _leave(" = %d", ret); |
| return ret; |
| } |
| |
| /* fill in the specific bits */ |
| candidate->addr.sin_family = AF_INET; |
| candidate->addr.sin_port = port; |
| candidate->addr.sin_addr.s_addr = addr; |
| |
| candidate->in_epoch = rxrpc_epoch; |
| candidate->out_epoch = rxrpc_epoch; |
| candidate->in_clientflag = 0; |
| candidate->out_clientflag = RXRPC_CLIENT_INITIATED; |
| candidate->service_id = htons(service_id); |
| |
| /* invent a unique connection ID */ |
| write_lock(&peer->conn_idlock); |
| |
| try_next_id: |
| connid = htonl(peer->conn_idcounter & RXRPC_CIDMASK); |
| peer->conn_idcounter += RXRPC_MAXCALLS; |
| |
| list_for_each(_p, &peer->conn_idlist) { |
| conn = list_entry(_p, struct rxrpc_connection, id_link); |
| if (connid == conn->conn_id) |
| goto try_next_id; |
| if (connid > conn->conn_id) |
| break; |
| } |
| |
| _debug("selected candidate conn ID %x.%u", |
| ntohl(peer->addr.s_addr), ntohl(connid)); |
| |
| candidate->conn_id = connid; |
| list_add_tail(&candidate->id_link, _p); |
| |
| write_unlock(&peer->conn_idlock); |
| |
| /* attach to peer */ |
| candidate->peer = peer; |
| |
| write_lock(&peer->conn_lock); |
| |
| /* search the peer's transport graveyard list */ |
| spin_lock(&peer->conn_gylock); |
| list_for_each(_p, &peer->conn_graveyard) { |
| conn = list_entry(_p, struct rxrpc_connection, link); |
| if (conn->addr.sin_port == candidate->addr.sin_port && |
| conn->security_ix == candidate->security_ix && |
| conn->service_id == candidate->service_id && |
| conn->in_clientflag == 0) |
| goto found_in_graveyard; |
| } |
| spin_unlock(&peer->conn_gylock); |
| |
| /* pick the new candidate */ |
| _debug("created connection: {%08x} [out]", ntohl(candidate->conn_id)); |
| atomic_inc(&peer->conn_count); |
| conn = candidate; |
| candidate = NULL; |
| |
| make_active: |
| list_add_tail(&conn->link, &peer->conn_active); |
| write_unlock(&peer->conn_lock); |
| |
| if (candidate) { |
| write_lock(&peer->conn_idlock); |
| list_del(&candidate->id_link); |
| write_unlock(&peer->conn_idlock); |
| |
| __RXACCT(atomic_dec(&rxrpc_connection_count)); |
| kfree(candidate); |
| } |
| else { |
| down_write(&rxrpc_conns_sem); |
| list_add_tail(&conn->proc_link, &rxrpc_conns); |
| up_write(&rxrpc_conns_sem); |
| } |
| |
| *_conn = conn; |
| _leave(" = 0 (%p)", conn); |
| |
| return 0; |
| |
| /* handle resurrecting a connection from the graveyard */ |
| found_in_graveyard: |
| _debug("resurrecting connection: {%08x} [out]", ntohl(conn->conn_id)); |
| rxrpc_get_connection(conn); |
| rxrpc_krxtimod_del_timer(&conn->timeout); |
| list_del_init(&conn->link); |
| spin_unlock(&peer->conn_gylock); |
| goto make_active; |
| } /* end rxrpc_create_connection() */ |
| |
| /*****************************************************************************/ |
| /* |
| * lookup the connection for an incoming packet |
| * - create a new connection record for unrecorded incoming connections |
| */ |
| int rxrpc_connection_lookup(struct rxrpc_peer *peer, |
| struct rxrpc_message *msg, |
| struct rxrpc_connection **_conn) |
| { |
| struct rxrpc_connection *conn, *candidate = NULL; |
| struct list_head *_p; |
| struct sk_buff *pkt = msg->pkt; |
| int ret, fresh = 0; |
| __be32 x_epoch, x_connid; |
| __be16 x_port, x_servid; |
| __u32 x_secix; |
| u8 x_clflag; |
| |
| _enter("%p{{%hu}},%u,%hu", |
| peer, |
| peer->trans->port, |
| ntohs(udp_hdr(pkt)->source), |
| ntohs(msg->hdr.serviceId)); |
| |
| x_port = udp_hdr(pkt)->source; |
| x_epoch = msg->hdr.epoch; |
| x_clflag = msg->hdr.flags & RXRPC_CLIENT_INITIATED; |
| x_connid = htonl(ntohl(msg->hdr.cid) & RXRPC_CIDMASK); |
| x_servid = msg->hdr.serviceId; |
| x_secix = msg->hdr.securityIndex; |
| |
| /* [common case] search the transport's active list first */ |
| read_lock(&peer->conn_lock); |
| list_for_each(_p, &peer->conn_active) { |
| conn = list_entry(_p, struct rxrpc_connection, link); |
| if (conn->addr.sin_port == x_port && |
| conn->in_epoch == x_epoch && |
| conn->conn_id == x_connid && |
| conn->security_ix == x_secix && |
| conn->service_id == x_servid && |
| conn->in_clientflag == x_clflag) |
| goto found_active; |
| } |
| read_unlock(&peer->conn_lock); |
| |
| /* [uncommon case] not active |
| * - create a candidate for a new record if an inbound connection |
| * - only examine the graveyard for an outbound connection |
| */ |
| if (x_clflag) { |
| ret = __rxrpc_create_connection(peer, &candidate); |
| if (ret < 0) { |
| _leave(" = %d", ret); |
| return ret; |
| } |
| |
| /* fill in the specifics */ |
| candidate->addr.sin_family = AF_INET; |
| candidate->addr.sin_port = x_port; |
| candidate->addr.sin_addr.s_addr = ip_hdr(pkt)->saddr; |
| candidate->in_epoch = x_epoch; |
| candidate->out_epoch = x_epoch; |
| candidate->in_clientflag = RXRPC_CLIENT_INITIATED; |
| candidate->out_clientflag = 0; |
| candidate->conn_id = x_connid; |
| candidate->service_id = x_servid; |
| candidate->security_ix = x_secix; |
| } |
| |
| /* search the active list again, just in case it appeared whilst we |
| * were busy */ |
| write_lock(&peer->conn_lock); |
| list_for_each(_p, &peer->conn_active) { |
| conn = list_entry(_p, struct rxrpc_connection, link); |
| if (conn->addr.sin_port == x_port && |
| conn->in_epoch == x_epoch && |
| conn->conn_id == x_connid && |
| conn->security_ix == x_secix && |
| conn->service_id == x_servid && |
| conn->in_clientflag == x_clflag) |
| goto found_active_second_chance; |
| } |
| |
| /* search the transport's graveyard list */ |
| spin_lock(&peer->conn_gylock); |
| list_for_each(_p, &peer->conn_graveyard) { |
| conn = list_entry(_p, struct rxrpc_connection, link); |
| if (conn->addr.sin_port == x_port && |
| conn->in_epoch == x_epoch && |
| conn->conn_id == x_connid && |
| conn->security_ix == x_secix && |
| conn->service_id == x_servid && |
| conn->in_clientflag == x_clflag) |
| goto found_in_graveyard; |
| } |
| spin_unlock(&peer->conn_gylock); |
| |
| /* outbound connections aren't created here */ |
| if (!x_clflag) { |
| write_unlock(&peer->conn_lock); |
| _leave(" = -ENOENT"); |
| return -ENOENT; |
| } |
| |
| /* we can now add the new candidate to the list */ |
| _debug("created connection: {%08x} [in]", ntohl(candidate->conn_id)); |
| rxrpc_get_peer(peer); |
| conn = candidate; |
| candidate = NULL; |
| atomic_inc(&peer->conn_count); |
| fresh = 1; |
| |
| make_active: |
| list_add_tail(&conn->link, &peer->conn_active); |
| |
| success_uwfree: |
| write_unlock(&peer->conn_lock); |
| |
| if (candidate) { |
| write_lock(&peer->conn_idlock); |
| list_del(&candidate->id_link); |
| write_unlock(&peer->conn_idlock); |
| |
| __RXACCT(atomic_dec(&rxrpc_connection_count)); |
| kfree(candidate); |
| } |
| |
| if (fresh) { |
| down_write(&rxrpc_conns_sem); |
| list_add_tail(&conn->proc_link, &rxrpc_conns); |
| up_write(&rxrpc_conns_sem); |
| } |
| |
| success: |
| *_conn = conn; |
| _leave(" = 0 (%p)", conn); |
| return 0; |
| |
| /* handle the connection being found in the active list straight off */ |
| found_active: |
| rxrpc_get_connection(conn); |
| read_unlock(&peer->conn_lock); |
| goto success; |
| |
| /* handle resurrecting a connection from the graveyard */ |
| found_in_graveyard: |
| _debug("resurrecting connection: {%08x} [in]", ntohl(conn->conn_id)); |
| rxrpc_get_peer(peer); |
| rxrpc_get_connection(conn); |
| rxrpc_krxtimod_del_timer(&conn->timeout); |
| list_del_init(&conn->link); |
| spin_unlock(&peer->conn_gylock); |
| goto make_active; |
| |
| /* handle finding the connection on the second time through the active |
| * list */ |
| found_active_second_chance: |
| rxrpc_get_connection(conn); |
| goto success_uwfree; |
| |
| } /* end rxrpc_connection_lookup() */ |
| |
| /*****************************************************************************/ |
| /* |
| * finish using a connection record |
| * - it will be transferred to the peer's connection graveyard when refcount |
| * reaches 0 |
| */ |
| void rxrpc_put_connection(struct rxrpc_connection *conn) |
| { |
| struct rxrpc_peer *peer; |
| |
| if (!conn) |
| return; |
| |
| _enter("%p{u=%d p=%hu}", |
| conn, atomic_read(&conn->usage), ntohs(conn->addr.sin_port)); |
| |
| peer = conn->peer; |
| spin_lock(&peer->conn_gylock); |
| |
| /* sanity check */ |
| if (atomic_read(&conn->usage) <= 0) |
| BUG(); |
| |
| if (likely(!atomic_dec_and_test(&conn->usage))) { |
| spin_unlock(&peer->conn_gylock); |
| _leave(""); |
| return; |
| } |
| |
| /* move to graveyard queue */ |
| _debug("burying connection: {%08x}", ntohl(conn->conn_id)); |
| list_move_tail(&conn->link, &peer->conn_graveyard); |
| |
| rxrpc_krxtimod_add_timer(&conn->timeout, rxrpc_conn_timeout * HZ); |
| |
| spin_unlock(&peer->conn_gylock); |
| |
| rxrpc_put_peer(conn->peer); |
| |
| _leave(" [killed]"); |
| } /* end rxrpc_put_connection() */ |
| |
| /*****************************************************************************/ |
| /* |
| * free a connection record |
| */ |
| static void rxrpc_conn_do_timeout(struct rxrpc_connection *conn) |
| { |
| struct rxrpc_peer *peer; |
| |
| _enter("%p{u=%d p=%hu}", |
| conn, atomic_read(&conn->usage), ntohs(conn->addr.sin_port)); |
| |
| peer = conn->peer; |
| |
| if (atomic_read(&conn->usage) < 0) |
| BUG(); |
| |
| /* remove from graveyard if still dead */ |
| spin_lock(&peer->conn_gylock); |
| if (atomic_read(&conn->usage) == 0) { |
| list_del_init(&conn->link); |
| } |
| else { |
| conn = NULL; |
| } |
| spin_unlock(&peer->conn_gylock); |
| |
| if (!conn) { |
| _leave(""); |
| return; /* resurrected */ |
| } |
| |
| _debug("--- Destroying Connection %p{%08x} ---", |
| conn, ntohl(conn->conn_id)); |
| |
| down_write(&rxrpc_conns_sem); |
| list_del(&conn->proc_link); |
| up_write(&rxrpc_conns_sem); |
| |
| write_lock(&peer->conn_idlock); |
| list_del(&conn->id_link); |
| write_unlock(&peer->conn_idlock); |
| |
| __RXACCT(atomic_dec(&rxrpc_connection_count)); |
| kfree(conn); |
| |
| /* if the graveyard is now empty, wake up anyone waiting for that */ |
| if (atomic_dec_and_test(&peer->conn_count)) |
| wake_up(&peer->conn_gy_waitq); |
| |
| _leave(" [destroyed]"); |
| } /* end rxrpc_conn_do_timeout() */ |
| |
| /*****************************************************************************/ |
| /* |
| * clear all connection records from a peer endpoint |
| */ |
| void rxrpc_conn_clearall(struct rxrpc_peer *peer) |
| { |
| DECLARE_WAITQUEUE(myself, current); |
| |
| struct rxrpc_connection *conn; |
| int err; |
| |
| _enter("%p", peer); |
| |
| /* there shouldn't be any active conns remaining */ |
| if (!list_empty(&peer->conn_active)) |
| BUG(); |
| |
| /* manually timeout all conns in the graveyard */ |
| spin_lock(&peer->conn_gylock); |
| while (!list_empty(&peer->conn_graveyard)) { |
| conn = list_entry(peer->conn_graveyard.next, |
| struct rxrpc_connection, link); |
| err = rxrpc_krxtimod_del_timer(&conn->timeout); |
| spin_unlock(&peer->conn_gylock); |
| |
| if (err == 0) |
| rxrpc_conn_do_timeout(conn); |
| |
| spin_lock(&peer->conn_gylock); |
| } |
| spin_unlock(&peer->conn_gylock); |
| |
| /* wait for the the conn graveyard to be completely cleared */ |
| set_current_state(TASK_UNINTERRUPTIBLE); |
| add_wait_queue(&peer->conn_gy_waitq, &myself); |
| |
| while (atomic_read(&peer->conn_count) != 0) { |
| schedule(); |
| set_current_state(TASK_UNINTERRUPTIBLE); |
| } |
| |
| remove_wait_queue(&peer->conn_gy_waitq, &myself); |
| set_current_state(TASK_RUNNING); |
| |
| _leave(""); |
| } /* end rxrpc_conn_clearall() */ |
| |
| /*****************************************************************************/ |
| /* |
| * allocate and prepare a message for sending out through the transport |
| * endpoint |
| */ |
| int rxrpc_conn_newmsg(struct rxrpc_connection *conn, |
| struct rxrpc_call *call, |
| uint8_t type, |
| int dcount, |
| struct kvec diov[], |
| gfp_t alloc_flags, |
| struct rxrpc_message **_msg) |
| { |
| struct rxrpc_message *msg; |
| int loop; |
| |
| _enter("%p{%d},%p,%u", conn, ntohs(conn->addr.sin_port), call, type); |
| |
| if (dcount > 3) { |
| _leave(" = -EINVAL"); |
| return -EINVAL; |
| } |
| |
| msg = kzalloc(sizeof(struct rxrpc_message), alloc_flags); |
| if (!msg) { |
| _leave(" = -ENOMEM"); |
| return -ENOMEM; |
| } |
| |
| atomic_set(&msg->usage, 1); |
| |
| INIT_LIST_HEAD(&msg->link); |
| |
| msg->state = RXRPC_MSG_PREPARED; |
| |
| msg->hdr.epoch = conn->out_epoch; |
| msg->hdr.cid = conn->conn_id | (call ? call->chan_ix : 0); |
| msg->hdr.callNumber = call ? call->call_id : 0; |
| msg->hdr.type = type; |
| msg->hdr.flags = conn->out_clientflag; |
| msg->hdr.securityIndex = conn->security_ix; |
| msg->hdr.serviceId = conn->service_id; |
| |
| /* generate sequence numbers for data packets */ |
| if (call) { |
| switch (type) { |
| case RXRPC_PACKET_TYPE_DATA: |
| msg->seq = ++call->snd_seq_count; |
| msg->hdr.seq = htonl(msg->seq); |
| break; |
| case RXRPC_PACKET_TYPE_ACK: |
| /* ACK sequence numbers are complicated. The following |
| * may be wrong: |
| * - jumbo packet ACKs should have a seq number |
| * - normal ACKs should not |
| */ |
| default: |
| break; |
| } |
| } |
| |
| msg->dcount = dcount + 1; |
| msg->dsize = sizeof(msg->hdr); |
| msg->data[0].iov_len = sizeof(msg->hdr); |
| msg->data[0].iov_base = &msg->hdr; |
| |
| for (loop=0; loop < dcount; loop++) { |
| msg->dsize += diov[loop].iov_len; |
| msg->data[loop+1].iov_len = diov[loop].iov_len; |
| msg->data[loop+1].iov_base = diov[loop].iov_base; |
| } |
| |
| __RXACCT(atomic_inc(&rxrpc_message_count)); |
| *_msg = msg; |
| _leave(" = 0 (%p) #%d", msg, atomic_read(&rxrpc_message_count)); |
| return 0; |
| } /* end rxrpc_conn_newmsg() */ |
| |
| /*****************************************************************************/ |
| /* |
| * free a message |
| */ |
| void __rxrpc_put_message(struct rxrpc_message *msg) |
| { |
| int loop; |
| |
| _enter("%p #%d", msg, atomic_read(&rxrpc_message_count)); |
| |
| if (msg->pkt) |
| kfree_skb(msg->pkt); |
| rxrpc_put_connection(msg->conn); |
| |
| for (loop = 0; loop < 8; loop++) |
| if (test_bit(loop, &msg->dfree)) |
| kfree(msg->data[loop].iov_base); |
| |
| __RXACCT(atomic_dec(&rxrpc_message_count)); |
| kfree(msg); |
| |
| _leave(""); |
| } /* end __rxrpc_put_message() */ |
| |
| /*****************************************************************************/ |
| /* |
| * send a message out through the transport endpoint |
| */ |
| int rxrpc_conn_sendmsg(struct rxrpc_connection *conn, |
| struct rxrpc_message *msg) |
| { |
| struct msghdr msghdr; |
| int ret; |
| |
| _enter("%p{%d}", conn, ntohs(conn->addr.sin_port)); |
| |
| /* fill in some fields in the header */ |
| spin_lock(&conn->lock); |
| msg->hdr.serial = htonl(++conn->serial_counter); |
| msg->rttdone = 0; |
| spin_unlock(&conn->lock); |
| |
| /* set up the message to be transmitted */ |
| msghdr.msg_name = &conn->addr; |
| msghdr.msg_namelen = sizeof(conn->addr); |
| msghdr.msg_control = NULL; |
| msghdr.msg_controllen = 0; |
| msghdr.msg_flags = MSG_CONFIRM | MSG_DONTWAIT; |
| |
| _net("Sending message type %d of %Zd bytes to %08x:%d", |
| msg->hdr.type, |
| msg->dsize, |
| ntohl(conn->addr.sin_addr.s_addr), |
| ntohs(conn->addr.sin_port)); |
| |
| /* send the message */ |
| ret = kernel_sendmsg(conn->trans->socket, &msghdr, |
| msg->data, msg->dcount, msg->dsize); |
| if (ret < 0) { |
| msg->state = RXRPC_MSG_ERROR; |
| } else { |
| msg->state = RXRPC_MSG_SENT; |
| ret = 0; |
| |
| spin_lock(&conn->lock); |
| do_gettimeofday(&conn->atime); |
| msg->stamp = conn->atime; |
| spin_unlock(&conn->lock); |
| } |
| |
| _leave(" = %d", ret); |
| |
| return ret; |
| } /* end rxrpc_conn_sendmsg() */ |
| |
| /*****************************************************************************/ |
| /* |
| * deal with a subsequent call packet |
| */ |
| int rxrpc_conn_receive_call_packet(struct rxrpc_connection *conn, |
| struct rxrpc_call *call, |
| struct rxrpc_message *msg) |
| { |
| struct rxrpc_message *pmsg; |
| struct dst_entry *dst; |
| struct list_head *_p; |
| unsigned cix, seq; |
| int ret = 0; |
| |
| _enter("%p,%p,%p", conn, call, msg); |
| |
| if (!call) { |
| cix = ntohl(msg->hdr.cid) & RXRPC_CHANNELMASK; |
| |
| spin_lock(&conn->lock); |
| call = conn->channels[cix]; |
| |
| if (!call || call->call_id != msg->hdr.callNumber) { |
| spin_unlock(&conn->lock); |
| rxrpc_trans_immediate_abort(conn->trans, msg, -ENOENT); |
| goto out; |
| } |
| else { |
| rxrpc_get_call(call); |
| spin_unlock(&conn->lock); |
| } |
| } |
| else { |
| rxrpc_get_call(call); |
| } |
| |
| _proto("Received packet %%%u [%u] on call %hu:%u:%u", |
| ntohl(msg->hdr.serial), |
| ntohl(msg->hdr.seq), |
| ntohs(msg->hdr.serviceId), |
| ntohl(conn->conn_id), |
| ntohl(call->call_id)); |
| |
| call->pkt_rcv_count++; |
| |
| dst = msg->pkt->dst; |
| if (dst && dst->dev) |
| conn->peer->if_mtu = |
| dst->dev->mtu - dst->dev->hard_header_len; |
| |
| /* queue on the call in seq order */ |
| rxrpc_get_message(msg); |
| seq = msg->seq; |
| |
| spin_lock(&call->lock); |
| list_for_each(_p, &call->rcv_receiveq) { |
| pmsg = list_entry(_p, struct rxrpc_message, link); |
| if (pmsg->seq > seq) |
| break; |
| } |
| list_add_tail(&msg->link, _p); |
| |
| /* reset the activity timeout */ |
| call->flags |= RXRPC_CALL_RCV_PKT; |
| mod_timer(&call->rcv_timeout,jiffies + rxrpc_call_rcv_timeout * HZ); |
| |
| spin_unlock(&call->lock); |
| |
| rxrpc_krxiod_queue_call(call); |
| |
| rxrpc_put_call(call); |
| out: |
| _leave(" = %d", ret); |
| return ret; |
| } /* end rxrpc_conn_receive_call_packet() */ |
| |
| /*****************************************************************************/ |
| /* |
| * handle an ICMP error being applied to a connection |
| */ |
| void rxrpc_conn_handle_error(struct rxrpc_connection *conn, |
| int local, int errno) |
| { |
| struct rxrpc_call *calls[4]; |
| int loop; |
| |
| _enter("%p{%d},%d", conn, ntohs(conn->addr.sin_port), errno); |
| |
| /* get a ref to all my calls in one go */ |
| memset(calls, 0, sizeof(calls)); |
| spin_lock(&conn->lock); |
| |
| for (loop = 3; loop >= 0; loop--) { |
| if (conn->channels[loop]) { |
| calls[loop] = conn->channels[loop]; |
| rxrpc_get_call(calls[loop]); |
| } |
| } |
| |
| spin_unlock(&conn->lock); |
| |
| /* now kick them all */ |
| for (loop = 3; loop >= 0; loop--) { |
| if (calls[loop]) { |
| rxrpc_call_handle_error(calls[loop], local, errno); |
| rxrpc_put_call(calls[loop]); |
| } |
| } |
| |
| _leave(""); |
| } /* end rxrpc_conn_handle_error() */ |