Skip to content
Snippets Groups Projects
tcp_output.c 73.3 KiB
Newer Older
  • Learn to ignore specific revisions
  • 		skb = tp->retransmit_skb_hint;
    
    		last_lost = TCP_SKB_CB(skb)->end_seq;
    		if (after(last_lost, tp->retransmit_high))
    			last_lost = tp->retransmit_high;
    	} else {
    
    		skb = tcp_write_queue_head(sk);
    
    		last_lost = tp->snd_una;
    	}
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	tcp_for_write_queue_from(skb, sk) {
    		__u8 sacked = TCP_SKB_CB(skb)->sacked;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		if (skb == tcp_send_head(sk))
    			break;
    		/* we could do better than to assign each time */
    
    		if (hole == NULL)
    			tp->retransmit_skb_hint = skb;
    
    
    		/* Assume this retransmit will generate
    		 * only one packet for congestion window
    		 * calculation purposes.  This works because
    		 * tcp_retransmit_skb() will chop up the
    		 * packet to be MSS sized and all the
    		 * packet counting works out.
    		 */
    		if (tcp_packets_in_flight(tp) >= tp->snd_cwnd)
    			return;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		if (fwd_rexmitting) {
    begin_fwd:
    			if (!before(TCP_SKB_CB(skb)->seq, tcp_highest_sack_seq(tp)))
    				break;
    			mib_idx = LINUX_MIB_TCPFORWARDRETRANS;
    
    		} else if (!before(TCP_SKB_CB(skb)->seq, tp->retransmit_high)) {
    
    			tp->retransmit_high = last_lost;
    
    			if (!tcp_can_forward_retransmit(sk))
    				break;
    			/* Backtrack if necessary to non-L'ed skb */
    			if (hole != NULL) {
    				skb = hole;
    				hole = NULL;
    			}
    			fwd_rexmitting = 1;
    			goto begin_fwd;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		} else if (!(sacked & TCPCB_LOST)) {
    
    			if (hole == NULL && !(sacked & (TCPCB_SACKED_RETRANS|TCPCB_SACKED_ACKED)))
    
    				hole = skb;
    			continue;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    			last_lost = TCP_SKB_CB(skb)->end_seq;
    
    			if (icsk->icsk_ca_state != TCP_CA_Loss)
    				mib_idx = LINUX_MIB_TCPFASTRETRANS;
    			else
    				mib_idx = LINUX_MIB_TCPSLOWSTARTRETRANS;
    		}
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		if (sacked & (TCPCB_SACKED_ACKED|TCPCB_SACKED_RETRANS))
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			continue;
    
    
    		if (tcp_retransmit_skb(sk, skb))
    
    			return;
    		NET_INC_STATS_BH(sock_net(sk), mib_idx);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		if (skb == tcp_write_queue_head(sk))
    
    			inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
    						  inet_csk(sk)->icsk_rto,
    						  TCP_RTO_MAX);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	}
    }
    
    /* Send a fin.  The caller locks the socket for us.  This cannot be
     * allowed to fail queueing a FIN frame under any circumstances.
     */
    void tcp_send_fin(struct sock *sk)
    {
    
    	struct tcp_sock *tp = tcp_sk(sk);
    
    	struct sk_buff *skb = tcp_write_queue_tail(sk);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	int mss_now;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	/* Optimization, tack on the FIN if we have a queue of
    	 * unsent frames.  But be careful about outgoing SACKS
    	 * and IP options.
    	 */
    
    	mss_now = tcp_current_mss(sk);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	if (tcp_send_head(sk) != NULL) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_FIN;
    		TCP_SKB_CB(skb)->end_seq++;
    		tp->write_seq++;
    	} else {
    		/* Socket is locked, keep trying until memory is available. */
    		for (;;) {
    
    			skb = alloc_skb_fclone(MAX_TCP_HEADER, GFP_KERNEL);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			if (skb)
    				break;
    			yield();
    		}
    
    		/* Reserve space for headers and prepare control bits. */
    		skb_reserve(skb, MAX_TCP_HEADER);
    		/* FIN eats a sequence byte, write_seq advanced by tcp_queue_skb(). */
    
    		tcp_init_nondata_skb(skb, tp->write_seq,
    				     TCPCB_FLAG_ACK | TCPCB_FLAG_FIN);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		tcp_queue_skb(sk, skb);
    	}
    
    	__tcp_push_pending_frames(sk, mss_now, TCP_NAGLE_OFF);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    /* We get here when a process closes a file descriptor (either due to
     * an explicit close() or as a byproduct of exit()'ing) and there
     * was unread data in the receive queue.  This behavior is recommended
    
     * by RFC 2525, section 2.17.  -DaveM
    
    Linus Torvalds's avatar
    Linus Torvalds committed
     */
    
    void tcp_send_active_reset(struct sock *sk, gfp_t priority)
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    {
    	struct sk_buff *skb;
    
    	/* NOTE: No TCP options attached and we never retransmit this. */
    	skb = alloc_skb(MAX_TCP_HEADER, priority);
    	if (!skb) {
    
    		NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTFAILED);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		return;
    	}
    
    	/* Reserve space for headers and prepare control bits. */
    	skb_reserve(skb, MAX_TCP_HEADER);
    
    	tcp_init_nondata_skb(skb, tcp_acceptable_seq(sk),
    			     TCPCB_FLAG_ACK | TCPCB_FLAG_RST);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	/* Send it off. */
    	TCP_SKB_CB(skb)->when = tcp_time_stamp;
    
    	if (tcp_transmit_skb(sk, skb, 0, priority))
    
    		NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTFAILED);
    
    	TCP_INC_STATS(sock_net(sk), TCP_MIB_OUTRSTS);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    /* WARNING: This routine must only be called when we have already sent
     * a SYN packet that crossed the incoming SYN that caused this routine
     * to get called. If this assumption fails then the initial rcv_wnd
     * and rcv_wscale values will not be correct.
     */
    int tcp_send_synack(struct sock *sk)
    {
    
    	struct sk_buff *skb;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	skb = tcp_write_queue_head(sk);
    
    	if (skb == NULL || !(TCP_SKB_CB(skb)->flags & TCPCB_FLAG_SYN)) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		printk(KERN_DEBUG "tcp_send_synack: wrong queue state\n");
    		return -EFAULT;
    	}
    
    	if (!(TCP_SKB_CB(skb)->flags & TCPCB_FLAG_ACK)) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		if (skb_cloned(skb)) {
    			struct sk_buff *nskb = skb_copy(skb, GFP_ATOMIC);
    			if (nskb == NULL)
    				return -ENOMEM;
    
    			tcp_unlink_write_queue(skb, sk);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			skb_header_release(nskb);
    
    			__tcp_add_write_queue_head(sk, nskb);
    
    			sk_wmem_free_skb(sk, skb);
    			sk->sk_wmem_queued += nskb->truesize;
    			sk_mem_charge(sk, nskb->truesize);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			skb = nskb;
    		}
    
    		TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_ACK;
    		TCP_ECN_send_synack(tcp_sk(sk), skb);
    	}
    	TCP_SKB_CB(skb)->when = tcp_time_stamp;
    
    	return tcp_transmit_skb(sk, skb, 1, GFP_ATOMIC);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    /*
     * Prepare a SYN-ACK.
     */
    
    struct sk_buff *tcp_make_synack(struct sock *sk, struct dst_entry *dst,
    				struct request_sock *req)
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    {
    
    	struct inet_request_sock *ireq = inet_rsk(req);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	struct tcp_sock *tp = tcp_sk(sk);
    	struct tcphdr *th;
    	int tcp_header_size;
    
    Adam Langley's avatar
    Adam Langley committed
    	struct tcp_out_options opts;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	struct sk_buff *skb;
    
    	struct tcp_md5sig_key *md5;
    	__u8 *md5_hash_location;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    	skb = sock_wmalloc(sk, MAX_TCP_HEADER + 15, 1, GFP_ATOMIC);
    	if (skb == NULL)
    		return NULL;
    
    	/* Reserve space for headers. */
    	skb_reserve(skb, MAX_TCP_HEADER);
    
    	skb->dst = dst_clone(dst);
    
    
    	mss = dst_metric(dst, RTAX_ADVMSS);
    	if (tp->rx_opt.user_mss && tp->rx_opt.user_mss < mss)
    		mss = tp->rx_opt.user_mss;
    
    
    Adam Langley's avatar
    Adam Langley committed
    	if (req->rcv_wnd == 0) { /* ignored for retransmitted syns */
    		__u8 rcv_wscale;
    		/* Set this up on the first call only */
    		req->window_clamp = tp->window_clamp ? : dst_metric(dst, RTAX_WINDOW);
    		/* tcp_full_space because it is guaranteed to be the first packet */
    		tcp_select_initial_window(tcp_full_space(sk),
    
    			mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0),
    
    Adam Langley's avatar
    Adam Langley committed
    			&req->rcv_wnd,
    			&req->window_clamp,
    			ireq->wscale_ok,
    			&rcv_wscale);
    		ireq->rcv_wscale = rcv_wscale;
    	}
    
    	memset(&opts, 0, sizeof(opts));
    
    #ifdef CONFIG_SYN_COOKIES
    	if (unlikely(req->cookie_ts))
    		TCP_SKB_CB(skb)->when = cookie_init_timestamp(req);
    	else
    #endif
    
    Adam Langley's avatar
    Adam Langley committed
    	TCP_SKB_CB(skb)->when = tcp_time_stamp;
    
    	tcp_header_size = tcp_synack_options(sk, req, mss,
    
    Adam Langley's avatar
    Adam Langley committed
    					     skb, &opts, &md5) +
    			  sizeof(struct tcphdr);
    
    	skb_push(skb, tcp_header_size);
    	skb_reset_transport_header(skb);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	memset(th, 0, sizeof(struct tcphdr));
    	th->syn = 1;
    	th->ack = 1;
    	TCP_ECN_make_synack(req, th);
    
    	th->source = ireq->loc_port;
    
    	/* Setting of flags are superfluous here for callers (and ECE is
    	 * not even correctly set)
    	 */
    	tcp_init_nondata_skb(skb, tcp_rsk(req)->snt_isn,
    			     TCPCB_FLAG_SYN | TCPCB_FLAG_ACK);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	th->seq = htonl(TCP_SKB_CB(skb)->seq);
    
    	th->ack_seq = htonl(tcp_rsk(req)->rcv_isn + 1);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    	/* RFC1323: The window in SYN & SYN/ACK segments is never scaled. */
    
    	th->window = htons(min(req->rcv_wnd, 65535U));
    
    Adam Langley's avatar
    Adam Langley committed
    	tcp_options_write((__be32 *)(th + 1), tp, &opts, &md5_hash_location);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	th->doff = (tcp_header_size >> 2);
    
    	TCP_INC_STATS(sock_net(sk), TCP_MIB_OUTSEGS);
    
    
    #ifdef CONFIG_TCP_MD5SIG
    	/* Okay, we have all we need - do the md5 hash if needed */
    	if (md5) {
    		tp->af_specific->calc_md5_hash(md5_hash_location,
    
    					       md5, NULL, req, skb);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	return skb;
    }
    
    
    Linus Torvalds's avatar
    Linus Torvalds committed
     * Do all connect socket setups that can be done AF independent.
    
    Stephen Hemminger's avatar
    Stephen Hemminger committed
    static void tcp_connect_init(struct sock *sk)
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    {
    	struct dst_entry *dst = __sk_dst_get(sk);
    	struct tcp_sock *tp = tcp_sk(sk);
    	__u8 rcv_wscale;
    
    	/* We'll fix this up when we get a response from the other end.
    	 * See tcp_input.c:tcp_rcv_state_process case TCP_SYN_SENT.
    	 */
    	tp->tcp_header_len = sizeof(struct tcphdr) +
    		(sysctl_tcp_timestamps ? TCPOLEN_TSTAMP_ALIGNED : 0);
    
    
    #ifdef CONFIG_TCP_MD5SIG
    	if (tp->af_specific->md5_lookup(sk, sk) != NULL)
    		tp->tcp_header_len += TCPOLEN_MD5SIG_ALIGNED;
    #endif
    
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	/* If user gave his TCP_MAXSEG, record it to clamp */
    	if (tp->rx_opt.user_mss)
    		tp->rx_opt.mss_clamp = tp->rx_opt.user_mss;
    	tp->max_window = 0;
    
    John Heffner's avatar
    John Heffner committed
    	tcp_mtup_init(sk);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tcp_sync_mss(sk, dst_mtu(dst));
    
    	if (!tp->window_clamp)
    		tp->window_clamp = dst_metric(dst, RTAX_WINDOW);
    	tp->advmss = dst_metric(dst, RTAX_ADVMSS);
    
    	if (tp->rx_opt.user_mss && tp->rx_opt.user_mss < tp->advmss)
    		tp->advmss = tp->rx_opt.user_mss;
    
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tcp_initialize_rcv_mss(sk);
    
    	tcp_select_initial_window(tcp_full_space(sk),
    				  tp->advmss - (tp->rx_opt.ts_recent_stamp ? tp->tcp_header_len - sizeof(struct tcphdr) : 0),
    				  &tp->rcv_wnd,
    				  &tp->window_clamp,
    				  sysctl_tcp_window_scaling,
    				  &rcv_wscale);
    
    	tp->rx_opt.rcv_wscale = rcv_wscale;
    	tp->rcv_ssthresh = tp->rcv_wnd;
    
    	sk->sk_err = 0;
    	sock_reset_flag(sk, SOCK_DONE);
    	tp->snd_wnd = 0;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tp->snd_una = tp->write_seq;
    	tp->snd_sml = tp->write_seq;
    
    	tp->snd_up = tp->write_seq;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tp->rcv_nxt = 0;
    	tp->rcv_wup = 0;
    	tp->copied_seq = 0;
    
    
    	inet_csk(sk)->icsk_rto = TCP_TIMEOUT_INIT;
    	inet_csk(sk)->icsk_retransmits = 0;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tcp_clear_retrans(tp);
    }
    
    /*
     * Build a SYN and send it off.
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    int tcp_connect(struct sock *sk)
    {
    	struct tcp_sock *tp = tcp_sk(sk);
    	struct sk_buff *buff;
    
    	tcp_connect_init(sk);
    
    
    	buff = alloc_skb_fclone(MAX_TCP_HEADER + 15, sk->sk_allocation);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	if (unlikely(buff == NULL))
    		return -ENOBUFS;
    
    	/* Reserve space for headers. */
    	skb_reserve(buff, MAX_TCP_HEADER);
    
    
    	tp->snd_nxt = tp->write_seq;
    
    	tcp_init_nondata_skb(buff, tp->write_seq++, TCPCB_FLAG_SYN);
    	TCP_ECN_send_syn(sk, buff);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    	/* Send it off. */
    	TCP_SKB_CB(buff)->when = tcp_time_stamp;
    	tp->retrans_stamp = TCP_SKB_CB(buff)->when;
    	skb_header_release(buff);
    
    	__tcp_add_write_queue_tail(sk, buff);
    
    	sk->sk_wmem_queued += buff->truesize;
    	sk_mem_charge(sk, buff->truesize);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	tp->packets_out += tcp_skb_pcount(buff);
    
    	tcp_transmit_skb(sk, buff, 1, GFP_KERNEL);
    
    
    	/* We change tp->snd_nxt after the tcp_transmit_skb() call
    	 * in order to make this packet get counted in tcpOutSegs.
    	 */
    	tp->snd_nxt = tp->write_seq;
    	tp->pushed_seq = tp->write_seq;
    
    	TCP_INC_STATS(sock_net(sk), TCP_MIB_ACTIVEOPENS);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    	/* Timer for repeating the SYN until an answer. */
    
    	inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
    				  inet_csk(sk)->icsk_rto, TCP_RTO_MAX);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	return 0;
    }
    
    /* Send out a delayed ack, the caller does the policy checking
     * to see if we should even be here.  See tcp_input.c:tcp_ack_snd_check()
     * for details.
     */
    void tcp_send_delayed_ack(struct sock *sk)
    {
    
    	struct inet_connection_sock *icsk = inet_csk(sk);
    	int ato = icsk->icsk_ack.ato;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	unsigned long timeout;
    
    	if (ato > TCP_DELACK_MIN) {
    
    		const struct tcp_sock *tp = tcp_sk(sk);
    
    		int max_ato = HZ / 2;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    		if (icsk->icsk_ack.pingpong ||
    		    (icsk->icsk_ack.pending & ICSK_ACK_PUSHED))
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			max_ato = TCP_DELACK_MAX;
    
    		/* Slow path, intersegment interval is "high". */
    
    		/* If some rtt estimate is known, use it to bound delayed ack.
    
    		 * Do not use inet_csk(sk)->icsk_rto here, use results of rtt measurements
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		 * directly.
    		 */
    		if (tp->srtt) {
    
    			int rtt = max(tp->srtt >> 3, TCP_DELACK_MIN);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    			if (rtt < max_ato)
    				max_ato = rtt;
    		}
    
    		ato = min(ato, max_ato);
    	}
    
    	/* Stay within the limit we were given */
    	timeout = jiffies + ato;
    
    	/* Use new timeout only if there wasn't a older one earlier. */
    
    	if (icsk->icsk_ack.pending & ICSK_ACK_TIMER) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		/* If delack timer was blocked or is about to expire,
    		 * send ACK now.
    		 */
    
    		if (icsk->icsk_ack.blocked ||
    		    time_before_eq(icsk->icsk_ack.timeout, jiffies + (ato >> 2))) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			tcp_send_ack(sk);
    			return;
    		}
    
    
    		if (!time_before(timeout, icsk->icsk_ack.timeout))
    			timeout = icsk->icsk_ack.timeout;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	}
    
    	icsk->icsk_ack.pending |= ICSK_ACK_SCHED | ICSK_ACK_TIMER;
    	icsk->icsk_ack.timeout = timeout;
    	sk_reset_timer(sk, &icsk->icsk_delack_timer, timeout);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    /* This routine sends an ack and also updates the window. */
    void tcp_send_ack(struct sock *sk)
    {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	/* If we have been reset, we may not send again. */
    	if (sk->sk_state == TCP_CLOSE)
    		return;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	/* We are not putting this on the write queue, so
    	 * tcp_transmit_skb() will set the ownership to this
    	 * sock.
    	 */
    	buff = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC);
    	if (buff == NULL) {
    		inet_csk_schedule_ack(sk);
    		inet_csk(sk)->icsk_ack.ato = TCP_ATO_MIN;
    		inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK,
    					  TCP_DELACK_MAX, TCP_RTO_MAX);
    		return;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	}
    
    
    	/* Reserve space for headers and prepare control bits. */
    	skb_reserve(buff, MAX_TCP_HEADER);
    
    	tcp_init_nondata_skb(buff, tcp_acceptable_seq(sk), TCPCB_FLAG_ACK);
    
    
    	/* Send it off, this clears delayed acks for us. */
    	TCP_SKB_CB(buff)->when = tcp_time_stamp;
    	tcp_transmit_skb(sk, buff, 0, GFP_ATOMIC);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    /* This routine sends a packet with an out of date sequence
     * number. It assumes the other end will try to ack it.
     *
     * Question: what should we make while urgent mode?
     * 4.4BSD forces sending single byte of data. We cannot send
     * out of window data, because we have SND.NXT==SND.MAX...
     *
     * Current solution: to send TWO zero-length segments in urgent mode:
     * one is with SEG.SEQ=SND.UNA to deliver urgent pointer, another is
     * out-of-date with SND.UNA-1 to probe window.
     */
    static int tcp_xmit_probe_skb(struct sock *sk, int urgent)
    {
    	struct tcp_sock *tp = tcp_sk(sk);
    	struct sk_buff *skb;
    
    	/* We don't queue it, tcp_transmit_skb() sets ownership. */
    	skb = alloc_skb(MAX_TCP_HEADER, GFP_ATOMIC);
    
    	if (skb == NULL)
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		return -1;
    
    	/* Reserve space for headers and set control bits. */
    	skb_reserve(skb, MAX_TCP_HEADER);
    	/* Use a previous sequence.  This should cause the other
    	 * end to send an ack.  Don't queue or clone SKB, just
    	 * send it.
    	 */
    
    	tcp_init_nondata_skb(skb, tp->snd_una - !urgent, TCPCB_FLAG_ACK);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	TCP_SKB_CB(skb)->when = tcp_time_stamp;
    
    	return tcp_transmit_skb(sk, skb, 0, GFP_ATOMIC);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    }
    
    int tcp_write_wakeup(struct sock *sk)
    {
    
    	struct tcp_sock *tp = tcp_sk(sk);
    	struct sk_buff *skb;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    
    
    	if (sk->sk_state == TCP_CLOSE)
    		return -1;
    
    	if ((skb = tcp_send_head(sk)) != NULL &&
    	    before(TCP_SKB_CB(skb)->seq, tcp_wnd_end(tp))) {
    		int err;
    
    		unsigned int mss = tcp_current_mss(sk);
    
    		unsigned int seg_size = tcp_wnd_end(tp) - TCP_SKB_CB(skb)->seq;
    
    		if (before(tp->pushed_seq, TCP_SKB_CB(skb)->end_seq))
    			tp->pushed_seq = TCP_SKB_CB(skb)->end_seq;
    
    		/* We are probing the opening of a window
    		 * but the window size is != 0
    		 * must have been a result SWS avoidance ( sender )
    		 */
    		if (seg_size < TCP_SKB_CB(skb)->end_seq - TCP_SKB_CB(skb)->seq ||
    		    skb->len > mss) {
    			seg_size = min(seg_size, mss);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    			TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH;
    
    			if (tcp_fragment(sk, skb, seg_size, mss))
    				return -1;
    		} else if (!tcp_skb_pcount(skb))
    			tcp_set_skb_tso_segs(sk, skb, mss);
    
    		TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH;
    		TCP_SKB_CB(skb)->when = tcp_time_stamp;
    		err = tcp_transmit_skb(sk, skb, 1, GFP_ATOMIC);
    		if (!err)
    			tcp_event_new_data_sent(sk, skb);
    		return err;
    	} else {
    
    		if (between(tp->snd_up, tp->snd_una + 1, tp->snd_una + 0xFFFF))
    
    			tcp_xmit_probe_skb(sk, 1);
    		return tcp_xmit_probe_skb(sk, 0);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	}
    }
    
    /* A window probe timeout has occurred.  If window is not closed send
     * a partial packet else a zero probe.
     */
    void tcp_send_probe0(struct sock *sk)
    {
    
    	struct inet_connection_sock *icsk = inet_csk(sk);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	struct tcp_sock *tp = tcp_sk(sk);
    	int err;
    
    	err = tcp_write_wakeup(sk);
    
    
    	if (tp->packets_out || !tcp_send_head(sk)) {
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		/* Cancel probe timer, if it is not required. */
    
    		icsk->icsk_backoff = 0;
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		return;
    	}
    
    	if (err <= 0) {
    
    		if (icsk->icsk_backoff < sysctl_tcp_retries2)
    			icsk->icsk_backoff++;
    
    		inet_csk_reset_xmit_timer(sk, ICSK_TIME_PROBE0,
    
    					  min(icsk->icsk_rto << icsk->icsk_backoff, TCP_RTO_MAX),
    					  TCP_RTO_MAX);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    	} else {
    		/* If packet was not sent due to local congestion,
    
    		 * do not backoff and do not remember icsk_probes_out.
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    		 * Let local senders to fight for local resources.
    		 *
    		 * Use accumulated backoff yet.
    		 */
    
    		if (!icsk->icsk_probes_out)
    			icsk->icsk_probes_out = 1;
    
    		inet_csk_reset_xmit_timer(sk, ICSK_TIME_PROBE0,
    
    					  min(icsk->icsk_rto << icsk->icsk_backoff,
    
    EXPORT_SYMBOL(tcp_select_initial_window);
    
    Linus Torvalds's avatar
    Linus Torvalds committed
    EXPORT_SYMBOL(tcp_connect);
    EXPORT_SYMBOL(tcp_make_synack);
    EXPORT_SYMBOL(tcp_simple_retransmit);
    EXPORT_SYMBOL(tcp_sync_mss);
    
    John Heffner's avatar
    John Heffner committed
    EXPORT_SYMBOL(tcp_mtup_init);