Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net
[cascardo/linux.git] / net / sctp / outqueue.c
index 72e54a4..5825853 100644 (file)
@@ -68,7 +68,7 @@ static void sctp_mark_missing(struct sctp_outq *q,
 
 static void sctp_generate_fwdtsn(struct sctp_outq *q, __u32 sack_ctsn);
 
-static int sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp);
+static void sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp);
 
 /* Add data to the front of the queue. */
 static inline void sctp_outq_head_data(struct sctp_outq *q,
@@ -285,10 +285,9 @@ void sctp_outq_free(struct sctp_outq *q)
 }
 
 /* Put a new chunk in an sctp_outq.  */
-int sctp_outq_tail(struct sctp_outq *q, struct sctp_chunk *chunk, gfp_t gfp)
+void sctp_outq_tail(struct sctp_outq *q, struct sctp_chunk *chunk, gfp_t gfp)
 {
        struct net *net = sock_net(q->asoc->base.sk);
-       int error = 0;
 
        pr_debug("%s: outq:%p, chunk:%p[%s]\n", __func__, q, chunk,
                 chunk && chunk->chunk_hdr ?
@@ -299,54 +298,26 @@ int sctp_outq_tail(struct sctp_outq *q, struct sctp_chunk *chunk, gfp_t gfp)
         * immediately.
         */
        if (sctp_chunk_is_data(chunk)) {
-               /* Is it OK to queue data chunks?  */
-               /* From 9. Termination of Association
-                *
-                * When either endpoint performs a shutdown, the
-                * association on each peer will stop accepting new
-                * data from its user and only deliver data in queue
-                * at the time of sending or receiving the SHUTDOWN
-                * chunk.
-                */
-               switch (q->asoc->state) {
-               case SCTP_STATE_CLOSED:
-               case SCTP_STATE_SHUTDOWN_PENDING:
-               case SCTP_STATE_SHUTDOWN_SENT:
-               case SCTP_STATE_SHUTDOWN_RECEIVED:
-               case SCTP_STATE_SHUTDOWN_ACK_SENT:
-                       /* Cannot send after transport endpoint shutdown */
-                       error = -ESHUTDOWN;
-                       break;
-
-               default:
-                       pr_debug("%s: outqueueing: outq:%p, chunk:%p[%s])\n",
-                                __func__, q, chunk, chunk && chunk->chunk_hdr ?
-                                sctp_cname(SCTP_ST_CHUNK(chunk->chunk_hdr->type)) :
-                                "illegal chunk");
-
-                       sctp_chunk_hold(chunk);
-                       sctp_outq_tail_data(q, chunk);
-                       if (chunk->asoc->prsctp_enable &&
-                           SCTP_PR_PRIO_ENABLED(chunk->sinfo.sinfo_flags))
-                               chunk->asoc->sent_cnt_removable++;
-                       if (chunk->chunk_hdr->flags & SCTP_DATA_UNORDERED)
-                               SCTP_INC_STATS(net, SCTP_MIB_OUTUNORDERCHUNKS);
-                       else
-                               SCTP_INC_STATS(net, SCTP_MIB_OUTORDERCHUNKS);
-                       break;
-               }
+               pr_debug("%s: outqueueing: outq:%p, chunk:%p[%s])\n",
+                        __func__, q, chunk, chunk && chunk->chunk_hdr ?
+                        sctp_cname(SCTP_ST_CHUNK(chunk->chunk_hdr->type)) :
+                        "illegal chunk");
+
+               sctp_outq_tail_data(q, chunk);
+               if (chunk->asoc->peer.prsctp_capable &&
+                   SCTP_PR_PRIO_ENABLED(chunk->sinfo.sinfo_flags))
+                       chunk->asoc->sent_cnt_removable++;
+               if (chunk->chunk_hdr->flags & SCTP_DATA_UNORDERED)
+                       SCTP_INC_STATS(net, SCTP_MIB_OUTUNORDERCHUNKS);
+               else
+                       SCTP_INC_STATS(net, SCTP_MIB_OUTORDERCHUNKS);
        } else {
                list_add_tail(&chunk->list, &q->control_chunk_list);
                SCTP_INC_STATS(net, SCTP_MIB_OUTCTRLCHUNKS);
        }
 
-       if (error < 0)
-               return error;
-
        if (!q->cork)
-               error = sctp_outq_flush(q, 0, gfp);
-
-       return error;
+               sctp_outq_flush(q, 0, gfp);
 }
 
 /* Insert a chunk into the sorted list based on the TSNs.  The retransmit list
@@ -383,7 +354,7 @@ static int sctp_prsctp_prune_sent(struct sctp_association *asoc,
 
        list_for_each_entry_safe(chk, temp, queue, transmitted_list) {
                if (!SCTP_PR_PRIO_ENABLED(chk->sinfo.sinfo_flags) ||
-                   chk->prsctp_param <= sinfo->sinfo_timetolive)
+                   chk->sinfo.sinfo_timetolive <= sinfo->sinfo_timetolive)
                        continue;
 
                list_del_init(&chk->transmitted_list);
@@ -418,7 +389,7 @@ static int sctp_prsctp_prune_unsent(struct sctp_association *asoc,
 
        list_for_each_entry_safe(chk, temp, queue, list) {
                if (!SCTP_PR_PRIO_ENABLED(chk->sinfo.sinfo_flags) ||
-                   chk->prsctp_param <= sinfo->sinfo_timetolive)
+                   chk->sinfo.sinfo_timetolive <= sinfo->sinfo_timetolive)
                        continue;
 
                list_del_init(&chk->list);
@@ -442,7 +413,7 @@ void sctp_prsctp_prune(struct sctp_association *asoc,
 {
        struct sctp_transport *transport;
 
-       if (!asoc->prsctp_enable || !asoc->sent_cnt_removable)
+       if (!asoc->peer.prsctp_capable || !asoc->sent_cnt_removable)
                return;
 
        msg_len = sctp_prsctp_prune_sent(asoc, sinfo,
@@ -559,7 +530,6 @@ void sctp_retransmit(struct sctp_outq *q, struct sctp_transport *transport,
                     sctp_retransmit_reason_t reason)
 {
        struct net *net = sock_net(q->asoc->base.sk);
-       int error = 0;
 
        switch (reason) {
        case SCTP_RTXR_T3_RTX:
@@ -603,10 +573,7 @@ void sctp_retransmit(struct sctp_outq *q, struct sctp_transport *transport,
         * will be flushed at the end.
         */
        if (reason != SCTP_RTXR_FAST_RTX)
-               error = sctp_outq_flush(q, /* rtx_timeout */ 1, GFP_ATOMIC);
-
-       if (error)
-               q->asoc->base.sk->sk_err = -error;
+               sctp_outq_flush(q, /* rtx_timeout */ 1, GFP_ATOMIC);
 }
 
 /*
@@ -778,12 +745,12 @@ redo:
 }
 
 /* Cork the outqueue so queued chunks are really queued. */
-int sctp_outq_uncork(struct sctp_outq *q, gfp_t gfp)
+void sctp_outq_uncork(struct sctp_outq *q, gfp_t gfp)
 {
        if (q->cork)
                q->cork = 0;
 
-       return sctp_outq_flush(q, 0, gfp);
+       sctp_outq_flush(q, 0, gfp);
 }
 
 
@@ -796,7 +763,7 @@ int sctp_outq_uncork(struct sctp_outq *q, gfp_t gfp)
  * locking concerns must be made.  Today we use the sock lock to protect
  * this function.
  */
-static int sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
+static void sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
 {
        struct sctp_packet *packet;
        struct sctp_packet singleton;
@@ -919,8 +886,10 @@ static int sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
                        sctp_packet_config(&singleton, vtag, 0);
                        sctp_packet_append_chunk(&singleton, chunk);
                        error = sctp_packet_transmit(&singleton, gfp);
-                       if (error < 0)
-                               return error;
+                       if (error < 0) {
+                               asoc->base.sk->sk_err = -error;
+                               return;
+                       }
                        break;
 
                case SCTP_CID_ABORT:
@@ -1018,6 +987,8 @@ static int sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
                retran:
                        error = sctp_outq_flush_rtx(q, packet,
                                                    rtx_timeout, &start_timer);
+                       if (error < 0)
+                               asoc->base.sk->sk_err = -error;
 
                        if (start_timer) {
                                sctp_transport_reset_t3_rtx(transport);
@@ -1055,7 +1026,7 @@ static int sctp_outq_flush(struct sctp_outq *q, int rtx_timeout, gfp_t gfp)
 
                                /* Mark as failed send. */
                                sctp_chunk_fail(chunk, SCTP_ERROR_INV_STRM);
-                               if (asoc->prsctp_enable &&
+                               if (asoc->peer.prsctp_capable &&
                                    SCTP_PR_PRIO_ENABLED(chunk->sinfo.sinfo_flags))
                                        asoc->sent_cnt_removable--;
                                sctp_chunk_free(chunk);
@@ -1192,14 +1163,15 @@ sctp_flush_out:
                                                      struct sctp_transport,
                                                      send_ready);
                packet = &t->packet;
-               if (!sctp_packet_empty(packet))
+               if (!sctp_packet_empty(packet)) {
                        error = sctp_packet_transmit(packet, gfp);
+                       if (error < 0)
+                               asoc->base.sk->sk_err = -error;
+               }
 
                /* Clear the burst limited state, if any */
                sctp_transport_burst_reset(t);
        }
-
-       return error;
 }
 
 /* Update unack_data based on the incoming SACK chunk */
@@ -1347,7 +1319,7 @@ int sctp_outq_sack(struct sctp_outq *q, struct sctp_chunk *chunk)
                tsn = ntohl(tchunk->subh.data_hdr->tsn);
                if (TSN_lte(tsn, ctsn)) {
                        list_del_init(&tchunk->transmitted_list);
-                       if (asoc->prsctp_enable &&
+                       if (asoc->peer.prsctp_capable &&
                            SCTP_PR_PRIO_ENABLED(chunk->sinfo.sinfo_flags))
                                asoc->sent_cnt_removable--;
                        sctp_chunk_free(tchunk);
@@ -1747,7 +1719,7 @@ static int sctp_acked(struct sctp_sackhdr *sack, __u32 tsn)
 {
        int i;
        sctp_sack_variable_t *frags;
-       __u16 gap;
+       __u16 tsn_offset, blocks;
        __u32 ctsn = ntohl(sack->cum_tsn_ack);
 
        if (TSN_lte(tsn, ctsn))
@@ -1766,10 +1738,11 @@ static int sctp_acked(struct sctp_sackhdr *sack, __u32 tsn)
         */
 
        frags = sack->variable;
-       gap = tsn - ctsn;
-       for (i = 0; i < ntohs(sack->num_gap_ack_blocks); ++i) {
-               if (TSN_lte(ntohs(frags[i].gab.start), gap) &&
-                   TSN_lte(gap, ntohs(frags[i].gab.end)))
+       blocks = ntohs(sack->num_gap_ack_blocks);
+       tsn_offset = tsn - ctsn;
+       for (i = 0; i < blocks; ++i) {
+               if (tsn_offset >= ntohs(frags[i].gab.start) &&
+                   tsn_offset <= ntohs(frags[i].gab.end))
                        goto pass;
        }