Lines Matching refs:skb_out

1207 	struct sk_buff *skb_out;  in cdc_ncm_fill_tx_frame()  local
1233 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame()
1236 if (!skb_out) { in cdc_ncm_fill_tx_frame()
1239 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1245 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1251 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1261 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame()
1264 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame()
1275 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame()
1281 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame()
1308 ndp.ndp16 = cdc_ncm_ndp16(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1310 ndp.ndp32 = cdc_ncm_ndp32(ctx, skb_out, sign, skb->len + ctx->tx_modulus + ctx->tx_remainder); in cdc_ncm_fill_tx_frame()
1313 cdc_ncm_align_tail(skb_out, ctx->tx_modulus, ctx->tx_remainder, ctx->tx_curr_size); in cdc_ncm_fill_tx_frame()
1317 skb_out->len + skb->len + delayed_ndp_size > ctx->tx_curr_size) { in cdc_ncm_fill_tx_frame()
1345 ndp.ndp16->dpe16[index].wDatagramIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1352 ndp.ndp32->dpe32[index].dwDatagramIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1355 skb_put_data(skb_out, skb->data, skb->len); in cdc_ncm_fill_tx_frame()
1380 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1386 ctx->tx_curr_skb = skb_out; in cdc_ncm_fill_tx_frame()
1402 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1403 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1404 nth.nth16->wNdpIndex = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1405 skb_put_data(skb_out, ctx->delayed_ndp16, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1410 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1411 cdc_ncm_align_tail(skb_out, ctx->tx_ndp_modulus, 0, ctx->tx_curr_size - ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1412 nth.nth32->dwNdpIndex = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1413 skb_put_data(skb_out, ctx->delayed_ndp32, ctx->max_ndp_size); in cdc_ncm_fill_tx_frame()
1429 skb_out->len > ctx->min_tx_pkt) { in cdc_ncm_fill_tx_frame()
1430 padding_count = ctx->tx_curr_size - skb_out->len; in cdc_ncm_fill_tx_frame()
1432 skb_put_zero(skb_out, padding_count); in cdc_ncm_fill_tx_frame()
1433 } else if (skb_out->len < ctx->tx_curr_size && in cdc_ncm_fill_tx_frame()
1434 (skb_out->len % dev->maxpacket) == 0) { in cdc_ncm_fill_tx_frame()
1435 skb_put_u8(skb_out, 0); /* force short packet */ in cdc_ncm_fill_tx_frame()
1440 nth.nth16 = (struct usb_cdc_ncm_nth16 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1441 nth.nth16->wBlockLength = cpu_to_le16(skb_out->len); in cdc_ncm_fill_tx_frame()
1443 nth.nth32 = (struct usb_cdc_ncm_nth32 *)skb_out->data; in cdc_ncm_fill_tx_frame()
1444 nth.nth32->dwBlockLength = cpu_to_le32(skb_out->len); in cdc_ncm_fill_tx_frame()
1451 ctx->tx_overhead += skb_out->len - ctx->tx_curr_frame_payload; in cdc_ncm_fill_tx_frame()
1458 usbnet_set_skb_tx_stats(skb_out, n, in cdc_ncm_fill_tx_frame()
1459 (long)ctx->tx_curr_frame_payload - skb_out->len); in cdc_ncm_fill_tx_frame()
1461 return skb_out; in cdc_ncm_fill_tx_frame()
1514 struct sk_buff *skb_out; in cdc_ncm_tx_fixup() local
1530 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1532 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP32_NOCRC_SIGN)); in cdc_ncm_tx_fixup()
1535 return skb_out; in cdc_ncm_tx_fixup()