/net/ipv6/ |
H A D | xfrm6_output.c | 23 int xfrm6_find_1stfragopt(struct xfrm_state *x, struct sk_buff *skb, argument 26 return ip6_find_1stfragopt(skb, prevhdr); 31 static int xfrm6_local_dontfrag(struct sk_buff *skb) argument 34 struct sock *sk = skb->sk; 46 static void xfrm6_local_rxpmtu(struct sk_buff *skb, u32 mtu) argument 49 struct sock *sk = skb->sk; 52 fl6.daddr = ipv6_hdr(skb)->daddr; 57 static void xfrm6_local_error(struct sk_buff *skb, u32 mtu) argument 60 struct sock *sk = skb->sk; 63 fl6.daddr = ipv6_hdr(skb) 68 xfrm6_tunnel_check_size(struct sk_buff *skb) argument 92 xfrm6_extract_output(struct xfrm_state *x, struct sk_buff *skb) argument 105 xfrm6_prepare_output(struct xfrm_state *x, struct sk_buff *skb) argument 125 xfrm6_output_finish(struct sk_buff *skb) argument 135 __xfrm6_output(struct sk_buff *skb) argument 157 xfrm6_output(struct sk_buff *skb) argument [all...] |
H A D | udp_offload.c | 19 static int udp6_ufo_send_check(struct sk_buff *skb) argument 25 if (skb->encapsulation) 28 if (!pskb_may_pull(skb, sizeof(*uh))) 31 ipv6h = ipv6_hdr(skb); 32 uh = udp_hdr(skb); 34 uh->check = ~csum_ipv6_magic(&ipv6h->saddr, &ipv6h->daddr, skb->len, 36 skb->csum_start = skb_transport_header(skb) - skb->head; 37 skb 42 udp6_ufo_fragment(struct sk_buff *skb, netdev_features_t features) argument [all...] |
H A D | tcpv6_offload.c | 18 static int tcp_v6_gso_send_check(struct sk_buff *skb) argument 23 if (!pskb_may_pull(skb, sizeof(*th))) 26 ipv6h = ipv6_hdr(skb); 27 th = tcp_hdr(skb); 30 skb->ip_summed = CHECKSUM_PARTIAL; 31 __tcp_v6_send_check(skb, &ipv6h->saddr, &ipv6h->daddr); 36 struct sk_buff *skb) 38 const struct ipv6hdr *iph = skb_gro_network_header(skb); 42 switch (skb->ip_summed) { 44 if (!tcp_v6_check(skb_gro_len(skb), 35 tcp6_gro_receive(struct sk_buff **head, struct sk_buff *skb) argument 71 tcp6_gro_complete(struct sk_buff *skb) argument [all...] |
H A D | xfrm6_mode_transport.c | 22 static int xfrm6_transport_output(struct xfrm_state *x, struct sk_buff *skb) argument 28 iph = ipv6_hdr(skb); 30 hdr_len = x->type->hdr_offset(x, skb, &prevhdr); 31 skb_set_mac_header(skb, (prevhdr - x->props.header_len) - skb->data); 32 skb_set_network_header(skb, -x->props.header_len); 33 skb->transport_header = skb->network_header + hdr_len; 34 __skb_pull(skb, hdr_len); 35 memmove(ipv6_hdr(skb), ip 47 xfrm6_transport_input(struct xfrm_state *x, struct sk_buff *skb) argument [all...] |
H A D | ip6_offload.c | 22 static int ipv6_gso_pull_exthdrs(struct sk_buff *skb, int proto) argument 40 if (unlikely(!pskb_may_pull(skb, 8))) 43 opth = (void *)skb->data; 46 if (unlikely(!pskb_may_pull(skb, len))) 50 __skb_pull(skb, len); 56 static int ipv6_gso_send_check(struct sk_buff *skb) argument 62 if (unlikely(!pskb_may_pull(skb, sizeof(*ipv6h)))) 65 ipv6h = ipv6_hdr(skb); 66 __skb_pull(skb, sizeof(*ipv6h)); 71 ipv6_gso_pull_exthdrs(skb, ipv6 83 ipv6_gso_segment(struct sk_buff *skb, netdev_features_t features) argument 144 ipv6_gro_receive(struct sk_buff **head, struct sk_buff *skb) argument 235 ipv6_gro_complete(struct sk_buff *skb) argument [all...] |
H A D | ip6_input.c | 50 int ip6_rcv_finish(struct sk_buff *skb) argument 52 if (sysctl_ip_early_demux && !skb_dst(skb)) { 55 ipprot = rcu_dereference(inet6_protos[ipv6_hdr(skb)->nexthdr]); 57 ipprot->early_demux(skb); 59 if (!skb_dst(skb)) 60 ip6_route_input(skb); 62 return dst_input(skb); 65 int ipv6_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev) argument 70 struct net *net = dev_net(skb->dev); 72 if (skb 197 ip6_input_finish(struct sk_buff *skb) argument 273 ip6_input(struct sk_buff *skb) argument 279 ip6_mc_input(struct sk_buff *skb) argument [all...] |
H A D | xfrm6_input.c | 19 int xfrm6_extract_input(struct xfrm_state *x, struct sk_buff *skb) argument 21 return xfrm6_extract_header(skb); 24 int xfrm6_rcv_spi(struct sk_buff *skb, int nexthdr, __be32 spi) argument 26 XFRM_SPI_SKB_CB(skb)->family = AF_INET6; 27 XFRM_SPI_SKB_CB(skb)->daddroff = offsetof(struct ipv6hdr, daddr); 28 return xfrm_input(skb, nexthdr, spi, 0); 32 int xfrm6_transport_finish(struct sk_buff *skb, int async) argument 34 skb_network_header(skb)[IP6CB(skb)->nhoff] = 35 XFRM_MODE_SKB_CB(skb) 50 xfrm6_rcv(struct sk_buff *skb) argument 58 xfrm6_input_addr(struct sk_buff *skb, xfrm_address_t *daddr, xfrm_address_t *saddr, u8 proto) argument [all...] |
H A D | xfrm6_mode_tunnel.c | 21 static inline void ipip6_ecn_decapsulate(struct sk_buff *skb) argument 23 const struct ipv6hdr *outer_iph = ipv6_hdr(skb); 24 struct ipv6hdr *inner_iph = ipipv6_hdr(skb); 34 static int xfrm6_mode_tunnel_output(struct xfrm_state *x, struct sk_buff *skb) argument 36 struct dst_entry *dst = skb_dst(skb); 40 skb_set_network_header(skb, -x->props.header_len); 41 skb->mac_header = skb->network_header + 43 skb->transport_header = skb 66 xfrm6_mode_tunnel_input(struct xfrm_state *x, struct sk_buff *skb) argument [all...] |
/net/bridge/ |
H A D | br_forward.c | 25 struct sk_buff *skb, 27 struct sk_buff *skb)); 31 const struct sk_buff *skb) 33 return (((p->flags & BR_HAIRPIN_MODE) || skb->dev != p->dev) && 34 br_allowed_egress(p->br, nbp_get_vlan_info(p), skb) && 38 static inline unsigned int packet_length(const struct sk_buff *skb) argument 40 return skb->len - (skb->protocol == htons(ETH_P_8021Q) ? VLAN_HLEN : 0); 43 int br_dev_queue_push_xmit(struct sk_buff *skb) argument 46 if (nf_bridge_maybe_copy_header(skb) || 30 should_deliver(const struct net_bridge_port *p, const struct sk_buff *skb) argument 58 br_forward_finish(struct sk_buff *skb) argument 65 __br_deliver(const struct net_bridge_port *to, struct sk_buff *skb) argument 87 __br_forward(const struct net_bridge_port *to, struct sk_buff *skb) argument 109 br_deliver(const struct net_bridge_port *to, struct sk_buff *skb) argument 120 br_forward(const struct net_bridge_port *to, struct sk_buff *skb, struct sk_buff *skb0) argument 134 deliver_clone(const struct net_bridge_port *prev, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) argument 151 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) argument 174 br_flood(struct net_bridge *br, struct sk_buff *skb, struct sk_buff *skb0, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) argument 206 br_flood_deliver(struct net_bridge *br, struct sk_buff *skb) argument 212 br_flood_forward(struct net_bridge *br, struct sk_buff *skb, struct sk_buff *skb2) argument 220 br_multicast_flood(struct net_bridge_mdb_entry *mdst, struct sk_buff *skb, struct sk_buff *skb0, void (*__packet_hook)( const struct net_bridge_port *p, struct sk_buff *skb)) argument 269 br_multicast_deliver(struct net_bridge_mdb_entry *mdst, struct sk_buff *skb) argument 276 br_multicast_forward(struct net_bridge_mdb_entry *mdst, struct sk_buff *skb, struct sk_buff *skb2) argument [all...] |
H A D | br_netfilter.c | 45 #define skb_origaddr(skb) (((struct bridge_skb_cb *) \ 46 (skb->nf_bridge->data))->daddr.ipv4) 47 #define store_orig_dstaddr(skb) (skb_origaddr(skb) = ip_hdr(skb)->daddr) 48 #define dnat_took_place(skb) (skb_origaddr(skb) != ip_hdr(skb)->daddr) 67 #define IS_IP(skb) \ 68 (!vlan_tx_tag_present(skb) 76 vlan_proto(const struct sk_buff *skb) argument 98 pppoe_proto(const struct sk_buff *skb) argument 114 fake_update_pmtu(struct dst_entry *dst, struct sock *sk, struct sk_buff *skb, u32 mtu) argument 119 fake_redirect(struct dst_entry *dst, struct sock *sk, struct sk_buff *skb) argument 129 fake_neigh_lookup(const struct dst_entry *dst, struct sk_buff *skb, const void *daddr) argument 190 nf_bridge_alloc(struct sk_buff *skb) argument 199 nf_bridge_unshare(struct sk_buff *skb) argument 216 nf_bridge_push_encap_header(struct sk_buff *skb) argument 224 nf_bridge_pull_encap_header(struct sk_buff *skb) argument 232 nf_bridge_pull_encap_header_rcsum(struct sk_buff *skb) argument 240 nf_bridge_save_header(struct sk_buff *skb) argument 248 nf_bridge_update_protocol(struct sk_buff *skb) argument 261 br_parse_ip_options(struct sk_buff *skb) argument 326 nf_bridge_copy_header(struct sk_buff *skb) argument 346 br_nf_pre_routing_finish_ipv6(struct sk_buff *skb) argument 378 br_nf_pre_routing_finish_bridge(struct sk_buff *skb) argument 455 br_nf_pre_routing_finish(struct sk_buff *skb) argument 530 brnf_get_logical_dev(struct sk_buff *skb, const struct net_device *dev) argument 545 setup_pre_routing(struct sk_buff *skb) argument 566 check_hbh_len(struct sk_buff *skb) argument 622 br_nf_pre_routing_ipv6(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 672 br_nf_pre_routing(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 730 br_nf_local_in(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 740 br_nf_forward_finish(struct sk_buff *skb) argument 768 br_nf_forward_ip(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 821 br_nf_forward_arp(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 857 br_nf_dev_queue_xmit(struct sk_buff *skb) argument 874 br_nf_dev_queue_xmit(struct sk_buff *skb) argument 881 br_nf_post_routing(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument 926 ip_sabotage_in(unsigned int hook, struct sk_buff *skb, const struct net_device *in, const struct net_device *out, int (*okfn)(struct sk_buff *)) argument [all...] |
H A D | br_input.c | 27 static int br_pass_frame_up(struct sk_buff *skb) argument 29 struct net_device *indev, *brdev = BR_INPUT_SKB_CB(skb)->brdev; 35 brstats->rx_bytes += skb->len; 43 !br_allowed_egress(br, br_get_vlan_info(br), skb)) { 44 kfree_skb(skb); 48 skb = br_handle_vlan(br, br_get_vlan_info(br), skb); 49 if (!skb) 52 indev = skb->dev; 53 skb 60 br_handle_frame_finish(struct sk_buff *skb) argument 139 br_handle_local_finish(struct sk_buff *skb) argument 156 struct sk_buff *skb = *pskb; local [all...] |
/net/ipv4/ |
H A D | xfrm4_mode_transport.c | 21 static int xfrm4_transport_output(struct xfrm_state *x, struct sk_buff *skb) argument 23 struct iphdr *iph = ip_hdr(skb); 26 skb_set_network_header(skb, -x->props.header_len); 27 skb->mac_header = skb->network_header + 29 skb->transport_header = skb->network_header + ihl; 30 __skb_pull(skb, ihl); 31 memmove(skb_network_header(skb), iph, ihl); 39 * On entry, skb 43 xfrm4_transport_input(struct xfrm_state *x, struct sk_buff *skb) argument [all...] |
H A D | ip_forward.c | 42 static int ip_forward_finish(struct sk_buff *skb) argument 44 struct ip_options *opt = &(IPCB(skb)->opt); 46 IP_INC_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTFORWDATAGRAMS); 47 IP_ADD_STATS_BH(dev_net(skb_dst(skb)->dev), IPSTATS_MIB_OUTOCTETS, skb->len); 50 ip_forward_options(skb); 52 return dst_output(skb); 55 int ip_forward(struct sk_buff *skb) argument 59 struct ip_options *opt = &(IPCB(skb)->opt); 61 if (skb_warn_if_lro(skb)) [all...] |
H A D | xfrm4_output.c | 21 static int xfrm4_tunnel_check_size(struct sk_buff *skb) argument 26 if (IPCB(skb)->flags & IPSKB_XFRM_TUNNEL_SIZE) 29 if (!(ip_hdr(skb)->frag_off & htons(IP_DF)) || skb->local_df) 32 dst = skb_dst(skb); 34 if (skb->len > mtu) { 35 if (skb->sk) 36 ip_local_error(skb->sk, EMSGSIZE, ip_hdr(skb)->daddr, 37 inet_sk(skb 47 xfrm4_extract_output(struct xfrm_state *x, struct sk_buff *skb) argument 60 xfrm4_prepare_output(struct xfrm_state *x, struct sk_buff *skb) argument 77 xfrm4_output_finish(struct sk_buff *skb) argument 92 xfrm4_output(struct sk_buff *skb) argument [all...] |
H A D | xfrm4_input.c | 20 int xfrm4_extract_input(struct xfrm_state *x, struct sk_buff *skb) argument 22 return xfrm4_extract_header(skb); 25 static inline int xfrm4_rcv_encap_finish(struct sk_buff *skb) argument 27 if (skb_dst(skb) == NULL) { 28 const struct iphdr *iph = ip_hdr(skb); 30 if (ip_route_input_noref(skb, iph->daddr, iph->saddr, 31 iph->tos, skb->dev)) 34 return dst_input(skb); 36 kfree_skb(skb); 40 int xfrm4_rcv_encap(struct sk_buff *skb, in argument 49 xfrm4_transport_finish(struct sk_buff *skb, int async) argument 76 xfrm4_udp_encap_rcv(struct sock *sk, struct sk_buff *skb) argument 162 xfrm4_rcv(struct sk_buff *skb) argument [all...] |
/net/rose/ |
H A D | rose_loopback.c | 35 int rose_loopback_queue(struct sk_buff *skb, struct rose_neigh *neigh) argument 39 skbn = skb_clone(skb, GFP_ATOMIC); 41 kfree_skb(skb); 68 struct sk_buff *skb; local 75 while ((skb = skb_dequeue(&loopback_queue)) != NULL) { 76 if (skb->len < ROSE_MIN_LEN) { 77 kfree_skb(skb); 80 lci_i = ((skb->data[0] << 8) & 0xF00) + ((skb->data[1] << 0) & 0x0FF); 81 frametype = skb 116 struct sk_buff *skb; local [all...] |
/net/dsa/ |
H A D | tag_trailer.c | 17 netdev_tx_t trailer_xmit(struct sk_buff *skb, struct net_device *dev) argument 25 dev->stats.tx_bytes += skb->len; 34 if (skb->len < 60) 35 padlen = 60 - skb->len; 37 nskb = alloc_skb(NET_IP_ALIGN + skb->len + padlen + 4, GFP_ATOMIC); 39 kfree_skb(skb); 45 skb_set_network_header(nskb, skb_network_header(skb) - skb->head); 46 skb_set_transport_header(nskb, skb_transport_header(skb) - skb 69 trailer_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev) argument [all...] |
H A D | tag_edsa.c | 20 netdev_tx_t edsa_xmit(struct sk_buff *skb, struct net_device *dev) argument 26 dev->stats.tx_bytes += skb->len; 34 if (skb->protocol == htons(ETH_P_8021Q)) { 35 if (skb_cow_head(skb, DSA_HLEN) < 0) 37 skb_push(skb, DSA_HLEN); 39 memmove(skb->data, skb->data + DSA_HLEN, 2 * ETH_ALEN); 44 edsa_header = skb->data + 2 * ETH_ALEN; 60 if (skb_cow_head(skb, EDSA_HLEN) < 0) 62 skb_push(skb, EDSA_HLE 92 edsa_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev) argument [all...] |
H A D | tag_dsa.c | 19 netdev_tx_t dsa_xmit(struct sk_buff *skb, struct net_device *dev) argument 25 dev->stats.tx_bytes += skb->len; 32 if (skb->protocol == htons(ETH_P_8021Q)) { 33 if (skb_cow_head(skb, 0) < 0) 39 dsa_header = skb->data + 2 * ETH_ALEN; 51 if (skb_cow_head(skb, DSA_HLEN) < 0) 53 skb_push(skb, DSA_HLEN); 55 memmove(skb->data, skb->data + DSA_HLEN, 2 * ETH_ALEN); 60 dsa_header = skb 79 dsa_rcv(struct sk_buff *skb, struct net_device *dev, struct packet_type *pt, struct net_device *orig_dev) argument [all...] |
/net/x25/ |
H A D | x25_dev.c | 17 * 2000-09-04 Henner Eisen Prevent freeing a dangling skb. 29 static int x25_receive_data(struct sk_buff *skb, struct x25_neigh *nb) argument 35 if (!pskb_may_pull(skb, X25_STD_MIN_LEN)) 38 frametype = skb->data[2]; 39 lci = ((skb->data[0] << 8) & 0xF00) + ((skb->data[1] << 0) & 0x0FF); 46 x25_link_control(skb, nb, frametype); 56 skb_reset_transport_header(skb); 59 queued = x25_process_rx_frame(sk, skb); 61 queued = !sk_add_backlog(sk, skb, s 97 x25_lapb_receive_frame(struct sk_buff *skb, struct net_device *dev, struct packet_type *ptype, struct net_device *orig_dev) argument 151 struct sk_buff *skb; local 180 struct sk_buff *skb; local 204 x25_send_frame(struct sk_buff *skb, struct x25_neigh *nb) argument [all...] |
/net/mac802154/ |
H A D | rx.c | 47 struct sk_buff *skb; member in struct:rx_work 54 mac802154_subif_rx(struct ieee802154_dev *hw, struct sk_buff *skb, u8 lqi) argument 58 mac_cb(skb)->lqi = lqi; 59 skb->protocol = htons(ETH_P_IEEE802154); 60 skb_reset_mac_header(skb); 62 BUILD_BUG_ON(sizeof(struct ieee802154_mac_cb) > sizeof(skb->cb)); 67 if (skb->len < 2) { 71 crc = crc_ccitt(0, skb->data, skb->len); 76 skb_trim(skb, sk 89 struct sk_buff *skb = rw->skb; local 96 ieee802154_rx_irqsafe(struct ieee802154_dev *dev, struct sk_buff *skb, u8 lqi) argument [all...] |
/net/lapb/ |
H A D | lapb_subr.c | 53 struct sk_buff *skb; local 63 skb = skb_dequeue(&lapb->ack_queue); 64 kfree_skb(skb); 71 struct sk_buff *skb, *skb_prev = NULL; local 78 while ((skb = skb_dequeue(&lapb->ack_queue)) != NULL) { 80 skb_queue_head(&lapb->write_queue, skb); 82 skb_append(skb_prev, skb, &lapb->write_queue); 83 skb_prev = skb; 111 int lapb_decode(struct lapb_cb *lapb, struct sk_buff *skb, argument 118 skb 231 struct sk_buff *skb; local 267 struct sk_buff *skb; local [all...] |
/net/llc/ |
H A D | llc_s_ac.c | 32 * @skb: the event to forward 37 int llc_sap_action_unitdata_ind(struct llc_sap *sap, struct sk_buff *skb) argument 39 llc_sap_rtn_pdu(sap, skb); 46 * @skb: the event to send 52 int llc_sap_action_send_ui(struct llc_sap *sap, struct sk_buff *skb) argument 54 struct llc_sap_state_ev *ev = llc_sap_ev(skb); 57 llc_pdu_header_init(skb, LLC_PDU_TYPE_U, ev->saddr.lsap, 59 llc_pdu_init_as_ui_cmd(skb); 60 rc = llc_mac_hdr_init(skb, ev->saddr.mac, ev->daddr.mac); 62 rc = dev_queue_xmit(skb); 75 llc_sap_action_send_xid_c(struct llc_sap *sap, struct sk_buff *skb) argument 97 llc_sap_action_send_xid_r(struct llc_sap *sap, struct sk_buff *skb) argument 129 llc_sap_action_send_test_c(struct llc_sap *sap, struct sk_buff *skb) argument 143 llc_sap_action_send_test_r(struct llc_sap *sap, struct sk_buff *skb) argument 177 llc_sap_action_report_status(struct llc_sap *sap, struct sk_buff *skb) argument 190 llc_sap_action_xid_ind(struct llc_sap *sap, struct sk_buff *skb) argument 204 llc_sap_action_test_ind(struct llc_sap *sap, struct sk_buff *skb) argument [all...] |
H A D | llc_s_ev.c | 23 int llc_sap_ev_activation_req(struct llc_sap *sap, struct sk_buff *skb) argument 25 struct llc_sap_state_ev *ev = llc_sap_ev(skb); 31 int llc_sap_ev_rx_ui(struct llc_sap *sap, struct sk_buff *skb) argument 33 struct llc_sap_state_ev *ev = llc_sap_ev(skb); 34 struct llc_pdu_un *pdu = llc_pdu_un_hdr(skb); 41 int llc_sap_ev_unitdata_req(struct llc_sap *sap, struct sk_buff *skb) argument 43 struct llc_sap_state_ev *ev = llc_sap_ev(skb); 51 int llc_sap_ev_xid_req(struct llc_sap *sap, struct sk_buff *skb) argument 53 struct llc_sap_state_ev *ev = llc_sap_ev(skb); 60 int llc_sap_ev_rx_xid_c(struct llc_sap *sap, struct sk_buff *skb) argument 70 llc_sap_ev_rx_xid_r(struct llc_sap *sap, struct sk_buff *skb) argument 80 llc_sap_ev_test_req(struct llc_sap *sap, struct sk_buff *skb) argument 89 llc_sap_ev_rx_test_c(struct llc_sap *sap, struct sk_buff *skb) argument 99 llc_sap_ev_rx_test_r(struct llc_sap *sap, struct sk_buff *skb) argument 109 llc_sap_ev_deactivation_req(struct llc_sap *sap, struct sk_buff *skb) argument [all...] |
/net/core/ |
H A D | timestamping.c | 30 static unsigned int classify(const struct sk_buff *skb) argument 32 if (likely(skb->dev && 33 skb->dev->phydev && 34 skb->dev->phydev->drv)) 35 return sk_run_filter(skb, ptp_filter); 40 void skb_clone_tx_timestamp(struct sk_buff *skb) argument 44 struct sock *sk = skb->sk; 50 type = classify(skb); 59 phydev = skb->dev->phydev; 63 clone = skb_clone(skb, GFP_ATOMI 78 skb_complete_tx_timestamp(struct sk_buff *skb, struct skb_shared_hwtstamps *hwtstamps) argument 104 skb_defer_rx_timestamp(struct sk_buff *skb) argument [all...] |