/net/rxrpc/ |
H A D | ar-peer.c | 33 * assess the MTU size for the network interface through which this peer is 36 static void rxrpc_assess_MTU_size(struct rxrpc_peer *peer) argument 41 peer->if_mtu = 1500; 44 peer->srx.transport.sin.sin_addr.s_addr, 0, 52 peer->if_mtu = dst_mtu(&rt->dst); 55 _leave(" [if_mtu %u]", peer->if_mtu); 59 * allocate a new peer 64 struct rxrpc_peer *peer; local 68 peer = kzalloc(sizeof(struct rxrpc_peer), gfp); 69 if (peer) { 108 struct rxrpc_peer *peer, *candidate; local 198 struct rxrpc_peer *peer; local 241 rxrpc_put_peer(struct rxrpc_peer *peer) argument 261 struct rxrpc_peer *peer = local [all...] |
H A D | ar-error.c | 33 struct rxrpc_peer *peer; local 55 peer = rxrpc_find_peer(local, addr, port); 56 if (IS_ERR(peer)) { 58 _leave(" [no peer]"); 62 trans = rxrpc_find_transport(local, peer); 64 rxrpc_put_peer(peer); 79 if (mtu > 0 && peer->if_mtu == 65535 && mtu < peer->if_mtu) { 80 peer->if_mtu = mtu; 92 if (mtu < peer [all...] |
H A D | ar-transport.c | 31 struct rxrpc_peer *peer, 41 trans->peer = peer; 52 if (peer->srx.transport.family == AF_INET) { 53 switch (peer->srx.transport_type) { 75 struct rxrpc_peer *peer, 85 &peer->srx.transport.sin.sin_addr, 86 ntohs(peer->srx.transport.sin.sin_port)); 91 if (trans->local == local && trans->peer == peer) 30 rxrpc_alloc_transport(struct rxrpc_local *local, struct rxrpc_peer *peer, gfp_t gfp) argument 74 rxrpc_get_transport(struct rxrpc_local *local, struct rxrpc_peer *peer, gfp_t gfp) argument 149 rxrpc_find_transport(struct rxrpc_local *local, struct rxrpc_peer *peer) argument [all...] |
H A D | Makefile | 17 ar-peer.o \
|
H A D | ar-proc.c | 69 &trans->peer->srx.transport.sin.sin_addr, 70 ntohs(trans->peer->srx.transport.sin.sin_port)); 152 &trans->peer->srx.transport.sin.sin_addr, 153 ntohs(trans->peer->srx.transport.sin.sin_port));
|
H A D | ar-ack.c | 285 * - these indicate packets the peer has or has not received, but hasn't yet 554 struct rxrpc_peer *peer; local 569 peer = call->conn->trans->peer; 570 if (mtu < peer->maxdata) { 571 spin_lock_bh(&peer->lock); 572 peer->maxdata = mtu; 573 peer->mtu = mtu + peer->hdrsize; 574 spin_unlock_bh(&peer [all...] |
H A D | ar-accept.c | 70 * accept an incoming call that needs peer, transport and/or connection setting 81 struct rxrpc_peer *peer; local 100 peer = rxrpc_get_peer(srx, GFP_NOIO); 101 if (IS_ERR(peer)) { 102 _debug("no peer"); 107 trans = rxrpc_get_transport(local, peer, GFP_NOIO); 108 rxrpc_put_peer(peer); 204 * accept incoming calls that need peer, transport and/or connection setting up
|
H A D | ar-output.c | 103 * abort a call, sending an ABORT packet to the peer 346 msg.msg_name = &trans->peer->srx.transport.sin; 347 msg.msg_namelen = sizeof(trans->peer->srx.transport.sin); 354 if (skb->len - sizeof(struct rxrpc_header) < trans->peer->maxdata) { 360 * message and update the peer record 369 _leave(" = %d [%u]", ret, trans->peer->maxdata); 391 _leave(" = %d [frag %u]", ret, trans->peer->maxdata); 582 max = call->conn->trans->peer->maxdata; 725 ret = call->conn->trans->peer->net_error;
|
/net/sctp/ |
H A D | associola.c | 210 asoc->peer.rwnd = SCTP_DEFAULT_MAXWINDOW; 221 asoc->peer.i.init_tag = 0; /* INIT needs a vtag of 0. */ 253 INIT_LIST_HEAD(&asoc->peer.transport_addr_list); 254 asoc->peer.transport_count = 0; 267 asoc->peer.sack_needed = 1; 268 asoc->peer.sack_cnt = 0; 269 asoc->peer.sack_generation = 1; 271 /* Assume that the peer will tell us if he recognizes ASCONF 276 asoc->peer.asconf_capable = 0; 278 asoc->peer 540 sctp_assoc_rm_peer(struct sctp_association *asoc, struct sctp_transport *peer) argument 637 struct sctp_transport *peer; local [all...] |
H A D | transport.c | 63 struct sctp_transport *peer, 68 peer->ipaddr = *addr; 69 peer->af_specific = sctp_get_af_specific(addr->sa.sa_family); 70 memset(&peer->saddr, 0, sizeof(union sctp_addr)); 72 peer->sack_generation = 0; 80 peer->rto = msecs_to_jiffies(net->sctp.rto_initial); 82 peer->last_time_heard = jiffies; 83 peer->last_time_ecne_reduced = jiffies; 85 peer->param_flags = SPP_HB_DISABLE | 90 peer 62 sctp_transport_init(struct net *net, struct sctp_transport *peer, const union sctp_addr *addr, gfp_t gfp) argument [all...] |
H A D | sm_make_chunk.c | 406 /* Tell peer that we'll do ECN only if peer advertised such cap. */ 407 if (asoc->peer.ecn_capable) 410 if (asoc->peer.prsctp_capable) 413 if (asoc->peer.asconf_capable) { 422 if (asoc->peer.auth_capable) { 466 if (asoc->peer.ecn_capable) 476 if (asoc->peer.prsctp_capable) 486 if (asoc->peer.auth_capable) { 510 * It is sent by the initiator of an association to its peer t 2980 struct sctp_transport *peer; local [all...] |
H A D | sm_sideeffect.c | 100 * its peer that it has reduced its congestion window. This 115 * and then ACK our peer that we we've done so by 165 struct sctp_transport *trans = asoc->peer.last_data_from; 171 asoc->peer.sack_needed = 1; 173 ctsn = sctp_tsnmap_get_ctsn(&asoc->peer.tsn_map); 174 max_tsn_seen = sctp_tsnmap_get_max_tsn_seen(&asoc->peer.tsn_map); 187 asoc->peer.sack_needed = 1; 197 if (!asoc->peer.sack_needed) { 198 asoc->peer.sack_cnt++; 207 if (asoc->peer 249 sctp_generate_t3_rtx_event(unsigned long peer) argument [all...] |
H A D | probe.c | 136 sp = asoc->peer.primary_path; 139 (!port || asoc->peer.port == port || 150 ep->base.bind_addr.port, asoc->peer.port, 151 asoc->pathmtu, asoc->peer.rwnd, asoc->unack_data); 153 list_for_each_entry(sp, &asoc->peer.transport_addr_list, 155 if (sp == asoc->peer.primary_path)
|
H A D | proc.c | 132 struct sctp_transport *peer; local 138 peer = asoc->peer.primary_path; 139 primary = &peer->saddr; 164 primary = &assoc->peer.primary_addr; 166 list_for_each_entry_rcu(transport, &assoc->peer.transport_addr_list, 348 assoc->peer.port); 457 list_for_each_entry_rcu(tsp, &assoc->peer.transport_addr_list,
|
H A D | outqueue.c | 230 list_for_each_entry(transport, &q->asoc->peer.transport_addr_list, 315 * association on each peer will stop accepting new 403 * stop considering it 'outstanding'. Our peer 412 q->asoc->peer.rwnd += sctp_data_size(chunk); 432 q->asoc->peer.rwnd += sctp_data_size(chunk); 489 if (transport == transport->asoc->peer.retran_path) 724 __u16 dport = asoc->peer.port; 725 __u32 vtag = asoc->peer.i.init_tag; 790 new_transport = asoc->peer.active_path; 810 new_transport = asoc->peer [all...] |
H A D | output.c | 216 /* if the peer did not request this chunk to be authenticated, 255 pkt->transport->asoc->peer.sack_generation) 266 asoc->peer.sack_needed = 0; 573 * any problems protocol-wise even if our peer ignores it. 588 if (asoc->peer.last_sent_to != tp) 592 asoc->peer.last_sent_to = tp; 663 * any destination transport address if its peer's rwnd indicates 664 * that the peer has no buffer space (i.e. rwnd is 0, see Section 673 rwnd = asoc->peer.rwnd; 740 u32 rwnd = asoc->peer [all...] |
H A D | auth.c | 242 /* Make a key vector based on peer's parameters */ 247 return sctp_auth_make_key_vector(asoc->peer.peer_random, 248 asoc->peer.peer_chunks, 249 asoc->peer.peer_hmacs, 400 /* If we don't support AUTH, or peer is not capable 403 if (!net->sctp.auth_enable || !asoc->peer.auth_capable) 540 hmacs = asoc->peer.peer_hmacs; 677 /* Check if peer requested that this chunk is authenticated */ 685 if (!net->sctp.auth_enable || !asoc->peer.auth_capable) 688 return __sctp_auth_cid(chunk, asoc->peer [all...] |
/net/ipv6/ |
H A D | output_core.c | 16 struct inet_peer *peer; local 20 peer = inet_getpeer_v6(net->ipv6.peers, &rt->rt6i_dst.addr, 1); 21 if (peer) { 22 fhdr->identification = htonl(inet_getid(peer, 0)); 23 inet_putpeer(peer);
|
/net/unix/ |
H A D | diag.c | 40 struct sock *peer; local 43 peer = unix_peer_get(sk); 44 if (peer) { 45 unix_state_lock(peer); 46 ino = sock_i_ino(peer); 47 unix_state_unlock(peer); 48 sock_put(peer); 74 struct sock *req, *peer; local 83 peer = unix_sk(req)->peer; [all...] |
/net/x25/ |
H A D | x25_forward.c | 98 struct net_device *peer = NULL; local 109 peer = frwd->dev2; 111 peer = frwd->dev1; 118 if ( (nb = x25_get_neigh(peer)) == NULL)
|
/net/wireless/ |
H A D | trace.h | 1322 const u8 *peer, const struct cfg80211_bitrate_mask *mask), 1323 TP_ARGS(wiphy, netdev, peer, mask), 1327 MAC_ENTRY(peer) 1332 MAC_ASSIGN(peer, peer); 1334 TP_printk(WIPHY_PR_FMT ", " NETDEV_PR_FMT ", peer: " MAC_PR_FMT, 1335 WIPHY_PR_ARG, NETDEV_PR_ARG, MAC_PR_ARG(peer)) 1446 u8 *peer, u8 action_code, u8 dialog_token, 1448 TP_ARGS(wiphy, netdev, peer, action_code, dialog_token, status_code, 1453 MAC_ENTRY(peer) [all...] |
H A D | rdev-ops.h | 543 struct net_device *dev, const u8 *peer, 547 trace_rdev_set_bitrate_mask(&rdev->wiphy, dev, peer, mask); 548 ret = rdev->ops->set_bitrate_mask(&rdev->wiphy, dev, peer, mask); 773 struct net_device *dev, u8 *peer, 778 trace_rdev_tdls_mgmt(&rdev->wiphy, dev, peer, action_code, 780 ret = rdev->ops->tdls_mgmt(&rdev->wiphy, dev, peer, action_code, 787 struct net_device *dev, u8 *peer, 791 trace_rdev_tdls_oper(&rdev->wiphy, dev, peer, oper); 792 ret = rdev->ops->tdls_oper(&rdev->wiphy, dev, peer, oper); 798 struct net_device *dev, const u8 *peer, 542 rdev_set_bitrate_mask(struct cfg80211_registered_device *rdev, struct net_device *dev, const u8 *peer, const struct cfg80211_bitrate_mask *mask) argument 772 rdev_tdls_mgmt(struct cfg80211_registered_device *rdev, struct net_device *dev, u8 *peer, u8 action_code, u8 dialog_token, u16 status_code, const u8 *buf, size_t len) argument 786 rdev_tdls_oper(struct cfg80211_registered_device *rdev, struct net_device *dev, u8 *peer, enum nl80211_tdls_operation oper) argument 797 rdev_probe_client(struct cfg80211_registered_device *rdev, struct net_device *dev, const u8 *peer, u64 *cookie) argument [all...] |
/net/ipv4/ |
H A D | route.c | 490 struct inet_peer *peer; local 492 peer = inet_getpeer_v4(net->ipv4.peers, iph->daddr, 1); 493 if (peer) { 494 iph->id = htons(inet_getid(peer, more)); 495 inet_putpeer(peer); 792 struct inet_peer *peer; local 806 peer = inet_getpeer_v4(net->ipv4.peers, ip_hdr(skb)->saddr, 1); 807 if (!peer) { 816 if (time_after(jiffies, peer->rate_last + ip_rt_redirect_silence)) 817 peer 855 struct inet_peer *peer; local [all...] |
H A D | inetpeer.c | 27 * We keep one entry for each peer IP address. The nodes contains long-living 28 * information about the peer which doesn't depend on routes. 555 bool inet_peer_xrlim_allow(struct inet_peer *peer, int timeout) argument 560 if (!peer) 563 token = peer->rate_tokens; 565 token += now - peer->rate_last; 566 peer->rate_last = now; 573 peer->rate_tokens = token;
|
H A D | ip_fragment.c | 79 struct inet_peer *peer; member in struct:ipq 150 qp->peer = sysctl_ipfrag_max_dist ? 159 if (qp->peer) 160 inet_putpeer(qp->peer); 275 struct inet_peer *peer = qp->peer; local 281 if (!peer || !max) 285 end = atomic_inc_return(&peer->rid);
|