ke 89 net/pf_table.c #define KENTRY_NETWORK(ke) ((ke)->pfrke_net < AF_BITS((ke)->pfrke_af)) ke 90 net/pf_table.c #define KENTRY_RNF_ROOT(ke) \ ke 91 net/pf_table.c ((((struct radix_node *)(ke))->rn_flags & RNF_ROOT) != 0) ke 135 net/pf_table.c struct pfr_kentry *ke); ke 776 net/pf_table.c struct pfr_kentry *ke; ke 790 net/pf_table.c ke = (struct pfr_kentry *)rn_lookup(&sa, &mask, head); ke 792 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 793 net/pf_table.c ke = NULL; ke 795 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&sa, head); ke 796 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 797 net/pf_table.c ke = NULL; ke 798 net/pf_table.c if (exact && ke && KENTRY_NETWORK(ke)) ke 799 net/pf_table.c ke = NULL; ke 801 net/pf_table.c return (ke); ke 807 net/pf_table.c struct pfr_kentry *ke; ke 810 net/pf_table.c ke = pool_get(&pfr_kentry_pl2, PR_NOWAIT); ke 812 net/pf_table.c ke = pool_get(&pfr_kentry_pl, PR_NOWAIT); ke 813 net/pf_table.c if (ke == NULL) ke 815 net/pf_table.c bzero(ke, sizeof(*ke)); ke 818 net/pf_table.c FILLIN_SIN(ke->pfrke_sa.sin, ad->pfra_ip4addr); ke 820 net/pf_table.c FILLIN_SIN6(ke->pfrke_sa.sin6, ad->pfra_ip6addr); ke 821 net/pf_table.c ke->pfrke_af = ad->pfra_af; ke 822 net/pf_table.c ke->pfrke_net = ad->pfra_net; ke 823 net/pf_table.c ke->pfrke_not = ad->pfra_not; ke 824 net/pf_table.c ke->pfrke_intrpool = intr; ke 825 net/pf_table.c return (ke); ke 840 net/pf_table.c pfr_destroy_kentry(struct pfr_kentry *ke) ke 842 net/pf_table.c if (ke->pfrke_intrpool) ke 843 net/pf_table.c pool_put(&pfr_kentry_pl2, ke); ke 845 net/pf_table.c pool_put(&pfr_kentry_pl, ke); ke 974 net/pf_table.c pfr_route_kentry(struct pfr_ktable *kt, struct pfr_kentry *ke) ke 981 net/pf_table.c bzero(ke->pfrke_node, sizeof(ke->pfrke_node)); ke 982 net/pf_table.c if (ke->pfrke_af == AF_INET) ke 984 net/pf_table.c else if (ke->pfrke_af == AF_INET6) ke 988 net/pf_table.c if (KENTRY_NETWORK(ke)) { ke 989 net/pf_table.c pfr_prepare_network(&mask, ke->pfrke_af, ke->pfrke_net); ke 990 net/pf_table.c rn = rn_addroute(&ke->pfrke_sa, &mask, head, ke->pfrke_node); ke 992 net/pf_table.c rn = rn_addroute(&ke->pfrke_sa, NULL, head, ke->pfrke_node); ke 999 net/pf_table.c pfr_unroute_kentry(struct pfr_ktable *kt, struct pfr_kentry *ke) ke 1006 net/pf_table.c if (ke->pfrke_af == AF_INET) ke 1008 net/pf_table.c else if (ke->pfrke_af == AF_INET6) ke 1012 net/pf_table.c if (KENTRY_NETWORK(ke)) { ke 1013 net/pf_table.c pfr_prepare_network(&mask, ke->pfrke_af, ke->pfrke_net); ke 1014 net/pf_table.c rn = rn_delete(&ke->pfrke_sa, &mask, head, NULL); ke 1016 net/pf_table.c rn = rn_delete(&ke->pfrke_sa, NULL, head, NULL); ke 1027 net/pf_table.c pfr_copyout_addr(struct pfr_addr *ad, struct pfr_kentry *ke) ke 1030 net/pf_table.c if (ke == NULL) ke 1032 net/pf_table.c ad->pfra_af = ke->pfrke_af; ke 1033 net/pf_table.c ad->pfra_net = ke->pfrke_net; ke 1034 net/pf_table.c ad->pfra_not = ke->pfrke_not; ke 1036 net/pf_table.c ad->pfra_ip4addr = ke->pfrke_sa.sin.sin_addr; ke 1038 net/pf_table.c ad->pfra_ip6addr = ke->pfrke_sa.sin6.sin6_addr; ke 1044 net/pf_table.c struct pfr_kentry *ke = (struct pfr_kentry *)rn; ke 1050 net/pf_table.c ke->pfrke_mark = 0; ke 1053 net/pf_table.c if (ke->pfrke_mark) ke 1057 net/pf_table.c SLIST_INSERT_HEAD(w->pfrw_workq, ke, pfrke_workq); ke 1064 net/pf_table.c pfr_copyout_addr(&ad, ke); ke 1074 net/pf_table.c pfr_copyout_addr(&as.pfras_a, ke); ke 1077 net/pf_table.c bcopy(ke->pfrke_packets, as.pfras_packets, ke 1079 net/pf_table.c bcopy(ke->pfrke_bytes, as.pfras_bytes, ke 1082 net/pf_table.c as.pfras_tzero = ke->pfrke_tzero; ke 1090 net/pf_table.c if (ke->pfrke_not) ke 1093 net/pf_table.c w->pfrw_kentry = ke; ke 1098 net/pf_table.c if (ke->pfrke_af == AF_INET) { ke 1101 net/pf_table.c pfr_prepare_network(&pfr_mask, AF_INET, ke->pfrke_net); ke 1103 net/pf_table.c &ke->pfrke_sa, AF_INET); ke 1106 net/pf_table.c } else if (ke->pfrke_af == AF_INET6){ ke 1109 net/pf_table.c pfr_prepare_network(&pfr_mask, AF_INET6, ke->pfrke_net); ke 1111 net/pf_table.c &ke->pfrke_sa, AF_INET6); ke 1973 net/pf_table.c struct pfr_kentry *ke = NULL; ke 1985 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin, kt->pfrkt_ip4); ke 1986 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 1987 net/pf_table.c ke = NULL; ke 1993 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin6, kt->pfrkt_ip6); ke 1994 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 1995 net/pf_table.c ke = NULL; ke 1999 net/pf_table.c match = (ke && !ke->pfrke_not); ke 2011 net/pf_table.c struct pfr_kentry *ke = NULL; ke 2022 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin, kt->pfrkt_ip4); ke 2023 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 2024 net/pf_table.c ke = NULL; ke 2030 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin6, kt->pfrkt_ip6); ke 2031 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke)) ke 2032 net/pf_table.c ke = NULL; ke 2038 net/pf_table.c if ((ke == NULL || ke->pfrke_not) != notrule) { ke 2045 net/pf_table.c if (ke != NULL && op_pass != PFR_OP_XPASS) { ke 2046 net/pf_table.c ke->pfrke_packets[dir_out][op_pass]++; ke 2047 net/pf_table.c ke->pfrke_bytes[dir_out][op_pass] += len; ke 2101 net/pf_table.c struct pfr_kentry *ke, *ke2; ke 2123 net/pf_table.c ke = pfr_kentry_byidx(kt, idx, af); ke 2124 net/pf_table.c if (ke == NULL) ke 2126 net/pf_table.c pfr_prepare_network(&pfr_mask, af, ke->pfrke_net); ke 2127 net/pf_table.c *raddr = SUNION2PF(&ke->pfrke_sa, af); ke 2144 net/pf_table.c if (!KENTRY_NETWORK(ke)) { ke 2159 net/pf_table.c if (ke2 == ke) { ke 104 nfs/nfs_kq.c struct kevq *ke; ke 112 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link) { ke 113 nfs/nfs_kq.c struct nfsnode *np = VTONFS(ke->vp); ke 117 nfs/nfs_kq.c VOP_PRINT(ke->vp); ke 120 nfs/nfs_kq.c if (nfs_getattrcache(ke->vp, &attr) != ENOENT) ke 127 nfs/nfs_kq.c ke->flags |= KEVQ_BUSY; ke 133 nfs/nfs_kq.c error = VOP_GETATTR(ke->vp, &attr, p->p_ucred, p); ke 136 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_DELETE); ke 144 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE ke 146 nfs/nfs_kq.c ke->omtime = attr.va_mtime; ke 147 nfs/nfs_kq.c } else if (attr.va_mtime.tv_sec != ke->omtime.tv_sec ke 148 nfs/nfs_kq.c || attr.va_mtime.tv_nsec != ke->omtime.tv_nsec) { ke 149 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE); ke 150 nfs/nfs_kq.c ke->omtime = attr.va_mtime; ke 153 nfs/nfs_kq.c if (attr.va_ctime.tv_sec != ke->octime.tv_sec ke 154 nfs/nfs_kq.c || attr.va_ctime.tv_nsec != ke->octime.tv_nsec) { ke 155 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_ATTRIB); ke 156 nfs/nfs_kq.c ke->octime = attr.va_ctime; ke 159 nfs/nfs_kq.c if (attr.va_nlink != ke->onlink) { ke 160 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_LINK); ke 161 nfs/nfs_kq.c ke->onlink = attr.va_nlink; ke 166 nfs/nfs_kq.c ke->flags &= ~KEVQ_BUSY; ke 167 nfs/nfs_kq.c if (ke->flags & KEVQ_WANT) { ke 168 nfs/nfs_kq.c ke->flags &= ~KEVQ_WANT; ke 169 nfs/nfs_kq.c wakeup(ke); ke 191 nfs/nfs_kq.c struct kevq *ke; ke 197 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link) { ke 198 nfs/nfs_kq.c if (ke->vp == vp) { ke 199 nfs/nfs_kq.c while (ke->flags & KEVQ_BUSY) { ke 200 nfs/nfs_kq.c ke->flags |= KEVQ_WANT; ke 202 nfs/nfs_kq.c (void) tsleep(ke, PSOCK, "nfskqdet", 0); ke 206 nfs/nfs_kq.c if (ke->usecount > 1) { ke 208 nfs/nfs_kq.c ke->usecount--; ke 211 nfs/nfs_kq.c SLIST_REMOVE(&kevlist, ke, kevq, kev_link); ke 212 nfs/nfs_kq.c FREE(ke, M_KEVENT); ke 265 nfs/nfs_kq.c struct kevq *ke; ke 313 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link) ke 314 nfs/nfs_kq.c if (ke->vp == vp) ke 317 nfs/nfs_kq.c if (ke) { ke 319 nfs/nfs_kq.c ke->usecount++; ke 322 nfs/nfs_kq.c MALLOC(ke, struct kevq *, ke 324 nfs/nfs_kq.c ke->vp = vp; ke 325 nfs/nfs_kq.c ke->usecount = 1; ke 326 nfs/nfs_kq.c ke->flags = 0; ke 327 nfs/nfs_kq.c ke->omtime = attr.va_mtime; ke 328 nfs/nfs_kq.c ke->octime = attr.va_ctime; ke 329 nfs/nfs_kq.c ke->onlink = attr.va_nlink; ke 330 nfs/nfs_kq.c SLIST_INSERT_HEAD(&kevlist, ke, kev_link);