ke 89 net/pf_table.c #define KENTRY_NETWORK(ke) ((ke)->pfrke_net < AF_BITS((ke)->pfrke_af))
ke 90 net/pf_table.c #define KENTRY_RNF_ROOT(ke) \
ke 91 net/pf_table.c ((((struct radix_node *)(ke))->rn_flags & RNF_ROOT) != 0)
ke 135 net/pf_table.c struct pfr_kentry *ke);
ke 776 net/pf_table.c struct pfr_kentry *ke;
ke 790 net/pf_table.c ke = (struct pfr_kentry *)rn_lookup(&sa, &mask, head);
ke 792 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 793 net/pf_table.c ke = NULL;
ke 795 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&sa, head);
ke 796 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 797 net/pf_table.c ke = NULL;
ke 798 net/pf_table.c if (exact && ke && KENTRY_NETWORK(ke))
ke 799 net/pf_table.c ke = NULL;
ke 801 net/pf_table.c return (ke);
ke 807 net/pf_table.c struct pfr_kentry *ke;
ke 810 net/pf_table.c ke = pool_get(&pfr_kentry_pl2, PR_NOWAIT);
ke 812 net/pf_table.c ke = pool_get(&pfr_kentry_pl, PR_NOWAIT);
ke 813 net/pf_table.c if (ke == NULL)
ke 815 net/pf_table.c bzero(ke, sizeof(*ke));
ke 818 net/pf_table.c FILLIN_SIN(ke->pfrke_sa.sin, ad->pfra_ip4addr);
ke 820 net/pf_table.c FILLIN_SIN6(ke->pfrke_sa.sin6, ad->pfra_ip6addr);
ke 821 net/pf_table.c ke->pfrke_af = ad->pfra_af;
ke 822 net/pf_table.c ke->pfrke_net = ad->pfra_net;
ke 823 net/pf_table.c ke->pfrke_not = ad->pfra_not;
ke 824 net/pf_table.c ke->pfrke_intrpool = intr;
ke 825 net/pf_table.c return (ke);
ke 840 net/pf_table.c pfr_destroy_kentry(struct pfr_kentry *ke)
ke 842 net/pf_table.c if (ke->pfrke_intrpool)
ke 843 net/pf_table.c pool_put(&pfr_kentry_pl2, ke);
ke 845 net/pf_table.c pool_put(&pfr_kentry_pl, ke);
ke 974 net/pf_table.c pfr_route_kentry(struct pfr_ktable *kt, struct pfr_kentry *ke)
ke 981 net/pf_table.c bzero(ke->pfrke_node, sizeof(ke->pfrke_node));
ke 982 net/pf_table.c if (ke->pfrke_af == AF_INET)
ke 984 net/pf_table.c else if (ke->pfrke_af == AF_INET6)
ke 988 net/pf_table.c if (KENTRY_NETWORK(ke)) {
ke 989 net/pf_table.c pfr_prepare_network(&mask, ke->pfrke_af, ke->pfrke_net);
ke 990 net/pf_table.c rn = rn_addroute(&ke->pfrke_sa, &mask, head, ke->pfrke_node);
ke 992 net/pf_table.c rn = rn_addroute(&ke->pfrke_sa, NULL, head, ke->pfrke_node);
ke 999 net/pf_table.c pfr_unroute_kentry(struct pfr_ktable *kt, struct pfr_kentry *ke)
ke 1006 net/pf_table.c if (ke->pfrke_af == AF_INET)
ke 1008 net/pf_table.c else if (ke->pfrke_af == AF_INET6)
ke 1012 net/pf_table.c if (KENTRY_NETWORK(ke)) {
ke 1013 net/pf_table.c pfr_prepare_network(&mask, ke->pfrke_af, ke->pfrke_net);
ke 1014 net/pf_table.c rn = rn_delete(&ke->pfrke_sa, &mask, head, NULL);
ke 1016 net/pf_table.c rn = rn_delete(&ke->pfrke_sa, NULL, head, NULL);
ke 1027 net/pf_table.c pfr_copyout_addr(struct pfr_addr *ad, struct pfr_kentry *ke)
ke 1030 net/pf_table.c if (ke == NULL)
ke 1032 net/pf_table.c ad->pfra_af = ke->pfrke_af;
ke 1033 net/pf_table.c ad->pfra_net = ke->pfrke_net;
ke 1034 net/pf_table.c ad->pfra_not = ke->pfrke_not;
ke 1036 net/pf_table.c ad->pfra_ip4addr = ke->pfrke_sa.sin.sin_addr;
ke 1038 net/pf_table.c ad->pfra_ip6addr = ke->pfrke_sa.sin6.sin6_addr;
ke 1044 net/pf_table.c struct pfr_kentry *ke = (struct pfr_kentry *)rn;
ke 1050 net/pf_table.c ke->pfrke_mark = 0;
ke 1053 net/pf_table.c if (ke->pfrke_mark)
ke 1057 net/pf_table.c SLIST_INSERT_HEAD(w->pfrw_workq, ke, pfrke_workq);
ke 1064 net/pf_table.c pfr_copyout_addr(&ad, ke);
ke 1074 net/pf_table.c pfr_copyout_addr(&as.pfras_a, ke);
ke 1077 net/pf_table.c bcopy(ke->pfrke_packets, as.pfras_packets,
ke 1079 net/pf_table.c bcopy(ke->pfrke_bytes, as.pfras_bytes,
ke 1082 net/pf_table.c as.pfras_tzero = ke->pfrke_tzero;
ke 1090 net/pf_table.c if (ke->pfrke_not)
ke 1093 net/pf_table.c w->pfrw_kentry = ke;
ke 1098 net/pf_table.c if (ke->pfrke_af == AF_INET) {
ke 1101 net/pf_table.c pfr_prepare_network(&pfr_mask, AF_INET, ke->pfrke_net);
ke 1103 net/pf_table.c &ke->pfrke_sa, AF_INET);
ke 1106 net/pf_table.c } else if (ke->pfrke_af == AF_INET6){
ke 1109 net/pf_table.c pfr_prepare_network(&pfr_mask, AF_INET6, ke->pfrke_net);
ke 1111 net/pf_table.c &ke->pfrke_sa, AF_INET6);
ke 1973 net/pf_table.c struct pfr_kentry *ke = NULL;
ke 1985 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin, kt->pfrkt_ip4);
ke 1986 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 1987 net/pf_table.c ke = NULL;
ke 1993 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin6, kt->pfrkt_ip6);
ke 1994 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 1995 net/pf_table.c ke = NULL;
ke 1999 net/pf_table.c match = (ke && !ke->pfrke_not);
ke 2011 net/pf_table.c struct pfr_kentry *ke = NULL;
ke 2022 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin, kt->pfrkt_ip4);
ke 2023 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 2024 net/pf_table.c ke = NULL;
ke 2030 net/pf_table.c ke = (struct pfr_kentry *)rn_match(&pfr_sin6, kt->pfrkt_ip6);
ke 2031 net/pf_table.c if (ke && KENTRY_RNF_ROOT(ke))
ke 2032 net/pf_table.c ke = NULL;
ke 2038 net/pf_table.c if ((ke == NULL || ke->pfrke_not) != notrule) {
ke 2045 net/pf_table.c if (ke != NULL && op_pass != PFR_OP_XPASS) {
ke 2046 net/pf_table.c ke->pfrke_packets[dir_out][op_pass]++;
ke 2047 net/pf_table.c ke->pfrke_bytes[dir_out][op_pass] += len;
ke 2101 net/pf_table.c struct pfr_kentry *ke, *ke2;
ke 2123 net/pf_table.c ke = pfr_kentry_byidx(kt, idx, af);
ke 2124 net/pf_table.c if (ke == NULL)
ke 2126 net/pf_table.c pfr_prepare_network(&pfr_mask, af, ke->pfrke_net);
ke 2127 net/pf_table.c *raddr = SUNION2PF(&ke->pfrke_sa, af);
ke 2144 net/pf_table.c if (!KENTRY_NETWORK(ke)) {
ke 2159 net/pf_table.c if (ke2 == ke) {
ke 104 nfs/nfs_kq.c struct kevq *ke;
ke 112 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link) {
ke 113 nfs/nfs_kq.c struct nfsnode *np = VTONFS(ke->vp);
ke 117 nfs/nfs_kq.c VOP_PRINT(ke->vp);
ke 120 nfs/nfs_kq.c if (nfs_getattrcache(ke->vp, &attr) != ENOENT)
ke 127 nfs/nfs_kq.c ke->flags |= KEVQ_BUSY;
ke 133 nfs/nfs_kq.c error = VOP_GETATTR(ke->vp, &attr, p->p_ucred, p);
ke 136 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_DELETE);
ke 144 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE
ke 146 nfs/nfs_kq.c ke->omtime = attr.va_mtime;
ke 147 nfs/nfs_kq.c } else if (attr.va_mtime.tv_sec != ke->omtime.tv_sec
ke 148 nfs/nfs_kq.c || attr.va_mtime.tv_nsec != ke->omtime.tv_nsec) {
ke 149 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_WRITE);
ke 150 nfs/nfs_kq.c ke->omtime = attr.va_mtime;
ke 153 nfs/nfs_kq.c if (attr.va_ctime.tv_sec != ke->octime.tv_sec
ke 154 nfs/nfs_kq.c || attr.va_ctime.tv_nsec != ke->octime.tv_nsec) {
ke 155 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_ATTRIB);
ke 156 nfs/nfs_kq.c ke->octime = attr.va_ctime;
ke 159 nfs/nfs_kq.c if (attr.va_nlink != ke->onlink) {
ke 160 nfs/nfs_kq.c VN_KNOTE(ke->vp, NOTE_LINK);
ke 161 nfs/nfs_kq.c ke->onlink = attr.va_nlink;
ke 166 nfs/nfs_kq.c ke->flags &= ~KEVQ_BUSY;
ke 167 nfs/nfs_kq.c if (ke->flags & KEVQ_WANT) {
ke 168 nfs/nfs_kq.c ke->flags &= ~KEVQ_WANT;
ke 169 nfs/nfs_kq.c wakeup(ke);
ke 191 nfs/nfs_kq.c struct kevq *ke;
ke 197 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link) {
ke 198 nfs/nfs_kq.c if (ke->vp == vp) {
ke 199 nfs/nfs_kq.c while (ke->flags & KEVQ_BUSY) {
ke 200 nfs/nfs_kq.c ke->flags |= KEVQ_WANT;
ke 202 nfs/nfs_kq.c (void) tsleep(ke, PSOCK, "nfskqdet", 0);
ke 206 nfs/nfs_kq.c if (ke->usecount > 1) {
ke 208 nfs/nfs_kq.c ke->usecount--;
ke 211 nfs/nfs_kq.c SLIST_REMOVE(&kevlist, ke, kevq, kev_link);
ke 212 nfs/nfs_kq.c FREE(ke, M_KEVENT);
ke 265 nfs/nfs_kq.c struct kevq *ke;
ke 313 nfs/nfs_kq.c SLIST_FOREACH(ke, &kevlist, kev_link)
ke 314 nfs/nfs_kq.c if (ke->vp == vp)
ke 317 nfs/nfs_kq.c if (ke) {
ke 319 nfs/nfs_kq.c ke->usecount++;
ke 322 nfs/nfs_kq.c MALLOC(ke, struct kevq *,
ke 324 nfs/nfs_kq.c ke->vp = vp;
ke 325 nfs/nfs_kq.c ke->usecount = 1;
ke 326 nfs/nfs_kq.c ke->flags = 0;
ke 327 nfs/nfs_kq.c ke->omtime = attr.va_mtime;
ke 328 nfs/nfs_kq.c ke->octime = attr.va_ctime;
ke 329 nfs/nfs_kq.c ke->onlink = attr.va_nlink;
ke 330 nfs/nfs_kq.c SLIST_INSERT_HEAD(&kevlist, ke, kev_link);