prev_entry 1014 drivers/net/ethernet/3com/3c515.c struct boom_tx_desc *prev_entry; prev_entry 1021 drivers/net/ethernet/3com/3c515.c prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; prev_entry 1023 drivers/net/ethernet/3com/3c515.c prev_entry = NULL; prev_entry 1040 drivers/net/ethernet/3com/3c515.c if (prev_entry) prev_entry 1041 drivers/net/ethernet/3com/3c515.c prev_entry->next = isa_virt_to_bus(&vp->tx_ring[entry]); prev_entry 1054 drivers/net/ethernet/3com/3c515.c if (prev_entry) prev_entry 1055 drivers/net/ethernet/3com/3c515.c prev_entry->status &= ~0x80000000; prev_entry 2116 drivers/net/ethernet/3com/3c59x.c struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; prev_entry 2214 drivers/net/ethernet/3com/3c59x.c prev_entry->next = cpu_to_le32(vp->tx_ring_dma + entry * sizeof(struct boom_tx_desc)); prev_entry 2230 drivers/net/ethernet/3com/3c59x.c prev_entry->status &= cpu_to_le32(~TxIntrUploaded); prev_entry 524 drivers/net/ethernet/natsemi/sonic.c int prev_entry = lp->eol_rx; prev_entry 580 drivers/net/ethernet/natsemi/sonic.c prev_entry = entry; prev_entry 586 drivers/net/ethernet/natsemi/sonic.c if (prev_entry != lp->eol_rx) { prev_entry 588 drivers/net/ethernet/natsemi/sonic.c sonic_rda_put(dev, prev_entry, SONIC_RD_LINK, SONIC_EOL | prev_entry 589 drivers/net/ethernet/natsemi/sonic.c sonic_rda_get(dev, prev_entry, SONIC_RD_LINK)); prev_entry 592 drivers/net/ethernet/natsemi/sonic.c lp->eol_rx = prev_entry; prev_entry 391 fs/btrfs/extent_io.c struct tree_entry *prev_entry = NULL; prev_entry 396 fs/btrfs/extent_io.c prev_entry = entry; prev_entry 413 fs/btrfs/extent_io.c while (prev && offset > prev_entry->end) { prev_entry 415 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev_entry 422 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev_entry 423 fs/btrfs/extent_io.c while (prev && offset < prev_entry->start) { prev_entry 425 fs/btrfs/extent_io.c prev_entry = rb_entry(prev, struct tree_entry, rb_node); prev_entry 153 fs/btrfs/extent_map.c struct extent_map *prev_entry = NULL; prev_entry 158 fs/btrfs/extent_map.c prev_entry = entry; prev_entry 170 fs/btrfs/extent_map.c while (prev && offset >= extent_map_end(prev_entry)) { prev_entry 172 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev_entry 179 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev_entry 180 fs/btrfs/extent_map.c while (prev && offset < prev_entry->start) { prev_entry 182 fs/btrfs/extent_map.c prev_entry = rb_entry(prev, struct extent_map, rb_node); prev_entry 74 fs/btrfs/ordered-data.c struct btrfs_ordered_extent *prev_entry = NULL; prev_entry 79 fs/btrfs/ordered-data.c prev_entry = entry; prev_entry 91 fs/btrfs/ordered-data.c while (prev && file_offset >= entry_end(prev_entry)) { prev_entry 95 fs/btrfs/ordered-data.c prev_entry = rb_entry(test, struct btrfs_ordered_extent, prev_entry 97 fs/btrfs/ordered-data.c if (file_offset < entry_end(prev_entry)) prev_entry 103 fs/btrfs/ordered-data.c prev_entry = rb_entry(prev, struct btrfs_ordered_extent, prev_entry 105 fs/btrfs/ordered-data.c while (prev && file_offset < entry_end(prev_entry)) { prev_entry 109 fs/btrfs/ordered-data.c prev_entry = rb_entry(test, struct btrfs_ordered_extent, prev_entry 98 fs/f2fs/extent_cache.c struct rb_entry **prev_entry, prev_entry 110 fs/f2fs/extent_cache.c *prev_entry = NULL; prev_entry 151 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); prev_entry 158 fs/f2fs/extent_cache.c *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); prev_entry 3539 fs/f2fs/f2fs.h struct rb_entry **prev_entry, struct rb_entry **next_entry, prev_entry 321 fs/nilfs2/alloc.c &cache->prev_entry, &cache->lock); prev_entry 335 fs/nilfs2/alloc.c &cache->prev_entry, &cache->lock); prev_entry 851 fs/nilfs2/alloc.c brelse(cache->prev_entry.bh); prev_entry 854 fs/nilfs2/alloc.c cache->prev_entry.bh = NULL; prev_entry 89 fs/nilfs2/alloc.h struct nilfs_bh_assoc prev_entry; prev_entry 240 net/ipv4/cipso_ipv4.c struct cipso_v4_map_cache_entry *prev_entry = NULL; prev_entry 258 net/ipv4/cipso_ipv4.c if (!prev_entry) { prev_entry 263 net/ipv4/cipso_ipv4.c if (prev_entry->activity > 0) prev_entry 264 net/ipv4/cipso_ipv4.c prev_entry->activity -= 1; prev_entry 265 net/ipv4/cipso_ipv4.c if (entry->activity > prev_entry->activity && prev_entry 266 net/ipv4/cipso_ipv4.c entry->activity - prev_entry->activity > prev_entry 270 net/ipv4/cipso_ipv4.c prev_entry->list.prev, prev_entry 271 net/ipv4/cipso_ipv4.c &prev_entry->list); prev_entry 277 net/ipv4/cipso_ipv4.c prev_entry = entry; prev_entry 202 net/ipv6/calipso.c struct calipso_map_cache_entry *prev_entry = NULL; prev_entry 220 net/ipv6/calipso.c if (!prev_entry) { prev_entry 225 net/ipv6/calipso.c if (prev_entry->activity > 0) prev_entry 226 net/ipv6/calipso.c prev_entry->activity -= 1; prev_entry 227 net/ipv6/calipso.c if (entry->activity > prev_entry->activity && prev_entry 228 net/ipv6/calipso.c entry->activity - prev_entry->activity > prev_entry 232 net/ipv6/calipso.c prev_entry->list.prev, prev_entry 233 net/ipv6/calipso.c &prev_entry->list); prev_entry 239 net/ipv6/calipso.c prev_entry = entry;