seqno 29 arch/powerpc/include/asm/hvsi.h __be16 seqno; seqno 68 arch/powerpc/include/asm/hvsi.h atomic_t seqno; /* packet sequence number */ seqno 69 arch/s390/include/uapi/asm/vtoc.h __u8 seqno; /* extent sequence number */ seqno 32 crypto/echainiv.c u64 seqno; seqno 66 crypto/echainiv.c seqno = be64_to_cpu(nseqno); seqno 77 crypto/echainiv.c a *= seqno; seqno 98 drivers/acpi/acpi_extlog.c static atomic_t seqno; seqno 108 drivers/acpi/acpi_extlog.c curr_seqno = atomic_inc_return(&seqno); seqno 541 drivers/acpi/apei/ghes.c static atomic_t seqno; seqno 552 drivers/acpi/apei/ghes.c curr_seqno = atomic_inc_return(&seqno); seqno 353 drivers/bluetooth/hci_bcsp.c u8 seqno; seqno 358 drivers/bluetooth/hci_bcsp.c seqno = bcsp->msgq_txseq; seqno 361 drivers/bluetooth/hci_bcsp.c if (bcsp->rxack == seqno) seqno 364 drivers/bluetooth/hci_bcsp.c seqno = (seqno - 1) & 0x07; seqno 367 drivers/bluetooth/hci_bcsp.c if (bcsp->rxack != seqno) seqno 372 drivers/bluetooth/hci_bcsp.c (seqno - 1) & 0x07); seqno 386 drivers/crypto/chelsio/chcr_ipsec.c __be64 seqno; seqno 412 drivers/crypto/chelsio/chcr_ipsec.c seqno = cpu_to_be64(seqlo + ((u64)xo->seq.hi << 32)); seqno 413 drivers/crypto/chelsio/chcr_ipsec.c memcpy(aadiv->seq_no, &seqno, 8); seqno 152 drivers/dma-buf/dma-fence-array.c u64 context, unsigned seqno, seqno 166 drivers/dma-buf/dma-fence-array.c context, seqno); seqno 89 drivers/dma-buf/dma-fence-chain.c int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t seqno) seqno 93 drivers/dma-buf/dma-fence-chain.c if (!seqno) seqno 97 drivers/dma-buf/dma-fence-chain.c if (!chain || chain->base.seqno < seqno) seqno 102 drivers/dma-buf/dma-fence-chain.c to_dma_fence_chain(*pfence)->prev_seqno < seqno) seqno 231 drivers/dma-buf/dma-fence-chain.c uint64_t seqno) seqno 243 drivers/dma-buf/dma-fence-chain.c if (prev_chain && __dma_fence_is_later(seqno, prev->seqno, prev->ops)) { seqno 245 drivers/dma-buf/dma-fence-chain.c chain->prev_seqno = prev->seqno; seqno 250 drivers/dma-buf/dma-fence-chain.c seqno = max(prev->seqno, seqno); seqno 254 drivers/dma-buf/dma-fence-chain.c &chain->lock, context, seqno); seqno 240 drivers/dma-buf/dma-fence.c fence->context, fence->seqno)) { seqno 647 drivers/dma-buf/dma-fence.c spinlock_t *lock, u64 context, u64 seqno) seqno 657 drivers/dma-buf/dma-fence.c fence->seqno = seqno; seqno 152 drivers/dma-buf/sw_sync.c return !__dma_fence_is_later(fence->seqno, parent->value, fence->ops); seqno 163 drivers/dma-buf/sw_sync.c snprintf(str, size, "%lld", fence->seqno); seqno 258 drivers/dma-buf/sw_sync.c cmp = value - other->base.seqno; seqno 142 drivers/dma-buf/sync_file.c fence->seqno); seqno 252 drivers/dma-buf/sync_file.c if (__dma_fence_is_later(pt_a->seqno, pt_b->seqno, seqno 227 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c unsigned seqno = ++adev->vm_manager.seqno[ring->idx]; seqno 236 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c seqno, true); seqno 392 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c i->fence->seqno, i->fence->context); seqno 171 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __field(unsigned int, seqno) seqno 181 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno = job->base.s_fence->finished.seqno; seqno 187 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno, __get_str(ring), __entry->num_ibs) seqno 197 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __field(unsigned int, seqno) seqno 206 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno = job->base.s_fence->finished.seqno; seqno 212 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno, __get_str(ring), __entry->num_ibs) seqno 475 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __field(unsigned, seqno) seqno 483 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno = fence->seqno; seqno 488 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h __entry->seqno) seqno 3012 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c adev->vm_manager.seqno[i] = 0; seqno 300 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h unsigned seqno[AMDGPU_MAX_RINGS]; seqno 941 drivers/gpu/drm/amd/amdkfd/kfd_device.c if (fence->seqno == p->last_eviction_seqno) seqno 944 drivers/gpu/drm/amd/amdkfd/kfd_device.c p->last_eviction_seqno = fence->seqno; seqno 1012 drivers/gpu/drm/amd/amdkfd/kfd_process.c WARN_ONCE(p->last_eviction_seqno != p->ef->seqno, seqno 220 drivers/gpu/drm/drm_dp_mst_topology.c buf[idx++] = (hdr->somt << 7) | (hdr->eomt << 6) | (hdr->seqno << 4); seqno 259 drivers/gpu/drm/drm_dp_mst_topology.c hdr->seqno = (buf[idx] >> 4) & 0x1; seqno 885 drivers/gpu/drm/drm_dp_mst_topology.c DRM_DEBUG_KMS("timedout msg send %p %d %d\n", txmsg, txmsg->state, txmsg->seqno); seqno 898 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[txmsg->seqno] = NULL; seqno 1963 drivers/gpu/drm/drm_dp_mst_topology.c if (txmsg->seqno == -1) { seqno 1969 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = mstb->last_seqno; seqno 1972 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = 0; seqno 1974 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = 1; seqno 1975 drivers/gpu/drm/drm_dp_mst_topology.c mstb->tx_slots[txmsg->seqno] = txmsg; seqno 1989 drivers/gpu/drm/drm_dp_mst_topology.c hdr->seqno = txmsg->seqno; seqno 2007 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = -1; seqno 2070 drivers/gpu/drm/drm_dp_mst_topology.c if (txmsg->seqno != -1) seqno 2071 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst->tx_slots[txmsg->seqno] = NULL; seqno 2089 drivers/gpu/drm/drm_dp_mst_topology.c if (txmsg->seqno != -1) { seqno 2090 drivers/gpu/drm/drm_dp_mst_topology.c WARN_ON((unsigned int)txmsg->seqno > seqno 2092 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->dst->tx_slots[txmsg->seqno] = NULL; seqno 2638 drivers/gpu/drm/drm_dp_mst_topology.c int req_type, int seqno, bool broadcast) seqno 2647 drivers/gpu/drm/drm_dp_mst_topology.c txmsg->seqno = seqno; seqno 2923 drivers/gpu/drm/drm_dp_mst_topology.c slot = mgr->down_rep_recv.initial_hdr.seqno; seqno 2932 drivers/gpu/drm/drm_dp_mst_topology.c mgr->down_rep_recv.initial_hdr.seqno, seqno 2977 drivers/gpu/drm/drm_dp_mst_topology.c bool seqno; seqno 2990 drivers/gpu/drm/drm_dp_mst_topology.c seqno = mgr->up_req_recv.initial_hdr.seqno; seqno 2994 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_up_ack_reply(mgr, mgr->mst_primary, msg.req_type, seqno, false); seqno 3011 drivers/gpu/drm/drm_dp_mst_topology.c drm_dp_send_up_ack_reply(mgr, mgr->mst_primary, msg.req_type, seqno, false); seqno 237 drivers/gpu/drm/drm_syncobj.c if (prev && prev->seqno >= point) seqno 1387 drivers/gpu/drm/drm_syncobj.c last_signaled->seqno : seqno 450 drivers/gpu/drm/etnaviv/etnaviv_gem.c fence->seqno); seqno 1030 drivers/gpu/drm/etnaviv/etnaviv_gpu.c return (s32)(f->gpu->completed_fence - f->base.seqno) >= 0; seqno 1434 drivers/gpu/drm/etnaviv/etnaviv_gpu.c if (fence_after(fence->seqno, gpu->completed_fence)) seqno 1435 drivers/gpu/drm/etnaviv/etnaviv_gpu.c gpu->completed_fence = fence->seqno; seqno 83 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); seqno 90 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c i915_seqno_passed(rq->fence.seqno, seqno 91 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_next_entry(rq, signal_link)->fence.seqno)) seqno 95 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c i915_seqno_passed(list_prev_entry(rq, signal_link)->fence.seqno, seqno 96 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c rq->fence.seqno)) seqno 308 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c if (i915_seqno_passed(rq->fence.seqno, it->fence.seqno)) seqno 365 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c rq->fence.context, rq->fence.seqno, seqno 1153 drivers/gpu/drm/i915/gt/intel_engine_cs.c rq->fence.context, rq->fence.seqno, seqno 720 drivers/gpu/drm/i915/gt/intel_lrc.c ports[0]->fence.seqno, seqno 725 drivers/gpu/drm/i915/gt/intel_lrc.c ports[1] ? ports[1]->fence.seqno : 0); seqno 1103 drivers/gpu/drm/i915/gt/intel_lrc.c last->fence.seqno, seqno 1130 drivers/gpu/drm/i915/gt/intel_lrc.c last->fence.seqno, seqno 1202 drivers/gpu/drm/i915/gt/intel_lrc.c rq->fence.seqno, seqno 1892 drivers/gpu/drm/i915/gt/intel_lrc.c *cs++ = rq->fence.seqno - 1; seqno 2945 drivers/gpu/drm/i915/gt/intel_lrc.c request->fence.seqno, seqno 2955 drivers/gpu/drm/i915/gt/intel_lrc.c request->fence.seqno, seqno 2974 drivers/gpu/drm/i915/gt/intel_lrc.c request->fence.seqno, seqno 3544 drivers/gpu/drm/i915/gt/intel_lrc.c rq->fence.context, rq->fence.seqno, seqno 3639 drivers/gpu/drm/i915/gt/intel_lrc.c rq->fence.seqno); seqno 122 drivers/gpu/drm/i915/gt/intel_reset.c rq->fence.seqno, seqno 719 drivers/gpu/drm/i915/gt/intel_reset.c engine->name, request->fence.context, request->fence.seqno); seqno 326 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 429 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 447 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 467 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 472 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 947 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 972 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = rq->fence.seqno; seqno 380 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb); seqno 382 drivers/gpu/drm/i915/gt/intel_timeline.c return tl->seqno += 1 + tl->has_initial_breadcrumb; seqno 387 drivers/gpu/drm/i915/gt/intel_timeline.c tl->seqno -= 1 + tl->has_initial_breadcrumb; seqno 393 drivers/gpu/drm/i915/gt/intel_timeline.c u32 *seqno) seqno 467 drivers/gpu/drm/i915/gt/intel_timeline.c *seqno = timeline_advance(tl); seqno 468 drivers/gpu/drm/i915/gt/intel_timeline.c GEM_BUG_ON(i915_seqno_passed(*tl->hwsp_seqno, *seqno)); seqno 482 drivers/gpu/drm/i915/gt/intel_timeline.c u32 *seqno) seqno 484 drivers/gpu/drm/i915/gt/intel_timeline.c *seqno = timeline_advance(tl); seqno 487 drivers/gpu/drm/i915/gt/intel_timeline.c if (unlikely(!*seqno && tl->hwsp_cacheline)) seqno 488 drivers/gpu/drm/i915/gt/intel_timeline.c return __intel_timeline_get_seqno(tl, rq, seqno); seqno 56 drivers/gpu/drm/i915/gt/intel_timeline.h u64 context, u32 seqno) seqno 58 drivers/gpu/drm/i915/gt/intel_timeline.h return i915_syncmap_set(&tl->sync, context, seqno); seqno 64 drivers/gpu/drm/i915/gt/intel_timeline.h return __intel_timeline_sync_set(tl, fence->context, fence->seqno); seqno 68 drivers/gpu/drm/i915/gt/intel_timeline.h u64 context, u32 seqno) seqno 70 drivers/gpu/drm/i915/gt/intel_timeline.h return i915_syncmap_is_later(&tl->sync, context, seqno); seqno 76 drivers/gpu/drm/i915/gt/intel_timeline.h return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno); seqno 83 drivers/gpu/drm/i915/gt/intel_timeline.h u32 *seqno); seqno 24 drivers/gpu/drm/i915/gt/intel_timeline_types.h u32 seqno; seqno 49 drivers/gpu/drm/i915/gt/selftest_hangcheck.c u32 *seqno; seqno 85 drivers/gpu/drm/i915/gt/selftest_hangcheck.c h->seqno = memset(vaddr, 0xff, PAGE_SIZE); seqno 194 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = rq->fence.seqno; seqno 208 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = rq->fence.seqno; seqno 221 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = rq->fence.seqno; seqno 233 drivers/gpu/drm/i915/gt/selftest_hangcheck.c *batch++ = rq->fence.seqno; seqno 272 drivers/gpu/drm/i915/gt/selftest_hangcheck.c return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]); seqno 294 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq->fence.seqno), seqno 297 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq->fence.seqno), seqno 612 drivers/gpu/drm/i915/gt/selftest_hangcheck.c __func__, rq->fence.seqno, hws_seqno(&h, rq)); seqno 700 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq->fence.seqno); seqno 874 drivers/gpu/drm/i915/gt/selftest_hangcheck.c __func__, rq->fence.seqno, hws_seqno(&h, rq)); seqno 1082 drivers/gpu/drm/i915/gt/selftest_hangcheck.c __func__, rq->fence.seqno, hws_seqno(&h, rq)); seqno 1271 drivers/gpu/drm/i915/gt/selftest_hangcheck.c __func__, rq->fence.seqno, hws_seqno(&h, rq)); seqno 1469 drivers/gpu/drm/i915/gt/selftest_hangcheck.c prev->fence.seqno, hws_seqno(&h, prev)); seqno 1575 drivers/gpu/drm/i915/gt/selftest_hangcheck.c __func__, rq->fence.seqno, hws_seqno(&h, rq)); seqno 1664 drivers/gpu/drm/i915/gt/selftest_hangcheck.c rq->fence.seqno, hws_seqno(&h, rq)); seqno 993 drivers/gpu/drm/i915/gt/selftest_lrc.c rq->fence.seqno = 1; seqno 994 drivers/gpu/drm/i915/gt/selftest_lrc.c BUILD_BUG_ON(sizeof(rq->fence.seqno) != 8); /* upper 32b == 0 */ seqno 995 drivers/gpu/drm/i915/gt/selftest_lrc.c rq->hwsp_seqno = (u32 *)&rq->fence.seqno + 1; seqno 1755 drivers/gpu/drm/i915/gt/selftest_lrc.c request[nc]->fence.seqno); seqno 1760 drivers/gpu/drm/i915/gt/selftest_lrc.c request[nc]->fence.seqno); seqno 1903 drivers/gpu/drm/i915/gt/selftest_lrc.c request[n]->fence.seqno); seqno 1908 drivers/gpu/drm/i915/gt/selftest_lrc.c request[n]->fence.seqno); seqno 161 drivers/gpu/drm/i915/gt/selftest_timeline.c u32 seqno; seqno 173 drivers/gpu/drm/i915/gt/selftest_timeline.c if (__intel_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) { seqno 175 drivers/gpu/drm/i915/gt/selftest_timeline.c name, p->name, ctx, p->seqno, yesno(p->expected)); seqno 180 drivers/gpu/drm/i915/gt/selftest_timeline.c ret = __intel_timeline_sync_set(tl, ctx, p->seqno); seqno 358 drivers/gpu/drm/i915/gt/selftest_timeline.c u32 seqno = prandom_u32_state(&prng); seqno 360 drivers/gpu/drm/i915/gt/selftest_timeline.c if (!__intel_timeline_sync_is_later(&tl, id, seqno)) seqno 361 drivers/gpu/drm/i915/gt/selftest_timeline.c __intel_timeline_sync_set(&tl, id, seqno); seqno 486 drivers/gpu/drm/i915/gt/selftest_timeline.c if (*tl->hwsp_seqno != tl->seqno) { seqno 488 drivers/gpu/drm/i915/gt/selftest_timeline.c *tl->hwsp_seqno, tl->seqno); seqno 679 drivers/gpu/drm/i915/gt/selftest_timeline.c u32 seqno[2]; seqno 690 drivers/gpu/drm/i915/gt/selftest_timeline.c tl->seqno = -4u; seqno 693 drivers/gpu/drm/i915/gt/selftest_timeline.c err = intel_timeline_get_seqno(tl, rq, &seqno[0]); seqno 700 drivers/gpu/drm/i915/gt/selftest_timeline.c seqno[0], tl->hwsp_offset); seqno 702 drivers/gpu/drm/i915/gt/selftest_timeline.c err = emit_ggtt_store_dw(rq, tl->hwsp_offset, seqno[0]); seqno 710 drivers/gpu/drm/i915/gt/selftest_timeline.c err = intel_timeline_get_seqno(tl, rq, &seqno[1]); seqno 717 drivers/gpu/drm/i915/gt/selftest_timeline.c seqno[1], tl->hwsp_offset); seqno 719 drivers/gpu/drm/i915/gt/selftest_timeline.c err = emit_ggtt_store_dw(rq, tl->hwsp_offset, seqno[1]); seqno 727 drivers/gpu/drm/i915/gt/selftest_timeline.c GEM_BUG_ON(seqno[1] >= seqno[0]); seqno 738 drivers/gpu/drm/i915/gt/selftest_timeline.c if (*hwsp_seqno[0] != seqno[0] || *hwsp_seqno[1] != seqno[1]) { seqno 741 drivers/gpu/drm/i915/gt/selftest_timeline.c seqno[0], seqno[1]); seqno 471 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c ring_tail, rq->fence.seqno); seqno 2321 drivers/gpu/drm/i915/i915_drv.h int __must_check i915_gem_set_global_seqno(struct drm_device *dev, u32 seqno); seqno 455 drivers/gpu/drm/i915/i915_gpu_error.c if (!erq->seqno) seqno 459 drivers/gpu/drm/i915/i915_gpu_error.c prefix, erq->pid, erq->context, erq->seqno, seqno 1176 drivers/gpu/drm/i915/i915_gpu_error.c erq->seqno = request->fence.seqno; seqno 147 drivers/gpu/drm/i915/i915_gpu_error.h u32 seqno; seqno 228 drivers/gpu/drm/i915/i915_request.c rq->fence.context, rq->fence.seqno, seqno 324 drivers/gpu/drm/i915/i915_request.c rq->fence.context, rq->fence.seqno, seqno 387 drivers/gpu/drm/i915/i915_request.c request->fence.context, request->fence.seqno, seqno 479 drivers/gpu/drm/i915/i915_request.c request->fence.context, request->fence.seqno, seqno 639 drivers/gpu/drm/i915/i915_request.c u32 seqno; seqno 686 drivers/gpu/drm/i915/i915_request.c ret = intel_timeline_get_seqno(tl, rq, &seqno); seqno 702 drivers/gpu/drm/i915/i915_request.c tl->fence_context, seqno); seqno 879 drivers/gpu/drm/i915/i915_request.c *cs++ = from->fence.seqno; seqno 1176 drivers/gpu/drm/i915/i915_request.c GEM_BUG_ON(timeline->seqno != rq->fence.seqno); seqno 1194 drivers/gpu/drm/i915/i915_request.c engine->name, rq->fence.context, rq->fence.seqno); seqno 358 drivers/gpu/drm/i915/i915_request.h u32 seqno; seqno 361 drivers/gpu/drm/i915/i915_request.h seqno = __hwsp_seqno(rq); seqno 364 drivers/gpu/drm/i915/i915_request.h return seqno; seqno 369 drivers/gpu/drm/i915/i915_request.h return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno - 1); seqno 428 drivers/gpu/drm/i915/i915_request.h return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno); seqno 433 drivers/gpu/drm/i915/i915_request.h rq->hwsp_seqno = (u32 *)&rq->fence.seqno; /* decouple from HWSP */ seqno 402 drivers/gpu/drm/i915/i915_sw_fence.c cb->dma->seqno, seqno 154 drivers/gpu/drm/i915/i915_syncmap.c bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno) seqno 195 drivers/gpu/drm/i915/i915_syncmap.c return seqno_later(__sync_seqno(p)[idx], seqno); seqno 214 drivers/gpu/drm/i915/i915_syncmap.c static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno) seqno 219 drivers/gpu/drm/i915/i915_syncmap.c __sync_seqno(p)[idx] = seqno; seqno 230 drivers/gpu/drm/i915/i915_syncmap.c static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno) seqno 335 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_seqno(p, id, seqno); seqno 353 drivers/gpu/drm/i915/i915_syncmap.c int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno) seqno 362 drivers/gpu/drm/i915/i915_syncmap.c __sync_set_seqno(p, id, seqno); seqno 366 drivers/gpu/drm/i915/i915_syncmap.c return __sync_set(root, id, seqno); seqno 34 drivers/gpu/drm/i915/i915_syncmap.h int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno); seqno 35 drivers/gpu/drm/i915/i915_syncmap.h bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno); seqno 672 drivers/gpu/drm/i915/i915_trace.h __field(u32, seqno) seqno 682 drivers/gpu/drm/i915/i915_trace.h __entry->seqno = rq->fence.seqno; seqno 688 drivers/gpu/drm/i915/i915_trace.h __entry->hw_id, __entry->ctx, __entry->seqno, seqno 702 drivers/gpu/drm/i915/i915_trace.h __field(u32, seqno) seqno 711 drivers/gpu/drm/i915/i915_trace.h __entry->seqno = rq->fence.seqno; seqno 716 drivers/gpu/drm/i915/i915_trace.h __entry->hw_id, __entry->ctx, __entry->seqno) seqno 745 drivers/gpu/drm/i915/i915_trace.h __field(u32, seqno) seqno 756 drivers/gpu/drm/i915/i915_trace.h __entry->seqno = rq->fence.seqno; seqno 763 drivers/gpu/drm/i915/i915_trace.h __entry->hw_id, __entry->ctx, __entry->seqno, seqno 777 drivers/gpu/drm/i915/i915_trace.h __field(u32, seqno) seqno 787 drivers/gpu/drm/i915/i915_trace.h __entry->seqno = rq->fence.seqno; seqno 793 drivers/gpu/drm/i915/i915_trace.h __entry->hw_id, __entry->ctx, __entry->seqno, seqno 836 drivers/gpu/drm/i915/i915_trace.h __field(u32, seqno) seqno 852 drivers/gpu/drm/i915/i915_trace.h __entry->seqno = rq->fence.seqno; seqno 858 drivers/gpu/drm/i915/i915_trace.h __entry->hw_id, __entry->ctx, __entry->seqno, seqno 385 drivers/gpu/drm/i915/selftests/i915_request.c rq->fence.context, rq->fence.seqno, seqno 401 drivers/gpu/drm/i915/selftests/i915_request.c rq->fence.context, rq->fence.seqno); seqno 146 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno) seqno 154 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (__sync_seqno(leaf)[idx] != seqno) { seqno 156 drivers/gpu/drm/i915/selftests/i915_syncmap.c __func__, idx, __sync_seqno(leaf)[idx], seqno); seqno 163 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) seqno 167 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = i915_syncmap_set(sync, context, seqno); seqno 189 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); seqno 193 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (!i915_syncmap_is_later(sync, context, seqno)) { seqno 195 drivers/gpu/drm/i915/selftests/i915_syncmap.c context, seqno); seqno 238 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno) seqno 242 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = i915_syncmap_set(sync, context, seqno); seqno 258 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno); seqno 262 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (!i915_syncmap_is_later(sync, context, seqno)) { seqno 264 drivers/gpu/drm/i915/selftests/i915_syncmap.c context, seqno); seqno 548 drivers/gpu/drm/i915/selftests/i915_syncmap.c u32 seqno; seqno 570 drivers/gpu/drm/i915/selftests/i915_syncmap.c seqno = 0; seqno 575 drivers/gpu/drm/i915/selftests/i915_syncmap.c u32 last_seqno = seqno; seqno 578 drivers/gpu/drm/i915/selftests/i915_syncmap.c seqno = prandom_u32_state(&prng); seqno 579 drivers/gpu/drm/i915/selftests/i915_syncmap.c expect = seqno_later(last_seqno, seqno); seqno 584 drivers/gpu/drm/i915/selftests/i915_syncmap.c if (i915_syncmap_is_later(&sync, context, seqno) != expect) { seqno 586 drivers/gpu/drm/i915/selftests/i915_syncmap.c context, last_seqno, seqno, expect); seqno 591 drivers/gpu/drm/i915/selftests/i915_syncmap.c err = i915_syncmap_set(&sync, context, seqno); seqno 41 drivers/gpu/drm/i915/selftests/igt_spinner.c spin->seqno = memset(vaddr, 0xff, PAGE_SIZE); seqno 138 drivers/gpu/drm/i915/selftests/igt_spinner.c *batch++ = rq->fence.seqno; seqno 173 drivers/gpu/drm/i915/selftests/igt_spinner.c u32 *seqno = spin->seqno + seqno_offset(rq->fence.context); seqno 175 drivers/gpu/drm/i915/selftests/igt_spinner.c return READ_ONCE(*seqno); seqno 198 drivers/gpu/drm/i915/selftests/igt_spinner.c rq->fence.seqno), seqno 201 drivers/gpu/drm/i915/selftests/igt_spinner.c rq->fence.seqno), seqno 24 drivers/gpu/drm/i915/selftests/igt_spinner.h void *seqno; seqno 102 drivers/gpu/drm/msm/adreno/a5xx_gpu.c ring->memptrs->fence = submit->seqno; seqno 182 drivers/gpu/drm/msm/adreno/a5xx_gpu.c OUT_RING(ring, submit->seqno); seqno 192 drivers/gpu/drm/msm/adreno/a5xx_gpu.c OUT_RING(ring, submit->seqno); seqno 964 drivers/gpu/drm/msm/adreno/a5xx_gpu.c ring ? ring->id : -1, ring ? ring->seqno : 0, seqno 85 drivers/gpu/drm/msm/adreno/a6xx_gpu.c unsigned int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT; seqno 135 drivers/gpu/drm/msm/adreno/a6xx_gpu.c OUT_RING(ring, submit->seqno); seqno 145 drivers/gpu/drm/msm/adreno/a6xx_gpu.c OUT_RING(ring, submit->seqno); seqno 677 drivers/gpu/drm/msm/adreno/a6xx_gpu.c ring ? ring->id : -1, ring ? ring->seqno : 0, seqno 353 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->memptrs->fence = ring->seqno; seqno 443 drivers/gpu/drm/msm/adreno/adreno_gpu.c OUT_RING(ring, submit->seqno); seqno 463 drivers/gpu/drm/msm/adreno/adreno_gpu.c OUT_RING(ring, submit->seqno); seqno 469 drivers/gpu/drm/msm/adreno/adreno_gpu.c OUT_RING(ring, submit->seqno); seqno 537 drivers/gpu/drm/msm/adreno/adreno_gpu.c state->ring[i].seqno = gpu->rb[i]->seqno; seqno 716 drivers/gpu/drm/msm/adreno/adreno_gpu.c drm_printf(p, " last-fence: %d\n", state->ring[i].seqno); seqno 772 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->seqno); seqno 114 drivers/gpu/drm/msm/msm_fence.c return fence_completed(f->fctx, f->base.seqno); seqno 796 drivers/gpu/drm/msm/msm_gem.c fence->seqno); seqno 137 drivers/gpu/drm/msm/msm_gem.h uint32_t seqno; /* Sequence number of the submit on the ring */ seqno 582 drivers/gpu/drm/msm/msm_gem_submit.c args->fence = submit->fence->seqno; seqno 395 drivers/gpu/drm/msm/msm_gpu.c if (submit->seqno > fence) seqno 399 drivers/gpu/drm/msm/msm_gpu.c submit->fence->seqno); seqno 411 drivers/gpu/drm/msm/msm_gpu.c if (submit->seqno == fence) seqno 529 drivers/gpu/drm/msm/msm_gpu.c } else if (fence < ring->seqno) { seqno 537 drivers/gpu/drm/msm/msm_gpu.c gpu->name, ring->seqno); seqno 543 drivers/gpu/drm/msm/msm_gpu.c if (ring->seqno > ring->hangcheck_fence) seqno 655 drivers/gpu/drm/msm/msm_gpu.c int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT; seqno 743 drivers/gpu/drm/msm/msm_gpu.c submit->seqno = ++ring->seqno; seqno 151 drivers/gpu/drm/msm/msm_gpu.h if (ring->seqno > ring->memptrs->fence) seqno 194 drivers/gpu/drm/msm/msm_gpu.h u32 seqno; seqno 40 drivers/gpu/drm/msm/msm_gpu_trace.h __field(u32, seqno) seqno 47 drivers/gpu/drm/msm/msm_gpu_trace.h __entry->seqno = submit->seqno; seqno 51 drivers/gpu/drm/msm/msm_gpu_trace.h __entry->id, __entry->pid, __entry->ringid, __entry->seqno, seqno 64 drivers/gpu/drm/msm/msm_gpu_trace.h __field(u32, seqno) seqno 74 drivers/gpu/drm/msm/msm_gpu_trace.h __entry->seqno = submit->seqno; seqno 81 drivers/gpu/drm/msm/msm_gpu_trace.h __entry->id, __entry->pid, __entry->ringid, __entry->seqno, seqno 374 drivers/gpu/drm/msm/msm_rd.c pid_nr(submit->pid), submit->seqno); seqno 377 drivers/gpu/drm/msm/msm_rd.c pid_nr(submit->pid), submit->seqno); seqno 43 drivers/gpu/drm/msm/msm_ringbuffer.h uint32_t seqno; seqno 135 drivers/gpu/drm/nouveau/nouveau_fence.c if ((int)(seq - fence->base.seqno) < 0) seqno 456 drivers/gpu/drm/nouveau/nouveau_fence.c ret = (int)(fctx->read(chan) - fence->base.seqno) >= 0; seqno 46 drivers/gpu/drm/nouveau/nv04_fence.c OUT_RING (chan, fence->base.seqno); seqno 36 drivers/gpu/drm/nouveau/nv10_fence.c OUT_RING (chan, fence->base.seqno); seqno 74 drivers/gpu/drm/nouveau/nv84_fence.c return fctx->base.emit32(chan, addr, fence->base.seqno); seqno 84 drivers/gpu/drm/nouveau/nv84_fence.c return fctx->base.sync32(chan, addr, fence->base.seqno); seqno 48 drivers/gpu/drm/panfrost/panfrost_job.c u64 seqno; seqno 95 drivers/gpu/drm/panfrost/panfrost_job.c fence->seqno = ++js->queue[js_num].emit_seqno; seqno 97 drivers/gpu/drm/panfrost/panfrost_job.c js->queue[js_num].fence_context, fence->seqno); seqno 145 drivers/gpu/drm/qxl/qxl_release.c release->base.seqno = ++qdev->release_seqno; seqno 451 drivers/gpu/drm/qxl/qxl_release.c release->id | 0xf0000000, release->base.seqno); seqno 127 drivers/gpu/drm/radeon/radeon_trace.h TP_PROTO(struct drm_device *dev, int ring, u32 seqno), seqno 129 drivers/gpu/drm/radeon/radeon_trace.h TP_ARGS(dev, ring, seqno), seqno 134 drivers/gpu/drm/radeon/radeon_trace.h __field(u32, seqno) seqno 140 drivers/gpu/drm/radeon/radeon_trace.h __entry->seqno = seqno; seqno 144 drivers/gpu/drm/radeon/radeon_trace.h __entry->dev, __entry->ring, __entry->seqno) seqno 149 drivers/gpu/drm/radeon/radeon_trace.h TP_PROTO(struct drm_device *dev, int ring, u32 seqno), seqno 151 drivers/gpu/drm/radeon/radeon_trace.h TP_ARGS(dev, ring, seqno) seqno 156 drivers/gpu/drm/radeon/radeon_trace.h TP_PROTO(struct drm_device *dev, int ring, u32 seqno), seqno 158 drivers/gpu/drm/radeon/radeon_trace.h TP_ARGS(dev, ring, seqno) seqno 163 drivers/gpu/drm/radeon/radeon_trace.h TP_PROTO(struct drm_device *dev, int ring, u32 seqno), seqno 165 drivers/gpu/drm/radeon/radeon_trace.h TP_ARGS(dev, ring, seqno) seqno 83 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __field(unsigned, seqno) seqno 91 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->seqno = fence->seqno; seqno 96 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h __entry->seqno) seqno 161 drivers/gpu/drm/v3d/v3d_drv.h u64 seqno; seqno 16 drivers/gpu/drm/v3d/v3d_fence.c fence->seqno = ++v3d->queue[queue].emit_seqno; seqno 18 drivers/gpu/drm/v3d/v3d_fence.c v3d->queue[queue].fence_context, fence->seqno); seqno 104 drivers/gpu/drm/v3d/v3d_irq.c trace_v3d_bcl_irq(&v3d->drm, fence->seqno); seqno 113 drivers/gpu/drm/v3d/v3d_irq.c trace_v3d_rcl_irq(&v3d->drm, fence->seqno); seqno 122 drivers/gpu/drm/v3d/v3d_irq.c trace_v3d_csd_irq(&v3d->drm, fence->seqno); seqno 158 drivers/gpu/drm/v3d/v3d_irq.c trace_v3d_tfu_irq(&v3d->drm, fence->seqno); seqno 120 drivers/gpu/drm/v3d/v3d_sched.c trace_v3d_submit_cl(dev, false, to_v3d_fence(fence)->seqno, seqno 169 drivers/gpu/drm/v3d/v3d_sched.c trace_v3d_submit_cl(dev, true, to_v3d_fence(fence)->seqno, seqno 200 drivers/gpu/drm/v3d/v3d_sched.c trace_v3d_submit_tfu(dev, to_v3d_fence(fence)->seqno); seqno 241 drivers/gpu/drm/v3d/v3d_sched.c trace_v3d_submit_csd(dev, to_v3d_fence(fence)->seqno); seqno 39 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno, seqno 41 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, is_render, seqno, ctnqba, ctnqea), seqno 46 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 54 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 62 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno, seqno 69 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 70 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 74 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 79 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 84 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 89 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 90 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 94 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 99 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 104 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 109 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 110 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 114 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 119 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 124 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 129 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 130 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 134 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 139 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 144 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 168 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 169 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 173 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 178 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 183 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 210 drivers/gpu/drm/v3d/v3d_trace.h uint64_t seqno), seqno 211 drivers/gpu/drm/v3d/v3d_trace.h TP_ARGS(dev, seqno), seqno 215 drivers/gpu/drm/v3d/v3d_trace.h __field(u64, seqno) seqno 220 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno = seqno; seqno 225 drivers/gpu/drm/v3d/v3d_trace.h __entry->seqno) seqno 956 drivers/gpu/drm/vc4/vc4_crtc.c vc4_queue_seqno_cb(dev, &flip_state->cb, bo->seqno, seqno 240 drivers/gpu/drm/vc4/vc4_drv.h uint64_t seqno; seqno 293 drivers/gpu/drm/vc4/vc4_drv.h uint64_t seqno; seqno 304 drivers/gpu/drm/vc4/vc4_drv.h uint64_t seqno; seqno 487 drivers/gpu/drm/vc4/vc4_drv.h uint64_t seqno; seqno 805 drivers/gpu/drm/vc4/vc4_drv.h int vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, seqno 809 drivers/gpu/drm/vc4/vc4_drv.h struct vc4_seqno_cb *cb, uint64_t seqno, seqno 41 drivers/gpu/drm/vc4/vc4_fence.c return vc4->finished_seqno >= f->seqno; seqno 381 drivers/gpu/drm/vc4/vc4_gem.c vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, uint64_t timeout_ns, seqno 389 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->finished_seqno >= seqno) seqno 397 drivers/gpu/drm/vc4/vc4_gem.c trace_vc4_wait_for_seqno_begin(dev, seqno, timeout_ns); seqno 408 drivers/gpu/drm/vc4/vc4_gem.c if (vc4->finished_seqno >= seqno) seqno 423 drivers/gpu/drm/vc4/vc4_gem.c trace_vc4_wait_for_seqno_end(dev, seqno); seqno 537 drivers/gpu/drm/vc4/vc4_gem.c vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno) seqno 544 drivers/gpu/drm/vc4/vc4_gem.c bo->seqno = seqno; seqno 550 drivers/gpu/drm/vc4/vc4_gem.c bo->seqno = seqno; seqno 555 drivers/gpu/drm/vc4/vc4_gem.c bo->write_seqno = seqno; seqno 671 drivers/gpu/drm/vc4/vc4_gem.c uint64_t seqno; seqno 682 drivers/gpu/drm/vc4/vc4_gem.c seqno = ++vc4->emit_seqno; seqno 683 drivers/gpu/drm/vc4/vc4_gem.c exec->seqno = seqno; seqno 686 drivers/gpu/drm/vc4/vc4_gem.c vc4->dma_fence_context, exec->seqno); seqno 687 drivers/gpu/drm/vc4/vc4_gem.c fence->seqno = exec->seqno; seqno 693 drivers/gpu/drm/vc4/vc4_gem.c vc4_update_bo_seqnos(exec, seqno); seqno 1009 drivers/gpu/drm/vc4/vc4_gem.c if (cb->seqno <= vc4->finished_seqno) { seqno 1026 drivers/gpu/drm/vc4/vc4_gem.c struct vc4_seqno_cb *cb, uint64_t seqno, seqno 1037 drivers/gpu/drm/vc4/vc4_gem.c if (seqno > vc4->finished_seqno) { seqno 1038 drivers/gpu/drm/vc4/vc4_gem.c cb->seqno = seqno; seqno 1063 drivers/gpu/drm/vc4/vc4_gem.c uint64_t seqno, seqno 1067 drivers/gpu/drm/vc4/vc4_gem.c int ret = vc4_wait_for_seqno(dev, seqno, *timeout_ns, true); seqno 1085 drivers/gpu/drm/vc4/vc4_gem.c return vc4_wait_for_seqno_ioctl_helper(dev, args->seqno, seqno 1108 drivers/gpu/drm/vc4/vc4_gem.c ret = vc4_wait_for_seqno_ioctl_helper(dev, bo->seqno, seqno 1257 drivers/gpu/drm/vc4/vc4_gem.c args->seqno = vc4->emit_seqno; seqno 18 drivers/gpu/drm/vc4/vc4_trace.h TP_PROTO(struct drm_device *dev, uint64_t seqno, uint64_t timeout), seqno 19 drivers/gpu/drm/vc4/vc4_trace.h TP_ARGS(dev, seqno, timeout), seqno 23 drivers/gpu/drm/vc4/vc4_trace.h __field(u64, seqno) seqno 29 drivers/gpu/drm/vc4/vc4_trace.h __entry->seqno = seqno; seqno 34 drivers/gpu/drm/vc4/vc4_trace.h __entry->dev, __entry->seqno, __entry->timeout) seqno 38 drivers/gpu/drm/vc4/vc4_trace.h TP_PROTO(struct drm_device *dev, uint64_t seqno), seqno 39 drivers/gpu/drm/vc4/vc4_trace.h TP_ARGS(dev, seqno), seqno 43 drivers/gpu/drm/vc4/vc4_trace.h __field(u64, seqno) seqno 48 drivers/gpu/drm/vc4/vc4_trace.h __entry->seqno = seqno; seqno 52 drivers/gpu/drm/vc4/vc4_trace.h __entry->dev, __entry->seqno) seqno 174 drivers/gpu/drm/vc4/vc4_v3d.c uint64_t seqno = 0; seqno 193 drivers/gpu/drm/vc4/vc4_v3d.c seqno = exec->seqno; seqno 196 drivers/gpu/drm/vc4/vc4_v3d.c if (seqno) { seqno 197 drivers/gpu/drm/vc4/vc4_v3d.c int ret = vc4_wait_for_seqno(dev, seqno, ~0ull, true); seqno 58 drivers/gpu/drm/vgem/vgem_fence.c snprintf(str, size, "%llu", fence->seqno); seqno 65 drivers/gpu/drm/vgem/vgem_fence.c dma_fence_is_signaled(fence) ? fence->seqno : 0); seqno 44 drivers/gpu/drm/virtio/virtgpu_fence.c if (atomic64_read(&fence->drv->last_seq) >= fence->f.seqno) seqno 51 drivers/gpu/drm/virtio/virtgpu_fence.c snprintf(str, size, "%llu", f->seqno); seqno 96 drivers/gpu/drm/virtio/virtgpu_fence.c fence->f.seqno = ++drv->sync_seq; seqno 104 drivers/gpu/drm/virtio/virtgpu_fence.c cmd_hdr->fence_id = cpu_to_le64(fence->f.seqno); seqno 117 drivers/gpu/drm/virtio/virtgpu_fence.c if (last_seq < fence->f.seqno) seqno 47 drivers/gpu/drm/virtio/virtgpu_object.c static atomic_t seqno = ATOMIC_INIT(0); seqno 48 drivers/gpu/drm/virtio/virtgpu_object.c int handle = atomic_inc_return(&seqno); seqno 887 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h uint32_t *seqno); seqno 1021 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h uint32_t seqno, bool interruptible, seqno 1026 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h uint32_t seqno); seqno 1030 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h uint32_t seqno, seqno 1052 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h uint32_t seqno); seqno 3436 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c fence_rep.seqno = fence->base.seqno; seqno 145 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c u32 seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); seqno 146 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c if (seqno - fence->base.seqno < VMW_FENCE_WRAP) seqno 340 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c struct vmw_fence_obj *fence, u32 seqno, seqno 346 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c fman->ctx, seqno); seqno 419 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c vmw_mmio_write(fence->base.seqno, seqno 456 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c goal_seqno - fence->base.seqno < VMW_FENCE_WRAP)) seqno 459 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c vmw_mmio_write(fence->base.seqno, fifo_mem + SVGA_FIFO_FENCE_GOAL); seqno 470 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c uint32_t seqno, new_seqno; seqno 473 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); seqno 476 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c if (seqno - fence->base.seqno < VMW_FENCE_WRAP) { seqno 493 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c needs_rerun = vmw_fence_goal_new_locked(fman, seqno); seqno 496 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c if (new_seqno != seqno) { seqno 497 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c seqno = new_seqno; seqno 551 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c uint32_t seqno, seqno 561 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c ret = vmw_fence_obj_init(fman, fence, seqno, seqno 602 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c uint32_t seqno, seqno 632 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c ret = vmw_fence_obj_init(fman, &ufence->fence, seqno, seqno 100 drivers/gpu/drm/vmwgfx/vmwgfx_fence.h uint32_t seqno, seqno 537 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c int vmw_fifo_send_fence(struct vmw_private *dev_priv, uint32_t *seqno) seqno 547 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c *seqno = atomic_read(&dev_priv->marker_seq); seqno 549 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c (void)vmw_fallback_wait(dev_priv, false, true, *seqno, seqno 555 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c *seqno = atomic_add_return(1, &dev_priv->marker_seq); seqno 556 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c } while (*seqno == 0); seqno 571 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c cmd_fence->fence = *seqno; seqno 573 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c (void) vmw_marker_push(&fifo_state->marker_queue, *seqno); seqno 111 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno) seqno 121 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c uint32_t seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE); seqno 123 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (dev_priv->last_read_seqno != seqno) { seqno 124 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c dev_priv->last_read_seqno = seqno; seqno 125 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c vmw_marker_pull(&fifo_state->marker_queue, seqno); seqno 131 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c uint32_t seqno) seqno 136 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) seqno 141 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) seqno 145 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c vmw_fifo_idle(dev_priv, seqno)) seqno 153 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c ret = ((atomic_read(&dev_priv->marker_seq) - seqno) seqno 162 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c uint32_t seqno, seqno 199 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (wait_condition(dev_priv, seqno)) seqno 286 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c bool lazy, uint32_t seqno, seqno 292 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP)) seqno 295 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c if (likely(vmw_seqno_passed(dev_priv, seqno))) seqno 301 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c return vmw_fallback_wait(dev_priv, lazy, true, seqno, seqno 305 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c return vmw_fallback_wait(dev_priv, lazy, false, seqno, seqno 313 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c vmw_seqno_passed(dev_priv, seqno), seqno 318 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c vmw_seqno_passed(dev_priv, seqno), seqno 33 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c uint32_t seqno; seqno 57 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c uint32_t seqno) seqno 64 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c marker->seqno = seqno; seqno 91 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c if (signaled_seqno - marker->seqno > (1 << 30)) seqno 132 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c uint32_t seqno; seqno 138 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c seqno = atomic_read(&dev_priv->marker_seq); seqno 142 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c seqno = marker->seqno; seqno 146 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c ret = vmw_wait_seqno(dev_priv, false, seqno, true, seqno 152 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c (void) vmw_marker_pull(queue, seqno); seqno 104 drivers/media/pci/saa7164/saa7164-bus.c dprintk(DBGLVL_BUS, " .seqno = %d\n", m->seqno); seqno 393 drivers/media/pci/saa7164/saa7164-bus.c (msg_tmp.seqno != msg->seqno) || (msg_tmp.size != msg->size)) { seqno 22 drivers/media/pci/saa7164/saa7164-cmd.c ret = dev->cmds[i].seqno; seqno 31 drivers/media/pci/saa7164/saa7164-cmd.c static void saa7164_cmd_free_seqno(struct saa7164_dev *dev, u8 seqno) seqno 34 drivers/media/pci/saa7164/saa7164-cmd.c if ((dev->cmds[seqno].inuse == 1) && seqno 35 drivers/media/pci/saa7164/saa7164-cmd.c (dev->cmds[seqno].seqno == seqno)) { seqno 36 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[seqno].inuse = 0; seqno 37 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[seqno].signalled = 0; seqno 38 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[seqno].timeout = 0; seqno 43 drivers/media/pci/saa7164/saa7164-cmd.c static void saa7164_cmd_timeout_seqno(struct saa7164_dev *dev, u8 seqno) seqno 46 drivers/media/pci/saa7164/saa7164-cmd.c if ((dev->cmds[seqno].inuse == 1) && seqno 47 drivers/media/pci/saa7164/saa7164-cmd.c (dev->cmds[seqno].seqno == seqno)) { seqno 48 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[seqno].timeout = 1; seqno 53 drivers/media/pci/saa7164/saa7164-cmd.c static u32 saa7164_cmd_timeout_get(struct saa7164_dev *dev, u8 seqno) seqno 58 drivers/media/pci/saa7164/saa7164-cmd.c if ((dev->cmds[seqno].inuse == 1) && seqno 59 drivers/media/pci/saa7164/saa7164-cmd.c (dev->cmds[seqno].seqno == seqno)) { seqno 60 drivers/media/pci/saa7164/saa7164-cmd.c ret = dev->cmds[seqno].timeout; seqno 86 drivers/media/pci/saa7164/saa7164-cmd.c q = &dev->cmds[tRsp.seqno].wait; seqno 87 drivers/media/pci/saa7164/saa7164-cmd.c timeout = saa7164_cmd_timeout_get(dev, tRsp.seqno); seqno 92 drivers/media/pci/saa7164/saa7164-cmd.c __func__, tRsp.seqno); seqno 93 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[tRsp.seqno].signalled = 1; seqno 139 drivers/media/pci/saa7164/saa7164-cmd.c q = &dev->cmds[tRsp.seqno].wait; seqno 140 drivers/media/pci/saa7164/saa7164-cmd.c timeout = saa7164_cmd_timeout_get(dev, tRsp.seqno); seqno 158 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_free_seqno(dev, tRsp.seqno); seqno 165 drivers/media/pci/saa7164/saa7164-cmd.c __func__, tRsp.seqno); seqno 166 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[tRsp.seqno].signalled = 1; seqno 245 drivers/media/pci/saa7164/saa7164-cmd.c static int saa7164_cmd_wait(struct saa7164_dev *dev, u8 seqno) seqno 255 drivers/media/pci/saa7164/saa7164-cmd.c dprintk(DBGLVL_CMD, "%s(seqno=%d)\n", __func__, seqno); seqno 258 drivers/media/pci/saa7164/saa7164-cmd.c if ((dev->cmds[seqno].inuse == 1) && seqno 259 drivers/media/pci/saa7164/saa7164-cmd.c (dev->cmds[seqno].seqno == seqno)) { seqno 260 drivers/media/pci/saa7164/saa7164-cmd.c q = &dev->cmds[seqno].wait; seqno 266 drivers/media/pci/saa7164/saa7164-cmd.c if (dev->cmds[seqno].signalled == 0) { seqno 270 drivers/media/pci/saa7164/saa7164-cmd.c __func__, seqno, dev->cmds[seqno].signalled); seqno 279 drivers/media/pci/saa7164/saa7164-cmd.c wait_event_timeout(*q, dev->cmds[seqno].signalled, seqno 285 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_timeout_seqno(dev, seqno); seqno 288 drivers/media/pci/saa7164/saa7164-cmd.c __func__, seqno, r, seqno 289 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[seqno].signalled); seqno 294 drivers/media/pci/saa7164/saa7164-cmd.c __func__, seqno); seqno 299 drivers/media/pci/saa7164/saa7164-cmd.c void saa7164_cmd_signal(struct saa7164_dev *dev, u8 seqno) seqno 309 drivers/media/pci/saa7164/saa7164-cmd.c dev->cmds[i].seqno, seqno 366 drivers/media/pci/saa7164/saa7164-cmd.c command_t.seqno = (u8)ret; seqno 373 drivers/media/pci/saa7164/saa7164-cmd.c __func__, pcommand_t->seqno); seqno 383 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_free_seqno(dev, pcommand_t->seqno); seqno 387 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_timeout_seqno(dev, pcommand_t->seqno); seqno 398 drivers/media/pci/saa7164/saa7164-cmd.c ret = saa7164_cmd_wait(dev, pcommand_t->seqno); seqno 405 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_timeout_seqno(dev, pcommand_t->seqno); seqno 426 drivers/media/pci/saa7164/saa7164-cmd.c __func__, presponse_t->seqno); seqno 435 drivers/media/pci/saa7164/saa7164-cmd.c if (presponse_t->seqno != pcommand_t->seqno) { seqno 439 drivers/media/pci/saa7164/saa7164-cmd.c presponse_t->seqno, pcommand_t->seqno); seqno 465 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_free_seqno(dev, pcommand_t->seqno); seqno 558 drivers/media/pci/saa7164/saa7164-cmd.c saa7164_cmd_free_seqno(dev, pcommand_t->seqno); seqno 1272 drivers/media/pci/saa7164/saa7164-core.c dev->cmds[i].seqno = i; seqno 87 drivers/media/pci/saa7164/saa7164-types.h u8 seqno; seqno 102 drivers/media/pci/saa7164/saa7164-types.h u8 seqno; seqno 513 drivers/media/pci/saa7164/saa7164.h void saa7164_cmd_signal(struct saa7164_dev *dev, u8 seqno); seqno 143 drivers/media/pci/tw5864/tw5864-core.c cur_frame->seqno = input->frame_seqno; seqno 1438 drivers/media/pci/tw5864/tw5864-video.c v4l2_buf->sequence = frame->seqno; seqno 139 drivers/media/pci/tw5864/tw5864.h unsigned int seqno; seqno 436 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u32 seqno; /* sequence number */ seqno 444 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h u32 seqno; /* next sequence number */ seqno 1018 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c be32_to_cpu(e->seqno), seqno 1121 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c __u32 seqno; seqno 1126 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c seqno = be32_to_cpu(e->seqno); seqno 1127 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c if (seqno < fseqno) { seqno 1128 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c fseqno = seqno; seqno 1178 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c entry->seqno, entry->timestamp, seqno 254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c entry->seqno = log->seqno++; seqno 3814 drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h __be32 seqno; seqno 1985 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c entry->seqno, entry->timestamp, seqno 279 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h u32 seqno; /* sequence number */ seqno 287 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h u32 seqno; /* next sequence number */ seqno 103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c entry->seqno = log->seqno++; seqno 34 drivers/net/ethernet/google/gve/gve.h u8 seqno; /* the next expected seqno for this desc*/ seqno 161 drivers/net/ethernet/google/gve/gve_rx.c rx->desc.seqno = 1; seqno 378 drivers/net/ethernet/google/gve/gve_rx.c return (GVE_SEQNO(flags_seq) == rx->desc.seqno); seqno 392 drivers/net/ethernet/google/gve/gve_rx.c while ((GVE_SEQNO(desc->flags_seq) == rx->desc.seqno) && seqno 400 drivers/net/ethernet/google/gve/gve_rx.c rx->desc.seqno); seqno 407 drivers/net/ethernet/google/gve/gve_rx.c rx->desc.seqno = gve_next_seqno(rx->desc.seqno); seqno 98 drivers/net/ethernet/intel/i40e/i40e_dcb.h __be32 seqno; seqno 101 drivers/net/ethernet/intel/ice/ice_dcb.h __be32 seqno; seqno 174 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c __be64 seqno; seqno 185 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c seqno = cpu_to_be64(xo->seq.low + ((u64)seq_hi << 32)); seqno 187 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c skb_store_bits(skb, iv_offset, &seqno, 8); seqno 194 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c __be64 seqno; seqno 197 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c seqno = cpu_to_be64(xo->seq.low + ((u64)xo->seq.hi << 32)); seqno 199 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c skb_store_bits(skb, iv_offset, &seqno, 8); seqno 159 drivers/net/ethernet/sfc/mcdi.c u32 xflags, seqno; seqno 165 drivers/net/ethernet/sfc/mcdi.c ++mcdi->seqno; seqno 168 drivers/net/ethernet/sfc/mcdi.c seqno = mcdi->seqno & SEQ_MASK; seqno 180 drivers/net/ethernet/sfc/mcdi.c MCDI_HEADER_SEQ, seqno, seqno 192 drivers/net/ethernet/sfc/mcdi.c MCDI_HEADER_SEQ, seqno, seqno 324 drivers/net/ethernet/sfc/mcdi.c } else if ((respseq ^ mcdi->seqno) & SEQ_MASK) { seqno 327 drivers/net/ethernet/sfc/mcdi.c respseq, mcdi->seqno); seqno 517 drivers/net/ethernet/sfc/mcdi.c ++mcdi->seqno; seqno 563 drivers/net/ethernet/sfc/mcdi.c static void efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno, seqno 571 drivers/net/ethernet/sfc/mcdi.c if ((seqno ^ mcdi->seqno) & SEQ_MASK) { seqno 578 drivers/net/ethernet/sfc/mcdi.c "seq 0x%x\n", seqno, mcdi->seqno); seqno 682 drivers/net/ethernet/sfc/mcdi.c ++mcdi->seqno; seqno 75 drivers/net/ethernet/sfc/mcdi.h unsigned int seqno; seqno 144 drivers/net/ppp/bsd_comp.c unsigned short seqno; /* sequence # of next packet */ seqno 317 drivers/net/ppp/bsd_comp.c db->seqno = 0; seqno 645 drivers/net/ppp/bsd_comp.c *wptr++ = db->seqno >> 8; seqno 646 drivers/net/ppp/bsd_comp.c *wptr++ = db->seqno; seqno 745 drivers/net/ppp/bsd_comp.c ++db->seqno; seqno 875 drivers/net/ppp/bsd_comp.c if (seq != db->seqno) seqno 880 drivers/net/ppp/bsd_comp.c db->unit, seq, db->seqno - 1); seqno 885 drivers/net/ppp/bsd_comp.c ++db->seqno; seqno 958 drivers/net/ppp/bsd_comp.c max_ent, explen, db->seqno); seqno 1122 drivers/net/ppp/bsd_comp.c db->unit, db->seqno - 1); seqno 25 drivers/net/ppp/ppp_deflate.c int seqno; seqno 150 drivers/net/ppp/ppp_deflate.c state->seqno = 0; seqno 170 drivers/net/ppp/ppp_deflate.c state->seqno = 0; seqno 213 drivers/net/ppp/ppp_deflate.c put_unaligned_be16(state->seqno, wptr); seqno 218 drivers/net/ppp/ppp_deflate.c ++state->seqno; seqno 365 drivers/net/ppp/ppp_deflate.c state->seqno = 0; seqno 386 drivers/net/ppp/ppp_deflate.c state->seqno = 0; seqno 428 drivers/net/ppp/ppp_deflate.c if (seq != (state->seqno & 0xffff)) { seqno 431 drivers/net/ppp/ppp_deflate.c state->unit, seq, state->seqno & 0xffff); seqno 434 drivers/net/ppp/ppp_deflate.c ++state->seqno; seqno 529 drivers/net/ppp/ppp_deflate.c ++state->seqno; seqno 270 drivers/net/wireless/ath/ath6kl/htc.h int seqno; seqno 520 drivers/net/wireless/ath/ath6kl/htc.h u8 seqno; seqno 458 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.seqno); seqno 523 drivers/net/wireless/ath/ath6kl/htc_mbox.c send_len, packet->info.tx.seqno, padded_len, seqno 642 drivers/net/wireless/ath/ath6kl/htc_mbox.c packet->info.tx.seqno = endpoint->seqno; seqno 643 drivers/net/wireless/ath/ath6kl/htc_mbox.c endpoint->seqno++; seqno 717 drivers/net/wireless/ath/ath6kl/htc_mbox.c cred_pad, packet->info.tx.seqno); seqno 728 drivers/net/wireless/ath/ath6kl/htc_mbox.c i, packet, packet->info.tx.seqno, len, rem_scat); seqno 926 drivers/net/wireless/ath/ath6kl/htc_mbox.c 0, packet->info.tx.seqno); seqno 166 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.seqno = ep->seqno; seqno 167 drivers/net/wireless/ath/ath6kl/htc_pipe.c ep->seqno++; seqno 192 drivers/net/wireless/ath/ath6kl/htc_pipe.c packet->info.tx.seqno = ep->seqno; seqno 195 drivers/net/wireless/ath/ath6kl/htc_pipe.c ep->seqno++; seqno 245 drivers/net/wireless/ath/ath6kl/htc_pipe.c htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno; seqno 214 drivers/net/wireless/ath/ath9k/ath9k.h u16 seqno; seqno 148 drivers/net/wireless/ath/ath9k/xmit.c static void ath_send_bar(struct ath_atx_tid *tid, u16 seqno) seqno 154 drivers/net/wireless/ath/ath9k/xmit.c seqno << IEEE80211_SEQ_SEQ_SHIFT); seqno 275 drivers/net/wireless/ath/ath9k/xmit.c u16 seqno = bf->bf_state.seqno; seqno 281 drivers/net/wireless/ath/ath9k/xmit.c index = ATH_BA_INDEX(tid->seq_start, seqno); seqno 298 drivers/net/wireless/ath/ath9k/xmit.c u16 seqno = bf->bf_state.seqno; seqno 304 drivers/net/wireless/ath/ath9k/xmit.c index = ATH_BA_INDEX(tid->seq_start, seqno); seqno 428 drivers/net/wireless/ath/ath9k/xmit.c ba_index = ATH_BA_INDEX(seq_st, bf->bf_state.seqno); seqno 525 drivers/net/wireless/ath/ath9k/xmit.c u16 seqno = bf->bf_state.seqno; seqno 534 drivers/net/wireless/ath/ath9k/xmit.c if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno) || seqno 541 drivers/net/wireless/ath/ath9k/xmit.c } else if (ATH_BA_ISSET(ba, ATH_BA_INDEX(seq_st, seqno))) { seqno 560 drivers/net/wireless/ath/ath9k/xmit.c ATH_BA_INDEX(seq_first, seqno)); seqno 612 drivers/net/wireless/ath/ath9k/xmit.c ATH_BA_INDEX(seq_first, seqno)); seqno 901 drivers/net/wireless/ath/ath9k/xmit.c u16 seqno; seqno 943 drivers/net/wireless/ath/ath9k/xmit.c seqno = bf->bf_state.seqno; seqno 946 drivers/net/wireless/ath/ath9k/xmit.c if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno)) { seqno 961 drivers/net/wireless/ath/ath9k/xmit.c if (tid->bar_index > ATH_BA_INDEX(tid->seq_start, seqno)) { seqno 2154 drivers/net/wireless/ath/ath9k/xmit.c u16 seqno; seqno 2166 drivers/net/wireless/ath/ath9k/xmit.c seqno = tid->seq_next; seqno 2175 drivers/net/wireless/ath/ath9k/xmit.c bf->bf_state.seqno = seqno; seqno 623 drivers/net/wireless/intel/iwlwifi/mvm/d3.c .value = cpu_to_le16(mvmvif->seqno), seqno 1509 drivers/net/wireless/intel/iwlwifi/mvm/d3.c mvmvif->seqno = le16_to_cpu(status->non_qos_seq_ctr) + 0x10; seqno 431 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h u16 seqno; seqno 317 drivers/net/wireless/marvell/mwl8k.c u16 seqno; seqno 1940 drivers/net/wireless/marvell/mwl8k.c wh->seq_ctrl |= cpu_to_le16(mwl8k_vif->seqno); seqno 1941 drivers/net/wireless/marvell/mwl8k.c mwl8k_vif->seqno += 0x10; seqno 4859 drivers/net/wireless/marvell/mwl8k.c mwl8k_vif->seqno = 0; seqno 77 drivers/net/wireless/mediatek/mt76/agg-rx.c mt76_rx_aggr_release_frames(tid, frames, status->seqno); seqno 118 drivers/net/wireless/mediatek/mt76/agg-rx.c u16 seqno; seqno 127 drivers/net/wireless/mediatek/mt76/agg-rx.c seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(bar->start_seq_num)); seqno 133 drivers/net/wireless/mediatek/mt76/agg-rx.c mt76_rx_aggr_release_frames(tid, frames, seqno); seqno 146 drivers/net/wireless/mediatek/mt76/agg-rx.c u16 seqno, head, size; seqno 177 drivers/net/wireless/mediatek/mt76/agg-rx.c seqno = status->seqno; seqno 179 drivers/net/wireless/mediatek/mt76/agg-rx.c sn_less = ieee80211_sn_less(seqno, head); seqno 194 drivers/net/wireless/mediatek/mt76/agg-rx.c if (seqno == head) { seqno 207 drivers/net/wireless/mediatek/mt76/agg-rx.c if (!ieee80211_sn_less(seqno, head + size)) { seqno 208 drivers/net/wireless/mediatek/mt76/agg-rx.c head = ieee80211_sn_inc(ieee80211_sn_sub(seqno, size)); seqno 212 drivers/net/wireless/mediatek/mt76/agg-rx.c idx = seqno % size; seqno 521 drivers/net/wireless/mediatek/mt76/mt76.h u16 seqno; seqno 544 drivers/net/wireless/mediatek/mt76/mt7603/mac.c status->seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl)); seqno 795 drivers/net/wireless/mediatek/mt76/mt7603/mac.c u16 seqno = 0; seqno 884 drivers/net/wireless/mediatek/mt76/mt7603/mac.c seqno = le16_to_cpu(hdr->seq_ctrl); seqno 886 drivers/net/wireless/mediatek/mt76/mt7603/mac.c seqno = le16_to_cpu(bar->start_seq_num); seqno 890 drivers/net/wireless/mediatek/mt76/mt7603/mac.c val |= FIELD_PREP(MT_TXD3_SEQ, seqno >> 4); seqno 213 drivers/net/wireless/mediatek/mt76/mt7615/mac.c status->seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl)); seqno 318 drivers/net/wireless/mediatek/mt76/mt7615/mac.c u16 seqno = 0; seqno 425 drivers/net/wireless/mediatek/mt76/mt7615/mac.c seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl)); seqno 430 drivers/net/wireless/mediatek/mt76/mt7615/mac.c seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(bar->start_seq_num)); seqno 433 drivers/net/wireless/mediatek/mt76/mt7615/mac.c val |= FIELD_PREP(MT_TXD3_SEQ, seqno); seqno 791 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c status->seqno = FIELD_GET(MT_RXWI_SN, tid_sn); seqno 374 drivers/net/wireless/ralink/rt2x00/rt2x00.h atomic_t seqno; seqno 183 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c u16 seqno; seqno 219 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c seqno = atomic_add_return(0x10, &intf->seqno); seqno 221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c seqno = atomic_read(&intf->seqno); seqno 224 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c hdr->seq_ctrl |= cpu_to_le16(seqno); seqno 539 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c priv->seqno += 0x10; seqno 541 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c hdr->seq_ctrl |= cpu_to_le16(priv->seqno); seqno 146 drivers/net/wireless/realtek/rtl818x/rtl8180/rtl8180.h u16 seqno; seqno 272 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c priv->seqno += 0x10; seqno 274 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c tx_hdr->seq_ctrl |= cpu_to_le16(priv->seqno); seqno 155 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h u16 seqno; seqno 37 drivers/nvme/host/fc.c u32 seqno; seqno 815 drivers/s390/net/qeth_core.h struct qeth_seqno seqno; seqno 772 drivers/s390/net/qeth_core_main.c __ipa_cmd(tmp)->hdr.seqno == cmd->hdr.seqno) { seqno 804 drivers/s390/net/qeth_core_main.c memcpy(&card->seqno.pdu_hdr_ack, seqno 1672 drivers/s390/net/qeth_core_main.c memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), &card->seqno.trans_hdr, seqno 1675 drivers/s390/net/qeth_core_main.c card->seqno.trans_hdr++; seqno 1693 drivers/s390/net/qeth_core_main.c &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH); seqno 1694 drivers/s390/net/qeth_core_main.c card->seqno.pdu_hdr++; seqno 1696 drivers/s390/net/qeth_core_main.c &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH); seqno 2722 drivers/s390/net/qeth_core_main.c __ipa_cmd(iob)->hdr.seqno = card->seqno.ipa++; seqno 454 drivers/s390/net/qeth_core_mpc.h __u32 seqno; seqno 762 drivers/s390/net/qeth_core_mpc.h __u16 seqno; seqno 75 drivers/tty/hvc/hvsi.c atomic_t seqno; /* HVSI packet sequence number */ seqno 211 drivers/tty/hvc/hvsi.c header->seqno); seqno 287 drivers/tty/hvc/hvsi.c packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno)); seqno 311 drivers/tty/hvc/hvsi.c hvsi_version_respond(hp, be16_to_cpu(query->hdr.seqno)); seqno 547 drivers/tty/hvc/hvsi.c packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno)); seqno 589 drivers/tty/hvc/hvsi.c packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno)); seqno 672 drivers/tty/hvc/hvsi.c packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno)); seqno 689 drivers/tty/hvc/hvsi.c packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno)); seqno 718 drivers/tty/hvc/hvsi.c atomic_set(&hp->seqno, 0); seqno 12 drivers/tty/hvc/hvsi_lib.c packet->seqno = cpu_to_be16(atomic_inc_return(&pv->seqno)); seqno 24 drivers/tty/hvc/hvsi_lib.c atomic_set(&pv->seqno, 0); seqno 101 drivers/tty/hvc/hvsi_lib.c r.query_seqno = pkt->hdr.seqno; seqno 2734 fs/ocfs2/dlmglue.c int ocfs2_orphan_scan_lock(struct ocfs2_super *osb, u32 *seqno) seqno 2754 fs/ocfs2/dlmglue.c *seqno = be32_to_cpu(lvb->lvb_os_seqno); seqno 2756 fs/ocfs2/dlmglue.c *seqno = osb->osb_orphan_scan.os_seqno + 1; seqno 2761 fs/ocfs2/dlmglue.c void ocfs2_orphan_scan_unlock(struct ocfs2_super *osb, u32 seqno) seqno 2770 fs/ocfs2/dlmglue.c lvb->lvb_os_seqno = cpu_to_be32(seqno); seqno 163 fs/ocfs2/dlmglue.h int ocfs2_orphan_scan_lock(struct ocfs2_super *osb, u32 *seqno); seqno 164 fs/ocfs2/dlmglue.h void ocfs2_orphan_scan_unlock(struct ocfs2_super *osb, u32 seqno); seqno 1925 fs/ocfs2/journal.c u32 seqno = 0; seqno 1935 fs/ocfs2/journal.c status = ocfs2_orphan_scan_lock(osb, &seqno); seqno 1946 fs/ocfs2/journal.c if (os->os_seqno != seqno) { seqno 1947 fs/ocfs2/journal.c os->os_seqno = seqno; seqno 1958 fs/ocfs2/journal.c seqno++; seqno 1962 fs/ocfs2/journal.c ocfs2_orphan_scan_unlock(osb, seqno); seqno 40 fs/xfs/libxfs/xfs_alloc_btree.c xfs_agnumber_t seqno = be32_to_cpu(agf->agf_seqno); seqno 42 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); seqno 117 fs/xfs/libxfs/xfs_alloc_btree.c xfs_agnumber_t seqno = be32_to_cpu(agf->agf_seqno); seqno 162 fs/xfs/libxfs/xfs_alloc_btree.c pag = xfs_perag_get(cur->bc_mp, seqno); seqno 39 fs/xfs/libxfs/xfs_refcount_btree.c xfs_agnumber_t seqno = be32_to_cpu(agf->agf_seqno); seqno 40 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); seqno 65 fs/xfs/libxfs/xfs_rmap_btree.c xfs_agnumber_t seqno = be32_to_cpu(agf->agf_seqno); seqno 67 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); seqno 177 include/drm/drm_dp_mst_helper.h bool seqno; seqno 397 include/drm/drm_dp_mst_helper.h int seqno; seqno 79 include/linux/dma-fence-array.h u64 context, unsigned seqno, seqno 67 include/linux/dma-fence-chain.h int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t seqno); seqno 71 include/linux/dma-fence-chain.h uint64_t seqno); seqno 92 include/linux/dma-fence.h u64 seqno; seqno 267 include/linux/dma-fence.h spinlock_t *lock, u64 context, u64 seqno); seqno 465 include/linux/dma-fence.h return __dma_fence_is_later(f1->seqno, f2->seqno, f1->ops); seqno 581 include/linux/dma-fence.h __ff->context, __ff->seqno, ##args); \ seqno 587 include/linux/dma-fence.h pr_warn("f %llu#%llu: " fmt, __ff->context, __ff->seqno,\ seqno 594 include/linux/dma-fence.h pr_err("f %llu#%llu: " fmt, __ff->context, __ff->seqno, \ seqno 89 include/linux/seqno-fence.h uint32_t seqno_ofs, uint32_t seqno, seqno 102 include/linux/seqno-fence.h dma_fence_init(&fence->base, &seqno_fence_ops, lock, context, seqno); seqno 16 include/net/esp.h __be64 seqno; seqno 22 include/trace/events/dma_fence.h __field(unsigned int, seqno) seqno 29 include/trace/events/dma_fence.h __entry->seqno = fence->seqno; seqno 34 include/trace/events/dma_fence.h __entry->seqno) seqno 193 include/trace/events/rpcgss.h __field(u32, seqno) seqno 202 include/trace/events/rpcgss.h __entry->seqno = rqst->rq_seqno; seqno 207 include/trace/events/rpcgss.h __entry->xid, __entry->seqno) seqno 224 include/trace/events/rpcgss.h __field(u32, seqno) seqno 233 include/trace/events/rpcgss.h __entry->seqno = task->tk_rqstp->rq_seqno; seqno 239 include/trace/events/rpcgss.h __entry->xid, __entry->seqno, __entry->seq_xmit, seqno 709 include/trace/events/sunrpc.h __field(u32, seqno) seqno 718 include/trace/events/sunrpc.h __entry->seqno = rqst->rq_seqno; seqno 725 include/trace/events/sunrpc.h __entry->seqno, __entry->status) seqno 740 include/trace/events/sunrpc.h __field(u32, seqno) seqno 749 include/trace/events/sunrpc.h __entry->seqno = task->tk_rqstp->rq_seqno; seqno 756 include/trace/events/sunrpc.h __entry->seqno, __entry->stage) seqno 181 include/uapi/drm/vc4_drm.h __u64 seqno; seqno 208 include/uapi/drm/vc4_drm.h __u64 seqno; seqno 359 include/uapi/drm/vmwgfx_drm.h __u32 seqno; seqno 213 include/uapi/linux/batadv_packet.h __be32 seqno; seqno 239 include/uapi/linux/batadv_packet.h __be32 seqno; seqno 259 include/uapi/linux/batadv_packet.h __be32 seqno; seqno 312 include/uapi/linux/batadv_packet.h __be16 seqno; seqno 342 include/uapi/linux/batadv_packet.h __be32 seqno; seqno 380 include/uapi/linux/batadv_packet.h __be16 seqno; seqno 463 include/uapi/linux/batadv_packet.h __be16 seqno; seqno 481 include/uapi/linux/batadv_packet.h __be32 seqno; seqno 48 include/uapi/linux/selinux_netlink.h __u32 seqno; seqno 373 net/batman-adv/bat_iv_ogm.c ntohl(batadv_ogm_packet->seqno), seqno 786 net/batman-adv/bat_iv_ogm.c u32 seqno; seqno 821 net/batman-adv/bat_iv_ogm.c seqno = (u32)atomic_read(&hard_iface->bat_iv.ogm_seqno); seqno 822 net/batman-adv/bat_iv_ogm.c batadv_ogm_packet->seqno = htonl(seqno); seqno 1225 net/batman-adv/bat_iv_ogm.c u32 seqno = ntohl(batadv_ogm_packet->seqno); seqno 1241 net/batman-adv/bat_iv_ogm.c seq_diff = seqno - orig_ifinfo->last_real_seqno; seqno 1262 net/batman-adv/bat_iv_ogm.c seqno); seqno 1291 net/batman-adv/bat_iv_ogm.c orig_ifinfo->last_real_seqno, seqno); seqno 1292 net/batman-adv/bat_iv_ogm.c orig_ifinfo->last_real_seqno = seqno; seqno 1435 net/batman-adv/bat_iv_ogm.c sameseq = orig_ifinfo->last_real_seqno == ntohl(ogm_packet->seqno); seqno 1543 net/batman-adv/bat_iv_ogm.c bit_pos -= ntohl(ogm_packet->seqno); seqno 1603 net/batman-adv/bat_iv_ogm.c ogm_packet->prev_sender, ntohl(ogm_packet->seqno), seqno 287 net/batman-adv/bat_v_elp.c elp_packet->seqno = htonl(atomic_read(&hard_iface->bat_v.elp_seqno)); seqno 479 net/batman-adv/bat_v_elp.c seqno_diff = ntohl(elp_packet->seqno) - elp_latest_seqno; seqno 489 net/batman-adv/bat_v_elp.c hardif_neigh->bat_v.elp_latest_seqno = ntohl(elp_packet->seqno); seqno 538 net/batman-adv/bat_v_elp.c ethhdr->h_source, ntohl(elp_packet->seqno), seqno 299 net/batman-adv/bat_v_ogm.c ogm_packet->seqno = htonl(atomic_read(&bat_priv->bat_v.ogm_seqno)); seqno 339 net/batman-adv/bat_v_ogm.c ogm_packet->orig, ntohl(ogm_packet->seqno), seqno 544 net/batman-adv/bat_v_ogm.c if (orig_ifinfo->last_seqno_forwarded == ntohl(ogm_received->seqno)) seqno 547 net/batman-adv/bat_v_ogm.c orig_ifinfo->last_seqno_forwarded = ntohl(ogm_received->seqno); seqno 622 net/batman-adv/bat_v_ogm.c seq_diff = ntohl(ogm2->seqno) - orig_ifinfo->last_real_seqno; seqno 648 net/batman-adv/bat_v_ogm.c orig_ifinfo->last_real_seqno = ntohl(ogm2->seqno); seqno 659 net/batman-adv/bat_v_ogm.c neigh_ifinfo->bat_v.last_seqno = ntohl(ogm2->seqno); seqno 878 net/batman-adv/bat_v_ogm.c ntohl(ogm_packet->seqno), ogm_throughput, ogm_packet->ttl, seqno 109 net/batman-adv/fragmentation.c u16 seqno) seqno 113 net/batman-adv/fragmentation.c if (chain->seqno == seqno) seqno 120 net/batman-adv/fragmentation.c chain->seqno = seqno; seqno 147 net/batman-adv/fragmentation.c u16 seqno, hdr_size = sizeof(struct batadv_frag_packet); seqno 158 net/batman-adv/fragmentation.c seqno = ntohs(frag_packet->seqno); seqno 159 net/batman-adv/fragmentation.c bucket = seqno % BATADV_FRAG_BUFFER_COUNT; seqno 174 net/batman-adv/fragmentation.c if (batadv_frag_init_chain(chain, seqno)) { seqno 480 net/batman-adv/fragmentation.c frag_header.seqno = htons(atomic_inc_return(&bat_priv->frag_seqno)); seqno 772 net/batman-adv/network-coding.c if (last_real_seqno != ntohl(ogm_packet->seqno)) seqno 1193 net/batman-adv/routing.c u32 seqno; seqno 1229 net/batman-adv/routing.c seqno = ntohl(bcast_packet->seqno); seqno 1232 net/batman-adv/routing.c seqno)) seqno 1235 net/batman-adv/routing.c seq_diff = seqno - orig_node->last_bcast_seqno; seqno 1247 net/batman-adv/routing.c orig_node->last_bcast_seqno = seqno; seqno 199 net/batman-adv/soft-interface.c u32 seqno; seqno 348 net/batman-adv/soft-interface.c seqno = atomic_inc_return(&bat_priv->bcast_seqno); seqno 349 net/batman-adv/soft-interface.c bcast_packet->seqno = htonl(seqno); seqno 575 net/batman-adv/tp_meter.c u32 seqno, size_t len, const u8 *session, seqno 602 net/batman-adv/tp_meter.c icmp->seqno = htonl(seqno); seqno 650 net/batman-adv/tp_meter.c if (batadv_seq_before(ntohl(icmp->seqno), seqno 670 net/batman-adv/tp_meter.c recv_ack = ntohl(icmp->seqno); seqno 1195 net/batman-adv/tp_meter.c icmp->seqno = htonl(seq); seqno 1240 net/batman-adv/tp_meter.c new->seqno = ntohl(icmp->seqno); seqno 1260 net/batman-adv/tp_meter.c if (new->seqno == un->seqno) { seqno 1269 net/batman-adv/tp_meter.c if (batadv_seq_before(new->seqno, un->seqno)) seqno 1310 net/batman-adv/tp_meter.c if (batadv_seq_before(tp_vars->last_recv, un->seqno)) seqno 1313 net/batman-adv/tp_meter.c to_ack = un->seqno + un->len - tp_vars->last_recv; seqno 1315 net/batman-adv/tp_meter.c if (batadv_seq_before(tp_vars->last_recv, un->seqno + un->len)) seqno 1390 net/batman-adv/tp_meter.c u32 seqno; seqno 1394 net/batman-adv/tp_meter.c seqno = ntohl(icmp->seqno); seqno 1398 net/batman-adv/tp_meter.c if (seqno == BATADV_TP_FIRST_SEQ) { seqno 1428 net/batman-adv/tp_meter.c if (batadv_seq_before(seqno, tp_vars->last_recv)) seqno 1432 net/batman-adv/tp_meter.c if (ntohl(icmp->seqno) != tp_vars->last_recv) { seqno 301 net/batman-adv/types.h u16 seqno; seqno 1365 net/batman-adv/types.h u32 seqno; seqno 51 net/dccp/ackvec.c int dccp_ackvec_update_records(struct dccp_ackvec *av, u64 seqno, u8 nonce_sum) seqno 59 net/dccp/ackvec.c avr->avr_ack_seqno = seqno; seqno 132 net/dccp/ackvec.c u64 seqno, enum dccp_ackvec_states state) seqno 163 net/dccp/ackvec.c (unsigned long long)seqno, state); seqno 197 net/dccp/ackvec.c u64 seqno, enum dccp_ackvec_states state) seqno 237 net/dccp/ackvec.c av->av_buf_ackno = seqno; seqno 248 net/dccp/ackvec.c u64 seqno = DCCP_SKB_CB(skb)->dccpd_seq; seqno 252 net/dccp/ackvec.c dccp_ackvec_add_new(av, 1, seqno, state); seqno 253 net/dccp/ackvec.c av->av_tail_ackno = seqno; seqno 256 net/dccp/ackvec.c s64 num_packets = dccp_delta_seqno(av->av_buf_ackno, seqno); seqno 264 net/dccp/ackvec.c av->av_buf_ackno = seqno; seqno 267 net/dccp/ackvec.c dccp_ackvec_add_new(av, num_packets, seqno, state); seqno 269 net/dccp/ackvec.c dccp_ackvec_update_old(av, num_packets, seqno, state); seqno 512 net/dccp/ccids/ccid2.c u64 ackno, seqno; seqno 518 net/dccp/ccids/ccid2.c seqno = DCCP_SKB_CB(skb)->dccpd_seq; seqno 527 net/dccp/ccids/ccid2.c hc->tx_rpseq = seqno; seqno 530 net/dccp/ccids/ccid2.c if (dccp_delta_seqno(hc->tx_rpseq, seqno) == 1) seqno 531 net/dccp/ccids/ccid2.c hc->tx_rpseq = seqno; seqno 533 net/dccp/ccids/ccid2.c else if (after48(seqno, hc->tx_rpseq)) { seqno 49 net/dccp/ccids/lib/packet_history.c int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno) seqno 55 net/dccp/ccids/lib/packet_history.c entry->seqno = seqno; seqno 38 net/dccp/ccids/lib/packet_history.h u64 seqno; seqno 43 net/dccp/ccids/lib/packet_history.h tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno) seqno 45 net/dccp/ccids/lib/packet_history.h while (head != NULL && head->seqno != seqno) seqno 50 net/dccp/ccids/lib/packet_history.h int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno); seqno 111 net/dccp/dccp.h static inline void dccp_set_seqno(u64 *seqno, u64 value) seqno 113 net/dccp/dccp.h *seqno = value & UINT48_MAX; seqno 116 net/dccp/dccp.h static inline void dccp_inc_seqno(u64 *seqno) seqno 118 net/dccp/dccp.h *seqno = ADD48(*seqno, 1); seqno 189 net/dccp/input.c u64 lswl, lawl, seqno = DCCP_SKB_CB(skb)->dccpd_seq, seqno 207 net/dccp/input.c dccp_delta_seqno(dp->dccps_swl, seqno) >= 0) seqno 208 net/dccp/input.c dccp_update_gsr(sk, seqno); seqno 234 net/dccp/input.c if (between48(seqno, lswl, dp->dccps_swh) && seqno 237 net/dccp/input.c dccp_update_gsr(sk, seqno); seqno 265 net/dccp/input.c (unsigned long long) lswl, (unsigned long long) seqno, seqno 275 net/dccp/input.c seqno = dp->dccps_gsr; seqno 276 net/dccp/input.c dccp_send_sync(sk, seqno, DCCP_PKT_SYNC); seqno 455 net/ipv4/esp4.c memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), seqno 526 net/ipv4/esp4.c esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + seqno 270 net/ipv4/esp4_offload.c esp.seqno = cpu_to_be64(seq + ((u64)xo->seq.hi << 32)); seqno 396 net/ipv6/esp6.c memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), seqno 465 net/ipv6/esp6.c esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + seqno 298 net/ipv6/esp6_offload.c esp.seqno = cpu_to_be64(xo->seq.low + ((u64)xo->seq.hi << 32)); seqno 104 net/smc/smc.h u16 seqno; /* connection seq # */ seqno 103 net/smc/smc_cdc.c conn->local_tx_ctrl.seqno = conn->tx_cdc_seq; seqno 374 net/smc/smc_cdc.c if (smc_cdc_before(ntohs(cdc->seqno), seqno 375 net/smc/smc_cdc.c conn->local_rx_ctrl.seqno)) seqno 44 net/smc/smc_cdc.h __be16 seqno; seqno 231 net/smc/smc_cdc.h peer->seqno = htons(local->seqno); seqno 264 net/smc/smc_cdc.h local->seqno = ntohs(peer->seqno); seqno 1958 net/sunrpc/auth_gss/auth_gss.c u32 len, offset, seqno, maj_stat; seqno 1971 net/sunrpc/auth_gss/auth_gss.c if (xdr_stream_decode_u32(xdr, &seqno)) seqno 1973 net/sunrpc/auth_gss/auth_gss.c if (seqno != rqstp->rq_seqno) seqno 2018 net/sunrpc/auth_gss/auth_gss.c trace_rpcgss_bad_seqno(task, rqstp->rq_seqno, seqno); seqno 663 net/tipc/group.c u32 event, u16 seqno, seqno 690 net/tipc/group.c msg_set_grp_bc_seqno(hdr, seqno); seqno 947 net/tipc/link.c u16 seqno = l->snd_nxt; seqno 981 net/tipc/link.c msg_set_seqno(hdr, seqno); seqno 1000 net/tipc/link.c seqno++; seqno 1021 net/tipc/link.c l->snd_nxt = seqno; seqno 1030 net/tipc/link.c u16 seqno = l->snd_nxt; seqno 1055 net/tipc/link.c msg_set_seqno(hdr, seqno); seqno 1060 net/tipc/link.c seqno++; seqno 1062 net/tipc/link.c l->snd_nxt = seqno; seqno 1273 net/tipc/link.c u16 seqno; seqno 1308 net/tipc/link.c seqno = buf_seqno(iskb); seqno 1309 net/tipc/link.c if (unlikely(less(seqno, l->drop_point))) { seqno 1313 net/tipc/link.c if (unlikely(seqno != l->drop_point)) { seqno 1314 net/tipc/link.c __tipc_skb_queue_sorted(fdefq, seqno, iskb); seqno 1353 net/tipc/link.c u16 len, expect, seqno = 0; seqno 1361 net/tipc/link.c seqno = buf_seqno(skb); seqno 1362 net/tipc/link.c if (unlikely(more(seqno, expect))) { seqno 1364 net/tipc/link.c ga->gacks[n].gap = htons(seqno - expect); seqno 1369 net/tipc/link.c } else if (unlikely(less(seqno, expect))) { seqno 1373 net/tipc/link.c expect = seqno + 1; seqno 1377 net/tipc/link.c ga->gacks[n].ack = htons(seqno); seqno 1409 net/tipc/link.c u16 seqno, n = 0; seqno 1413 net/tipc/link.c seqno = buf_seqno(skb); seqno 1416 net/tipc/link.c if (less_eq(seqno, acked)) { seqno 1420 net/tipc/link.c } else if (less_eq(seqno, acked + gap)) { seqno 1538 net/tipc/link.c u16 seqno, rcv_nxt, win_lim; seqno 1550 net/tipc/link.c seqno = msg_seqno(hdr); seqno 1561 net/tipc/link.c if (unlikely(less(seqno, rcv_nxt) || more(seqno, win_lim))) { seqno 1574 net/tipc/link.c if (unlikely(seqno != rcv_nxt)) { seqno 1575 net/tipc/link.c __tipc_skb_queue_sorted(defq, seqno, skb); seqno 1722 net/tipc/link.c u16 pktlen, pktcnt, seqno = l->snd_nxt; seqno 1786 net/tipc/link.c msg_set_seqno(hdr, seqno++); seqno 760 net/tipc/msg.c void __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno, seqno 765 net/tipc/msg.c if (skb_queue_empty(list) || less(seqno, buf_seqno(skb_peek(list)))) { seqno 770 net/tipc/msg.c if (more(seqno, buf_seqno(skb_peek_tail(list)))) { seqno 776 net/tipc/msg.c if (more(seqno, buf_seqno(_skb))) seqno 778 net/tipc/msg.c if (seqno == buf_seqno(_skb)) seqno 1073 net/tipc/msg.h void __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno, seqno 1192 net/tipc/msg.h u16 seqno) seqno 1196 net/tipc/msg.h if (skb && less_eq(buf_seqno(skb), seqno)) { seqno 572 security/selinux/avc.c int seqno, int is_insert) seqno 580 security/selinux/avc.c if (seqno < avc->avc_cache.latest_notif) { seqno 582 security/selinux/avc.c seqno, avc->avc_cache.latest_notif); seqno 586 security/selinux/avc.c if (seqno > avc->avc_cache.latest_notif) seqno 587 security/selinux/avc.c avc->avc_cache.latest_notif = seqno; seqno 623 security/selinux/avc.c if (avc_latest_notif_update(avc, avd->seqno, 1)) seqno 830 security/selinux/avc.c u32 tsid, u16 tclass, u32 seqno, seqno 873 security/selinux/avc.c seqno == pos->ae.avd.seqno){ seqno 964 security/selinux/avc.c int avc_ss_reset(struct selinux_avc *avc, u32 seqno) seqno 981 security/selinux/avc.c avc_latest_notif_update(avc, seqno, 0); seqno 1021 security/selinux/avc.c xperm, ssid, tsid, tclass, avd->seqno, NULL, flags); seqno 1084 security/selinux/avc.c driver, xperm, ssid, tsid, tclass, avd.seqno, seqno 13 security/selinux/include/avc_ss.h int avc_ss_reset(struct selinux_avc *avc, u32 seqno); seqno 195 security/selinux/include/security.h u32 seqno; seqno 384 security/selinux/include/security.h int seqno); seqno 391 security/selinux/include/security.h extern void selnl_notify_policyload(u32 seqno); seqno 58 security/selinux/netlink.c msg->seqno = *((u32 *)data); seqno 103 security/selinux/netlink.c void selnl_notify_policyload(u32 seqno) seqno 105 security/selinux/netlink.c selnl_notify(SELNL_MSG_POLICYLOAD, &seqno); seqno 856 security/selinux/selinuxfs.c avd.seqno, avd.flags); seqno 940 security/selinux/ss/services.c avd->seqno = state->ss->latest_granting; seqno 2093 security/selinux/ss/services.c u32 seqno; seqno 2138 security/selinux/ss/services.c seqno = ++state->ss->latest_granting; seqno 2140 security/selinux/ss/services.c avc_ss_reset(state->avc, seqno); seqno 2141 security/selinux/ss/services.c selnl_notify_policyload(seqno); seqno 2142 security/selinux/ss/services.c selinux_status_update_policyload(state, seqno); seqno 2212 security/selinux/ss/services.c seqno = ++state->ss->latest_granting; seqno 2221 security/selinux/ss/services.c avc_ss_reset(state->avc, seqno); seqno 2222 security/selinux/ss/services.c selnl_notify_policyload(seqno); seqno 2223 security/selinux/ss/services.c selinux_status_update_policyload(state, seqno); seqno 2857 security/selinux/ss/services.c int lenp, seqno = 0; seqno 2892 security/selinux/ss/services.c seqno = ++state->ss->latest_granting; seqno 2897 security/selinux/ss/services.c avc_ss_reset(state->avc, seqno); seqno 2898 security/selinux/ss/services.c selnl_notify_policyload(seqno); seqno 2899 security/selinux/ss/services.c selinux_status_update_policyload(state, seqno); seqno 106 security/selinux/ss/status.c int seqno) seqno 117 security/selinux/ss/status.c status->policyload = seqno;