seqno              29 arch/powerpc/include/asm/hvsi.h 	__be16 seqno;
seqno              68 arch/powerpc/include/asm/hvsi.h 	atomic_t	seqno;		/* packet sequence number */
seqno              69 arch/s390/include/uapi/asm/vtoc.h 	__u8 seqno;			/* extent sequence number */
seqno              32 crypto/echainiv.c 	u64 seqno;
seqno              66 crypto/echainiv.c 	seqno = be64_to_cpu(nseqno);
seqno              77 crypto/echainiv.c 		a *= seqno;
seqno              98 drivers/acpi/acpi_extlog.c 	static atomic_t seqno;
seqno             108 drivers/acpi/acpi_extlog.c 	curr_seqno = atomic_inc_return(&seqno);
seqno             541 drivers/acpi/apei/ghes.c 	static atomic_t seqno;
seqno             552 drivers/acpi/apei/ghes.c 	curr_seqno = atomic_inc_return(&seqno);
seqno             353 drivers/bluetooth/hci_bcsp.c 	u8 seqno;
seqno             358 drivers/bluetooth/hci_bcsp.c 	seqno = bcsp->msgq_txseq;
seqno             361 drivers/bluetooth/hci_bcsp.c 		if (bcsp->rxack == seqno)
seqno             364 drivers/bluetooth/hci_bcsp.c 		seqno = (seqno - 1) & 0x07;
seqno             367 drivers/bluetooth/hci_bcsp.c 	if (bcsp->rxack != seqno)
seqno             372 drivers/bluetooth/hci_bcsp.c 	       (seqno - 1) & 0x07);
seqno             386 drivers/crypto/chelsio/chcr_ipsec.c 	__be64 seqno;
seqno             412 drivers/crypto/chelsio/chcr_ipsec.c 	seqno = cpu_to_be64(seqlo + ((u64)xo->seq.hi << 32));
seqno             413 drivers/crypto/chelsio/chcr_ipsec.c 	memcpy(aadiv->seq_no, &seqno, 8);
seqno             152 drivers/dma-buf/dma-fence-array.c 					       u64 context, unsigned seqno,
seqno             166 drivers/dma-buf/dma-fence-array.c 		       context, seqno);
seqno              89 drivers/dma-buf/dma-fence-chain.c int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t seqno)
seqno              93 drivers/dma-buf/dma-fence-chain.c 	if (!seqno)
seqno              97 drivers/dma-buf/dma-fence-chain.c 	if (!chain || chain->base.seqno < seqno)
seqno             102 drivers/dma-buf/dma-fence-chain.c 		    to_dma_fence_chain(*pfence)->prev_seqno < seqno)
seqno             231 drivers/dma-buf/dma-fence-chain.c 			  uint64_t seqno)
seqno             243 drivers/dma-buf/dma-fence-chain.c 	if (prev_chain && __dma_fence_is_later(seqno, prev->seqno, prev->ops)) {
seqno             245 drivers/dma-buf/dma-fence-chain.c 		chain->prev_seqno = prev->seqno;
seqno             250 drivers/dma-buf/dma-fence-chain.c 			seqno = max(prev->seqno, seqno);
seqno             254 drivers/dma-buf/dma-fence-chain.c 		       &chain->lock, context, seqno);
seqno             240 drivers/dma-buf/dma-fence.c 		 fence->context, fence->seqno)) {
seqno             647 drivers/dma-buf/dma-fence.c 	       spinlock_t *lock, u64 context, u64 seqno)
seqno             657 drivers/dma-buf/dma-fence.c 	fence->seqno = seqno;
seqno             152 drivers/dma-buf/sw_sync.c 	return !__dma_fence_is_later(fence->seqno, parent->value, fence->ops);
seqno             163 drivers/dma-buf/sw_sync.c 	snprintf(str, size, "%lld", fence->seqno);
seqno             258 drivers/dma-buf/sw_sync.c 			cmp = value - other->base.seqno;
seqno             142 drivers/dma-buf/sync_file.c 			 fence->seqno);
seqno             252 drivers/dma-buf/sync_file.c 			if (__dma_fence_is_later(pt_a->seqno, pt_b->seqno,
seqno             227 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 		unsigned seqno = ++adev->vm_manager.seqno[ring->idx];
seqno             236 drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c 					       seqno, true);
seqno             392 drivers/gpu/drm/amd/amdgpu/amdgpu_sa.c 				   i->fence->seqno, i->fence->context);
seqno             171 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			     __field(unsigned int, seqno)
seqno             181 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->seqno = job->base.s_fence->finished.seqno;
seqno             187 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 		      __entry->seqno, __get_str(ring), __entry->num_ibs)
seqno             197 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			     __field(unsigned int, seqno)
seqno             206 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->seqno = job->base.s_fence->finished.seqno;
seqno             212 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 		      __entry->seqno, __get_str(ring), __entry->num_ibs)
seqno             475 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			     __field(unsigned, seqno)
seqno             483 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 			   __entry->seqno = fence->seqno;
seqno             488 drivers/gpu/drm/amd/amdgpu/amdgpu_trace.h 		      __entry->seqno)
seqno            3012 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c 		adev->vm_manager.seqno[i] = 0;
seqno             300 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h 	unsigned				seqno[AMDGPU_MAX_RINGS];
seqno             941 drivers/gpu/drm/amd/amdkfd/kfd_device.c 	if (fence->seqno == p->last_eviction_seqno)
seqno             944 drivers/gpu/drm/amd/amdkfd/kfd_device.c 	p->last_eviction_seqno = fence->seqno;
seqno            1012 drivers/gpu/drm/amd/amdkfd/kfd_process.c 	WARN_ONCE(p->last_eviction_seqno != p->ef->seqno,
seqno             220 drivers/gpu/drm/drm_dp_mst_topology.c 	buf[idx++] = (hdr->somt << 7) | (hdr->eomt << 6) | (hdr->seqno << 4);
seqno             259 drivers/gpu/drm/drm_dp_mst_topology.c 	hdr->seqno = (buf[idx] >> 4) & 0x1;
seqno             885 drivers/gpu/drm/drm_dp_mst_topology.c 		DRM_DEBUG_KMS("timedout msg send %p %d %d\n", txmsg, txmsg->state, txmsg->seqno);
seqno             898 drivers/gpu/drm/drm_dp_mst_topology.c 			mstb->tx_slots[txmsg->seqno] = NULL;
seqno            1963 drivers/gpu/drm/drm_dp_mst_topology.c 	if (txmsg->seqno == -1) {
seqno            1969 drivers/gpu/drm/drm_dp_mst_topology.c 			txmsg->seqno = mstb->last_seqno;
seqno            1972 drivers/gpu/drm/drm_dp_mst_topology.c 			txmsg->seqno = 0;
seqno            1974 drivers/gpu/drm/drm_dp_mst_topology.c 			txmsg->seqno = 1;
seqno            1975 drivers/gpu/drm/drm_dp_mst_topology.c 		mstb->tx_slots[txmsg->seqno] = txmsg;
seqno            1989 drivers/gpu/drm/drm_dp_mst_topology.c 	hdr->seqno = txmsg->seqno;
seqno            2007 drivers/gpu/drm/drm_dp_mst_topology.c 		txmsg->seqno = -1;
seqno            2070 drivers/gpu/drm/drm_dp_mst_topology.c 		if (txmsg->seqno != -1)
seqno            2071 drivers/gpu/drm/drm_dp_mst_topology.c 			txmsg->dst->tx_slots[txmsg->seqno] = NULL;
seqno            2089 drivers/gpu/drm/drm_dp_mst_topology.c 	if (txmsg->seqno != -1) {
seqno            2090 drivers/gpu/drm/drm_dp_mst_topology.c 		WARN_ON((unsigned int)txmsg->seqno >
seqno            2092 drivers/gpu/drm/drm_dp_mst_topology.c 		txmsg->dst->tx_slots[txmsg->seqno] = NULL;
seqno            2638 drivers/gpu/drm/drm_dp_mst_topology.c 				    int req_type, int seqno, bool broadcast)
seqno            2647 drivers/gpu/drm/drm_dp_mst_topology.c 	txmsg->seqno = seqno;
seqno            2923 drivers/gpu/drm/drm_dp_mst_topology.c 		slot = mgr->down_rep_recv.initial_hdr.seqno;
seqno            2932 drivers/gpu/drm/drm_dp_mst_topology.c 			       mgr->down_rep_recv.initial_hdr.seqno,
seqno            2977 drivers/gpu/drm/drm_dp_mst_topology.c 		bool seqno;
seqno            2990 drivers/gpu/drm/drm_dp_mst_topology.c 		seqno = mgr->up_req_recv.initial_hdr.seqno;
seqno            2994 drivers/gpu/drm/drm_dp_mst_topology.c 			drm_dp_send_up_ack_reply(mgr, mgr->mst_primary, msg.req_type, seqno, false);
seqno            3011 drivers/gpu/drm/drm_dp_mst_topology.c 			drm_dp_send_up_ack_reply(mgr, mgr->mst_primary, msg.req_type, seqno, false);
seqno             237 drivers/gpu/drm/drm_syncobj.c 	if (prev && prev->seqno >= point)
seqno            1387 drivers/gpu/drm/drm_syncobj.c 				last_signaled->seqno :
seqno             450 drivers/gpu/drm/etnaviv/etnaviv_gem.c 			   fence->seqno);
seqno            1030 drivers/gpu/drm/etnaviv/etnaviv_gpu.c 	return (s32)(f->gpu->completed_fence - f->base.seqno) >= 0;
seqno            1434 drivers/gpu/drm/etnaviv/etnaviv_gpu.c 			if (fence_after(fence->seqno, gpu->completed_fence))
seqno            1435 drivers/gpu/drm/etnaviv/etnaviv_gpu.c 				gpu->completed_fence = fence->seqno;
seqno              83 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno);
seqno              90 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	    i915_seqno_passed(rq->fence.seqno,
seqno              91 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 			      list_next_entry(rq, signal_link)->fence.seqno))
seqno              95 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	    i915_seqno_passed(list_prev_entry(rq, signal_link)->fence.seqno,
seqno              96 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 			      rq->fence.seqno))
seqno             308 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 			if (i915_seqno_passed(rq->fence.seqno, it->fence.seqno))
seqno             365 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 				   rq->fence.context, rq->fence.seqno,
seqno            1153 drivers/gpu/drm/i915/gt/intel_engine_cs.c 		   rq->fence.context, rq->fence.seqno,
seqno             720 drivers/gpu/drm/i915/gt/intel_lrc.c 		  ports[0]->fence.seqno,
seqno             725 drivers/gpu/drm/i915/gt/intel_lrc.c 		  ports[1] ? ports[1]->fence.seqno : 0);
seqno            1103 drivers/gpu/drm/i915/gt/intel_lrc.c 				  last->fence.seqno,
seqno            1130 drivers/gpu/drm/i915/gt/intel_lrc.c 				  last->fence.seqno,
seqno            1202 drivers/gpu/drm/i915/gt/intel_lrc.c 				  rq->fence.seqno,
seqno            1892 drivers/gpu/drm/i915/gt/intel_lrc.c 	*cs++ = rq->fence.seqno - 1;
seqno            2945 drivers/gpu/drm/i915/gt/intel_lrc.c 				  request->fence.seqno,
seqno            2955 drivers/gpu/drm/i915/gt/intel_lrc.c 				      request->fence.seqno,
seqno            2974 drivers/gpu/drm/i915/gt/intel_lrc.c 				      request->fence.seqno,
seqno            3544 drivers/gpu/drm/i915/gt/intel_lrc.c 		  rq->fence.context, rq->fence.seqno,
seqno            3639 drivers/gpu/drm/i915/gt/intel_lrc.c 		  rq->fence.seqno);
seqno             122 drivers/gpu/drm/i915/gt/intel_reset.c 		  rq->fence.seqno,
seqno             719 drivers/gpu/drm/i915/gt/intel_reset.c 		  engine->name, request->fence.context, request->fence.seqno);
seqno             326 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = rq->fence.seqno;
seqno             429 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = rq->fence.seqno;
seqno             447 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = rq->fence.seqno;
seqno             467 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = rq->fence.seqno;
seqno             472 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 		*cs++ = rq->fence.seqno;
seqno             947 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = rq->fence.seqno;
seqno             972 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 		*cs++ = rq->fence.seqno;
seqno             380 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb);
seqno             382 drivers/gpu/drm/i915/gt/intel_timeline.c 	return tl->seqno += 1 + tl->has_initial_breadcrumb;
seqno             387 drivers/gpu/drm/i915/gt/intel_timeline.c 	tl->seqno -= 1 + tl->has_initial_breadcrumb;
seqno             393 drivers/gpu/drm/i915/gt/intel_timeline.c 			   u32 *seqno)
seqno             467 drivers/gpu/drm/i915/gt/intel_timeline.c 	*seqno = timeline_advance(tl);
seqno             468 drivers/gpu/drm/i915/gt/intel_timeline.c 	GEM_BUG_ON(i915_seqno_passed(*tl->hwsp_seqno, *seqno));
seqno             482 drivers/gpu/drm/i915/gt/intel_timeline.c 			     u32 *seqno)
seqno             484 drivers/gpu/drm/i915/gt/intel_timeline.c 	*seqno = timeline_advance(tl);
seqno             487 drivers/gpu/drm/i915/gt/intel_timeline.c 	if (unlikely(!*seqno && tl->hwsp_cacheline))
seqno             488 drivers/gpu/drm/i915/gt/intel_timeline.c 		return __intel_timeline_get_seqno(tl, rq, seqno);
seqno              56 drivers/gpu/drm/i915/gt/intel_timeline.h 					    u64 context, u32 seqno)
seqno              58 drivers/gpu/drm/i915/gt/intel_timeline.h 	return i915_syncmap_set(&tl->sync, context, seqno);
seqno              64 drivers/gpu/drm/i915/gt/intel_timeline.h 	return __intel_timeline_sync_set(tl, fence->context, fence->seqno);
seqno              68 drivers/gpu/drm/i915/gt/intel_timeline.h 						  u64 context, u32 seqno)
seqno              70 drivers/gpu/drm/i915/gt/intel_timeline.h 	return i915_syncmap_is_later(&tl->sync, context, seqno);
seqno              76 drivers/gpu/drm/i915/gt/intel_timeline.h 	return __intel_timeline_sync_is_later(tl, fence->context, fence->seqno);
seqno              83 drivers/gpu/drm/i915/gt/intel_timeline.h 			     u32 *seqno);
seqno              24 drivers/gpu/drm/i915/gt/intel_timeline_types.h 	u32 seqno;
seqno              49 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	u32 *seqno;
seqno              85 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	h->seqno = memset(vaddr, 0xff, PAGE_SIZE);
seqno             194 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = rq->fence.seqno;
seqno             208 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = rq->fence.seqno;
seqno             221 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = rq->fence.seqno;
seqno             233 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		*batch++ = rq->fence.seqno;
seqno             272 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 	return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]);
seqno             294 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 					       rq->fence.seqno),
seqno             297 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 					    rq->fence.seqno),
seqno             612 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 					       __func__, rq->fence.seqno, hws_seqno(&h, rq));
seqno             700 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 			  rq->fence.seqno);
seqno             874 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 					       __func__, rq->fence.seqno, hws_seqno(&h, rq));
seqno            1082 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		       __func__, rq->fence.seqno, hws_seqno(&h, rq));
seqno            1271 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		       __func__, rq->fence.seqno, hws_seqno(&h, rq));
seqno            1469 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 				       prev->fence.seqno, hws_seqno(&h, prev));
seqno            1575 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		       __func__, rq->fence.seqno, hws_seqno(&h, rq));
seqno            1664 drivers/gpu/drm/i915/gt/selftest_hangcheck.c 		       rq->fence.seqno, hws_seqno(&h, rq));
seqno             993 drivers/gpu/drm/i915/gt/selftest_lrc.c 	rq->fence.seqno = 1;
seqno             994 drivers/gpu/drm/i915/gt/selftest_lrc.c 	BUILD_BUG_ON(sizeof(rq->fence.seqno) != 8); /* upper 32b == 0 */
seqno             995 drivers/gpu/drm/i915/gt/selftest_lrc.c 	rq->hwsp_seqno = (u32 *)&rq->fence.seqno + 1;
seqno            1755 drivers/gpu/drm/i915/gt/selftest_lrc.c 				       request[nc]->fence.seqno);
seqno            1760 drivers/gpu/drm/i915/gt/selftest_lrc.c 					  request[nc]->fence.seqno);
seqno            1903 drivers/gpu/drm/i915/gt/selftest_lrc.c 			       request[n]->fence.seqno);
seqno            1908 drivers/gpu/drm/i915/gt/selftest_lrc.c 				  request[n]->fence.seqno);
seqno             161 drivers/gpu/drm/i915/gt/selftest_timeline.c 	u32 seqno;
seqno             173 drivers/gpu/drm/i915/gt/selftest_timeline.c 	if (__intel_timeline_sync_is_later(tl, ctx, p->seqno) != p->expected) {
seqno             175 drivers/gpu/drm/i915/gt/selftest_timeline.c 		       name, p->name, ctx, p->seqno, yesno(p->expected));
seqno             180 drivers/gpu/drm/i915/gt/selftest_timeline.c 		ret = __intel_timeline_sync_set(tl, ctx, p->seqno);
seqno             358 drivers/gpu/drm/i915/gt/selftest_timeline.c 		u32 seqno = prandom_u32_state(&prng);
seqno             360 drivers/gpu/drm/i915/gt/selftest_timeline.c 		if (!__intel_timeline_sync_is_later(&tl, id, seqno))
seqno             361 drivers/gpu/drm/i915/gt/selftest_timeline.c 			__intel_timeline_sync_set(&tl, id, seqno);
seqno             486 drivers/gpu/drm/i915/gt/selftest_timeline.c 	if (*tl->hwsp_seqno != tl->seqno) {
seqno             488 drivers/gpu/drm/i915/gt/selftest_timeline.c 		       *tl->hwsp_seqno, tl->seqno);
seqno             679 drivers/gpu/drm/i915/gt/selftest_timeline.c 		u32 seqno[2];
seqno             690 drivers/gpu/drm/i915/gt/selftest_timeline.c 		tl->seqno = -4u;
seqno             693 drivers/gpu/drm/i915/gt/selftest_timeline.c 		err = intel_timeline_get_seqno(tl, rq, &seqno[0]);
seqno             700 drivers/gpu/drm/i915/gt/selftest_timeline.c 			 seqno[0], tl->hwsp_offset);
seqno             702 drivers/gpu/drm/i915/gt/selftest_timeline.c 		err = emit_ggtt_store_dw(rq, tl->hwsp_offset, seqno[0]);
seqno             710 drivers/gpu/drm/i915/gt/selftest_timeline.c 		err = intel_timeline_get_seqno(tl, rq, &seqno[1]);
seqno             717 drivers/gpu/drm/i915/gt/selftest_timeline.c 			 seqno[1], tl->hwsp_offset);
seqno             719 drivers/gpu/drm/i915/gt/selftest_timeline.c 		err = emit_ggtt_store_dw(rq, tl->hwsp_offset, seqno[1]);
seqno             727 drivers/gpu/drm/i915/gt/selftest_timeline.c 		GEM_BUG_ON(seqno[1] >= seqno[0]);
seqno             738 drivers/gpu/drm/i915/gt/selftest_timeline.c 		if (*hwsp_seqno[0] != seqno[0] || *hwsp_seqno[1] != seqno[1]) {
seqno             741 drivers/gpu/drm/i915/gt/selftest_timeline.c 			       seqno[0], seqno[1]);
seqno             471 drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c 			   ring_tail, rq->fence.seqno);
seqno            2321 drivers/gpu/drm/i915/i915_drv.h int __must_check i915_gem_set_global_seqno(struct drm_device *dev, u32 seqno);
seqno             455 drivers/gpu/drm/i915/i915_gpu_error.c 	if (!erq->seqno)
seqno             459 drivers/gpu/drm/i915/i915_gpu_error.c 		   prefix, erq->pid, erq->context, erq->seqno,
seqno            1176 drivers/gpu/drm/i915/i915_gpu_error.c 	erq->seqno = request->fence.seqno;
seqno             147 drivers/gpu/drm/i915/i915_gpu_error.h 			u32 seqno;
seqno             228 drivers/gpu/drm/i915/i915_request.c 		  rq->fence.context, rq->fence.seqno,
seqno             324 drivers/gpu/drm/i915/i915_request.c 		  rq->fence.context, rq->fence.seqno,
seqno             387 drivers/gpu/drm/i915/i915_request.c 		  request->fence.context, request->fence.seqno,
seqno             479 drivers/gpu/drm/i915/i915_request.c 		  request->fence.context, request->fence.seqno,
seqno             639 drivers/gpu/drm/i915/i915_request.c 	u32 seqno;
seqno             686 drivers/gpu/drm/i915/i915_request.c 	ret = intel_timeline_get_seqno(tl, rq, &seqno);
seqno             702 drivers/gpu/drm/i915/i915_request.c 		       tl->fence_context, seqno);
seqno             879 drivers/gpu/drm/i915/i915_request.c 	*cs++ = from->fence.seqno;
seqno            1176 drivers/gpu/drm/i915/i915_request.c 	GEM_BUG_ON(timeline->seqno != rq->fence.seqno);
seqno            1194 drivers/gpu/drm/i915/i915_request.c 		  engine->name, rq->fence.context, rq->fence.seqno);
seqno             358 drivers/gpu/drm/i915/i915_request.h 	u32 seqno;
seqno             361 drivers/gpu/drm/i915/i915_request.h 	seqno = __hwsp_seqno(rq);
seqno             364 drivers/gpu/drm/i915/i915_request.h 	return seqno;
seqno             369 drivers/gpu/drm/i915/i915_request.h 	return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno - 1);
seqno             428 drivers/gpu/drm/i915/i915_request.h 	return i915_seqno_passed(hwsp_seqno(rq), rq->fence.seqno);
seqno             433 drivers/gpu/drm/i915/i915_request.h 	rq->hwsp_seqno = (u32 *)&rq->fence.seqno; /* decouple from HWSP */
seqno             402 drivers/gpu/drm/i915/i915_sw_fence.c 		  cb->dma->seqno,
seqno             154 drivers/gpu/drm/i915/i915_syncmap.c bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno)
seqno             195 drivers/gpu/drm/i915/i915_syncmap.c 	return seqno_later(__sync_seqno(p)[idx], seqno);
seqno             214 drivers/gpu/drm/i915/i915_syncmap.c static inline void __sync_set_seqno(struct i915_syncmap *p, u64 id, u32 seqno)
seqno             219 drivers/gpu/drm/i915/i915_syncmap.c 	__sync_seqno(p)[idx] = seqno;
seqno             230 drivers/gpu/drm/i915/i915_syncmap.c static noinline int __sync_set(struct i915_syncmap **root, u64 id, u32 seqno)
seqno             335 drivers/gpu/drm/i915/i915_syncmap.c 	__sync_set_seqno(p, id, seqno);
seqno             353 drivers/gpu/drm/i915/i915_syncmap.c int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno)
seqno             362 drivers/gpu/drm/i915/i915_syncmap.c 		__sync_set_seqno(p, id, seqno);
seqno             366 drivers/gpu/drm/i915/i915_syncmap.c 	return __sync_set(root, id, seqno);
seqno              34 drivers/gpu/drm/i915/i915_syncmap.h int i915_syncmap_set(struct i915_syncmap **root, u64 id, u32 seqno);
seqno              35 drivers/gpu/drm/i915/i915_syncmap.h bool i915_syncmap_is_later(struct i915_syncmap **root, u64 id, u32 seqno);
seqno             672 drivers/gpu/drm/i915/i915_trace.h 			     __field(u32, seqno)
seqno             682 drivers/gpu/drm/i915/i915_trace.h 			   __entry->seqno = rq->fence.seqno;
seqno             688 drivers/gpu/drm/i915/i915_trace.h 		      __entry->hw_id, __entry->ctx, __entry->seqno,
seqno             702 drivers/gpu/drm/i915/i915_trace.h 			     __field(u32, seqno)
seqno             711 drivers/gpu/drm/i915/i915_trace.h 			   __entry->seqno = rq->fence.seqno;
seqno             716 drivers/gpu/drm/i915/i915_trace.h 		      __entry->hw_id, __entry->ctx, __entry->seqno)
seqno             745 drivers/gpu/drm/i915/i915_trace.h 			     __field(u32, seqno)
seqno             756 drivers/gpu/drm/i915/i915_trace.h 			   __entry->seqno = rq->fence.seqno;
seqno             763 drivers/gpu/drm/i915/i915_trace.h 		      __entry->hw_id, __entry->ctx, __entry->seqno,
seqno             777 drivers/gpu/drm/i915/i915_trace.h 			     __field(u32, seqno)
seqno             787 drivers/gpu/drm/i915/i915_trace.h 			   __entry->seqno = rq->fence.seqno;
seqno             793 drivers/gpu/drm/i915/i915_trace.h 			      __entry->hw_id, __entry->ctx, __entry->seqno,
seqno             836 drivers/gpu/drm/i915/i915_trace.h 			     __field(u32, seqno)
seqno             852 drivers/gpu/drm/i915/i915_trace.h 			   __entry->seqno = rq->fence.seqno;
seqno             858 drivers/gpu/drm/i915/i915_trace.h 		      __entry->hw_id, __entry->ctx, __entry->seqno,
seqno             385 drivers/gpu/drm/i915/selftests/i915_request.c 			       rq->fence.context, rq->fence.seqno,
seqno             401 drivers/gpu/drm/i915/selftests/i915_request.c 				       rq->fence.context, rq->fence.seqno);
seqno             146 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_seqno(struct i915_syncmap *leaf, unsigned int idx, u32 seqno)
seqno             154 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	if (__sync_seqno(leaf)[idx] != seqno) {
seqno             156 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		       __func__, idx, __sync_seqno(leaf)[idx], seqno);
seqno             163 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno)
seqno             167 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	err = i915_syncmap_set(sync, context, seqno);
seqno             189 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno);
seqno             193 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	if (!i915_syncmap_is_later(sync, context, seqno)) {
seqno             195 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		       context, seqno);
seqno             238 drivers/gpu/drm/i915/selftests/i915_syncmap.c static int check_leaf(struct i915_syncmap **sync, u64 context, u32 seqno)
seqno             242 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	err = i915_syncmap_set(sync, context, seqno);
seqno             258 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	err = check_seqno((*sync), ilog2((*sync)->bitmap), seqno);
seqno             262 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	if (!i915_syncmap_is_later(sync, context, seqno)) {
seqno             264 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		       context, seqno);
seqno             548 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	u32 seqno;
seqno             570 drivers/gpu/drm/i915/selftests/i915_syncmap.c 	seqno = 0;
seqno             575 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		u32 last_seqno = seqno;
seqno             578 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		seqno = prandom_u32_state(&prng);
seqno             579 drivers/gpu/drm/i915/selftests/i915_syncmap.c 		expect = seqno_later(last_seqno, seqno);
seqno             584 drivers/gpu/drm/i915/selftests/i915_syncmap.c 			if (i915_syncmap_is_later(&sync, context, seqno) != expect) {
seqno             586 drivers/gpu/drm/i915/selftests/i915_syncmap.c 				       context, last_seqno, seqno, expect);
seqno             591 drivers/gpu/drm/i915/selftests/i915_syncmap.c 			err = i915_syncmap_set(&sync, context, seqno);
seqno              41 drivers/gpu/drm/i915/selftests/igt_spinner.c 	spin->seqno = memset(vaddr, 0xff, PAGE_SIZE);
seqno             138 drivers/gpu/drm/i915/selftests/igt_spinner.c 	*batch++ = rq->fence.seqno;
seqno             173 drivers/gpu/drm/i915/selftests/igt_spinner.c 	u32 *seqno = spin->seqno + seqno_offset(rq->fence.context);
seqno             175 drivers/gpu/drm/i915/selftests/igt_spinner.c 	return READ_ONCE(*seqno);
seqno             198 drivers/gpu/drm/i915/selftests/igt_spinner.c 					       rq->fence.seqno),
seqno             201 drivers/gpu/drm/i915/selftests/igt_spinner.c 					    rq->fence.seqno),
seqno              24 drivers/gpu/drm/i915/selftests/igt_spinner.h 	void *seqno;
seqno             102 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 	ring->memptrs->fence = submit->seqno;
seqno             182 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 	OUT_RING(ring, submit->seqno);
seqno             192 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 	OUT_RING(ring, submit->seqno);
seqno             964 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 		ring ? ring->id : -1, ring ? ring->seqno : 0,
seqno              85 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	unsigned int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT;
seqno             135 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	OUT_RING(ring, submit->seqno);
seqno             145 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	OUT_RING(ring, submit->seqno);
seqno             677 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 		ring ? ring->id : -1, ring ? ring->seqno : 0,
seqno             353 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		ring->memptrs->fence = ring->seqno;
seqno             443 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	OUT_RING(ring, submit->seqno);
seqno             463 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		OUT_RING(ring, submit->seqno);
seqno             469 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		OUT_RING(ring, submit->seqno);
seqno             537 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		state->ring[i].seqno = gpu->rb[i]->seqno;
seqno             716 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		drm_printf(p, "    last-fence: %d\n", state->ring[i].seqno);
seqno             772 drivers/gpu/drm/msm/adreno/adreno_gpu.c 			ring->seqno);
seqno             114 drivers/gpu/drm/msm/msm_fence.c 	return fence_completed(f->fctx, f->base.seqno);
seqno             796 drivers/gpu/drm/msm/msm_gem.c 				fence->seqno);
seqno             137 drivers/gpu/drm/msm/msm_gem.h 	uint32_t seqno;		/* Sequence number of the submit on the ring */
seqno             582 drivers/gpu/drm/msm/msm_gem_submit.c 	args->fence = submit->fence->seqno;
seqno             395 drivers/gpu/drm/msm/msm_gpu.c 		if (submit->seqno > fence)
seqno             399 drivers/gpu/drm/msm/msm_gpu.c 			submit->fence->seqno);
seqno             411 drivers/gpu/drm/msm/msm_gpu.c 		if (submit->seqno == fence)
seqno             529 drivers/gpu/drm/msm/msm_gpu.c 	} else if (fence < ring->seqno) {
seqno             537 drivers/gpu/drm/msm/msm_gpu.c 				gpu->name, ring->seqno);
seqno             543 drivers/gpu/drm/msm/msm_gpu.c 	if (ring->seqno > ring->hangcheck_fence)
seqno             655 drivers/gpu/drm/msm/msm_gpu.c 	int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT;
seqno             743 drivers/gpu/drm/msm/msm_gpu.c 	submit->seqno = ++ring->seqno;
seqno             151 drivers/gpu/drm/msm/msm_gpu.h 		if (ring->seqno > ring->memptrs->fence)
seqno             194 drivers/gpu/drm/msm/msm_gpu.h 		u32 seqno;
seqno              40 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __field(u32, seqno)
seqno              47 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __entry->seqno = submit->seqno;
seqno              51 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __entry->id, __entry->pid, __entry->ringid, __entry->seqno,
seqno              64 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __field(u32, seqno)
seqno              74 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __entry->seqno = submit->seqno;
seqno              81 drivers/gpu/drm/msm/msm_gpu_trace.h 		    __entry->id, __entry->pid, __entry->ringid, __entry->seqno,
seqno             374 drivers/gpu/drm/msm/msm_rd.c 				pid_nr(submit->pid), submit->seqno);
seqno             377 drivers/gpu/drm/msm/msm_rd.c 				pid_nr(submit->pid), submit->seqno);
seqno              43 drivers/gpu/drm/msm/msm_ringbuffer.h 	uint32_t seqno;
seqno             135 drivers/gpu/drm/nouveau/nouveau_fence.c 		if ((int)(seq - fence->base.seqno) < 0)
seqno             456 drivers/gpu/drm/nouveau/nouveau_fence.c 		ret = (int)(fctx->read(chan) - fence->base.seqno) >= 0;
seqno              46 drivers/gpu/drm/nouveau/nv04_fence.c 		OUT_RING  (chan, fence->base.seqno);
seqno              36 drivers/gpu/drm/nouveau/nv10_fence.c 		OUT_RING  (chan, fence->base.seqno);
seqno              74 drivers/gpu/drm/nouveau/nv84_fence.c 	return fctx->base.emit32(chan, addr, fence->base.seqno);
seqno              84 drivers/gpu/drm/nouveau/nv84_fence.c 	return fctx->base.sync32(chan, addr, fence->base.seqno);
seqno              48 drivers/gpu/drm/panfrost/panfrost_job.c 	u64 seqno;
seqno              95 drivers/gpu/drm/panfrost/panfrost_job.c 	fence->seqno = ++js->queue[js_num].emit_seqno;
seqno              97 drivers/gpu/drm/panfrost/panfrost_job.c 		       js->queue[js_num].fence_context, fence->seqno);
seqno             145 drivers/gpu/drm/qxl/qxl_release.c 	release->base.seqno = ++qdev->release_seqno;
seqno             451 drivers/gpu/drm/qxl/qxl_release.c 		       release->id | 0xf0000000, release->base.seqno);
seqno             127 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
seqno             129 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_ARGS(dev, ring, seqno),
seqno             134 drivers/gpu/drm/radeon/radeon_trace.h 			     __field(u32, seqno)
seqno             140 drivers/gpu/drm/radeon/radeon_trace.h 			   __entry->seqno = seqno;
seqno             144 drivers/gpu/drm/radeon/radeon_trace.h 		      __entry->dev, __entry->ring, __entry->seqno)
seqno             149 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
seqno             151 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_ARGS(dev, ring, seqno)
seqno             156 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
seqno             158 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_ARGS(dev, ring, seqno)
seqno             163 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_PROTO(struct drm_device *dev, int ring, u32 seqno),
seqno             165 drivers/gpu/drm/radeon/radeon_trace.h 	    TP_ARGS(dev, ring, seqno)
seqno              83 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			     __field(unsigned, seqno)
seqno              91 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 			   __entry->seqno = fence->seqno;
seqno              96 drivers/gpu/drm/scheduler/gpu_scheduler_trace.h 		      __entry->seqno)
seqno             161 drivers/gpu/drm/v3d/v3d_drv.h 	u64 seqno;
seqno              16 drivers/gpu/drm/v3d/v3d_fence.c 	fence->seqno = ++v3d->queue[queue].emit_seqno;
seqno              18 drivers/gpu/drm/v3d/v3d_fence.c 		       v3d->queue[queue].fence_context, fence->seqno);
seqno             104 drivers/gpu/drm/v3d/v3d_irq.c 		trace_v3d_bcl_irq(&v3d->drm, fence->seqno);
seqno             113 drivers/gpu/drm/v3d/v3d_irq.c 		trace_v3d_rcl_irq(&v3d->drm, fence->seqno);
seqno             122 drivers/gpu/drm/v3d/v3d_irq.c 		trace_v3d_csd_irq(&v3d->drm, fence->seqno);
seqno             158 drivers/gpu/drm/v3d/v3d_irq.c 		trace_v3d_tfu_irq(&v3d->drm, fence->seqno);
seqno             120 drivers/gpu/drm/v3d/v3d_sched.c 	trace_v3d_submit_cl(dev, false, to_v3d_fence(fence)->seqno,
seqno             169 drivers/gpu/drm/v3d/v3d_sched.c 	trace_v3d_submit_cl(dev, true, to_v3d_fence(fence)->seqno,
seqno             200 drivers/gpu/drm/v3d/v3d_sched.c 	trace_v3d_submit_tfu(dev, to_v3d_fence(fence)->seqno);
seqno             241 drivers/gpu/drm/v3d/v3d_sched.c 	trace_v3d_submit_csd(dev, to_v3d_fence(fence)->seqno);
seqno              39 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno,
seqno              41 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, is_render, seqno, ctnqba, ctnqea),
seqno              46 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno              54 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno              62 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno,
seqno              69 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno              70 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno              74 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno              79 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno              84 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno              89 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno              90 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno              94 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno              99 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno             104 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno             109 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno             110 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno             114 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno             119 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno             124 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno             129 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno             130 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno             134 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno             139 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno             144 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno             168 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno             169 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno             173 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno             178 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno             183 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno             210 drivers/gpu/drm/v3d/v3d_trace.h 		     uint64_t seqno),
seqno             211 drivers/gpu/drm/v3d/v3d_trace.h 	    TP_ARGS(dev, seqno),
seqno             215 drivers/gpu/drm/v3d/v3d_trace.h 			     __field(u64, seqno)
seqno             220 drivers/gpu/drm/v3d/v3d_trace.h 			   __entry->seqno = seqno;
seqno             225 drivers/gpu/drm/v3d/v3d_trace.h 		      __entry->seqno)
seqno             956 drivers/gpu/drm/vc4/vc4_crtc.c 	vc4_queue_seqno_cb(dev, &flip_state->cb, bo->seqno,
seqno             240 drivers/gpu/drm/vc4/vc4_drv.h 	uint64_t seqno;
seqno             293 drivers/gpu/drm/vc4/vc4_drv.h 	uint64_t seqno;
seqno             304 drivers/gpu/drm/vc4/vc4_drv.h 	uint64_t seqno;
seqno             487 drivers/gpu/drm/vc4/vc4_drv.h 	uint64_t seqno;
seqno             805 drivers/gpu/drm/vc4/vc4_drv.h int vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno,
seqno             809 drivers/gpu/drm/vc4/vc4_drv.h 		       struct vc4_seqno_cb *cb, uint64_t seqno,
seqno              41 drivers/gpu/drm/vc4/vc4_fence.c 	return vc4->finished_seqno >= f->seqno;
seqno             381 drivers/gpu/drm/vc4/vc4_gem.c vc4_wait_for_seqno(struct drm_device *dev, uint64_t seqno, uint64_t timeout_ns,
seqno             389 drivers/gpu/drm/vc4/vc4_gem.c 	if (vc4->finished_seqno >= seqno)
seqno             397 drivers/gpu/drm/vc4/vc4_gem.c 	trace_vc4_wait_for_seqno_begin(dev, seqno, timeout_ns);
seqno             408 drivers/gpu/drm/vc4/vc4_gem.c 		if (vc4->finished_seqno >= seqno)
seqno             423 drivers/gpu/drm/vc4/vc4_gem.c 	trace_vc4_wait_for_seqno_end(dev, seqno);
seqno             537 drivers/gpu/drm/vc4/vc4_gem.c vc4_update_bo_seqnos(struct vc4_exec_info *exec, uint64_t seqno)
seqno             544 drivers/gpu/drm/vc4/vc4_gem.c 		bo->seqno = seqno;
seqno             550 drivers/gpu/drm/vc4/vc4_gem.c 		bo->seqno = seqno;
seqno             555 drivers/gpu/drm/vc4/vc4_gem.c 		bo->write_seqno = seqno;
seqno             671 drivers/gpu/drm/vc4/vc4_gem.c 	uint64_t seqno;
seqno             682 drivers/gpu/drm/vc4/vc4_gem.c 	seqno = ++vc4->emit_seqno;
seqno             683 drivers/gpu/drm/vc4/vc4_gem.c 	exec->seqno = seqno;
seqno             686 drivers/gpu/drm/vc4/vc4_gem.c 		       vc4->dma_fence_context, exec->seqno);
seqno             687 drivers/gpu/drm/vc4/vc4_gem.c 	fence->seqno = exec->seqno;
seqno             693 drivers/gpu/drm/vc4/vc4_gem.c 	vc4_update_bo_seqnos(exec, seqno);
seqno            1009 drivers/gpu/drm/vc4/vc4_gem.c 		if (cb->seqno <= vc4->finished_seqno) {
seqno            1026 drivers/gpu/drm/vc4/vc4_gem.c 		       struct vc4_seqno_cb *cb, uint64_t seqno,
seqno            1037 drivers/gpu/drm/vc4/vc4_gem.c 	if (seqno > vc4->finished_seqno) {
seqno            1038 drivers/gpu/drm/vc4/vc4_gem.c 		cb->seqno = seqno;
seqno            1063 drivers/gpu/drm/vc4/vc4_gem.c 				uint64_t seqno,
seqno            1067 drivers/gpu/drm/vc4/vc4_gem.c 	int ret = vc4_wait_for_seqno(dev, seqno, *timeout_ns, true);
seqno            1085 drivers/gpu/drm/vc4/vc4_gem.c 	return vc4_wait_for_seqno_ioctl_helper(dev, args->seqno,
seqno            1108 drivers/gpu/drm/vc4/vc4_gem.c 	ret = vc4_wait_for_seqno_ioctl_helper(dev, bo->seqno,
seqno            1257 drivers/gpu/drm/vc4/vc4_gem.c 	args->seqno = vc4->emit_seqno;
seqno              18 drivers/gpu/drm/vc4/vc4_trace.h 	    TP_PROTO(struct drm_device *dev, uint64_t seqno, uint64_t timeout),
seqno              19 drivers/gpu/drm/vc4/vc4_trace.h 	    TP_ARGS(dev, seqno, timeout),
seqno              23 drivers/gpu/drm/vc4/vc4_trace.h 			     __field(u64, seqno)
seqno              29 drivers/gpu/drm/vc4/vc4_trace.h 			   __entry->seqno = seqno;
seqno              34 drivers/gpu/drm/vc4/vc4_trace.h 		      __entry->dev, __entry->seqno, __entry->timeout)
seqno              38 drivers/gpu/drm/vc4/vc4_trace.h 	    TP_PROTO(struct drm_device *dev, uint64_t seqno),
seqno              39 drivers/gpu/drm/vc4/vc4_trace.h 	    TP_ARGS(dev, seqno),
seqno              43 drivers/gpu/drm/vc4/vc4_trace.h 			     __field(u64, seqno)
seqno              48 drivers/gpu/drm/vc4/vc4_trace.h 			   __entry->seqno = seqno;
seqno              52 drivers/gpu/drm/vc4/vc4_trace.h 		      __entry->dev, __entry->seqno)
seqno             174 drivers/gpu/drm/vc4/vc4_v3d.c 	uint64_t seqno = 0;
seqno             193 drivers/gpu/drm/vc4/vc4_v3d.c 		seqno = exec->seqno;
seqno             196 drivers/gpu/drm/vc4/vc4_v3d.c 	if (seqno) {
seqno             197 drivers/gpu/drm/vc4/vc4_v3d.c 		int ret = vc4_wait_for_seqno(dev, seqno, ~0ull, true);
seqno              58 drivers/gpu/drm/vgem/vgem_fence.c 	snprintf(str, size, "%llu", fence->seqno);
seqno              65 drivers/gpu/drm/vgem/vgem_fence.c 		 dma_fence_is_signaled(fence) ? fence->seqno : 0);
seqno              44 drivers/gpu/drm/virtio/virtgpu_fence.c 	if (atomic64_read(&fence->drv->last_seq) >= fence->f.seqno)
seqno              51 drivers/gpu/drm/virtio/virtgpu_fence.c 	snprintf(str, size, "%llu", f->seqno);
seqno              96 drivers/gpu/drm/virtio/virtgpu_fence.c 	fence->f.seqno = ++drv->sync_seq;
seqno             104 drivers/gpu/drm/virtio/virtgpu_fence.c 	cmd_hdr->fence_id = cpu_to_le64(fence->f.seqno);
seqno             117 drivers/gpu/drm/virtio/virtgpu_fence.c 		if (last_seq < fence->f.seqno)
seqno              47 drivers/gpu/drm/virtio/virtgpu_object.c 		static atomic_t seqno = ATOMIC_INIT(0);
seqno              48 drivers/gpu/drm/virtio/virtgpu_object.c 		int handle = atomic_inc_return(&seqno);
seqno             887 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 			       uint32_t *seqno);
seqno            1021 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 			  uint32_t seqno, bool interruptible,
seqno            1026 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 				uint32_t seqno);
seqno            1030 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 			     uint32_t seqno,
seqno            1052 drivers/gpu/drm/vmwgfx/vmwgfx_drv.h 			   uint32_t seqno);
seqno            3436 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		fence_rep.seqno = fence->base.seqno;
seqno             145 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	u32 seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE);
seqno             146 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	if (seqno - fence->base.seqno < VMW_FENCE_WRAP)
seqno             340 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 			      struct vmw_fence_obj *fence, u32 seqno,
seqno             346 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		       fman->ctx, seqno);
seqno             419 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 			vmw_mmio_write(fence->base.seqno,
seqno             456 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		   goal_seqno - fence->base.seqno < VMW_FENCE_WRAP))
seqno             459 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	vmw_mmio_write(fence->base.seqno, fifo_mem + SVGA_FIFO_FENCE_GOAL);
seqno             470 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	uint32_t seqno, new_seqno;
seqno             473 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE);
seqno             476 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		if (seqno - fence->base.seqno < VMW_FENCE_WRAP) {
seqno             493 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	needs_rerun = vmw_fence_goal_new_locked(fman, seqno);
seqno             496 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		if (new_seqno != seqno) {
seqno             497 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 			seqno = new_seqno;
seqno             551 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 		     uint32_t seqno,
seqno             561 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	ret = vmw_fence_obj_init(fman, fence, seqno,
seqno             602 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 			  uint32_t seqno,
seqno             632 drivers/gpu/drm/vmwgfx/vmwgfx_fence.c 	ret = vmw_fence_obj_init(fman, &ufence->fence, seqno,
seqno             100 drivers/gpu/drm/vmwgfx/vmwgfx_fence.h 			    uint32_t seqno,
seqno             537 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c int vmw_fifo_send_fence(struct vmw_private *dev_priv, uint32_t *seqno)
seqno             547 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 		*seqno = atomic_read(&dev_priv->marker_seq);
seqno             549 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 		(void)vmw_fallback_wait(dev_priv, false, true, *seqno,
seqno             555 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 		*seqno = atomic_add_return(1, &dev_priv->marker_seq);
seqno             556 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 	} while (*seqno == 0);
seqno             571 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 	cmd_fence->fence = *seqno;
seqno             573 drivers/gpu/drm/vmwgfx/vmwgfx_fifo.c 	(void) vmw_marker_push(&fifo_state->marker_queue, *seqno);
seqno             111 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c static bool vmw_fifo_idle(struct vmw_private *dev_priv, uint32_t seqno)
seqno             121 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	uint32_t seqno = vmw_mmio_read(fifo_mem + SVGA_FIFO_FENCE);
seqno             123 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	if (dev_priv->last_read_seqno != seqno) {
seqno             124 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		dev_priv->last_read_seqno = seqno;
seqno             125 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		vmw_marker_pull(&fifo_state->marker_queue, seqno);
seqno             131 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 			 uint32_t seqno)
seqno             136 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP))
seqno             141 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP))
seqno             145 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	    vmw_fifo_idle(dev_priv, seqno))
seqno             153 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	ret = ((atomic_read(&dev_priv->marker_seq) - seqno)
seqno             162 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		      uint32_t seqno,
seqno             199 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		if (wait_condition(dev_priv, seqno))
seqno             286 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		      bool lazy, uint32_t seqno,
seqno             292 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	if (likely(dev_priv->last_read_seqno - seqno < VMW_FENCE_WRAP))
seqno             295 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 	if (likely(vmw_seqno_passed(dev_priv, seqno)))
seqno             301 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		return vmw_fallback_wait(dev_priv, lazy, true, seqno,
seqno             305 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		return vmw_fallback_wait(dev_priv, lazy, false, seqno,
seqno             313 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		     vmw_seqno_passed(dev_priv, seqno),
seqno             318 drivers/gpu/drm/vmwgfx/vmwgfx_irq.c 		     vmw_seqno_passed(dev_priv, seqno),
seqno              33 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	uint32_t seqno;
seqno              57 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 		   uint32_t seqno)
seqno              64 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	marker->seqno = seqno;
seqno              91 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 		if (signaled_seqno - marker->seqno > (1 << 30))
seqno             132 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 	uint32_t seqno;
seqno             138 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 			seqno = atomic_read(&dev_priv->marker_seq);
seqno             142 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 			seqno = marker->seqno;
seqno             146 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 		ret = vmw_wait_seqno(dev_priv, false, seqno, true,
seqno             152 drivers/gpu/drm/vmwgfx/vmwgfx_marker.c 		(void) vmw_marker_pull(queue, seqno);
seqno             104 drivers/media/pci/saa7164/saa7164-bus.c 	dprintk(DBGLVL_BUS, " .seqno            = %d\n",   m->seqno);
seqno             393 drivers/media/pci/saa7164/saa7164-bus.c 		(msg_tmp.seqno != msg->seqno) || (msg_tmp.size != msg->size)) {
seqno              22 drivers/media/pci/saa7164/saa7164-cmd.c 			ret = dev->cmds[i].seqno;
seqno              31 drivers/media/pci/saa7164/saa7164-cmd.c static void saa7164_cmd_free_seqno(struct saa7164_dev *dev, u8 seqno)
seqno              34 drivers/media/pci/saa7164/saa7164-cmd.c 	if ((dev->cmds[seqno].inuse == 1) &&
seqno              35 drivers/media/pci/saa7164/saa7164-cmd.c 		(dev->cmds[seqno].seqno == seqno)) {
seqno              36 drivers/media/pci/saa7164/saa7164-cmd.c 		dev->cmds[seqno].inuse = 0;
seqno              37 drivers/media/pci/saa7164/saa7164-cmd.c 		dev->cmds[seqno].signalled = 0;
seqno              38 drivers/media/pci/saa7164/saa7164-cmd.c 		dev->cmds[seqno].timeout = 0;
seqno              43 drivers/media/pci/saa7164/saa7164-cmd.c static void saa7164_cmd_timeout_seqno(struct saa7164_dev *dev, u8 seqno)
seqno              46 drivers/media/pci/saa7164/saa7164-cmd.c 	if ((dev->cmds[seqno].inuse == 1) &&
seqno              47 drivers/media/pci/saa7164/saa7164-cmd.c 		(dev->cmds[seqno].seqno == seqno)) {
seqno              48 drivers/media/pci/saa7164/saa7164-cmd.c 		dev->cmds[seqno].timeout = 1;
seqno              53 drivers/media/pci/saa7164/saa7164-cmd.c static u32 saa7164_cmd_timeout_get(struct saa7164_dev *dev, u8 seqno)
seqno              58 drivers/media/pci/saa7164/saa7164-cmd.c 	if ((dev->cmds[seqno].inuse == 1) &&
seqno              59 drivers/media/pci/saa7164/saa7164-cmd.c 		(dev->cmds[seqno].seqno == seqno)) {
seqno              60 drivers/media/pci/saa7164/saa7164-cmd.c 		ret = dev->cmds[seqno].timeout;
seqno              86 drivers/media/pci/saa7164/saa7164-cmd.c 		q = &dev->cmds[tRsp.seqno].wait;
seqno              87 drivers/media/pci/saa7164/saa7164-cmd.c 		timeout = saa7164_cmd_timeout_get(dev, tRsp.seqno);
seqno              92 drivers/media/pci/saa7164/saa7164-cmd.c 				__func__, tRsp.seqno);
seqno              93 drivers/media/pci/saa7164/saa7164-cmd.c 			dev->cmds[tRsp.seqno].signalled = 1;
seqno             139 drivers/media/pci/saa7164/saa7164-cmd.c 		q = &dev->cmds[tRsp.seqno].wait;
seqno             140 drivers/media/pci/saa7164/saa7164-cmd.c 		timeout = saa7164_cmd_timeout_get(dev, tRsp.seqno);
seqno             158 drivers/media/pci/saa7164/saa7164-cmd.c 				saa7164_cmd_free_seqno(dev, tRsp.seqno);
seqno             165 drivers/media/pci/saa7164/saa7164-cmd.c 			__func__, tRsp.seqno);
seqno             166 drivers/media/pci/saa7164/saa7164-cmd.c 		dev->cmds[tRsp.seqno].signalled = 1;
seqno             245 drivers/media/pci/saa7164/saa7164-cmd.c static int saa7164_cmd_wait(struct saa7164_dev *dev, u8 seqno)
seqno             255 drivers/media/pci/saa7164/saa7164-cmd.c 	dprintk(DBGLVL_CMD, "%s(seqno=%d)\n", __func__, seqno);
seqno             258 drivers/media/pci/saa7164/saa7164-cmd.c 	if ((dev->cmds[seqno].inuse == 1) &&
seqno             259 drivers/media/pci/saa7164/saa7164-cmd.c 		(dev->cmds[seqno].seqno == seqno)) {
seqno             260 drivers/media/pci/saa7164/saa7164-cmd.c 		q = &dev->cmds[seqno].wait;
seqno             266 drivers/media/pci/saa7164/saa7164-cmd.c 		if (dev->cmds[seqno].signalled == 0) {
seqno             270 drivers/media/pci/saa7164/saa7164-cmd.c 				__func__, seqno, dev->cmds[seqno].signalled);
seqno             279 drivers/media/pci/saa7164/saa7164-cmd.c 			wait_event_timeout(*q, dev->cmds[seqno].signalled,
seqno             285 drivers/media/pci/saa7164/saa7164-cmd.c 				saa7164_cmd_timeout_seqno(dev, seqno);
seqno             288 drivers/media/pci/saa7164/saa7164-cmd.c 				__func__, seqno, r,
seqno             289 drivers/media/pci/saa7164/saa7164-cmd.c 				dev->cmds[seqno].signalled);
seqno             294 drivers/media/pci/saa7164/saa7164-cmd.c 			__func__, seqno);
seqno             299 drivers/media/pci/saa7164/saa7164-cmd.c void saa7164_cmd_signal(struct saa7164_dev *dev, u8 seqno)
seqno             309 drivers/media/pci/saa7164/saa7164-cmd.c 				dev->cmds[i].seqno,
seqno             366 drivers/media/pci/saa7164/saa7164-cmd.c 	command_t.seqno = (u8)ret;
seqno             373 drivers/media/pci/saa7164/saa7164-cmd.c 		__func__, pcommand_t->seqno);
seqno             383 drivers/media/pci/saa7164/saa7164-cmd.c 			saa7164_cmd_free_seqno(dev, pcommand_t->seqno);
seqno             387 drivers/media/pci/saa7164/saa7164-cmd.c 			saa7164_cmd_timeout_seqno(dev, pcommand_t->seqno);
seqno             398 drivers/media/pci/saa7164/saa7164-cmd.c 		ret = saa7164_cmd_wait(dev, pcommand_t->seqno);
seqno             405 drivers/media/pci/saa7164/saa7164-cmd.c 			saa7164_cmd_timeout_seqno(dev, pcommand_t->seqno);
seqno             426 drivers/media/pci/saa7164/saa7164-cmd.c 			__func__, presponse_t->seqno);
seqno             435 drivers/media/pci/saa7164/saa7164-cmd.c 		if (presponse_t->seqno != pcommand_t->seqno) {
seqno             439 drivers/media/pci/saa7164/saa7164-cmd.c 				presponse_t->seqno, pcommand_t->seqno);
seqno             465 drivers/media/pci/saa7164/saa7164-cmd.c 			saa7164_cmd_free_seqno(dev, pcommand_t->seqno);
seqno             558 drivers/media/pci/saa7164/saa7164-cmd.c 	saa7164_cmd_free_seqno(dev, pcommand_t->seqno);
seqno            1272 drivers/media/pci/saa7164/saa7164-core.c 		dev->cmds[i].seqno = i;
seqno              87 drivers/media/pci/saa7164/saa7164-types.h 	u8	seqno;
seqno             102 drivers/media/pci/saa7164/saa7164-types.h 	u8 seqno;
seqno             513 drivers/media/pci/saa7164/saa7164.h void saa7164_cmd_signal(struct saa7164_dev *dev, u8 seqno);
seqno             143 drivers/media/pci/tw5864/tw5864-core.c 		cur_frame->seqno = input->frame_seqno;
seqno            1438 drivers/media/pci/tw5864/tw5864-video.c 	v4l2_buf->sequence = frame->seqno;
seqno             139 drivers/media/pci/tw5864/tw5864.h 	unsigned int seqno;
seqno             436 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	u32 seqno;			/* sequence number */
seqno             444 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 	u32 seqno;			/* next sequence number */
seqno            1018 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			   be32_to_cpu(e->seqno),
seqno            1121 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		__u32 seqno;
seqno            1126 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		seqno = be32_to_cpu(e->seqno);
seqno            1127 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		if (seqno < fseqno) {
seqno            1128 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 			fseqno = seqno;
seqno            1178 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c 		   entry->seqno, entry->timestamp,
seqno             254 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	entry->seqno = log->seqno++;
seqno            3814 drivers/net/ethernet/chelsio/cxgb4/t4fw_api.h 	__be32	seqno;
seqno            1985 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c 		   entry->seqno, entry->timestamp,
seqno             279 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h 	u32 seqno;			/* sequence number */
seqno             287 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h 	u32 seqno;			/* next sequence number */
seqno             103 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c 	entry->seqno = log->seqno++;
seqno              34 drivers/net/ethernet/google/gve/gve.h 	u8 seqno; /* the next expected seqno for this desc*/
seqno             161 drivers/net/ethernet/google/gve/gve_rx.c 	rx->desc.seqno = 1;
seqno             378 drivers/net/ethernet/google/gve/gve_rx.c 	return (GVE_SEQNO(flags_seq) == rx->desc.seqno);
seqno             392 drivers/net/ethernet/google/gve/gve_rx.c 	while ((GVE_SEQNO(desc->flags_seq) == rx->desc.seqno) &&
seqno             400 drivers/net/ethernet/google/gve/gve_rx.c 			   rx->desc.seqno);
seqno             407 drivers/net/ethernet/google/gve/gve_rx.c 		rx->desc.seqno = gve_next_seqno(rx->desc.seqno);
seqno              98 drivers/net/ethernet/intel/i40e/i40e_dcb.h 	__be32 seqno;
seqno             101 drivers/net/ethernet/intel/ice/ice_dcb.h 	__be32 seqno;
seqno             174 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	__be64 seqno;
seqno             185 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	seqno = cpu_to_be64(xo->seq.low + ((u64)seq_hi << 32));
seqno             187 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	skb_store_bits(skb, iv_offset, &seqno, 8);
seqno             194 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	__be64 seqno;
seqno             197 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	seqno = cpu_to_be64(xo->seq.low + ((u64)xo->seq.hi << 32));
seqno             199 drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_rxtx.c 	skb_store_bits(skb, iv_offset, &seqno, 8);
seqno             159 drivers/net/ethernet/sfc/mcdi.c 	u32 xflags, seqno;
seqno             165 drivers/net/ethernet/sfc/mcdi.c 	++mcdi->seqno;
seqno             168 drivers/net/ethernet/sfc/mcdi.c 	seqno = mcdi->seqno & SEQ_MASK;
seqno             180 drivers/net/ethernet/sfc/mcdi.c 				     MCDI_HEADER_SEQ, seqno,
seqno             192 drivers/net/ethernet/sfc/mcdi.c 				     MCDI_HEADER_SEQ, seqno,
seqno             324 drivers/net/ethernet/sfc/mcdi.c 	} else if ((respseq ^ mcdi->seqno) & SEQ_MASK) {
seqno             327 drivers/net/ethernet/sfc/mcdi.c 			  respseq, mcdi->seqno);
seqno             517 drivers/net/ethernet/sfc/mcdi.c 		++mcdi->seqno;
seqno             563 drivers/net/ethernet/sfc/mcdi.c static void efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno,
seqno             571 drivers/net/ethernet/sfc/mcdi.c 	if ((seqno ^ mcdi->seqno) & SEQ_MASK) {
seqno             578 drivers/net/ethernet/sfc/mcdi.c 				  "seq 0x%x\n", seqno, mcdi->seqno);
seqno             682 drivers/net/ethernet/sfc/mcdi.c 		++mcdi->seqno;
seqno              75 drivers/net/ethernet/sfc/mcdi.h 	unsigned int seqno;
seqno             144 drivers/net/ppp/bsd_comp.c     unsigned short seqno;		/* sequence # of next packet */
seqno             317 drivers/net/ppp/bsd_comp.c     db->seqno       = 0;
seqno             645 drivers/net/ppp/bsd_comp.c 	*wptr++ = db->seqno >> 8;
seqno             646 drivers/net/ppp/bsd_comp.c 	*wptr++ = db->seqno;
seqno             745 drivers/net/ppp/bsd_comp.c     ++db->seqno;
seqno             875 drivers/net/ppp/bsd_comp.c     if (seq != db->seqno)
seqno             880 drivers/net/ppp/bsd_comp.c 		   db->unit, seq, db->seqno - 1);
seqno             885 drivers/net/ppp/bsd_comp.c     ++db->seqno;
seqno             958 drivers/net/ppp/bsd_comp.c 		       max_ent, explen, db->seqno);
seqno            1122 drivers/net/ppp/bsd_comp.c 		   db->unit, db->seqno - 1);
seqno              25 drivers/net/ppp/ppp_deflate.c     int		seqno;
seqno             150 drivers/net/ppp/ppp_deflate.c 	state->seqno = 0;
seqno             170 drivers/net/ppp/ppp_deflate.c 	state->seqno = 0;
seqno             213 drivers/net/ppp/ppp_deflate.c 	put_unaligned_be16(state->seqno, wptr);
seqno             218 drivers/net/ppp/ppp_deflate.c 	++state->seqno;
seqno             365 drivers/net/ppp/ppp_deflate.c 	state->seqno = 0;
seqno             386 drivers/net/ppp/ppp_deflate.c 	state->seqno = 0;
seqno             428 drivers/net/ppp/ppp_deflate.c 	if (seq != (state->seqno & 0xffff)) {
seqno             431 drivers/net/ppp/ppp_deflate.c 			       state->unit, seq, state->seqno & 0xffff);
seqno             434 drivers/net/ppp/ppp_deflate.c 	++state->seqno;
seqno             529 drivers/net/ppp/ppp_deflate.c 	++state->seqno;
seqno             270 drivers/net/wireless/ath/ath6kl/htc.h 	int seqno;
seqno             520 drivers/net/wireless/ath/ath6kl/htc.h 	u8 seqno;
seqno             458 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   packet->info.tx.seqno);
seqno             523 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		   send_len, packet->info.tx.seqno, padded_len,
seqno             642 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		packet->info.tx.seqno = endpoint->seqno;
seqno             643 drivers/net/wireless/ath/ath6kl/htc_mbox.c 		endpoint->seqno++;
seqno             717 drivers/net/wireless/ath/ath6kl/htc_mbox.c 				       cred_pad, packet->info.tx.seqno);
seqno             728 drivers/net/wireless/ath/ath6kl/htc_mbox.c 			   i, packet, packet->info.tx.seqno, len, rem_scat);
seqno             926 drivers/net/wireless/ath/ath6kl/htc_mbox.c 					       0, packet->info.tx.seqno);
seqno             166 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.seqno = ep->seqno;
seqno             167 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		ep->seqno++;
seqno             192 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		packet->info.tx.seqno = ep->seqno;
seqno             195 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		ep->seqno++;
seqno             245 drivers/net/wireless/ath/ath6kl/htc_pipe.c 		htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
seqno             214 drivers/net/wireless/ath/ath9k/ath9k.h 	u16 seqno;
seqno             148 drivers/net/wireless/ath/ath9k/xmit.c static void ath_send_bar(struct ath_atx_tid *tid, u16 seqno)
seqno             154 drivers/net/wireless/ath/ath9k/xmit.c 			   seqno << IEEE80211_SEQ_SEQ_SHIFT);
seqno             275 drivers/net/wireless/ath/ath9k/xmit.c 	u16 seqno = bf->bf_state.seqno;
seqno             281 drivers/net/wireless/ath/ath9k/xmit.c 	index  = ATH_BA_INDEX(tid->seq_start, seqno);
seqno             298 drivers/net/wireless/ath/ath9k/xmit.c 	u16 seqno = bf->bf_state.seqno;
seqno             304 drivers/net/wireless/ath/ath9k/xmit.c 	index  = ATH_BA_INDEX(tid->seq_start, seqno);
seqno             428 drivers/net/wireless/ath/ath9k/xmit.c 		ba_index = ATH_BA_INDEX(seq_st, bf->bf_state.seqno);
seqno             525 drivers/net/wireless/ath/ath9k/xmit.c 		u16 seqno = bf->bf_state.seqno;
seqno             534 drivers/net/wireless/ath/ath9k/xmit.c 		if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno) ||
seqno             541 drivers/net/wireless/ath/ath9k/xmit.c 		} else if (ATH_BA_ISSET(ba, ATH_BA_INDEX(seq_st, seqno))) {
seqno             560 drivers/net/wireless/ath/ath9k/xmit.c 				ATH_BA_INDEX(seq_first, seqno));
seqno             612 drivers/net/wireless/ath/ath9k/xmit.c 						ATH_BA_INDEX(seq_first, seqno));
seqno             901 drivers/net/wireless/ath/ath9k/xmit.c 	u16 seqno;
seqno             943 drivers/net/wireless/ath/ath9k/xmit.c 		seqno = bf->bf_state.seqno;
seqno             946 drivers/net/wireless/ath/ath9k/xmit.c 		if (!BAW_WITHIN(tid->seq_start, tid->baw_size, seqno)) {
seqno             961 drivers/net/wireless/ath/ath9k/xmit.c 		if (tid->bar_index > ATH_BA_INDEX(tid->seq_start, seqno)) {
seqno            2154 drivers/net/wireless/ath/ath9k/xmit.c 	u16 seqno;
seqno            2166 drivers/net/wireless/ath/ath9k/xmit.c 		seqno = tid->seq_next;
seqno            2175 drivers/net/wireless/ath/ath9k/xmit.c 		bf->bf_state.seqno = seqno;
seqno             623 drivers/net/wireless/intel/iwlwifi/mvm/d3.c 		.value = cpu_to_le16(mvmvif->seqno),
seqno            1509 drivers/net/wireless/intel/iwlwifi/mvm/d3.c 	mvmvif->seqno = le16_to_cpu(status->non_qos_seq_ctr) + 0x10;
seqno             431 drivers/net/wireless/intel/iwlwifi/mvm/mvm.h 	u16 seqno;
seqno             317 drivers/net/wireless/marvell/mwl8k.c 	u16 seqno;
seqno            1940 drivers/net/wireless/marvell/mwl8k.c 		wh->seq_ctrl |= cpu_to_le16(mwl8k_vif->seqno);
seqno            1941 drivers/net/wireless/marvell/mwl8k.c 		mwl8k_vif->seqno += 0x10;
seqno            4859 drivers/net/wireless/marvell/mwl8k.c 	mwl8k_vif->seqno = 0;
seqno              77 drivers/net/wireless/mediatek/mt76/agg-rx.c 		mt76_rx_aggr_release_frames(tid, frames, status->seqno);
seqno             118 drivers/net/wireless/mediatek/mt76/agg-rx.c 	u16 seqno;
seqno             127 drivers/net/wireless/mediatek/mt76/agg-rx.c 	seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(bar->start_seq_num));
seqno             133 drivers/net/wireless/mediatek/mt76/agg-rx.c 	mt76_rx_aggr_release_frames(tid, frames, seqno);
seqno             146 drivers/net/wireless/mediatek/mt76/agg-rx.c 	u16 seqno, head, size;
seqno             177 drivers/net/wireless/mediatek/mt76/agg-rx.c 	seqno = status->seqno;
seqno             179 drivers/net/wireless/mediatek/mt76/agg-rx.c 	sn_less = ieee80211_sn_less(seqno, head);
seqno             194 drivers/net/wireless/mediatek/mt76/agg-rx.c 	if (seqno == head) {
seqno             207 drivers/net/wireless/mediatek/mt76/agg-rx.c 	if (!ieee80211_sn_less(seqno, head + size)) {
seqno             208 drivers/net/wireless/mediatek/mt76/agg-rx.c 		head = ieee80211_sn_inc(ieee80211_sn_sub(seqno, size));
seqno             212 drivers/net/wireless/mediatek/mt76/agg-rx.c 	idx = seqno % size;
seqno             521 drivers/net/wireless/mediatek/mt76/mt76.h 	u16 seqno;
seqno             544 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	status->seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl));
seqno             795 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	u16 seqno = 0;
seqno             884 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		seqno = le16_to_cpu(hdr->seq_ctrl);
seqno             886 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		seqno = le16_to_cpu(bar->start_seq_num);
seqno             890 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	val |= FIELD_PREP(MT_TXD3_SEQ, seqno >> 4);
seqno             213 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	status->seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl));
seqno             318 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	u16 seqno = 0;
seqno             425 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(hdr->seq_ctrl));
seqno             430 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		seqno = IEEE80211_SEQ_TO_SN(le16_to_cpu(bar->start_seq_num));
seqno             433 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	val |= FIELD_PREP(MT_TXD3_SEQ, seqno);
seqno             791 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 	status->seqno = FIELD_GET(MT_RXWI_SN, tid_sn);
seqno             374 drivers/net/wireless/ralink/rt2x00/rt2x00.h 	atomic_t seqno;
seqno             183 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	u16 seqno;
seqno             219 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		seqno = atomic_add_return(0x10, &intf->seqno);
seqno             221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 		seqno = atomic_read(&intf->seqno);
seqno             224 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c 	hdr->seq_ctrl |= cpu_to_le16(seqno);
seqno             539 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 			priv->seqno += 0x10;
seqno             541 drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c 		hdr->seq_ctrl |= cpu_to_le16(priv->seqno);
seqno             146 drivers/net/wireless/realtek/rtl818x/rtl8180/rtl8180.h 	u16 seqno;
seqno             272 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 			priv->seqno += 0x10;
seqno             274 drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c 		tx_hdr->seq_ctrl |= cpu_to_le16(priv->seqno);
seqno             155 drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h 	u16 seqno;
seqno              37 drivers/nvme/host/fc.c 	u32			seqno;
seqno             815 drivers/s390/net/qeth_core.h 	struct qeth_seqno seqno;
seqno             772 drivers/s390/net/qeth_core_main.c 		    __ipa_cmd(tmp)->hdr.seqno == cmd->hdr.seqno) {
seqno             804 drivers/s390/net/qeth_core_main.c 	memcpy(&card->seqno.pdu_hdr_ack,
seqno            1672 drivers/s390/net/qeth_core_main.c 	memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), &card->seqno.trans_hdr,
seqno            1675 drivers/s390/net/qeth_core_main.c 		card->seqno.trans_hdr++;
seqno            1693 drivers/s390/net/qeth_core_main.c 	       &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH);
seqno            1694 drivers/s390/net/qeth_core_main.c 	card->seqno.pdu_hdr++;
seqno            1696 drivers/s390/net/qeth_core_main.c 	       &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH);
seqno            2722 drivers/s390/net/qeth_core_main.c 	__ipa_cmd(iob)->hdr.seqno = card->seqno.ipa++;
seqno             454 drivers/s390/net/qeth_core_mpc.h 	__u32 seqno;
seqno             762 drivers/s390/net/qeth_core_mpc.h 	__u16  seqno;
seqno              75 drivers/tty/hvc/hvsi.c 	atomic_t seqno; /* HVSI packet sequence number */
seqno             211 drivers/tty/hvc/hvsi.c 			header->seqno);
seqno             287 drivers/tty/hvc/hvsi.c 	packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno));
seqno             311 drivers/tty/hvc/hvsi.c 			hvsi_version_respond(hp, be16_to_cpu(query->hdr.seqno));
seqno             547 drivers/tty/hvc/hvsi.c 	packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno));
seqno             589 drivers/tty/hvc/hvsi.c 	packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno));
seqno             672 drivers/tty/hvc/hvsi.c 	packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno));
seqno             689 drivers/tty/hvc/hvsi.c 	packet.hdr.seqno = cpu_to_be16(atomic_inc_return(&hp->seqno));
seqno             718 drivers/tty/hvc/hvsi.c 	atomic_set(&hp->seqno, 0);
seqno              12 drivers/tty/hvc/hvsi_lib.c 	packet->seqno = cpu_to_be16(atomic_inc_return(&pv->seqno));
seqno              24 drivers/tty/hvc/hvsi_lib.c 	atomic_set(&pv->seqno, 0);
seqno             101 drivers/tty/hvc/hvsi_lib.c 	r.query_seqno = pkt->hdr.seqno;
seqno            2734 fs/ocfs2/dlmglue.c int ocfs2_orphan_scan_lock(struct ocfs2_super *osb, u32 *seqno)
seqno            2754 fs/ocfs2/dlmglue.c 		*seqno = be32_to_cpu(lvb->lvb_os_seqno);
seqno            2756 fs/ocfs2/dlmglue.c 		*seqno = osb->osb_orphan_scan.os_seqno + 1;
seqno            2761 fs/ocfs2/dlmglue.c void ocfs2_orphan_scan_unlock(struct ocfs2_super *osb, u32 seqno)
seqno            2770 fs/ocfs2/dlmglue.c 		lvb->lvb_os_seqno = cpu_to_be32(seqno);
seqno             163 fs/ocfs2/dlmglue.h int ocfs2_orphan_scan_lock(struct ocfs2_super *osb, u32 *seqno);
seqno             164 fs/ocfs2/dlmglue.h void ocfs2_orphan_scan_unlock(struct ocfs2_super *osb, u32 seqno);
seqno            1925 fs/ocfs2/journal.c 	u32 seqno = 0;
seqno            1935 fs/ocfs2/journal.c 	status = ocfs2_orphan_scan_lock(osb, &seqno);
seqno            1946 fs/ocfs2/journal.c 	if (os->os_seqno != seqno) {
seqno            1947 fs/ocfs2/journal.c 		os->os_seqno = seqno;
seqno            1958 fs/ocfs2/journal.c 	seqno++;
seqno            1962 fs/ocfs2/journal.c 	ocfs2_orphan_scan_unlock(osb, seqno);
seqno              40 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_agnumber_t		seqno = be32_to_cpu(agf->agf_seqno);
seqno              42 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
seqno             117 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_agnumber_t		seqno = be32_to_cpu(agf->agf_seqno);
seqno             162 fs/xfs/libxfs/xfs_alloc_btree.c 	pag = xfs_perag_get(cur->bc_mp, seqno);
seqno              39 fs/xfs/libxfs/xfs_refcount_btree.c 	xfs_agnumber_t		seqno = be32_to_cpu(agf->agf_seqno);
seqno              40 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
seqno              65 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_agnumber_t		seqno = be32_to_cpu(agf->agf_seqno);
seqno              67 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
seqno             177 include/drm/drm_dp_mst_helper.h 	bool seqno;
seqno             397 include/drm/drm_dp_mst_helper.h 	int seqno;
seqno              79 include/linux/dma-fence-array.h 					       u64 context, unsigned seqno,
seqno              67 include/linux/dma-fence-chain.h int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t seqno);
seqno              71 include/linux/dma-fence-chain.h 			  uint64_t seqno);
seqno              92 include/linux/dma-fence.h 	u64 seqno;
seqno             267 include/linux/dma-fence.h 		    spinlock_t *lock, u64 context, u64 seqno);
seqno             465 include/linux/dma-fence.h 	return __dma_fence_is_later(f1->seqno, f2->seqno, f1->ops);
seqno             581 include/linux/dma-fence.h 				__ff->context, __ff->seqno, ##args);	\
seqno             587 include/linux/dma-fence.h 		pr_warn("f %llu#%llu: " fmt, __ff->context, __ff->seqno,\
seqno             594 include/linux/dma-fence.h 		pr_err("f %llu#%llu: " fmt, __ff->context, __ff->seqno,	\
seqno              89 include/linux/seqno-fence.h 		 uint32_t seqno_ofs, uint32_t seqno,
seqno             102 include/linux/seqno-fence.h 	dma_fence_init(&fence->base, &seqno_fence_ops, lock, context, seqno);
seqno              16 include/net/esp.h 	__be64	seqno;
seqno              22 include/trace/events/dma_fence.h 		__field(unsigned int, seqno)
seqno              29 include/trace/events/dma_fence.h 		__entry->seqno = fence->seqno;
seqno              34 include/trace/events/dma_fence.h 		  __entry->seqno)
seqno             193 include/trace/events/rpcgss.h 		__field(u32, seqno)
seqno             202 include/trace/events/rpcgss.h 		__entry->seqno = rqst->rq_seqno;
seqno             207 include/trace/events/rpcgss.h 		__entry->xid, __entry->seqno)
seqno             224 include/trace/events/rpcgss.h 		__field(u32, seqno)
seqno             233 include/trace/events/rpcgss.h 		__entry->seqno = task->tk_rqstp->rq_seqno;
seqno             239 include/trace/events/rpcgss.h 		__entry->xid, __entry->seqno, __entry->seq_xmit,
seqno             709 include/trace/events/sunrpc.h 		__field(u32, seqno)
seqno             718 include/trace/events/sunrpc.h 		__entry->seqno = rqst->rq_seqno;
seqno             725 include/trace/events/sunrpc.h 		__entry->seqno, __entry->status)
seqno             740 include/trace/events/sunrpc.h 		__field(u32, seqno)
seqno             749 include/trace/events/sunrpc.h 		__entry->seqno = task->tk_rqstp->rq_seqno;
seqno             756 include/trace/events/sunrpc.h 		__entry->seqno, __entry->stage)
seqno             181 include/uapi/drm/vc4_drm.h 	__u64 seqno;
seqno             208 include/uapi/drm/vc4_drm.h 	__u64 seqno;
seqno             359 include/uapi/drm/vmwgfx_drm.h 	__u32 seqno;
seqno             213 include/uapi/linux/batadv_packet.h 	__be32 seqno;
seqno             239 include/uapi/linux/batadv_packet.h 	__be32 seqno;
seqno             259 include/uapi/linux/batadv_packet.h 	__be32 seqno;
seqno             312 include/uapi/linux/batadv_packet.h 	__be16 seqno;
seqno             342 include/uapi/linux/batadv_packet.h 	__be32 seqno;
seqno             380 include/uapi/linux/batadv_packet.h 	__be16 seqno;
seqno             463 include/uapi/linux/batadv_packet.h 	__be16 seqno;
seqno             481 include/uapi/linux/batadv_packet.h 	__be32 seqno;
seqno              48 include/uapi/linux/selinux_netlink.h 	__u32	seqno;
seqno             373 net/batman-adv/bat_iv_ogm.c 			   ntohl(batadv_ogm_packet->seqno),
seqno             786 net/batman-adv/bat_iv_ogm.c 	u32 seqno;
seqno             821 net/batman-adv/bat_iv_ogm.c 	seqno = (u32)atomic_read(&hard_iface->bat_iv.ogm_seqno);
seqno             822 net/batman-adv/bat_iv_ogm.c 	batadv_ogm_packet->seqno = htonl(seqno);
seqno            1225 net/batman-adv/bat_iv_ogm.c 	u32 seqno = ntohl(batadv_ogm_packet->seqno);
seqno            1241 net/batman-adv/bat_iv_ogm.c 	seq_diff = seqno - orig_ifinfo->last_real_seqno;
seqno            1262 net/batman-adv/bat_iv_ogm.c 					 seqno);
seqno            1291 net/batman-adv/bat_iv_ogm.c 			   orig_ifinfo->last_real_seqno, seqno);
seqno            1292 net/batman-adv/bat_iv_ogm.c 		orig_ifinfo->last_real_seqno = seqno;
seqno            1435 net/batman-adv/bat_iv_ogm.c 	sameseq = orig_ifinfo->last_real_seqno == ntohl(ogm_packet->seqno);
seqno            1543 net/batman-adv/bat_iv_ogm.c 	bit_pos -= ntohl(ogm_packet->seqno);
seqno            1603 net/batman-adv/bat_iv_ogm.c 		   ogm_packet->prev_sender, ntohl(ogm_packet->seqno),
seqno             287 net/batman-adv/bat_v_elp.c 	elp_packet->seqno = htonl(atomic_read(&hard_iface->bat_v.elp_seqno));
seqno             479 net/batman-adv/bat_v_elp.c 	seqno_diff = ntohl(elp_packet->seqno) - elp_latest_seqno;
seqno             489 net/batman-adv/bat_v_elp.c 	hardif_neigh->bat_v.elp_latest_seqno = ntohl(elp_packet->seqno);
seqno             538 net/batman-adv/bat_v_elp.c 		   ethhdr->h_source, ntohl(elp_packet->seqno),
seqno             299 net/batman-adv/bat_v_ogm.c 	ogm_packet->seqno = htonl(atomic_read(&bat_priv->bat_v.ogm_seqno));
seqno             339 net/batman-adv/bat_v_ogm.c 			   ogm_packet->orig, ntohl(ogm_packet->seqno),
seqno             544 net/batman-adv/bat_v_ogm.c 	if (orig_ifinfo->last_seqno_forwarded == ntohl(ogm_received->seqno))
seqno             547 net/batman-adv/bat_v_ogm.c 	orig_ifinfo->last_seqno_forwarded = ntohl(ogm_received->seqno);
seqno             622 net/batman-adv/bat_v_ogm.c 	seq_diff = ntohl(ogm2->seqno) - orig_ifinfo->last_real_seqno;
seqno             648 net/batman-adv/bat_v_ogm.c 	orig_ifinfo->last_real_seqno = ntohl(ogm2->seqno);
seqno             659 net/batman-adv/bat_v_ogm.c 	neigh_ifinfo->bat_v.last_seqno = ntohl(ogm2->seqno);
seqno             878 net/batman-adv/bat_v_ogm.c 		   ntohl(ogm_packet->seqno), ogm_throughput, ogm_packet->ttl,
seqno             109 net/batman-adv/fragmentation.c 				   u16 seqno)
seqno             113 net/batman-adv/fragmentation.c 	if (chain->seqno == seqno)
seqno             120 net/batman-adv/fragmentation.c 	chain->seqno = seqno;
seqno             147 net/batman-adv/fragmentation.c 	u16 seqno, hdr_size = sizeof(struct batadv_frag_packet);
seqno             158 net/batman-adv/fragmentation.c 	seqno = ntohs(frag_packet->seqno);
seqno             159 net/batman-adv/fragmentation.c 	bucket = seqno % BATADV_FRAG_BUFFER_COUNT;
seqno             174 net/batman-adv/fragmentation.c 	if (batadv_frag_init_chain(chain, seqno)) {
seqno             480 net/batman-adv/fragmentation.c 	frag_header.seqno = htons(atomic_inc_return(&bat_priv->frag_seqno));
seqno             772 net/batman-adv/network-coding.c 	if (last_real_seqno != ntohl(ogm_packet->seqno))
seqno            1193 net/batman-adv/routing.c 	u32 seqno;
seqno            1229 net/batman-adv/routing.c 	seqno = ntohl(bcast_packet->seqno);
seqno            1232 net/batman-adv/routing.c 			    seqno))
seqno            1235 net/batman-adv/routing.c 	seq_diff = seqno - orig_node->last_bcast_seqno;
seqno            1247 net/batman-adv/routing.c 		orig_node->last_bcast_seqno = seqno;
seqno             199 net/batman-adv/soft-interface.c 	u32 seqno;
seqno             348 net/batman-adv/soft-interface.c 		seqno = atomic_inc_return(&bat_priv->bcast_seqno);
seqno             349 net/batman-adv/soft-interface.c 		bcast_packet->seqno = htonl(seqno);
seqno             575 net/batman-adv/tp_meter.c 			      u32 seqno, size_t len, const u8 *session,
seqno             602 net/batman-adv/tp_meter.c 	icmp->seqno = htonl(seqno);
seqno             650 net/batman-adv/tp_meter.c 	if (batadv_seq_before(ntohl(icmp->seqno),
seqno             670 net/batman-adv/tp_meter.c 	recv_ack = ntohl(icmp->seqno);
seqno            1195 net/batman-adv/tp_meter.c 	icmp->seqno = htonl(seq);
seqno            1240 net/batman-adv/tp_meter.c 	new->seqno = ntohl(icmp->seqno);
seqno            1260 net/batman-adv/tp_meter.c 		if (new->seqno == un->seqno) {
seqno            1269 net/batman-adv/tp_meter.c 		if (batadv_seq_before(new->seqno, un->seqno))
seqno            1310 net/batman-adv/tp_meter.c 		if (batadv_seq_before(tp_vars->last_recv, un->seqno))
seqno            1313 net/batman-adv/tp_meter.c 		to_ack = un->seqno + un->len - tp_vars->last_recv;
seqno            1315 net/batman-adv/tp_meter.c 		if (batadv_seq_before(tp_vars->last_recv, un->seqno + un->len))
seqno            1390 net/batman-adv/tp_meter.c 	u32 seqno;
seqno            1394 net/batman-adv/tp_meter.c 	seqno = ntohl(icmp->seqno);
seqno            1398 net/batman-adv/tp_meter.c 	if (seqno == BATADV_TP_FIRST_SEQ) {
seqno            1428 net/batman-adv/tp_meter.c 	if (batadv_seq_before(seqno, tp_vars->last_recv))
seqno            1432 net/batman-adv/tp_meter.c 	if (ntohl(icmp->seqno) != tp_vars->last_recv) {
seqno             301 net/batman-adv/types.h 	u16 seqno;
seqno            1365 net/batman-adv/types.h 	u32 seqno;
seqno              51 net/dccp/ackvec.c int dccp_ackvec_update_records(struct dccp_ackvec *av, u64 seqno, u8 nonce_sum)
seqno              59 net/dccp/ackvec.c 	avr->avr_ack_seqno  = seqno;
seqno             132 net/dccp/ackvec.c 				   u64 seqno, enum dccp_ackvec_states state)
seqno             163 net/dccp/ackvec.c 					      (unsigned long long)seqno, state);
seqno             197 net/dccp/ackvec.c 				u64 seqno, enum dccp_ackvec_states state)
seqno             237 net/dccp/ackvec.c 	av->av_buf_ackno	    = seqno;
seqno             248 net/dccp/ackvec.c 	u64 seqno = DCCP_SKB_CB(skb)->dccpd_seq;
seqno             252 net/dccp/ackvec.c 		dccp_ackvec_add_new(av, 1, seqno, state);
seqno             253 net/dccp/ackvec.c 		av->av_tail_ackno = seqno;
seqno             256 net/dccp/ackvec.c 		s64 num_packets = dccp_delta_seqno(av->av_buf_ackno, seqno);
seqno             264 net/dccp/ackvec.c 			av->av_buf_ackno = seqno;
seqno             267 net/dccp/ackvec.c 			dccp_ackvec_add_new(av, num_packets, seqno, state);
seqno             269 net/dccp/ackvec.c 			dccp_ackvec_update_old(av, num_packets, seqno, state);
seqno             512 net/dccp/ccids/ccid2.c 	u64 ackno, seqno;
seqno             518 net/dccp/ccids/ccid2.c 	seqno = DCCP_SKB_CB(skb)->dccpd_seq;
seqno             527 net/dccp/ccids/ccid2.c 		hc->tx_rpseq    = seqno;
seqno             530 net/dccp/ccids/ccid2.c 		if (dccp_delta_seqno(hc->tx_rpseq, seqno) == 1)
seqno             531 net/dccp/ccids/ccid2.c 			hc->tx_rpseq = seqno;
seqno             533 net/dccp/ccids/ccid2.c 		else if (after48(seqno, hc->tx_rpseq)) {
seqno              49 net/dccp/ccids/lib/packet_history.c int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno)
seqno              55 net/dccp/ccids/lib/packet_history.c 	entry->seqno = seqno;
seqno              38 net/dccp/ccids/lib/packet_history.h 	u64			  seqno;
seqno              43 net/dccp/ccids/lib/packet_history.h 	tfrc_tx_hist_find_entry(struct tfrc_tx_hist_entry *head, u64 seqno)
seqno              45 net/dccp/ccids/lib/packet_history.h 	while (head != NULL && head->seqno != seqno)
seqno              50 net/dccp/ccids/lib/packet_history.h int tfrc_tx_hist_add(struct tfrc_tx_hist_entry **headp, u64 seqno);
seqno             111 net/dccp/dccp.h static inline void dccp_set_seqno(u64 *seqno, u64 value)
seqno             113 net/dccp/dccp.h 	*seqno = value & UINT48_MAX;
seqno             116 net/dccp/dccp.h static inline void dccp_inc_seqno(u64 *seqno)
seqno             118 net/dccp/dccp.h 	*seqno = ADD48(*seqno, 1);
seqno             189 net/dccp/input.c 	u64 lswl, lawl, seqno = DCCP_SKB_CB(skb)->dccpd_seq,
seqno             207 net/dccp/input.c 		    dccp_delta_seqno(dp->dccps_swl, seqno) >= 0)
seqno             208 net/dccp/input.c 			dccp_update_gsr(sk, seqno);
seqno             234 net/dccp/input.c 	if (between48(seqno, lswl, dp->dccps_swh) &&
seqno             237 net/dccp/input.c 		dccp_update_gsr(sk, seqno);
seqno             265 net/dccp/input.c 			  (unsigned long long) lswl, (unsigned long long) seqno,
seqno             275 net/dccp/input.c 			seqno = dp->dccps_gsr;
seqno             276 net/dccp/input.c 		dccp_send_sync(sk, seqno, DCCP_PKT_SYNC);
seqno             455 net/ipv4/esp4.c 	memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8),
seqno             526 net/ipv4/esp4.c 	esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low +
seqno             270 net/ipv4/esp4_offload.c 	esp.seqno = cpu_to_be64(seq + ((u64)xo->seq.hi << 32));
seqno             396 net/ipv6/esp6.c 	memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8),
seqno             465 net/ipv6/esp6.c 	esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low +
seqno             298 net/ipv6/esp6_offload.c 	esp.seqno = cpu_to_be64(xo->seq.low + ((u64)xo->seq.hi << 32));
seqno             104 net/smc/smc.h  	u16				seqno;	/* connection seq # */
seqno             103 net/smc/smc_cdc.c 	conn->local_tx_ctrl.seqno = conn->tx_cdc_seq;
seqno             374 net/smc/smc_cdc.c 		if (smc_cdc_before(ntohs(cdc->seqno),
seqno             375 net/smc/smc_cdc.c 				   conn->local_rx_ctrl.seqno))
seqno              44 net/smc/smc_cdc.h 	__be16				seqno;
seqno             231 net/smc/smc_cdc.h 	peer->seqno = htons(local->seqno);
seqno             264 net/smc/smc_cdc.h 	local->seqno = ntohs(peer->seqno);
seqno            1958 net/sunrpc/auth_gss/auth_gss.c 	u32 len, offset, seqno, maj_stat;
seqno            1971 net/sunrpc/auth_gss/auth_gss.c 	if (xdr_stream_decode_u32(xdr, &seqno))
seqno            1973 net/sunrpc/auth_gss/auth_gss.c 	if (seqno != rqstp->rq_seqno)
seqno            2018 net/sunrpc/auth_gss/auth_gss.c 	trace_rpcgss_bad_seqno(task, rqstp->rq_seqno, seqno);
seqno             663 net/tipc/group.c 				    u32 event, u16 seqno,
seqno             690 net/tipc/group.c 	msg_set_grp_bc_seqno(hdr, seqno);
seqno             947 net/tipc/link.c 	u16 seqno = l->snd_nxt;
seqno             981 net/tipc/link.c 		msg_set_seqno(hdr, seqno);
seqno            1000 net/tipc/link.c 			seqno++;
seqno            1021 net/tipc/link.c 	l->snd_nxt = seqno;
seqno            1030 net/tipc/link.c 	u16 seqno = l->snd_nxt;
seqno            1055 net/tipc/link.c 		msg_set_seqno(hdr, seqno);
seqno            1060 net/tipc/link.c 		seqno++;
seqno            1062 net/tipc/link.c 	l->snd_nxt = seqno;
seqno            1273 net/tipc/link.c 	u16 seqno;
seqno            1308 net/tipc/link.c 		seqno = buf_seqno(iskb);
seqno            1309 net/tipc/link.c 		if (unlikely(less(seqno, l->drop_point))) {
seqno            1313 net/tipc/link.c 		if (unlikely(seqno != l->drop_point)) {
seqno            1314 net/tipc/link.c 			__tipc_skb_queue_sorted(fdefq, seqno, iskb);
seqno            1353 net/tipc/link.c 	u16 len, expect, seqno = 0;
seqno            1361 net/tipc/link.c 		seqno = buf_seqno(skb);
seqno            1362 net/tipc/link.c 		if (unlikely(more(seqno, expect))) {
seqno            1364 net/tipc/link.c 			ga->gacks[n].gap = htons(seqno - expect);
seqno            1369 net/tipc/link.c 		} else if (unlikely(less(seqno, expect))) {
seqno            1373 net/tipc/link.c 		expect = seqno + 1;
seqno            1377 net/tipc/link.c 	ga->gacks[n].ack = htons(seqno);
seqno            1409 net/tipc/link.c 	u16 seqno, n = 0;
seqno            1413 net/tipc/link.c 		seqno = buf_seqno(skb);
seqno            1416 net/tipc/link.c 		if (less_eq(seqno, acked)) {
seqno            1420 net/tipc/link.c 		} else if (less_eq(seqno, acked + gap)) {
seqno            1538 net/tipc/link.c 	u16 seqno, rcv_nxt, win_lim;
seqno            1550 net/tipc/link.c 		seqno = msg_seqno(hdr);
seqno            1561 net/tipc/link.c 		if (unlikely(less(seqno, rcv_nxt) || more(seqno, win_lim))) {
seqno            1574 net/tipc/link.c 		if (unlikely(seqno != rcv_nxt)) {
seqno            1575 net/tipc/link.c 			__tipc_skb_queue_sorted(defq, seqno, skb);
seqno            1722 net/tipc/link.c 	u16 pktlen, pktcnt, seqno = l->snd_nxt;
seqno            1786 net/tipc/link.c 			msg_set_seqno(hdr, seqno++);
seqno             760 net/tipc/msg.c void __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno,
seqno             765 net/tipc/msg.c 	if (skb_queue_empty(list) || less(seqno, buf_seqno(skb_peek(list)))) {
seqno             770 net/tipc/msg.c 	if (more(seqno, buf_seqno(skb_peek_tail(list)))) {
seqno             776 net/tipc/msg.c 		if (more(seqno, buf_seqno(_skb)))
seqno             778 net/tipc/msg.c 		if (seqno == buf_seqno(_skb))
seqno            1073 net/tipc/msg.h void __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno,
seqno            1192 net/tipc/msg.h 						 u16 seqno)
seqno            1196 net/tipc/msg.h 	if (skb && less_eq(buf_seqno(skb), seqno)) {
seqno             572 security/selinux/avc.c 				   int seqno, int is_insert)
seqno             580 security/selinux/avc.c 		if (seqno < avc->avc_cache.latest_notif) {
seqno             582 security/selinux/avc.c 			       seqno, avc->avc_cache.latest_notif);
seqno             586 security/selinux/avc.c 		if (seqno > avc->avc_cache.latest_notif)
seqno             587 security/selinux/avc.c 			avc->avc_cache.latest_notif = seqno;
seqno             623 security/selinux/avc.c 	if (avc_latest_notif_update(avc, avd->seqno, 1))
seqno             830 security/selinux/avc.c 			   u32 tsid, u16 tclass, u32 seqno,
seqno             873 security/selinux/avc.c 		    seqno == pos->ae.avd.seqno){
seqno             964 security/selinux/avc.c int avc_ss_reset(struct selinux_avc *avc, u32 seqno)
seqno             981 security/selinux/avc.c 	avc_latest_notif_update(avc, seqno, 0);
seqno            1021 security/selinux/avc.c 			xperm, ssid, tsid, tclass, avd->seqno, NULL, flags);
seqno            1084 security/selinux/avc.c 				driver, xperm, ssid, tsid, tclass, avd.seqno,
seqno              13 security/selinux/include/avc_ss.h int avc_ss_reset(struct selinux_avc *avc, u32 seqno);
seqno             195 security/selinux/include/security.h 	u32 seqno;
seqno             384 security/selinux/include/security.h 					     int seqno);
seqno             391 security/selinux/include/security.h extern void selnl_notify_policyload(u32 seqno);
seqno              58 security/selinux/netlink.c 		msg->seqno = *((u32 *)data);
seqno             103 security/selinux/netlink.c void selnl_notify_policyload(u32 seqno)
seqno             105 security/selinux/netlink.c 	selnl_notify(SELNL_MSG_POLICYLOAD, &seqno);
seqno             856 security/selinux/selinuxfs.c 			  avd.seqno, avd.flags);
seqno             940 security/selinux/ss/services.c 	avd->seqno = state->ss->latest_granting;
seqno            2093 security/selinux/ss/services.c 	u32 seqno;
seqno            2138 security/selinux/ss/services.c 		seqno = ++state->ss->latest_granting;
seqno            2140 security/selinux/ss/services.c 		avc_ss_reset(state->avc, seqno);
seqno            2141 security/selinux/ss/services.c 		selnl_notify_policyload(seqno);
seqno            2142 security/selinux/ss/services.c 		selinux_status_update_policyload(state, seqno);
seqno            2212 security/selinux/ss/services.c 	seqno = ++state->ss->latest_granting;
seqno            2221 security/selinux/ss/services.c 	avc_ss_reset(state->avc, seqno);
seqno            2222 security/selinux/ss/services.c 	selnl_notify_policyload(seqno);
seqno            2223 security/selinux/ss/services.c 	selinux_status_update_policyload(state, seqno);
seqno            2857 security/selinux/ss/services.c 	int lenp, seqno = 0;
seqno            2892 security/selinux/ss/services.c 	seqno = ++state->ss->latest_granting;
seqno            2897 security/selinux/ss/services.c 		avc_ss_reset(state->avc, seqno);
seqno            2898 security/selinux/ss/services.c 		selnl_notify_policyload(seqno);
seqno            2899 security/selinux/ss/services.c 		selinux_status_update_policyload(state, seqno);
seqno             106 security/selinux/ss/status.c 				      int seqno)
seqno             117 security/selinux/ss/status.c 		status->policyload = seqno;