Home
last modified time | relevance | path

Searched refs:context (Results 1 – 200 of 1624) sorted by relevance

123456789

/linux-4.1.27/drivers/staging/media/lirc/
Dlirc_imon.c189 static void free_imon_context(struct imon_context *context) in free_imon_context() argument
191 struct device *dev = context->driver->dev; in free_imon_context()
193 usb_free_urb(context->tx_urb); in free_imon_context()
194 usb_free_urb(context->rx_urb); in free_imon_context()
195 lirc_buffer_free(context->driver->rbuf); in free_imon_context()
196 kfree(context->driver->rbuf); in free_imon_context()
197 kfree(context->driver); in free_imon_context()
198 kfree(context); in free_imon_context()
203 static void deregister_from_lirc(struct imon_context *context) in deregister_from_lirc() argument
206 int minor = context->driver->minor; in deregister_from_lirc()
[all …]
Dlirc_sasem.c165 static void delete_context(struct sasem_context *context) in delete_context() argument
167 usb_free_urb(context->tx_urb); /* VFD */ in delete_context()
168 usb_free_urb(context->rx_urb); /* IR */ in delete_context()
169 lirc_buffer_free(context->driver->rbuf); in delete_context()
170 kfree(context->driver->rbuf); in delete_context()
171 kfree(context->driver); in delete_context()
172 kfree(context); in delete_context()
175 static void deregister_from_lirc(struct sasem_context *context) in deregister_from_lirc() argument
178 int minor = context->driver->minor; in deregister_from_lirc()
182 dev_err(&context->dev->dev, in deregister_from_lirc()
[all …]
/linux-4.1.27/drivers/misc/vmw_vmci/
Dvmci_context.c48 static void ctx_signal_notify(struct vmci_ctx *context) in ctx_signal_notify() argument
50 *context->notify = true; in ctx_signal_notify()
53 static void ctx_clear_notify(struct vmci_ctx *context) in ctx_clear_notify() argument
55 *context->notify = false; in ctx_clear_notify()
62 static void ctx_clear_notify_call(struct vmci_ctx *context) in ctx_clear_notify_call() argument
64 if (context->pending_datagrams == 0 && in ctx_clear_notify_call()
65 vmci_handle_arr_get_size(context->pending_doorbell_array) == 0) in ctx_clear_notify_call()
66 ctx_clear_notify(context); in ctx_clear_notify_call()
73 void vmci_ctx_check_signal_notify(struct vmci_ctx *context) in vmci_ctx_check_signal_notify() argument
75 spin_lock(&context->lock); in vmci_ctx_check_signal_notify()
[all …]
Dvmci_route.c50 if (VMCI_INVALID_ID == dst->context) in vmci_route()
54 if (VMCI_HYPERVISOR_CONTEXT_ID == dst->context) { in vmci_route()
72 if (VMCI_HOST_CONTEXT_ID == src->context) in vmci_route()
83 if (VMCI_INVALID_ID == src->context && in vmci_route()
85 src->context = vmci_get_context_id(); in vmci_route()
93 if (VMCI_HOST_CONTEXT_ID == dst->context) { in vmci_route()
102 if (src->context == VMCI_HYPERVISOR_CONTEXT_ID) { in vmci_route()
121 if (VMCI_INVALID_ID == src->context) in vmci_route()
122 src->context = vmci_get_context_id(); in vmci_route()
138 if (VMCI_INVALID_ID == src->context) { in vmci_route()
[all …]
Dvmci_context.h137 void vmci_ctx_destroy(struct vmci_ctx *context);
139 bool vmci_ctx_supports_host_qp(struct vmci_ctx *context);
141 int vmci_ctx_dequeue_datagram(struct vmci_ctx *context,
145 void vmci_ctx_put(struct vmci_ctx *context);
155 int vmci_ctx_qp_create(struct vmci_ctx *context, struct vmci_handle handle);
156 int vmci_ctx_qp_destroy(struct vmci_ctx *context, struct vmci_handle handle);
157 bool vmci_ctx_qp_exists(struct vmci_ctx *context, struct vmci_handle handle);
159 void vmci_ctx_check_signal_notify(struct vmci_ctx *context);
160 void vmci_ctx_unset_notify(struct vmci_ctx *context);
175 static inline u32 vmci_ctx_get_id(struct vmci_ctx *context) in vmci_ctx_get_id() argument
[all …]
Dvmci_host.c93 struct vmci_ctx *context; member
146 vmci_ctx_destroy(vmci_host_dev->context); in vmci_host_close()
147 vmci_host_dev->context = NULL; in vmci_host_close()
171 struct vmci_ctx *context = vmci_host_dev->context; in vmci_host_poll() local
177 poll_wait(filp, &context->host_context.wait_queue, in vmci_host_poll()
180 spin_lock(&context->lock); in vmci_host_poll()
181 if (context->pending_datagrams > 0 || in vmci_host_poll()
183 context->pending_doorbell_array) > 0) { in vmci_host_poll()
186 spin_unlock(&context->lock); in vmci_host_poll()
224 static int vmci_host_setup_notify(struct vmci_ctx *context, in vmci_host_setup_notify() argument
[all …]
Dvmci_datagram.c100 handle.context, handle.resource, result); in dg_create_handle()
174 if (dg->dst.context == VMCI_HYPERVISOR_CONTEXT_ID) in dg_dispatch_as_host()
178 if (dg->src.context != context_id) { in dg_dispatch_as_host()
180 context_id, dg->src.context, dg->src.resource); in dg_dispatch_as_host()
189 dg->src.context, dg->src.resource); in dg_dispatch_as_host()
194 if (dg->dst.context == VMCI_HOST_CONTEXT_ID) { in dg_dispatch_as_host()
199 if (dg->src.context == VMCI_HYPERVISOR_CONTEXT_ID && in dg_dispatch_as_host()
208 dg->dst.context, dg->dst.resource); in dg_dispatch_as_host()
225 dg->src.context == VMCI_HOST_CONTEXT_ID) { in dg_dispatch_as_host()
261 if (context_id != dg->dst.context) { in dg_dispatch_as_host()
[all …]
Dvmci_doorbell.c95 if (priv_flags == NULL || handle.context == VMCI_INVALID_ID) in vmci_dbell_get_priv_flags()
98 if (handle.context == VMCI_HOST_CONTEXT_ID) { in vmci_dbell_get_priv_flags()
110 } else if (handle.context == VMCI_HYPERVISOR_CONTEXT_ID) { in vmci_dbell_get_priv_flags()
117 *priv_flags = vmci_context_get_priv_flags(handle.context); in vmci_dbell_get_priv_flags()
307 handle.context, handle.resource); in vmci_dbell_host_context_notify()
315 handle.context, handle.resource); in vmci_dbell_host_context_notify()
446 if (handle->context == VMCI_HOST_CONTEXT_ID || in vmci_doorbell_create()
448 vmci_get_context_id() == handle->context)) { in vmci_doorbell_create()
454 handle->context, handle->resource); in vmci_doorbell_create()
476 new_handle.context, new_handle.resource, result); in vmci_doorbell_create()
[all …]
Dvmci_queue_pair.c1004 handle.context, handle.resource, result); in qp_guest_endpoint_create()
1286 if (queue_pair_entry->qp.handle.context != context_id || in qp_alloc_guest_work()
1373 struct vmci_ctx *context, in qp_broker_create() argument
1378 const u32 context_id = vmci_ctx_get_id(context); in qp_broker_create()
1392 if (handle.context != context_id && handle.context != peer) in qp_broker_create()
1409 if (vmci_ctx_get_id(context) == VMCI_HOST_CONTEXT_ID && !is_local) { in qp_broker_create()
1435 !!(context->priv_flags & VMCI_PRIVILEGE_FLAG_RESTRICTED); in qp_broker_create()
1503 handle.context, handle.resource, result); in qp_broker_create()
1515 vmci_ctx_qp_create(context, entry->qp.handle); in qp_broker_create()
1601 struct vmci_ctx *context, in qp_broker_attach() argument
[all …]
Dvmci_queue_pair.h154 struct vmci_ctx *context);
157 struct vmci_ctx *context);
158 int vmci_qp_broker_detach(struct vmci_handle handle, struct vmci_ctx *context);
169 struct vmci_ctx *context, u64 guest_mem);
171 struct vmci_ctx *context, u32 gid);
/linux-4.1.27/security/selinux/ss/
Dmls.h27 int mls_compute_context_len(struct context *context);
28 void mls_sid_to_context(struct context *context, char **scontext);
29 int mls_context_isvalid(struct policydb *p, struct context *c);
36 struct context *context,
40 int mls_from_string(char *str, struct context *context, gfp_t gfp_mask);
42 int mls_range_set(struct context *context, struct mls_range *range);
46 struct context *context);
48 int mls_compute_sid(struct context *scontext,
49 struct context *tcontext,
52 struct context *newcontext,
[all …]
Dmls.c35 int mls_compute_context_len(struct context *context) in mls_compute_context_len() argument
47 int index_sens = context->range.level[l].sens; in mls_compute_context_len()
53 e = &context->range.level[l].cat; in mls_compute_context_len()
72 if (mls_level_eq(&context->range.level[0], in mls_compute_context_len()
73 &context->range.level[1])) in mls_compute_context_len()
88 void mls_sid_to_context(struct context *context, in mls_sid_to_context() argument
106 context->range.level[l].sens - 1)); in mls_sid_to_context()
112 e = &context->range.level[l].cat; in mls_sid_to_context()
148 if (mls_level_eq(&context->range.level[0], in mls_sid_to_context()
149 &context->range.level[1])) in mls_sid_to_context()
[all …]
Dsidtab.c33 int sidtab_insert(struct sidtab *s, u32 sid, struct context *context) in sidtab_insert() argument
62 if (context_cpy(&newnode->context, context)) { in sidtab_insert()
85 static struct context *sidtab_search_core(struct sidtab *s, u32 sid, int force) in sidtab_search_core()
98 if (force && cur && sid == cur->sid && cur->context.len) in sidtab_search_core()
99 return &cur->context; in sidtab_search_core()
101 if (cur == NULL || sid != cur->sid || cur->context.len) { in sidtab_search_core()
112 return &cur->context; in sidtab_search_core()
115 struct context *sidtab_search(struct sidtab *s, u32 sid) in sidtab_search()
120 struct context *sidtab_search_force(struct sidtab *s, u32 sid) in sidtab_search_force()
127 struct context *context, in sidtab_map() argument
[all …]
Dsidtab.h14 struct context context; /* security context structure */ member
35 int sidtab_insert(struct sidtab *s, u32 sid, struct context *context);
36 struct context *sidtab_search(struct sidtab *s, u32 sid);
37 struct context *sidtab_search_force(struct sidtab *s, u32 sid);
41 struct context *context,
46 struct context *context,
Dcontext.h26 struct context { struct
35 static inline void mls_context_init(struct context *c) in mls_context_init()
40 static inline int mls_context_cpy(struct context *dst, struct context *src) in mls_context_cpy()
60 static inline int mls_context_cpy_low(struct context *dst, struct context *src) in mls_context_cpy_low()
80 static inline int mls_context_cpy_high(struct context *dst, struct context *src) in mls_context_cpy_high()
97 static inline int mls_context_cmp(struct context *c1, struct context *c2) in mls_context_cmp()
105 static inline void mls_context_destroy(struct context *c) in mls_context_destroy()
112 static inline void context_init(struct context *c) in context_init()
117 static inline int context_cpy(struct context *dst, struct context *src) in context_cpy()
141 static inline void context_destroy(struct context *c) in context_destroy()
[all …]
Dservices.c92 static int context_struct_to_string(struct context *context, char **scontext,
95 static void context_struct_compute_av(struct context *scontext,
96 struct context *tcontext,
271 static int constraint_expr_eval(struct context *scontext, in constraint_expr_eval()
272 struct context *tcontext, in constraint_expr_eval()
273 struct context *xcontext, in constraint_expr_eval()
277 struct context *c; in constraint_expr_eval()
461 static void security_dump_masked_av(struct context *scontext, in security_dump_masked_av()
462 struct context *tcontext, in security_dump_masked_av()
539 static void type_attribute_bounds_av(struct context *scontext, in type_attribute_bounds_av()
[all …]
/linux-4.1.27/kernel/
Dauditsc.c825 struct audit_context *context = tsk->audit_context; in audit_take_context() local
827 if (!context) in audit_take_context()
829 context->return_valid = return_valid; in audit_take_context()
845 context->return_code = -EINTR; in audit_take_context()
847 context->return_code = return_code; in audit_take_context()
849 if (context->in_syscall && !context->dummy) { in audit_take_context()
850 audit_filter_syscall(tsk, context, &audit_filter_list[AUDIT_FILTER_EXIT]); in audit_take_context()
851 audit_filter_inodes(tsk, context); in audit_take_context()
855 return context; in audit_take_context()
858 static inline void audit_proctitle_free(struct audit_context *context) in audit_proctitle_free() argument
[all …]
/linux-4.1.27/fs/xfs/
Dxfs_attr_list.c69 xfs_attr_shortform_list(xfs_attr_list_context_t *context) in xfs_attr_shortform_list() argument
79 ASSERT(context != NULL); in xfs_attr_shortform_list()
80 dp = context->dp; in xfs_attr_shortform_list()
87 cursor = context->cursor; in xfs_attr_shortform_list()
90 trace_xfs_attr_list_sf(context); in xfs_attr_shortform_list()
101 if (context->bufsize == 0 || in xfs_attr_shortform_list()
103 (dp->i_afp->if_bytes + sf->hdr.count * 16) < context->bufsize)) { in xfs_attr_shortform_list()
105 error = context->put_listent(context, in xfs_attr_shortform_list()
116 if (context->seen_enough) in xfs_attr_shortform_list()
123 trace_xfs_attr_list_sf_all(context); in xfs_attr_shortform_list()
[all …]
Dxfs_xattr.c131 struct xfs_attr_list_context *context, in xfs_xattr_put_listent() argument
142 ASSERT(context->count >= 0); in xfs_xattr_put_listent()
151 arraytop = context->count + prefix_len + namelen + 1; in xfs_xattr_put_listent()
152 if (arraytop > context->firstu) { in xfs_xattr_put_listent()
153 context->count = -1; /* insufficient space */ in xfs_xattr_put_listent()
156 offset = (char *)context->alist + context->count; in xfs_xattr_put_listent()
162 context->count += prefix_len + namelen + 1; in xfs_xattr_put_listent()
168 struct xfs_attr_list_context *context, in xfs_xattr_put_listent_sizes() argument
175 context->count += xfs_xattr_prefix_len(flags) + namelen + 1; in xfs_xattr_put_listent_sizes()
198 struct xfs_attr_list_context context; in xfs_vn_listxattr() local
[all …]
/linux-4.1.27/drivers/infiniband/core/
Dumem_odp.c82 static void ib_ucontext_notifier_start_account(struct ib_ucontext *context) in ib_ucontext_notifier_start_account() argument
84 atomic_inc(&context->notifier_count); in ib_ucontext_notifier_start_account()
91 static void ib_ucontext_notifier_end_account(struct ib_ucontext *context) in ib_ucontext_notifier_end_account() argument
93 int zero_notifiers = atomic_dec_and_test(&context->notifier_count); in ib_ucontext_notifier_end_account()
96 !list_empty(&context->no_private_counters)) { in ib_ucontext_notifier_end_account()
103 down_write(&context->umem_rwsem); in ib_ucontext_notifier_end_account()
107 if (!atomic_read(&context->notifier_count)) { in ib_ucontext_notifier_end_account()
109 &context->no_private_counters, in ib_ucontext_notifier_end_account()
119 up_write(&context->umem_rwsem); in ib_ucontext_notifier_end_account()
135 item->context->invalidate_range(item, ib_umem_start(item), in ib_umem_notifier_release_trampoline()
[all …]
Ddevice.c205 struct ib_client_data *context; in add_client_context() local
208 context = kmalloc(sizeof *context, GFP_KERNEL); in add_client_context()
209 if (!context) { in add_client_context()
215 context->client = client; in add_client_context()
216 context->data = NULL; in add_client_context()
219 list_add(&context->list, &device->client_data_list); in add_client_context()
340 struct ib_client_data *context, *tmp; in ib_unregister_device() local
359 list_for_each_entry_safe(context, tmp, &device->client_data_list, list) in ib_unregister_device()
360 kfree(context); in ib_unregister_device()
407 struct ib_client_data *context, *tmp; in ib_unregister_client() local
[all …]
Dumem.c83 struct ib_umem *ib_umem_get(struct ib_ucontext *context, unsigned long addr, in ib_umem_get() argument
120 umem->context = context; in ib_umem_get()
137 ret = ib_umem_odp_get(context, umem); in ib_umem_get()
214 umem->nmap = ib_dma_map_sg_attrs(context->device, in ib_umem_get()
230 __ib_umem_release(context->device, umem, 0); in ib_umem_get()
262 struct ib_ucontext *context = umem->context; in ib_umem_release() local
272 __ib_umem_release(umem->context->device, umem, 1); in ib_umem_release()
293 if (context->closing) { in ib_umem_release()
Dsa_query.c87 void *context; member
93 void *context; member
99 void *context; member
105 void *context; member
638 query->mad_buf->context[0] = query; in send_mad()
683 query->callback(status, &rec, query->context); in ib_sa_path_rec_callback()
685 query->callback(status, NULL, query->context); in ib_sa_path_rec_callback()
725 void *context), in ib_sa_path_rec_get() argument
726 void *context, in ib_sa_path_rec_get()
754 query->context = context; in ib_sa_path_rec_get()
[all …]
/linux-4.1.27/drivers/tty/serial/
Dtilegx.c49 gxio_uart_context_t context; member
71 gxio_uart_context_t *context = &tile_uart->context; in receive_chars() local
74 count.word = gxio_uart_read(context, UART_FIFO_COUNT); in receive_chars()
76 c = (char)gxio_uart_read(context, UART_RECEIVE_DATA); in receive_chars()
89 gxio_uart_context_t *context = &tile_uart->context; in handle_receive() local
98 gxio_uart_write(context, UART_INTERRUPT_STATUS, in handle_receive()
117 static int tilegx_putchar(gxio_uart_context_t *context, char c) in tilegx_putchar() argument
120 flag.word = gxio_uart_read(context, UART_FLAG); in tilegx_putchar()
124 gxio_uart_write(context, UART_TRANSMIT_DATA, (unsigned long)c); in tilegx_putchar()
137 gxio_uart_context_t *context = &tile_uart->context; in handle_transmit() local
[all …]
/linux-4.1.27/arch/tile/gxio/
Dmpipe.c58 int gxio_mpipe_init(gxio_mpipe_context_t *context, unsigned int mpipe_index) in gxio_mpipe_init() argument
71 context->fd = fd; in gxio_mpipe_init()
81 context->mmio_cfg_base = (void __force *) in gxio_mpipe_init()
84 if (context->mmio_cfg_base == NULL) in gxio_mpipe_init()
87 context->mmio_fast_base = (void __force *) in gxio_mpipe_init()
90 if (context->mmio_fast_base == NULL) in gxio_mpipe_init()
95 context->__stacks.stacks[i] = 255; in gxio_mpipe_init()
97 context->instance = mpipe_index; in gxio_mpipe_init()
102 iounmap((void __force __iomem *)(context->mmio_cfg_base)); in gxio_mpipe_init()
104 hv_dev_close(context->fd); in gxio_mpipe_init()
[all …]
Duart.c28 int gxio_uart_init(gxio_uart_context_t *context, int uart_index) in gxio_uart_init() argument
42 context->fd = fd; in gxio_uart_init()
45 context->mmio_base = (void __force *) in gxio_uart_init()
48 if (context->mmio_base == NULL) { in gxio_uart_init()
49 hv_dev_close(context->fd); in gxio_uart_init()
50 context->fd = -1; in gxio_uart_init()
59 int gxio_uart_destroy(gxio_uart_context_t *context) in gxio_uart_destroy() argument
61 iounmap((void __force __iomem *)(context->mmio_base)); in gxio_uart_destroy()
62 hv_dev_close(context->fd); in gxio_uart_destroy()
64 context->mmio_base = NULL; in gxio_uart_destroy()
[all …]
Diorpc_mpipe.c24 int gxio_mpipe_alloc_buffer_stacks(gxio_mpipe_context_t *context, in gxio_mpipe_alloc_buffer_stacks() argument
35 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_alloc_buffer_stacks()
48 int gxio_mpipe_init_buffer_stack_aux(gxio_mpipe_context_t *context, in gxio_mpipe_init_buffer_stack_aux() argument
69 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_init_buffer_stack_aux()
83 int gxio_mpipe_alloc_notif_rings(gxio_mpipe_context_t *context, in gxio_mpipe_alloc_notif_rings() argument
94 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_alloc_notif_rings()
105 int gxio_mpipe_init_notif_ring_aux(gxio_mpipe_context_t *context, void *mem_va, in gxio_mpipe_init_notif_ring_aux() argument
124 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_init_notif_ring_aux()
136 int gxio_mpipe_request_notif_ring_interrupt(gxio_mpipe_context_t *context, in gxio_mpipe_request_notif_ring_interrupt() argument
150 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_request_notif_ring_interrupt()
[all …]
Dusb_host.c29 int gxio_usb_host_init(gxio_usb_host_context_t *context, int usb_index, in gxio_usb_host_init() argument
50 context->fd = fd; in gxio_usb_host_init()
53 context->mmio_base = in gxio_usb_host_init()
56 if (context->mmio_base == NULL) { in gxio_usb_host_init()
57 hv_dev_close(context->fd); in gxio_usb_host_init()
66 int gxio_usb_host_destroy(gxio_usb_host_context_t *context) in gxio_usb_host_destroy() argument
68 iounmap((void __force __iomem *)(context->mmio_base)); in gxio_usb_host_destroy()
69 hv_dev_close(context->fd); in gxio_usb_host_destroy()
71 context->mmio_base = NULL; in gxio_usb_host_destroy()
72 context->fd = -1; in gxio_usb_host_destroy()
[all …]
Diorpc_trio.c24 int gxio_trio_alloc_asids(gxio_trio_context_t *context, unsigned int count, in gxio_trio_alloc_asids() argument
34 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_trio_alloc_asids()
47 int gxio_trio_alloc_memory_maps(gxio_trio_context_t *context, in gxio_trio_alloc_memory_maps() argument
58 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_trio_alloc_memory_maps()
70 int gxio_trio_alloc_scatter_queues(gxio_trio_context_t *context, in gxio_trio_alloc_scatter_queues() argument
81 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_trio_alloc_scatter_queues()
94 int gxio_trio_alloc_pio_regions(gxio_trio_context_t *context, in gxio_trio_alloc_pio_regions() argument
105 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_trio_alloc_pio_regions()
118 int gxio_trio_init_pio_region_aux(gxio_trio_context_t *context, in gxio_trio_init_pio_region_aux() argument
130 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_trio_init_pio_region_aux()
[all …]
Diorpc_mpipe_info.c22 int gxio_mpipe_info_instance_aux(gxio_mpipe_info_context_t *context, in gxio_mpipe_info_instance_aux() argument
30 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_info_instance_aux()
41 int gxio_mpipe_info_enumerate_aux(gxio_mpipe_info_context_t *context, in gxio_mpipe_info_enumerate_aux() argument
51 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), in gxio_mpipe_info_enumerate_aux()
66 int gxio_mpipe_info_get_mmio_base(gxio_mpipe_info_context_t *context, in gxio_mpipe_info_get_mmio_base() argument
74 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), in gxio_mpipe_info_get_mmio_base()
88 int gxio_mpipe_info_check_mmio_offset(gxio_mpipe_info_context_t *context, in gxio_mpipe_info_check_mmio_offset() argument
97 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_mpipe_info_check_mmio_offset()
Diorpc_usb_host.c22 int gxio_usb_host_cfg_interrupt(gxio_usb_host_context_t *context, int inter_x, in gxio_usb_host_cfg_interrupt() argument
33 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_usb_host_cfg_interrupt()
44 int gxio_usb_host_register_client_memory(gxio_usb_host_context_t *context, in gxio_usb_host_register_client_memory() argument
53 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_usb_host_register_client_memory()
64 int gxio_usb_host_get_mmio_base(gxio_usb_host_context_t *context, HV_PTE *base) in gxio_usb_host_get_mmio_base() argument
71 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), in gxio_usb_host_get_mmio_base()
85 int gxio_usb_host_check_mmio_offset(gxio_usb_host_context_t *context, in gxio_usb_host_check_mmio_offset() argument
94 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_usb_host_check_mmio_offset()
Diorpc_uart.c22 int gxio_uart_cfg_interrupt(gxio_uart_context_t *context, int inter_x, in gxio_uart_cfg_interrupt() argument
33 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_uart_cfg_interrupt()
43 int gxio_uart_get_mmio_base(gxio_uart_context_t *context, HV_PTE *base) in gxio_uart_get_mmio_base() argument
50 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), in gxio_uart_get_mmio_base()
64 int gxio_uart_check_mmio_offset(gxio_uart_context_t *context, in gxio_uart_check_mmio_offset() argument
73 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, in gxio_uart_check_mmio_offset()
Dtrio.c28 int gxio_trio_init(gxio_trio_context_t *context, unsigned int trio_index) in gxio_trio_init() argument
36 context->fd = -1; in gxio_trio_init()
44 context->fd = fd; in gxio_trio_init()
/linux-4.1.27/arch/tile/include/gxio/
Diorpc_mpipe.h59 int gxio_mpipe_alloc_buffer_stacks(gxio_mpipe_context_t *context,
63 int gxio_mpipe_init_buffer_stack_aux(gxio_mpipe_context_t *context,
69 int gxio_mpipe_alloc_notif_rings(gxio_mpipe_context_t *context,
73 int gxio_mpipe_init_notif_ring_aux(gxio_mpipe_context_t *context, void *mem_va,
77 int gxio_mpipe_request_notif_ring_interrupt(gxio_mpipe_context_t *context,
82 int gxio_mpipe_enable_notif_ring_interrupt(gxio_mpipe_context_t *context,
85 int gxio_mpipe_alloc_notif_groups(gxio_mpipe_context_t *context,
89 int gxio_mpipe_init_notif_group(gxio_mpipe_context_t *context,
93 int gxio_mpipe_alloc_buckets(gxio_mpipe_context_t *context, unsigned int count,
96 int gxio_mpipe_init_bucket(gxio_mpipe_context_t *context, unsigned int bucket,
[all …]
Dmpipe.h347 extern int gxio_mpipe_init(gxio_mpipe_context_t *context,
360 extern int gxio_mpipe_destroy(gxio_mpipe_context_t *context);
378 extern int gxio_mpipe_alloc_buffer_stacks(gxio_mpipe_context_t *context,
439 extern int gxio_mpipe_init_buffer_stack(gxio_mpipe_context_t *context,
457 static inline void gxio_mpipe_push_buffer(gxio_mpipe_context_t *context, in gxio_mpipe_push_buffer() argument
478 __gxio_mmio_write(context->mmio_fast_base + offset.word, val.word); in gxio_mpipe_push_buffer()
487 static inline void *gxio_mpipe_pop_buffer(gxio_mpipe_context_t *context, in gxio_mpipe_pop_buffer() argument
509 __gxio_mmio_read(context->mmio_fast_base + in gxio_mpipe_pop_buffer()
541 extern int gxio_mpipe_alloc_notif_rings(gxio_mpipe_context_t *context,
558 extern int gxio_mpipe_init_notif_ring(gxio_mpipe_context_t *context,
[all …]
Diorpc_trio.h49 int gxio_trio_alloc_asids(gxio_trio_context_t *context, unsigned int count,
53 int gxio_trio_alloc_memory_maps(gxio_trio_context_t *context,
58 int gxio_trio_alloc_scatter_queues(gxio_trio_context_t *context,
62 int gxio_trio_alloc_pio_regions(gxio_trio_context_t *context,
66 int gxio_trio_init_pio_region_aux(gxio_trio_context_t *context,
71 int gxio_trio_init_memory_map_mmu_aux(gxio_trio_context_t *context,
78 int gxio_trio_get_port_property(gxio_trio_context_t *context,
81 int gxio_trio_config_legacy_intr(gxio_trio_context_t *context, int inter_x,
85 int gxio_trio_config_msi_intr(gxio_trio_context_t *context, int inter_x,
92 int gxio_trio_set_mps_mrs(gxio_trio_context_t *context, uint16_t mps,
[all …]
Duart.h58 extern int gxio_uart_cfg_interrupt(gxio_uart_context_t *context,
74 extern int gxio_uart_init(gxio_uart_context_t *context, int uart_index);
88 extern int gxio_uart_destroy(gxio_uart_context_t *context);
95 extern void gxio_uart_write(gxio_uart_context_t *context, uint64_t offset,
103 extern uint64_t gxio_uart_read(gxio_uart_context_t *context, uint64_t offset);
Dusb_host.h56 extern int gxio_usb_host_init(gxio_usb_host_context_t *context, int usb_index,
71 extern int gxio_usb_host_destroy(gxio_usb_host_context_t *context);
78 extern void *gxio_usb_host_get_reg_start(gxio_usb_host_context_t *context);
85 extern size_t gxio_usb_host_get_reg_len(gxio_usb_host_context_t *context);
Diorpc_mpipe_info.h36 int gxio_mpipe_info_instance_aux(gxio_mpipe_info_context_t *context,
39 int gxio_mpipe_info_enumerate_aux(gxio_mpipe_info_context_t *context,
44 int gxio_mpipe_info_get_mmio_base(gxio_mpipe_info_context_t *context,
47 int gxio_mpipe_info_check_mmio_offset(gxio_mpipe_info_context_t *context,
Diorpc_usb_host.h34 int gxio_usb_host_cfg_interrupt(gxio_usb_host_context_t *context, int inter_x,
37 int gxio_usb_host_register_client_memory(gxio_usb_host_context_t *context,
40 int gxio_usb_host_get_mmio_base(gxio_usb_host_context_t *context,
43 int gxio_usb_host_check_mmio_offset(gxio_usb_host_context_t *context,
Diorpc_uart.h32 int gxio_uart_cfg_interrupt(gxio_uart_context_t *context, int inter_x,
35 int gxio_uart_get_mmio_base(gxio_uart_context_t *context, HV_PTE *base);
37 int gxio_uart_check_mmio_offset(gxio_uart_context_t *context,
Dtrio.h214 extern int gxio_trio_init(gxio_trio_context_t *context,
249 extern int gxio_trio_init_memory_map(gxio_trio_context_t *context,
294 extern int gxio_trio_alloc_asids(gxio_trio_context_t *context,
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
Den_resources.c42 int user_prio, struct mlx4_qp_context *context) in mlx4_en_fill_qp_context() argument
47 memset(context, 0, sizeof *context); in mlx4_en_fill_qp_context()
48 context->flags = cpu_to_be32(7 << 16 | rss << MLX4_RSS_QPC_FLAG_OFFSET); in mlx4_en_fill_qp_context()
49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context()
50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context()
52 context->rq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context()
54 context->sq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context()
56 context->params2 |= MLX4_QP_BIT_FPP; in mlx4_en_fill_qp_context()
59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context()
61 context->usr_page = cpu_to_be32(mdev->priv_uar.index); in mlx4_en_fill_qp_context()
[all …]
Dintf.c44 void *context; member
60 dev_ctx->context = intf->add(&priv->dev); in mlx4_add_device()
62 if (dev_ctx->context) { in mlx4_add_device()
80 intf->remove(&priv->dev, dev_ctx->context); in mlx4_remove_device()
164 dev_ctx->intf->remove(dev, dev_ctx->context); in mlx4_do_bond()
165 dev_ctx->context = dev_ctx->intf->add(dev); in mlx4_do_bond()
189 dev_ctx->intf->event(dev, dev_ctx->context, type, param); in mlx4_dispatch_event()
240 result = dev_ctx->intf->get_dev(dev, dev_ctx->context, port); in mlx4_get_protocol_dev()
/linux-4.1.27/arch/s390/include/asm/
Dmmu_context.h18 spin_lock_init(&mm->context.list_lock); in init_new_context()
19 INIT_LIST_HEAD(&mm->context.pgtable_list); in init_new_context()
20 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context()
21 cpumask_clear(&mm->context.cpu_attach_mask); in init_new_context()
22 atomic_set(&mm->context.attach_count, 0); in init_new_context()
23 mm->context.flush_mm = 0; in init_new_context()
25 mm->context.alloc_pgste = page_table_allocate_pgste; in init_new_context()
26 mm->context.has_pgste = 0; in init_new_context()
27 mm->context.use_skey = 0; in init_new_context()
29 if (mm->context.asce_limit == 0) { in init_new_context()
[all …]
Dtlbflush.h67 atomic_add(0x10000, &mm->context.attach_count); in __tlb_flush_full()
77 &mm->context.cpu_attach_mask); in __tlb_flush_full()
79 atomic_sub(0x10000, &mm->context.attach_count); in __tlb_flush_full()
92 count = atomic_add_return(0x10000, &mm->context.attach_count); in __tlb_flush_asce()
104 &mm->context.cpu_attach_mask); in __tlb_flush_asce()
106 atomic_sub(0x10000, &mm->context.attach_count); in __tlb_flush_asce()
114 init_mm.context.asce_bits); in __tlb_flush_kernel()
137 init_mm.context.asce_bits); in __tlb_flush_kernel()
150 if (MACHINE_HAS_IDTE && list_empty(&mm->context.gmap_list)) in __tlb_flush_mm()
152 mm->context.asce_bits); in __tlb_flush_mm()
[all …]
Dmmu.h26 .context.list_lock = __SPIN_LOCK_UNLOCKED(name.context.list_lock), \
27 .context.pgtable_list = LIST_HEAD_INIT(name.context.pgtable_list), \
28 .context.gmap_list = LIST_HEAD_INIT(name.context.gmap_list),
/linux-4.1.27/arch/sparc/mm/
Dtsb.c77 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user()
79 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user()
80 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user()
86 if (mm->context.tsb_block[MM_TSB_HUGE].tsb) { in flush_tsb_user()
87 base = (unsigned long) mm->context.tsb_block[MM_TSB_HUGE].tsb; in flush_tsb_user()
88 nentries = mm->context.tsb_block[MM_TSB_HUGE].tsb_nentries; in flush_tsb_user()
94 spin_unlock_irqrestore(&mm->context.lock, flags); in flush_tsb_user()
101 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user_page()
103 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user_page()
104 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user_page()
[all …]
/linux-4.1.27/drivers/usb/image/
Dmicrotek.c190 MTS_DEBUG("transfer = 0x%x context = 0x%x\n",(int)transfer,(int)context ); \
191 …MTS_DEBUG("status = 0x%x data-length = 0x%x sent = 0x%x\n",transfer->status,(int)context->data_len…
192 mts_debug_dump(context->instance);\
207 struct mts_transfer_context* context = (struct mts_transfer_context*)transfer->context; \
384 context->instance->usb_dev, in mts_int_submit_urb()
389 context in mts_int_submit_urb()
395 context->srb->result = DID_ERROR << 16; in mts_int_submit_urb()
406 if ( likely(context->final_callback != NULL) ) in mts_transfer_cleanup()
407 context->final_callback(context->srb); in mts_transfer_cleanup()
414 context->srb->result &= MTS_SCSI_ERR_MASK; in mts_transfer_done()
[all …]
/linux-4.1.27/drivers/pci/hotplug/
Dacpiphp_glue.c64 static void hotplug_event(u32 type, struct acpiphp_context *context);
75 struct acpiphp_context *context; in acpiphp_init_context() local
77 context = kzalloc(sizeof(*context), GFP_KERNEL); in acpiphp_init_context()
78 if (!context) in acpiphp_init_context()
81 context->refcount = 1; in acpiphp_init_context()
82 context->hp.notify = acpiphp_hotplug_notify; in acpiphp_init_context()
83 context->hp.fixup = acpiphp_post_dock_fixup; in acpiphp_init_context()
84 acpi_set_hp_context(adev, &context->hp); in acpiphp_init_context()
85 return context; in acpiphp_init_context()
96 struct acpiphp_context *context; in acpiphp_get_context() local
[all …]
/linux-4.1.27/arch/ia64/include/asm/
Dmmu_context.h81 nv_mm_context_t context = mm->context; in get_mmu_context() local
83 if (likely(context)) in get_mmu_context()
88 context = mm->context; in get_mmu_context()
89 if (context == 0) { in get_mmu_context()
99 mm->context = context = ia64_ctx.next++; in get_mmu_context()
100 __set_bit(context, ia64_ctx.bitmap); in get_mmu_context()
110 return context; in get_mmu_context()
120 mm->context = 0; in init_new_context()
131 reload_context (nv_mm_context_t context) in reload_context() argument
138 rid = context << 3; /* make space for encoding the region number */ in reload_context()
[all …]
/linux-4.1.27/arch/parisc/include/asm/
Dmmu_context.h26 mm->context = alloc_sid(); in init_new_context()
33 free_sid(mm->context); in destroy_context()
34 mm->context = 0; in destroy_context()
37 static inline unsigned long __space_to_prot(mm_context_t context) in __space_to_prot() argument
40 return context << 1; in __space_to_prot()
42 return context >> (SPACEID_SHIFT - 1); in __space_to_prot()
46 static inline void load_context(mm_context_t context) in load_context() argument
48 mtsp(context, 3); in load_context()
49 mtctl(__space_to_prot(context), 8); in load_context()
57 load_context(next->context); in switch_mm()
[all …]
Dtlbflush.h34 __flush_tlb_range((vma)->vm_mm->context, start, end)
69 if (mm->context != 0) in flush_tlb_mm()
70 free_sid(mm->context); in flush_tlb_mm()
71 mm->context = alloc_sid(); in flush_tlb_mm()
73 load_context(mm->context); in flush_tlb_mm()
83 sid = vma->vm_mm->context; in flush_tlb_page()
/linux-4.1.27/fs/ocfs2/
Dmove_extents.c59 struct ocfs2_move_extents_context *context, in __ocfs2_move_extent() argument
64 struct inode *inode = context->inode; in __ocfs2_move_extent()
69 u64 ino = ocfs2_metadata_cache_owner(context->et.et_ci); in __ocfs2_move_extent()
85 path = ocfs2_new_path_from_et(&context->et); in __ocfs2_move_extent()
120 context->et.et_root_bh, in __ocfs2_move_extent()
127 ret = ocfs2_split_extent(handle, &context->et, path, index, in __ocfs2_move_extent()
128 &replace_rec, context->meta_ac, in __ocfs2_move_extent()
129 &context->dealloc); in __ocfs2_move_extent()
135 ocfs2_journal_dirty(handle, context->et.et_root_bh); in __ocfs2_move_extent()
137 context->new_phys_cpos = new_p_cpos; in __ocfs2_move_extent()
[all …]
Drefcounttree.c64 int (*get_clusters)(struct ocfs2_cow_context *context,
3130 struct ocfs2_cow_context *context, in ocfs2_replace_clusters() argument
3136 struct ocfs2_caching_info *ci = context->data_et.et_ci; in ocfs2_replace_clusters()
3144 ret = context->cow_duplicate_clusters(handle, context->inode, in ocfs2_replace_clusters()
3152 ret = ocfs2_clear_ext_refcount(handle, &context->data_et, in ocfs2_replace_clusters()
3154 context->meta_ac, &context->dealloc); in ocfs2_replace_clusters()
3211 static int ocfs2_di_get_clusters(struct ocfs2_cow_context *context, in ocfs2_di_get_clusters() argument
3216 return ocfs2_get_clusters(context->inode, v_cluster, p_cluster, in ocfs2_di_get_clusters()
3221 struct ocfs2_cow_context *context, in ocfs2_make_clusters_writable() argument
3231 struct ocfs2_caching_info *ref_ci = &context->ref_tree->rf_ci; in ocfs2_make_clusters_writable()
[all …]
/linux-4.1.27/arch/m68k/include/asm/
Dmmu_context.h34 if (mm->context != NO_CONTEXT) in get_mmu_context()
47 mm->context = ctx; in get_mmu_context()
54 #define init_new_context(tsk, mm) (((mm)->context = NO_CONTEXT), 0)
61 if (mm->context != NO_CONTEXT) { in destroy_context()
62 clear_bit(mm->context, context_map); in destroy_context()
63 mm->context = NO_CONTEXT; in destroy_context()
68 static inline void set_context(mm_context_t context, pgd_t *pgd) in set_context() argument
70 __asm__ __volatile__ ("movec %0,%%asid" : : "d" (context)); in set_context()
77 set_context(tsk->mm->context, next->pgd); in switch_mm()
88 set_context(mm->context, mm->pgd); in activate_mm()
[all …]
/linux-4.1.27/drivers/gpu/drm/
Ddrm_lock.c43 static int drm_lock_take(struct drm_lock_data *lock_data, unsigned int context);
69 if (lock->context == DRM_KERNEL_CONTEXT) { in drm_legacy_lock()
71 task_pid_nr(current), lock->context); in drm_legacy_lock()
76 lock->context, task_pid_nr(current), in drm_legacy_lock()
92 if (drm_lock_take(&master->lock, lock->context)) { in drm_legacy_lock()
113 DRM_DEBUG("%d %s\n", lock->context, in drm_legacy_lock()
126 dev->sigdata.context = lock->context; in drm_legacy_lock()
135 lock->context); in drm_legacy_lock()
162 if (lock->context == DRM_KERNEL_CONTEXT) { in drm_legacy_unlock()
164 task_pid_nr(current), lock->context); in drm_legacy_unlock()
[all …]
/linux-4.1.27/arch/x86/kernel/
Dldt.c32 pc = &current->active_mm->context; in flush_ldt()
83 smp_store_release(&current_mm->context.ldt, ldt); in install_ldt()
112 mutex_init(&mm->context.lock); in init_new_context()
115 mm->context.ldt = NULL; in init_new_context()
119 mutex_lock(&old_mm->context.lock); in init_new_context()
120 if (!old_mm->context.ldt) { in init_new_context()
121 mm->context.ldt = NULL; in init_new_context()
125 new_ldt = alloc_ldt_struct(old_mm->context.ldt->size); in init_new_context()
131 memcpy(new_ldt->entries, old_mm->context.ldt->entries, in init_new_context()
135 mm->context.ldt = new_ldt; in init_new_context()
[all …]
Dstep.c33 mutex_lock(&child->mm->context.lock); in convert_ip_to_linear()
34 if (unlikely(!child->mm->context.ldt || in convert_ip_to_linear()
35 seg >= child->mm->context.ldt->size)) in convert_ip_to_linear()
38 desc = &child->mm->context.ldt->entries[seg]; in convert_ip_to_linear()
46 mutex_unlock(&child->mm->context.lock); in convert_ip_to_linear()
/linux-4.1.27/arch/blackfin/include/asm/
Dmmu_context.h64 mm->context.l1_stack_save = current_l1_stack_save = (void*)sp_base; in activate_l1stack()
82 if (prev_mm->context.page_rwx_mask == current_rwx_mask[cpu]) { in __switch_mm()
84 set_mask_dcplbs(next_mm->context.page_rwx_mask, cpu); in __switch_mm()
90 if (!next_mm->context.l1_stack_save) in __switch_mm()
92 if (next_mm->context.l1_stack_save == current_l1_stack_save) in __switch_mm()
97 current_l1_stack_save = next_mm->context.l1_stack_save; in __switch_mm()
123 unsigned long *mask = mm->context.page_rwx_mask; in protect_page()
154 if (mm->context.page_rwx_mask == current_rwx_mask[cpu]) { in update_protections()
156 set_mask_dcplbs(mm->context.page_rwx_mask, cpu); in update_protections()
177 mm->context.page_rwx_mask = (unsigned long *)p; in init_new_context()
[all …]
/linux-4.1.27/arch/sparc/include/asm/
Dmmu_context_64.h38 &mm->context.tsb_block[0], in tsb_context_switch()
40 (mm->context.tsb_block[1].tsb ? in tsb_context_switch()
41 &mm->context.tsb_block[1] : in tsb_context_switch()
46 , __pa(&mm->context.tsb_descr[0])); in tsb_context_switch()
68 : "r" (CTX_HWBITS((__mm)->context)), \
82 spin_lock_irqsave(&mm->context.lock, flags); in switch_mm()
83 ctx_valid = CTX_VALID(mm->context); in switch_mm()
127 __flush_tlb_mm(CTX_HWBITS(mm->context), in switch_mm()
130 spin_unlock_irqrestore(&mm->context.lock, flags); in switch_mm()
141 spin_lock_irqsave(&mm->context.lock, flags); in activate_mm()
[all …]
/linux-4.1.27/drivers/media/usb/as102/
Das10x_cmd_cfg.c43 sizeof(pcmd->body.context.req)); in as10x_cmd_get_context()
46 pcmd->body.context.req.proc_id = cpu_to_le16(CONTROL_PROC_CONTEXT); in as10x_cmd_get_context()
47 pcmd->body.context.req.tag = cpu_to_le16(tag); in as10x_cmd_get_context()
48 pcmd->body.context.req.type = cpu_to_le16(GET_CONTEXT_DATA); in as10x_cmd_get_context()
54 sizeof(pcmd->body.context.req) in as10x_cmd_get_context()
57 sizeof(prsp->body.context.rsp) in as10x_cmd_get_context()
72 *pvalue = le32_to_cpu((__force __le32)prsp->body.context.rsp.reg_val.u.value32); in as10x_cmd_get_context()
99 sizeof(pcmd->body.context.req)); in as10x_cmd_set_context()
102 pcmd->body.context.req.proc_id = cpu_to_le16(CONTROL_PROC_CONTEXT); in as10x_cmd_set_context()
104 pcmd->body.context.req.reg_val.u.value32 = (__force u32)cpu_to_le32(value); in as10x_cmd_set_context()
[all …]
/linux-4.1.27/arch/powerpc/mm/
Dmmu_context_hash64.c77 mm->context.id = index; in init_new_context()
79 mm->context.cop_lockp = kmalloc(sizeof(spinlock_t), GFP_KERNEL); in init_new_context()
80 if (!mm->context.cop_lockp) { in init_new_context()
83 mm->context.id = MMU_NO_CONTEXT; in init_new_context()
86 spin_lock_init(mm->context.cop_lockp); in init_new_context()
90 mm->context.pte_frag = NULL; in init_new_context()
110 pte_frag = mm->context.pte_frag; in destroy_pagetable_page()
137 drop_cop(mm->context.acop, mm); in destroy_context()
138 kfree(mm->context.cop_lockp); in destroy_context()
139 mm->context.cop_lockp = NULL; in destroy_context()
[all …]
Dmmu_context_nohash.c100 if (mm->context.active) { in steal_context_smp()
109 mm->context.id = MMU_NO_CONTEXT; in steal_context_smp()
152 mm->context.id = MMU_NO_CONTEXT; in steal_all_contexts()
157 mm->context.active = 0; in steal_all_contexts()
190 mm->context.id = MMU_NO_CONTEXT; in steal_context_up()
212 nact += context_mm[id]->context.active; in context_check_map()
238 cpu, next, next->context.active, next->context.id); in switch_mmu_context()
242 next->context.active++; in switch_mmu_context()
244 pr_hardcont(" (old=0x%p a=%d)", prev, prev->context.active); in switch_mmu_context()
245 WARN_ON(prev->context.active < 1); in switch_mmu_context()
[all …]
Dicswx_pid.c62 if (mm->context.cop_pid == COP_PID_NONE) { in get_cop_pid()
66 mm->context.cop_pid = pid; in get_cop_pid()
68 return mm->context.cop_pid; in get_cop_pid()
75 if ((!mm->context.acop) && (mm->context.cop_pid != COP_PID_NONE)) { in disable_cop_pid()
76 free_pid = mm->context.cop_pid; in disable_cop_pid()
77 mm->context.cop_pid = COP_PID_NONE; in disable_cop_pid()
Dicswx.c71 mtspr(SPRN_PID, next->context.cop_pid); in switch_cop()
73 mtspr(SPRN_ACOP, next->context.acop); in switch_cop()
97 spin_lock(mm->context.cop_lockp); in use_cop()
104 mm->context.acop |= acop; in use_cop()
117 spin_unlock(mm->context.cop_lockp); in use_cop()
141 spin_lock(mm->context.cop_lockp); in drop_cop()
143 mm->context.acop &= ~acop; in drop_cop()
159 spin_unlock(mm->context.cop_lockp); in drop_cop()
243 if ((acop_copro_type_bit(ct) & current->active_mm->context.acop) != 0) { in acop_handle_fault()
Dslice.c158 lpsizes = mm->context.low_slices_psize; in slice_mask_for_size()
163 hpsizes = mm->context.high_slices_psize; in slice_mask_for_size()
189 get_paca()->context = current->active_mm->context; in slice_flush_segments()
212 lpsizes = mm->context.low_slices_psize; in slice_convert()
219 mm->context.low_slices_psize = lpsizes; in slice_convert()
221 hpsizes = mm->context.high_slices_psize; in slice_convert()
232 mm->context.low_slices_psize, in slice_convert()
233 mm->context.high_slices_psize); in slice_convert()
554 current->mm->context.user_psize, 0); in arch_get_unmapped_area()
564 current->mm->context.user_psize, 1); in arch_get_unmapped_area_topdown()
[all …]
Dmmu_context_hash32.c84 mm->context.id = __init_new_context(); in init_new_context()
104 if (mm->context.id != NO_CONTEXT) { in destroy_context()
105 __destroy_context(mm->context.id); in destroy_context()
106 mm->context.id = NO_CONTEXT; in destroy_context()
/linux-4.1.27/drivers/base/regmap/
Dregmap-i2c.c19 static int regmap_smbus_byte_reg_read(void *context, unsigned int reg, in regmap_smbus_byte_reg_read() argument
22 struct device *dev = context; in regmap_smbus_byte_reg_read()
38 static int regmap_smbus_byte_reg_write(void *context, unsigned int reg, in regmap_smbus_byte_reg_write() argument
41 struct device *dev = context; in regmap_smbus_byte_reg_write()
55 static int regmap_smbus_word_reg_read(void *context, unsigned int reg, in regmap_smbus_word_reg_read() argument
58 struct device *dev = context; in regmap_smbus_word_reg_read()
74 static int regmap_smbus_word_reg_write(void *context, unsigned int reg, in regmap_smbus_word_reg_write() argument
77 struct device *dev = context; in regmap_smbus_word_reg_write()
91 static int regmap_smbus_word_read_swapped(void *context, unsigned int reg, in regmap_smbus_word_read_swapped() argument
94 struct device *dev = context; in regmap_smbus_word_read_swapped()
[all …]
Dregmap-spmi.c25 static int regmap_spmi_base_read(void *context, in regmap_spmi_base_read() argument
35 err = spmi_register_read(context, addr++, val++); in regmap_spmi_base_read()
40 static int regmap_spmi_base_gather_write(void *context, in regmap_spmi_base_gather_write() argument
55 err = spmi_register_zero_write(context, *data); in regmap_spmi_base_gather_write()
65 err = spmi_register_write(context, addr, *data); in regmap_spmi_base_gather_write()
78 static int regmap_spmi_base_write(void *context, const void *data, in regmap_spmi_base_write() argument
82 return regmap_spmi_base_gather_write(context, data, 1, data + 1, in regmap_spmi_base_write()
125 static int regmap_spmi_ext_read(void *context, in regmap_spmi_ext_read() argument
144 err = spmi_ext_register_read(context, addr, val, len); in regmap_spmi_ext_read()
156 err = spmi_ext_register_readl(context, addr, val, len); in regmap_spmi_ext_read()
[all …]
Dregmap-spi.c32 static int regmap_spi_write(void *context, const void *data, size_t count) in regmap_spi_write() argument
34 struct device *dev = context; in regmap_spi_write()
40 static int regmap_spi_gather_write(void *context, in regmap_spi_gather_write() argument
44 struct device *dev = context; in regmap_spi_gather_write()
57 static int regmap_spi_async_write(void *context, in regmap_spi_async_write() argument
65 struct device *dev = context; in regmap_spi_async_write()
79 async->m.context = async; in regmap_spi_async_write()
95 static int regmap_spi_read(void *context, in regmap_spi_read() argument
99 struct device *dev = context; in regmap_spi_read()
Dregmap-mmio.c88 static int regmap_mmio_gather_write(void *context, in regmap_mmio_gather_write() argument
92 struct regmap_mmio_context *ctx = context; in regmap_mmio_gather_write()
137 static int regmap_mmio_write(void *context, const void *data, size_t count) in regmap_mmio_write() argument
139 struct regmap_mmio_context *ctx = context; in regmap_mmio_write()
144 return regmap_mmio_gather_write(context, data, ctx->reg_bytes, in regmap_mmio_write()
148 static int regmap_mmio_read(void *context, in regmap_mmio_read() argument
152 struct regmap_mmio_context *ctx = context; in regmap_mmio_read()
197 static void regmap_mmio_free_context(void *context) in regmap_mmio_free_context() argument
199 struct regmap_mmio_context *ctx = context; in regmap_mmio_free_context()
205 kfree(context); in regmap_mmio_free_context()
Dregmap-ac97.c56 static int regmap_ac97_reg_read(void *context, unsigned int reg, in regmap_ac97_reg_read() argument
59 struct snd_ac97 *ac97 = context; in regmap_ac97_reg_read()
66 static int regmap_ac97_reg_write(void *context, unsigned int reg, in regmap_ac97_reg_write() argument
69 struct snd_ac97 *ac97 = context; in regmap_ac97_reg_write()
/linux-4.1.27/tools/perf/scripts/python/
Dnetdev-times.py227 def irq__softirq_entry(name, context, cpu, sec, nsec, pid, comm, callchain, vec): argument
230 event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec)
233 def irq__softirq_exit(name, context, cpu, sec, nsec, pid, comm, callchain, vec): argument
236 event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec)
239 def irq__softirq_raise(name, context, cpu, sec, nsec, pid, comm, callchain, vec): argument
242 event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec)
245 def irq__irq_handler_entry(name, context, cpu, sec, nsec, pid, comm, argument
247 event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm,
251 def irq__irq_handler_exit(name, context, cpu, sec, nsec, pid, comm, callchain, irq, ret): argument
252 event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, irq, ret)
[all …]
Dcheck-perf-trace.py28 def irq__softirq_entry(event_name, context, common_cpu, argument
34 print_uncommon(context)
39 def kmem__kmalloc(event_name, context, common_cpu, argument
46 print_uncommon(context)
54 def trace_unhandled(event_name, context, event_fields_dict): argument
65 def print_uncommon(context): argument
67 % (common_pc(context), trace_flag_str(common_flags(context)), \
68 common_lock_depth(context))
Dsched-migration.py370 def sched__sched_stat_runtime(event_name, context, common_cpu, argument
375 def sched__sched_stat_iowait(event_name, context, common_cpu, argument
380 def sched__sched_stat_sleep(event_name, context, common_cpu, argument
385 def sched__sched_stat_wait(event_name, context, common_cpu, argument
390 def sched__sched_process_fork(event_name, context, common_cpu, argument
395 def sched__sched_process_wait(event_name, context, common_cpu, argument
400 def sched__sched_process_exit(event_name, context, common_cpu, argument
405 def sched__sched_process_free(event_name, context, common_cpu, argument
410 def sched__sched_migrate_task(event_name, context, common_cpu, argument
418 def sched__sched_switch(event_name, context, common_cpu, argument
[all …]
/linux-4.1.27/crypto/asymmetric_keys/
Dpkcs7_parser.c163 int pkcs7_note_OID(void *context, size_t hdrlen, in pkcs7_note_OID() argument
167 struct pkcs7_parse_context *ctx = context; in pkcs7_note_OID()
182 int pkcs7_sig_note_digest_algo(void *context, size_t hdrlen, in pkcs7_sig_note_digest_algo() argument
186 struct pkcs7_parse_context *ctx = context; in pkcs7_sig_note_digest_algo()
211 int pkcs7_sig_note_pkey_algo(void *context, size_t hdrlen, in pkcs7_sig_note_pkey_algo() argument
215 struct pkcs7_parse_context *ctx = context; in pkcs7_sig_note_pkey_algo()
231 int pkcs7_extract_cert(void *context, size_t hdrlen, in pkcs7_extract_cert() argument
235 struct pkcs7_parse_context *ctx = context; in pkcs7_extract_cert()
271 int pkcs7_note_certificate_list(void *context, size_t hdrlen, in pkcs7_note_certificate_list() argument
275 struct pkcs7_parse_context *ctx = context; in pkcs7_note_certificate_list()
[all …]
Dx509_cert_parser.c139 int x509_note_OID(void *context, size_t hdrlen, in x509_note_OID() argument
143 struct x509_parse_context *ctx = context; in x509_note_OID()
159 int x509_note_tbs_certificate(void *context, size_t hdrlen, in x509_note_tbs_certificate() argument
163 struct x509_parse_context *ctx = context; in x509_note_tbs_certificate()
176 int x509_note_pkey_algo(void *context, size_t hdrlen, in x509_note_pkey_algo() argument
180 struct x509_parse_context *ctx = context; in x509_note_pkey_algo()
228 int x509_note_signature(void *context, size_t hdrlen, in x509_note_signature() argument
232 struct x509_parse_context *ctx = context; in x509_note_signature()
250 int x509_note_serial(void *context, size_t hdrlen, in x509_note_serial() argument
254 struct x509_parse_context *ctx = context; in x509_note_serial()
[all …]
Dmscode_parser.c46 int mscode_note_content_type(void *context, size_t hdrlen, in mscode_note_content_type() argument
78 int mscode_note_digest_algo(void *context, size_t hdrlen, in mscode_note_digest_algo() argument
82 struct pefile_context *ctx = context; in mscode_note_digest_algo()
117 int mscode_note_digest(void *context, size_t hdrlen, in mscode_note_digest() argument
121 struct pefile_context *ctx = context; in mscode_note_digest()
/linux-4.1.27/tools/perf/scripts/python/Perf-Trace-Util/
DContext.c31 PyObject *context; in perf_trace_context_common_pc() local
34 if (!PyArg_ParseTuple(args, "O", &context)) in perf_trace_context_common_pc()
37 scripting_context = PyCObject_AsVoidPtr(context); in perf_trace_context_common_pc()
47 PyObject *context; in perf_trace_context_common_flags() local
50 if (!PyArg_ParseTuple(args, "O", &context)) in perf_trace_context_common_flags()
53 scripting_context = PyCObject_AsVoidPtr(context); in perf_trace_context_common_flags()
63 PyObject *context; in perf_trace_context_common_lock_depth() local
66 if (!PyArg_ParseTuple(args, "O", &context)) in perf_trace_context_common_lock_depth()
69 scripting_context = PyCObject_AsVoidPtr(context); in perf_trace_context_common_lock_depth()
/linux-4.1.27/drivers/staging/vt6656/
Dusbpipe.c107 struct vnt_private *priv = urb->context; in vnt_start_interrupt_urb_complete()
169 struct vnt_rcb *rcb = urb->context; in vnt_submit_rx_urb_complete()
249 struct vnt_usb_send_context *context = urb->context; in vnt_tx_context_complete() local
250 struct vnt_private *priv = context->priv; in vnt_tx_context_complete()
254 dev_dbg(&priv->usb->dev, "Write %d bytes\n", context->buf_len); in vnt_tx_context_complete()
259 context->in_use = false; in vnt_tx_context_complete()
267 if (context->type == CONTEXT_DATA_PACKET) in vnt_tx_context_complete()
270 if (urb->status || context->type == CONTEXT_BEACON_PACKET) { in vnt_tx_context_complete()
271 if (context->skb) in vnt_tx_context_complete()
272 ieee80211_free_txskb(priv->hw, context->skb); in vnt_tx_context_complete()
[all …]
Dint.c72 struct vnt_usb_send_context *context; in vnt_int_report_rate() local
81 context = priv->tx_context[pkt_no]; in vnt_int_report_rate()
83 if (!context->skb) in vnt_int_report_rate()
86 info = IEEE80211_SKB_CB(context->skb); in vnt_int_report_rate()
89 if (context->fb_option && !(tsr & (TSR_TMO | TSR_RETRYTMO))) { in vnt_int_report_rate()
99 if (context->fb_option == AUTO_FB_0) in vnt_int_report_rate()
101 else if (context->fb_option == AUTO_FB_1) in vnt_int_report_rate()
119 ieee80211_tx_status_irqsafe(priv->hw, context->skb); in vnt_int_report_rate()
121 context->in_use = false; in vnt_int_report_rate()
/linux-4.1.27/drivers/acpi/
Dbus.c136 void *context) in acpi_bus_private_data_handler() argument
181 struct acpi_osc_context *context, char *error) in acpi_print_osc_error() argument
193 for (i = 0; i < context->cap.length; i += sizeof(u32)) in acpi_print_osc_error()
194 printk("%x ", *((u32 *)(context->cap.pointer + i))); in acpi_print_osc_error()
221 acpi_status acpi_run_osc(acpi_handle handle, struct acpi_osc_context *context) in acpi_run_osc() argument
231 if (!context) in acpi_run_osc()
233 if (ACPI_FAILURE(acpi_str_to_uuid(context->uuid_str, uuid))) in acpi_run_osc()
235 context->ret.length = ACPI_ALLOCATE_BUFFER; in acpi_run_osc()
236 context->ret.pointer = NULL; in acpi_run_osc()
245 in_params[1].integer.value = context->rev; in acpi_run_osc()
[all …]
Dsbshc.c32 void *context; member
201 smbus_alarm_callback callback, void *context) in acpi_smbus_register_callback() argument
205 hc->context = context; in acpi_smbus_register_callback()
216 hc->context = NULL; in acpi_smbus_unregister_callback()
223 static inline void acpi_smbus_callback(void *context) in acpi_smbus_callback() argument
225 struct acpi_smb_hc *hc = context; in acpi_smbus_callback()
227 hc->callback(hc->context); in acpi_smbus_callback()
230 static int smbus_alarm(void *context) in smbus_alarm() argument
232 struct acpi_smb_hc *hc = context; in smbus_alarm()
/linux-4.1.27/arch/nios2/mm/
Dmmu_context.c53 static void set_context(mm_context_t context) in set_context() argument
55 set_mmu_pid(CTX_PID(context)); in set_context()
89 if (unlikely(CTX_VERSION(next->context) != in switch_mm()
91 next->context = get_new_context(); in switch_mm()
97 set_context(next->context); in switch_mm()
108 next->context = get_new_context(); in activate_mm()
109 set_context(next->context); in activate_mm()
113 unsigned long get_pid_from_context(mm_context_t *context) in get_pid_from_context() argument
115 return CTX_PID((*context)); in get_pid_from_context()
/linux-4.1.27/drivers/infiniband/hw/mthca/
Dmthca_provider.c301 struct mthca_ucontext *context; in mthca_alloc_ucontext() local
315 context = kmalloc(sizeof *context, GFP_KERNEL); in mthca_alloc_ucontext()
316 if (!context) in mthca_alloc_ucontext()
319 err = mthca_uar_alloc(to_mdev(ibdev), &context->uar); in mthca_alloc_ucontext()
321 kfree(context); in mthca_alloc_ucontext()
325 context->db_tab = mthca_init_user_db_tab(to_mdev(ibdev)); in mthca_alloc_ucontext()
326 if (IS_ERR(context->db_tab)) { in mthca_alloc_ucontext()
327 err = PTR_ERR(context->db_tab); in mthca_alloc_ucontext()
328 mthca_uar_free(to_mdev(ibdev), &context->uar); in mthca_alloc_ucontext()
329 kfree(context); in mthca_alloc_ucontext()
[all …]
Dmthca_srq.c98 struct mthca_tavor_srq_context *context) in mthca_tavor_init_srq_context() argument
100 memset(context, 0, sizeof *context); in mthca_tavor_init_srq_context()
102 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); in mthca_tavor_init_srq_context()
103 context->state_pd = cpu_to_be32(pd->pd_num); in mthca_tavor_init_srq_context()
104 context->lkey = cpu_to_be32(srq->mr.ibmr.lkey); in mthca_tavor_init_srq_context()
107 context->uar = in mthca_tavor_init_srq_context()
108 cpu_to_be32(to_mucontext(pd->ibpd.uobject->context)->uar.index); in mthca_tavor_init_srq_context()
110 context->uar = cpu_to_be32(dev->driver_uar.index); in mthca_tavor_init_srq_context()
116 struct mthca_arbel_srq_context *context) in mthca_arbel_init_srq_context() argument
120 memset(context, 0, sizeof *context); in mthca_arbel_init_srq_context()
[all …]
/linux-4.1.27/tools/perf/scripts/perl/
Dcheck-perf-trace.pl30 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
37 print_uncommon($context);
45 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
53 print_uncommon($context);
65 my ($context) = @_;
68 common_pc($context), trace_flag_str(common_flags($context)),
69 common_lock_depth($context));
94 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
Drwtop.pl37 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
55 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
68 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
81 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
199 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
Drw-by-pid.pl26 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
42 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
53 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
64 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
180 my ($event_name, $context, $common_cpu, $common_secs, $common_nsecs,
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Ddoorbell.c46 int mlx5_ib_db_map_user(struct mlx5_ib_ucontext *context, unsigned long virt, in mlx5_ib_db_map_user() argument
52 mutex_lock(&context->db_page_mutex); in mlx5_ib_db_map_user()
54 list_for_each_entry(page, &context->db_page_list, list) in mlx5_ib_db_map_user()
66 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx5_ib_db_map_user()
74 list_add(&page->list, &context->db_page_list); in mlx5_ib_db_map_user()
82 mutex_unlock(&context->db_page_mutex); in mlx5_ib_db_map_user()
87 void mlx5_ib_db_unmap_user(struct mlx5_ib_ucontext *context, struct mlx5_db *db) in mlx5_ib_db_unmap_user() argument
89 mutex_lock(&context->db_page_mutex); in mlx5_ib_db_unmap_user()
97 mutex_unlock(&context->db_page_mutex); in mlx5_ib_db_unmap_user()
Dqp.c606 struct mlx5_ib_ucontext *context; in create_user_qp() local
622 context = to_mucontext(pd->uobject->context); in create_user_qp()
626 uuarn = alloc_uuar(&context->uuari, MLX5_IB_LATENCY_CLASS_HIGH); in create_user_qp()
630 uuarn = alloc_uuar(&context->uuari, MLX5_IB_LATENCY_CLASS_MEDIUM); in create_user_qp()
634 uuarn = alloc_uuar(&context->uuari, MLX5_IB_LATENCY_CLASS_LOW); in create_user_qp()
642 uar_index = uuarn_to_uar_index(&context->uuari, uuarn); in create_user_qp()
654 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, in create_user_qp()
693 err = mlx5_ib_db_map_user(context, ucmd.db_addr, &qp->db); in create_user_qp()
709 mlx5_ib_db_unmap_user(context, &qp->db); in create_user_qp()
719 free_uuar(&context->uuari, uuarn); in create_user_qp()
[all …]
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Ddoorbell.c44 int mlx4_ib_db_map_user(struct mlx4_ib_ucontext *context, unsigned long virt, in mlx4_ib_db_map_user() argument
50 mutex_lock(&context->db_page_mutex); in mlx4_ib_db_map_user()
52 list_for_each_entry(page, &context->db_page_list, list) in mlx4_ib_db_map_user()
64 page->umem = ib_umem_get(&context->ibucontext, virt & PAGE_MASK, in mlx4_ib_db_map_user()
72 list_add(&page->list, &context->db_page_list); in mlx4_ib_db_map_user()
80 mutex_unlock(&context->db_page_mutex); in mlx4_ib_db_map_user()
85 void mlx4_ib_db_unmap_user(struct mlx4_ib_ucontext *context, struct mlx4_db *db) in mlx4_ib_db_unmap_user() argument
87 mutex_lock(&context->db_page_mutex); in mlx4_ib_db_unmap_user()
95 mutex_unlock(&context->db_page_mutex); in mlx4_ib_db_unmap_user()
Dqp.c724 qp->umem = ib_umem_get(pd->uobject->context, ucmd.buf_addr, in create_qp_common()
741 err = mlx4_ib_db_map_user(to_mucontext(pd->uobject->context), in create_qp_common()
875 mlx4_ib_db_unmap_user(to_mucontext(pd->uobject->context), &qp->db); in create_qp_common()
1049 mlx4_ib_db_unmap_user(to_mucontext(qp->ibqp.uobject->context), in destroy_qp_common()
1424 struct mlx4_qp_context *context) in handle_eth_ud_smac_index() argument
1431 context->pri_path.sched_queue = MLX4_IB_DEFAULT_SCHED_QUEUE | ((qp->port - 1) << 6); in handle_eth_ud_smac_index()
1438 context->pri_path.grh_mylmc = 0x80 | (u8) smac_index; in handle_eth_ud_smac_index()
1454 struct mlx4_qp_context *context; in __mlx4_ib_modify_qp() local
1466 context = kzalloc(sizeof *context, GFP_KERNEL); in __mlx4_ib_modify_qp()
1467 if (!context) in __mlx4_ib_modify_qp()
[all …]
/linux-4.1.27/arch/microblaze/include/asm/
Dmmu_context_mm.h54 extern void set_context(mm_context_t context, pgd_t *pgd);
85 if (mm->context != NO_CONTEXT) in get_mmu_context()
96 mm->context = ctx; in get_mmu_context()
103 # define init_new_context(tsk, mm) (((mm)->context = NO_CONTEXT), 0)
110 if (mm->context != NO_CONTEXT) { in destroy_context()
111 clear_bit(mm->context, context_map); in destroy_context()
112 mm->context = NO_CONTEXT; in destroy_context()
122 set_context(next->context, next->pgd); in switch_mm()
134 set_context(mm->context, mm->pgd); in activate_mm()
/linux-4.1.27/arch/m68k/sun3/
Dmmu_emu.c208 void clear_context(unsigned long context) in clear_context() argument
213 if(context) { in clear_context()
214 if(!ctx_alloc[context]) in clear_context()
217 ctx_alloc[context]->context = SUN3_INVALID_CONTEXT; in clear_context()
218 ctx_alloc[context] = (struct mm_struct *)0; in clear_context()
224 sun3_put_context(context); in clear_context()
227 if((pmeg_ctx[i] == context) && (pmeg_alloc[i] == 1)) { in clear_context()
282 inline void mmu_emu_map_pmeg (int context, int vaddr) in mmu_emu_map_pmeg() argument
297 curr_pmeg, context, vaddr); in mmu_emu_map_pmeg()
304 sun3_put_context(context); in mmu_emu_map_pmeg()
[all …]
/linux-4.1.27/tools/perf/scripts/perl/Perf-Trace-Util/
DContext.xs32 common_pc(context)
33 struct scripting_context * context
36 common_flags(context)
37 struct scripting_context * context
40 common_lock_depth(context)
41 struct scripting_context * context
DContext.c58 struct scripting_context * context = INT2PTR(struct scripting_context *,SvIV(ST(0))); in XS() local
62 RETVAL = common_pc(context); in XS()
81 struct scripting_context * context = INT2PTR(struct scripting_context *,SvIV(ST(0))); in XS() local
85 RETVAL = common_flags(context); in XS()
104 struct scripting_context * context = INT2PTR(struct scripting_context *,SvIV(ST(0))); in XS() local
108 RETVAL = common_lock_depth(context); in XS()
/linux-4.1.27/drivers/gpu/drm/tegra/
Ddrm.c246 static void tegra_drm_context_free(struct tegra_drm_context *context) in tegra_drm_context_free() argument
248 context->client->ops->close_channel(context); in tegra_drm_context_free()
249 kfree(context); in tegra_drm_context_free()
318 int tegra_drm_submit(struct tegra_drm_context *context, in tegra_drm_submit() argument
339 job = host1x_job_alloc(context->channel, args->num_cmdbufs, in tegra_drm_submit()
346 job->client = (u32)args->context; in tegra_drm_submit()
347 job->class = context->client->base.class; in tegra_drm_submit()
391 job->is_addr_reg = context->client->ops->is_addr_reg; in tegra_drm_submit()
399 err = host1x_job_pin(job, context->client->base.dev); in tegra_drm_submit()
421 static struct tegra_drm_context *tegra_drm_get_context(__u64 context) in tegra_drm_get_context() argument
[all …]
/linux-4.1.27/drivers/isdn/hisax/
Dhfc_usb.c260 hfcusb_data *hfc = (hfcusb_data *) urb->context; in ctrl_complete()
429 usb_complete_t complete, void *context) in fill_isoc_urb() argument
438 urb->context = context; in fill_isoc_urb()
544 iso_urb_struct *context_iso_urb = (iso_urb_struct *) urb->context; in tx_iso_complete()
594 tx_iso_complete, urb->context); in tx_iso_complete()
682 iso_urb_struct *context_iso_urb = (iso_urb_struct *) urb->context; in rx_iso_complete()
762 rx_iso_complete, urb->context); in rx_iso_complete()
855 usb_fifo *fifo = (usb_fifo *) urb->context; in rx_int_complete()
1260 hfcusb_data *context; in hfc_usb_probe() local
1359 if (!(context = kzalloc(sizeof(hfcusb_data), GFP_KERNEL))) in hfc_usb_probe()
[all …]
Dst5481_usb.c62 ctrl_complete_t complete, void *context) in usb_ctrl_msg() argument
80 ctrl_msg->context = context; in usb_ctrl_msg()
90 ctrl_complete_t complete, void *context) in st5481_usb_device_ctrl_msg() argument
94 value, 0, complete, context); in st5481_usb_device_ctrl_msg()
102 ctrl_complete_t complete, void *context) in st5481_usb_pipe_reset() argument
108 0, pipe, complete, context); in st5481_usb_pipe_reset()
130 struct st5481_adapter *adapter = urb->context; in usb_ctrl_complete()
156 ctrl_msg->complete(ctrl_msg->context); in usb_ctrl_complete()
179 struct st5481_adapter *adapter = urb->context; in usb_int_complete()
407 void *context) in fill_isoc_urb() argument
[all …]
/linux-4.1.27/drivers/net/phy/
Dat803x.c66 struct at803x_context *context) in at803x_context_save() argument
68 context->bmcr = phy_read(phydev, MII_BMCR); in at803x_context_save()
69 context->advertise = phy_read(phydev, MII_ADVERTISE); in at803x_context_save()
70 context->control1000 = phy_read(phydev, MII_CTRL1000); in at803x_context_save()
71 context->int_enable = phy_read(phydev, AT803X_INTR_ENABLE); in at803x_context_save()
72 context->smart_speed = phy_read(phydev, AT803X_SMART_SPEED); in at803x_context_save()
73 context->led_control = phy_read(phydev, AT803X_LED_CONTROL); in at803x_context_save()
78 const struct at803x_context *context) in at803x_context_restore() argument
80 phy_write(phydev, MII_BMCR, context->bmcr); in at803x_context_restore()
81 phy_write(phydev, MII_ADVERTISE, context->advertise); in at803x_context_restore()
[all …]
/linux-4.1.27/arch/xtensa/include/asm/
Dmmu_context.h80 mm->context.asid[cpu] = asid; in get_new_mmu_context()
81 mm->context.cpu = cpu; in get_new_mmu_context()
91 unsigned long asid = mm->context.asid[cpu]; in get_mmu_context()
102 set_rasid_register(ASID_INSERT(mm->context.asid[cpu])); in activate_context()
117 mm->context.asid[cpu] = NO_CONTEXT; in init_new_context()
119 mm->context.cpu = -1; in init_new_context()
127 int migrated = next->context.cpu != cpu; in switch_mm()
131 next->context.cpu = cpu; in switch_mm()
/linux-4.1.27/arch/arm64/mm/
Dcontext.c43 mm->context.id = 0; in __init_new_context()
44 raw_spin_lock_init(&mm->context.id_lock); in __init_new_context()
68 raw_spin_lock_irqsave(&mm->context.id_lock, flags); in set_mm_context()
69 if (likely((mm->context.id ^ cpu_last_asid) >> MAX_ASID_BITS)) { in set_mm_context()
74 mm->context.id = asid; in set_mm_context()
77 raw_spin_unlock_irqrestore(&mm->context.id_lock, flags); in set_mm_context()
117 mm->context.id = asid; in set_mm_context()
134 if (!unlikely((mm->context.id ^ cpu_last_asid) >> MAX_ASID_BITS)) { in __new_context()
/linux-4.1.27/arch/hexagon/mm/
Dvm_tlb.c42 if (mm->context.ptbase == current->active_mm->context.ptbase) in flush_tlb_range()
71 if (current->active_mm->context.ptbase == mm->context.ptbase) in flush_tlb_mm()
82 if (mm->context.ptbase == current->active_mm->context.ptbase) in flush_tlb_page()
/linux-4.1.27/drivers/gpio/
Dgpio-omap.c56 struct gpio_regs context; member
110 bank->context.oe = l; in omap_set_gpio_direction()
123 bank->context.dataout |= l; in omap_set_gpio_dataout_reg()
126 bank->context.dataout &= ~l; in omap_set_gpio_dataout_reg()
146 bank->context.dataout = l; in omap_set_gpio_dataout_mask()
254 bank->context.debounce = debounce; in omap2_set_gpio_debounce()
255 bank->context.debounce_en = val; in omap2_set_gpio_debounce()
280 bank->context.debounce_en &= ~gpio_bit; in omap_clear_gpio_debounce()
281 writel_relaxed(bank->context.debounce_en, in omap_clear_gpio_debounce()
285 bank->context.debounce = 0; in omap_clear_gpio_debounce()
[all …]
/linux-4.1.27/arch/cris/mm/
Dtlb.c52 old_mm->context.page_id = NO_CONTEXT; in alloc_context()
57 mm->context.page_id = map_replace_ptr; in alloc_context()
73 if(mm->context.page_id == NO_CONTEXT) in get_mmu_context()
88 if(mm->context.page_id != NO_CONTEXT) { in destroy_context()
89 D(printk("destroy_context %d (%p)\n", mm->context.page_id, mm)); in destroy_context()
91 page_id_map[mm->context.page_id] = NULL; in destroy_context()
/linux-4.1.27/include/trace/events/
Dfence.h21 __field(unsigned int, context)
33 __entry->context = fence->context;
38 __entry->waiting_context = f1->context;
45 __get_str(driver), __get_str(timeline), __entry->context,
60 __field(unsigned int, context)
67 __entry->context = fence->context;
72 __get_str(driver), __get_str(timeline), __entry->context,
/linux-4.1.27/arch/arm/include/asm/
Dmmu_context.h29 #define init_new_context(tsk,mm) ({ atomic64_set(&mm->context.id, 0); 0; })
48 if (unlikely(mm->context.vmalloc_seq != init_mm.context.vmalloc_seq)) in check_and_switch_context()
59 mm->context.switch_pending = 1; in check_and_switch_context()
70 if (mm && mm->context.switch_pending) { in finish_arch_post_lock_switch()
78 if (mm->context.switch_pending) { in finish_arch_post_lock_switch()
79 mm->context.switch_pending = 0; in finish_arch_post_lock_switch()
/linux-4.1.27/drivers/uwb/i1480/dfu/
Ddfu.c53 const char *cmd, u8 context, u8 expected_type, in i1480_rceb_check() argument
58 if (rceb->bEventContext != context) { in i1480_rceb_check()
62 rceb->bEventContext, context); in i1480_rceb_check()
100 u8 context; in i1480_cmd() local
105 get_random_bytes(&context, 1); in i1480_cmd()
106 } while (context == 0x00 || context == 0xff); in i1480_cmd()
107 cmd->bCommandContext = context; in i1480_cmd()
147 result = i1480_rceb_check(i1480, i1480->evt_buf, cmd_name, context, in i1480_cmd()
/linux-4.1.27/drivers/acpi/acpica/
Devxface.c61 acpi_gpe_handler address, void *context);
93 acpi_notify_handler handler, void *context) in acpi_install_notify_handler() argument
132 acpi_gbl_global_notify[i].context = context; in acpi_install_notify_handler()
202 handler_obj->notify.context = context; in acpi_install_notify_handler()
287 acpi_gbl_global_notify[i].context = NULL; in ACPI_EXPORT_SYMBOL()
423 acpi_status acpi_install_sci_handler(acpi_sci_handler address, void *context) in ACPI_EXPORT_SYMBOL()
444 new_sci_handler->context = context; in ACPI_EXPORT_SYMBOL()
568 acpi_install_global_event_handler(acpi_gbl_event_handler handler, void *context) in ACPI_EXPORT_SYMBOL()
593 acpi_gbl_global_event_handler_context = context; in ACPI_EXPORT_SYMBOL()
619 acpi_event_handler handler, void *context) in ACPI_EXPORT_SYMBOL()
[all …]
Devregion.c62 u32 level, void *context, void **return_value);
144 struct acpi_connection_info *context; in acpi_ev_address_space_dispatch() local
168 context = handler_desc->address_space.context; in acpi_ev_address_space_dispatch()
199 context, &region_context); in acpi_ev_address_space_dispatch()
252 context && field_obj) { in acpi_ev_address_space_dispatch()
256 context->connection = field_obj->field.resource_buffer; in acpi_ev_address_space_dispatch()
257 context->length = field_obj->field.resource_length; in acpi_ev_address_space_dispatch()
258 context->access_length = field_obj->field.access_length; in acpi_ev_address_space_dispatch()
261 context && field_obj) { in acpi_ev_address_space_dispatch()
265 context->connection = field_obj->field.resource_buffer; in acpi_ev_address_space_dispatch()
[all …]
Devsci.c53 static u32 ACPI_SYSTEM_XFACE acpi_ev_sci_xrupt_handler(void *context);
90 int_status |= sci_handler->address(sci_handler->context); in acpi_ev_sci_dispatch()
112 static u32 ACPI_SYSTEM_XFACE acpi_ev_sci_xrupt_handler(void *context) in acpi_ev_sci_xrupt_handler() argument
114 struct acpi_gpe_xrupt_info *gpe_xrupt_list = context; in acpi_ev_sci_xrupt_handler()
156 u32 ACPI_SYSTEM_XFACE acpi_ev_gpe_xrupt_handler(void *context) in acpi_ev_gpe_xrupt_handler() argument
158 struct acpi_gpe_xrupt_info *gpe_xrupt_list = context; in acpi_ev_gpe_xrupt_handler()
Dacevents.h117 void *context);
138 u32 level, void *context, void **return_value);
144 acpi_ev_walk_gpe_list(acpi_gpe_callback gpe_walk_callback, void *context);
148 struct acpi_gpe_block_info *gpe_block, void *context);
159 void *context);
174 acpi_adr_space_setup setup, void *context);
244 u32 ACPI_SYSTEM_XFACE acpi_ev_gpe_xrupt_handler(void *context);
Dnsinit.c56 u32 level, void *context, void **return_value);
60 u32 nesting_level, void *context, void **return_value);
64 u32 nesting_level, void *context, void **return_value);
242 u32 level, void *context, void **return_value) in acpi_ns_init_one_object() argument
247 (struct acpi_init_walk_info *)context; in acpi_ns_init_one_object()
384 u32 nesting_level, void *context, void **return_value) in acpi_ns_find_ini_methods() argument
387 ACPI_CAST_PTR(struct acpi_device_walk_info, context); in acpi_ns_find_ini_methods()
449 u32 nesting_level, void *context, void **return_value) in acpi_ns_init_one_device() argument
452 ACPI_CAST_PTR(struct acpi_device_walk_info, context); in acpi_ns_init_one_device()
Devmisc.c53 static void ACPI_SYSTEM_XFACE acpi_ev_notify_dispatch(void *context);
195 static void ACPI_SYSTEM_XFACE acpi_ev_notify_dispatch(void *context) in acpi_ev_notify_dispatch() argument
197 union acpi_generic_state *info = (union acpi_generic_state *)context; in acpi_ev_notify_dispatch()
207 info->notify.global->context); in acpi_ev_notify_dispatch()
216 handler_obj->notify.context); in acpi_ev_notify_dispatch()
/linux-4.1.27/arch/metag/include/asm/
Dmmu_context.h26 mm->context.pgd_base = (unsigned long) mm->pgd; in init_new_context()
29 INIT_LIST_HEAD(&mm->context.tcm); in init_new_context()
43 list_for_each_entry_safe(pos, n, &mm->context.tcm, list) { in destroy_context()
81 if (prev->context.pgd_base != (unsigned long) prev->pgd) { in switch_mmu()
83 ((pgd_t *) prev->context.pgd_base)[i] = prev->pgd[i]; in switch_mmu()
88 prev->pgd = (pgd_t *) prev->context.pgd_base; in switch_mmu()
91 next->pgd[i] = ((pgd_t *) next->context.pgd_base)[i]; in switch_mmu()
/linux-4.1.27/arch/x86/kvm/
Dmmu.c3518 struct kvm_mmu *context) in nonpaging_init_context() argument
3520 context->page_fault = nonpaging_page_fault; in nonpaging_init_context()
3521 context->gva_to_gpa = nonpaging_gva_to_gpa; in nonpaging_init_context()
3522 context->sync_page = nonpaging_sync_page; in nonpaging_init_context()
3523 context->invlpg = nonpaging_invlpg; in nonpaging_init_context()
3524 context->update_pte = nonpaging_update_pte; in nonpaging_init_context()
3525 context->root_level = 0; in nonpaging_init_context()
3526 context->shadow_root_level = PT32E_ROOT_LEVEL; in nonpaging_init_context()
3527 context->root_hpa = INVALID_PAGE; in nonpaging_init_context()
3528 context->direct_map = true; in nonpaging_init_context()
[all …]
/linux-4.1.27/arch/frv/mm/
Dmmu-context.c31 memset(&mm->context, 0, sizeof(mm->context)); in init_new_context()
32 INIT_LIST_HEAD(&mm->context.id_link); in init_new_context()
33 mm->context.itlb_cached_pge = 0xffffffffUL; in init_new_context()
34 mm->context.dtlb_cached_pge = 0xffffffffUL; in init_new_context()
132 mm_context_t *ctx = &mm->context; in destroy_context()
157 buffer += sprintf(buffer, "CXNR: %u\n", mm->context.id); in proc_pid_status_frv_cxnr()
203 cxn_pinned = get_cxn(&mm->context); in cxn_pin_by_pid()
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_mmap.c50 struct ipath_ibdev *dev = to_idev(ip->context->device); in ipath_release_mmap_info()
89 int ipath_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in ipath_mmap() argument
91 struct ipath_ibdev *dev = to_idev(context->device); in ipath_mmap()
106 if (context != ip->context || (__u64) offset != ip->offset) in ipath_mmap()
133 struct ib_ucontext *context, in ipath_create_mmap_info() argument
152 ip->context = context; in ipath_create_mmap_info()
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_mmap.c50 struct qib_ibdev *dev = to_idev(ip->context->device); in qib_release_mmap_info()
89 int qib_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in qib_mmap() argument
91 struct qib_ibdev *dev = to_idev(context->device); in qib_mmap()
106 if (context != ip->context || (__u64) offset != ip->offset) in qib_mmap()
133 struct ib_ucontext *context, in qib_create_mmap_info() argument
152 ip->context = context; in qib_create_mmap_info()
/linux-4.1.27/drivers/md/persistent-data/
Ddm-btree.h41 void *context; member
61 void (*inc)(void *context, const void *value);
68 void (*dec)(void *context, const void *value);
75 int (*equal)(void *context, const void *value1, const void *value2);
159 int (*fn)(void *context, uint64_t *keys, void *leaf),
160 void *context);
Ddm-array.c116 fn(info->value_type.context, element_at(info, ab, i)); in on_entries()
188 vt->inc(vt->context, value); in fill_ablock()
212 vt->dec(vt->context, element_at(info, ab, i - 1)); in trim_ablock()
562 static void block_inc(void *context, const void *value) in block_inc() argument
565 struct dm_array_info *info = context; in block_inc()
571 static void block_dec(void *context, const void *value) in block_dec() argument
579 struct dm_array_info *info = context; in block_dec()
610 static int block_equal(void *context, const void *value1, const void *value2) in block_equal() argument
627 bvt->context = info; in dm_array_info_init()
746 (!vt->equal || !vt->equal(vt->context, old_value, value))) { in array_set_value()
[all …]
/linux-4.1.27/arch/hexagon/include/asm/
Dmmu_context.h75 if (next->context.generation < prev->context.generation) { in switch_mm()
79 next->context.generation = prev->context.generation; in switch_mm()
82 __vmnewmap((void *)next->context.ptbase); in switch_mm()
Dpgalloc.h49 mm->context.generation = kmap_generation; in pgd_alloc()
52 mm->context.ptbase = __pa(pgd); in pgd_alloc()
125 mm->context.generation = kmap_generation; in pmd_populate_kernel()
126 current->active_mm->context.generation = kmap_generation; in pmd_populate_kernel()
/linux-4.1.27/arch/um/kernel/skas/
Dmmu.c53 struct mm_context *to_mm = &mm->context; in init_new_context()
63 from_mm = &current->mm->context; in init_new_context()
100 ret = init_stub_pte(mm, STUB_DATA, mm->context.id.stack); in uml_setup_stubs()
104 mm->context.stub_pages[0] = virt_to_page(&__syscall_stub_start); in uml_setup_stubs()
105 mm->context.stub_pages[1] = virt_to_page(mm->context.id.stack); in uml_setup_stubs()
111 mm->context.stub_pages); in uml_setup_stubs()
139 struct mm_context *mmu = &mm->context; in destroy_context()
/linux-4.1.27/tools/perf/util/
Dtrace-event-parse.c31 static int get_common_field(struct scripting_context *context, in get_common_field() argument
34 struct pevent *pevent = context->pevent; in get_common_field()
50 return pevent_read_number(pevent, context->event_data + *offset, *size); in get_common_field()
53 int common_lock_depth(struct scripting_context *context) in common_lock_depth() argument
59 ret = get_common_field(context, &size, &offset, in common_lock_depth()
67 int common_flags(struct scripting_context *context) in common_flags() argument
73 ret = get_common_field(context, &size, &offset, in common_flags()
81 int common_pc(struct scripting_context *context) in common_pc() argument
87 ret = get_common_field(context, &size, &offset, in common_pc()
/linux-4.1.27/arch/score/include/asm/
Dmmu_context.h57 mm->context = asid; in get_new_mmu_context()
68 mm->context = 0; in init_new_context()
78 if ((next->context ^ asid_cache) & ASID_VERSION_MASK) in switch_mm()
81 pevn_set(next->context); in switch_mm()
108 pevn_set(next->context); in activate_mm()
/linux-4.1.27/drivers/net/ieee802154/
Dat86rf230.c75 void (*complete)(void *context);
304 const u8 state, void (*complete)(void *context),
467 at86rf230_async_error_recover(void *context) in at86rf230_async_error_recover() argument
469 struct at86rf230_state_change *ctx = context; in at86rf230_async_error_recover()
491 void (*complete)(void *context), in at86rf230_async_read_reg() argument
519 at86rf230_async_state_assert(void *context) in at86rf230_async_state_assert() argument
521 struct at86rf230_state_change *ctx = context; in at86rf230_async_state_assert()
575 ctx->complete(context); in at86rf230_async_state_assert()
593 at86rf230_async_state_delay(void *context) in at86rf230_async_state_delay() argument
595 struct at86rf230_state_change *ctx = context; in at86rf230_async_state_delay()
[all …]
/linux-4.1.27/arch/arm/plat-omap/
Ddmtimer.c97 timer->context.twer); in omap_timer_restore_context()
99 timer->context.tcrr); in omap_timer_restore_context()
101 timer->context.tldr); in omap_timer_restore_context()
103 timer->context.tmar); in omap_timer_restore_context()
105 timer->context.tsicr); in omap_timer_restore_context()
106 writel_relaxed(timer->context.tier, timer->irq_ena); in omap_timer_restore_context()
108 timer->context.tclr); in omap_timer_restore_context()
452 timer->context.tclr = l; in omap_dm_timer_start()
474 timer->context.tclr = in omap_dm_timer_stop()
557 timer->context.tclr = l; in omap_dm_timer_set_load()
[all …]
/linux-4.1.27/include/linux/
Dssbi.h24 ssbi_reg_read(void *context, unsigned int reg, unsigned int *val) in ssbi_reg_read() argument
29 ret = ssbi_read(context, reg, &v, 1); in ssbi_reg_read()
37 ssbi_reg_write(void *context, unsigned int reg, unsigned int val) in ssbi_reg_write() argument
40 return ssbi_write(context, reg, &v, 1); in ssbi_reg_write()
Dhw_breakpoint.h49 void *context,
62 void *context,
68 void *context);
94 void *context, in register_user_hw_breakpoint() argument
102 void *context, in register_wide_hw_breakpoint_cpu() argument
107 void *context) { return NULL; } in register_wide_hw_breakpoint() argument
Dfence.h78 unsigned context, seqno; member
179 spinlock_t *lock, unsigned context, unsigned seqno);
293 if (WARN_ON(f1->context != f2->context)) in fence_later()
343 __ff->context, __ff->seqno, ##args); \
349 pr_warn("f %u#%u: " fmt, __ff->context, __ff->seqno, \
356 pr_err("f %u#%u: " fmt, __ff->context, __ff->seqno, \
Ddm-kcopyd.h62 void *context);
66 unsigned flags, dm_kcopyd_notify_fn fn, void *context);
80 dm_kcopyd_notify_fn fn, void *context);
85 unsigned flags, dm_kcopyd_notify_fn fn, void *context);
Dfirmware.h46 const char *name, struct device *device, gfp_t gfp, void *context,
47 void (*cont)(const struct firmware *fw, void *context));
61 const char *name, struct device *device, gfp_t gfp, void *context, in request_firmware_nowait() argument
62 void (*cont)(const struct firmware *fw, void *context)) in request_firmware_nowait()
Ddm-region-hash.h36 void *context, void (*dispatch_bios)(void *context,
38 void (*wakeup_workers)(void *context),
39 void (*wakeup_all_recovery_waiters)(void *context),
Dvexpress.h39 struct regmap * (*regmap_init)(struct device *dev, void *context);
40 void (*regmap_exit)(struct regmap *regmap, void *context);
44 struct vexpress_config_bridge_ops *ops, void *context);
/linux-4.1.27/arch/frv/include/asm/
Dmmu_context.h38 change_mm_context(&prev->context, &next->context, next->pgd); \
43 change_mm_context(&prev->context, &next->context, next->pgd); \
/linux-4.1.27/sound/firewire/
Damdtp.c84 s->context = ERR_PTR(-1); in amdtp_stream_init()
602 if (IS_ERR(s->context)) in queue_packet()
610 err = fw_iso_context_queue(s->context, &p, &s->buffer.iso_buffer, in queue_packet()
785 static void out_stream_callback(struct fw_iso_context *context, u32 cycle, in out_stream_callback() argument
803 fw_iso_context_queue_flush(s->context); in out_stream_callback()
806 static void in_stream_callback(struct fw_iso_context *context, u32 cycle, in in_stream_callback() argument
847 fw_iso_context_queue_flush(s->sync_slave->context); in in_stream_callback()
849 fw_iso_context_queue_flush(s->context); in in_stream_callback()
853 static void slave_stream_callback(struct fw_iso_context *context, u32 cycle, in slave_stream_callback() argument
861 static void amdtp_stream_first_callback(struct fw_iso_context *context, in amdtp_stream_first_callback() argument
[all …]
Disight.c58 struct fw_iso_context *context; member
165 static void isight_packet(struct fw_iso_context *context, u32 cycle, in isight_packet() argument
196 err = fw_iso_context_queue(isight->context, &audio_packet, in isight_packet()
205 fw_iso_context_queue_flush(isight->context); in isight_packet()
317 if (!isight->context) in isight_stop_streaming()
320 fw_iso_context_stop(isight->context); in isight_stop_streaming()
321 fw_iso_context_destroy(isight->context); in isight_stop_streaming()
322 isight->context = NULL; in isight_stop_streaming()
348 if (isight->context) { in isight_start_streaming()
367 isight->context = fw_iso_context_create(isight->device->card, in isight_start_streaming()
[all …]
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dprovider.c90 static int c4iw_dealloc_ucontext(struct ib_ucontext *context) in c4iw_dealloc_ucontext() argument
92 struct c4iw_dev *rhp = to_c4iw_dev(context->device); in c4iw_dealloc_ucontext()
93 struct c4iw_ucontext *ucontext = to_c4iw_ucontext(context); in c4iw_dealloc_ucontext()
96 PDBG("%s context %p\n", __func__, context); in c4iw_dealloc_ucontext()
107 struct c4iw_ucontext *context; in c4iw_alloc_ucontext() local
115 context = kzalloc(sizeof(*context), GFP_KERNEL); in c4iw_alloc_ucontext()
116 if (!context) { in c4iw_alloc_ucontext()
121 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext()
122 INIT_LIST_HEAD(&context->mmaps); in c4iw_alloc_ucontext()
123 spin_lock_init(&context->mmap_lock); in c4iw_alloc_ucontext()
[all …]
/linux-4.1.27/drivers/md/
Ddm-round-robin.c71 ps->context = s; in rr_create()
77 struct selector *s = (struct selector *) ps->context; in rr_destroy()
82 ps->context = NULL; in rr_destroy()
114 struct selector *s = (struct selector *) ps->context; in rr_add_path()
149 struct selector *s = (struct selector *) ps->context; in rr_fail_path()
157 struct selector *s = (struct selector *) ps->context; in rr_reinstate_path()
168 struct selector *s = (struct selector *) ps->context; in rr_select_path()
Ddm-queue-length.c60 ps->context = s; in ql_create()
76 struct selector *s = ps->context; in ql_destroy()
81 ps->context = NULL; in ql_destroy()
112 struct selector *s = ps->context; in ql_add_path()
152 struct selector *s = ps->context; in ql_fail_path()
160 struct selector *s = ps->context; in ql_reinstate_path()
174 struct selector *s = ps->context; in ql_select_path()
Ddm-region-hash.c88 void *context; member
92 void (*dispatch_bios)(void *context, struct bio_list *bios);
95 void (*wakeup_workers)(void *context);
98 void (*wakeup_all_recovery_waiters)(void *context);
136 return reg->rh->context; in dm_rh_region_context()
161 void *context, void (*dispatch_bios)(void *context, in dm_region_hash_create() argument
163 void (*wakeup_workers)(void *context), in dm_region_hash_create() argument
164 void (*wakeup_all_recovery_waiters)(void *context), in dm_region_hash_create() argument
188 rh->context = context; in dm_region_hash_create()
379 rh->dispatch_bios(rh->context, &reg->delayed_bios); in complete_resync_work()
[all …]
Ddm-kcopyd.c350 void *context; member
438 void *context = job->context; in run_complete_job() local
452 fn(read_err, write_err, context); in run_complete_job()
460 static void complete_io(unsigned long error, void *context) in complete_io() argument
462 struct kcopyd_job *job = (struct kcopyd_job *) context; in complete_io()
504 .notify.context = job, in run_io_job()
618 void *context) in segment_complete() argument
623 struct kcopyd_job *sub_job = (struct kcopyd_job *) context; in segment_complete()
666 sub_job->context = sub_job; in segment_complete()
703 unsigned int flags, dm_kcopyd_notify_fn fn, void *context) in dm_kcopyd_copy() argument
[all …]
Ddm-log-userspace-base.c322 log->context = lc; in userspace_ctr()
330 struct log_c *lc = log->context; in userspace_dtr()
357 struct log_c *lc = log->context; in userspace_presuspend()
368 struct log_c *lc = log->context; in userspace_postsuspend()
385 struct log_c *lc = log->context; in userspace_resume()
396 struct log_c *lc = log->context; in userspace_get_region_size()
415 struct log_c *lc = log->context; in userspace_is_clean()
443 struct log_c *lc = log->context; in userspace_in_sync()
568 struct log_c *lc = log->context; in userspace_flush()
653 struct log_c *lc = log->context; in userspace_mark_region()
[all …]
Ddm-cache-metadata.h75 typedef int (*load_discard_fn)(void *context, sector_t discard_block_size,
78 load_discard_fn fn, void *context);
86 typedef int (*load_mapping_fn)(void *context, dm_oblock_t oblock,
92 void *context);
Ddm-log.c509 log->context = lc; in create_log_context()
529 struct log_c *lc = (struct log_c *) log->context; in core_dtr()
566 struct log_c *lc = (struct log_c *) log->context; in disk_dtr()
587 struct log_c *lc = (struct log_c *) log->context; in disk_resume()
649 struct log_c *lc = (struct log_c *) log->context; in core_get_region_size()
655 struct log_c *lc = (struct log_c *) log->context; in core_resume()
662 struct log_c *lc = (struct log_c *) log->context; in core_is_clean()
668 struct log_c *lc = (struct log_c *) log->context; in core_in_sync()
681 struct log_c *lc = log->context; in disk_flush()
720 struct log_c *lc = (struct log_c *) log->context; in core_mark_region()
[all …]
Ddm-service-time.c56 ps->context = s; in st_create()
72 struct selector *s = ps->context; in st_destroy()
77 ps->context = NULL; in st_destroy()
109 struct selector *s = ps->context; in st_add_path()
165 struct selector *s = ps->context; in st_fail_path()
173 struct selector *s = ps->context; in st_reinstate_path()
261 struct selector *s = ps->context; in st_select_path()
Ddm-io.c38 void *context; member
118 void *context = io->context; in complete_io() local
125 fn(error_bits, context); in complete_io()
395 static void sync_io_complete(unsigned long error, void *context) in sync_io_complete() argument
397 struct sync_io *sio = context; in sync_io_complete()
422 io->context = &sio; in sync_io()
439 io_notify_fn fn, void *context) in async_io() argument
445 fn(1, context); in async_io()
454 io->context = context; in async_io()
523 &dp, io_req->notify.fn, io_req->notify.context); in dm_io()
Ddm-snap-transient.c28 kfree(store->context); in transient_dtr()
42 struct transient_c *tc = store->context; in transient_prepare_exception()
68 *sectors_allocated = ((struct transient_c *) store->context)->next_free; in transient_usage()
83 store->context = tc; in transient_ctr()
/linux-4.1.27/arch/s390/mm/
Dpgtable.c72 if (mm->context.asce_limit < limit) { in crst_table_upgrade()
74 if (mm->context.asce_limit <= (1UL << 31)) { in crst_table_upgrade()
76 mm->context.asce_limit = 1UL << 42; in crst_table_upgrade()
77 mm->context.asce_bits = _ASCE_TABLE_LENGTH | in crst_table_upgrade()
82 mm->context.asce_limit = 1UL << 53; in crst_table_upgrade()
83 mm->context.asce_bits = _ASCE_TABLE_LENGTH | in crst_table_upgrade()
90 mm->task_size = mm->context.asce_limit; in crst_table_upgrade()
97 if (mm->context.asce_limit < limit) in crst_table_upgrade()
112 while (mm->context.asce_limit > limit) { in crst_table_downgrade()
116 mm->context.asce_limit = 1UL << 42; in crst_table_downgrade()
[all …]
/linux-4.1.27/drivers/net/wireless/mwifiex/
Dusb.c159 struct urb_context *context = (struct urb_context *)urb->context; in mwifiex_usb_rx_complete() local
160 struct mwifiex_adapter *adapter = context->adapter; in mwifiex_usb_rx_complete()
161 struct sk_buff *skb = context->skb; in mwifiex_usb_rx_complete()
172 if (card->rx_cmd_ep == context->ep) in mwifiex_usb_rx_complete()
182 if (card->rx_cmd_ep != context->ep) in mwifiex_usb_rx_complete()
191 status = mwifiex_usb_recv(adapter, skb, context->ep); in mwifiex_usb_rx_complete()
202 if (card->rx_cmd_ep == context->ep) in mwifiex_usb_rx_complete()
210 if (card->rx_cmd_ep != context->ep) in mwifiex_usb_rx_complete()
223 if (card->rx_cmd_ep != context->ep) in mwifiex_usb_rx_complete()
230 if (card->rx_cmd_ep == context->ep) in mwifiex_usb_rx_complete()
[all …]
/linux-4.1.27/arch/xtensa/mm/
Dtlb.c70 mm->context.asid[cpu] = NO_CONTEXT; in local_flush_tlb_mm()
74 mm->context.asid[cpu] = NO_CONTEXT; in local_flush_tlb_mm()
75 mm->context.cpu = -1; in local_flush_tlb_mm()
95 if (mm->context.asid[cpu] == NO_CONTEXT) in local_flush_tlb_range()
100 (unsigned long)mm->context.asid[cpu], start, end); in local_flush_tlb_range()
107 set_rasid_register(ASID_INSERT(mm->context.asid[cpu])); in local_flush_tlb_range()
135 if (mm->context.asid[cpu] == NO_CONTEXT) in local_flush_tlb_page()
141 set_rasid_register(ASID_INSERT(mm->context.asid[cpu])); in local_flush_tlb_page()
/linux-4.1.27/arch/arm/mm/
Dcontext.c62 context_id = mm->context.id.counter; in a15_erratum_get_cpumask()
180 u64 asid = atomic64_read(&mm->context.id); in new_context()
232 if (unlikely(mm->context.vmalloc_seq != init_mm.context.vmalloc_seq)) in check_and_switch_context()
242 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
249 asid = atomic64_read(&mm->context.id); in check_and_switch_context()
252 atomic64_set(&mm->context.id, asid); in check_and_switch_context()
/linux-4.1.27/drivers/net/can/usb/peak_usb/
Dpcan_usb_core.c190 struct peak_usb_device *dev = urb->context; in peak_usb_read_bulk_callback()
255 struct peak_tx_urb_context *context = urb->context; in peak_usb_write_bulk_callback() local
259 BUG_ON(!context); in peak_usb_write_bulk_callback()
261 dev = context->dev; in peak_usb_write_bulk_callback()
274 netdev->stats.tx_bytes += context->data_len; in peak_usb_write_bulk_callback()
293 can_get_echo_skb(netdev, context->echo_index); in peak_usb_write_bulk_callback()
294 context->echo_index = PCAN_USB_MAX_TX_URBS; in peak_usb_write_bulk_callback()
308 struct peak_tx_urb_context *context = NULL; in peak_usb_ndo_start_xmit() local
321 context = dev->tx_contexts + i; in peak_usb_ndo_start_xmit()
325 if (!context) { in peak_usb_ndo_start_xmit()
[all …]
/linux-4.1.27/scripts/coccinelle/api/
Dresource_size.cocci16 virtual context
22 // For context mode
25 @r_context depends on context && !patch && !org@
35 @r_patch depends on !context && patch && !org@
47 @r_org depends on !context && !patch && (org || report)@
54 @rbad_org depends on !context && !patch && (org || report)@
Dplatform_no_drv_owner.cocci7 virtual context
22 @fix1 depends on match1 && patch && !context && !org && !report@
42 @fix2 depends on match2 && patch && !context && !org && !report@
53 @fix1_context depends on match1 && !patch && (context || org || report)@
64 @fix2_context depends on match2 && !patch && (context || org || report)@
Derr_cast.cocci15 virtual context
21 @ depends on context && !patch && !org && !report@
27 @ depends on !context && patch && !org && !report @
34 @r depends on !context && !patch && (org || report)@
/linux-4.1.27/drivers/irqchip/
Dirq-metag-ext.c626 struct meta_intc_context *context; in meta_intc_suspend() local
630 context = kzalloc(sizeof(*context), GFP_ATOMIC); in meta_intc_suspend()
631 if (!context) in meta_intc_suspend()
650 context->vectors[hw] = metag_in32(vec_addr); in meta_intc_suspend()
659 context->levels[bank] = metag_in32(level_addr); in meta_intc_suspend()
662 context->masks[bank] = metag_in32(mask_addr); in meta_intc_suspend()
672 context->txvecint[i][j] = metag_in32(T0VECINT_BHALT + in meta_intc_suspend()
677 meta_intc_context = context; in meta_intc_suspend()
693 struct meta_intc_context *context = meta_intc_context; in meta_intc_resume() local
715 metag_out32(context->vectors[hw], vec_addr); in meta_intc_resume()
[all …]
/linux-4.1.27/include/linux/i2c/
Dpcf857x.h37 void *context);
40 void *context);
41 void *context; member
Dmax732x.h13 void *context; /* param to setup/teardown */ member
17 void *context);
20 void *context);
/linux-4.1.27/include/linux/platform_data/
Dpca953x.h19 void *context; /* param to setup/teardown */ member
23 void *context);
26 void *context);
/linux-4.1.27/drivers/staging/i2o/
Diop.c117 entry->context = atomic_read(&c->context_list_counter); in i2o_cntxt_list_add()
123 osm_debug("%s: Add context to list %p -> %d\n", c->name, ptr, context); in i2o_cntxt_list_add()
125 return entry->context; in i2o_cntxt_list_add()
141 u32 context = 0; in i2o_cntxt_list_remove() local
148 context = entry->context; in i2o_cntxt_list_remove()
154 if (!context) in i2o_cntxt_list_remove()
159 context, ptr); in i2o_cntxt_list_remove()
161 return context; in i2o_cntxt_list_remove()
172 void *i2o_cntxt_list_get(struct i2o_controller *c, u32 context) in i2o_cntxt_list_get() argument
180 if (entry->context == context) { in i2o_cntxt_list_get()
[all …]
Ddriver.c111 drv->context = i; in i2o_driver_register()
116 osm_debug("driver %s gets context id %d\n", drv->name, drv->context); in i2o_driver_register()
166 i2o_drivers[drv->context] = NULL; in i2o_driver_unregister()
192 u32 context = le32_to_cpu(msg->u.s.icntxt); in i2o_driver_dispatch() local
195 if (unlikely(context >= i2o_max_drivers)) { in i2o_driver_dispatch()
197 context); in i2o_driver_dispatch()
202 drv = i2o_drivers[context]; in i2o_driver_dispatch()
207 context); in i2o_driver_dispatch()
Dexec-osm.c148 msg->u.s.icntxt = cpu_to_le32(i2o_exec_driver.context); in i2o_msg_post_wait_mem()
217 struct i2o_message *msg, u32 context) in i2o_msg_post_wait_complete() argument
231 if (wait->tcntxt == context) { in i2o_msg_post_wait_complete()
264 context); in i2o_msg_post_wait_complete()
406 msg->u.s.icntxt = cpu_to_le32(i2o_exec_driver.context); in i2o_exec_lct_notify()
463 u32 context; in i2o_exec_reply() local
476 context = readl(&pmsg->u.s.tcntxt); in i2o_exec_reply()
483 context = le32_to_cpu(msg->u.s.tcntxt); in i2o_exec_reply()
485 if (context & 0x80000000) in i2o_exec_reply()
486 return i2o_msg_post_wait_complete(c, m, msg, context); in i2o_exec_reply()
/linux-4.1.27/drivers/video/fbdev/omap2/
Dvrfb.c40 #define SMS_ROT_CONTROL(context) (0x0 + 0x10 * context) argument
41 #define SMS_ROT_SIZE(context) (0x4 + 0x10 * context) argument
42 #define SMS_ROT_PHYSICAL_BA(context) (0x8 + 0x10 * context) argument
188 u8 ctx = vrfb->context; in omap_vrfb_setup()
264 int ctx = vrfb->context; in omap_vrfb_release_ctx()
284 vrfb->context = 0xff; in omap_vrfb_release_ctx()
317 vrfb->context = ctx; in omap_vrfb_request_ctx()
/linux-4.1.27/drivers/bus/
Dvexpress-config.c23 void *context; member
89 bridge->ops->regmap_exit(regmap, bridge->context); in vexpress_config_devres_release()
110 regmap = bridge->ops->regmap_init(dev, bridge->context); in devm_regmap_init_vexpress_config()
124 struct vexpress_config_bridge_ops *ops, void *context) in vexpress_config_bridge_register() argument
149 bridge->context = context; in vexpress_config_bridge_register()
/linux-4.1.27/drivers/hv/
Dhv_util.c59 static void shutdown_onchannelcallback(void *context);
64 static void timesync_onchannelcallback(void *context);
69 static void heartbeat_onchannelcallback(void *context);
102 static void shutdown_onchannelcallback(void *context) in shutdown_onchannelcallback() argument
104 struct vmbus_channel *channel = context; in shutdown_onchannelcallback()
233 static void timesync_onchannelcallback(void *context) in timesync_onchannelcallback() argument
235 struct vmbus_channel *channel = context; in timesync_onchannelcallback()
276 static void heartbeat_onchannelcallback(void *context) in heartbeat_onchannelcallback() argument
278 struct vmbus_channel *channel = context; in heartbeat_onchannelcallback()
/linux-4.1.27/drivers/isdn/hardware/eicon/
Distream.c39 int diva_istream_write(void *context,
46 int diva_istream_read(void *context,
70 int diva_istream_write(void *context, in diva_istream_write() argument
77 ADAPTER *a = (ADAPTER *)context; in diva_istream_write()
148 int diva_istream_read(void *context, in diva_istream_read() argument
155 ADAPTER *a = (ADAPTER *)context; in diva_istream_read()
Ddadapter.c36 void IDI_CALL_ENTITY_T *context; member
85 void IDI_CALL_ENTITY_T *context);
226 (void IDI_CALL_ENTITY_T *)pinfo->context); in diva_dadapter_request()
270 void IDI_CALL_ENTITY_T *context) { in diva_register_adapter_callback() argument
278 NotificationTable[i].context = context; in diva_register_adapter_callback()
296 NotificationTable[handle].context = NULL; in diva_remove_adapter_callback()
322 (*(nfy.callback))(nfy.context, d, removal); in diva_notify_adapter_change()
/linux-4.1.27/Documentation/
Dunshare.txt61 2.1 Per-security context namespaces
65 such as per-user and/or per-security context instance of /tmp, /var/tmp or
66 per-security context instance of a user's home directory, isolate user
83 disassociate parts of the context during the servicing of the
109 allowed incremental context unsharing in future without an ABI change.
112 new context flags without requiring a rebuild of old applications.
113 If and when new context flags are added, unshare design should allow
119 unshare - disassociate parts of the process execution context
128 context that are currently being shared with other processes. Part
129 of execution context, such as the namespace, is shared by default
[all …]
/linux-4.1.27/drivers/usb/musb/
Dmusb_core.c2269 musb->context.frame = musb_readw(musb_base, MUSB_FRAME); in musb_save_context()
2270 musb->context.testmode = musb_readb(musb_base, MUSB_TESTMODE); in musb_save_context()
2271 musb->context.busctl = musb_read_ulpi_buscontrol(musb->mregs); in musb_save_context()
2272 musb->context.power = musb_readb(musb_base, MUSB_POWER); in musb_save_context()
2273 musb->context.intrusbe = musb_readb(musb_base, MUSB_INTRUSBE); in musb_save_context()
2274 musb->context.index = musb_readb(musb_base, MUSB_INDEX); in musb_save_context()
2275 musb->context.devctl = musb_readb(musb_base, MUSB_DEVCTL); in musb_save_context()
2289 musb->context.index_regs[i].txmaxp = in musb_save_context()
2291 musb->context.index_regs[i].txcsr = in musb_save_context()
2293 musb->context.index_regs[i].rxmaxp = in musb_save_context()
[all …]
/linux-4.1.27/drivers/uwb/
Dneh.c109 u8 context; member
176 neh->context = result; in __uwb_rc_ctx_get()
186 if (neh->context == 0) in __uwb_rc_ctx_put()
188 if (test_bit(neh->context, rc->ctx_bm) == 0) { in __uwb_rc_ctx_put()
190 neh->context); in __uwb_rc_ctx_put()
193 clear_bit(neh->context, rc->ctx_bm); in __uwb_rc_ctx_put()
194 neh->context = 0; in __uwb_rc_ctx_put()
239 cmd->bCommandContext = neh->context; in uwb_rc_neh_add()
295 if (neh->context) in uwb_rc_neh_arm()
311 && neh->context == rceb->bEventContext; in uwb_rc_neh_match()
[all …]
/linux-4.1.27/drivers/media/firewire/
Dfiredtv-fw.c77 struct fw_iso_context *context; member
93 return fw_iso_context_queue(ctx->context, &p, &ctx->buffer, in queue_iso()
97 static void handle_iso(struct fw_iso_context *context, u32 cycle, in handle_iso() argument
127 fw_iso_context_queue_flush(ctx->context); in handle_iso()
141 ctx->context = fw_iso_context_create(device->card, in fdtv_start_iso()
144 if (IS_ERR(ctx->context)) { in fdtv_start_iso()
145 err = PTR_ERR(ctx->context); in fdtv_start_iso()
166 err = fw_iso_context_start(ctx->context, -1, 0, in fdtv_start_iso()
177 fw_iso_context_destroy(ctx->context); in fdtv_start_iso()
188 fw_iso_context_stop(ctx->context); in fdtv_stop_iso()
[all …]
/linux-4.1.27/arch/arm64/include/asm/
Dmmu.h26 .context.id_lock = __RAW_SPIN_LOCK_UNLOCKED(name.context.id_lock),
28 #define ASID(mm) ((mm)->context.id & 0xffff)
/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_ib_verbs.c419 struct ib_ucontext *context, in usnic_ib_alloc_pd() argument
438 pd, context, ibdev->name); in usnic_ib_alloc_pd()
466 ucontext = to_uucontext(pd->uobject->context); in usnic_ib_create_qp()
574 int vector, struct ib_ucontext *context, in usnic_ib_create_cq() argument
629 usnic_uiom_reg_release(mr->umem, ibmr->pd->uobject->context->closing); in usnic_ib_dereg_mr()
637 struct usnic_ib_ucontext *context; in usnic_ib_alloc_ucontext() local
641 context = kmalloc(sizeof(*context), GFP_KERNEL); in usnic_ib_alloc_ucontext()
642 if (!context) in usnic_ib_alloc_ucontext()
645 INIT_LIST_HEAD(&context->qp_grp_list); in usnic_ib_alloc_ucontext()
647 list_add_tail(&context->link, &us_ibdev->ctx_list); in usnic_ib_alloc_ucontext()
[all …]
/linux-4.1.27/arch/avr32/include/asm/
Dmmu_context.h45 if (((mm->context ^ mc) & MMU_CONTEXT_VERSION_MASK) == 0) in get_mmu_context()
64 mm->context = mc; in get_mmu_context()
74 mm->context = NO_CONTEXT; in init_new_context()
105 set_asid(mm->context & MMU_CONTEXT_ASID_MASK); in activate_context()
/linux-4.1.27/arch/x86/math-emu/
Dfpu_system.h29 mutex_lock(&current->mm->context.lock); in FPU_get_ldt_descriptor()
30 if (current->mm->context.ldt && seg < current->mm->context.ldt->size) in FPU_get_ldt_descriptor()
31 ret = current->mm->context.ldt->entries[seg]; in FPU_get_ldt_descriptor()
32 mutex_unlock(&current->mm->context.lock); in FPU_get_ldt_descriptor()
/linux-4.1.27/arch/score/mm/
Dtlb-score.c70 pevn_set(mm->context & ASID_MASK); in drop_mmu_context()
76 if (mm->context != 0) in local_flush_tlb_mm()
84 unsigned long vma_mm_context = mm->context; in local_flush_tlb_range()
85 if (mm->context != 0) { in local_flush_tlb_range()
161 if (vma && vma->vm_mm->context != 0) { in local_flush_tlb_page()
164 unsigned long vma_ASID = vma->vm_mm->context; in local_flush_tlb_page()
/linux-4.1.27/drivers/misc/
Dvexpress-syscfg.c117 static int vexpress_syscfg_read(void *context, unsigned int index, in vexpress_syscfg_read() argument
120 struct vexpress_syscfg_func *func = context; in vexpress_syscfg_read()
125 static int vexpress_syscfg_write(void *context, unsigned int index, in vexpress_syscfg_write() argument
128 struct vexpress_syscfg_func *func = context; in vexpress_syscfg_write()
146 void *context) in vexpress_syscfg_regmap_init() argument
149 struct vexpress_syscfg *syscfg = context; in vexpress_syscfg_regmap_init()
227 static void vexpress_syscfg_regmap_exit(struct regmap *regmap, void *context) in vexpress_syscfg_regmap_exit() argument
229 struct vexpress_syscfg *syscfg = context; in vexpress_syscfg_regmap_exit()
/linux-4.1.27/scripts/coccinelle/null/
Dderef_null.cocci13 virtual context
51 @r depends on !context && (org || report) exists@
91 @script:python depends on !context && !org && report@
101 @script:python depends on !context && org && !report@
112 @s depends on !context && (org || report) exists@
151 @script:python depends on !context && !org && report@
160 @script:python depends on !context && org && !report@
170 // For context mode
172 @depends on context && !org && !report exists@
245 @depends on context && !org && !report exists@
/linux-4.1.27/Documentation/infiniband/
Dcore_locking.txt8 Sleeping and interrupt context
25 which may not sleep and must be callable from any context.
40 are therefore safe to call from any context.
47 the midlayer is also safe to call from any context.
92 The context in which completion event and asynchronous event
94 may be process context, softirq context, or interrupt context.
105 ib_unregister_device() from process context. It must not hold any
/linux-4.1.27/arch/cris/arch-v32/mm/
Dtlb.c88 page_id = mm->context.page_id; in __flush_tlb_mm()
130 page_id = vma->vm_mm->context.page_id; in __flush_tlb_page()
171 mm->context.page_id = NO_CONTEXT; in init_new_context()
200 SPEC_REG_WR(SPEC_REG_PID, next->context.page_id | in switch_mm()
203 SPEC_REG_WR(SPEC_REG_PID, next->context.page_id); in switch_mm()
/linux-4.1.27/arch/cris/arch-v10/mm/
Dtlb.c67 int page_id = mm->context.page_id; in flush_tlb_mm()
102 int page_id = mm->context.page_id; in flush_tlb_page()
145 mm->context.page_id = NO_CONTEXT; in init_new_context()
170 next->context, next)); in switch_mm()
173 page_id, next->context.page_id); in switch_mm()
/linux-4.1.27/include/rdma/
Dib_sa.h306 void *context),
307 void *context,
318 void *context),
319 void *context,
327 void *context; member
365 void *context);
427 void *context),
428 void *context,
/linux-4.1.27/arch/x86/kernel/cpu/mcheck/
Dmce-severity.c32 enum context { IN_KERNEL = 1, IN_USER = 2 }; enum
43 unsigned char context; member
49 #define KERNEL .context = IN_KERNEL
50 #define USER .context = IN_USER
195 enum context ctx = error_context(m); in mce_severity_amd()
246 enum context ctx = error_context(m); in mce_severity_intel()
258 if (s->context && ctx != s->context) in mce_severity_intel()
/linux-4.1.27/drivers/firewire/
Dohci.c112 struct context;
114 typedef int (*descriptor_callback_t)(struct context *ctx,
130 struct context { struct
178 struct context context; argument
219 struct context at_request_ctx;
220 struct context at_response_ctx;
1069 struct context *ctx = (struct context *) data; in context_tasklet()
1114 static int context_add_buffer(struct context *ctx) in context_add_buffer()
1143 static int context_init(struct context *ctx, struct fw_ohci *ohci, in context_init()
1176 static void context_release(struct context *ctx) in context_release()
[all …]
/linux-4.1.27/Documentation/powerpc/
Dcxl.txt16 Coherent in this context means that the accelerator and CPUs can
59 the fault. The context to which this fault is serviced is based on
69 When using dedicated mode only one MMU context is supported. In
76 support fewer). In this mode, the AFU sends a 16 bit context ID
77 with each of its requests. This tells the PSL which context is
80 determine the userspace context associated with an operation.
88 just a per context portion. The hardware is self describing, hence
89 the kernel can determine the offset and size of the per context
107 The WED is a 64-bit parameter passed to the AFU when a context is
121 master context and /dev/cxl/afu0.0s will correspond to a slave
[all …]
/linux-4.1.27/include/uapi/drm/
Dtegra_drm.h65 __u64 context; member
69 __u64 context; member
73 __u64 context; member
79 __u64 context; member
117 __u64 context; member
/linux-4.1.27/drivers/usb/misc/
Dusbtest.c232 complete(urb->context); in simple_callback()
425 urb->context = &completion; in simple_io()
1036 struct ctrl_ctx *ctx = urb->context; in ctrl_complete()
1151 struct ctrl_ctx context; in test_ctrl_queue() local
1157 spin_lock_init(&context.lock); in test_ctrl_queue()
1158 context.dev = dev; in test_ctrl_queue()
1159 init_completion(&context.complete); in test_ctrl_queue()
1160 context.count = param->sglen * param->iterations; in test_ctrl_queue()
1161 context.pending = 0; in test_ctrl_queue()
1162 context.status = -ENOMEM; in test_ctrl_queue()
[all …]
/linux-4.1.27/arch/um/kernel/
Dexec.c27 ret = unmap(&current->mm->context.id, 0, STUB_START, 0, &data); in flush_thread()
28 ret = ret || unmap(&current->mm->context.id, STUB_END, in flush_thread()
38 __switch_mm(&current->mm->context.id); in flush_thread()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/gr/fuc/
Dhub.fuc67 // 31:0: total PGRAPH context size
134 // context size calculation, reserve first 256 bytes for use by fuc
144 // calculate size of mmio context data
168 // context data in GPCn_CC_SCRATCH[1], and starting its FUC (which
172 // when it has completed, and return the size of its context data
191 // wait for it to complete, and adjust context size
212 // save context size, and tell host we're ready
229 // context switch, requested by GPU?
269 // ack the context switch request
276 // request to set current channel? (*not* a context switch)
[all …]
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dnv40.c73 u32 context, chid = chan->base.chid; in nv40_fifo_object_attach() local
77 context = nv_gpuobj(object)->addr >> 4; in nv40_fifo_object_attach()
79 context = 0x00000004; /* just non-zero */ in nv40_fifo_object_attach()
84 context |= 0x00000000; in nv40_fifo_object_attach()
87 context |= 0x00100000; in nv40_fifo_object_attach()
90 context |= 0x00200000; in nv40_fifo_object_attach()
96 context |= chid << 23; in nv40_fifo_object_attach()
99 ret = nvkm_ramht_insert(priv->ramht, chid, handle, context); in nv40_fifo_object_attach()
/linux-4.1.27/drivers/net/can/usb/
Dusb_8dev.c516 struct usb_8dev_priv *priv = urb->context; in usb_8dev_read_bulk_callback()
576 struct usb_8dev_tx_urb_context *context = urb->context; in usb_8dev_write_bulk_callback() local
580 BUG_ON(!context); in usb_8dev_write_bulk_callback()
582 priv = context->priv; in usb_8dev_write_bulk_callback()
599 netdev->stats.tx_bytes += context->dlc; in usb_8dev_write_bulk_callback()
601 can_get_echo_skb(netdev, context->echo_index); in usb_8dev_write_bulk_callback()
606 context->echo_index = MAX_TX_URBS; in usb_8dev_write_bulk_callback()
620 struct usb_8dev_tx_urb_context *context = NULL; in usb_8dev_start_xmit() local
661 context = &priv->tx_contexts[i]; in usb_8dev_start_xmit()
669 if (!context) in usb_8dev_start_xmit()
[all …]
Dems_usb.c274 struct ems_usb *dev = urb->context; in ems_usb_read_interrupt_callback()
422 struct ems_usb *dev = urb->context; in ems_usb_read_bulk_callback()
508 struct ems_tx_urb_context *context = urb->context; in ems_usb_write_bulk_callback() local
512 BUG_ON(!context); in ems_usb_write_bulk_callback()
514 dev = context->dev; in ems_usb_write_bulk_callback()
533 netdev->stats.tx_bytes += context->dlc; in ems_usb_write_bulk_callback()
535 can_get_echo_skb(netdev, context->echo_index); in ems_usb_write_bulk_callback()
538 context->echo_index = MAX_TX_URBS; in ems_usb_write_bulk_callback()
745 struct ems_tx_urb_context *context = NULL; in ems_usb_start_xmit() local
794 context = &dev->tx_contexts[i]; in ems_usb_start_xmit()
[all …]
/linux-4.1.27/arch/tile/kernel/
Dpci_gx.c149 gxio_trio_context_t *context = &trio_contexts[trio_index]; in tile_pcie_open() local
154 ret = gxio_trio_init(context, trio_index); in tile_pcie_open()
159 ret = gxio_trio_alloc_asids(context, 1, 0, 0); in tile_pcie_open()
166 context->asid = ret; in tile_pcie_open()
174 ret = gxio_trio_alloc_pio_regions(context, 1, 0, 0); in tile_pcie_open()
181 context->pio_cfg_index = ret; in tile_pcie_open()
187 ret = gxio_trio_init_pio_region_aux(context, context->pio_cfg_index, in tile_pcie_open()
197 ret = gxio_trio_get_port_property(context, &pcie_ports[trio_index]); in tile_pcie_open()
204 context->mmio_base_mac = in tile_pcie_open()
205 iorpc_ioremap(context->fd, 0, HV_TRIO_CONFIG_IOREMAP_SIZE); in tile_pcie_open()
[all …]
/linux-4.1.27/include/acpi/
Dacpixf.h460 handler, void *context))
476 void *context,
481 void *context,
553 void *context))
560 void *context))
567 *context))
579 void *context))
587 void *context))
598 void *context))
613 void *context))
[all …]
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_uverbs.c65 int ehca_dealloc_ucontext(struct ib_ucontext *context) in ehca_dealloc_ucontext() argument
67 kfree(container_of(context, struct ehca_ucontext, ib_ucontext)); in ehca_dealloc_ucontext()
248 int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in ehca_mmap() argument
269 if (!cq->ib_cq.uobject || cq->ib_cq.uobject->context != context) in ehca_mmap()
291 if (!uobject || uobject->context != context) in ehca_mmap()
/linux-4.1.27/arch/m32r/include/asm/
Dmmu_context.h26 #define mm_context(mm) mm->context
30 #define mm_context(mm) mm->context[smp_processor_id()]
78 mm->context = NO_CONTEXT; in init_new_context()
84 mm->context[i] = NO_CONTEXT; in init_new_context()

123456789