Lines Matching refs:thr
560 struct task_struct *thr; /* thread */ member
583 d->thr = NULL; in crc32_threadfn()
602 struct task_struct *thr; /* thread */ member
626 d->thr = NULL; in lzo_compress_threadfn()
661 unsigned thr, run_threads, nr_threads; in save_image_lzo() local
688 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
689 memset(&data[thr], 0, offsetof(struct cmp_data, go)); in save_image_lzo()
702 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
703 init_waitqueue_head(&data[thr].go); in save_image_lzo()
704 init_waitqueue_head(&data[thr].done); in save_image_lzo()
706 data[thr].thr = kthread_run(lzo_compress_threadfn, in save_image_lzo()
707 &data[thr], in save_image_lzo()
708 "image_compress/%u", thr); in save_image_lzo()
709 if (IS_ERR(data[thr].thr)) { in save_image_lzo()
710 data[thr].thr = NULL; in save_image_lzo()
726 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
727 crc->unc[thr] = data[thr].unc; in save_image_lzo()
728 crc->unc_len[thr] = &data[thr].unc_len; in save_image_lzo()
731 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in save_image_lzo()
732 if (IS_ERR(crc->thr)) { in save_image_lzo()
733 crc->thr = NULL; in save_image_lzo()
755 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
764 memcpy(data[thr].unc + off, in save_image_lzo()
777 data[thr].unc_len = off; in save_image_lzo()
779 atomic_set(&data[thr].ready, 1); in save_image_lzo()
780 wake_up(&data[thr].go); in save_image_lzo()
783 if (!thr) in save_image_lzo()
786 crc->run_threads = thr; in save_image_lzo()
790 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in save_image_lzo()
791 wait_event(data[thr].done, in save_image_lzo()
792 atomic_read(&data[thr].stop)); in save_image_lzo()
793 atomic_set(&data[thr].stop, 0); in save_image_lzo()
795 ret = data[thr].ret; in save_image_lzo()
802 if (unlikely(!data[thr].cmp_len || in save_image_lzo()
803 data[thr].cmp_len > in save_image_lzo()
804 lzo1x_worst_compress(data[thr].unc_len))) { in save_image_lzo()
811 *(size_t *)data[thr].cmp = data[thr].cmp_len; in save_image_lzo()
822 off < LZO_HEADER + data[thr].cmp_len; in save_image_lzo()
824 memcpy(page, data[thr].cmp + off, PAGE_SIZE); in save_image_lzo()
846 if (crc->thr) in save_image_lzo()
847 kthread_stop(crc->thr); in save_image_lzo()
851 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
852 if (data[thr].thr) in save_image_lzo()
853 kthread_stop(data[thr].thr); in save_image_lzo()
1093 struct task_struct *thr; /* thread */ member
1116 d->thr = NULL; in lzo_decompress_threadfn()
1151 unsigned i, thr, run_threads, nr_threads; in load_image_lzo() local
1181 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1182 memset(&data[thr], 0, offsetof(struct dec_data, go)); in load_image_lzo()
1195 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1196 init_waitqueue_head(&data[thr].go); in load_image_lzo()
1197 init_waitqueue_head(&data[thr].done); in load_image_lzo()
1199 data[thr].thr = kthread_run(lzo_decompress_threadfn, in load_image_lzo()
1200 &data[thr], in load_image_lzo()
1201 "image_decompress/%u", thr); in load_image_lzo()
1202 if (IS_ERR(data[thr].thr)) { in load_image_lzo()
1203 data[thr].thr = NULL; in load_image_lzo()
1219 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1220 crc->unc[thr] = data[thr].unc; in load_image_lzo()
1221 crc->unc_len[thr] = &data[thr].unc_len; in load_image_lzo()
1224 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in load_image_lzo()
1225 if (IS_ERR(crc->thr)) { in load_image_lzo()
1226 crc->thr = NULL; in load_image_lzo()
1321 for (thr = 0; have && thr < nr_threads; thr++) { in load_image_lzo()
1322 data[thr].cmp_len = *(size_t *)page[pg]; in load_image_lzo()
1323 if (unlikely(!data[thr].cmp_len || in load_image_lzo()
1324 data[thr].cmp_len > in load_image_lzo()
1332 need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER, in load_image_lzo()
1343 off < LZO_HEADER + data[thr].cmp_len; in load_image_lzo()
1345 memcpy(data[thr].cmp + off, in load_image_lzo()
1353 atomic_set(&data[thr].ready, 1); in load_image_lzo()
1354 wake_up(&data[thr].go); in load_image_lzo()
1370 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in load_image_lzo()
1371 wait_event(data[thr].done, in load_image_lzo()
1372 atomic_read(&data[thr].stop)); in load_image_lzo()
1373 atomic_set(&data[thr].stop, 0); in load_image_lzo()
1375 ret = data[thr].ret; in load_image_lzo()
1383 if (unlikely(!data[thr].unc_len || in load_image_lzo()
1384 data[thr].unc_len > LZO_UNC_SIZE || in load_image_lzo()
1385 data[thr].unc_len & (PAGE_SIZE - 1))) { in load_image_lzo()
1393 off < data[thr].unc_len; off += PAGE_SIZE) { in load_image_lzo()
1395 data[thr].unc + off, PAGE_SIZE); in load_image_lzo()
1406 crc->run_threads = thr + 1; in load_image_lzo()
1414 crc->run_threads = thr; in load_image_lzo()
1445 if (crc->thr) in load_image_lzo()
1446 kthread_stop(crc->thr); in load_image_lzo()
1450 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1451 if (data[thr].thr) in load_image_lzo()
1452 kthread_stop(data[thr].thr); in load_image_lzo()