Lines Matching refs:thr
486 struct task_struct *thr; /* thread */ member
509 d->thr = NULL; in crc32_threadfn()
528 struct task_struct *thr; /* thread */ member
552 d->thr = NULL; in lzo_compress_threadfn()
587 unsigned thr, run_threads, nr_threads; in save_image_lzo() local
612 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
613 memset(&data[thr], 0, offsetof(struct cmp_data, go)); in save_image_lzo()
626 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
627 init_waitqueue_head(&data[thr].go); in save_image_lzo()
628 init_waitqueue_head(&data[thr].done); in save_image_lzo()
630 data[thr].thr = kthread_run(lzo_compress_threadfn, in save_image_lzo()
631 &data[thr], in save_image_lzo()
632 "image_compress/%u", thr); in save_image_lzo()
633 if (IS_ERR(data[thr].thr)) { in save_image_lzo()
634 data[thr].thr = NULL; in save_image_lzo()
650 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
651 crc->unc[thr] = data[thr].unc; in save_image_lzo()
652 crc->unc_len[thr] = &data[thr].unc_len; in save_image_lzo()
655 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in save_image_lzo()
656 if (IS_ERR(crc->thr)) { in save_image_lzo()
657 crc->thr = NULL; in save_image_lzo()
680 for (thr = 0; thr < nr_threads; thr++) { in save_image_lzo()
689 memcpy(data[thr].unc + off, in save_image_lzo()
702 data[thr].unc_len = off; in save_image_lzo()
704 atomic_set(&data[thr].ready, 1); in save_image_lzo()
705 wake_up(&data[thr].go); in save_image_lzo()
708 if (!thr) in save_image_lzo()
711 crc->run_threads = thr; in save_image_lzo()
715 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in save_image_lzo()
716 wait_event(data[thr].done, in save_image_lzo()
717 atomic_read(&data[thr].stop)); in save_image_lzo()
718 atomic_set(&data[thr].stop, 0); in save_image_lzo()
720 ret = data[thr].ret; in save_image_lzo()
727 if (unlikely(!data[thr].cmp_len || in save_image_lzo()
728 data[thr].cmp_len > in save_image_lzo()
729 lzo1x_worst_compress(data[thr].unc_len))) { in save_image_lzo()
736 *(size_t *)data[thr].cmp = data[thr].cmp_len; in save_image_lzo()
747 off < LZO_HEADER + data[thr].cmp_len; in save_image_lzo()
749 memcpy(page, data[thr].cmp + off, PAGE_SIZE); in save_image_lzo()
771 if (crc->thr) in save_image_lzo()
772 kthread_stop(crc->thr); in save_image_lzo()
776 for (thr = 0; thr < nr_threads; thr++) in save_image_lzo()
777 if (data[thr].thr) in save_image_lzo()
778 kthread_stop(data[thr].thr); in save_image_lzo()
1017 struct task_struct *thr; /* thread */ member
1040 d->thr = NULL; in lzo_decompress_threadfn()
1075 unsigned i, thr, run_threads, nr_threads; in load_image_lzo() local
1103 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1104 memset(&data[thr], 0, offsetof(struct dec_data, go)); in load_image_lzo()
1117 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1118 init_waitqueue_head(&data[thr].go); in load_image_lzo()
1119 init_waitqueue_head(&data[thr].done); in load_image_lzo()
1121 data[thr].thr = kthread_run(lzo_decompress_threadfn, in load_image_lzo()
1122 &data[thr], in load_image_lzo()
1123 "image_decompress/%u", thr); in load_image_lzo()
1124 if (IS_ERR(data[thr].thr)) { in load_image_lzo()
1125 data[thr].thr = NULL; in load_image_lzo()
1141 for (thr = 0; thr < nr_threads; thr++) { in load_image_lzo()
1142 crc->unc[thr] = data[thr].unc; in load_image_lzo()
1143 crc->unc_len[thr] = &data[thr].unc_len; in load_image_lzo()
1146 crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); in load_image_lzo()
1147 if (IS_ERR(crc->thr)) { in load_image_lzo()
1148 crc->thr = NULL; in load_image_lzo()
1244 for (thr = 0; have && thr < nr_threads; thr++) { in load_image_lzo()
1245 data[thr].cmp_len = *(size_t *)page[pg]; in load_image_lzo()
1246 if (unlikely(!data[thr].cmp_len || in load_image_lzo()
1247 data[thr].cmp_len > in load_image_lzo()
1255 need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER, in load_image_lzo()
1266 off < LZO_HEADER + data[thr].cmp_len; in load_image_lzo()
1268 memcpy(data[thr].cmp + off, in load_image_lzo()
1276 atomic_set(&data[thr].ready, 1); in load_image_lzo()
1277 wake_up(&data[thr].go); in load_image_lzo()
1293 for (run_threads = thr, thr = 0; thr < run_threads; thr++) { in load_image_lzo()
1294 wait_event(data[thr].done, in load_image_lzo()
1295 atomic_read(&data[thr].stop)); in load_image_lzo()
1296 atomic_set(&data[thr].stop, 0); in load_image_lzo()
1298 ret = data[thr].ret; in load_image_lzo()
1306 if (unlikely(!data[thr].unc_len || in load_image_lzo()
1307 data[thr].unc_len > LZO_UNC_SIZE || in load_image_lzo()
1308 data[thr].unc_len & (PAGE_SIZE - 1))) { in load_image_lzo()
1316 off < data[thr].unc_len; off += PAGE_SIZE) { in load_image_lzo()
1318 data[thr].unc + off, PAGE_SIZE); in load_image_lzo()
1329 crc->run_threads = thr + 1; in load_image_lzo()
1337 crc->run_threads = thr; in load_image_lzo()
1368 if (crc->thr) in load_image_lzo()
1369 kthread_stop(crc->thr); in load_image_lzo()
1373 for (thr = 0; thr < nr_threads; thr++) in load_image_lzo()
1374 if (data[thr].thr) in load_image_lzo()
1375 kthread_stop(data[thr].thr); in load_image_lzo()