wl_wrk 576 drivers/mtd/ubi/wl.c static int erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk, wl_wrk 593 drivers/mtd/ubi/wl.c struct ubi_work *wl_wrk; wl_wrk 600 drivers/mtd/ubi/wl.c wl_wrk = kmalloc(sizeof(struct ubi_work), GFP_NOFS); wl_wrk 601 drivers/mtd/ubi/wl.c if (!wl_wrk) wl_wrk 604 drivers/mtd/ubi/wl.c wl_wrk->func = &erase_worker; wl_wrk 605 drivers/mtd/ubi/wl.c wl_wrk->e = e; wl_wrk 606 drivers/mtd/ubi/wl.c wl_wrk->vol_id = vol_id; wl_wrk 607 drivers/mtd/ubi/wl.c wl_wrk->lnum = lnum; wl_wrk 608 drivers/mtd/ubi/wl.c wl_wrk->torture = torture; wl_wrk 611 drivers/mtd/ubi/wl.c __schedule_ubi_work(ubi, wl_wrk); wl_wrk 613 drivers/mtd/ubi/wl.c schedule_ubi_work(ubi, wl_wrk); wl_wrk 617 drivers/mtd/ubi/wl.c static int __erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk); wl_wrk 630 drivers/mtd/ubi/wl.c struct ubi_work wl_wrk; wl_wrk 634 drivers/mtd/ubi/wl.c wl_wrk.e = e; wl_wrk 635 drivers/mtd/ubi/wl.c wl_wrk.vol_id = vol_id; wl_wrk 636 drivers/mtd/ubi/wl.c wl_wrk.lnum = lnum; wl_wrk 637 drivers/mtd/ubi/wl.c wl_wrk.torture = torture; wl_wrk 639 drivers/mtd/ubi/wl.c return __erase_worker(ubi, &wl_wrk); wl_wrk 1082 drivers/mtd/ubi/wl.c static int __erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk) wl_wrk 1084 drivers/mtd/ubi/wl.c struct ubi_wl_entry *e = wl_wrk->e; wl_wrk 1086 drivers/mtd/ubi/wl.c int vol_id = wl_wrk->vol_id; wl_wrk 1087 drivers/mtd/ubi/wl.c int lnum = wl_wrk->lnum; wl_wrk 1091 drivers/mtd/ubi/wl.c pnum, e->ec, wl_wrk->vol_id, wl_wrk->lnum); wl_wrk 1093 drivers/mtd/ubi/wl.c err = sync_erase(ubi, e, wl_wrk->torture); wl_wrk 1196 drivers/mtd/ubi/wl.c static int erase_worker(struct ubi_device *ubi, struct ubi_work *wl_wrk, wl_wrk 1202 drivers/mtd/ubi/wl.c struct ubi_wl_entry *e = wl_wrk->e; wl_wrk 1205 drivers/mtd/ubi/wl.c kfree(wl_wrk); wl_wrk 1210 drivers/mtd/ubi/wl.c ret = __erase_worker(ubi, wl_wrk); wl_wrk 1211 drivers/mtd/ubi/wl.c kfree(wl_wrk);