Lines Matching refs:aeb
122 struct ubi_ainf_peb *aeb; in add_to_list() local
134 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in add_to_list()
135 if (!aeb) in add_to_list()
138 aeb->pnum = pnum; in add_to_list()
139 aeb->vol_id = vol_id; in add_to_list()
140 aeb->lnum = lnum; in add_to_list()
141 aeb->ec = ec; in add_to_list()
143 list_add(&aeb->u.list, list); in add_to_list()
145 list_add_tail(&aeb->u.list, list); in add_to_list()
162 struct ubi_ainf_peb *aeb; in add_corrupted() local
166 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in add_corrupted()
167 if (!aeb) in add_corrupted()
171 aeb->pnum = pnum; in add_corrupted()
172 aeb->ec = ec; in add_corrupted()
173 list_add(&aeb->u.list, &ai->corr); in add_corrupted()
324 int ubi_compare_lebs(struct ubi_device *ubi, const struct ubi_ainf_peb *aeb, in ubi_compare_lebs() argument
332 if (sqnum2 == aeb->sqnum) { in ubi_compare_lebs()
346 second_is_newer = (sqnum2 > aeb->sqnum); in ubi_compare_lebs()
365 if (!aeb->copy_flag) { in ubi_compare_lebs()
376 pnum = aeb->pnum; in ubi_compare_lebs()
455 struct ubi_ainf_peb *aeb; in ubi_add_to_av() local
481 aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); in ubi_add_to_av()
482 if (lnum != aeb->lnum) { in ubi_add_to_av()
483 if (lnum < aeb->lnum) in ubi_add_to_av()
496 aeb->pnum, aeb->sqnum, aeb->ec); in ubi_add_to_av()
511 if (aeb->sqnum == sqnum && sqnum != 0) { in ubi_add_to_av()
514 ubi_dump_aeb(aeb, 0); in ubi_add_to_av()
523 cmp_res = ubi_compare_lebs(ubi, aeb, pnum, vid_hdr); in ubi_add_to_av()
536 err = add_to_list(ai, aeb->pnum, aeb->vol_id, in ubi_add_to_av()
537 aeb->lnum, aeb->ec, cmp_res & 4, in ubi_add_to_av()
542 aeb->ec = ec; in ubi_add_to_av()
543 aeb->pnum = pnum; in ubi_add_to_av()
544 aeb->vol_id = vol_id; in ubi_add_to_av()
545 aeb->lnum = lnum; in ubi_add_to_av()
546 aeb->scrub = ((cmp_res & 2) || bitflips); in ubi_add_to_av()
547 aeb->copy_flag = vid_hdr->copy_flag; in ubi_add_to_av()
548 aeb->sqnum = sqnum; in ubi_add_to_av()
574 aeb = kmem_cache_alloc(ai->aeb_slab_cache, GFP_KERNEL); in ubi_add_to_av()
575 if (!aeb) in ubi_add_to_av()
578 aeb->ec = ec; in ubi_add_to_av()
579 aeb->pnum = pnum; in ubi_add_to_av()
580 aeb->vol_id = vol_id; in ubi_add_to_av()
581 aeb->lnum = lnum; in ubi_add_to_av()
582 aeb->scrub = bitflips; in ubi_add_to_av()
583 aeb->copy_flag = vid_hdr->copy_flag; in ubi_add_to_av()
584 aeb->sqnum = sqnum; in ubi_add_to_av()
592 rb_link_node(&aeb->u.rb, parent, p); in ubi_add_to_av()
593 rb_insert_color(&aeb->u.rb, &av->root); in ubi_add_to_av()
634 struct ubi_ainf_peb *aeb; in ubi_remove_av() local
639 aeb = rb_entry(rb, struct ubi_ainf_peb, u.rb); in ubi_remove_av()
640 rb_erase(&aeb->u.rb, &av->root); in ubi_remove_av()
641 list_add_tail(&aeb->u.list, &ai->erase); in ubi_remove_av()
713 struct ubi_ainf_peb *aeb, *tmp_aeb; in ubi_early_get_peb() local
716 aeb = list_entry(ai->free.next, struct ubi_ainf_peb, u.list); in ubi_early_get_peb()
717 list_del(&aeb->u.list); in ubi_early_get_peb()
718 dbg_bld("return free PEB %d, EC %d", aeb->pnum, aeb->ec); in ubi_early_get_peb()
719 return aeb; in ubi_early_get_peb()
728 list_for_each_entry_safe(aeb, tmp_aeb, &ai->erase, u.list) { in ubi_early_get_peb()
729 if (aeb->ec == UBI_UNKNOWN) in ubi_early_get_peb()
730 aeb->ec = ai->mean_ec; in ubi_early_get_peb()
732 err = early_erase_peb(ubi, ai, aeb->pnum, aeb->ec+1); in ubi_early_get_peb()
736 aeb->ec += 1; in ubi_early_get_peb()
737 list_del(&aeb->u.list); in ubi_early_get_peb()
738 dbg_bld("return PEB %d, EC %d", aeb->pnum, aeb->ec); in ubi_early_get_peb()
739 return aeb; in ubi_early_get_peb()
1070 struct ubi_ainf_peb *aeb; in late_analysis() local
1085 list_for_each_entry(aeb, &ai->corr, u.list) in late_analysis()
1086 pr_cont(" %d", aeb->pnum); in late_analysis()
1139 struct ubi_ainf_peb *aeb; in destroy_av() local
1148 aeb = rb_entry(this, struct ubi_ainf_peb, u.rb); in destroy_av()
1151 if (this->rb_left == &aeb->u.rb) in destroy_av()
1157 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_av()
1169 struct ubi_ainf_peb *aeb, *aeb_tmp; in destroy_ai() local
1173 list_for_each_entry_safe(aeb, aeb_tmp, &ai->alien, u.list) { in destroy_ai()
1174 list_del(&aeb->u.list); in destroy_ai()
1175 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1177 list_for_each_entry_safe(aeb, aeb_tmp, &ai->erase, u.list) { in destroy_ai()
1178 list_del(&aeb->u.list); in destroy_ai()
1179 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1181 list_for_each_entry_safe(aeb, aeb_tmp, &ai->corr, u.list) { in destroy_ai()
1182 list_del(&aeb->u.list); in destroy_ai()
1183 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1185 list_for_each_entry_safe(aeb, aeb_tmp, &ai->free, u.list) { in destroy_ai()
1186 list_del(&aeb->u.list); in destroy_ai()
1187 kmem_cache_free(ai->aeb_slab_cache, aeb); in destroy_ai()
1234 struct ubi_ainf_peb *aeb; in scan_all() local
1270 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) in scan_all()
1271 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1272 aeb->ec = ai->mean_ec; in scan_all()
1275 list_for_each_entry(aeb, &ai->free, u.list) { in scan_all()
1276 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1277 aeb->ec = ai->mean_ec; in scan_all()
1280 list_for_each_entry(aeb, &ai->corr, u.list) in scan_all()
1281 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1282 aeb->ec = ai->mean_ec; in scan_all()
1284 list_for_each_entry(aeb, &ai->erase, u.list) in scan_all()
1285 if (aeb->ec == UBI_UNKNOWN) in scan_all()
1286 aeb->ec = ai->mean_ec; in scan_all()
1510 struct ubi_ainf_peb *aeb, *last_aeb; in self_check_ai() local
1562 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) { in self_check_ai()
1565 last_aeb = aeb; in self_check_ai()
1568 if (aeb->pnum < 0 || aeb->ec < 0) { in self_check_ai()
1573 if (aeb->ec < ai->min_ec) { in self_check_ai()
1575 ai->min_ec, aeb->ec); in self_check_ai()
1579 if (aeb->ec > ai->max_ec) { in self_check_ai()
1581 ai->max_ec, aeb->ec); in self_check_ai()
1585 if (aeb->pnum >= ubi->peb_count) { in self_check_ai()
1587 aeb->pnum, ubi->peb_count); in self_check_ai()
1592 if (aeb->lnum >= av->used_ebs) { in self_check_ai()
1603 if (aeb->lnum > av->highest_lnum) { in self_check_ai()
1618 aeb = last_aeb; in self_check_ai()
1620 if (aeb->lnum != av->highest_lnum) { in self_check_ai()
1635 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) { in self_check_ai()
1640 last_aeb = aeb; in self_check_ai()
1642 err = ubi_io_read_vid_hdr(ubi, aeb->pnum, vidh, 1); in self_check_ai()
1658 if (aeb->sqnum != be64_to_cpu(vidh->sqnum)) { in self_check_ai()
1659 ubi_err(ubi, "bad sqnum %llu", aeb->sqnum); in self_check_ai()
1673 if (aeb->lnum != be32_to_cpu(vidh->lnum)) { in self_check_ai()
1674 ubi_err(ubi, "bad lnum %d", aeb->lnum); in self_check_ai()
1722 ubi_rb_for_each_entry(rb2, aeb, &av->root, u.rb) in self_check_ai()
1723 buf[aeb->pnum] = 1; in self_check_ai()
1725 list_for_each_entry(aeb, &ai->free, u.list) in self_check_ai()
1726 buf[aeb->pnum] = 1; in self_check_ai()
1728 list_for_each_entry(aeb, &ai->corr, u.list) in self_check_ai()
1729 buf[aeb->pnum] = 1; in self_check_ai()
1731 list_for_each_entry(aeb, &ai->erase, u.list) in self_check_ai()
1732 buf[aeb->pnum] = 1; in self_check_ai()
1734 list_for_each_entry(aeb, &ai->alien, u.list) in self_check_ai()
1735 buf[aeb->pnum] = 1; in self_check_ai()
1750 ubi_err(ubi, "bad attaching information about LEB %d", aeb->lnum); in self_check_ai()
1751 ubi_dump_aeb(aeb, 0); in self_check_ai()