/linux-4.1.27/sound/pci/ctxfi/ |
H A D | cthardware.h | 90 int (*src_rsc_put_ctrl_blk)(void *blk); 91 int (*src_set_state)(void *blk, unsigned int state); 92 int (*src_set_bm)(void *blk, unsigned int bm); 93 int (*src_set_rsr)(void *blk, unsigned int rsr); 94 int (*src_set_sf)(void *blk, unsigned int sf); 95 int (*src_set_wr)(void *blk, unsigned int wr); 96 int (*src_set_pm)(void *blk, unsigned int pm); 97 int (*src_set_rom)(void *blk, unsigned int rom); 98 int (*src_set_vo)(void *blk, unsigned int vo); 99 int (*src_set_st)(void *blk, unsigned int st); 100 int (*src_set_ie)(void *blk, unsigned int ie); 101 int (*src_set_ilsz)(void *blk, unsigned int ilsz); 102 int (*src_set_bp)(void *blk, unsigned int bp); 103 int (*src_set_cisz)(void *blk, unsigned int cisz); 104 int (*src_set_ca)(void *blk, unsigned int ca); 105 int (*src_set_sa)(void *blk, unsigned int sa); 106 int (*src_set_la)(void *blk, unsigned int la); 107 int (*src_set_pitch)(void *blk, unsigned int pitch); 108 int (*src_set_clear_zbufs)(void *blk, unsigned int clear); 109 int (*src_set_dirty)(void *blk, unsigned int flags); 110 int (*src_set_dirty_all)(void *blk); 111 int (*src_commit_write)(struct hw *hw, unsigned int idx, void *blk); 112 int (*src_get_ca)(struct hw *hw, unsigned int idx, void *blk); 113 unsigned int (*src_get_dirty)(void *blk); 116 int (*src_mgr_put_ctrl_blk)(void *blk); 118 int (*src_mgr_enbs_src)(void *blk, unsigned int idx); 120 int (*src_mgr_enb_src)(void *blk, unsigned int idx); 122 int (*src_mgr_dsb_src)(void *blk, unsigned int idx); 123 int (*src_mgr_commit_write)(struct hw *hw, void *blk); 127 int (*srcimp_mgr_put_ctrl_blk)(void *blk); 128 int (*srcimp_mgr_set_imaparc)(void *blk, unsigned int slot); 129 int (*srcimp_mgr_set_imapuser)(void *blk, unsigned int user); 130 int (*srcimp_mgr_set_imapnxt)(void *blk, unsigned int next); 131 int (*srcimp_mgr_set_imapaddr)(void *blk, unsigned int addr); 132 int (*srcimp_mgr_commit_write)(struct hw *hw, void *blk); 136 int (*amixer_rsc_put_ctrl_blk)(void *blk); 138 int (*amixer_mgr_put_ctrl_blk)(void *blk); 139 int (*amixer_set_mode)(void *blk, unsigned int mode); 140 int (*amixer_set_iv)(void *blk, unsigned int iv); 141 int (*amixer_set_x)(void *blk, unsigned int x); 142 int (*amixer_set_y)(void *blk, unsigned int y); 143 int (*amixer_set_sadr)(void *blk, unsigned int sadr); 144 int (*amixer_set_se)(void *blk, unsigned int se); 145 int (*amixer_set_dirty)(void *blk, unsigned int flags); 146 int (*amixer_set_dirty_all)(void *blk); 147 int (*amixer_commit_write)(struct hw *hw, unsigned int idx, void *blk); 148 int (*amixer_get_y)(void *blk); 149 unsigned int (*amixer_get_dirty)(void *blk); 153 int (*dai_put_ctrl_blk)(void *blk); 154 int (*dai_srt_set_srco)(void *blk, unsigned int src); 155 int (*dai_srt_set_srcm)(void *blk, unsigned int src); 156 int (*dai_srt_set_rsr)(void *blk, unsigned int rsr); 157 int (*dai_srt_set_drat)(void *blk, unsigned int drat); 158 int (*dai_srt_set_ec)(void *blk, unsigned int ec); 159 int (*dai_srt_set_et)(void *blk, unsigned int et); 160 int (*dai_commit_write)(struct hw *hw, unsigned int idx, void *blk); 162 int (*dao_put_ctrl_blk)(void *blk); 163 int (*dao_set_spos)(void *blk, unsigned int spos); 164 int (*dao_commit_write)(struct hw *hw, unsigned int idx, void *blk); 165 int (*dao_get_spos)(void *blk, unsigned int *spos); 168 int (*daio_mgr_put_ctrl_blk)(void *blk); 169 int (*daio_mgr_enb_dai)(void *blk, unsigned int idx); 170 int (*daio_mgr_dsb_dai)(void *blk, unsigned int idx); 171 int (*daio_mgr_enb_dao)(void *blk, unsigned int idx); 172 int (*daio_mgr_dsb_dao)(void *blk, unsigned int idx); 173 int (*daio_mgr_dao_init)(void *blk, unsigned int idx, 175 int (*daio_mgr_set_imaparc)(void *blk, unsigned int slot); 176 int (*daio_mgr_set_imapnxt)(void *blk, unsigned int next); 177 int (*daio_mgr_set_imapaddr)(void *blk, unsigned int addr); 178 int (*daio_mgr_commit_write)(struct hw *hw, void *blk);
|
H A D | cthw20k1.c | 167 struct src_rsc_ctrl_blk *blk; src_get_rsc_ctrl_blk() local 170 blk = kzalloc(sizeof(*blk), GFP_KERNEL); src_get_rsc_ctrl_blk() 171 if (!blk) src_get_rsc_ctrl_blk() 174 *rblk = blk; src_get_rsc_ctrl_blk() 179 static int src_put_rsc_ctrl_blk(void *blk) src_put_rsc_ctrl_blk() argument 181 kfree((struct src_rsc_ctrl_blk *)blk); src_put_rsc_ctrl_blk() 186 static int src_set_state(void *blk, unsigned int state) src_set_state() argument 188 struct src_rsc_ctrl_blk *ctl = blk; src_set_state() 195 static int src_set_bm(void *blk, unsigned int bm) src_set_bm() argument 197 struct src_rsc_ctrl_blk *ctl = blk; src_set_bm() 204 static int src_set_rsr(void *blk, unsigned int rsr) src_set_rsr() argument 206 struct src_rsc_ctrl_blk *ctl = blk; src_set_rsr() 213 static int src_set_sf(void *blk, unsigned int sf) src_set_sf() argument 215 struct src_rsc_ctrl_blk *ctl = blk; src_set_sf() 222 static int src_set_wr(void *blk, unsigned int wr) src_set_wr() argument 224 struct src_rsc_ctrl_blk *ctl = blk; src_set_wr() 231 static int src_set_pm(void *blk, unsigned int pm) src_set_pm() argument 233 struct src_rsc_ctrl_blk *ctl = blk; src_set_pm() 240 static int src_set_rom(void *blk, unsigned int rom) src_set_rom() argument 242 struct src_rsc_ctrl_blk *ctl = blk; src_set_rom() 249 static int src_set_vo(void *blk, unsigned int vo) src_set_vo() argument 251 struct src_rsc_ctrl_blk *ctl = blk; src_set_vo() 258 static int src_set_st(void *blk, unsigned int st) src_set_st() argument 260 struct src_rsc_ctrl_blk *ctl = blk; src_set_st() 267 static int src_set_ie(void *blk, unsigned int ie) src_set_ie() argument 269 struct src_rsc_ctrl_blk *ctl = blk; src_set_ie() 276 static int src_set_ilsz(void *blk, unsigned int ilsz) src_set_ilsz() argument 278 struct src_rsc_ctrl_blk *ctl = blk; src_set_ilsz() 285 static int src_set_bp(void *blk, unsigned int bp) src_set_bp() argument 287 struct src_rsc_ctrl_blk *ctl = blk; src_set_bp() 294 static int src_set_cisz(void *blk, unsigned int cisz) src_set_cisz() argument 296 struct src_rsc_ctrl_blk *ctl = blk; src_set_cisz() 303 static int src_set_ca(void *blk, unsigned int ca) src_set_ca() argument 305 struct src_rsc_ctrl_blk *ctl = blk; src_set_ca() 312 static int src_set_sa(void *blk, unsigned int sa) src_set_sa() argument 314 struct src_rsc_ctrl_blk *ctl = blk; src_set_sa() 321 static int src_set_la(void *blk, unsigned int la) src_set_la() argument 323 struct src_rsc_ctrl_blk *ctl = blk; src_set_la() 330 static int src_set_pitch(void *blk, unsigned int pitch) src_set_pitch() argument 332 struct src_rsc_ctrl_blk *ctl = blk; src_set_pitch() 339 static int src_set_clear_zbufs(void *blk, unsigned int clear) src_set_clear_zbufs() argument 341 ((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0); src_set_clear_zbufs() 345 static int src_set_dirty(void *blk, unsigned int flags) src_set_dirty() argument 347 ((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); src_set_dirty() 351 static int src_set_dirty_all(void *blk) src_set_dirty_all() argument 353 ((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); src_set_dirty_all() 369 static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) src_commit_write() argument 371 struct src_rsc_ctrl_blk *ctl = blk; src_commit_write() 426 static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) src_get_ca() argument 428 struct src_rsc_ctrl_blk *ctl = blk; src_get_ca() 436 static unsigned int src_get_dirty(void *blk) src_get_dirty() argument 438 return ((struct src_rsc_ctrl_blk *)blk)->dirty.data; src_get_dirty() 446 static int src_mgr_enbs_src(void *blk, unsigned int idx) src_mgr_enbs_src() argument 448 ((struct src_mgr_ctrl_blk *)blk)->enbsa = ~(0x0); src_mgr_enbs_src() 449 ((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1; src_mgr_enbs_src() 450 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); src_mgr_enbs_src() 454 static int src_mgr_enb_src(void *blk, unsigned int idx) src_mgr_enb_src() argument 456 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); src_mgr_enb_src() 457 ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); src_mgr_enb_src() 461 static int src_mgr_dsb_src(void *blk, unsigned int idx) src_mgr_dsb_src() argument 463 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); src_mgr_dsb_src() 464 ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); src_mgr_dsb_src() 468 static int src_mgr_commit_write(struct hw *hw, void *blk) src_mgr_commit_write() argument 470 struct src_mgr_ctrl_blk *ctl = blk; src_mgr_commit_write() 493 struct src_mgr_ctrl_blk *blk; src_mgr_get_ctrl_blk() local 496 blk = kzalloc(sizeof(*blk), GFP_KERNEL); src_mgr_get_ctrl_blk() 497 if (!blk) src_mgr_get_ctrl_blk() 500 *rblk = blk; src_mgr_get_ctrl_blk() 505 static int src_mgr_put_ctrl_blk(void *blk) src_mgr_put_ctrl_blk() argument 507 kfree((struct src_mgr_ctrl_blk *)blk); src_mgr_put_ctrl_blk() 514 struct srcimp_mgr_ctrl_blk *blk; srcimp_mgr_get_ctrl_blk() local 517 blk = kzalloc(sizeof(*blk), GFP_KERNEL); srcimp_mgr_get_ctrl_blk() 518 if (!blk) srcimp_mgr_get_ctrl_blk() 521 *rblk = blk; srcimp_mgr_get_ctrl_blk() 526 static int srcimp_mgr_put_ctrl_blk(void *blk) srcimp_mgr_put_ctrl_blk() argument 528 kfree((struct srcimp_mgr_ctrl_blk *)blk); srcimp_mgr_put_ctrl_blk() 533 static int srcimp_mgr_set_imaparc(void *blk, unsigned int slot) srcimp_mgr_set_imaparc() argument 535 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imaparc() 542 static int srcimp_mgr_set_imapuser(void *blk, unsigned int user) srcimp_mgr_set_imapuser() argument 544 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imapuser() 551 static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) srcimp_mgr_set_imapnxt() argument 553 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imapnxt() 560 static int srcimp_mgr_set_imapaddr(void *blk, unsigned int addr) srcimp_mgr_set_imapaddr() argument 562 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imapaddr() 569 static int srcimp_mgr_commit_write(struct hw *hw, void *blk) srcimp_mgr_commit_write() argument 571 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_commit_write() 610 static int amixer_set_mode(void *blk, unsigned int mode) amixer_set_mode() argument 612 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_mode() 619 static int amixer_set_iv(void *blk, unsigned int iv) amixer_set_iv() argument 625 static int amixer_set_x(void *blk, unsigned int x) amixer_set_x() argument 627 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_x() 634 static int amixer_set_y(void *blk, unsigned int y) amixer_set_y() argument 636 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_y() 643 static int amixer_set_sadr(void *blk, unsigned int sadr) amixer_set_sadr() argument 645 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_sadr() 652 static int amixer_set_se(void *blk, unsigned int se) amixer_set_se() argument 654 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_se() 661 static int amixer_set_dirty(void *blk, unsigned int flags) amixer_set_dirty() argument 663 ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); amixer_set_dirty() 667 static int amixer_set_dirty_all(void *blk) amixer_set_dirty_all() argument 669 ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); amixer_set_dirty_all() 673 static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) amixer_commit_write() argument 675 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_commit_write() 687 static int amixer_get_y(void *blk) amixer_get_y() argument 689 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_get_y() 694 static unsigned int amixer_get_dirty(void *blk) amixer_get_dirty() argument 696 return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data; amixer_get_dirty() 701 struct amixer_rsc_ctrl_blk *blk; amixer_rsc_get_ctrl_blk() local 704 blk = kzalloc(sizeof(*blk), GFP_KERNEL); amixer_rsc_get_ctrl_blk() 705 if (!blk) amixer_rsc_get_ctrl_blk() 708 *rblk = blk; amixer_rsc_get_ctrl_blk() 713 static int amixer_rsc_put_ctrl_blk(void *blk) amixer_rsc_put_ctrl_blk() argument 715 kfree((struct amixer_rsc_ctrl_blk *)blk); amixer_rsc_put_ctrl_blk() 722 /*amixer_mgr_ctrl_blk_t *blk;*/ amixer_mgr_get_ctrl_blk() 725 /*blk = kzalloc(sizeof(*blk), GFP_KERNEL); amixer_mgr_get_ctrl_blk() 726 if (!blk) amixer_mgr_get_ctrl_blk() 729 *rblk = blk;*/ amixer_mgr_get_ctrl_blk() 734 static int amixer_mgr_put_ctrl_blk(void *blk) amixer_mgr_put_ctrl_blk() argument 736 /*kfree((amixer_mgr_ctrl_blk_t *)blk);*/ amixer_mgr_put_ctrl_blk() 834 static int dai_srt_set_srcr(void *blk, unsigned int src) dai_srt_set_srcr() argument 836 struct dai_ctrl_blk *ctl = blk; dai_srt_set_srcr() 843 static int dai_srt_set_srcl(void *blk, unsigned int src) dai_srt_set_srcl() argument 845 struct dai_ctrl_blk *ctl = blk; dai_srt_set_srcl() 852 static int dai_srt_set_rsr(void *blk, unsigned int rsr) dai_srt_set_rsr() argument 854 struct dai_ctrl_blk *ctl = blk; dai_srt_set_rsr() 861 static int dai_srt_set_drat(void *blk, unsigned int drat) dai_srt_set_drat() argument 863 struct dai_ctrl_blk *ctl = blk; dai_srt_set_drat() 870 static int dai_srt_set_ec(void *blk, unsigned int ec) dai_srt_set_ec() argument 872 struct dai_ctrl_blk *ctl = blk; dai_srt_set_ec() 879 static int dai_srt_set_et(void *blk, unsigned int et) dai_srt_set_et() argument 881 struct dai_ctrl_blk *ctl = blk; dai_srt_set_et() 888 static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) dai_commit_write() argument 890 struct dai_ctrl_blk *ctl = blk; dai_commit_write() 908 struct dai_ctrl_blk *blk; dai_get_ctrl_blk() local 911 blk = kzalloc(sizeof(*blk), GFP_KERNEL); dai_get_ctrl_blk() 912 if (!blk) dai_get_ctrl_blk() 915 *rblk = blk; dai_get_ctrl_blk() 920 static int dai_put_ctrl_blk(void *blk) dai_put_ctrl_blk() argument 922 kfree((struct dai_ctrl_blk *)blk); dai_put_ctrl_blk() 927 static int dao_set_spos(void *blk, unsigned int spos) dao_set_spos() argument 929 ((struct dao_ctrl_blk *)blk)->spos = spos; dao_set_spos() 930 ((struct dao_ctrl_blk *)blk)->dirty.bf.spos = 1; dao_set_spos() 934 static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) dao_commit_write() argument 936 struct dao_ctrl_blk *ctl = blk; dao_commit_write() 949 static int dao_get_spos(void *blk, unsigned int *spos) dao_get_spos() argument 951 *spos = ((struct dao_ctrl_blk *)blk)->spos; dao_get_spos() 957 struct dao_ctrl_blk *blk; dao_get_ctrl_blk() local 960 blk = kzalloc(sizeof(*blk), GFP_KERNEL); dao_get_ctrl_blk() 961 if (!blk) dao_get_ctrl_blk() 964 *rblk = blk; dao_get_ctrl_blk() 969 static int dao_put_ctrl_blk(void *blk) dao_put_ctrl_blk() argument 971 kfree((struct dao_ctrl_blk *)blk); dao_put_ctrl_blk() 976 static int daio_mgr_enb_dai(void *blk, unsigned int idx) daio_mgr_enb_dai() argument 978 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_enb_dai() 993 static int daio_mgr_dsb_dai(void *blk, unsigned int idx) daio_mgr_dsb_dai() argument 995 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dsb_dai() 1010 static int daio_mgr_enb_dao(void *blk, unsigned int idx) daio_mgr_enb_dao() argument 1012 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_enb_dao() 1027 static int daio_mgr_dsb_dao(void *blk, unsigned int idx) daio_mgr_dsb_dao() argument 1029 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dsb_dao() 1044 static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) daio_mgr_dao_init() argument 1046 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dao_init() 1081 static int daio_mgr_set_imaparc(void *blk, unsigned int slot) daio_mgr_set_imaparc() argument 1083 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_set_imaparc() 1090 static int daio_mgr_set_imapnxt(void *blk, unsigned int next) daio_mgr_set_imapnxt() argument 1092 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_set_imapnxt() 1099 static int daio_mgr_set_imapaddr(void *blk, unsigned int addr) daio_mgr_set_imapaddr() argument 1101 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_set_imapaddr() 1108 static int daio_mgr_commit_write(struct hw *hw, void *blk) daio_mgr_commit_write() argument 1110 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_commit_write() 1151 struct daio_mgr_ctrl_blk *blk; daio_mgr_get_ctrl_blk() local 1154 blk = kzalloc(sizeof(*blk), GFP_KERNEL); daio_mgr_get_ctrl_blk() 1155 if (!blk) daio_mgr_get_ctrl_blk() 1158 blk->i2sctl = hw_read_20kx(hw, I2SCTL); daio_mgr_get_ctrl_blk() 1159 blk->spoctl = hw_read_20kx(hw, SPOCTL); daio_mgr_get_ctrl_blk() 1160 blk->spictl = hw_read_20kx(hw, SPICTL); daio_mgr_get_ctrl_blk() 1162 *rblk = blk; daio_mgr_get_ctrl_blk() 1167 static int daio_mgr_put_ctrl_blk(void *blk) daio_mgr_put_ctrl_blk() argument 1169 kfree((struct daio_mgr_ctrl_blk *)blk); daio_mgr_put_ctrl_blk()
|
H A D | cthw20k2.c | 167 struct src_rsc_ctrl_blk *blk; src_get_rsc_ctrl_blk() local 170 blk = kzalloc(sizeof(*blk), GFP_KERNEL); src_get_rsc_ctrl_blk() 171 if (!blk) src_get_rsc_ctrl_blk() 174 *rblk = blk; src_get_rsc_ctrl_blk() 179 static int src_put_rsc_ctrl_blk(void *blk) src_put_rsc_ctrl_blk() argument 181 kfree(blk); src_put_rsc_ctrl_blk() 186 static int src_set_state(void *blk, unsigned int state) src_set_state() argument 188 struct src_rsc_ctrl_blk *ctl = blk; src_set_state() 195 static int src_set_bm(void *blk, unsigned int bm) src_set_bm() argument 197 struct src_rsc_ctrl_blk *ctl = blk; src_set_bm() 204 static int src_set_rsr(void *blk, unsigned int rsr) src_set_rsr() argument 206 struct src_rsc_ctrl_blk *ctl = blk; src_set_rsr() 213 static int src_set_sf(void *blk, unsigned int sf) src_set_sf() argument 215 struct src_rsc_ctrl_blk *ctl = blk; src_set_sf() 222 static int src_set_wr(void *blk, unsigned int wr) src_set_wr() argument 224 struct src_rsc_ctrl_blk *ctl = blk; src_set_wr() 231 static int src_set_pm(void *blk, unsigned int pm) src_set_pm() argument 233 struct src_rsc_ctrl_blk *ctl = blk; src_set_pm() 240 static int src_set_rom(void *blk, unsigned int rom) src_set_rom() argument 242 struct src_rsc_ctrl_blk *ctl = blk; src_set_rom() 249 static int src_set_vo(void *blk, unsigned int vo) src_set_vo() argument 251 struct src_rsc_ctrl_blk *ctl = blk; src_set_vo() 258 static int src_set_st(void *blk, unsigned int st) src_set_st() argument 260 struct src_rsc_ctrl_blk *ctl = blk; src_set_st() 267 static int src_set_ie(void *blk, unsigned int ie) src_set_ie() argument 269 struct src_rsc_ctrl_blk *ctl = blk; src_set_ie() 276 static int src_set_ilsz(void *blk, unsigned int ilsz) src_set_ilsz() argument 278 struct src_rsc_ctrl_blk *ctl = blk; src_set_ilsz() 285 static int src_set_bp(void *blk, unsigned int bp) src_set_bp() argument 287 struct src_rsc_ctrl_blk *ctl = blk; src_set_bp() 294 static int src_set_cisz(void *blk, unsigned int cisz) src_set_cisz() argument 296 struct src_rsc_ctrl_blk *ctl = blk; src_set_cisz() 303 static int src_set_ca(void *blk, unsigned int ca) src_set_ca() argument 305 struct src_rsc_ctrl_blk *ctl = blk; src_set_ca() 312 static int src_set_sa(void *blk, unsigned int sa) src_set_sa() argument 314 struct src_rsc_ctrl_blk *ctl = blk; src_set_sa() 321 static int src_set_la(void *blk, unsigned int la) src_set_la() argument 323 struct src_rsc_ctrl_blk *ctl = blk; src_set_la() 330 static int src_set_pitch(void *blk, unsigned int pitch) src_set_pitch() argument 332 struct src_rsc_ctrl_blk *ctl = blk; src_set_pitch() 339 static int src_set_clear_zbufs(void *blk, unsigned int clear) src_set_clear_zbufs() argument 341 ((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0); src_set_clear_zbufs() 345 static int src_set_dirty(void *blk, unsigned int flags) src_set_dirty() argument 347 ((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); src_set_dirty() 351 static int src_set_dirty_all(void *blk) src_set_dirty_all() argument 353 ((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); src_set_dirty_all() 369 static int src_commit_write(struct hw *hw, unsigned int idx, void *blk) src_commit_write() argument 371 struct src_rsc_ctrl_blk *ctl = blk; src_commit_write() 426 static int src_get_ca(struct hw *hw, unsigned int idx, void *blk) src_get_ca() argument 428 struct src_rsc_ctrl_blk *ctl = blk; src_get_ca() 436 static unsigned int src_get_dirty(void *blk) src_get_dirty() argument 438 return ((struct src_rsc_ctrl_blk *)blk)->dirty.data; src_get_dirty() 446 static int src_mgr_enbs_src(void *blk, unsigned int idx) src_mgr_enbs_src() argument 448 ((struct src_mgr_ctrl_blk *)blk)->enbsa |= (0x1 << ((idx%128)/4)); src_mgr_enbs_src() 449 ((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1; src_mgr_enbs_src() 450 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); src_mgr_enbs_src() 454 static int src_mgr_enb_src(void *blk, unsigned int idx) src_mgr_enb_src() argument 456 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] |= (0x1 << (idx%32)); src_mgr_enb_src() 457 ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); src_mgr_enb_src() 461 static int src_mgr_dsb_src(void *blk, unsigned int idx) src_mgr_dsb_src() argument 463 ((struct src_mgr_ctrl_blk *)blk)->enb[idx/32] &= ~(0x1 << (idx%32)); src_mgr_dsb_src() 464 ((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32)); src_mgr_dsb_src() 468 static int src_mgr_commit_write(struct hw *hw, void *blk) src_mgr_commit_write() argument 470 struct src_mgr_ctrl_blk *ctl = blk; src_mgr_commit_write() 493 struct src_mgr_ctrl_blk *blk; src_mgr_get_ctrl_blk() local 496 blk = kzalloc(sizeof(*blk), GFP_KERNEL); src_mgr_get_ctrl_blk() 497 if (!blk) src_mgr_get_ctrl_blk() 500 *rblk = blk; src_mgr_get_ctrl_blk() 505 static int src_mgr_put_ctrl_blk(void *blk) src_mgr_put_ctrl_blk() argument 507 kfree(blk); src_mgr_put_ctrl_blk() 514 struct srcimp_mgr_ctrl_blk *blk; srcimp_mgr_get_ctrl_blk() local 517 blk = kzalloc(sizeof(*blk), GFP_KERNEL); srcimp_mgr_get_ctrl_blk() 518 if (!blk) srcimp_mgr_get_ctrl_blk() 521 *rblk = blk; srcimp_mgr_get_ctrl_blk() 526 static int srcimp_mgr_put_ctrl_blk(void *blk) srcimp_mgr_put_ctrl_blk() argument 528 kfree(blk); srcimp_mgr_put_ctrl_blk() 533 static int srcimp_mgr_set_imaparc(void *blk, unsigned int slot) srcimp_mgr_set_imaparc() argument 535 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imaparc() 542 static int srcimp_mgr_set_imapuser(void *blk, unsigned int user) srcimp_mgr_set_imapuser() argument 544 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imapuser() 551 static int srcimp_mgr_set_imapnxt(void *blk, unsigned int next) srcimp_mgr_set_imapnxt() argument 553 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_set_imapnxt() 560 static int srcimp_mgr_set_imapaddr(void *blk, unsigned int addr) srcimp_mgr_set_imapaddr() argument 562 ((struct srcimp_mgr_ctrl_blk *)blk)->srcimap.idx = addr; srcimp_mgr_set_imapaddr() 563 ((struct srcimp_mgr_ctrl_blk *)blk)->dirty.bf.srcimap = 1; srcimp_mgr_set_imapaddr() 567 static int srcimp_mgr_commit_write(struct hw *hw, void *blk) srcimp_mgr_commit_write() argument 569 struct srcimp_mgr_ctrl_blk *ctl = blk; srcimp_mgr_commit_write() 609 static int amixer_set_mode(void *blk, unsigned int mode) amixer_set_mode() argument 611 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_mode() 618 static int amixer_set_iv(void *blk, unsigned int iv) amixer_set_iv() argument 620 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_iv() 627 static int amixer_set_x(void *blk, unsigned int x) amixer_set_x() argument 629 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_x() 636 static int amixer_set_y(void *blk, unsigned int y) amixer_set_y() argument 638 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_y() 645 static int amixer_set_sadr(void *blk, unsigned int sadr) amixer_set_sadr() argument 647 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_sadr() 654 static int amixer_set_se(void *blk, unsigned int se) amixer_set_se() argument 656 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_set_se() 663 static int amixer_set_dirty(void *blk, unsigned int flags) amixer_set_dirty() argument 665 ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff); amixer_set_dirty() 669 static int amixer_set_dirty_all(void *blk) amixer_set_dirty_all() argument 671 ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0); amixer_set_dirty_all() 675 static int amixer_commit_write(struct hw *hw, unsigned int idx, void *blk) amixer_commit_write() argument 677 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_commit_write() 689 static int amixer_get_y(void *blk) amixer_get_y() argument 691 struct amixer_rsc_ctrl_blk *ctl = blk; amixer_get_y() 696 static unsigned int amixer_get_dirty(void *blk) amixer_get_dirty() argument 698 return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data; amixer_get_dirty() 703 struct amixer_rsc_ctrl_blk *blk; amixer_rsc_get_ctrl_blk() local 706 blk = kzalloc(sizeof(*blk), GFP_KERNEL); amixer_rsc_get_ctrl_blk() 707 if (!blk) amixer_rsc_get_ctrl_blk() 710 *rblk = blk; amixer_rsc_get_ctrl_blk() 715 static int amixer_rsc_put_ctrl_blk(void *blk) amixer_rsc_put_ctrl_blk() argument 717 kfree(blk); amixer_rsc_put_ctrl_blk() 729 static int amixer_mgr_put_ctrl_blk(void *blk) amixer_mgr_put_ctrl_blk() argument 824 static int dai_srt_set_srco(void *blk, unsigned int src) dai_srt_set_srco() argument 826 struct dai_ctrl_blk *ctl = blk; dai_srt_set_srco() 833 static int dai_srt_set_srcm(void *blk, unsigned int src) dai_srt_set_srcm() argument 835 struct dai_ctrl_blk *ctl = blk; dai_srt_set_srcm() 842 static int dai_srt_set_rsr(void *blk, unsigned int rsr) dai_srt_set_rsr() argument 844 struct dai_ctrl_blk *ctl = blk; dai_srt_set_rsr() 851 static int dai_srt_set_drat(void *blk, unsigned int drat) dai_srt_set_drat() argument 853 struct dai_ctrl_blk *ctl = blk; dai_srt_set_drat() 860 static int dai_srt_set_ec(void *blk, unsigned int ec) dai_srt_set_ec() argument 862 struct dai_ctrl_blk *ctl = blk; dai_srt_set_ec() 869 static int dai_srt_set_et(void *blk, unsigned int et) dai_srt_set_et() argument 871 struct dai_ctrl_blk *ctl = blk; dai_srt_set_et() 878 static int dai_commit_write(struct hw *hw, unsigned int idx, void *blk) dai_commit_write() argument 880 struct dai_ctrl_blk *ctl = blk; dai_commit_write() 892 struct dai_ctrl_blk *blk; dai_get_ctrl_blk() local 895 blk = kzalloc(sizeof(*blk), GFP_KERNEL); dai_get_ctrl_blk() 896 if (!blk) dai_get_ctrl_blk() 899 *rblk = blk; dai_get_ctrl_blk() 904 static int dai_put_ctrl_blk(void *blk) dai_put_ctrl_blk() argument 906 kfree(blk); dai_put_ctrl_blk() 911 static int dao_set_spos(void *blk, unsigned int spos) dao_set_spos() argument 913 ((struct dao_ctrl_blk *)blk)->atxcsl = spos; dao_set_spos() 914 ((struct dao_ctrl_blk *)blk)->dirty.bf.atxcsl = 1; dao_set_spos() 918 static int dao_commit_write(struct hw *hw, unsigned int idx, void *blk) dao_commit_write() argument 920 struct dao_ctrl_blk *ctl = blk; dao_commit_write() 934 static int dao_get_spos(void *blk, unsigned int *spos) dao_get_spos() argument 936 *spos = ((struct dao_ctrl_blk *)blk)->atxcsl; dao_get_spos() 942 struct dao_ctrl_blk *blk; dao_get_ctrl_blk() local 945 blk = kzalloc(sizeof(*blk), GFP_KERNEL); dao_get_ctrl_blk() 946 if (!blk) dao_get_ctrl_blk() 949 *rblk = blk; dao_get_ctrl_blk() 954 static int dao_put_ctrl_blk(void *blk) dao_put_ctrl_blk() argument 956 kfree(blk); dao_put_ctrl_blk() 961 static int daio_mgr_enb_dai(void *blk, unsigned int idx) daio_mgr_enb_dai() argument 963 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_enb_dai() 970 static int daio_mgr_dsb_dai(void *blk, unsigned int idx) daio_mgr_dsb_dai() argument 972 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dsb_dai() 980 static int daio_mgr_enb_dao(void *blk, unsigned int idx) daio_mgr_enb_dao() argument 982 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_enb_dao() 989 static int daio_mgr_dsb_dao(void *blk, unsigned int idx) daio_mgr_dsb_dao() argument 991 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dsb_dao() 998 static int daio_mgr_dao_init(void *blk, unsigned int idx, unsigned int conf) daio_mgr_dao_init() argument 1000 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_dao_init() 1036 static int daio_mgr_set_imaparc(void *blk, unsigned int slot) daio_mgr_set_imaparc() argument 1038 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_set_imaparc() 1045 static int daio_mgr_set_imapnxt(void *blk, unsigned int next) daio_mgr_set_imapnxt() argument 1047 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_set_imapnxt() 1054 static int daio_mgr_set_imapaddr(void *blk, unsigned int addr) daio_mgr_set_imapaddr() argument 1056 ((struct daio_mgr_ctrl_blk *)blk)->daoimap.idx = addr; daio_mgr_set_imapaddr() 1057 ((struct daio_mgr_ctrl_blk *)blk)->dirty.bf.daoimap = 1; daio_mgr_set_imapaddr() 1061 static int daio_mgr_commit_write(struct hw *hw, void *blk) daio_mgr_commit_write() argument 1063 struct daio_mgr_ctrl_blk *ctl = blk; daio_mgr_commit_write() 1092 struct daio_mgr_ctrl_blk *blk; daio_mgr_get_ctrl_blk() local 1096 blk = kzalloc(sizeof(*blk), GFP_KERNEL); daio_mgr_get_ctrl_blk() 1097 if (!blk) daio_mgr_get_ctrl_blk() 1101 blk->txctl[i] = hw_read_20kx(hw, AUDIO_IO_TX_CTL+(0x40*i)); daio_mgr_get_ctrl_blk() 1102 blk->rxctl[i] = hw_read_20kx(hw, AUDIO_IO_RX_CTL+(0x40*i)); daio_mgr_get_ctrl_blk() 1105 *rblk = blk; daio_mgr_get_ctrl_blk() 1110 static int daio_mgr_put_ctrl_blk(void *blk) daio_mgr_put_ctrl_blk() argument 1112 kfree(blk); daio_mgr_put_ctrl_blk()
|
/linux-4.1.27/sound/pci/emu10k1/ |
H A D | memory.c | 77 static int synth_alloc_pages(struct snd_emu10k1 *hw, struct snd_emu10k1_memblk *blk); 78 static int synth_free_pages(struct snd_emu10k1 *hw, struct snd_emu10k1_memblk *blk); 84 static void emu10k1_memblk_init(struct snd_emu10k1_memblk *blk) emu10k1_memblk_init() argument 86 blk->mapped_page = -1; emu10k1_memblk_init() 87 INIT_LIST_HEAD(&blk->mapped_link); emu10k1_memblk_init() 88 INIT_LIST_HEAD(&blk->mapped_order_link); emu10k1_memblk_init() 89 blk->map_locked = 0; emu10k1_memblk_init() 91 blk->first_page = get_aligned_page(blk->mem.offset); emu10k1_memblk_init() 92 blk->last_page = get_aligned_page(blk->mem.offset + blk->mem.size - 1); emu10k1_memblk_init() 93 blk->pages = blk->last_page - blk->first_page + 1; emu10k1_memblk_init() 112 struct snd_emu10k1_memblk *blk = get_emu10k1_memblk(pos, mapped_link); search_empty_map_area() local 113 if (blk->mapped_page < 0) search_empty_map_area() 115 size = blk->mapped_page - page; search_empty_map_area() 126 page = blk->mapped_page + blk->pages; search_empty_map_area() 142 static int map_memblk(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) map_memblk() argument 147 page = search_empty_map_area(emu, blk->pages, &next); map_memblk() 151 list_add_tail(&blk->mapped_link, next); map_memblk() 153 list_add_tail(&blk->mapped_order_link, &emu->mapped_order_link_head); map_memblk() 154 blk->mapped_page = page; map_memblk() 156 for (pg = blk->first_page; pg <= blk->last_page; pg++) { map_memblk() 169 static int unmap_memblk(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) unmap_memblk() argument 176 if ((p = blk->mapped_link.prev) != &emu->mapped_link_head) { unmap_memblk() 181 if ((p = blk->mapped_link.next) != &emu->mapped_link_head) { unmap_memblk() 188 list_del(&blk->mapped_link); unmap_memblk() 189 list_del(&blk->mapped_order_link); unmap_memblk() 191 mpage = blk->mapped_page; unmap_memblk() 192 for (pg = blk->first_page; pg <= blk->last_page; pg++) { unmap_memblk() 196 blk->mapped_page = -1; unmap_memblk() 209 struct snd_emu10k1_memblk *blk; search_empty() local 215 blk = get_emu10k1_memblk(p, mem.list); search_empty() 216 if (page + psize <= blk->first_page) search_empty() 218 page = blk->last_page + 1; search_empty() 225 blk = (struct snd_emu10k1_memblk *)__snd_util_memblk_new(emu->memhdr, psize << PAGE_SHIFT, p->prev); search_empty() 226 if (blk == NULL) search_empty() 228 blk->mem.offset = aligned_page_offset(page); /* set aligned offset */ search_empty() 229 emu10k1_memblk_init(blk); search_empty() 230 return blk; search_empty() 258 int snd_emu10k1_memblk_map(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) snd_emu10k1_memblk_map() argument 267 if (blk->mapped_page >= 0) { snd_emu10k1_memblk_map() 269 list_move_tail(&blk->mapped_order_link, snd_emu10k1_memblk_map() 274 if ((err = map_memblk(emu, blk)) < 0) { snd_emu10k1_memblk_map() 284 if (size >= blk->pages) { snd_emu10k1_memblk_map() 286 err = map_memblk(emu, blk); snd_emu10k1_memblk_map() 305 struct snd_emu10k1_memblk *blk; snd_emu10k1_alloc_pages() local 320 blk = search_empty(emu, runtime->dma_bytes + idx); snd_emu10k1_alloc_pages() 321 if (blk == NULL) { snd_emu10k1_alloc_pages() 329 for (page = blk->first_page; page <= blk->last_page; page++, idx++) { snd_emu10k1_alloc_pages() 347 blk->map_locked = 1; /* do not unmap this block! */ snd_emu10k1_alloc_pages() 348 err = snd_emu10k1_memblk_map(emu, blk); snd_emu10k1_alloc_pages() 350 __snd_util_mem_free(hdr, (struct snd_util_memblk *)blk); snd_emu10k1_alloc_pages() 355 return (struct snd_util_memblk *)blk; snd_emu10k1_alloc_pages() 362 int snd_emu10k1_free_pages(struct snd_emu10k1 *emu, struct snd_util_memblk *blk) snd_emu10k1_free_pages() argument 364 if (snd_BUG_ON(!emu || !blk)) snd_emu10k1_free_pages() 366 return snd_emu10k1_synth_free(emu, blk); snd_emu10k1_free_pages() 381 struct snd_emu10k1_memblk *blk; snd_emu10k1_synth_alloc() local 385 blk = (struct snd_emu10k1_memblk *)__snd_util_mem_alloc(hdr, size); snd_emu10k1_synth_alloc() 386 if (blk == NULL) { snd_emu10k1_synth_alloc() 390 if (synth_alloc_pages(hw, blk)) { snd_emu10k1_synth_alloc() 391 __snd_util_mem_free(hdr, (struct snd_util_memblk *)blk); snd_emu10k1_synth_alloc() 395 snd_emu10k1_memblk_map(hw, blk); snd_emu10k1_synth_alloc() 397 return (struct snd_util_memblk *)blk; snd_emu10k1_synth_alloc() 409 struct snd_emu10k1_memblk *blk = (struct snd_emu10k1_memblk *)memblk; snd_emu10k1_synth_free() local 414 if (blk->mapped_page >= 0) snd_emu10k1_synth_free() 415 unmap_memblk(emu, blk); snd_emu10k1_synth_free() 417 synth_free_pages(emu, blk); snd_emu10k1_synth_free() 427 struct snd_emu10k1_memblk *blk, get_single_page_range() 433 first_page = blk->first_page; get_single_page_range() 434 if ((p = blk->mem.list.prev) != &hdr->block) { get_single_page_range() 439 last_page = blk->last_page; get_single_page_range() 440 if ((p = blk->mem.list.next) != &hdr->block) { get_single_page_range() 465 static int synth_alloc_pages(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) synth_alloc_pages() argument 469 emu10k1_memblk_init(blk); synth_alloc_pages() 470 get_single_page_range(emu->memhdr, blk, &first_page, &last_page); synth_alloc_pages() 497 static int synth_free_pages(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk) synth_free_pages() argument 501 get_single_page_range(emu->memhdr, blk, &first_page, &last_page); synth_free_pages() 523 * bzero(blk + offset, size) 525 int snd_emu10k1_synth_bzero(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, snd_emu10k1_synth_bzero() argument 530 struct snd_emu10k1_memblk *p = (struct snd_emu10k1_memblk *)blk; snd_emu10k1_synth_bzero() 532 offset += blk->offset & (PAGE_SIZE - 1); snd_emu10k1_synth_bzero() 553 * copy_from_user(blk + offset, data, size) 555 int snd_emu10k1_synth_copy_from_user(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, snd_emu10k1_synth_copy_from_user() argument 560 struct snd_emu10k1_memblk *p = (struct snd_emu10k1_memblk *)blk; snd_emu10k1_synth_copy_from_user() 562 offset += blk->offset & (PAGE_SIZE - 1); snd_emu10k1_synth_copy_from_user() 426 get_single_page_range(struct snd_util_memhdr *hdr, struct snd_emu10k1_memblk *blk, int *first_page_ret, int *last_page_ret) get_single_page_range() argument
|
/linux-4.1.27/arch/powerpc/lib/ |
H A D | rheap.c | 47 rh_block_t *block, *blk; grow() local 73 for (i = 0, blk = block; i < info->max_blocks; i++, blk++) grow() 74 fixup(blks, blke, delta, &blk->list); grow() 91 blk = block + info->max_blocks - new_blocks; grow() 92 for (i = 0; i < new_blocks; i++, blk++) grow() 93 list_add(&blk->list, &info->empty_list); grow() 123 rh_block_t *blk; get_slot() local 133 blk = list_entry(info->empty_list.next, rh_block_t, list); get_slot() 134 list_del_init(&blk->list); get_slot() 138 blk->start = 0; get_slot() 139 blk->size = 0; get_slot() 140 blk->owner = NULL; get_slot() 142 return blk; get_slot() 145 static inline void release_slot(rh_info_t * info, rh_block_t * blk) release_slot() argument 147 list_add(&blk->list, &info->empty_list); release_slot() 153 rh_block_t *blk; attach_free_block() local 173 blk = list_entry(l, rh_block_t, list); attach_free_block() 175 bs = blk->start; attach_free_block() 176 be = bs + blk->size; attach_free_block() 179 next = blk; attach_free_block() 182 before = blk; attach_free_block() 185 after = blk; attach_free_block() 234 rh_block_t *blk; attach_taken_block() local 239 blk = list_entry(l, rh_block_t, list); attach_taken_block() 240 if (blk->start > blkn->start) { attach_taken_block() 241 list_add_tail(&blkn->list, &blk->list); attach_taken_block() 304 rh_block_t *blk; rh_init() local 323 for (i = 0, blk = block; i < max_blocks; i++, blk++) rh_init() 324 list_add(&blk->list, &info->empty_list); rh_init() 331 rh_block_t *blk; rh_attach_region() local 358 blk = get_slot(info); rh_attach_region() 359 blk->start = start; rh_attach_region() 360 blk->size = size; rh_attach_region() 361 blk->owner = NULL; rh_attach_region() 363 attach_free_block(info, blk); rh_attach_region() 373 rh_block_t *blk, *newblk; rh_detach_region() local 394 blk = NULL; rh_detach_region() 396 blk = list_entry(l, rh_block_t, list); rh_detach_region() 398 bs = blk->start; rh_detach_region() 399 be = blk->start + blk->size; rh_detach_region() 402 blk = NULL; rh_detach_region() 405 if (blk == NULL) rh_detach_region() 411 list_del(&blk->list); rh_detach_region() 412 release_slot(info, blk); rh_detach_region() 416 /* blk still in free list, with updated start and/or size */ rh_detach_region() 419 blk->start += size; rh_detach_region() 420 blk->size -= size; rh_detach_region() 424 blk->size = s - bs; rh_detach_region() 431 list_add(&newblk->list, &blk->list); rh_detach_region() 445 rh_block_t *blk; rh_alloc_align() local 459 blk = NULL; rh_alloc_align() 461 blk = list_entry(l, rh_block_t, list); rh_alloc_align() 462 if (size <= blk->size) { rh_alloc_align() 463 start = (blk->start + alignment - 1) & ~(alignment - 1); rh_alloc_align() 464 if (start + size <= blk->start + blk->size) rh_alloc_align() 467 blk = NULL; rh_alloc_align() 470 if (blk == NULL) rh_alloc_align() 474 if (blk->size == size) { rh_alloc_align() 476 list_del(&blk->list); rh_alloc_align() 477 newblk = blk; rh_alloc_align() 481 sp_size = start - blk->start; rh_alloc_align() 486 spblk->start = blk->start; rh_alloc_align() 488 /* add before the blk */ rh_alloc_align() 489 list_add(&spblk->list, blk->list.prev); rh_alloc_align() 495 /* blk still in free list, with updated start and size rh_alloc_align() 497 blk->start = start + size; rh_alloc_align() 498 blk->size -= sp_size + size; rh_alloc_align() 499 /* No fragment in the end, remove blk */ rh_alloc_align() 500 if (blk->size == 0) { rh_alloc_align() 501 list_del(&blk->list); rh_alloc_align() 502 release_slot(info, blk); rh_alloc_align() 530 rh_block_t *blk, *newblk1, *newblk2; rh_alloc_fixed() local 551 blk = NULL; rh_alloc_fixed() 553 blk = list_entry(l, rh_block_t, list); rh_alloc_fixed() 555 bs = blk->start; rh_alloc_fixed() 556 be = blk->start + blk->size; rh_alloc_fixed() 559 blk = NULL; rh_alloc_fixed() 562 if (blk == NULL) rh_alloc_fixed() 568 list_del(&blk->list); rh_alloc_fixed() 569 blk->owner = owner; rh_alloc_fixed() 571 start = blk->start; rh_alloc_fixed() 572 attach_taken_block(info, blk); rh_alloc_fixed() 578 /* blk still in free list, with updated start and/or size */ rh_alloc_fixed() 581 blk->start += size; rh_alloc_fixed() 582 blk->size -= size; rh_alloc_fixed() 586 blk->size = s - bs; rh_alloc_fixed() 593 list_add(&newblk2->list, &blk->list); rh_alloc_fixed() 614 rh_block_t *blk, *blk2; rh_free() local 619 blk = NULL; rh_free() 624 blk = blk2; rh_free() 627 if (blk == NULL || start > (blk->start + blk->size)) rh_free() 631 list_del(&blk->list); rh_free() 634 size = blk->size; rh_free() 635 attach_free_block(info, blk); rh_free() 643 rh_block_t *blk; rh_get_stats() local 665 blk = list_entry(l, rh_block_t, list); list_for_each() 667 stats->start = blk->start; list_for_each() 668 stats->size = blk->size; list_for_each() 669 stats->owner = blk->owner; list_for_each() 681 rh_block_t *blk, *blk2; rh_set_owner() local 686 blk = NULL; rh_set_owner() 691 blk = blk2; rh_set_owner() 694 if (blk == NULL || start > (blk->start + blk->size)) rh_set_owner() 697 blk->owner = owner; rh_set_owner() 698 size = blk->size; rh_set_owner() 740 void rh_dump_blk(rh_info_t * info, rh_block_t * blk) rh_dump_blk() argument 743 "blk @0x%p: 0x%lx-0x%lx (%u)\n", rh_dump_blk() 744 blk, blk->start, blk->start + blk->size, blk->size); rh_dump_blk()
|
/linux-4.1.27/sound/synth/ |
H A D | util_mem.c | 75 struct snd_util_memblk *blk; __snd_util_mem_alloc() local 92 blk = get_memblk(p); __snd_util_mem_alloc() 93 if (blk->offset - prev_offset >= units) __snd_util_mem_alloc() 95 prev_offset = blk->offset + blk->size; __snd_util_mem_alloc() 113 struct snd_util_memblk *blk; __snd_util_memblk_new() local 115 blk = kmalloc(sizeof(struct snd_util_memblk) + hdr->block_extra_size, __snd_util_memblk_new() 117 if (blk == NULL) __snd_util_memblk_new() 121 blk->offset = 0; __snd_util_memblk_new() 124 blk->offset = p->offset + p->size; __snd_util_memblk_new() 126 blk->size = units; __snd_util_memblk_new() 127 list_add(&blk->list, prev); __snd_util_memblk_new() 130 return blk; __snd_util_memblk_new() 140 struct snd_util_memblk *blk; snd_util_mem_alloc() local 142 blk = __snd_util_mem_alloc(hdr, size); snd_util_mem_alloc() 144 return blk; snd_util_mem_alloc() 153 __snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk) __snd_util_mem_free() argument 155 list_del(&blk->list); __snd_util_mem_free() 157 hdr->used -= blk->size; __snd_util_mem_free() 158 kfree(blk); __snd_util_mem_free() 164 int snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk) snd_util_mem_free() argument 166 if (snd_BUG_ON(!hdr || !blk)) snd_util_mem_free() 170 __snd_util_mem_free(hdr, blk); snd_util_mem_free()
|
/linux-4.1.27/block/partitions/ |
H A D | amiga.c | 34 int start_sect, nr_sects, blk, part, res = 0; amiga_partition() local 39 for (blk = 0; ; blk++, put_dev_sector(sect)) { amiga_partition() 40 if (blk == RDB_ALLOCATION_LIMIT) amiga_partition() 42 data = read_part_sector(state, blk, §); amiga_partition() 46 bdevname(state->bdev, b), blk); amiga_partition() 63 blk); amiga_partition() 68 bdevname(state->bdev, b), blk); amiga_partition() 81 blk = be32_to_cpu(rdb->rdb_PartitionList); amiga_partition() 83 for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) { amiga_partition() 84 blk *= blksize; /* Read in terms partition table understands */ amiga_partition() 85 data = read_part_sector(state, blk, §); amiga_partition() 89 bdevname(state->bdev, b), blk); amiga_partition() 94 blk = be32_to_cpu(pb->pb_Next); amiga_partition()
|
H A D | ibm.c | 138 sector_t blk; find_vol1_partitions() local 154 blk = cchhb2blk(&label->vol.vtoc, geo) + 1; find_vol1_partitions() 156 data = read_part_sector(state, blk * secperblk, §); find_vol1_partitions() 165 blk++; find_vol1_partitions() 166 data = read_part_sector(state, blk * secperblk, §); find_vol1_partitions() 183 blk++; find_vol1_partitions() 184 data = read_part_sector(state, blk * secperblk, §); find_vol1_partitions()
|
/linux-4.1.27/sound/pci/trident/ |
H A D | trident_memory.c | 133 #define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page) 134 #define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page) 142 struct snd_util_memblk *blk; search_empty() local 149 blk = list_entry(p, struct snd_util_memblk, list); search_empty() 150 if (page + psize <= firstpg(blk)) search_empty() 152 page = lastpg(blk) + 1; search_empty() 159 blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev); search_empty() 160 if (blk == NULL) search_empty() 162 blk->offset = aligned_page_offset(page); /* set aligned offset */ search_empty() 163 firstpg(blk) = page; search_empty() 164 lastpg(blk) = page + psize - 1; search_empty() 165 return blk; search_empty() 193 struct snd_util_memblk *blk; snd_trident_alloc_sg_pages() local 208 blk = search_empty(hdr, runtime->dma_bytes); snd_trident_alloc_sg_pages() 209 if (blk == NULL) { snd_trident_alloc_sg_pages() 216 for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { snd_trident_alloc_sg_pages() 222 __snd_util_mem_free(hdr, blk); snd_trident_alloc_sg_pages() 229 return blk; snd_trident_alloc_sg_pages() 240 struct snd_util_memblk *blk; snd_trident_alloc_cont_pages() local 255 blk = search_empty(hdr, runtime->dma_bytes); snd_trident_alloc_cont_pages() 256 if (blk == NULL) { snd_trident_alloc_cont_pages() 264 for (page = firstpg(blk); page <= lastpg(blk); page++, snd_trident_alloc_cont_pages() 267 __snd_util_mem_free(hdr, blk); snd_trident_alloc_cont_pages() 274 return blk; snd_trident_alloc_cont_pages() 297 struct snd_util_memblk *blk) snd_trident_free_pages() 302 if (snd_BUG_ON(!trident || !blk)) snd_trident_free_pages() 308 for (page = firstpg(blk); page <= lastpg(blk); page++) snd_trident_free_pages() 311 __snd_util_mem_free(hdr, blk); snd_trident_free_pages() 296 snd_trident_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk) snd_trident_free_pages() argument
|
H A D | trident.h | 438 int snd_trident_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk); 440 int snd_trident_synth_free(struct snd_trident *trident, struct snd_util_memblk *blk); 441 int snd_trident_synth_copy_from_user(struct snd_trident *trident, struct snd_util_memblk *blk,
|
/linux-4.1.27/arch/arm64/crypto/ |
H A D | Makefile | 26 obj-$(CONFIG_CRYPTO_AES_ARM64_CE_BLK) += aes-ce-blk.o 27 aes-ce-blk-y := aes-glue-ce.o aes-ce.o 29 obj-$(CONFIG_CRYPTO_AES_ARM64_NEON_BLK) += aes-neon-blk.o 30 aes-neon-blk-y := aes-glue-neon.o aes-neon.o
|
/linux-4.1.27/fs/gfs2/ |
H A D | recovery.h | 17 static inline void gfs2_replay_incr_blk(struct gfs2_sbd *sdp, unsigned int *blk) gfs2_replay_incr_blk() argument 19 if (++*blk == sdp->sd_jdesc->jd_blocks) gfs2_replay_incr_blk() 20 *blk = 0; gfs2_replay_incr_blk() 23 extern int gfs2_replay_read_block(struct gfs2_jdesc *jd, unsigned int blk,
|
H A D | recovery.c | 32 int gfs2_replay_read_block(struct gfs2_jdesc *jd, unsigned int blk, gfs2_replay_read_block() argument 42 error = gfs2_extent_map(&ip->i_inode, blk, &new, &dblock, &extlen); gfs2_replay_read_block() 139 * @blk: the block to look at 150 static int get_log_header(struct gfs2_jdesc *jd, unsigned int blk, get_log_header() argument 159 error = gfs2_replay_read_block(jd, blk, &bh); get_log_header() 170 if (error || lh.lh_blkno != blk || lh.lh_hash != hash) get_log_header() 181 * @blk: the segment to start searching from 191 static int find_good_lh(struct gfs2_jdesc *jd, unsigned int *blk, find_good_lh() argument 194 unsigned int orig_blk = *blk; find_good_lh() 198 error = get_log_header(jd, *blk, head); find_good_lh() 202 if (++*blk == jd->jd_blocks) find_good_lh() 203 *blk = 0; find_good_lh() 205 if (*blk == orig_blk) { find_good_lh() 225 unsigned int blk = head->lh_blkno; jhead_scan() local 230 if (++blk == jd->jd_blocks) jhead_scan() 231 blk = 0; jhead_scan() 233 error = get_log_header(jd, blk, &lh); jhead_scan()
|
H A D | rgrp.h | 31 extern struct gfs2_rgrpd *gfs2_blk2rgrpd(struct gfs2_sbd *sdp, u64 blk, bool exact);
|
H A D | dir.c | 1885 u64 blk, nblk; leaf_dealloc() local 1911 for (blk = leaf_no; blk; blk = nblk) { leaf_dealloc() 1912 if (blk != leaf_no) { leaf_dealloc() 1913 error = get_leaf(dip, blk, &bh); leaf_dealloc() 1919 if (blk != leaf_no) leaf_dealloc() 1922 gfs2_rlist_add(dip, &rlist, blk); leaf_dealloc() 1946 for (blk = leaf_no; blk; blk = nblk) { leaf_dealloc() 1947 if (blk != leaf_no) { leaf_dealloc() 1948 error = get_leaf(dip, blk, &bh); leaf_dealloc() 1954 if (blk != leaf_no) leaf_dealloc() 1957 gfs2_free_meta(dip, blk, 1); leaf_dealloc()
|
H A D | rgrp.c | 178 * @blk: absolute file system block number of the new reservation 186 static inline int rs_cmp(u64 blk, u32 len, struct gfs2_blkreserv *rs) rs_cmp() argument 190 if (blk >= startblk + rs->rs_free) rs_cmp() 192 if (blk + len - 1 < startblk) rs_cmp() 497 * @blk: The data block number 503 struct gfs2_rgrpd *gfs2_blk2rgrpd(struct gfs2_sbd *sdp, u64 blk, bool exact) gfs2_blk2rgrpd() argument 513 if (blk < cur->rd_addr) gfs2_blk2rgrpd() 515 else if (blk >= cur->rd_data0 + cur->rd_data) gfs2_blk2rgrpd() 520 if (blk < cur->rd_addr) gfs2_blk2rgrpd() 522 if (blk >= cur->rd_data0 + cur->rd_data) gfs2_blk2rgrpd() 1272 u64 blk; gfs2_rgrp_send_discards() local 1293 blk = offset + ((bi->bi_start + x) * GFS2_NBBY); gfs2_rgrp_send_discards() 1298 if ((start + nr_blks) != blk) { gfs2_rgrp_send_discards() 1309 start = blk; gfs2_rgrp_send_discards() 1314 blk++; gfs2_rgrp_send_discards()
|
H A D | xattr.c | 1004 u64 blk; ea_set_block() local 1006 error = gfs2_alloc_blocks(ip, &blk, &n, 0, NULL); ea_set_block() 1009 gfs2_trans_add_unrevoke(sdp, blk, 1); ea_set_block() 1010 indbh = gfs2_meta_new(ip->i_gl, blk); ea_set_block() 1017 ip->i_eattr = blk; ea_set_block()
|
H A D | quota.c | 837 * 1 blk for unstuffing inode if stuffed. We add this extra do_sync()
|
/linux-4.1.27/fs/quota/ |
H A D | quota_tree.c | 52 static ssize_t read_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) read_blk() argument 58 info->dqi_usable_bs, blk << info->dqi_blocksize_bits); read_blk() 61 static ssize_t write_blk(struct qtree_mem_dqinfo *info, uint blk, char *buf) write_blk() argument 67 info->dqi_usable_bs, blk << info->dqi_blocksize_bits); write_blk() 81 int ret, blk; get_free_dqblk() local 86 blk = info->dqi_free_blk; get_free_dqblk() 87 ret = read_blk(info, blk, buf); get_free_dqblk() 98 blk = info->dqi_blocks++; get_free_dqblk() 101 ret = blk; get_free_dqblk() 108 static int put_free_dqblk(struct qtree_mem_dqinfo *info, char *buf, uint blk) put_free_dqblk() argument 116 err = write_blk(info, blk, buf); put_free_dqblk() 119 info->dqi_free_blk = blk; put_free_dqblk() 126 uint blk) remove_free_dqentry() 162 if (write_blk(info, blk, buf) < 0) remove_free_dqentry() 164 "with free entries", blk); remove_free_dqentry() 173 uint blk) insert_free_dqentry() 183 err = write_blk(info, blk, buf); insert_free_dqentry() 191 cpu_to_le32(blk); insert_free_dqentry() 197 info->dqi_free_entry = blk; insert_free_dqentry() 221 uint blk, i; find_free_dqentry() local 233 blk = info->dqi_free_entry; find_free_dqentry() 234 *err = read_blk(info, blk, buf); find_free_dqentry() 238 blk = get_free_dqblk(info); find_free_dqentry() 239 if ((int)blk < 0) { find_free_dqentry() 240 *err = blk; find_free_dqentry() 247 info->dqi_free_entry = blk; find_free_dqentry() 252 *err = remove_free_dqentry(info, buf, blk); find_free_dqentry() 255 "from entry free list", blk); find_free_dqentry() 274 *err = write_blk(info, blk, buf); find_free_dqentry() 277 blk); find_free_dqentry() 280 dquot->dq_off = (blk << info->dqi_blocksize_bits) + find_free_dqentry() 284 return blk; find_free_dqentry() 407 uint blk) free_dqentry() 415 if (dquot->dq_off >> info->dqi_blocksize_bits != blk) { free_dqentry() 417 "other block (%u) than it should (%u)", blk, free_dqentry() 421 ret = read_blk(info, blk, buf); free_dqentry() 424 blk); free_dqentry() 430 ret = remove_free_dqentry(info, buf, blk); free_dqentry() 432 ret = put_free_dqblk(info, buf, blk); free_dqentry() 435 "(%u) to free list", blk); free_dqentry() 445 ret = insert_free_dqentry(info, buf, blk); free_dqentry() 448 "data block (%u) to free entry list", blk); free_dqentry() 452 ret = write_blk(info, blk, buf); free_dqentry() 455 "data block %u", blk); free_dqentry() 468 uint *blk, int depth) remove_tree() 477 ret = read_blk(info, *blk, buf); remove_tree() 480 *blk); remove_tree() 498 && *blk != QT_TREEOFF) { remove_tree() 499 put_free_dqblk(info, buf, *blk); remove_tree() 500 *blk = 0; remove_tree() 502 ret = write_blk(info, *blk, buf); remove_tree() 506 *blk); remove_tree() 527 struct dquot *dquot, uint blk) find_block_dqentry() 536 ret = read_blk(info, blk, buf); find_block_dqentry() 539 "block %u", blk); find_block_dqentry() 555 ret = (blk << info->dqi_blocksize_bits) + sizeof(struct find_block_dqentry() 565 struct dquot *dquot, uint blk, int depth) find_tree_dqentry() 573 ret = read_blk(info, blk, buf); find_tree_dqentry() 576 blk); find_tree_dqentry() 580 blk = le32_to_cpu(ref[get_index(info, dquot->dq_id, depth)]); find_tree_dqentry() 581 if (!blk) /* No reference? */ find_tree_dqentry() 584 ret = find_tree_dqentry(info, dquot, blk, depth+1); find_tree_dqentry() 586 ret = find_block_dqentry(info, dquot, blk); find_tree_dqentry() 125 remove_free_dqentry(struct qtree_mem_dqinfo *info, char *buf, uint blk) remove_free_dqentry() argument 172 insert_free_dqentry(struct qtree_mem_dqinfo *info, char *buf, uint blk) insert_free_dqentry() argument 406 free_dqentry(struct qtree_mem_dqinfo *info, struct dquot *dquot, uint blk) free_dqentry() argument 467 remove_tree(struct qtree_mem_dqinfo *info, struct dquot *dquot, uint *blk, int depth) remove_tree() argument 526 find_block_dqentry(struct qtree_mem_dqinfo *info, struct dquot *dquot, uint blk) find_block_dqentry() argument 564 find_tree_dqentry(struct qtree_mem_dqinfo *info, struct dquot *dquot, uint blk, int depth) find_tree_dqentry() argument
|
/linux-4.1.27/fs/affs/ |
H A D | bitmap.c | 43 u32 blk, bmap, bit, mask, tmp; affs_free_block() local 51 blk = block - sbi->s_reserved; affs_free_block() 52 bmap = blk / sbi->s_bmap_bits; affs_free_block() 53 bit = blk % sbi->s_bmap_bits; affs_free_block() 121 u32 blk, bmap, bit, mask, mask2, tmp; affs_alloc_block() local 143 blk = goal - sbi->s_reserved; affs_alloc_block() 144 bmap = blk / sbi->s_bmap_bits; affs_alloc_block() 166 blk = bmap * sbi->s_bmap_bits; affs_alloc_block() 181 bit = blk % sbi->s_bmap_bits; affs_alloc_block() 185 blk &= ~31UL; affs_alloc_block() 193 blk += 32; affs_alloc_block() 206 blk += bit + sbi->s_reserved; affs_alloc_block() 208 AFFS_I(inode)->i_lastalloc = blk; affs_alloc_block() 230 pr_debug("%d\n", blk); affs_alloc_block() 231 return blk; affs_alloc_block() 248 u32 size, blk, end, offset, mask; affs_init_bitmap() local 274 blk = sb->s_blocksize / 4 - 49; affs_init_bitmap() 275 end = blk + 25; affs_init_bitmap() 280 bm->bm_key = be32_to_cpu(bmap_blk[blk]); affs_init_bitmap() 293 pr_debug("read bitmap block %d: %d\n", blk, bm->bm_key); affs_init_bitmap() 299 if (++blk < end || i == 1) affs_init_bitmap() 303 bmap_bh = affs_bread(sb, be32_to_cpu(bmap_blk[blk])); affs_init_bitmap() 310 blk = 0; affs_init_bitmap()
|
H A D | file.c | 838 u32 last_blk, blkcnt, blk; affs_truncate() local 895 blk = last_blk; affs_truncate() 898 blk++; affs_truncate() 903 if (size > blkcnt - blk + i) affs_truncate() 904 size = blkcnt - blk + i; affs_truncate() 905 for (; i < size; i++, blk++) { affs_truncate() 941 if (size > blkcnt - blk) affs_truncate() 942 size = blkcnt - blk; affs_truncate() 943 for (i = 0; i < size; i++, blk++) affs_truncate()
|
H A D | affs.h | 18 #define AFFS_BLOCK(sb, bh, blk) (AFFS_HEAD(bh)->table[AFFS_SB(sb)->s_hashsize-1-(blk)])
|
/linux-4.1.27/block/ |
H A D | blk-mq-cpu.c | 2 * CPU notifier helper code for blk-mq 15 #include <linux/blk-mq.h> 16 #include "blk-mq.h"
|
H A D | blk-mq-cpumap.c | 13 #include <linux/blk-mq.h> 14 #include "blk.h" 15 #include "blk-mq.h"
|
H A D | blk-timeout.c | 9 #include "blk.h" 10 #include "blk-mq.h" 104 * the blk layer. blk_rq_timed_out() 196 /* blk-mq has its own handler, so we don't need ->rq_timed_out_fn */ blk_add_timer()
|
H A D | blk-iopoll.c | 12 #include <linux/blk-iopoll.h> 15 #include "blk.h" 27 * raise of the blk iopoll softirq. The driver must already have gotten a
|
H A D | blk-exec.c | 8 #include <linux/blk-mq.h> 11 #include "blk.h"
|
H A D | blk-mq-tag.h | 4 #include "blk-mq.h"
|
H A D | blk-mq-tag.c | 17 #include <linux/blk-mq.h> 18 #include "blk.h" 19 #include "blk-mq.h" 20 #include "blk-mq-tag.h" 179 * last_tag caches, which blk-mq stores in the blk_mq_ctx software queue 575 pr_err("blk-mq: tag depth too large\n"); blk_mq_init_tags()
|
H A D | blk-mq-sysfs.c | 12 #include <linux/blk-mq.h> 13 #include "blk-mq.h" 14 #include "blk-mq-tag.h"
|
H A D | blk-sysfs.c | 10 #include <linux/blk-mq.h> 12 #include "blk.h" 13 #include "blk-cgroup.h" 14 #include "blk-mq.h"
|
H A D | blk.h | 5 #include <linux/blk-mq.h> 6 #include "blk-mq.h"
|
H A D | blk-flush.c | 72 #include <linux/blk-mq.h> 74 #include "blk.h" 75 #include "blk-mq.h"
|
H A D | blk-map.c | 10 #include "blk.h"
|
H A D | blk-softirq.c | 13 #include "blk.h"
|
H A D | blk-integrity.c | 2 * blk-integrity.c - Block layer data integrity extensions 30 #include "blk.h"
|
H A D | blk-merge.c | 10 #include "blk.h" 349 * blk-mq uses req->special to carry normal driver per-request payload, it
|
H A D | blk-mq.c | 27 #include <linux/blk-mq.h> 28 #include "blk.h" 29 #include "blk-mq.h" 30 #include "blk-mq-tag.h" 830 pr_err("blk-mq: bad return on queue: %d\n", ret); __blk_mq_run_hw_queue() 2165 pr_err("blk-mq: failed to allocate request map\n"); blk_mq_alloc_rq_maps() 2170 pr_info("blk-mq: reduced tag depth (%u -> %u)\n", blk_mq_alloc_rq_maps() 2197 pr_info("blk-mq: reduced tag depth to %u\n", blk_mq_alloc_tag_set()
|
H A D | blk-lib.c | 10 #include "blk.h"
|
H A D | elevator.c | 41 #include "blk.h" 42 #include "blk-cgroup.h"
|
H A D | blk-ioc.c | 11 #include "blk.h"
|
H A D | blk-tag.c | 10 #include "blk.h"
|
H A D | blk-throttle.c | 12 #include "blk-cgroup.h" 13 #include "blk.h" 995 * more than once as a throttled bio will go through blk-throtl the throtl_charge_bio() 1388 * blk-throttle. tg_set_conf() 1587 * As multiple blk-throtls may stack in the same issue path, we blk_throtl_bio()
|
H A D | blk-cgroup.c | 22 #include "blk-cgroup.h" 23 #include "blk.h"
|
H A D | blk-cgroup.h | 90 /* association between a blk cgroup and a request queue */
|
H A D | blk-core.c | 19 #include <linux/blk-mq.h> 39 #include "blk.h" 40 #include "blk-cgroup.h" 41 #include "blk-mq.h"
|
H A D | blk-settings.c | 15 #include "blk.h"
|
H A D | genhd.c | 23 #include "blk.h"
|
/linux-4.1.27/drivers/net/wireless/orinoco/ |
H A D | hermes_dld.c | 96 dblock_addr(const struct dblock *blk) dblock_addr() argument 98 return le32_to_cpu(blk->addr); dblock_addr() 102 dblock_len(const struct dblock *blk) dblock_len() argument 104 return le16_to_cpu(blk->len); dblock_len() 256 const struct dblock *blk = (const struct dblock *) first_block; hermes_blocks_length() local 260 end -= sizeof(*blk); hermes_blocks_length() 264 while (((void *) blk <= end) && hermes_blocks_length() 265 (dblock_addr(blk) != BLOCK_END)) { hermes_blocks_length() 266 len = dblock_len(blk); hermes_blocks_length() 267 total_len += sizeof(*blk) + len; hermes_blocks_length() 268 blk = (struct dblock *) &blk->data[len]; hermes_blocks_length() 279 const struct dblock *blk; hermes_program() local 284 blk = (const struct dblock *) first_block; hermes_program() 286 if ((void *) blk > (end - sizeof(*blk))) hermes_program() 289 blkaddr = dblock_addr(blk); hermes_program() 290 blklen = dblock_len(blk); hermes_program() 293 (((void *) blk + blklen) <= end)) { hermes_program() 297 err = hw->ops->program(hw, blk->data, blkaddr, blklen); hermes_program() 301 blk = (const struct dblock *) &blk->data[blklen]; hermes_program() 303 if ((void *) blk > (end - sizeof(*blk))) hermes_program() 306 blkaddr = dblock_addr(blk); hermes_program() 307 blklen = dblock_len(blk); hermes_program()
|
/linux-4.1.27/arch/x86/mm/ |
H A D | numa.c | 150 mi->blk[mi->nr_blks].start = start; numa_add_memblk_to() 151 mi->blk[mi->nr_blks].end = end; numa_add_memblk_to() 152 mi->blk[mi->nr_blks].nid = nid; numa_add_memblk_to() 162 * Remove @idx'th numa_memblk from @mi by shifting @mi->blk[] and 168 memmove(&mi->blk[idx], &mi->blk[idx + 1], numa_remove_memblk_from() 169 (mi->nr_blks - idx) * sizeof(mi->blk[0])); numa_remove_memblk_from() 243 struct numa_memblk *bi = &mi->blk[i]; numa_cleanup_meminfo() 256 struct numa_memblk *bi = &mi->blk[i]; numa_cleanup_meminfo() 259 struct numa_memblk *bj = &mi->blk[j]; numa_cleanup_meminfo() 289 struct numa_memblk *bk = &mi->blk[k]; numa_cleanup_meminfo() 308 for (i = mi->nr_blks; i < ARRAY_SIZE(mi->blk); i++) { numa_cleanup_meminfo() 309 mi->blk[i].start = mi->blk[i].end = 0; numa_cleanup_meminfo() 310 mi->blk[i].nid = NUMA_NO_NODE; numa_cleanup_meminfo() 324 for (i = 0; i < ARRAY_SIZE(mi->blk); i++) numa_nodemask_from_meminfo() 325 if (mi->blk[i].start != mi->blk[i].end && numa_nodemask_from_meminfo() 326 mi->blk[i].nid != NUMA_NO_NODE) numa_nodemask_from_meminfo() 327 node_set(mi->blk[i].nid, *nodemask); numa_nodemask_from_meminfo() 446 u64 s = mi->blk[i].start >> PAGE_SHIFT; numa_meminfo_cover_memory() 447 u64 e = mi->blk[i].end >> PAGE_SHIFT; numa_meminfo_cover_memory() 449 numaram -= __absent_pages_in_range(mi->blk[i].nid, s, e); numa_meminfo_cover_memory() 479 struct numa_memblk *mb = &numa_meminfo.blk[i]; numa_clear_kernel_node_hotplug() 498 nid = numa_meminfo.blk[i].nid; for_each_memblock() 502 start = numa_meminfo.blk[i].start; for_each_memblock() 503 end = numa_meminfo.blk[i].end; for_each_memblock() 521 struct numa_memblk *mb = &mi->blk[i]; numa_register_memblks() 557 if (nid != mi->blk[i].nid) for_each_node_mask() 559 start = min(mi->blk[i].start, start); for_each_node_mask() 560 end = max(mi->blk[i].end, end); for_each_node_mask() 873 int nid = mi->blk[0].nid; memory_add_physaddr_to_nid() 877 if (mi->blk[i].start <= start && mi->blk[i].end > start) memory_add_physaddr_to_nid() 878 nid = mi->blk[i].nid; memory_add_physaddr_to_nid()
|
H A D | numa_internal.h | 15 struct numa_memblk blk[NR_NODE_MEMBLKS]; member in struct:numa_meminfo
|
H A D | numa_emulation.c | 26 if (mi->blk[i].nid == nid) emu_find_memblk_by_nid() 49 struct numa_memblk *eb = &ei->blk[ei->nr_blks]; emu_setup_memblk() 50 struct numa_memblk *pb = &pi->blk[phys_blk]; emu_setup_memblk() 120 node_set(pi->blk[i].nid, physnode_mask); split_nodes_interleave() 137 start = pi->blk[phys_blk].start; for_each_node_mask() 138 limit = pi->blk[phys_blk].end; for_each_node_mask() 235 node_set(pi->blk[i].nid, physnode_mask); split_nodes_size_interleave() 252 start = pi->blk[phys_blk].start; for_each_node_mask() 253 limit = pi->blk[phys_blk].end; for_each_node_mask()
|
/linux-4.1.27/drivers/media/platform/vivid/ |
H A D | vivid-radio-tx.c | 40 unsigned blk; vivid_radio_tx_write() local 62 blk = ts.tv_sec * 100 + ts.tv_nsec / 10000000; vivid_radio_tx_write() 63 blk = (blk * VIVID_RDS_GEN_BLOCKS) / 500; vivid_radio_tx_write() 64 if (blk - VIVID_RDS_GEN_BLOCKS >= dev->radio_tx_rds_last_block) vivid_radio_tx_write() 65 dev->radio_tx_rds_last_block = blk - VIVID_RDS_GEN_BLOCKS + 1; vivid_radio_tx_write() 73 if (blk == dev->radio_tx_rds_last_block || vivid_radio_tx_write() 85 for (i = 0; i < size && blk > dev->radio_tx_rds_last_block; vivid_radio_tx_write()
|
H A D | vivid-radio-rx.c | 42 unsigned blk; vivid_radio_rx_read() local 74 blk = ts.tv_sec * 100 + ts.tv_nsec / 10000000; vivid_radio_rx_read() 75 blk = (blk * VIVID_RDS_GEN_BLOCKS) / 500; vivid_radio_rx_read() 76 if (blk >= dev->radio_rx_rds_last_block + VIVID_RDS_GEN_BLOCKS) vivid_radio_rx_read() 77 dev->radio_rx_rds_last_block = blk - VIVID_RDS_GEN_BLOCKS + 1; vivid_radio_rx_read() 85 if (blk == dev->radio_rx_rds_last_block || !dev->radio_rx_rds_enabled || vivid_radio_rx_read() 101 for (i = 0; i < size && blk > dev->radio_rx_rds_last_block; vivid_radio_rx_read()
|
/linux-4.1.27/include/sound/ |
H A D | util_mem.h | 34 #define snd_util_memblk_argptr(blk) (void*)((char*)(blk) + sizeof(struct snd_util_memblk)) 54 int snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk); 59 void __snd_util_mem_free(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk);
|
H A D | emu10k1.h | 1571 #define snd_emu10k1_memblk_offset(blk) (((blk)->mapped_page << PAGE_SHIFT) | ((blk)->mem.offset & (PAGE_SIZE - 1))) 1879 int snd_emu10k1_free_pages(struct snd_emu10k1 *emu, struct snd_util_memblk *blk); 1881 int snd_emu10k1_synth_free(struct snd_emu10k1 *emu, struct snd_util_memblk *blk); 1882 int snd_emu10k1_synth_bzero(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, int offset, int size); 1883 int snd_emu10k1_synth_copy_from_user(struct snd_emu10k1 *emu, struct snd_util_memblk *blk, int offset, const char __user *data, int size); 1884 int snd_emu10k1_memblk_map(struct snd_emu10k1 *emu, struct snd_emu10k1_memblk *blk);
|
/linux-4.1.27/fs/xfs/libxfs/ |
H A D | xfs_attr.c | 880 xfs_da_state_blk_t *blk; xfs_attr_node_addname() local 904 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr_node_addname() 905 ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); xfs_attr_node_addname() 932 retval = xfs_attr3_leaf_add(blk->bp, state->args); xfs_attr_node_addname() 1081 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr_node_addname() 1082 ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); xfs_attr_node_addname() 1083 error = xfs_attr3_leaf_remove(blk->bp, args); xfs_attr_node_addname() 1149 xfs_da_state_blk_t *blk; xfs_attr_node_removename() local 1179 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr_node_removename() 1180 ASSERT(blk->bp != NULL); xfs_attr_node_removename() 1181 ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); xfs_attr_node_removename() 1215 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr_node_removename() 1216 ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); xfs_attr_node_removename() 1217 retval = xfs_attr3_leaf_remove(blk->bp, args); xfs_attr_node_removename() 1260 ASSERT(state->path.blk[0].bp); xfs_attr_node_removename() 1261 state->path.blk[0].bp = NULL; xfs_attr_node_removename() 1310 xfs_da_state_blk_t *blk; xfs_attr_fillstate() local 1321 for (blk = path->blk, level = 0; level < path->active; blk++, level++) { xfs_attr_fillstate() 1322 if (blk->bp) { xfs_attr_fillstate() 1323 blk->disk_blkno = XFS_BUF_ADDR(blk->bp); xfs_attr_fillstate() 1324 blk->bp = NULL; xfs_attr_fillstate() 1326 blk->disk_blkno = 0; xfs_attr_fillstate() 1336 for (blk = path->blk, level = 0; level < path->active; blk++, level++) { xfs_attr_fillstate() 1337 if (blk->bp) { xfs_attr_fillstate() 1338 blk->disk_blkno = XFS_BUF_ADDR(blk->bp); xfs_attr_fillstate() 1339 blk->bp = NULL; xfs_attr_fillstate() 1341 blk->disk_blkno = 0; xfs_attr_fillstate() 1358 xfs_da_state_blk_t *blk; xfs_attr_refillstate() local 1369 for (blk = path->blk, level = 0; level < path->active; blk++, level++) { xfs_attr_refillstate() 1370 if (blk->disk_blkno) { xfs_attr_refillstate() 1373 blk->blkno, blk->disk_blkno, xfs_attr_refillstate() 1374 &blk->bp, XFS_ATTR_FORK); xfs_attr_refillstate() 1378 blk->bp = NULL; xfs_attr_refillstate() 1388 for (blk = path->blk, level = 0; level < path->active; blk++, level++) { xfs_attr_refillstate() 1389 if (blk->disk_blkno) { xfs_attr_refillstate() 1392 blk->blkno, blk->disk_blkno, xfs_attr_refillstate() 1393 &blk->bp, XFS_ATTR_FORK); xfs_attr_refillstate() 1397 blk->bp = NULL; xfs_attr_refillstate() 1415 xfs_da_state_blk_t *blk; xfs_attr_node_get() local 1432 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr_node_get() 1433 ASSERT(blk->bp != NULL); xfs_attr_node_get() 1434 ASSERT(blk->magic == XFS_ATTR_LEAF_MAGIC); xfs_attr_node_get() 1439 retval = xfs_attr3_leaf_getvalue(blk->bp, args); xfs_attr_node_get() 1450 xfs_trans_brelse(args->trans, state->path.blk[i].bp); xfs_attr_node_get() 1451 state->path.blk[i].bp = NULL; xfs_attr_node_get()
|
H A D | xfs_da_btree.c | 113 state->altpath.blk[i].bp = NULL; xfs_da_state_kill_altpath() 369 ASSERT(state->path.blk[max].magic == XFS_ATTR_LEAF_MAGIC || xfs_da3_split() 370 state->path.blk[max].magic == XFS_DIR2_LEAFN_MAGIC); xfs_da3_split() 372 addblk = &state->path.blk[max]; /* initial dummy value */ xfs_da3_split() 374 oldblk = &state->path.blk[i]; xfs_da3_split() 375 newblk = &state->altpath.blk[i]; xfs_da3_split() 446 oldblk = &state->path.blk[0]; xfs_da3_split() 940 drop_blk = &state->path.blk[ state->path.active-1 ]; xfs_da3_join() 941 save_blk = &state->altpath.blk[ state->path.active-1 ]; xfs_da3_join() 942 ASSERT(state->path.blk[0].magic == XFS_DA_NODE_MAGIC); xfs_da3_join() 1008 error = xfs_da3_root_join(state, &state->path.blk[0]); xfs_da3_join() 1116 struct xfs_da_state_blk *blk; xfs_da3_node_toosmall() local 1135 blk = &state->path.blk[ state->path.active-1 ]; xfs_da3_node_toosmall() 1136 info = blk->bp->b_addr; xfs_da3_node_toosmall() 1140 *action = 0; /* blk over 50%, don't try to join */ xfs_da3_node_toosmall() 1141 return 0; /* blk over 50%, don't try to join */ xfs_da3_node_toosmall() 1180 /* start with smaller blk num */ xfs_da3_node_toosmall() 1212 if (blkno < blk->blkno) { xfs_da3_node_toosmall() 1261 struct xfs_da_state_blk *blk; xfs_da3_fixhashpath() local 1272 blk = &path->blk[ level ]; xfs_da3_fixhashpath() 1273 switch (blk->magic) { xfs_da3_fixhashpath() 1275 lasthash = xfs_attr_leaf_lasthash(blk->bp, &count); xfs_da3_fixhashpath() 1280 lasthash = xfs_dir2_leafn_lasthash(dp, blk->bp, &count); xfs_da3_fixhashpath() 1285 lasthash = xfs_da3_node_lasthash(dp, blk->bp, &count); xfs_da3_fixhashpath() 1290 for (blk--, level--; level >= 0; blk--, level--) { xfs_da3_fixhashpath() 1293 node = blk->bp->b_addr; xfs_da3_fixhashpath() 1296 if (be32_to_cpu(btree[blk->index].hashval) == lasthash) xfs_da3_fixhashpath() 1298 blk->hashval = lasthash; xfs_da3_fixhashpath() 1299 btree[blk->index].hashval = cpu_to_be32(lasthash); xfs_da3_fixhashpath() 1300 xfs_trans_log_buf(state->args->trans, blk->bp, xfs_da3_fixhashpath() 1301 XFS_DA_LOGRANGE(node, &btree[blk->index], xfs_da3_fixhashpath() 1450 struct xfs_da_state_blk *blk; xfs_da3_node_lookup_int() local 1473 for (blk = &state->path.blk[0], state->path.active = 1; xfs_da3_node_lookup_int() 1475 blk++, state->path.active++) { xfs_da3_node_lookup_int() 1479 blk->blkno = blkno; xfs_da3_node_lookup_int() 1481 -1, &blk->bp, args->whichfork); xfs_da3_node_lookup_int() 1483 blk->blkno = 0; xfs_da3_node_lookup_int() 1487 curr = blk->bp->b_addr; xfs_da3_node_lookup_int() 1488 blk->magic = be16_to_cpu(curr->magic); xfs_da3_node_lookup_int() 1490 if (blk->magic == XFS_ATTR_LEAF_MAGIC || xfs_da3_node_lookup_int() 1491 blk->magic == XFS_ATTR3_LEAF_MAGIC) { xfs_da3_node_lookup_int() 1492 blk->magic = XFS_ATTR_LEAF_MAGIC; xfs_da3_node_lookup_int() 1493 blk->hashval = xfs_attr_leaf_lasthash(blk->bp, NULL); xfs_da3_node_lookup_int() 1497 if (blk->magic == XFS_DIR2_LEAFN_MAGIC || xfs_da3_node_lookup_int() 1498 blk->magic == XFS_DIR3_LEAFN_MAGIC) { xfs_da3_node_lookup_int() 1499 blk->magic = XFS_DIR2_LEAFN_MAGIC; xfs_da3_node_lookup_int() 1500 blk->hashval = xfs_dir2_leafn_lasthash(args->dp, xfs_da3_node_lookup_int() 1501 blk->bp, NULL); xfs_da3_node_lookup_int() 1505 blk->magic = XFS_DA_NODE_MAGIC; xfs_da3_node_lookup_int() 1511 node = blk->bp->b_addr; xfs_da3_node_lookup_int() 1516 blk->hashval = be32_to_cpu(btree[max - 1].hashval); xfs_da3_node_lookup_int() 1554 blk->index = max - 1; xfs_da3_node_lookup_int() 1557 blk->index = probe; xfs_da3_node_lookup_int() 1569 if (blk->magic == XFS_DIR2_LEAFN_MAGIC) { xfs_da3_node_lookup_int() 1570 retval = xfs_dir2_leafn_lookup_int(blk->bp, args, xfs_da3_node_lookup_int() 1571 &blk->index, state); xfs_da3_node_lookup_int() 1572 } else if (blk->magic == XFS_ATTR_LEAF_MAGIC) { xfs_da3_node_lookup_int() 1573 retval = xfs_attr3_leaf_lookup_int(blk->bp, args); xfs_da3_node_lookup_int() 1574 blk->index = args->index; xfs_da3_node_lookup_int() 1575 args->blkno = blk->blkno; xfs_da3_node_lookup_int() 1581 (blk->hashval == args->hashval)) { xfs_da3_node_lookup_int() 1588 } else if (blk->magic == XFS_ATTR_LEAF_MAGIC) { xfs_da3_node_lookup_int() 1819 struct xfs_da_state_blk *blk; xfs_da3_path_shift() local 1842 for (blk = &path->blk[level]; level >= 0; blk--, level--) { xfs_da3_path_shift() 1843 node = blk->bp->b_addr; xfs_da3_path_shift() 1847 if (forward && (blk->index < nodehdr.count - 1)) { xfs_da3_path_shift() 1848 blk->index++; xfs_da3_path_shift() 1849 blkno = be32_to_cpu(btree[blk->index].before); xfs_da3_path_shift() 1851 } else if (!forward && (blk->index > 0)) { xfs_da3_path_shift() 1852 blk->index--; xfs_da3_path_shift() 1853 blkno = be32_to_cpu(btree[blk->index].before); xfs_da3_path_shift() 1867 for (blk++, level++; level < path->active; blk++, level++) { xfs_da3_path_shift() 1873 xfs_trans_brelse(args->trans, blk->bp); xfs_da3_path_shift() 1878 blk->blkno = blkno; xfs_da3_path_shift() 1880 &blk->bp, args->whichfork); xfs_da3_path_shift() 1883 info = blk->bp->b_addr; xfs_da3_path_shift() 1899 blk->magic = XFS_DA_NODE_MAGIC; xfs_da3_path_shift() 1903 blk->hashval = be32_to_cpu(btree[nodehdr.count - 1].hashval); xfs_da3_path_shift() 1905 blk->index = 0; xfs_da3_path_shift() 1907 blk->index = nodehdr.count - 1; xfs_da3_path_shift() 1908 blkno = be32_to_cpu(btree[blk->index].before); xfs_da3_path_shift() 1912 blk->magic = XFS_ATTR_LEAF_MAGIC; xfs_da3_path_shift() 1914 blk->index = 0; xfs_da3_path_shift() 1915 blk->hashval = xfs_attr_leaf_lasthash(blk->bp, NULL); xfs_da3_path_shift() 1919 blk->magic = XFS_DIR2_LEAFN_MAGIC; xfs_da3_path_shift() 1921 blk->index = 0; xfs_da3_path_shift() 1922 blk->hashval = xfs_dir2_leafn_lasthash(args->dp, xfs_da3_path_shift() 1923 blk->bp, NULL); xfs_da3_path_shift()
|
H A D | xfs_da_btree.h | 78 int index; /* index of attr of interest in blk */ 83 int index2; /* index of 2nd attr in blk */ 120 int magic; /* blk's magic number, ie: blk type */ 125 xfs_da_state_blk_t blk[XFS_DA_NODE_MAXDEPTH]; member in struct:xfs_da_state_path
|
H A D | xfs_dir2_node.c | 1378 xfs_da_state_blk_t *blk; /* leaf block */ xfs_dir2_leafn_toosmall() local 1397 blk = &state->path.blk[state->path.active - 1]; xfs_dir2_leafn_toosmall() 1398 leaf = blk->bp->b_addr; xfs_dir2_leafn_toosmall() 1401 xfs_dir3_leaf_check(dp, blk->bp); xfs_dir2_leafn_toosmall() 1487 if (blkno < blk->blkno) xfs_dir2_leafn_toosmall() 1570 xfs_da_state_blk_t *blk; /* leaf block for insert */ xfs_dir2_node_addname() local 1602 blk = &state->path.blk[state->path.active - 1]; xfs_dir2_node_addname() 1603 ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); xfs_dir2_node_addname() 1607 rval = xfs_dir2_leafn_add(blk->bp, args, blk->index); xfs_dir2_node_addname() 2046 xfs_trans_brelse(args->trans, state->path.blk[i].bp); xfs_dir2_node_lookup() 2047 state->path.blk[i].bp = NULL; xfs_dir2_node_lookup() 2067 struct xfs_da_state_blk *blk; /* leaf block */ xfs_dir2_node_removename() local 2092 blk = &state->path.blk[state->path.active - 1]; xfs_dir2_node_removename() 2093 ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); xfs_dir2_node_removename() 2099 error = xfs_dir2_leafn_remove(args, blk->bp, blk->index, xfs_dir2_node_removename() 2129 xfs_da_state_blk_t *blk; /* leaf block */ xfs_dir2_node_replace() local 2173 blk = &state->path.blk[state->path.active - 1]; xfs_dir2_node_replace() 2174 ASSERT(blk->magic == XFS_DIR2_LEAFN_MAGIC); xfs_dir2_node_replace() 2175 leaf = blk->bp->b_addr; xfs_dir2_node_replace() 2177 lep = &ents[blk->index]; xfs_dir2_node_replace() 2209 xfs_trans_brelse(args->trans, state->path.blk[i].bp); xfs_dir2_node_replace() 2210 state->path.blk[i].bp = NULL; xfs_dir2_node_replace()
|
H A D | xfs_attr_leaf.c | 1746 struct xfs_da_state_blk *blk; xfs_attr3_leaf_toosmall() local 1763 blk = &state->path.blk[ state->path.active-1 ]; xfs_attr3_leaf_toosmall() 1764 leaf = blk->bp->b_addr; xfs_attr3_leaf_toosmall() 1770 *action = 0; /* blk over 50%, don't try to join */ xfs_attr3_leaf_toosmall() 1806 /* start with smaller blk num */ xfs_attr3_leaf_toosmall() 1844 if (blkno < blk->blkno) { xfs_attr3_leaf_toosmall()
|
H A D | xfs_btree.h | 219 xfs_fsblock_t firstblock; /* 1st blk allocated */
|
H A D | xfs_dir2_leaf.c | 1737 lbp = state->path.blk[0].bp; xfs_dir2_node_to_leaf() 1817 state->path.blk[0].bp = NULL; xfs_dir2_node_to_leaf()
|
H A D | xfs_da_format.h | 654 __u8 holes; /* != 0 if blk needs compaction */
|
/linux-4.1.27/arch/powerpc/sysdev/ |
H A D | ppc4xx_ocm.c | 87 struct ocm_block *blk, *tmp; ocm_free_region() local 93 list_for_each_entry_safe(blk, tmp, &ocm_reg->list, list) { ocm_free_region() 94 if (blk->addr == addr) { ocm_free_region() 96 ocm_reg->memfree += blk->size; ocm_free_region() 98 list_del(&blk->list); ocm_free_region() 99 kfree(blk); ocm_free_region() 235 struct ocm_block *blk, *tmp; ocm_debugfs_show() local 257 list_for_each_entry_safe(blk, tmp, &ocm->nc.list, list) { ocm_debugfs_show() 259 blk->size, blk->owner); ocm_debugfs_show() 269 list_for_each_entry_safe(blk, tmp, &ocm->c.list, list) { ocm_debugfs_show() 271 blk->size, blk->owner); ocm_debugfs_show()
|
/linux-4.1.27/fs/adfs/ |
H A D | dir_f.c | 147 int blk = 0; adfs_dir_read() local 161 for (blk = 0; blk < size; blk++) { adfs_dir_read() 164 phys = __adfs_block_map(sb, object_id, blk); adfs_dir_read() 167 object_id, blk); adfs_dir_read() 171 dir->bh[blk] = sb_bread(sb, phys); adfs_dir_read() 172 if (!dir->bh[blk]) adfs_dir_read() 190 dir->nr_buffers = blk; adfs_dir_read() 198 for (blk -= 1; blk >= 0; blk -= 1) adfs_dir_read() 199 brelse(dir->bh[blk]); adfs_dir_read()
|
H A D | dir_fplus.c | 21 unsigned int blk, size; adfs_fplus_read() local 74 for (blk = 1; blk < size; blk++) { adfs_fplus_read() 75 block = __adfs_block_map(sb, id, blk); adfs_fplus_read() 77 adfs_error(sb, "dir object %X has a hole at offset %d", id, blk); adfs_fplus_read() 81 dir->bh_fplus[blk] = sb_bread(sb, block); adfs_fplus_read() 82 if (!dir->bh_fplus[blk]) { adfs_fplus_read() 84 id, blk, block); adfs_fplus_read()
|
/linux-4.1.27/fs/ext4/ |
H A D | block_validity.c | 227 unsigned int blk; ext4_check_blockref() local 230 blk = le32_to_cpu(*bref++); ext4_check_blockref() 231 if (blk && ext4_check_blockref() 233 blk, 1))) { ext4_check_blockref() 234 es->s_last_error_block = cpu_to_le64(blk); ext4_check_blockref() 235 ext4_error_inode(inode, function, line, blk, ext4_check_blockref()
|
H A D | balloc.c | 322 ext4_fsblk_t blk; ext4_valid_block_bitmap() local 337 blk = ext4_block_bitmap(sb, desc); ext4_valid_block_bitmap() 338 offset = blk - group_first_block; ext4_valid_block_bitmap() 341 return blk; ext4_valid_block_bitmap() 344 blk = ext4_inode_bitmap(sb, desc); ext4_valid_block_bitmap() 345 offset = blk - group_first_block; ext4_valid_block_bitmap() 348 return blk; ext4_valid_block_bitmap() 351 blk = ext4_inode_table(sb, desc); ext4_valid_block_bitmap() 352 offset = blk - group_first_block; ext4_valid_block_bitmap() 359 return blk; ext4_valid_block_bitmap() 368 ext4_fsblk_t blk; ext4_validate_block_bitmap() local 376 blk = ext4_valid_block_bitmap(sb, desc, block_group, bh); ext4_validate_block_bitmap() 377 if (unlikely(blk != 0)) { ext4_validate_block_bitmap() 380 block_group, blk); ext4_validate_block_bitmap()
|
H A D | resize.c | 355 ext4_fsblk_t blk) bclean() 360 bh = sb_getblk(sb, blk); bclean() 714 const ext4_fsblk_t blk = primary->b_blocknr; verify_reserved_gdb() local 724 grp * EXT4_BLOCKS_PER_GROUP(sb) + blk){ verify_reserved_gdb() 727 blk, grp, verify_reserved_gdb() 730 blk); verify_reserved_gdb() 941 ext4_fsblk_t blk; reserve_backup_gdb() local 958 blk = EXT4_SB(sb)->s_sbh->b_blocknr + 1 + EXT4_SB(sb)->s_gdb_count; reserve_backup_gdb() 964 for (res = 0; res < reserved_gdb; res++, blk++) { reserve_backup_gdb() 965 if (le32_to_cpu(*data) != blk) { reserve_backup_gdb() 968 blk, reserve_backup_gdb() 973 primary[res] = sb_bread(sb, blk); reserve_backup_gdb() 1001 blk = group * EXT4_BLOCKS_PER_GROUP(sb); reserve_backup_gdb() 1007 blk + primary[i]->b_blocknr); */ reserve_backup_gdb() 1008 data[gdbackups] = cpu_to_le32(blk + primary[i]->b_blocknr); reserve_backup_gdb() 354 bclean(handle_t *handle, struct super_block *sb, ext4_fsblk_t blk) bclean() argument
|
H A D | migrate.c | 624 ext4_fsblk_t blk; ext4_ind_migrate() local 662 blk = len = start = end = 0; ext4_ind_migrate() 665 blk = ext4_ext_pblock(ex); ext4_ind_migrate() 677 ei->i_data[i] = cpu_to_le32(blk++); ext4_ind_migrate()
|
H A D | ext4.h | 256 #define EXT4_B2C(sbi, blk) ((blk) >> (sbi)->s_cluster_bits) 2498 struct ext4_group_desc *bg, ext4_fsblk_t blk); 2500 struct ext4_group_desc *bg, ext4_fsblk_t blk); 2502 struct ext4_group_desc *bg, ext4_fsblk_t blk); 2553 ext4_fsblk_t blk) ext4_blocks_count_set() 2555 es->s_blocks_count_lo = cpu_to_le32((u32)blk); ext4_blocks_count_set() 2556 es->s_blocks_count_hi = cpu_to_le32(blk >> 32); ext4_blocks_count_set() 2560 ext4_fsblk_t blk) ext4_free_blocks_count_set() 2562 es->s_free_blocks_count_lo = cpu_to_le32((u32)blk); ext4_free_blocks_count_set() 2563 es->s_free_blocks_count_hi = cpu_to_le32(blk >> 32); ext4_free_blocks_count_set() 2567 ext4_fsblk_t blk) ext4_r_blocks_count_set() 2569 es->s_r_blocks_count_lo = cpu_to_le32((u32)blk); ext4_r_blocks_count_set() 2570 es->s_r_blocks_count_hi = cpu_to_le32(blk >> 32); ext4_r_blocks_count_set() 2552 ext4_blocks_count_set(struct ext4_super_block *es, ext4_fsblk_t blk) ext4_blocks_count_set() argument 2559 ext4_free_blocks_count_set(struct ext4_super_block *es, ext4_fsblk_t blk) ext4_free_blocks_count_set() argument 2566 ext4_r_blocks_count_set(struct ext4_super_block *es, ext4_fsblk_t blk) ext4_r_blocks_count_set() argument
|
H A D | ialloc.c | 1262 ext4_fsblk_t blk; ext4_init_inode_table() local 1309 blk = ext4_inode_table(sb, gdp) + used_blks; ext4_init_inode_table() 1328 ret = sb_issue_zeroout(sb, blk, num, GFP_NOFS); ext4_init_inode_table()
|
H A D | super.c | 235 struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_block_bitmap_set() 237 bg->bg_block_bitmap_lo = cpu_to_le32((u32)blk); ext4_block_bitmap_set() 239 bg->bg_block_bitmap_hi = cpu_to_le32(blk >> 32); ext4_block_bitmap_set() 243 struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_inode_bitmap_set() 245 bg->bg_inode_bitmap_lo = cpu_to_le32((u32)blk); ext4_inode_bitmap_set() 247 bg->bg_inode_bitmap_hi = cpu_to_le32(blk >> 32); ext4_inode_bitmap_set() 251 struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_inode_table_set() 253 bg->bg_inode_table_lo = cpu_to_le32((u32)blk); ext4_inode_table_set() 255 bg->bg_inode_table_hi = cpu_to_le32(blk >> 32); ext4_inode_table_set() 5428 ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb); ext4_quota_read() local 5443 bh = ext4_bread(NULL, inode, blk, 0); ext4_quota_read() 5454 blk++; ext4_quota_read() 5465 ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb); ext4_quota_write() local 5487 bh = ext4_bread(handle, inode, blk, 1); ext4_quota_write() 234 ext4_block_bitmap_set(struct super_block *sb, struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_block_bitmap_set() argument 242 ext4_inode_bitmap_set(struct super_block *sb, struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_inode_bitmap_set() argument 250 ext4_inode_table_set(struct super_block *sb, struct ext4_group_desc *bg, ext4_fsblk_t blk) ext4_inode_table_set() argument
|
H A D | indirect.c | 545 ext4_fsblk_t blk; ext4_ind_map_blocks() local 547 blk = le32_to_cpu(*(chain[depth-1].p + count)); ext4_ind_map_blocks() 549 if (blk == first_block + count) ext4_ind_map_blocks()
|
/linux-4.1.27/include/scsi/ |
H A D | scsi_tcq.h | 31 * Only works with tags allocated by the generic blk layer. 58 * We always have a shared tag map around when using blk-mq. scsi_init_shared_tag_map() 84 * Only works with tags allocated by the generic blk layer.
|
H A D | scsi_host.h | 10 #include <linux/blk-mq.h> 504 /* temporary flag to disable blk-mq I/O path */
|
/linux-4.1.27/arch/mips/alchemy/common/ |
H A D | vss.c | 17 #define VSS_ADDR(blk) (KSEG1ADDR(AU1300_VSS_PHYS_ADDR) + (blk * 0x0c))
|
/linux-4.1.27/fs/xfs/ |
H A D | xfs_qm.h | 47 * Currently, it is just one file system block, and a 4K blk contains 30 79 xfs_qcnt_t qi_bhardlimit; /* default data blk hard limit */ 80 xfs_qcnt_t qi_bsoftlimit; /* default data blk soft limit */ 83 xfs_qcnt_t qi_rtbhardlimit;/* default realtime blk hard limit */ 84 xfs_qcnt_t qi_rtbsoftlimit;/* default realtime blk soft limit */
|
H A D | xfs_quota.h | 61 long qt_bcount_delta; /* dquot blk count changes */ 62 long qt_delbcnt_delta; /* delayed dquot blk count changes */ 66 long qt_rtbcount_delta;/* dquot realtime blk changes */ 67 long qt_delrtb_delta; /* delayed RT blk count changes */
|
H A D | xfs_trans_dquot.c | 84 * Carry forward whatever is left of the quota blk reservation to 102 * Because the quota blk reservation is carried forward, xfs_trans_dup_dqinfo() 225 * regular disk blk reservation xfs_trans_mod_dquot() 582 * if the blk reservation is for RT or regular blocks.
|
H A D | xfs_log_recover.c | 1837 void *blk = bp->b_addr; xlog_recover_get_buf_lsn() local 1845 magic32 = be32_to_cpu(*(__be32 *)blk); xlog_recover_get_buf_lsn() 1853 struct xfs_btree_block *btb = blk; xlog_recover_get_buf_lsn() 1861 struct xfs_btree_block *btb = blk; xlog_recover_get_buf_lsn() 1868 lsn = be64_to_cpu(((struct xfs_agf *)blk)->agf_lsn); xlog_recover_get_buf_lsn() 1869 uuid = &((struct xfs_agf *)blk)->agf_uuid; xlog_recover_get_buf_lsn() 1872 lsn = be64_to_cpu(((struct xfs_agfl *)blk)->agfl_lsn); xlog_recover_get_buf_lsn() 1873 uuid = &((struct xfs_agfl *)blk)->agfl_uuid; xlog_recover_get_buf_lsn() 1876 lsn = be64_to_cpu(((struct xfs_agi *)blk)->agi_lsn); xlog_recover_get_buf_lsn() 1877 uuid = &((struct xfs_agi *)blk)->agi_uuid; xlog_recover_get_buf_lsn() 1880 lsn = be64_to_cpu(((struct xfs_dsymlink_hdr *)blk)->sl_lsn); xlog_recover_get_buf_lsn() 1881 uuid = &((struct xfs_dsymlink_hdr *)blk)->sl_uuid; xlog_recover_get_buf_lsn() 1886 lsn = be64_to_cpu(((struct xfs_dir3_blk_hdr *)blk)->lsn); xlog_recover_get_buf_lsn() 1887 uuid = &((struct xfs_dir3_blk_hdr *)blk)->uuid; xlog_recover_get_buf_lsn() 1899 lsn = be64_to_cpu(((struct xfs_dsb *)blk)->sb_lsn); xlog_recover_get_buf_lsn() 1900 uuid = &((struct xfs_dsb *)blk)->sb_uuid; xlog_recover_get_buf_lsn() 1912 magicda = be16_to_cpu(((struct xfs_da_blkinfo *)blk)->magic); xlog_recover_get_buf_lsn() 1917 lsn = be64_to_cpu(((struct xfs_da3_blkinfo *)blk)->lsn); xlog_recover_get_buf_lsn() 1918 uuid = &((struct xfs_da3_blkinfo *)blk)->uuid; xlog_recover_get_buf_lsn() 1941 magic16 = be16_to_cpu(*(__be16 *)blk); xlog_recover_get_buf_lsn()
|
H A D | xfs_dir2_readdir.c | 484 xfs_dir2_off_t newoff; /* new curoff after new blk */ xfs_dir2_leaf_getdents()
|
/linux-4.1.27/drivers/gpu/drm/msm/mdp/mdp5/ |
H A D | mdp5_smp.c | 143 int blk = find_first_zero_bit(smp->state, cnt); smp_request_block() local 144 set_bit(blk, ps->pending); smp_request_block() 145 set_bit(blk, smp->state); smp_request_block() 150 int blk = find_first_bit(ps->pending, cnt); smp_request_block() local 151 clear_bit(blk, ps->pending); smp_request_block() 238 u32 blk, val; update_smp_state() local 240 for_each_set_bit(blk, *assigned, cnt) { update_smp_state() 241 int idx = blk / 3; update_smp_state() 242 int fld = blk % 3; update_smp_state()
|
/linux-4.1.27/drivers/crypto/ |
H A D | geode-aes.c | 171 op->fallback.blk->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; geode_setkey_blk() 172 op->fallback.blk->base.crt_flags |= (tfm->crt_flags & CRYPTO_TFM_REQ_MASK); geode_setkey_blk() 174 ret = crypto_blkcipher_setkey(op->fallback.blk, key, len); geode_setkey_blk() 177 tfm->crt_flags |= (op->fallback.blk->base.crt_flags & CRYPTO_TFM_RES_MASK); geode_setkey_blk() 191 desc->tfm = op->fallback.blk; fallback_blk_dec() 207 desc->tfm = op->fallback.blk; fallback_blk_enc() 371 op->fallback.blk = crypto_alloc_blkcipher(name, 0, fallback_init_blk() 374 if (IS_ERR(op->fallback.blk)) { fallback_init_blk() 376 return PTR_ERR(op->fallback.blk); fallback_init_blk() 386 crypto_free_blkcipher(op->fallback.blk); fallback_exit_blk() 387 op->fallback.blk = NULL; fallback_exit_blk()
|
H A D | geode-aes.h | 67 struct crypto_blkcipher *blk; member in union:geode_aes_op::__anon3809
|
H A D | mxs-dcp.c | 487 struct crypto_ablkcipher *blk; mxs_dcp_aes_fallback_init() local 489 blk = crypto_alloc_ablkcipher(name, 0, flags); mxs_dcp_aes_fallback_init() 490 if (IS_ERR(blk)) mxs_dcp_aes_fallback_init() 491 return PTR_ERR(blk); mxs_dcp_aes_fallback_init() 493 actx->fallback = blk; mxs_dcp_aes_fallback_init()
|
/linux-4.1.27/drivers/md/persistent-data/ |
H A D | dm-space-map-common.c | 282 struct dm_block *blk; sm_ll_lookup_bitmap() local 290 &dm_sm_bitmap_validator, &blk); sm_ll_lookup_bitmap() 294 *result = sm_lookup_bitmap(dm_bitmap_data(blk), b); sm_ll_lookup_bitmap() 296 return dm_tm_unlock(ll->tm, blk); sm_ll_lookup_bitmap() 342 struct dm_block *blk; sm_ll_find_free_block() local 354 &dm_sm_bitmap_validator, &blk); sm_ll_find_free_block() 360 r = sm_find_free(dm_bitmap_data(blk), sm_ll_find_free_block() 368 dm_tm_unlock(ll->tm, blk); sm_ll_find_free_block() 372 dm_tm_unlock(ll->tm, blk); sm_ll_find_free_block() 376 r = dm_tm_unlock(ll->tm, blk); sm_ll_find_free_block()
|
H A D | dm-transaction-manager.c | 330 struct dm_block **blk) dm_tm_read_lock() 333 int r = dm_bm_read_try_lock(tm->real->bm, b, v, blk); dm_tm_read_lock() 341 return dm_bm_read_lock(tm->bm, b, v, blk); dm_tm_read_lock() 328 dm_tm_read_lock(struct dm_transaction_manager *tm, dm_block_t b, struct dm_block_validator *v, struct dm_block **blk) dm_tm_read_lock() argument
|
/linux-4.1.27/arch/arm/mach-ixp4xx/ |
H A D | ixp4xx_npe.c | 515 } *blk; npe_load_firmware() local 619 for (i = 0, blk = image->blocks; i < blocks; i++, blk++) { npe_load_firmware() 620 if (blk->offset > image->size - sizeof(struct dl_codeblock) / 4 npe_load_firmware() 621 || blk->offset < table_end) { npe_load_firmware() 623 "firmware block #%i\n", blk->offset, i); npe_load_firmware() 627 cb = (struct dl_codeblock*)&image->data[blk->offset]; npe_load_firmware() 628 if (blk->type == FW_BLOCK_TYPE_INSTR) { npe_load_firmware() 632 } else if (blk->type == FW_BLOCK_TYPE_DATA) { npe_load_firmware() 638 "type 0x%X\n", i, blk->type); npe_load_firmware() 641 if (blk->offset + sizeof(*cb) / 4 + cb->size > image->size) { npe_load_firmware() 645 blk->type == FW_BLOCK_TYPE_INSTR ? 'I' : 'D', npe_load_firmware() 663 blk->type == FW_BLOCK_TYPE_INSTR ? 'I' : 'D', npe_load_firmware()
|
/linux-4.1.27/drivers/mtd/ |
H A D | ftl.c | 729 uint32_t blk; find_free() local 764 for (blk = 0; blk < part->BlocksPerUnit; blk++) find_free() 765 if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break; find_free() 766 if (blk == part->BlocksPerUnit) { find_free() 775 pr_debug("ftl_cs: found free block at %d in %d\n", blk, eun); find_free() 776 return blk; find_free() 836 uint32_t bsize, blk, le_virt_addr; set_bam_entry() local 848 blk = (log_addr % bsize) / SECTOR_SIZE; set_bam_entry() 849 offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) + set_bam_entry() 876 if (le32_to_cpu(part->bam_cache[blk]) != old_addr) { set_bam_entry() 883 le32_to_cpu(part->bam_cache[blk]), old_addr); set_bam_entry() 888 part->bam_cache[blk] = le_virt_addr; set_bam_entry() 904 uint32_t bsize, log_addr, virt_addr, old_addr, blk; ftl_write() local 932 blk = find_free(part); ftl_write() 933 if (blk == 0) { ftl_write() 942 log_addr = part->bam_index * bsize + blk * SECTOR_SIZE; ftl_write() 949 blk * SECTOR_SIZE); ftl_write()
|
/linux-4.1.27/arch/s390/crypto/ |
H A D | aes_s390.c | 45 struct crypto_blkcipher *blk; member in union:s390_aes_ctx::__anon2423 238 sctx->fallback.blk->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; setkey_fallback_blk() 239 sctx->fallback.blk->base.crt_flags |= (tfm->crt_flags & setkey_fallback_blk() 242 ret = crypto_blkcipher_setkey(sctx->fallback.blk, key, len); setkey_fallback_blk() 245 tfm->crt_flags |= (sctx->fallback.blk->base.crt_flags & setkey_fallback_blk() 260 desc->tfm = sctx->fallback.blk; fallback_blk_dec() 277 desc->tfm = sctx->fallback.blk; fallback_blk_enc() 371 sctx->fallback.blk = crypto_alloc_blkcipher(name, 0, fallback_init_blk() 374 if (IS_ERR(sctx->fallback.blk)) { fallback_init_blk() 377 return PTR_ERR(sctx->fallback.blk); fallback_init_blk() 387 crypto_free_blkcipher(sctx->fallback.blk); fallback_exit_blk() 388 sctx->fallback.blk = NULL; fallback_exit_blk()
|
/linux-4.1.27/arch/sparc/include/uapi/asm/ |
H A D | asi.h | 238 #define ASI_BLK_AIUPL 0x78 /* Primary, user, little, blk ld/st*/ 239 #define ASI_BLK_AIUSL 0x79 /* Secondary, user, little, blk ld/st*/ 268 #define ASI_BLK_COMMIT_P 0xe0 /* Primary, blk store commit */ 269 #define ASI_BLK_COMMIT_S 0xe1 /* Secondary, blk store commit */ 276 #define ASI_BLK_P 0xf0 /* Primary, blk ld/st */ 277 #define ASI_BLK_S 0xf1 /* Secondary, blk ld/st */ 286 #define ASI_BLK_PL 0xf8 /* Primary, blk ld/st, little */ 287 #define ASI_BLK_SL 0xf9 /* Secondary, blk ld/st, little */
|
/linux-4.1.27/drivers/edac/ |
H A D | edac_device.c | 78 struct edac_device_block *dev_blk, *blk_p, *blk; edac_device_alloc_ctl_info() local 176 blk = &blk_p[block]; edac_device_alloc_ctl_info() 177 blk->instance = inst; edac_device_alloc_ctl_info() 178 snprintf(blk->name, sizeof(blk->name), edac_device_alloc_ctl_info() 182 instance, inst, block, blk, blk->name); edac_device_alloc_ctl_info() 191 blk->nr_attribs = nr_attrib; edac_device_alloc_ctl_info() 193 blk->block_attributes = attrib_p; edac_device_alloc_ctl_info() 196 blk->block_attributes); edac_device_alloc_ctl_info() 213 attrib->block = blk; /* up link */ edac_device_alloc_ctl_info()
|
/linux-4.1.27/sound/soc/codecs/ |
H A D | wm_adsp.c | 1179 struct wmfw_coeff_item *blk; wm_adsp_load_coeff() local 1235 pos - firmware->size > sizeof(*blk)) { wm_adsp_load_coeff() 1236 blk = (void*)(&firmware->data[pos]); wm_adsp_load_coeff() 1238 type = le16_to_cpu(blk->type); wm_adsp_load_coeff() 1239 offset = le16_to_cpu(blk->offset); wm_adsp_load_coeff() 1242 file, blocks, le32_to_cpu(blk->id), wm_adsp_load_coeff() 1243 (le32_to_cpu(blk->ver) >> 16) & 0xff, wm_adsp_load_coeff() 1244 (le32_to_cpu(blk->ver) >> 8) & 0xff, wm_adsp_load_coeff() 1245 le32_to_cpu(blk->ver) & 0xff); wm_adsp_load_coeff() 1247 file, blocks, le32_to_cpu(blk->len), offset, type); wm_adsp_load_coeff() 1260 if (le32_to_cpu(blk->id) == dsp->fw_id && wm_adsp_load_coeff() 1281 file, blocks, le32_to_cpu(blk->len), wm_adsp_load_coeff() 1282 type, le32_to_cpu(blk->id)); wm_adsp_load_coeff() 1293 if (le32_to_cpu(blk->id) == alg_region->alg && wm_adsp_load_coeff() 1305 type, le32_to_cpu(blk->id)); wm_adsp_load_coeff() 1315 buf = wm_adsp_buf_alloc(blk->data, wm_adsp_load_coeff() 1316 le32_to_cpu(blk->len), wm_adsp_load_coeff() 1325 file, blocks, le32_to_cpu(blk->len), wm_adsp_load_coeff() 1328 le32_to_cpu(blk->len)); wm_adsp_load_coeff() 1336 pos += (le32_to_cpu(blk->len) + sizeof(*blk) + 3) & ~0x03; wm_adsp_load_coeff()
|
/linux-4.1.27/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_hw.h | 50 * Bit 16: tx_reset_pb => 1:reset frame xmit protocol blk, 0:no-op 51 * Bit 17: rx_reset_pb => 1:reset frame recv protocol blk, 0:no-op 52 * Bit 18: tx_reset_mac => 1:reset data/ctl multiplexer blk, 0:no-op 53 * Bit 19: rx_reset_mac => 1:reset ctl frames & timers blk, 0:no-op
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
H A D | isph3a_aewb.c | 34 u32 blk; h3a_aewb_setup_regs() local 58 blk = conf->blk_ver_win_start << ISPH3A_AEWINBLK_WINSV_SHIFT; h3a_aewb_setup_regs() 59 blk |= ((conf->blk_win_height >> 1) - 1) << ISPH3A_AEWINBLK_WINH_SHIFT; h3a_aewb_setup_regs() 69 isp_reg_writel(aewb->isp, blk, OMAP3_ISP_IOMEM_H3A, ISPH3A_AEWINBLK); h3a_aewb_setup_regs()
|
/linux-4.1.27/drivers/pci/hotplug/ |
H A D | ibmphp_ebda.c | 284 next_offset = readw (io_mem + offset); /* offset of next blk */ ibmphp_access_ebda() 287 if (next_offset == 0) /* 0 indicate it's last blk */ ibmphp_access_ebda() 289 blk_id = readw (io_mem + offset); /* this blk id */ ibmphp_access_ebda() 298 debug ("hot blk id: %x\n", blk_id); ibmphp_access_ebda() 304 debug ("hot blk format: %x\n", format); ibmphp_access_ebda() 305 /* hot swap sub blk */ ibmphp_access_ebda() 309 re = readw (io_mem + sub_addr); /* next sub blk */ ibmphp_access_ebda() 312 rc_id = readw (io_mem + sub_addr); /* sub blk id */ ibmphp_access_ebda() 317 /* rc sub blk signature */ ibmphp_access_ebda() 328 hpc_list_ptr->phys_addr = sub_addr; /* offset of RSRC_CONTROLLER blk */ ibmphp_access_ebda() 330 debug ("hot blk format: %x\n", format); ibmphp_access_ebda() 334 sub_addr = base + re; /* re sub blk */ ibmphp_access_ebda() 336 rc = readw (io_mem + sub_addr); /* next sub blk */ ibmphp_access_ebda() 339 re_id = readw (io_mem + sub_addr); /* sub blk id */ ibmphp_access_ebda() 348 sub_addr += 2; /* offset of RSRC_ENTRIES blk */ ibmphp_access_ebda() 367 debug ("rio blk id: %x\n", blk_id); ibmphp_access_ebda()
|
/linux-4.1.27/drivers/block/ |
H A D | loop.h | 14 #include <linux/blk-mq.h>
|
H A D | ida_ioctl.h | 55 __u32 blk; member in struct:__anon3601
|
H A D | cpqarray.c | 158 unsigned int blk, 929 c->req.hdr.blk = blk_rq_pos(creq); do_ida_request() 1019 cmd->req.hdr.blk, cmd->req.hdr.blk_cnt, complete_command() 1250 c->req.hdr.blk = io->blk; ida_ctlr_ioctl() 1263 c->req.hdr.blk = pci_map_single(h->pci_dev, &(io->c), ida_ctlr_ioctl() 1327 pci_unmap_single(h->pci_dev, c->req.hdr.blk, ida_ctlr_ioctl() 1412 unsigned int blk, sendcmd() 1443 c->req.hdr.blk = blk; sendcmd() 1407 sendcmd( __u8 cmd, int ctlr, void *buff, size_t size, unsigned int blk, unsigned int blkcnt, unsigned int log_unit ) sendcmd() argument
|
H A D | virtio_blk.c | 14 #include <linux/blk-mq.h> 446 * virtio blk but don't ever use it for any new driver. 868 virtblk_wq = alloc_workqueue("virtio-blk", 0, 0); init()
|
H A D | ida_cmd.h | 64 __u32 blk; member in struct:__anon3582
|
H A D | null_blk.c | 9 #include <linux/blk-mq.h>
|
H A D | osdblk.c | 97 struct request *rq; /* blk layer request */
|
H A D | mg_disk.c | 1055 /* unregister blk device */ mg_remove()
|
H A D | sunvdc.c | 167 /* restart blk queue when ring is half emptied. also called after vdc_blk_queue_start()
|
H A D | nvme-core.c | 18 #include <linux/blk-mq.h> 222 * blk-mq map the tags so they match up with the nvme queue tags. */ nvme_init_hctx()
|
/linux-4.1.27/fs/squashfs/ |
H A D | namei.c | 213 unsigned int blk, off, ino_num; squashfs_lookup() local 215 blk = le32_to_cpu(dirh.start_block); squashfs_lookup() 219 ino = SQUASHFS_MKINODE(blk, off); squashfs_lookup() 223 blk, off, ino_num); squashfs_lookup()
|
H A D | export.c | 55 int blk = SQUASHFS_LOOKUP_BLOCK(ino_num - 1); squashfs_inode_lookup() local 57 u64 start = le64_to_cpu(msblk->inode_lookup_table[blk]); squashfs_inode_lookup()
|
/linux-4.1.27/drivers/base/regmap/ |
H A D | regcache-rbtree.c | 287 u8 *blk; regcache_rbtree_insert_to_block() local 293 blk = krealloc(rbnode->block, regcache_rbtree_insert_to_block() 296 if (!blk) regcache_rbtree_insert_to_block() 304 kfree(blk); regcache_rbtree_insert_to_block() 317 memmove(blk + offset * map->cache_word_size, regcache_rbtree_insert_to_block() 318 blk, rbnode->blklen * map->cache_word_size); regcache_rbtree_insert_to_block() 323 rbnode->block = blk; regcache_rbtree_insert_to_block()
|
/linux-4.1.27/fs/ufs/ |
H A D | util.h | 342 u64 blk) ubh_get_data_ptr() 345 return ubh_get_addr64(ubh, blk); ubh_get_data_ptr() 347 return ubh_get_addr32(ubh, blk); ubh_get_data_ptr() 553 unsigned blk) ufs_get_direct_data_ptr() 555 BUG_ON(blk > UFS_TIND_BLOCK); ufs_get_direct_data_ptr() 557 (void *)&ufsi->i_u1.u2_i_data[blk] : ufs_get_direct_data_ptr() 558 (void *)&ufsi->i_u1.i_data[blk]; ufs_get_direct_data_ptr() 340 ubh_get_data_ptr(struct ufs_sb_private_info *uspi, struct ufs_buffer_head *ubh, u64 blk) ubh_get_data_ptr() argument 551 ufs_get_direct_data_ptr(struct ufs_sb_private_info *uspi, struct ufs_inode_info *ufsi, unsigned blk) ufs_get_direct_data_ptr() argument
|
H A D | ufs_fs.h | 217 #define ufs_cgsblock(c) (ufs_cgstart(c) + uspi->s_sblkno) /* super blk */ 219 #define ufs_cgimin(c) (ufs_cgstart(c) + uspi->s_iblkno) /* inode blk */ 256 #define ufs_lblktosize(blk) ((blk) << uspi->s_bshift) 372 __fs32 fs_bmask; /* ``blkoff'' calc of blk offsets */ 407 __fs32 fs_csaddr; /* blk addr of cyl grp summary area */ 455 __fs64 fs_csaddr; /* blk addr of cyl grp summary area */ 745 __u32 s_bmask; /* `blkoff'' calc of blk offsets */ 760 __u64 s_csaddr; /* blk addr of cyl grp summary area */ 919 __fs64 fs_csaddr; /* blk addr of cyl grp summary area */
|
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/ |
H A D | brw_test.c | 363 srpc_bulk_t *blk = rpc->srpc_bulk; brw_server_rpc_done() local 365 if (blk == NULL) brw_server_rpc_done() 370 blk->bk_sink ? "from" : "to", brw_server_rpc_done() 374 blk->bk_niov, blk->bk_sink ? "from" : "to", brw_server_rpc_done()
|
/linux-4.1.27/fs/ext3/ |
H A D | resize.c | 113 ext3_fsblk_t blk) bclean() 118 bh = sb_getblk(sb, blk); bclean() 389 const ext3_fsblk_t blk = primary->b_blocknr; verify_reserved_gdb() local 399 if (le32_to_cpu(*p++) != grp * EXT3_BLOCKS_PER_GROUP(sb) + blk){ verify_reserved_gdb() 403 blk, grp, verify_reserved_gdb() 404 grp * EXT3_BLOCKS_PER_GROUP(sb) + blk); verify_reserved_gdb() 587 ext3_fsblk_t blk; reserve_backup_gdb() local 604 blk = EXT3_SB(sb)->s_sbh->b_blocknr + 1 + EXT3_SB(sb)->s_gdb_count; reserve_backup_gdb() 610 for (res = 0; res < reserved_gdb; res++, blk++) { reserve_backup_gdb() 611 if (le32_to_cpu(*data) != blk) { reserve_backup_gdb() 615 blk, reserve_backup_gdb() 620 primary[res] = sb_bread(sb, blk); reserve_backup_gdb() 652 blk = input->group * EXT3_BLOCKS_PER_GROUP(sb); reserve_backup_gdb() 658 blk + primary[i]->b_blocknr); */ reserve_backup_gdb() 659 data[gdbackups] = cpu_to_le32(blk + primary[i]->b_blocknr); reserve_backup_gdb() 112 bclean(handle_t *handle, struct super_block *sb, ext3_fsblk_t blk) bclean() argument
|
H A D | dir.c | 116 unsigned long blk = ctx->pos >> EXT3_BLOCK_SIZE_BITS(sb); ext3_readdir() local 121 err = ext3_get_blocks_handle(NULL, inode, blk, 1, &map_bh, 0); ext3_readdir() 131 bh = ext3_bread(NULL, inode, blk, 0, &err); ext3_readdir()
|
H A D | super.c | 3020 sector_t blk = off >> EXT3_BLOCK_SIZE_BITS(sb); ext3_quota_read() local 3036 bh = ext3_bread(NULL, inode, blk, 0, &err); ext3_quota_read() 3047 blk++; ext3_quota_read() 3058 sector_t blk = off >> EXT3_BLOCK_SIZE_BITS(sb); ext3_quota_write() local 3083 bh = ext3_bread(handle, inode, blk, 1, &err); ext3_quota_write()
|
/linux-4.1.27/drivers/net/wireless/rsi/ |
H A D | rsi_91x_usb_ops.c | 54 "%s: Unable to load %s blk\n", __func__, rsi_copy_to_card() 69 "%s: Unable to load %s blk\n", __func__, rsi_copy_to_card()
|
H A D | rsi_91x_sdio_ops.c | 103 "%s: Unable to load %s blk\n", __func__, rsi_copy_to_card()
|
H A D | rsi_91x_sdio.c | 370 "%s: Operational blk length is %d\n", __func__, length); rsi_setblocklength()
|
H A D | rsi_91x_mac80211.c | 368 rsi_dbg(INFO_ZONE, "blk data q %d\n", channel); rsi_channel_change()
|
/linux-4.1.27/drivers/scsi/ |
H A D | sr.h | 46 /* GET_EVENT spurious event handling, blk layer guarantees exclusion */
|
H A D | osst.c | 1385 "%s:D: Positioned at ppos %d, frame %d, lbn %d, file %d, blk %d, %cptr %d, eof %d\n", osst_get_sector() 1457 "%s:D: Now positioned at ppos %d, frame %d, lbn %d, file %d, blk %d, rptr %d, eof %d\n", osst_seek_sector() 1863 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks_bwd\n", name); osst_space_over_filemarks_backward() 1897 "%s:D: Couldn't get logical blk num in space_filemarks\n", name); osst_space_over_filemarks_backward() 1924 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks\n", name); osst_space_over_filemarks_backward() 1961 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks_fwd\n", name); osst_space_over_filemarks_forward_slow() 1968 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks\n", name); osst_space_over_filemarks_forward_slow() 2016 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks_fwd\n", name); osst_space_over_filemarks_forward_fast() 2055 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks\n", osst_space_over_filemarks_forward_fast() 2096 "%s:D: Couldn't get logical blk num in space_filemarks_fwd_fast\n", osst_space_over_filemarks_forward_fast() 2128 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in space_filemarks\n", osst_space_over_filemarks_forward_fast() 2626 printk(OSST_DEB_MSG "%s:D: Couldn't get logical blk num in verify_position\n", name); osst_verify_position() 4945 int i, cmd_nr, cmd_type, blk, retval = 0; osst_ioctl() 5252 blk = osst_get_frame_position(STp, &SRpnt); osst_ioctl() 5254 blk = osst_get_sector(STp, &SRpnt); osst_ioctl() 5255 if (blk < 0) { osst_ioctl() 5256 retval = blk; osst_ioctl() 5259 mt_pos.mt_blkno = blk; osst_ioctl() 4941 int i, cmd_nr, cmd_type, blk, retval = 0; osst_ioctl() local
|
H A D | st.c | 3106 DEBC_printk(STp, "Got tape pos. blk %d part %d.\n", set_location() 3123 unsigned int blk; set_location() 3143 if (get_location(STp, &blk, &p, 1)) set_location() 3147 STps->last_block_visited = blk; set_location() 3150 blk, STp->partition); set_location() 3368 unsigned int blk; st_ioctl() 3675 if ((i = get_location(STp, &blk, &bt, 0)) < 0) { st_ioctl() 3679 mt_pos.mt_blkno = blk; st_ioctl() 3110 unsigned int blk; set_location() local 3353 unsigned int blk; st_ioctl() local
|
H A D | ipr.h | 35 #include <linux/blk-iopoll.h>
|
/linux-4.1.27/fs/f2fs/ |
H A D | data.c | 262 map_bh(bh_result, sb, ei->blk + pgofs - ei->fofs); f2fs_map_bh() 284 start_blkaddr = fi->ext.blk; lookup_extent_info() 308 start_blkaddr = fi->ext.blk; update_extent_info() 309 end_blkaddr = fi->ext.blk + fi->ext.len - 1; update_extent_info() 319 fi->ext.blk = blkaddr; update_extent_info() 328 fi->ext.blk--; update_extent_info() 346 fi->ext.blk = start_blkaddr + fofs - start_fofs + 1; update_extent_info() 479 en->ei.blk = prev->ei.blk; __try_back_merge() 522 en->ei.blk = ei->blk; __insert_extent_tree() 591 le32_to_cpu(i_ext->blk), le32_to_cpu(i_ext->len)); f2fs_init_extent_tree() 666 set_extent_info(&ei, dei.fofs, dei.blk, f2fs_update_extent_tree() 675 fofs - dei.fofs + dei.blk, endofs - fofs); f2fs_update_extent_tree() 934 dn.data_blkaddr = ei.blk + index - ei.fofs; find_data_page() 998 dn.data_blkaddr = ei.blk + index - ei.fofs; get_lock_data_page()
|
H A D | f2fs.h | 298 u32 blk; /* start block address of the extent */ member in struct:extent_info 355 ext->blk = le32_to_cpu(i_ext.blk); get_extent_info() 363 i_ext->blk = cpu_to_le32(ext->blk); set_raw_extent() 368 u32 blk, unsigned int len) set_extent_info() 371 ei->blk = blk; set_extent_info() 378 return (ei1->fofs == ei2->fofs && ei1->blk == ei2->blk && __is_extent_same() 386 back->blk + back->len == front->blk); __is_extent_mergeable() 367 set_extent_info(struct extent_info *ei, unsigned int fofs, u32 blk, unsigned int len) set_extent_info() argument
|
/linux-4.1.27/drivers/usb/storage/ |
H A D | ene_ub6250.c | 1203 u16 blk, index = 0; ms_lib_check_disableblock() local 1215 blk = be16_to_cpu(PageBuf[index]); ms_lib_check_disableblock() 1216 if (blk == MS_LB_NOT_USED) ms_lib_check_disableblock() 1218 if (blk == info->MS_Lib.Log2PhyMap[0]) { ms_lib_check_disableblock() 1373 u16 blk; ms_libsearch_block_from_physical() local 1381 for (blk = phyblk + 1; blk != phyblk; blk++) { ms_libsearch_block_from_physical() 1382 if ((blk & MS_PHYSICAL_BLOCKS_PER_SEGMENT_MASK) == 0) ms_libsearch_block_from_physical() 1383 blk -= MS_PHYSICAL_BLOCKS_PER_SEGMENT; ms_libsearch_block_from_physical() 1385 Newblk = info->MS_Lib.Phy2LogMap[blk]; ms_libsearch_block_from_physical() 1386 if (info->MS_Lib.Phy2LogMap[blk] == MS_LB_NOT_USED_ERASED) { ms_libsearch_block_from_physical() 1387 return blk; ms_libsearch_block_from_physical() 1388 } else if (info->MS_Lib.Phy2LogMap[blk] == MS_LB_NOT_USED) { ms_libsearch_block_from_physical() 1389 switch (ms_lib_read_extra(us, blk, 0, &extdat)) { ms_libsearch_block_from_physical() 1399 ms_lib_setacquired_errorblock(us, blk); ms_libsearch_block_from_physical() 1404 ms_lib_setacquired_errorblock(us, blk); ms_libsearch_block_from_physical() 1408 switch (ms_lib_erase_phyblock(us, blk)) { ms_libsearch_block_from_physical() 1410 return blk; ms_libsearch_block_from_physical() 1415 ms_lib_error_phyblock(us, blk); ms_libsearch_block_from_physical()
|
/linux-4.1.27/drivers/scsi/megaraid/ |
H A D | megaraid_sas_fp.c | 326 u64 blk, debugBlk; MR_GetSpanBlock() local 327 blk = mega_div64_32((row-le64_to_cpu(quad->logStart)), le32_to_cpu(quad->diff)); MR_GetSpanBlock() 328 debugBlk = blk; MR_GetSpanBlock() 330 blk = (blk + le64_to_cpu(quad->offsetInSpan)) << raid->stripeShift; MR_GetSpanBlock() 331 *span_blk = blk; MR_GetSpanBlock() 475 u64 blk; mr_spanset_get_span_block() local 476 blk = mega_div64_32 mr_spanset_get_span_block() 479 blk = (blk + le64_to_cpu(quad->offsetInSpan)) mr_spanset_get_span_block() 481 *span_blk = blk; mr_spanset_get_span_block()
|
/linux-4.1.27/include/trace/events/ |
H A D | ext3.h | 321 TP_PROTO(struct super_block *sb, unsigned long blk, 324 TP_ARGS(sb, blk, count), 328 __field( unsigned long, blk ) 335 __entry->blk = blk; 339 TP_printk("dev %d,%d blk %lu count %lu", 341 __entry->blk, __entry->count)
|
H A D | f2fs.h | 1108 __field(u32, blk) 1117 __entry->blk = en->ei.blk; 1122 "ext_info(fofs: %u, blk: %u, len: %u)", 1126 __entry->blk, 1140 __field(u32, blk) 1147 __entry->blk = blkaddr; 1153 __entry->blk)
|
H A D | ext4.h | 587 TP_PROTO(struct super_block *sb, unsigned long long blk, 590 TP_ARGS(sb, blk, count), 594 __field( __u64, blk ) 601 __entry->blk = blk; 605 TP_printk("dev %d,%d blk %llu count %llu", 607 __entry->blk, __entry->count) 1992 "(blk = %u)",
|
/linux-4.1.27/drivers/media/platform/s5p-tv/ |
H A D | mixer_video.c | 264 unsigned long mxr_get_plane_size(const struct mxr_block *blk, mxr_get_plane_size() argument 267 unsigned int bl_width = divup(width, blk->width); mxr_get_plane_size() 268 unsigned int bl_height = divup(height, blk->height); mxr_get_plane_size() 270 return bl_width * bl_height * blk->size; mxr_get_plane_size() 286 const struct mxr_block *blk = &fmt->plane[i]; mxr_mplane_fill() local 287 u32 bl_width = divup(width, blk->width); mxr_mplane_fill() 288 u32 bl_height = divup(height, blk->height); mxr_mplane_fill() 289 u32 sizeimage = bl_width * bl_height * blk->size; mxr_mplane_fill() 290 u32 bytesperline = bl_width * blk->size / blk->height; mxr_mplane_fill()
|
H A D | mixer.h | 311 unsigned long mxr_get_plane_size(const struct mxr_block *blk,
|
/linux-4.1.27/fs/jfs/ |
H A D | super.c | 745 sector_t blk = off >> sb->s_blocksize_bits; jfs_quota_read() local 765 err = jfs_get_block(inode, blk, &tmp_bh, 0); jfs_quota_read() 780 blk++; jfs_quota_read() 790 sector_t blk = off >> sb->s_blocksize_bits; jfs_quota_write() local 805 err = jfs_get_block(inode, blk, &tmp_bh, 1); jfs_quota_write() 826 blk++; jfs_quota_write()
|
/linux-4.1.27/include/linux/ |
H A D | f2fs_fs.h | 20 #define F2FS_LOG_SECTORS_PER_BLOCK 3 /* log number for sector/blk */ 30 #define F2FS_BLK_TO_BYTES(blk) ((blk) << F2FS_BLKSIZE_BITS) 156 __le32 blk; /* start block address of the extent */ member in struct:f2fs_extent
|
H A D | nvme.h | 21 #include <linux/blk-mq.h>
|
H A D | quota.h | 274 #define DQ_BLKS_B 1 /* uid/gid has been warned about blk limit */
|
H A D | cgroup.h | 173 * - blkcg: blk-throttle becomes properly hierarchical.
|
H A D | genhd.h | 415 /* block/blk-core.c */
|
H A D | blkdev.h | 628 * request_fn defined, or is blk-mq based. 1093 struct list_head mq_list; /* blk-mq requests */ blk_post_runtime_resume()
|
/linux-4.1.27/drivers/staging/ft1000/ft1000-pcmcia/ |
H A D | ft1000_hw.c | 1609 struct pseudo_hdr blk; ft1000_copy_down_pkt() member in union:__anon9792 1637 pseudo.blk.length = len; ft1000_copy_down_pkt() 1639 pseudo.blk.length = ntohs(len); ft1000_copy_down_pkt() 1641 pseudo.blk.source = DSPID; /* Need to swap to get in correct ft1000_copy_down_pkt() 1643 pseudo.blk.destination = HOSTID; ft1000_copy_down_pkt() 1644 pseudo.blk.portdest = NETWORKID; /* Need to swap to get in ft1000_copy_down_pkt() 1646 pseudo.blk.portsrc = DSPAIRID; ft1000_copy_down_pkt() 1647 pseudo.blk.sh_str_id = 0; ft1000_copy_down_pkt() 1648 pseudo.blk.control = 0; ft1000_copy_down_pkt() 1649 pseudo.blk.rsvd1 = 0; ft1000_copy_down_pkt() 1650 pseudo.blk.seq_num = 0; ft1000_copy_down_pkt() 1651 pseudo.blk.rsvd2 = pcmcia->packetseqnum++; ft1000_copy_down_pkt() 1652 pseudo.blk.qos_class = 0; ft1000_copy_down_pkt() 1654 pseudo.blk.checksum = pseudo.buff[0]; ft1000_copy_down_pkt() 1656 pseudo.blk.checksum ^= pseudo.buff[i]; ft1000_copy_down_pkt()
|
/linux-4.1.27/drivers/lguest/ |
H A D | interrupts_and_traps.c | 200 DECLARE_BITMAP(blk, LGUEST_IRQS); interrupt_pending() 208 * wants blocked: the result ends up in "blk". interrupt_pending() 210 if (copy_from_user(&blk, cpu->lg->lguest_data->blocked_interrupts, interrupt_pending() 211 sizeof(blk))) interrupt_pending() 213 bitmap_andnot(blk, cpu->irqs_pending, blk, LGUEST_IRQS); interrupt_pending() 216 irq = find_first_bit(blk, LGUEST_IRQS); interrupt_pending() 217 *more = find_next_bit(blk, LGUEST_IRQS, irq+1); interrupt_pending()
|
/linux-4.1.27/drivers/scsi/device_handler/ |
H A D | scsi_dh_hp_sw.c | 61 * @errors: blk error code 163 * @errors: blk error code
|
/linux-4.1.27/drivers/accessibility/braille/ |
H A D | braille_console.c | 156 static int keyboard_notifier_call(struct notifier_block *blk, keyboard_notifier_call() argument 275 static int vt_notifier_call(struct notifier_block *blk, vt_notifier_call() argument
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_hdr.h | 812 * Bit 16: tx_reset_pb => 1:reset frame xmit protocol blk, 0:no-op 813 * Bit 17: rx_reset_pb => 1:reset frame recv protocol blk, 0:no-op 814 * Bit 18: tx_reset_mac => 1:reset data/ctl multiplexer blk, 0:no-op 815 * Bit 19: rx_reset_mac => 1:reset ctl frames & timers blk, 0:no-op
|
/linux-4.1.27/drivers/media/platform/ti-vpe/ |
H A D | vpdma.c | 481 struct vpdma_buf *blk, u32 dest_offset) vpdma_add_cfd_block() 484 int len = blk->size; vpdma_add_cfd_block() 486 WARN_ON(blk->dma_addr & VPDMA_DESC_ALIGN); vpdma_add_cfd_block() 493 cfd->payload_addr = (u32) blk->dma_addr; vpdma_add_cfd_block() 480 vpdma_add_cfd_block(struct vpdma_desc_list *list, int client, struct vpdma_buf *blk, u32 dest_offset) vpdma_add_cfd_block() argument
|
H A D | vpdma.h | 184 struct vpdma_buf *blk, u32 dest_offset);
|
/linux-4.1.27/drivers/media/usb/cx231xx/ |
H A D | cx231xx-pcb-cfg.h | 92 present (requires i2s blk) */
|
/linux-4.1.27/drivers/scsi/be2iscsi/ |
H A D | be.h | 23 #include <linux/blk-iopoll.h>
|
/linux-4.1.27/drivers/staging/skein/ |
H A D | skein_base.h | 161 #define SKEIN_T1_POS_FIRST SKEIN_T1_BIT(126) /* 126 first blk flag */ 162 #define SKEIN_T1_POS_FINAL SKEIN_T1_BIT(127) /* 127 final blk flag */
|
/linux-4.1.27/fs/ubifs/ |
H A D | journal.c | 704 dbg_jnlk(key, "ino %lu, blk %u, len %d, key ", ubifs_jnl_write_data() 1155 unsigned int blk; ubifs_jnl_truncate() local 1180 blk = new_size >> UBIFS_BLOCK_SHIFT; ubifs_jnl_truncate() 1181 data_key_init(c, &key, inum, blk); ubifs_jnl_truncate() 1244 blk = (new_size >> UBIFS_BLOCK_SHIFT) + (bit ? 1 : 0); ubifs_jnl_truncate() 1245 data_key_init(c, &key, inum, blk); ubifs_jnl_truncate() 1248 blk = (old_size >> UBIFS_BLOCK_SHIFT) - (bit ? 0 : 1); ubifs_jnl_truncate() 1249 data_key_init(c, &to_key, inum, blk); ubifs_jnl_truncate()
|
/linux-4.1.27/fs/ocfs2/ |
H A D | quota_global.c | 172 sector_t blk = off >> sb->s_blocksize_bits; ocfs2_quota_read() local 186 err = ocfs2_extent_map_get_blocks(gqinode, blk, &pblock, ocfs2_quota_read() 207 blk++; ocfs2_quota_read() 221 sector_t blk = off >> sb->s_blocksize_bits; ocfs2_quota_write() local 250 err = ocfs2_extent_map_get_blocks(gqinode, blk, &pblock, &pcount, NULL); ocfs2_quota_write()
|
H A D | ocfs2_trace.h | 692 TP_PROTO(struct super_block *sb, unsigned long long blk, 694 TP_ARGS(sb, blk, count), 698 __field(unsigned long long, blk) 704 __entry->blk = blk; 709 __entry->blk, __entry->count)
|
H A D | dir.c | 1840 unsigned long offset, blk, last_ra_blk = 0; ocfs2_dir_foreach_blk_el() local 1853 blk = ctx->pos >> sb->s_blocksize_bits; ocfs2_dir_foreach_blk_el() 1854 if (ocfs2_read_dir_block(inode, blk, &bh, 0)) { ocfs2_dir_foreach_blk_el() 1866 || (((last_ra_blk - blk) << 9) <= (ra_sectors / 2))) { ocfs2_dir_foreach_blk_el() 1870 if (!ocfs2_read_dir_block(inode, ++blk, &tmp, ocfs2_dir_foreach_blk_el() 1874 last_ra_blk = blk; ocfs2_dir_foreach_blk_el() 4363 u64 blk; ocfs2_dx_dir_remove_index() local 4408 blk = le64_to_cpu(dx_root->dr_blkno); ocfs2_dx_dir_remove_index() 4413 bg_blkno = ocfs2_which_suballoc_group(blk, bit); ocfs2_dx_dir_remove_index()
|
H A D | move_extents.c | 499 * moving goal is not allowd to start with a group desc blok(#0 blk) ocfs2_validate_and_adjust_move_goal()
|
H A D | xattr.c | 2481 u64 blk, bg_blkno; ocfs2_xattr_free_block() local 2497 blk = le64_to_cpu(xb->xb_blkno); ocfs2_xattr_free_block() 2502 bg_blkno = ocfs2_which_suballoc_group(blk, bit); ocfs2_xattr_free_block() 4610 * Move some xattrs in old bucket(blk) to new bucket(new_blk). 4621 u64 blk, ocfs2_divide_xattr_bucket() 4633 trace_ocfs2_divide_xattr_bucket_begin((unsigned long long)blk, ocfs2_divide_xattr_bucket() 4644 ret = ocfs2_read_xattr_bucket(s_bucket, blk); ocfs2_divide_xattr_bucket() 4619 ocfs2_divide_xattr_bucket(struct inode *inode, handle_t *handle, u64 blk, u64 new_blk, u32 *first_hash, int new_bucket_head) ocfs2_divide_xattr_bucket() argument
|
H A D | suballoc.c | 2537 u64 blk = le64_to_cpu(di->i_blkno); ocfs2_free_dinode() local 2539 u64 bg_blkno = ocfs2_which_suballoc_group(blk, bit); ocfs2_free_dinode()
|
/linux-4.1.27/fs/ext2/ |
H A D | super.c | 1446 sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb); ext2_quota_read() local 1466 err = ext2_get_block(inode, blk, &tmp_bh, 0); ext2_quota_read() 1481 blk++; ext2_quota_read() 1491 sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb); ext2_quota_write() local 1505 err = ext2_get_block(inode, blk, &tmp_bh, 1); ext2_quota_write() 1526 blk++; ext2_quota_write()
|
H A D | inode.c | 649 ext2_fsblk_t blk; ext2_get_blocks() local 662 blk = le32_to_cpu(*(chain[depth-1].p + count)); ext2_get_blocks() 663 if (blk == first_block + count) ext2_get_blocks()
|
/linux-4.1.27/arch/parisc/kernel/ |
H A D | cache.c | 157 printk("dc_conf = 0x%lx alias %d blk %d line %d shift %d\n", parisc_cache_init() 175 printk("ic_conf = 0x%lx alias %d blk %d line %d shift %d\n", parisc_cache_init()
|
/linux-4.1.27/drivers/md/ |
H A D | dm.c | 26 #include <linux/blk-mq.h> 229 /* for blk-mq request-based DM support */ 1085 /* stacked on blk-mq queue(s) */ free_rq_clone() 1091 * NOTE: for the blk-mq queue stacked on request_fn queue(s) case: free_rq_clone() 1092 * no need to call free_clone_request() because we leverage blk-mq by free_rq_clone() 1093 * allocating the clone at the end of the blk-mq pdu (see: clone_rq) free_rq_clone() 1314 * Called with the clone's queue lock held (for non-blk-mq) 2245 * Initialize aspects of queue that aren't relevant for blk-mq dm_init_old_md_queue() 2748 /* Clone the request if underlying devices aren't blk-mq */ dm_mq_queue_rq() 2786 /* make the memory for non-blk-mq clone part of the pdu */ dm_init_request_based_blk_mq_queue() 2844 DMWARN("Cannot initialize queue for request-based blk-mq mapped device"); dm_setup_md_queue()
|
H A D | dm-table.c | 21 #include <linux/blk-mq.h> 902 /* verify _all_ devices in the table are blk-mq devices */ list_for_each_entry() 906 " are blk-mq request-stackable"); list_for_each_entry()
|
H A D | dm-io.c | 506 * q->unplug_delay, which defaults to 3ms in blk-settings.c.
|
H A D | dm-mpath.c | 429 /* blk-mq request-based interface */ __multipath_map()
|
/linux-4.1.27/drivers/net/wireless/ti/wl18xx/ |
H A D | acx.c | 36 wl1271_debug(DEBUG_ACX, "acx cfg bitmap %d blk %d spare %d field %d", wl18xx_acx_host_if_cfg_bitmap()
|
/linux-4.1.27/drivers/parisc/ |
H A D | iosapic_private.h | 182 char *si_pib_base; /* intr blk base address */
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | thread_info_64.h | 99 #define FAULT_CODE_BLKCOMMIT 0x10 /* Use blk-commit ASI in copy_page */
|
/linux-4.1.27/include/uapi/linux/ |
H A D | if_packet.h | 192 * you can see which blk[s] is[are] outstanding etc.
|
H A D | dqblk_xfs.h | 47 * It is 64-bit aligned, and all the blk units are in BBs (Basic Blocks) of
|
/linux-4.1.27/arch/s390/include/uapi/asm/ |
H A D | vtoc.h | 96 __u8 DS1NOBDB; /* no. of bytes used in last direction blk */
|
/linux-4.1.27/fs/freevxfs/ |
H A D | vxfs.h | 111 int32_t vs_inopb; /* number of inodes per blk */
|
/linux-4.1.27/scripts/ |
H A D | checkpatch.pl | 1011 my $blk = ''; 1029 #warn "CSB: blk<$blk> remain<$remain>\n"; 1038 $blk .= $lines[$line] . "\n"; 1039 $len = length($blk); 1044 #warn "CSB: blk<$blk> off<$off> len<$len>\n"; 1048 if ($level == 0 && substr($blk, $off) =~ /^.\s*#\s*define/) { 1054 $c = substr($blk, $off, 1); 1055 $remainder = substr($blk, $off); 1082 #warn "[" . substr($blk, $soff, $coff - $soff + 1) . "]\n"; 1108 if (substr($blk, $off + 1, 1) eq ';') { 1130 my $statement = substr($blk, $soff, $off - $soff + 1); 1131 my $condition = substr($blk, $soff, $coff - $soff + 1); 1218 my $blk = ''; 1229 $blk .= $rawlines[$line];
|
/linux-4.1.27/fs/reiserfs/ |
H A D | super.c | 2407 unsigned long blk = off >> sb->s_blocksize_bits; reiserfs_quota_read() local 2428 err = reiserfs_get_block(inode, blk, &tmp_bh, 0); reiserfs_quota_read() 2444 blk++; reiserfs_quota_read() 2457 unsigned long blk = off >> sb->s_blocksize_bits; reiserfs_quota_write() local 2473 err = reiserfs_get_block(inode, blk, &tmp_bh, GET_BLOCK_CREATE); reiserfs_quota_write() 2500 blk++; reiserfs_quota_write()
|
/linux-4.1.27/drivers/mtd/ubi/ |
H A D | block.c | 48 #include <linux/blk-mq.h> 462 /* Finally destroy the blk queue */ ubiblock_cleanup()
|
/linux-4.1.27/drivers/net/ethernet/chelsio/cxgb/ |
H A D | vsc7326_reg.h | 13 #define CRA(blk,sub,adr) ((((blk) & 0x7) << 13) | (((sub) & 0xf) << 9) | (((adr) & 0xff) << 1))
|
/linux-4.1.27/fs/ |
H A D | ioctl.c | 225 static inline loff_t blk_to_logical(struct inode *inode, sector_t blk) blk_to_logical() argument 227 return (blk << inode->i_blkbits); blk_to_logical()
|
/linux-4.1.27/include/target/ |
H A D | target_core_base.h | 94 /* Emulation for Thin Provisioning UNMAP using block/blk-lib.c:blkdev_issue_discard() */ 98 * block/blk-lib.c:blkdev_issue_discard()
|
/linux-4.1.27/drivers/gpu/drm/i2c/ |
H A D | tda998x_drv.c | 1039 static int read_edid_block(void *data, u8 *buf, unsigned int blk, size_t length) read_edid_block() argument 1045 offset = (blk & 1) ? 128 : 0; read_edid_block() 1046 segptr = blk / 2; read_edid_block() 1088 blk, ret); read_edid_block()
|
/linux-4.1.27/include/linux/usb/ |
H A D | wusb.h | 111 } blk[WUIE_ELT_MAX]; member in struct:wuie_connect_ack
|
/linux-4.1.27/arch/arm/mach-omap1/ |
H A D | clock.h | 56 * @find_idlest: function returning the IDLEST register for the clock's IP blk
|
/linux-4.1.27/include/linux/bcma/ |
H A D | bcma_driver_pci.h | 78 #define BCMA_CORE_PCI_MDIODATA_BLK_ADDR 0x1F /* blk address for serdes */
|
/linux-4.1.27/drivers/usb/wusbcore/ |
H A D | devconnect.c | 141 cack_ie->blk[cnt].CDID = dev_itr->cdid; wusbhc_fill_cack_ie() 142 cack_ie->blk[cnt].bDeviceAddress = dev_itr->addr; wusbhc_fill_cack_ie() 147 + cnt * sizeof(cack_ie->blk[0]); wusbhc_fill_cack_ie()
|
/linux-4.1.27/drivers/cdrom/ |
H A D | cdrom.c | 3062 struct cdrom_blk blk; mmc_ioctl_cdrom_play_blk() local 3064 if (copy_from_user(&blk, (struct cdrom_blk __user *)arg, sizeof(blk))) mmc_ioctl_cdrom_play_blk() 3067 cgc->cmd[2] = (blk.from >> 24) & 0xff; mmc_ioctl_cdrom_play_blk() 3068 cgc->cmd[3] = (blk.from >> 16) & 0xff; mmc_ioctl_cdrom_play_blk() 3069 cgc->cmd[4] = (blk.from >> 8) & 0xff; mmc_ioctl_cdrom_play_blk() 3070 cgc->cmd[5] = blk.from & 0xff; mmc_ioctl_cdrom_play_blk() 3071 cgc->cmd[7] = (blk.len >> 8) & 0xff; mmc_ioctl_cdrom_play_blk() 3072 cgc->cmd[8] = blk.len & 0xff; mmc_ioctl_cdrom_play_blk()
|
/linux-4.1.27/arch/ia64/sn/kernel/ |
H A D | io_common.c | 484 * This is needed to avoid bounce limit checks in the blk layer sn_io_early_init()
|
/linux-4.1.27/drivers/staging/lustre/lustre/include/ |
H A D | lu_object.h | 424 /** blk bits of the object*/ 426 /** blk size of the object*/
|
/linux-4.1.27/drivers/mtd/onenand/ |
H A D | onenand_base.c | 275 unsigned boundary, blk, die = 0; flexonenand_block() local 284 blk = addr >> (this->erase_shift - 1); flexonenand_block() 285 if (blk > boundary) flexonenand_block() 286 blk = (blk + boundary + 1) >> 1; flexonenand_block() 288 blk += die ? this->density_mask : 0; flexonenand_block() 289 return blk; flexonenand_block()
|
/linux-4.1.27/drivers/s390/block/ |
H A D | dasd_fba.c | 181 "and %d B/blk%s\n", dasd_fba_check_characteristics()
|
/linux-4.1.27/drivers/net/fddi/ |
H A D | defxx.h | 1728 char *rcv_block_virt; /* LLC host receive queue buf blk virt */ 1729 dma_addr_t rcv_block_phys; /* LLC host receive queue buf blk phys */
|
/linux-4.1.27/drivers/net/ethernet/broadcom/ |
H A D | cnic.c | 1649 int blk = (cid - ethdev->starting_cid) / cp->cids_per_blk; cnic_get_bnx2x_ctx() local 1658 if (cp->ctx_arr[blk].mapping & mask) cnic_get_bnx2x_ctx() 1660 (cp->ctx_arr[blk].mapping & mask); cnic_get_bnx2x_ctx() 1662 ctx_map = cp->ctx_arr[blk].mapping + align_off + cnic_get_bnx2x_ctx() 1664 ctx = cp->ctx_arr[blk].ctx + align_off + cnic_get_bnx2x_ctx()
|
/linux-4.1.27/drivers/scsi/lpfc/ |
H A D | lpfc_scsi.c | 3097 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err() 3109 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err() 3127 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err() 3144 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err() 3161 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err() 3205 " 0x%x lba 0x%llx blk cnt 0x%x " lpfc_parse_bg_err()
|
/linux-4.1.27/drivers/mmc/host/ |
H A D | sdhci-esdhc-imx.c | 90 * open ended multi-blk IO. Otherwise the TC INT wouldn't
|
/linux-4.1.27/drivers/net/ethernet/neterion/ |
H A D | s2io.h | 605 * Rxd in this blk */
|
H A D | s2io.c | 2665 static void free_rxd_blk(struct s2io_nic *sp, int ring_no, int blk) free_rxd_blk() argument 2679 rx_blocks[blk].rxds[j].virt_addr; free_rxd_blk() 2726 int i, blk = 0, buf_cnt = 0; free_rx_buffers() local 2733 for (blk = 0; blk < rx_ring_sz[i]; blk++) free_rx_buffers() 2734 free_rxd_blk(sp, i, blk); free_rx_buffers()
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
H A D | qla_target.h | 593 uint16_t runt_guard; /* reported runt blk guard */
|
/linux-4.1.27/drivers/target/ |
H A D | target_core_pscsi.c | 860 * in block/blk-core.c:blk_make_request() pscsi_get_bio()
|
/linux-4.1.27/drivers/scsi/fnic/ |
H A D | fnic_scsi.c | 2268 * this design issue, but that won't work for blk-mq. fnic_device_reset() 2272 * blk-mq case, or we'll have to bite the bullet and fnic_device_reset()
|
/linux-4.1.27/drivers/media/radio/wl128x/ |
H A D | fmdrv_common.c | 737 /* Transform the blk type into index sequence (0, 1, 2, 3, 4) */ fm_irq_handle_rdsdata_getcmd_resp()
|
/linux-4.1.27/drivers/char/pcmcia/ |
H A D | cm4000_cs.c | 117 wait_queue_head_t readq; /* used by write to wake blk.read */
|
/linux-4.1.27/drivers/usb/gadget/udc/ |
H A D | s3c2410_udc.c | 1098 dprintk(DEBUG_NORMAL, "enable %s(%d) ep%x%s-blk max %02x\n", s3c2410_udc_ep_enable()
|