object 443 arch/arm/include/asm/assembler.h .type \name , #object object 20 arch/parisc/math-emu/cnv_float.h #define Dintp1(object) (object) object 21 arch/parisc/math-emu/cnv_float.h #define Dintp2(object) (object) object 23 arch/parisc/math-emu/cnv_float.h #define Duintp1(object) (object) object 24 arch/parisc/math-emu/cnv_float.h #define Duintp2(object) (object) object 26 arch/parisc/math-emu/cnv_float.h #define Qintp0(object) (object) object 27 arch/parisc/math-emu/cnv_float.h #define Qintp1(object) (object) object 28 arch/parisc/math-emu/cnv_float.h #define Qintp2(object) (object) object 29 arch/parisc/math-emu/cnv_float.h #define Qintp3(object) (object) object 18 arch/parisc/math-emu/dbl_float.h #define Dbl_sign(object) Dsign(object) object 19 arch/parisc/math-emu/dbl_float.h #define Dbl_exponent(object) Dexponent(object) object 20 arch/parisc/math-emu/dbl_float.h #define Dbl_signexponent(object) Dsignexponent(object) object 21 arch/parisc/math-emu/dbl_float.h #define Dbl_mantissap1(object) Dmantissap1(object) object 22 arch/parisc/math-emu/dbl_float.h #define Dbl_mantissap2(object) Dmantissap2(object) object 23 arch/parisc/math-emu/dbl_float.h #define Dbl_exponentmantissap1(object) Dexponentmantissap1(object) object 24 arch/parisc/math-emu/dbl_float.h #define Dbl_allp1(object) Dallp1(object) object 25 arch/parisc/math-emu/dbl_float.h #define Dbl_allp2(object) Dallp2(object) object 48 arch/parisc/math-emu/float.h #define Sall(object) (object) object 49 arch/parisc/math-emu/float.h #define Ssign(object) Bitfield_extract( 0, 1,object) object 50 arch/parisc/math-emu/float.h #define Ssignedsign(object) Bitfield_signed_extract( 0, 1,object) object 51 arch/parisc/math-emu/float.h #define Sexponent(object) Bitfield_extract( 1, 8,object) object 52 arch/parisc/math-emu/float.h #define Smantissa(object) Bitfield_mask( 9, 23,object) object 53 arch/parisc/math-emu/float.h #define Ssignaling(object) Bitfield_extract( 9, 1,object) object 54 arch/parisc/math-emu/float.h #define Ssignalingnan(object) Bitfield_extract( 1, 9,object) object 55 arch/parisc/math-emu/float.h #define Shigh2mantissa(object) Bitfield_extract( 9, 2,object) object 56 arch/parisc/math-emu/float.h #define Sexponentmantissa(object) Bitfield_mask( 1, 31,object) object 57 arch/parisc/math-emu/float.h #define Ssignexponent(object) Bitfield_extract( 0, 9,object) object 58 arch/parisc/math-emu/float.h #define Shidden(object) Bitfield_extract( 8, 1,object) object 59 arch/parisc/math-emu/float.h #define Shiddenoverflow(object) Bitfield_extract( 7, 1,object) object 60 arch/parisc/math-emu/float.h #define Shiddenhigh7mantissa(object) Bitfield_extract( 8, 8,object) object 61 arch/parisc/math-emu/float.h #define Shiddenhigh3mantissa(object) Bitfield_extract( 8, 4,object) object 62 arch/parisc/math-emu/float.h #define Slow(object) Bitfield_mask( 31, 1,object) object 63 arch/parisc/math-emu/float.h #define Slow4(object) Bitfield_mask( 28, 4,object) object 64 arch/parisc/math-emu/float.h #define Slow31(object) Bitfield_mask( 1, 31,object) object 65 arch/parisc/math-emu/float.h #define Shigh31(object) Bitfield_extract( 0, 31,object) object 66 arch/parisc/math-emu/float.h #define Ssignedhigh31(object) Bitfield_signed_extract( 0, 31,object) object 67 arch/parisc/math-emu/float.h #define Shigh4(object) Bitfield_extract( 0, 4,object) object 68 arch/parisc/math-emu/float.h #define Sbit24(object) Bitfield_extract( 24, 1,object) object 69 arch/parisc/math-emu/float.h #define Sbit28(object) Bitfield_extract( 28, 1,object) object 70 arch/parisc/math-emu/float.h #define Sbit29(object) Bitfield_extract( 29, 1,object) object 71 arch/parisc/math-emu/float.h #define Sbit30(object) Bitfield_extract( 30, 1,object) object 72 arch/parisc/math-emu/float.h #define Sbit31(object) Bitfield_mask( 31, 1,object) object 74 arch/parisc/math-emu/float.h #define Deposit_ssign(object,value) Bitfield_deposit(value,0,1,object) object 75 arch/parisc/math-emu/float.h #define Deposit_sexponent(object,value) Bitfield_deposit(value,1,8,object) object 76 arch/parisc/math-emu/float.h #define Deposit_smantissa(object,value) Bitfield_deposit(value,9,23,object) object 77 arch/parisc/math-emu/float.h #define Deposit_shigh2mantissa(object,value) Bitfield_deposit(value,9,2,object) object 78 arch/parisc/math-emu/float.h #define Deposit_sexponentmantissa(object,value) \ object 79 arch/parisc/math-emu/float.h Bitfield_deposit(value,1,31,object) object 80 arch/parisc/math-emu/float.h #define Deposit_ssignexponent(object,value) Bitfield_deposit(value,0,9,object) object 81 arch/parisc/math-emu/float.h #define Deposit_slow(object,value) Bitfield_deposit(value,31,1,object) object 82 arch/parisc/math-emu/float.h #define Deposit_shigh4(object,value) Bitfield_deposit(value,0,4,object) object 84 arch/parisc/math-emu/float.h #define Is_ssign(object) Bitfield_mask( 0, 1,object) object 85 arch/parisc/math-emu/float.h #define Is_ssignaling(object) Bitfield_mask( 9, 1,object) object 86 arch/parisc/math-emu/float.h #define Is_shidden(object) Bitfield_mask( 8, 1,object) object 87 arch/parisc/math-emu/float.h #define Is_shiddenoverflow(object) Bitfield_mask( 7, 1,object) object 88 arch/parisc/math-emu/float.h #define Is_slow(object) Bitfield_mask( 31, 1,object) object 89 arch/parisc/math-emu/float.h #define Is_sbit24(object) Bitfield_mask( 24, 1,object) object 90 arch/parisc/math-emu/float.h #define Is_sbit28(object) Bitfield_mask( 28, 1,object) object 91 arch/parisc/math-emu/float.h #define Is_sbit29(object) Bitfield_mask( 29, 1,object) object 92 arch/parisc/math-emu/float.h #define Is_sbit30(object) Bitfield_mask( 30, 1,object) object 93 arch/parisc/math-emu/float.h #define Is_sbit31(object) Bitfield_mask( 31, 1,object) object 105 arch/parisc/math-emu/float.h #define Dallp1(object) (object) object 106 arch/parisc/math-emu/float.h #define Dsign(object) Bitfield_extract( 0, 1,object) object 107 arch/parisc/math-emu/float.h #define Dsignedsign(object) Bitfield_signed_extract( 0, 1,object) object 108 arch/parisc/math-emu/float.h #define Dexponent(object) Bitfield_extract( 1, 11,object) object 109 arch/parisc/math-emu/float.h #define Dmantissap1(object) Bitfield_mask( 12, 20,object) object 110 arch/parisc/math-emu/float.h #define Dsignaling(object) Bitfield_extract( 12, 1,object) object 111 arch/parisc/math-emu/float.h #define Dsignalingnan(object) Bitfield_extract( 1, 12,object) object 112 arch/parisc/math-emu/float.h #define Dhigh2mantissa(object) Bitfield_extract( 12, 2,object) object 113 arch/parisc/math-emu/float.h #define Dexponentmantissap1(object) Bitfield_mask( 1, 31,object) object 114 arch/parisc/math-emu/float.h #define Dsignexponent(object) Bitfield_extract( 0, 12,object) object 115 arch/parisc/math-emu/float.h #define Dhidden(object) Bitfield_extract( 11, 1,object) object 116 arch/parisc/math-emu/float.h #define Dhiddenoverflow(object) Bitfield_extract( 10, 1,object) object 117 arch/parisc/math-emu/float.h #define Dhiddenhigh7mantissa(object) Bitfield_extract( 11, 8,object) object 118 arch/parisc/math-emu/float.h #define Dhiddenhigh3mantissa(object) Bitfield_extract( 11, 4,object) object 119 arch/parisc/math-emu/float.h #define Dlowp1(object) Bitfield_mask( 31, 1,object) object 120 arch/parisc/math-emu/float.h #define Dlow31p1(object) Bitfield_mask( 1, 31,object) object 121 arch/parisc/math-emu/float.h #define Dhighp1(object) Bitfield_extract( 0, 1,object) object 122 arch/parisc/math-emu/float.h #define Dhigh4p1(object) Bitfield_extract( 0, 4,object) object 123 arch/parisc/math-emu/float.h #define Dhigh31p1(object) Bitfield_extract( 0, 31,object) object 124 arch/parisc/math-emu/float.h #define Dsignedhigh31p1(object) Bitfield_signed_extract( 0, 31,object) object 125 arch/parisc/math-emu/float.h #define Dbit3p1(object) Bitfield_extract( 3, 1,object) object 127 arch/parisc/math-emu/float.h #define Deposit_dsign(object,value) Bitfield_deposit(value,0,1,object) object 128 arch/parisc/math-emu/float.h #define Deposit_dexponent(object,value) Bitfield_deposit(value,1,11,object) object 129 arch/parisc/math-emu/float.h #define Deposit_dmantissap1(object,value) Bitfield_deposit(value,12,20,object) object 130 arch/parisc/math-emu/float.h #define Deposit_dhigh2mantissa(object,value) Bitfield_deposit(value,12,2,object) object 131 arch/parisc/math-emu/float.h #define Deposit_dexponentmantissap1(object,value) \ object 132 arch/parisc/math-emu/float.h Bitfield_deposit(value,1,31,object) object 133 arch/parisc/math-emu/float.h #define Deposit_dsignexponent(object,value) Bitfield_deposit(value,0,12,object) object 134 arch/parisc/math-emu/float.h #define Deposit_dlowp1(object,value) Bitfield_deposit(value,31,1,object) object 135 arch/parisc/math-emu/float.h #define Deposit_dhigh4p1(object,value) Bitfield_deposit(value,0,4,object) object 137 arch/parisc/math-emu/float.h #define Is_dsign(object) Bitfield_mask( 0, 1,object) object 138 arch/parisc/math-emu/float.h #define Is_dsignaling(object) Bitfield_mask( 12, 1,object) object 139 arch/parisc/math-emu/float.h #define Is_dhidden(object) Bitfield_mask( 11, 1,object) object 140 arch/parisc/math-emu/float.h #define Is_dhiddenoverflow(object) Bitfield_mask( 10, 1,object) object 141 arch/parisc/math-emu/float.h #define Is_dlowp1(object) Bitfield_mask( 31, 1,object) object 142 arch/parisc/math-emu/float.h #define Is_dhighp1(object) Bitfield_mask( 0, 1,object) object 143 arch/parisc/math-emu/float.h #define Is_dbit3p1(object) Bitfield_mask( 3, 1,object) object 145 arch/parisc/math-emu/float.h #define Dallp2(object) (object) object 146 arch/parisc/math-emu/float.h #define Dmantissap2(object) (object) object 147 arch/parisc/math-emu/float.h #define Dlowp2(object) Bitfield_mask( 31, 1,object) object 148 arch/parisc/math-emu/float.h #define Dlow4p2(object) Bitfield_mask( 28, 4,object) object 149 arch/parisc/math-emu/float.h #define Dlow31p2(object) Bitfield_mask( 1, 31,object) object 150 arch/parisc/math-emu/float.h #define Dhighp2(object) Bitfield_extract( 0, 1,object) object 151 arch/parisc/math-emu/float.h #define Dhigh31p2(object) Bitfield_extract( 0, 31,object) object 152 arch/parisc/math-emu/float.h #define Dbit2p2(object) Bitfield_extract( 2, 1,object) object 153 arch/parisc/math-emu/float.h #define Dbit3p2(object) Bitfield_extract( 3, 1,object) object 154 arch/parisc/math-emu/float.h #define Dbit21p2(object) Bitfield_extract( 21, 1,object) object 155 arch/parisc/math-emu/float.h #define Dbit28p2(object) Bitfield_extract( 28, 1,object) object 156 arch/parisc/math-emu/float.h #define Dbit29p2(object) Bitfield_extract( 29, 1,object) object 157 arch/parisc/math-emu/float.h #define Dbit30p2(object) Bitfield_extract( 30, 1,object) object 158 arch/parisc/math-emu/float.h #define Dbit31p2(object) Bitfield_mask( 31, 1,object) object 160 arch/parisc/math-emu/float.h #define Deposit_dlowp2(object,value) Bitfield_deposit(value,31,1,object) object 162 arch/parisc/math-emu/float.h #define Is_dlowp2(object) Bitfield_mask( 31, 1,object) object 163 arch/parisc/math-emu/float.h #define Is_dhighp2(object) Bitfield_mask( 0, 1,object) object 164 arch/parisc/math-emu/float.h #define Is_dbit2p2(object) Bitfield_mask( 2, 1,object) object 165 arch/parisc/math-emu/float.h #define Is_dbit3p2(object) Bitfield_mask( 3, 1,object) object 166 arch/parisc/math-emu/float.h #define Is_dbit21p2(object) Bitfield_mask( 21, 1,object) object 167 arch/parisc/math-emu/float.h #define Is_dbit28p2(object) Bitfield_mask( 28, 1,object) object 168 arch/parisc/math-emu/float.h #define Is_dbit29p2(object) Bitfield_mask( 29, 1,object) object 169 arch/parisc/math-emu/float.h #define Is_dbit30p2(object) Bitfield_mask( 30, 1,object) object 170 arch/parisc/math-emu/float.h #define Is_dbit31p2(object) Bitfield_mask( 31, 1,object) object 251 arch/parisc/math-emu/float.h #define Extall(object) (object) object 252 arch/parisc/math-emu/float.h #define Extsign(object) Bitfield_extract( 0, 1,object) object 253 arch/parisc/math-emu/float.h #define Exthigh31(object) Bitfield_extract( 0, 31,object) object 254 arch/parisc/math-emu/float.h #define Extlow31(object) Bitfield_extract( 1, 31,object) object 255 arch/parisc/math-emu/float.h #define Extlow(object) Bitfield_extract( 31, 1,object) object 261 arch/parisc/math-emu/float.h #define Sextallp1(object) (object) object 262 arch/parisc/math-emu/float.h #define Sextallp2(object) (object) object 263 arch/parisc/math-emu/float.h #define Sextlowp1(object) Bitfield_extract( 31, 1,object) object 264 arch/parisc/math-emu/float.h #define Sexthighp2(object) Bitfield_extract( 0, 1,object) object 265 arch/parisc/math-emu/float.h #define Sextlow31p2(object) Bitfield_extract( 1, 31,object) object 266 arch/parisc/math-emu/float.h #define Sexthiddenoverflow(object) Bitfield_extract( 4, 1,object) object 267 arch/parisc/math-emu/float.h #define Is_sexthiddenoverflow(object) Bitfield_mask( 4, 1,object) object 273 arch/parisc/math-emu/float.h #define Dextallp1(object) (object) object 274 arch/parisc/math-emu/float.h #define Dextallp2(object) (object) object 275 arch/parisc/math-emu/float.h #define Dextallp3(object) (object) object 276 arch/parisc/math-emu/float.h #define Dextallp4(object) (object) object 277 arch/parisc/math-emu/float.h #define Dextlowp2(object) Bitfield_extract( 31, 1,object) object 278 arch/parisc/math-emu/float.h #define Dexthighp3(object) Bitfield_extract( 0, 1,object) object 279 arch/parisc/math-emu/float.h #define Dextlow31p3(object) Bitfield_extract( 1, 31,object) object 280 arch/parisc/math-emu/float.h #define Dexthiddenoverflow(object) Bitfield_extract( 10, 1,object) object 281 arch/parisc/math-emu/float.h #define Is_dexthiddenoverflow(object) Bitfield_mask( 10, 1,object) object 282 arch/parisc/math-emu/float.h #define Deposit_dextlowp4(object,value) Bitfield_deposit(value,31,1,object) object 410 arch/parisc/math-emu/float.h #define Cbit(object) Bitfield_extract( 5, 1,object) object 411 arch/parisc/math-emu/float.h #define Tbit(object) Bitfield_extract( 25, 1,object) object 412 arch/parisc/math-emu/float.h #define Roundingmode(object) Bitfield_extract( 21, 2,object) object 413 arch/parisc/math-emu/float.h #define Invalidtrap(object) Bitfield_extract( 27, 1,object) object 414 arch/parisc/math-emu/float.h #define Divisionbyzerotrap(object) Bitfield_extract( 28, 1,object) object 415 arch/parisc/math-emu/float.h #define Overflowtrap(object) Bitfield_extract( 29, 1,object) object 416 arch/parisc/math-emu/float.h #define Underflowtrap(object) Bitfield_extract( 30, 1,object) object 417 arch/parisc/math-emu/float.h #define Inexacttrap(object) Bitfield_extract( 31, 1,object) object 418 arch/parisc/math-emu/float.h #define Invalidflag(object) Bitfield_extract( 0, 1,object) object 419 arch/parisc/math-emu/float.h #define Divisionbyzeroflag(object) Bitfield_extract( 1, 1,object) object 420 arch/parisc/math-emu/float.h #define Overflowflag(object) Bitfield_extract( 2, 1,object) object 421 arch/parisc/math-emu/float.h #define Underflowflag(object) Bitfield_extract( 3, 1,object) object 422 arch/parisc/math-emu/float.h #define Inexactflag(object) Bitfield_extract( 4, 1,object) object 423 arch/parisc/math-emu/float.h #define Allflags(object) Bitfield_extract( 0, 5,object) object 455 arch/parisc/math-emu/float.h #define Allexception(object) (object) object 456 arch/parisc/math-emu/float.h #define Exceptiontype(object) Bitfield_extract( 0, 6,object) object 457 arch/parisc/math-emu/float.h #define Instructionfield(object) Bitfield_mask( 6,26,object) object 458 arch/parisc/math-emu/float.h #define Parmfield(object) Bitfield_extract( 23, 3,object) object 459 arch/parisc/math-emu/float.h #define Rabit(object) Bitfield_extract( 24, 1,object) object 460 arch/parisc/math-emu/float.h #define Ibit(object) Bitfield_extract( 25, 1,object) object 462 arch/parisc/math-emu/float.h #define Set_exceptiontype(object,value) Bitfield_deposit(value, 0, 6,object) object 463 arch/parisc/math-emu/float.h #define Set_parmfield(object,value) Bitfield_deposit(value, 23, 3,object) object 464 arch/parisc/math-emu/float.h #define Set_exceptiontype_and_instr_field(exception,instruction,object) \ object 465 arch/parisc/math-emu/float.h object = exception << 26 | instruction object 474 arch/parisc/math-emu/float.h #define Greaterthanbit(object) Bitfield_extract( 27, 1,object) object 475 arch/parisc/math-emu/float.h #define Lessthanbit(object) Bitfield_extract( 28, 1,object) object 476 arch/parisc/math-emu/float.h #define Equalbit(object) Bitfield_extract( 29, 1,object) object 477 arch/parisc/math-emu/float.h #define Unorderedbit(object) Bitfield_extract( 30, 1,object) object 478 arch/parisc/math-emu/float.h #define Exceptionbit(object) Bitfield_extract( 31, 1,object) object 40 arch/parisc/math-emu/fpbits.h #define Bitfield_extract(start, length, object) \ object 41 arch/parisc/math-emu/fpbits.h ((object) >> (HOSTWDSZ - (start) - (length)) & \ object 44 arch/parisc/math-emu/fpbits.h #define Bitfield_signed_extract(start, length, object) \ object 45 arch/parisc/math-emu/fpbits.h ((int)((object) << start) >> (HOSTWDSZ - (length))) object 47 arch/parisc/math-emu/fpbits.h #define Bitfield_mask(start, len, object) \ object 48 arch/parisc/math-emu/fpbits.h ((object) & (((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) object 50 arch/parisc/math-emu/fpbits.h #define Bitfield_deposit(value,start,len,object) object = \ object 51 arch/parisc/math-emu/fpbits.h ((object) & ~(((unsigned)-1 >> (HOSTWDSZ-len)) << (HOSTWDSZ-start-len))) | \ object 19 arch/parisc/math-emu/sgl_float.h #define Sgl_sign(object) Ssign(object) object 20 arch/parisc/math-emu/sgl_float.h #define Sgl_exponent(object) Sexponent(object) object 21 arch/parisc/math-emu/sgl_float.h #define Sgl_signexponent(object) Ssignexponent(object) object 22 arch/parisc/math-emu/sgl_float.h #define Sgl_mantissa(object) Smantissa(object) object 23 arch/parisc/math-emu/sgl_float.h #define Sgl_exponentmantissa(object) Sexponentmantissa(object) object 24 arch/parisc/math-emu/sgl_float.h #define Sgl_all(object) Sall(object) object 231 drivers/acpi/acpi_processor.c union acpi_object object = { 0 }; object 232 drivers/acpi/acpi_processor.c struct acpi_buffer buffer = { sizeof(union acpi_object), &object }; object 263 drivers/acpi/acpi_processor.c pr->acpi_id = object.processor.proc_id; object 335 drivers/acpi/acpi_processor.c if (!object.processor.pblk_address) object 337 drivers/acpi/acpi_processor.c else if (object.processor.pblk_length != 6) object 339 drivers/acpi/acpi_processor.c object.processor.pblk_length); object 341 drivers/acpi/acpi_processor.c pr->throttling.address = object.processor.pblk_address; object 345 drivers/acpi/acpi_processor.c pr->pblk = object.processor.pblk_address; object 642 drivers/acpi/acpi_processor.c union acpi_object object = { 0 }; object 643 drivers/acpi/acpi_processor.c struct acpi_buffer buffer = { sizeof(union acpi_object), &object }; object 654 drivers/acpi/acpi_processor.c uid = object.processor.proc_id; object 99 drivers/acpi/acpica/acdebug.h acpi_status acpi_db_convert_to_package(char *string, union acpi_object *object); object 103 drivers/acpi/acpica/acdebug.h char *string, union acpi_object *object); object 289 drivers/acpi/acpica/acdispat.h acpi_ds_obj_stack_push(void *object, struct acpi_walk_state *walk_state); object 329 drivers/acpi/acpica/acdispat.h acpi_ds_result_pop(union acpi_operand_object **object, object 333 drivers/acpi/acpica/acdispat.h acpi_ds_result_push(union acpi_operand_object *object, object 391 drivers/acpi/acpica/acinterp.h acpi_ex_dump_object_descriptor(union acpi_operand_object *object, u32 flags); object 134 drivers/acpi/acpica/aclocal.h union acpi_operand_object *object; /* Interpreter object */ object 567 drivers/acpi/acpica/aclocal.h ACPI_STATE_COMMON union acpi_operand_object *object; object 296 drivers/acpi/acpica/acnamesp.h union acpi_operand_object *object, acpi_object_type type); object 310 drivers/acpi/acpica/acobject.h void *object; /* name_op=>HANDLE to obj, index_op=>union acpi_operand_object */ object 433 drivers/acpi/acpica/acobject.h union acpi_operand_object object; object 210 drivers/acpi/acpica/acutils.h const char *acpi_ut_get_node_name(void *object); object 212 drivers/acpi/acpica/acutils.h const char *acpi_ut_get_descriptor_name(void *object); object 214 drivers/acpi/acpica/acutils.h const char *acpi_ut_get_reference_name(union acpi_operand_object *object); object 272 drivers/acpi/acpica/acutils.h acpi_ut_update_object_reference(union acpi_operand_object *object, u16 action); object 350 drivers/acpi/acpica/acutils.h void acpi_ut_add_reference(union acpi_operand_object *object); object 352 drivers/acpi/acpica/acutils.h void acpi_ut_remove_reference(union acpi_operand_object *object); object 354 drivers/acpi/acpica/acutils.h void acpi_ut_delete_internal_package_object(union acpi_operand_object *object); object 356 drivers/acpi/acpica/acutils.h void acpi_ut_delete_internal_simple_object(union acpi_operand_object *object); object 432 drivers/acpi/acpica/acutils.h void acpi_ut_delete_object_desc(union acpi_operand_object *object); object 434 drivers/acpi/acpica/acutils.h u8 acpi_ut_valid_internal_object(void *object); object 501 drivers/acpi/acpica/acutils.h *object, u16 action); object 508 drivers/acpi/acpica/acutils.h acpi_ut_create_update_state_and_push(union acpi_operand_object *object, object 519 drivers/acpi/acpica/dbcmds.c status = acpi_rs_create_resource_list(node->object, &return_buffer); object 538 drivers/acpi/acpica/dbcmds.c acpi_ut_debug_dump_buffer((u8 *)node->object->buffer.pointer, object 539 drivers/acpi/acpica/dbcmds.c node->object->buffer.length, object 101 drivers/acpi/acpica/dbconvert.c acpi_db_convert_to_buffer(char *string, union acpi_object *object) object 141 drivers/acpi/acpica/dbconvert.c object->type = ACPI_TYPE_BUFFER; object 142 drivers/acpi/acpica/dbconvert.c object->buffer.pointer = buffer; object 143 drivers/acpi/acpica/dbconvert.c object->buffer.length = length; object 161 drivers/acpi/acpica/dbconvert.c acpi_status acpi_db_convert_to_package(char *string, union acpi_object *object) object 193 drivers/acpi/acpica/dbconvert.c object->type = ACPI_TYPE_PACKAGE; object 194 drivers/acpi/acpica/dbconvert.c object->package.count = i; object 195 drivers/acpi/acpica/dbconvert.c object->package.elements = elements; object 219 drivers/acpi/acpica/dbconvert.c char *string, union acpi_object *object) object 226 drivers/acpi/acpica/dbconvert.c object->type = ACPI_TYPE_STRING; object 227 drivers/acpi/acpica/dbconvert.c object->string.pointer = string; object 228 drivers/acpi/acpica/dbconvert.c object->string.length = (u32)strlen(string); object 233 drivers/acpi/acpica/dbconvert.c status = acpi_db_convert_to_buffer(string, object); object 238 drivers/acpi/acpica/dbconvert.c status = acpi_db_convert_to_package(string, object); object 243 drivers/acpi/acpica/dbconvert.c object->type = ACPI_TYPE_INTEGER; object 244 drivers/acpi/acpica/dbconvert.c status = acpi_ut_strtoul64(string, &object->integer.value); object 138 drivers/acpi/acpica/dbmethod.c obj_desc = node->object; object 181 drivers/acpi/acpica/dbmethod.c obj_desc = walk_state->arguments[index].object; object 204 drivers/acpi/acpica/dbmethod.c obj_desc = walk_state->local_variables[index].object; object 282 drivers/acpi/acpica/dbmethod.c obj_desc = method->object; object 651 drivers/acpi/acpica/dbnames.c union acpi_operand_object *object; object 671 drivers/acpi/acpica/dbnames.c node = (struct acpi_namespace_node *)node->object; object 688 drivers/acpi/acpica/dbnames.c object = acpi_ns_get_attached_object(node); object 689 drivers/acpi/acpica/dbnames.c if (object) { object 691 drivers/acpi/acpica/dbnames.c if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) { object 694 drivers/acpi/acpica/dbnames.c object, acpi_ut_get_descriptor_name(object)); object 255 drivers/acpi/acpica/dbobject.c [obj_desc->reference.value].object; object 268 drivers/acpi/acpica/dbobject.c [obj_desc->reference.value].object; object 282 drivers/acpi/acpica/dbobject.c object); object 284 drivers/acpi/acpica/dbobject.c (obj_desc->reference.object); object 313 drivers/acpi/acpica/dbobject.c if (!obj_desc->reference.object) { object 322 drivers/acpi/acpica/dbobject.c (obj_desc->reference.object)) { object 326 drivers/acpi/acpica/dbobject.c object); object 332 drivers/acpi/acpica/dbobject.c (obj_desc->reference.object); object 418 drivers/acpi/acpica/dbobject.c obj_desc = walk_state->local_variables[i].object; object 433 drivers/acpi/acpica/dbobject.c obj_desc = walk_state->local_variables[i].object; object 489 drivers/acpi/acpica/dbobject.c obj_desc = walk_state->arguments[i].object; object 502 drivers/acpi/acpica/dbobject.c node->object->method.param_count); object 505 drivers/acpi/acpica/dbobject.c obj_desc = walk_state->arguments[i].object; object 251 drivers/acpi/acpica/dbtest.c obj_desc = node->object; object 521 drivers/acpi/acpica/dsfield.c if (info.region_node->object->region.space_id == object 523 drivers/acpi/acpica/dsfield.c && !(region_node->object->field.internal_pcc_buffer = object 524 drivers/acpi/acpica/dsfield.c ACPI_ALLOCATE_ZEROED(info.region_node->object->region. object 25 drivers/acpi/acpica/dsmthdat.c union acpi_operand_object *object, object 111 drivers/acpi/acpica/dsmthdat.c if (walk_state->local_variables[index].object) { object 115 drivers/acpi/acpica/dsmthdat.c object)); object 127 drivers/acpi/acpica/dsmthdat.c if (walk_state->arguments[index].object) { object 130 drivers/acpi/acpica/dsmthdat.c walk_state->arguments[index].object)); object 281 drivers/acpi/acpica/dsmthdat.c union acpi_operand_object *object, object 290 drivers/acpi/acpica/dsmthdat.c "NewObj %p Type %2.2X, Refs=%u [%s]\n", object, object 291 drivers/acpi/acpica/dsmthdat.c type, object->common.reference_count, object 292 drivers/acpi/acpica/dsmthdat.c acpi_ut_get_type_name(object->common.type))); object 307 drivers/acpi/acpica/dsmthdat.c acpi_ut_add_reference(object); object 311 drivers/acpi/acpica/dsmthdat.c node->object = object; object 340 drivers/acpi/acpica/dsmthdat.c union acpi_operand_object *object; object 360 drivers/acpi/acpica/dsmthdat.c object = node->object; object 364 drivers/acpi/acpica/dsmthdat.c if (!object) { object 376 drivers/acpi/acpica/dsmthdat.c object = acpi_ut_create_integer_object((u64) 0); object 377 drivers/acpi/acpica/dsmthdat.c if (!object) { object 381 drivers/acpi/acpica/dsmthdat.c node->object = object; object 416 drivers/acpi/acpica/dsmthdat.c *dest_desc = object; object 417 drivers/acpi/acpica/dsmthdat.c acpi_ut_add_reference(object); object 444 drivers/acpi/acpica/dsmthdat.c union acpi_operand_object *object; object 457 drivers/acpi/acpica/dsmthdat.c object = acpi_ns_get_attached_object(node); object 464 drivers/acpi/acpica/dsmthdat.c node->object = NULL; object 466 drivers/acpi/acpica/dsmthdat.c if ((object) && object 467 drivers/acpi/acpica/dsmthdat.c (ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_OPERAND)) { object 473 drivers/acpi/acpica/dsmthdat.c acpi_ut_remove_reference(object); object 597 drivers/acpi/acpica/dsmthdat.c object, object 656 drivers/acpi/acpica/dsmthdat.c union acpi_operand_object *object; object 669 drivers/acpi/acpica/dsmthdat.c object = acpi_ns_get_attached_object(node); object 670 drivers/acpi/acpica/dsmthdat.c if (!object) { object 679 drivers/acpi/acpica/dsmthdat.c return_VALUE(object->type); object 481 drivers/acpi/acpica/dsobject.c object)); object 501 drivers/acpi/acpica/dsobject.c object)); object 514 drivers/acpi/acpica/dsobject.c obj_desc->reference.object = object 515 drivers/acpi/acpica/dsobject.c op->common.node->object; object 468 drivers/acpi/acpica/dspkginit.c resolved_node->object); object 516 drivers/acpi/acpica/dspkginit.c acpi_ut_remove_reference(original_node->object); object 481 drivers/acpi/acpica/dswexec.c node->object); object 562 drivers/acpi/acpica/dswexec.c node->object, object 633 drivers/acpi/acpica/dswload2.c if (node->object) { object 38 drivers/acpi/acpica/dswstate.c acpi_ds_result_pop(union acpi_operand_object **object, object 74 drivers/acpi/acpica/dswstate.c *object = state->results.obj_desc[index]; object 75 drivers/acpi/acpica/dswstate.c if (!*object) { object 91 drivers/acpi/acpica/dswstate.c "Obj=%p [%s] Index=%X State=%p Num=%X\n", *object, object 92 drivers/acpi/acpica/dswstate.c acpi_ut_get_object_type_name(*object), object 112 drivers/acpi/acpica/dswstate.c acpi_ds_result_push(union acpi_operand_object *object, object 147 drivers/acpi/acpica/dswstate.c if (!object) { object 150 drivers/acpi/acpica/dswstate.c object, walk_state, walk_state->result_count)); object 157 drivers/acpi/acpica/dswstate.c state->results.obj_desc[index] = object; object 161 drivers/acpi/acpica/dswstate.c object, object 164 drivers/acpi/acpica/dswstate.c object), walk_state, object 276 drivers/acpi/acpica/dswstate.c acpi_ds_obj_stack_push(void *object, struct acpi_walk_state *walk_state) object 285 drivers/acpi/acpica/dswstate.c object, walk_state, walk_state->num_operands)); object 291 drivers/acpi/acpica/dswstate.c walk_state->operands[walk_state->operand_index] = object; object 299 drivers/acpi/acpica/dswstate.c object, object 302 drivers/acpi/acpica/dswstate.c object), walk_state, object 822 drivers/acpi/acpica/evregion.c (next_node->object) && object 823 drivers/acpi/acpica/evregion.c (next_node->object->region.space_id == ACPI_ADR_SPACE_EC)) { object 886 drivers/acpi/acpica/evxfgpe.c if (node->object) { object 52 drivers/acpi/acpica/excreate.c target_node->object); object 80 drivers/acpi/acpica/excreate.c alias_node->object = object 87 drivers/acpi/acpica/excreate.c alias_node->object = object 234 drivers/acpi/acpica/exdebug.c node)->object, object 239 drivers/acpi/acpica/exdebug.c } else if (source_desc->reference.object) { object 241 drivers/acpi/acpica/exdebug.c (source_desc->reference.object) == object 250 drivers/acpi/acpica/exdebug.c object), level + 4, 0); object 252 drivers/acpi/acpica/exdebug.c object_desc = source_desc->reference.object; object 198 drivers/acpi/acpica/exdump.c {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(reference.object), "Object Desc"}, object 274 drivers/acpi/acpica/exdump.c {ACPI_EXD_LIST, ACPI_EXD_NSOFFSET(object), "Object List"}, object 645 drivers/acpi/acpica/exdump.c acpi_os_printf("%p\n", obj_desc->reference.object); object 656 drivers/acpi/acpica/exdump.c acpi_os_printf("%p [%s]\n", obj_desc->reference.object, object 662 drivers/acpi/acpica/exdump.c object)->common. object 991 drivers/acpi/acpica/exdump.c } else if (obj_desc->reference.object) { object 995 drivers/acpi/acpica/exdump.c obj_desc->reference.object); object 1006 drivers/acpi/acpica/exdump.c object)-> object 1012 drivers/acpi/acpica/exdump.c obj_desc->reference.object); object 1145 drivers/acpi/acpica/exdump.c obj_desc = ((struct acpi_namespace_node *)obj_desc)->object; object 61 drivers/acpi/acpica/exmisc.c referenced_obj = obj_desc->reference.object; object 95 drivers/acpi/acpica/exmisc.c reference_obj->reference.object = referenced_obj; object 824 drivers/acpi/acpica/exoparg1.c operand[0]->reference.object; object 925 drivers/acpi/acpica/exoparg1.c operand[0]->reference.object; object 982 drivers/acpi/acpica/exoparg1.c return_desc = operand[0]->reference.object; object 404 drivers/acpi/acpica/exoparg2.c return_desc->reference.object = operand[0]; object 439 drivers/acpi/acpica/exprep.c second_desc = info->connection_node->object; object 73 drivers/acpi/acpica/exresnte.c node = ACPI_CAST_PTR(struct acpi_namespace_node, node->object); object 231 drivers/acpi/acpica/exresolv.c (stack_desc->reference.node)->object; object 382 drivers/acpi/acpica/exresolv.c node = obj_desc->reference.object; object 23 drivers/acpi/acpica/exresop.c acpi_object_type this_type, void *object); object 41 drivers/acpi/acpica/exresop.c acpi_object_type this_type, void *object) object 59 drivers/acpi/acpica/exresop.c (((union acpi_operand_object *)object)->common.flags & object 69 drivers/acpi/acpica/exresop.c acpi_ut_get_type_name(this_type), object)); object 128 drivers/acpi/acpica/exstore.c object, walk_state, object 246 drivers/acpi/acpica/exstore.c index_desc->reference.object)->common. object 257 drivers/acpi/acpica/exstore.c index_desc->reference.object)->common. object 279 drivers/acpi/acpica/exstore.c obj_desc = index_desc->reference.object; object 658 drivers/acpi/acpica/nsaccess.c if (!this_node->object) { object 664 drivers/acpi/acpica/nsaccess.c this_node->object)->type)) { object 667 drivers/acpi/acpica/nsaccess.c this_node->object; object 91 drivers/acpi/acpica/nsalloc.c obj_desc = node->object; object 149 drivers/acpi/acpica/nsarguments.c aml_param_count = node->object->method.param_count; object 227 drivers/acpi/acpica/nsarguments.c aml_param_count = node->object->method.param_count; object 477 drivers/acpi/acpica/nsconvert.c new_object->reference.object = node->object; object 484 drivers/acpi/acpica/nsconvert.c acpi_ut_add_reference(node->object); object 76 drivers/acpi/acpica/nseval.c info->node->object); object 38 drivers/acpi/acpica/nsobject.c union acpi_operand_object *object, acpi_object_type type) object 57 drivers/acpi/acpica/nsobject.c if (!object && (ACPI_TYPE_ANY != type)) { object 77 drivers/acpi/acpica/nsobject.c if (node->object == object) { object 80 drivers/acpi/acpica/nsobject.c object, node)); object 87 drivers/acpi/acpica/nsobject.c if (!object) { object 96 drivers/acpi/acpica/nsobject.c else if ((ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED) && object 97 drivers/acpi/acpica/nsobject.c ((struct acpi_namespace_node *)object)->object) { object 102 drivers/acpi/acpica/nsobject.c obj_desc = ((struct acpi_namespace_node *)object)->object; object 103 drivers/acpi/acpica/nsobject.c object_type = ((struct acpi_namespace_node *)object)->type; object 111 drivers/acpi/acpica/nsobject.c obj_desc = (union acpi_operand_object *)object; object 123 drivers/acpi/acpica/nsobject.c if (node->object) { object 145 drivers/acpi/acpica/nsobject.c last_obj_desc->common.next_object = node->object; object 149 drivers/acpi/acpica/nsobject.c node->object = obj_desc; object 174 drivers/acpi/acpica/nsobject.c obj_desc = node->object; object 195 drivers/acpi/acpica/nsobject.c node->object = NULL; object 200 drivers/acpi/acpica/nsobject.c node->object = obj_desc->common.next_object; object 204 drivers/acpi/acpica/nsobject.c if (node->object && object 205 drivers/acpi/acpica/nsobject.c (node->object->common.type != ACPI_TYPE_LOCAL_DATA)) { object 206 drivers/acpi/acpica/nsobject.c node->object = node->object->common.next_object; object 257 drivers/acpi/acpica/nsobject.c if (!node->object || object 258 drivers/acpi/acpica/nsobject.c ((ACPI_GET_DESCRIPTOR_TYPE(node->object) != ACPI_DESC_TYPE_OPERAND) object 259 drivers/acpi/acpica/nsobject.c && (ACPI_GET_DESCRIPTOR_TYPE(node->object) != object 261 drivers/acpi/acpica/nsobject.c || ((node->object)->common.type == ACPI_TYPE_LOCAL_DATA)) { object 265 drivers/acpi/acpica/nsobject.c return_PTR(node->object); object 323 drivers/acpi/acpica/nsobject.c obj_desc = node->object; object 349 drivers/acpi/acpica/nsobject.c node->object = data_desc; object 377 drivers/acpi/acpica/nsobject.c obj_desc = node->object; object 385 drivers/acpi/acpica/nsobject.c node->object = obj_desc->common.next_object; object 420 drivers/acpi/acpica/nsobject.c obj_desc = node->object; object 102 drivers/acpi/acpica/nssearch.c node->object); object 312 drivers/acpi/acpica/nssearch.c acpi_ut_remove_reference((*return_node)->object); object 313 drivers/acpi/acpica/nssearch.c (*return_node)->object = NULL; object 498 drivers/acpi/acpica/nsxfeval.c node = info->return_object->reference.object; object 500 drivers/acpi/acpica/nsxfeval.c obj_desc = node->object; object 268 drivers/acpi/acpica/nsxfname.c param_count = node->object->method.param_count; object 327 drivers/acpi/acpica/tbxfload.c acpi_status acpi_unload_parent_table(acpi_handle object) object 330 drivers/acpi/acpica/tbxfload.c ACPI_CAST_PTR(struct acpi_namespace_node, object); object 339 drivers/acpi/acpica/tbxfload.c if (!object) { object 153 drivers/acpi/acpica/utcache.c acpi_status acpi_os_release_object(struct acpi_memory_list *cache, void *object) object 159 drivers/acpi/acpica/utcache.c if (!cache || !object) { object 166 drivers/acpi/acpica/utcache.c ACPI_FREE(object); object 180 drivers/acpi/acpica/utcache.c memset(object, 0xCA, cache->object_size); object 181 drivers/acpi/acpica/utcache.c ACPI_SET_DESCRIPTOR_TYPE(object, ACPI_DESC_TYPE_CACHED); object 185 drivers/acpi/acpica/utcache.c ACPI_SET_DESCRIPTOR_PTR(object, cache->list_head); object 186 drivers/acpi/acpica/utcache.c cache->list_head = object; object 211 drivers/acpi/acpica/utcache.c void *object; object 232 drivers/acpi/acpica/utcache.c object = cache->list_head; object 233 drivers/acpi/acpica/utcache.c cache->list_head = ACPI_GET_DESCRIPTOR_PTR(object); object 240 drivers/acpi/acpica/utcache.c ACPI_GET_FUNCTION_NAME, object, object 250 drivers/acpi/acpica/utcache.c memset(object, 0, cache->object_size); object 271 drivers/acpi/acpica/utcache.c object = ACPI_ALLOCATE_ZEROED(cache->object_size); object 272 drivers/acpi/acpica/utcache.c if (!object) { object 277 drivers/acpi/acpica/utcache.c return_PTR(object); object 507 drivers/acpi/acpica/utcopy.c internal_object->reference.object = object 734 drivers/acpi/acpica/utcopy.c acpi_ut_add_reference(source_desc->reference.object); object 238 drivers/acpi/acpica/utdecode.c const char *acpi_ut_get_node_name(void *object) object 240 drivers/acpi/acpica/utdecode.c struct acpi_namespace_node *node = (struct acpi_namespace_node *)object; object 244 drivers/acpi/acpica/utdecode.c if (!object) { object 250 drivers/acpi/acpica/utdecode.c if ((object == ACPI_ROOT_OBJECT) || (object == acpi_gbl_root_node)) { object 304 drivers/acpi/acpica/utdecode.c const char *acpi_ut_get_descriptor_name(void *object) object 307 drivers/acpi/acpica/utdecode.c if (!object) { object 311 drivers/acpi/acpica/utdecode.c if (ACPI_GET_DESCRIPTOR_TYPE(object) > ACPI_DESC_TYPE_MAX) { object 315 drivers/acpi/acpica/utdecode.c return (acpi_gbl_desc_type_names[ACPI_GET_DESCRIPTOR_TYPE(object)]); object 342 drivers/acpi/acpica/utdecode.c const char *acpi_ut_get_reference_name(union acpi_operand_object *object) object 345 drivers/acpi/acpica/utdecode.c if (!object) { object 349 drivers/acpi/acpica/utdecode.c if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) { object 353 drivers/acpi/acpica/utdecode.c if (object->common.type != ACPI_TYPE_LOCAL_REFERENCE) { object 357 drivers/acpi/acpica/utdecode.c if (object->reference.class > ACPI_REFCLASS_MAX) { object 361 drivers/acpi/acpica/utdecode.c return (acpi_gbl_ref_class_names[object->reference.class]); object 18 drivers/acpi/acpica/utdelete.c static void acpi_ut_delete_internal_obj(union acpi_operand_object *object); object 21 drivers/acpi/acpica/utdelete.c acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action); object 36 drivers/acpi/acpica/utdelete.c static void acpi_ut_delete_internal_obj(union acpi_operand_object *object) object 45 drivers/acpi/acpica/utdelete.c ACPI_FUNCTION_TRACE_PTR(ut_delete_internal_obj, object); object 47 drivers/acpi/acpica/utdelete.c if (!object) { object 55 drivers/acpi/acpica/utdelete.c switch (object->common.type) { object 59 drivers/acpi/acpica/utdelete.c "**** String %p, ptr %p\n", object, object 60 drivers/acpi/acpica/utdelete.c object->string.pointer)); object 64 drivers/acpi/acpica/utdelete.c if (!(object->common.flags & AOPOBJ_STATIC_POINTER)) { object 68 drivers/acpi/acpica/utdelete.c obj_pointer = object->string.pointer; object 75 drivers/acpi/acpica/utdelete.c "**** Buffer %p, ptr %p\n", object, object 76 drivers/acpi/acpica/utdelete.c object->buffer.pointer)); object 80 drivers/acpi/acpica/utdelete.c if (!(object->common.flags & AOPOBJ_STATIC_POINTER)) { object 84 drivers/acpi/acpica/utdelete.c obj_pointer = object->buffer.pointer; object 92 drivers/acpi/acpica/utdelete.c object->package.count)); object 101 drivers/acpi/acpica/utdelete.c obj_pointer = object->package.elements; object 110 drivers/acpi/acpica/utdelete.c if (object->device.gpe_block) { object 111 drivers/acpi/acpica/utdelete.c (void)acpi_ev_delete_gpe_block(object->device. object 122 drivers/acpi/acpica/utdelete.c handler_desc = object->common_notify.handler; object 134 drivers/acpi/acpica/utdelete.c object, object->mutex.os_mutex)); object 136 drivers/acpi/acpica/utdelete.c if (object == acpi_gbl_global_lock_mutex) { object 145 drivers/acpi/acpica/utdelete.c acpi_os_delete_mutex(object->mutex.os_mutex); object 148 drivers/acpi/acpica/utdelete.c acpi_ex_unlink_mutex(object); object 149 drivers/acpi/acpica/utdelete.c acpi_os_delete_mutex(object->mutex.os_mutex); object 157 drivers/acpi/acpica/utdelete.c object, object->event.os_semaphore)); object 159 drivers/acpi/acpica/utdelete.c (void)acpi_os_delete_semaphore(object->event.os_semaphore); object 160 drivers/acpi/acpica/utdelete.c object->event.os_semaphore = NULL; object 166 drivers/acpi/acpica/utdelete.c "***** Method %p\n", object)); object 170 drivers/acpi/acpica/utdelete.c if (object->method.mutex) { object 171 drivers/acpi/acpica/utdelete.c acpi_os_delete_mutex(object->method.mutex->mutex. object 173 drivers/acpi/acpica/utdelete.c acpi_ut_delete_object_desc(object->method.mutex); object 174 drivers/acpi/acpica/utdelete.c object->method.mutex = NULL; object 177 drivers/acpi/acpica/utdelete.c if (object->method.node) { object 178 drivers/acpi/acpica/utdelete.c object->method.node = NULL; object 185 drivers/acpi/acpica/utdelete.c "***** Region %p\n", object)); object 191 drivers/acpi/acpica/utdelete.c if (!(object->region.node->flags & ANOBJ_TEMPORARY)) { object 192 drivers/acpi/acpica/utdelete.c acpi_ut_remove_address_range(object->region.space_id, object 193 drivers/acpi/acpica/utdelete.c object->region.node); object 196 drivers/acpi/acpica/utdelete.c second_desc = acpi_ns_get_secondary_object(object); object 203 drivers/acpi/acpica/utdelete.c handler_desc = object->region.handler; object 214 drivers/acpi/acpica/utdelete.c if (next_desc == object) { object 242 drivers/acpi/acpica/utdelete.c address_space.setup(object, object 260 drivers/acpi/acpica/utdelete.c if (object->field.internal_pcc_buffer) { object 261 drivers/acpi/acpica/utdelete.c ACPI_FREE(object->field.internal_pcc_buffer); object 269 drivers/acpi/acpica/utdelete.c "***** Buffer Field %p\n", object)); object 271 drivers/acpi/acpica/utdelete.c second_desc = acpi_ns_get_secondary_object(object); object 280 drivers/acpi/acpica/utdelete.c "***** Bank Field %p\n", object)); object 282 drivers/acpi/acpica/utdelete.c second_desc = acpi_ns_get_secondary_object(object); object 305 drivers/acpi/acpica/utdelete.c ACPI_GET_FUNCTION_NAME, object, object 306 drivers/acpi/acpica/utdelete.c acpi_ut_get_object_type_name(object))); object 308 drivers/acpi/acpica/utdelete.c acpi_ut_delete_object_desc(object); object 357 drivers/acpi/acpica/utdelete.c acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action) object 366 drivers/acpi/acpica/utdelete.c if (!object) { object 375 drivers/acpi/acpica/utdelete.c original_count = object->common.reference_count; object 383 drivers/acpi/acpica/utdelete.c object->common.reference_count = new_count; object 391 drivers/acpi/acpica/utdelete.c object)); object 396 drivers/acpi/acpica/utdelete.c object, object->common.type, object 397 drivers/acpi/acpica/utdelete.c acpi_ut_get_object_type_name(object), object 408 drivers/acpi/acpica/utdelete.c object->common.reference_count = new_count; object 416 drivers/acpi/acpica/utdelete.c object)); object 421 drivers/acpi/acpica/utdelete.c ACPI_GET_FUNCTION_NAME, object, object 422 drivers/acpi/acpica/utdelete.c object->common.type, new_count)); object 427 drivers/acpi/acpica/utdelete.c acpi_ut_delete_internal_obj(object); object 447 drivers/acpi/acpica/utdelete.c new_count, object, object->common.type, message)); object 473 drivers/acpi/acpica/utdelete.c acpi_ut_update_object_reference(union acpi_operand_object *object, u16 action) object 484 drivers/acpi/acpica/utdelete.c while (object) { object 488 drivers/acpi/acpica/utdelete.c if (ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED) { object 490 drivers/acpi/acpica/utdelete.c "Object %p is NS handle\n", object)); object 498 drivers/acpi/acpica/utdelete.c switch (object->common.type) { object 509 drivers/acpi/acpica/utdelete.c object->common_notify.notify_list[i]; object 525 drivers/acpi/acpica/utdelete.c for (i = 0; i < object->package.count; i++) { object 530 drivers/acpi/acpica/utdelete.c next_object = object->package.elements[i]; object 568 drivers/acpi/acpica/utdelete.c next_object = object->buffer_field.buffer_obj; object 573 drivers/acpi/acpica/utdelete.c next_object = object->field.region_obj; object 578 drivers/acpi/acpica/utdelete.c next_object = object->bank_field.bank_obj; object 580 drivers/acpi/acpica/utdelete.c acpi_ut_create_update_state_and_push(object-> object 592 drivers/acpi/acpica/utdelete.c next_object = object->index_field.index_obj; object 594 drivers/acpi/acpica/utdelete.c acpi_ut_create_update_state_and_push(object-> object 610 drivers/acpi/acpica/utdelete.c if ((object->reference.class == ACPI_REFCLASS_INDEX) || object 611 drivers/acpi/acpica/utdelete.c (object->reference.class == ACPI_REFCLASS_NAME)) { object 612 drivers/acpi/acpica/utdelete.c next_object = object->reference.object; object 627 drivers/acpi/acpica/utdelete.c acpi_ut_update_ref_count(object, action); object 628 drivers/acpi/acpica/utdelete.c object = NULL; object 633 drivers/acpi/acpica/utdelete.c object = next_object; object 637 drivers/acpi/acpica/utdelete.c object = state->update.object; object 672 drivers/acpi/acpica/utdelete.c void acpi_ut_add_reference(union acpi_operand_object *object) object 679 drivers/acpi/acpica/utdelete.c if (!acpi_ut_valid_internal_object(object)) { object 685 drivers/acpi/acpica/utdelete.c object, object->common.reference_count)); object 689 drivers/acpi/acpica/utdelete.c (void)acpi_ut_update_object_reference(object, REF_INCREMENT); object 705 drivers/acpi/acpica/utdelete.c void acpi_ut_remove_reference(union acpi_operand_object *object) object 714 drivers/acpi/acpica/utdelete.c if (!object || object 715 drivers/acpi/acpica/utdelete.c (ACPI_GET_DESCRIPTOR_TYPE(object) == ACPI_DESC_TYPE_NAMED)) { object 721 drivers/acpi/acpica/utdelete.c if (!acpi_ut_valid_internal_object(object)) { object 727 drivers/acpi/acpica/utdelete.c ACPI_GET_FUNCTION_NAME, object, object 728 drivers/acpi/acpica/utdelete.c object->common.reference_count)); object 735 drivers/acpi/acpica/utdelete.c (void)acpi_ut_update_object_reference(object, REF_DECREMENT); object 190 drivers/acpi/acpica/utinit.c acpi_gbl_root_node_struct.object = NULL; object 158 drivers/acpi/acpica/utmisc.c acpi_ut_create_update_state_and_push(union acpi_operand_object *object, object 168 drivers/acpi/acpica/utmisc.c if (!object) { object 172 drivers/acpi/acpica/utmisc.c state = acpi_ut_create_update_state(object, action); object 60 drivers/acpi/acpica/utobject.c union acpi_operand_object *object; object 68 drivers/acpi/acpica/utobject.c object = object 71 drivers/acpi/acpica/utobject.c if (!object) { object 74 drivers/acpi/acpica/utobject.c kmemleak_not_leak(object); object 87 drivers/acpi/acpica/utobject.c acpi_ut_delete_object_desc(object); object 96 drivers/acpi/acpica/utobject.c object->common.next_object = second_object; object 107 drivers/acpi/acpica/utobject.c object->common.type = (u8) type; object 111 drivers/acpi/acpica/utobject.c object->common.reference_count = 1; object 115 drivers/acpi/acpica/utobject.c return_PTR(object); object 305 drivers/acpi/acpica/utobject.c u8 acpi_ut_valid_internal_object(void *object) object 312 drivers/acpi/acpica/utobject.c if (!object) { object 319 drivers/acpi/acpica/utobject.c switch (ACPI_GET_DESCRIPTOR_TYPE(object)) { object 330 drivers/acpi/acpica/utobject.c object, acpi_ut_get_descriptor_name(object))); object 355 drivers/acpi/acpica/utobject.c union acpi_operand_object *object; object 359 drivers/acpi/acpica/utobject.c object = acpi_os_acquire_object(acpi_gbl_operand_cache); object 360 drivers/acpi/acpica/utobject.c if (!object) { object 369 drivers/acpi/acpica/utobject.c ACPI_SET_DESCRIPTOR_TYPE(object, ACPI_DESC_TYPE_OPERAND); object 372 drivers/acpi/acpica/utobject.c object, (u32) sizeof(union acpi_operand_object))); object 374 drivers/acpi/acpica/utobject.c return_PTR(object); object 389 drivers/acpi/acpica/utobject.c void acpi_ut_delete_object_desc(union acpi_operand_object *object) object 391 drivers/acpi/acpica/utobject.c ACPI_FUNCTION_TRACE_PTR(ut_delete_object_desc, object); object 395 drivers/acpi/acpica/utobject.c if (ACPI_GET_DESCRIPTOR_TYPE(object) != ACPI_DESC_TYPE_OPERAND) { object 397 drivers/acpi/acpica/utobject.c "%p is not an ACPI Operand object [%s]", object, object 398 drivers/acpi/acpica/utobject.c acpi_ut_get_descriptor_name(object))); object 402 drivers/acpi/acpica/utobject.c (void)acpi_os_release_object(acpi_gbl_operand_cache, object); object 396 drivers/acpi/acpica/utosi.c string_desc = walk_state->arguments[0].object; object 157 drivers/acpi/acpica/utstate.c *object, u16 action) object 173 drivers/acpi/acpica/utstate.c state->update.object = object; object 654 drivers/acpi/acpica/uttrack.c (descriptor->object.common. object 656 drivers/acpi/acpica/uttrack.c descriptor->object.common. object 20 drivers/acpi/dptf/dptf_power.c #define DPTF_POWER_SHOW(name, object) \ object 29 drivers/acpi/dptf/dptf_power.c status = acpi_evaluate_integer(acpi_dev->handle, #object,\ object 1694 drivers/acpi/osl.c acpi_status acpi_os_release_object(acpi_cache_t * cache, void *object) object 1696 drivers/acpi/osl.c kmem_cache_free(cache, object); object 29 drivers/acpi/processor_pdc.c union acpi_object object = { 0 }; object 30 drivers/acpi/processor_pdc.c struct acpi_buffer buffer = { sizeof(union acpi_object), &object }; object 41 drivers/acpi/processor_pdc.c acpi_id = object.processor.proc_id; object 2032 drivers/android/binder.c struct binder_object *object) object 2038 drivers/android/binder.c read_size = min_t(size_t, sizeof(*object), buffer->data_size - offset); object 2040 drivers/android/binder.c binder_alloc_copy_from_buffer(&proc->alloc, object, buffer, object 2045 drivers/android/binder.c hdr = &object->hdr; object 2097 drivers/android/binder.c struct binder_object *object, object 2115 drivers/android/binder.c object_size = binder_get_object(proc, b, object_offset, object); object 2116 drivers/android/binder.c if (!object_size || object->hdr.type != BINDER_TYPE_PTR) object 2121 drivers/android/binder.c return &object->bbo; object 2286 drivers/android/binder.c struct binder_object object; object 2293 drivers/android/binder.c object_offset, &object); object 2299 drivers/android/binder.c hdr = &object.hdr; object 2692 drivers/android/binder.c struct binder_object object; object 2699 drivers/android/binder.c parent = binder_validate_ptr(target_proc, b, &object, bp->parent, object 3227 drivers/android/binder.c struct binder_object object; object 3241 drivers/android/binder.c object_offset, &object); object 3254 drivers/android/binder.c hdr = &object.hdr; object 360 drivers/base/cacheinfo.c #define show_one(file_name, object) \ object 365 drivers/base/cacheinfo.c return sprintf(buf, "%u\n", this_leaf->object); \ object 670 drivers/cpufreq/cpufreq.c #define show_one(file_name, object) \ object 674 drivers/cpufreq/cpufreq.c return sprintf(buf, "%u\n", policy->object); \ object 707 drivers/cpufreq/cpufreq.c #define store_one(file_name, object) \ object 718 drivers/cpufreq/cpufreq.c ret = freq_qos_update_request(policy->object##_freq_req, val);\ object 935 drivers/cpufreq/intel_pstate.c #define show_one(file_name, object) \ object 939 drivers/cpufreq/intel_pstate.c return sprintf(buf, "%u\n", global.object); \ object 62 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c static uint32_t get_src_obj_list(struct bios_parser *bp, ATOM_OBJECT *object, object 70 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object); object 76 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object); object 194 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 200 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = get_bios_object(bp, object_id); object 202 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) { object 207 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c number = get_src_obj_list(bp, object, &id); object 222 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 230 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = get_bios_object(bp, id); object 232 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) object 235 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c offset = le16_to_cpu(object->usRecordOffset) object 268 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 274 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = get_bios_object(bp, id); object 276 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) object 279 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c record = get_hpd_record(bp, object); object 292 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object, object 298 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c offset = le16_to_cpu(object->usRecordOffset) object 334 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 342 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = get_bios_object(bp, connector_object_id); object 344 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) { object 349 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (bios_parser_get_device_tag_record(bp, object, &record) object 866 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object) object 871 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) { object 876 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c offset = le16_to_cpu(object->usRecordOffset) object 1470 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 1476 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = get_bios_object(bp, object_id); object 1478 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) object 1481 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c record = get_encoder_cap_record(bp, object); object 1506 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object) object 1511 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) { object 1516 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c offset = le16_to_cpu(object->usRecordOffset) object 1993 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c static uint32_t get_src_obj_list(struct bios_parser *bp, ATOM_OBJECT *object, object 1999 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c if (!object) { object 2004 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c offset = le16_to_cpu(object->usSrcDstTableOffset) object 2703 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c ATOM_OBJECT *object; object 2708 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = NULL; object 2728 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c object = &object_table->asObjects[i]; object 2729 drivers/gpu/drm/amd/display/dc/bios/bios_parser.c record_offset = object->usRecordOffset + object 103 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object); object 107 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object); object 294 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object; object 302 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c object = get_bios_object(bp, id); object 304 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) object 307 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c offset = object->disp_recordoffset + bp->object_info_tbl_offset; object 420 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object; object 426 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c object = get_bios_object(bp, id); object 428 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) object 431 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c record = get_hpd_record(bp, object); object 444 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object) object 449 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) { object 454 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c offset = le16_to_cpu(object->disp_recordoffset) object 618 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object; object 624 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c object = get_bios_object(bp, connector_object_id); object 626 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) { object 632 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c info->dev_id = device_type_from_device_id(object->device_tag); object 1295 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object; object 1301 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c object = get_bios_object(bp, object_id); object 1303 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) object 1306 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c record = get_encoder_cap_record(bp, object); object 1327 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object) object 1332 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c if (!object) { object 1337 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c offset = object->encoder_recordoffset + bp->object_info_tbl_offset; object 1686 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c struct atom_display_object_path_v2 *object; object 1702 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c object = &v1_4->display_path[i]; object 1704 drivers/gpu/drm/amd/display/dc/bios/bios_parser2.c (object->disp_recordoffset) + object 150 drivers/gpu/drm/drm_crtc_internal.h struct drm_mode_object *object); object 201 drivers/gpu/drm/drm_lease.c int object; object 215 drivers/gpu/drm/drm_lease.c idr_for_each_entry(leases, entry, object) { object 217 drivers/gpu/drm/drm_lease.c if (!idr_find(&dev->mode_config.object_idr, object)) object 219 drivers/gpu/drm/drm_lease.c else if (_drm_has_leased(lessor, object)) object 223 drivers/gpu/drm/drm_lease.c DRM_DEBUG_LEASE("object %d failed %d\n", object, error); object 302 drivers/gpu/drm/drm_lease.c int object; object 316 drivers/gpu/drm/drm_lease.c idr_for_each_entry(&master->leases, entry, object) object 317 drivers/gpu/drm/drm_lease.c idr_remove(&master->leases, object); object 665 drivers/gpu/drm/drm_lease.c int object; object 687 drivers/gpu/drm/drm_lease.c idr_for_each_entry(object_idr, entry, object) { object 689 drivers/gpu/drm/drm_lease.c DRM_DEBUG_LEASE("adding object %d\n", object); object 690 drivers/gpu/drm/drm_lease.c ret = put_user(object, object_ids + count); object 105 drivers/gpu/drm/drm_mode_object.c struct drm_mode_object *object) object 107 drivers/gpu/drm/drm_mode_object.c WARN_ON(!dev->driver->load && dev->registered && !object->free_cb); object 110 drivers/gpu/drm/drm_mode_object.c if (object->id) { object 111 drivers/gpu/drm/drm_mode_object.c idr_remove(&dev->mode_config.object_idr, object->id); object 112 drivers/gpu/drm/drm_mode_object.c object->id = 0; object 240 drivers/gpu/drm/etnaviv/etnaviv_gem.c struct etnaviv_gem_object *etnaviv_obj = mapping->object; object 302 drivers/gpu/drm/etnaviv/etnaviv_gem.c mapping->object = etnaviv_obj; object 27 drivers/gpu/drm/etnaviv/etnaviv_gem.h struct etnaviv_gem_object *object; object 132 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct etnaviv_gem_object *etnaviv_obj = mapping->object; object 199 drivers/gpu/drm/nouveau/dispnv04/arb.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 1079 drivers/gpu/drm/nouveau/dispnv04/crtc.c struct nouveau_cli *cli = (void *)notify->object->client; object 68 drivers/gpu/drm/nouveau/dispnv04/dac.c struct nvif_object *device = &drm->client.device.object; object 135 drivers/gpu/drm/nouveau/dispnv04/dac.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 238 drivers/gpu/drm/nouveau/dispnv04/dac.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 284 drivers/gpu/drm/nouveau/dispnv04/dfp.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 489 drivers/gpu/drm/nouveau/dispnv04/dfp.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 184 drivers/gpu/drm/nouveau/dispnv04/disp.c nvif_object_unmap(&drm->client.device.object); object 204 drivers/gpu/drm/nouveau/dispnv04/disp.c nvif_object_map(&drm->client.device.object, NULL, 0); object 167 drivers/gpu/drm/nouveau/dispnv04/hw.c struct nvif_object *device = &drm->client.device.object; object 667 drivers/gpu/drm/nouveau/dispnv04/hw.c struct nvif_object *device = &drm->client.device.object; object 777 drivers/gpu/drm/nouveau/dispnv04/hw.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 796 drivers/gpu/drm/nouveau/dispnv04/hw.c struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 62 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 73 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 82 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 93 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 122 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 130 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 167 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 183 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 196 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 203 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 211 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 225 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 261 drivers/gpu/drm/nouveau/dispnv04/hw.h struct nvif_object *device = &nouveau_drm(dev)->client.device.object; object 120 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nvif_object *dev = &drm->client.device.object; object 193 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object; object 217 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nvif_object *dev = &nouveau_drm(plane->base.dev)->client.device.object; object 368 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object; object 442 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nvif_object *dev = &nouveau_drm(plane->dev)->client.device.object; object 134 drivers/gpu/drm/nouveau/dispnv04/tvnv17.h nvif_wr32(&device->object, reg, val); object 140 drivers/gpu/drm/nouveau/dispnv04/tvnv17.h return nvif_rd32(&device->object, reg); object 46 drivers/gpu/drm/nouveau/dispnv50/base.c cid = nvif_mclass(&disp->disp->object, bases); object 275 drivers/gpu/drm/nouveau/dispnv50/base507c.c ret = nv50_dmac_create(&drm->client.device, &disp->disp->object, object 64 drivers/gpu/drm/nouveau/dispnv50/core.c cid = nvif_mclass(&disp->disp->object, cores); object 100 drivers/gpu/drm/nouveau/dispnv50/core507d.c ret = nv50_dmac_create(&drm->client.device, &disp->disp->object, object 46 drivers/gpu/drm/nouveau/dispnv50/curs.c cid = nvif_mclass(&disp->disp->object, curses); object 126 drivers/gpu/drm/nouveau/dispnv50/curs507a.c ret = nvif_object_init(&disp->disp->object, 0, oclass, &args, object 138 drivers/gpu/drm/nouveau/dispnv50/disp.c struct nouveau_cli *cli = (void *)device->object.client; object 161 drivers/gpu/drm/nouveau/dispnv50/disp.c dmac->ptr = dmac->push.object.map.ptr; object 163 drivers/gpu/drm/nouveau/dispnv50/disp.c args->pushbuf = nvif_handle(&dmac->push.object); object 209 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_wr32(&device->object, 0x070000, 0x00000001); object 211 drivers/gpu/drm/nouveau/dispnv50/disp.c if (!(nvif_rd32(&device->object, 0x070000) & 0x00000002)) object 272 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); object 293 drivers/gpu/drm/nouveau/dispnv50/disp.c ret = nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); object 422 drivers/gpu/drm/nouveau/dispnv50/disp.c ret = nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); object 498 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); object 528 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &args, object 551 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &args, sizeof(args)); object 628 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &args, size); object 768 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&drm->display->disp.object, 0, &args, sizeof(args)); object 1249 drivers/gpu/drm/nouveau/dispnv50/disp.c struct nvif_object *disp = &drm->display->disp.object; object 1531 drivers/gpu/drm/nouveau/dispnv50/disp.c nvif_mthd(&disp->disp->object, 0, &lvds, sizeof(lvds)); object 1614 drivers/gpu/drm/nouveau/dispnv50/disp.c if (disp->disp->object.oclass < GF110_DISP) { object 2365 drivers/gpu/drm/nouveau/dispnv50/disp.c if (disp->disp->object.oclass >= GV100_DISP) object 2366 drivers/gpu/drm/nouveau/dispnv50/disp.c crtcs = nvif_rd32(&device->object, 0x610060) & 0xff; object 2368 drivers/gpu/drm/nouveau/dispnv50/disp.c if (disp->disp->object.oclass >= GF110_DISP) object 2369 drivers/gpu/drm/nouveau/dispnv50/disp.c crtcs = nvif_rd32(&device->object, 0x612004) & 0xf; object 493 drivers/gpu/drm/nouveau/dispnv50/head.c if (disp->disp->object.oclass < GV100_DISP) { object 514 drivers/gpu/drm/nouveau/dispnv50/head.c if (disp->disp->object.oclass >= GF110_DISP) object 36 drivers/gpu/drm/nouveau/dispnv50/lut.c void __iomem *mem = lut->mem[buffer].object.map.ptr; object 70 drivers/gpu/drm/nouveau/dispnv50/lut.c const u32 size = disp->disp->object.oclass < GF110_DISP ? 257 : 1025; object 44 drivers/gpu/drm/nouveau/dispnv50/oimm.c cid = nvif_mclass(&disp->disp->object, oimms); object 36 drivers/gpu/drm/nouveau/dispnv50/oimm507b.c ret = nvif_object_init(&disp->disp->object, 0, oclass, &args, object 46 drivers/gpu/drm/nouveau/dispnv50/ovly.c cid = nvif_mclass(&disp->disp->object, ovlys); object 187 drivers/gpu/drm/nouveau/dispnv50/ovly507e.c ret = nv50_dmac_create(&drm->client.device, &disp->disp->object, object 41 drivers/gpu/drm/nouveau/dispnv50/wimm.c cid = nvif_mclass(&disp->disp->object, wimms); object 70 drivers/gpu/drm/nouveau/dispnv50/wimmc37b.c ret = nv50_dmac_create(&drm->client.device, &disp->disp->object, object 36 drivers/gpu/drm/nouveau/dispnv50/wndw.c nvif_object_fini(&ctxdma->object); object 60 drivers/gpu/drm/nouveau/dispnv50/wndw.c if (ctxdma->object.handle == handle) object 92 drivers/gpu/drm/nouveau/dispnv50/wndw.c &args, argc, &ctxdma->object); object 503 drivers/gpu/drm/nouveau/dispnv50/wndw.c asyw->image.handle[0] = ctxdma->object.handle; object 715 drivers/gpu/drm/nouveau/dispnv50/wndw.c cid = nvif_mclass(&disp->disp->object, wndws); object 12 drivers/gpu/drm/nouveau/dispnv50/wndw.h struct nvif_object object; object 294 drivers/gpu/drm/nouveau/dispnv50/wndwc37e.c ret = nv50_dmac_create(&drm->client.device, &disp->disp->object, object 8 drivers/gpu/drm/nouveau/include/nvif/client.h struct nvif_object object; object 26 drivers/gpu/drm/nouveau/include/nvif/client.h (struct nvkm_client *)_client->object.priv; \ object 10 drivers/gpu/drm/nouveau/include/nvif/device.h struct nvif_object object; object 61 drivers/gpu/drm/nouveau/include/nvif/device.h struct nvkm_object object; \ object 63 drivers/gpu/drm/nouveau/include/nvif/device.h } *_udevice = _device->object.priv; \ object 7 drivers/gpu/drm/nouveau/include/nvif/disp.h struct nvif_object object; object 31 drivers/gpu/drm/nouveau/include/nvif/ioctl.h __u64 object; object 57 drivers/gpu/drm/nouveau/include/nvif/ioctl.h __u64 object; object 6 drivers/gpu/drm/nouveau/include/nvif/mem.h struct nvif_object object; object 6 drivers/gpu/drm/nouveau/include/nvif/mmu.h struct nvif_object object; object 6 drivers/gpu/drm/nouveau/include/nvif/notify.h struct nvif_object *object; object 40 drivers/gpu/drm/nouveau/include/nvif/object.h #define nvif_object(a) (a)->object object 79 drivers/gpu/drm/nouveau/include/nvif/object.h struct nvif_object *object = (o); \ object 85 drivers/gpu/drm/nouveau/include/nvif/object.h cnt = nvif_object_sclass_get(object, &sclass); \ object 8 drivers/gpu/drm/nouveau/include/nvif/user.h struct nvif_object object; object 19 drivers/gpu/drm/nouveau/include/nvif/vmm.h struct nvif_object object; object 4 drivers/gpu/drm/nouveau/include/nvkm/core/client.h #define nvkm_client(p) container_of((p), struct nvkm_client, object) object 8 drivers/gpu/drm/nouveau/include/nvkm/core/client.h struct nvkm_object object; object 19 drivers/gpu/drm/nouveau/include/nvkm/core/object.h u64 object; object 75 drivers/gpu/drm/nouveau/include/nvkm/core/object.h struct nvkm_object *nvkm_object_search(struct nvkm_client *, u64 object, object 26 drivers/gpu/drm/nouveau/include/nvkm/core/oclass.h u64 object; object 10 drivers/gpu/drm/nouveau/include/nvkm/core/oproxy.h struct nvkm_object *object; object 12 drivers/gpu/drm/nouveau/include/nvkm/engine/dma.h struct nvkm_object object; object 24 drivers/gpu/drm/nouveau/include/nvkm/engine/dma.h struct nvkm_dmaobj *nvkm_dmaobj_search(struct nvkm_client *, u64 object); object 12 drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h struct nvkm_object *object; object 21 drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h struct nvkm_object object; object 58 drivers/gpu/drm/nouveau/nouveau_abi16.c if (nvif_device_init(&cli->base.object, 0, NV_DEVICE, object 84 drivers/gpu/drm/nouveau/nouveau_abi16.c struct nouveau_cli *cli = (void *)abi16->device.object.client; object 117 drivers/gpu/drm/nouveau/nouveau_abi16.c nvif_object_fini(&ntfy->object); object 161 drivers/gpu/drm/nouveau/nouveau_abi16.c struct nouveau_cli *cli = (void *)abi16->device.object.client; object 395 drivers/gpu/drm/nouveau/nouveau_abi16.c args->v0.object = nvif_handle(&chan->chan->user); object 400 drivers/gpu/drm/nouveau/nouveau_abi16.c args->v0.object = nvif_handle(&abi16->device.object); object 439 drivers/gpu/drm/nouveau/nouveau_abi16.c client = abi16->device.object.client; object 506 drivers/gpu/drm/nouveau/nouveau_abi16.c NULL, 0, &ntfy->object); object 533 drivers/gpu/drm/nouveau/nouveau_abi16.c client = abi16->device.object.client; object 574 drivers/gpu/drm/nouveau/nouveau_abi16.c &ntfy->object); object 607 drivers/gpu/drm/nouveau/nouveau_abi16.c if (ntfy->object.handle == fini->handle) { object 16 drivers/gpu/drm/nouveau/nouveau_abi16.h struct nvif_object object; object 70 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 82 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 104 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 120 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 135 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 156 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 174 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 203 drivers/gpu/drm/nouveau/nouveau_backlight.c struct nvif_object *device = &drm->client.device.object; object 216 drivers/gpu/drm/nouveau/nouveau_bios.c struct nvif_object *device = &drm->client.device.object; object 320 drivers/gpu/drm/nouveau/nouveau_bios.c struct nvif_object *device = &drm->client.device.object; object 636 drivers/gpu/drm/nouveau/nouveau_bios.c struct nvif_object *device = &drm->client.device.object; object 1921 drivers/gpu/drm/nouveau/nouveau_bios.c struct nvif_object *device = &drm->client.device.object; object 284 drivers/gpu/drm/nouveau/nouveau_bo.c if (mmu->object.oclass >= NVIF_CLASS_MMU_GF100) object 1481 drivers/gpu/drm/nouveau/nouveau_bo.c switch (mem->mem.object.oclass) { object 1500 drivers/gpu/drm/nouveau/nouveau_bo.c ret = nvif_object_map_handle(&mem->mem.object, object 1526 drivers/gpu/drm/nouveau/nouveau_bo.c nvif_object_unmap_handle(&mem->mem.object); object 1529 drivers/gpu/drm/nouveau/nouveau_bo.c nvif_object_unmap_handle(&mem->mem.object); object 126 drivers/gpu/drm/nouveau/nouveau_chan.c struct nouveau_cli *cli = (void *)device->object.client; object 215 drivers/gpu/drm/nouveau/nouveau_chan.c ret = nvif_object_init(&device->object, 0, NV_DMA_FROM_MEMORY, object 263 drivers/gpu/drm/nouveau/nouveau_chan.c args.volta.vmm = nvif_handle(&chan->vmm->vmm.object); object 272 drivers/gpu/drm/nouveau/nouveau_chan.c args.kepler.vmm = nvif_handle(&chan->vmm->vmm.object); object 280 drivers/gpu/drm/nouveau/nouveau_chan.c args.fermi.vmm = nvif_handle(&chan->vmm->vmm.object); object 287 drivers/gpu/drm/nouveau/nouveau_chan.c args.nv50.vmm = nvif_handle(&chan->vmm->vmm.object); object 291 drivers/gpu/drm/nouveau/nouveau_chan.c ret = nvif_object_init(&device->object, 0, *oclass++, object 342 drivers/gpu/drm/nouveau/nouveau_chan.c ret = nvif_object_init(&device->object, 0, *oclass++, object 478 drivers/gpu/drm/nouveau/nouveau_chan.c struct nouveau_cli *cli = (void *)device->object.client; object 524 drivers/gpu/drm/nouveau/nouveau_chan.c struct nvif_object *device = &drm->client.device.object; object 154 drivers/gpu/drm/nouveau/nouveau_connector.c if (disp->disp.object.oclass < NV50_DISP) object 271 drivers/gpu/drm/nouveau/nouveau_connector.c if (nouveau_display(connector->dev)->disp.object.oclass < NV50_DISP) { object 331 drivers/gpu/drm/nouveau/nouveau_connector.c if (disp->disp.object.oclass < NV50_DISP) object 1385 drivers/gpu/drm/nouveau/nouveau_connector.c if ((disp->disp.object.oclass >= G82_DISP) object 1407 drivers/gpu/drm/nouveau/nouveau_connector.c if (disp->disp.object.oclass < NV50_DISP) { object 1438 drivers/gpu/drm/nouveau/nouveau_connector.c ret = nvif_notify_init(&disp->disp.object, nouveau_connector_hotplug, object 61 drivers/gpu/drm/nouveau/nouveau_debugfs.c nvif_rd32(&drm->client.device.object, 0x101000)); object 264 drivers/gpu/drm/nouveau/nouveau_debugfs.c ret = nvif_object_init(&drm->client.device.object, 0, object 117 drivers/gpu/drm/nouveau/nouveau_display.c ret = nvif_mthd(&disp->disp.object, 0, &args, sizeof(args)); object 176 drivers/gpu/drm/nouveau/nouveau_display.c ret = nvif_notify_init(&disp->disp.object, object 466 drivers/gpu/drm/nouveau/nouveau_display.c if (disp->disp.object.oclass < NV50_DISP) object 469 drivers/gpu/drm/nouveau/nouveau_display.c if (disp->disp.object.oclass < GF110_DISP) object 548 drivers/gpu/drm/nouveau/nouveau_display.c if (disp->disp.object.oclass < NV50_DISP) object 241 drivers/gpu/drm/nouveau/nouveau_drm.c ret = nvif_device_init(&cli->base.object, 0, NV_DEVICE, object 251 drivers/gpu/drm/nouveau/nouveau_drm.c ret = nvif_mclass(&cli->device.object, mmus); object 257 drivers/gpu/drm/nouveau/nouveau_drm.c ret = nvif_mmu_init(&cli->device.object, mmus[ret].oclass, &cli->mmu); object 263 drivers/gpu/drm/nouveau/nouveau_drm.c ret = nvif_mclass(&cli->mmu.object, vmms); object 275 drivers/gpu/drm/nouveau/nouveau_drm.c ret = nvif_mclass(&cli->mmu.object, mems); object 440 drivers/gpu/drm/nouveau/nouveau_drm.c ret = n = nvif_object_sclass_get(&device->object, &sclass); object 530 drivers/gpu/drm/nouveau/nouveau_drm.c nvif_mask(&drm->client.device.object, 0x00088080, 0x00000800, 0x00000000); object 1022 drivers/gpu/drm/nouveau/nouveau_drm.c nvif_mask(&device->object, 0x088488, (1 << 25), (1 << 25)); object 71 drivers/gpu/drm/nouveau/nouveau_gem.c if (vmm->vmm.object.oclass < NVIF_CLASS_VMM_NV50) object 146 drivers/gpu/drm/nouveau/nouveau_gem.c if (vmm->vmm.object.oclass < NVIF_CLASS_VMM_NV50) object 236 drivers/gpu/drm/nouveau/nouveau_gem.c if (vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50) { object 344 drivers/gpu/drm/nouveau/nouveau_gem.c if (chan->vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50) { object 439 drivers/gpu/drm/nouveau/nouveau_gem.c if (chan->vmm->vmm.object.oclass >= NVIF_CLASS_VMM_NV50) { object 41 drivers/gpu/drm/nouveau/nouveau_led.c struct nvif_object *device = &drm->client.device.object; object 58 drivers/gpu/drm/nouveau/nouveau_led.c struct nvif_object *device = &drm->client.device.object; object 47 drivers/gpu/drm/nouveau/nouveau_mem.c switch (vmm->object.oclass) { object 76 drivers/gpu/drm/nouveau/nouveau_mem.c super = vmm->object.client->super; object 77 drivers/gpu/drm/nouveau/nouveau_mem.c vmm->object.client->super = true; object 80 drivers/gpu/drm/nouveau/nouveau_mem.c vmm->object.client->super = super; object 114 drivers/gpu/drm/nouveau/nouveau_mem.c if (mmu->object.oclass >= NVIF_CLASS_MMU_GF100) object 64 drivers/gpu/drm/nouveau/nouveau_nvif.c return nvkm_object_init(&client->object); object 71 drivers/gpu/drm/nouveau/nouveau_nvif.c return nvkm_object_fini(&client->object, true); object 46 drivers/gpu/drm/nouveau/nouveau_svm.c struct nvif_object object; object 248 drivers/gpu/drm/nouveau/nouveau_svm.c bool super = svmm->vmm->vmm.object.client->super; object 249 drivers/gpu/drm/nouveau/nouveau_svm.c svmm->vmm->vmm.object.client->super = true; object 250 drivers/gpu/drm/nouveau/nouveau_svm.c nvif_object_mthd(&svmm->vmm->vmm.object, NVIF_VMM_V0_PFNCLR, object 255 drivers/gpu/drm/nouveau/nouveau_svm.c svmm->vmm->vmm.object.client->super = super; object 338 drivers/gpu/drm/nouveau/nouveau_svm.c ret = nvif_vmm_init(&cli->mmu, cli->vmm.vmm.object.oclass, true, object 384 drivers/gpu/drm/nouveau/nouveau_svm.c WARN_ON(nvif_object_mthd(&svm->drm->client.vmm.vmm.object, object 400 drivers/gpu/drm/nouveau/nouveau_svm.c WARN_ON(nvif_object_mthd(&svm->drm->client.vmm.vmm.object, object 439 drivers/gpu/drm/nouveau/nouveau_svm.c struct nvif_object *memory = &buffer->object; object 530 drivers/gpu/drm/nouveau/nouveau_svm.c struct nvif_object *device = &svm->drm->client.device.object; object 706 drivers/gpu/drm/nouveau/nouveau_svm.c svmm->vmm->vmm.object.client->super = true; object 707 drivers/gpu/drm/nouveau/nouveau_svm.c ret = nvif_object_ioctl(&svmm->vmm->vmm.object, object 711 drivers/gpu/drm/nouveau/nouveau_svm.c svmm->vmm->vmm.object.client->super = false; object 752 drivers/gpu/drm/nouveau/nouveau_svm.c struct nvif_object *device = &svm->drm->client.device.object; object 774 drivers/gpu/drm/nouveau/nouveau_svm.c nvif_object_fini(&buffer->object); object 782 drivers/gpu/drm/nouveau/nouveau_svm.c struct nvif_object *device = &drm->client.device.object; object 789 drivers/gpu/drm/nouveau/nouveau_svm.c &buffer->object); object 795 drivers/gpu/drm/nouveau/nouveau_svm.c nvif_object_map(&buffer->object, NULL, 0); object 800 drivers/gpu/drm/nouveau/nouveau_svm.c ret = nvif_notify_init(&buffer->object, nouveau_svm_fault, true, object 865 drivers/gpu/drm/nouveau/nouveau_svm.c ret = nvif_mclass(&drm->client.device.object, buffers); object 257 drivers/gpu/drm/nouveau/nouveau_usif.c usif_object_dtor(struct usif_object *object) object 259 drivers/gpu/drm/nouveau/nouveau_usif.c list_del(&object->head); object 260 drivers/gpu/drm/nouveau/nouveau_usif.c kfree(object); object 271 drivers/gpu/drm/nouveau/nouveau_usif.c struct usif_object *object; object 274 drivers/gpu/drm/nouveau/nouveau_usif.c if (!(object = kmalloc(sizeof(*object), GFP_KERNEL))) object 276 drivers/gpu/drm/nouveau/nouveau_usif.c list_add(&object->head, &cli->objects); object 279 drivers/gpu/drm/nouveau/nouveau_usif.c object->route = args->v0.route; object 280 drivers/gpu/drm/nouveau/nouveau_usif.c object->token = args->v0.token; object 282 drivers/gpu/drm/nouveau/nouveau_usif.c args->v0.token = (unsigned long)(void *)object; object 284 drivers/gpu/drm/nouveau/nouveau_usif.c args->v0.token = object->token; object 285 drivers/gpu/drm/nouveau/nouveau_usif.c args->v0.route = object->route; object 289 drivers/gpu/drm/nouveau/nouveau_usif.c usif_object_dtor(object); object 303 drivers/gpu/drm/nouveau/nouveau_usif.c struct usif_object *object; object 315 drivers/gpu/drm/nouveau/nouveau_usif.c if (argv->v0.object == 0ULL && object 357 drivers/gpu/drm/nouveau/nouveau_usif.c object = (void *)(unsigned long)argv->v0.token; object 358 drivers/gpu/drm/nouveau/nouveau_usif.c argv->v0.route = object->route; object 359 drivers/gpu/drm/nouveau/nouveau_usif.c argv->v0.token = object->token; object 361 drivers/gpu/drm/nouveau/nouveau_usif.c list_del(&object->head); object 362 drivers/gpu/drm/nouveau/nouveau_usif.c kfree(object); object 381 drivers/gpu/drm/nouveau/nouveau_usif.c struct usif_object *object, *otemp; object 388 drivers/gpu/drm/nouveau/nouveau_usif.c list_for_each_entry_safe(object, otemp, &cli->objects, head) { object 389 drivers/gpu/drm/nouveau/nouveau_usif.c usif_object_dtor(object); object 17 drivers/gpu/drm/nouveau/nouveau_vga.c struct nvif_object *device = &drm->client.device.object; object 35 drivers/gpu/drm/nouveau/nvif/client.c return client->driver->ioctl(client->object.priv, client->super, data, size, NULL); object 41 drivers/gpu/drm/nouveau/nvif/client.c return client->driver->suspend(client->object.priv); object 47 drivers/gpu/drm/nouveau/nvif/client.c return client->driver->resume(client->object.priv); object 53 drivers/gpu/drm/nouveau/nvif/client.c nvif_object_fini(&client->object); object 56 drivers/gpu/drm/nouveau/nvif/client.c client->driver->fini(client->object.priv); object 73 drivers/gpu/drm/nouveau/nvif/client.c ret = nvif_object_init(parent != client ? &parent->object : NULL, object 75 drivers/gpu/drm/nouveau/nvif/client.c &client->object); object 79 drivers/gpu/drm/nouveau/nvif/client.c client->object.client = client; object 80 drivers/gpu/drm/nouveau/nvif/client.c client->object.handle = ~0; object 31 drivers/gpu/drm/nouveau/nvif/device.c int ret = nvif_object_mthd(&device->object, NV_DEVICE_V0_TIME, object 43 drivers/gpu/drm/nouveau/nvif/device.c nvif_object_fini(&device->object); object 51 drivers/gpu/drm/nouveau/nvif/device.c &device->object); object 56 drivers/gpu/drm/nouveau/nvif/device.c ret = nvif_object_mthd(&device->object, NV_DEVICE_V0_INFO, object 30 drivers/gpu/drm/nouveau/nvif/disp.c nvif_object_fini(&disp->object); object 54 drivers/gpu/drm/nouveau/nvif/disp.c int cid = nvif_sclass(&device->object, disps, oclass); object 55 drivers/gpu/drm/nouveau/nvif/disp.c disp->object.client = NULL; object 59 drivers/gpu/drm/nouveau/nvif/disp.c return nvif_object_init(&device->object, 0, disps[cid].oclass, object 60 drivers/gpu/drm/nouveau/nvif/disp.c NULL, 0, &disp->object); object 48 drivers/gpu/drm/nouveau/nvif/driver.c &client->object.priv); object 51 drivers/gpu/drm/nouveau/nvif/driver.c client->driver->fini(client->object.priv); object 27 drivers/gpu/drm/nouveau/nvif/fifo.c struct nvif_object *object = &device->object; object 48 drivers/gpu/drm/nouveau/nvif/fifo.c ret = nvif_object_mthd(object, NV_DEVICE_V0_INFO, a, sizeof(*a)); object 73 drivers/gpu/drm/nouveau/nvif/fifo.c struct nvif_object *object = &device->object; object 90 drivers/gpu/drm/nouveau/nvif/fifo.c ret = nvif_object_mthd(object, NV_DEVICE_V0_INFO, &a, sizeof(a)); object 33 drivers/gpu/drm/nouveau/nvif/mem.c ret = nvif_object_map(&mem->object, NULL, 0); object 43 drivers/gpu/drm/nouveau/nvif/mem.c nvif_object_fini(&mem->object); object 54 drivers/gpu/drm/nouveau/nvif/mem.c mem->object.client = NULL; object 70 drivers/gpu/drm/nouveau/nvif/mem.c ret = nvif_object_init(&mmu->object, 0, oclass, args, object 71 drivers/gpu/drm/nouveau/nvif/mem.c sizeof(*args) + argc, &mem->object); object 91 drivers/gpu/drm/nouveau/nvif/mem.c mem->object.client = NULL; object 33 drivers/gpu/drm/nouveau/nvif/mmu.c nvif_object_fini(&mmu->object); object 54 drivers/gpu/drm/nouveau/nvif/mmu.c &mmu->object); object 63 drivers/gpu/drm/nouveau/nvif/mmu.c ret = nvif_mclass(&mmu->object, mems); object 83 drivers/gpu/drm/nouveau/nvif/mmu.c ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_HEAP, object 94 drivers/gpu/drm/nouveau/nvif/mmu.c ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_TYPE, object 120 drivers/gpu/drm/nouveau/nvif/mmu.c ret = nvif_object_mthd(&mmu->object, NVIF_MMU_V0_KIND, object 35 drivers/gpu/drm/nouveau/nvif/notify.c struct nvif_object *object = notify->object; object 47 drivers/gpu/drm/nouveau/nvif/notify.c return nvif_object_ioctl(object, &args, sizeof(args), NULL); object 53 drivers/gpu/drm/nouveau/nvif/notify.c if (likely(notify->object) && object 66 drivers/gpu/drm/nouveau/nvif/notify.c struct nvif_object *object = notify->object; object 78 drivers/gpu/drm/nouveau/nvif/notify.c return nvif_object_ioctl(object, &args, sizeof(args), NULL); object 84 drivers/gpu/drm/nouveau/nvif/notify.c if (likely(notify->object) && object 127 drivers/gpu/drm/nouveau/nvif/notify.c struct nvif_client *client = notify->object->client; object 147 drivers/gpu/drm/nouveau/nvif/notify.c struct nvif_object *object = notify->object; object 156 drivers/gpu/drm/nouveau/nvif/notify.c if (ret >= 0 && object) { object 157 drivers/gpu/drm/nouveau/nvif/notify.c ret = nvif_object_ioctl(object, &args, sizeof(args), NULL); object 158 drivers/gpu/drm/nouveau/nvif/notify.c notify->object = NULL; object 165 drivers/gpu/drm/nouveau/nvif/notify.c nvif_notify_init(struct nvif_object *object, int (*func)(struct nvif_notify *), object 176 drivers/gpu/drm/nouveau/nvif/notify.c notify->object = object; object 202 drivers/gpu/drm/nouveau/nvif/notify.c ret = nvif_object_ioctl(object, args, sizeof(*args) + size, NULL); object 31 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_ioctl(struct nvif_object *object, void *data, u32 size, void **hack) object 33 drivers/gpu/drm/nouveau/nvif/object.c struct nvif_client *client = object->client; object 39 drivers/gpu/drm/nouveau/nvif/object.c if (object != &client->object) object 40 drivers/gpu/drm/nouveau/nvif/object.c args->v0.object = nvif_handle(object); object 42 drivers/gpu/drm/nouveau/nvif/object.c args->v0.object = 0; object 47 drivers/gpu/drm/nouveau/nvif/object.c return client->driver->ioctl(client->object.priv, client->super, object 59 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_sclass_get(struct nvif_object *object, struct nvif_sclass **psclass) object 77 drivers/gpu/drm/nouveau/nvif/object.c ret = nvif_object_ioctl(object, args, size, NULL); object 103 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_rd(struct nvif_object *object, int size, u64 addr) object 113 drivers/gpu/drm/nouveau/nvif/object.c int ret = nvif_object_ioctl(object, &args, sizeof(args), NULL); object 122 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_wr(struct nvif_object *object, int size, u64 addr, u32 data) object 133 drivers/gpu/drm/nouveau/nvif/object.c int ret = nvif_object_ioctl(object, &args, sizeof(args), NULL); object 140 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_mthd(struct nvif_object *object, u32 mthd, void *data, u32 size) object 161 drivers/gpu/drm/nouveau/nvif/object.c ret = nvif_object_ioctl(object, args, sizeof(*args) + size, NULL); object 169 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_unmap_handle(struct nvif_object *object) object 178 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_ioctl(object, &args, sizeof(args), NULL); object 182 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_map_handle(struct nvif_object *object, void *argv, u32 argc, object 197 drivers/gpu/drm/nouveau/nvif/object.c ret = nvif_object_ioctl(object, args, argn, NULL); object 206 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_unmap(struct nvif_object *object) object 208 drivers/gpu/drm/nouveau/nvif/object.c struct nvif_client *client = object->client; object 209 drivers/gpu/drm/nouveau/nvif/object.c if (object->map.ptr) { object 210 drivers/gpu/drm/nouveau/nvif/object.c if (object->map.size) { object 211 drivers/gpu/drm/nouveau/nvif/object.c client->driver->unmap(client, object->map.ptr, object 212 drivers/gpu/drm/nouveau/nvif/object.c object->map.size); object 213 drivers/gpu/drm/nouveau/nvif/object.c object->map.size = 0; object 215 drivers/gpu/drm/nouveau/nvif/object.c object->map.ptr = NULL; object 216 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_unmap_handle(object); object 221 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_map(struct nvif_object *object, void *argv, u32 argc) object 223 drivers/gpu/drm/nouveau/nvif/object.c struct nvif_client *client = object->client; object 225 drivers/gpu/drm/nouveau/nvif/object.c int ret = nvif_object_map_handle(object, argv, argc, &handle, &length); object 228 drivers/gpu/drm/nouveau/nvif/object.c object->map.ptr = client->driver->map(client, object 231 drivers/gpu/drm/nouveau/nvif/object.c if (ret = -ENOMEM, object->map.ptr) { object 232 drivers/gpu/drm/nouveau/nvif/object.c object->map.size = length; object 236 drivers/gpu/drm/nouveau/nvif/object.c object->map.ptr = (void *)(unsigned long)handle; object 239 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_unmap_handle(object); object 245 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_fini(struct nvif_object *object) object 254 drivers/gpu/drm/nouveau/nvif/object.c if (!object->client) object 257 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_unmap(object); object 258 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_ioctl(object, &args, sizeof(args), NULL); object 259 drivers/gpu/drm/nouveau/nvif/object.c object->client = NULL; object 264 drivers/gpu/drm/nouveau/nvif/object.c void *data, u32 size, struct nvif_object *object) object 272 drivers/gpu/drm/nouveau/nvif/object.c object->client = NULL; object 273 drivers/gpu/drm/nouveau/nvif/object.c object->handle = handle; object 274 drivers/gpu/drm/nouveau/nvif/object.c object->oclass = oclass; object 275 drivers/gpu/drm/nouveau/nvif/object.c object->map.ptr = NULL; object 276 drivers/gpu/drm/nouveau/nvif/object.c object->map.size = 0; object 280 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_fini(object); object 288 drivers/gpu/drm/nouveau/nvif/object.c args->new.token = nvif_handle(object); object 289 drivers/gpu/drm/nouveau/nvif/object.c args->new.object = nvif_handle(object); object 295 drivers/gpu/drm/nouveau/nvif/object.c &object->priv); object 299 drivers/gpu/drm/nouveau/nvif/object.c object->client = parent->client; object 303 drivers/gpu/drm/nouveau/nvif/object.c nvif_object_fini(object); object 31 drivers/gpu/drm/nouveau/nvif/user.c nvif_object_fini(&device->user.object); object 52 drivers/gpu/drm/nouveau/nvif/user.c cid = nvif_mclass(&device->object, users); object 56 drivers/gpu/drm/nouveau/nvif/user.c ret = nvif_object_init(&device->object, 0, users[cid].oclass, NULL, 0, object 57 drivers/gpu/drm/nouveau/nvif/user.c &device->user.object); object 61 drivers/gpu/drm/nouveau/nvif/user.c nvif_object_map(&device->user.object, NULL, 0); object 27 drivers/gpu/drm/nouveau/nvif/userc361.c nvif_wr32(&user->object, 0x90, token); object 30 drivers/gpu/drm/nouveau/nvif/vmm.c return nvif_object_mthd(&vmm->object, NVIF_VMM_V0_UNMAP, object 53 drivers/gpu/drm/nouveau/nvif/vmm.c args->memory = nvif_handle(&mem->object); object 57 drivers/gpu/drm/nouveau/nvif/vmm.c ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_MAP, object 68 drivers/gpu/drm/nouveau/nvif/vmm.c WARN_ON(nvif_object_mthd(&vmm->object, NVIF_VMM_V0_PUT, object 98 drivers/gpu/drm/nouveau/nvif/vmm.c ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_GET, object 111 drivers/gpu/drm/nouveau/nvif/vmm.c nvif_object_fini(&vmm->object); object 122 drivers/gpu/drm/nouveau/nvif/vmm.c vmm->object.client = NULL; object 133 drivers/gpu/drm/nouveau/nvif/vmm.c ret = nvif_object_init(&mmu->object, 0, oclass, args, argn, object 134 drivers/gpu/drm/nouveau/nvif/vmm.c &vmm->object); object 152 drivers/gpu/drm/nouveau/nvif/vmm.c ret = nvif_object_mthd(&vmm->object, NVIF_VMM_V0_PAGE, object 53 drivers/gpu/drm/nouveau/nvkm/core/client.c client->object.client = oclass->client; object 54 drivers/gpu/drm/nouveau/nvkm/core/client.c client->object.handle = oclass->handle; object 55 drivers/gpu/drm/nouveau/nvkm/core/client.c client->object.route = oclass->route; object 56 drivers/gpu/drm/nouveau/nvkm/core/client.c client->object.token = oclass->token; object 57 drivers/gpu/drm/nouveau/nvkm/core/client.c client->object.object = oclass->object; object 59 drivers/gpu/drm/nouveau/nvkm/core/client.c *pobject = &client->object; object 128 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_client_notify_new(struct nvkm_object *object, object 131 drivers/gpu/drm/nouveau/nvkm/core/client.c struct nvkm_client *client = object->client; object 151 drivers/gpu/drm/nouveau/nvkm/core/client.c nvif_ioctl(object, "notify new size %d\n", size); object 153 drivers/gpu/drm/nouveau/nvkm/core/client.c nvif_ioctl(object, "notify new vers %d reply %d route %02x " object 165 drivers/gpu/drm/nouveau/nvkm/core/client.c ret = nvkm_notify_init(object, event, nvkm_client_notify, object 182 drivers/gpu/drm/nouveau/nvkm/core/client.c struct nvkm_object *object; object 184 drivers/gpu/drm/nouveau/nvkm/core/client.c object = nvkm_object_search(client, handle, &nvkm_client); object 185 drivers/gpu/drm/nouveau/nvkm/core/client.c if (IS_ERR(object)) object 186 drivers/gpu/drm/nouveau/nvkm/core/client.c return (void *)object; object 188 drivers/gpu/drm/nouveau/nvkm/core/client.c return nvkm_client(object); object 199 drivers/gpu/drm/nouveau/nvkm/core/client.c nvif_ioctl(&client->object, "client devlist size %d\n", size); object 201 drivers/gpu/drm/nouveau/nvkm/core/client.c nvif_ioctl(&client->object, "client devlist vers %d count %d\n", object 218 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_client_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 220 drivers/gpu/drm/nouveau/nvkm/core/client.c struct nvkm_client *client = nvkm_client(object); object 238 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_client_child_get(struct nvkm_object *object, int index, object 256 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_client_fini(struct nvkm_object *object, bool suspend) object 258 drivers/gpu/drm/nouveau/nvkm/core/client.c struct nvkm_client *client = nvkm_client(object); object 261 drivers/gpu/drm/nouveau/nvkm/core/client.c nvif_debug(object, "%s notify\n", name[suspend]); object 268 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_client_dtor(struct nvkm_object *object) object 270 drivers/gpu/drm/nouveau/nvkm/core/client.c struct nvkm_client *client = nvkm_client(object); object 298 drivers/gpu/drm/nouveau/nvkm/core/client.c nvkm_object_ctor(&nvkm_client, &oclass, &client->object); object 33 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 40 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "nop size %d\n", size); object 42 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "nop vers %lld\n", args->v0.version); object 51 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 59 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "sclass size %d\n", size); object 61 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "sclass vers %d count %d\n", object 66 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c while (object->func->sclass && object 67 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c object->func->sclass(object, i, &oclass) >= 0) { object 89 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object = NULL; object 98 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c args->v0.route, args->v0.token, args->v0.object); object 112 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c oclass.object = args->v0.object; object 126 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = oclass.ctor(&oclass, data, size, &object); object 129 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_init(object); object 131 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c list_add(&object->head, &parent->tree); object 132 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c if (nvkm_object_insert(object)) { object 133 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c client->data = object; object 138 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvkm_object_fini(object, false); object 141 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvkm_object_del(&object); object 147 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 154 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "delete size %d\n", size); object 156 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "delete\n"); object 157 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvkm_object_fini(object, false); object 158 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvkm_object_del(&object); object 166 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 173 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "mthd size %d\n", size); object 175 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "mthd vers %d mthd %02x\n", object 177 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_mthd(object, args->v0.method, data, size); object 186 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 198 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "rd size %d\n", size); object 200 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "rd vers %d size %d addr %016llx\n", object 204 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_rd08(object, args->v0.addr, &v.b08); object 208 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_rd16(object, args->v0.addr, &v.b16); object 212 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_rd32(object, args->v0.addr, &v.b32); object 226 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 233 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "wr size %d\n", size); object 235 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, object 243 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c case 1: return nvkm_object_wr08(object, args->v0.addr, args->v0.data); object 244 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c case 2: return nvkm_object_wr16(object, args->v0.addr, args->v0.data); object 245 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c case 4: return nvkm_object_wr32(object, args->v0.addr, args->v0.data); object 255 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 263 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "map size %d\n", size); object 265 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "map vers %d\n", args->v0.version); object 266 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_map(object, data, size, &type, object 280 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 287 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "unmap size %d\n", size); object 289 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "unmap\n"); object 290 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_unmap(object); object 298 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 306 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy new size %d\n", size); object 308 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy new vers %d event %02x\n", object 310 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_object_ntfy(object, args->v0.event, &event); object 312 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_client_notify_new(object, event, data, size); object 325 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 332 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy del size %d\n", size); object 334 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy del vers %d index %d\n", object 344 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 351 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy get size %d\n", size); object 353 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy get vers %d index %d\n", object 363 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object, void *data, u32 size) object 370 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy put size %d\n", size); object 372 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "ntfy put vers %d index %d\n", object 404 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object; object 407 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c object = nvkm_object_search(client, handle, NULL); object 408 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c if (IS_ERR(object)) { object 409 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(&client->object, "object not found\n"); object 410 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c return PTR_ERR(object); object 413 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c if (owner != NVIF_IOCTL_V0_OWNER_ANY && owner != object->route) { object 414 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(&client->object, "route != owner\n"); object 417 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c *route = object->route; object 418 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c *token = object->token; object 422 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_ioctl_v0[type].func(client, object, data, size); object 432 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c struct nvkm_object *object = &client->object; object 439 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "size %d\n", size); object 442 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, object 444 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c args->v0.version, args->v0.type, args->v0.object, object 446 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = nvkm_ioctl_path(client, args->v0.object, args->v0.type, object 452 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c nvif_ioctl(object, "return %d\n", ret); object 132 drivers/gpu/drm/nouveau/nvkm/core/notify.c nvkm_notify_init(struct nvkm_object *object, struct nvkm_event *event, object 140 drivers/gpu/drm/nouveau/nvkm/core/notify.c ret = event->func->ctor(object, data, size, notify); object 32 drivers/gpu/drm/nouveau/nvkm/core/object.c struct nvkm_object *object; object 37 drivers/gpu/drm/nouveau/nvkm/core/object.c object = rb_entry(node, typeof(*object), node); object 38 drivers/gpu/drm/nouveau/nvkm/core/object.c if (handle < object->object) object 41 drivers/gpu/drm/nouveau/nvkm/core/object.c if (handle > object->object) object 48 drivers/gpu/drm/nouveau/nvkm/core/object.c object = &client->object; object 52 drivers/gpu/drm/nouveau/nvkm/core/object.c if (unlikely(func && object->func != func)) object 54 drivers/gpu/drm/nouveau/nvkm/core/object.c return object; object 58 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_remove(struct nvkm_object *object) object 60 drivers/gpu/drm/nouveau/nvkm/core/object.c if (!RB_EMPTY_NODE(&object->node)) object 61 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_erase(&object->node, &object->client->objroot); object 65 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_insert(struct nvkm_object *object) object 67 drivers/gpu/drm/nouveau/nvkm/core/object.c struct rb_node **ptr = &object->client->objroot.rb_node; object 73 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->object < this->object) object 76 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->object > this->object) object 82 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_link_node(&object->node, parent, ptr); object 83 drivers/gpu/drm/nouveau/nvkm/core/object.c rb_insert_color(&object->node, &object->client->objroot); object 88 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 90 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->mthd)) object 91 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->mthd(object, mthd, data, size); object 96 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_ntfy(struct nvkm_object *object, u32 mthd, object 99 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->ntfy)) object 100 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->ntfy(object, mthd, pevent); object 105 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_map(struct nvkm_object *object, void *argv, u32 argc, object 108 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->map)) object 109 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->map(object, argv, argc, type, addr, size); object 114 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_unmap(struct nvkm_object *object) object 116 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->unmap)) object 117 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->unmap(object); object 122 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_rd08(struct nvkm_object *object, u64 addr, u8 *data) object 124 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->rd08)) object 125 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->rd08(object, addr, data); object 130 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_rd16(struct nvkm_object *object, u64 addr, u16 *data) object 132 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->rd16)) object 133 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->rd16(object, addr, data); object 138 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_rd32(struct nvkm_object *object, u64 addr, u32 *data) object 140 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->rd32)) object 141 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->rd32(object, addr, data); object 146 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_wr08(struct nvkm_object *object, u64 addr, u8 data) object 148 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->wr08)) object 149 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->wr08(object, addr, data); object 154 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_wr16(struct nvkm_object *object, u64 addr, u16 data) object 156 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->wr16)) object 157 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->wr16(object, addr, data); object 162 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_wr32(struct nvkm_object *object, u64 addr, u32 data) object 164 drivers/gpu/drm/nouveau/nvkm/core/object.c if (likely(object->func->wr32)) object 165 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->wr32(object, addr, data); object 170 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *gpuobj, object 173 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->bind) object 174 drivers/gpu/drm/nouveau/nvkm/core/object.c return object->func->bind(object, gpuobj, align, pgpuobj); object 179 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_fini(struct nvkm_object *object, bool suspend) object 186 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "%s children...\n", action); object 188 drivers/gpu/drm/nouveau/nvkm/core/object.c list_for_each_entry(child, &object->tree, head) { object 194 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "%s running...\n", action); object 195 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->fini) { object 196 drivers/gpu/drm/nouveau/nvkm/core/object.c ret = object->func->fini(object, suspend); object 198 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_error(object, "%s failed with %d\n", action, ret); object 205 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "%s completed in %lldus\n", action, time); object 209 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->init) { object 210 drivers/gpu/drm/nouveau/nvkm/core/object.c int rret = object->func->init(object); object 212 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_fatal(object, "failed to restart, %d\n", rret); object 215 drivers/gpu/drm/nouveau/nvkm/core/object.c list_for_each_entry_continue_reverse(child, &object->tree, head) { object 222 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_init(struct nvkm_object *object) object 228 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "init running...\n"); object 230 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->init) { object 231 drivers/gpu/drm/nouveau/nvkm/core/object.c ret = object->func->init(object); object 236 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "init children...\n"); object 237 drivers/gpu/drm/nouveau/nvkm/core/object.c list_for_each_entry(child, &object->tree, head) { object 244 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "init completed in %lldus\n", time); object 248 drivers/gpu/drm/nouveau/nvkm/core/object.c list_for_each_entry_continue_reverse(child, &object->tree, head) object 251 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_error(object, "init failed with %d\n", ret); object 252 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->fini) object 253 drivers/gpu/drm/nouveau/nvkm/core/object.c object->func->fini(object, false); object 258 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_dtor(struct nvkm_object *object) object 261 drivers/gpu/drm/nouveau/nvkm/core/object.c void *data = object; object 264 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "destroy children...\n"); object 266 drivers/gpu/drm/nouveau/nvkm/core/object.c list_for_each_entry_safe(child, ctemp, &object->tree, head) { object 270 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "destroy running...\n"); object 271 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_unmap(object); object 272 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object->func->dtor) object 273 drivers/gpu/drm/nouveau/nvkm/core/object.c data = object->func->dtor(object); object 274 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_engine_unref(&object->engine); object 276 drivers/gpu/drm/nouveau/nvkm/core/object.c nvif_debug(object, "destroy completed in %lldus...\n", time); object 283 drivers/gpu/drm/nouveau/nvkm/core/object.c struct nvkm_object *object = *pobject; object 284 drivers/gpu/drm/nouveau/nvkm/core/object.c if (object && !WARN_ON(!object->func)) { object 285 drivers/gpu/drm/nouveau/nvkm/core/object.c *pobject = nvkm_object_dtor(object); object 286 drivers/gpu/drm/nouveau/nvkm/core/object.c nvkm_object_remove(object); object 287 drivers/gpu/drm/nouveau/nvkm/core/object.c list_del(&object->head); object 295 drivers/gpu/drm/nouveau/nvkm/core/object.c const struct nvkm_oclass *oclass, struct nvkm_object *object) object 297 drivers/gpu/drm/nouveau/nvkm/core/object.c object->func = func; object 298 drivers/gpu/drm/nouveau/nvkm/core/object.c object->client = oclass->client; object 299 drivers/gpu/drm/nouveau/nvkm/core/object.c object->engine = nvkm_engine_ref(oclass->engine); object 300 drivers/gpu/drm/nouveau/nvkm/core/object.c object->oclass = oclass->base.oclass; object 301 drivers/gpu/drm/nouveau/nvkm/core/object.c object->handle = oclass->handle; object 302 drivers/gpu/drm/nouveau/nvkm/core/object.c object->route = oclass->route; object 303 drivers/gpu/drm/nouveau/nvkm/core/object.c object->token = oclass->token; object 304 drivers/gpu/drm/nouveau/nvkm/core/object.c object->object = oclass->object; object 305 drivers/gpu/drm/nouveau/nvkm/core/object.c INIT_LIST_HEAD(&object->head); object 306 drivers/gpu/drm/nouveau/nvkm/core/object.c INIT_LIST_HEAD(&object->tree); object 307 drivers/gpu/drm/nouveau/nvkm/core/object.c RB_CLEAR_NODE(&object->node); object 308 drivers/gpu/drm/nouveau/nvkm/core/object.c WARN_ON(IS_ERR(object->engine)); object 27 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 29 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_mthd(nvkm_oproxy(object)->object, mthd, data, size); object 33 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_ntfy(struct nvkm_object *object, u32 mthd, object 36 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_ntfy(nvkm_oproxy(object)->object, mthd, pevent); object 40 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_map(struct nvkm_object *object, void *argv, u32 argc, object 43 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c struct nvkm_oproxy *oproxy = nvkm_oproxy(object); object 44 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_map(oproxy->object, argv, argc, type, addr, size); object 48 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_unmap(struct nvkm_object *object) object 50 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_unmap(nvkm_oproxy(object)->object); object 54 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_rd08(struct nvkm_object *object, u64 addr, u8 *data) object 56 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_rd08(nvkm_oproxy(object)->object, addr, data); object 60 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_rd16(struct nvkm_object *object, u64 addr, u16 *data) object 62 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_rd16(nvkm_oproxy(object)->object, addr, data); object 66 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_rd32(struct nvkm_object *object, u64 addr, u32 *data) object 68 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_rd32(nvkm_oproxy(object)->object, addr, data); object 72 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_wr08(struct nvkm_object *object, u64 addr, u8 data) object 74 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_wr08(nvkm_oproxy(object)->object, addr, data); object 78 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_wr16(struct nvkm_object *object, u64 addr, u16 data) object 80 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_wr16(nvkm_oproxy(object)->object, addr, data); object 84 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_wr32(struct nvkm_object *object, u64 addr, u32 data) object 86 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_wr32(nvkm_oproxy(object)->object, addr, data); object 90 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 93 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return nvkm_object_bind(nvkm_oproxy(object)->object, object 98 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_sclass(struct nvkm_object *object, int index, object 101 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c struct nvkm_oproxy *oproxy = nvkm_oproxy(object); object 102 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c oclass->parent = oproxy->object; object 103 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c if (!oproxy->object->func->sclass) object 105 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c return oproxy->object->func->sclass(oproxy->object, index, oclass); object 109 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_fini(struct nvkm_object *object, bool suspend) object 111 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c struct nvkm_oproxy *oproxy = nvkm_oproxy(object); object 120 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c if (oproxy->object->func->fini) { object 121 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c ret = oproxy->object->func->fini(oproxy->object, suspend); object 136 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_init(struct nvkm_object *object) object 138 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c struct nvkm_oproxy *oproxy = nvkm_oproxy(object); object 147 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c if (oproxy->object->func->init) { object 148 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c ret = oproxy->object->func->init(oproxy->object); object 163 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_oproxy_dtor(struct nvkm_object *object) object 165 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c struct nvkm_oproxy *oproxy = nvkm_oproxy(object); object 168 drivers/gpu/drm/nouveau/nvkm/core/oproxy.c nvkm_object_del(&oproxy->object); object 60 drivers/gpu/drm/nouveau/nvkm/core/ramht.c nvkm_ramht_update(struct nvkm_ramht *ramht, int co, struct nvkm_object *object, object 71 drivers/gpu/drm/nouveau/nvkm/core/ramht.c if (object) { object 72 drivers/gpu/drm/nouveau/nvkm/core/ramht.c ret = nvkm_object_bind(object, ramht->parent, 16, &data->inst); object 107 drivers/gpu/drm/nouveau/nvkm/core/ramht.c nvkm_ramht_insert(struct nvkm_ramht *ramht, struct nvkm_object *object, object 118 drivers/gpu/drm/nouveau/nvkm/core/ramht.c return nvkm_ramht_update(ramht, co, object, chid, object 60 drivers/gpu/drm/nouveau/nvkm/engine/ce/gt215.c chan ? chan->object.client->name : "unknown", object 34 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c g84_cipher_oclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 37 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, object 41 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c nvkm_wo32(*pgpuobj, 0x00, object->oclass); object 56 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c g84_cipher_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 59 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c return nvkm_gpuobj_new(object->engine->subdev.device, 256, object 99 drivers/gpu/drm/nouveau/nvkm/engine/cipher/g84.c chan ? chan->object.client->name : "unknown", object 2614 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c nvkm_device_event_ctor(struct nvkm_object *object, void *data, u32 size, object 43 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, "control pstate info size %d\n", size); object 45 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, "control pstate info vers %d\n", object 81 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, "control pstate attr size %d\n", size); object 83 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, object 149 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, "control pstate user size %d\n", size); object 151 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvif_ioctl(&ctrl->object, object 170 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvkm_control_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 172 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c struct nvkm_control *ctrl = nvkm_control(object); object 199 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c *pobject = &ctrl->object; object 202 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c nvkm_object_ctor(&nvkm_control, oclass, &ctrl->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.h #define nvkm_control(p) container_of((p), struct nvkm_control, object) object 8 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.h struct nvkm_object object; object 24 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c #define nvkm_udevice(p) container_of((p), struct nvkm_udevice, object) object 38 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_object object; object 105 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_object *object = &udev->object; object 115 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device info size %d\n", size); object 117 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device info vers %d count %d\n", object 127 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device info vers %d\n", args->v0.version); object 201 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_object *object = &udev->object; object 208 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device time size %d\n", size); object 210 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device time vers %d\n", args->v0.version); object 218 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 220 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 221 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvif_ioctl(object, "device mthd %08x\n", mthd); object 234 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_rd08(struct nvkm_object *object, u64 addr, u8 *data) object 236 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 242 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_rd16(struct nvkm_object *object, u64 addr, u16 *data) object 244 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 250 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_rd32(struct nvkm_object *object, u64 addr, u32 *data) object 252 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 258 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_wr08(struct nvkm_object *object, u64 addr, u8 data) object 260 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 266 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_wr16(struct nvkm_object *object, u64 addr, u16 data) object 268 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 274 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_wr32(struct nvkm_object *object, u64 addr, u32 data) object 276 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 282 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_map(struct nvkm_object *object, void *argv, u32 argc, object 285 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 294 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_fini(struct nvkm_object *object, bool suspend) object 296 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 315 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_init(struct nvkm_object *object) object 317 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 345 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_udevice_child_get(struct nvkm_object *object, int index, object 348 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_udevice *udev = nvkm_udevice(object); object 416 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c struct nvkm_object *parent = &client->object; object 436 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c nvkm_object_ctor(func, oclass, &udev->object); object 437 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c *pobject = &udev->object; object 61 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c nvkm_disp_vblank_ctor(struct nvkm_object *object, void *data, u32 size, object 98 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c nvkm_disp_hpd_ctor(struct nvkm_object *object, void *data, u32 size, object 132 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c nvkm_disp_ntfy(struct nvkm_object *object, u32 type, struct nvkm_event **event) object 134 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c struct nvkm_disp *disp = nvkm_disp(object->engine); object 186 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c return sclass->ctor(disp, oclass, data, size, &oproxy->object); object 134 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_uevent_ctor(struct nvkm_object *object, void *data, u32 size, object 137 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 177 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_rd32(struct nvkm_object *object, u64 addr, u32 *data) object 179 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 187 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_wr32(struct nvkm_object *object, u64 addr, u32 data) object 189 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 197 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_ntfy(struct nvkm_object *object, u32 type, object 200 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 213 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_map(struct nvkm_object *object, void *argv, u32 argc, object 216 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 233 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan_object *object = object 234 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c container_of(base, typeof(*object), oproxy); object 235 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nvkm_ramht_remove(object->disp->ramht, object->hash); object 251 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan_object *object; object 254 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c if (!(object = kzalloc(sizeof(*object), GFP_KERNEL))) object 256 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nvkm_oproxy_ctor(&nv50_disp_chan_child_func_, oclass, &object->oproxy); object 257 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c object->disp = disp; object 258 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c *pobject = &object->oproxy.base; object 260 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c ret = sclass->ctor(device, oclass, argv, argc, &object->oproxy.object); object 264 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c object->hash = chan->func->bind(chan, object->oproxy.object, object 266 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c if (object->hash < 0) object 267 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c return object->hash; object 273 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_child_get(struct nvkm_object *object, int index, object 276 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 298 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_fini(struct nvkm_object *object, bool suspend) object 300 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 307 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_init(struct nvkm_object *object) object 309 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 315 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nv50_disp_chan_dtor(struct nvkm_object *object) object 317 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c struct nv50_disp_chan *chan = nv50_disp_chan(object); object 348 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c *pobject = &chan->object; object 350 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c nvkm_object_ctor(&nv50_disp_chan, oclass, &chan->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h #define nv50_disp_chan(p) container_of((p), struct nv50_disp_chan, object) object 20 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.h struct nvkm_object object; object 31 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c struct nvkm_object *object, u32 handle) object 33 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgf119.c return nvkm_ramht_insert(chan->disp->ramht, object, object 42 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgv100.c struct nvkm_object *object, u32 handle) object 44 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacgv100.c return nvkm_ramht_insert(chan->disp->ramht, object, object 71 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c struct nvkm_object *object, u32 handle) object 73 drivers/gpu/drm/nouveau/nvkm/engine/disp/dmacnv50.c return nvkm_ramht_insert(chan->disp->ramht, object, object 43 drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c nvkm_head_mthd_scanoutpos(struct nvkm_object *object, object 51 drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c nvif_ioctl(object, "head scanoutpos size %d\n", size); object 53 drivers/gpu/drm/nouveau/nvkm/engine/disp/head.c nvif_ioctl(object, "head scanoutpos vers %d\n", object 24 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c #define nv04_disp_root(p) container_of((p), struct nv04_disp_root, object) object 35 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c struct nvkm_object object; object 40 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c nv04_disp_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 42 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c struct nv04_disp_root *root = nv04_disp_root(object); object 49 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c nvif_ioctl(object, "disp mthd size %d\n", size); object 51 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c nvif_ioctl(object, "disp mthd vers %d mthd %02x head %d\n", object 63 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c return nvkm_head_mthd_scanoutpos(object, head, data, size); object 86 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c *pobject = &root->object; object 88 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c nvkm_object_ctor(&nv04_disp_root, oclass, &root->object); object 37 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nv50_disp_root_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 43 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c struct nv50_disp_root *root = nv50_disp_root(object); object 53 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp mthd size %d\n", size); object 55 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp mthd vers %d mthd %02x head %d\n", object 61 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp mthd vers %d mthd %02x " object 89 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c return nvkm_head_mthd_scanoutpos(object, head, data, size); object 142 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor hda eld size %d\n", size); object 144 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor hda eld vers %d\n", object 176 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor hdmi ctrl size %d\n", size); object 178 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor hdmi ctrl vers %d state %d " object 219 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor lvds script size %d\n", size); object 221 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor lvds script " object 236 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor dp mst link size %d\n", size); object 238 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor dp mst link vers %d state %d\n", object 251 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor dp mst vcpi size %d\n", size); object 253 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvif_ioctl(object, "disp sor dp mst vcpi vers %d " object 287 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nv50_disp_root_child_get_(struct nvkm_object *object, int index, object 290 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c struct nv50_disp_root *root = nv50_disp_root(object); object 303 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nv50_disp_root_dtor_(struct nvkm_object *object) object 305 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c struct nv50_disp_root *root = nv50_disp_root(object); object 327 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c *pobject = &root->object; object 329 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c nvkm_object_ctor(&nv50_disp_root_, oclass, &root->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.h #define nv50_disp_root(p) container_of((p), struct nv50_disp_root, object) object 11 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.h struct nvkm_object object; object 42 drivers/gpu/drm/nouveau/nvkm/engine/dma/base.c *pobject = &dmaobj->object; object 38 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c struct nvkm_object *object; object 40 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c object = nvkm_object_search(client, handle, &nvkm_dmaobj_func); object 41 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c if (IS_ERR(object)) object 42 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c return (void *)object; object 44 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c return nvkm_dmaobj(object); object 84 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.c nvkm_object_ctor(&nvkm_dmaobj_func, oclass, &dmaobj->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/dma/user.h #define nvkm_dmaobj(p) container_of((p), struct nvkm_dmaobj, object) object 46 drivers/gpu/drm/nouveau/nvkm/engine/falcon.c nvkm_falcon_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 49 drivers/gpu/drm/nouveau/nvkm/engine/falcon.c return nvkm_gpuobj_new(object->engine->subdev.device, 256, object 129 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c nvkm_fifo_kevent_ctor(struct nvkm_object *object, void *data, u32 size, object 132 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 148 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c nvkm_fifo_cevent_ctor(struct nvkm_object *object, void *data, u32 size, object 186 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c nvkm_fifo_uevent_ctor(struct nvkm_object *object, void *data, u32 size, object 41 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan_object *object = object 42 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c container_of(base, typeof(*object), oproxy); object 43 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_engine *engine = object->oproxy.object->engine; object 44 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = object->chan; object 55 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvif_error(&chan->object, object 61 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (engn->object) { object 62 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c ret = nvkm_object_fini(engn->object, suspend); object 67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvif_trace(&chan->object, "detached %s\n", name); object 74 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan_object *object = object 75 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c container_of(base, typeof(*object), oproxy); object 76 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_engine *engine = object->oproxy.object->engine; object 77 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = object->chan; object 85 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (engn->object) { object 86 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c ret = nvkm_object_init(engn->object); object 94 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvif_error(&chan->object, object 100 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvif_trace(&chan->object, "attached %s\n", name); object 107 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan_object *object = object 108 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c container_of(base, typeof(*object), oproxy); object 109 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_engine *engine = object->oproxy.base.engine; object 110 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = object->chan; object 114 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c chan->func->object_dtor(chan, object->hash); object 119 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_object_del(&engn->object); object 139 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan_object *object; object 142 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (!(object = kzalloc(sizeof(*object), GFP_KERNEL))) object 144 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_oproxy_ctor(&nvkm_fifo_chan_child_func, oclass, &object->oproxy); object 145 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c object->chan = chan; object 146 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c *pobject = &object->oproxy.base; object 159 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c &engn->object); object 163 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c NULL, 0, &engn->object); object 170 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c engn->object); object 180 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c .object = oclass->object, object 182 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c .parent = engn->object ? object 183 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c engn->object : object 186 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c }, data, size, &object->oproxy.object); object 191 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c object->hash = object 192 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c chan->func->object_ctor(chan, object->oproxy.object); object 193 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (object->hash < 0) object 194 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c return object->hash; object 201 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_child_get(struct nvkm_object *object, int index, object 204 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 246 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_ntfy(struct nvkm_object *object, u32 type, object 249 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 256 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_map(struct nvkm_object *object, void *argv, u32 argc, object 259 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 267 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_rd32(struct nvkm_object *object, u64 addr, u32 *data) object 269 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 282 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_wr32(struct nvkm_object *object, u64 addr, u32 data) object 284 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 297 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_fini(struct nvkm_object *object, bool suspend) object 299 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 305 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_init(struct nvkm_object *object) object 307 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 313 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_fifo_chan_dtor(struct nvkm_object *object) object 315 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c struct nvkm_fifo_chan *chan = nvkm_fifo_chan(object); object 365 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c nvkm_object_ctor(&nvkm_fifo_chan_func, oclass, &chan->object); object 382 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c ret = nvkm_object_bind(&dmaobj->object, chan->inst, -16, object 4 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.h #define nvkm_fifo_chan(p) container_of((p), struct nvkm_fifo_chan, object) object 115 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c chan->base.chid, chan->base.object.client->name); object 162 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c struct nvkm_object *object) object 170 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c return nvkm_object_bind(object, NULL, 0, &chan->engn[engn]); object 175 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c struct nvkm_object *object) object 178 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c u32 handle = object->handle; object 181 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c switch (object->engine->subdev.index) { object 201 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chang84.c return nvkm_ramht_insert(chan->ramht, object, 0, 4, handle, context); object 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c chan->base.chid, chan->base.object.client->name); object 139 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c struct nvkm_object *object) object 147 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c return nvkm_object_bind(object, NULL, 0, &chan->engn[engn]); object 159 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c struct nvkm_object *object) object 162 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c u32 handle = object->handle; object 165 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c switch (object->engine->subdev.index) { object 175 drivers/gpu/drm/nouveau/nvkm/engine/fifo/channv50.c return nvkm_ramht_insert(chan->ramht, object, 0, 4, handle, context); object 58 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c *pobject = &chan->base.object; object 48 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c struct nvkm_object *object) object 53 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c u32 handle = object->handle; object 56 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c switch (object->engine->subdev.index) { object 67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, object 190 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c *pobject = &chan->base.object; object 61 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv10.c *pobject = &chan->base.object; object 61 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv17.c *pobject = &chan->base.object; object 130 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c struct nvkm_object *object) object 139 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c return nvkm_object_bind(object, NULL, 0, &chan->engn[engn]); object 144 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c struct nvkm_object *object) object 149 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c u32 handle = object->handle; object 152 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c switch (object->engine->subdev.index) { object 163 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c hash = nvkm_ramht_insert(imem->ramht, object, chan->base.chid, 4, object 208 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c *pobject = &chan->base.object; object 58 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c *pobject = &chan->base.object; object 303 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c info->inst, chan ? chan->object.client->name : "unknown"); object 433 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c chan ? chan->object.client->name : "unknown", object 528 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c info->inst, chan ? chan->object.client->name : "unknown"); object 715 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c chan ? chan->object.client->name : "unknown", object 60 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c *pobject = &chan->base.object; object 87 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogf100.c chan->base.chid, chan->base.object.client->name); object 136 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogf100.c struct nvkm_object *object) object 145 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogf100.c ret = nvkm_object_bind(object, NULL, 0, &chan->engn[engn].inst); object 240 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogf100.c *pobject = &chan->base.object; object 43 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogk104.c struct nvkm_client *client = chan->base.object.client; object 161 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogk104.c struct nvkm_object *object) object 170 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogk104.c ret = nvkm_object_bind(object, NULL, 0, &chan->engn[engn].inst); object 270 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogk104.c *pobject = &chan->base.object; object 150 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogv100.c *pobject = &chan->base.object; object 60 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c *pobject = &chan->base.object; object 167 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c chid, chan ? chan->object.client->name : "unknown", object 201 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c name = chan ? chan->object.client->name : "unknown"; object 25 drivers/gpu/drm/nouveau/nvkm/engine/fifo/usergv100.c gv100_fifo_user_map(struct nvkm_object *object, void *argv, u32 argc, object 28 drivers/gpu/drm/nouveau/nvkm/engine/fifo/usergv100.c struct nvkm_device *device = object->engine->subdev.device; object 25 drivers/gpu/drm/nouveau/nvkm/engine/fifo/usertu102.c tu102_fifo_user_map(struct nvkm_object *object, void *argv, u32 argc, object 28 drivers/gpu/drm/nouveau/nvkm/engine/fifo/usertu102.c struct nvkm_device *device = object->engine->subdev.device; object 153 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c #define gf100_gr_object(p) container_of((p), struct gf100_gr_object, object) object 156 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct nvkm_object object; object 161 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c gf100_fermi_mthd_zbc_color(struct nvkm_object *object, void *data, u32 size) object 163 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct gf100_gr *gr = gf100_gr(nvkm_gr(object->engine)); object 207 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c gf100_fermi_mthd_zbc_depth(struct nvkm_object *object, void *data, u32 size) object 209 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct gf100_gr *gr = gf100_gr(nvkm_gr(object->engine)); object 231 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c gf100_fermi_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 233 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c nvif_ioctl(object, "fermi mthd %08x\n", mthd); object 236 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c return gf100_fermi_mthd_zbc_color(object, data, size); object 238 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c return gf100_fermi_mthd_zbc_depth(object, data, size); object 286 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct gf100_gr_object *object; object 288 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c if (!(object = kzalloc(sizeof(*object), GFP_KERNEL))) object 290 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c *pobject = &object->object; object 293 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c &gf100_gr_object_func, oclass, &object->object); object 294 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c object->chan = chan; object 320 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c gf100_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 323 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct gf100_gr_chan *chan = gf100_gr_chan(object); object 355 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c gf100_gr_chan_dtor(struct nvkm_object *object) object 357 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c struct gf100_gr_chan *chan = gf100_gr_chan(object); object 392 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c nvkm_object_ctor(&gf100_gr_chan, oclass, &chan->object); object 395 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c *pobject = &chan->object; object 1566 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c name = chan->object.client->name; object 252 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h #define gf100_gr_chan(p) container_of((p), struct gf100_gr_chan, object) object 256 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.h struct nvkm_object object; object 357 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c #define nv04_gr_chan(p) container_of((p), struct nv04_gr_chan, object) object 360 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c struct nvkm_object object; object 1043 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nv04_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 1046 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, align, object 1050 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nvkm_wo32(*pgpuobj, 0x00, object->oclass); object 1148 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nv04_gr_chan_dtor(struct nvkm_object *object) object 1150 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c struct nv04_gr_chan *chan = nv04_gr_chan(object); object 1161 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nv04_gr_chan_fini(struct nvkm_object *object, bool suspend) object 1163 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c struct nv04_gr_chan *chan = nv04_gr_chan(object); object 1193 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c nvkm_object_ctor(&nv04_gr_chan, oclass, &chan->object); object 1196 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c *pobject = &chan->object; object 1320 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c chan ? chan->object.client->name : "unknown", object 397 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c #define nv10_gr_chan(p) container_of((p), struct nv10_gr_chan, object) object 400 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nvkm_object object; object 433 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nvkm_device *device = chan->object.engine->subdev.device; object 506 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nvkm_device *device = chan->object.engine->subdev.device; object 954 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c nv10_gr_chan_fini(struct nvkm_object *object, bool suspend) object 956 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nv10_gr_chan *chan = nv10_gr_chan(object); object 971 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c nv10_gr_chan_dtor(struct nvkm_object *object) object 973 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c struct nv10_gr_chan *chan = nv10_gr_chan(object); object 1012 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c nvkm_object_ctor(&nv10_gr_chan, oclass, &chan->object); object 1015 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c *pobject = &chan->object; object 1128 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv10.c chan ? chan->object.client->name : "unknown", object 17 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c nv20_gr_chan_init(struct nvkm_object *object) object 19 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c struct nv20_gr_chan *chan = nv20_gr_chan(object); object 30 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c nv20_gr_chan_fini(struct nvkm_object *object, bool suspend) object 32 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c struct nv20_gr_chan *chan = nv20_gr_chan(object); object 60 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c nv20_gr_chan_dtor(struct nvkm_object *object) object 62 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c struct nv20_gr_chan *chan = nv20_gr_chan(object); object 84 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c nvkm_object_ctor(&nv20_gr_chan, oclass, &chan->object); object 87 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c *pobject = &chan->object; object 212 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.c chan ? chan->object.client->name : "unknown", object 22 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.h #define nv20_gr_chan(p) container_of((p), struct nv20_gr_chan, object) object 26 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv20.h struct nvkm_object object; object 30 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv25.c nvkm_object_ctor(&nv25_gr_chan, oclass, &chan->object); object 33 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv25.c *pobject = &chan->object; object 30 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv2a.c nvkm_object_ctor(&nv2a_gr_chan, oclass, &chan->object); object 33 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv2a.c *pobject = &chan->object; object 31 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv30.c nvkm_object_ctor(&nv30_gr_chan, oclass, &chan->object); object 34 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv30.c *pobject = &chan->object; object 30 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv34.c nvkm_object_ctor(&nv34_gr_chan, oclass, &chan->object); object 33 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv34.c *pobject = &chan->object; object 30 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv35.c nvkm_object_ctor(&nv35_gr_chan, oclass, &chan->object); object 33 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv35.c *pobject = &chan->object; object 44 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nv40_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 47 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 20, align, object 51 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nvkm_wo32(*pgpuobj, 0x00, object->oclass); object 74 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nv40_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 77 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c struct nv40_gr_chan *chan = nv40_gr_chan(object); object 92 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nv40_gr_chan_fini(struct nvkm_object *object, bool suspend) object 94 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c struct nv40_gr_chan *chan = nv40_gr_chan(object); object 130 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nv40_gr_chan_dtor(struct nvkm_object *object) object 132 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c struct nv40_gr_chan *chan = nv40_gr_chan(object); object 157 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c nvkm_object_ctor(&nv40_gr_chan, oclass, &chan->object); object 160 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c *pobject = &chan->object; object 279 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.c chan ? chan->fifo->object.client->name : "unknown", object 19 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.h #define nv40_gr_chan(p) container_of((p), struct nv40_gr_chan, object) object 23 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv40.h struct nvkm_object object; object 43 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c nv50_gr_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 46 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, object 50 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c nvkm_wo32(*pgpuobj, 0x00, object->oclass); object 69 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c nv50_gr_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 72 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c struct nv50_gr *gr = nv50_gr_chan(object)->gr; object 97 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c nvkm_object_ctor(&nv50_gr_chan, oclass, &chan->object); object 99 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c *pobject = &chan->object; object 642 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.c name = chan->object.client->name; object 22 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.h #define nv50_gr_chan(p) container_of((p), struct nv50_gr_chan, object) object 26 drivers/gpu/drm/nouveau/nvkm/engine/gr/nv50.h struct nvkm_object object; object 39 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c nv31_mpeg_object_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 42 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, align, object 46 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c nvkm_wo32(*pgpuobj, 0x00, object->oclass); object 65 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c nv31_mpeg_chan_dtor(struct nvkm_object *object) object 67 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c struct nv31_mpeg_chan *chan = nv31_mpeg_chan(object); object 95 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c nvkm_object_ctor(&nv31_mpeg_chan, oclass, &chan->object); object 98 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c *pobject = &chan->object; object 217 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.c mpeg->chan ? mpeg->chan->object.client->name : object 21 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.h #define nv31_mpeg_chan(p) container_of((p), struct nv31_mpeg_chan, object) object 25 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv31.h struct nvkm_object object; object 41 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c #define nv44_mpeg_chan(p) container_of((p), struct nv44_mpeg_chan, object) object 44 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c struct nvkm_object object; object 52 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c nv44_mpeg_chan_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 55 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object); object 56 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c int ret = nvkm_gpuobj_new(chan->object.engine->subdev.device, 264 * 4, object 68 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c nv44_mpeg_chan_fini(struct nvkm_object *object, bool suspend) object 71 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object); object 84 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c nv44_mpeg_chan_dtor(struct nvkm_object *object) object 86 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c struct nv44_mpeg_chan *chan = nv44_mpeg_chan(object); object 113 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c nvkm_object_ctor(&nv44_mpeg_chan, oclass, &chan->object); object 116 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c *pobject = &chan->object; object 186 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv44.c chan ? chan->object.client->name : "unknown", object 37 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv50.c nv50_mpeg_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 40 drivers/gpu/drm/nouveau/nvkm/engine/mpeg/nv50.c int ret = nvkm_gpuobj_new(object->engine->subdev.device, 128 * 4, object 213 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &dom->object; object 217 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom init size %d\n", size); object 219 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom init\n"); object 243 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &dom->object; object 247 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom sample size %d\n", size); object 249 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom sample\n"); object 267 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &dom->object; object 271 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom read size %d\n", size); object 273 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfdom read vers %d\n", args->v0.version); object 293 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_perfdom_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 295 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_perfdom *dom = nvkm_perfdom(object); object 310 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_perfdom_dtor(struct nvkm_object *object) object 312 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_perfdom *dom = nvkm_perfdom(object); object 418 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_object_ctor(&nvkm_perfdom, oclass, &dom->object); object 420 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c *pobject = &dom->object; object 440 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &perfmon->object; object 446 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfmon query domain size %d\n", size); object 448 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfmon domain vers %d iter %02x\n", object 488 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &perfmon->object; object 497 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfmon query signal size %d\n", size); object 499 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, object 542 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_object *object = &perfmon->object; object 550 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, "perfmon query source size %d\n", size); object 552 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvif_ioctl(object, object 588 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_perfmon_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 590 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_perfmon *perfmon = nvkm_perfmon(object); object 613 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_perfmon_child_get(struct nvkm_object *object, int index, object 627 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_perfmon_dtor(struct nvkm_object *object) object 629 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c struct nvkm_perfmon *perfmon = nvkm_perfmon(object); object 632 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c if (pm->perfmon == &perfmon->object) object 653 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c nvkm_object_ctor(&nvkm_perfmon, oclass, &perfmon->object); object 655 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c *pobject = &perfmon->object; object 70 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h #define nvkm_perfdom(p) container_of((p), struct nvkm_perfdom, object) object 74 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h struct nvkm_object object; object 99 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h #define nvkm_perfmon(p) container_of((p), struct nvkm_perfmon, object) object 102 drivers/gpu/drm/nouveau/nvkm/engine/pm/priv.h struct nvkm_object object; object 59 drivers/gpu/drm/nouveau/nvkm/engine/sec/g98.c chan ? chan->object.client->name : "unknown", object 50 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c nvkm_sw_chan_event_ctor(struct nvkm_object *object, void *data, u32 size, object 73 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c nvkm_sw_chan_dtor(struct nvkm_object *object) object 75 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c struct nvkm_sw_chan *chan = nvkm_sw_chan(object); object 102 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c nvkm_object_ctor(&nvkm_sw_chan, oclass, &chan->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.h #define nvkm_sw_chan(p) container_of((p), struct nvkm_sw_chan, object) object 12 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.h struct nvkm_object object; object 60 drivers/gpu/drm/nouveau/nvkm/engine/sw/gf100.c struct nvkm_engine *engine = chan->base.object.engine; object 115 drivers/gpu/drm/nouveau/nvkm/engine/sw/gf100.c *pobject = &chan->base.object; object 117 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv04.c *pobject = &chan->base.object; object 46 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv10.c *pobject = &chan->object; object 65 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv50.c struct nvkm_engine *engine = chan->base.object.engine; object 109 drivers/gpu/drm/nouveau/nvkm/engine/sw/nv50.c *pobject = &chan->base.object; object 30 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c nvkm_nvsw_mthd_(struct nvkm_object *object, u32 mthd, void *data, u32 size) object 32 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c struct nvkm_nvsw *nvsw = nvkm_nvsw(object); object 39 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c nvkm_nvsw_ntfy_(struct nvkm_object *object, u32 mthd, object 42 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c struct nvkm_nvsw *nvsw = nvkm_nvsw(object); object 68 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c *pobject = &nvsw->object; object 70 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.c nvkm_object_ctor(&nvkm_nvsw_, oclass, &nvsw->object); object 4 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.h #define nvkm_nvsw(p) container_of((p), struct nvkm_nvsw, object) object 8 drivers/gpu/drm/nouveau/nvkm/engine/sw/nvsw.h struct nvkm_object object; object 44 drivers/gpu/drm/nouveau/nvkm/engine/xtensa.c nvkm_xtensa_cclass_bind(struct nvkm_object *object, struct nvkm_gpuobj *parent, object 47 drivers/gpu/drm/nouveau/nvkm/engine/xtensa.c return nvkm_gpuobj_new(object->engine->subdev.device, 0x10000, align, object 42 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c nvkm_fault_ntfy_ctor(struct nvkm_object *object, void *argv, u32 argc, object 45 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); object 203 drivers/gpu/drm/nouveau/nvkm/subdev/fault/gv100.c return nvkm_notify_init(&fault->buffer[0]->object, &fault->event, object 3 drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h #define nvkm_fault_buffer(p) container_of((p), struct nvkm_fault_buffer, object) object 11 drivers/gpu/drm/nouveau/nvkm/subdev/fault/priv.h struct nvkm_object object; object 31 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_ufault_map(struct nvkm_object *object, void *argv, u32 argc, object 34 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); object 43 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_ufault_ntfy(struct nvkm_object *object, u32 type, object 46 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); object 55 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_ufault_fini(struct nvkm_object *object, bool suspend) object 57 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); object 63 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_ufault_init(struct nvkm_object *object) object 65 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); object 71 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_ufault_dtor(struct nvkm_object *object) object 103 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c nvkm_object_ctor(&nvkm_ufault, oclass, &buffer->object); object 104 drivers/gpu/drm/nouveau/nvkm/subdev/fault/user.c *pobject = &buffer->object; object 188 drivers/gpu/drm/nouveau/nvkm/subdev/fb/nv50.c chan ? chan->object.client->name : "unknown", object 127 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c nvkm_gpio_intr_ctor(struct nvkm_object *object, void *data, u32 size, object 108 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c nvkm_i2c_intr_ctor(struct nvkm_object *object, void *data, u32 size, object 37 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c struct nvkm_client *master = client->object.client; object 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c struct nvkm_object *object; object 42 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c object = nvkm_object_search(client, handle, &nvkm_umem); object 43 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c if (IS_ERR(object)) { object 47 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c if (umem->object.object == handle) { object 55 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c umem = nvkm_umem(object); object 64 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c nvkm_umem_unmap(struct nvkm_object *object) object 66 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c struct nvkm_umem *umem = nvkm_umem(object); object 87 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c nvkm_umem_map(struct nvkm_object *object, void *argv, u32 argc, object 90 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c struct nvkm_umem *umem = nvkm_umem(object); object 125 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c nvkm_umem_dtor(struct nvkm_object *object) object 127 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c struct nvkm_umem *umem = nvkm_umem(object); object 128 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c spin_lock(&umem->object.client->lock); object 130 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c spin_unlock(&umem->object.client->lock); object 167 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c nvkm_object_ctor(&nvkm_umem, oclass, &umem->object); object 172 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c *pobject = &umem->object; object 184 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c spin_lock(&umem->object.client->lock); object 185 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c list_add(&umem->head, &umem->object.client->umem); object 186 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.c spin_unlock(&umem->object.client->lock); object 3 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h #define nvkm_umem(p) container_of((p), struct nvkm_umem, object) object 8 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/umem.h struct nvkm_object object; object 32 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c nvkm_ummu_sclass(struct nvkm_object *object, int index, object 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu; object 131 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c nvkm_ummu_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc) object 133 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c struct nvkm_ummu *ummu = nvkm_ummu(object); object 174 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c nvkm_object_ctor(&nvkm_ummu, oclass, &ummu->object); object 176 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c *pobject = &ummu->object; object 3 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.h #define nvkm_ummu(p) container_of((p), struct nvkm_ummu, object) object 8 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.h struct nvkm_object object; object 36 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_object *object; object 38 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c object = nvkm_object_search(client, handle, &nvkm_uvmm); object 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c if (IS_ERR(object)) object 40 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c return (void *)object; object 42 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c return nvkm_uvmm(object)->vmm; object 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 77 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 111 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 154 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 233 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 271 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_client *client = uvmm->object.client; object 333 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc) object 335 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_uvmm *uvmm = nvkm_uvmm(object); object 347 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c uvmm->object.client, object 358 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_uvmm_dtor(struct nvkm_object *object) object 360 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c struct nvkm_uvmm *uvmm = nvkm_uvmm(object); object 395 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object); object 396 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c *pobject = &uvmm->object; object 3 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.h #define nvkm_uvmm(p) container_of((p), struct nvkm_uvmm, object) object 8 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.h struct nvkm_object object; object 376 drivers/iio/temperature/mlx90632.c static s32 mlx90632_calc_temp_object_iteration(s32 prev_object_temp, s64 object, object 395 drivers/iio/temperature/mlx90632.c ir_Alpha = div64_s64((s64)object * 10000000LL, Alpha_corr); object 405 drivers/iio/temperature/mlx90632.c static s32 mlx90632_calc_temp_object(s64 object, s64 ambient, s32 Ea, s32 Eb, object 419 drivers/iio/temperature/mlx90632.c temp = mlx90632_calc_temp_object_iteration(temp, object, TAdut, object 433 drivers/iio/temperature/mlx90632.c s64 object, ambient; object 475 drivers/iio/temperature/mlx90632.c object = mlx90632_preprocess_temp_obj(object_new_raw, object 480 drivers/iio/temperature/mlx90632.c *val = mlx90632_calc_temp_object(object, ambient, Ea, Eb, Fa, Fb, Ga, object 138 drivers/infiniband/core/rdma_core.c if (uobj->object) { object 151 drivers/infiniband/core/rdma_core.c uobj->object = NULL; object 612 drivers/infiniband/core/rdma_core.c uobj->object = filp; object 672 drivers/infiniband/core/rdma_core.c uobj->object = NULL; object 687 drivers/infiniband/core/rdma_core.c struct file *filp = uobj->object; object 738 drivers/infiniband/core/rdma_core.c WARN_ON(entry->object); object 438 drivers/infiniband/core/uverbs_cmd.c uobj->object = pd; object 620 drivers/infiniband/core/uverbs_cmd.c obj->uobject.object = xrcd; object 762 drivers/infiniband/core/uverbs_cmd.c uobj->object = mr; object 814 drivers/infiniband/core/uverbs_cmd.c mr = uobj->object; object 918 drivers/infiniband/core/uverbs_cmd.c uobj->object = mw; object 1037 drivers/infiniband/core/uverbs_cmd.c obj->uobject.object = cq; object 1327 drivers/infiniband/core/uverbs_cmd.c xrcd = (struct ib_xrcd *)xrcd_uobj->object; object 1448 drivers/infiniband/core/uverbs_cmd.c obj->uevent.uobject.object = qp; object 1577 drivers/infiniband/core/uverbs_cmd.c xrcd = (struct ib_xrcd *)xrcd_uobj->object; object 1598 drivers/infiniband/core/uverbs_cmd.c obj->uevent.uobject.object = qp; object 2447 drivers/infiniband/core/uverbs_cmd.c uobj->object = ah; object 2943 drivers/infiniband/core/uverbs_cmd.c obj->uevent.uobject.object = wq; object 2953 drivers/infiniband/core/uverbs_cmd.c obj->uevent.uobject.object = wq; object 3121 drivers/infiniband/core/uverbs_cmd.c uobj->object = rwq_ind_tbl; object 3381 drivers/infiniband/core/uverbs_cmd.c attr.ext.xrc.xrcd = (struct ib_xrcd *)xrcd_uobj->object; object 3446 drivers/infiniband/core/uverbs_cmd.c obj->uevent.uobject.object = srq; object 46 drivers/infiniband/core/uverbs_std_types.c return rdma_destroy_ah_user((struct ib_ah *)uobject->object, object 55 drivers/infiniband/core/uverbs_std_types.c struct ib_flow *flow = (struct ib_flow *)uobject->object; object 75 drivers/infiniband/core/uverbs_std_types.c return uverbs_dealloc_mw((struct ib_mw *)uobject->object); object 82 drivers/infiniband/core/uverbs_std_types.c struct ib_qp *qp = uobject->object; object 116 drivers/infiniband/core/uverbs_std_types.c struct ib_rwq_ind_table *rwq_ind_tbl = uobject->object; object 132 drivers/infiniband/core/uverbs_std_types.c struct ib_wq *wq = uobject->object; object 149 drivers/infiniband/core/uverbs_std_types.c struct ib_srq *srq = uobject->object; object 174 drivers/infiniband/core/uverbs_std_types.c struct ib_xrcd *xrcd = uobject->object; object 194 drivers/infiniband/core/uverbs_std_types.c struct ib_pd *pd = uobject->object; object 42 drivers/infiniband/core/uverbs_std_types_counters.c struct ib_counters *counters = uobject->object; object 77 drivers/infiniband/core/uverbs_std_types_counters.c uobj->object = counters; object 41 drivers/infiniband/core/uverbs_std_types_cq.c struct ib_cq *cq = uobject->object; object 132 drivers/infiniband/core/uverbs_std_types_cq.c obj->uobject.object = cq; object 41 drivers/infiniband/core/uverbs_std_types_dm.c struct ib_dm *dm = uobject->object; object 84 drivers/infiniband/core/uverbs_std_types_dm.c uobj->object = dm; object 41 drivers/infiniband/core/uverbs_std_types_flow_action.c struct ib_flow_action *action = uobject->object; object 342 drivers/infiniband/core/uverbs_std_types_flow_action.c struct ib_flow_action *action = uobj->object; object 41 drivers/infiniband/core/uverbs_std_types_mr.c return ib_dereg_mr_user((struct ib_mr *)uobject->object, object 137 drivers/infiniband/core/uverbs_std_types_mr.c uobj->object = mr; object 579 drivers/infiniband/hw/mlx5/devx.c to_mcq(uobj->object)->mcq.cqn) == object 584 drivers/infiniband/hw/mlx5/devx.c struct mlx5_core_srq *srq = &(to_msrq(uobj->object)->msrq); object 602 drivers/infiniband/hw/mlx5/devx.c to_msrq(uobj->object)->msrq.srqn) == object 608 drivers/infiniband/hw/mlx5/devx.c struct mlx5_ib_qp *qp = to_mqp(uobj->object); object 638 drivers/infiniband/hw/mlx5/devx.c to_mrwq(uobj->object)->core_qp.qpn) == object 643 drivers/infiniband/hw/mlx5/devx.c to_mrwq_ind_table(uobj->object)->rqtn) == object 647 drivers/infiniband/hw/mlx5/devx.c return ((struct devx_obj *)uobj->object)->obj_id == obj_id; object 1336 drivers/infiniband/hw/mlx5/devx.c struct devx_obj *obj = uobject->object; object 1466 drivers/infiniband/hw/mlx5/devx.c uobj->object = obj; object 1678 drivers/infiniband/hw/mlx5/devx.c fput(fd_uobj->object); object 1749 drivers/infiniband/hw/mlx5/devx.c get_file(fd_uobj->object); object 1764 drivers/infiniband/hw/mlx5/devx.c fput(fd_uobj->object); object 1945 drivers/infiniband/hw/mlx5/devx.c obj = (struct devx_obj *)devx_uobj->object; object 2035 drivers/infiniband/hw/mlx5/devx.c event_sub->filp = fd_uobj->object; object 2210 drivers/infiniband/hw/mlx5/devx.c uobj->object = obj; object 2231 drivers/infiniband/hw/mlx5/devx.c struct devx_umem *obj = uobject->object; object 152 drivers/infiniband/hw/mlx5/flow.c devx_obj = arr_flow_actions[0]->object; object 176 drivers/infiniband/hw/mlx5/flow.c to_mflow_act(arr_flow_actions[i]->object); object 182 drivers/infiniband/hw/mlx5/flow.c arr_flow_actions[i]->object); object 218 drivers/infiniband/hw/mlx5/flow.c struct mlx5_ib_flow_matcher *obj = uobject->object; object 319 drivers/infiniband/hw/mlx5/flow.c uobj->object = obj; object 261 drivers/input/gameport/gameport.c void *object; object 300 drivers/input/gameport/gameport.c if (event->object == e->object) { object 335 drivers/input/gameport/gameport.c gameport_add_port(event->object); object 339 drivers/input/gameport/gameport.c gameport_attach_driver(event->object); object 352 drivers/input/gameport/gameport.c static int gameport_queue_event(void *object, struct module *owner, object 369 drivers/input/gameport/gameport.c if (event->object == object) { object 392 drivers/input/gameport/gameport.c event->object = object; object 407 drivers/input/gameport/gameport.c static void gameport_remove_pending_events(void *object) object 415 drivers/input/gameport/gameport.c if (event->object == object) { object 442 drivers/input/gameport/gameport.c gameport = event->object; object 588 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; object 592 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; object 596 drivers/input/serio/hil_mlc.c mlc->imatch = node->object.packet; object 640 drivers/input/serio/hil_mlc.c BUG_ON(node->object.func == NULL); object 641 drivers/input/serio/hil_mlc.c rc = node->object.func(mlc, node->arg); object 673 drivers/input/serio/hil_mlc.c pack = node->object.packet; object 679 drivers/input/serio/hil_mlc.c pack = node->object.packet; object 685 drivers/input/serio/hil_mlc.c pack = node->object.packet; object 142 drivers/input/serio/serio.c void *object; object 173 drivers/input/serio/serio.c static void serio_remove_duplicate_events(void *object, object 182 drivers/input/serio/serio.c if (object == e->object) { object 210 drivers/input/serio/serio.c serio_add_port(event->object); object 214 drivers/input/serio/serio.c serio_reconnect_port(event->object); object 218 drivers/input/serio/serio.c serio_disconnect_port(event->object); object 219 drivers/input/serio/serio.c serio_find_driver(event->object); object 223 drivers/input/serio/serio.c serio_reconnect_subtree(event->object); object 227 drivers/input/serio/serio.c serio_attach_driver(event->object); object 231 drivers/input/serio/serio.c serio_remove_duplicate_events(event->object, event->type); object 240 drivers/input/serio/serio.c static int serio_queue_event(void *object, struct module *owner, object 257 drivers/input/serio/serio.c if (event->object == object) { object 280 drivers/input/serio/serio.c event->object = object; object 295 drivers/input/serio/serio.c static void serio_remove_pending_events(void *object) object 303 drivers/input/serio/serio.c if (event->object == object) { object 328 drivers/input/serio/serio.c serio = event->object; object 696 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 700 drivers/input/touchscreen/atmel_mxt_ts.c object = data->object_table + i; object 701 drivers/input/touchscreen/atmel_mxt_ts.c if (object->type == type) object 702 drivers/input/touchscreen/atmel_mxt_ts.c return object; object 745 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 748 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, type); object 749 drivers/input/touchscreen/atmel_mxt_ts.c if (!object || offset >= mxt_obj_size(object)) object 752 drivers/input/touchscreen/atmel_mxt_ts.c reg = object->start_address; object 1289 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 1310 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, type); object 1311 drivers/input/touchscreen/atmel_mxt_ts.c if (!object) { object 1326 drivers/input/touchscreen/atmel_mxt_ts.c if (size > mxt_obj_size(object)) { object 1333 drivers/input/touchscreen/atmel_mxt_ts.c size - mxt_obj_size(object), type); object 1334 drivers/input/touchscreen/atmel_mxt_ts.c } else if (mxt_obj_size(object) > size) { object 1345 drivers/input/touchscreen/atmel_mxt_ts.c mxt_obj_size(object) - size, type); object 1348 drivers/input/touchscreen/atmel_mxt_ts.c if (instance >= mxt_obj_instances(object)) { object 1353 drivers/input/touchscreen/atmel_mxt_ts.c reg = object->start_address + mxt_obj_size(object) * instance; object 1366 drivers/input/touchscreen/atmel_mxt_ts.c if (i > mxt_obj_size(object)) object 1375 drivers/input/touchscreen/atmel_mxt_ts.c reg, object->type, byte_offset); object 1627 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object = object_table + i; object 1630 drivers/input/touchscreen/atmel_mxt_ts.c le16_to_cpus(&object->start_address); object 1632 drivers/input/touchscreen/atmel_mxt_ts.c if (object->num_report_ids) { object 1634 drivers/input/touchscreen/atmel_mxt_ts.c reportid += object->num_report_ids * object 1635 drivers/input/touchscreen/atmel_mxt_ts.c mxt_obj_instances(object); object 1644 drivers/input/touchscreen/atmel_mxt_ts.c object->type, object->start_address, object 1645 drivers/input/touchscreen/atmel_mxt_ts.c mxt_obj_size(object), mxt_obj_instances(object), object 1648 drivers/input/touchscreen/atmel_mxt_ts.c switch (object->type) { object 1657 drivers/input/touchscreen/atmel_mxt_ts.c data->T5_msg_size = mxt_obj_size(object); object 1660 drivers/input/touchscreen/atmel_mxt_ts.c data->T5_msg_size = mxt_obj_size(object) - 1; object 1662 drivers/input/touchscreen/atmel_mxt_ts.c data->T5_address = object->start_address; object 1666 drivers/input/touchscreen/atmel_mxt_ts.c data->T6_address = object->start_address; object 1669 drivers/input/touchscreen/atmel_mxt_ts.c data->T7_address = object->start_address; object 1672 drivers/input/touchscreen/atmel_mxt_ts.c data->T71_address = object->start_address; object 1679 drivers/input/touchscreen/atmel_mxt_ts.c object->num_report_ids - 1; object 1680 drivers/input/touchscreen/atmel_mxt_ts.c data->num_touchids = object->num_report_ids; object 1683 drivers/input/touchscreen/atmel_mxt_ts.c data->T44_address = object->start_address; object 1693 drivers/input/touchscreen/atmel_mxt_ts.c data->num_touchids = object->num_report_ids - 2; object 1697 drivers/input/touchscreen/atmel_mxt_ts.c end_address = object->start_address object 1698 drivers/input/touchscreen/atmel_mxt_ts.c + mxt_obj_size(object) * mxt_obj_instances(object) - 1; object 1815 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 1817 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, MXT_TOUCH_MULTI_T9); object 1818 drivers/input/touchscreen/atmel_mxt_ts.c if (!object) object 1822 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T9_XSIZE, object 1828 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T9_YSIZE, object 1834 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T9_RANGE, object 1843 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T9_ORIENT, object 1859 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 1864 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, MXT_TOUCH_MULTITOUCHSCREEN_T100); object 1865 drivers/input/touchscreen/atmel_mxt_ts.c if (!object) object 1870 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_XRANGE, object 1878 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_YRANGE, object 1886 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_XSIZE, object 1892 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_YSIZE, object 1899 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_CFG1, object 1910 drivers/input/touchscreen/atmel_mxt_ts.c object->start_address + MXT_T100_TCHAUX, object 2555 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 2558 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, MXT_GEN_COMMAND_T6); object 2559 drivers/input/touchscreen/atmel_mxt_ts.c if (!object) object 2562 drivers/input/touchscreen/atmel_mxt_ts.c dbg->diag_cmd_address = object->start_address + MXT_COMMAND_DIAGNOSTIC; object 2564 drivers/input/touchscreen/atmel_mxt_ts.c object = mxt_get_object(data, MXT_DEBUG_DIAGNOSTIC_T37); object 2565 drivers/input/touchscreen/atmel_mxt_ts.c if (!object) object 2568 drivers/input/touchscreen/atmel_mxt_ts.c if (mxt_obj_size(object) != sizeof(struct t37_debug)) { object 2573 drivers/input/touchscreen/atmel_mxt_ts.c dbg->t37_address = object->start_address; object 2687 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object, int instance, object 2692 drivers/input/touchscreen/atmel_mxt_ts.c if (mxt_obj_instances(object) > 1) object 2696 drivers/input/touchscreen/atmel_mxt_ts.c for (i = 0; i < mxt_obj_size(object); i++) object 2708 drivers/input/touchscreen/atmel_mxt_ts.c struct mxt_object *object; object 2721 drivers/input/touchscreen/atmel_mxt_ts.c object = data->object_table + i; object 2723 drivers/input/touchscreen/atmel_mxt_ts.c if (!mxt_object_readable(object->type)) object 2727 drivers/input/touchscreen/atmel_mxt_ts.c "T%u:\n", object->type); object 2729 drivers/input/touchscreen/atmel_mxt_ts.c for (j = 0; j < mxt_obj_instances(object); j++) { object 2730 drivers/input/touchscreen/atmel_mxt_ts.c u16 size = mxt_obj_size(object); object 2731 drivers/input/touchscreen/atmel_mxt_ts.c u16 addr = object->start_address + j * size; object 2737 drivers/input/touchscreen/atmel_mxt_ts.c count = mxt_show_instance(buf, count, object, j, obuf); object 2915 drivers/input/touchscreen/atmel_mxt_ts.c static DEVICE_ATTR(object, S_IRUGO, mxt_object_show, NULL); object 385 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c field = qp_read_field(d->dev, d->object, desc->i, &is_str); object 389 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c field = eq_read_field(d->dev, d->object, desc->i); object 393 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c field = cq_read_field(d->dev, d->object, desc->i); object 428 drivers/net/ethernet/mellanox/mlx5/core/debugfs.c d->object = data; object 763 drivers/platform/x86/sony-laptop.c union acpi_object *object = NULL; object 766 drivers/platform/x86/sony-laptop.c object = __call_snc_method(handle, name, &v); object 768 drivers/platform/x86/sony-laptop.c object = __call_snc_method(handle, name, NULL); object 770 drivers/platform/x86/sony-laptop.c if (!object) object 773 drivers/platform/x86/sony-laptop.c if (object->type != ACPI_TYPE_INTEGER) { object 775 drivers/platform/x86/sony-laptop.c ACPI_TYPE_INTEGER, object->type); object 776 drivers/platform/x86/sony-laptop.c kfree(object); object 781 drivers/platform/x86/sony-laptop.c *result = object->integer.value; object 783 drivers/platform/x86/sony-laptop.c kfree(object); object 793 drivers/platform/x86/sony-laptop.c union acpi_object *object = __call_snc_method(handle, name, value); object 795 drivers/platform/x86/sony-laptop.c if (!object) object 798 drivers/platform/x86/sony-laptop.c if (object->type == ACPI_TYPE_BUFFER) { object 799 drivers/platform/x86/sony-laptop.c len = MIN(buflen, object->buffer.length); object 800 drivers/platform/x86/sony-laptop.c memcpy(buffer, object->buffer.pointer, len); object 802 drivers/platform/x86/sony-laptop.c } else if (object->type == ACPI_TYPE_INTEGER) { object 803 drivers/platform/x86/sony-laptop.c len = MIN(buflen, sizeof(object->integer.value)); object 804 drivers/platform/x86/sony-laptop.c memcpy(buffer, &object->integer.value, len); object 808 drivers/platform/x86/sony-laptop.c ACPI_TYPE_BUFFER, object->type); object 812 drivers/platform/x86/sony-laptop.c kfree(object); object 541 drivers/platform/x86/thinkpad_acpi.c #define TPACPI_HANDLE(object, parent, paths...) \ object 542 drivers/platform/x86/thinkpad_acpi.c static acpi_handle object##_handle; \ object 543 drivers/platform/x86/thinkpad_acpi.c static const acpi_handle * const object##_parent __initconst = \ object 545 drivers/platform/x86/thinkpad_acpi.c static char *object##_paths[] __initdata = { paths } object 690 drivers/platform/x86/thinkpad_acpi.c #define TPACPI_ACPIHANDLE_INIT(object) \ object 691 drivers/platform/x86/thinkpad_acpi.c drv_acpi_handle_init(#object, &object##_handle, *object##_parent, \ object 692 drivers/platform/x86/thinkpad_acpi.c object##_paths, ARRAY_SIZE(object##_paths)) object 37 drivers/xen/xen-acpi-cpuhotplug.c union acpi_object object = { 0 }; object 38 drivers/xen/xen-acpi-cpuhotplug.c struct acpi_buffer buffer = { sizeof(union acpi_object), &object }; object 49 drivers/xen/xen-acpi-cpuhotplug.c pr->acpi_id = object.processor.proc_id; object 331 drivers/xen/xen-acpi-processor.c union acpi_object object = { 0 }; object 332 drivers/xen/xen-acpi-processor.c struct acpi_buffer buffer = { sizeof(union acpi_object), &object }; object 344 drivers/xen/xen-acpi-processor.c acpi_id = object.processor.proc_id; object 345 drivers/xen/xen-acpi-processor.c pblk = object.processor.pblk_address; object 27 fs/cachefiles/interface.c struct cachefiles_object *object; object 43 fs/cachefiles/interface.c object = kmem_cache_alloc(cachefiles_object_jar, cachefiles_gfp); object 44 fs/cachefiles/interface.c if (!object) object 47 fs/cachefiles/interface.c ASSERTCMP(object->backer, ==, NULL); object 49 fs/cachefiles/interface.c BUG_ON(test_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)); object 50 fs/cachefiles/interface.c atomic_set(&object->usage, 1); object 52 fs/cachefiles/interface.c fscache_object_init(&object->fscache, cookie, &cache->cache); object 54 fs/cachefiles/interface.c object->type = cookie->def->type; object 77 fs/cachefiles/interface.c key = cachefiles_cook_key(buffer, keylen + 2, object->type); object 97 fs/cachefiles/interface.c object->lookup_data = lookup_data; object 99 fs/cachefiles/interface.c _leave(" = %p [%p]", &object->fscache, lookup_data); object 100 fs/cachefiles/interface.c return &object->fscache; object 105 fs/cachefiles/interface.c BUG_ON(test_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)); object 106 fs/cachefiles/interface.c kmem_cache_free(cachefiles_object_jar, object); object 122 fs/cachefiles/interface.c struct cachefiles_object *parent, *object; object 132 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 133 fs/cachefiles/interface.c lookup_data = object->lookup_data; object 139 fs/cachefiles/interface.c ret = cachefiles_walk_to_object(parent, object, object 146 fs/cachefiles/interface.c object->fscache.cookie->def->type != FSCACHE_COOKIE_TYPE_INDEX) object 147 fs/cachefiles/interface.c cachefiles_attr_changed(&object->fscache); object 152 fs/cachefiles/interface.c fscache_object_lookup_error(&object->fscache); object 164 fs/cachefiles/interface.c struct cachefiles_object *object; object 166 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 168 fs/cachefiles/interface.c _enter("{OBJ%x,%p}", object->fscache.debug_id, object->lookup_data); object 170 fs/cachefiles/interface.c if (object->lookup_data) { object 171 fs/cachefiles/interface.c kfree(object->lookup_data->key); object 172 fs/cachefiles/interface.c kfree(object->lookup_data->auxdata); object 173 fs/cachefiles/interface.c kfree(object->lookup_data); object 174 fs/cachefiles/interface.c object->lookup_data = NULL; object 185 fs/cachefiles/interface.c struct cachefiles_object *object = object 189 fs/cachefiles/interface.c _enter("{OBJ%x,%d}", _object->debug_id, atomic_read(&object->usage)); object 192 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); object 195 fs/cachefiles/interface.c u = atomic_inc_return(&object->usage); object 196 fs/cachefiles/interface.c trace_cachefiles_ref(object, _object->cookie, object 198 fs/cachefiles/interface.c return &object->fscache; object 206 fs/cachefiles/interface.c struct cachefiles_object *object; object 216 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 217 fs/cachefiles/interface.c cache = container_of(object->fscache.cache, struct cachefiles_cache, object 225 fs/cachefiles/interface.c cookie = object->fscache.cookie; object 251 fs/cachefiles/interface.c cachefiles_update_object_xattr(object, auxdata); object 263 fs/cachefiles/interface.c struct cachefiles_object *object; object 271 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 274 fs/cachefiles/interface.c object->fscache.debug_id, atomic_read(&object->usage)); object 276 fs/cachefiles/interface.c cache = container_of(object->fscache.cache, object 280 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); object 288 fs/cachefiles/interface.c if (object->dentry) { object 290 fs/cachefiles/interface.c if (test_bit(FSCACHE_OBJECT_RETIRED, &object->fscache.flags) && object 293 fs/cachefiles/interface.c _debug("- retire object OBJ%x", object->fscache.debug_id); object 294 fs/cachefiles/interface.c inode = d_backing_inode(object->dentry); object 299 fs/cachefiles/interface.c cachefiles_delete_object(cache, object); object 304 fs/cachefiles/interface.c if (object->backer != object->dentry) object 305 fs/cachefiles/interface.c dput(object->backer); object 306 fs/cachefiles/interface.c object->backer = NULL; object 310 fs/cachefiles/interface.c if (test_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)) object 311 fs/cachefiles/interface.c cachefiles_mark_object_inactive(cache, object, i_blocks); object 313 fs/cachefiles/interface.c dput(object->dentry); object 314 fs/cachefiles/interface.c object->dentry = NULL; object 325 fs/cachefiles/interface.c struct cachefiles_object *object; object 331 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 334 fs/cachefiles/interface.c object->fscache.debug_id, atomic_read(&object->usage)); object 337 fs/cachefiles/interface.c ASSERT((atomic_read(&object->usage) & 0xffff0000) != 0x6b6b0000); object 340 fs/cachefiles/interface.c ASSERTIFCMP(object->fscache.parent, object 341 fs/cachefiles/interface.c object->fscache.parent->n_children, >, 0); object 343 fs/cachefiles/interface.c u = atomic_dec_return(&object->usage); object 344 fs/cachefiles/interface.c trace_cachefiles_ref(object, _object->cookie, object 348 fs/cachefiles/interface.c _debug("- kill object OBJ%x", object->fscache.debug_id); object 350 fs/cachefiles/interface.c ASSERT(!test_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)); object 351 fs/cachefiles/interface.c ASSERTCMP(object->fscache.parent, ==, NULL); object 352 fs/cachefiles/interface.c ASSERTCMP(object->backer, ==, NULL); object 353 fs/cachefiles/interface.c ASSERTCMP(object->dentry, ==, NULL); object 354 fs/cachefiles/interface.c ASSERTCMP(object->fscache.n_ops, ==, 0); object 355 fs/cachefiles/interface.c ASSERTCMP(object->fscache.n_children, ==, 0); object 357 fs/cachefiles/interface.c if (object->lookup_data) { object 358 fs/cachefiles/interface.c kfree(object->lookup_data->key); object 359 fs/cachefiles/interface.c kfree(object->lookup_data->auxdata); object 360 fs/cachefiles/interface.c kfree(object->lookup_data); object 361 fs/cachefiles/interface.c object->lookup_data = NULL; object 364 fs/cachefiles/interface.c cache = object->fscache.cache; object 365 fs/cachefiles/interface.c fscache_object_destroy(&object->fscache); object 366 fs/cachefiles/interface.c kmem_cache_free(cachefiles_object_jar, object); object 407 fs/cachefiles/interface.c struct cachefiles_object *object; object 412 fs/cachefiles/interface.c _enter("{OBJ%x}", op->object->debug_id); object 414 fs/cachefiles/interface.c object = container_of(op->object, struct cachefiles_object, fscache); object 415 fs/cachefiles/interface.c cache = container_of(object->fscache.cache, object 419 fs/cachefiles/interface.c ret = cachefiles_check_auxdata(object); object 432 fs/cachefiles/interface.c struct cachefiles_object *object; object 445 fs/cachefiles/interface.c object = container_of(_object, struct cachefiles_object, fscache); object 446 fs/cachefiles/interface.c cache = container_of(object->fscache.cache, object 449 fs/cachefiles/interface.c if (ni_size == object->i_size) object 452 fs/cachefiles/interface.c if (!object->backer) object 455 fs/cachefiles/interface.c ASSERT(d_is_reg(object->backer)); object 457 fs/cachefiles/interface.c fscache_set_store_limit(&object->fscache, ni_size); object 459 fs/cachefiles/interface.c oi_size = i_size_read(d_backing_inode(object->backer)); object 464 fs/cachefiles/interface.c inode_lock(d_inode(object->backer)); object 473 fs/cachefiles/interface.c ret = notify_change(object->backer, &newattrs, NULL); object 480 fs/cachefiles/interface.c ret = notify_change(object->backer, &newattrs, NULL); object 483 fs/cachefiles/interface.c inode_unlock(d_inode(object->backer)); object 487 fs/cachefiles/interface.c fscache_set_store_limit(&object->fscache, 0); object 488 fs/cachefiles/interface.c cachefiles_io_error_obj(object, "Size set failed"); object 501 fs/cachefiles/interface.c struct cachefiles_object *object; object 508 fs/cachefiles/interface.c object = container_of(op->object, struct cachefiles_object, fscache); object 509 fs/cachefiles/interface.c cache = container_of(object->fscache.cache, object 512 fs/cachefiles/interface.c ni_size = op->object->store_limit_l; object 515 fs/cachefiles/interface.c op->object->debug_id, (unsigned long long)ni_size); object 517 fs/cachefiles/interface.c if (object->backer) { object 518 fs/cachefiles/interface.c ASSERT(d_is_reg(object->backer)); object 520 fs/cachefiles/interface.c fscache_set_store_limit(&object->fscache, ni_size); object 522 fs/cachefiles/interface.c path.dentry = object->backer; object 532 fs/cachefiles/interface.c fscache_set_store_limit(&object->fscache, 0); object 534 fs/cachefiles/interface.c cachefiles_io_error_obj(object, object 108 fs/cachefiles/internal.h struct cachefiles_object *object; object 162 fs/cachefiles/internal.h struct cachefiles_object *object, object 165 fs/cachefiles/internal.h struct cachefiles_object *object); object 167 fs/cachefiles/internal.h struct cachefiles_object *object, object 243 fs/cachefiles/internal.h extern int cachefiles_check_object_type(struct cachefiles_object *object); object 244 fs/cachefiles/internal.h extern int cachefiles_set_object_xattr(struct cachefiles_object *object, object 246 fs/cachefiles/internal.h extern int cachefiles_update_object_xattr(struct cachefiles_object *object, object 248 fs/cachefiles/internal.h extern int cachefiles_check_auxdata(struct cachefiles_object *object); object 249 fs/cachefiles/internal.h extern int cachefiles_check_object_xattr(struct cachefiles_object *object, object 266 fs/cachefiles/internal.h #define cachefiles_io_error_obj(object, FMT, ...) \ object 270 fs/cachefiles/internal.h ___cache = container_of((object)->fscache.cache, \ object 42 fs/cachefiles/main.c struct cachefiles_object *object = _object; object 44 fs/cachefiles/main.c memset(object, 0, sizeof(*object)); object 45 fs/cachefiles/main.c spin_lock_init(&object->work_lock); object 27 fs/cachefiles/namei.c void __cachefiles_printk_object(struct cachefiles_object *object, object 34 fs/cachefiles/namei.c pr_err("%sobject: OBJ%x\n", prefix, object->fscache.debug_id); object 36 fs/cachefiles/namei.c prefix, object->fscache.state->name, object 37 fs/cachefiles/namei.c object->fscache.flags, work_busy(&object->fscache.work), object 38 fs/cachefiles/namei.c object->fscache.events, object->fscache.event_mask); object 40 fs/cachefiles/namei.c prefix, object->fscache.n_ops, object->fscache.n_in_progress, object 41 fs/cachefiles/namei.c object->fscache.n_exclusive); object 43 fs/cachefiles/namei.c prefix, object->fscache.parent); object 45 fs/cachefiles/namei.c spin_lock(&object->fscache.lock); object 46 fs/cachefiles/namei.c cookie = object->fscache.cookie; object 50 fs/cachefiles/namei.c object->fscache.cookie, object 51 fs/cachefiles/namei.c object->fscache.cookie->parent, object 52 fs/cachefiles/namei.c object->fscache.cookie->netfs_data, object 53 fs/cachefiles/namei.c object->fscache.cookie->flags); object 63 fs/cachefiles/namei.c spin_unlock(&object->fscache.lock); object 69 fs/cachefiles/namei.c static noinline void cachefiles_printk_object(struct cachefiles_object *object, object 72 fs/cachefiles/namei.c if (object) object 73 fs/cachefiles/namei.c __cachefiles_printk_object(object, ""); object 88 fs/cachefiles/namei.c struct cachefiles_object *object; object 97 fs/cachefiles/namei.c object = rb_entry(p, struct cachefiles_object, active_node); object 98 fs/cachefiles/namei.c if (object->dentry > dentry) object 100 fs/cachefiles/namei.c else if (object->dentry < dentry) object 114 fs/cachefiles/namei.c object->fscache.debug_id, object 115 fs/cachefiles/namei.c object->fscache.state->name, object 118 fs/cachefiles/namei.c trace_cachefiles_mark_buried(object, dentry, why); object 120 fs/cachefiles/namei.c if (fscache_object_is_live(&object->fscache)) { object 123 fs/cachefiles/namei.c cachefiles_printk_object(object, NULL); object 126 fs/cachefiles/namei.c fscache_object_mark_killed(&object->fscache, why); object 137 fs/cachefiles/namei.c struct cachefiles_object *object) object 143 fs/cachefiles/namei.c _enter(",%p", object); object 148 fs/cachefiles/namei.c dentry = object->dentry; object 149 fs/cachefiles/namei.c trace_cachefiles_mark_active(object, dentry); object 151 fs/cachefiles/namei.c if (test_and_set_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags)) { object 153 fs/cachefiles/namei.c cachefiles_printk_object(object, NULL); object 163 fs/cachefiles/namei.c ASSERT(xobject != object); object 173 fs/cachefiles/namei.c rb_link_node(&object->active_node, _parent, _p); object 174 fs/cachefiles/namei.c rb_insert_color(&object->active_node, &cache->active_nodes); object 183 fs/cachefiles/namei.c trace_cachefiles_wait_active(object, dentry, xobject); object 184 fs/cachefiles/namei.c clear_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags); object 189 fs/cachefiles/namei.c cachefiles_printk_object(object, xobject); object 205 fs/cachefiles/namei.c object->fscache.debug_id, object 227 fs/cachefiles/namei.c object->fscache.debug_id, object 235 fs/cachefiles/namei.c cachefiles_printk_object(object, xobject); object 257 fs/cachefiles/namei.c struct cachefiles_object *object, object 260 fs/cachefiles/namei.c struct dentry *dentry = object->dentry; object 263 fs/cachefiles/namei.c trace_cachefiles_mark_inactive(object, dentry, inode); object 266 fs/cachefiles/namei.c rb_erase(&object->active_node, &cache->active_nodes); object 267 fs/cachefiles/namei.c clear_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags); object 270 fs/cachefiles/namei.c wake_up_bit(&object->flags, CACHEFILES_OBJECT_ACTIVE); object 288 fs/cachefiles/namei.c struct cachefiles_object *object, object 313 fs/cachefiles/namei.c trace_cachefiles_unlink(object, rep, why); object 415 fs/cachefiles/namei.c trace_cachefiles_rename(object, rep, grave, why); object 436 fs/cachefiles/namei.c struct cachefiles_object *object) object 441 fs/cachefiles/namei.c _enter(",OBJ%x{%p}", object->fscache.debug_id, object->dentry); object 443 fs/cachefiles/namei.c ASSERT(object->dentry); object 444 fs/cachefiles/namei.c ASSERT(d_backing_inode(object->dentry)); object 445 fs/cachefiles/namei.c ASSERT(object->dentry->d_parent); object 447 fs/cachefiles/namei.c dir = dget_parent(object->dentry); object 451 fs/cachefiles/namei.c if (test_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->fscache.flags)) { object 460 fs/cachefiles/namei.c if (dir == object->dentry->d_parent) { object 461 fs/cachefiles/namei.c ret = cachefiles_bury_object(cache, object, dir, object 462 fs/cachefiles/namei.c object->dentry, false, object 483 fs/cachefiles/namei.c struct cachefiles_object *object, object 497 fs/cachefiles/namei.c object->fscache.debug_id, key); object 534 fs/cachefiles/namei.c trace_cachefiles_lookup(object, next, NULL); object 539 fs/cachefiles/namei.c trace_cachefiles_lookup(object, next, inode); object 543 fs/cachefiles/namei.c object->new = !inode; object 549 fs/cachefiles/namei.c fscache_object_lookup_negative(&object->fscache); object 552 fs/cachefiles/namei.c if (key || object->type == FSCACHE_COOKIE_TYPE_INDEX) { object 567 fs/cachefiles/namei.c trace_cachefiles_mkdir(object, next, ret); object 602 fs/cachefiles/namei.c trace_cachefiles_create(object, next, ret); object 632 fs/cachefiles/namei.c object->dentry = next; object 636 fs/cachefiles/namei.c if (!object->new) { object 639 fs/cachefiles/namei.c ret = cachefiles_check_object_xattr(object, auxdata); object 643 fs/cachefiles/namei.c object->dentry = NULL; object 645 fs/cachefiles/namei.c ret = cachefiles_bury_object(cache, object, dir, next, object 655 fs/cachefiles/namei.c fscache_object_retrying_stale(&object->fscache); object 661 fs/cachefiles/namei.c ret = cachefiles_mark_object_active(cache, object); object 672 fs/cachefiles/namei.c if (object->new) { object 674 fs/cachefiles/namei.c ret = cachefiles_set_object_xattr(object, auxdata); object 687 fs/cachefiles/namei.c if (object->type != FSCACHE_COOKIE_TYPE_INDEX) { object 688 fs/cachefiles/namei.c if (d_is_reg(object->dentry)) { object 692 fs/cachefiles/namei.c aops = d_backing_inode(object->dentry)->i_mapping->a_ops; object 695 fs/cachefiles/namei.c if (object->dentry->d_sb->s_blocksize > PAGE_SIZE) object 698 fs/cachefiles/namei.c object->backer = object->dentry; object 704 fs/cachefiles/namei.c object->new = 0; object 705 fs/cachefiles/namei.c fscache_obtained_object(&object->fscache); object 707 fs/cachefiles/namei.c _leave(" = 0 [%lu]", d_backing_inode(object->dentry)->i_ino); object 711 fs/cachefiles/namei.c fscache_object_mark_killed(&object->fscache, FSCACHE_OBJECT_NO_SPACE); object 725 fs/cachefiles/namei.c cache, object, d_backing_inode(object->dentry)->i_blocks); object 727 fs/cachefiles/namei.c dput(object->dentry); object 728 fs/cachefiles/namei.c object->dentry = NULL; object 867 fs/cachefiles/namei.c struct cachefiles_object *object; object 904 fs/cachefiles/namei.c object = rb_entry(_n, struct cachefiles_object, active_node); object 906 fs/cachefiles/namei.c if (object->dentry > victim) object 908 fs/cachefiles/namei.c else if (object->dentry < victim) object 25 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 60 fs/cachefiles/rdwr.c object = container_of(op->op.object, struct cachefiles_object, fscache); object 61 fs/cachefiles/rdwr.c spin_lock(&object->work_lock); object 64 fs/cachefiles/rdwr.c spin_unlock(&object->work_lock); object 77 fs/cachefiles/rdwr.c static int cachefiles_read_reissue(struct cachefiles_object *object, object 80 fs/cachefiles/rdwr.c struct address_space *bmapping = d_backing_inode(object->backer)->i_mapping; object 85 fs/cachefiles/rdwr.c d_backing_inode(object->backer)->i_ino, object 141 fs/cachefiles/rdwr.c spin_lock_irq(&object->work_lock); object 143 fs/cachefiles/rdwr.c spin_unlock_irq(&object->work_lock); object 155 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 160 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 163 fs/cachefiles/rdwr.c _enter("{ino=%lu}", d_backing_inode(object->backer)->i_ino); object 166 fs/cachefiles/rdwr.c spin_lock_irq(&object->work_lock); object 173 fs/cachefiles/rdwr.c spin_unlock_irq(&object->work_lock); object 179 fs/cachefiles/rdwr.c &object->fscache.cookie->flags)) { object 188 fs/cachefiles/rdwr.c error = cachefiles_read_reissue(object, monitor); object 194 fs/cachefiles/rdwr.c object, object 218 fs/cachefiles/rdwr.c spin_lock_irq(&object->work_lock); object 221 fs/cachefiles/rdwr.c spin_unlock_irq(&object->work_lock); object 229 fs/cachefiles/rdwr.c static int cachefiles_read_backing_file_one(struct cachefiles_object *object, object 253 fs/cachefiles/rdwr.c bmapping = d_backing_inode(object->backer)->i_mapping; object 362 fs/cachefiles/rdwr.c cachefiles_io_error_obj(object, "Page read error on backing file"); object 396 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 403 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 405 fs/cachefiles/rdwr.c cache = container_of(object->fscache.cache, object 408 fs/cachefiles/rdwr.c _enter("{%p},{%lx},,,", object, page->index); object 410 fs/cachefiles/rdwr.c if (!object->backer) object 413 fs/cachefiles/rdwr.c inode = d_backing_inode(object->backer); object 442 fs/cachefiles/rdwr.c ret = cachefiles_read_backing_file_one(object, op, page); object 465 fs/cachefiles/rdwr.c static int cachefiles_read_backing_file(struct cachefiles_object *object, object 470 fs/cachefiles/rdwr.c struct address_space *bmapping = d_backing_inode(object->backer)->i_mapping; object 671 fs/cachefiles/rdwr.c cachefiles_io_error_obj(object, "Page read error on backing file"); object 687 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 696 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 698 fs/cachefiles/rdwr.c cache = container_of(object->fscache.cache, object 702 fs/cachefiles/rdwr.c object->fscache.debug_id, atomic_read(&op->op.usage), object 705 fs/cachefiles/rdwr.c if (!object->backer) object 712 fs/cachefiles/rdwr.c inode = d_backing_inode(object->backer); object 772 fs/cachefiles/rdwr.c ret2 = cachefiles_read_backing_file(object, op, &backpages); object 801 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 805 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 807 fs/cachefiles/rdwr.c cache = container_of(object->fscache.cache, object 810 fs/cachefiles/rdwr.c _enter("%p,{%lx},", object, page->index); object 839 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 845 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 847 fs/cachefiles/rdwr.c cache = container_of(object->fscache.cache, object 850 fs/cachefiles/rdwr.c _enter("%p,,,%d,", object, *nr_pages); object 882 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 894 fs/cachefiles/rdwr.c object = container_of(op->op.object, object 897 fs/cachefiles/rdwr.c _enter("%p,%p{%lx},,,", object, page, page->index); object 899 fs/cachefiles/rdwr.c if (!object->backer) { object 904 fs/cachefiles/rdwr.c ASSERT(d_is_reg(object->backer)); object 906 fs/cachefiles/rdwr.c cache = container_of(object->fscache.cache, object 914 fs/cachefiles/rdwr.c eof = object->fscache.store_limit_l; object 921 fs/cachefiles/rdwr.c path.dentry = object->backer; object 952 fs/cachefiles/rdwr.c cachefiles_io_error_obj(object, object 964 fs/cachefiles/rdwr.c __releases(&object->fscache.cookie->lock) object 966 fs/cachefiles/rdwr.c struct cachefiles_object *object; object 968 fs/cachefiles/rdwr.c object = container_of(_object, struct cachefiles_object, fscache); object 970 fs/cachefiles/rdwr.c _enter("%p,{%lu}", object, page->index); object 972 fs/cachefiles/rdwr.c spin_unlock(&object->fscache.cookie->lock); object 25 fs/cachefiles/xattr.c int cachefiles_check_object_type(struct cachefiles_object *object) object 27 fs/cachefiles/xattr.c struct dentry *dentry = object->dentry; object 34 fs/cachefiles/xattr.c if (!object->fscache.cookie) object 37 fs/cachefiles/xattr.c snprintf(type, 3, "%02x", object->fscache.cookie->def->type); object 39 fs/cachefiles/xattr.c _enter("%p{%s}", object, type); object 99 fs/cachefiles/xattr.c int cachefiles_set_object_xattr(struct cachefiles_object *object, object 102 fs/cachefiles/xattr.c struct dentry *dentry = object->dentry; object 107 fs/cachefiles/xattr.c _enter("%p,#%d", object, auxdata->len); object 112 fs/cachefiles/xattr.c clear_bit(FSCACHE_COOKIE_AUX_UPDATED, &object->fscache.cookie->flags); object 118 fs/cachefiles/xattr.c object, object 128 fs/cachefiles/xattr.c int cachefiles_update_object_xattr(struct cachefiles_object *object, object 131 fs/cachefiles/xattr.c struct dentry *dentry = object->dentry; object 137 fs/cachefiles/xattr.c _enter("%p,#%d", object, auxdata->len); object 142 fs/cachefiles/xattr.c clear_bit(FSCACHE_COOKIE_AUX_UPDATED, &object->fscache.cookie->flags); object 148 fs/cachefiles/xattr.c object, object 158 fs/cachefiles/xattr.c int cachefiles_check_auxdata(struct cachefiles_object *object) object 162 fs/cachefiles/xattr.c struct dentry *dentry = object->dentry; object 168 fs/cachefiles/xattr.c ASSERT(object->fscache.cookie->def->check_aux); object 178 fs/cachefiles/xattr.c auxbuf->type != object->fscache.cookie->def->type) object 182 fs/cachefiles/xattr.c validity = fscache_check_aux(&object->fscache, &auxbuf->data, xlen, object 197 fs/cachefiles/xattr.c int cachefiles_check_object_xattr(struct cachefiles_object *object, object 201 fs/cachefiles/xattr.c struct dentry *dentry = object->dentry; object 204 fs/cachefiles/xattr.c _enter("%p,#%d", object, auxdata->len); object 226 fs/cachefiles/xattr.c cachefiles_io_error_obj(object, object 242 fs/cachefiles/xattr.c if (object->fscache.cookie->def->check_aux) { object 249 fs/cachefiles/xattr.c object->fscache.cookie->def->name, dlen); object 251 fs/cachefiles/xattr.c result = fscache_check_aux(&object->fscache, object 277 fs/cachefiles/xattr.c cachefiles_io_error_obj(object, object 94 fs/fscache/cache.c struct fscache_object *object; object 110 fs/fscache/cache.c object = hlist_entry(cookie->backing_objects.first, object 113 fs/fscache/cache.c cache = object->cache; object 114 fs/fscache/cache.c if (fscache_object_is_dying(object) || object 328 fs/fscache/cache.c struct fscache_object *object; object 334 fs/fscache/cache.c object = list_entry(cache->object_list.next, object 336 fs/fscache/cache.c list_move_tail(&object->cache_link, dying_objects); object 338 fs/fscache/cache.c _debug("withdraw %p", object->cookie); object 343 fs/fscache/cache.c fscache_raise_event(object, FSCACHE_OBJECT_EV_KILL); object 28 fs/fscache/cookie.c struct fscache_object *object); object 32 fs/fscache/cookie.c struct hlist_node *object; object 43 fs/fscache/cookie.c object = READ_ONCE(cookie->backing_objects.first); object 44 fs/fscache/cookie.c if (object) object 46 fs/fscache/cookie.c prefix, hlist_entry(object, struct fscache_object, cookie_link)); object 396 fs/fscache/cookie.c struct fscache_object *object; object 442 fs/fscache/cookie.c object = hlist_entry(cookie->backing_objects.first, object 445 fs/fscache/cookie.c fscache_set_store_limit(object, object_size); object 449 fs/fscache/cookie.c fscache_raise_event(object, FSCACHE_OBJECT_EV_NEW_CHILD); object 480 fs/fscache/cookie.c struct fscache_object *object; object 486 fs/fscache/cookie.c hlist_for_each_entry(object, &cookie->backing_objects, object 488 fs/fscache/cookie.c if (object->cache == cache) object 496 fs/fscache/cookie.c object = cache->ops->alloc_object(cache, cookie); object 498 fs/fscache/cookie.c if (IS_ERR(object)) { object 500 fs/fscache/cookie.c ret = PTR_ERR(object); object 504 fs/fscache/cookie.c ASSERTCMP(object->cookie, ==, cookie); object 507 fs/fscache/cookie.c object->debug_id = atomic_inc_return(&fscache_object_debug_id); object 510 fs/fscache/cookie.c object->debug_id, cookie->def->name, object->events); object 519 fs/fscache/cookie.c if (fscache_attach_object(cookie, object) < 0) { object 521 fs/fscache/cookie.c cache->ops->put_object(object, fscache_obj_put_attach_fail); object 530 fs/fscache/cookie.c if (fscache_object_is_dying(object) || object 531 fs/fscache/cookie.c fscache_cache_is_broken(object)) { object 541 fs/fscache/cookie.c cache->ops->put_object(object, fscache_obj_put_alloc_fail); object 552 fs/fscache/cookie.c struct fscache_object *object) object 555 fs/fscache/cookie.c struct fscache_cache *cache = object->cache; object 558 fs/fscache/cookie.c _enter("{%s},{OBJ%x}", cookie->def->name, object->debug_id); object 560 fs/fscache/cookie.c ASSERTCMP(object->cookie, ==, cookie); object 568 fs/fscache/cookie.c if (p->cache == object->cache) { object 579 fs/fscache/cookie.c if (p->cache == object->cache) { object 585 fs/fscache/cookie.c object->parent = p; object 595 fs/fscache/cookie.c if (list_empty(&object->cache_link)) { object 597 fs/fscache/cookie.c list_add(&object->cache_link, &cache->object_list); object 602 fs/fscache/cookie.c hlist_add_head(&object->cookie_link, &cookie->backing_objects); object 604 fs/fscache/cookie.c fscache_objlist_add(object); object 618 fs/fscache/cookie.c struct fscache_object *object; object 641 fs/fscache/cookie.c object = hlist_entry(cookie->backing_objects.first, object 644 fs/fscache/cookie.c if (fscache_object_is_live(object)) object 646 fs/fscache/cookie.c object, FSCACHE_OBJECT_EV_INVALIDATE); object 675 fs/fscache/cookie.c struct fscache_object *object; object 695 fs/fscache/cookie.c hlist_for_each_entry(object, object 697 fs/fscache/cookie.c fscache_raise_event(object, FSCACHE_OBJECT_EV_UPDATE); object 713 fs/fscache/cookie.c struct fscache_object *object; object 746 fs/fscache/cookie.c hlist_for_each_entry(object, &cookie->backing_objects, cookie_link) { object 748 fs/fscache/cookie.c set_bit(FSCACHE_OBJECT_RETIRED, &object->flags); object 749 fs/fscache/cookie.c clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); object 750 fs/fscache/cookie.c fscache_raise_event(object, FSCACHE_OBJECT_EV_KILL); object 892 fs/fscache/cookie.c struct fscache_object *object; object 923 fs/fscache/cookie.c object = hlist_entry(cookie->backing_objects.first, object 925 fs/fscache/cookie.c if (test_bit(FSCACHE_IOERROR, &object->cache->flags)) object 931 fs/fscache/cookie.c if (fscache_submit_op(object, op) < 0) object 937 fs/fscache/cookie.c ret = fscache_wait_for_operation_activation(object, op, NULL, NULL); object 940 fs/fscache/cookie.c ret = object->cache->ops->check_consistency(op); object 119 fs/fscache/internal.h #define fscache_objlist_add(object) do {} while(0) object 120 fs/fscache/internal.h #define fscache_objlist_remove(object) do {} while(0) object 306 fs/fscache/internal.h static inline void fscache_raise_event(struct fscache_object *object, object 312 fs/fscache/internal.h object->debug_id, object->event_mask, (1 << event)); object 314 fs/fscache/internal.h if (!test_and_set_bit(event, &object->events) && object 315 fs/fscache/internal.h test_bit(event, &object->event_mask)) object 316 fs/fscache/internal.h fscache_enqueue_object(object); object 149 fs/fscache/object.c static inline void fscache_done_parent_op(struct fscache_object *object) object 151 fs/fscache/object.c struct fscache_object *parent = object->parent; object 154 fs/fscache/object.c object->debug_id, parent->debug_id, parent->n_ops); object 167 fs/fscache/object.c static void fscache_object_sm_dispatcher(struct fscache_object *object) object 175 fs/fscache/object.c ASSERT(object != NULL); object 178 fs/fscache/object.c object->debug_id, object->state->name, object->events); object 180 fs/fscache/object.c event_mask = object->event_mask; object 182 fs/fscache/object.c object->event_mask = 0; /* Mask normal event handling */ object 183 fs/fscache/object.c state = object->state; object 185 fs/fscache/object.c events = object->events; object 188 fs/fscache/object.c if (events & object->oob_event_mask) { object 190 fs/fscache/object.c object->debug_id, events & object->oob_event_mask); object 192 fs/fscache/object.c for (t = object->oob_table; t->events; t++) { object 197 fs/fscache/object.c __clear_bit(event, &object->oob_event_mask); object 198 fs/fscache/object.c clear_bit(event, &object->events); object 212 fs/fscache/object.c trace_fscache_osm(object, state, object 214 fs/fscache/object.c clear_bit(event, &object->events); object 216 fs/fscache/object.c object->debug_id, event, object 218 fs/fscache/object.c object->state = state = new_state; object 231 fs/fscache/object.c _debug("{OBJ%x} exec %s", object->debug_id, state->name); object 233 fs/fscache/object.c trace_fscache_osm(object, state, false, oob, event); object 234 fs/fscache/object.c new_state = state->work(object, event); object 237 fs/fscache/object.c _debug("{OBJ%x} %s notrans", object->debug_id, state->name); object 242 fs/fscache/object.c fscache_enqueue_object(object); object 243 fs/fscache/object.c event_mask = object->oob_event_mask; object 248 fs/fscache/object.c object->debug_id, state->name, new_state->name); object 249 fs/fscache/object.c object->state = state = new_state; object 260 fs/fscache/object.c event_mask = object->oob_event_mask; object 265 fs/fscache/object.c object->event_mask = event_mask; object 267 fs/fscache/object.c events = object->events; object 278 fs/fscache/object.c struct fscache_object *object = object 282 fs/fscache/object.c _enter("{OBJ%x}", object->debug_id); object 285 fs/fscache/object.c fscache_object_sm_dispatcher(object); object 287 fs/fscache/object.c fscache_put_object(object, fscache_obj_put_work); object 301 fs/fscache/object.c void fscache_object_init(struct fscache_object *object, object 309 fs/fscache/object.c object->state = STATE(WAIT_FOR_INIT); object 310 fs/fscache/object.c object->oob_table = fscache_osm_init_oob; object 311 fs/fscache/object.c object->flags = 1 << FSCACHE_OBJECT_IS_LIVE; object 312 fs/fscache/object.c spin_lock_init(&object->lock); object 313 fs/fscache/object.c INIT_LIST_HEAD(&object->cache_link); object 314 fs/fscache/object.c INIT_HLIST_NODE(&object->cookie_link); object 315 fs/fscache/object.c INIT_WORK(&object->work, fscache_object_work_func); object 316 fs/fscache/object.c INIT_LIST_HEAD(&object->dependents); object 317 fs/fscache/object.c INIT_LIST_HEAD(&object->dep_link); object 318 fs/fscache/object.c INIT_LIST_HEAD(&object->pending_ops); object 319 fs/fscache/object.c object->n_children = 0; object 320 fs/fscache/object.c object->n_ops = object->n_in_progress = object->n_exclusive = 0; object 321 fs/fscache/object.c object->events = 0; object 322 fs/fscache/object.c object->store_limit = 0; object 323 fs/fscache/object.c object->store_limit_l = 0; object 324 fs/fscache/object.c object->cache = cache; object 325 fs/fscache/object.c object->cookie = cookie; object 327 fs/fscache/object.c object->parent = NULL; object 329 fs/fscache/object.c RB_CLEAR_NODE(&object->objlist_link); object 332 fs/fscache/object.c object->oob_event_mask = 0; object 333 fs/fscache/object.c for (t = object->oob_table; t->events; t++) object 334 fs/fscache/object.c object->oob_event_mask |= t->events; object 335 fs/fscache/object.c object->event_mask = object->oob_event_mask; object 336 fs/fscache/object.c for (t = object->state->transitions; t->events; t++) object 337 fs/fscache/object.c object->event_mask |= t->events; object 345 fs/fscache/object.c static inline void fscache_mark_object_dead(struct fscache_object *object) object 347 fs/fscache/object.c spin_lock(&object->lock); object 348 fs/fscache/object.c clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags); object 349 fs/fscache/object.c spin_unlock(&object->lock); object 355 fs/fscache/object.c static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object, object 358 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 360 fs/fscache/object.c object->oob_event_mask = 0; object 361 fs/fscache/object.c fscache_dequeue_object(object); object 372 fs/fscache/object.c static const struct fscache_state *fscache_initialise_object(struct fscache_object *object, object 378 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 380 fs/fscache/object.c ASSERT(list_empty(&object->dep_link)); object 382 fs/fscache/object.c parent = object->parent; object 406 fs/fscache/object.c object->cache->ops->grab_object(object, fscache_obj_get_add_to_deps)) { object 407 fs/fscache/object.c list_add(&object->dep_link, &parent->dependents); object 427 fs/fscache/object.c static const struct fscache_state *fscache_parent_ready(struct fscache_object *object, object 430 fs/fscache/object.c struct fscache_object *parent = object->parent; object 432 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 439 fs/fscache/object.c object->lookup_jif = jiffies; object 451 fs/fscache/object.c static const struct fscache_state *fscache_look_up_object(struct fscache_object *object, object 454 fs/fscache/object.c struct fscache_cookie *cookie = object->cookie; object 455 fs/fscache/object.c struct fscache_object *parent = object->parent; object 458 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 460 fs/fscache/object.c object->oob_table = fscache_osm_lookup_oob; object 470 fs/fscache/object.c test_bit(FSCACHE_IOERROR, &object->cache->flags) || object 471 fs/fscache/object.c !fscache_use_cookie(object)) { object 477 fs/fscache/object.c cookie->def->name, object->cache->tag->name); object 481 fs/fscache/object.c ret = object->cache->ops->lookup_object(object); object 484 fs/fscache/object.c fscache_unuse_cookie(object); object 510 fs/fscache/object.c void fscache_object_lookup_negative(struct fscache_object *object) object 512 fs/fscache/object.c struct fscache_cookie *cookie = object->cookie; object 514 fs/fscache/object.c _enter("{OBJ%x,%s}", object->debug_id, object->state->name); object 516 fs/fscache/object.c if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) { object 543 fs/fscache/object.c void fscache_obtained_object(struct fscache_object *object) object 545 fs/fscache/object.c struct fscache_cookie *cookie = object->cookie; object 547 fs/fscache/object.c _enter("{OBJ%x,%s}", object->debug_id, object->state->name); object 551 fs/fscache/object.c if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) { object 567 fs/fscache/object.c set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags); object 575 fs/fscache/object.c static const struct fscache_state *fscache_object_available(struct fscache_object *object, object 578 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 580 fs/fscache/object.c object->oob_table = fscache_osm_run_oob; object 582 fs/fscache/object.c spin_lock(&object->lock); object 584 fs/fscache/object.c fscache_done_parent_op(object); object 585 fs/fscache/object.c if (object->n_in_progress == 0) { object 586 fs/fscache/object.c if (object->n_ops > 0) { object 587 fs/fscache/object.c ASSERTCMP(object->n_ops, >=, object->n_obj_ops); object 588 fs/fscache/object.c fscache_start_operations(object); object 590 fs/fscache/object.c ASSERT(list_empty(&object->pending_ops)); object 593 fs/fscache/object.c spin_unlock(&object->lock); object 596 fs/fscache/object.c object->cache->ops->lookup_complete(object); object 599 fs/fscache/object.c fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif); object 609 fs/fscache/object.c static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object, object 612 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 614 fs/fscache/object.c if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY)) object 622 fs/fscache/object.c static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object, object 627 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 629 fs/fscache/object.c object->oob_event_mask = 0; object 632 fs/fscache/object.c object->cache->ops->lookup_complete(object); object 635 fs/fscache/object.c set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags); object 637 fs/fscache/object.c cookie = object->cookie; object 642 fs/fscache/object.c fscache_done_parent_op(object); object 650 fs/fscache/object.c static const struct fscache_state *fscache_kill_object(struct fscache_object *object, object 654 fs/fscache/object.c object->debug_id, object->n_ops, object->n_children, event); object 656 fs/fscache/object.c fscache_mark_object_dead(object); object 657 fs/fscache/object.c object->oob_event_mask = 0; object 659 fs/fscache/object.c if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) { object 661 fs/fscache/object.c clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); object 662 fs/fscache/object.c fscache_cancel_all_ops(object); object 665 fs/fscache/object.c if (list_empty(&object->dependents) && object 666 fs/fscache/object.c object->n_ops == 0 && object 667 fs/fscache/object.c object->n_children == 0) object 670 fs/fscache/object.c if (object->n_in_progress == 0) { object 671 fs/fscache/object.c spin_lock(&object->lock); object 672 fs/fscache/object.c if (object->n_ops > 0 && object->n_in_progress == 0) object 673 fs/fscache/object.c fscache_start_operations(object); object 674 fs/fscache/object.c spin_unlock(&object->lock); object 677 fs/fscache/object.c if (!list_empty(&object->dependents)) object 686 fs/fscache/object.c static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object, object 689 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 691 fs/fscache/object.c if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL)) object 699 fs/fscache/object.c static const struct fscache_state *fscache_drop_object(struct fscache_object *object, object 702 fs/fscache/object.c struct fscache_object *parent = object->parent; object 703 fs/fscache/object.c struct fscache_cookie *cookie = object->cookie; object 704 fs/fscache/object.c struct fscache_cache *cache = object->cache; object 707 fs/fscache/object.c _enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event); object 710 fs/fscache/object.c ASSERT(!hlist_unhashed(&object->cookie_link)); object 714 fs/fscache/object.c fscache_update_aux_data(object); object 721 fs/fscache/object.c hlist_del_init(&object->cookie_link); object 736 fs/fscache/object.c spin_lock(&object->lock); object 737 fs/fscache/object.c spin_unlock(&object->lock); object 741 fs/fscache/object.c list_del_init(&object->cache_link); object 745 fs/fscache/object.c cache->ops->drop_object(object); object 758 fs/fscache/object.c object->parent = NULL; object 762 fs/fscache/object.c fscache_put_object(object, fscache_obj_put_drop_obj); object 772 fs/fscache/object.c static int fscache_get_object(struct fscache_object *object, object 778 fs/fscache/object.c ret = object->cache->ops->grab_object(object, why) ? 0 : -EAGAIN; object 786 fs/fscache/object.c static void fscache_put_object(struct fscache_object *object, object 790 fs/fscache/object.c object->cache->ops->put_object(object, why); object 800 fs/fscache/object.c void fscache_object_destroy(struct fscache_object *object) object 802 fs/fscache/object.c fscache_objlist_remove(object); object 805 fs/fscache/object.c fscache_cookie_put(object->cookie, fscache_cookie_put_object); object 806 fs/fscache/object.c object->cookie = NULL; object 813 fs/fscache/object.c void fscache_enqueue_object(struct fscache_object *object) object 815 fs/fscache/object.c _enter("{OBJ%x}", object->debug_id); object 817 fs/fscache/object.c if (fscache_get_object(object, fscache_obj_get_queue) >= 0) { object 821 fs/fscache/object.c if (queue_work(fscache_object_wq, &object->work)) { object 825 fs/fscache/object.c fscache_put_object(object, fscache_obj_put_queue); object 867 fs/fscache/object.c static bool fscache_enqueue_dependents(struct fscache_object *object, int event) object 872 fs/fscache/object.c _enter("{OBJ%x}", object->debug_id); object 874 fs/fscache/object.c if (list_empty(&object->dependents)) object 877 fs/fscache/object.c spin_lock(&object->lock); object 879 fs/fscache/object.c while (!list_empty(&object->dependents)) { object 880 fs/fscache/object.c dep = list_entry(object->dependents.next, object 887 fs/fscache/object.c if (!list_empty(&object->dependents) && need_resched()) { object 893 fs/fscache/object.c spin_unlock(&object->lock); object 900 fs/fscache/object.c static void fscache_dequeue_object(struct fscache_object *object) object 902 fs/fscache/object.c _enter("{OBJ%x}", object->debug_id); object 904 fs/fscache/object.c if (!list_empty(&object->dep_link)) { object 905 fs/fscache/object.c spin_lock(&object->parent->lock); object 906 fs/fscache/object.c list_del_init(&object->dep_link); object 907 fs/fscache/object.c spin_unlock(&object->parent->lock); object 924 fs/fscache/object.c enum fscache_checkaux fscache_check_aux(struct fscache_object *object, object 930 fs/fscache/object.c if (!object->cookie->def->check_aux) { object 935 fs/fscache/object.c result = object->cookie->def->check_aux(object->cookie->netfs_data, object 964 fs/fscache/object.c static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object, object 968 fs/fscache/object.c struct fscache_cookie *cookie = object->cookie; object 970 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 975 fs/fscache/object.c if (!fscache_use_cookie(object)) { object 976 fs/fscache/object.c ASSERT(radix_tree_empty(&object->cookie->stores)); object 977 fs/fscache/object.c set_bit(FSCACHE_OBJECT_RETIRED, &object->flags); object 984 fs/fscache/object.c clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); object 985 fs/fscache/object.c fscache_cancel_all_ops(object); object 992 fs/fscache/object.c fscache_operation_init(cookie, op, object->cache->ops->invalidate_object, object 1000 fs/fscache/object.c if (fscache_submit_exclusive_op(object, op) < 0) object 1020 fs/fscache/object.c fscache_mark_object_dead(object); object 1021 fs/fscache/object.c fscache_unuse_cookie(object); object 1026 fs/fscache/object.c fscache_mark_object_dead(object); object 1028 fs/fscache/object.c fscache_unuse_cookie(object); object 1034 fs/fscache/object.c static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object, object 1041 fs/fscache/object.c s = _fscache_invalidate_object(object, event); object 1049 fs/fscache/object.c static void fscache_update_aux_data(struct fscache_object *object) object 1053 fs/fscache/object.c object->cache->ops->update_object(object); object 1060 fs/fscache/object.c static const struct fscache_state *fscache_update_object(struct fscache_object *object, object 1063 fs/fscache/object.c _enter("{OBJ%x},%d", object->debug_id, event); object 1065 fs/fscache/object.c fscache_update_aux_data(object); object 1078 fs/fscache/object.c void fscache_object_retrying_stale(struct fscache_object *object) object 1092 fs/fscache/object.c void fscache_object_mark_killed(struct fscache_object *object, object 1095 fs/fscache/object.c if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) { object 1097 fs/fscache/object.c object->cache->identifier); object 1124 fs/fscache/object.c static const struct fscache_state *fscache_object_dead(struct fscache_object *object, object 1128 fs/fscache/object.c &object->flags)) object 60 fs/fscache/operation.c struct fscache_cookie *cookie = op->object->cookie; object 63 fs/fscache/operation.c op->object->debug_id, op->debug_id, atomic_read(&op->usage)); object 67 fs/fscache/operation.c ASSERT(fscache_object_is_available(op->object)); object 96 fs/fscache/operation.c static void fscache_run_op(struct fscache_object *object, object 102 fs/fscache/operation.c object->n_in_progress++; object 108 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_run); object 115 fs/fscache/operation.c static void fscache_report_unexpected_submission(struct fscache_object *object, object 128 fs/fscache/operation.c op->debug_id, object->debug_id, object->state->name); object 129 fs/fscache/operation.c kdebug("objstate=%s [%s]", object->state->name, ostate->name); object 130 fs/fscache/operation.c kdebug("objflags=%lx", object->flags); object 131 fs/fscache/operation.c kdebug("objevent=%lx [%lx]", object->events, object->event_mask); object 133 fs/fscache/operation.c object->n_ops, object->n_in_progress, object->n_exclusive); object 135 fs/fscache/operation.c if (!list_empty(&object->pending_ops)) { object 137 fs/fscache/operation.c list_for_each_entry(p, &object->pending_ops, pend_link) { object 138 fs/fscache/operation.c ASSERTCMP(p->object, ==, object); object 154 fs/fscache/operation.c int fscache_submit_exclusive_op(struct fscache_object *object, object 161 fs/fscache/operation.c _enter("{OBJ%x OP%x},", object->debug_id, op->debug_id); object 163 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_submit_ex); object 168 fs/fscache/operation.c spin_lock(&object->lock); object 169 fs/fscache/operation.c ASSERTCMP(object->n_ops, >=, object->n_in_progress); object 170 fs/fscache/operation.c ASSERTCMP(object->n_ops, >=, object->n_exclusive); object 173 fs/fscache/operation.c ostate = object->state; object 177 fs/fscache/operation.c flags = READ_ONCE(object->flags); object 183 fs/fscache/operation.c } else if (unlikely(fscache_cache_is_broken(object))) { object 188 fs/fscache/operation.c op->object = object; object 189 fs/fscache/operation.c object->n_ops++; object 190 fs/fscache/operation.c object->n_exclusive++; /* reads and writes must wait */ object 192 fs/fscache/operation.c if (object->n_in_progress > 0) { object 194 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 196 fs/fscache/operation.c } else if (!list_empty(&object->pending_ops)) { object 198 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 200 fs/fscache/operation.c fscache_start_operations(object); object 202 fs/fscache/operation.c ASSERTCMP(object->n_in_progress, ==, 0); object 203 fs/fscache/operation.c fscache_run_op(object, op); object 207 fs/fscache/operation.c clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); object 210 fs/fscache/operation.c op->object = object; object 211 fs/fscache/operation.c object->n_ops++; object 212 fs/fscache/operation.c object->n_exclusive++; /* reads and writes must wait */ object 214 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 222 fs/fscache/operation.c fscache_report_unexpected_submission(object, op, ostate); object 228 fs/fscache/operation.c spin_unlock(&object->lock); object 240 fs/fscache/operation.c int fscache_submit_op(struct fscache_object *object, object 248 fs/fscache/operation.c object->debug_id, op->debug_id, atomic_read(&op->usage)); object 250 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_submit); object 255 fs/fscache/operation.c spin_lock(&object->lock); object 256 fs/fscache/operation.c ASSERTCMP(object->n_ops, >=, object->n_in_progress); object 257 fs/fscache/operation.c ASSERTCMP(object->n_ops, >=, object->n_exclusive); object 260 fs/fscache/operation.c ostate = object->state; object 264 fs/fscache/operation.c flags = READ_ONCE(object->flags); object 270 fs/fscache/operation.c } else if (unlikely(fscache_cache_is_broken(object))) { object 275 fs/fscache/operation.c op->object = object; object 276 fs/fscache/operation.c object->n_ops++; object 278 fs/fscache/operation.c if (object->n_exclusive > 0) { object 280 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 282 fs/fscache/operation.c } else if (!list_empty(&object->pending_ops)) { object 284 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 286 fs/fscache/operation.c fscache_start_operations(object); object 288 fs/fscache/operation.c ASSERTCMP(object->n_exclusive, ==, 0); object 289 fs/fscache/operation.c fscache_run_op(object, op); object 293 fs/fscache/operation.c op->object = object; object 294 fs/fscache/operation.c object->n_ops++; object 296 fs/fscache/operation.c list_add_tail(&op->pend_link, &object->pending_ops); object 304 fs/fscache/operation.c fscache_report_unexpected_submission(object, op, ostate); object 305 fs/fscache/operation.c ASSERT(!fscache_object_is_active(object)); object 311 fs/fscache/operation.c spin_unlock(&object->lock); object 319 fs/fscache/operation.c void fscache_abort_object(struct fscache_object *object) object 321 fs/fscache/operation.c _enter("{OBJ%x}", object->debug_id); object 323 fs/fscache/operation.c fscache_raise_event(object, FSCACHE_OBJECT_EV_ERROR); object 330 fs/fscache/operation.c void fscache_start_operations(struct fscache_object *object) object 335 fs/fscache/operation.c while (!list_empty(&object->pending_ops) && !stop) { object 336 fs/fscache/operation.c op = list_entry(object->pending_ops.next, object 340 fs/fscache/operation.c if (object->n_in_progress > 0) object 345 fs/fscache/operation.c fscache_run_op(object, op); object 351 fs/fscache/operation.c ASSERTCMP(object->n_in_progress, <=, object->n_ops); object 354 fs/fscache/operation.c object->n_in_progress, object->debug_id); object 363 fs/fscache/operation.c struct fscache_object *object = op->object; object 367 fs/fscache/operation.c _enter("OBJ%x OP%x}", op->object->debug_id, op->debug_id); object 369 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_cancel); object 375 fs/fscache/operation.c spin_lock(&object->lock); object 387 fs/fscache/operation.c object->n_exclusive--; object 392 fs/fscache/operation.c ASSERTCMP(object->n_in_progress, >, 0); object 394 fs/fscache/operation.c object->n_exclusive--; object 395 fs/fscache/operation.c object->n_in_progress--; object 396 fs/fscache/operation.c if (object->n_in_progress == 0) object 397 fs/fscache/operation.c fscache_start_operations(object); object 403 fs/fscache/operation.c object->n_exclusive--; object 411 fs/fscache/operation.c spin_unlock(&object->lock); object 419 fs/fscache/operation.c void fscache_cancel_all_ops(struct fscache_object *object) object 423 fs/fscache/operation.c _enter("OBJ%x", object->debug_id); object 425 fs/fscache/operation.c spin_lock(&object->lock); object 427 fs/fscache/operation.c while (!list_empty(&object->pending_ops)) { object 428 fs/fscache/operation.c op = list_entry(object->pending_ops.next, object 433 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_cancel_all); object 440 fs/fscache/operation.c object->n_exclusive--; object 444 fs/fscache/operation.c cond_resched_lock(&object->lock); object 447 fs/fscache/operation.c spin_unlock(&object->lock); object 456 fs/fscache/operation.c struct fscache_object *object = op->object; object 458 fs/fscache/operation.c _enter("OBJ%x", object->debug_id); object 461 fs/fscache/operation.c ASSERTCMP(object->n_in_progress, >, 0); object 463 fs/fscache/operation.c object->n_exclusive, >, 0); object 465 fs/fscache/operation.c object->n_in_progress, ==, 1); object 467 fs/fscache/operation.c spin_lock(&object->lock); object 470 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_completed); object 474 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_cancelled); object 479 fs/fscache/operation.c object->n_exclusive--; object 480 fs/fscache/operation.c object->n_in_progress--; object 481 fs/fscache/operation.c if (object->n_in_progress == 0) object 482 fs/fscache/operation.c fscache_start_operations(object); object 484 fs/fscache/operation.c spin_unlock(&object->lock); object 495 fs/fscache/operation.c struct fscache_object *object; object 499 fs/fscache/operation.c op->object ? op->object->debug_id : 0, object 507 fs/fscache/operation.c trace_fscache_op(op->object ? op->object->cookie : NULL, op, fscache_op_put); object 522 fs/fscache/operation.c object = op->object; object 523 fs/fscache/operation.c if (likely(object)) { object 525 fs/fscache/operation.c atomic_dec(&object->n_reads); object 527 fs/fscache/operation.c fscache_unuse_cookie(object); object 532 fs/fscache/operation.c if (!spin_trylock(&object->lock)) { object 536 fs/fscache/operation.c cache = object->cache; object 545 fs/fscache/operation.c ASSERTCMP(object->n_ops, >, 0); object 546 fs/fscache/operation.c object->n_ops--; object 547 fs/fscache/operation.c if (object->n_ops == 0) object 548 fs/fscache/operation.c fscache_raise_event(object, FSCACHE_OBJECT_EV_CLEARED); object 550 fs/fscache/operation.c spin_unlock(&object->lock); object 564 fs/fscache/operation.c struct fscache_object *object; object 583 fs/fscache/operation.c object = op->object; object 584 fs/fscache/operation.c trace_fscache_op(object->cookie, op, fscache_op_gc); object 586 fs/fscache/operation.c spin_lock(&object->lock); object 589 fs/fscache/operation.c object->debug_id, op->debug_id); object 595 fs/fscache/operation.c ASSERTCMP(object->n_ops, >, 0); object 596 fs/fscache/operation.c object->n_ops--; object 597 fs/fscache/operation.c if (object->n_ops == 0) object 598 fs/fscache/operation.c fscache_raise_event(object, FSCACHE_OBJECT_EV_CLEARED); object 600 fs/fscache/operation.c spin_unlock(&object->lock); object 622 fs/fscache/operation.c op->object->debug_id, op->debug_id, atomic_read(&op->usage)); object 624 fs/fscache/operation.c trace_fscache_op(op->object->cookie, op, fscache_op_work); object 146 fs/fscache/page.c static void fscache_end_page_write(struct fscache_object *object, object 152 fs/fscache/page.c spin_lock(&object->lock); object 153 fs/fscache/page.c cookie = object->cookie; object 179 fs/fscache/page.c spin_unlock(&object->lock); object 189 fs/fscache/page.c struct fscache_object *object = op->object; object 192 fs/fscache/page.c _enter("{OBJ%x OP%x}", object->debug_id, op->debug_id); object 196 fs/fscache/page.c if (fscache_object_is_active(object)) { object 198 fs/fscache/page.c ret = object->cache->ops->attr_changed(object); object 201 fs/fscache/page.c fscache_abort_object(object); object 216 fs/fscache/page.c struct fscache_object *object; object 243 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 247 fs/fscache/page.c if (fscache_submit_exclusive_op(object, op) < 0) object 373 fs/fscache/page.c int fscache_wait_for_operation_activation(struct fscache_object *object, object 388 fs/fscache/page.c trace_fscache_op(object->cookie, op, fscache_op_signal); object 407 fs/fscache/page.c if (unlikely(fscache_object_is_dying(object) || object 408 fs/fscache/page.c fscache_cache_is_broken(object))) { object 410 fs/fscache/page.c trace_fscache_op(object->cookie, op, fscache_op_signal); object 436 fs/fscache/page.c struct fscache_object *object; object 472 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 475 fs/fscache/page.c ASSERT(test_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)); object 478 fs/fscache/page.c atomic_inc(&object->n_reads); object 481 fs/fscache/page.c if (fscache_submit_op(object, &op->op) < 0) object 490 fs/fscache/page.c object, &op->op, object 497 fs/fscache/page.c if (test_bit(FSCACHE_COOKIE_NO_DATA_YET, &object->cookie->flags)) { object 499 fs/fscache/page.c ret = object->cache->ops->allocate_page(op, page, gfp); object 505 fs/fscache/page.c ret = object->cache->ops->read_or_alloc_page(op, page, gfp); object 526 fs/fscache/page.c atomic_dec(&object->n_reads); object 567 fs/fscache/page.c struct fscache_object *object; object 601 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 605 fs/fscache/page.c atomic_inc(&object->n_reads); object 608 fs/fscache/page.c if (fscache_submit_op(object, &op->op) < 0) object 617 fs/fscache/page.c object, &op->op, object 624 fs/fscache/page.c if (test_bit(FSCACHE_COOKIE_NO_DATA_YET, &object->cookie->flags)) { object 626 fs/fscache/page.c ret = object->cache->ops->allocate_pages( object 631 fs/fscache/page.c ret = object->cache->ops->read_or_alloc_pages( object 653 fs/fscache/page.c atomic_dec(&object->n_reads); object 680 fs/fscache/page.c struct fscache_object *object; object 713 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 717 fs/fscache/page.c if (fscache_submit_op(object, &op->op) < 0) object 724 fs/fscache/page.c object, &op->op, object 732 fs/fscache/page.c ret = object->cache->ops->allocate_page(op, page, gfp); object 792 fs/fscache/page.c struct fscache_object *object = op->op.object; object 802 fs/fscache/page.c spin_lock(&object->lock); object 803 fs/fscache/page.c cookie = object->cookie; object 805 fs/fscache/page.c if (!fscache_object_is_active(object)) { object 810 fs/fscache/page.c spin_unlock(&object->lock); object 823 fs/fscache/page.c spin_unlock(&object->lock); object 826 fs/fscache/page.c _op->flags, _op->state, object->state->short_name, object 827 fs/fscache/page.c object->flags); object 853 fs/fscache/page.c spin_unlock(&object->lock); object 860 fs/fscache/page.c ret = object->cache->ops->write_page(op, page); object 863 fs/fscache/page.c fscache_end_page_write(object, page); object 865 fs/fscache/page.c fscache_abort_object(object); object 877 fs/fscache/page.c fscache_end_page_write(object, page); object 885 fs/fscache/page.c clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); object 886 fs/fscache/page.c spin_unlock(&object->lock); object 966 fs/fscache/page.c struct fscache_object *object; object 1004 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 1006 fs/fscache/page.c if (test_bit(FSCACHE_IOERROR, &object->cache->flags)) object 1013 fs/fscache/page.c spin_lock(&object->lock); object 1015 fs/fscache/page.c if (object->store_limit_l != object_size) object 1016 fs/fscache/page.c fscache_set_store_limit(object, object_size); object 1020 fs/fscache/page.c _debug("store limit %llx", (unsigned long long) object->store_limit); object 1038 fs/fscache/page.c if (test_and_set_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags)) object 1042 fs/fscache/page.c spin_unlock(&object->lock); object 1045 fs/fscache/page.c op->store_limit = object->store_limit; object 1048 fs/fscache/page.c if (fscache_submit_op(object, &op->op) < 0) object 1065 fs/fscache/page.c spin_unlock(&object->lock); object 1085 fs/fscache/page.c spin_unlock(&object->lock); object 1110 fs/fscache/page.c struct fscache_object *object; object 1133 fs/fscache/page.c object = hlist_entry(cookie->backing_objects.first, object 1142 fs/fscache/page.c object->cache->ops->uncache_page) { object 1145 fs/fscache/page.c object->cache->ops->uncache_page(object, page); object 1167 fs/fscache/page.c struct fscache_cookie *cookie = op->op.object->cookie; object 274 fs/gfs2/glock.h static inline void glock_set_object(struct gfs2_glock *gl, void *object) object 279 fs/gfs2/glock.h gl->gl_object = object; object 301 fs/gfs2/glock.h static inline void glock_clear_object(struct gfs2_glock *gl, void *object) object 304 fs/gfs2/glock.h if (gl->gl_object == object) object 219 include/acpi/acpiosxf.h acpi_status acpi_os_release_object(acpi_cache_t * cache, void *object); object 464 include/acpi/acpixf.h acpi_unload_parent_table(acpi_handle object)) object 519 include/acpi/acpixf.h acpi_get_name(acpi_handle object, u32 name_type, object 526 include/acpi/acpixf.h acpi_attach_data(acpi_handle object, object 530 include/acpi/acpixf.h acpi_detach_data(acpi_handle object, object 533 include/acpi/acpixf.h acpi_get_data(acpi_handle object, object 544 include/acpi/acpixf.h acpi_evaluate_object(acpi_handle object, object 551 include/acpi/acpixf.h acpi_evaluate_object_typed(acpi_handle object, object 560 include/acpi/acpixf.h acpi_get_object_info(acpi_handle object, object 572 include/acpi/acpixf.h acpi_get_type(acpi_handle object, object 576 include/acpi/acpixf.h acpi_get_parent(acpi_handle object, object 951 include/acpi/acpixf.h acpi_get_data_full(acpi_handle object, object 1054 include/acpi/actypes.h void (*acpi_object_handler) (acpi_handle object, void *data); object 1057 include/acpi/actypes.h acpi_status (*acpi_init_handler) (acpi_handle object, u32 function); object 1110 include/acpi/actypes.h acpi_status (*acpi_walk_callback) (acpi_handle object, object 35 include/linux/assoc_array.h unsigned long (*get_object_key_chunk)(const void *object, int level); object 38 include/linux/assoc_array.h bool (*compare_object)(const void *object, const void *index_key); object 43 include/linux/assoc_array.h int (*diff_objects)(const void *object, const void *index_key); object 46 include/linux/assoc_array.h void (*free_object)(void *object); object 61 include/linux/assoc_array.h int (*iterator)(const void *object, object 72 include/linux/assoc_array.h void *object); object 74 include/linux/assoc_array.h void *object); object 84 include/linux/assoc_array.h bool (*iterator)(void *object, void *iterator_data), object 32 include/linux/debugobjects.h void *object; object 100 include/linux/fscache-cache.h struct fscache_object *object; /* object to be operated upon */ object 236 include/linux/fscache-cache.h int (*lookup_object)(struct fscache_object *object); object 239 include/linux/fscache-cache.h void (*lookup_complete)(struct fscache_object *object); object 242 include/linux/fscache-cache.h struct fscache_object *(*grab_object)(struct fscache_object *object, object 246 include/linux/fscache-cache.h int (*pin_object)(struct fscache_object *object); object 249 include/linux/fscache-cache.h void (*unpin_object)(struct fscache_object *object); object 256 include/linux/fscache-cache.h void (*update_object)(struct fscache_object *object); object 263 include/linux/fscache-cache.h void (*drop_object)(struct fscache_object *object); object 266 include/linux/fscache-cache.h void (*put_object)(struct fscache_object *object, object 274 include/linux/fscache-cache.h int (*attr_changed)(struct fscache_object *object); object 277 include/linux/fscache-cache.h int (*reserve_space)(struct fscache_object *object, loff_t i_size); object 302 include/linux/fscache-cache.h void (*uncache_page)(struct fscache_object *object, object 338 include/linux/fscache-cache.h const struct fscache_state *(*work)(struct fscache_object *object, object 395 include/linux/fscache-cache.h extern void fscache_object_lookup_negative(struct fscache_object *object); object 396 include/linux/fscache-cache.h extern void fscache_obtained_object(struct fscache_object *object); object 398 include/linux/fscache-cache.h static inline bool fscache_object_is_live(struct fscache_object *object) object 400 include/linux/fscache-cache.h return test_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags); object 403 include/linux/fscache-cache.h static inline bool fscache_object_is_dying(struct fscache_object *object) object 405 include/linux/fscache-cache.h return !fscache_object_is_live(object); object 408 include/linux/fscache-cache.h static inline bool fscache_object_is_available(struct fscache_object *object) object 410 include/linux/fscache-cache.h return test_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags); object 413 include/linux/fscache-cache.h static inline bool fscache_cache_is_broken(struct fscache_object *object) object 415 include/linux/fscache-cache.h return test_bit(FSCACHE_IOERROR, &object->cache->flags); object 418 include/linux/fscache-cache.h static inline bool fscache_object_is_active(struct fscache_object *object) object 420 include/linux/fscache-cache.h return fscache_object_is_available(object) && object 421 include/linux/fscache-cache.h fscache_object_is_live(object) && object 422 include/linux/fscache-cache.h !fscache_cache_is_broken(object); object 444 include/linux/fscache-cache.h static inline void fscache_object_lookup_error(struct fscache_object *object) object 446 include/linux/fscache-cache.h set_bit(FSCACHE_OBJECT_EV_ERROR, &object->events); object 461 include/linux/fscache-cache.h void fscache_set_store_limit(struct fscache_object *object, loff_t i_size) object 463 include/linux/fscache-cache.h object->store_limit_l = i_size; object 464 include/linux/fscache-cache.h object->store_limit = i_size >> PAGE_SHIFT; object 466 include/linux/fscache-cache.h object->store_limit++; object 496 include/linux/fscache-cache.h static inline bool fscache_use_cookie(struct fscache_object *object) object 498 include/linux/fscache-cache.h struct fscache_cookie *cookie = object->cookie; object 519 include/linux/fscache-cache.h static inline void fscache_unuse_cookie(struct fscache_object *object) object 521 include/linux/fscache-cache.h struct fscache_cookie *cookie = object->cookie; object 549 include/linux/fscache-cache.h extern enum fscache_checkaux fscache_check_aux(struct fscache_object *object, object 554 include/linux/fscache-cache.h extern void fscache_object_retrying_stale(struct fscache_object *object); object 562 include/linux/fscache-cache.h extern void fscache_object_mark_killed(struct fscache_object *object, object 97 include/linux/hil_mlc.h } object; object 50 include/linux/kasan.h void kasan_unpoison_object_data(struct kmem_cache *cache, void *object); object 51 include/linux/kasan.h void kasan_poison_object_data(struct kmem_cache *cache, void *object); object 53 include/linux/kasan.h const void *object); object 59 include/linux/kasan.h void * __must_check kasan_kmalloc(struct kmem_cache *s, const void *object, object 61 include/linux/kasan.h void * __must_check kasan_krealloc(const void *object, size_t new_size, object 64 include/linux/kasan.h void * __must_check kasan_slab_alloc(struct kmem_cache *s, void *object, object 66 include/linux/kasan.h bool kasan_slab_free(struct kmem_cache *s, void *object, unsigned long ip); object 108 include/linux/kasan.h void *object) {} object 110 include/linux/kasan.h void *object) {} object 112 include/linux/kasan.h const void *object) object 114 include/linux/kasan.h return (void *)object; object 123 include/linux/kasan.h static inline void *kasan_kmalloc(struct kmem_cache *s, const void *object, object 126 include/linux/kasan.h return (void *)object; object 128 include/linux/kasan.h static inline void *kasan_krealloc(const void *object, size_t new_size, object 131 include/linux/kasan.h return (void *)object; object 134 include/linux/kasan.h static inline void *kasan_slab_alloc(struct kmem_cache *s, void *object, object 137 include/linux/kasan.h return object; object 139 include/linux/kasan.h static inline bool kasan_slab_free(struct kmem_cache *s, void *object, object 198 include/linux/mlx5/driver.h void *object; object 95 include/linux/slab_def.h void *object = x - (x - page->s_mem) % cache->size; object 98 include/linux/slab_def.h if (unlikely(object > last_object)) object 101 include/linux/slab_def.h return object; object 170 include/linux/slub_def.h u8 *object, char *reason); object 176 include/linux/slub_def.h void *object = x - (x - page_address(page)) % cache->size; object 179 include/linux/slub_def.h void *result = (unlikely(object > last_object)) ? last_object : object; object 248 include/media/media-request.h int (*prepare)(struct media_request_object *object); object 249 include/media/media-request.h void (*unprepare)(struct media_request_object *object); object 250 include/media/media-request.h void (*queue)(struct media_request_object *object); object 251 include/media/media-request.h void (*unbind)(struct media_request_object *object); object 252 include/media/media-request.h void (*release)(struct media_request_object *object); object 1482 include/rdma/ib_verbs.h void *object; /* containing object */ object 715 include/rdma/uverbs_ioctl.h return attr->obj_attr.uobject->object; object 65 include/rdma/uverbs_std_types.h return uobj->object; object 147 include/rdma/uverbs_std_types.h uobj->object = action; object 176 include/rdma/uverbs_std_types.h uobj->object = ibflow; object 329 include/trace/events/fscache.h TP_PROTO(struct fscache_object *object, object 333 include/trace/events/fscache.h TP_ARGS(object, state, wait, oob, event_num), object 337 include/trace/events/fscache.h __field(struct fscache_object *, object ) object 345 include/trace/events/fscache.h __entry->cookie = object->cookie; object 346 include/trace/events/fscache.h __entry->object = object; object 355 include/trace/events/fscache.h __entry->object, object 134 lib/assoc_array.c int (*iterator)(const void *object, object 967 lib/assoc_array.c void *object) object 979 lib/assoc_array.c BUG_ON(assoc_array_ptr_is_meta(object)); object 986 lib/assoc_array.c edit->leaf = assoc_array_leaf_to_ptr(object); object 1031 lib/assoc_array.c void assoc_array_insert_set_object(struct assoc_array_edit *edit, void *object) object 1033 lib/assoc_array.c BUG_ON(!object); object 1034 lib/assoc_array.c edit->leaf = assoc_array_leaf_to_ptr(object); object 1456 lib/assoc_array.c bool (*iterator)(void *object, void *iterator_data), object 197 lib/debugobjects.c if (obj->object == addr) object 276 lib/debugobjects.c obj->object = addr; object 483 lib/debugobjects.c descr->debug_hint(obj->object) : NULL; object 627 lib/debugobjects.c struct debug_obj o = { .object = addr, object 734 lib/debugobjects.c struct debug_obj o = { .object = addr, object 856 lib/debugobjects.c struct debug_obj o = { .object = addr, object 922 lib/debugobjects.c struct debug_obj o = { .object = addr, object 958 lib/debugobjects.c oaddr = (unsigned long) obj->object; object 305 mm/kasan/common.c const void *object) object 307 mm/kasan/common.c return (void *)object + cache->kasan_info.alloc_meta_offset; object 311 mm/kasan/common.c const void *object) object 314 mm/kasan/common.c return (void *)object + cache->kasan_info.free_meta_offset; object 319 mm/kasan/common.c void *object, u8 tag) object 324 mm/kasan/common.c alloc_meta = get_alloc_info(cache, object); object 345 mm/kasan/common.c void kasan_unpoison_object_data(struct kmem_cache *cache, void *object) object 347 mm/kasan/common.c kasan_unpoison_shadow(object, cache->object_size); object 350 mm/kasan/common.c void kasan_poison_object_data(struct kmem_cache *cache, void *object) object 352 mm/kasan/common.c kasan_poison_shadow(object, object 371 mm/kasan/common.c static u8 assign_tag(struct kmem_cache *cache, const void *object, object 381 mm/kasan/common.c return get_tag(object); object 393 mm/kasan/common.c return (u8)obj_to_index(cache, virt_to_page(object), (void *)object); object 399 mm/kasan/common.c return init ? random_tag() : get_tag(object); object 404 mm/kasan/common.c const void *object) object 409 mm/kasan/common.c return (void *)object; object 411 mm/kasan/common.c alloc_info = get_alloc_info(cache, object); object 415 mm/kasan/common.c object = set_tag(object, object 416 mm/kasan/common.c assign_tag(cache, object, true, false)); object 418 mm/kasan/common.c return (void *)object; object 436 mm/kasan/common.c static bool __kasan_slab_free(struct kmem_cache *cache, void *object, object 444 mm/kasan/common.c tag = get_tag(object); object 445 mm/kasan/common.c tagged_object = object; object 446 mm/kasan/common.c object = reset_tag(object); object 448 mm/kasan/common.c if (unlikely(nearest_obj(cache, virt_to_head_page(object), object) != object 449 mm/kasan/common.c object)) { object 458 mm/kasan/common.c shadow_byte = READ_ONCE(*(s8 *)kasan_mem_to_shadow(object)); object 465 mm/kasan/common.c kasan_poison_shadow(object, rounded_up_size, KASAN_KMALLOC_FREE); object 471 mm/kasan/common.c kasan_set_free_info(cache, object, tag); object 473 mm/kasan/common.c quarantine_put(get_free_info(cache, object), cache); object 478 mm/kasan/common.c bool kasan_slab_free(struct kmem_cache *cache, void *object, unsigned long ip) object 480 mm/kasan/common.c return __kasan_slab_free(cache, object, ip, true); object 483 mm/kasan/common.c static void *__kasan_kmalloc(struct kmem_cache *cache, const void *object, object 493 mm/kasan/common.c if (unlikely(object == NULL)) object 496 mm/kasan/common.c redzone_start = round_up((unsigned long)(object + size), object 498 mm/kasan/common.c redzone_end = round_up((unsigned long)object + cache->object_size, object 502 mm/kasan/common.c tag = assign_tag(cache, object, false, keep_tag); object 505 mm/kasan/common.c kasan_unpoison_shadow(set_tag(object, tag), size); object 510 mm/kasan/common.c set_track(&get_alloc_info(cache, object)->alloc_track, flags); object 512 mm/kasan/common.c return set_tag(object, tag); object 515 mm/kasan/common.c void * __must_check kasan_slab_alloc(struct kmem_cache *cache, void *object, object 518 mm/kasan/common.c return __kasan_kmalloc(cache, object, cache->object_size, flags, false); object 521 mm/kasan/common.c void * __must_check kasan_kmalloc(struct kmem_cache *cache, const void *object, object 524 mm/kasan/common.c return __kasan_kmalloc(cache, object, size, flags, true); object 553 mm/kasan/common.c void * __must_check kasan_krealloc(const void *object, size_t size, gfp_t flags) object 557 mm/kasan/common.c if (unlikely(object == ZERO_SIZE_PTR)) object 558 mm/kasan/common.c return (void *)object; object 560 mm/kasan/common.c page = virt_to_head_page(object); object 563 mm/kasan/common.c return kasan_kmalloc_large(object, size, flags); object 565 mm/kasan/common.c return __kasan_kmalloc(page->slab_cache, object, size, object 124 mm/kasan/kasan.h const void *object); object 126 mm/kasan/kasan.h const void *object); object 157 mm/kasan/kasan.h void kasan_report_invalid_free(void *object, unsigned long ip); object 142 mm/kasan/quarantine.c void *object = qlink_to_object(qlink, cache); object 148 mm/kasan/quarantine.c ___cache_free(cache, object, _THIS_IP_); object 122 mm/kasan/report.c static void describe_object_addr(struct kmem_cache *cache, void *object, object 126 mm/kasan/report.c unsigned long object_addr = (unsigned long)object; object 132 mm/kasan/report.c object, cache->name, cache->object_size); object 155 mm/kasan/report.c void *object, u8 tag) object 160 mm/kasan/report.c alloc_meta = get_alloc_info(cache, object); object 174 mm/kasan/report.c static void describe_object(struct kmem_cache *cache, void *object, object 177 mm/kasan/report.c struct kasan_alloc_meta *alloc_info = get_alloc_info(cache, object); object 184 mm/kasan/report.c free_track = kasan_get_free_track(cache, object, tag); object 189 mm/kasan/report.c describe_object_addr(cache, object, addr); object 379 mm/kasan/report.c void *object = nearest_obj(cache, page, addr); object 381 mm/kasan/report.c describe_object(cache, object, addr, tag); object 458 mm/kasan/report.c void kasan_report_invalid_free(void *object, unsigned long ip) object 461 mm/kasan/report.c u8 tag = get_tag(object); object 463 mm/kasan/report.c object = reset_tag(object); object 466 mm/kasan/report.c print_tags(tag, object); object 468 mm/kasan/report.c print_address_description(object, tag); object 470 mm/kasan/report.c print_shadow_for_address(object); object 44 mm/kasan/tags_report.c void *object; object 53 mm/kasan/tags_report.c object = nearest_obj(cache, page, (void *)addr); object 54 mm/kasan/tags_report.c alloc_meta = get_alloc_info(cache, object); object 281 mm/kmemleak.c struct kmemleak_object *object) object 283 mm/kmemleak.c const u8 *ptr = (const u8 *)object->pointer; object 287 mm/kmemleak.c len = min_t(size_t, object->size, HEX_MAX_LINES * HEX_ROW_SIZE); object 306 mm/kmemleak.c static bool color_white(const struct kmemleak_object *object) object 308 mm/kmemleak.c return object->count != KMEMLEAK_BLACK && object 309 mm/kmemleak.c object->count < object->min_count; object 312 mm/kmemleak.c static bool color_gray(const struct kmemleak_object *object) object 314 mm/kmemleak.c return object->min_count != KMEMLEAK_BLACK && object 315 mm/kmemleak.c object->count >= object->min_count; object 323 mm/kmemleak.c static bool unreferenced_object(struct kmemleak_object *object) object 325 mm/kmemleak.c return (color_white(object) && object->flags & OBJECT_ALLOCATED) && object 326 mm/kmemleak.c time_before_eq(object->jiffies + jiffies_min_age, object 335 mm/kmemleak.c struct kmemleak_object *object) object 338 mm/kmemleak.c unsigned int msecs_age = jiffies_to_msecs(jiffies - object->jiffies); object 341 mm/kmemleak.c object->pointer, object->size); object 343 mm/kmemleak.c object->comm, object->pid, object->jiffies, object 345 mm/kmemleak.c hex_dump_object(seq, object); object 348 mm/kmemleak.c for (i = 0; i < object->trace_len; i++) { object 349 mm/kmemleak.c void *ptr = (void *)object->trace[i]; object 359 mm/kmemleak.c static void dump_object_info(struct kmemleak_object *object) object 362 mm/kmemleak.c object->pointer, object->size); object 364 mm/kmemleak.c object->comm, object->pid, object->jiffies); object 365 mm/kmemleak.c pr_notice(" min_count = %d\n", object->min_count); object 366 mm/kmemleak.c pr_notice(" count = %d\n", object->count); object 367 mm/kmemleak.c pr_notice(" flags = 0x%x\n", object->flags); object 368 mm/kmemleak.c pr_notice(" checksum = %u\n", object->checksum); object 370 mm/kmemleak.c stack_trace_print(object->trace, object->trace_len, 4); object 384 mm/kmemleak.c struct kmemleak_object *object = object 386 mm/kmemleak.c if (ptr < object->pointer) object 387 mm/kmemleak.c rb = object->rb_node.rb_left; object 388 mm/kmemleak.c else if (object->pointer + object->size <= ptr) object 389 mm/kmemleak.c rb = object->rb_node.rb_right; object 390 mm/kmemleak.c else if (object->pointer == ptr || alias) object 391 mm/kmemleak.c return object; object 395 mm/kmemleak.c dump_object_info(object); object 408 mm/kmemleak.c static int get_object(struct kmemleak_object *object) object 410 mm/kmemleak.c return atomic_inc_not_zero(&object->use_count); object 419 mm/kmemleak.c struct kmemleak_object *object; object 423 mm/kmemleak.c object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp)); object 424 mm/kmemleak.c if (object) object 425 mm/kmemleak.c return object; object 430 mm/kmemleak.c object = list_first_entry_or_null(&mem_pool_free_list, object 431 mm/kmemleak.c typeof(*object), object_list); object 432 mm/kmemleak.c if (object) object 433 mm/kmemleak.c list_del(&object->object_list); object 435 mm/kmemleak.c object = &mem_pool[--mem_pool_free_count]; object 440 mm/kmemleak.c return object; object 446 mm/kmemleak.c static void mem_pool_free(struct kmemleak_object *object) object 450 mm/kmemleak.c if (object < mem_pool || object >= mem_pool + ARRAY_SIZE(mem_pool)) { object 451 mm/kmemleak.c kmem_cache_free(object_cache, object); object 457 mm/kmemleak.c list_add(&object->object_list, &mem_pool_free_list); object 468 mm/kmemleak.c struct kmemleak_object *object = object 475 mm/kmemleak.c hlist_for_each_entry_safe(area, tmp, &object->area_list, node) { object 479 mm/kmemleak.c mem_pool_free(object); object 489 mm/kmemleak.c static void put_object(struct kmemleak_object *object) object 491 mm/kmemleak.c if (!atomic_dec_and_test(&object->use_count)) object 495 mm/kmemleak.c WARN_ON(object->flags & OBJECT_ALLOCATED); object 503 mm/kmemleak.c call_rcu(&object->rcu, free_object_rcu); object 505 mm/kmemleak.c free_object_rcu(&object->rcu); object 514 mm/kmemleak.c struct kmemleak_object *object; object 518 mm/kmemleak.c object = lookup_object(ptr, alias); object 522 mm/kmemleak.c if (object && !get_object(object)) object 523 mm/kmemleak.c object = NULL; object 526 mm/kmemleak.c return object; object 533 mm/kmemleak.c static void __remove_object(struct kmemleak_object *object) object 535 mm/kmemleak.c rb_erase(&object->rb_node, &object_tree_root); object 536 mm/kmemleak.c list_del_rcu(&object->object_list); object 547 mm/kmemleak.c struct kmemleak_object *object; object 550 mm/kmemleak.c object = lookup_object(ptr, alias); object 551 mm/kmemleak.c if (object) object 552 mm/kmemleak.c __remove_object(object); object 555 mm/kmemleak.c return object; object 574 mm/kmemleak.c struct kmemleak_object *object, *parent; object 578 mm/kmemleak.c object = mem_pool_alloc(gfp); object 579 mm/kmemleak.c if (!object) { object 585 mm/kmemleak.c INIT_LIST_HEAD(&object->object_list); object 586 mm/kmemleak.c INIT_LIST_HEAD(&object->gray_list); object 587 mm/kmemleak.c INIT_HLIST_HEAD(&object->area_list); object 588 mm/kmemleak.c spin_lock_init(&object->lock); object 589 mm/kmemleak.c atomic_set(&object->use_count, 1); object 590 mm/kmemleak.c object->flags = OBJECT_ALLOCATED; object 591 mm/kmemleak.c object->pointer = ptr; object 592 mm/kmemleak.c object->size = size; object 593 mm/kmemleak.c object->excess_ref = 0; object 594 mm/kmemleak.c object->min_count = min_count; object 595 mm/kmemleak.c object->count = 0; /* white color initially */ object 596 mm/kmemleak.c object->jiffies = jiffies; object 597 mm/kmemleak.c object->checksum = 0; object 601 mm/kmemleak.c object->pid = 0; object 602 mm/kmemleak.c strncpy(object->comm, "hardirq", sizeof(object->comm)); object 604 mm/kmemleak.c object->pid = 0; object 605 mm/kmemleak.c strncpy(object->comm, "softirq", sizeof(object->comm)); object 607 mm/kmemleak.c object->pid = current->pid; object 614 mm/kmemleak.c strncpy(object->comm, current->comm, sizeof(object->comm)); object 618 mm/kmemleak.c object->trace_len = __save_stack_trace(object->trace); object 642 mm/kmemleak.c kmem_cache_free(object_cache, object); object 643 mm/kmemleak.c object = NULL; object 647 mm/kmemleak.c rb_link_node(&object->rb_node, rb_parent, link); object 648 mm/kmemleak.c rb_insert_color(&object->rb_node, &object_tree_root); object 650 mm/kmemleak.c list_add_tail_rcu(&object->object_list, &object_list); object 653 mm/kmemleak.c return object; object 659 mm/kmemleak.c static void __delete_object(struct kmemleak_object *object) object 663 mm/kmemleak.c WARN_ON(!(object->flags & OBJECT_ALLOCATED)); object 664 mm/kmemleak.c WARN_ON(atomic_read(&object->use_count) < 1); object 670 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 671 mm/kmemleak.c object->flags &= ~OBJECT_ALLOCATED; object 672 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 673 mm/kmemleak.c put_object(object); object 682 mm/kmemleak.c struct kmemleak_object *object; object 684 mm/kmemleak.c object = find_and_remove_object(ptr, 0); object 685 mm/kmemleak.c if (!object) { object 692 mm/kmemleak.c __delete_object(object); object 702 mm/kmemleak.c struct kmemleak_object *object; object 705 mm/kmemleak.c object = find_and_remove_object(ptr, 1); object 706 mm/kmemleak.c if (!object) { object 719 mm/kmemleak.c start = object->pointer; object 720 mm/kmemleak.c end = object->pointer + object->size; object 722 mm/kmemleak.c create_object(start, ptr - start, object->min_count, object 725 mm/kmemleak.c create_object(ptr + size, end - ptr - size, object->min_count, object 728 mm/kmemleak.c __delete_object(object); object 731 mm/kmemleak.c static void __paint_it(struct kmemleak_object *object, int color) object 733 mm/kmemleak.c object->min_count = color; object 735 mm/kmemleak.c object->flags |= OBJECT_NO_SCAN; object 738 mm/kmemleak.c static void paint_it(struct kmemleak_object *object, int color) object 742 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 743 mm/kmemleak.c __paint_it(object, color); object 744 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 749 mm/kmemleak.c struct kmemleak_object *object; object 751 mm/kmemleak.c object = find_and_get_object(ptr, 0); object 752 mm/kmemleak.c if (!object) { object 759 mm/kmemleak.c paint_it(object, color); object 760 mm/kmemleak.c put_object(object); object 788 mm/kmemleak.c struct kmemleak_object *object; object 791 mm/kmemleak.c object = find_and_get_object(ptr, 1); object 792 mm/kmemleak.c if (!object) { object 801 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 805 mm/kmemleak.c object->flags |= OBJECT_FULL_SCAN; object 809 mm/kmemleak.c size = object->pointer + object->size - ptr; object 810 mm/kmemleak.c } else if (ptr + size > object->pointer + object->size) { object 812 mm/kmemleak.c dump_object_info(object); object 821 mm/kmemleak.c hlist_add_head(&area->node, &object->area_list); object 823 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 824 mm/kmemleak.c put_object(object); object 836 mm/kmemleak.c struct kmemleak_object *object; object 838 mm/kmemleak.c object = find_and_get_object(ptr, 0); object 839 mm/kmemleak.c if (!object) { object 845 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 846 mm/kmemleak.c object->excess_ref = excess_ref; object 847 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 848 mm/kmemleak.c put_object(object); object 859 mm/kmemleak.c struct kmemleak_object *object; object 861 mm/kmemleak.c object = find_and_get_object(ptr, 0); object 862 mm/kmemleak.c if (!object) { object 867 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 868 mm/kmemleak.c object->flags |= OBJECT_NO_SCAN; object 869 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 870 mm/kmemleak.c put_object(object); object 1012 mm/kmemleak.c struct kmemleak_object *object; object 1020 mm/kmemleak.c object = find_and_get_object((unsigned long)ptr, 1); object 1021 mm/kmemleak.c if (!object) { object 1029 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1030 mm/kmemleak.c object->trace_len = __save_stack_trace(object->trace); object 1031 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1033 mm/kmemleak.c put_object(object); object 1167 mm/kmemleak.c static bool update_checksum(struct kmemleak_object *object) object 1169 mm/kmemleak.c u32 old_csum = object->checksum; object 1172 mm/kmemleak.c object->checksum = crc32(0, (void *)object->pointer, object->size); object 1175 mm/kmemleak.c return object->checksum != old_csum; object 1181 mm/kmemleak.c static void update_refs(struct kmemleak_object *object) object 1183 mm/kmemleak.c if (!color_white(object)) { object 1194 mm/kmemleak.c object->count++; object 1195 mm/kmemleak.c if (color_gray(object)) { object 1197 mm/kmemleak.c WARN_ON(!get_object(object)); object 1198 mm/kmemleak.c list_add_tail(&object->gray_list, &gray_list); object 1238 mm/kmemleak.c struct kmemleak_object *object; object 1259 mm/kmemleak.c object = lookup_object(pointer, 1); object 1260 mm/kmemleak.c if (!object) object 1262 mm/kmemleak.c if (object == scanned) object 1271 mm/kmemleak.c spin_lock_nested(&object->lock, SINGLE_DEPTH_NESTING); object 1273 mm/kmemleak.c if (color_gray(object)) { object 1274 mm/kmemleak.c excess_ref = object->excess_ref; object 1278 mm/kmemleak.c update_refs(object); object 1280 mm/kmemleak.c spin_unlock(&object->lock); object 1283 mm/kmemleak.c object = lookup_object(excess_ref, 0); object 1284 mm/kmemleak.c if (!object) object 1286 mm/kmemleak.c if (object == scanned) object 1289 mm/kmemleak.c spin_lock_nested(&object->lock, SINGLE_DEPTH_NESTING); object 1290 mm/kmemleak.c update_refs(object); object 1291 mm/kmemleak.c spin_unlock(&object->lock); object 1318 mm/kmemleak.c static void scan_object(struct kmemleak_object *object) object 1327 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1328 mm/kmemleak.c if (object->flags & OBJECT_NO_SCAN) object 1330 mm/kmemleak.c if (!(object->flags & OBJECT_ALLOCATED)) object 1333 mm/kmemleak.c if (hlist_empty(&object->area_list) || object 1334 mm/kmemleak.c object->flags & OBJECT_FULL_SCAN) { object 1335 mm/kmemleak.c void *start = (void *)object->pointer; object 1336 mm/kmemleak.c void *end = (void *)(object->pointer + object->size); object 1341 mm/kmemleak.c scan_block(start, next, object); object 1347 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1349 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1350 mm/kmemleak.c } while (object->flags & OBJECT_ALLOCATED); object 1352 mm/kmemleak.c hlist_for_each_entry(area, &object->area_list, node) object 1355 mm/kmemleak.c object); object 1357 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1366 mm/kmemleak.c struct kmemleak_object *object, *tmp; object 1373 mm/kmemleak.c object = list_entry(gray_list.next, typeof(*object), gray_list); object 1374 mm/kmemleak.c while (&object->gray_list != &gray_list) { object 1379 mm/kmemleak.c scan_object(object); object 1381 mm/kmemleak.c tmp = list_entry(object->gray_list.next, typeof(*object), object 1385 mm/kmemleak.c list_del(&object->gray_list); object 1386 mm/kmemleak.c put_object(object); object 1388 mm/kmemleak.c object = tmp; object 1401 mm/kmemleak.c struct kmemleak_object *object; object 1409 mm/kmemleak.c list_for_each_entry_rcu(object, &object_list, object_list) { object 1410 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1416 mm/kmemleak.c if (atomic_read(&object->use_count) > 1) { object 1418 mm/kmemleak.c atomic_read(&object->use_count)); object 1419 mm/kmemleak.c dump_object_info(object); object 1423 mm/kmemleak.c object->count = 0; object 1424 mm/kmemleak.c if (color_gray(object) && get_object(object)) object 1425 mm/kmemleak.c list_add_tail(&object->gray_list, &gray_list); object 1427 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1494 mm/kmemleak.c list_for_each_entry_rcu(object, &object_list, object_list) { object 1495 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1496 mm/kmemleak.c if (color_white(object) && (object->flags & OBJECT_ALLOCATED) object 1497 mm/kmemleak.c && update_checksum(object) && get_object(object)) { object 1499 mm/kmemleak.c object->count = object->min_count; object 1500 mm/kmemleak.c list_add_tail(&object->gray_list, &gray_list); object 1502 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1521 mm/kmemleak.c list_for_each_entry_rcu(object, &object_list, object_list) { object 1522 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1523 mm/kmemleak.c if (unreferenced_object(object) && object 1524 mm/kmemleak.c !(object->flags & OBJECT_REPORTED)) { object 1525 mm/kmemleak.c object->flags |= OBJECT_REPORTED; object 1528 mm/kmemleak.c print_unreferenced(NULL, object); object 1532 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1616 mm/kmemleak.c struct kmemleak_object *object; object 1625 mm/kmemleak.c list_for_each_entry_rcu(object, &object_list, object_list) { object 1628 mm/kmemleak.c if (get_object(object)) object 1631 mm/kmemleak.c object = NULL; object 1633 mm/kmemleak.c return object; object 1681 mm/kmemleak.c struct kmemleak_object *object = v; object 1684 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1685 mm/kmemleak.c if ((object->flags & OBJECT_REPORTED) && unreferenced_object(object)) object 1686 mm/kmemleak.c print_unreferenced(seq, object); object 1687 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1706 mm/kmemleak.c struct kmemleak_object *object; object 1711 mm/kmemleak.c object = find_and_get_object(addr, 0); object 1712 mm/kmemleak.c if (!object) { object 1717 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1718 mm/kmemleak.c dump_object_info(object); object 1719 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1721 mm/kmemleak.c put_object(object); object 1733 mm/kmemleak.c struct kmemleak_object *object; object 1737 mm/kmemleak.c list_for_each_entry_rcu(object, &object_list, object_list) { object 1738 mm/kmemleak.c spin_lock_irqsave(&object->lock, flags); object 1739 mm/kmemleak.c if ((object->flags & OBJECT_REPORTED) && object 1740 mm/kmemleak.c unreferenced_object(object)) object 1741 mm/kmemleak.c __paint_it(object, KMEMLEAK_GREY); object 1742 mm/kmemleak.c spin_unlock_irqrestore(&object->lock, flags); object 1845 mm/kmemleak.c struct kmemleak_object *object, *tmp; object 1851 mm/kmemleak.c list_for_each_entry_safe(object, tmp, &object_list, object_list) { object 1852 mm/kmemleak.c __remove_object(object); object 1853 mm/kmemleak.c __delete_object(object); object 278 mm/slub.c static inline void *get_freepointer(struct kmem_cache *s, void *object) object 280 mm/slub.c return freelist_dereference(s, object + s->offset); object 283 mm/slub.c static void prefetch_freepointer(const struct kmem_cache *s, void *object) object 285 mm/slub.c prefetch(object + s->offset); object 288 mm/slub.c static inline void *get_freepointer_safe(struct kmem_cache *s, void *object) object 294 mm/slub.c return get_freepointer(s, object); object 296 mm/slub.c freepointer_addr = (unsigned long)object + s->offset; object 301 mm/slub.c static inline void set_freepointer(struct kmem_cache *s, void *object, void *fp) object 303 mm/slub.c unsigned long freeptr_addr = (unsigned long)object + s->offset; object 306 mm/slub.c BUG_ON(object == fp); /* naive detection of double free or corruption */ object 509 mm/slub.c struct page *page, void *object) object 513 mm/slub.c if (!object) object 517 mm/slub.c object = kasan_reset_tag(object); object 518 mm/slub.c object = restore_red_left(s, object); object 519 mm/slub.c if (object < base || object >= base + page->objects * s->size || object 520 mm/slub.c (object - base) % s->size) { object 536 mm/slub.c static struct track *get_track(struct kmem_cache *s, void *object, object 542 mm/slub.c p = object + s->offset + sizeof(void *); object 544 mm/slub.c p = object + s->inuse; object 549 mm/slub.c static void set_track(struct kmem_cache *s, void *object, object 552 mm/slub.c struct track *p = get_track(s, object, alloc); object 574 mm/slub.c static void init_tracking(struct kmem_cache *s, void *object) object 579 mm/slub.c set_track(s, object, TRACK_FREE, 0UL); object 580 mm/slub.c set_track(s, object, TRACK_ALLOC, 0UL); object 602 mm/slub.c static void print_tracking(struct kmem_cache *s, void *object) object 608 mm/slub.c print_track("Allocated", get_track(s, object, TRACK_ALLOC), pr_time); object 609 mm/slub.c print_track("Freed", get_track(s, object, TRACK_FREE), pr_time); object 690 mm/slub.c u8 *object, char *reason) object 693 mm/slub.c print_trailer(s, page, object); object 710 mm/slub.c static void init_object(struct kmem_cache *s, void *object, u8 val) object 712 mm/slub.c u8 *p = object; object 734 mm/slub.c u8 *object, char *what, object 753 mm/slub.c print_trailer(s, page, object); object 855 mm/slub.c void *object, u8 val) object 857 mm/slub.c u8 *p = object; object 858 mm/slub.c u8 *endobject = object + s->object_size; object 861 mm/slub.c if (!check_bytes_and_report(s, page, object, "Redzone", object 862 mm/slub.c object - s->red_left_pad, val, s->red_left_pad)) object 865 mm/slub.c if (!check_bytes_and_report(s, page, object, "Redzone", object 945 mm/slub.c void *object = NULL; object 953 mm/slub.c if (object) { object 954 mm/slub.c object_err(s, page, object, object 956 mm/slub.c set_freepointer(s, object, NULL); object 966 mm/slub.c object = fp; object 967 mm/slub.c fp = get_freepointer(s, object); object 990 mm/slub.c static void trace(struct kmem_cache *s, struct page *page, void *object, object 997 mm/slub.c object, page->inuse, object 1001 mm/slub.c print_section(KERN_INFO, "Object ", (void *)object, object 1068 mm/slub.c void *object) object 1073 mm/slub.c init_object(s, object, SLUB_RED_INACTIVE); object 1074 mm/slub.c init_tracking(s, object); object 1089 mm/slub.c struct page *page, void *object) object 1094 mm/slub.c if (!check_valid_pointer(s, page, object)) { object 1095 mm/slub.c object_err(s, page, object, "Freelist Pointer check fails"); object 1099 mm/slub.c if (!check_object(s, page, object, SLUB_RED_INACTIVE)) object 1107 mm/slub.c void *object, unsigned long addr) object 1110 mm/slub.c if (!alloc_consistency_checks(s, page, object)) object 1116 mm/slub.c set_track(s, object, TRACK_ALLOC, addr); object 1117 mm/slub.c trace(s, page, object, 1); object 1118 mm/slub.c init_object(s, object, SLUB_RED_ACTIVE); object 1136 mm/slub.c struct page *page, void *object, unsigned long addr) object 1138 mm/slub.c if (!check_valid_pointer(s, page, object)) { object 1139 mm/slub.c slab_err(s, page, "Invalid object pointer 0x%p", object); object 1143 mm/slub.c if (on_freelist(s, page, object)) { object 1144 mm/slub.c object_err(s, page, object, "Object already free"); object 1148 mm/slub.c if (!check_object(s, page, object, SLUB_RED_ACTIVE)) object 1154 mm/slub.c object); object 1157 mm/slub.c object); object 1160 mm/slub.c object_err(s, page, object, object 1174 mm/slub.c void *object = head; object 1191 mm/slub.c if (!free_consistency_checks(s, page, object, addr)) object 1196 mm/slub.c set_track(s, object, TRACK_FREE, addr); object 1197 mm/slub.c trace(s, page, object, 0); object 1199 mm/slub.c init_object(s, object, SLUB_RED_INACTIVE); object 1202 mm/slub.c if (object != tail) { object 1203 mm/slub.c object = get_freepointer(s, object); object 1216 mm/slub.c slab_fix(s, "Object at 0x%p not freed", object); object 1343 mm/slub.c struct page *page, void *object) {} object 1348 mm/slub.c struct page *page, void *object, unsigned long addr) { return 0; } object 1358 mm/slub.c void *object, u8 val) { return 1; } object 1431 mm/slub.c void *object; object 1441 mm/slub.c object = next; object 1442 mm/slub.c next = get_freepointer(s, object); object 1449 mm/slub.c memset(object, 0, s->object_size); object 1452 mm/slub.c memset((char *)object + s->inuse, 0, object 1457 mm/slub.c if (!slab_free_hook(s, object)) { object 1459 mm/slub.c set_freepointer(s, object, *head); object 1460 mm/slub.c *head = object; object 1462 mm/slub.c *tail = object; object 1464 mm/slub.c } while (object != old_tail); object 1473 mm/slub.c void *object) object 1475 mm/slub.c setup_object_debug(s, page, object); object 1476 mm/slub.c object = kasan_init_slab_obj(s, object); object 1478 mm/slub.c kasan_unpoison_object_data(s, object); object 1479 mm/slub.c s->ctor(object); object 1480 mm/slub.c kasan_poison_object_data(s, object); object 1482 mm/slub.c return object; object 1839 mm/slub.c void *object = NULL; object 1859 mm/slub.c t = acquire_slab(s, n, page, object == NULL, &objects); object 1864 mm/slub.c if (!object) { object 1867 mm/slub.c object = t; object 1878 mm/slub.c return object; object 1892 mm/slub.c void *object; object 1927 mm/slub.c object = get_partial_node(s, n, c, flags); object 1928 mm/slub.c if (object) { object 1936 mm/slub.c return object; object 1951 mm/slub.c void *object; object 1957 mm/slub.c object = get_partial_node(s, get_node(s, searchnode), c, flags); object 1958 mm/slub.c if (object || node != NUMA_NO_NODE) object 1959 mm/slub.c return object; object 2685 mm/slub.c void *object; object 2727 mm/slub.c object = c->freelist; object 2729 mm/slub.c if (unlikely(!object || !node_match(page, node))) { object 2730 mm/slub.c object = __slab_alloc(s, gfpflags, node, addr, c); object 2733 mm/slub.c void *next_object = get_freepointer_safe(s, object); object 2751 mm/slub.c object, tid, object 2761 mm/slub.c maybe_wipe_obj_freeptr(s, object); object 2763 mm/slub.c if (unlikely(slab_want_init_on_alloc(gfpflags, s)) && object) object 2764 mm/slub.c memset(object, 0, s->object_size); object 2766 mm/slub.c slab_post_alloc_hook(s, gfpflags, 1, &object); object 2768 mm/slub.c return object; object 3061 mm/slub.c void *object; object 3068 mm/slub.c object = p[--size]; object 3070 mm/slub.c } while (!object && size); object 3072 mm/slub.c if (!object) object 3075 mm/slub.c page = virt_to_head_page(object); object 3080 mm/slub.c kfree_hook(object); object 3088 mm/slub.c df->s = cache_from_obj(s, object); /* Support for memcg */ object 3093 mm/slub.c set_freepointer(df->s, object, NULL); object 3094 mm/slub.c df->tail = object; object 3095 mm/slub.c df->freelist = object; object 3100 mm/slub.c object = p[--size]; object 3101 mm/slub.c if (!object) object 3105 mm/slub.c if (df->page == virt_to_head_page(object)) { object 3107 mm/slub.c set_freepointer(df->s, object, df->freelist); object 3108 mm/slub.c df->freelist = object; object 3164 mm/slub.c void *object = c->freelist; object 3166 mm/slub.c if (unlikely(!object)) { object 3190 mm/slub.c c->freelist = get_freepointer(s, object); object 3191 mm/slub.c p[i] = object; object 3936 mm/slub.c size_t __ksize(const void *object) object 3940 mm/slub.c if (unlikely(object == ZERO_SIZE_PTR)) object 3943 mm/slub.c page = virt_to_head_page(object); object 3957 mm/slub.c void *object = (void *)x; object 3969 mm/slub.c kfree_hook(object); object 3975 mm/slub.c slab_free(page->slab_cache, page, object, NULL, 1, _RET_IP_); object 1700 net/ceph/messenger.c int end, int size, void *object) object 1705 net/ceph/messenger.c int ret = ceph_tcp_recvmsg(con->sock, object + have, left); object 53 samples/bpf/fds_example.c static int bpf_prog_create(const char *object) object 64 samples/bpf/fds_example.c if (object) { object 65 samples/bpf/fds_example.c assert(!bpf_prog_load(object, BPF_PROG_TYPE_UNSPEC, object 109 samples/bpf/fds_example.c static int bpf_do_prog(const char *file, uint32_t flags, const char *object) object 114 samples/bpf/fds_example.c fd = bpf_prog_create(object); object 140 samples/bpf/fds_example.c const char *file = NULL, *object = NULL; object 173 samples/bpf/fds_example.c object = optarg; object 187 samples/bpf/fds_example.c return bpf_do_prog(file, flags, object); object 459 scripts/kconfig/gconf.c void on_window1_destroy(GtkObject * object, gpointer user_data) object 130 security/keys/internal.h int (*iterator)(const void *object, void *iterator_data); object 42 security/keys/keyring.c void *object = assoc_array_ptr_to_leaf(x); object 43 security/keys/keyring.c return (struct key *)((unsigned long)object & ~KEYRING_PTR_SUBTYPE); object 303 security/keys/keyring.c static unsigned long keyring_get_object_key_chunk(const void *object, int level) object 305 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 309 security/keys/keyring.c static bool keyring_compare_object(const void *object, const void *data) object 312 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 325 security/keys/keyring.c static int keyring_diff_objects(const void *object, const void *data) object 327 security/keys/keyring.c const struct key *key_a = keyring_ptr_to_key(object); object 387 security/keys/keyring.c static void keyring_free_object(void *object) object 389 security/keys/keyring.c key_put(keyring_ptr_to_key(object)); object 458 security/keys/keyring.c static int keyring_read_iterator(const void *object, void *data) object 461 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 574 security/keys/keyring.c static int keyring_search_iterator(const void *object, void *iterator_data) object 577 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 649 security/keys/keyring.c const void *object; object 651 security/keys/keyring.c object = assoc_array_find(&keyring->keys, object 654 security/keys/keyring.c return object ? ctx->iterator(object, ctx) : 0; object 1100 security/keys/keyring.c const void *object; object 1107 security/keys/keyring.c object = assoc_array_find(&keyring->keys, &keyring_assoc_array_ops, object 1110 security/keys/keyring.c if (object) object 1117 security/keys/keyring.c key = keyring_ptr_to_key(object); object 1187 security/keys/keyring.c static int keyring_detect_cycle_iterator(const void *object, object 1191 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 1684 security/keys/keyring.c static bool keyring_gc_select_iterator(void *object, void *iterator_data) object 1686 security/keys/keyring.c struct key *key = keyring_ptr_to_key(object); object 1695 security/keys/keyring.c static int keyring_gc_check_iterator(const void *object, void *iterator_data) object 1697 security/keys/keyring.c const struct key *key = keyring_ptr_to_key(object); object 276 security/smack/smack.h char *object; object 116 security/smack/smack_access.c int smk_access(struct smack_known *subject, struct smack_known *object, object 137 security/smack/smack_access.c if (object == &smack_known_web || subject == &smack_known_web) object 142 security/smack/smack_access.c if (object == &smack_known_star) object 148 security/smack/smack_access.c if (subject->smk_known == object->smk_known) object 156 security/smack/smack_access.c if (object == &smack_known_floor) object 169 security/smack/smack_access.c may = smk_access_entry(subject->smk_known, object->smk_known, object 191 security/smack/smack_access.c if (object == smack_unconfined) object 200 security/smack/smack_access.c smack_log(subject->smk_known, object->smk_known, object 322 security/smack/smack_access.c audit_log_untrustedstring(ab, sad->object); object 365 security/smack/smack_access.c sad->object = object_label; object 2492 security/smack/smack_lsm.c struct smack_known *object, object 2510 security/smack/smack_lsm.c rc = smk_access(subject, object, MAY_WRITE, &ad); object 2511 security/smack/smack_lsm.c rc = smk_bu_note("IPv6 check", subject, object, MAY_WRITE, rc); object 2619 security/smack/smack_lsm.c struct smack_known *object; object 2623 security/smack/smack_lsm.c object = ssp->smk_in; object 2626 security/smack/smack_lsm.c object = smack_ipv6host_label(address); object 2632 security/smack/smack_lsm.c if (skp != NULL && object != NULL) object 2633 security/smack/smack_lsm.c return smk_ipv6_check(skp, object, address, act); object 2636 security/smack/smack_lsm.c if (object == NULL) object 2637 security/smack/smack_lsm.c object = smack_net_ambient; object 2643 security/smack/smack_lsm.c return smk_ipv6_check(skp, object, address, act); object 2656 security/smack/smack_lsm.c object = spp->smk_in; object 2663 security/smack/smack_lsm.c return smk_ipv6_check(skp, object, address, act); object 314 security/smack/smackfs.c static int smk_fill_rule(const char *subject, const char *object, object 327 security/smack/smackfs.c rule->smk_object = smk_import_entry(object, len); object 340 security/smack/smackfs.c cp = smk_parse_smack(object, len); object 325 sound/pci/asihpi/hpi6000.c if (phm->object != HPI_OBJ_SUBSYSTEM) { object 328 sound/pci/asihpi/hpi6000.c hpi_init_response(phr, phm->object, phm->function, object 337 sound/pci/asihpi/hpi6000.c hpi_init_response(phr, phm->object, phm->function, object 346 sound/pci/asihpi/hpi6000.c hpi_init_response(phr, phm->object, phm->function, object 351 sound/pci/asihpi/hpi6000.c switch (phm->object) { object 531 sound/pci/asihpi/hpi6000.c hm.object = HPI_OBJ_ADAPTER; object 1710 sound/pci/asihpi/hpi6000.c switch (phm->object) { object 1747 sound/pci/asihpi/hpi6000.c hm.object = phm->u.d.u.stream.object_type; object 377 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, object 379 sound/pci/asihpi/hpi6205.c HPI_DEBUG_LOG(WARNING, " %d,%d dsp crashed.\n", phm->object, object 391 sound/pci/asihpi/hpi6205.c switch (phm->object) { object 428 sound/pci/asihpi/hpi6205.c if (phm->object != HPI_OBJ_SUBSYSTEM) { object 440 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, object 633 sound/pci/asihpi/hpi6205.c hm.object = HPI_OBJ_ADAPTER; object 737 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 906 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 978 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 1023 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 1191 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 1240 sound/pci/asihpi/hpi6205.c hpi_init_response(phr, phm->object, phm->function, 0); object 991 sound/pci/asihpi/hpi_internal.h u16 object; /* HPI_OBJ_* */ object 1002 sound/pci/asihpi/hpi_internal.h u16 object; /* HPI_OBJ_* */ object 1053 sound/pci/asihpi/hpi_internal.h u16 object; /* HPI_OBJ_* */ object 1064 sound/pci/asihpi/hpi_internal.h u16 object; /* HPI_OBJ_* */ object 42 sound/pci/asihpi/hpicmn.c if (phr->object != phm->object) { object 44 sound/pci/asihpi/hpicmn.c phr->object); object 520 sound/pci/asihpi/hpicmn.c phr->object = phm->object; object 700 sound/pci/asihpi/hpicmn.c switch (phm->object) { object 27 sound/pci/asihpi/hpimsginit.c static void hpi_init_message(struct hpi_message *phm, u16 object, object 32 sound/pci/asihpi/hpimsginit.c if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { object 33 sound/pci/asihpi/hpimsginit.c object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); object 34 sound/pci/asihpi/hpimsginit.c size = msg_size[object]; object 46 sound/pci/asihpi/hpimsginit.c phm->object = object; object 56 sound/pci/asihpi/hpimsginit.c void hpi_init_response(struct hpi_response *phr, u16 object, u16 function, object 61 sound/pci/asihpi/hpimsginit.c if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { object 62 sound/pci/asihpi/hpimsginit.c object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); object 63 sound/pci/asihpi/hpimsginit.c size = res_size[object]; object 71 sound/pci/asihpi/hpimsginit.c phr->object = object; object 79 sound/pci/asihpi/hpimsginit.c struct hpi_response *phr, u16 object, u16 function) object 81 sound/pci/asihpi/hpimsginit.c hpi_init_message(phm, object, function); object 84 sound/pci/asihpi/hpimsginit.c hpi_init_response(phr, object, function, object 89 sound/pci/asihpi/hpimsginit.c u16 object, u16 function) object 92 sound/pci/asihpi/hpimsginit.c if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { object 95 sound/pci/asihpi/hpimsginit.c phm->object = object; object 103 sound/pci/asihpi/hpimsginit.c u16 object, u16 function) object 105 sound/pci/asihpi/hpimsginit.c (void)object; object 115 sound/pci/asihpi/hpimsginit.c struct hpi_response_header *phr, u16 res_size, u16 object, object 118 sound/pci/asihpi/hpimsginit.c hpi_init_messageV1(phm, msg_size, object, function); object 119 sound/pci/asihpi/hpimsginit.c hpi_init_responseV1(phr, res_size, object, function); object 22 sound/pci/asihpi/hpimsginit.h void hpi_init_response(struct hpi_response *phr, u16 object, u16 function, object 26 sound/pci/asihpi/hpimsginit.h struct hpi_response *phr, u16 object, u16 function); object 29 sound/pci/asihpi/hpimsginit.h u16 object, u16 function); object 32 sound/pci/asihpi/hpimsginit.h struct hpi_response_header *phr, u16 res_size, u16 object, object 68 sound/pci/asihpi/hpimsgx.c hpi_init_response(phr, phm->object, phm->function, object 310 sound/pci/asihpi/hpimsgx.c hpi_init_response(phr, phm->object, phm->function, object 317 sound/pci/asihpi/hpimsgx.c hpi_init_response(phr, phm->object, phm->function, object 322 sound/pci/asihpi/hpimsgx.c switch (phm->object) { object 703 sound/pci/asihpi/hpimsgx.c hpi_init_response(&hr, phm->object, phm->function, object 62 sound/pci/asihpi/hpioctl.c && (phm->object != HPI_OBJ_SUBSYSTEM)) object 173 sound/pci/asihpi/hpioctl.c if (hm->h.object == HPI_OBJ_SUBSYSTEM) { object 187 sound/pci/asihpi/hpioctl.c hpi_init_response(&hr->r0, hm->h.object, object 601 sound/soc/sof/topology.c int (*get_token)(void *elem, void *object, u32 offset, u32 size); object 606 sound/soc/sof/topology.c static int get_token_u32(void *elem, void *object, u32 offset, u32 size) object 609 sound/soc/sof/topology.c u32 *val = (u32 *)((u8 *)object + offset); object 615 sound/soc/sof/topology.c static int get_token_u16(void *elem, void *object, u32 offset, u32 size) object 618 sound/soc/sof/topology.c u16 *val = (u16 *)((u8 *)object + offset); object 624 sound/soc/sof/topology.c static int get_token_comp_format(void *elem, void *object, u32 offset, u32 size) object 627 sound/soc/sof/topology.c u32 *val = (u32 *)((u8 *)object + offset); object 633 sound/soc/sof/topology.c static int get_token_dai_type(void *elem, void *object, u32 offset, u32 size) object 636 sound/soc/sof/topology.c u32 *val = (u32 *)((u8 *)object + offset); object 642 sound/soc/sof/topology.c static int get_token_process_type(void *elem, void *object, u32 offset, object 646 sound/soc/sof/topology.c u32 *val = (u32 *)((u8 *)object + offset); object 844 sound/soc/sof/topology.c void *object, object 867 sound/soc/sof/topology.c tokens[j].get_token(elem, object, tokens[j].offset, object 874 sound/soc/sof/topology.c void *object, object 897 sound/soc/sof/topology.c tokens[j].get_token(elem, object, tokens[j].offset, object 904 sound/soc/sof/topology.c void *object, object 968 sound/soc/sof/topology.c tokens[j].get_token(elem, object, object 976 sound/soc/sof/topology.c void *object, object 1006 sound/soc/sof/topology.c sof_parse_uuid_tokens(scomp, object, tokens, count, object 1010 sound/soc/sof/topology.c sof_parse_string_tokens(scomp, object, tokens, count, object 1017 sound/soc/sof/topology.c sof_parse_word_tokens(scomp, object, tokens, count, object 89 tools/lib/bpf/libbpf.h LIBBPF_API int bpf_object__pin(struct bpf_object *object, const char *path); object 90 tools/lib/bpf/libbpf.h LIBBPF_API void bpf_object__close(struct bpf_object *object); object 683 tools/power/x86/intel-speed-select/isst-config.c #define _get_tdp_level(desc, suffix, object, help) \ object 684 tools/power/x86/intel-speed-select/isst-config.c static void get_tdp_##object(void) \ object 698 tools/power/x86/intel-speed-select/isst-config.c &ctdp, desc, &ctdp.object); \ object 703 tools/power/x86/intel-speed-select/isst-config.c &ctdp.object); \ object 199 tools/usb/usbip/libsrc/usbip_common.c #define READ_ATTR(object, type, dev, name, format) \ object 201 tools/usb/usbip/libsrc/usbip_common.c (object)->name = (type) read_attr_value(dev, to_string(name), \