pd               1447 arch/alpha/kernel/setup.c 	struct platform_device *pd;
pd               1450 arch/alpha/kernel/setup.c 	pd = platform_device_alloc("pcspkr", -1);
pd               1451 arch/alpha/kernel/setup.c 	if (!pd)
pd               1454 arch/alpha/kernel/setup.c 	ret = platform_device_add(pd);
pd               1456 arch/alpha/kernel/setup.c 		platform_device_put(pd);
pd                805 arch/arm/common/sa1111.c 	struct sa1111_platform_data *pd = me->platform_data;
pd                811 arch/arm/common/sa1111.c 	if (!pd)
pd                831 arch/arm/common/sa1111.c 	sachip->pdata = pd;
pd                867 arch/arm/common/sa1111.c 	ret = sa1111_setup_irq(sachip, pd->irq_base);
pd                909 arch/arm/common/sa1111.c 	if (pd)
pd                910 arch/arm/common/sa1111.c 		has_devs &= ~pd->disable_devs;
pd                 91 arch/arm/mach-imx/mach-mx35_3ds.c 				struct plat_lcd_data *pd, unsigned int power)
pd               1070 arch/arm/mach-pxa/devices.c 	struct platform_device *pd;
pd               1072 arch/arm/mach-pxa/devices.c 	pd = platform_device_alloc("pxa2xx-spi", id);
pd               1073 arch/arm/mach-pxa/devices.c 	if (pd == NULL) {
pd               1079 arch/arm/mach-pxa/devices.c 	pd->dev.platform_data = info;
pd               1080 arch/arm/mach-pxa/devices.c 	platform_device_add(pd);
pd                532 arch/arm/mach-pxa/hx4700.c static void hx4700_lcd_set_power(struct plat_lcd_data *pd, unsigned int power)
pd                 38 arch/arm/mach-rockchip/platsmp.c static int pmu_power_domain_is_on(int pd)
pd                 47 arch/arm/mach-rockchip/platsmp.c 	return !(val & BIT(pd));
pd                 64 arch/arm/mach-rockchip/platsmp.c static int pmu_set_power_domain(int pd, bool on)
pd                 66 arch/arm/mach-rockchip/platsmp.c 	u32 val = (on) ? 0 : BIT(pd);
pd                 67 arch/arm/mach-rockchip/platsmp.c 	struct reset_control *rstc = rockchip_get_core_reset(pd);
pd                 72 arch/arm/mach-rockchip/platsmp.c 		       __func__, pd);
pd                 85 arch/arm/mach-rockchip/platsmp.c 		ret = regmap_update_bits(pmu, PMU_PWRDN_CON, BIT(pd), val);
pd                 94 arch/arm/mach-rockchip/platsmp.c 			ret = pmu_power_domain_is_on(pd);
pd                534 arch/arm/mach-s3c24xx/mach-h1940.c static void h1940_lcd_power_set(struct plat_lcd_data *pd,
pd                103 arch/arm/mach-s3c64xx/mach-anw6410.c static void anw6410_lcd_power_set(struct plat_lcd_data *pd,
pd                129 arch/arm/mach-s3c64xx/mach-crag6410.c static void crag6410_lcd_power_set(struct plat_lcd_data *pd, unsigned int power)
pd                197 arch/arm/mach-s3c64xx/mach-mini6410.c static void mini6410_lcd_power_set(struct plat_lcd_data *pd,
pd                233 arch/arm/mach-s3c64xx/mach-smartq.c static void smartq_lcd_power_set(struct plat_lcd_data *pd, unsigned int power)
pd                118 arch/arm/mach-s3c64xx/mach-smdk6410.c static void smdk6410_lcd_power_set(struct plat_lcd_data *pd,
pd                 38 arch/arm/mach-s3c64xx/pm.c 	struct generic_pm_domain pd;
pd                 43 arch/arm/mach-s3c64xx/pm.c 	struct s3c64xx_pm_domain *pd;
pd                 46 arch/arm/mach-s3c64xx/pm.c 	pd = container_of(domain, struct s3c64xx_pm_domain, pd);
pd                 49 arch/arm/mach-s3c64xx/pm.c 	val &= ~(pd->ena);
pd                 57 arch/arm/mach-s3c64xx/pm.c 	struct s3c64xx_pm_domain *pd;
pd                 61 arch/arm/mach-s3c64xx/pm.c 	pd = container_of(domain, struct s3c64xx_pm_domain, pd);
pd                 64 arch/arm/mach-s3c64xx/pm.c 	val |= pd->ena;
pd                 68 arch/arm/mach-s3c64xx/pm.c 	if (pd->pwr_stat) {
pd                 71 arch/arm/mach-s3c64xx/pm.c 			if (__raw_readl(S3C64XX_BLK_PWR_STAT) & pd->pwr_stat)
pd                 76 arch/arm/mach-s3c64xx/pm.c 			pr_err("Failed to start domain %s\n", pd->name);
pd                 87 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                 97 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                107 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                117 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                127 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                137 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                146 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                156 arch/arm/mach-s3c64xx/pm.c 	.pd = {
pd                315 arch/arm/mach-s3c64xx/pm.c 		pm_genpd_init(&s3c64xx_always_on_pm_domains[i]->pd,
pd                319 arch/arm/mach-s3c64xx/pm.c 		pm_genpd_init(&s3c64xx_pm_domains[i]->pd, NULL, false);
pd                323 arch/arm/mach-s3c64xx/pm.c 		pm_genpd_add_device(&s3c64xx_pm_f.pd, &s3c_device_fb.dev);
pd                160 arch/arm/plat-samsung/devs.c void __init s3c_fb_set_platdata(struct s3c_fb_platdata *pd)
pd                162 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(pd, sizeof(struct s3c_fb_platdata),
pd                176 arch/arm/plat-samsung/devs.c void __init s3c_hwmon_set_platdata(struct s3c_hwmon_pdata *pd)
pd                178 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(pd, sizeof(struct s3c_hwmon_pdata),
pd                209 arch/arm/plat-samsung/devs.c void s3c_sdhci0_set_platdata(struct s3c_sdhci_platdata *pd)
pd                211 arch/arm/plat-samsung/devs.c 	s3c_sdhci_set_platdata(pd, &s3c_hsmmc0_def_platdata);
pd                239 arch/arm/plat-samsung/devs.c void s3c_sdhci1_set_platdata(struct s3c_sdhci_platdata *pd)
pd                241 arch/arm/plat-samsung/devs.c 	s3c_sdhci_set_platdata(pd, &s3c_hsmmc1_def_platdata);
pd                271 arch/arm/plat-samsung/devs.c void s3c_sdhci2_set_platdata(struct s3c_sdhci_platdata *pd)
pd                273 arch/arm/plat-samsung/devs.c 	s3c_sdhci_set_platdata(pd, &s3c_hsmmc2_def_platdata);
pd                301 arch/arm/plat-samsung/devs.c void s3c_sdhci3_set_platdata(struct s3c_sdhci_platdata *pd)
pd                303 arch/arm/plat-samsung/devs.c 	s3c_sdhci_set_platdata(pd, &s3c_hsmmc3_def_platdata);
pd                328 arch/arm/plat-samsung/devs.c void __init s3c_i2c0_set_platdata(struct s3c2410_platform_i2c *pd)
pd                332 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                333 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                334 arch/arm/plat-samsung/devs.c 		pd->bus_num = 0;
pd                337 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c0);
pd                356 arch/arm/plat-samsung/devs.c void __init s3c_i2c1_set_platdata(struct s3c2410_platform_i2c *pd)
pd                360 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                361 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                362 arch/arm/plat-samsung/devs.c 		pd->bus_num = 1;
pd                365 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c1);
pd                385 arch/arm/plat-samsung/devs.c void __init s3c_i2c2_set_platdata(struct s3c2410_platform_i2c *pd)
pd                389 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                390 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                391 arch/arm/plat-samsung/devs.c 		pd->bus_num = 2;
pd                394 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c2);
pd                414 arch/arm/plat-samsung/devs.c void __init s3c_i2c3_set_platdata(struct s3c2410_platform_i2c *pd)
pd                418 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                419 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                420 arch/arm/plat-samsung/devs.c 		pd->bus_num = 3;
pd                423 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c3);
pd                443 arch/arm/plat-samsung/devs.c void __init s3c_i2c4_set_platdata(struct s3c2410_platform_i2c *pd)
pd                447 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                448 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                449 arch/arm/plat-samsung/devs.c 		pd->bus_num = 4;
pd                452 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c4);
pd                472 arch/arm/plat-samsung/devs.c void __init s3c_i2c5_set_platdata(struct s3c2410_platform_i2c *pd)
pd                476 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                477 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                478 arch/arm/plat-samsung/devs.c 		pd->bus_num = 5;
pd                481 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c5);
pd                501 arch/arm/plat-samsung/devs.c void __init s3c_i2c6_set_platdata(struct s3c2410_platform_i2c *pd)
pd                505 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                506 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                507 arch/arm/plat-samsung/devs.c 		pd->bus_num = 6;
pd                510 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c6);
pd                530 arch/arm/plat-samsung/devs.c void __init s3c_i2c7_set_platdata(struct s3c2410_platform_i2c *pd)
pd                534 arch/arm/plat-samsung/devs.c 	if (!pd) {
pd                535 arch/arm/plat-samsung/devs.c 		pd = &default_i2c_data;
pd                536 arch/arm/plat-samsung/devs.c 		pd->bus_num = 7;
pd                539 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c7);
pd                601 arch/arm/plat-samsung/devs.c void __init samsung_keypad_set_platdata(struct samsung_keypad_platdata *pd)
pd                605 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &samsung_device_keypad);
pd                631 arch/arm/plat-samsung/devs.c void __init s3c24xx_fb_set_platdata(struct s3c2410fb_mach_info *pd)
pd                635 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_lcd);
pd                637 arch/arm/plat-samsung/devs.c 		npd->displays = kmemdup(pd->displays,
pd                795 arch/arm/plat-samsung/devs.c void __init samsung_pwm_set_platdata(struct samsung_pwm_variant *pd)
pd                797 arch/arm/plat-samsung/devs.c 	samsung_device_pwm.dev.platform_data = pd;
pd                921 arch/arm/plat-samsung/devs.c void __init s3c64xx_ts_set_platdata(struct s3c2410_ts_mach_info *pd)
pd                923 arch/arm/plat-samsung/devs.c 	if (!pd)
pd                924 arch/arm/plat-samsung/devs.c 		pd = &default_ts_data;
pd                926 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(pd, sizeof(struct s3c2410_ts_mach_info),
pd                981 arch/arm/plat-samsung/devs.c void __init s3c24xx_udc_set_platdata(struct s3c2410_udc_mach_info *pd)
pd                983 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(pd, sizeof(*pd), &s3c_device_usbgadget);
pd               1006 arch/arm/plat-samsung/devs.c void __init dwc2_hsotg_set_platdata(struct dwc2_hsotg_plat *pd)
pd               1010 arch/arm/plat-samsung/devs.c 	npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_usb_hsotg);
pd               1038 arch/arm/plat-samsung/devs.c void __init s3c24xx_hsudc_set_platdata(struct s3c24xx_hsudc_platdata *pd)
pd               1040 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(pd, sizeof(*pd), &s3c_device_usb_hsudc);
pd               1080 arch/arm/plat-samsung/devs.c 	struct s3c64xx_spi_info pd;
pd               1088 arch/arm/plat-samsung/devs.c 	pd.num_cs = num_cs;
pd               1089 arch/arm/plat-samsung/devs.c 	pd.src_clk_nr = src_clk_nr;
pd               1090 arch/arm/plat-samsung/devs.c 	pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi0_cfg_gpio;
pd               1092 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi0);
pd               1116 arch/arm/plat-samsung/devs.c 	struct s3c64xx_spi_info pd;
pd               1124 arch/arm/plat-samsung/devs.c 	pd.num_cs = num_cs;
pd               1125 arch/arm/plat-samsung/devs.c 	pd.src_clk_nr = src_clk_nr;
pd               1126 arch/arm/plat-samsung/devs.c 	pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi1_cfg_gpio;
pd               1128 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi1);
pd               1152 arch/arm/plat-samsung/devs.c 	struct s3c64xx_spi_info pd;
pd               1160 arch/arm/plat-samsung/devs.c 	pd.num_cs = num_cs;
pd               1161 arch/arm/plat-samsung/devs.c 	pd.src_clk_nr = src_clk_nr;
pd               1162 arch/arm/plat-samsung/devs.c 	pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi2_cfg_gpio;
pd               1164 arch/arm/plat-samsung/devs.c 	s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi2);
pd                 93 arch/arm/plat-samsung/include/plat/devs.h extern void *s3c_set_platdata(void *pd, size_t pdsize,
pd                 22 arch/arm/plat-samsung/include/plat/fb.h extern void s3c_fb_set_platdata(struct s3c_fb_platdata *pd);
pd                 22 arch/arm/plat-samsung/include/plat/keypad.h extern void samsung_keypad_set_platdata(struct samsung_keypad_platdata *pd);
pd                 14 arch/arm/plat-samsung/include/plat/pwm-core.h extern void samsung_pwm_set_platdata(struct samsung_pwm_variant *pd);
pd                 16 arch/arm/plat-samsung/include/plat/pwm-core.h static inline void samsung_pwm_set_platdata(struct samsung_pwm_variant *pd) { }
pd                 24 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci_set_platdata(struct s3c_sdhci_platdata *pd,
pd                 35 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci0_set_platdata(struct s3c_sdhci_platdata *pd);
pd                 36 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci1_set_platdata(struct s3c_sdhci_platdata *pd);
pd                 37 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci2_set_platdata(struct s3c_sdhci_platdata *pd);
pd                 38 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci3_set_platdata(struct s3c_sdhci_platdata *pd);
pd                 15 arch/arm/plat-samsung/platformdata.c void __init *s3c_set_platdata(void *pd, size_t pdsize,
pd                 20 arch/arm/plat-samsung/platformdata.c 	if (!pd) {
pd                 26 arch/arm/plat-samsung/platformdata.c 	npd = kmemdup(pd, pdsize, GFP_KERNEL);
pd                 34 arch/arm/plat-samsung/platformdata.c void s3c_sdhci_set_platdata(struct s3c_sdhci_platdata *pd,
pd                 37 arch/arm/plat-samsung/platformdata.c 	set->cd_type = pd->cd_type;
pd                 38 arch/arm/plat-samsung/platformdata.c 	set->ext_cd_init = pd->ext_cd_init;
pd                 39 arch/arm/plat-samsung/platformdata.c 	set->ext_cd_cleanup = pd->ext_cd_cleanup;
pd                 40 arch/arm/plat-samsung/platformdata.c 	set->ext_cd_gpio = pd->ext_cd_gpio;
pd                 41 arch/arm/plat-samsung/platformdata.c 	set->ext_cd_gpio_invert = pd->ext_cd_gpio_invert;
pd                 43 arch/arm/plat-samsung/platformdata.c 	if (pd->max_width)
pd                 44 arch/arm/plat-samsung/platformdata.c 		set->max_width = pd->max_width;
pd                 45 arch/arm/plat-samsung/platformdata.c 	if (pd->cfg_gpio)
pd                 46 arch/arm/plat-samsung/platformdata.c 		set->cfg_gpio = pd->cfg_gpio;
pd                 47 arch/arm/plat-samsung/platformdata.c 	if (pd->host_caps)
pd                 48 arch/arm/plat-samsung/platformdata.c 		set->host_caps |= pd->host_caps;
pd                 49 arch/arm/plat-samsung/platformdata.c 	if (pd->host_caps2)
pd                 50 arch/arm/plat-samsung/platformdata.c 		set->host_caps2 |= pd->host_caps2;
pd                 51 arch/arm/plat-samsung/platformdata.c 	if (pd->pm_caps)
pd                 52 arch/arm/plat-samsung/platformdata.c 		set->pm_caps |= pd->pm_caps;
pd                137 arch/microblaze/mm/pgtable.c 	pmd_t *pd;
pd                141 arch/microblaze/mm/pgtable.c 	pd = pmd_offset(pgd_offset_k(va), va);
pd                143 arch/microblaze/mm/pgtable.c 	pg = pte_alloc_kernel(pd, va); /* from powerpc - pgtable.c */
pd                 99 arch/mips/alchemy/devboards/platform.c 	struct platform_device *pd;
pd                111 arch/mips/alchemy/devboards/platform.c 	pd = platform_device_alloc("db1xxx_pcmcia", id);
pd                112 arch/mips/alchemy/devboards/platform.c 	if (!pd) {
pd                153 arch/mips/alchemy/devboards/platform.c 	pd->resource = sr;
pd                154 arch/mips/alchemy/devboards/platform.c 	pd->num_resources = cnt;
pd                156 arch/mips/alchemy/devboards/platform.c 	ret = platform_device_add(pd);
pd                160 arch/mips/alchemy/devboards/platform.c 	platform_device_put(pd);
pd                173 arch/mips/alchemy/devboards/platform.c 	struct platform_device *pd;
pd                194 arch/mips/alchemy/devboards/platform.c 	pd = platform_device_alloc("physmap-flash", 0);
pd                195 arch/mips/alchemy/devboards/platform.c 	if (!pd)
pd                248 arch/mips/alchemy/devboards/platform.c 	pd->dev.platform_data = pfd;
pd                249 arch/mips/alchemy/devboards/platform.c 	pd->resource = res;
pd                250 arch/mips/alchemy/devboards/platform.c 	pd->num_resources = 1;
pd                252 arch/mips/alchemy/devboards/platform.c 	ret = platform_device_add(pd);
pd                256 arch/mips/alchemy/devboards/platform.c 	platform_device_put(pd);
pd                211 arch/mips/bcm63xx/dev-enet.c 				 const struct bcm63xx_enet_platform_data *pd)
pd                247 arch/mips/bcm63xx/dev-enet.c 	memcpy(dpd, pd, sizeof(*pd));
pd                291 arch/mips/bcm63xx/dev-enet.c bcm63xx_enetsw_register(const struct bcm63xx_enetsw_platform_data *pd)
pd                310 arch/mips/bcm63xx/dev-enet.c 	memcpy(bcm63xx_enetsw_device.dev.platform_data, pd, sizeof(*pd));
pd                 57 arch/mips/bcm63xx/dev-pcmcia.c static struct bcm63xx_pcmcia_platform_data pd;
pd                 65 arch/mips/bcm63xx/dev-pcmcia.c 		.platform_data = &pd,
pd                114 arch/mips/bcm63xx/dev-pcmcia.c 		pd.ready_gpio = 22;
pd                118 arch/mips/bcm63xx/dev-pcmcia.c 		pd.ready_gpio = 18;
pd                 36 arch/mips/bcm63xx/dev-usb-usbd.c int __init bcm63xx_usbd_register(const struct bcm63xx_usbd_platform_data *pd)
pd                 62 arch/mips/bcm63xx/dev-usb-usbd.c 	platform_device_add_data(&bcm63xx_usbd_device, pd, sizeof(*pd));
pd                316 arch/mips/cavium-octeon/octeon-platform.c 	struct platform_device *pd;
pd                324 arch/mips/cavium-octeon/octeon-platform.c 	pd = of_find_device_by_node(ehci_node);
pd                326 arch/mips/cavium-octeon/octeon-platform.c 	if (!pd)
pd                329 arch/mips/cavium-octeon/octeon-platform.c 	pd->dev.platform_data = &octeon_ehci_pdata;
pd                330 arch/mips/cavium-octeon/octeon-platform.c 	octeon_ehci_hw_start(&pd->dev);
pd                379 arch/mips/cavium-octeon/octeon-platform.c 	struct platform_device *pd;
pd                387 arch/mips/cavium-octeon/octeon-platform.c 	pd = of_find_device_by_node(ohci_node);
pd                389 arch/mips/cavium-octeon/octeon-platform.c 	if (!pd)
pd                392 arch/mips/cavium-octeon/octeon-platform.c 	pd->dev.platform_data = &octeon_ohci_pdata;
pd                393 arch/mips/cavium-octeon/octeon-platform.c 	octeon_ohci_hw_start(&pd->dev);
pd                404 arch/mips/cavium-octeon/octeon-platform.c 	struct platform_device *pd;
pd                419 arch/mips/cavium-octeon/octeon-platform.c 	pd = platform_device_alloc("octeon_rng", -1);
pd                420 arch/mips/cavium-octeon/octeon-platform.c 	if (!pd) {
pd                425 arch/mips/cavium-octeon/octeon-platform.c 	ret = platform_device_add_resources(pd, rng_resources,
pd                430 arch/mips/cavium-octeon/octeon-platform.c 	ret = platform_device_add(pd);
pd                436 arch/mips/cavium-octeon/octeon-platform.c 	platform_device_put(pd);
pd                 20 arch/mips/cobalt/buttons.c 	struct platform_device *pd;
pd                 23 arch/mips/cobalt/buttons.c 	pd = platform_device_alloc("Cobalt buttons", -1);
pd                 24 arch/mips/cobalt/buttons.c 	if (!pd)
pd                 27 arch/mips/cobalt/buttons.c 	error = platform_device_add_resources(pd, &cobalt_buttons_resource, 1);
pd                 31 arch/mips/cobalt/buttons.c 	error = platform_device_add(pd);
pd                 38 arch/mips/cobalt/buttons.c 	platform_device_put(pd);
pd                103 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_enet.h 				 const struct bcm63xx_enet_platform_data *pd);
pd                105 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_enet.h int bcm63xx_enetsw_register(const struct bcm63xx_enetsw_platform_data *pd);
pd                 16 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_usb_usbd.h int bcm63xx_usbd_register(const struct bcm63xx_usbd_platform_data *pd);
pd                 90 arch/mips/netlogic/common/irq.c 	struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d);
pd                 92 arch/mips/netlogic/common/irq.c 	BUG_ON(!pd);
pd                 93 arch/mips/netlogic/common/irq.c 	spin_lock_irqsave(&pd->node->piclock, flags);
pd                 94 arch/mips/netlogic/common/irq.c 	nlm_pic_enable_irt(pd->node->picbase, pd->irt);
pd                 95 arch/mips/netlogic/common/irq.c 	spin_unlock_irqrestore(&pd->node->piclock, flags);
pd                100 arch/mips/netlogic/common/irq.c 	struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d);
pd                103 arch/mips/netlogic/common/irq.c 	BUG_ON(!pd);
pd                104 arch/mips/netlogic/common/irq.c 	spin_lock_irqsave(&pd->node->piclock, flags);
pd                105 arch/mips/netlogic/common/irq.c 	nlm_pic_disable_irt(pd->node->picbase, pd->irt);
pd                106 arch/mips/netlogic/common/irq.c 	spin_unlock_irqrestore(&pd->node->piclock, flags);
pd                111 arch/mips/netlogic/common/irq.c 	struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d);
pd                113 arch/mips/netlogic/common/irq.c 	clear_c0_eimr(pd->picirq);
pd                114 arch/mips/netlogic/common/irq.c 	ack_c0_eirr(pd->picirq);
pd                119 arch/mips/netlogic/common/irq.c 	struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d);
pd                121 arch/mips/netlogic/common/irq.c 	BUG_ON(!pd);
pd                123 arch/mips/netlogic/common/irq.c 	if (pd->extra_ack)
pd                124 arch/mips/netlogic/common/irq.c 		pd->extra_ack(d);
pd                127 arch/mips/netlogic/common/irq.c 	set_c0_eimr(pd->picirq);
pd                130 arch/mips/netlogic/common/irq.c 	nlm_pic_ack(pd->node->picbase, pd->irt);
pd                364 arch/mips/pci/pci-alchemy.c 	struct alchemy_pci_platdata *pd = pdev->dev.platform_data;
pd                373 arch/mips/pci/pci-alchemy.c 	if (!pd) {
pd                442 arch/mips/pci/pci-alchemy.c 	if (pd->board_map_irq)
pd                443 arch/mips/pci/pci-alchemy.c 		ctx->board_map_irq = pd->board_map_irq;
pd                445 arch/mips/pci/pci-alchemy.c 	if (pd->board_pci_idsel)
pd                446 arch/mips/pci/pci-alchemy.c 		ctx->board_pci_idsel = pd->board_pci_idsel;
pd                474 arch/mips/pci/pci-alchemy.c 	val &= ~pd->pci_cfg_clr;
pd                475 arch/mips/pci/pci-alchemy.c 	val |= pd->pci_cfg_set;
pd                 54 arch/mips/sgi-ip32/ip32-platform.c 	struct platform_device *pd;
pd                 57 arch/mips/sgi-ip32/ip32-platform.c 	pd = platform_device_alloc("meth", -1);
pd                 58 arch/mips/sgi-ip32/ip32-platform.c 	if (!pd)
pd                 61 arch/mips/sgi-ip32/ip32-platform.c 	ret = platform_device_add(pd);
pd                 63 arch/mips/sgi-ip32/ip32-platform.c 		platform_device_put(pd);
pd                 72 arch/mips/sgi-ip32/ip32-platform.c 	struct platform_device *pd;
pd                 75 arch/mips/sgi-ip32/ip32-platform.c 	pd = platform_device_alloc("sgio2audio", -1);
pd                 76 arch/mips/sgi-ip32/ip32-platform.c 	if (!pd)
pd                 79 arch/mips/sgi-ip32/ip32-platform.c 	ret = platform_device_add(pd);
pd                 81 arch/mips/sgi-ip32/ip32-platform.c 		platform_device_put(pd);
pd                 70 arch/powerpc/include/asm/pgtable-types.h typedef struct { unsigned long pd; } hugepd_t;
pd                 74 arch/powerpc/include/asm/pgtable-types.h 	return x.pd;
pd                575 arch/powerpc/kernel/setup-common.c 	struct platform_device *pd;
pd                583 arch/powerpc/kernel/setup-common.c 	pd = platform_device_alloc("pcspkr", -1);
pd                584 arch/powerpc/kernel/setup-common.c 	if (!pd)
pd                587 arch/powerpc/kernel/setup-common.c 	ret = platform_device_add(pd);
pd                589 arch/powerpc/kernel/setup-common.c 		platform_device_put(pd);
pd                 61 arch/powerpc/mm/pgtable_32.c 	pmd_t *pd;
pd                 66 arch/powerpc/mm/pgtable_32.c 	pd = pmd_offset(pud_offset(pgd_offset_k(va), va), va);
pd                 69 arch/powerpc/mm/pgtable_32.c 		pg = pte_alloc_kernel(pd, va);
pd                 71 arch/powerpc/mm/pgtable_32.c 		pg = early_pte_alloc_kernel(pd, va);
pd                 96 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd = irq_data_get_irq_chip_data(d);
pd                100 arch/powerpc/platforms/ps3/interrupt.c 		pd->thread_id, d->irq);
pd                103 arch/powerpc/platforms/ps3/interrupt.c 	clear_bit(63 - d->irq, &pd->bmp.mask);
pd                104 arch/powerpc/platforms/ps3/interrupt.c 	lv1_did_update_interrupt_mask(pd->ppe_id, pd->thread_id);
pd                117 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd = irq_data_get_irq_chip_data(d);
pd                121 arch/powerpc/platforms/ps3/interrupt.c 		pd->thread_id, d->irq);
pd                124 arch/powerpc/platforms/ps3/interrupt.c 	set_bit(63 - d->irq, &pd->bmp.mask);
pd                125 arch/powerpc/platforms/ps3/interrupt.c 	lv1_did_update_interrupt_mask(pd->ppe_id, pd->thread_id);
pd                138 arch/powerpc/platforms/ps3/interrupt.c 	const struct ps3_private *pd = irq_data_get_irq_chip_data(d);
pd                142 arch/powerpc/platforms/ps3/interrupt.c 	if (!test_bit(63 - d->irq, &pd->ipi_mask))
pd                143 arch/powerpc/platforms/ps3/interrupt.c 		lv1_end_of_interrupt_ext(pd->ppe_id, pd->thread_id, d->irq);
pd                172 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd;
pd                179 arch/powerpc/platforms/ps3/interrupt.c 	pd = &per_cpu(ps3_private, cpu);
pd                193 arch/powerpc/platforms/ps3/interrupt.c 	result = irq_set_chip_data(*virq, pd);
pd                220 arch/powerpc/platforms/ps3/interrupt.c 	const struct ps3_private *pd = irq_get_chip_data(virq);
pd                223 arch/powerpc/platforms/ps3/interrupt.c 		__LINE__, pd->ppe_id, pd->thread_id, virq);
pd                246 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd;
pd                255 arch/powerpc/platforms/ps3/interrupt.c 	pd = irq_get_chip_data(*virq);
pd                259 arch/powerpc/platforms/ps3/interrupt.c 	result = lv1_connect_irq_plug_ext(pd->ppe_id, pd->thread_id, *virq,
pd                290 arch/powerpc/platforms/ps3/interrupt.c 	const struct ps3_private *pd = irq_get_chip_data(virq);
pd                293 arch/powerpc/platforms/ps3/interrupt.c 		__LINE__, pd->ppe_id, pd->thread_id, virq);
pd                297 arch/powerpc/platforms/ps3/interrupt.c 	result = lv1_disconnect_irq_plug_ext(pd->ppe_id, pd->thread_id, virq);
pd                634 arch/powerpc/platforms/ps3/interrupt.c static void _dump_bmp(struct ps3_private* pd, const char* func, int line)
pd                638 arch/powerpc/platforms/ps3/interrupt.c 	spin_lock_irqsave(&pd->bmp_lock, flags);
pd                639 arch/powerpc/platforms/ps3/interrupt.c 	_dump_64_bmp("stat", &pd->bmp.status, pd->thread_id, func, line);
pd                640 arch/powerpc/platforms/ps3/interrupt.c 	_dump_64_bmp("mask", (u64*)&pd->bmp.mask, pd->thread_id, func, line);
pd                641 arch/powerpc/platforms/ps3/interrupt.c 	spin_unlock_irqrestore(&pd->bmp_lock, flags);
pd                645 arch/powerpc/platforms/ps3/interrupt.c static void __maybe_unused _dump_mask(struct ps3_private *pd,
pd                650 arch/powerpc/platforms/ps3/interrupt.c 	spin_lock_irqsave(&pd->bmp_lock, flags);
pd                651 arch/powerpc/platforms/ps3/interrupt.c 	_dump_64_bmp("mask", (u64*)&pd->bmp.mask, pd->thread_id, func, line);
pd                652 arch/powerpc/platforms/ps3/interrupt.c 	spin_unlock_irqrestore(&pd->bmp_lock, flags);
pd                655 arch/powerpc/platforms/ps3/interrupt.c static void dump_bmp(struct ps3_private* pd) {};
pd                683 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd = &per_cpu(ps3_private, cpu);
pd                685 arch/powerpc/platforms/ps3/interrupt.c 	set_bit(63 - virq, &pd->ipi_debug_brk_mask);
pd                688 arch/powerpc/platforms/ps3/interrupt.c 		cpu, virq, pd->ipi_debug_brk_mask);
pd                693 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd = &per_cpu(ps3_private, cpu);
pd                695 arch/powerpc/platforms/ps3/interrupt.c 	set_bit(63 - virq, &pd->ipi_mask);
pd                698 arch/powerpc/platforms/ps3/interrupt.c 		cpu, virq, pd->ipi_mask);
pd                703 arch/powerpc/platforms/ps3/interrupt.c 	struct ps3_private *pd = this_cpu_ptr(&ps3_private);
pd                704 arch/powerpc/platforms/ps3/interrupt.c 	u64 x = (pd->bmp.status & pd->bmp.mask);
pd                709 arch/powerpc/platforms/ps3/interrupt.c 	if (x & pd->ipi_debug_brk_mask)
pd                710 arch/powerpc/platforms/ps3/interrupt.c 		x &= pd->ipi_debug_brk_mask;
pd                717 arch/powerpc/platforms/ps3/interrupt.c 			__LINE__, pd->thread_id);
pd                733 arch/powerpc/platforms/ps3/interrupt.c 	if (test_bit(63 - plug, &pd->ipi_mask))
pd                734 arch/powerpc/platforms/ps3/interrupt.c 		lv1_end_of_interrupt_ext(pd->ppe_id, pd->thread_id, plug);
pd                749 arch/powerpc/platforms/ps3/interrupt.c 		struct ps3_private *pd = &per_cpu(ps3_private, cpu);
pd                751 arch/powerpc/platforms/ps3/interrupt.c 		lv1_get_logical_ppe_id(&pd->ppe_id);
pd                752 arch/powerpc/platforms/ps3/interrupt.c 		pd->thread_id = get_hard_smp_processor_id(cpu);
pd                753 arch/powerpc/platforms/ps3/interrupt.c 		spin_lock_init(&pd->bmp_lock);
pd                756 arch/powerpc/platforms/ps3/interrupt.c 			__func__, __LINE__, pd->ppe_id, pd->thread_id,
pd                757 arch/powerpc/platforms/ps3/interrupt.c 			ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
pd                759 arch/powerpc/platforms/ps3/interrupt.c 		result = lv1_configure_irq_state_bitmap(pd->ppe_id,
pd                760 arch/powerpc/platforms/ps3/interrupt.c 			pd->thread_id, ps3_mm_phys_to_lpar(__pa(&pd->bmp)));
pd               1294 arch/powerpc/sysdev/fsl_pci.c 	struct mpc85xx_edac_pci_plat_data pd = {
pd               1303 arch/powerpc/sysdev/fsl_pci.c 						   &pd, sizeof(pd));
pd                 23 arch/powerpc/sysdev/rtc_cmos_setup.c 	struct platform_device *pd;
pd                 62 arch/powerpc/sysdev/rtc_cmos_setup.c 	pd = platform_device_register_simple("rtc_cmos", -1,
pd                 65 arch/powerpc/sysdev/rtc_cmos_setup.c 	return PTR_ERR_OR_ZERO(pd);
pd                 39 arch/s390/include/asm/nmi.h 		u64 pd :  1; /* 01 instruction-processing damage */
pd                 26 arch/s390/include/asm/vx-insn.h 	\opd = 255
pd                 28 arch/s390/include/asm/vx-insn.h 		\opd = 0
pd                 31 arch/s390/include/asm/vx-insn.h 		\opd = 1
pd                 34 arch/s390/include/asm/vx-insn.h 		\opd = 2
pd                 37 arch/s390/include/asm/vx-insn.h 		\opd = 3
pd                 40 arch/s390/include/asm/vx-insn.h 		\opd = 4
pd                 43 arch/s390/include/asm/vx-insn.h 		\opd = 5
pd                 46 arch/s390/include/asm/vx-insn.h 		\opd = 6
pd                 49 arch/s390/include/asm/vx-insn.h 		\opd = 7
pd                 52 arch/s390/include/asm/vx-insn.h 		\opd = 8
pd                 55 arch/s390/include/asm/vx-insn.h 		\opd = 9
pd                 58 arch/s390/include/asm/vx-insn.h 		\opd = 10
pd                 61 arch/s390/include/asm/vx-insn.h 		\opd = 11
pd                 64 arch/s390/include/asm/vx-insn.h 		\opd = 12
pd                 67 arch/s390/include/asm/vx-insn.h 		\opd = 13
pd                 70 arch/s390/include/asm/vx-insn.h 		\opd = 14
pd                 73 arch/s390/include/asm/vx-insn.h 		\opd = 15
pd                 75 arch/s390/include/asm/vx-insn.h 	.if \opd == 255
pd                 76 arch/s390/include/asm/vx-insn.h 		\opd = \gr
pd                 90 arch/s390/include/asm/vx-insn.h 	\opd = 255
pd                 92 arch/s390/include/asm/vx-insn.h 		\opd = 0
pd                 95 arch/s390/include/asm/vx-insn.h 		\opd = 1
pd                 98 arch/s390/include/asm/vx-insn.h 		\opd = 2
pd                101 arch/s390/include/asm/vx-insn.h 		\opd = 3
pd                104 arch/s390/include/asm/vx-insn.h 		\opd = 4
pd                107 arch/s390/include/asm/vx-insn.h 		\opd = 5
pd                110 arch/s390/include/asm/vx-insn.h 		\opd = 6
pd                113 arch/s390/include/asm/vx-insn.h 		\opd = 7
pd                116 arch/s390/include/asm/vx-insn.h 		\opd = 8
pd                119 arch/s390/include/asm/vx-insn.h 		\opd = 9
pd                122 arch/s390/include/asm/vx-insn.h 		\opd = 10
pd                125 arch/s390/include/asm/vx-insn.h 		\opd = 11
pd                128 arch/s390/include/asm/vx-insn.h 		\opd = 12
pd                131 arch/s390/include/asm/vx-insn.h 		\opd = 13
pd                134 arch/s390/include/asm/vx-insn.h 		\opd = 14
pd                137 arch/s390/include/asm/vx-insn.h 		\opd = 15
pd                140 arch/s390/include/asm/vx-insn.h 		\opd = 16
pd                143 arch/s390/include/asm/vx-insn.h 		\opd = 17
pd                146 arch/s390/include/asm/vx-insn.h 		\opd = 18
pd                149 arch/s390/include/asm/vx-insn.h 		\opd = 19
pd                152 arch/s390/include/asm/vx-insn.h 		\opd = 20
pd                155 arch/s390/include/asm/vx-insn.h 		\opd = 21
pd                158 arch/s390/include/asm/vx-insn.h 		\opd = 22
pd                161 arch/s390/include/asm/vx-insn.h 		\opd = 23
pd                164 arch/s390/include/asm/vx-insn.h 		\opd = 24
pd                167 arch/s390/include/asm/vx-insn.h 		\opd = 25
pd                170 arch/s390/include/asm/vx-insn.h 		\opd = 26
pd                173 arch/s390/include/asm/vx-insn.h 		\opd = 27
pd                176 arch/s390/include/asm/vx-insn.h 		\opd = 28
pd                179 arch/s390/include/asm/vx-insn.h 		\opd = 29
pd                182 arch/s390/include/asm/vx-insn.h 		\opd = 30
pd                185 arch/s390/include/asm/vx-insn.h 		\opd = 31
pd                187 arch/s390/include/asm/vx-insn.h 	.if \opd == 255
pd                188 arch/s390/include/asm/vx-insn.h 		\opd = \vxr
pd                361 arch/s390/kernel/nmi.c 	if (mci.pd && !test_cpu_flag(CIF_MCCK_GUEST)) {
pd               1057 arch/sh/boards/mach-ecovec24/setup.c static void __init sh_eth_init(struct sh_eth_plat_data *pd)
pd               1068 arch/sh/boards/mach-ecovec24/setup.c 	for (i = 0; i < sizeof(pd->mac_addr); i++) {
pd               1069 arch/sh/boards/mach-ecovec24/setup.c 		pd->mac_addr[i] = mac_read(a, 0x10 + i);
pd               1076 arch/sh/boards/mach-ecovec24/setup.c static void __init sh_eth_init(struct sh_eth_plat_data *pd)
pd                 19 arch/x86/include/asm/inst.h 	\opd = REG_NUM_INVALID
pd                 21 arch/x86/include/asm/inst.h 	\opd = 0
pd                 24 arch/x86/include/asm/inst.h 	\opd = 1
pd                 27 arch/x86/include/asm/inst.h 	\opd = 2
pd                 30 arch/x86/include/asm/inst.h 	\opd = 3
pd                 33 arch/x86/include/asm/inst.h 	\opd = 4
pd                 36 arch/x86/include/asm/inst.h 	\opd = 5
pd                 39 arch/x86/include/asm/inst.h 	\opd = 6
pd                 42 arch/x86/include/asm/inst.h 	\opd = 7
pd                 46 arch/x86/include/asm/inst.h 	\opd = 8
pd                 49 arch/x86/include/asm/inst.h 	\opd = 9
pd                 52 arch/x86/include/asm/inst.h 	\opd = 10
pd                 55 arch/x86/include/asm/inst.h 	\opd = 11
pd                 58 arch/x86/include/asm/inst.h 	\opd = 12
pd                 61 arch/x86/include/asm/inst.h 	\opd = 13
pd                 64 arch/x86/include/asm/inst.h 	\opd = 14
pd                 67 arch/x86/include/asm/inst.h 	\opd = 15
pd                 73 arch/x86/include/asm/inst.h 	\opd = REG_NUM_INVALID
pd                 76 arch/x86/include/asm/inst.h 	\opd = 0
pd                 79 arch/x86/include/asm/inst.h 	\opd = 1
pd                 82 arch/x86/include/asm/inst.h 	\opd = 2
pd                 85 arch/x86/include/asm/inst.h 	\opd = 3
pd                 88 arch/x86/include/asm/inst.h 	\opd = 4
pd                 91 arch/x86/include/asm/inst.h 	\opd = 5
pd                 94 arch/x86/include/asm/inst.h 	\opd = 6
pd                 97 arch/x86/include/asm/inst.h 	\opd = 7
pd                100 arch/x86/include/asm/inst.h 	\opd = 8
pd                103 arch/x86/include/asm/inst.h 	\opd = 9
pd                106 arch/x86/include/asm/inst.h 	\opd = 10
pd                109 arch/x86/include/asm/inst.h 	\opd = 11
pd                112 arch/x86/include/asm/inst.h 	\opd = 12
pd                115 arch/x86/include/asm/inst.h 	\opd = 13
pd                118 arch/x86/include/asm/inst.h 	\opd = 14
pd                121 arch/x86/include/asm/inst.h 	\opd = 15
pd                127 arch/x86/include/asm/inst.h 	\opd = REG_NUM_INVALID
pd                129 arch/x86/include/asm/inst.h 	\opd = 0
pd                132 arch/x86/include/asm/inst.h 	\opd = 1
pd                135 arch/x86/include/asm/inst.h 	\opd = 2
pd                138 arch/x86/include/asm/inst.h 	\opd = 3
pd                141 arch/x86/include/asm/inst.h 	\opd = 4
pd                144 arch/x86/include/asm/inst.h 	\opd = 5
pd                147 arch/x86/include/asm/inst.h 	\opd = 6
pd                150 arch/x86/include/asm/inst.h 	\opd = 7
pd                153 arch/x86/include/asm/inst.h 	\opd = 8
pd                156 arch/x86/include/asm/inst.h 	\opd = 9
pd                159 arch/x86/include/asm/inst.h 	\opd = 10
pd                162 arch/x86/include/asm/inst.h 	\opd = 11
pd                165 arch/x86/include/asm/inst.h 	\opd = 12
pd                168 arch/x86/include/asm/inst.h 	\opd = 13
pd                171 arch/x86/include/asm/inst.h 	\opd = 14
pd                174 arch/x86/include/asm/inst.h 	\opd = 15
pd                  8 arch/x86/kernel/pcspeaker.c 	struct platform_device *pd;
pd                 10 arch/x86/kernel/pcspeaker.c 	pd = platform_device_register_simple("pcspkr", -1, NULL, 0);
pd                 12 arch/x86/kernel/pcspeaker.c 	return PTR_ERR_OR_ZERO(pd);
pd                 41 arch/x86/kernel/sysfb.c 	struct platform_device *pd;
pd                 64 arch/x86/kernel/sysfb.c 	pd = platform_device_register_resndata(NULL, name, 0,
pd                 66 arch/x86/kernel/sysfb.c 	return PTR_ERR_OR_ZERO(pd);
pd                 63 arch/x86/kernel/sysfb_simplefb.c 	struct platform_device *pd;
pd                108 arch/x86/kernel/sysfb_simplefb.c 	pd = platform_device_register_resndata(NULL, "simple-framebuffer", 0,
pd                110 arch/x86/kernel/sysfb_simplefb.c 	return PTR_ERR_OR_ZERO(pd);
pd               5892 arch/x86/kvm/x86.c static int kernel_pio(struct kvm_vcpu *vcpu, void *pd)
pd               5899 arch/x86/kvm/x86.c 					    vcpu->arch.pio.size, pd);
pd               5903 arch/x86/kvm/x86.c 					     pd);
pd               5906 arch/x86/kvm/x86.c 		pd += vcpu->arch.pio.size;
pd                 90 block/bfq-cgroup.c static u64 blkg_prfill_stat(struct seq_file *sf, struct blkg_policy_data *pd,
pd                 93 block/bfq-cgroup.c 	return __blkg_prfill_u64(sf, pd, bfq_stat_read((void *)pd + off));
pd                279 block/bfq-cgroup.c static struct bfq_group *pd_to_bfqg(struct blkg_policy_data *pd)
pd                281 block/bfq-cgroup.c 	return pd ? container_of(pd, struct bfq_group, pd) : NULL;
pd                286 block/bfq-cgroup.c 	return pd_to_blkg(&bfqg->pd);
pd                473 block/bfq-cgroup.c 	return cpd ? container_of(cpd, struct bfq_group_data, pd) : NULL;
pd                488 block/bfq-cgroup.c 	return &bgd->pd;
pd                520 block/bfq-cgroup.c 	return &bfqg->pd;
pd                523 block/bfq-cgroup.c static void bfq_pd_init(struct blkg_policy_data *pd)
pd                525 block/bfq-cgroup.c 	struct blkcg_gq *blkg = pd_to_blkg(pd);
pd                542 block/bfq-cgroup.c static void bfq_pd_free(struct blkg_policy_data *pd)
pd                544 block/bfq-cgroup.c 	struct bfq_group *bfqg = pd_to_bfqg(pd);
pd                550 block/bfq-cgroup.c static void bfq_pd_reset_stats(struct blkg_policy_data *pd)
pd                552 block/bfq-cgroup.c 	struct bfq_group *bfqg = pd_to_bfqg(pd);
pd                858 block/bfq-cgroup.c static void bfq_pd_offline(struct blkg_policy_data *pd)
pd                861 block/bfq-cgroup.c 	struct bfq_group *bfqg = pd_to_bfqg(pd);
pd                949 block/bfq-cgroup.c 				     struct blkg_policy_data *pd, int off)
pd                951 block/bfq-cgroup.c 	struct bfq_group *bfqg = pd_to_bfqg(pd);
pd                955 block/bfq-cgroup.c 	return __blkg_prfill_u64(sf, pd, bfqg->entity.dev_weight);
pd               1102 block/bfq-cgroup.c 				      struct blkg_policy_data *pd, int off)
pd               1104 block/bfq-cgroup.c 	struct blkcg_gq *blkg = pd_to_blkg(pd);
pd               1123 block/bfq-cgroup.c 	return __blkg_prfill_u64(sf, pd, sum);
pd               1127 block/bfq-cgroup.c 					struct blkg_policy_data *pd, int off)
pd               1131 block/bfq-cgroup.c 	blkg_rwstat_recursive_sum(pd_to_blkg(pd), &blkcg_policy_bfq, off, &sum);
pd               1132 block/bfq-cgroup.c 	return __blkg_prfill_rwstat(sf, pd, &sum);
pd               1151 block/bfq-cgroup.c static u64 bfqg_prfill_sectors(struct seq_file *sf, struct blkg_policy_data *pd,
pd               1154 block/bfq-cgroup.c 	u64 sum = blkg_rwstat_total(&pd->blkg->stat_bytes);
pd               1156 block/bfq-cgroup.c 	return __blkg_prfill_u64(sf, pd, sum >> 9);
pd               1167 block/bfq-cgroup.c 					 struct blkg_policy_data *pd, int off)
pd               1171 block/bfq-cgroup.c 	blkg_rwstat_recursive_sum(pd->blkg, NULL,
pd               1174 block/bfq-cgroup.c 	return __blkg_prfill_u64(sf, pd,
pd               1187 block/bfq-cgroup.c 				      struct blkg_policy_data *pd, int off)
pd               1189 block/bfq-cgroup.c 	struct bfq_group *bfqg = pd_to_bfqg(pd);
pd               1197 block/bfq-cgroup.c 	__blkg_prfill_u64(sf, pd, v);
pd                853 block/bfq-iosched.h 	struct blkcg_policy_data pd;
pd                892 block/bfq-iosched.h 	struct blkg_policy_data pd;
pd                 80 block/blk-cgroup.c 		if (blkg->pd[i])
pd                 81 block/blk-cgroup.c 			blkcg_policy[i]->pd_free_fn(blkg->pd[i]);
pd                172 block/blk-cgroup.c 		struct blkg_policy_data *pd;
pd                178 block/blk-cgroup.c 		pd = pol->pd_alloc_fn(gfp_mask, q, blkcg);
pd                179 block/blk-cgroup.c 		if (!pd)
pd                182 block/blk-cgroup.c 		blkg->pd[i] = pd;
pd                183 block/blk-cgroup.c 		pd->blkg = blkg;
pd                184 block/blk-cgroup.c 		pd->plid = i;
pd                278 block/blk-cgroup.c 		if (blkg->pd[i] && pol->pd_init_fn)
pd                279 block/blk-cgroup.c 			pol->pd_init_fn(blkg->pd[i]);
pd                292 block/blk-cgroup.c 			if (blkg->pd[i] && pol->pd_online_fn)
pd                293 block/blk-cgroup.c 				pol->pd_online_fn(blkg->pd[i]);
pd                409 block/blk-cgroup.c 		if (blkg->pd[i] && pol->pd_offline_fn)
pd                410 block/blk-cgroup.c 			pol->pd_offline_fn(blkg->pd[i]);
pd                484 block/blk-cgroup.c 			if (blkg->pd[i] && pol->pd_reset_stats_fn)
pd                485 block/blk-cgroup.c 				pol->pd_reset_stats_fn(blkg->pd[i]);
pd                533 block/blk-cgroup.c 			total += prfill(sf, blkg->pd[pol->plid], data);
pd                551 block/blk-cgroup.c u64 __blkg_prfill_u64(struct seq_file *sf, struct blkg_policy_data *pd, u64 v)
pd                553 block/blk-cgroup.c 	const char *dname = blkg_dev_name(pd->blkg);
pd                571 block/blk-cgroup.c u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd,
pd                581 block/blk-cgroup.c 	const char *dname = blkg_dev_name(pd->blkg);
pd                608 block/blk-cgroup.c u64 blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd,
pd                613 block/blk-cgroup.c 	blkg_rwstat_read((void *)pd + off, &rwstat);
pd                614 block/blk-cgroup.c 	return __blkg_prfill_rwstat(sf, pd, &rwstat);
pd                619 block/blk-cgroup.c 				    struct blkg_policy_data *pd, int off)
pd                623 block/blk-cgroup.c 	blkg_rwstat_read((void *)pd->blkg + off, &rwstat);
pd                624 block/blk-cgroup.c 	return __blkg_prfill_rwstat(sf, pd, &rwstat);
pd                662 block/blk-cgroup.c 					      struct blkg_policy_data *pd,
pd                667 block/blk-cgroup.c 	blkg_rwstat_recursive_sum(pd->blkg, NULL, off, &rwstat);
pd                668 block/blk-cgroup.c 	return __blkg_prfill_rwstat(sf, pd, &rwstat);
pd                986 block/blk-cgroup.c 			if (!blkg->pd[i] || !pol->pd_stat_fn)
pd                989 block/blk-cgroup.c 			written = pol->pd_stat_fn(blkg->pd[i], buf+off, size-off);
pd               1381 block/blk-cgroup.c 		struct blkg_policy_data *pd;
pd               1383 block/blk-cgroup.c 		if (blkg->pd[pol->plid])
pd               1388 block/blk-cgroup.c 			pd = pd_prealloc;
pd               1391 block/blk-cgroup.c 			pd = pol->pd_alloc_fn(GFP_NOWAIT | __GFP_NOWARN, q,
pd               1395 block/blk-cgroup.c 		if (!pd) {
pd               1417 block/blk-cgroup.c 		blkg->pd[pol->plid] = pd;
pd               1418 block/blk-cgroup.c 		pd->blkg = blkg;
pd               1419 block/blk-cgroup.c 		pd->plid = pol->plid;
pd               1425 block/blk-cgroup.c 			pol->pd_init_fn(blkg->pd[pol->plid]);
pd               1444 block/blk-cgroup.c 		if (blkg->pd[pol->plid]) {
pd               1445 block/blk-cgroup.c 			pol->pd_free_fn(blkg->pd[pol->plid]);
pd               1446 block/blk-cgroup.c 			blkg->pd[pol->plid] = NULL;
pd               1479 block/blk-cgroup.c 		if (blkg->pd[pol->plid]) {
pd               1481 block/blk-cgroup.c 				pol->pd_offline_fn(blkg->pd[pol->plid]);
pd               1482 block/blk-cgroup.c 			pol->pd_free_fn(blkg->pd[pol->plid]);
pd               1483 block/blk-cgroup.c 			blkg->pd[pol->plid] = NULL;
pd                431 block/blk-iocost.c 	struct blkg_policy_data		pd;
pd                634 block/blk-iocost.c static struct ioc_gq *pd_to_iocg(struct blkg_policy_data *pd)
pd                636 block/blk-iocost.c 	return pd ? container_of(pd, struct ioc_gq, pd) : NULL;
pd                646 block/blk-iocost.c 	return pd_to_blkg(&iocg->pd);
pd               2013 block/blk-iocost.c 	return &iocg->pd;
pd               2016 block/blk-iocost.c static void ioc_pd_init(struct blkg_policy_data *pd)
pd               2018 block/blk-iocost.c 	struct ioc_gq *iocg = pd_to_iocg(pd);
pd               2019 block/blk-iocost.c 	struct blkcg_gq *blkg = pd_to_blkg(&iocg->pd);
pd               2053 block/blk-iocost.c static void ioc_pd_free(struct blkg_policy_data *pd)
pd               2055 block/blk-iocost.c 	struct ioc_gq *iocg = pd_to_iocg(pd);
pd               2072 block/blk-iocost.c static u64 ioc_weight_prfill(struct seq_file *sf, struct blkg_policy_data *pd,
pd               2075 block/blk-iocost.c 	const char *dname = blkg_dev_name(pd->blkg);
pd               2076 block/blk-iocost.c 	struct ioc_gq *iocg = pd_to_iocg(pd);
pd               2158 block/blk-iocost.c static u64 ioc_qos_prfill(struct seq_file *sf, struct blkg_policy_data *pd,
pd               2161 block/blk-iocost.c 	const char *dname = blkg_dev_name(pd->blkg);
pd               2162 block/blk-iocost.c 	struct ioc *ioc = pd_to_iocg(pd)->ioc;
pd               2330 block/blk-iocost.c 				 struct blkg_policy_data *pd, int off)
pd               2332 block/blk-iocost.c 	const char *dname = blkg_dev_name(pd->blkg);
pd               2333 block/blk-iocost.c 	struct ioc *ioc = pd_to_iocg(pd)->ioc;
pd                134 block/blk-iolatency.c 	struct blkg_policy_data pd;
pd                177 block/blk-iolatency.c static inline struct iolatency_grp *pd_to_lat(struct blkg_policy_data *pd)
pd                179 block/blk-iolatency.c 	return pd ? container_of(pd, struct iolatency_grp, pd) : NULL;
pd                189 block/blk-iolatency.c 	return pd_to_blkg(&iolat->pd);
pd                868 block/blk-iolatency.c 				  struct blkg_policy_data *pd, int off)
pd                870 block/blk-iolatency.c 	struct iolatency_grp *iolat = pd_to_lat(pd);
pd                871 block/blk-iolatency.c 	const char *dname = blkg_dev_name(pd->blkg);
pd                913 block/blk-iolatency.c static size_t iolatency_pd_stat(struct blkg_policy_data *pd, char *buf,
pd                916 block/blk-iolatency.c 	struct iolatency_grp *iolat = pd_to_lat(pd);
pd                952 block/blk-iolatency.c 	return &iolat->pd;
pd                955 block/blk-iolatency.c static void iolatency_pd_init(struct blkg_policy_data *pd)
pd                957 block/blk-iolatency.c 	struct iolatency_grp *iolat = pd_to_lat(pd);
pd               1000 block/blk-iolatency.c static void iolatency_pd_offline(struct blkg_policy_data *pd)
pd               1002 block/blk-iolatency.c 	struct iolatency_grp *iolat = pd_to_lat(pd);
pd               1015 block/blk-iolatency.c static void iolatency_pd_free(struct blkg_policy_data *pd)
pd               1017 block/blk-iolatency.c 	struct iolatency_grp *iolat = pd_to_lat(pd);
pd                108 block/blk-throttle.c 	struct blkg_policy_data pd;
pd                227 block/blk-throttle.c static inline struct throtl_grp *pd_to_tg(struct blkg_policy_data *pd)
pd                229 block/blk-throttle.c 	return pd ? container_of(pd, struct throtl_grp, pd) : NULL;
pd                239 block/blk-throttle.c 	return pd_to_blkg(&tg->pd);
pd                515 block/blk-throttle.c 	return &tg->pd;
pd                518 block/blk-throttle.c static void throtl_pd_init(struct blkg_policy_data *pd)
pd                520 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd                562 block/blk-throttle.c static void throtl_pd_online(struct blkg_policy_data *pd)
pd                564 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd                594 block/blk-throttle.c static void throtl_pd_offline(struct blkg_policy_data *pd)
pd                596 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd                609 block/blk-throttle.c static void throtl_pd_free(struct blkg_policy_data *pd)
pd                611 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd               1330 block/blk-throttle.c static u64 tg_prfill_conf_u64(struct seq_file *sf, struct blkg_policy_data *pd,
pd               1333 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd               1338 block/blk-throttle.c 	return __blkg_prfill_u64(sf, pd, v);
pd               1341 block/blk-throttle.c static u64 tg_prfill_conf_uint(struct seq_file *sf, struct blkg_policy_data *pd,
pd               1344 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd               1349 block/blk-throttle.c 	return __blkg_prfill_u64(sf, pd, v);
pd               1515 block/blk-throttle.c static u64 tg_prfill_limit(struct seq_file *sf, struct blkg_policy_data *pd,
pd               1518 block/blk-throttle.c 	struct throtl_grp *tg = pd_to_tg(pd);
pd               1519 block/blk-throttle.c 	const char *dname = blkg_dev_name(pd->blkg);
pd                 36 drivers/ata/pata_pxa.c 	struct pata_pxa_data *pd = d;
pd                 39 drivers/ata/pata_pxa.c 	status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, NULL);
pd                 41 drivers/ata/pata_pxa.c 		complete(&pd->dma_done);
pd                 49 drivers/ata/pata_pxa.c 	struct pata_pxa_data *pd = qc->ap->private_data;
pd                 57 drivers/ata/pata_pxa.c 	tx = dmaengine_prep_slave_sg(pd->dma_chan, qc->sg, qc->n_elem, dir,
pd                 64 drivers/ata/pata_pxa.c 	tx->callback_param = pd;
pd                 65 drivers/ata/pata_pxa.c 	pd->dma_cookie = dmaengine_submit(tx);
pd                 82 drivers/ata/pata_pxa.c 	struct pata_pxa_data *pd = qc->ap->private_data;
pd                 83 drivers/ata/pata_pxa.c 	init_completion(&pd->dma_done);
pd                 84 drivers/ata/pata_pxa.c 	dma_async_issue_pending(pd->dma_chan);
pd                 92 drivers/ata/pata_pxa.c 	struct pata_pxa_data *pd = qc->ap->private_data;
pd                 95 drivers/ata/pata_pxa.c 	status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, NULL);
pd                 97 drivers/ata/pata_pxa.c 	    wait_for_completion_timeout(&pd->dma_done, HZ))
pd                100 drivers/ata/pata_pxa.c 	dmaengine_terminate_all(pd->dma_chan);
pd                109 drivers/ata/pata_pxa.c 	struct pata_pxa_data *pd = ap->private_data;
pd                114 drivers/ata/pata_pxa.c 	status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, &state);
pd               1427 drivers/base/platform.c 	struct platform_device *pd;
pd               1429 drivers/base/platform.c 	list_for_each_entry(pd, &early_platform_device_list, dev.devres_head)
pd               1430 drivers/base/platform.c 		if (platform_match(&pd->dev, &epdrv->pdrv->driver))
pd               1431 drivers/base/platform.c 			if (pd->id == id)
pd               1432 drivers/base/platform.c 				return pd;
pd               1445 drivers/base/platform.c 	struct platform_device *pd;
pd               1447 drivers/base/platform.c 	list_for_each_entry(pd, &early_platform_device_list, dev.devres_head)
pd               1448 drivers/base/platform.c 		if (platform_match(&pd->dev, &epdrv->pdrv->driver))
pd               1449 drivers/base/platform.c 			if (pd->id >= id)
pd               1582 drivers/base/platform.c 	struct platform_device *pd, *pd2;
pd               1585 drivers/base/platform.c 	list_for_each_entry_safe(pd, pd2, &early_platform_device_list,
pd               1587 drivers/base/platform.c 		list_del(&pd->dev.devres_head);
pd               1588 drivers/base/platform.c 		memset(&pd->dev.devres_head, 0, sizeof(pd->dev.devres_head));
pd                200 drivers/base/power/common.c void dev_pm_domain_set(struct device *dev, struct dev_pm_domain *pd)
pd                202 drivers/base/power/common.c 	if (dev->pm_domain == pd)
pd                205 drivers/base/power/common.c 	WARN(pd && device_is_bound(dev),
pd                207 drivers/base/power/common.c 	dev->pm_domain = pd;
pd               2360 drivers/base/power/domain.c 	struct generic_pm_domain *pd;
pd               2364 drivers/base/power/domain.c 	pd = dev_to_genpd(dev);
pd               2365 drivers/base/power/domain.c 	if (IS_ERR(pd))
pd               2368 drivers/base/power/domain.c 	dev_dbg(dev, "removing from PM domain %s\n", pd->name);
pd               2371 drivers/base/power/domain.c 		ret = genpd_remove_device(pd, dev);
pd               2381 drivers/base/power/domain.c 			pd->name, ret);
pd               2386 drivers/base/power/domain.c 	genpd_queue_power_off_work(pd);
pd               2395 drivers/base/power/domain.c 	struct generic_pm_domain *pd;
pd               2397 drivers/base/power/domain.c 	pd = dev_to_genpd(dev);
pd               2398 drivers/base/power/domain.c 	if (IS_ERR(pd))
pd               2401 drivers/base/power/domain.c 	genpd_queue_power_off_work(pd);
pd               2408 drivers/base/power/domain.c 	struct generic_pm_domain *pd;
pd               2417 drivers/base/power/domain.c 	pd = genpd_get_from_provider(&pd_args);
pd               2419 drivers/base/power/domain.c 	if (IS_ERR(pd)) {
pd               2422 drivers/base/power/domain.c 			__func__, PTR_ERR(pd));
pd               2426 drivers/base/power/domain.c 	dev_dbg(dev, "adding to PM domain %s\n", pd->name);
pd               2428 drivers/base/power/domain.c 	ret = genpd_add_device(pd, dev, base_dev);
pd               2434 drivers/base/power/domain.c 				pd->name, ret);
pd               2442 drivers/base/power/domain.c 		genpd_lock(pd);
pd               2443 drivers/base/power/domain.c 		ret = genpd_power_on(pd, 0);
pd               2444 drivers/base/power/domain.c 		genpd_unlock(pd);
pd               2448 drivers/base/power/domain.c 		genpd_remove_device(pd, dev);
pd                120 drivers/base/power/domain_governor.c static bool __default_power_down_ok(struct dev_pm_domain *pd,
pd                123 drivers/base/power/domain_governor.c 	struct generic_pm_domain *genpd = pd_to_genpd(pd);
pd                209 drivers/base/power/domain_governor.c static bool default_power_down_ok(struct dev_pm_domain *pd)
pd                211 drivers/base/power/domain_governor.c 	struct generic_pm_domain *genpd = pd_to_genpd(pd);
pd                234 drivers/base/power/domain_governor.c 	while (!__default_power_down_ok(pd, genpd->state_idx)) {
pd                252 drivers/base/power/domain_governor.c static bool cpu_power_down_ok(struct dev_pm_domain *pd)
pd                254 drivers/base/power/domain_governor.c 	struct generic_pm_domain *genpd = pd_to_genpd(pd);
pd                261 drivers/base/power/domain_governor.c 	if (!default_power_down_ok(pd))
pd               1383 drivers/block/drbd/drbd_worker.c static int pd_send_unplug_remote(struct drbd_peer_device *pd)
pd               1385 drivers/block/drbd/drbd_worker.c 	struct drbd_socket *sock = &pd->connection->data;
pd               1386 drivers/block/drbd/drbd_worker.c 	if (!drbd_prepare_command(pd, sock))
pd               1388 drivers/block/drbd/drbd_worker.c 	return drbd_send_command(pd, sock, P_UNPLUG_REMOTE, 0, NULL, 0);
pd                243 drivers/block/paride/pd.c static struct pd_unit pd[PD_UNITS];
pd                404 drivers/block/paride/pd.c 		disk = pd[pd_queue].gd;
pd                899 drivers/block/paride/pd.c 	p->first_minor = (disk - pd) << PD_BITS;
pd                944 drivers/block/paride/pd.c 		struct pd_unit *disk = pd + unit;
pd                965 drivers/block/paride/pd.c 		disk = pd;
pd                974 drivers/block/paride/pd.c 		for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) {
pd                987 drivers/block/paride/pd.c 	for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) {
pd               1027 drivers/block/paride/pd.c 	for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) {
pd                 75 drivers/block/pktcdvd.c #define pkt_err(pd, fmt, ...)						\
pd                 76 drivers/block/pktcdvd.c 	pr_err("%s: " fmt, pd->name, ##__VA_ARGS__)
pd                 77 drivers/block/pktcdvd.c #define pkt_notice(pd, fmt, ...)					\
pd                 78 drivers/block/pktcdvd.c 	pr_notice("%s: " fmt, pd->name, ##__VA_ARGS__)
pd                 79 drivers/block/pktcdvd.c #define pkt_info(pd, fmt, ...)						\
pd                 80 drivers/block/pktcdvd.c 	pr_info("%s: " fmt, pd->name, ##__VA_ARGS__)
pd                 82 drivers/block/pktcdvd.c #define pkt_dbg(level, pd, fmt, ...)					\
pd                 86 drivers/block/pktcdvd.c 			  pd->name, __func__, ##__VA_ARGS__);		\
pd                 88 drivers/block/pktcdvd.c 		pr_notice("%s: " fmt, pd->name, ##__VA_ARGS__);		\
pd                111 drivers/block/pktcdvd.c static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd)
pd                113 drivers/block/pktcdvd.c 	return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1);
pd                119 drivers/block/pktcdvd.c static struct pktcdvd_kobj* pkt_kobj_create(struct pktcdvd_device *pd,
pd                130 drivers/block/pktcdvd.c 	p->pd = pd;
pd                210 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = to_pktcdvdkobj(kobj)->pd;
pd                214 drivers/block/pktcdvd.c 		n = sprintf(data, "%lu\n", pd->stats.pkt_started);
pd                217 drivers/block/pktcdvd.c 		n = sprintf(data, "%lu\n", pd->stats.pkt_ended);
pd                220 drivers/block/pktcdvd.c 		n = sprintf(data, "%lu\n", pd->stats.secs_w >> 1);
pd                223 drivers/block/pktcdvd.c 		n = sprintf(data, "%lu\n", pd->stats.secs_r >> 1);
pd                226 drivers/block/pktcdvd.c 		n = sprintf(data, "%lu\n", pd->stats.secs_rg >> 1);
pd                229 drivers/block/pktcdvd.c 		spin_lock(&pd->lock);
pd                230 drivers/block/pktcdvd.c 		v = pd->bio_queue_size;
pd                231 drivers/block/pktcdvd.c 		spin_unlock(&pd->lock);
pd                235 drivers/block/pktcdvd.c 		spin_lock(&pd->lock);
pd                236 drivers/block/pktcdvd.c 		v = pd->write_congestion_off;
pd                237 drivers/block/pktcdvd.c 		spin_unlock(&pd->lock);
pd                241 drivers/block/pktcdvd.c 		spin_lock(&pd->lock);
pd                242 drivers/block/pktcdvd.c 		v = pd->write_congestion_on;
pd                243 drivers/block/pktcdvd.c 		spin_unlock(&pd->lock);
pd                270 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = to_pktcdvdkobj(kobj)->pd;
pd                274 drivers/block/pktcdvd.c 		pd->stats.pkt_started = 0;
pd                275 drivers/block/pktcdvd.c 		pd->stats.pkt_ended = 0;
pd                276 drivers/block/pktcdvd.c 		pd->stats.secs_w = 0;
pd                277 drivers/block/pktcdvd.c 		pd->stats.secs_rg = 0;
pd                278 drivers/block/pktcdvd.c 		pd->stats.secs_r = 0;
pd                282 drivers/block/pktcdvd.c 		spin_lock(&pd->lock);
pd                283 drivers/block/pktcdvd.c 		pd->write_congestion_off = val;
pd                284 drivers/block/pktcdvd.c 		init_write_congestion_marks(&pd->write_congestion_off,
pd                285 drivers/block/pktcdvd.c 					&pd->write_congestion_on);
pd                286 drivers/block/pktcdvd.c 		spin_unlock(&pd->lock);
pd                290 drivers/block/pktcdvd.c 		spin_lock(&pd->lock);
pd                291 drivers/block/pktcdvd.c 		pd->write_congestion_on = val;
pd                292 drivers/block/pktcdvd.c 		init_write_congestion_marks(&pd->write_congestion_off,
pd                293 drivers/block/pktcdvd.c 					&pd->write_congestion_on);
pd                294 drivers/block/pktcdvd.c 		spin_unlock(&pd->lock);
pd                314 drivers/block/pktcdvd.c static void pkt_sysfs_dev_new(struct pktcdvd_device *pd)
pd                317 drivers/block/pktcdvd.c 		pd->dev = device_create(class_pktcdvd, NULL, MKDEV(0, 0), NULL,
pd                318 drivers/block/pktcdvd.c 					"%s", pd->name);
pd                319 drivers/block/pktcdvd.c 		if (IS_ERR(pd->dev))
pd                320 drivers/block/pktcdvd.c 			pd->dev = NULL;
pd                322 drivers/block/pktcdvd.c 	if (pd->dev) {
pd                323 drivers/block/pktcdvd.c 		pd->kobj_stat = pkt_kobj_create(pd, "stat",
pd                324 drivers/block/pktcdvd.c 					&pd->dev->kobj,
pd                326 drivers/block/pktcdvd.c 		pd->kobj_wqueue = pkt_kobj_create(pd, "write_queue",
pd                327 drivers/block/pktcdvd.c 					&pd->dev->kobj,
pd                332 drivers/block/pktcdvd.c static void pkt_sysfs_dev_remove(struct pktcdvd_device *pd)
pd                334 drivers/block/pktcdvd.c 	pkt_kobj_remove(pd->kobj_stat);
pd                335 drivers/block/pktcdvd.c 	pkt_kobj_remove(pd->kobj_wqueue);
pd                337 drivers/block/pktcdvd.c 		device_unregister(pd->dev);
pd                360 drivers/block/pktcdvd.c 		struct pktcdvd_device *pd = pkt_devs[idx];
pd                361 drivers/block/pktcdvd.c 		if (!pd)
pd                364 drivers/block/pktcdvd.c 			pd->name,
pd                365 drivers/block/pktcdvd.c 			MAJOR(pd->pkt_dev), MINOR(pd->pkt_dev),
pd                366 drivers/block/pktcdvd.c 			MAJOR(pd->bdev->bd_dev),
pd                367 drivers/block/pktcdvd.c 			MINOR(pd->bdev->bd_dev));
pd                473 drivers/block/pktcdvd.c static void pkt_debugfs_dev_new(struct pktcdvd_device *pd)
pd                477 drivers/block/pktcdvd.c 	pd->dfs_d_root = debugfs_create_dir(pd->name, pkt_debugfs_root);
pd                478 drivers/block/pktcdvd.c 	if (!pd->dfs_d_root)
pd                481 drivers/block/pktcdvd.c 	pd->dfs_f_info = debugfs_create_file("info", 0444,
pd                482 drivers/block/pktcdvd.c 					     pd->dfs_d_root, pd, &debug_fops);
pd                485 drivers/block/pktcdvd.c static void pkt_debugfs_dev_remove(struct pktcdvd_device *pd)
pd                489 drivers/block/pktcdvd.c 	debugfs_remove(pd->dfs_f_info);
pd                490 drivers/block/pktcdvd.c 	debugfs_remove(pd->dfs_d_root);
pd                491 drivers/block/pktcdvd.c 	pd->dfs_f_info = NULL;
pd                492 drivers/block/pktcdvd.c 	pd->dfs_d_root = NULL;
pd                509 drivers/block/pktcdvd.c static void pkt_bio_finished(struct pktcdvd_device *pd)
pd                511 drivers/block/pktcdvd.c 	BUG_ON(atomic_read(&pd->cdrw.pending_bios) <= 0);
pd                512 drivers/block/pktcdvd.c 	if (atomic_dec_and_test(&pd->cdrw.pending_bios)) {
pd                513 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "queue empty\n");
pd                514 drivers/block/pktcdvd.c 		atomic_set(&pd->iosched.attention, 1);
pd                515 drivers/block/pktcdvd.c 		wake_up(&pd->wqueue);
pd                591 drivers/block/pktcdvd.c static void pkt_shrink_pktlist(struct pktcdvd_device *pd)
pd                595 drivers/block/pktcdvd.c 	BUG_ON(!list_empty(&pd->cdrw.pkt_active_list));
pd                597 drivers/block/pktcdvd.c 	list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_free_list, list) {
pd                600 drivers/block/pktcdvd.c 	INIT_LIST_HEAD(&pd->cdrw.pkt_free_list);
pd                603 drivers/block/pktcdvd.c static int pkt_grow_pktlist(struct pktcdvd_device *pd, int nr_packets)
pd                607 drivers/block/pktcdvd.c 	BUG_ON(!list_empty(&pd->cdrw.pkt_free_list));
pd                610 drivers/block/pktcdvd.c 		pkt = pkt_alloc_packet_data(pd->settings.size >> 2);
pd                612 drivers/block/pktcdvd.c 			pkt_shrink_pktlist(pd);
pd                616 drivers/block/pktcdvd.c 		pkt->pd = pd;
pd                617 drivers/block/pktcdvd.c 		list_add(&pkt->list, &pd->cdrw.pkt_free_list);
pd                631 drivers/block/pktcdvd.c static void pkt_rbtree_erase(struct pktcdvd_device *pd, struct pkt_rb_node *node)
pd                633 drivers/block/pktcdvd.c 	rb_erase(&node->rb_node, &pd->bio_queue);
pd                634 drivers/block/pktcdvd.c 	mempool_free(node, &pd->rb_pool);
pd                635 drivers/block/pktcdvd.c 	pd->bio_queue_size--;
pd                636 drivers/block/pktcdvd.c 	BUG_ON(pd->bio_queue_size < 0);
pd                642 drivers/block/pktcdvd.c static struct pkt_rb_node *pkt_rbtree_find(struct pktcdvd_device *pd, sector_t s)
pd                644 drivers/block/pktcdvd.c 	struct rb_node *n = pd->bio_queue.rb_node;
pd                649 drivers/block/pktcdvd.c 		BUG_ON(pd->bio_queue_size > 0);
pd                676 drivers/block/pktcdvd.c static void pkt_rbtree_insert(struct pktcdvd_device *pd, struct pkt_rb_node *node)
pd                678 drivers/block/pktcdvd.c 	struct rb_node **p = &pd->bio_queue.rb_node;
pd                692 drivers/block/pktcdvd.c 	rb_insert_color(&node->rb_node, &pd->bio_queue);
pd                693 drivers/block/pktcdvd.c 	pd->bio_queue_size++;
pd                700 drivers/block/pktcdvd.c static int pkt_generic_packet(struct pktcdvd_device *pd, struct packet_command *cgc)
pd                702 drivers/block/pktcdvd.c 	struct request_queue *q = bdev_get_queue(pd->bdev);
pd                725 drivers/block/pktcdvd.c 	blk_execute_rq(rq->q, pd->bdev->bd_disk, rq, 0);
pd                748 drivers/block/pktcdvd.c static void pkt_dump_sense(struct pktcdvd_device *pd,
pd                754 drivers/block/pktcdvd.c 		pkt_err(pd, "%*ph - sense %02x.%02x.%02x (%s)\n",
pd                759 drivers/block/pktcdvd.c 		pkt_err(pd, "%*ph - no sense\n", CDROM_PACKET_SIZE, cgc->cmd);
pd                765 drivers/block/pktcdvd.c static int pkt_flush_cache(struct pktcdvd_device *pd)
pd                780 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, &cgc);
pd                786 drivers/block/pktcdvd.c static noinline_for_stack int pkt_set_speed(struct pktcdvd_device *pd,
pd                801 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd                803 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd                812 drivers/block/pktcdvd.c static void pkt_queue_bio(struct pktcdvd_device *pd, struct bio *bio)
pd                814 drivers/block/pktcdvd.c 	spin_lock(&pd->iosched.lock);
pd                816 drivers/block/pktcdvd.c 		bio_list_add(&pd->iosched.read_queue, bio);
pd                818 drivers/block/pktcdvd.c 		bio_list_add(&pd->iosched.write_queue, bio);
pd                819 drivers/block/pktcdvd.c 	spin_unlock(&pd->iosched.lock);
pd                821 drivers/block/pktcdvd.c 	atomic_set(&pd->iosched.attention, 1);
pd                822 drivers/block/pktcdvd.c 	wake_up(&pd->wqueue);
pd                841 drivers/block/pktcdvd.c static void pkt_iosched_process_queue(struct pktcdvd_device *pd)
pd                844 drivers/block/pktcdvd.c 	if (atomic_read(&pd->iosched.attention) == 0)
pd                846 drivers/block/pktcdvd.c 	atomic_set(&pd->iosched.attention, 0);
pd                852 drivers/block/pktcdvd.c 		spin_lock(&pd->iosched.lock);
pd                853 drivers/block/pktcdvd.c 		reads_queued = !bio_list_empty(&pd->iosched.read_queue);
pd                854 drivers/block/pktcdvd.c 		writes_queued = !bio_list_empty(&pd->iosched.write_queue);
pd                855 drivers/block/pktcdvd.c 		spin_unlock(&pd->iosched.lock);
pd                860 drivers/block/pktcdvd.c 		if (pd->iosched.writing) {
pd                862 drivers/block/pktcdvd.c 			spin_lock(&pd->iosched.lock);
pd                863 drivers/block/pktcdvd.c 			bio = bio_list_peek(&pd->iosched.write_queue);
pd                864 drivers/block/pktcdvd.c 			spin_unlock(&pd->iosched.lock);
pd                866 drivers/block/pktcdvd.c 				    pd->iosched.last_write))
pd                869 drivers/block/pktcdvd.c 				if (atomic_read(&pd->cdrw.pending_bios) > 0) {
pd                870 drivers/block/pktcdvd.c 					pkt_dbg(2, pd, "write, waiting\n");
pd                873 drivers/block/pktcdvd.c 				pkt_flush_cache(pd);
pd                874 drivers/block/pktcdvd.c 				pd->iosched.writing = 0;
pd                878 drivers/block/pktcdvd.c 				if (atomic_read(&pd->cdrw.pending_bios) > 0) {
pd                879 drivers/block/pktcdvd.c 					pkt_dbg(2, pd, "read, waiting\n");
pd                882 drivers/block/pktcdvd.c 				pd->iosched.writing = 1;
pd                886 drivers/block/pktcdvd.c 		spin_lock(&pd->iosched.lock);
pd                887 drivers/block/pktcdvd.c 		if (pd->iosched.writing)
pd                888 drivers/block/pktcdvd.c 			bio = bio_list_pop(&pd->iosched.write_queue);
pd                890 drivers/block/pktcdvd.c 			bio = bio_list_pop(&pd->iosched.read_queue);
pd                891 drivers/block/pktcdvd.c 		spin_unlock(&pd->iosched.lock);
pd                897 drivers/block/pktcdvd.c 			pd->iosched.successive_reads +=
pd                900 drivers/block/pktcdvd.c 			pd->iosched.successive_reads = 0;
pd                901 drivers/block/pktcdvd.c 			pd->iosched.last_write = bio_end_sector(bio);
pd                903 drivers/block/pktcdvd.c 		if (pd->iosched.successive_reads >= HI_SPEED_SWITCH) {
pd                904 drivers/block/pktcdvd.c 			if (pd->read_speed == pd->write_speed) {
pd                905 drivers/block/pktcdvd.c 				pd->read_speed = MAX_SPEED;
pd                906 drivers/block/pktcdvd.c 				pkt_set_speed(pd, pd->write_speed, pd->read_speed);
pd                909 drivers/block/pktcdvd.c 			if (pd->read_speed != pd->write_speed) {
pd                910 drivers/block/pktcdvd.c 				pd->read_speed = pd->write_speed;
pd                911 drivers/block/pktcdvd.c 				pkt_set_speed(pd, pd->write_speed, pd->read_speed);
pd                915 drivers/block/pktcdvd.c 		atomic_inc(&pd->cdrw.pending_bios);
pd                924 drivers/block/pktcdvd.c static int pkt_set_segment_merging(struct pktcdvd_device *pd, struct request_queue *q)
pd                926 drivers/block/pktcdvd.c 	if ((pd->settings.size << 9) / CD_FRAMESIZE
pd                931 drivers/block/pktcdvd.c 		clear_bit(PACKET_MERGE_SEGS, &pd->flags);
pd                933 drivers/block/pktcdvd.c 	} else if ((pd->settings.size << 9) / PAGE_SIZE
pd                939 drivers/block/pktcdvd.c 		set_bit(PACKET_MERGE_SEGS, &pd->flags);
pd                942 drivers/block/pktcdvd.c 		pkt_err(pd, "cdrom max_phys_segments too small\n");
pd                950 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = pkt->pd;
pd                951 drivers/block/pktcdvd.c 	BUG_ON(!pd);
pd                953 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "bio=%p sec0=%llx sec=%llx err=%d\n",
pd                961 drivers/block/pktcdvd.c 		wake_up(&pd->wqueue);
pd                963 drivers/block/pktcdvd.c 	pkt_bio_finished(pd);
pd                969 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = pkt->pd;
pd                970 drivers/block/pktcdvd.c 	BUG_ON(!pd);
pd                972 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "id=%d, err=%d\n", pkt->id, bio->bi_status);
pd                974 drivers/block/pktcdvd.c 	pd->stats.pkt_ended++;
pd                976 drivers/block/pktcdvd.c 	pkt_bio_finished(pd);
pd                979 drivers/block/pktcdvd.c 	wake_up(&pd->wqueue);
pd                985 drivers/block/pktcdvd.c static void pkt_gather_data(struct pktcdvd_device *pd, struct packet_data *pkt)
pd               1006 drivers/block/pktcdvd.c 		pd->stats.secs_w += num_frames * (CD_FRAMESIZE >> 9);
pd               1015 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "zone %llx cached\n",
pd               1032 drivers/block/pktcdvd.c 		bio_set_dev(bio, pd->bdev);
pd               1038 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "Adding frame %d, page:%p offs:%d\n",
pd               1045 drivers/block/pktcdvd.c 		pkt_queue_bio(pd, bio);
pd               1050 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "need %d frames for zone %llx\n",
pd               1052 drivers/block/pktcdvd.c 	pd->stats.pkt_started++;
pd               1053 drivers/block/pktcdvd.c 	pd->stats.secs_rg += frames_read * (CD_FRAMESIZE >> 9);
pd               1060 drivers/block/pktcdvd.c static struct packet_data *pkt_get_packet_data(struct pktcdvd_device *pd, int zone)
pd               1064 drivers/block/pktcdvd.c 	list_for_each_entry(pkt, &pd->cdrw.pkt_free_list, list) {
pd               1065 drivers/block/pktcdvd.c 		if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) {
pd               1076 drivers/block/pktcdvd.c static void pkt_put_packet_data(struct pktcdvd_device *pd, struct packet_data *pkt)
pd               1079 drivers/block/pktcdvd.c 		list_add(&pkt->list, &pd->cdrw.pkt_free_list);
pd               1081 drivers/block/pktcdvd.c 		list_add_tail(&pkt->list, &pd->cdrw.pkt_free_list);
pd               1100 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = rq->rq_disk->private_data;
pd               1106 drivers/block/pktcdvd.c 	pkt_bdev = bdget(kdev_t_to_nr(pd->pkt_dev));
pd               1126 drivers/block/pktcdvd.c 	bio_set_dev(pkt->bio, pd->bdev);
pd               1151 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "pkt %2d : s=%6llx %s -> %s\n",
pd               1162 drivers/block/pktcdvd.c static int pkt_handle_queue(struct pktcdvd_device *pd)
pd               1171 drivers/block/pktcdvd.c 	atomic_set(&pd->scan_queue, 0);
pd               1173 drivers/block/pktcdvd.c 	if (list_empty(&pd->cdrw.pkt_free_list)) {
pd               1174 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "no pkt\n");
pd               1181 drivers/block/pktcdvd.c 	spin_lock(&pd->lock);
pd               1182 drivers/block/pktcdvd.c 	first_node = pkt_rbtree_find(pd, pd->current_sector);
pd               1184 drivers/block/pktcdvd.c 		n = rb_first(&pd->bio_queue);
pd               1191 drivers/block/pktcdvd.c 		zone = get_zone(bio->bi_iter.bi_sector, pd);
pd               1192 drivers/block/pktcdvd.c 		list_for_each_entry(p, &pd->cdrw.pkt_active_list, list) {
pd               1202 drivers/block/pktcdvd.c 			n = rb_first(&pd->bio_queue);
pd               1209 drivers/block/pktcdvd.c 	spin_unlock(&pd->lock);
pd               1211 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "no bio\n");
pd               1215 drivers/block/pktcdvd.c 	pkt = pkt_get_packet_data(pd, zone);
pd               1217 drivers/block/pktcdvd.c 	pd->current_sector = zone + pd->settings.size;
pd               1219 drivers/block/pktcdvd.c 	BUG_ON(pkt->frames != pd->settings.size >> 2);
pd               1226 drivers/block/pktcdvd.c 	spin_lock(&pd->lock);
pd               1227 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "looking for zone %llx\n", (unsigned long long)zone);
pd               1228 drivers/block/pktcdvd.c 	while ((node = pkt_rbtree_find(pd, zone)) != NULL) {
pd               1230 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "found zone=%llx\n", (unsigned long long)
pd               1231 drivers/block/pktcdvd.c 			get_zone(bio->bi_iter.bi_sector, pd));
pd               1232 drivers/block/pktcdvd.c 		if (get_zone(bio->bi_iter.bi_sector, pd) != zone)
pd               1234 drivers/block/pktcdvd.c 		pkt_rbtree_erase(pd, node);
pd               1242 drivers/block/pktcdvd.c 	wakeup = (pd->write_congestion_on > 0
pd               1243 drivers/block/pktcdvd.c 	 		&& pd->bio_queue_size <= pd->write_congestion_off);
pd               1244 drivers/block/pktcdvd.c 	spin_unlock(&pd->lock);
pd               1246 drivers/block/pktcdvd.c 		clear_bdi_congested(pd->disk->queue->backing_dev_info,
pd               1254 drivers/block/pktcdvd.c 	spin_lock(&pd->cdrw.active_list_lock);
pd               1255 drivers/block/pktcdvd.c 	list_add(&pkt->list, &pd->cdrw.pkt_active_list);
pd               1256 drivers/block/pktcdvd.c 	spin_unlock(&pd->cdrw.active_list_lock);
pd               1265 drivers/block/pktcdvd.c static void pkt_start_write(struct pktcdvd_device *pd, struct packet_data *pkt)
pd               1271 drivers/block/pktcdvd.c 	bio_set_dev(pkt->w_bio, pd->bdev);
pd               1283 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "vcnt=%d\n", pkt->w_bio->bi_vcnt);
pd               1294 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "Writing %d frames for zone %llx\n",
pd               1297 drivers/block/pktcdvd.c 	if (test_bit(PACKET_MERGE_SEGS, &pd->flags) || (pkt->write_size < pkt->frames))
pd               1305 drivers/block/pktcdvd.c 	pkt_queue_bio(pd, pkt->w_bio);
pd               1322 drivers/block/pktcdvd.c static void pkt_run_state_machine(struct pktcdvd_device *pd, struct packet_data *pkt)
pd               1324 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "pkt %d\n", pkt->id);
pd               1333 drivers/block/pktcdvd.c 			pkt_gather_data(pd, pkt);
pd               1344 drivers/block/pktcdvd.c 				pkt_start_write(pd, pkt);
pd               1361 drivers/block/pktcdvd.c 				pkt_start_write(pd, pkt);
pd               1363 drivers/block/pktcdvd.c 				pkt_dbg(2, pd, "No recovery possible\n");
pd               1379 drivers/block/pktcdvd.c static void pkt_handle_packets(struct pktcdvd_device *pd)
pd               1386 drivers/block/pktcdvd.c 	list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               1389 drivers/block/pktcdvd.c 			pkt_run_state_machine(pd, pkt);
pd               1396 drivers/block/pktcdvd.c 	spin_lock(&pd->cdrw.active_list_lock);
pd               1397 drivers/block/pktcdvd.c 	list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_active_list, list) {
pd               1400 drivers/block/pktcdvd.c 			pkt_put_packet_data(pd, pkt);
pd               1402 drivers/block/pktcdvd.c 			atomic_set(&pd->scan_queue, 1);
pd               1405 drivers/block/pktcdvd.c 	spin_unlock(&pd->cdrw.active_list_lock);
pd               1408 drivers/block/pktcdvd.c static void pkt_count_states(struct pktcdvd_device *pd, int *states)
pd               1416 drivers/block/pktcdvd.c 	spin_lock(&pd->cdrw.active_list_lock);
pd               1417 drivers/block/pktcdvd.c 	list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               1420 drivers/block/pktcdvd.c 	spin_unlock(&pd->cdrw.active_list_lock);
pd               1429 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = foobar;
pd               1442 drivers/block/pktcdvd.c 		add_wait_queue(&pd->wqueue, &wait);
pd               1447 drivers/block/pktcdvd.c 			if (atomic_read(&pd->scan_queue) > 0)
pd               1451 drivers/block/pktcdvd.c 			list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               1457 drivers/block/pktcdvd.c 			if (atomic_read(&pd->iosched.attention) != 0)
pd               1463 drivers/block/pktcdvd.c 				pkt_count_states(pd, states);
pd               1464 drivers/block/pktcdvd.c 				pkt_dbg(2, pd, "i:%d ow:%d rw:%d ww:%d rec:%d fin:%d\n",
pd               1470 drivers/block/pktcdvd.c 			list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               1475 drivers/block/pktcdvd.c 			pkt_dbg(2, pd, "sleeping\n");
pd               1477 drivers/block/pktcdvd.c 			pkt_dbg(2, pd, "wake up\n");
pd               1482 drivers/block/pktcdvd.c 			list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               1497 drivers/block/pktcdvd.c 		remove_wait_queue(&pd->wqueue, &wait);
pd               1506 drivers/block/pktcdvd.c 		while (pkt_handle_queue(pd))
pd               1512 drivers/block/pktcdvd.c 		pkt_handle_packets(pd);
pd               1517 drivers/block/pktcdvd.c 		pkt_iosched_process_queue(pd);
pd               1523 drivers/block/pktcdvd.c static void pkt_print_settings(struct pktcdvd_device *pd)
pd               1525 drivers/block/pktcdvd.c 	pkt_info(pd, "%s packets, %u blocks, Mode-%c disc\n",
pd               1526 drivers/block/pktcdvd.c 		 pd->settings.fp ? "Fixed" : "Variable",
pd               1527 drivers/block/pktcdvd.c 		 pd->settings.size >> 2,
pd               1528 drivers/block/pktcdvd.c 		 pd->settings.block_mode == 8 ? '1' : '2');
pd               1531 drivers/block/pktcdvd.c static int pkt_mode_sense(struct pktcdvd_device *pd, struct packet_command *cgc, int page_code, int page_control)
pd               1540 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, cgc);
pd               1543 drivers/block/pktcdvd.c static int pkt_mode_select(struct pktcdvd_device *pd, struct packet_command *cgc)
pd               1552 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, cgc);
pd               1555 drivers/block/pktcdvd.c static int pkt_get_disc_info(struct pktcdvd_device *pd, disc_information *di)
pd               1566 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               1580 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, &cgc);
pd               1583 drivers/block/pktcdvd.c static int pkt_get_track_info(struct pktcdvd_device *pd, __u16 track, __u8 type, track_information *ti)
pd               1596 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               1607 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, &cgc);
pd               1610 drivers/block/pktcdvd.c static noinline_for_stack int pkt_get_last_written(struct pktcdvd_device *pd,
pd               1618 drivers/block/pktcdvd.c 	ret = pkt_get_disc_info(pd, &di);
pd               1623 drivers/block/pktcdvd.c 	ret = pkt_get_track_info(pd, last_track, 1, &ti);
pd               1630 drivers/block/pktcdvd.c 		ret = pkt_get_track_info(pd, last_track, 1, &ti);
pd               1651 drivers/block/pktcdvd.c static noinline_for_stack int pkt_set_write_settings(struct pktcdvd_device *pd)
pd               1660 drivers/block/pktcdvd.c 	if ((pd->mmc3_profile == 0x1a) || (pd->mmc3_profile == 0x12))
pd               1666 drivers/block/pktcdvd.c 	ret = pkt_mode_sense(pd, &cgc, GPMODE_WRITE_PARMS_PAGE, 0);
pd               1668 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               1673 drivers/block/pktcdvd.c 	pd->mode_offset = (buffer[6] << 8) | (buffer[7] & 0xff);
pd               1682 drivers/block/pktcdvd.c 	ret = pkt_mode_sense(pd, &cgc, GPMODE_WRITE_PARMS_PAGE, 0);
pd               1684 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               1691 drivers/block/pktcdvd.c 	wp = (write_param_page *) &buffer[sizeof(struct mode_page_header) + pd->mode_offset];
pd               1693 drivers/block/pktcdvd.c 	wp->fp = pd->settings.fp;
pd               1694 drivers/block/pktcdvd.c 	wp->track_mode = pd->settings.track_mode;
pd               1695 drivers/block/pktcdvd.c 	wp->write_type = pd->settings.write_type;
pd               1696 drivers/block/pktcdvd.c 	wp->data_block_type = pd->settings.block_mode;
pd               1719 drivers/block/pktcdvd.c 		pkt_err(pd, "write mode wrong %d\n", wp->data_block_type);
pd               1722 drivers/block/pktcdvd.c 	wp->packet_size = cpu_to_be32(pd->settings.size >> 2);
pd               1725 drivers/block/pktcdvd.c 	ret = pkt_mode_select(pd, &cgc);
pd               1727 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               1731 drivers/block/pktcdvd.c 	pkt_print_settings(pd);
pd               1738 drivers/block/pktcdvd.c static int pkt_writable_track(struct pktcdvd_device *pd, track_information *ti)
pd               1740 drivers/block/pktcdvd.c 	switch (pd->mmc3_profile) {
pd               1764 drivers/block/pktcdvd.c 	pkt_err(pd, "bad state %d-%d-%d\n", ti->rt, ti->blank, ti->packet);
pd               1771 drivers/block/pktcdvd.c static int pkt_writable_disc(struct pktcdvd_device *pd, disc_information *di)
pd               1773 drivers/block/pktcdvd.c 	switch (pd->mmc3_profile) {
pd               1782 drivers/block/pktcdvd.c 			pkt_dbg(2, pd, "Wrong disc profile (%x)\n",
pd               1783 drivers/block/pktcdvd.c 				pd->mmc3_profile);
pd               1792 drivers/block/pktcdvd.c 		pkt_notice(pd, "unknown disc - no track?\n");
pd               1797 drivers/block/pktcdvd.c 		pkt_err(pd, "wrong disc type (%x)\n", di->disc_type);
pd               1802 drivers/block/pktcdvd.c 		pkt_notice(pd, "disc not erasable\n");
pd               1807 drivers/block/pktcdvd.c 		pkt_err(pd, "can't write to last track (reserved)\n");
pd               1814 drivers/block/pktcdvd.c static noinline_for_stack int pkt_probe_settings(struct pktcdvd_device *pd)
pd               1825 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               1826 drivers/block/pktcdvd.c 	pd->mmc3_profile = ret ? 0xffff : buf[6] << 8 | buf[7];
pd               1831 drivers/block/pktcdvd.c 	ret = pkt_get_disc_info(pd, &di);
pd               1833 drivers/block/pktcdvd.c 		pkt_err(pd, "failed get_disc\n");
pd               1837 drivers/block/pktcdvd.c 	if (!pkt_writable_disc(pd, &di))
pd               1840 drivers/block/pktcdvd.c 	pd->type = di.erasable ? PACKET_CDRW : PACKET_CDR;
pd               1843 drivers/block/pktcdvd.c 	ret = pkt_get_track_info(pd, track, 1, &ti);
pd               1845 drivers/block/pktcdvd.c 		pkt_err(pd, "failed get_track\n");
pd               1849 drivers/block/pktcdvd.c 	if (!pkt_writable_track(pd, &ti)) {
pd               1850 drivers/block/pktcdvd.c 		pkt_err(pd, "can't write to this track\n");
pd               1858 drivers/block/pktcdvd.c 	pd->settings.size = be32_to_cpu(ti.fixed_packet_size) << 2;
pd               1859 drivers/block/pktcdvd.c 	if (pd->settings.size == 0) {
pd               1860 drivers/block/pktcdvd.c 		pkt_notice(pd, "detected zero packet size!\n");
pd               1863 drivers/block/pktcdvd.c 	if (pd->settings.size > PACKET_MAX_SECTORS) {
pd               1864 drivers/block/pktcdvd.c 		pkt_err(pd, "packet size is too big\n");
pd               1867 drivers/block/pktcdvd.c 	pd->settings.fp = ti.fp;
pd               1868 drivers/block/pktcdvd.c 	pd->offset = (be32_to_cpu(ti.track_start) << 2) & (pd->settings.size - 1);
pd               1871 drivers/block/pktcdvd.c 		pd->nwa = be32_to_cpu(ti.next_writable);
pd               1872 drivers/block/pktcdvd.c 		set_bit(PACKET_NWA_VALID, &pd->flags);
pd               1881 drivers/block/pktcdvd.c 		pd->lra = be32_to_cpu(ti.last_rec_address);
pd               1882 drivers/block/pktcdvd.c 		set_bit(PACKET_LRA_VALID, &pd->flags);
pd               1884 drivers/block/pktcdvd.c 		pd->lra = 0xffffffff;
pd               1885 drivers/block/pktcdvd.c 		set_bit(PACKET_LRA_VALID, &pd->flags);
pd               1891 drivers/block/pktcdvd.c 	pd->settings.link_loss = 7;
pd               1892 drivers/block/pktcdvd.c 	pd->settings.write_type = 0;	/* packet */
pd               1893 drivers/block/pktcdvd.c 	pd->settings.track_mode = ti.track_mode;
pd               1900 drivers/block/pktcdvd.c 			pd->settings.block_mode = PACKET_BLOCK_MODE1;
pd               1903 drivers/block/pktcdvd.c 			pd->settings.block_mode = PACKET_BLOCK_MODE2;
pd               1906 drivers/block/pktcdvd.c 			pkt_err(pd, "unknown data mode\n");
pd               1915 drivers/block/pktcdvd.c static noinline_for_stack int pkt_write_caching(struct pktcdvd_device *pd,
pd               1925 drivers/block/pktcdvd.c 	cgc.buflen = pd->mode_offset + 12;
pd               1932 drivers/block/pktcdvd.c 	ret = pkt_mode_sense(pd, &cgc, GPMODE_WCACHING_PAGE, 0);
pd               1936 drivers/block/pktcdvd.c 	buf[pd->mode_offset + 10] |= (!!set << 2);
pd               1939 drivers/block/pktcdvd.c 	ret = pkt_mode_select(pd, &cgc);
pd               1941 drivers/block/pktcdvd.c 		pkt_err(pd, "write caching control failed\n");
pd               1942 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               1944 drivers/block/pktcdvd.c 		pkt_notice(pd, "enabled write caching\n");
pd               1948 drivers/block/pktcdvd.c static int pkt_lock_door(struct pktcdvd_device *pd, int lockflag)
pd               1955 drivers/block/pktcdvd.c 	return pkt_generic_packet(pd, &cgc);
pd               1961 drivers/block/pktcdvd.c static noinline_for_stack int pkt_get_max_speed(struct pktcdvd_device *pd,
pd               1970 drivers/block/pktcdvd.c 	cap_buf = &buf[sizeof(struct mode_page_header) + pd->mode_offset];
pd               1974 drivers/block/pktcdvd.c 	ret = pkt_mode_sense(pd, &cgc, GPMODE_CAPABILITIES_PAGE, 0);
pd               1976 drivers/block/pktcdvd.c 		cgc.buflen = pd->mode_offset + cap_buf[1] + 2 +
pd               1978 drivers/block/pktcdvd.c 		ret = pkt_mode_sense(pd, &cgc, GPMODE_CAPABILITIES_PAGE, 0);
pd               1980 drivers/block/pktcdvd.c 			pkt_dump_sense(pd, &cgc);
pd               2022 drivers/block/pktcdvd.c static noinline_for_stack int pkt_media_speed(struct pktcdvd_device *pd,
pd               2037 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               2039 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               2052 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               2054 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               2059 drivers/block/pktcdvd.c 		pkt_notice(pd, "disc type is not CD-RW\n");
pd               2063 drivers/block/pktcdvd.c 		pkt_notice(pd, "A1 values on media are not valid, maybe not CDRW?\n");
pd               2083 drivers/block/pktcdvd.c 			pkt_notice(pd, "unknown disc sub-type %d\n", st);
pd               2087 drivers/block/pktcdvd.c 		pkt_info(pd, "maximum media speed: %d\n", *speed);
pd               2090 drivers/block/pktcdvd.c 		pkt_notice(pd, "unknown speed %d for sub-type %d\n", sp, st);
pd               2095 drivers/block/pktcdvd.c static noinline_for_stack int pkt_perform_opc(struct pktcdvd_device *pd)
pd               2101 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "Performing OPC\n");
pd               2108 drivers/block/pktcdvd.c 	ret = pkt_generic_packet(pd, &cgc);
pd               2110 drivers/block/pktcdvd.c 		pkt_dump_sense(pd, &cgc);
pd               2114 drivers/block/pktcdvd.c static int pkt_open_write(struct pktcdvd_device *pd)
pd               2119 drivers/block/pktcdvd.c 	ret = pkt_probe_settings(pd);
pd               2121 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "failed probe\n");
pd               2125 drivers/block/pktcdvd.c 	ret = pkt_set_write_settings(pd);
pd               2127 drivers/block/pktcdvd.c 		pkt_dbg(1, pd, "failed saving write settings\n");
pd               2131 drivers/block/pktcdvd.c 	pkt_write_caching(pd, USE_WCACHING);
pd               2133 drivers/block/pktcdvd.c 	ret = pkt_get_max_speed(pd, &write_speed);
pd               2136 drivers/block/pktcdvd.c 	switch (pd->mmc3_profile) {
pd               2140 drivers/block/pktcdvd.c 			pkt_dbg(1, pd, "write speed %ukB/s\n", write_speed);
pd               2143 drivers/block/pktcdvd.c 			ret = pkt_media_speed(pd, &media_write_speed);
pd               2147 drivers/block/pktcdvd.c 			pkt_dbg(1, pd, "write speed %ux\n", write_speed / 176);
pd               2152 drivers/block/pktcdvd.c 	ret = pkt_set_speed(pd, write_speed, read_speed);
pd               2154 drivers/block/pktcdvd.c 		pkt_dbg(1, pd, "couldn't set write speed\n");
pd               2157 drivers/block/pktcdvd.c 	pd->write_speed = write_speed;
pd               2158 drivers/block/pktcdvd.c 	pd->read_speed = read_speed;
pd               2160 drivers/block/pktcdvd.c 	ret = pkt_perform_opc(pd);
pd               2162 drivers/block/pktcdvd.c 		pkt_dbg(1, pd, "Optimum Power Calibration failed\n");
pd               2171 drivers/block/pktcdvd.c static int pkt_open_dev(struct pktcdvd_device *pd, fmode_t write)
pd               2182 drivers/block/pktcdvd.c 	bdget(pd->bdev->bd_dev);
pd               2183 drivers/block/pktcdvd.c 	ret = blkdev_get(pd->bdev, FMODE_READ | FMODE_EXCL, pd);
pd               2187 drivers/block/pktcdvd.c 	ret = pkt_get_last_written(pd, &lba);
pd               2189 drivers/block/pktcdvd.c 		pkt_err(pd, "pkt_get_last_written failed\n");
pd               2193 drivers/block/pktcdvd.c 	set_capacity(pd->disk, lba << 2);
pd               2194 drivers/block/pktcdvd.c 	set_capacity(pd->bdev->bd_disk, lba << 2);
pd               2195 drivers/block/pktcdvd.c 	bd_set_size(pd->bdev, (loff_t)lba << 11);
pd               2197 drivers/block/pktcdvd.c 	q = bdev_get_queue(pd->bdev);
pd               2199 drivers/block/pktcdvd.c 		ret = pkt_open_write(pd);
pd               2206 drivers/block/pktcdvd.c 		blk_queue_max_hw_sectors(q, pd->settings.size);
pd               2207 drivers/block/pktcdvd.c 		set_bit(PACKET_WRITABLE, &pd->flags);
pd               2209 drivers/block/pktcdvd.c 		pkt_set_speed(pd, MAX_SPEED, MAX_SPEED);
pd               2210 drivers/block/pktcdvd.c 		clear_bit(PACKET_WRITABLE, &pd->flags);
pd               2213 drivers/block/pktcdvd.c 	ret = pkt_set_segment_merging(pd, q);
pd               2218 drivers/block/pktcdvd.c 		if (!pkt_grow_pktlist(pd, CONFIG_CDROM_PKTCDVD_BUFFERS)) {
pd               2219 drivers/block/pktcdvd.c 			pkt_err(pd, "not enough memory for buffers\n");
pd               2223 drivers/block/pktcdvd.c 		pkt_info(pd, "%lukB available on disc\n", lba << 1);
pd               2229 drivers/block/pktcdvd.c 	blkdev_put(pd->bdev, FMODE_READ | FMODE_EXCL);
pd               2238 drivers/block/pktcdvd.c static void pkt_release_dev(struct pktcdvd_device *pd, int flush)
pd               2240 drivers/block/pktcdvd.c 	if (flush && pkt_flush_cache(pd))
pd               2241 drivers/block/pktcdvd.c 		pkt_dbg(1, pd, "not flushing cache\n");
pd               2243 drivers/block/pktcdvd.c 	pkt_lock_door(pd, 0);
pd               2245 drivers/block/pktcdvd.c 	pkt_set_speed(pd, MAX_SPEED, MAX_SPEED);
pd               2246 drivers/block/pktcdvd.c 	blkdev_put(pd->bdev, FMODE_READ | FMODE_EXCL);
pd               2248 drivers/block/pktcdvd.c 	pkt_shrink_pktlist(pd);
pd               2262 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = NULL;
pd               2267 drivers/block/pktcdvd.c 	pd = pkt_find_dev_from_minor(MINOR(bdev->bd_dev));
pd               2268 drivers/block/pktcdvd.c 	if (!pd) {
pd               2272 drivers/block/pktcdvd.c 	BUG_ON(pd->refcnt < 0);
pd               2274 drivers/block/pktcdvd.c 	pd->refcnt++;
pd               2275 drivers/block/pktcdvd.c 	if (pd->refcnt > 1) {
pd               2277 drivers/block/pktcdvd.c 		    !test_bit(PACKET_WRITABLE, &pd->flags)) {
pd               2282 drivers/block/pktcdvd.c 		ret = pkt_open_dev(pd, mode & FMODE_WRITE);
pd               2297 drivers/block/pktcdvd.c 	pd->refcnt--;
pd               2306 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = disk->private_data;
pd               2310 drivers/block/pktcdvd.c 	pd->refcnt--;
pd               2311 drivers/block/pktcdvd.c 	BUG_ON(pd->refcnt < 0);
pd               2312 drivers/block/pktcdvd.c 	if (pd->refcnt == 0) {
pd               2313 drivers/block/pktcdvd.c 		int flush = test_bit(PACKET_WRITABLE, &pd->flags);
pd               2314 drivers/block/pktcdvd.c 		pkt_release_dev(pd, flush);
pd               2324 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = psd->pd;
pd               2330 drivers/block/pktcdvd.c 	pkt_bio_finished(pd);
pd               2333 drivers/block/pktcdvd.c static void pkt_make_request_read(struct pktcdvd_device *pd, struct bio *bio)
pd               2338 drivers/block/pktcdvd.c 	psd->pd = pd;
pd               2340 drivers/block/pktcdvd.c 	bio_set_dev(cloned_bio, pd->bdev);
pd               2343 drivers/block/pktcdvd.c 	pd->stats.secs_r += bio_sectors(bio);
pd               2344 drivers/block/pktcdvd.c 	pkt_queue_bio(pd, cloned_bio);
pd               2349 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = q->queuedata;
pd               2355 drivers/block/pktcdvd.c 	zone = get_zone(bio->bi_iter.bi_sector, pd);
pd               2361 drivers/block/pktcdvd.c 	spin_lock(&pd->cdrw.active_list_lock);
pd               2363 drivers/block/pktcdvd.c 	list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) {
pd               2374 drivers/block/pktcdvd.c 					wake_up(&pd->wqueue);
pd               2377 drivers/block/pktcdvd.c 				spin_unlock(&pd->cdrw.active_list_lock);
pd               2385 drivers/block/pktcdvd.c 	spin_unlock(&pd->cdrw.active_list_lock);
pd               2392 drivers/block/pktcdvd.c 	spin_lock(&pd->lock);
pd               2393 drivers/block/pktcdvd.c 	if (pd->write_congestion_on > 0
pd               2394 drivers/block/pktcdvd.c 	    && pd->bio_queue_size >= pd->write_congestion_on) {
pd               2397 drivers/block/pktcdvd.c 			spin_unlock(&pd->lock);
pd               2399 drivers/block/pktcdvd.c 			spin_lock(&pd->lock);
pd               2400 drivers/block/pktcdvd.c 		} while(pd->bio_queue_size > pd->write_congestion_off);
pd               2402 drivers/block/pktcdvd.c 	spin_unlock(&pd->lock);
pd               2407 drivers/block/pktcdvd.c 	node = mempool_alloc(&pd->rb_pool, GFP_NOIO);
pd               2409 drivers/block/pktcdvd.c 	spin_lock(&pd->lock);
pd               2410 drivers/block/pktcdvd.c 	BUG_ON(pd->bio_queue_size < 0);
pd               2411 drivers/block/pktcdvd.c 	was_empty = (pd->bio_queue_size == 0);
pd               2412 drivers/block/pktcdvd.c 	pkt_rbtree_insert(pd, node);
pd               2413 drivers/block/pktcdvd.c 	spin_unlock(&pd->lock);
pd               2418 drivers/block/pktcdvd.c 	atomic_set(&pd->scan_queue, 1);
pd               2421 drivers/block/pktcdvd.c 		wake_up(&pd->wqueue);
pd               2422 drivers/block/pktcdvd.c 	} else if (!list_empty(&pd->cdrw.pkt_free_list) && !blocked_bio) {
pd               2427 drivers/block/pktcdvd.c 		wake_up(&pd->wqueue);
pd               2433 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd;
pd               2439 drivers/block/pktcdvd.c 	pd = q->queuedata;
pd               2440 drivers/block/pktcdvd.c 	if (!pd) {
pd               2445 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "start = %6llx stop = %6llx\n",
pd               2453 drivers/block/pktcdvd.c 		pkt_make_request_read(pd, bio);
pd               2457 drivers/block/pktcdvd.c 	if (!test_bit(PACKET_WRITABLE, &pd->flags)) {
pd               2458 drivers/block/pktcdvd.c 		pkt_notice(pd, "WRITE for ro device (%llu)\n",
pd               2464 drivers/block/pktcdvd.c 		pkt_err(pd, "wrong bio size\n");
pd               2469 drivers/block/pktcdvd.c 		sector_t zone = get_zone(bio->bi_iter.bi_sector, pd);
pd               2470 drivers/block/pktcdvd.c 		sector_t last_zone = get_zone(bio_end_sector(bio) - 1, pd);
pd               2473 drivers/block/pktcdvd.c 			BUG_ON(last_zone != zone + pd->settings.size);
pd               2492 drivers/block/pktcdvd.c static void pkt_init_queue(struct pktcdvd_device *pd)
pd               2494 drivers/block/pktcdvd.c 	struct request_queue *q = pd->disk->queue;
pd               2499 drivers/block/pktcdvd.c 	q->queuedata = pd;
pd               2504 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = m->private;
pd               2509 drivers/block/pktcdvd.c 	seq_printf(m, "Writer %s mapped to %s:\n", pd->name,
pd               2510 drivers/block/pktcdvd.c 		   bdevname(pd->bdev, bdev_buf));
pd               2513 drivers/block/pktcdvd.c 	seq_printf(m, "\tpacket size:\t\t%dkB\n", pd->settings.size / 2);
pd               2515 drivers/block/pktcdvd.c 	if (pd->settings.write_type == 0)
pd               2521 drivers/block/pktcdvd.c 	seq_printf(m, "\tpacket type:\t\t%s\n", pd->settings.fp ? "Fixed" : "Variable");
pd               2522 drivers/block/pktcdvd.c 	seq_printf(m, "\tlink loss:\t\t%d\n", pd->settings.link_loss);
pd               2524 drivers/block/pktcdvd.c 	seq_printf(m, "\ttrack mode:\t\t%d\n", pd->settings.track_mode);
pd               2526 drivers/block/pktcdvd.c 	if (pd->settings.block_mode == PACKET_BLOCK_MODE1)
pd               2528 drivers/block/pktcdvd.c 	else if (pd->settings.block_mode == PACKET_BLOCK_MODE2)
pd               2535 drivers/block/pktcdvd.c 	seq_printf(m, "\tpackets started:\t%lu\n", pd->stats.pkt_started);
pd               2536 drivers/block/pktcdvd.c 	seq_printf(m, "\tpackets ended:\t\t%lu\n", pd->stats.pkt_ended);
pd               2537 drivers/block/pktcdvd.c 	seq_printf(m, "\twritten:\t\t%lukB\n", pd->stats.secs_w >> 1);
pd               2538 drivers/block/pktcdvd.c 	seq_printf(m, "\tread gather:\t\t%lukB\n", pd->stats.secs_rg >> 1);
pd               2539 drivers/block/pktcdvd.c 	seq_printf(m, "\tread:\t\t\t%lukB\n", pd->stats.secs_r >> 1);
pd               2542 drivers/block/pktcdvd.c 	seq_printf(m, "\treference count:\t%d\n", pd->refcnt);
pd               2543 drivers/block/pktcdvd.c 	seq_printf(m, "\tflags:\t\t\t0x%lx\n", pd->flags);
pd               2544 drivers/block/pktcdvd.c 	seq_printf(m, "\tread speed:\t\t%ukB/s\n", pd->read_speed);
pd               2545 drivers/block/pktcdvd.c 	seq_printf(m, "\twrite speed:\t\t%ukB/s\n", pd->write_speed);
pd               2546 drivers/block/pktcdvd.c 	seq_printf(m, "\tstart offset:\t\t%lu\n", pd->offset);
pd               2547 drivers/block/pktcdvd.c 	seq_printf(m, "\tmode page offset:\t%u\n", pd->mode_offset);
pd               2550 drivers/block/pktcdvd.c 	seq_printf(m, "\tbios queued:\t\t%d\n", pd->bio_queue_size);
pd               2551 drivers/block/pktcdvd.c 	seq_printf(m, "\tbios pending:\t\t%d\n", atomic_read(&pd->cdrw.pending_bios));
pd               2552 drivers/block/pktcdvd.c 	seq_printf(m, "\tcurrent sector:\t\t0x%llx\n", (unsigned long long)pd->current_sector);
pd               2554 drivers/block/pktcdvd.c 	pkt_count_states(pd, states);
pd               2559 drivers/block/pktcdvd.c 			pd->write_congestion_off,
pd               2560 drivers/block/pktcdvd.c 			pd->write_congestion_on);
pd               2564 drivers/block/pktcdvd.c static int pkt_new_dev(struct pktcdvd_device *pd, dev_t dev)
pd               2571 drivers/block/pktcdvd.c 	if (pd->pkt_dev == dev) {
pd               2572 drivers/block/pktcdvd.c 		pkt_err(pd, "recursive setup not allowed\n");
pd               2580 drivers/block/pktcdvd.c 			pkt_err(pd, "%s already setup\n",
pd               2585 drivers/block/pktcdvd.c 			pkt_err(pd, "can't chain pktcdvd devices\n");
pd               2604 drivers/block/pktcdvd.c 	pd->bdev = bdev;
pd               2607 drivers/block/pktcdvd.c 	pkt_init_queue(pd);
pd               2609 drivers/block/pktcdvd.c 	atomic_set(&pd->cdrw.pending_bios, 0);
pd               2610 drivers/block/pktcdvd.c 	pd->cdrw.thread = kthread_run(kcdrwd, pd, "%s", pd->name);
pd               2611 drivers/block/pktcdvd.c 	if (IS_ERR(pd->cdrw.thread)) {
pd               2612 drivers/block/pktcdvd.c 		pkt_err(pd, "can't start kernel thread\n");
pd               2617 drivers/block/pktcdvd.c 	proc_create_single_data(pd->name, 0, pkt_proc, pkt_seq_show, pd);
pd               2618 drivers/block/pktcdvd.c 	pkt_dbg(1, pd, "writer mapped to %s\n", bdevname(bdev, b));
pd               2630 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = bdev->bd_disk->private_data;
pd               2633 drivers/block/pktcdvd.c 	pkt_dbg(2, pd, "cmd %x, dev %d:%d\n",
pd               2643 drivers/block/pktcdvd.c 		if (pd->refcnt == 1)
pd               2644 drivers/block/pktcdvd.c 			pkt_lock_door(pd, 0);
pd               2654 drivers/block/pktcdvd.c 		ret = __blkdev_driver_ioctl(pd->bdev, mode, cmd, arg);
pd               2658 drivers/block/pktcdvd.c 		pkt_dbg(2, pd, "Unknown ioctl (%x)\n", cmd);
pd               2669 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd = disk->private_data;
pd               2672 drivers/block/pktcdvd.c 	if (!pd)
pd               2674 drivers/block/pktcdvd.c 	if (!pd->bdev)
pd               2676 drivers/block/pktcdvd.c 	attached_disk = pd->bdev->bd_disk;
pd               2702 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd;
pd               2716 drivers/block/pktcdvd.c 	pd = kzalloc(sizeof(struct pktcdvd_device), GFP_KERNEL);
pd               2717 drivers/block/pktcdvd.c 	if (!pd)
pd               2720 drivers/block/pktcdvd.c 	ret = mempool_init_kmalloc_pool(&pd->rb_pool, PKT_RB_POOL_SIZE,
pd               2725 drivers/block/pktcdvd.c 	INIT_LIST_HEAD(&pd->cdrw.pkt_free_list);
pd               2726 drivers/block/pktcdvd.c 	INIT_LIST_HEAD(&pd->cdrw.pkt_active_list);
pd               2727 drivers/block/pktcdvd.c 	spin_lock_init(&pd->cdrw.active_list_lock);
pd               2729 drivers/block/pktcdvd.c 	spin_lock_init(&pd->lock);
pd               2730 drivers/block/pktcdvd.c 	spin_lock_init(&pd->iosched.lock);
pd               2731 drivers/block/pktcdvd.c 	bio_list_init(&pd->iosched.read_queue);
pd               2732 drivers/block/pktcdvd.c 	bio_list_init(&pd->iosched.write_queue);
pd               2733 drivers/block/pktcdvd.c 	sprintf(pd->name, DRIVER_NAME"%d", idx);
pd               2734 drivers/block/pktcdvd.c 	init_waitqueue_head(&pd->wqueue);
pd               2735 drivers/block/pktcdvd.c 	pd->bio_queue = RB_ROOT;
pd               2737 drivers/block/pktcdvd.c 	pd->write_congestion_on  = write_congestion_on;
pd               2738 drivers/block/pktcdvd.c 	pd->write_congestion_off = write_congestion_off;
pd               2744 drivers/block/pktcdvd.c 	pd->disk = disk;
pd               2749 drivers/block/pktcdvd.c 	strcpy(disk->disk_name, pd->name);
pd               2751 drivers/block/pktcdvd.c 	disk->private_data = pd;
pd               2756 drivers/block/pktcdvd.c 	pd->pkt_dev = MKDEV(pktdev_major, idx);
pd               2757 drivers/block/pktcdvd.c 	ret = pkt_new_dev(pd, dev);
pd               2762 drivers/block/pktcdvd.c 	disk->events = pd->bdev->bd_disk->events;
pd               2766 drivers/block/pktcdvd.c 	pkt_sysfs_dev_new(pd);
pd               2767 drivers/block/pktcdvd.c 	pkt_debugfs_dev_new(pd);
pd               2769 drivers/block/pktcdvd.c 	pkt_devs[idx] = pd;
pd               2771 drivers/block/pktcdvd.c 		*pkt_dev = pd->pkt_dev;
pd               2779 drivers/block/pktcdvd.c 	mempool_exit(&pd->rb_pool);
pd               2780 drivers/block/pktcdvd.c 	kfree(pd);
pd               2792 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd;
pd               2799 drivers/block/pktcdvd.c 		pd = pkt_devs[idx];
pd               2800 drivers/block/pktcdvd.c 		if (pd && (pd->pkt_dev == pkt_dev))
pd               2809 drivers/block/pktcdvd.c 	if (pd->refcnt > 0) {
pd               2813 drivers/block/pktcdvd.c 	if (!IS_ERR(pd->cdrw.thread))
pd               2814 drivers/block/pktcdvd.c 		kthread_stop(pd->cdrw.thread);
pd               2818 drivers/block/pktcdvd.c 	pkt_debugfs_dev_remove(pd);
pd               2819 drivers/block/pktcdvd.c 	pkt_sysfs_dev_remove(pd);
pd               2821 drivers/block/pktcdvd.c 	blkdev_put(pd->bdev, FMODE_READ | FMODE_NDELAY);
pd               2823 drivers/block/pktcdvd.c 	remove_proc_entry(pd->name, pkt_proc);
pd               2824 drivers/block/pktcdvd.c 	pkt_dbg(1, pd, "writer unmapped\n");
pd               2826 drivers/block/pktcdvd.c 	del_gendisk(pd->disk);
pd               2827 drivers/block/pktcdvd.c 	blk_cleanup_queue(pd->disk->queue);
pd               2828 drivers/block/pktcdvd.c 	put_disk(pd->disk);
pd               2830 drivers/block/pktcdvd.c 	mempool_exit(&pd->rb_pool);
pd               2831 drivers/block/pktcdvd.c 	kfree(pd);
pd               2843 drivers/block/pktcdvd.c 	struct pktcdvd_device *pd;
pd               2847 drivers/block/pktcdvd.c 	pd = pkt_find_dev_from_minor(ctrl_cmd->dev_index);
pd               2848 drivers/block/pktcdvd.c 	if (pd) {
pd               2849 drivers/block/pktcdvd.c 		ctrl_cmd->dev = new_encode_dev(pd->bdev->bd_dev);
pd               2850 drivers/block/pktcdvd.c 		ctrl_cmd->pkt_dev = new_encode_dev(pd->pkt_dev);
pd                 85 drivers/cdrom/gdrom.c static struct platform_device *pd;
pd                850 drivers/cdrom/gdrom.c 	pd = platform_device_register_simple(GDROM_DEV_NAME, -1, NULL, 0);
pd                851 drivers/cdrom/gdrom.c 	if (IS_ERR(pd)) {
pd                853 drivers/cdrom/gdrom.c 		return PTR_ERR(pd);
pd                860 drivers/cdrom/gdrom.c 	platform_device_unregister(pd);
pd                 83 drivers/clk/davinci/psc.c 	u32 pd;
pd                121 drivers/clk/davinci/psc.c 	regmap_read(lpsc->regmap, PDSTAT(lpsc->pd), &pdstat);
pd                123 drivers/clk/davinci/psc.c 		regmap_write_bits(lpsc->regmap, PDCTL(lpsc->pd), PDCTL_NEXT,
pd                126 drivers/clk/davinci/psc.c 		regmap_write(lpsc->regmap, PTCMD, BIT(lpsc->pd));
pd                129 drivers/clk/davinci/psc.c 					 epcpr & BIT(lpsc->pd), 0, 0);
pd                131 drivers/clk/davinci/psc.c 		regmap_write_bits(lpsc->regmap, PDCTL(lpsc->pd), PDCTL_EPCGOOD,
pd                134 drivers/clk/davinci/psc.c 		regmap_write(lpsc->regmap, PTCMD, BIT(lpsc->pd));
pd                138 drivers/clk/davinci/psc.c 				 !(ptstat & BIT(lpsc->pd)), 0, 0);
pd                236 drivers/clk/davinci/psc.c 			  u32 md, u32 pd, u32 flags)
pd                263 drivers/clk/davinci/psc.c 	lpsc->pd = pd;
pd                410 drivers/clk/davinci/psc.c 						 regmap, info->md, info->pd,
pd                 65 drivers/clk/davinci/psc.h 	u32 pd;
pd                 75 drivers/clk/davinci/psc.h 	.pd	= (d),		\
pd                 58 drivers/clk/imx/clk-pllv1.c 	unsigned int mfi, mfn, mfd, pd;
pd                 77 drivers/clk/imx/clk-pllv1.c 	pd =  (reg >> 26) & 0xf;
pd                 96 drivers/clk/imx/clk-pllv1.c 	rate /= pd + 1;
pd               1526 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd               1535 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd               1544 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd               1553 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd               1562 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd               1571 drivers/clk/qcom/camcc-sdm845.c 	.pd = {
pd                572 drivers/clk/qcom/dispcc-sdm845.c 	.pd = {
pd               3250 drivers/clk/qcom/gcc-apq8084.c 	.pd = {
pd               3258 drivers/clk/qcom/gcc-apq8084.c 	.pd = {
pd               3266 drivers/clk/qcom/gcc-apq8084.c 	.pd = {
pd               3274 drivers/clk/qcom/gcc-apq8084.c 	.pd = {
pd               3052 drivers/clk/qcom/gcc-msm8916.c 	.pd = {
pd               3060 drivers/clk/qcom/gcc-msm8916.c 	.pd = {
pd               3068 drivers/clk/qcom/gcc-msm8916.c 	.pd = {
pd               3076 drivers/clk/qcom/gcc-msm8916.c 	.pd = {
pd               3084 drivers/clk/qcom/gcc-msm8916.c 	.pd = {
pd               2427 drivers/clk/qcom/gcc-msm8974.c 	.pd = {
pd               3207 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3216 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3225 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3234 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3243 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3251 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3259 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3267 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               3275 drivers/clk/qcom/gcc-msm8996.c 	.pd = {
pd               2644 drivers/clk/qcom/gcc-msm8998.c 	.pd = {
pd               2654 drivers/clk/qcom/gcc-msm8998.c 	.pd = {
pd               2664 drivers/clk/qcom/gcc-msm8998.c 	.pd = {
pd               2228 drivers/clk/qcom/gcc-sdm660.c 	.pd = {
pd               2238 drivers/clk/qcom/gcc-sdm660.c 	.pd = {
pd               2248 drivers/clk/qcom/gcc-sdm660.c 	.pd = {
pd               3200 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3209 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3218 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3227 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3236 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3245 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3254 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3263 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3272 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3281 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3290 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3299 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd               3308 drivers/clk/qcom/gcc-sdm845.c 	.pd = {
pd                 43 drivers/clk/qcom/gdsc.c #define domain_to_gdsc(domain) container_of(domain, struct gdsc, pd)
pd                145 drivers/clk/qcom/gdsc.c 	WARN(ret, "%s status stuck at 'o%s'", sc->pd.name, status ? "ff" : "n");
pd                332 drivers/clk/qcom/gdsc.c 		gdsc_enable(&sc->pd);
pd                337 drivers/clk/qcom/gdsc.c 			gdsc_enable(&sc->pd);
pd                339 drivers/clk/qcom/gdsc.c 		sc->pd.flags |= GENPD_FLAG_ALWAYS_ON;
pd                347 drivers/clk/qcom/gdsc.c 	if (!sc->pd.power_off)
pd                348 drivers/clk/qcom/gdsc.c 		sc->pd.power_off = gdsc_disable;
pd                349 drivers/clk/qcom/gdsc.c 	if (!sc->pd.power_on)
pd                350 drivers/clk/qcom/gdsc.c 		sc->pd.power_on = gdsc_enable;
pd                351 drivers/clk/qcom/gdsc.c 	pm_genpd_init(&sc->pd, NULL, !on);
pd                383 drivers/clk/qcom/gdsc.c 		data->domains[i] = &scs[i]->pd;
pd                391 drivers/clk/qcom/gdsc.c 			pm_genpd_add_subdomain(scs[i]->parent, &scs[i]->pd);
pd                409 drivers/clk/qcom/gdsc.c 			pm_genpd_remove_subdomain(scs[i]->parent, &scs[i]->pd);
pd                 29 drivers/clk/qcom/gdsc.h 	struct generic_pm_domain	pd;
pd                127 drivers/clk/qcom/gpucc-sdm845.c 	.pd = {
pd                162 drivers/clk/qcom/gpucc-sdm845.c 	.pd = {
pd               3059 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3067 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3075 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3085 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3093 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3103 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3113 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               3121 drivers/clk/qcom/mmcc-apq8084.c 	.pd = {
pd               2342 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2352 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2362 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2372 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2382 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2390 drivers/clk/qcom/mmcc-msm8974.c 	.pd = {
pd               2393 drivers/clk/qcom/mmcc-msm8974.c 	.parent = &oxili_gdsc.pd,
pd               2899 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2909 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2919 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2929 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2940 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2943 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &mmagic_video_gdsc.pd,
pd               2951 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2954 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &venus_gdsc.pd,
pd               2963 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2966 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &venus_gdsc.pd,
pd               2975 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2978 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &mmagic_camss_gdsc.pd,
pd               2986 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               2989 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &camss_gdsc.pd,
pd               2997 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3000 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &camss_gdsc.pd,
pd               3008 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3011 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &camss_gdsc.pd,
pd               3019 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3022 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &camss_gdsc.pd,
pd               3030 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3033 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &camss_gdsc.pd,
pd               3041 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3044 drivers/clk/qcom/mmcc-msm8996.c 	.parent = &mmagic_mdss_gdsc.pd,
pd               3051 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd               3063 drivers/clk/qcom/mmcc-msm8996.c 	.pd = {
pd                248 drivers/clk/qcom/videocc-sdm845.c 	.pd = {
pd                259 drivers/clk/qcom/videocc-sdm845.c 	.pd = {
pd                270 drivers/clk/qcom/videocc-sdm845.c 	.pd = {
pd                324 drivers/clk/renesas/clk-mstp.c 	struct generic_pm_domain *pd;
pd                332 drivers/clk/renesas/clk-mstp.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                333 drivers/clk/renesas/clk-mstp.c 	if (!pd)
pd                336 drivers/clk/renesas/clk-mstp.c 	pd->name = np->name;
pd                337 drivers/clk/renesas/clk-mstp.c 	pd->flags = GENPD_FLAG_PM_CLK | GENPD_FLAG_ALWAYS_ON |
pd                339 drivers/clk/renesas/clk-mstp.c 	pd->attach_dev = cpg_mstp_attach_dev;
pd                340 drivers/clk/renesas/clk-mstp.c 	pd->detach_dev = cpg_mstp_detach_dev;
pd                341 drivers/clk/renesas/clk-mstp.c 	pm_genpd_init(pd, &pm_domain_always_on_gov, false);
pd                343 drivers/clk/renesas/clk-mstp.c 	of_genpd_add_provider_simple(np, pd);
pd                380 drivers/clk/renesas/r9a06g032-clocks.c static int r9a06g032_attach_dev(struct generic_pm_domain *pd,
pd                391 drivers/clk/renesas/r9a06g032-clocks.c 		if (clkspec.np != pd->dev.of_node)
pd                417 drivers/clk/renesas/r9a06g032-clocks.c 	struct generic_pm_domain *pd;
pd                419 drivers/clk/renesas/r9a06g032-clocks.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                420 drivers/clk/renesas/r9a06g032-clocks.c 	if (!pd)
pd                423 drivers/clk/renesas/r9a06g032-clocks.c 	pd->name = np->name;
pd                424 drivers/clk/renesas/r9a06g032-clocks.c 	pd->flags = GENPD_FLAG_PM_CLK | GENPD_FLAG_ALWAYS_ON |
pd                426 drivers/clk/renesas/r9a06g032-clocks.c 	pd->attach_dev = r9a06g032_attach_dev;
pd                427 drivers/clk/renesas/r9a06g032-clocks.c 	pd->detach_dev = r9a06g032_detach_dev;
pd                428 drivers/clk/renesas/r9a06g032-clocks.c 	pm_genpd_init(pd, &pm_domain_always_on_gov, false);
pd                430 drivers/clk/renesas/r9a06g032-clocks.c 	of_genpd_add_provider_simple(np, pd);
pd                459 drivers/clk/renesas/renesas-cpg-mssr.c 			       struct cpg_mssr_clk_domain *pd)
pd                463 drivers/clk/renesas/renesas-cpg-mssr.c 	if (clkspec->np != pd->genpd.dev.of_node || clkspec->args_count != 2)
pd                468 drivers/clk/renesas/renesas-cpg-mssr.c 		for (i = 0; i < pd->num_core_pm_clks; i++)
pd                469 drivers/clk/renesas/renesas-cpg-mssr.c 			if (clkspec->args[1] == pd->core_pm_clks[i])
pd                483 drivers/clk/renesas/renesas-cpg-mssr.c 	struct cpg_mssr_clk_domain *pd = cpg_mssr_clk_domain;
pd                490 drivers/clk/renesas/renesas-cpg-mssr.c 	if (!pd) {
pd                497 drivers/clk/renesas/renesas-cpg-mssr.c 		if (cpg_mssr_is_pm_clk(&clkspec, pd))
pd                542 drivers/clk/renesas/renesas-cpg-mssr.c 	struct cpg_mssr_clk_domain *pd;
pd                545 drivers/clk/renesas/renesas-cpg-mssr.c 	pd = devm_kzalloc(dev, sizeof(*pd) + pm_size, GFP_KERNEL);
pd                546 drivers/clk/renesas/renesas-cpg-mssr.c 	if (!pd)
pd                549 drivers/clk/renesas/renesas-cpg-mssr.c 	pd->num_core_pm_clks = num_core_pm_clks;
pd                550 drivers/clk/renesas/renesas-cpg-mssr.c 	memcpy(pd->core_pm_clks, core_pm_clks, pm_size);
pd                552 drivers/clk/renesas/renesas-cpg-mssr.c 	genpd = &pd->genpd;
pd                559 drivers/clk/renesas/renesas-cpg-mssr.c 	cpg_mssr_clk_domain = pd;
pd                163 drivers/clk/sifive/fu540-prci.c 	void (*enable_bypass)(struct __prci_data *pd);
pd                164 drivers/clk/sifive/fu540-prci.c 	void (*disable_bypass)(struct __prci_data *pd);
pd                186 drivers/clk/sifive/fu540-prci.c 	struct __prci_data *pd;
pd                208 drivers/clk/sifive/fu540-prci.c static u32 __prci_readl(struct __prci_data *pd, u32 offs)
pd                210 drivers/clk/sifive/fu540-prci.c 	return readl_relaxed(pd->va + offs);
pd                213 drivers/clk/sifive/fu540-prci.c static void __prci_writel(u32 v, u32 offs, struct __prci_data *pd)
pd                215 drivers/clk/sifive/fu540-prci.c 	writel_relaxed(v, pd->va + offs);
pd                303 drivers/clk/sifive/fu540-prci.c static void __prci_wrpll_read_cfg(struct __prci_data *pd,
pd                306 drivers/clk/sifive/fu540-prci.c 	__prci_wrpll_unpack(&pwd->c, __prci_readl(pd, pwd->cfg0_offs));
pd                323 drivers/clk/sifive/fu540-prci.c static void __prci_wrpll_write_cfg(struct __prci_data *pd,
pd                327 drivers/clk/sifive/fu540-prci.c 	__prci_writel(__prci_wrpll_pack(c), pwd->cfg0_offs, pd);
pd                343 drivers/clk/sifive/fu540-prci.c static void __prci_coreclksel_use_hfclk(struct __prci_data *pd)
pd                347 drivers/clk/sifive/fu540-prci.c 	r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET);
pd                349 drivers/clk/sifive/fu540-prci.c 	__prci_writel(r, PRCI_CORECLKSEL_OFFSET, pd);
pd                351 drivers/clk/sifive/fu540-prci.c 	r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); /* barrier */
pd                363 drivers/clk/sifive/fu540-prci.c static void __prci_coreclksel_use_corepll(struct __prci_data *pd)
pd                367 drivers/clk/sifive/fu540-prci.c 	r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET);
pd                369 drivers/clk/sifive/fu540-prci.c 	__prci_writel(r, PRCI_CORECLKSEL_OFFSET, pd);
pd                371 drivers/clk/sifive/fu540-prci.c 	r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); /* barrier */
pd                411 drivers/clk/sifive/fu540-prci.c 	struct __prci_data *pd = pc->pd;
pd                419 drivers/clk/sifive/fu540-prci.c 		pwd->enable_bypass(pd);
pd                421 drivers/clk/sifive/fu540-prci.c 	__prci_wrpll_write_cfg(pd, pwd, &pwd->c);
pd                426 drivers/clk/sifive/fu540-prci.c 		pwd->disable_bypass(pd);
pd                447 drivers/clk/sifive/fu540-prci.c 	struct __prci_data *pd = pc->pd;
pd                451 drivers/clk/sifive/fu540-prci.c 	v = __prci_readl(pd, PRCI_CLKMUXSTATUSREG_OFFSET);
pd                519 drivers/clk/sifive/fu540-prci.c static int __prci_register_clocks(struct device *dev, struct __prci_data *pd)
pd                542 drivers/clk/sifive/fu540-prci.c 		pic->pd = pd;
pd                545 drivers/clk/sifive/fu540-prci.c 			__prci_wrpll_read_cfg(pd, pic->pwd);
pd                561 drivers/clk/sifive/fu540-prci.c 		pd->hw_clks.hws[i] = &pic->hw;
pd                564 drivers/clk/sifive/fu540-prci.c 	pd->hw_clks.num = i;
pd                567 drivers/clk/sifive/fu540-prci.c 					&pd->hw_clks);
pd                586 drivers/clk/sifive/fu540-prci.c 	struct __prci_data *pd;
pd                589 drivers/clk/sifive/fu540-prci.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                590 drivers/clk/sifive/fu540-prci.c 	if (!pd)
pd                594 drivers/clk/sifive/fu540-prci.c 	pd->va = devm_ioremap_resource(dev, res);
pd                595 drivers/clk/sifive/fu540-prci.c 	if (IS_ERR(pd->va))
pd                596 drivers/clk/sifive/fu540-prci.c 		return PTR_ERR(pd->va);
pd                598 drivers/clk/sifive/fu540-prci.c 	r = __prci_register_clocks(dev, pd);
pd                 64 drivers/clocksource/timer-stm32.c 	struct stm32_timer_private *pd = to->private_data;
pd                 66 drivers/clocksource/timer-stm32.c 	pd->bits = bits;
pd                 80 drivers/clocksource/timer-stm32.c 	struct stm32_timer_private *pd = to->private_data;
pd                 82 drivers/clocksource/timer-stm32.c 	return pd->bits;
pd                198 drivers/crypto/amcc/crypto4xx_core.c 		struct ce_pd *pd = &dev->pdr[i];
pd                201 drivers/crypto/amcc/crypto4xx_core.c 		pd->sa = dev->shadow_sa_pool_pa +
pd                446 drivers/crypto/amcc/crypto4xx_core.c 				      struct ce_pd *pd,
pd                520 drivers/crypto/amcc/crypto4xx_core.c 				     struct ce_pd *pd)
pd                529 drivers/crypto/amcc/crypto4xx_core.c 		crypto4xx_copy_pkt_to_dst(dev, pd, pd_uinfo,
pd                572 drivers/crypto/amcc/crypto4xx_core.c 				struct ce_pd *pd)
pd                583 drivers/crypto/amcc/crypto4xx_core.c 		crypto4xx_copy_pkt_to_dst(dev, pd, pd_uinfo,
pd                584 drivers/crypto/amcc/crypto4xx_core.c 					  pd->pd_ctl_len.bf.pkt_len,
pd                587 drivers/crypto/amcc/crypto4xx_core.c 		dma_unmap_page(dev->core_dev->device, pd->dest, dst->length,
pd                612 drivers/crypto/amcc/crypto4xx_core.c 	if (pd->pd_ctl.bf.status & 0xff) {
pd                614 drivers/crypto/amcc/crypto4xx_core.c 			if (pd->pd_ctl.bf.status & 2)
pd                616 drivers/crypto/amcc/crypto4xx_core.c 			if (pd->pd_ctl.bf.status & 4)
pd                618 drivers/crypto/amcc/crypto4xx_core.c 			if (pd->pd_ctl.bf.status & 8)
pd                621 drivers/crypto/amcc/crypto4xx_core.c 				pd->pd_ctl.bf.status & 0xff);
pd                623 drivers/crypto/amcc/crypto4xx_core.c 				pd->pd_ctl.bf.pd_pad_ctl);
pd                636 drivers/crypto/amcc/crypto4xx_core.c 	struct ce_pd *pd = &dev->pdr[idx];
pd                641 drivers/crypto/amcc/crypto4xx_core.c 		crypto4xx_cipher_done(dev, pd_uinfo, pd);
pd                644 drivers/crypto/amcc/crypto4xx_core.c 		crypto4xx_aead_done(dev, pd_uinfo, pd);
pd                692 drivers/crypto/amcc/crypto4xx_core.c 	struct ce_pd *pd;
pd                808 drivers/crypto/amcc/crypto4xx_core.c 	pd = &dev->pdr[pd_entry];
pd                809 drivers/crypto/amcc/crypto4xx_core.c 	pd->sa_len = sa_len;
pd                835 drivers/crypto/amcc/crypto4xx_core.c 		pd->src = gd_dma;
pd                859 drivers/crypto/amcc/crypto4xx_core.c 		pd->src = (u32)dma_map_page(dev->core_dev->device, sg_page(src),
pd                878 drivers/crypto/amcc/crypto4xx_core.c 		pd->dest = (u32)dma_map_page(dev->core_dev->device,
pd                891 drivers/crypto/amcc/crypto4xx_core.c 		pd->dest = sd_dma;
pd                918 drivers/crypto/amcc/crypto4xx_core.c 	pd->pd_ctl.w = PD_CTL_HOST_READY |
pd                922 drivers/crypto/amcc/crypto4xx_core.c 	pd->pd_ctl_len.w = 0x00400000 | (assoclen + datalen);
pd               1079 drivers/crypto/amcc/crypto4xx_core.c 	struct ce_pd *pd;
pd               1085 drivers/crypto/amcc/crypto4xx_core.c 		pd = &core_dev->dev->pdr[tail];
pd               1087 drivers/crypto/amcc/crypto4xx_core.c 		     ((READ_ONCE(pd->pd_ctl.w) &
pd                285 drivers/dma/amba-pl08x.c 	struct pl08x_platform_data *pd;
pd                335 drivers/dma/amba-pl08x.c 	const struct pl08x_platform_data *pd = plchan->host->pd;
pd                338 drivers/dma/amba-pl08x.c 	if (plchan->mux_use++ == 0 && pd->get_xfer_signal) {
pd                339 drivers/dma/amba-pl08x.c 		ret = pd->get_xfer_signal(plchan->cd);
pd                352 drivers/dma/amba-pl08x.c 	const struct pl08x_platform_data *pd = plchan->host->pd;
pd                357 drivers/dma/amba-pl08x.c 		if (--plchan->mux_use == 0 && pd->put_xfer_signal) {
pd                358 drivers/dma/amba-pl08x.c 			pd->put_xfer_signal(plchan->cd, plchan->signal);
pd                463 drivers/dma/amba-pl08x.c 		switch (pl08x->pd->memcpy_burst_size) {
pd                500 drivers/dma/amba-pl08x.c 		if (pl08x->pd->memcpy_prot_buff)
pd                502 drivers/dma/amba-pl08x.c 		if (pl08x->pd->memcpy_prot_cache)
pd               1766 drivers/dma/amba-pl08x.c 	switch (pl08x->pd->memcpy_burst_size) {
pd               1805 drivers/dma/amba-pl08x.c 	switch (pl08x->pd->memcpy_bus_width) {
pd               1825 drivers/dma/amba-pl08x.c 	if (pl08x->pd->memcpy_prot_buff)
pd               1827 drivers/dma/amba-pl08x.c 	if (pl08x->pd->memcpy_prot_cache)
pd               1849 drivers/dma/amba-pl08x.c 	switch (pl08x->pd->memcpy_bus_width) {
pd               2392 drivers/dma/amba-pl08x.c 			chan->cd = &pl08x->pd->slave_channels[i];
pd               2407 drivers/dma/amba-pl08x.c 			chan->cd->periph_buses = pl08x->pd->mem_buses;
pd               2576 drivers/dma/amba-pl08x.c 	struct pl08x_platform_data *pd;
pd               2582 drivers/dma/amba-pl08x.c 	pd = devm_kzalloc(&adev->dev, sizeof(*pd), GFP_KERNEL);
pd               2583 drivers/dma/amba-pl08x.c 	if (!pd)
pd               2588 drivers/dma/amba-pl08x.c 		pd->lli_buses |= PL08X_AHB1;
pd               2590 drivers/dma/amba-pl08x.c 		pd->lli_buses |= PL08X_AHB2;
pd               2591 drivers/dma/amba-pl08x.c 	if (!pd->lli_buses) {
pd               2593 drivers/dma/amba-pl08x.c 		pd->lli_buses |= PL08X_AHB1 | PL08X_AHB2;
pd               2598 drivers/dma/amba-pl08x.c 		pd->mem_buses |= PL08X_AHB1;
pd               2600 drivers/dma/amba-pl08x.c 		pd->mem_buses |= PL08X_AHB2;
pd               2601 drivers/dma/amba-pl08x.c 	if (!pd->mem_buses) {
pd               2603 drivers/dma/amba-pl08x.c 		pd->mem_buses |= PL08X_AHB1 | PL08X_AHB2;
pd               2617 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_1;
pd               2620 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_4;
pd               2623 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_8;
pd               2626 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_16;
pd               2629 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_32;
pd               2632 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_64;
pd               2635 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_128;
pd               2638 drivers/dma/amba-pl08x.c 		pd->memcpy_burst_size = PL08X_BURST_SZ_256;
pd               2652 drivers/dma/amba-pl08x.c 		pd->memcpy_bus_width = PL08X_BUS_WIDTH_8_BITS;
pd               2655 drivers/dma/amba-pl08x.c 		pd->memcpy_bus_width = PL08X_BUS_WIDTH_16_BITS;
pd               2658 drivers/dma/amba-pl08x.c 		pd->memcpy_bus_width = PL08X_BUS_WIDTH_32_BITS;
pd               2676 drivers/dma/amba-pl08x.c 		pd->slave_channels = chanp;
pd               2684 drivers/dma/amba-pl08x.c 		pd->num_slave_channels = pl08x->vd->signals;
pd               2687 drivers/dma/amba-pl08x.c 	pl08x->pd = pd;
pd               2810 drivers/dma/amba-pl08x.c 	pl08x->pd = dev_get_platdata(&adev->dev);
pd               2811 drivers/dma/amba-pl08x.c 	if (!pl08x->pd) {
pd               2822 drivers/dma/amba-pl08x.c 		pl08x->slave.filter.map = pl08x->pd->slave_map;
pd               2823 drivers/dma/amba-pl08x.c 		pl08x->slave.filter.mapcnt = pl08x->pd->slave_map_len;
pd               2831 drivers/dma/amba-pl08x.c 		pl08x->lli_buses = pl08x->pd->lli_buses;
pd               2832 drivers/dma/amba-pl08x.c 		pl08x->mem_buses = pl08x->pd->mem_buses;
pd               2934 drivers/dma/amba-pl08x.c 					pl08x->pd->num_slave_channels, true);
pd                284 drivers/dma/imx-sdma.c 	u32  pd;
pd                132 drivers/dma/pch_dma.c #define dma_readl(pd, name) \
pd                133 drivers/dma/pch_dma.c 	readl((pd)->membase + PCH_DMA_##name)
pd                134 drivers/dma/pch_dma.c #define dma_writel(pd, name, val) \
pd                135 drivers/dma/pch_dma.c 	writel((val), (pd)->membase + PCH_DMA_##name)
pd                179 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(chan->device);
pd                188 drivers/dma/pch_dma.c 	val = dma_readl(pd, CTL2);
pd                195 drivers/dma/pch_dma.c 	dma_writel(pd, CTL2, val);
pd                204 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(chan->device);
pd                210 drivers/dma/pch_dma.c 		val = dma_readl(pd, CTL0);
pd                225 drivers/dma/pch_dma.c 		dma_writel(pd, CTL0, val);
pd                228 drivers/dma/pch_dma.c 		val = dma_readl(pd, CTL3);
pd                242 drivers/dma/pch_dma.c 		dma_writel(pd, CTL3, val);
pd                251 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(chan->device);
pd                261 drivers/dma/pch_dma.c 		val = dma_readl(pd, CTL0);
pd                265 drivers/dma/pch_dma.c 		dma_writel(pd, CTL0, val);
pd                272 drivers/dma/pch_dma.c 		val = dma_readl(pd, CTL3);
pd                276 drivers/dma/pch_dma.c 		dma_writel(pd, CTL3, val);
pd                285 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(pd_chan->chan.device);
pd                288 drivers/dma/pch_dma.c 	val = dma_readl(pd, STS0);
pd                295 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(pd_chan->chan.device);
pd                298 drivers/dma/pch_dma.c 	val = dma_readl(pd, STS2);
pd                429 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(chan->device);
pd                432 drivers/dma/pch_dma.c 	desc = dma_pool_zalloc(pd->pool, flags, &addr);
pd                530 drivers/dma/pch_dma.c 	struct pch_dma *pd = to_pd(chan->device);
pd                544 drivers/dma/pch_dma.c 		dma_pool_free(pd->pool, desc, desc->txd.phys);
pd                694 drivers/dma/pch_dma.c 	struct pch_dma *pd = (struct pch_dma *)devid;
pd                702 drivers/dma/pch_dma.c 	sts0 = dma_readl(pd, STS0);
pd                703 drivers/dma/pch_dma.c 	sts2 = dma_readl(pd, STS2);
pd                705 drivers/dma/pch_dma.c 	dev_dbg(pd->dma.dev, "pd_irq sts0: %x\n", sts0);
pd                707 drivers/dma/pch_dma.c 	for (i = 0; i < pd->dma.chancnt; i++) {
pd                708 drivers/dma/pch_dma.c 		pd_chan = &pd->channels[i];
pd                731 drivers/dma/pch_dma.c 		dma_writel(pd, STS0, sts0);
pd                733 drivers/dma/pch_dma.c 		dma_writel(pd, STS2, sts2);
pd                739 drivers/dma/pch_dma.c static void pch_dma_save_regs(struct pch_dma *pd)
pd                745 drivers/dma/pch_dma.c 	pd->regs.dma_ctl0 = dma_readl(pd, CTL0);
pd                746 drivers/dma/pch_dma.c 	pd->regs.dma_ctl1 = dma_readl(pd, CTL1);
pd                747 drivers/dma/pch_dma.c 	pd->regs.dma_ctl2 = dma_readl(pd, CTL2);
pd                748 drivers/dma/pch_dma.c 	pd->regs.dma_ctl3 = dma_readl(pd, CTL3);
pd                750 drivers/dma/pch_dma.c 	list_for_each_entry_safe(chan, _c, &pd->dma.channels, device_node) {
pd                753 drivers/dma/pch_dma.c 		pd->ch_regs[i].dev_addr = channel_readl(pd_chan, DEV_ADDR);
pd                754 drivers/dma/pch_dma.c 		pd->ch_regs[i].mem_addr = channel_readl(pd_chan, MEM_ADDR);
pd                755 drivers/dma/pch_dma.c 		pd->ch_regs[i].size = channel_readl(pd_chan, SIZE);
pd                756 drivers/dma/pch_dma.c 		pd->ch_regs[i].next = channel_readl(pd_chan, NEXT);
pd                762 drivers/dma/pch_dma.c static void pch_dma_restore_regs(struct pch_dma *pd)
pd                768 drivers/dma/pch_dma.c 	dma_writel(pd, CTL0, pd->regs.dma_ctl0);
pd                769 drivers/dma/pch_dma.c 	dma_writel(pd, CTL1, pd->regs.dma_ctl1);
pd                770 drivers/dma/pch_dma.c 	dma_writel(pd, CTL2, pd->regs.dma_ctl2);
pd                771 drivers/dma/pch_dma.c 	dma_writel(pd, CTL3, pd->regs.dma_ctl3);
pd                773 drivers/dma/pch_dma.c 	list_for_each_entry_safe(chan, _c, &pd->dma.channels, device_node) {
pd                776 drivers/dma/pch_dma.c 		channel_writel(pd_chan, DEV_ADDR, pd->ch_regs[i].dev_addr);
pd                777 drivers/dma/pch_dma.c 		channel_writel(pd_chan, MEM_ADDR, pd->ch_regs[i].mem_addr);
pd                778 drivers/dma/pch_dma.c 		channel_writel(pd_chan, SIZE, pd->ch_regs[i].size);
pd                779 drivers/dma/pch_dma.c 		channel_writel(pd_chan, NEXT, pd->ch_regs[i].next);
pd                787 drivers/dma/pch_dma.c 	struct pch_dma *pd = pci_get_drvdata(pdev);
pd                789 drivers/dma/pch_dma.c 	if (pd)
pd                790 drivers/dma/pch_dma.c 		pch_dma_save_regs(pd);
pd                801 drivers/dma/pch_dma.c 	struct pch_dma *pd = pci_get_drvdata(pdev);
pd                813 drivers/dma/pch_dma.c 	if (pd)
pd                814 drivers/dma/pch_dma.c 		pch_dma_restore_regs(pd);
pd                823 drivers/dma/pch_dma.c 	struct pch_dma *pd;
pd                830 drivers/dma/pch_dma.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                831 drivers/dma/pch_dma.c 	if (!pd)
pd                834 drivers/dma/pch_dma.c 	pci_set_drvdata(pdev, pd);
pd                860 drivers/dma/pch_dma.c 	regs = pd->membase = pci_iomap(pdev, 1, 0);
pd                861 drivers/dma/pch_dma.c 	if (!pd->membase) {
pd                868 drivers/dma/pch_dma.c 	pd->dma.dev = &pdev->dev;
pd                870 drivers/dma/pch_dma.c 	err = request_irq(pdev->irq, pd_irq, IRQF_SHARED, DRV_NAME, pd);
pd                876 drivers/dma/pch_dma.c 	pd->pool = dma_pool_create("pch_dma_desc_pool", &pdev->dev,
pd                878 drivers/dma/pch_dma.c 	if (!pd->pool) {
pd                885 drivers/dma/pch_dma.c 	INIT_LIST_HEAD(&pd->dma.channels);
pd                888 drivers/dma/pch_dma.c 		struct pch_dma_chan *pd_chan = &pd->channels[i];
pd                890 drivers/dma/pch_dma.c 		pd_chan->chan.device = &pd->dma;
pd                903 drivers/dma/pch_dma.c 		list_add_tail(&pd_chan->chan.device_node, &pd->dma.channels);
pd                906 drivers/dma/pch_dma.c 	dma_cap_zero(pd->dma.cap_mask);
pd                907 drivers/dma/pch_dma.c 	dma_cap_set(DMA_PRIVATE, pd->dma.cap_mask);
pd                908 drivers/dma/pch_dma.c 	dma_cap_set(DMA_SLAVE, pd->dma.cap_mask);
pd                910 drivers/dma/pch_dma.c 	pd->dma.device_alloc_chan_resources = pd_alloc_chan_resources;
pd                911 drivers/dma/pch_dma.c 	pd->dma.device_free_chan_resources = pd_free_chan_resources;
pd                912 drivers/dma/pch_dma.c 	pd->dma.device_tx_status = pd_tx_status;
pd                913 drivers/dma/pch_dma.c 	pd->dma.device_issue_pending = pd_issue_pending;
pd                914 drivers/dma/pch_dma.c 	pd->dma.device_prep_slave_sg = pd_prep_slave_sg;
pd                915 drivers/dma/pch_dma.c 	pd->dma.device_terminate_all = pd_device_terminate_all;
pd                917 drivers/dma/pch_dma.c 	err = dma_async_device_register(&pd->dma);
pd                926 drivers/dma/pch_dma.c 	dma_pool_destroy(pd->pool);
pd                928 drivers/dma/pch_dma.c 	free_irq(pdev->irq, pd);
pd                930 drivers/dma/pch_dma.c 	pci_iounmap(pdev, pd->membase);
pd                936 drivers/dma/pch_dma.c 	kfree(pd);
pd                942 drivers/dma/pch_dma.c 	struct pch_dma *pd = pci_get_drvdata(pdev);
pd                946 drivers/dma/pch_dma.c 	if (pd) {
pd                947 drivers/dma/pch_dma.c 		dma_async_device_unregister(&pd->dma);
pd                949 drivers/dma/pch_dma.c 		free_irq(pdev->irq, pd);
pd                951 drivers/dma/pch_dma.c 		list_for_each_entry_safe(chan, _c, &pd->dma.channels,
pd                958 drivers/dma/pch_dma.c 		dma_pool_destroy(pd->pool);
pd                959 drivers/dma/pch_dma.c 		pci_iounmap(pdev, pd->membase);
pd                962 drivers/dma/pch_dma.c 		kfree(pd);
pd               3000 drivers/dma/pl330.c 	struct dma_device *pd;
pd               3015 drivers/dma/pl330.c 	pd = &pl330->ddma;
pd               3016 drivers/dma/pl330.c 	pd->dev = &adev->dev;
pd               3086 drivers/dma/pl330.c 	INIT_LIST_HEAD(&pd->channels);
pd               3108 drivers/dma/pl330.c 		pch->chan.device = pd;
pd               3113 drivers/dma/pl330.c 		list_add_tail(&pch->chan.device_node, &pd->channels);
pd               3116 drivers/dma/pl330.c 	dma_cap_set(DMA_MEMCPY, pd->cap_mask);
pd               3118 drivers/dma/pl330.c 		dma_cap_set(DMA_SLAVE, pd->cap_mask);
pd               3119 drivers/dma/pl330.c 		dma_cap_set(DMA_CYCLIC, pd->cap_mask);
pd               3120 drivers/dma/pl330.c 		dma_cap_set(DMA_PRIVATE, pd->cap_mask);
pd               3123 drivers/dma/pl330.c 	pd->device_alloc_chan_resources = pl330_alloc_chan_resources;
pd               3124 drivers/dma/pl330.c 	pd->device_free_chan_resources = pl330_free_chan_resources;
pd               3125 drivers/dma/pl330.c 	pd->device_prep_dma_memcpy = pl330_prep_dma_memcpy;
pd               3126 drivers/dma/pl330.c 	pd->device_prep_dma_cyclic = pl330_prep_dma_cyclic;
pd               3127 drivers/dma/pl330.c 	pd->device_tx_status = pl330_tx_status;
pd               3128 drivers/dma/pl330.c 	pd->device_prep_slave_sg = pl330_prep_slave_sg;
pd               3129 drivers/dma/pl330.c 	pd->device_config = pl330_config;
pd               3130 drivers/dma/pl330.c 	pd->device_pause = pl330_pause;
pd               3131 drivers/dma/pl330.c 	pd->device_terminate_all = pl330_terminate_all;
pd               3132 drivers/dma/pl330.c 	pd->device_issue_pending = pl330_issue_pending;
pd               3133 drivers/dma/pl330.c 	pd->src_addr_widths = PL330_DMA_BUSWIDTHS;
pd               3134 drivers/dma/pl330.c 	pd->dst_addr_widths = PL330_DMA_BUSWIDTHS;
pd               3135 drivers/dma/pl330.c 	pd->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
pd               3136 drivers/dma/pl330.c 	pd->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
pd               3137 drivers/dma/pl330.c 	pd->max_burst = ((pl330->quirks & PL330_QUIRK_BROKEN_NO_FLUSHP) ?
pd               3140 drivers/dma/pl330.c 	ret = dma_async_device_register(pd);
pd                284 drivers/firewire/net.c static bool fwnet_frag_overlap(struct fwnet_partial_datagram *pd,
pd                290 drivers/firewire/net.c 	list_for_each_entry(fi, &pd->fi_list, fi_link)
pd                299 drivers/firewire/net.c 	struct fwnet_partial_datagram *pd, unsigned offset, unsigned len)
pd                304 drivers/firewire/net.c 	list = &pd->fi_list;
pd                305 drivers/firewire/net.c 	list_for_each_entry(fi, &pd->fi_list, fi_link) {
pd                401 drivers/firewire/net.c 	struct fwnet_partial_datagram *pd;
pd                403 drivers/firewire/net.c 	list_for_each_entry(pd, &peer->pd_list, pd_link)
pd                404 drivers/firewire/net.c 		if (pd->datagram_label == datagram_label)
pd                405 drivers/firewire/net.c 			return pd;
pd                424 drivers/firewire/net.c 		struct fwnet_partial_datagram *pd, void *frag_buf,
pd                427 drivers/firewire/net.c 	if (fwnet_frag_new(pd, frag_off, frag_len) == NULL)
pd                430 drivers/firewire/net.c 	memcpy(pd->pbuf + frag_off, frag_buf, frag_len);
pd                436 drivers/firewire/net.c 	list_move_tail(&pd->pd_link, &peer->pd_list);
pd                441 drivers/firewire/net.c static bool fwnet_pd_is_complete(struct fwnet_partial_datagram *pd)
pd                445 drivers/firewire/net.c 	fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link);
pd                447 drivers/firewire/net.c 	return fi->len == pd->datagram_size;
pd                579 drivers/firewire/net.c 	struct fwnet_partial_datagram *pd;
pd                643 drivers/firewire/net.c 	pd = fwnet_pd_find(peer, datagram_label);
pd                644 drivers/firewire/net.c 	if (pd == NULL) {
pd                651 drivers/firewire/net.c 		pd = fwnet_pd_new(net, peer, datagram_label,
pd                653 drivers/firewire/net.c 		if (pd == NULL) {
pd                659 drivers/firewire/net.c 		if (fwnet_frag_overlap(pd, fg_off, len) ||
pd                660 drivers/firewire/net.c 		    pd->datagram_size != dg_size) {
pd                665 drivers/firewire/net.c 			fwnet_pd_delete(pd);
pd                666 drivers/firewire/net.c 			pd = fwnet_pd_new(net, peer, datagram_label,
pd                668 drivers/firewire/net.c 			if (pd == NULL) {
pd                674 drivers/firewire/net.c 			if (!fwnet_pd_update(peer, pd, buf, fg_off, len)) {
pd                680 drivers/firewire/net.c 				fwnet_pd_delete(pd);
pd                689 drivers/firewire/net.c 		pd->ether_type = ether_type;
pd                691 drivers/firewire/net.c 	if (fwnet_pd_is_complete(pd)) {
pd                692 drivers/firewire/net.c 		ether_type = pd->ether_type;
pd                694 drivers/firewire/net.c 		skb = skb_get(pd->skb);
pd                695 drivers/firewire/net.c 		fwnet_pd_delete(pd);
pd               1549 drivers/firewire/net.c 	struct fwnet_partial_datagram *pd, *pd_next;
pd               1557 drivers/firewire/net.c 	list_for_each_entry_safe(pd, pd_next, &peer->pd_list, pd_link)
pd               1558 drivers/firewire/net.c 		fwnet_pd_delete(pd);
pd               2767 drivers/firewire/ohci.c 	struct descriptor *pd;
pd               2770 drivers/firewire/ohci.c 	for (pd = d; pd <= last; pd++)
pd               2771 drivers/firewire/ohci.c 		if (pd->transfer_status)
pd               2773 drivers/firewire/ohci.c 	if (pd > last)
pd               2848 drivers/firewire/ohci.c 					  struct descriptor *pd)
pd               2854 drivers/firewire/ohci.c 	if (pd->control & cpu_to_le16(DESCRIPTOR_BRANCH_ALWAYS))
pd               2858 drivers/firewire/ohci.c 	pd += 2;
pd               2865 drivers/firewire/ohci.c 	if ((le32_to_cpu(pd->data_address) & PAGE_MASK) ==
pd               2867 drivers/firewire/ohci.c 		if (pd->control & cpu_to_le16(DESCRIPTOR_BRANCH_ALWAYS))
pd               2869 drivers/firewire/ohci.c 		pd++;
pd               2873 drivers/firewire/ohci.c 		buffer_dma = le32_to_cpu(pd->data_address);
pd               2877 drivers/firewire/ohci.c 					      le16_to_cpu(pd->req_count),
pd               2879 drivers/firewire/ohci.c 		control = pd->control;
pd               2880 drivers/firewire/ohci.c 		pd++;
pd               2890 drivers/firewire/ohci.c 	struct descriptor *pd;
pd               2893 drivers/firewire/ohci.c 	for (pd = d; pd <= last; pd++)
pd               2894 drivers/firewire/ohci.c 		if (pd->transfer_status)
pd               2896 drivers/firewire/ohci.c 	if (pd > last)
pd               2911 drivers/firewire/ohci.c 	*ctx_hdr = cpu_to_be32((le16_to_cpu(pd->transfer_status) << 16) |
pd               2912 drivers/firewire/ohci.c 			       le16_to_cpu(pd->res_count));
pd               3197 drivers/firewire/ohci.c 	struct descriptor *d, *last, *pd;
pd               3260 drivers/firewire/ohci.c 	pd = d + z - payload_z;
pd               3268 drivers/firewire/ohci.c 		pd[i].req_count    = cpu_to_le16(length);
pd               3271 drivers/firewire/ohci.c 		pd[i].data_address = cpu_to_le32(page_bus + offset);
pd               3302 drivers/firewire/ohci.c 	struct descriptor *d, *pd;
pd               3339 drivers/firewire/ohci.c 		pd = d;
pd               3341 drivers/firewire/ohci.c 			pd++;
pd               3342 drivers/firewire/ohci.c 			pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
pd               3349 drivers/firewire/ohci.c 			pd->req_count = cpu_to_le16(length);
pd               3350 drivers/firewire/ohci.c 			pd->res_count = pd->req_count;
pd               3351 drivers/firewire/ohci.c 			pd->transfer_status = 0;
pd               3354 drivers/firewire/ohci.c 			pd->data_address = cpu_to_le32(page_bus + offset);
pd               3365 drivers/firewire/ohci.c 		pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
pd               3369 drivers/firewire/ohci.c 			pd->control |= cpu_to_le16(DESCRIPTOR_IRQ_ALWAYS);
pd                 27 drivers/firmware/arm_scmi/scmi_pm_domain.c 	struct scmi_pm_domain *pd = to_scmi_pd(domain);
pd                 28 drivers/firmware/arm_scmi/scmi_pm_domain.c 	const struct scmi_power_ops *ops = pd->handle->power_ops;
pd                 35 drivers/firmware/arm_scmi/scmi_pm_domain.c 	ret = ops->state_set(pd->handle, pd->domain, state);
pd                 37 drivers/firmware/arm_scmi/scmi_pm_domain.c 		ret = ops->state_get(pd->handle, pd->domain, &ret_state);
pd                272 drivers/firmware/efi/arm-init.c 	void *pd;
pd                277 drivers/firmware/efi/arm-init.c 	pd = platform_device_register_data(NULL, "efi-framebuffer", 0,
pd                279 drivers/firmware/efi/arm-init.c 	return PTR_ERR_OR_ZERO(pd);
pd                 68 drivers/firmware/imx/scu-pd.c 	struct generic_pm_domain pd;
pd                173 drivers/firmware/imx/scu-pd.c 	return container_of(genpd, struct imx_sc_pm_domain, pd);
pd                180 drivers/firmware/imx/scu-pd.c 	struct imx_sc_pm_domain *pd;
pd                183 drivers/firmware/imx/scu-pd.c 	pd = to_imx_sc_pd(domain);
pd                190 drivers/firmware/imx/scu-pd.c 	msg.resource = pd->rsrc;
pd                196 drivers/firmware/imx/scu-pd.c 			power_on ? "up" : "off", pd->rsrc, ret);
pd                223 drivers/firmware/imx/scu-pd.c 			domain = &sc_pd->pd;
pd                243 drivers/firmware/imx/scu-pd.c 	sc_pd->pd.power_off = imx_sc_pd_power_off;
pd                244 drivers/firmware/imx/scu-pd.c 	sc_pd->pd.power_on = imx_sc_pd_power_on;
pd                253 drivers/firmware/imx/scu-pd.c 	sc_pd->pd.name = sc_pd->name;
pd                263 drivers/firmware/imx/scu-pd.c 	ret = pm_genpd_init(&sc_pd->pd, NULL, true);
pd                302 drivers/firmware/imx/scu-pd.c 			domains[count++] = &sc_pd->pd;
pd                303 drivers/firmware/imx/scu-pd.c 			dev_dbg(dev, "added power domain %s\n", sc_pd->pd.name);
pd                 34 drivers/firmware/scpi_pm_domain.c static int scpi_pd_power(struct scpi_pm_domain *pd, bool power_on)
pd                 44 drivers/firmware/scpi_pm_domain.c 	ret = pd->ops->device_set_power_state(pd->domain, state);
pd                 48 drivers/firmware/scpi_pm_domain.c 	return !(state == pd->ops->device_get_power_state(pd->domain));
pd                 53 drivers/firmware/scpi_pm_domain.c 	struct scpi_pm_domain *pd = to_scpi_pd(domain);
pd                 55 drivers/firmware/scpi_pm_domain.c 	return scpi_pd_power(pd, true);
pd                 60 drivers/firmware/scpi_pm_domain.c 	struct scpi_pm_domain *pd = to_scpi_pd(domain);
pd                 62 drivers/firmware/scpi_pm_domain.c 	return scpi_pd_power(pd, false);
pd                134 drivers/firmware/trusted_foundations.c void register_trusted_foundations(struct trusted_foundations_platform_data *pd)
pd                312 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo *pd = vm->root.base.bo;
pd                313 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
pd                327 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	ret = amdgpu_amdkfd_validate(&param, pd);
pd                336 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		ret = amdgpu_bo_kmap(pd, NULL);
pd                348 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo *pd = vm->root.base.bo;
pd                349 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
pd                812 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 		struct amdgpu_bo *pd = peer_vm->root.base.bo;
pd                815 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 					sync, pd->tbo.base.resv,
pd                995 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo *pd = vm->root.base.bo;
pd               1001 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_bo_reserve(pd, false);
pd               1002 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_bo_fence(pd, NULL, false);
pd               1003 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	amdgpu_bo_unreserve(pd);
pd               1062 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_bo *pd = avm->root.base.bo;
pd               1063 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c 	struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev);
pd                 71 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c 	struct amdgpu_bo_list_entry pd;
pd                 81 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c 	amdgpu_vm_get_pd_bo(vm, &list, &pd);
pd               1048 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	struct amdgpu_bo *pd;
pd               1073 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	pd = amdgpu_bo_ref(fpriv->vm.root.base.bo);
pd               1079 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 		amdgpu_pasid_free_delayed(pd->tbo.base.resv, pasid);
pd               1080 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c 	amdgpu_bo_unref(&pd);
pd                128 drivers/gpu/drm/gma500/mmu.c void psb_mmu_set_pd_context(struct psb_mmu_pd *pd, int hw_context)
pd                130 drivers/gpu/drm/gma500/mmu.c 	struct drm_device *dev = pd->driver->dev;
pd                135 drivers/gpu/drm/gma500/mmu.c 	down_write(&pd->driver->sem);
pd                136 drivers/gpu/drm/gma500/mmu.c 	PSB_WSGX32(page_to_pfn(pd->p) << PAGE_SHIFT, offset);
pd                138 drivers/gpu/drm/gma500/mmu.c 	psb_mmu_flush_pd_locked(pd->driver, 1);
pd                139 drivers/gpu/drm/gma500/mmu.c 	pd->hw_context = hw_context;
pd                140 drivers/gpu/drm/gma500/mmu.c 	up_write(&pd->driver->sem);
pd                168 drivers/gpu/drm/gma500/mmu.c 	struct psb_mmu_pd *pd = kmalloc(sizeof(*pd), GFP_KERNEL);
pd                172 drivers/gpu/drm/gma500/mmu.c 	if (!pd)
pd                175 drivers/gpu/drm/gma500/mmu.c 	pd->p = alloc_page(GFP_DMA32);
pd                176 drivers/gpu/drm/gma500/mmu.c 	if (!pd->p)
pd                178 drivers/gpu/drm/gma500/mmu.c 	pd->dummy_pt = alloc_page(GFP_DMA32);
pd                179 drivers/gpu/drm/gma500/mmu.c 	if (!pd->dummy_pt)
pd                181 drivers/gpu/drm/gma500/mmu.c 	pd->dummy_page = alloc_page(GFP_DMA32);
pd                182 drivers/gpu/drm/gma500/mmu.c 	if (!pd->dummy_page)
pd                186 drivers/gpu/drm/gma500/mmu.c 		pd->invalid_pde = psb_mmu_mask_pte(page_to_pfn(pd->dummy_pt),
pd                188 drivers/gpu/drm/gma500/mmu.c 		pd->invalid_pte = psb_mmu_mask_pte(page_to_pfn(pd->dummy_page),
pd                191 drivers/gpu/drm/gma500/mmu.c 		pd->invalid_pde = 0;
pd                192 drivers/gpu/drm/gma500/mmu.c 		pd->invalid_pte = 0;
pd                195 drivers/gpu/drm/gma500/mmu.c 	v = kmap(pd->dummy_pt);
pd                197 drivers/gpu/drm/gma500/mmu.c 		v[i] = pd->invalid_pte;
pd                199 drivers/gpu/drm/gma500/mmu.c 	kunmap(pd->dummy_pt);
pd                201 drivers/gpu/drm/gma500/mmu.c 	v = kmap(pd->p);
pd                203 drivers/gpu/drm/gma500/mmu.c 		v[i] = pd->invalid_pde;
pd                205 drivers/gpu/drm/gma500/mmu.c 	kunmap(pd->p);
pd                207 drivers/gpu/drm/gma500/mmu.c 	clear_page(kmap(pd->dummy_page));
pd                208 drivers/gpu/drm/gma500/mmu.c 	kunmap(pd->dummy_page);
pd                210 drivers/gpu/drm/gma500/mmu.c 	pd->tables = vmalloc_user(sizeof(struct psb_mmu_pt *) * 1024);
pd                211 drivers/gpu/drm/gma500/mmu.c 	if (!pd->tables)
pd                214 drivers/gpu/drm/gma500/mmu.c 	pd->hw_context = -1;
pd                215 drivers/gpu/drm/gma500/mmu.c 	pd->pd_mask = PSB_PTE_VALID;
pd                216 drivers/gpu/drm/gma500/mmu.c 	pd->driver = driver;
pd                218 drivers/gpu/drm/gma500/mmu.c 	return pd;
pd                221 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->dummy_page);
pd                223 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->dummy_pt);
pd                225 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->p);
pd                227 drivers/gpu/drm/gma500/mmu.c 	kfree(pd);
pd                237 drivers/gpu/drm/gma500/mmu.c void psb_mmu_free_pagedir(struct psb_mmu_pd *pd)
pd                239 drivers/gpu/drm/gma500/mmu.c 	struct psb_mmu_driver *driver = pd->driver;
pd                246 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1) {
pd                247 drivers/gpu/drm/gma500/mmu.c 		PSB_WSGX32(0, PSB_CR_BIF_DIR_LIST_BASE0 + pd->hw_context * 4);
pd                255 drivers/gpu/drm/gma500/mmu.c 		pt = pd->tables[i];
pd                260 drivers/gpu/drm/gma500/mmu.c 	vfree(pd->tables);
pd                261 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->dummy_page);
pd                262 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->dummy_pt);
pd                263 drivers/gpu/drm/gma500/mmu.c 	__free_page(pd->p);
pd                264 drivers/gpu/drm/gma500/mmu.c 	kfree(pd);
pd                268 drivers/gpu/drm/gma500/mmu.c static struct psb_mmu_pt *psb_mmu_alloc_pt(struct psb_mmu_pd *pd)
pd                272 drivers/gpu/drm/gma500/mmu.c 	uint32_t clflush_add = pd->driver->clflush_add >> PAGE_SHIFT;
pd                274 drivers/gpu/drm/gma500/mmu.c 	spinlock_t *lock = &pd->driver->lock;
pd                294 drivers/gpu/drm/gma500/mmu.c 		*ptes++ = pd->invalid_pte;
pd                297 drivers/gpu/drm/gma500/mmu.c 	if (pd->driver->has_clflush && pd->hw_context != -1) {
pd                310 drivers/gpu/drm/gma500/mmu.c 	pt->pd = pd;
pd                316 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pt *psb_mmu_pt_alloc_map_lock(struct psb_mmu_pd *pd,
pd                322 drivers/gpu/drm/gma500/mmu.c 	spinlock_t *lock = &pd->driver->lock;
pd                325 drivers/gpu/drm/gma500/mmu.c 	pt = pd->tables[index];
pd                328 drivers/gpu/drm/gma500/mmu.c 		pt = psb_mmu_alloc_pt(pd);
pd                333 drivers/gpu/drm/gma500/mmu.c 		if (pd->tables[index]) {
pd                337 drivers/gpu/drm/gma500/mmu.c 			pt = pd->tables[index];
pd                341 drivers/gpu/drm/gma500/mmu.c 		v = kmap_atomic(pd->p);
pd                342 drivers/gpu/drm/gma500/mmu.c 		pd->tables[index] = pt;
pd                343 drivers/gpu/drm/gma500/mmu.c 		v[index] = (page_to_pfn(pt->p) << 12) | pd->pd_mask;
pd                347 drivers/gpu/drm/gma500/mmu.c 		if (pd->hw_context != -1) {
pd                348 drivers/gpu/drm/gma500/mmu.c 			psb_mmu_clflush(pd->driver, (void *)&v[index]);
pd                349 drivers/gpu/drm/gma500/mmu.c 			atomic_set(&pd->driver->needs_tlbflush, 1);
pd                356 drivers/gpu/drm/gma500/mmu.c static struct psb_mmu_pt *psb_mmu_pt_map_lock(struct psb_mmu_pd *pd,
pd                361 drivers/gpu/drm/gma500/mmu.c 	spinlock_t *lock = &pd->driver->lock;
pd                364 drivers/gpu/drm/gma500/mmu.c 	pt = pd->tables[index];
pd                375 drivers/gpu/drm/gma500/mmu.c 	struct psb_mmu_pd *pd = pt->pd;
pd                380 drivers/gpu/drm/gma500/mmu.c 		v = kmap_atomic(pd->p);
pd                381 drivers/gpu/drm/gma500/mmu.c 		v[pt->index] = pd->invalid_pde;
pd                382 drivers/gpu/drm/gma500/mmu.c 		pd->tables[pt->index] = NULL;
pd                384 drivers/gpu/drm/gma500/mmu.c 		if (pd->hw_context != -1) {
pd                385 drivers/gpu/drm/gma500/mmu.c 			psb_mmu_clflush(pd->driver, (void *)&v[pt->index]);
pd                386 drivers/gpu/drm/gma500/mmu.c 			atomic_set(&pd->driver->needs_tlbflush, 1);
pd                389 drivers/gpu/drm/gma500/mmu.c 		spin_unlock(&pd->driver->lock);
pd                393 drivers/gpu/drm/gma500/mmu.c 	spin_unlock(&pd->driver->lock);
pd                405 drivers/gpu/drm/gma500/mmu.c 	pt->v[psb_mmu_pt_index(addr)] = pt->pd->invalid_pte;
pd                410 drivers/gpu/drm/gma500/mmu.c 	struct psb_mmu_pd *pd;
pd                413 drivers/gpu/drm/gma500/mmu.c 	pd = driver->default_pd;
pd                416 drivers/gpu/drm/gma500/mmu.c 	return pd;
pd                422 drivers/gpu/drm/gma500/mmu.c 	struct psb_mmu_pd *pd;
pd                424 drivers/gpu/drm/gma500/mmu.c 	pd = psb_mmu_get_default_pd(driver);
pd                425 drivers/gpu/drm/gma500/mmu.c 	return page_to_pfn(pd->p) << PAGE_SHIFT;
pd                499 drivers/gpu/drm/gma500/mmu.c static void psb_mmu_flush_ptes(struct psb_mmu_pd *pd, unsigned long address,
pd                511 drivers/gpu/drm/gma500/mmu.c 	unsigned long clflush_add = pd->driver->clflush_add;
pd                512 drivers/gpu/drm/gma500/mmu.c 	unsigned long clflush_mask = pd->driver->clflush_mask;
pd                514 drivers/gpu/drm/gma500/mmu.c 	if (!pd->driver->has_clflush)
pd                532 drivers/gpu/drm/gma500/mmu.c 			pt = psb_mmu_pt_map_lock(pd, addr);
pd                547 drivers/gpu/drm/gma500/mmu.c static void psb_mmu_flush_ptes(struct psb_mmu_pd *pd, unsigned long address,
pd                555 drivers/gpu/drm/gma500/mmu.c void psb_mmu_remove_pfn_sequence(struct psb_mmu_pd *pd,
pd                564 drivers/gpu/drm/gma500/mmu.c 	down_read(&pd->driver->sem);
pd                571 drivers/gpu/drm/gma500/mmu.c 		pt = psb_mmu_pt_alloc_map_lock(pd, addr);
pd                583 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                584 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1);
pd                586 drivers/gpu/drm/gma500/mmu.c 	up_read(&pd->driver->sem);
pd                588 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                589 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush(pd->driver);
pd                594 drivers/gpu/drm/gma500/mmu.c void psb_mmu_remove_pages(struct psb_mmu_pd *pd, unsigned long address,
pd                616 drivers/gpu/drm/gma500/mmu.c 	down_read(&pd->driver->sem);
pd                627 drivers/gpu/drm/gma500/mmu.c 			pt = psb_mmu_pt_map_lock(pd, addr);
pd                640 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                641 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush_ptes(pd, f_address, num_pages,
pd                644 drivers/gpu/drm/gma500/mmu.c 	up_read(&pd->driver->sem);
pd                646 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                647 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush(pd->driver);
pd                650 drivers/gpu/drm/gma500/mmu.c int psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd, uint32_t start_pfn,
pd                662 drivers/gpu/drm/gma500/mmu.c 	down_read(&pd->driver->sem);
pd                669 drivers/gpu/drm/gma500/mmu.c 		pt = psb_mmu_pt_alloc_map_lock(pd, addr);
pd                685 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                686 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1);
pd                688 drivers/gpu/drm/gma500/mmu.c 	up_read(&pd->driver->sem);
pd                690 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                691 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush(pd->driver);
pd                696 drivers/gpu/drm/gma500/mmu.c int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
pd                724 drivers/gpu/drm/gma500/mmu.c 	down_read(&pd->driver->sem);
pd                733 drivers/gpu/drm/gma500/mmu.c 			pt = psb_mmu_pt_alloc_map_lock(pd, addr);
pd                751 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                752 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush_ptes(pd, f_address, num_pages,
pd                755 drivers/gpu/drm/gma500/mmu.c 	up_read(&pd->driver->sem);
pd                757 drivers/gpu/drm/gma500/mmu.c 	if (pd->hw_context != -1)
pd                758 drivers/gpu/drm/gma500/mmu.c 		psb_mmu_flush(pd->driver);
pd                763 drivers/gpu/drm/gma500/mmu.c int psb_mmu_virtual_to_pfn(struct psb_mmu_pd *pd, uint32_t virtual,
pd                769 drivers/gpu/drm/gma500/mmu.c 	spinlock_t *lock = &pd->driver->lock;
pd                771 drivers/gpu/drm/gma500/mmu.c 	down_read(&pd->driver->sem);
pd                772 drivers/gpu/drm/gma500/mmu.c 	pt = psb_mmu_pt_map_lock(pd, virtual);
pd                777 drivers/gpu/drm/gma500/mmu.c 		v = kmap_atomic(pd->p);
pd                782 drivers/gpu/drm/gma500/mmu.c 		if (tmp != pd->invalid_pde || !(tmp & PSB_PTE_VALID) ||
pd                783 drivers/gpu/drm/gma500/mmu.c 		    !(pd->invalid_pte & PSB_PTE_VALID)) {
pd                788 drivers/gpu/drm/gma500/mmu.c 		*pfn = pd->invalid_pte >> PAGE_SHIFT;
pd                800 drivers/gpu/drm/gma500/mmu.c 	up_read(&pd->driver->sem);
pd                 36 drivers/gpu/drm/gma500/mmu.h 	struct psb_mmu_pd *pd;
pd                 65 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_free_pagedir(struct psb_mmu_pd *pd);
pd                 67 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_remove_pfn_sequence(struct psb_mmu_pd *pd,
pd                 70 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd,
pd                 74 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_virtual_to_pfn(struct psb_mmu_pd *pd, uint32_t virtual,
pd                 76 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_set_pd_context(struct psb_mmu_pd *pd, int hw_context);
pd                 77 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages,
pd                 81 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_remove_pages(struct psb_mmu_pd *pd,
pd                994 drivers/gpu/drm/i915/gem/i915_gem_context.c 		const dma_addr_t pd_daddr = px_dma(ppgtt->pd);
pd                 58 drivers/gpu/drm/i915/gt/intel_lrc_reg.h 	const u64 addr__ = px_dma(ppgtt->pd); \
pd               1545 drivers/gpu/drm/i915/gt/intel_ringbuffer.c 	*cs++ = px_base(ppgtt->pd)->ggtt_offset << 10;
pd                372 drivers/gpu/drm/i915/gvt/scheduler.c 		px_dma(ppgtt->pd) = mm->ppgtt_mm.shadow_pdps[0];
pd                375 drivers/gpu/drm/i915/gvt/scheduler.c 			struct i915_page_directory * const pd =
pd                376 drivers/gpu/drm/i915/gvt/scheduler.c 				i915_pd_entry(ppgtt->pd, i);
pd                380 drivers/gpu/drm/i915/gvt/scheduler.c 			if (!pd)
pd                382 drivers/gpu/drm/i915/gvt/scheduler.c 			px_dma(pd) = mm->ppgtt_mm.shadow_pdps[i];
pd               1152 drivers/gpu/drm/i915/gvt/scheduler.c 		px_dma(ppgtt->pd) = s->i915_context_pml4;
pd               1155 drivers/gpu/drm/i915/gvt/scheduler.c 			struct i915_page_directory * const pd =
pd               1156 drivers/gpu/drm/i915/gvt/scheduler.c 				i915_pd_entry(ppgtt->pd, i);
pd               1158 drivers/gpu/drm/i915/gvt/scheduler.c 			px_dma(pd) = s->i915_context_pdps[i];
pd               1213 drivers/gpu/drm/i915/gvt/scheduler.c 		s->i915_context_pml4 = px_dma(ppgtt->pd);
pd               1216 drivers/gpu/drm/i915/gvt/scheduler.c 			struct i915_page_directory * const pd =
pd               1217 drivers/gpu/drm/i915/gvt/scheduler.c 				i915_pd_entry(ppgtt->pd, i);
pd               1219 drivers/gpu/drm/i915/gvt/scheduler.c 			s->i915_context_pdps[i] = px_dma(pd);
pd                724 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory *pd;
pd                726 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd = kzalloc(sz, I915_GFP_ALLOW_FAIL);
pd                727 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(!pd))
pd                730 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_lock_init(&pd->lock);
pd                731 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pd;
pd                736 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory *pd;
pd                738 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd = __alloc_pd(sizeof(*pd));
pd                739 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(!pd))
pd                742 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(setup_page_dma(vm, px_base(pd)))) {
pd                743 drivers/gpu/drm/i915/i915_gem_gtt.c 		kfree(pd);
pd                747 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pd;
pd                750 drivers/gpu/drm/i915/i915_gem_gtt.c static void free_pd(struct i915_address_space *vm, struct i915_page_dma *pd)
pd                752 drivers/gpu/drm/i915/i915_gem_gtt.c 	cleanup_page_dma(vm, pd);
pd                753 drivers/gpu/drm/i915/i915_gem_gtt.c 	kfree(pd);
pd                770 drivers/gpu/drm/i915/i915_gem_gtt.c __set_pd_entry(struct i915_page_directory * const pd,
pd                776 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry));
pd                778 drivers/gpu/drm/i915/i915_gem_gtt.c 	atomic_inc(px_used(pd));
pd                779 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd->entry[idx] = to;
pd                780 drivers/gpu/drm/i915/i915_gem_gtt.c 	write_dma_entry(px_base(pd), idx, encode(to->daddr, I915_CACHE_LLC));
pd                783 drivers/gpu/drm/i915/i915_gem_gtt.c #define set_pd_entry(pd, idx, to) \
pd                784 drivers/gpu/drm/i915/i915_gem_gtt.c 	__set_pd_entry((pd), (idx), px_base(to), gen8_pde_encode)
pd                787 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(struct i915_page_directory * const pd,
pd                791 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(atomic_read(px_used(pd)) == 0);
pd                793 drivers/gpu/drm/i915/i915_gem_gtt.c 	write_dma_entry(px_base(pd), idx, scratch->encode);
pd                794 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd->entry[idx] = NULL;
pd                795 drivers/gpu/drm/i915/i915_gem_gtt.c 	atomic_dec(px_used(pd));
pd                799 drivers/gpu/drm/i915/i915_gem_gtt.c release_pd_entry(struct i915_page_directory * const pd,
pd                809 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_lock(&pd->lock);
pd                811 drivers/gpu/drm/i915/i915_gem_gtt.c 		clear_pd_entry(pd, idx, scratch);
pd                814 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_unlock(&pd->lock);
pd                837 drivers/gpu/drm/i915/i915_gem_gtt.c 		atomic_inc(px_used(ppgtt->pd)); /* never remove */
pd                839 drivers/gpu/drm/i915/i915_gem_gtt.c 		atomic_dec(px_used(ppgtt->pd));
pd                844 drivers/gpu/drm/i915/i915_gem_gtt.c 		const u64 daddr = px_dma(ppgtt->pd);
pd                923 drivers/gpu/drm/i915/i915_gem_gtt.c 		return ppgtt->pd;
pd                925 drivers/gpu/drm/i915/i915_gem_gtt.c 		return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm->top));
pd                935 drivers/gpu/drm/i915/i915_gem_gtt.c 				 struct i915_page_directory *pd,
pd                939 drivers/gpu/drm/i915/i915_gem_gtt.c 		void **pde = pd->entry;
pd                949 drivers/gpu/drm/i915/i915_gem_gtt.c 	free_px(vm, pd);
pd                959 drivers/gpu/drm/i915/i915_gem_gtt.c 	__gen8_ppgtt_cleanup(vm, ppgtt->pd, gen8_pd_top_count(vm), vm->top);
pd                964 drivers/gpu/drm/i915/i915_gem_gtt.c 			      struct i915_page_directory * const pd,
pd                975 drivers/gpu/drm/i915/i915_gem_gtt.c 	    idx, len, atomic_read(px_used(pd)));
pd                976 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(!len || len >= atomic_read(px_used(pd)));
pd                979 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_page_table *pt = pd->entry[idx];
pd                985 drivers/gpu/drm/i915/i915_gem_gtt.c 			clear_pd_entry(pd, idx, scratch);
pd               1015 drivers/gpu/drm/i915/i915_gem_gtt.c 		if (release_pd_entry(pd, idx, pt, scratch))
pd               1033 drivers/gpu/drm/i915/i915_gem_gtt.c 	__gen8_ppgtt_clear(vm, i915_vm_to_ppgtt(vm)->pd,
pd               1038 drivers/gpu/drm/i915/i915_gem_gtt.c 			      struct i915_page_directory * const pd,
pd               1051 drivers/gpu/drm/i915/i915_gem_gtt.c 	    idx, len, atomic_read(px_used(pd)));
pd               1054 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_lock(&pd->lock);
pd               1055 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(!atomic_read(px_used(pd))); /* Must be pinned! */
pd               1057 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_page_table *pt = pd->entry[idx];
pd               1060 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_unlock(&pd->lock);
pd               1090 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_lock(&pd->lock);
pd               1091 drivers/gpu/drm/i915/i915_gem_gtt.c 			if (likely(!pd->entry[idx]))
pd               1092 drivers/gpu/drm/i915/i915_gem_gtt.c 				set_pd_entry(pd, idx, pt);
pd               1094 drivers/gpu/drm/i915/i915_gem_gtt.c 				alloc = pt, pt = pd->entry[idx];
pd               1099 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_unlock(&pd->lock);
pd               1104 drivers/gpu/drm/i915/i915_gem_gtt.c 				if (release_pd_entry(pd, idx, pt, scratch))
pd               1109 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_lock(&pd->lock);
pd               1126 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_unlock(&pd->lock);
pd               1148 drivers/gpu/drm/i915/i915_gem_gtt.c 	err = __gen8_ppgtt_alloc(vm, i915_vm_to_ppgtt(vm)->pd,
pd               1151 drivers/gpu/drm/i915/i915_gem_gtt.c 		__gen8_ppgtt_clear(vm, i915_vm_to_ppgtt(vm)->pd,
pd               1174 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory *pd;
pd               1178 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd = i915_pd_entry(pdp, gen8_pd_index(idx, 2));
pd               1179 drivers/gpu/drm/i915/i915_gem_gtt.c 	vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1)));
pd               1202 drivers/gpu/drm/i915/i915_gem_gtt.c 				pd = pdp->entry[gen8_pd_index(idx, 2)];
pd               1206 drivers/gpu/drm/i915/i915_gem_gtt.c 			vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1)));
pd               1228 drivers/gpu/drm/i915/i915_gem_gtt.c 		struct i915_page_directory * const pd =
pd               1244 drivers/gpu/drm/i915/i915_gem_gtt.c 			vaddr = kmap_atomic_px(pd);
pd               1247 drivers/gpu/drm/i915/i915_gem_gtt.c 				i915_pt_entry(pd, __gen8_pte_index(start, 1));
pd               1303 drivers/gpu/drm/i915/i915_gem_gtt.c 			vaddr = kmap_atomic_px(pd);
pd               1321 drivers/gpu/drm/i915/i915_gem_gtt.c 				vaddr = kmap_atomic_px(i915_pt_entry(pd, maybe_64K));
pd               1409 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory *pd = ppgtt->pd;
pd               1423 drivers/gpu/drm/i915/i915_gem_gtt.c 		set_pd_entry(pd, idx, pde);
pd               1451 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory *pd;
pd               1453 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(count > ARRAY_SIZE(pd->entry));
pd               1455 drivers/gpu/drm/i915/i915_gem_gtt.c 	pd = __alloc_pd(offsetof(typeof(*pd), entry[count]));
pd               1456 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(!pd))
pd               1459 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (unlikely(setup_page_dma(vm, px_base(pd)))) {
pd               1460 drivers/gpu/drm/i915/i915_gem_gtt.c 		kfree(pd);
pd               1464 drivers/gpu/drm/i915/i915_gem_gtt.c 	fill_page_dma(px_base(pd), vm->scratch[vm->top].encode, count);
pd               1465 drivers/gpu/drm/i915/i915_gem_gtt.c 	atomic_inc(px_used(pd)); /* mark as pinned */
pd               1466 drivers/gpu/drm/i915/i915_gem_gtt.c 	return pd;
pd               1506 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt->pd = gen8_alloc_top_pd(&ppgtt->vm);
pd               1507 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (IS_ERR(ppgtt->pd)) {
pd               1508 drivers/gpu/drm/i915/i915_gem_gtt.c 		err = PTR_ERR(ppgtt->pd);
pd               1532 drivers/gpu/drm/i915/i915_gem_gtt.c 	__gen8_ppgtt_cleanup(&ppgtt->vm, ppgtt->pd,
pd               1616 drivers/gpu/drm/i915/i915_gem_gtt.c 			i915_pt_entry(ppgtt->base.pd, pde++);
pd               1649 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory * const pd = ppgtt->pd;
pd               1657 drivers/gpu/drm/i915/i915_gem_gtt.c 	GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]);
pd               1659 drivers/gpu/drm/i915/i915_gem_gtt.c 	vaddr = kmap_atomic_px(i915_pt_entry(pd, act_pt));
pd               1676 drivers/gpu/drm/i915/i915_gem_gtt.c 			vaddr = kmap_atomic_px(i915_pt_entry(pd, ++act_pt));
pd               1689 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory * const pd = ppgtt->base.pd;
pd               1699 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_lock(&pd->lock);
pd               1700 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_for_each_pde(pt, pd, start, length, pde) {
pd               1704 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_unlock(&pd->lock);
pd               1716 drivers/gpu/drm/i915/i915_gem_gtt.c 			spin_lock(&pd->lock);
pd               1717 drivers/gpu/drm/i915/i915_gem_gtt.c 			if (pd->entry[pde] == &vm->scratch[1]) {
pd               1718 drivers/gpu/drm/i915/i915_gem_gtt.c 				pd->entry[pde] = pt;
pd               1726 drivers/gpu/drm/i915/i915_gem_gtt.c 				pt = pd->entry[pde];
pd               1732 drivers/gpu/drm/i915/i915_gem_gtt.c 	spin_unlock(&pd->lock);
pd               1753 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory * const pd = ppgtt->base.pd;
pd               1770 drivers/gpu/drm/i915/i915_gem_gtt.c 	memset_p(pd->entry, &vm->scratch[1], I915_PDES);
pd               1777 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory * const pd = ppgtt->base.pd;
pd               1783 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_for_all_pdes(pt, pd, pde)
pd               1800 drivers/gpu/drm/i915/i915_gem_gtt.c 	kfree(ppgtt->base.pd);
pd               1826 drivers/gpu/drm/i915/i915_gem_gtt.c 	px_base(ppgtt->base.pd)->ggtt_offset = ggtt_offset * sizeof(gen6_pte_t);
pd               1829 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_for_all_pdes(pt, ppgtt->base.pd, pde)
pd               1841 drivers/gpu/drm/i915/i915_gem_gtt.c 	struct i915_page_directory * const pd = ppgtt->base.pd;
pd               1851 drivers/gpu/drm/i915/i915_gem_gtt.c 	gen6_for_all_pdes(pt, ppgtt->base.pd, pde) {
pd               1856 drivers/gpu/drm/i915/i915_gem_gtt.c 		pd->entry[pde] = scratch;
pd               1979 drivers/gpu/drm/i915/i915_gem_gtt.c 	ppgtt->base.pd = __alloc_pd(sizeof(*ppgtt->base.pd));
pd               1980 drivers/gpu/drm/i915/i915_gem_gtt.c 	if (!ppgtt->base.pd) {
pd               2000 drivers/gpu/drm/i915/i915_gem_gtt.c 	kfree(ppgtt->base.pd);
pd                426 drivers/gpu/drm/i915/i915_gem_gtt.h 	struct i915_page_directory *pd;
pd                455 drivers/gpu/drm/i915/i915_gem_gtt.h #define gen6_for_each_pde(pt, pd, start, length, iter)			\
pd                458 drivers/gpu/drm/i915/i915_gem_gtt.h 		     (pt = i915_pt_entry(pd, iter), true);		\
pd                463 drivers/gpu/drm/i915/i915_gem_gtt.h #define gen6_for_all_pdes(pt, pd, iter)					\
pd                466 drivers/gpu/drm/i915/i915_gem_gtt.h 		     (pt = i915_pt_entry(pd, iter), true);		\
pd                517 drivers/gpu/drm/i915/i915_gem_gtt.h i915_pt_entry(const struct i915_page_directory * const pd,
pd                520 drivers/gpu/drm/i915/i915_gem_gtt.h 	return pd->entry[n];
pd                533 drivers/gpu/drm/i915/i915_gem_gtt.h 	struct i915_page_dma *pt = ppgtt->pd->entry[n];
pd                 92 drivers/gpu/drm/lima/lima_mmu.c 	mmu_write(LIMA_MMU_DTE_ADDR, dev->empty_vm->pd.dma);
pd                113 drivers/gpu/drm/lima/lima_mmu.c 		mmu_write(LIMA_MMU_DTE_ADDR, vm->pd.dma);
pd                137 drivers/gpu/drm/lima/lima_mmu.c 		mmu_write(LIMA_MMU_DTE_ADDR, dev->empty_vm->pd.dma);
pd                 59 drivers/gpu/drm/lima/lima_vm.c 			u32 *pd;
pd                 72 drivers/gpu/drm/lima/lima_vm.c 			pd = vm->pd.cpu + (pbe << LIMA_VM_NUM_PT_PER_BT_SHIFT);
pd                 74 drivers/gpu/drm/lima/lima_vm.c 				pd[j] = pts | LIMA_VM_FLAG_PRESENT;
pd                212 drivers/gpu/drm/lima/lima_vm.c 	vm->pd.cpu = dma_alloc_wc(dev->dev, LIMA_PAGE_SIZE, &vm->pd.dma,
pd                214 drivers/gpu/drm/lima/lima_vm.c 	if (!vm->pd.cpu)
pd                230 drivers/gpu/drm/lima/lima_vm.c 	dma_free_wc(dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma);
pd                249 drivers/gpu/drm/lima/lima_vm.c 	if (vm->pd.cpu)
pd                250 drivers/gpu/drm/lima/lima_vm.c 		dma_free_wc(vm->dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma);
pd                258 drivers/gpu/drm/lima/lima_vm.c 	u32 *pd, *pt;
pd                260 drivers/gpu/drm/lima/lima_vm.c 	if (!vm->pd.cpu)
pd                263 drivers/gpu/drm/lima/lima_vm.c 	pd = vm->pd.cpu;
pd                272 drivers/gpu/drm/lima/lima_vm.c 			printk(KERN_INFO "lima vm pd %03x:%08x\n", idx, pd[idx]);
pd                 37 drivers/gpu/drm/lima/lima_vm.h 	struct lima_vm_page pd;
pd                157 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c 	s32 hb_en, hb_en_ckln, pd_ckln, pd;
pd                171 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c 	pd = timing->hs_prep_dly;
pd                173 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c 	val = (hb_en << 2) + (pd << 1);
pd                166 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c static int pll_get_post_div(struct hdmi_8996_post_divider *pd, u64 bclk)
pd                207 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 		pd->vco_freq = vco_optimal;
pd                208 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 		pd->tx_band_sel = tx_band_sel[vco_optimal_index % 4];
pd                209 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 		pd->vco_ratio = ratio[vco_optimal_index / 4];
pd                210 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 		pd->hsclk_divsel = hs_divsel[vco_optimal_index / 4];
pd                221 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	struct hdmi_8996_post_divider pd;
pd                244 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	ret = pll_get_post_div(&pd, bclk);
pd                248 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	dec_start = pd.vco_freq;
pd                252 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	frac_start = pd.vco_freq * (1 << 20);
pd                265 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	fdata = pd.vco_freq;
pd                266 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	do_div(fdata, pd.vco_ratio);
pd                270 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	DBG("VCO freq: %llu", pd.vco_freq);
pd                274 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	DBG("HSCLK_SEL: %d", pd.hsclk_divsel);
pd                281 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	DBG("TX_BAND: %d", pd.tx_band_sel);
pd                290 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 	cfg->com_hsclk_sel = (0x20 | pd.hsclk_divsel);
pd                318 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c 		cfg->tx_lx_tx_band[i] = pd.tx_band_sel + 4;
pd                 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h 	struct nvkm_vmm_pt *pd;
pd                 53 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv04.c 			device->mmu->vmm->pd->pt[0]->memory;
pd                 37 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 	int pd, lf, xd, vh, vr, vo, l3;
pd                 45 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		pd =  ram->next->bios.ramcfg_11_01_80;
pd                104 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 		pd = 1; /* binary driver does this.. bug? */
pd                107 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c 	ram->mr[6] |= (pd & 0x01) << 0;
pd                 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv41.c 	nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr);
pd                 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c 	struct nvkm_memory *pt = mmu->vmm->pd->pt[0]->memory;
pd                523 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	it.pt[it.max] = vmm->pd;
pd               1012 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vmm->pd) {
pd               1013 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_mmu_ptc_put(vmm->mmu, true, &vmm->pd->pt[0]);
pd               1014 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		nvkm_vmm_pt_del(&vmm->pd);
pd               1072 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	vmm->pd = nvkm_vmm_pt_new(desc, false, NULL);
pd               1073 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (!vmm->pd)
pd               1075 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	vmm->pd->refs[0] = 1;
pd               1083 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vmm->pd->pt[0] = nvkm_mmu_ptc_get(mmu, size, desc->align, true);
pd               1084 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (!vmm->pd->pt[0])
pd                109 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	struct nvkm_mmu_pt *pd = pgd->pt[0];
pd                141 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	nvkm_kmap(pd->memory);
pd                142 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	VMM_WO064(pd, vmm, pdei * 8, data);
pd                143 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	nvkm_done(pd->memory);
pd                192 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
pd                205 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 		switch (nvkm_memory_target(pd->memory)) {
pd                213 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 		addr |= (vmm->pd->pt[0]->addr >> 12) << 4;
pd                343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	struct nvkm_mmu_pt *pd = vmm->pd->pt[0];
pd                345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	switch (nvkm_memory_target(pd->memory)) {
pd                355 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	base |= pd->addr;
pd                233 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	struct nvkm_mmu_pt *pd = pgd->pt[0];
pd                241 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	nvkm_kmap(pd->memory);
pd                242 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	VMM_WO128(pd, vmm, pdei * 0x10, data[0], data[1]);
pd                243 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	nvkm_done(pd->memory);
pd                273 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	struct nvkm_mmu_pt *pd = pgd->pt[0];
pd                279 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	nvkm_kmap(pd->memory);
pd                280 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	VMM_WO064(pd, vmm, pdei * 8, data);
pd                281 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	nvkm_done(pd->memory);
pd                135 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c 	mem = vmm->pd->pt[0]->memory;
pd                354 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 		if (!nv50_vmm_pde(vmm, vmm->pd->pde[pdei], &data)) {
pd                 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c 	nvkm_wr32(device, 0xb830a0, vmm->pd->pt[0]->addr >> 8);
pd                644 drivers/gpu/drm/radeon/radeon_vm.c 	struct radeon_bo *pd = vm->page_directory;
pd                645 drivers/gpu/drm/radeon/radeon_vm.c 	uint64_t pd_addr = radeon_bo_gpu_offset(pd);
pd                705 drivers/gpu/drm/radeon/radeon_vm.c 		radeon_sync_resv(rdev, &ib.sync, pd->tbo.base.resv, true);
pd                713 drivers/gpu/drm/radeon/radeon_vm.c 		radeon_bo_fence(pd, ib.fence, false);
pd                259 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c static u16 sun6i_dsi_crc_repeat(u8 pd, u8 *buffer, size_t len)
pd                261 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c 	memset(buffer, pd, len);
pd                284 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c static u32 sun6i_dsi_build_blk1_pkt(u16 pd, u8 *buffer, size_t len)
pd                286 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c 	u32 val = SUN6I_DSI_BLK_PD(pd);
pd                288 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c 	return val | SUN6I_DSI_BLK_PF(sun6i_dsi_crc_repeat(pd, buffer, len));
pd                326 drivers/hsi/controllers/omap_ssi_core.c static int ssi_get_iomem(struct platform_device *pd,
pd                331 drivers/hsi/controllers/omap_ssi_core.c 	struct hsi_controller *ssi = platform_get_drvdata(pd);
pd                333 drivers/hsi/controllers/omap_ssi_core.c 	mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name);
pd                347 drivers/hsi/controllers/omap_ssi_core.c 						struct platform_device *pd)
pd                362 drivers/hsi/controllers/omap_ssi_core.c 	ssi->device.parent = &pd->dev;
pd                366 drivers/hsi/controllers/omap_ssi_core.c 	err = ssi_get_iomem(pd, "sys", &omap_ssi->sys, NULL);
pd                369 drivers/hsi/controllers/omap_ssi_core.c 	err = ssi_get_iomem(pd, "gdd", &omap_ssi->gdd, NULL);
pd                372 drivers/hsi/controllers/omap_ssi_core.c 	err = platform_get_irq_byname(pd, "gdd_mpu");
pd                395 drivers/hsi/controllers/omap_ssi_core.c 		dev_err(&pd->dev, "Could not acquire clock \"ssi_ssr_fck\": %li\n",
pd                480 drivers/hsi/controllers/omap_ssi_core.c static int ssi_probe(struct platform_device *pd)
pd                483 drivers/hsi/controllers/omap_ssi_core.c 	struct device_node *np = pd->dev.of_node;
pd                490 drivers/hsi/controllers/omap_ssi_core.c 		dev_err(&pd->dev, "missing device tree data\n");
pd                498 drivers/hsi/controllers/omap_ssi_core.c 		dev_err(&pd->dev, "No memory for controller\n");
pd                502 drivers/hsi/controllers/omap_ssi_core.c 	platform_set_drvdata(pd, ssi);
pd                504 drivers/hsi/controllers/omap_ssi_core.c 	err = ssi_add_controller(ssi, pd);
pd                508 drivers/hsi/controllers/omap_ssi_core.c 	pm_runtime_enable(&pd->dev);
pd                523 drivers/hsi/controllers/omap_ssi_core.c 		childpdev = of_platform_device_create(child, NULL, &pd->dev);
pd                526 drivers/hsi/controllers/omap_ssi_core.c 			dev_err(&pd->dev, "failed to create ssi controller port\n");
pd                531 drivers/hsi/controllers/omap_ssi_core.c 	dev_info(&pd->dev, "ssi controller %d initialized (%d ports)!\n",
pd                535 drivers/hsi/controllers/omap_ssi_core.c 	device_for_each_child(&pd->dev, NULL, ssi_remove_ports);
pd                539 drivers/hsi/controllers/omap_ssi_core.c 	platform_set_drvdata(pd, NULL);
pd                540 drivers/hsi/controllers/omap_ssi_core.c 	pm_runtime_disable(&pd->dev);
pd                545 drivers/hsi/controllers/omap_ssi_core.c static int ssi_remove(struct platform_device *pd)
pd                547 drivers/hsi/controllers/omap_ssi_core.c 	struct hsi_controller *ssi = platform_get_drvdata(pd);
pd                550 drivers/hsi/controllers/omap_ssi_core.c 	device_for_each_child(&pd->dev, NULL, ssi_remove_ports);
pd                556 drivers/hsi/controllers/omap_ssi_core.c 	platform_set_drvdata(pd, NULL);
pd                558 drivers/hsi/controllers/omap_ssi_core.c 	pm_runtime_disable(&pd->dev);
pd               1035 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_irq(struct hsi_port *port, struct platform_device *pd)
pd               1040 drivers/hsi/controllers/omap_ssi_port.c 	err = platform_get_irq(pd, 0);
pd               1052 drivers/hsi/controllers/omap_ssi_port.c static int ssi_wake_irq(struct hsi_port *port, struct platform_device *pd)
pd               1092 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_get_iomem(struct platform_device *pd,
pd               1095 drivers/hsi/controllers/omap_ssi_port.c 	struct hsi_port *port = platform_get_drvdata(pd);
pd               1100 drivers/hsi/controllers/omap_ssi_port.c 	mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name);
pd               1102 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "IO memory region missing (%s)\n", name);
pd               1106 drivers/hsi/controllers/omap_ssi_port.c 					resource_size(mem), dev_name(&pd->dev));
pd               1108 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "%s IO memory region request failed\n",
pd               1114 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "%s IO remap failed\n", mem->name);
pd               1125 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_probe(struct platform_device *pd)
pd               1127 drivers/hsi/controllers/omap_ssi_port.c 	struct device_node *np = pd->dev.of_node;
pd               1130 drivers/hsi/controllers/omap_ssi_port.c 	struct hsi_controller *ssi = dev_get_drvdata(pd->dev.parent);
pd               1136 drivers/hsi/controllers/omap_ssi_port.c 	dev_dbg(&pd->dev, "init ssi port...\n");
pd               1139 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "ssi controller not initialized!\n");
pd               1150 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "port id out of range!\n");
pd               1158 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "missing device tree data\n");
pd               1163 drivers/hsi/controllers/omap_ssi_port.c 	cawake_gpio = devm_gpiod_get(&pd->dev, "ti,ssi-cawake", GPIOD_IN);
pd               1166 drivers/hsi/controllers/omap_ssi_port.c 		dev_err(&pd->dev, "couldn't get cawake gpio (err=%d)!\n", err);
pd               1176 drivers/hsi/controllers/omap_ssi_port.c 	omap_port->pdev = &pd->dev;
pd               1192 drivers/hsi/controllers/omap_ssi_port.c 	platform_set_drvdata(pd, port);
pd               1194 drivers/hsi/controllers/omap_ssi_port.c 	err = ssi_port_get_iomem(pd, "tx", &omap_port->sst_base,
pd               1198 drivers/hsi/controllers/omap_ssi_port.c 	err = ssi_port_get_iomem(pd, "rx", &omap_port->ssr_base,
pd               1203 drivers/hsi/controllers/omap_ssi_port.c 	err = ssi_port_irq(port, pd);
pd               1206 drivers/hsi/controllers/omap_ssi_port.c 	err = ssi_wake_irq(port, pd);
pd               1229 drivers/hsi/controllers/omap_ssi_port.c 	dev_info(&pd->dev, "ssi port %u successfully initialized\n", port_id);
pd               1237 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_remove(struct platform_device *pd)
pd               1239 drivers/hsi/controllers/omap_ssi_port.c 	struct hsi_port *port = platform_get_drvdata(pd);
pd               1260 drivers/hsi/controllers/omap_ssi_port.c 	platform_set_drvdata(pd, NULL);
pd               1262 drivers/hsi/controllers/omap_ssi_port.c 	pm_runtime_dont_use_autosuspend(&pd->dev);
pd               1263 drivers/hsi/controllers/omap_ssi_port.c 	pm_runtime_disable(&pd->dev);
pd                648 drivers/hwmon/coretemp.c 	struct platform_data *pd;
pd                668 drivers/hwmon/coretemp.c 	pd = platform_get_drvdata(pdev);
pd                669 drivers/hwmon/coretemp.c 	tdata = pd->core_data[indx];
pd                671 drivers/hwmon/coretemp.c 	cpumask_clear_cpu(cpu, &pd->cpumask);
pd                678 drivers/hwmon/coretemp.c 	target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu));
pd                680 drivers/hwmon/coretemp.c 		coretemp_remove_core(pd, indx);
pd                692 drivers/hwmon/coretemp.c 	if (cpumask_empty(&pd->cpumask)) {
pd                702 drivers/hwmon/coretemp.c 	tdata = pd->core_data[PKG_SYSFS_ATTR_NO];
pd                704 drivers/hwmon/coretemp.c 		target = cpumask_first(&pd->cpumask);
pd                876 drivers/i2c/busses/i2c-mv64xxx.c mv64xxx_i2c_probe(struct platform_device *pd)
pd                879 drivers/i2c/busses/i2c-mv64xxx.c 	struct mv64xxx_i2c_pdata	*pdata = dev_get_platdata(&pd->dev);
pd                883 drivers/i2c/busses/i2c-mv64xxx.c 	if ((!pdata && !pd->dev.of_node))
pd                886 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data = devm_kzalloc(&pd->dev, sizeof(struct mv64xxx_i2c_data),
pd                891 drivers/i2c/busses/i2c-mv64xxx.c 	r = platform_get_resource(pd, IORESOURCE_MEM, 0);
pd                892 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->reg_base = devm_ioremap_resource(&pd->dev, r);
pd                903 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->clk = devm_clk_get(&pd->dev, NULL);
pd                909 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->reg_clk = devm_clk_get(&pd->dev, "reg");
pd                916 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->irq = platform_get_irq(pd, 0);
pd                924 drivers/i2c/busses/i2c-mv64xxx.c 	} else if (pd->dev.of_node) {
pd                925 drivers/i2c/busses/i2c-mv64xxx.c 		rc = mv64xxx_of_config(drv_data, &pd->dev);
pd                934 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->adapter.dev.parent = &pd->dev;
pd                938 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->adapter.nr = pd->id;
pd                939 drivers/i2c/busses/i2c-mv64xxx.c 	drv_data->adapter.dev.of_node = pd->dev.of_node;
pd                940 drivers/i2c/busses/i2c-mv64xxx.c 	platform_set_drvdata(pd, drv_data);
pd                 37 drivers/i2c/busses/i2c-pca-isa.c static void pca_isa_writebyte(void *pd, int reg, int val)
pd                 47 drivers/i2c/busses/i2c-pca-isa.c static int pca_isa_readbyte(void *pd, int reg)
pd                 59 drivers/i2c/busses/i2c-pca-isa.c static int pca_isa_waitforcompletion(void *pd)
pd                 66 drivers/i2c/busses/i2c-pca-isa.c 				pca_isa_readbyte(pd, I2C_PCA_CON)
pd                 73 drivers/i2c/busses/i2c-pca-isa.c 			if (pca_isa_readbyte(pd, I2C_PCA_CON)
pd                 83 drivers/i2c/busses/i2c-pca-isa.c static void pca_isa_resetchip(void *pd)
pd                 42 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte8(void *pd, int reg)
pd                 44 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 48 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte16(void *pd, int reg)
pd                 50 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 54 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte32(void *pd, int reg)
pd                 56 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 60 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte8(void *pd, int reg, int val)
pd                 62 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 66 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte16(void *pd, int reg, int val)
pd                 68 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 72 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte32(void *pd, int reg, int val)
pd                 74 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                 79 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_waitforcompletion(void *pd)
pd                 81 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                104 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_dummyreset(void *pd)
pd                106 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                111 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_resetchip(void *pd)
pd                113 drivers/i2c/busses/i2c-pca-platform.c 	struct i2c_pca_pf_data *i2c = pd;
pd                434 drivers/i2c/busses/i2c-sh7760.c 	struct sh7760_i2c_platdata *pd;
pd                439 drivers/i2c/busses/i2c-sh7760.c 	pd = dev_get_platdata(&pdev->dev);
pd                440 drivers/i2c/busses/i2c-sh7760.c 	if (!pd) {
pd                497 drivers/i2c/busses/i2c-sh7760.c 	ret = calc_CCR(pd->speed_khz * 1000);
pd                500 drivers/i2c/busses/i2c-sh7760.c 			pd->speed_khz);
pd                519 drivers/i2c/busses/i2c-sh7760.c 		 pd->speed_khz, res->start, id->irq);
pd                143 drivers/i2c/busses/i2c-sh_mobile.c 	int (*setup)(struct sh_mobile_i2c_data *pd);
pd                187 drivers/i2c/busses/i2c-sh_mobile.c static void iic_wr(struct sh_mobile_i2c_data *pd, int offs, unsigned char data)
pd                190 drivers/i2c/busses/i2c-sh_mobile.c 		data |= pd->icic;
pd                192 drivers/i2c/busses/i2c-sh_mobile.c 	iowrite8(data, pd->reg + offs);
pd                195 drivers/i2c/busses/i2c-sh_mobile.c static unsigned char iic_rd(struct sh_mobile_i2c_data *pd, int offs)
pd                197 drivers/i2c/busses/i2c-sh_mobile.c 	return ioread8(pd->reg + offs);
pd                200 drivers/i2c/busses/i2c-sh_mobile.c static void iic_set_clr(struct sh_mobile_i2c_data *pd, int offs,
pd                203 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, offs, (iic_rd(pd, offs) | set) & ~clr);
pd                240 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_check_timing(struct sh_mobile_i2c_data *pd)
pd                242 drivers/i2c/busses/i2c-sh_mobile.c 	u16 max_val = pd->flags & IIC_FLAG_HAS_ICIC67 ? 0x1ff : 0xff;
pd                244 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->iccl > max_val || pd->icch > max_val) {
pd                245 drivers/i2c/busses/i2c-sh_mobile.c 		dev_err(pd->dev, "timing values out of range: L/H=0x%x/0x%x\n",
pd                246 drivers/i2c/busses/i2c-sh_mobile.c 			pd->iccl, pd->icch);
pd                251 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->iccl & 0x100)
pd                252 drivers/i2c/busses/i2c-sh_mobile.c 		pd->icic |= ICIC_ICCLB8;
pd                254 drivers/i2c/busses/i2c-sh_mobile.c 		pd->icic &= ~ICIC_ICCLB8;
pd                257 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->icch & 0x100)
pd                258 drivers/i2c/busses/i2c-sh_mobile.c 		pd->icic |= ICIC_ICCHB8;
pd                260 drivers/i2c/busses/i2c-sh_mobile.c 		pd->icic &= ~ICIC_ICCHB8;
pd                262 drivers/i2c/busses/i2c-sh_mobile.c 	dev_dbg(pd->dev, "timing values: L/H=0x%x/0x%x\n", pd->iccl, pd->icch);
pd                266 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_init(struct sh_mobile_i2c_data *pd)
pd                271 drivers/i2c/busses/i2c-sh_mobile.c 	i2c_clk_khz = clk_get_rate(pd->clk) / 1000 / pd->clks_per_count;
pd                273 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->bus_speed == STANDARD_MODE) {
pd                277 drivers/i2c/busses/i2c-sh_mobile.c 	} else if (pd->bus_speed == FAST_MODE) {
pd                282 drivers/i2c/busses/i2c-sh_mobile.c 		dev_err(pd->dev, "unrecognized bus speed %lu Hz\n",
pd                283 drivers/i2c/busses/i2c-sh_mobile.c 			pd->bus_speed);
pd                287 drivers/i2c/busses/i2c-sh_mobile.c 	pd->iccl = sh_mobile_i2c_iccl(i2c_clk_khz, tLOW, tf);
pd                288 drivers/i2c/busses/i2c-sh_mobile.c 	pd->icch = sh_mobile_i2c_icch(i2c_clk_khz, tHIGH, tf);
pd                290 drivers/i2c/busses/i2c-sh_mobile.c 	return sh_mobile_i2c_check_timing(pd);
pd                293 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_v2_init(struct sh_mobile_i2c_data *pd)
pd                298 drivers/i2c/busses/i2c-sh_mobile.c 	clks_per_cycle = clk_get_rate(pd->clk) / pd->bus_speed;
pd                299 drivers/i2c/busses/i2c-sh_mobile.c 	pd->iccl = DIV_ROUND_UP(clks_per_cycle * 5 / 9 - 1, pd->clks_per_count);
pd                300 drivers/i2c/busses/i2c-sh_mobile.c 	pd->icch = DIV_ROUND_UP(clks_per_cycle * 4 / 9 - 5, pd->clks_per_count);
pd                302 drivers/i2c/busses/i2c-sh_mobile.c 	return sh_mobile_i2c_check_timing(pd);
pd                305 drivers/i2c/busses/i2c-sh_mobile.c static unsigned char i2c_op(struct sh_mobile_i2c_data *pd, enum sh_mobile_i2c_op op)
pd                310 drivers/i2c/busses/i2c-sh_mobile.c 	dev_dbg(pd->dev, "op %d\n", op);
pd                312 drivers/i2c/busses/i2c-sh_mobile.c 	spin_lock_irqsave(&pd->lock, flags);
pd                316 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_ICE | ICCR_TRS | ICCR_BBSY);
pd                319 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICIC, ICIC_WAITE | ICIC_ALE | ICIC_TACKE);
pd                320 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICDR, i2c_8bit_addr_from_msg(pd->msg));
pd                323 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICDR, pd->msg->buf[pd->pos]);
pd                326 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, pd->send_stop ? ICCR_ICE | ICCR_TRS
pd                330 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP);
pd                333 drivers/i2c/busses/i2c-sh_mobile.c 		ret = iic_rd(pd, ICDR);
pd                336 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICIC,
pd                338 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK);
pd                341 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICIC,
pd                343 drivers/i2c/busses/i2c-sh_mobile.c 		ret = iic_rd(pd, ICDR);
pd                344 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK);
pd                348 drivers/i2c/busses/i2c-sh_mobile.c 	spin_unlock_irqrestore(&pd->lock, flags);
pd                350 drivers/i2c/busses/i2c-sh_mobile.c 	dev_dbg(pd->dev, "op %d, data out 0x%02x\n", op, ret);
pd                354 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_isr_tx(struct sh_mobile_i2c_data *pd)
pd                356 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->pos == pd->msg->len) {
pd                357 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_op(pd, OP_TX_STOP);
pd                361 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->pos == -1)
pd                362 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_op(pd, OP_TX_FIRST);
pd                364 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_op(pd, OP_TX);
pd                366 drivers/i2c/busses/i2c-sh_mobile.c 	pd->pos++;
pd                370 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_isr_rx(struct sh_mobile_i2c_data *pd)
pd                376 drivers/i2c/busses/i2c-sh_mobile.c 	real_pos = pd->pos - 2;
pd                378 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->pos == -1) {
pd                379 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_op(pd, OP_TX_FIRST);
pd                380 drivers/i2c/busses/i2c-sh_mobile.c 	} else if (pd->pos == 0) {
pd                381 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_op(pd, OP_TX_TO_RX);
pd                382 drivers/i2c/busses/i2c-sh_mobile.c 	} else if (pd->pos == pd->msg->len) {
pd                383 drivers/i2c/busses/i2c-sh_mobile.c 		if (pd->stop_after_dma) {
pd                385 drivers/i2c/busses/i2c-sh_mobile.c 			i2c_op(pd, OP_RX_STOP);
pd                386 drivers/i2c/busses/i2c-sh_mobile.c 			pd->pos++;
pd                391 drivers/i2c/busses/i2c-sh_mobile.c 			i2c_op(pd, OP_RX_STOP);
pd                393 drivers/i2c/busses/i2c-sh_mobile.c 			data = i2c_op(pd, OP_RX_STOP_DATA);
pd                395 drivers/i2c/busses/i2c-sh_mobile.c 		data = i2c_op(pd, OP_RX);
pd                399 drivers/i2c/busses/i2c-sh_mobile.c 		pd->msg->buf[real_pos] = data;
pd                401 drivers/i2c/busses/i2c-sh_mobile.c 	pd->pos++;
pd                402 drivers/i2c/busses/i2c-sh_mobile.c 	return pd->pos == (pd->msg->len + 2);
pd                407 drivers/i2c/busses/i2c-sh_mobile.c 	struct sh_mobile_i2c_data *pd = dev_id;
pd                411 drivers/i2c/busses/i2c-sh_mobile.c 	sr = iic_rd(pd, ICSR);
pd                412 drivers/i2c/busses/i2c-sh_mobile.c 	pd->sr |= sr; /* remember state */
pd                414 drivers/i2c/busses/i2c-sh_mobile.c 	dev_dbg(pd->dev, "i2c_isr 0x%02x 0x%02x %s %d %d!\n", sr, pd->sr,
pd                415 drivers/i2c/busses/i2c-sh_mobile.c 	       (pd->msg->flags & I2C_M_RD) ? "read" : "write",
pd                416 drivers/i2c/busses/i2c-sh_mobile.c 	       pd->pos, pd->msg->len);
pd                419 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->dma_direction == DMA_TO_DEVICE && pd->pos == 0)
pd                420 drivers/i2c/busses/i2c-sh_mobile.c 		iic_set_clr(pd, ICIC, ICIC_TDMAE, 0);
pd                423 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICSR, sr & ~(ICSR_AL | ICSR_TACK));
pd                424 drivers/i2c/busses/i2c-sh_mobile.c 	else if (pd->msg->flags & I2C_M_RD)
pd                425 drivers/i2c/busses/i2c-sh_mobile.c 		wakeup = sh_mobile_i2c_isr_rx(pd);
pd                427 drivers/i2c/busses/i2c-sh_mobile.c 		wakeup = sh_mobile_i2c_isr_tx(pd);
pd                430 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->dma_direction == DMA_FROM_DEVICE && pd->pos == 1)
pd                431 drivers/i2c/busses/i2c-sh_mobile.c 		iic_set_clr(pd, ICIC, ICIC_RDMAE, 0);
pd                434 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICSR, sr & ~ICSR_WAIT);
pd                437 drivers/i2c/busses/i2c-sh_mobile.c 		pd->sr |= SW_DONE;
pd                438 drivers/i2c/busses/i2c-sh_mobile.c 		wake_up(&pd->wait);
pd                442 drivers/i2c/busses/i2c-sh_mobile.c 	iic_rd(pd, ICSR);
pd                447 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_dma_unmap(struct sh_mobile_i2c_data *pd)
pd                449 drivers/i2c/busses/i2c-sh_mobile.c 	struct dma_chan *chan = pd->dma_direction == DMA_FROM_DEVICE
pd                450 drivers/i2c/busses/i2c-sh_mobile.c 				? pd->dma_rx : pd->dma_tx;
pd                452 drivers/i2c/busses/i2c-sh_mobile.c 	dma_unmap_single(chan->device->dev, sg_dma_address(&pd->sg),
pd                453 drivers/i2c/busses/i2c-sh_mobile.c 			 pd->msg->len, pd->dma_direction);
pd                455 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dma_direction = DMA_NONE;
pd                458 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_cleanup_dma(struct sh_mobile_i2c_data *pd)
pd                460 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->dma_direction == DMA_NONE)
pd                462 drivers/i2c/busses/i2c-sh_mobile.c 	else if (pd->dma_direction == DMA_FROM_DEVICE)
pd                463 drivers/i2c/busses/i2c-sh_mobile.c 		dmaengine_terminate_all(pd->dma_rx);
pd                464 drivers/i2c/busses/i2c-sh_mobile.c 	else if (pd->dma_direction == DMA_TO_DEVICE)
pd                465 drivers/i2c/busses/i2c-sh_mobile.c 		dmaengine_terminate_all(pd->dma_tx);
pd                467 drivers/i2c/busses/i2c-sh_mobile.c 	sh_mobile_i2c_dma_unmap(pd);
pd                472 drivers/i2c/busses/i2c-sh_mobile.c 	struct sh_mobile_i2c_data *pd = data;
pd                474 drivers/i2c/busses/i2c-sh_mobile.c 	sh_mobile_i2c_dma_unmap(pd);
pd                475 drivers/i2c/busses/i2c-sh_mobile.c 	pd->pos = pd->msg->len;
pd                476 drivers/i2c/busses/i2c-sh_mobile.c 	pd->stop_after_dma = true;
pd                478 drivers/i2c/busses/i2c-sh_mobile.c 	iic_set_clr(pd, ICIC, 0, ICIC_TDMAE | ICIC_RDMAE);
pd                517 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_xfer_dma(struct sh_mobile_i2c_data *pd)
pd                519 drivers/i2c/busses/i2c-sh_mobile.c 	bool read = pd->msg->flags & I2C_M_RD;
pd                521 drivers/i2c/busses/i2c-sh_mobile.c 	struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx;
pd                528 drivers/i2c/busses/i2c-sh_mobile.c 			chan = pd->dma_rx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_DEV_TO_MEM,
pd                529 drivers/i2c/busses/i2c-sh_mobile.c 									   pd->res->start + ICDR);
pd                531 drivers/i2c/busses/i2c-sh_mobile.c 			chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV,
pd                532 drivers/i2c/busses/i2c-sh_mobile.c 									   pd->res->start + ICDR);
pd                538 drivers/i2c/busses/i2c-sh_mobile.c 	dma_addr = dma_map_single(chan->device->dev, pd->dma_buf, pd->msg->len, dir);
pd                540 drivers/i2c/busses/i2c-sh_mobile.c 		dev_dbg(pd->dev, "dma map failed, using PIO\n");
pd                544 drivers/i2c/busses/i2c-sh_mobile.c 	sg_dma_len(&pd->sg) = pd->msg->len;
pd                545 drivers/i2c/busses/i2c-sh_mobile.c 	sg_dma_address(&pd->sg) = dma_addr;
pd                547 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dma_direction = dir;
pd                549 drivers/i2c/busses/i2c-sh_mobile.c 	txdesc = dmaengine_prep_slave_sg(chan, &pd->sg, 1,
pd                553 drivers/i2c/busses/i2c-sh_mobile.c 		dev_dbg(pd->dev, "dma prep slave sg failed, using PIO\n");
pd                554 drivers/i2c/busses/i2c-sh_mobile.c 		sh_mobile_i2c_cleanup_dma(pd);
pd                559 drivers/i2c/busses/i2c-sh_mobile.c 	txdesc->callback_param = pd;
pd                563 drivers/i2c/busses/i2c-sh_mobile.c 		dev_dbg(pd->dev, "submitting dma failed, using PIO\n");
pd                564 drivers/i2c/busses/i2c-sh_mobile.c 		sh_mobile_i2c_cleanup_dma(pd);
pd                571 drivers/i2c/busses/i2c-sh_mobile.c static void start_ch(struct sh_mobile_i2c_data *pd, struct i2c_msg *usr_msg,
pd                576 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_SCP);
pd                579 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP);
pd                582 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCL, pd->iccl & 0xff);
pd                583 drivers/i2c/busses/i2c-sh_mobile.c 		iic_wr(pd, ICCH, pd->icch & 0xff);
pd                586 drivers/i2c/busses/i2c-sh_mobile.c 	pd->msg = usr_msg;
pd                587 drivers/i2c/busses/i2c-sh_mobile.c 	pd->pos = -1;
pd                588 drivers/i2c/busses/i2c-sh_mobile.c 	pd->sr = 0;
pd                590 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dma_buf = i2c_get_dma_safe_msg_buf(pd->msg, 8);
pd                591 drivers/i2c/busses/i2c-sh_mobile.c 	if (pd->dma_buf)
pd                592 drivers/i2c/busses/i2c-sh_mobile.c 		sh_mobile_i2c_xfer_dma(pd);
pd                595 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICIC, ICIC_DTEE | ICIC_WAITE | ICIC_ALE | ICIC_TACKE);
pd                598 drivers/i2c/busses/i2c-sh_mobile.c static int poll_dte(struct sh_mobile_i2c_data *pd)
pd                603 drivers/i2c/busses/i2c-sh_mobile.c 		u_int8_t val = iic_rd(pd, ICSR);
pd                617 drivers/i2c/busses/i2c-sh_mobile.c static int poll_busy(struct sh_mobile_i2c_data *pd)
pd                622 drivers/i2c/busses/i2c-sh_mobile.c 		u_int8_t val = iic_rd(pd, ICSR);
pd                624 drivers/i2c/busses/i2c-sh_mobile.c 		dev_dbg(pd->dev, "val 0x%02x pd->sr 0x%02x\n", val, pd->sr);
pd                632 drivers/i2c/busses/i2c-sh_mobile.c 			val |= pd->sr;
pd                650 drivers/i2c/busses/i2c-sh_mobile.c 	struct sh_mobile_i2c_data *pd = i2c_get_adapdata(adapter);
pd                657 drivers/i2c/busses/i2c-sh_mobile.c 	pm_runtime_get_sync(pd->dev);
pd                661 drivers/i2c/busses/i2c-sh_mobile.c 		bool do_start = pd->send_stop || !i;
pd                663 drivers/i2c/busses/i2c-sh_mobile.c 		pd->send_stop = i == num - 1 || msg->flags & I2C_M_STOP;
pd                664 drivers/i2c/busses/i2c-sh_mobile.c 		pd->stop_after_dma = false;
pd                666 drivers/i2c/busses/i2c-sh_mobile.c 		start_ch(pd, msg, do_start);
pd                669 drivers/i2c/busses/i2c-sh_mobile.c 			i2c_op(pd, OP_START);
pd                672 drivers/i2c/busses/i2c-sh_mobile.c 		timeout = wait_event_timeout(pd->wait,
pd                673 drivers/i2c/busses/i2c-sh_mobile.c 				       pd->sr & (ICSR_TACK | SW_DONE),
pd                677 drivers/i2c/busses/i2c-sh_mobile.c 		i2c_put_dma_safe_msg_buf(pd->dma_buf, pd->msg, pd->stop_after_dma);
pd                680 drivers/i2c/busses/i2c-sh_mobile.c 			dev_err(pd->dev, "Transfer request timed out\n");
pd                681 drivers/i2c/busses/i2c-sh_mobile.c 			if (pd->dma_direction != DMA_NONE)
pd                682 drivers/i2c/busses/i2c-sh_mobile.c 				sh_mobile_i2c_cleanup_dma(pd);
pd                688 drivers/i2c/busses/i2c-sh_mobile.c 		if (pd->send_stop)
pd                689 drivers/i2c/busses/i2c-sh_mobile.c 			err = poll_busy(pd);
pd                691 drivers/i2c/busses/i2c-sh_mobile.c 			err = poll_dte(pd);
pd                697 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICCR, ICCR_SCP);
pd                700 drivers/i2c/busses/i2c-sh_mobile.c 	pm_runtime_put_sync(pd->dev);
pd                722 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_r8a7740_workaround(struct sh_mobile_i2c_data *pd)
pd                724 drivers/i2c/busses/i2c-sh_mobile.c 	iic_set_clr(pd, ICCR, ICCR_ICE, 0);
pd                725 drivers/i2c/busses/i2c-sh_mobile.c 	iic_rd(pd, ICCR); /* dummy read */
pd                727 drivers/i2c/busses/i2c-sh_mobile.c 	iic_set_clr(pd, ICSTART, ICSTART_ICSTART, 0);
pd                728 drivers/i2c/busses/i2c-sh_mobile.c 	iic_rd(pd, ICSTART); /* dummy read */
pd                732 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICCR, ICCR_SCP);
pd                733 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICSTART, 0);
pd                737 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICCR, ICCR_TRS);
pd                739 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICCR, 0);
pd                741 drivers/i2c/busses/i2c-sh_mobile.c 	iic_wr(pd, ICCR, ICCR_TRS);
pd                744 drivers/i2c/busses/i2c-sh_mobile.c 	return sh_mobile_i2c_init(pd);
pd                786 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_release_dma(struct sh_mobile_i2c_data *pd)
pd                788 drivers/i2c/busses/i2c-sh_mobile.c 	if (!IS_ERR(pd->dma_tx)) {
pd                789 drivers/i2c/busses/i2c-sh_mobile.c 		dma_release_channel(pd->dma_tx);
pd                790 drivers/i2c/busses/i2c-sh_mobile.c 		pd->dma_tx = ERR_PTR(-EPROBE_DEFER);
pd                793 drivers/i2c/busses/i2c-sh_mobile.c 	if (!IS_ERR(pd->dma_rx)) {
pd                794 drivers/i2c/busses/i2c-sh_mobile.c 		dma_release_channel(pd->dma_rx);
pd                795 drivers/i2c/busses/i2c-sh_mobile.c 		pd->dma_rx = ERR_PTR(-EPROBE_DEFER);
pd                799 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_hook_irqs(struct platform_device *dev, struct sh_mobile_i2c_data *pd)
pd                808 drivers/i2c/busses/i2c-sh_mobile.c 					  0, dev_name(&dev->dev), pd);
pd                822 drivers/i2c/busses/i2c-sh_mobile.c 	struct sh_mobile_i2c_data *pd;
pd                829 drivers/i2c/busses/i2c-sh_mobile.c 	pd = devm_kzalloc(&dev->dev, sizeof(struct sh_mobile_i2c_data), GFP_KERNEL);
pd                830 drivers/i2c/busses/i2c-sh_mobile.c 	if (!pd)
pd                833 drivers/i2c/busses/i2c-sh_mobile.c 	pd->clk = devm_clk_get(&dev->dev, NULL);
pd                834 drivers/i2c/busses/i2c-sh_mobile.c 	if (IS_ERR(pd->clk)) {
pd                836 drivers/i2c/busses/i2c-sh_mobile.c 		return PTR_ERR(pd->clk);
pd                839 drivers/i2c/busses/i2c-sh_mobile.c 	ret = sh_mobile_i2c_hook_irqs(dev, pd);
pd                843 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dev = &dev->dev;
pd                844 drivers/i2c/busses/i2c-sh_mobile.c 	platform_set_drvdata(dev, pd);
pd                848 drivers/i2c/busses/i2c-sh_mobile.c 	pd->res = res;
pd                849 drivers/i2c/busses/i2c-sh_mobile.c 	pd->reg = devm_ioremap_resource(&dev->dev, res);
pd                850 drivers/i2c/busses/i2c-sh_mobile.c 	if (IS_ERR(pd->reg))
pd                851 drivers/i2c/busses/i2c-sh_mobile.c 		return PTR_ERR(pd->reg);
pd                854 drivers/i2c/busses/i2c-sh_mobile.c 	pd->bus_speed = (ret || !bus_speed) ? STANDARD_MODE : bus_speed;
pd                855 drivers/i2c/busses/i2c-sh_mobile.c 	pd->clks_per_count = 1;
pd                859 drivers/i2c/busses/i2c-sh_mobile.c 		pd->flags |= IIC_FLAG_HAS_ICIC67;
pd                866 drivers/i2c/busses/i2c-sh_mobile.c 		pd->clks_per_count = config->clks_per_count;
pd                867 drivers/i2c/busses/i2c-sh_mobile.c 		ret = config->setup(pd);
pd                869 drivers/i2c/busses/i2c-sh_mobile.c 		ret = sh_mobile_i2c_init(pd);
pd                877 drivers/i2c/busses/i2c-sh_mobile.c 	sg_init_table(&pd->sg, 1);
pd                878 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dma_direction = DMA_NONE;
pd                879 drivers/i2c/busses/i2c-sh_mobile.c 	pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER);
pd                882 drivers/i2c/busses/i2c-sh_mobile.c 	adap = &pd->adap;
pd                883 drivers/i2c/busses/i2c-sh_mobile.c 	i2c_set_adapdata(adap, pd);
pd                895 drivers/i2c/busses/i2c-sh_mobile.c 	spin_lock_init(&pd->lock);
pd                896 drivers/i2c/busses/i2c-sh_mobile.c 	init_waitqueue_head(&pd->wait);
pd                900 drivers/i2c/busses/i2c-sh_mobile.c 		sh_mobile_i2c_release_dma(pd);
pd                904 drivers/i2c/busses/i2c-sh_mobile.c 	dev_info(&dev->dev, "I2C adapter %d, bus speed %lu Hz\n", adap->nr, pd->bus_speed);
pd                911 drivers/i2c/busses/i2c-sh_mobile.c 	struct sh_mobile_i2c_data *pd = platform_get_drvdata(dev);
pd                913 drivers/i2c/busses/i2c-sh_mobile.c 	i2c_del_adapter(&pd->adap);
pd                914 drivers/i2c/busses/i2c-sh_mobile.c 	sh_mobile_i2c_release_dma(pd);
pd                 36 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd = pw;
pd                 37 drivers/i2c/busses/i2c-simtec.c 	writeb(CMD_SET_SDA | (state ? STATE_SDA : 0), pd->reg);
pd                 42 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd = pw;
pd                 43 drivers/i2c/busses/i2c-simtec.c 	writeb(CMD_SET_SCL | (state ? STATE_SCL : 0), pd->reg);
pd                 48 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd = pw;
pd                 49 drivers/i2c/busses/i2c-simtec.c 	return readb(pd->reg) & STATE_SDA ? 1 : 0;
pd                 54 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd = pw;
pd                 55 drivers/i2c/busses/i2c-simtec.c 	return readb(pd->reg) & STATE_SCL ? 1 : 0;
pd                 62 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd;
pd                 67 drivers/i2c/busses/i2c-simtec.c 	pd = kzalloc(sizeof(struct simtec_i2c_data), GFP_KERNEL);
pd                 68 drivers/i2c/busses/i2c-simtec.c 	if (pd == NULL)
pd                 71 drivers/i2c/busses/i2c-simtec.c 	platform_set_drvdata(dev, pd);
pd                 82 drivers/i2c/busses/i2c-simtec.c 	pd->ioarea = request_mem_region(res->start, size, dev->name);
pd                 83 drivers/i2c/busses/i2c-simtec.c 	if (pd->ioarea == NULL) {
pd                 89 drivers/i2c/busses/i2c-simtec.c 	pd->reg = ioremap(res->start, size);
pd                 90 drivers/i2c/busses/i2c-simtec.c 	if (pd->reg == NULL) {
pd                 98 drivers/i2c/busses/i2c-simtec.c 	pd->adap.owner = THIS_MODULE;
pd                 99 drivers/i2c/busses/i2c-simtec.c 	pd->adap.algo_data = &pd->bit;
pd                100 drivers/i2c/busses/i2c-simtec.c 	pd->adap.dev.parent = &dev->dev;
pd                102 drivers/i2c/busses/i2c-simtec.c 	strlcpy(pd->adap.name, "Simtec I2C", sizeof(pd->adap.name));
pd                104 drivers/i2c/busses/i2c-simtec.c 	pd->bit.data = pd;
pd                105 drivers/i2c/busses/i2c-simtec.c 	pd->bit.setsda = simtec_i2c_setsda;
pd                106 drivers/i2c/busses/i2c-simtec.c 	pd->bit.setscl = simtec_i2c_setscl;
pd                107 drivers/i2c/busses/i2c-simtec.c 	pd->bit.getsda = simtec_i2c_getsda;
pd                108 drivers/i2c/busses/i2c-simtec.c 	pd->bit.getscl = simtec_i2c_getscl;
pd                109 drivers/i2c/busses/i2c-simtec.c 	pd->bit.timeout = HZ;
pd                110 drivers/i2c/busses/i2c-simtec.c 	pd->bit.udelay = 20;
pd                112 drivers/i2c/busses/i2c-simtec.c 	ret = i2c_bit_add_bus(&pd->adap);
pd                119 drivers/i2c/busses/i2c-simtec.c 	iounmap(pd->reg);
pd                122 drivers/i2c/busses/i2c-simtec.c 	release_mem_region(pd->ioarea->start, size);
pd                125 drivers/i2c/busses/i2c-simtec.c 	kfree(pd);
pd                131 drivers/i2c/busses/i2c-simtec.c 	struct simtec_i2c_data *pd = platform_get_drvdata(dev);
pd                133 drivers/i2c/busses/i2c-simtec.c 	i2c_del_adapter(&pd->adap);
pd                135 drivers/i2c/busses/i2c-simtec.c 	iounmap(pd->reg);
pd                136 drivers/i2c/busses/i2c-simtec.c 	release_mem_region(pd->ioarea->start, resource_size(pd->ioarea));
pd                137 drivers/i2c/busses/i2c-simtec.c 	kfree(pd);
pd                391 drivers/iio/dac/mcp4725.c 	u8 pd;
pd                470 drivers/iio/dac/mcp4725.c 	pd = (inbuf[0] >> 1) & 0x3;
pd                471 drivers/iio/dac/mcp4725.c 	data->powerdown = pd > 0;
pd                472 drivers/iio/dac/mcp4725.c 	data->powerdown_mode = pd ? pd - 1 : 2; /* largest resistor to gnd */
pd                102 drivers/infiniband/core/agent.c 	ah = ib_create_ah_from_wc(agent->qp->pd, wc, grh, port_num);
pd                347 drivers/infiniband/core/cm.c 	ah = rdma_create_ah(mad_agent->qp->pd, &av->ah_attr, 0);
pd                392 drivers/infiniband/core/cm.c 	ah = ib_create_ah_from_wc(port->mad_agent->qp->pd, mad_recv_wc->wc,
pd                950 drivers/infiniband/core/cma.c int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd,
pd                958 drivers/infiniband/core/cma.c 	if (id->device != pd->device)
pd                962 drivers/infiniband/core/cma.c 	qp = ib_create_qp(pd, qp_init_attr);
pd                321 drivers/infiniband/core/core_priv.h 					  struct ib_pd *pd,
pd                333 drivers/infiniband/core/core_priv.h 	qp = dev->ops.create_qp(pd, attr, udata);
pd                338 drivers/infiniband/core/core_priv.h 	qp->pd = pd;
pd                194 drivers/infiniband/core/fmr_pool.c struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd             *pd,
pd                206 drivers/infiniband/core/fmr_pool.c 	device = pd->device;
pd                284 drivers/infiniband/core/fmr_pool.c 			fmr->fmr = ib_alloc_fmr(pd, params->access, &fmr_attr);
pd               1097 drivers/infiniband/core/mad.c 	mad_send_wr->sg_list[0].lkey = mad_agent->qp->pd->local_dma_lkey;
pd               1106 drivers/infiniband/core/mad.c 	mad_send_wr->sg_list[1].lkey = mad_agent->qp->pd->local_dma_lkey;
pd               2936 drivers/infiniband/core/mad.c 	sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey;
pd               3165 drivers/infiniband/core/mad.c 	qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr);
pd               3227 drivers/infiniband/core/mad.c 	port_priv->pd = ib_alloc_pd(device, 0);
pd               3228 drivers/infiniband/core/mad.c 	if (IS_ERR(port_priv->pd)) {
pd               3230 drivers/infiniband/core/mad.c 		ret = PTR_ERR(port_priv->pd);
pd               3285 drivers/infiniband/core/mad.c 	ib_dealloc_pd(port_priv->pd);
pd               3316 drivers/infiniband/core/mad.c 	ib_dealloc_pd(port_priv->pd);
pd                204 drivers/infiniband/core/mad_priv.h 	struct ib_pd *pd;
pd                163 drivers/infiniband/core/mad_rmpp.c 	ah = ib_create_ah_from_wc(agent->qp->pd, recv_wc->wc,
pd                295 drivers/infiniband/core/mad_rmpp.c 	rmpp_recv->ah = ib_create_ah_from_wc(agent->agent.qp->pd,
pd                 45 drivers/infiniband/core/mr_pool.c 			mr = ib_alloc_mr_integrity(qp->pd, max_num_sg,
pd                 48 drivers/infiniband/core/mr_pool.c 			mr = ib_alloc_mr(qp->pd, type, max_num_sg);
pd                473 drivers/infiniband/core/nldev.c 	    nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PDN, qp->pd->res.id))
pd                582 drivers/infiniband/core/nldev.c 	struct ib_device *dev = mr->pd->device;
pd                599 drivers/infiniband/core/nldev.c 	    nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PDN, mr->pd->res.id))
pd                616 drivers/infiniband/core/nldev.c 	struct ib_pd *pd = container_of(res, struct ib_pd, res);
pd                617 drivers/infiniband/core/nldev.c 	struct ib_device *dev = pd->device;
pd                621 drivers/infiniband/core/nldev.c 				pd->local_dma_lkey))
pd                623 drivers/infiniband/core/nldev.c 		if ((pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) &&
pd                625 drivers/infiniband/core/nldev.c 				pd->unsafe_global_rkey))
pd                629 drivers/infiniband/core/nldev.c 			      atomic_read(&pd->usecnt), RDMA_NLDEV_ATTR_PAD))
pd                637 drivers/infiniband/core/nldev.c 			pd->uobject->context->res.id))
pd               1417 drivers/infiniband/core/nldev.c RES_GET_FUNCS(pd, RDMA_RESTRACK_PD);
pd                131 drivers/infiniband/core/restrack.c 	struct ib_pd *pd;
pd                135 drivers/infiniband/core/restrack.c 		pd = container_of(res, struct ib_qp, res)->pd;
pd                136 drivers/infiniband/core/restrack.c 		if (!pd) {
pd                143 drivers/infiniband/core/restrack.c 		pd = container_of(res, struct ib_mr, res)->pd;
pd                147 drivers/infiniband/core/restrack.c 		pd = NULL;
pd                151 drivers/infiniband/core/restrack.c 	if (pd)
pd                152 drivers/infiniband/core/restrack.c 		res->kern_name = pd->res.kern_name;
pd                 89 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pd                123 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pd                224 drivers/infiniband/core/rw.c 			sge->lkey = qp->pd->local_dma_lkey;
pd                253 drivers/infiniband/core/rw.c 	ctx->single.sge.lkey = qp->pd->local_dma_lkey;
pd                307 drivers/infiniband/core/rw.c 	struct ib_device *dev = qp->pd->device;
pd                377 drivers/infiniband/core/rw.c 	struct ib_device *dev = qp->pd->device;
pd                378 drivers/infiniband/core/rw.c 	u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device,
pd                599 drivers/infiniband/core/rw.c 	rdma_rw_unmap_sg(qp->pd->device, sg, sg_cnt, dir);
pd                626 drivers/infiniband/core/rw.c 	ib_dma_unmap_sg(qp->pd->device, sg, sg_cnt, dir);
pd                628 drivers/infiniband/core/rw.c 		ib_dma_unmap_sg(qp->pd->device, prot_sg, prot_sg_cnt, dir);
pd                691 drivers/infiniband/core/rw.c 	struct ib_device *dev = qp->pd->device;
pd               2271 drivers/infiniband/core/sa_query.c 	new_ah->ah = rdma_create_ah(port->agent->qp->pd, &ah_attr,
pd                547 drivers/infiniband/core/user_mad.c 	ah = rdma_create_user_ah(agent->qp->pd, &ah_attr, NULL);
pd                410 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd                  *pd;
pd                422 drivers/infiniband/core/uverbs_cmd.c 	pd = rdma_zalloc_drv_obj(ib_dev, ib_pd);
pd                423 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd                428 drivers/infiniband/core/uverbs_cmd.c 	pd->device  = ib_dev;
pd                429 drivers/infiniband/core/uverbs_cmd.c 	pd->uobject = uobj;
pd                430 drivers/infiniband/core/uverbs_cmd.c 	pd->__internal_mr = NULL;
pd                431 drivers/infiniband/core/uverbs_cmd.c 	atomic_set(&pd->usecnt, 0);
pd                432 drivers/infiniband/core/uverbs_cmd.c 	pd->res.type = RDMA_RESTRACK_PD;
pd                434 drivers/infiniband/core/uverbs_cmd.c 	ret = ib_dev->ops.alloc_pd(pd, &attrs->driver_udata);
pd                438 drivers/infiniband/core/uverbs_cmd.c 	uobj->object = pd;
pd                441 drivers/infiniband/core/uverbs_cmd.c 	rdma_restrack_uadd(&pd->res);
pd                450 drivers/infiniband/core/uverbs_cmd.c 	ib_dealloc_pd_user(pd, uverbs_get_cleared_udata(attrs));
pd                451 drivers/infiniband/core/uverbs_cmd.c 	pd = NULL;
pd                453 drivers/infiniband/core/uverbs_cmd.c 	kfree(pd);
pd                709 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd                *pd;
pd                729 drivers/infiniband/core/uverbs_cmd.c 	pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
pd                730 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd                736 drivers/infiniband/core/uverbs_cmd.c 		if (!(pd->device->attrs.device_cap_flags &
pd                744 drivers/infiniband/core/uverbs_cmd.c 	mr = pd->device->ops.reg_user_mr(pd, cmd.start, cmd.length, cmd.hca_va,
pd                752 drivers/infiniband/core/uverbs_cmd.c 	mr->device  = pd->device;
pd                753 drivers/infiniband/core/uverbs_cmd.c 	mr->pd      = pd;
pd                758 drivers/infiniband/core/uverbs_cmd.c 	atomic_inc(&pd->usecnt);
pd                773 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd                781 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd                792 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd                *pd = NULL;
pd                828 drivers/infiniband/core/uverbs_cmd.c 		pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle,
pd                830 drivers/infiniband/core/uverbs_cmd.c 		if (!pd) {
pd                836 drivers/infiniband/core/uverbs_cmd.c 	old_pd = mr->pd;
pd                839 drivers/infiniband/core/uverbs_cmd.c 					    cmd.access_flags, pd,
pd                845 drivers/infiniband/core/uverbs_cmd.c 		atomic_inc(&pd->usecnt);
pd                846 drivers/infiniband/core/uverbs_cmd.c 		mr->pd = pd;
pd                858 drivers/infiniband/core/uverbs_cmd.c 		uobj_put_obj_read(pd);
pd                883 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd                  *pd;
pd                896 drivers/infiniband/core/uverbs_cmd.c 	pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
pd                897 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd                907 drivers/infiniband/core/uverbs_cmd.c 	mw = pd->device->ops.alloc_mw(pd, cmd.mw_type, &attrs->driver_udata);
pd                913 drivers/infiniband/core/uverbs_cmd.c 	mw->device  = pd->device;
pd                914 drivers/infiniband/core/uverbs_cmd.c 	mw->pd      = pd;
pd                916 drivers/infiniband/core/uverbs_cmd.c 	atomic_inc(&pd->usecnt);
pd                928 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd                934 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               1274 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd			*pd = NULL;
pd               1365 drivers/infiniband/core/uverbs_cmd.c 		pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle,
pd               1367 drivers/infiniband/core/uverbs_cmd.c 		if (!pd || (!scq && has_sq)) {
pd               1372 drivers/infiniband/core/uverbs_cmd.c 		device = pd->device;
pd               1419 drivers/infiniband/core/uverbs_cmd.c 		qp = ib_create_qp(pd, &attr);
pd               1421 drivers/infiniband/core/uverbs_cmd.c 		qp = _ib_create_qp(device, pd, &attr, &attrs->driver_udata,
pd               1434 drivers/infiniband/core/uverbs_cmd.c 		atomic_inc(&pd->usecnt);
pd               1471 drivers/infiniband/core/uverbs_cmd.c 	if (pd)
pd               1472 drivers/infiniband/core/uverbs_cmd.c 		uobj_put_obj_read(pd);
pd               1489 drivers/infiniband/core/uverbs_cmd.c 	if (pd)
pd               1490 drivers/infiniband/core/uverbs_cmd.c 		uobj_put_obj_read(pd);
pd               2396 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd			*pd;
pd               2415 drivers/infiniband/core/uverbs_cmd.c 	pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
pd               2416 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd               2439 drivers/infiniband/core/uverbs_cmd.c 	ah = rdma_create_user_ah(pd, &attr, &attrs->driver_udata);
pd               2455 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               2463 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               2897 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd *pd;
pd               2914 drivers/infiniband/core/uverbs_cmd.c 	pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs);
pd               2915 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd               2936 drivers/infiniband/core/uverbs_cmd.c 	wq = pd->device->ops.create_wq(pd, &wq_init_attr, &attrs->driver_udata);
pd               2946 drivers/infiniband/core/uverbs_cmd.c 	wq->pd = pd;
pd               2947 drivers/infiniband/core/uverbs_cmd.c 	wq->device = pd->device;
pd               2950 drivers/infiniband/core/uverbs_cmd.c 	atomic_inc(&pd->usecnt);
pd               2965 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               2974 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               3358 drivers/infiniband/core/uverbs_cmd.c 	struct ib_pd                    *pd;
pd               3400 drivers/infiniband/core/uverbs_cmd.c 	pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle, attrs);
pd               3401 drivers/infiniband/core/uverbs_cmd.c 	if (!pd) {
pd               3422 drivers/infiniband/core/uverbs_cmd.c 	srq->device        = pd->device;
pd               3423 drivers/infiniband/core/uverbs_cmd.c 	srq->pd            = pd;
pd               3429 drivers/infiniband/core/uverbs_cmd.c 	ret = pd->device->ops.create_srq(srq, &attr, udata);
pd               3443 drivers/infiniband/core/uverbs_cmd.c 	atomic_inc(&pd->usecnt);
pd               3466 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd               3476 drivers/infiniband/core/uverbs_cmd.c 	uobj_put_obj_read(pd);
pd                107 drivers/infiniband/core/uverbs_main.c 	struct ib_pd *pd = mw->pd;
pd                112 drivers/infiniband/core/uverbs_main.c 		atomic_dec(&pd->usecnt);
pd                194 drivers/infiniband/core/uverbs_std_types.c 	struct ib_pd *pd = uobject->object;
pd                197 drivers/infiniband/core/uverbs_std_types.c 	ret = ib_destroy_usecnt(&pd->usecnt, why, uobject);
pd                201 drivers/infiniband/core/uverbs_std_types.c 	ib_dealloc_pd_user(pd, &attrs->driver_udata);
pd                 48 drivers/infiniband/core/uverbs_std_types_mr.c 	struct ib_pd *pd =
pd                 51 drivers/infiniband/core/uverbs_std_types_mr.c 	struct ib_device *ib_dev = pd->device;
pd                 77 drivers/infiniband/core/uverbs_std_types_mr.c 	return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge,
pd                 89 drivers/infiniband/core/uverbs_std_types_mr.c 	struct ib_pd *pd =
pd                 91 drivers/infiniband/core/uverbs_std_types_mr.c 	struct ib_device *ib_dev = pd->device;
pd                125 drivers/infiniband/core/uverbs_std_types_mr.c 	mr = pd->device->ops.reg_dm_mr(pd, dm, &attr, attrs);
pd                129 drivers/infiniband/core/uverbs_std_types_mr.c 	mr->device  = pd->device;
pd                130 drivers/infiniband/core/uverbs_std_types_mr.c 	mr->pd      = pd;
pd                134 drivers/infiniband/core/uverbs_std_types_mr.c 	atomic_inc(&pd->usecnt);
pd                257 drivers/infiniband/core/verbs.c 	struct ib_pd *pd;
pd                261 drivers/infiniband/core/verbs.c 	pd = rdma_zalloc_drv_obj(device, ib_pd);
pd                262 drivers/infiniband/core/verbs.c 	if (!pd)
pd                265 drivers/infiniband/core/verbs.c 	pd->device = device;
pd                266 drivers/infiniband/core/verbs.c 	pd->uobject = NULL;
pd                267 drivers/infiniband/core/verbs.c 	pd->__internal_mr = NULL;
pd                268 drivers/infiniband/core/verbs.c 	atomic_set(&pd->usecnt, 0);
pd                269 drivers/infiniband/core/verbs.c 	pd->flags = flags;
pd                271 drivers/infiniband/core/verbs.c 	pd->res.type = RDMA_RESTRACK_PD;
pd                272 drivers/infiniband/core/verbs.c 	rdma_restrack_set_task(&pd->res, caller);
pd                274 drivers/infiniband/core/verbs.c 	ret = device->ops.alloc_pd(pd, NULL);
pd                276 drivers/infiniband/core/verbs.c 		kfree(pd);
pd                279 drivers/infiniband/core/verbs.c 	rdma_restrack_kadd(&pd->res);
pd                282 drivers/infiniband/core/verbs.c 		pd->local_dma_lkey = device->local_dma_lkey;
pd                294 drivers/infiniband/core/verbs.c 		mr = pd->device->ops.get_dma_mr(pd, mr_access_flags);
pd                296 drivers/infiniband/core/verbs.c 			ib_dealloc_pd(pd);
pd                300 drivers/infiniband/core/verbs.c 		mr->device	= pd->device;
pd                301 drivers/infiniband/core/verbs.c 		mr->pd		= pd;
pd                306 drivers/infiniband/core/verbs.c 		pd->__internal_mr = mr;
pd                309 drivers/infiniband/core/verbs.c 			pd->local_dma_lkey = pd->__internal_mr->lkey;
pd                312 drivers/infiniband/core/verbs.c 			pd->unsafe_global_rkey = pd->__internal_mr->rkey;
pd                315 drivers/infiniband/core/verbs.c 	return pd;
pd                328 drivers/infiniband/core/verbs.c void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata)
pd                332 drivers/infiniband/core/verbs.c 	if (pd->__internal_mr) {
pd                333 drivers/infiniband/core/verbs.c 		ret = pd->device->ops.dereg_mr(pd->__internal_mr, NULL);
pd                335 drivers/infiniband/core/verbs.c 		pd->__internal_mr = NULL;
pd                340 drivers/infiniband/core/verbs.c 	WARN_ON(atomic_read(&pd->usecnt));
pd                342 drivers/infiniband/core/verbs.c 	rdma_restrack_del(&pd->res);
pd                343 drivers/infiniband/core/verbs.c 	pd->device->ops.dealloc_pd(pd, udata);
pd                344 drivers/infiniband/core/verbs.c 	kfree(pd);
pd                497 drivers/infiniband/core/verbs.c static struct ib_ah *_rdma_create_ah(struct ib_pd *pd,
pd                502 drivers/infiniband/core/verbs.c 	struct ib_device *device = pd->device;
pd                518 drivers/infiniband/core/verbs.c 	ah->pd = pd;
pd                528 drivers/infiniband/core/verbs.c 	atomic_inc(&pd->usecnt);
pd                543 drivers/infiniband/core/verbs.c struct ib_ah *rdma_create_ah(struct ib_pd *pd, struct rdma_ah_attr *ah_attr,
pd                550 drivers/infiniband/core/verbs.c 	ret = rdma_fill_sgid_attr(pd->device, ah_attr, &old_sgid_attr);
pd                554 drivers/infiniband/core/verbs.c 	ah = _rdma_create_ah(pd, ah_attr, flags, NULL);
pd                574 drivers/infiniband/core/verbs.c struct ib_ah *rdma_create_user_ah(struct ib_pd *pd,
pd                582 drivers/infiniband/core/verbs.c 	err = rdma_fill_sgid_attr(pd->device, ah_attr, &old_sgid_attr);
pd                587 drivers/infiniband/core/verbs.c 		err = ib_resolve_eth_dmac(pd->device, ah_attr);
pd                594 drivers/infiniband/core/verbs.c 	ah = _rdma_create_ah(pd, ah_attr, RDMA_CREATE_AH_SLEEPABLE, udata);
pd                896 drivers/infiniband/core/verbs.c struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc,
pd                903 drivers/infiniband/core/verbs.c 	ret = ib_init_ah_attr_from_wc(pd->device, port_num, wc, grh, &ah_attr);
pd                907 drivers/infiniband/core/verbs.c 	ah = rdma_create_ah(pd, &ah_attr, RDMA_CREATE_AH_SLEEPABLE);
pd                949 drivers/infiniband/core/verbs.c 	struct ib_pd *pd;
pd                953 drivers/infiniband/core/verbs.c 	pd = ah->pd;
pd                956 drivers/infiniband/core/verbs.c 	atomic_dec(&pd->usecnt);
pd                967 drivers/infiniband/core/verbs.c struct ib_srq *ib_create_srq(struct ib_pd *pd,
pd                973 drivers/infiniband/core/verbs.c 	if (!pd->device->ops.create_srq)
pd                976 drivers/infiniband/core/verbs.c 	srq = rdma_zalloc_drv_obj(pd->device, ib_srq);
pd                980 drivers/infiniband/core/verbs.c 	srq->device = pd->device;
pd                981 drivers/infiniband/core/verbs.c 	srq->pd = pd;
pd                994 drivers/infiniband/core/verbs.c 	atomic_inc(&pd->usecnt);
pd                996 drivers/infiniband/core/verbs.c 	ret = pd->device->ops.create_srq(srq, srq_init_attr, NULL);
pd                998 drivers/infiniband/core/verbs.c 		atomic_dec(&srq->pd->usecnt);
pd               1036 drivers/infiniband/core/verbs.c 	atomic_dec(&srq->pd->usecnt);
pd               1132 drivers/infiniband/core/verbs.c 	qp->pd = NULL;
pd               1148 drivers/infiniband/core/verbs.c struct ib_qp *ib_create_qp_user(struct ib_pd *pd,
pd               1152 drivers/infiniband/core/verbs.c 	struct ib_device *device = pd ? pd->device : qp_init_attr->xrcd->device;
pd               1175 drivers/infiniband/core/verbs.c 	qp = _ib_create_qp(device, pd, qp_init_attr, NULL, NULL);
pd               1211 drivers/infiniband/core/verbs.c 	atomic_inc(&pd->usecnt);
pd               1860 drivers/infiniband/core/verbs.c 	struct ib_pd *pd;
pd               1875 drivers/infiniband/core/verbs.c 	pd   = qp->pd;
pd               1895 drivers/infiniband/core/verbs.c 		if (pd)
pd               1896 drivers/infiniband/core/verbs.c 			atomic_dec(&pd->usecnt);
pd               1983 drivers/infiniband/core/verbs.c 	struct ib_pd *pd = mr->pd;
pd               1991 drivers/infiniband/core/verbs.c 		atomic_dec(&pd->usecnt);
pd               2014 drivers/infiniband/core/verbs.c struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               2019 drivers/infiniband/core/verbs.c 	if (!pd->device->ops.alloc_mr)
pd               2025 drivers/infiniband/core/verbs.c 	mr = pd->device->ops.alloc_mr(pd, mr_type, max_num_sg, udata);
pd               2027 drivers/infiniband/core/verbs.c 		mr->device  = pd->device;
pd               2028 drivers/infiniband/core/verbs.c 		mr->pd      = pd;
pd               2031 drivers/infiniband/core/verbs.c 		atomic_inc(&pd->usecnt);
pd               2055 drivers/infiniband/core/verbs.c struct ib_mr *ib_alloc_mr_integrity(struct ib_pd *pd,
pd               2062 drivers/infiniband/core/verbs.c 	if (!pd->device->ops.alloc_mr_integrity ||
pd               2063 drivers/infiniband/core/verbs.c 	    !pd->device->ops.map_mr_sg_pi)
pd               2073 drivers/infiniband/core/verbs.c 	mr = pd->device->ops.alloc_mr_integrity(pd, max_num_data_sg,
pd               2080 drivers/infiniband/core/verbs.c 	mr->device = pd->device;
pd               2081 drivers/infiniband/core/verbs.c 	mr->pd = pd;
pd               2084 drivers/infiniband/core/verbs.c 	atomic_inc(&pd->usecnt);
pd               2097 drivers/infiniband/core/verbs.c struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd,
pd               2103 drivers/infiniband/core/verbs.c 	if (!pd->device->ops.alloc_fmr)
pd               2106 drivers/infiniband/core/verbs.c 	fmr = pd->device->ops.alloc_fmr(pd, mr_access_flags, fmr_attr);
pd               2108 drivers/infiniband/core/verbs.c 		fmr->device = pd->device;
pd               2109 drivers/infiniband/core/verbs.c 		fmr->pd     = pd;
pd               2110 drivers/infiniband/core/verbs.c 		atomic_inc(&pd->usecnt);
pd               2131 drivers/infiniband/core/verbs.c 	struct ib_pd *pd;
pd               2134 drivers/infiniband/core/verbs.c 	pd = fmr->pd;
pd               2137 drivers/infiniband/core/verbs.c 		atomic_dec(&pd->usecnt);
pd               2273 drivers/infiniband/core/verbs.c struct ib_wq *ib_create_wq(struct ib_pd *pd,
pd               2278 drivers/infiniband/core/verbs.c 	if (!pd->device->ops.create_wq)
pd               2281 drivers/infiniband/core/verbs.c 	wq = pd->device->ops.create_wq(pd, wq_attr, NULL);
pd               2287 drivers/infiniband/core/verbs.c 		wq->device = pd->device;
pd               2288 drivers/infiniband/core/verbs.c 		wq->pd = pd;
pd               2290 drivers/infiniband/core/verbs.c 		atomic_inc(&pd->usecnt);
pd               2306 drivers/infiniband/core/verbs.c 	struct ib_pd *pd = wq->pd;
pd               2312 drivers/infiniband/core/verbs.c 	atomic_dec(&pd->usecnt);
pd                405 drivers/infiniband/hw/bnxt_re/ib_verbs.c static void bnxt_re_create_fence_wqe(struct bnxt_re_pd *pd)
pd                407 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_fence_data *fence = &pd->fence;
pd                433 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct ib_pd *ib_pd = qp->ib_qp.pd;
pd                434 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd                435 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_fence_data *fence = &pd->fence;
pd                446 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		wqe.bind.r_key, qp->qplib_qp.id, pd);
pd                457 drivers/infiniband/hw/bnxt_re/ib_verbs.c static void bnxt_re_destroy_fence_mr(struct bnxt_re_pd *pd)
pd                459 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_fence_data *fence = &pd->fence;
pd                460 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                484 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_create_fence_mr(struct bnxt_re_pd *pd)
pd                487 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_fence_data *fence = &pd->fence;
pd                488 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                515 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mr->qplib_mr.pd = &pd->qplib_pd;
pd                538 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mw = bnxt_re_alloc_mw(&pd->ib_pd, IB_MW_TYPE_1, NULL);
pd                541 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			"Failed to create fence-MW for PD: %p\n", pd);
pd                547 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	bnxt_re_create_fence_wqe(pd);
pd                551 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	bnxt_re_destroy_fence_mr(pd);
pd                558 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd                559 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                561 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	bnxt_re_destroy_fence_mr(pd);
pd                563 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (pd->qplib_pd.id)
pd                565 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				      &pd->qplib_pd);
pd                574 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ibpd, struct bnxt_re_pd, ib_pd);
pd                577 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	pd->rdev = rdev;
pd                578 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	if (bnxt_qplib_alloc_pd(&rdev->qplib_res.pd_tbl, &pd->qplib_pd)) {
pd                599 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		resp.pdid = pd->qplib_pd.id;
pd                613 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		if (bnxt_re_create_fence_mr(pd))
pd                619 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			      &pd->qplib_pd);
pd                655 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct ib_pd *ib_pd = ib_ah->pd;
pd                656 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd                658 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                670 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	ah->qplib_ah.pd = &pd->qplib_pd;
pd                836 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_init_user_qp(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd,
pd                892 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				(struct bnxt_re_pd *pd,
pd                896 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                906 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	ah->qplib_ah.pd = &pd->qplib_pd;
pd                939 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				(struct bnxt_re_pd *pd,
pd                943 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd                956 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	qp->qplib_qp.pd = &pd->qplib_pd;
pd               1005 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               1006 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               1026 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	qp->qplib_qp.pd = &pd->qplib_pd;
pd               1127 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		rdev->qp1_sqp = bnxt_re_create_shadow_qp(pd, &rdev->qplib_res,
pd               1135 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		rdev->sqp_ah = bnxt_re_create_shadow_qp_ah(pd, &rdev->qplib_res,
pd               1166 drivers/infiniband/hw/bnxt_re/ib_verbs.c 			rc = bnxt_re_init_user_qp(rdev, pd, qp, udata);
pd               1309 drivers/infiniband/hw/bnxt_re/ib_verbs.c 				 struct bnxt_re_pd *pd,
pd               1343 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct ib_pd *ib_pd = ib_srq->pd;
pd               1344 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               1345 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               1364 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	srq->qplib_srq.pd = &pd->qplib_pd;
pd               1381 drivers/infiniband/hw/bnxt_re/ib_verbs.c 		rc = bnxt_re_init_user_srq(rdev, pd, srq, udata);
pd               3277 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               3278 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               3288 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mr->qplib_mr.pd = &pd->qplib_pd;
pd               3368 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               3369 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               3385 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mr->qplib_mr.pd = &pd->qplib_pd;
pd               3424 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               3425 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               3433 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mw->qplib_mw.pd = &pd->qplib_pd;
pd               3505 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
pd               3506 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	struct bnxt_re_dev *rdev = pd->rdev;
pd               3523 drivers/infiniband/hw/bnxt_re/ib_verbs.c 	mr->qplib_mr.pd = &pd->qplib_pd;
pd                164 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                165 drivers/infiniband/hw/bnxt_re/ib_verbs.h void bnxt_re_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                181 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_qp *bnxt_re_create_qp(struct ib_pd *pd,
pd                198 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_mr *bnxt_re_get_dma_mr(struct ib_pd *pd, int mr_access_flags);
pd                208 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_mr *bnxt_re_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                582 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	req.pd_id = cpu_to_le32(srq->pd->id);
pd                834 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	req.pd_id = cpu_to_le32(qp->pd->id);
pd               1058 drivers/infiniband/hw/bnxt_re/qplib_fp.c 	req.pd_id = cpu_to_le32(qp->pd->id);
pd                 43 drivers/infiniband/hw/bnxt_re/qplib_fp.h 	struct bnxt_qplib_pd		*pd;
pd                254 drivers/infiniband/hw/bnxt_re/qplib_fp.h 	struct bnxt_qplib_pd		*pd;
pd                575 drivers/infiniband/hw/bnxt_re/qplib_res.c int bnxt_qplib_alloc_pd(struct bnxt_qplib_pd_tbl *pdt, struct bnxt_qplib_pd *pd)
pd                585 drivers/infiniband/hw/bnxt_re/qplib_res.c 	pd->id = bit_num;
pd                591 drivers/infiniband/hw/bnxt_re/qplib_res.c 			  struct bnxt_qplib_pd *pd)
pd                593 drivers/infiniband/hw/bnxt_re/qplib_res.c 	if (test_and_set_bit(pd->id, pdt->tbl)) {
pd                595 drivers/infiniband/hw/bnxt_re/qplib_res.c 			 pd->id);
pd                598 drivers/infiniband/hw/bnxt_re/qplib_res.c 	pd->id = 0;
pd                245 drivers/infiniband/hw/bnxt_re/qplib_res.h 			struct bnxt_qplib_pd *pd);
pd                248 drivers/infiniband/hw/bnxt_re/qplib_res.h 			  struct bnxt_qplib_pd *pd);
pd                521 drivers/infiniband/hw/bnxt_re/qplib_sp.c 	req.pd_id = cpu_to_le32(ah->pd->id);
pd                604 drivers/infiniband/hw/bnxt_re/qplib_sp.c 	req.pd_id = cpu_to_le32(mrw->pd->id);
pd                 94 drivers/infiniband/hw/bnxt_re/qplib_sp.h 	struct bnxt_qplib_pd		*pd;
pd                109 drivers/infiniband/hw/bnxt_re/qplib_sp.h 	struct bnxt_qplib_pd		*pd;
pd                293 drivers/infiniband/hw/cxgb3/iwch_provider.c static void iwch_deallocate_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                298 drivers/infiniband/hw/cxgb3/iwch_provider.c 	php = to_iwch_pd(pd);
pd                300 drivers/infiniband/hw/cxgb3/iwch_provider.c 	pr_debug("%s ibpd %p pdid 0x%x\n", __func__, pd, php->pdid);
pd                304 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_allocate_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                306 drivers/infiniband/hw/cxgb3/iwch_provider.c 	struct iwch_pd *php = to_iwch_pd(pd);
pd                307 drivers/infiniband/hw/cxgb3/iwch_provider.c 	struct ib_device *ibdev = pd->device;
pd                355 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_get_dma_mr(struct ib_pd *pd, int acc)
pd                359 drivers/infiniband/hw/cxgb3/iwch_provider.c 	struct iwch_pd *php = to_iwch_pd(pd);
pd                365 drivers/infiniband/hw/cxgb3/iwch_provider.c 	pr_debug("%s ib_pd %p\n", __func__, pd);
pd                433 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                444 drivers/infiniband/hw/cxgb3/iwch_provider.c 	pr_debug("%s ib_pd %p\n", __func__, pd);
pd                446 drivers/infiniband/hw/cxgb3/iwch_provider.c 	php = to_iwch_pd(pd);
pd                531 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mw *iwch_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd                544 drivers/infiniband/hw/cxgb3/iwch_provider.c 	php = to_iwch_pd(pd);
pd                585 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                599 drivers/infiniband/hw/cxgb3/iwch_provider.c 	php = to_iwch_pd(pd);
pd                695 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_qp *iwch_create_qp(struct ib_pd *pd,
pd                708 drivers/infiniband/hw/cxgb3/iwch_provider.c 	pr_debug("%s ib_pd %p\n", __func__, pd);
pd                711 drivers/infiniband/hw/cxgb3/iwch_provider.c 	php = to_iwch_pd(pd);
pd                769 drivers/infiniband/hw/cxgb3/iwch_provider.c 	qhp->attr.pd = php->pdid;
pd                150 drivers/infiniband/hw/cxgb3/iwch_provider.h 	u32 pd;	/* IN */
pd                824 drivers/infiniband/hw/cxgb3/iwch_qp.c 	init_attr.pdid = qhp->attr.pd;
pd                476 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.pd.total, dev->rdev.stats.pd.cur,
pd                477 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.pd.max, dev->rdev.stats.pd.fail);
pd                524 drivers/infiniband/hw/cxgb4/device.c 	dev->rdev.stats.pd.max = 0;
pd                525 drivers/infiniband/hw/cxgb4/device.c 	dev->rdev.stats.pd.fail = 0;
pd                842 drivers/infiniband/hw/cxgb4/device.c 	rdev->stats.pd.total = T4_MAX_NUM_PD;
pd                130 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	struct c4iw_stat pd;
pd                470 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	u32 pd;
pd                982 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                988 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd                990 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start,
pd                993 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc);
pd               1006 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_qp *c4iw_create_qp(struct ib_pd *pd,
pd                447 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc)
pd                455 drivers/infiniband/hw/cxgb4/mem.c 	pr_debug("ib_pd %p\n", pd);
pd                456 drivers/infiniband/hw/cxgb4/mem.c 	php = to_c4iw_pd(pd);
pd                508 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                519 drivers/infiniband/hw/cxgb4/mem.c 	pr_debug("ib_pd %p\n", pd);
pd                527 drivers/infiniband/hw/cxgb4/mem.c 	php = to_c4iw_pd(pd);
pd                615 drivers/infiniband/hw/cxgb4/mem.c struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd                628 drivers/infiniband/hw/cxgb4/mem.c 	php = to_c4iw_pd(pd);
pd                693 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                704 drivers/infiniband/hw/cxgb4/mem.c 	php = to_c4iw_pd(pd);
pd                193 drivers/infiniband/hw/cxgb4/provider.c static void c4iw_deallocate_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                198 drivers/infiniband/hw/cxgb4/provider.c 	php = to_c4iw_pd(pd);
pd                200 drivers/infiniband/hw/cxgb4/provider.c 	pr_debug("ibpd %p pdid 0x%x\n", pd, php->pdid);
pd                203 drivers/infiniband/hw/cxgb4/provider.c 	rhp->rdev.stats.pd.cur--;
pd                207 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_allocate_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                209 drivers/infiniband/hw/cxgb4/provider.c 	struct c4iw_pd *php = to_c4iw_pd(pd);
pd                210 drivers/infiniband/hw/cxgb4/provider.c 	struct ib_device *ibdev = pd->device;
pd                231 drivers/infiniband/hw/cxgb4/provider.c 	rhp->rdev.stats.pd.cur++;
pd                232 drivers/infiniband/hw/cxgb4/provider.c 	if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max)
pd                233 drivers/infiniband/hw/cxgb4/provider.c 		rhp->rdev.stats.pd.max = rhp->rdev.stats.pd.cur;
pd               1799 drivers/infiniband/hw/cxgb4/qp.c 	wqe->u.init.pdid = cpu_to_be32(qhp->attr.pd);
pd               2111 drivers/infiniband/hw/cxgb4/qp.c struct ib_qp *c4iw_create_qp(struct ib_pd *pd, struct ib_qp_init_attr *attrs,
pd               2127 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("ib_pd %p\n", pd);
pd               2132 drivers/infiniband/hw/cxgb4/qp.c 	php = to_c4iw_pd(pd);
pd               2196 drivers/infiniband/hw/cxgb4/qp.c 	qhp->attr.pd = php->pdid;
pd               2672 drivers/infiniband/hw/cxgb4/qp.c 	struct ib_pd *pd = ib_srq->pd;
pd               2683 drivers/infiniband/hw/cxgb4/qp.c 	pr_debug("%s ib_pd %p\n", __func__, pd);
pd               2685 drivers/infiniband/hw/cxgb4/qp.c 	php = to_c4iw_pd(pd);
pd                105 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                287 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                310 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                329 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                701 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                711 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h 	u16 pd;
pd                 26 drivers/infiniband/hw/efa/efa_com_cmd.c 	create_qp_cmd.pd = params->pd;
pd                228 drivers/infiniband/hw/efa/efa_com_cmd.c 	mr_cmd.pd = params->pd;
pd                308 drivers/infiniband/hw/efa/efa_com_cmd.c 	ah_cmd.pd = params->pdn;
pd                337 drivers/infiniband/hw/efa/efa_com_cmd.c 	ah_cmd.pd = params->pdn;
pd                347 drivers/infiniband/hw/efa/efa_com_cmd.c 				      ah_cmd.ah, ah_cmd.pd, err);
pd                648 drivers/infiniband/hw/efa/efa_com_cmd.c 	result->pdn = resp.pd;
pd                662 drivers/infiniband/hw/efa/efa_com_cmd.c 	cmd.pd = params->pdn;
pd                672 drivers/infiniband/hw/efa/efa_com_cmd.c 				      cmd.pd, err);
pd                 27 drivers/infiniband/hw/efa/efa_com_cmd.h 	u16 pd;
pd                177 drivers/infiniband/hw/efa/efa_com_cmd.h 	u16 pd;
pd                433 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_pd *pd = to_epd(ibpd);
pd                448 drivers/infiniband/hw/efa/efa_verbs.c 	pd->pdn = result.pdn;
pd                461 drivers/infiniband/hw/efa/efa_verbs.c 	ibdev_dbg(&dev->ibdev, "Allocated pd[%d]\n", pd->pdn);
pd                475 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_pd *pd = to_epd(ibpd);
pd                477 drivers/infiniband/hw/efa/efa_verbs.c 	ibdev_dbg(&dev->ibdev, "Dealloc pd[%d]\n", pd->pdn);
pd                478 drivers/infiniband/hw/efa/efa_verbs.c 	efa_pd_dealloc(dev, pd->pdn);
pd                490 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_dev *dev = to_edev(ibqp->pd->device);
pd                691 drivers/infiniband/hw/efa/efa_verbs.c 	create_qp_params.pd = to_epd(ibpd)->pdn;
pd               1434 drivers/infiniband/hw/efa/efa_verbs.c 	params.pd = to_epd(ibpd)->pdn;
pd               1675 drivers/infiniband/hw/efa/efa_verbs.c 		.pdn = to_epd(ah->ibah.pd)->pdn,
pd               1709 drivers/infiniband/hw/efa/efa_verbs.c 	params.pdn = to_epd(ibah->pd)->pdn;
pd               1741 drivers/infiniband/hw/efa/efa_verbs.c 	struct efa_dev *dev = to_edev(ibah->pd->device);
pd                308 drivers/infiniband/hw/hfi1/mad.c 		ah = rdma_create_ah(qp0->ibqp.pd, &attr, 0);
pd               1782 drivers/infiniband/hw/hfi1/mad.c 	u8 *pd = data;
pd               1789 drivers/infiniband/hw/hfi1/mad.c 			if ((pd[i] & 0x1f) == 0xf)
pd               1790 drivers/infiniband/hw/hfi1/mad.c 				pd[i] = ILLEGAL_VL;
pd               1792 drivers/infiniband/hw/hfi1/mad.c 			if ((pd[i] & 0x1f) == ILLEGAL_VL)
pd               1793 drivers/infiniband/hw/hfi1/mad.c 				pd[i] = 0xf;
pd                390 drivers/infiniband/hw/hns/hns_roce_device.h 	u32			pd;   /* PD num of MR */
pd               1185 drivers/infiniband/hw/hns/hns_roce_device.h int hns_roce_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd               1186 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd               1188 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_get_dma_mr(struct ib_pd *pd, int acc);
pd               1189 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd               1193 drivers/infiniband/hw/hns/hns_roce_device.h 			   u64 virt_addr, int mr_access_flags, struct ib_pd *pd,
pd               1195 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1205 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mw *hns_roce_alloc_mw(struct ib_pd *pd, enum ib_mw_type,
pd                677 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 						    struct ib_pd *pd)
pd                689 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	qp = hns_roce_create_qp(pd, &init_attr, NULL);
pd                709 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	struct ib_pd *pd;
pd                745 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	pd = rdma_zalloc_drv_obj(ibdev, ib_pd);
pd                746 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	if (!pd) {
pd                751 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	pd->device  = ibdev;
pd                752 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	ret = hns_roce_alloc_pd(pd, NULL);
pd                756 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	free_mr->mr_free_pd = to_hr_pd(pd);
pd                796 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		free_mr->mr_free_qp[i] = hns_roce_v1_create_lp_qp(hr_dev, pd);
pd                810 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		hr_qp->ibqp.pd		= pd;
pd                862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	hns_roce_dealloc_pd(pd, NULL);
pd                865 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 	kfree(pd);
pd               1875 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 		       MPT_BYTE_28_PD_S, mr->pd);
pd               2605 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       QP1C_BYTES_4_PD_S, to_hr_pd(ibqp->pd)->pdn);
pd               2791 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       to_hr_pd(ibqp->pd)->pdn);
pd               2857 drivers/infiniband/hw/hns/hns_roce_hw_v1.c 			       to_hr_pd(ibqp->pd)->pdn);
pd                375 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 				       to_hr_pd(ibqp->pd)->pdn);
pd               2287 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_MPT_BYTE_4_PD_S, mr->pd);
pd               2336 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		mr->pd = pdn;
pd               2385 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_MPT_BYTE_4_PD_S, mr->pd);
pd               3282 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_QPC_BYTE_16_PD_S, to_hr_pd(ibqp->pd)->pdn);
pd               3602 drivers/infiniband/hw/hns/hns_roce_hw_v2.c 		       V2_QPC_BYTE_16_PD_S, to_hr_pd(ibqp->pd)->pdn);
pd                571 drivers/infiniband/hw/hns/hns_roce_mr.c static int hns_roce_mr_alloc(struct hns_roce_dev *hr_dev, u32 pd, u64 iova,
pd                586 drivers/infiniband/hw/hns/hns_roce_mr.c 	mr->pd = pd;				/* MR num */
pd                985 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_get_dma_mr(struct ib_pd *pd, int acc)
pd                997 drivers/infiniband/hw/hns/hns_roce_mr.c 	ret = hns_roce_mr_alloc(to_hr_dev(pd->device), to_hr_pd(pd)->pdn, 0,
pd               1002 drivers/infiniband/hw/hns/hns_roce_mr.c 	ret = hns_roce_mr_enable(to_hr_dev(pd->device), mr);
pd               1012 drivers/infiniband/hw/hns/hns_roce_mr.c 	hns_roce_mr_free(to_hr_dev(pd->device), mr);
pd               1130 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd               1134 drivers/infiniband/hw/hns/hns_roce_mr.c 	struct hns_roce_dev *hr_dev = to_hr_dev(pd->device);
pd               1180 drivers/infiniband/hw/hns/hns_roce_mr.c 	ret = hns_roce_mr_alloc(hr_dev, to_hr_pd(pd)->pdn, virt_addr, length,
pd               1286 drivers/infiniband/hw/hns/hns_roce_mr.c 			   u64 virt_addr, int mr_access_flags, struct ib_pd *pd,
pd               1318 drivers/infiniband/hw/hns/hns_roce_mr.c 		pdn = to_hr_pd(pd)->pdn;
pd               1374 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1377 drivers/infiniband/hw/hns/hns_roce_mr.c 	struct hns_roce_dev *hr_dev = to_hr_dev(pd->device);
pd               1403 drivers/infiniband/hw/hns/hns_roce_mr.c 	ret = hns_roce_mr_alloc(hr_dev, to_hr_pd(pd)->pdn, 0, length,
pd               1418 drivers/infiniband/hw/hns/hns_roce_mr.c 	hns_roce_mr_free(to_hr_dev(pd->device), mr);
pd                 65 drivers/infiniband/hw/hns/hns_roce_pd.c 	struct hns_roce_pd *pd = to_hr_pd(ibpd);
pd                 68 drivers/infiniband/hw/hns/hns_roce_pd.c 	ret = hns_roce_pd_alloc(to_hr_dev(ib_dev), &pd->pdn);
pd                 75 drivers/infiniband/hw/hns/hns_roce_pd.c 		struct hns_roce_ib_alloc_pd_resp uresp = {.pdn = pd->pdn};
pd                 78 drivers/infiniband/hw/hns/hns_roce_pd.c 			hns_roce_pd_free(to_hr_dev(ib_dev), pd->pdn);
pd                 87 drivers/infiniband/hw/hns/hns_roce_pd.c void hns_roce_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                 89 drivers/infiniband/hw/hns/hns_roce_pd.c 	hns_roce_pd_free(to_hr_dev(pd->device), to_hr_pd(pd)->pdn);
pd               1016 drivers/infiniband/hw/hns/hns_roce_qp.c struct ib_qp *hns_roce_create_qp(struct ib_pd *pd,
pd               1020 drivers/infiniband/hw/hns/hns_roce_qp.c 	struct hns_roce_dev *hr_dev = to_hr_dev(pd->device);
pd               1032 drivers/infiniband/hw/hns/hns_roce_qp.c 		ret = hns_roce_create_qp_common(hr_dev, pd, init_attr, udata, 0,
pd               1067 drivers/infiniband/hw/hns/hns_roce_qp.c 		ret = hns_roce_create_qp_common(hr_dev, pd, init_attr, udata,
pd                252 drivers/infiniband/hw/hns/hns_roce_srq.c static int hns_roce_create_idx_que(struct ib_pd *pd, struct hns_roce_srq *srq,
pd                255 drivers/infiniband/hw/hns/hns_roce_srq.c 	struct hns_roce_dev *hr_dev = to_hr_dev(pd->device);
pd                296 drivers/infiniband/hw/hns/hns_roce_srq.c 	ret = hns_roce_create_idx_que(srq->ibsrq.pd, srq, page_shift);
pd                413 drivers/infiniband/hw/hns/hns_roce_srq.c 	ret = hns_roce_srq_alloc(hr_dev, to_hr_pd(ib_srq->pd)->pdn, cqn, 0,
pd               3720 drivers/infiniband/hw/i40iw/i40iw_cm.c 		ibmr->pd = &iwpd->ibpd;
pd                418 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 			     struct i40iw_sc_pd *pd,
pd                422 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	pd->size = sizeof(*pd);
pd                423 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	pd->pd_id = pd_id;
pd                424 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	pd->abi_ver = abi_ver;
pd                425 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	pd->dev = dev;
pd               2364 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	qp->dev = info->pd->dev;
pd               2373 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	qp->pd = info->pd;
pd               2375 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	offset = (qp->pd->dev->is_pf) ? I40E_PFPE_WQEALLOC : I40E_VFPE_WQEALLOC1;
pd               2376 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	if (i40iw_get_hw_addr(qp->pd->dev))
pd               2377 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 		wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) +
pd               2381 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	info->qp_uk_init_info.abi_ver = qp->pd->abi_ver;
pd               2387 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	pble_obj_cnt = info->pd->dev->hmc_info->hmc_obj[I40IW_HMC_IW_PBLE].cnt;
pd               2401 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	switch (qp->pd->abi_ver) {
pd               2452 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	cqp = qp->pd->dev->cqp;
pd               2500 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	cqp = qp->pd->dev->cqp;
pd               2567 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	cqp = qp->pd->dev->cqp;
pd               2620 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	cqp = qp->pd->dev->cqp;
pd               2677 drivers/infiniband/hw/i40iw/i40iw_ctrl.c 	cqp = qp->pd->dev->cqp;
pd                577 drivers/infiniband/hw/i40iw/i40iw_puda.c 	qp->pd = &rsrc->sc_pd;
pd                603 drivers/infiniband/hw/i40iw/i40iw_puda.c 	if (qp->pd->dev->is_pf)
pd                604 drivers/infiniband/hw/i40iw/i40iw_puda.c 		ukqp->wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) +
pd                607 drivers/infiniband/hw/i40iw/i40iw_puda.c 		ukqp->wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) +
pd                376 drivers/infiniband/hw/i40iw/i40iw_type.h 	struct i40iw_sc_pd *pd;
pd                844 drivers/infiniband/hw/i40iw/i40iw_type.h 	struct i40iw_sc_pd *pd;
pd                285 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                287 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(pd);
pd                288 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_device *iwdev = to_iwdev(pd->device);
pd                617 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	init_info.pd = &iwpd->sc_pd;
pd               1512 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(iwmr->ibmr.pd);
pd               1550 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1553 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(pd);
pd               1554 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_device *iwdev = to_iwdev(pd->device);
pd               1575 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.pd = pd;
pd               1576 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.device = pd->device;
pd               1678 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(iwmr->ibmr.pd);
pd               1735 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_reg_user_mr(struct ib_pd *pd,
pd               1742 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(pd);
pd               1743 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_device *iwdev = to_iwdev(pd->device);
pd               1784 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.pd = pd;
pd               1785 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.device = pd->device;
pd               1885 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_mr *i40iw_reg_phys_mr(struct ib_pd *pd,
pd               1891 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_pd *iwpd = to_iwpd(pd);
pd               1892 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct i40iw_device *iwdev = to_iwdev(pd->device);
pd               1903 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.pd = pd;
pd               1904 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	iwmr->ibmr.device = pd->device;
pd               1940 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_get_dma_mr(struct ib_pd *pd, int acc)
pd               1944 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	return i40iw_reg_phys_mr(pd, 0, 0, acc, &kva);
pd               1986 drivers/infiniband/hw/i40iw/i40iw_verbs.c 	struct ib_pd *ibpd = ib_mr->pd;
pd                 48 drivers/infiniband/hw/mlx4/ah.c 	ah->av.ib.port_pd = cpu_to_be32(to_mpd(ib_ah->pd)->pdn |
pd                118 drivers/infiniband/hw/mlx4/ah.c 	ah->av.eth.port_pd = cpu_to_be32(to_mpd(ib_ah->pd)->pdn |
pd                204 drivers/infiniband/hw/mlx4/mad.c 	new_ah = rdma_create_ah(dev->send_agent[port_num - 1][0]->qp->pd,
pd                570 drivers/infiniband/hw/mlx4/mad.c 	ah = rdma_create_ah(tun_ctx->pd, &attr, 0);
pd                638 drivers/infiniband/hw/mlx4/mad.c 	list.lkey = tun_ctx->pd->local_dma_lkey;
pd               1324 drivers/infiniband/hw/mlx4/mad.c 	sg_list.lkey = ctx->pd->local_dma_lkey;
pd               1396 drivers/infiniband/hw/mlx4/mad.c 	ah = rdma_zalloc_drv_obj(sqp_ctx->pd->device, ib_ah);
pd               1400 drivers/infiniband/hw/mlx4/mad.c 	ah->device = sqp_ctx->pd->device;
pd               1401 drivers/infiniband/hw/mlx4/mad.c 	ah->pd = sqp_ctx->pd;
pd               1437 drivers/infiniband/hw/mlx4/mad.c 	list.lkey = sqp_ctx->pd->local_dma_lkey;
pd               1834 drivers/infiniband/hw/mlx4/mad.c 	tun_qp->qp = ib_create_qp(ctx->pd, &qp_init_attr.init_attr);
pd               2020 drivers/infiniband/hw/mlx4/mad.c 	ctx->pd = ib_alloc_pd(ctx->ib_dev, 0);
pd               2021 drivers/infiniband/hw/mlx4/mad.c 	if (IS_ERR(ctx->pd)) {
pd               2022 drivers/infiniband/hw/mlx4/mad.c 		ret = PTR_ERR(ctx->pd);
pd               2070 drivers/infiniband/hw/mlx4/mad.c 	ib_dealloc_pd(ctx->pd);
pd               2071 drivers/infiniband/hw/mlx4/mad.c 	ctx->pd = NULL;
pd               2105 drivers/infiniband/hw/mlx4/mad.c 		ib_dealloc_pd(ctx->pd);
pd               2106 drivers/infiniband/hw/mlx4/mad.c 		ctx->pd = NULL;
pd               2237 drivers/infiniband/hw/mlx4/mad.c 		ib_dealloc_pd(sqp_ctx->pd);
pd               2238 drivers/infiniband/hw/mlx4/mad.c 		sqp_ctx->pd = NULL;
pd               1196 drivers/infiniband/hw/mlx4/main.c 	struct mlx4_ib_pd *pd = to_mpd(ibpd);
pd               1200 drivers/infiniband/hw/mlx4/main.c 	err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn);
pd               1204 drivers/infiniband/hw/mlx4/main.c 	if (udata && ib_copy_to_udata(udata, &pd->pdn, sizeof(__u32))) {
pd               1205 drivers/infiniband/hw/mlx4/main.c 		mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn);
pd               1211 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd               1213 drivers/infiniband/hw/mlx4/main.c 	mlx4_pd_free(to_mdev(pd->device)->dev, to_mpd(pd)->pdn);
pd               1234 drivers/infiniband/hw/mlx4/main.c 	xrcd->pd = ib_alloc_pd(ibdev, 0);
pd               1235 drivers/infiniband/hw/mlx4/main.c 	if (IS_ERR(xrcd->pd)) {
pd               1236 drivers/infiniband/hw/mlx4/main.c 		err = PTR_ERR(xrcd->pd);
pd               1250 drivers/infiniband/hw/mlx4/main.c 	ib_dealloc_pd(xrcd->pd);
pd               1261 drivers/infiniband/hw/mlx4/main.c 	ib_dealloc_pd(to_mxrcd(xrcd)->pd);
pd                100 drivers/infiniband/hw/mlx4/mlx4_ib.h 	struct ib_pd	       *pd;
pd                459 drivers/infiniband/hw/mlx4/mlx4_ib.h 	struct ib_pd *pd;
pd                730 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc);
pd                733 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                737 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd                740 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                771 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd,
pd                797 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int mr_access_flags,
pd                896 drivers/infiniband/hw/mlx4/mlx4_ib.h 			  int mr_access_flags, struct ib_pd *pd,
pd                906 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_wq *mlx4_ib_create_wq(struct ib_pd *pd,
pd                 58 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc)
pd                 67 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0,
pd                 72 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr);
pd                 82 drivers/infiniband/hw/mlx4/mr.c 	(void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
pd                404 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                408 drivers/infiniband/hw/mlx4/mr.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd                427 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, virt_addr, length,
pd                448 drivers/infiniband/hw/mlx4/mr.c 	(void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr);
pd                461 drivers/infiniband/hw/mlx4/mr.c 			  int mr_access_flags, struct ib_pd *pd,
pd                481 drivers/infiniband/hw/mlx4/mr.c 					   to_mpd(pd)->pdn);
pd                614 drivers/infiniband/hw/mlx4/mr.c struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd                617 drivers/infiniband/hw/mlx4/mr.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd                625 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_mw_alloc(dev->dev, to_mpd(pd)->pdn,
pd                657 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                660 drivers/infiniband/hw/mlx4/mr.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd                672 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, 0, 0, 0,
pd                677 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_alloc_priv_pages(pd->device, mr, max_num_sg);
pd                692 drivers/infiniband/hw/mlx4/mr.c 	mr->ibmr.device = pd->device;
pd                701 drivers/infiniband/hw/mlx4/mr.c struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int acc,
pd                704 drivers/infiniband/hw/mlx4/mr.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd                712 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_fmr_alloc(dev->dev, to_mpd(pd)->pdn, convert_access(acc),
pd                718 drivers/infiniband/hw/mlx4/mr.c 	err = mlx4_fmr_enable(to_mdev(pd->device)->dev, &fmr->mfmr);
pd                727 drivers/infiniband/hw/mlx4/mr.c 	(void) mlx4_mr_free(to_mdev(pd->device)->dev, &fmr->mfmr.mr);
pd                699 drivers/infiniband/hw/mlx4/qp.c static struct ib_qp *_mlx4_ib_create_qp_rss(struct ib_pd *pd,
pd                764 drivers/infiniband/hw/mlx4/qp.c 	err = create_qp_rss(to_mdev(pd->device), init_attr, &ucmd, qp);
pd                858 drivers/infiniband/hw/mlx4/qp.c static int create_rq(struct ib_pd *pd, struct ib_qp_init_attr *init_attr,
pd                861 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd                990 drivers/infiniband/hw/mlx4/qp.c static int create_qp_common(struct ib_pd *pd, struct ib_qp_init_attr *init_attr,
pd                994 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_dev *dev = to_mdev(pd->device);
pd               1197 drivers/infiniband/hw/mlx4/qp.c 			if (alloc_proxy_bufs(pd->device, qp)) {
pd               1273 drivers/infiniband/hw/mlx4/qp.c 		free_proxy_bufs(pd->device, qp);
pd               1360 drivers/infiniband/hw/mlx4/qp.c 		return to_mpd(to_mxrcd(qp->ibqp.xrcd)->pd);
pd               1362 drivers/infiniband/hw/mlx4/qp.c 		return to_mpd(qp->ibqp.pd);
pd               1532 drivers/infiniband/hw/mlx4/qp.c static struct ib_qp *_mlx4_ib_create_qp(struct ib_pd *pd,
pd               1542 drivers/infiniband/hw/mlx4/qp.c 		return _mlx4_ib_create_qp_rss(pd, init_attr, udata);
pd               1578 drivers/infiniband/hw/mlx4/qp.c 		pd = to_mxrcd(init_attr->xrcd)->pd;
pd               1583 drivers/infiniband/hw/mlx4/qp.c 		if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC))
pd               1598 drivers/infiniband/hw/mlx4/qp.c 		err = create_qp_common(pd, init_attr, udata, 0, &qp);
pd               1618 drivers/infiniband/hw/mlx4/qp.c 			int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev,
pd               1625 drivers/infiniband/hw/mlx4/qp.c 			sqpn = get_sqp_num(to_mdev(pd->device), init_attr);
pd               1628 drivers/infiniband/hw/mlx4/qp.c 		err = create_qp_common(pd, init_attr, udata, sqpn, &qp);
pd               1645 drivers/infiniband/hw/mlx4/qp.c struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd,
pd               1648 drivers/infiniband/hw/mlx4/qp.c 	struct ib_device *device = pd ? pd->device : init_attr->xrcd->device;
pd               1652 drivers/infiniband/hw/mlx4/qp.c 	ibqp = _mlx4_ib_create_qp(pd, init_attr, udata);
pd               1663 drivers/infiniband/hw/mlx4/qp.c 			sqp->roce_v2_gsi = ib_create_qp(pd, init_attr);
pd               2143 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_ib_pd *pd;
pd               2163 drivers/infiniband/hw/mlx4/qp.c 		pd	    = to_mpd(ibwq->pd);
pd               2173 drivers/infiniband/hw/mlx4/qp.c 		pd	    = get_pd(qp);
pd               2375 drivers/infiniband/hw/mlx4/qp.c 	context->pd = cpu_to_be32(pd->pdn);
pd               4129 drivers/infiniband/hw/mlx4/qp.c struct ib_wq *mlx4_ib_create_wq(struct ib_pd *pd,
pd               4133 drivers/infiniband/hw/mlx4/qp.c 	struct mlx4_dev *dev = to_mdev(pd->device)->dev;
pd               4188 drivers/infiniband/hw/mlx4/qp.c 	err = create_rq(pd, &ib_qp_init_attr, udata, qp);
pd                179 drivers/infiniband/hw/mlx4/srq.c 	err = mlx4_srq_alloc(dev->dev, to_mpd(ib_srq->pd)->pdn, cqn, xrcdn,
pd                272 drivers/infiniband/hw/mlx5/cmd.c 	MLX5_SET(dealloc_pd_in, in, pd, pdn);
pd               1265 drivers/infiniband/hw/mlx5/devx.c 	mkey->pd = MLX5_GET(mkc, mkc, pd);
pd                115 drivers/infiniband/hw/mlx5/gsi.c struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd,
pd                118 drivers/infiniband/hw/mlx5/gsi.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                122 drivers/infiniband/hw/mlx5/gsi.c 	const int num_pkeys = pd->device->attrs.max_pkeys;
pd                171 drivers/infiniband/hw/mlx5/gsi.c 	gsi->cq = ib_alloc_cq(pd->device, gsi, init_attr->cap.max_send_wr, 0,
pd                187 drivers/infiniband/hw/mlx5/gsi.c 	gsi->rx_qp = ib_create_qp(pd, &hw_init_attr);
pd                253 drivers/infiniband/hw/mlx5/gsi.c 	struct ib_pd *pd = gsi->rx_qp->pd;
pd                269 drivers/infiniband/hw/mlx5/gsi.c 	return ib_create_qp(pd, &init_attr);
pd               2469 drivers/infiniband/hw/mlx5/main.c 	struct mlx5_ib_pd *pd = to_mpd(ibpd);
pd               2487 drivers/infiniband/hw/mlx5/main.c 	pd->pdn = MLX5_GET(alloc_pd_out, out, pd);
pd               2488 drivers/infiniband/hw/mlx5/main.c 	pd->uid = uid;
pd               2490 drivers/infiniband/hw/mlx5/main.c 		resp.pdn = pd->pdn;
pd               2492 drivers/infiniband/hw/mlx5/main.c 			mlx5_cmd_dealloc_pd(to_mdev(ibdev)->mdev, pd->pdn, uid);
pd               2500 drivers/infiniband/hw/mlx5/main.c static void mlx5_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd               2502 drivers/infiniband/hw/mlx5/main.c 	struct mlx5_ib_dev *mdev = to_mdev(pd->device);
pd               2503 drivers/infiniband/hw/mlx5/main.c 	struct mlx5_ib_pd *mpd = to_mpd(pd);
pd               4347 drivers/infiniband/hw/mlx5/main.c 	uid = ibqp->pd ?
pd               4348 drivers/infiniband/hw/mlx5/main.c 		to_mpd(ibqp->pd)->uid : 0;
pd               4369 drivers/infiniband/hw/mlx5/main.c 	uid = ibqp->pd ?
pd               4370 drivers/infiniband/hw/mlx5/main.c 		to_mpd(ibqp->pd)->uid : 0;
pd               4796 drivers/infiniband/hw/mlx5/main.c 	if (dev->umrc.pd)
pd               4797 drivers/infiniband/hw/mlx5/main.c 		ib_dealloc_pd(dev->umrc.pd);
pd               4808 drivers/infiniband/hw/mlx5/main.c 	struct ib_pd *pd;
pd               4820 drivers/infiniband/hw/mlx5/main.c 	pd = ib_alloc_pd(&dev->ib_dev, 0);
pd               4821 drivers/infiniband/hw/mlx5/main.c 	if (IS_ERR(pd)) {
pd               4823 drivers/infiniband/hw/mlx5/main.c 		ret = PTR_ERR(pd);
pd               4841 drivers/infiniband/hw/mlx5/main.c 	qp = mlx5_ib_create_qp(pd, init_attr, NULL);
pd               4883 drivers/infiniband/hw/mlx5/main.c 	dev->umrc.pd = pd;
pd               4906 drivers/infiniband/hw/mlx5/main.c 	ib_dealloc_pd(pd);
pd               4907 drivers/infiniband/hw/mlx5/main.c 	dev->umrc.pd = NULL;
pd               5002 drivers/infiniband/hw/mlx5/main.c 	devr->s0->pd		= devr->p0;
pd               5026 drivers/infiniband/hw/mlx5/main.c 	devr->s1->pd		= devr->p0;
pd                480 drivers/infiniband/hw/mlx5/mlx5_ib.h 	struct ib_pd		       *pd;
pd                652 drivers/infiniband/hw/mlx5/mlx5_ib.h 	struct ib_pd	*pd;
pd               1120 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd,
pd               1147 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc);
pd               1148 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd               1151 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_advise_mr(struct ib_pd *pd,
pd               1157 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd               1162 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_mr *mlx5_ib_alloc_implicit_mr(struct mlx5_ib_pd *pd,
pd               1168 drivers/infiniband/hw/mlx5/mlx5_ib.h 			  struct ib_pd *pd, struct ib_udata *udata);
pd               1170 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1172 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_alloc_mr_integrity(struct ib_pd *pd,
pd               1229 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd,
pd               1245 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_reg_dm_mr(struct ib_pd *pd, struct ib_dm *dm,
pd               1261 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_advise_mr_prefetch(struct ib_pd *pd,
pd               1280 drivers/infiniband/hw/mlx5/mlx5_ib.h mlx5_ib_advise_mr_prefetch(struct ib_pd *pd,
pd               1314 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd,
pd                682 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc)
pd                684 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                712 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
pd                850 drivers/infiniband/hw/mlx5/mr.c 				  struct ib_pd *pd, struct ib_umem *umem,
pd                854 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                874 drivers/infiniband/hw/mlx5/mr.c 	mr->ibmr.pd = pd;
pd                880 drivers/infiniband/hw/mlx5/mr.c 	mr->mmkey.pd = to_mpd(pd)->pdn;
pd                985 drivers/infiniband/hw/mlx5/mr.c 	sg.lkey = dev->umrc.pd->local_dma_lkey;
pd                995 drivers/infiniband/hw/mlx5/mr.c 	wr.pd = mr->ibmr.pd;
pd               1050 drivers/infiniband/hw/mlx5/mr.c static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd,
pd               1056 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1069 drivers/infiniband/hw/mlx5/mr.c 	mr->ibmr.pd = pd;
pd               1101 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
pd               1147 drivers/infiniband/hw/mlx5/mr.c static struct ib_mr *mlx5_ib_get_dm_mr(struct ib_pd *pd, u64 start_addr,
pd               1150 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1179 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
pd               1203 drivers/infiniband/hw/mlx5/mr.c int mlx5_ib_advise_mr(struct ib_pd *pd,
pd               1214 drivers/infiniband/hw/mlx5/mr.c 	return mlx5_ib_advise_mr_prefetch(pd, advice, flags,
pd               1218 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_reg_dm_mr(struct ib_pd *pd, struct ib_dm *dm,
pd               1246 drivers/infiniband/hw/mlx5/mr.c 	return mlx5_ib_get_dm_mr(pd, start_addr, attr->length,
pd               1250 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd               1254 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1276 drivers/infiniband/hw/mlx5/mr.c 		mr = mlx5_ib_alloc_implicit_mr(to_mpd(pd), udata, access_flags);
pd               1291 drivers/infiniband/hw/mlx5/mr.c 		mr = alloc_mr_from_cache(pd, umem, virt_addr, length, ncont,
pd               1308 drivers/infiniband/hw/mlx5/mr.c 		mr = reg_create(NULL, pd, virt_addr, length, umem, ncont,
pd               1362 drivers/infiniband/hw/mlx5/mr.c 	umrwr.pd = dev->umrc.pd;
pd               1369 drivers/infiniband/hw/mlx5/mr.c static int rereg_umr(struct ib_pd *pd, struct mlx5_ib_mr *mr,
pd               1372 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1382 drivers/infiniband/hw/mlx5/mr.c 		umrwr.pd = pd;
pd               1398 drivers/infiniband/hw/mlx5/mr.c 	struct ib_pd *pd = (flags & IB_MR_REREG_PD) ? new_pd : ib_mr->pd;
pd               1456 drivers/infiniband/hw/mlx5/mr.c 		mr = reg_create(ib_mr, pd, addr, len, mr->umem, ncont,
pd               1470 drivers/infiniband/hw/mlx5/mr.c 		mr->ibmr.pd = pd;
pd               1474 drivers/infiniband/hw/mlx5/mr.c 		mr->mmkey.pd = to_mpd(pd)->pdn;
pd               1485 drivers/infiniband/hw/mlx5/mr.c 			err = rereg_umr(pd, mr, access_flags, flags);
pd               1642 drivers/infiniband/hw/mlx5/mr.c static void mlx5_set_umr_free_mkey(struct ib_pd *pd, u32 *in, int ndescs,
pd               1651 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
pd               1659 drivers/infiniband/hw/mlx5/mr.c static int _mlx5_alloc_mkey_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr,
pd               1663 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1670 drivers/infiniband/hw/mlx5/mr.c 	err = mlx5_alloc_priv_descs(pd->device, mr, ndescs, desc_size);
pd               1674 drivers/infiniband/hw/mlx5/mr.c 	mlx5_set_umr_free_mkey(pd, in, ndescs, access_mode, page_shift);
pd               1691 drivers/infiniband/hw/mlx5/mr.c static struct mlx5_ib_mr *mlx5_ib_alloc_pi_mr(struct ib_pd *pd,
pd               1706 drivers/infiniband/hw/mlx5/mr.c 	mr->ibmr.pd = pd;
pd               1707 drivers/infiniband/hw/mlx5/mr.c 	mr->ibmr.device = pd->device;
pd               1718 drivers/infiniband/hw/mlx5/mr.c 	err = _mlx5_alloc_mkey_descs(pd, mr, ndescs, desc_size, page_shift,
pd               1735 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_mem_reg_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr,
pd               1738 drivers/infiniband/hw/mlx5/mr.c 	return _mlx5_alloc_mkey_descs(pd, mr, ndescs, sizeof(struct mlx5_mtt),
pd               1743 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_sg_gaps_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr,
pd               1746 drivers/infiniband/hw/mlx5/mr.c 	return _mlx5_alloc_mkey_descs(pd, mr, ndescs, sizeof(struct mlx5_klm),
pd               1750 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_integrity_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr,
pd               1754 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1764 drivers/infiniband/hw/mlx5/mr.c 	err = mlx5_core_create_psv(dev->mdev, to_mpd(pd)->pdn, 2, psv_index);
pd               1775 drivers/infiniband/hw/mlx5/mr.c 	mr->klm_mr = mlx5_ib_alloc_pi_mr(pd, max_num_sg, max_num_meta_sg,
pd               1782 drivers/infiniband/hw/mlx5/mr.c 	mr->mtt_mr = mlx5_ib_alloc_pi_mr(pd, max_num_sg, max_num_meta_sg,
pd               1795 drivers/infiniband/hw/mlx5/mr.c 	err = _mlx5_alloc_mkey_descs(pd, mr, 4, sizeof(struct mlx5_klm), 0,
pd               1821 drivers/infiniband/hw/mlx5/mr.c static struct ib_mr *__mlx5_ib_alloc_mr(struct ib_pd *pd,
pd               1825 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1842 drivers/infiniband/hw/mlx5/mr.c 	mr->ibmr.device = pd->device;
pd               1847 drivers/infiniband/hw/mlx5/mr.c 		err = mlx5_alloc_mem_reg_descs(pd, mr, ndescs, in, inlen);
pd               1850 drivers/infiniband/hw/mlx5/mr.c 		err = mlx5_alloc_sg_gaps_descs(pd, mr, ndescs, in, inlen);
pd               1853 drivers/infiniband/hw/mlx5/mr.c 		err = mlx5_alloc_integrity_descs(pd, mr, max_num_sg,
pd               1875 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1878 drivers/infiniband/hw/mlx5/mr.c 	return __mlx5_ib_alloc_mr(pd, mr_type, max_num_sg, 0);
pd               1881 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_alloc_mr_integrity(struct ib_pd *pd,
pd               1884 drivers/infiniband/hw/mlx5/mr.c 	return __mlx5_ib_alloc_mr(pd, IB_MR_TYPE_INTEGRITY, max_num_sg,
pd               1888 drivers/infiniband/hw/mlx5/mr.c struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type,
pd               1891 drivers/infiniband/hw/mlx5/mr.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1929 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn);
pd               2073 drivers/infiniband/hw/mlx5/mr.c 	u32 lkey = mr->ibmr.pd->local_dma_lkey;
pd                166 drivers/infiniband/hw/mlx5/odp.c 	struct ib_pd *pd = mr->ibmr.pd;
pd                167 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                267 drivers/infiniband/hw/mlx5/odp.c 	if (!mr || !mr->ibmr.pd)
pd                419 drivers/infiniband/hw/mlx5/odp.c static struct mlx5_ib_mr *implicit_mr_alloc(struct ib_pd *pd,
pd                423 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                433 drivers/infiniband/hw/mlx5/odp.c 	mr->ibmr.pd = pd;
pd                478 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(mr->ibmr.pd->device);
pd                502 drivers/infiniband/hw/mlx5/odp.c 		mtt = implicit_mr_alloc(mr->ibmr.pd, odp, 0,
pd                549 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_mr *mlx5_ib_alloc_implicit_mr(struct mlx5_ib_pd *pd,
pd                560 drivers/infiniband/hw/mlx5/odp.c 	imr = implicit_mr_alloc(&pd->ibpd, umem_odp, 1, access_flags);
pd                775 drivers/infiniband/hw/mlx5/odp.c 					 struct ib_pd *pd, u32 key,
pd                812 drivers/infiniband/hw/mlx5/odp.c 		if (!smp_load_acquire(&mr->live) || !mr->ibmr.pd) {
pd                820 drivers/infiniband/hw/mlx5/odp.c 			    mr->ibmr.pd != pd) {
pd               1627 drivers/infiniband/hw/mlx5/odp.c 	struct ib_pd *pd;
pd               1655 drivers/infiniband/hw/mlx5/odp.c static bool num_pending_prefetch_inc(struct ib_pd *pd,
pd               1658 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1685 drivers/infiniband/hw/mlx5/odp.c 		if (mr->ibmr.pd != pd) {
pd               1699 drivers/infiniband/hw/mlx5/odp.c static int mlx5_ib_prefetch_sg_list(struct ib_pd *pd, u32 pf_flags,
pd               1704 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1710 drivers/infiniband/hw/mlx5/odp.c 		ret = pagefault_single_data_segment(dev, pd, sg->lkey, sg->addr,
pd               1726 drivers/infiniband/hw/mlx5/odp.c 	if (ib_device_try_get(w->pd->device)) {
pd               1727 drivers/infiniband/hw/mlx5/odp.c 		mlx5_ib_prefetch_sg_list(w->pd, w->pf_flags, w->sg_list,
pd               1729 drivers/infiniband/hw/mlx5/odp.c 		ib_device_put(w->pd->device);
pd               1732 drivers/infiniband/hw/mlx5/odp.c 	num_pending_prefetch_dec(to_mdev(w->pd->device), w->sg_list,
pd               1737 drivers/infiniband/hw/mlx5/odp.c int mlx5_ib_advise_mr_prefetch(struct ib_pd *pd,
pd               1741 drivers/infiniband/hw/mlx5/odp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               1751 drivers/infiniband/hw/mlx5/odp.c 		return mlx5_ib_prefetch_sg_list(pd, pf_flags, sg_list,
pd               1764 drivers/infiniband/hw/mlx5/odp.c 	work->pd = pd;
pd               1772 drivers/infiniband/hw/mlx5/odp.c 	valid_req = num_pending_prefetch_inc(pd, sg_list, num_sge);
pd                778 drivers/infiniband/hw/mlx5/qp.c static void destroy_user_rq(struct mlx5_ib_dev *dev, struct ib_pd *pd,
pd                794 drivers/infiniband/hw/mlx5/qp.c static int create_user_rq(struct mlx5_ib_dev *dev, struct ib_pd *pd,
pd                855 drivers/infiniband/hw/mlx5/qp.c static int create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd,
pd                937 drivers/infiniband/hw/mlx5/qp.c 	       attr->qp_type != IB_QPT_XRC_INI) ? to_mpd(pd)->uid : 0;
pd                985 drivers/infiniband/hw/mlx5/qp.c static void destroy_qp_user(struct mlx5_ib_dev *dev, struct ib_pd *pd,
pd               1193 drivers/infiniband/hw/mlx5/qp.c 				    struct ib_pd *pd)
pd               1198 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_tis_in, in, uid, to_mpd(pd)->uid);
pd               1207 drivers/infiniband/hw/mlx5/qp.c 				      struct mlx5_ib_sq *sq, struct ib_pd *pd)
pd               1209 drivers/infiniband/hw/mlx5/qp.c 	mlx5_cmd_destroy_tis(dev->mdev, sq->tisn, to_mpd(pd)->uid);
pd               1222 drivers/infiniband/hw/mlx5/qp.c 				   struct ib_pd *pd)
pd               1250 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_sq_in, in, uid, to_mpd(pd)->uid);
pd               1266 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd));
pd               1318 drivers/infiniband/hw/mlx5/qp.c 				   size_t qpinlen, struct ib_pd *pd)
pd               1339 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid);
pd               1357 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd));
pd               1390 drivers/infiniband/hw/mlx5/qp.c 				      struct ib_pd *pd)
pd               1395 drivers/infiniband/hw/mlx5/qp.c 	mlx5_cmd_destroy_tir(dev->mdev, rq->tirn, to_mpd(pd)->uid);
pd               1401 drivers/infiniband/hw/mlx5/qp.c 				    struct ib_pd *pd,
pd               1415 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid);
pd               1443 drivers/infiniband/hw/mlx5/qp.c 			destroy_raw_packet_qp_tir(dev, rq, 0, pd);
pd               1452 drivers/infiniband/hw/mlx5/qp.c 				struct ib_pd *pd,
pd               1463 drivers/infiniband/hw/mlx5/qp.c 	u16 uid = to_mpd(pd)->uid;
pd               1467 drivers/infiniband/hw/mlx5/qp.c 		err = create_raw_packet_qp_tis(dev, qp, sq, tdn, pd);
pd               1471 drivers/infiniband/hw/mlx5/qp.c 		err = create_raw_packet_qp_sq(dev, udata, sq, in, pd);
pd               1493 drivers/infiniband/hw/mlx5/qp.c 		err = create_raw_packet_qp_rq(dev, rq, in, inlen, pd);
pd               1498 drivers/infiniband/hw/mlx5/qp.c 			dev, rq, tdn, &qp->flags_en, pd, out,
pd               1534 drivers/infiniband/hw/mlx5/qp.c 	destroy_raw_packet_qp_tir(dev, rq, qp->flags_en, pd);
pd               1542 drivers/infiniband/hw/mlx5/qp.c 	destroy_raw_packet_qp_tis(dev, sq, pd);
pd               1555 drivers/infiniband/hw/mlx5/qp.c 		destroy_raw_packet_qp_tir(dev, rq, qp->flags_en, qp->ibqp.pd);
pd               1561 drivers/infiniband/hw/mlx5/qp.c 		destroy_raw_packet_qp_tis(dev, sq, qp->ibqp.pd);
pd               1583 drivers/infiniband/hw/mlx5/qp.c 			     to_mpd(qp->ibqp.pd)->uid);
pd               1587 drivers/infiniband/hw/mlx5/qp.c 				 struct ib_pd *pd,
pd               1684 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid);
pd               1802 drivers/infiniband/hw/mlx5/qp.c 					     to_mpd(pd)->uid);
pd               1945 drivers/infiniband/hw/mlx5/qp.c static int create_qp_common(struct mlx5_ib_dev *dev, struct ib_pd *pd,
pd               1978 drivers/infiniband/hw/mlx5/qp.c 		err = create_rss_raw_qp_tir(dev, qp, pd, init_attr, udata);
pd               2127 drivers/infiniband/hw/mlx5/qp.c 	if (pd) {
pd               2147 drivers/infiniband/hw/mlx5/qp.c 			err = create_user_qp(dev, pd, qp, udata, init_attr, &in,
pd               2177 drivers/infiniband/hw/mlx5/qp.c 		MLX5_SET(qpc, qpc, pd, to_mpd(pd ? pd : devr->p0)->pdn);
pd               2284 drivers/infiniband/hw/mlx5/qp.c 		err = create_raw_packet_qp(dev, qp, in, inlen, pd, udata,
pd               2321 drivers/infiniband/hw/mlx5/qp.c 		destroy_qp_user(dev, pd, qp, base, udata);
pd               2390 drivers/infiniband/hw/mlx5/qp.c 	return to_mpd(qp->ibqp.pd);
pd               2540 drivers/infiniband/hw/mlx5/qp.c static struct ib_qp *mlx5_ib_create_dct(struct ib_pd *pd,
pd               2569 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_dct_in, qp->dct.in, uid, to_mpd(pd)->uid);
pd               2572 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(dctc, dctc, pd, to_mpd(pd)->pdn);
pd               2627 drivers/infiniband/hw/mlx5/qp.c struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd,
pd               2640 drivers/infiniband/hw/mlx5/qp.c 	if (pd) {
pd               2641 drivers/infiniband/hw/mlx5/qp.c 		dev = to_mdev(pd->device);
pd               2679 drivers/infiniband/hw/mlx5/qp.c 			return mlx5_ib_create_dct(pd, init_attr, &ucmd, udata);
pd               2709 drivers/infiniband/hw/mlx5/qp.c 		err = create_qp_common(dev, pd, init_attr, udata, qp);
pd               2733 drivers/infiniband/hw/mlx5/qp.c 		return mlx5_ib_gsi_create_qp(pd, init_attr);
pd               2855 drivers/infiniband/hw/mlx5/qp.c 				      struct ib_pd *pd)
pd               2868 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid);
pd               2882 drivers/infiniband/hw/mlx5/qp.c 					 struct ib_pd *pd)
pd               2895 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid);
pd               2980 drivers/infiniband/hw/mlx5/qp.c 						  sl & 0xf, qp->ibqp.pd);
pd               3151 drivers/infiniband/hw/mlx5/qp.c 	const struct mlx5_modify_raw_qp_param *raw_qp_param, struct ib_pd *pd)
pd               3164 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(modify_rq_in, in, uid, to_mpd(pd)->uid);
pd               3193 drivers/infiniband/hw/mlx5/qp.c 	const struct mlx5_modify_raw_qp_param *raw_qp_param, struct ib_pd *pd)
pd               3210 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(modify_sq_in, in, uid, to_mpd(pd)->uid);
pd               3316 drivers/infiniband/hw/mlx5/qp.c 					       qp->ibqp.pd);
pd               3327 drivers/infiniband/hw/mlx5/qp.c 							    qp->ibqp.pd);
pd               3338 drivers/infiniband/hw/mlx5/qp.c 					      raw_qp_param, qp->ibqp.pd);
pd               3357 drivers/infiniband/hw/mlx5/qp.c 				    struct mlx5_ib_pd *pd,
pd               3457 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_pd *pd;
pd               3475 drivers/infiniband/hw/mlx5/qp.c 	pd = get_pd(qp);
pd               3504 drivers/infiniband/hw/mlx5/qp.c 				tx_affinity = get_tx_affinity(dev, pd, base, p,
pd               3566 drivers/infiniband/hw/mlx5/qp.c 	context->flags_pd = cpu_to_be32(pd ? pd->pdn : to_mpd(dev->devr.p0)->pdn);
pd               4379 drivers/infiniband/hw/mlx5/qp.c 	if (umrwr->pd)
pd               4380 drivers/infiniband/hw/mlx5/qp.c 		seg->flags_pd = cpu_to_be32(to_mpd(umrwr->pd)->pdn);
pd               4394 drivers/infiniband/hw/mlx5/qp.c 			     struct mlx5_ib_pd *pd)
pd               4400 drivers/infiniband/hw/mlx5/qp.c 	dseg->lkey = cpu_to_be32(pd->ibpd.local_dma_lkey);
pd               4814 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_pd *pd = to_mpd(qp->ibqp.pd);
pd               4815 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->ibpd.device);
pd               4857 drivers/infiniband/hw/mlx5/qp.c 		set_reg_data_seg(*seg, mr, pd);
pd               5128 drivers/infiniband/hw/mlx5/qp.c 						mr->ibmr.pd->local_dma_lkey;
pd               5918 drivers/infiniband/hw/mlx5/qp.c static int  create_rq(struct mlx5_ib_rwq *rwq, struct ib_pd *pd,
pd               5930 drivers/infiniband/hw/mlx5/qp.c 	dev = to_mdev(pd->device);
pd               5937 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid);
pd               5968 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(wq, wq, pd, to_mpd(pd)->pdn);
pd               6040 drivers/infiniband/hw/mlx5/qp.c static int prepare_user_rq(struct ib_pd *pd,
pd               6045 drivers/infiniband/hw/mlx5/qp.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd               6110 drivers/infiniband/hw/mlx5/qp.c 	err = create_user_rq(dev, pd, udata, rwq, &ucmd);
pd               6120 drivers/infiniband/hw/mlx5/qp.c struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd,
pd               6141 drivers/infiniband/hw/mlx5/qp.c 	dev = to_mdev(pd->device);
pd               6147 drivers/infiniband/hw/mlx5/qp.c 		err = prepare_user_rq(pd, init_attr, udata, rwq);
pd               6150 drivers/infiniband/hw/mlx5/qp.c 		err = create_rq(rwq, pd, init_attr);
pd               6177 drivers/infiniband/hw/mlx5/qp.c 	destroy_user_rq(dev, pd, rwq, udata);
pd               6189 drivers/infiniband/hw/mlx5/qp.c 	destroy_user_rq(dev, wq->pd, rwq, udata);
pd               6244 drivers/infiniband/hw/mlx5/qp.c 	rwq_ind_tbl->uid = to_mpd(init_attr->ind_tbl[0]->pd)->uid;
pd               6327 drivers/infiniband/hw/mlx5/qp.c 	MLX5_SET(modify_rq_in, in, uid, to_mpd(wq->pd)->uid);
pd                 44 drivers/infiniband/hw/mlx5/srq.c static int create_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq,
pd                 48 drivers/infiniband/hw/mlx5/srq.c 	struct mlx5_ib_dev *dev = to_mdev(pd->device);
pd                115 drivers/infiniband/hw/mlx5/srq.c 	in->uid = (in->type != IB_SRQT_XRC) ?  to_mpd(pd)->uid : 0;
pd                197 drivers/infiniband/hw/mlx5/srq.c static void destroy_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq,
pd                262 drivers/infiniband/hw/mlx5/srq.c 		err = create_srq_user(ib_srq->pd, srq, &in, udata, buf_size);
pd                299 drivers/infiniband/hw/mlx5/srq.c 	in.pd = to_mpd(ib_srq->pd)->pdn;
pd                329 drivers/infiniband/hw/mlx5/srq.c 		destroy_srq_user(ib_srq->pd, srq, udata);
pd                 26 drivers/infiniband/hw/mlx5/srq.h 	u32 pd;
pd                 36 drivers/infiniband/hw/mlx5/srq_cmd.c 	MLX5_SET(wq,   wq, pd,		  in->pd);
pd                 49 drivers/infiniband/hw/mlx5/srq_cmd.c 	MLX5_SET(srqc,	 srqc, pd,	      in->pd);
pd                 64 drivers/infiniband/hw/mlx5/srq_cmd.c 	in->pd		  = MLX5_GET(wq,   wq, pd);
pd                 77 drivers/infiniband/hw/mlx5/srq_cmd.c 	in->pd		  = MLX5_GET(srqc,   srqc, pd);
pd                196 drivers/infiniband/hw/mthca/mthca_allocator.c 		    union mthca_buf *buf, int *is_direct, struct mthca_pd *pd,
pd                262 drivers/infiniband/hw/mthca/mthca_allocator.c 	err = mthca_mr_alloc_phys(dev, pd->pd_num,
pd                154 drivers/infiniband/hw/mthca/mthca_av.c 		    struct mthca_pd *pd,
pd                170 drivers/infiniband/hw/mthca/mthca_av.c 	} else if (!atomic_read(&pd->sqp_count) &&
pd                197 drivers/infiniband/hw/mthca/mthca_av.c 	ah->key = pd->ntmr.ibmr.lkey;
pd                199 drivers/infiniband/hw/mthca/mthca_av.c 	av->port_pd = cpu_to_be32(pd->pd_num |
pd                 70 drivers/infiniband/hw/mthca/mthca_cq.c 	__be32 pd;
pd                833 drivers/infiniband/hw/mthca/mthca_cq.c 	cq_context->pd              = cpu_to_be32(pdn);
pd                425 drivers/infiniband/hw/mthca/mthca_dev.h 		    union mthca_buf *buf, int *is_direct, struct mthca_pd *pd,
pd                462 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd);
pd                463 drivers/infiniband/hw/mthca/mthca_dev.h void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd);
pd                471 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift,
pd                473 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd,
pd                475 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd,
pd                481 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd,
pd                512 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd,
pd                545 drivers/infiniband/hw/mthca/mthca_dev.h 		   struct mthca_pd *pd,
pd                554 drivers/infiniband/hw/mthca/mthca_dev.h 		    struct mthca_pd *pd,
pd                565 drivers/infiniband/hw/mthca/mthca_dev.h 		    struct mthca_pd *pd,
pd                 91 drivers/infiniband/hw/mthca/mthca_mad.c 	new_ah = rdma_create_ah(dev->send_agent[port_num - 1][0]->qp->pd,
pd                 54 drivers/infiniband/hw/mthca/mthca_mr.c 	__be32 pd;
pd                429 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift,
pd                468 drivers/infiniband/hw/mthca/mthca_mr.c 	mpt_entry->pd        = cpu_to_be32(pd);
pd                512 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd,
pd                516 drivers/infiniband/hw/mthca/mthca_mr.c 	return mthca_mr_alloc(dev, pd, 12, 0, ~0ULL, access, mr);
pd                519 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd,
pd                536 drivers/infiniband/hw/mthca/mthca_mr.c 	err = mthca_mr_alloc(dev, pd, buffer_size_shift, iova,
pd                567 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd,
pd                638 drivers/infiniband/hw/mthca/mthca_mr.c 	mpt_entry->pd        = cpu_to_be32(pd);
pd                 39 drivers/infiniband/hw/mthca/mthca_pd.c int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd)
pd                 43 drivers/infiniband/hw/mthca/mthca_pd.c 	pd->privileged = privileged;
pd                 45 drivers/infiniband/hw/mthca/mthca_pd.c 	atomic_set(&pd->sqp_count, 0);
pd                 46 drivers/infiniband/hw/mthca/mthca_pd.c 	pd->pd_num = mthca_alloc(&dev->pd_table.alloc);
pd                 47 drivers/infiniband/hw/mthca/mthca_pd.c 	if (pd->pd_num == -1)
pd                 51 drivers/infiniband/hw/mthca/mthca_pd.c 		err = mthca_mr_alloc_notrans(dev, pd->pd_num,
pd                 54 drivers/infiniband/hw/mthca/mthca_pd.c 					     &pd->ntmr);
pd                 56 drivers/infiniband/hw/mthca/mthca_pd.c 			mthca_free(&dev->pd_table.alloc, pd->pd_num);
pd                 62 drivers/infiniband/hw/mthca/mthca_pd.c void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd)
pd                 64 drivers/infiniband/hw/mthca/mthca_pd.c 	if (pd->privileged)
pd                 65 drivers/infiniband/hw/mthca/mthca_pd.c 		mthca_free_mr(dev, &pd->ntmr);
pd                 66 drivers/infiniband/hw/mthca/mthca_pd.c 	mthca_free(&dev->pd_table.alloc, pd->pd_num);
pd                369 drivers/infiniband/hw/mthca/mthca_provider.c 	struct mthca_pd *pd = to_mpd(ibpd);
pd                372 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd);
pd                377 drivers/infiniband/hw/mthca/mthca_provider.c 		if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) {
pd                378 drivers/infiniband/hw/mthca/mthca_provider.c 			mthca_pd_free(to_mdev(ibdev), pd);
pd                386 drivers/infiniband/hw/mthca/mthca_provider.c static void mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                388 drivers/infiniband/hw/mthca/mthca_provider.c 	mthca_pd_free(to_mdev(pd->device), to_mpd(pd));
pd                397 drivers/infiniband/hw/mthca/mthca_provider.c 	return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd), ah_attr,
pd                434 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd),
pd                468 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_qp *mthca_create_qp(struct ib_pd *pd,
pd                496 drivers/infiniband/hw/mthca/mthca_provider.c 			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
pd                504 drivers/infiniband/hw/mthca/mthca_provider.c 			err = mthca_map_user_db(to_mdev(pd->device), &context->uar,
pd                508 drivers/infiniband/hw/mthca/mthca_provider.c 				mthca_unmap_user_db(to_mdev(pd->device),
pd                521 drivers/infiniband/hw/mthca/mthca_provider.c 		err = mthca_alloc_qp(to_mdev(pd->device), to_mpd(pd),
pd                528 drivers/infiniband/hw/mthca/mthca_provider.c 			mthca_unmap_user_db(to_mdev(pd->device),
pd                532 drivers/infiniband/hw/mthca/mthca_provider.c 			mthca_unmap_user_db(to_mdev(pd->device),
pd                554 drivers/infiniband/hw/mthca/mthca_provider.c 		err = mthca_alloc_sqp(to_mdev(pd->device), to_mpd(pd),
pd                831 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc)
pd                840 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_mr_alloc_notrans(to_mdev(pd->device),
pd                841 drivers/infiniband/hw/mthca/mthca_provider.c 				     to_mpd(pd)->pd_num,
pd                854 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                857 drivers/infiniband/hw/mthca/mthca_provider.c 	struct mthca_dev *dev = to_mdev(pd->device);
pd                932 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length,
pd                962 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_fmr *mthca_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
pd                973 drivers/infiniband/hw/mthca/mthca_provider.c 	err = mthca_fmr_alloc(to_mdev(pd->device), to_mpd(pd)->pd_num,
pd                134 drivers/infiniband/hw/mthca/mthca_qp.c 	__be32 pd;
pd                709 drivers/infiniband/hw/mthca/mthca_qp.c 	qp_context->pd         = cpu_to_be32(to_mpd(ibqp->pd)->pd_num);
pd                956 drivers/infiniband/hw/mthca/mthca_qp.c static inline int mthca_max_inline_data(struct mthca_pd *pd, int max_data_size)
pd                959 drivers/infiniband/hw/mthca/mthca_qp.c 	return pd->ibpd.uobject ? max_data_size - MTHCA_INLINE_HEADER_SIZE : 0;
pd                963 drivers/infiniband/hw/mthca/mthca_qp.c 				 struct mthca_pd *pd,
pd                970 drivers/infiniband/hw/mthca/mthca_qp.c 	qp->max_inline_data = mthca_max_inline_data(pd, max_data_size);
pd                988 drivers/infiniband/hw/mthca/mthca_qp.c 			       struct mthca_pd *pd,
pd               1069 drivers/infiniband/hw/mthca/mthca_qp.c 			      &qp->queue, &qp->is_direct, pd, 0, &qp->mr);
pd               1160 drivers/infiniband/hw/mthca/mthca_qp.c 				 struct mthca_pd *pd,
pd               1188 drivers/infiniband/hw/mthca/mthca_qp.c 	ret = mthca_alloc_wqe_buf(dev, pd, qp, udata);
pd               1194 drivers/infiniband/hw/mthca/mthca_qp.c 	mthca_adjust_qp_caps(dev, pd, qp);
pd               1250 drivers/infiniband/hw/mthca/mthca_qp.c 			     struct mthca_pd *pd, struct mthca_qp *qp)
pd               1259 drivers/infiniband/hw/mthca/mthca_qp.c 	    cap->max_inline_data > mthca_max_inline_data(pd, max_data_size))
pd               1289 drivers/infiniband/hw/mthca/mthca_qp.c 		   struct mthca_pd *pd,
pd               1307 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_set_qp_size(dev, cap, pd, qp);
pd               1318 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
pd               1364 drivers/infiniband/hw/mthca/mthca_qp.c 		    struct mthca_pd *pd,
pd               1378 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_set_qp_size(dev, cap, pd, &sqp->qp);
pd               1402 drivers/infiniband/hw/mthca/mthca_qp.c 	err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq,
pd               1407 drivers/infiniband/hw/mthca/mthca_qp.c 	atomic_inc(&pd->sqp_count);
pd               1489 drivers/infiniband/hw/mthca/mthca_qp.c 		atomic_dec(&(to_mpd(qp->ibqp.pd)->sqp_count));
pd               1559 drivers/infiniband/hw/mthca/mthca_qp.c 	data->lkey       = cpu_to_be32(to_mpd(sqp->qp.ibqp.pd)->ntmr.ibmr.lkey);
pd                 98 drivers/infiniband/hw/mthca/mthca_srq.c 					 struct mthca_pd *pd,
pd                109 drivers/infiniband/hw/mthca/mthca_srq.c 	context->state_pd    = cpu_to_be32(pd->pd_num);
pd                119 drivers/infiniband/hw/mthca/mthca_srq.c 					 struct mthca_pd *pd,
pd                144 drivers/infiniband/hw/mthca/mthca_srq.c 	context->eq_pd = cpu_to_be32(MTHCA_EQ_ASYNC << 24 | pd->pd_num);
pd                154 drivers/infiniband/hw/mthca/mthca_srq.c static int mthca_alloc_srq_buf(struct mthca_dev *dev, struct mthca_pd *pd,
pd                171 drivers/infiniband/hw/mthca/mthca_srq.c 			      &srq->queue, &srq->is_direct, pd, 1, &srq->mr);
pd                206 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd,
pd                262 drivers/infiniband/hw/mthca/mthca_srq.c 	err = mthca_alloc_srq_buf(dev, pd, srq, udata);
pd                272 drivers/infiniband/hw/mthca/mthca_srq.c 		mthca_arbel_init_srq_context(dev, pd, srq, mailbox->buf, udata);
pd                274 drivers/infiniband/hw/mthca/mthca_srq.c 		mthca_tavor_init_srq_context(dev, pd, srq, mailbox->buf, udata);
pd                386 drivers/infiniband/hw/ocrdma/ocrdma.h 	struct ocrdma_pd *pd;
pd                423 drivers/infiniband/hw/ocrdma/ocrdma.h 	struct ocrdma_pd *pd;
pd                167 drivers/infiniband/hw/ocrdma/ocrdma_ah.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibah->pd);
pd                189 drivers/infiniband/hw/ocrdma/ocrdma_ah.c 	status = set_av_attr(dev, ah, attr, &sgid_attr->gid, pd->id,
pd                195 drivers/infiniband/hw/ocrdma/ocrdma_ah.c 	if ((pd->uctx) && (pd->uctx->ah_tbl.va)) {
pd                196 drivers/infiniband/hw/ocrdma/ocrdma_ah.c 		ahid_addr = pd->uctx->ah_tbl.va + rdma_ah_get_dlid(attr);
pd               1463 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_mbx_alloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd)
pd               1472 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	if (pd->dpp_enabled)
pd               1478 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	pd->id = rsp->dpp_page_pdid & OCRDMA_ALLOC_PD_RSP_PDID_MASK;
pd               1480 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pd->dpp_enabled = true;
pd               1481 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pd->dpp_page = rsp->dpp_page_pdid >>
pd               1484 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pd->dpp_enabled = false;
pd               1485 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		pd->num_dpp_qp = 0;
pd               1492 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_mbx_dealloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd)
pd               1500 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->id = pd->id;
pd               2188 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pd *pd = qp->pd;
pd               2189 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd               2243 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pd *pd = qp->pd;
pd               2244 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd               2284 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 					 struct ocrdma_pd *pd,
pd               2288 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	pd->num_dpp_qp--;
pd               2302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pd *pd = qp->pd;
pd               2303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd               2373 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pd *pd = qp->pd;
pd               2374 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd               2419 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->type_pgsz_pdn |= (pd->id << OCRDMA_CREATE_QP_REQ_PD_ID_SHIFT) &
pd               2440 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	if (pd->dpp_enabled && attrs->cap.max_inline_data && pd->num_dpp_qp &&
pd               2442 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		ocrdma_set_create_qp_dpp_cmd(cmd, pd, qp, enable_dpp_cq,
pd               2739 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 		qp->pd->num_dpp_qp++;
pd               2745 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			  struct ocrdma_pd *pd)
pd               2760 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	cmd->pgsz_pdid = pd->id & OCRDMA_CREATE_SRQ_PD_ID_MASK;
pd               2822 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_pd *pd = srq->pd;
pd               2823 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd                321 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int ocrdma_get_pd_num(struct ocrdma_dev *dev, struct ocrdma_pd *pd)
pd                327 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (pd->dpp_enabled) {
pd                331 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->id = dev->pd_mgr->pd_dpp_start + pd_idx;
pd                332 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->dpp_page = dev->pd_mgr->dpp_page_index + pd_idx;
pd                336 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->id = dev->pd_mgr->pd_norm_start + pd_idx;
pd                337 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->dpp_enabled = false;
pd                344 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->id = dev->pd_mgr->pd_norm_start + pd_idx;
pd                363 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int _ocrdma_alloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd,
pd                370 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		pd->dpp_enabled =
pd                372 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		pd->num_dpp_qp =
pd                373 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->dpp_enabled ? (dev->nic_info.db_page_size /
pd                378 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		return ocrdma_get_pd_num(dev, pd);
pd                381 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_mbx_alloc_pd(dev, pd);
pd                383 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (pd->dpp_enabled) {
pd                384 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->dpp_enabled = false;
pd                385 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			pd->num_dpp_qp = 0;
pd                395 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				 struct ocrdma_pd *pd)
pd                397 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	return (uctx->cntxt_pd == pd);
pd                401 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			      struct ocrdma_pd *pd)
pd                404 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_put_pd_num(dev, pd->id, pd->dpp_enabled);
pd                406 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_mbx_dealloc_pd(dev, pd);
pd                414 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ib_pd *pd;
pd                417 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd = rdma_zalloc_drv_obj(ibdev, ib_pd);
pd                418 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (!pd)
pd                421 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd->device  = ibdev;
pd                422 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	uctx->cntxt_pd = get_ocrdma_pd(pd);
pd                438 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = uctx->cntxt_pd;
pd                439 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd                443 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		       __func__, dev->id, pd->id);
pd                447 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	_ocrdma_dealloc_pd(dev, pd);
pd                452 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = NULL;
pd                457 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		pd = uctx->cntxt_pd;
pd                461 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	return pd;
pd                588 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int ocrdma_copy_pd_uresp(struct ocrdma_dev *dev, struct ocrdma_pd *pd,
pd                600 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	rsp.id = pd->id;
pd                601 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	rsp.dpp_enabled = pd->dpp_enabled;
pd                602 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	db_page_addr = ocrdma_get_db_addr(dev, pd->id);
pd                609 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (pd->dpp_enabled) {
pd                611 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				(pd->id * PAGE_SIZE);
pd                624 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd->uctx = uctx;
pd                628 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (pd->dpp_enabled)
pd                629 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(pd->uctx, dpp_page_addr, PAGE_SIZE);
pd                631 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_del_mmap(pd->uctx, db_page_addr, db_page_size);
pd                639 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd;
pd                646 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		pd = ocrdma_get_ucontext_pd(uctx);
pd                647 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (pd) {
pd                653 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd = get_ocrdma_pd(ibpd);
pd                654 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = _ocrdma_alloc_pd(dev, pd, uctx, udata);
pd                660 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		status = ocrdma_copy_pd_uresp(dev, pd, udata);
pd                670 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		_ocrdma_dealloc_pd(dev, pd);
pd                677 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibpd);
pd                682 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	uctx = pd->uctx;
pd                685 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * PAGE_SIZE);
pd                686 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (pd->dpp_enabled)
pd                687 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			ocrdma_del_mmap(pd->uctx, dpp_db, PAGE_SIZE);
pd                688 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		usr_db = ocrdma_get_db_addr(dev, pd->id);
pd                689 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(pd->uctx, usr_db, dev->nic_info.db_page_size);
pd                691 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		if (is_ucontext_pd(uctx, pd)) {
pd                696 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	_ocrdma_dealloc_pd(dev, pd);
pd                727 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibpd);
pd                739 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_alloc_lkey(dev, mr, pd->id, acc, 0,
pd                867 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd;
pd                870 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd = get_ocrdma_pd(ibpd);
pd                901 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_reg_mr(dev, &mr->hwmr, pd->id, acc);
pd               1191 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = qp->pd;
pd               1192 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device);
pd               1196 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * dev->nic_info.db_page_size);
pd               1225 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_add_mmap(pd->uctx, uresp.sq_page_addr[0],
pd               1231 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		status = ocrdma_add_mmap(pd->uctx, uresp.rq_page_addr[0],
pd               1238 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_del_mmap(pd->uctx, uresp.sq_page_addr[0], uresp.sq_page_size);
pd               1244 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			     struct ocrdma_pd *pd)
pd               1248 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * dev->nic_info.db_page_size) +
pd               1251 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * dev->nic_info.db_page_size) +
pd               1255 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * dev->nic_info.db_page_size) +
pd               1258 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			(pd->id * dev->nic_info.db_page_size) +
pd               1279 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 				      struct ocrdma_pd *pd,
pd               1282 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	qp->pd = pd;
pd               1311 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibpd);
pd               1331 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_set_qp_init_params(qp, pd, attrs);
pd               1353 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	ocrdma_set_qp_db(dev, qp, pd);
pd               1686 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd;
pd               1696 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	pd = qp->pd;
pd               1725 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (!pd->uctx) {
pd               1731 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (pd->uctx) {
pd               1732 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa,
pd               1735 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 			ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa,
pd               1759 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	    (srq->pd->id * dev->nic_info.db_page_size);
pd               1773 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_add_mmap(srq->pd->uctx, uresp.rq_page_addr[0],
pd               1784 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibsrq->pd);
pd               1794 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	srq->pd = pd;
pd               1795 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	srq->db = dev->nic_info.db + (pd->id * dev->nic_info.db_page_size);
pd               1796 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_mbx_create_srq(dev, srq, init_attr, pd);
pd               1877 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	if (srq->pd->uctx)
pd               1878 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa,
pd               2916 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_pd *pd = get_ocrdma_pd(ibpd);
pd               2947 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	status = ocrdma_reg_mr(dev, &mr->hwmr, pd->id, 0);
pd                 71 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h int ocrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                 72 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h void ocrdma_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                105 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h struct ib_mr *ocrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                232 drivers/infiniband/hw/qedr/qedr.h 	struct qedr_pd *pd;
pd                407 drivers/infiniband/hw/qedr/qedr.h 	struct qedr_pd *pd;
pd                438 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd                454 drivers/infiniband/hw/qedr/verbs.c 	pd->pd_id = pd_id;
pd                470 drivers/infiniband/hw/qedr/verbs.c 		pd->uctx = context;
pd                471 drivers/infiniband/hw/qedr/verbs.c 		pd->uctx->pd = pd;
pd                480 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd                482 drivers/infiniband/hw/qedr/verbs.c 	DP_DEBUG(dev, QEDR_MSG_INIT, "Deallocating PD %d\n", pd->pd_id);
pd                483 drivers/infiniband/hw/qedr/verbs.c 	dev->ops->rdma_dealloc_pd(dev->rdma_ctx, pd->pd_id);
pd               1193 drivers/infiniband/hw/qedr/verbs.c 				      struct qedr_pd *pd,
pd               1201 drivers/infiniband/hw/qedr/verbs.c 	qp->pd = pd;
pd               1222 drivers/infiniband/hw/qedr/verbs.c 		 pd->pd_id, qp->qp_type, qp->max_inline_data,
pd               1364 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibsrq->pd);
pd               1374 drivers/infiniband/hw/qedr/verbs.c 		 (udata) ? "User lib" : "kernel", pd);
pd               1416 drivers/infiniband/hw/qedr/verbs.c 	in_params.pd_id = pd->pd_id;
pd               1515 drivers/infiniband/hw/qedr/verbs.c 			      struct qedr_pd *pd,
pd               1527 drivers/infiniband/hw/qedr/verbs.c 	params->pd = pd->pd_id;
pd               1528 drivers/infiniband/hw/qedr/verbs.c 	params->dpi = pd->uctx ? pd->uctx->dpi : dev->dpi;
pd               1602 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd               1630 drivers/infiniband/hw/qedr/verbs.c 	qedr_init_common_qp_in_params(dev, pd, qp, attrs, false, &in_params);
pd               1814 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd               1863 drivers/infiniband/hw/qedr/verbs.c 	qedr_init_common_qp_in_params(dev, pd, qp, attrs, true, &in_params);
pd               1889 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd               1895 drivers/infiniband/hw/qedr/verbs.c 		 udata ? "user library" : "kernel", pd);
pd               1903 drivers/infiniband/hw/qedr/verbs.c 		 udata ? "user library" : "kernel", attrs->event_handler, pd,
pd               1915 drivers/infiniband/hw/qedr/verbs.c 	qedr_set_common_qp_params(dev, qp, pd, attrs);
pd               2607 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd;
pd               2610 drivers/infiniband/hw/qedr/verbs.c 	pd = get_qedr_pd(ibpd);
pd               2613 drivers/infiniband/hw/qedr/verbs.c 		 pd->pd_id, start, len, usr_addr, acc);
pd               2646 drivers/infiniband/hw/qedr/verbs.c 	mr->hw_mr.pd = pd->pd_id;
pd               2714 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd               2720 drivers/infiniband/hw/qedr/verbs.c 		 "qedr_alloc_frmr pd = %d max_page_list_len= %d\n", pd->pd_id,
pd               2743 drivers/infiniband/hw/qedr/verbs.c 	mr->hw_mr.pd = pd->pd_id;
pd               2856 drivers/infiniband/hw/qedr/verbs.c 	struct qedr_pd *pd = get_qedr_pd(ibpd);
pd               2874 drivers/infiniband/hw/qedr/verbs.c 	mr->hw_mr.pd = pd->pd_id;
pd                 50 drivers/infiniband/hw/qedr/verbs.h int qedr_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                 51 drivers/infiniband/hw/qedr/verbs.h void qedr_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                 87 drivers/infiniband/hw/qedr/verbs.h struct ib_mr *qedr_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                 40 drivers/infiniband/hw/qib/qib_user_sdma.h int qib_user_sdma_writev(struct qib_ctxtdata *pd,
pd               1355 drivers/infiniband/hw/qib/qib_verbs.c 		ah = rdma_create_ah(qp0->ibqp.pd, &attr, 0);
pd                 96 drivers/infiniband/hw/usnic/usnic_ib.h 	struct usnic_ib_pd		*pd;
pd                587 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 				struct usnic_ib_pd *pd,
pd                597 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 		err = usnic_uiom_attach_dev_to_pd(pd->umem_pd, &pdev->dev);
pd                603 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 		vf->pd = pd;
pd                607 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 	WARN_ON(vf->pd != pd);
pd                616 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 	struct usnic_ib_pd *pd;
pd                620 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 	pd = qp_grp->vf->pd;
pd                623 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 		qp_grp->vf->pd = NULL;
pd                624 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 		usnic_uiom_detach_dev_from_pd(pd->umem_pd, &pdev->dev);
pd                670 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 			struct usnic_ib_pd *pd,
pd                702 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c 	err = qp_grp_and_vf_bind(vf, pd, qp_grp);
pd                 94 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.h 			struct usnic_ib_pd *pd,
pd                173 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 				struct usnic_ib_pd *pd,
pd                192 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 		dev_list = usnic_uiom_get_dev_list(pd->umem_pd);
pd                206 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 								vf, pd,
pd                226 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 							pd, res_spec,
pd                453 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	struct usnic_ib_pd *pd = to_upd(ibpd);
pd                456 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	umem_pd = pd->umem_pd = usnic_uiom_alloc_pd();
pd                464 drivers/infiniband/hw/usnic/usnic_ib_verbs.c void usnic_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                466 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	usnic_uiom_dealloc_pd((to_upd(pd))->umem_pd);
pd                469 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct ib_qp *usnic_ib_create_qp(struct ib_pd *pd,
pd                485 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	us_ibdev = to_usdev(pd->device);
pd                515 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	qp_grp = find_free_vf_and_create_qp_grp(us_ibdev, to_upd(pd),
pd                605 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct ib_mr *usnic_ib_reg_mr(struct ib_pd *pd, u64 start, u64 length,
pd                619 drivers/infiniband/hw/usnic/usnic_ib_verbs.c 	mr->umem = usnic_uiom_reg_get(to_upd(pd)->umem_pd, start, length,
pd                 54 drivers/infiniband/hw/usnic/usnic_ib_verbs.h void usnic_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                 55 drivers/infiniband/hw/usnic/usnic_ib_verbs.h struct ib_qp *usnic_ib_create_qp(struct ib_pd *pd,
pd                 64 drivers/infiniband/hw/usnic/usnic_ib_verbs.h struct ib_mr *usnic_ib_reg_mr(struct ib_pd *pd, u64 start, u64 length,
pd                196 drivers/infiniband/hw/usnic/usnic_uiom.c 						struct usnic_uiom_pd *pd)
pd                207 drivers/infiniband/hw/usnic/usnic_uiom.c 			iommu_unmap(pd->domain, va, PAGE_SIZE);
pd                214 drivers/infiniband/hw/usnic/usnic_uiom.c static void __usnic_uiom_reg_release(struct usnic_uiom_pd *pd,
pd                228 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock(&pd->lock);
pd                229 drivers/infiniband/hw/usnic/usnic_uiom.c 	usnic_uiom_remove_interval(&pd->root, vpn_start,
pd                231 drivers/infiniband/hw/usnic/usnic_uiom.c 	usnic_uiom_unmap_sorted_intervals(&rm_intervals, pd);
pd                241 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                255 drivers/infiniband/hw/usnic/usnic_uiom.c 	struct usnic_uiom_pd *pd = uiomr->pd;
pd                284 drivers/infiniband/hw/usnic/usnic_uiom.c 				err = iommu_map(pd->domain, va_start, pa_start,
pd                301 drivers/infiniband/hw/usnic/usnic_uiom.c 				err = iommu_map(pd->domain, va_start, pa_start,
pd                330 drivers/infiniband/hw/usnic/usnic_uiom.c 	usnic_uiom_unmap_sorted_intervals(intervals, pd);
pd                334 drivers/infiniband/hw/usnic/usnic_uiom.c struct usnic_uiom_reg *usnic_uiom_reg_get(struct usnic_uiom_pd *pd,
pd                367 drivers/infiniband/hw/usnic/usnic_uiom.c 	uiomr->pd = pd;
pd                377 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock(&pd->lock);
pd                381 drivers/infiniband/hw/usnic/usnic_uiom.c 						&pd->root,
pd                397 drivers/infiniband/hw/usnic/usnic_uiom.c 	err = usnic_uiom_insert_interval(&pd->root, vpn_start, vpn_last,
pd                406 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                411 drivers/infiniband/hw/usnic/usnic_uiom.c 	usnic_uiom_unmap_sorted_intervals(&sorted_diff_intervals, pd);
pd                416 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                436 drivers/infiniband/hw/usnic/usnic_uiom.c 	__usnic_uiom_reg_release(uiomr->pd, uiomr, 1);
pd                444 drivers/infiniband/hw/usnic/usnic_uiom.c 	struct usnic_uiom_pd *pd;
pd                447 drivers/infiniband/hw/usnic/usnic_uiom.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                448 drivers/infiniband/hw/usnic/usnic_uiom.c 	if (!pd)
pd                451 drivers/infiniband/hw/usnic/usnic_uiom.c 	pd->domain = domain = iommu_domain_alloc(&pci_bus_type);
pd                454 drivers/infiniband/hw/usnic/usnic_uiom.c 		kfree(pd);
pd                458 drivers/infiniband/hw/usnic/usnic_uiom.c 	iommu_set_fault_handler(pd->domain, usnic_uiom_dma_fault, NULL);
pd                460 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock_init(&pd->lock);
pd                461 drivers/infiniband/hw/usnic/usnic_uiom.c 	INIT_LIST_HEAD(&pd->devs);
pd                463 drivers/infiniband/hw/usnic/usnic_uiom.c 	return pd;
pd                466 drivers/infiniband/hw/usnic/usnic_uiom.c void usnic_uiom_dealloc_pd(struct usnic_uiom_pd *pd)
pd                468 drivers/infiniband/hw/usnic/usnic_uiom.c 	iommu_domain_free(pd->domain);
pd                469 drivers/infiniband/hw/usnic/usnic_uiom.c 	kfree(pd);
pd                472 drivers/infiniband/hw/usnic/usnic_uiom.c int usnic_uiom_attach_dev_to_pd(struct usnic_uiom_pd *pd, struct device *dev)
pd                482 drivers/infiniband/hw/usnic/usnic_uiom.c 	err = iommu_attach_device(pd->domain, dev);
pd                493 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock(&pd->lock);
pd                494 drivers/infiniband/hw/usnic/usnic_uiom.c 	list_add_tail(&uiom_dev->link, &pd->devs);
pd                495 drivers/infiniband/hw/usnic/usnic_uiom.c 	pd->dev_cnt++;
pd                496 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                501 drivers/infiniband/hw/usnic/usnic_uiom.c 	iommu_detach_device(pd->domain, dev);
pd                507 drivers/infiniband/hw/usnic/usnic_uiom.c void usnic_uiom_detach_dev_from_pd(struct usnic_uiom_pd *pd, struct device *dev)
pd                512 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock(&pd->lock);
pd                513 drivers/infiniband/hw/usnic/usnic_uiom.c 	list_for_each_entry(uiom_dev, &pd->devs, link) {
pd                523 drivers/infiniband/hw/usnic/usnic_uiom.c 		spin_unlock(&pd->lock);
pd                528 drivers/infiniband/hw/usnic/usnic_uiom.c 	pd->dev_cnt--;
pd                529 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                531 drivers/infiniband/hw/usnic/usnic_uiom.c 	return iommu_detach_device(pd->domain, dev);
pd                534 drivers/infiniband/hw/usnic/usnic_uiom.c struct device **usnic_uiom_get_dev_list(struct usnic_uiom_pd *pd)
pd                540 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_lock(&pd->lock);
pd                541 drivers/infiniband/hw/usnic/usnic_uiom.c 	devs = kcalloc(pd->dev_cnt + 1, sizeof(*devs), GFP_ATOMIC);
pd                547 drivers/infiniband/hw/usnic/usnic_uiom.c 	list_for_each_entry(uiom_dev, &pd->devs, link) {
pd                551 drivers/infiniband/hw/usnic/usnic_uiom.c 	spin_unlock(&pd->lock);
pd                 66 drivers/infiniband/hw/usnic/usnic_uiom.h 	struct usnic_uiom_pd		*pd;
pd                 84 drivers/infiniband/hw/usnic/usnic_uiom.h void usnic_uiom_dealloc_pd(struct usnic_uiom_pd *pd);
pd                 85 drivers/infiniband/hw/usnic/usnic_uiom.h int usnic_uiom_attach_dev_to_pd(struct usnic_uiom_pd *pd, struct device *dev);
pd                 86 drivers/infiniband/hw/usnic/usnic_uiom.h void usnic_uiom_detach_dev_from_pd(struct usnic_uiom_pd *pd,
pd                 88 drivers/infiniband/hw/usnic/usnic_uiom.h struct device **usnic_uiom_get_dev_list(struct usnic_uiom_pd *pd);
pd                 90 drivers/infiniband/hw/usnic/usnic_uiom.h struct usnic_uiom_reg *usnic_uiom_reg_get(struct usnic_uiom_pd *pd,
pd                 58 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_get_dma_mr(struct ib_pd *pd, int acc)
pd                 60 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	struct pvrdma_dev *dev = to_vdev(pd->device);
pd                 81 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	cmd->pd_handle = to_vpd(pd)->pd_handle;
pd                111 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                115 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	struct pvrdma_dev *dev = to_vdev(pd->device);
pd                169 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	cmd->pd_handle = to_vpd(pd)->pd_handle;
pd                204 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                207 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	struct pvrdma_dev *dev = to_vdev(pd->device);
pd                240 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c 	cmd->pd_handle = to_vpd(pd)->pd_handle;
pd                188 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct ib_qp *pvrdma_create_qp(struct ib_pd *pd,
pd                193 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	struct pvrdma_dev *dev = to_vdev(pd->device);
pd                230 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 		    init_attr->port_num > pd->device->phys_port_cnt ||
pd                293 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			ret = pvrdma_set_sq_size(to_vdev(pd->device),
pd                298 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 			ret = pvrdma_set_rq_size(to_vdev(pd->device),
pd                349 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c 	cmd->pd_handle = to_vpd(pd)->pd_handle;
pd                177 drivers/infiniband/hw/vmw_pvrdma/pvrdma_srq.c 	cmd->pd_handle = to_vpd(ibsrq->pd)->pd_handle;
pd                430 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	struct pvrdma_pd *pd = to_vpd(ibpd);
pd                455 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	pd->privileged = !udata;
pd                456 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	pd->pd_handle = resp->pd_handle;
pd                457 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	pd->pdn = resp->pd_handle;
pd                464 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 			pvrdma_dealloc_pd(&pd->ibpd, udata);
pd                484 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c void pvrdma_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                486 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	struct pvrdma_dev *dev = to_vdev(pd->device);
pd                492 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	cmd->pd_handle = to_vpd(pd)->pd_handle;
pd                531 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c 	ah->av.port_pd = to_vpd(ibah->pd)->pd_handle | (port_num << 24);
pd                401 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd                403 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_get_dma_mr(struct ib_pd *pd, int acc);
pd                404 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                408 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                428 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_qp *pvrdma_create_qp(struct ib_pd *pd,
pd                136 drivers/infiniband/sw/rdmavt/mr.c static int rvt_init_mregion(struct rvt_mregion *mr, struct ib_pd *pd,
pd                140 drivers/infiniband/sw/rdmavt/mr.c 	struct rvt_dev_info *dev = ib_to_rvt(pd->device);
pd                158 drivers/infiniband/sw/rdmavt/mr.c 	mr->pd = pd;
pd                184 drivers/infiniband/sw/rdmavt/mr.c 	struct rvt_dev_info *dev = ib_to_rvt(mr->pd->device);
pd                253 drivers/infiniband/sw/rdmavt/mr.c 	struct rvt_dev_info *dev = ib_to_rvt(mr->pd->device);
pd                280 drivers/infiniband/sw/rdmavt/mr.c static struct rvt_mr *__rvt_alloc_mr(int count, struct ib_pd *pd)
pd                292 drivers/infiniband/sw/rdmavt/mr.c 	rval = rvt_init_mregion(&mr->mr, pd, count, 0);
pd                331 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_get_dma_mr(struct ib_pd *pd, int acc)
pd                337 drivers/infiniband/sw/rdmavt/mr.c 	if (ibpd_to_rvtpd(pd)->user)
pd                346 drivers/infiniband/sw/rdmavt/mr.c 	rval = rvt_init_mregion(&mr->mr, pd, 0, 0);
pd                380 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                399 drivers/infiniband/sw/rdmavt/mr.c 	mr = __rvt_alloc_mr(n, pd);
pd                456 drivers/infiniband/sw/rdmavt/mr.c 	if (mr->pd != qp->ibqp.pd)
pd                470 drivers/infiniband/sw/rdmavt/mr.c 	struct rvt_dev_info *rdi = ib_to_rvt(mr->pd->device);
pd                489 drivers/infiniband/sw/rdmavt/mr.c 	struct rvt_dev_info *rdi = ib_to_rvt(mr->pd->device);
pd                502 drivers/infiniband/sw/rdmavt/mr.c 			   t, mr, mr->pd, mr->lkey,
pd                579 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                587 drivers/infiniband/sw/rdmavt/mr.c 	mr = __rvt_alloc_mr(max_num_sg, pd);
pd                663 drivers/infiniband/sw/rdmavt/mr.c 	if (qp->ibqp.pd != mr->mr.pd)
pd                703 drivers/infiniband/sw/rdmavt/mr.c 	if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd))
pd                724 drivers/infiniband/sw/rdmavt/mr.c struct ib_fmr *rvt_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
pd                738 drivers/infiniband/sw/rdmavt/mr.c 	rval = rvt_init_mregion(&fmr->mr, pd, fmr_attr->max_pages,
pd                914 drivers/infiniband/sw/rdmavt/mr.c int rvt_lkey_ok(struct rvt_lkey_table *rkt, struct rvt_pd *pd,
pd                927 drivers/infiniband/sw/rdmavt/mr.c 		struct rvt_dev_info *dev = ib_to_rvt(pd->ibpd.device);
pd                929 drivers/infiniband/sw/rdmavt/mr.c 		if (pd->user)
pd                959 drivers/infiniband/sw/rdmavt/mr.c 		     mr->lkey != sge->lkey || mr->pd != &pd->ibpd))
pd               1039 drivers/infiniband/sw/rdmavt/mr.c 		struct rvt_pd *pd = ibpd_to_rvtpd(qp->ibqp.pd);
pd               1040 drivers/infiniband/sw/rdmavt/mr.c 		struct rvt_dev_info *rdi = ib_to_rvt(pd->ibpd.device);
pd               1042 drivers/infiniband/sw/rdmavt/mr.c 		if (pd->user)
pd               1067 drivers/infiniband/sw/rdmavt/mr.c 		     mr->lkey != rkey || qp->ibqp.pd != mr->pd))
pd                 77 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_get_dma_mr(struct ib_pd *pd, int acc);
pd                 78 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
pd                 82 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd                 86 drivers/infiniband/sw/rdmavt/mr.h struct ib_fmr *rvt_alloc_fmr(struct ib_pd *pd, int mr_access_flags,
pd                 64 drivers/infiniband/sw/rdmavt/pd.c 	struct rvt_pd *pd = ibpd_to_rvtpd(ibpd);
pd                 85 drivers/infiniband/sw/rdmavt/pd.c 	pd->user = !!udata;
pd                 53 drivers/infiniband/sw/rdmavt/pd.h int rvt_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
pd               1914 drivers/infiniband/sw/rdmavt/qp.c 	    ibpd_to_rvtpd(qp->ibqp.pd)->user)
pd               1928 drivers/infiniband/sw/rdmavt/qp.c 		if (qp->ibqp.pd != ud_wr(wr)->ah->pd)
pd               2009 drivers/infiniband/sw/rdmavt/qp.c 	struct rvt_pd *pd;
pd               2076 drivers/infiniband/sw/rdmavt/qp.c 	pd = ibpd_to_rvtpd(qp->ibqp.pd);
pd               2094 drivers/infiniband/sw/rdmavt/qp.c 			ret = rvt_lkey_ok(rkt, pd, &wqe->sg_list[j], last_sge,
pd               2311 drivers/infiniband/sw/rdmavt/qp.c 	struct rvt_pd *pd;
pd               2316 drivers/infiniband/sw/rdmavt/qp.c 	pd = ibpd_to_rvtpd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd);
pd               2324 drivers/infiniband/sw/rdmavt/qp.c 		ret = rvt_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge,
pd                 66 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ENTRY(ib_to_rvt(mr->pd->device))
pd                 79 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ASSIGN(ib_to_rvt(mr->pd->device));
pd                127 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ENTRY(ib_to_rvt(sge->mr->pd->device))
pd                142 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ASSIGN(ib_to_rvt(sge->mr->pd->device));
pd                154 drivers/infiniband/sw/rdmavt/trace_mr.h 		__entry->user = ibpd_to_rvtpd(sge->mr->pd)->user;
pd                189 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device))
pd                198 drivers/infiniband/sw/rdmavt/trace_mr.h 		RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device))
pd                373 drivers/infiniband/sw/rxe/rxe_comp.c 	ret = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE,
pd                393 drivers/infiniband/sw/rxe/rxe_comp.c 	ret = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE,
pd                106 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_dma(struct rxe_pd *pd,
pd                109 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_user(struct rxe_pd *pd, u64 start,
pd                113 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_fast(struct rxe_pd *pd,
pd                119 drivers/infiniband/sw/rxe/rxe_loc.h int copy_data(struct rxe_pd *pd, int access,
pd                130 drivers/infiniband/sw/rxe/rxe_loc.h struct rxe_mem *lookup_mem(struct rxe_pd *pd, int access, u32 key,
pd                157 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_qp_from_init(struct rxe_dev *rxe, struct rxe_qp *qp, struct rxe_pd *pd,
pd                147 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_dma(struct rxe_pd *pd,
pd                152 drivers/infiniband/sw/rxe/rxe_mr.c 	mem->pd			= pd;
pd                160 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_user(struct rxe_pd *pd, u64 start,
pd                222 drivers/infiniband/sw/rxe/rxe_mr.c 	mem->pd			= pd;
pd                238 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_fast(struct rxe_pd *pd,
pd                252 drivers/infiniband/sw/rxe/rxe_mr.c 	mem->pd			= pd;
pd                372 drivers/infiniband/sw/rxe/rxe_mr.c 			*crcp = rxe_crc32(to_rdev(mem->pd->ibpd.device),
pd                406 drivers/infiniband/sw/rxe/rxe_mr.c 			crc = rxe_crc32(to_rdev(mem->pd->ibpd.device),
pd                436 drivers/infiniband/sw/rxe/rxe_mr.c 	struct rxe_pd		*pd,
pd                461 drivers/infiniband/sw/rxe/rxe_mr.c 		mem = lookup_mem(pd, access, sge->lkey, lookup_local);
pd                486 drivers/infiniband/sw/rxe/rxe_mr.c 				mem = lookup_mem(pd, access, sge->lkey,
pd                568 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_mem *lookup_mem(struct rxe_pd *pd, int access, u32 key,
pd                572 drivers/infiniband/sw/rxe/rxe_mr.c 	struct rxe_dev *rxe = to_rdev(pd->ibpd.device);
pd                581 drivers/infiniband/sw/rxe/rxe_mr.c 		     mem->pd != pd ||
pd                336 drivers/infiniband/sw/rxe/rxe_qp.c int rxe_qp_from_init(struct rxe_dev *rxe, struct rxe_qp *qp, struct rxe_pd *pd,
pd                347 drivers/infiniband/sw/rxe/rxe_qp.c 	rxe_add_ref(pd);
pd                353 drivers/infiniband/sw/rxe/rxe_qp.c 	qp->pd			= pd;
pd                380 drivers/infiniband/sw/rxe/rxe_qp.c 	rxe_drop_ref(pd);
pd                824 drivers/infiniband/sw/rxe/rxe_qp.c 	if (qp->pd)
pd                825 drivers/infiniband/sw/rxe/rxe_qp.c 		rxe_drop_ref(qp->pd);
pd                496 drivers/infiniband/sw/rxe/rxe_req.c 			err = copy_data(qp->pd, 0, &wqe->dma,
pd                459 drivers/infiniband/sw/rxe/rxe_resp.c 	mem = lookup_mem(qp->pd, access, rkey, lookup_remote);
pd                512 drivers/infiniband/sw/rxe/rxe_resp.c 	err = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE, &qp->resp.wqe->dma,
pd                182 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                184 drivers/infiniband/sw/rxe/rxe_verbs.c 	return rxe_add_to_pool(&rxe->pd_pool, &pd->pelem);
pd                189 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                191 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(pd);
pd                297 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibsrq->pd);
pd                315 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_add_ref(pd);
pd                316 drivers/infiniband/sw/rxe/rxe_verbs.c 	srq->pd = pd;
pd                325 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(pd);
pd                383 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(srq->pd);
pd                417 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                447 drivers/infiniband/sw/rxe/rxe_verbs.c 	err = rxe_qp_from_init(rxe, qp, pd, init, uresp, ibpd, udata);
pd                897 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                909 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_add_ref(pd);
pd                911 drivers/infiniband/sw/rxe/rxe_verbs.c 	err = rxe_mem_init_dma(pd, access, mr);
pd                918 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(pd);
pd                933 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                944 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_add_ref(pd);
pd                946 drivers/infiniband/sw/rxe/rxe_verbs.c 	err = rxe_mem_init_user(pd, start, length, iova,
pd                954 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(pd);
pd                966 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(mr->pd);
pd                976 drivers/infiniband/sw/rxe/rxe_verbs.c 	struct rxe_pd *pd = to_rpd(ibpd);
pd                991 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_add_ref(pd);
pd                993 drivers/infiniband/sw/rxe/rxe_verbs.c 	err = rxe_mem_init_fast(pd, max_num_sg, mr);
pd               1000 drivers/infiniband/sw/rxe/rxe_verbs.c 	rxe_drop_ref(pd);
pd                 76 drivers/infiniband/sw/rxe/rxe_verbs.h 	struct rxe_pd		*pd;
pd                125 drivers/infiniband/sw/rxe/rxe_verbs.h 	struct rxe_pd		*pd;
pd                243 drivers/infiniband/sw/rxe/rxe_verbs.h 	struct rxe_pd		*pd;
pd                325 drivers/infiniband/sw/rxe/rxe_verbs.h 	struct rxe_pd		*pd;
pd                433 drivers/infiniband/sw/rxe/rxe_verbs.h static inline struct rxe_pd *to_rpd(struct ib_pd *pd)
pd                435 drivers/infiniband/sw/rxe/rxe_verbs.h 	return pd ? container_of(pd, struct rxe_pd, ibpd) : NULL;
pd                173 drivers/infiniband/sw/siw/siw.h 	struct ib_pd *pd;
pd                443 drivers/infiniband/sw/siw/siw.h 	struct ib_pd *pd;
pd                728 drivers/infiniband/sw/siw/siw.h #define siw_dbg_pd(pd, fmt, ...)                                               \
pd                729 drivers/infiniband/sw/siw/siw.h 	ibdev_dbg(pd->device, "PD[%u] %s: " fmt, pd->res.id, __func__,         \
pd                 89 drivers/infiniband/sw/siw/siw_mem.c int siw_mr_add_mem(struct siw_mr *mr, struct ib_pd *pd, void *mem_obj,
pd                 92 drivers/infiniband/sw/siw/siw_mem.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                105 drivers/infiniband/sw/siw/siw_mem.c 	mem->pd = pd;
pd                168 drivers/infiniband/sw/siw/siw_mem.c int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr,
pd                172 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "STag 0x%08x invalid\n", mem->stag);
pd                175 drivers/infiniband/sw/siw/siw_mem.c 	if (mem->pd != pd) {
pd                176 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "STag 0x%08x: PD mismatch\n", mem->stag);
pd                183 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "permissions 0x%08x < 0x%08x\n",
pd                191 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "MEM interval len %d\n", len);
pd                192 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "[0x%pK, 0x%pK] out of bounds\n",
pd                195 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "[0x%pK, 0x%pK] STag=0x%08x\n",
pd                222 drivers/infiniband/sw/siw/siw_mem.c int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[],
pd                225 drivers/infiniband/sw/siw/siw_mem.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                236 drivers/infiniband/sw/siw/siw_mem.c 			siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey);
pd                248 drivers/infiniband/sw/siw/siw_mem.c 	rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len);
pd                292 drivers/infiniband/sw/siw/siw_mem.c int siw_invalidate_stag(struct ib_pd *pd, u32 stag)
pd                294 drivers/infiniband/sw/siw/siw_mem.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                299 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "STag 0x%08x unknown\n", stag);
pd                302 drivers/infiniband/sw/siw/siw_mem.c 	if (unlikely(mem->pd != pd)) {
pd                303 drivers/infiniband/sw/siw/siw_mem.c 		siw_dbg_pd(pd, "PD mismatch for STag 0x%08x\n", stag);
pd                313 drivers/infiniband/sw/siw/siw_mem.c 	siw_dbg_pd(pd, "STag 0x%08x now invalid\n", stag);
pd                 15 drivers/infiniband/sw/siw/siw_mem.h int siw_invalidate_stag(struct ib_pd *pd, u32 stag);
pd                 16 drivers/infiniband/sw/siw/siw_mem.h int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr,
pd                 18 drivers/infiniband/sw/siw/siw_mem.h int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge,
pd                 22 drivers/infiniband/sw/siw/siw_mem.h int siw_mr_add_mem(struct siw_mr *mr, struct ib_pd *pd, void *mem_obj,
pd                454 drivers/infiniband/sw/siw/siw_qp_rx.c 		struct ib_pd *pd;
pd                474 drivers/infiniband/sw/siw/siw_qp_rx.c 		pd = qp->srq == NULL ? qp->pd : qp->srq->base_srq.pd;
pd                476 drivers/infiniband/sw/siw/siw_qp_rx.c 		rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE,
pd                588 drivers/infiniband/sw/siw/siw_qp_rx.c 	rv = siw_check_mem(qp->pd, mem, srx->ddp_to + srx->fpdu_part_rcvd,
pd                825 drivers/infiniband/sw/siw/siw_qp_rx.c 		rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0,
pd               1234 drivers/infiniband/sw/siw/siw_qp_rx.c 			rv = siw_invalidate_stag(qp->pd, srx->inval_stag);
pd               1272 drivers/infiniband/sw/siw/siw_qp_rx.c 			rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey);
pd                751 drivers/infiniband/sw/siw/siw_qp_tx.c static int siw_check_sgl_tx(struct ib_pd *pd, struct siw_wqe *wqe,
pd                765 drivers/infiniband/sw/siw/siw_qp_tx.c 			int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0,
pd                806 drivers/infiniband/sw/siw/siw_qp_tx.c 				rv = siw_check_sgl_tx(qp->pd, wqe, 0);
pd                919 drivers/infiniband/sw/siw/siw_qp_tx.c static int siw_fastreg_mr(struct ib_pd *pd, struct siw_sqe *sqe)
pd                922 drivers/infiniband/sw/siw/siw_qp_tx.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                926 drivers/infiniband/sw/siw/siw_qp_tx.c 	siw_dbg_pd(pd, "STag 0x%08x\n", sqe->rkey);
pd                944 drivers/infiniband/sw/siw/siw_qp_tx.c 	if (unlikely(mem->pd != pd)) {
pd                972 drivers/infiniband/sw/siw/siw_qp_tx.c 		rv = siw_fastreg_mr(qp->pd, &wqe->sqe);
pd                976 drivers/infiniband/sw/siw/siw_qp_tx.c 		rv = siw_invalidate_stag(qp->pd, wqe->sqe.rkey);
pd                265 drivers/infiniband/sw/siw/siw_verbs.c int siw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                267 drivers/infiniband/sw/siw/siw_verbs.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                273 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "now %d PD's(s)\n", atomic_read(&sdev->num_pd));
pd                278 drivers/infiniband/sw/siw/siw_verbs.c void siw_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata)
pd                280 drivers/infiniband/sw/siw/siw_verbs.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd                282 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "free PD\n");
pd                306 drivers/infiniband/sw/siw/siw_verbs.c struct ib_qp *siw_create_qp(struct ib_pd *pd,
pd                312 drivers/infiniband/sw/siw/siw_verbs.c 	struct ib_device *base_dev = pd->device;
pd                416 drivers/infiniband/sw/siw/siw_verbs.c 	qp->pd = pd;
pd               1295 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_reg_user_mr(struct ib_pd *pd, u64 start, u64 len,
pd               1301 drivers/infiniband/sw/siw/siw_verbs.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd               1306 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "start: 0x%pK, va: 0x%pK, len: %llu\n",
pd               1311 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "too many mr's\n");
pd               1325 drivers/infiniband/sw/siw/siw_verbs.c 			siw_dbg_pd(pd, "pages req %lu, max %lu, lock %lu\n",
pd               1335 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "getting user memory failed: %d\n", rv);
pd               1344 drivers/infiniband/sw/siw/siw_verbs.c 	rv = siw_mr_add_mem(mr, pd, umem, start, len, rights);
pd               1390 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               1393 drivers/infiniband/sw/siw/siw_verbs.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd               1399 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "too many mr's\n");
pd               1404 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "mr type %d unsupported\n", mr_type);
pd               1409 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "too many sge's: %d\n", max_sge);
pd               1416 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "pbl allocation failed: %d\n", rv);
pd               1425 drivers/infiniband/sw/siw/siw_verbs.c 	rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0);
pd               1431 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
pd               1445 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "failed: %d\n", rv);
pd               1524 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_get_dma_mr(struct ib_pd *pd, int rights)
pd               1526 drivers/infiniband/sw/siw/siw_verbs.c 	struct siw_device *sdev = to_siw_dev(pd->device);
pd               1531 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(pd, "too many mr's\n");
pd               1540 drivers/infiniband/sw/siw/siw_verbs.c 	rv = siw_mr_add_mem(mr, pd, NULL, 0, ULONG_MAX, rights);
pd               1546 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag);
pd               1581 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(base_srq->pd, "too many SRQ's\n");
pd               1632 drivers/infiniband/sw/siw/siw_verbs.c 	siw_dbg_pd(base_srq->pd, "[SRQ]: success\n");
pd               1752 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(base_srq->pd,
pd               1769 drivers/infiniband/sw/siw/siw_verbs.c 			siw_dbg_pd(base_srq->pd, "SRQ full\n");
pd               1774 drivers/infiniband/sw/siw/siw_verbs.c 			siw_dbg_pd(base_srq->pd,
pd               1794 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(base_srq->pd, "[SRQ]: error %d\n", rv);
pd               1847 drivers/infiniband/sw/siw/siw_verbs.c 		siw_dbg_pd(srq->base_srq.pd,
pd                364 drivers/infiniband/ulp/ipoib/ipoib.h 	struct ib_pd	 *pd;
pd                487 drivers/infiniband/ulp/ipoib/ipoib.h 				 struct ib_pd *pd, struct rdma_ah_attr *attr);
pd                270 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	return ib_create_qp(priv->pd, &attr);
pd                336 drivers/infiniband/ulp/ipoib/ipoib_cm.c 		sge[i].lkey = priv->pd->local_dma_lkey;
pd               1077 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	tx_qp = ib_create_qp(priv->pd, &attr);
pd               1567 drivers/infiniband/ulp/ipoib/ipoib_cm.c 	priv->cm.srq = ib_create_srq(priv->pd, &srq_init_attr);
pd                 56 drivers/infiniband/ulp/ipoib/ipoib_ib.c 				 struct ib_pd *pd, struct rdma_ah_attr *attr)
pd                 69 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	vah = rdma_create_ah(pd, attr, RDMA_CREATE_AH_SLEEPABLE);
pd               1316 drivers/infiniband/ulp/ipoib/ipoib_ib.c 	if (priv->pd) {
pd               1317 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		ib_dealloc_pd(priv->pd);
pd               1318 drivers/infiniband/ulp/ipoib/ipoib_ib.c 		priv->pd = NULL;
pd                770 drivers/infiniband/ulp/ipoib/ipoib_main.c 			ah = ipoib_create_ah(dev, priv->pd, &av);
pd               1765 drivers/infiniband/ulp/ipoib/ipoib_main.c 	priv->pd = ib_alloc_pd(priv->ca, 0);
pd               1766 drivers/infiniband/ulp/ipoib/ipoib_main.c 	if (IS_ERR(priv->pd)) {
pd               1800 drivers/infiniband/ulp/ipoib/ipoib_main.c 	if (priv->pd) {
pd               1801 drivers/infiniband/ulp/ipoib/ipoib_main.c 		ib_dealloc_pd(priv->pd);
pd               1802 drivers/infiniband/ulp/ipoib/ipoib_main.c 		priv->pd = NULL;
pd                289 drivers/infiniband/ulp/ipoib/ipoib_multicast.c 	ah = ipoib_create_ah(dev, priv->pd, &av);
pd                209 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->qp = ib_create_qp(priv->pd, &init_attr);
pd                219 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 		priv->tx_sge[i].lkey = priv->pd->local_dma_lkey;
pd                225 drivers/infiniband/ulp/ipoib/ipoib_verbs.c 	priv->rx_sge[0].lkey = priv->pd->local_dma_lkey;
pd                214 drivers/infiniband/ulp/iser/iscsi_iser.c 	tx_desc->tx_sg[0].lkey   = device->pd->local_dma_lkey;
pd                361 drivers/infiniband/ulp/iser/iscsi_iser.h 	struct ib_pd	             *pd;
pd                280 drivers/infiniband/ulp/iser/iser_initiator.c 		rx_sg->lkey = device->pd->local_dma_lkey;
pd                534 drivers/infiniband/ulp/iser/iser_initiator.c 		tx_dsg->lkey = device->pd->local_dma_lkey;
pd                196 drivers/infiniband/ulp/iser/iser_memory.c 	reg->sge.lkey = device->pd->local_dma_lkey;
pd                202 drivers/infiniband/ulp/iser/iser_memory.c 	if (device->pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY)
pd                203 drivers/infiniband/ulp/iser/iser_memory.c 		reg->rkey = device->pd->unsafe_global_rkey;
pd                 91 drivers/infiniband/ulp/iser/iser_verbs.c 	device->pd = ib_alloc_pd(ib_dev,
pd                 93 drivers/infiniband/ulp/iser/iser_verbs.c 	if (IS_ERR(device->pd))
pd                119 drivers/infiniband/ulp/iser/iser_verbs.c 	ib_dealloc_pd(device->pd);
pd                143 drivers/infiniband/ulp/iser/iser_verbs.c 	ib_dealloc_pd(device->pd);
pd                147 drivers/infiniband/ulp/iser/iser_verbs.c 	device->pd = NULL;
pd                195 drivers/infiniband/ulp/iser/iser_verbs.c 	fmr_pool = ib_create_fmr_pool(device->pd, &params);
pd                238 drivers/infiniband/ulp/iser/iser_verbs.c 			 struct ib_pd *pd,
pd                256 drivers/infiniband/ulp/iser/iser_verbs.c 	desc->rsc.mr = ib_alloc_mr(pd, mr_type, size);
pd                264 drivers/infiniband/ulp/iser/iser_verbs.c 		desc->rsc.sig_mr = ib_alloc_mr_integrity(pd, size, size);
pd                314 drivers/infiniband/ulp/iser/iser_verbs.c 		desc = iser_create_fastreg_desc(device, device->pd,
pd                420 drivers/infiniband/ulp/iser/iser_verbs.c 	ret = rdma_create_qp(ib_conn->cma_id, device->pd, &init_attr);
pd                981 drivers/infiniband/ulp/iser/iser_verbs.c 	desc->sge.lkey = ib_conn->device->pd->local_dma_lkey;
pd                138 drivers/infiniband/ulp/isert/ib_isert.c 	ret = rdma_create_qp(cma_id, device->pd, &attr);
pd                195 drivers/infiniband/ulp/isert/ib_isert.c 		rx_sg->lkey = device->pd->local_dma_lkey;
pd                302 drivers/infiniband/ulp/isert/ib_isert.c 	device->pd = ib_alloc_pd(ib_dev, 0);
pd                303 drivers/infiniband/ulp/isert/ib_isert.c 	if (IS_ERR(device->pd)) {
pd                304 drivers/infiniband/ulp/isert/ib_isert.c 		ret = PTR_ERR(device->pd);
pd                329 drivers/infiniband/ulp/isert/ib_isert.c 	ib_dealloc_pd(device->pd);
pd                891 drivers/infiniband/ulp/isert/ib_isert.c 	if (tx_desc->tx_sg[0].lkey != device->pd->local_dma_lkey) {
pd                892 drivers/infiniband/ulp/isert/ib_isert.c 		tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey;
pd                929 drivers/infiniband/ulp/isert/ib_isert.c 	tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey;
pd                969 drivers/infiniband/ulp/isert/ib_isert.c 	sge.lkey = isert_conn->device->pd->local_dma_lkey;
pd               1018 drivers/infiniband/ulp/isert/ib_isert.c 		tx_dsg->lkey	= isert_conn->device->pd->local_dma_lkey;
pd               1880 drivers/infiniband/ulp/isert/ib_isert.c 		tx_dsg->lkey	= device->pd->local_dma_lkey;
pd               2008 drivers/infiniband/ulp/isert/ib_isert.c 	tx_dsg->lkey	= device->pd->local_dma_lkey;
pd               2051 drivers/infiniband/ulp/isert/ib_isert.c 		tx_dsg->lkey	= device->pd->local_dma_lkey;
pd                188 drivers/infiniband/ulp/isert/ib_isert.h 	struct ib_pd		*pd;
pd                636 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c 	ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc,
pd                780 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c 	ah = rdma_create_ah(port->mad_agent->qp->pd, &ah_attr, 0);
pd                406 drivers/infiniband/ulp/srp/ib_srp.c 	return ib_create_fmr_pool(dev->pd, &fmr_param);
pd                436 drivers/infiniband/ulp/srp/ib_srp.c 					      struct ib_pd *pd, int pool_size,
pd                462 drivers/infiniband/ulp/srp/ib_srp.c 		mr = ib_alloc_mr(pd, mr_type, max_page_list_len);
pd                529 drivers/infiniband/ulp/srp/ib_srp.c 	return srp_create_fr_pool(dev->dev, dev->pd, target->mr_pool_size,
pd                593 drivers/infiniband/ulp/srp/ib_srp.c 		ret = rdma_create_qp(ch->rdma_cm.cm_id, dev->pd, init_attr);
pd                596 drivers/infiniband/ulp/srp/ib_srp.c 		qp = ib_create_qp(dev->pd, init_attr);
pd               3804 drivers/infiniband/ulp/srp/ib_srp.c 	target->lkey		= host->srp_dev->pd->local_dma_lkey;
pd               4176 drivers/infiniband/ulp/srp/ib_srp.c 	srp_dev->pd  = ib_alloc_pd(device, flags);
pd               4177 drivers/infiniband/ulp/srp/ib_srp.c 	if (IS_ERR(srp_dev->pd))
pd               4181 drivers/infiniband/ulp/srp/ib_srp.c 		srp_dev->global_rkey = srp_dev->pd->unsafe_global_rkey;
pd               4233 drivers/infiniband/ulp/srp/ib_srp.c 	ib_dealloc_pd(srp_dev->pd);
pd                104 drivers/infiniband/ulp/srp/ib_srp.h 	struct ib_pd	       *pd;
pd                483 drivers/infiniband/ulp/srpt/ib_srpt.c 	ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc,
pd               1833 drivers/infiniband/ulp/srpt/ib_srpt.c 		ret = rdma_create_qp(ch->rdma_cm.cm_id, sdev->pd, qp_init);
pd               1836 drivers/infiniband/ulp/srpt/ib_srpt.c 		ch->qp = ib_create_qp(sdev->pd, qp_init);
pd               3050 drivers/infiniband/ulp/srpt/ib_srpt.c 	srq = ib_create_srq(sdev->pd, &srq_attr);
pd               3125 drivers/infiniband/ulp/srpt/ib_srpt.c 	sdev->pd = ib_alloc_pd(device, 0);
pd               3126 drivers/infiniband/ulp/srpt/ib_srpt.c 	if (IS_ERR(sdev->pd))
pd               3129 drivers/infiniband/ulp/srpt/ib_srpt.c 	sdev->lkey = sdev->pd->local_dma_lkey;
pd               3207 drivers/infiniband/ulp/srpt/ib_srpt.c 	ib_dealloc_pd(sdev->pd);
pd               3259 drivers/infiniband/ulp/srpt/ib_srpt.c 	ib_dealloc_pd(sdev->pd);
pd                424 drivers/infiniband/ulp/srpt/ib_srpt.h 	struct ib_pd		*pd;
pd                 90 drivers/input/joystick/db9.c 	struct pardevice *pd;
pd                358 drivers/input/joystick/db9.c 	struct parport *port = db9->pd->port;
pd                510 drivers/input/joystick/db9.c 	struct parport *port = db9->pd->port;
pd                518 drivers/input/joystick/db9.c 		parport_claim(db9->pd);
pd                534 drivers/input/joystick/db9.c 	struct parport *port = db9->pd->port;
pd                541 drivers/input/joystick/db9.c 		parport_release(db9->pd);
pd                550 drivers/input/joystick/db9.c 	struct pardevice *pd;
pd                587 drivers/input/joystick/db9.c 	pd = parport_register_dev_model(pp, "db9", &db9_parport_cb, port_idx);
pd                588 drivers/input/joystick/db9.c 	if (!pd) {
pd                598 drivers/input/joystick/db9.c 	db9->pd = pd;
pd                612 drivers/input/joystick/db9.c 			 "%s/input%d", db9->pd->port->name, i);
pd                650 drivers/input/joystick/db9.c 	parport_unregister_device(pd);
pd                671 drivers/input/joystick/db9.c 	parport_unregister_device(db9->pd);
pd                 74 drivers/input/joystick/gamecon.c 	struct pardevice *pd;
pd                136 drivers/input/joystick/gamecon.c 	struct parport *port = gc->pd->port;
pd                149 drivers/input/joystick/gamecon.c 	struct parport *port = gc->pd->port;
pd                191 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, GC_N64_POWER_R);
pd                193 drivers/input/joystick/gamecon.c 		data[i] = parport_read_status(gc->pd->port);
pd                194 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, GC_N64_POWER_R | GC_N64_CLOCK);
pd                345 drivers/input/joystick/gamecon.c 	parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK | GC_NES_LATCH);
pd                347 drivers/input/joystick/gamecon.c 	parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK);
pd                351 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, GC_NES_POWER);
pd                352 drivers/input/joystick/gamecon.c 		data[i] = parport_read_status(gc->pd->port) ^ 0x7f;
pd                354 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK);
pd                465 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, ~(1 << i));
pd                466 drivers/input/joystick/gamecon.c 		data[i] = parport_read_status(gc->pd->port) ^ 0x7f;
pd                552 drivers/input/joystick/gamecon.c 	struct parport *port = gc->pd->port;
pd                571 drivers/input/joystick/gamecon.c 		parport_write_data(gc->pd->port, cmd | GC_PSX_CLOCK | GC_PSX_POWER);
pd                590 drivers/input/joystick/gamecon.c 	parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_SELECT | GC_PSX_POWER);
pd                593 drivers/input/joystick/gamecon.c 	parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_POWER);
pd                622 drivers/input/joystick/gamecon.c 	parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_SELECT | GC_PSX_POWER);
pd                778 drivers/input/joystick/gamecon.c 		parport_claim(gc->pd);
pd                779 drivers/input/joystick/gamecon.c 		parport_write_control(gc->pd->port, 0x04);
pd                794 drivers/input/joystick/gamecon.c 		parport_write_control(gc->pd->port, 0x00);
pd                795 drivers/input/joystick/gamecon.c 		parport_release(gc->pd);
pd                821 drivers/input/joystick/gamecon.c 		 "%s/input%d", gc->pd->port->name, idx);
pd                924 drivers/input/joystick/gamecon.c 	struct pardevice *pd;
pd                948 drivers/input/joystick/gamecon.c 	pd = parport_register_dev_model(pp, "gamecon", &gc_parport_cb,
pd                950 drivers/input/joystick/gamecon.c 	if (!pd) {
pd                962 drivers/input/joystick/gamecon.c 	gc->pd = pd;
pd                991 drivers/input/joystick/gamecon.c 	parport_unregister_device(pd);
pd               1013 drivers/input/joystick/gamecon.c 	parport_unregister_device(gc->pd);
pd                 61 drivers/input/joystick/turbografx.c 	struct pardevice *pd;
pd                 87 drivers/input/joystick/turbografx.c 			parport_write_data(tgfx->pd->port, ~(1 << i));
pd                 88 drivers/input/joystick/turbografx.c 			data1 = parport_read_status(tgfx->pd->port) ^ 0x7f;
pd                 89 drivers/input/joystick/turbografx.c 			data2 = parport_read_control(tgfx->pd->port) ^ 0x04;	/* CAVEAT parport */
pd                116 drivers/input/joystick/turbografx.c 		parport_claim(tgfx->pd);
pd                117 drivers/input/joystick/turbografx.c 		parport_write_control(tgfx->pd->port, 0x04);
pd                132 drivers/input/joystick/turbografx.c 		parport_write_control(tgfx->pd->port, 0x00);
pd                133 drivers/input/joystick/turbografx.c 		parport_release(tgfx->pd);
pd                148 drivers/input/joystick/turbografx.c 	struct pardevice *pd;
pd                171 drivers/input/joystick/turbografx.c 	pd = parport_register_dev_model(pp, "turbografx", &tgfx_parport_cb,
pd                173 drivers/input/joystick/turbografx.c 	if (!pd) {
pd                185 drivers/input/joystick/turbografx.c 	tgfx->pd = pd;
pd                208 drivers/input/joystick/turbografx.c 			 "%s/input%d", tgfx->pd->port->name, i);
pd                250 drivers/input/joystick/turbografx.c 	parport_unregister_device(pd);
pd                272 drivers/input/joystick/turbografx.c 	parport_unregister_device(tgfx->pd);
pd                 86 drivers/input/keyboard/snvs_pwrkey.c 	struct pwrkey_drv_data *pd = pdata;
pd                 88 drivers/input/keyboard/snvs_pwrkey.c 	del_timer_sync(&pd->check_timer);
pd                 69 drivers/input/misc/soc_button_array.c 	struct platform_device *pd;
pd                130 drivers/input/misc/soc_button_array.c 	pd = platform_device_register_resndata(&pdev->dev, "gpio-keys",
pd                134 drivers/input/misc/soc_button_array.c 	error = PTR_ERR_OR_ZERO(pd);
pd                141 drivers/input/misc/soc_button_array.c 	return pd;
pd                326 drivers/input/misc/soc_button_array.c 	struct platform_device *pd;
pd                358 drivers/input/misc/soc_button_array.c 		pd = soc_button_device_create(pdev, button_info, i == 0);
pd                359 drivers/input/misc/soc_button_array.c 		if (IS_ERR(pd)) {
pd                360 drivers/input/misc/soc_button_array.c 			error = PTR_ERR(pd);
pd                368 drivers/input/misc/soc_button_array.c 		priv->children[i] = pd;
pd                 54 drivers/iommu/tegra-smmu.c 	struct page *pd;
pd                290 drivers/iommu/tegra-smmu.c 	as->pd = alloc_page(GFP_KERNEL | __GFP_DMA | __GFP_ZERO);
pd                291 drivers/iommu/tegra-smmu.c 	if (!as->pd) {
pd                298 drivers/iommu/tegra-smmu.c 		__free_page(as->pd);
pd                306 drivers/iommu/tegra-smmu.c 		__free_page(as->pd);
pd                414 drivers/iommu/tegra-smmu.c 	as->pd_dma = dma_map_page(smmu->dev, as->pd, 0, SMMU_SIZE_PD,
pd                525 drivers/iommu/tegra-smmu.c 	u32 *pd = page_address(as->pd);
pd                526 drivers/iommu/tegra-smmu.c 	unsigned long offset = pd_index * sizeof(*pd);
pd                529 drivers/iommu/tegra-smmu.c 	pd[pd_index] = value;
pd                533 drivers/iommu/tegra-smmu.c 					 sizeof(*pd), DMA_TO_DEVICE);
pd                554 drivers/iommu/tegra-smmu.c 	u32 *pd;
pd                560 drivers/iommu/tegra-smmu.c 	pd = page_address(as->pd);
pd                561 drivers/iommu/tegra-smmu.c 	*dmap = smmu_pde_to_dma(smmu, pd[pd_index]);
pd                601 drivers/iommu/tegra-smmu.c 		u32 *pd = page_address(as->pd);
pd                603 drivers/iommu/tegra-smmu.c 		*dmap = smmu_pde_to_dma(smmu, pd[pde]);
pd                627 drivers/iommu/tegra-smmu.c 		u32 *pd = page_address(as->pd);
pd                628 drivers/iommu/tegra-smmu.c 		dma_addr_t pte_dma = smmu_pde_to_dma(smmu, pd[pde]);
pd                 38 drivers/leds/leds-s3c24xx.c 	struct s3c24xx_led_platdata *pd = led->pdata;
pd                 39 drivers/leds/leds-s3c24xx.c 	int state = (value ? 1 : 0) ^ (pd->flags & S3C24XX_LEDF_ACTLOW);
pd                 44 drivers/leds/leds-s3c24xx.c 	gpio_set_value(pd->gpio, state);
pd                 46 drivers/leds/leds-s3c24xx.c 	if (pd->flags & S3C24XX_LEDF_TRISTATE) {
pd                 48 drivers/leds/leds-s3c24xx.c 			gpio_direction_output(pd->gpio, state);
pd                 50 drivers/leds/leds-s3c24xx.c 			gpio_direction_input(pd->gpio);
pd                415 drivers/media/i2c/adp1653.c 	struct adp1653_platform_data *pd;
pd                418 drivers/media/i2c/adp1653.c 	pd = devm_kzalloc(&client->dev, sizeof(*pd), GFP_KERNEL);
pd                419 drivers/media/i2c/adp1653.c 	if (!pd)
pd                421 drivers/media/i2c/adp1653.c 	flash->platform_data = pd;
pd                428 drivers/media/i2c/adp1653.c 				 &pd->max_flash_timeout))
pd                432 drivers/media/i2c/adp1653.c 				 &pd->max_flash_intensity))
pd                435 drivers/media/i2c/adp1653.c 	pd->max_flash_intensity /= 1000;
pd                438 drivers/media/i2c/adp1653.c 				 &pd->max_torch_intensity))
pd                441 drivers/media/i2c/adp1653.c 	pd->max_torch_intensity /= 1000;
pd                449 drivers/media/i2c/adp1653.c 				 &pd->max_indicator_intensity))
pd                454 drivers/media/i2c/adp1653.c 	pd->enable_gpio = devm_gpiod_get(&client->dev, "enable", GPIOD_OUT_LOW);
pd                455 drivers/media/i2c/adp1653.c 	if (IS_ERR(pd->enable_gpio)) {
pd                457 drivers/media/i2c/adp1653.c 		return PTR_ERR(pd->enable_gpio);
pd                128 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = vb2_get_drv_priv(vq);
pd                129 drivers/media/pci/dt3155/dt3155.c 	unsigned size = pd->width * pd->height;
pd                142 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = vb2_get_drv_priv(vb->vb2_queue);
pd                144 drivers/media/pci/dt3155/dt3155.c 	vb2_set_plane_payload(vb, 0, pd->width * pd->height);
pd                150 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = vb2_get_drv_priv(q);
pd                151 drivers/media/pci/dt3155/dt3155.c 	struct vb2_buffer *vb = &pd->curr_buf->vb2_buf;
pd                154 drivers/media/pci/dt3155/dt3155.c 	pd->sequence = 0;
pd                156 drivers/media/pci/dt3155/dt3155.c 	iowrite32(dma_addr, pd->regs + EVEN_DMA_START);
pd                157 drivers/media/pci/dt3155/dt3155.c 	iowrite32(dma_addr + pd->width, pd->regs + ODD_DMA_START);
pd                158 drivers/media/pci/dt3155/dt3155.c 	iowrite32(pd->width, pd->regs + EVEN_DMA_STRIDE);
pd                159 drivers/media/pci/dt3155/dt3155.c 	iowrite32(pd->width, pd->regs + ODD_DMA_STRIDE);
pd                162 drivers/media/pci/dt3155/dt3155.c 			FLD_END_EVEN | FLD_END_ODD, pd->regs + INT_CSR);
pd                165 drivers/media/pci/dt3155/dt3155.c 							pd->regs + CSR1);
pd                166 drivers/media/pci/dt3155/dt3155.c 	wait_i2c_reg(pd->regs);
pd                167 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, CONFIG, pd->config);
pd                168 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, EVEN_CSR, CSR_ERROR | CSR_DONE);
pd                169 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, ODD_CSR, CSR_ERROR | CSR_DONE);
pd                172 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, CSR2, pd->csr2 | BUSY_EVEN | BUSY_ODD);
pd                178 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = vb2_get_drv_priv(q);
pd                181 drivers/media/pci/dt3155/dt3155.c 	spin_lock_irq(&pd->lock);
pd                183 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg_nowait(pd->regs, CSR2, pd->csr2);
pd                185 drivers/media/pci/dt3155/dt3155.c 		  FLD_DN_ODD | FLD_DN_EVEN, pd->regs + CSR1);
pd                187 drivers/media/pci/dt3155/dt3155.c 	iowrite32(FLD_START | FLD_END_EVEN | FLD_END_ODD, pd->regs + INT_CSR);
pd                188 drivers/media/pci/dt3155/dt3155.c 	spin_unlock_irq(&pd->lock);
pd                197 drivers/media/pci/dt3155/dt3155.c 	spin_lock_irq(&pd->lock);
pd                198 drivers/media/pci/dt3155/dt3155.c 	if (pd->curr_buf) {
pd                199 drivers/media/pci/dt3155/dt3155.c 		vb2_buffer_done(&pd->curr_buf->vb2_buf, VB2_BUF_STATE_ERROR);
pd                200 drivers/media/pci/dt3155/dt3155.c 		pd->curr_buf = NULL;
pd                203 drivers/media/pci/dt3155/dt3155.c 	while (!list_empty(&pd->dmaq)) {
pd                204 drivers/media/pci/dt3155/dt3155.c 		vb = list_first_entry(&pd->dmaq, typeof(*vb), done_entry);
pd                208 drivers/media/pci/dt3155/dt3155.c 	spin_unlock_irq(&pd->lock);
pd                214 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = vb2_get_drv_priv(vb->vb2_queue);
pd                217 drivers/media/pci/dt3155/dt3155.c 	spin_lock_irq(&pd->lock);
pd                218 drivers/media/pci/dt3155/dt3155.c 	if (pd->curr_buf)
pd                219 drivers/media/pci/dt3155/dt3155.c 		list_add_tail(&vb->done_entry, &pd->dmaq);
pd                221 drivers/media/pci/dt3155/dt3155.c 		pd->curr_buf = vbuf;
pd                222 drivers/media/pci/dt3155/dt3155.c 	spin_unlock_irq(&pd->lock);
pd                295 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                299 drivers/media/pci/dt3155/dt3155.c 	sprintf(cap->bus_info, "PCI:%s", pci_name(pd->pdev));
pd                314 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                316 drivers/media/pci/dt3155/dt3155.c 	f->fmt.pix.width = pd->width;
pd                317 drivers/media/pci/dt3155/dt3155.c 	f->fmt.pix.height = pd->height;
pd                328 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                330 drivers/media/pci/dt3155/dt3155.c 	*norm = pd->std;
pd                336 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                338 drivers/media/pci/dt3155/dt3155.c 	if (pd->std == norm)
pd                340 drivers/media/pci/dt3155/dt3155.c 	if (vb2_is_busy(&pd->vidq))
pd                342 drivers/media/pci/dt3155/dt3155.c 	pd->std = norm;
pd                343 drivers/media/pci/dt3155/dt3155.c 	if (pd->std & V4L2_STD_525_60) {
pd                344 drivers/media/pci/dt3155/dt3155.c 		pd->csr2 = VT_60HZ;
pd                345 drivers/media/pci/dt3155/dt3155.c 		pd->width = 640;
pd                346 drivers/media/pci/dt3155/dt3155.c 		pd->height = 480;
pd                348 drivers/media/pci/dt3155/dt3155.c 		pd->csr2 = VT_50HZ;
pd                349 drivers/media/pci/dt3155/dt3155.c 		pd->width = 768;
pd                350 drivers/media/pci/dt3155/dt3155.c 		pd->height = 576;
pd                373 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                375 drivers/media/pci/dt3155/dt3155.c 	*i = pd->input;
pd                381 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = video_drvdata(filp);
pd                385 drivers/media/pci/dt3155/dt3155.c 	pd->input = i;
pd                386 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG);
pd                387 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_CMD, (i << 6) | (i << 4) | SYNC_LVL_3);
pd                412 drivers/media/pci/dt3155/dt3155.c static int dt3155_init_board(struct dt3155_priv *pd)
pd                414 drivers/media/pci/dt3155/dt3155.c 	struct pci_dev *pdev = pd->pdev;
pd                422 drivers/media/pci/dt3155/dt3155.c 			FLD_DN_ODD | FLD_DN_EVEN, pd->regs + CSR1);
pd                426 drivers/media/pci/dt3155/dt3155.c 	iowrite32(FIFO_EN | SRST, pd->regs + CSR1);
pd                427 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0xEEEEEE01, pd->regs + EVEN_PIXEL_FMT);
pd                428 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0xEEEEEE01, pd->regs + ODD_PIXEL_FMT);
pd                429 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0x00000020, pd->regs + FIFO_TRIGER);
pd                430 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0x00000103, pd->regs + XFER_MODE);
pd                431 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0, pd->regs + RETRY_WAIT_CNT);
pd                432 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0, pd->regs + INT_CSR);
pd                433 drivers/media/pci/dt3155/dt3155.c 	iowrite32(1, pd->regs + EVEN_FLD_MASK);
pd                434 drivers/media/pci/dt3155/dt3155.c 	iowrite32(1, pd->regs + ODD_FLD_MASK);
pd                435 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0, pd->regs + MASK_LENGTH);
pd                436 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0x0005007C, pd->regs + FIFO_FLAG_CNT);
pd                437 drivers/media/pci/dt3155/dt3155.c 	iowrite32(0x01010101, pd->regs + IIC_CLK_DUR);
pd                440 drivers/media/pci/dt3155/dt3155.c 	read_i2c_reg(pd->regs, DT_ID, &tmp);
pd                445 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, 0);
pd                447 drivers/media/pci/dt3155/dt3155.c 		write_i2c_reg(pd->regs, AD_LUT, i);
pd                451 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG);
pd                452 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_CMD, VIDEO_CNL_1 | SYNC_CNL_1 | SYNC_LVL_3);
pd                453 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, AD_POS_REF);
pd                454 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_CMD, 34);
pd                455 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, AD_NEG_REF);
pd                456 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_CMD, 0);
pd                459 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, CONFIG, pd->config | PM_LUT_PGM);
pd                461 drivers/media/pci/dt3155/dt3155.c 		write_i2c_reg(pd->regs, PM_LUT_ADDR, i);
pd                462 drivers/media/pci/dt3155/dt3155.c 		write_i2c_reg(pd->regs, PM_LUT_DATA, i);
pd                464 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, CONFIG, pd->config | PM_LUT_PGM | PM_LUT_SEL);
pd                466 drivers/media/pci/dt3155/dt3155.c 		write_i2c_reg(pd->regs, PM_LUT_ADDR, i);
pd                467 drivers/media/pci/dt3155/dt3155.c 		write_i2c_reg(pd->regs, PM_LUT_DATA, i);
pd                469 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, CONFIG, pd->config); /*  ACQ_MODE_EVEN  */
pd                472 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG);
pd                473 drivers/media/pci/dt3155/dt3155.c 	write_i2c_reg(pd->regs, AD_CMD, VIDEO_CNL_1 | SYNC_CNL_1 | SYNC_LVL_3);
pd                477 drivers/media/pci/dt3155/dt3155.c 			pd->regs + INT_CSR);
pd                496 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd;
pd                501 drivers/media/pci/dt3155/dt3155.c 	pd = devm_kzalloc(&pdev->dev, sizeof(*pd), GFP_KERNEL);
pd                502 drivers/media/pci/dt3155/dt3155.c 	if (!pd)
pd                505 drivers/media/pci/dt3155/dt3155.c 	err = v4l2_device_register(&pdev->dev, &pd->v4l2_dev);
pd                508 drivers/media/pci/dt3155/dt3155.c 	pd->vdev = dt3155_vdev;
pd                509 drivers/media/pci/dt3155/dt3155.c 	pd->vdev.v4l2_dev = &pd->v4l2_dev;
pd                510 drivers/media/pci/dt3155/dt3155.c 	video_set_drvdata(&pd->vdev, pd);  /* for use in video_fops */
pd                511 drivers/media/pci/dt3155/dt3155.c 	pd->pdev = pdev;
pd                512 drivers/media/pci/dt3155/dt3155.c 	pd->std = V4L2_STD_625_50;
pd                513 drivers/media/pci/dt3155/dt3155.c 	pd->csr2 = VT_50HZ;
pd                514 drivers/media/pci/dt3155/dt3155.c 	pd->width = 768;
pd                515 drivers/media/pci/dt3155/dt3155.c 	pd->height = 576;
pd                516 drivers/media/pci/dt3155/dt3155.c 	INIT_LIST_HEAD(&pd->dmaq);
pd                517 drivers/media/pci/dt3155/dt3155.c 	mutex_init(&pd->mux);
pd                518 drivers/media/pci/dt3155/dt3155.c 	pd->vdev.lock = &pd->mux; /* for locking v4l2_file_operations */
pd                519 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
pd                520 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
pd                521 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ;
pd                522 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.ops = &q_ops;
pd                523 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.mem_ops = &vb2_dma_contig_memops;
pd                524 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.drv_priv = pd;
pd                525 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.min_buffers_needed = 2;
pd                526 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.gfp_flags = GFP_DMA32;
pd                527 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.lock = &pd->mux; /* for locking v4l2_file_operations */
pd                528 drivers/media/pci/dt3155/dt3155.c 	pd->vidq.dev = &pdev->dev;
pd                529 drivers/media/pci/dt3155/dt3155.c 	pd->vdev.queue = &pd->vidq;
pd                530 drivers/media/pci/dt3155/dt3155.c 	err = vb2_queue_init(&pd->vidq);
pd                533 drivers/media/pci/dt3155/dt3155.c 	spin_lock_init(&pd->lock);
pd                534 drivers/media/pci/dt3155/dt3155.c 	pd->config = ACQ_MODE_EVEN;
pd                541 drivers/media/pci/dt3155/dt3155.c 	pd->regs = pci_iomap(pdev, 0, pci_resource_len(pd->pdev, 0));
pd                542 drivers/media/pci/dt3155/dt3155.c 	if (!pd->regs) {
pd                546 drivers/media/pci/dt3155/dt3155.c 	err = dt3155_init_board(pd);
pd                549 drivers/media/pci/dt3155/dt3155.c 	err = request_irq(pd->pdev->irq, dt3155_irq_handler_even,
pd                550 drivers/media/pci/dt3155/dt3155.c 					IRQF_SHARED, DT3155_NAME, pd);
pd                553 drivers/media/pci/dt3155/dt3155.c 	err = video_register_device(&pd->vdev, VFL_TYPE_GRABBER, -1);
pd                556 drivers/media/pci/dt3155/dt3155.c 	dev_info(&pdev->dev, "/dev/video%i is ready\n", pd->vdev.minor);
pd                560 drivers/media/pci/dt3155/dt3155.c 	free_irq(pd->pdev->irq, pd);
pd                562 drivers/media/pci/dt3155/dt3155.c 	pci_iounmap(pdev, pd->regs);
pd                568 drivers/media/pci/dt3155/dt3155.c 	v4l2_device_unregister(&pd->v4l2_dev);
pd                575 drivers/media/pci/dt3155/dt3155.c 	struct dt3155_priv *pd = container_of(v4l2_dev, struct dt3155_priv,
pd                578 drivers/media/pci/dt3155/dt3155.c 	video_unregister_device(&pd->vdev);
pd                579 drivers/media/pci/dt3155/dt3155.c 	free_irq(pd->pdev->irq, pd);
pd                580 drivers/media/pci/dt3155/dt3155.c 	vb2_queue_release(&pd->vidq);
pd                581 drivers/media/pci/dt3155/dt3155.c 	v4l2_device_unregister(&pd->v4l2_dev);
pd                582 drivers/media/pci/dt3155/dt3155.c 	pci_iounmap(pdev, pd->regs);
pd                387 drivers/media/platform/exynos4-is/media-dev.c 	struct fimc_source_info *pd = &fmd->sensor[index].pdata;
pd                408 drivers/media/platform/exynos4-is/media-dev.c 	pd->mux_id = (endpoint.base.port - 1) & 0x1;
pd                420 drivers/media/platform/exynos4-is/media-dev.c 			pd->sensor_bus_type = FIMC_BUS_TYPE_ITU_601;
pd                422 drivers/media/platform/exynos4-is/media-dev.c 			pd->sensor_bus_type = FIMC_BUS_TYPE_ITU_656;
pd                423 drivers/media/platform/exynos4-is/media-dev.c 		pd->flags = endpoint.bus.parallel.flags;
pd                429 drivers/media/platform/exynos4-is/media-dev.c 		pd->sensor_bus_type = FIMC_BUS_TYPE_MIPI_CSI2;
pd                445 drivers/media/platform/exynos4-is/media-dev.c 		pd->fimc_bus_type = FIMC_BUS_TYPE_ISP_WRITEBACK;
pd                447 drivers/media/platform/exynos4-is/media-dev.c 		pd->fimc_bus_type = pd->sensor_bus_type;
pd               1612 drivers/media/radio/si4713/si4713.c 	sdev->pd = si4713_pdev;
pd               1631 drivers/media/radio/si4713/si4713.c 	platform_device_unregister(sdev->pd);
pd                242 drivers/media/radio/si4713/si4713.h 	struct platform_device *pd;
pd                555 drivers/media/tuners/tda18271-common.c 	u8 d, pd;
pd                558 drivers/media/tuners/tda18271-common.c 	int ret = tda18271_lookup_pll_map(fe, MAIN_PLL, &freq, &pd, &d);
pd                562 drivers/media/tuners/tda18271-common.c 	regs[R_MPD]   = (0x7f & pd);
pd                578 drivers/media/tuners/tda18271-common.c 	u8 d, pd;
pd                581 drivers/media/tuners/tda18271-common.c 	int ret = tda18271_lookup_pll_map(fe, CAL_PLL, &freq, &pd, &d);
pd                585 drivers/media/tuners/tda18271-common.c 	regs[R_CPD]   = pd;
pd                 13 drivers/media/tuners/tda18271-maps.c 	u8 pd; /* post div */
pd                 25 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  32000, .pd = 0x5f, .d = 0xf0 },
pd                 26 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  35000, .pd = 0x5e, .d = 0xe0 },
pd                 27 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  37000, .pd = 0x5d, .d = 0xd0 },
pd                 28 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  41000, .pd = 0x5c, .d = 0xc0 },
pd                 29 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  44000, .pd = 0x5b, .d = 0xb0 },
pd                 30 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  49000, .pd = 0x5a, .d = 0xa0 },
pd                 31 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  54000, .pd = 0x59, .d = 0x90 },
pd                 32 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  61000, .pd = 0x58, .d = 0x80 },
pd                 33 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  65000, .pd = 0x4f, .d = 0x78 },
pd                 34 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  70000, .pd = 0x4e, .d = 0x70 },
pd                 35 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  75000, .pd = 0x4d, .d = 0x68 },
pd                 36 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  82000, .pd = 0x4c, .d = 0x60 },
pd                 37 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  89000, .pd = 0x4b, .d = 0x58 },
pd                 38 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  98000, .pd = 0x4a, .d = 0x50 },
pd                 39 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 109000, .pd = 0x49, .d = 0x48 },
pd                 40 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 123000, .pd = 0x48, .d = 0x40 },
pd                 41 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 131000, .pd = 0x3f, .d = 0x3c },
pd                 42 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 141000, .pd = 0x3e, .d = 0x38 },
pd                 43 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 151000, .pd = 0x3d, .d = 0x34 },
pd                 44 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 164000, .pd = 0x3c, .d = 0x30 },
pd                 45 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 179000, .pd = 0x3b, .d = 0x2c },
pd                 46 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 197000, .pd = 0x3a, .d = 0x28 },
pd                 47 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 219000, .pd = 0x39, .d = 0x24 },
pd                 48 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 246000, .pd = 0x38, .d = 0x20 },
pd                 49 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 263000, .pd = 0x2f, .d = 0x1e },
pd                 50 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 282000, .pd = 0x2e, .d = 0x1c },
pd                 51 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 303000, .pd = 0x2d, .d = 0x1a },
pd                 52 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 329000, .pd = 0x2c, .d = 0x18 },
pd                 53 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 359000, .pd = 0x2b, .d = 0x16 },
pd                 54 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 395000, .pd = 0x2a, .d = 0x14 },
pd                 55 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 438000, .pd = 0x29, .d = 0x12 },
pd                 56 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 493000, .pd = 0x28, .d = 0x10 },
pd                 57 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 526000, .pd = 0x1f, .d = 0x0f },
pd                 58 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 564000, .pd = 0x1e, .d = 0x0e },
pd                 59 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 607000, .pd = 0x1d, .d = 0x0d },
pd                 60 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 658000, .pd = 0x1c, .d = 0x0c },
pd                 61 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 718000, .pd = 0x1b, .d = 0x0b },
pd                 62 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 790000, .pd = 0x1a, .d = 0x0a },
pd                 63 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 877000, .pd = 0x19, .d = 0x09 },
pd                 64 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 987000, .pd = 0x18, .d = 0x08 },
pd                 65 drivers/media/tuners/tda18271-maps.c 	{ .lomax =      0, .pd = 0x00, .d = 0x00 }, /* end */
pd                 69 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  33125, .pd = 0x57, .d = 0xf0 },
pd                 70 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  35500, .pd = 0x56, .d = 0xe0 },
pd                 71 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  38188, .pd = 0x55, .d = 0xd0 },
pd                 72 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  41375, .pd = 0x54, .d = 0xc0 },
pd                 73 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  45125, .pd = 0x53, .d = 0xb0 },
pd                 74 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  49688, .pd = 0x52, .d = 0xa0 },
pd                 75 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  55188, .pd = 0x51, .d = 0x90 },
pd                 76 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  62125, .pd = 0x50, .d = 0x80 },
pd                 77 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  66250, .pd = 0x47, .d = 0x78 },
pd                 78 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  71000, .pd = 0x46, .d = 0x70 },
pd                 79 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  76375, .pd = 0x45, .d = 0x68 },
pd                 80 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  82750, .pd = 0x44, .d = 0x60 },
pd                 81 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  90250, .pd = 0x43, .d = 0x58 },
pd                 82 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  99375, .pd = 0x42, .d = 0x50 },
pd                 83 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 110375, .pd = 0x41, .d = 0x48 },
pd                 84 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 124250, .pd = 0x40, .d = 0x40 },
pd                 85 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 132500, .pd = 0x37, .d = 0x3c },
pd                 86 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 142000, .pd = 0x36, .d = 0x38 },
pd                 87 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 152750, .pd = 0x35, .d = 0x34 },
pd                 88 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 165500, .pd = 0x34, .d = 0x30 },
pd                 89 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 180500, .pd = 0x33, .d = 0x2c },
pd                 90 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 198750, .pd = 0x32, .d = 0x28 },
pd                 91 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 220750, .pd = 0x31, .d = 0x24 },
pd                 92 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 248500, .pd = 0x30, .d = 0x20 },
pd                 93 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 265000, .pd = 0x27, .d = 0x1e },
pd                 94 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 284000, .pd = 0x26, .d = 0x1c },
pd                 95 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 305500, .pd = 0x25, .d = 0x1a },
pd                 96 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 331000, .pd = 0x24, .d = 0x18 },
pd                 97 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 361000, .pd = 0x23, .d = 0x16 },
pd                 98 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 397500, .pd = 0x22, .d = 0x14 },
pd                 99 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 441500, .pd = 0x21, .d = 0x12 },
pd                100 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 497000, .pd = 0x20, .d = 0x10 },
pd                101 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 530000, .pd = 0x17, .d = 0x0f },
pd                102 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 568000, .pd = 0x16, .d = 0x0e },
pd                103 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 611000, .pd = 0x15, .d = 0x0d },
pd                104 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 662000, .pd = 0x14, .d = 0x0c },
pd                105 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 722000, .pd = 0x13, .d = 0x0b },
pd                106 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 795000, .pd = 0x12, .d = 0x0a },
pd                107 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 883000, .pd = 0x11, .d = 0x09 },
pd                108 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 994000, .pd = 0x10, .d = 0x08 },
pd                109 drivers/media/tuners/tda18271-maps.c 	{ .lomax =      0, .pd = 0x00, .d = 0x00 }, /* end */
pd                113 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   33000, .pd = 0xdd, .d = 0xd0 },
pd                114 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   36000, .pd = 0xdc, .d = 0xc0 },
pd                115 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   40000, .pd = 0xdb, .d = 0xb0 },
pd                116 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   44000, .pd = 0xda, .d = 0xa0 },
pd                117 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   49000, .pd = 0xd9, .d = 0x90 },
pd                118 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   55000, .pd = 0xd8, .d = 0x80 },
pd                119 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   63000, .pd = 0xd3, .d = 0x70 },
pd                120 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   67000, .pd = 0xcd, .d = 0x68 },
pd                121 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   73000, .pd = 0xcc, .d = 0x60 },
pd                122 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   80000, .pd = 0xcb, .d = 0x58 },
pd                123 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   88000, .pd = 0xca, .d = 0x50 },
pd                124 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   98000, .pd = 0xc9, .d = 0x48 },
pd                125 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  110000, .pd = 0xc8, .d = 0x40 },
pd                126 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  126000, .pd = 0xc3, .d = 0x38 },
pd                127 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  135000, .pd = 0xbd, .d = 0x34 },
pd                128 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  147000, .pd = 0xbc, .d = 0x30 },
pd                129 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  160000, .pd = 0xbb, .d = 0x2c },
pd                130 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  176000, .pd = 0xba, .d = 0x28 },
pd                131 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  196000, .pd = 0xb9, .d = 0x24 },
pd                132 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  220000, .pd = 0xb8, .d = 0x20 },
pd                133 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  252000, .pd = 0xb3, .d = 0x1c },
pd                134 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  271000, .pd = 0xad, .d = 0x1a },
pd                135 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  294000, .pd = 0xac, .d = 0x18 },
pd                136 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  321000, .pd = 0xab, .d = 0x16 },
pd                137 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  353000, .pd = 0xaa, .d = 0x14 },
pd                138 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  392000, .pd = 0xa9, .d = 0x12 },
pd                139 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  441000, .pd = 0xa8, .d = 0x10 },
pd                140 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  505000, .pd = 0xa3, .d = 0x0e },
pd                141 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  543000, .pd = 0x9d, .d = 0x0d },
pd                142 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  589000, .pd = 0x9c, .d = 0x0c },
pd                143 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  642000, .pd = 0x9b, .d = 0x0b },
pd                144 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  707000, .pd = 0x9a, .d = 0x0a },
pd                145 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  785000, .pd = 0x99, .d = 0x09 },
pd                146 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  883000, .pd = 0x98, .d = 0x08 },
pd                147 drivers/media/tuners/tda18271-maps.c 	{ .lomax = 1010000, .pd = 0x93, .d = 0x07 },
pd                148 drivers/media/tuners/tda18271-maps.c 	{ .lomax =       0, .pd = 0x00, .d = 0x00 }, /* end */
pd                152 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   33813, .pd = 0xdd, .d = 0xd0 },
pd                153 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   36625, .pd = 0xdc, .d = 0xc0 },
pd                154 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   39938, .pd = 0xdb, .d = 0xb0 },
pd                155 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   43938, .pd = 0xda, .d = 0xa0 },
pd                156 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   48813, .pd = 0xd9, .d = 0x90 },
pd                157 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   54938, .pd = 0xd8, .d = 0x80 },
pd                158 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   62813, .pd = 0xd3, .d = 0x70 },
pd                159 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   67625, .pd = 0xcd, .d = 0x68 },
pd                160 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   73250, .pd = 0xcc, .d = 0x60 },
pd                161 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   79875, .pd = 0xcb, .d = 0x58 },
pd                162 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   87875, .pd = 0xca, .d = 0x50 },
pd                163 drivers/media/tuners/tda18271-maps.c 	{ .lomax =   97625, .pd = 0xc9, .d = 0x48 },
pd                164 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  109875, .pd = 0xc8, .d = 0x40 },
pd                165 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  125625, .pd = 0xc3, .d = 0x38 },
pd                166 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  135250, .pd = 0xbd, .d = 0x34 },
pd                167 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  146500, .pd = 0xbc, .d = 0x30 },
pd                168 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  159750, .pd = 0xbb, .d = 0x2c },
pd                169 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  175750, .pd = 0xba, .d = 0x28 },
pd                170 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  195250, .pd = 0xb9, .d = 0x24 },
pd                171 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  219750, .pd = 0xb8, .d = 0x20 },
pd                172 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  251250, .pd = 0xb3, .d = 0x1c },
pd                173 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  270500, .pd = 0xad, .d = 0x1a },
pd                174 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  293000, .pd = 0xac, .d = 0x18 },
pd                175 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  319500, .pd = 0xab, .d = 0x16 },
pd                176 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  351500, .pd = 0xaa, .d = 0x14 },
pd                177 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  390500, .pd = 0xa9, .d = 0x12 },
pd                178 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  439500, .pd = 0xa8, .d = 0x10 },
pd                179 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  502500, .pd = 0xa3, .d = 0x0e },
pd                180 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  541000, .pd = 0x9d, .d = 0x0d },
pd                181 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  586000, .pd = 0x9c, .d = 0x0c },
pd                182 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  639000, .pd = 0x9b, .d = 0x0b },
pd                183 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  703000, .pd = 0x9a, .d = 0x0a },
pd                184 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  781000, .pd = 0x99, .d = 0x09 },
pd                185 drivers/media/tuners/tda18271-maps.c 	{ .lomax =  879000, .pd = 0x98, .d = 0x08 },
pd                186 drivers/media/tuners/tda18271-maps.c 	{ .lomax =       0, .pd = 0x00, .d = 0x00 }, /* end */
pd               1093 drivers/media/tuners/tda18271-maps.c 	*post_div = map[i].pd;
pd               1201 drivers/memory/emif.c 	struct emif_platform_data *pd = emif->plat_data;
pd               1203 drivers/memory/emif.c 	pd->timings		= lpddr2_jedec_timings;
pd               1204 drivers/memory/emif.c 	pd->timings_arr_size	= ARRAY_SIZE(lpddr2_jedec_timings);
pd               1347 drivers/memory/emif.c 	struct emif_platform_data	*pd = NULL;
pd               1355 drivers/memory/emif.c 	pd	= devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd               1358 drivers/memory/emif.c 	if (!emif || !pd || !dev_info) {
pd               1364 drivers/memory/emif.c 	emif->plat_data		= pd;
pd               1365 drivers/memory/emif.c 	pd->device_info		= dev_info;
pd               1375 drivers/memory/emif.c 	of_property_read_u32(np_emif, "phy-type", &pd->phy_type);
pd               1378 drivers/memory/emif.c 		pd->hw_caps |= EMIF_HW_CAPS_LL_INTERFACE;
pd               1381 drivers/memory/emif.c 	if (!is_dev_data_valid(pd->device_info->type, pd->device_info->density,
pd               1382 drivers/memory/emif.c 			pd->device_info->io_width, pd->phy_type, pd->ip_rev,
pd               1431 drivers/memory/emif.c 	struct emif_platform_data	*pd;
pd               1435 drivers/memory/emif.c 	pd = pdev->dev.platform_data;
pd               1438 drivers/memory/emif.c 	if (!(pd && pd->device_info && is_dev_data_valid(pd->device_info->type,
pd               1439 drivers/memory/emif.c 			pd->device_info->density, pd->device_info->io_width,
pd               1440 drivers/memory/emif.c 			pd->phy_type, pd->ip_rev, dev))) {
pd               1446 drivers/memory/emif.c 	temp	= devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd               1449 drivers/memory/emif.c 	if (!emif || !pd || !dev_info) {
pd               1454 drivers/memory/emif.c 	memcpy(temp, pd, sizeof(*pd));
pd               1455 drivers/memory/emif.c 	pd = temp;
pd               1456 drivers/memory/emif.c 	memcpy(dev_info, pd->device_info, sizeof(*dev_info));
pd               1458 drivers/memory/emif.c 	pd->device_info		= dev_info;
pd               1459 drivers/memory/emif.c 	emif->plat_data		= pd;
pd               1474 drivers/memory/emif.c 		pd->timings = NULL;
pd               1475 drivers/memory/emif.c 		pd->min_tck = NULL;
pd               1486 drivers/memory/emif.c 	cust_cfgs = pd->custom_configs;
pd               1494 drivers/memory/emif.c 		pd->custom_configs = temp;
pd               1501 drivers/memory/emif.c 	size = sizeof(struct lpddr2_timings) * pd->timings_arr_size;
pd               1502 drivers/memory/emif.c 	if (pd->timings) {
pd               1505 drivers/memory/emif.c 			memcpy(temp, pd->timings, size);
pd               1506 drivers/memory/emif.c 			pd->timings = temp;
pd               1516 drivers/memory/emif.c 	if (pd->min_tck) {
pd               1517 drivers/memory/emif.c 		temp = devm_kzalloc(dev, sizeof(*pd->min_tck), GFP_KERNEL);
pd               1519 drivers/memory/emif.c 			memcpy(temp, pd->min_tck, sizeof(*pd->min_tck));
pd               1520 drivers/memory/emif.c 			pd->min_tck = temp;
pd               1524 drivers/memory/emif.c 			pd->min_tck = &lpddr2_jedec_min_tck;
pd               1527 drivers/memory/emif.c 		pd->min_tck = &lpddr2_jedec_min_tck;
pd                134 drivers/mfd/max8997.c 	struct max8997_platform_data *pd;
pd                136 drivers/mfd/max8997.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                137 drivers/mfd/max8997.c 	if (!pd)
pd                140 drivers/mfd/max8997.c 	pd->ono = irq_of_parse_and_map(dev->of_node, 1);
pd                142 drivers/mfd/max8997.c 	return pd;
pd                139 drivers/mfd/max8998.c 	struct max8998_platform_data *pd;
pd                141 drivers/mfd/max8998.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                142 drivers/mfd/max8998.c 	if (!pd)
pd                145 drivers/mfd/max8998.c 	pd->ono = irq_of_parse_and_map(dev->of_node, 1);
pd                152 drivers/mfd/max8998.c 	return pd;
pd                297 drivers/mfd/sec-core.c 	struct sec_platform_data *pd;
pd                299 drivers/mfd/sec-core.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                300 drivers/mfd/sec-core.c 	if (!pd)
pd                309 drivers/mfd/sec-core.c 	pd->manual_poweroff = of_property_read_bool(dev->of_node,
pd                311 drivers/mfd/sec-core.c 	pd->disable_wrstbi = of_property_read_bool(dev->of_node,
pd                313 drivers/mfd/sec-core.c 	return pd;
pd               1445 drivers/mfd/sm501.c 	struct sm501_platdata *pd = sm->platdata;
pd               1447 drivers/mfd/sm501.c 	if (!pd)
pd               1450 drivers/mfd/sm501.c 	if (pd->get_power) {
pd               1451 drivers/mfd/sm501.c 		if (pd->get_power(sm->dev) == on) {
pd               1457 drivers/mfd/sm501.c 	if (pd->set_power) {
pd               1460 drivers/mfd/sm501.c 		pd->set_power(sm->dev, on);
pd                278 drivers/misc/eeprom/eeprom_93xx46.c 	struct eeprom_93xx46_platform_data *pd = edev->pdata;
pd                319 drivers/misc/eeprom/eeprom_93xx46.c 	if (pd->finish)
pd                320 drivers/misc/eeprom/eeprom_93xx46.c 		pd->finish(edev);
pd                375 drivers/misc/eeprom/eeprom_93xx46.c 	struct eeprom_93xx46_platform_data *pd;
pd                379 drivers/misc/eeprom/eeprom_93xx46.c 	pd = devm_kzalloc(&spi->dev, sizeof(*pd), GFP_KERNEL);
pd                380 drivers/misc/eeprom/eeprom_93xx46.c 	if (!pd)
pd                390 drivers/misc/eeprom/eeprom_93xx46.c 		pd->flags |= EE_ADDR8;
pd                392 drivers/misc/eeprom/eeprom_93xx46.c 		pd->flags |= EE_ADDR16;
pd                399 drivers/misc/eeprom/eeprom_93xx46.c 		pd->flags |= EE_READONLY;
pd                401 drivers/misc/eeprom/eeprom_93xx46.c 	pd->select = devm_gpiod_get_optional(&spi->dev, "select",
pd                403 drivers/misc/eeprom/eeprom_93xx46.c 	if (IS_ERR(pd->select))
pd                404 drivers/misc/eeprom/eeprom_93xx46.c 		return PTR_ERR(pd->select);
pd                406 drivers/misc/eeprom/eeprom_93xx46.c 	pd->prepare = select_assert;
pd                407 drivers/misc/eeprom/eeprom_93xx46.c 	pd->finish = select_deassert;
pd                408 drivers/misc/eeprom/eeprom_93xx46.c 	gpiod_direction_output(pd->select, 0);
pd                413 drivers/misc/eeprom/eeprom_93xx46.c 		pd->quirks = data->quirks;
pd                416 drivers/misc/eeprom/eeprom_93xx46.c 	spi->dev.platform_data = pd;
pd                423 drivers/misc/eeprom/eeprom_93xx46.c 	struct eeprom_93xx46_platform_data *pd;
pd                433 drivers/misc/eeprom/eeprom_93xx46.c 	pd = spi->dev.platform_data;
pd                434 drivers/misc/eeprom/eeprom_93xx46.c 	if (!pd) {
pd                443 drivers/misc/eeprom/eeprom_93xx46.c 	if (pd->flags & EE_ADDR8)
pd                445 drivers/misc/eeprom/eeprom_93xx46.c 	else if (pd->flags & EE_ADDR16)
pd                455 drivers/misc/eeprom/eeprom_93xx46.c 	edev->pdata = pd;
pd                460 drivers/misc/eeprom/eeprom_93xx46.c 	edev->nvmem_config.read_only = pd->flags & EE_READONLY;
pd                477 drivers/misc/eeprom/eeprom_93xx46.c 		(pd->flags & EE_ADDR8) ? 8 : 16,
pd                478 drivers/misc/eeprom/eeprom_93xx46.c 		(pd->flags & EE_READONLY) ? "(readonly)" : "");
pd                480 drivers/misc/eeprom/eeprom_93xx46.c 	if (!(pd->flags & EE_READONLY)) {
pd                201 drivers/misc/fastrpc.c 	int pd;
pd                897 drivers/misc/fastrpc.c 	msg->ctx = ctx->ctxid | fl->pd;
pd               1009 drivers/misc/fastrpc.c 	fl->pd = 1;
pd               1252 drivers/misc/fastrpc.c 	fl->pd = 0;
pd                223 drivers/mmc/host/sh_mmcif.c 	struct platform_device *pd;
pd                258 drivers/mmc/host/sh_mmcif.c #define sh_mmcif_host_to_dev(host) (&host->pd->dev)
pd                407 drivers/mmc/host/sh_mmcif.c 	res = platform_get_resource(host->pd, IORESOURCE_MEM, 0);
pd               1370 drivers/mmc/host/sh_mmcif.c 	struct sh_mmcif_plat_data *pd = dev->platform_data;
pd               1375 drivers/mmc/host/sh_mmcif.c 	if (!pd)
pd               1379 drivers/mmc/host/sh_mmcif.c 		mmc->ocr_avail = pd->ocr;
pd               1380 drivers/mmc/host/sh_mmcif.c 	else if (pd->ocr)
pd               1390 drivers/mmc/host/sh_mmcif.c 	struct sh_mmcif_plat_data *pd = dev->platform_data;
pd               1420 drivers/mmc/host/sh_mmcif.c 	host->pd = pdev;
pd               1431 drivers/mmc/host/sh_mmcif.c 	if (pd && pd->caps)
pd               1432 drivers/mmc/host/sh_mmcif.c 		mmc->caps |= pd->caps;
pd                378 drivers/mtd/nand/raw/au1550nd.c 	struct au1550nd_platdata *pd;
pd                385 drivers/mtd/nand/raw/au1550nd.c 	pd = dev_get_platdata(&pdev->dev);
pd                386 drivers/mtd/nand/raw/au1550nd.c 	if (!pd) {
pd                436 drivers/mtd/nand/raw/au1550nd.c 	if (pd->devwidth)
pd                439 drivers/mtd/nand/raw/au1550nd.c 	this->legacy.read_byte = (pd->devwidth) ? au_read_byte16 : au_read_byte;
pd                440 drivers/mtd/nand/raw/au1550nd.c 	ctx->write_byte = (pd->devwidth) ? au_write_byte16 : au_write_byte;
pd                441 drivers/mtd/nand/raw/au1550nd.c 	this->legacy.write_buf = (pd->devwidth) ? au_write_buf16 : au_write_buf;
pd                442 drivers/mtd/nand/raw/au1550nd.c 	this->legacy.read_buf = (pd->devwidth) ? au_read_buf16 : au_read_buf;
pd                450 drivers/mtd/nand/raw/au1550nd.c 	mtd_device_register(mtd, pd->parts, pd->num_parts);
pd               1470 drivers/net/can/grcan.c 	static void grcan_sanitize_##name(struct platform_device *pd)	\
pd               1475 drivers/net/can/grcan.c 			dev_err(&pd->dev,				\
pd                 41 drivers/net/dsa/sja1105/sja1105_clocking.c 	u64 pd;
pd                 57 drivers/net/dsa/sja1105/sja1105_clocking.c 	u64 pd;
pd                 87 drivers/net/dsa/sja1105/sja1105_clocking.c 	u64 pd;
pd                 98 drivers/net/dsa/sja1105/sja1105_clocking.c 	sja1105_packing(buf, &idiv->pd,         0,  0, size, op);
pd                118 drivers/net/dsa/sja1105/sja1105_clocking.c 	idiv.pd        = enabled ? 0 : 1; /* Power down? */
pd                134 drivers/net/dsa/sja1105/sja1105_clocking.c 	sja1105_packing(buf, &cmd->pd,         0,  0, size, op);
pd                167 drivers/net/dsa/sja1105/sja1105_clocking.c 	mii_tx_clk.pd        = 0;  /* Power Down off => enabled */
pd                192 drivers/net/dsa/sja1105/sja1105_clocking.c 	mii_rx_clk.pd        = 0;  /* Power Down off => enabled */
pd                217 drivers/net/dsa/sja1105/sja1105_clocking.c 	mii_ext_tx_clk.pd        = 0; /* Power Down off => enabled */
pd                242 drivers/net/dsa/sja1105/sja1105_clocking.c 	mii_ext_rx_clk.pd        = 0; /* Power Down off => enabled */
pd                313 drivers/net/dsa/sja1105/sja1105_clocking.c 	sja1105_packing(buf, &cmd->pd,         0,  0, size, op);
pd                337 drivers/net/dsa/sja1105/sja1105_clocking.c 	txc.pd = 0;
pd                547 drivers/net/dsa/sja1105/sja1105_clocking.c 	ref_clk.pd        = 0;      /* Power Down off => enabled */
pd                565 drivers/net/dsa/sja1105/sja1105_clocking.c 	ext_tx_clk.pd        = 0;   /* Power Down off => enabled */
pd                595 drivers/net/dsa/sja1105/sja1105_clocking.c 	pll.pd        = 0x1;
pd                606 drivers/net/dsa/sja1105/sja1105_clocking.c 	pll.pd = 0x0;
pd               1081 drivers/net/ethernet/amd/au1000_eth.c 	struct au1000_eth_platform_data *pd;
pd               1193 drivers/net/ethernet/amd/au1000_eth.c 	pd = dev_get_platdata(&pdev->dev);
pd               1194 drivers/net/ethernet/amd/au1000_eth.c 	if (!pd) {
pd               1199 drivers/net/ethernet/amd/au1000_eth.c 		if (is_valid_ether_addr(pd->mac)) {
pd               1200 drivers/net/ethernet/amd/au1000_eth.c 			memcpy(dev->dev_addr, pd->mac, ETH_ALEN);
pd               1206 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy_static_config = pd->phy_static_config;
pd               1207 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy_search_highest_addr = pd->phy_search_highest_addr;
pd               1208 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy1_search_mac0 = pd->phy1_search_mac0;
pd               1209 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy_addr = pd->phy_addr;
pd               1210 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy_busid = pd->phy_busid;
pd               1211 drivers/net/ethernet/amd/au1000_eth.c 		aup->phy_irq = pd->phy_irq;
pd               1695 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	struct bcm63xx_enet_platform_data *pd;
pd               1745 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	pd = dev_get_platdata(&pdev->dev);
pd               1746 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	if (pd) {
pd               1747 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN);
pd               1748 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->has_phy = pd->has_phy;
pd               1749 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->phy_id = pd->phy_id;
pd               1750 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->has_phy_interrupt = pd->has_phy_interrupt;
pd               1751 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->phy_interrupt = pd->phy_interrupt;
pd               1752 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->use_external_mii = !pd->use_internal_phy;
pd               1753 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->pause_auto = pd->pause_auto;
pd               1754 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->pause_rx = pd->pause_rx;
pd               1755 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->pause_tx = pd->pause_tx;
pd               1756 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->force_duplex_full = pd->force_duplex_full;
pd               1757 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->force_speed_100 = pd->force_speed_100;
pd               1758 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_en_mask = pd->dma_chan_en_mask;
pd               1759 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_int_mask = pd->dma_chan_int_mask;
pd               1760 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_width = pd->dma_chan_width;
pd               1761 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_has_sram = pd->dma_has_sram;
pd               1762 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_desc_shift = pd->dma_desc_shift;
pd               1763 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->rx_chan = pd->rx_chan;
pd               1764 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->tx_chan = pd->tx_chan;
pd               1816 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		if (pd && pd->mii_config &&
pd               1817 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		    pd->mii_config(dev, 1, bcm_enet_mdio_read_mii,
pd               1899 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		struct bcm63xx_enet_platform_data *pd;
pd               1901 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		pd = dev_get_platdata(&pdev->dev);
pd               1902 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		if (pd && pd->mii_config)
pd               1903 drivers/net/ethernet/broadcom/bcm63xx_enet.c 			pd->mii_config(dev, 0, bcm_enet_mdio_read_mii,
pd               2643 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	struct bcm63xx_enetsw_platform_data *pd;
pd               2670 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	pd = dev_get_platdata(&pdev->dev);
pd               2671 drivers/net/ethernet/broadcom/bcm63xx_enet.c 	if (pd) {
pd               2672 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN);
pd               2673 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		memcpy(priv->used_ports, pd->used_ports,
pd               2674 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		       sizeof(pd->used_ports));
pd               2675 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->num_ports = pd->num_ports;
pd               2676 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_has_sram = pd->dma_has_sram;
pd               2677 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_en_mask = pd->dma_chan_en_mask;
pd               2678 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_int_mask = pd->dma_chan_int_mask;
pd               2679 drivers/net/ethernet/broadcom/bcm63xx_enet.c 		priv->dma_chan_width = pd->dma_chan_width;
pd               5736 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 			u16 pd, gp_status4;
pd               5749 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 						&pd);
pd               5750 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c 				if (pd & (1<<15))
pd               3439 drivers/net/ethernet/broadcom/genet/bcmgenet.c 	struct bcmgenet_platform_data *pd = pdev->dev.platform_data;
pd               3481 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		macaddr = pd->mac_address;
pd               3524 drivers/net/ethernet/broadcom/genet/bcmgenet.c 		priv->version = pd->genet_version;
pd                394 drivers/net/ethernet/broadcom/genet/bcmmii.c 	struct bcmgenet_platform_data *pd = kdev->platform_data;
pd                396 drivers/net/ethernet/broadcom/genet/bcmmii.c 	if (pd->phy_interface != PHY_INTERFACE_MODE_MOCA && pd->mdio_enabled) {
pd                400 drivers/net/ethernet/broadcom/genet/bcmmii.c 		if (pd->phy_address >= 0 && pd->phy_address < PHY_MAX_ADDR)
pd                401 drivers/net/ethernet/broadcom/genet/bcmmii.c 			ppd->phy_mask = 1 << pd->phy_address;
pd                530 drivers/net/ethernet/broadcom/genet/bcmmii.c 	struct bcmgenet_platform_data *pd = kdev->platform_data;
pd                538 drivers/net/ethernet/broadcom/genet/bcmmii.c 	if (pd->phy_interface != PHY_INTERFACE_MODE_MOCA && pd->mdio_enabled) {
pd                540 drivers/net/ethernet/broadcom/genet/bcmmii.c 			 mdio_bus_id, pd->phy_address);
pd                545 drivers/net/ethernet/broadcom/genet/bcmmii.c 		phydev = phy_attach(priv->dev, phy_name, pd->phy_interface);
pd                557 drivers/net/ethernet/broadcom/genet/bcmmii.c 			.speed = pd->phy_speed,
pd                558 drivers/net/ethernet/broadcom/genet/bcmmii.c 			.duplex = pd->phy_duplex,
pd                574 drivers/net/ethernet/broadcom/genet/bcmmii.c 	priv->phy_interface = pd->phy_interface;
pd               12003 drivers/net/ethernet/broadcom/tg3.c 	u8  *pd;
pd               12047 drivers/net/ethernet/broadcom/tg3.c 	pd = &data[eeprom->len];
pd               12056 drivers/net/ethernet/broadcom/tg3.c 		memcpy(pd + i, &val, 4);
pd               12070 drivers/net/ethernet/broadcom/tg3.c 		pd = &data[eeprom->len];
pd               12076 drivers/net/ethernet/broadcom/tg3.c 		memcpy(pd, &val, b_count);
pd               1475 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 		struct phy_device *pd;
pd               1501 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 			pd = of_phy_find_device(phy_np);
pd               1502 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 			if (!pd)
pd               1504 drivers/net/ethernet/cavium/thunder/thunder_bgx.c 			bgx->lmac[lmac].phydev = pd;
pd                372 drivers/net/ethernet/emulex/benet/be_cmds.h 	u8 pd[10];		/* dword 1*/
pd                464 drivers/net/ethernet/emulex/benet/be_cmds.h 	u8 pd[10];		/* dword 1*/
pd                385 drivers/net/ethernet/ibm/ehea/ehea.h 	u32 pd;                    /* protection domain */
pd               1512 drivers/net/ethernet/ibm/ehea/ehea_main.c 	pr->qp = ehea_create_qp(adapter, adapter->pd, init_attr);
pd               3399 drivers/net/ethernet/ibm/ehea/ehea_main.c 	adapter->pd = EHEA_PD_ID;
pd                199 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 			     struct ehea_qp_init_attr *init_attr, const u32 pd,
pd                216 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 	u64 r9_reg = EHEA_BMASK_SET(H_ALL_RES_QP_PD, pd)
pd                434 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 			const u64 vaddr_in, const u32 access_ctrl, const u32 pd,
pd                446 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 				 pd,				 /* R8 */
pd                479 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 			     const u32 pd, u64 *mr_handle, u32 *lkey)
pd                491 drivers/net/ethernet/ibm/ehea/ehea_phyp.c 				 pd,				   /* R9 */
pd                368 drivers/net/ethernet/ibm/ehea/ehea_phyp.h 			     const u32 pd,
pd                394 drivers/net/ethernet/ibm/ehea/ehea_phyp.h 			     const u32 pd, u64 *mr_handle, u32 *lkey);
pd                401 drivers/net/ethernet/ibm/ehea/ehea_phyp.h 			const u64 vaddr_in, const u32 access_ctrl, const u32 pd,
pd                398 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 			       u32 pd, struct ehea_qp_init_attr *init_attr)
pd                413 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 	hret = ehea_h_alloc_resource_qp(adapter->handle, init_attr, pd,
pd                877 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 					ehea_mr_len, acc_ctrl, adapter->pd,
pd                941 drivers/net/ethernet/ibm/ehea/ehea_qmr.c 				   adapter->pd, shared_mr);
pd                369 drivers/net/ethernet/ibm/ehea/ehea_qmr.h struct ehea_qp *ehea_create_qp(struct ehea_adapter *adapter, u32 pd,
pd               1484 drivers/net/ethernet/ibm/emac/core.c 				  u32 pd, int len, int last, u16 base_ctrl)
pd               1499 drivers/net/ethernet/ibm/emac/core.c 		dev->tx_desc[slot].data_ptr = pd;
pd               1507 drivers/net/ethernet/ibm/emac/core.c 		pd += chunk;
pd               1521 drivers/net/ethernet/ibm/emac/core.c 	u32 pd;
pd               1543 drivers/net/ethernet/ibm/emac/core.c 	dev->tx_desc[slot].data_ptr = pd =
pd               1548 drivers/net/ethernet/ibm/emac/core.c 		slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags,
pd               1558 drivers/net/ethernet/ibm/emac/core.c 		pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len,
pd               1561 drivers/net/ethernet/ibm/emac/core.c 		slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1,
pd               2786 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_shared_platform_data *pd;
pd               2794 drivers/net/ethernet/marvell/mv643xx_eth.c 	pd = devm_kzalloc(&pdev->dev, sizeof(*pd), GFP_KERNEL);
pd               2795 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (!pd)
pd               2797 drivers/net/ethernet/marvell/mv643xx_eth.c 	pdev->dev.platform_data = pd;
pd               2799 drivers/net/ethernet/marvell/mv643xx_eth.c 	mv643xx_eth_property(np, "tx-checksum-limit", pd->tx_csum_limit);
pd               2834 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_shared_platform_data *pd;
pd               2871 drivers/net/ethernet/marvell/mv643xx_eth.c 	pd = dev_get_platdata(&pdev->dev);
pd               2873 drivers/net/ethernet/marvell/mv643xx_eth.c 	msp->tx_csum_limit = (pd != NULL && pd->tx_csum_limit) ?
pd               2874 drivers/net/ethernet/marvell/mv643xx_eth.c 					pd->tx_csum_limit : 9 * 1024;
pd               2925 drivers/net/ethernet/marvell/mv643xx_eth.c 		       struct mv643xx_eth_platform_data *pd)
pd               2930 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (is_valid_ether_addr(pd->mac_addr))
pd               2931 drivers/net/ethernet/marvell/mv643xx_eth.c 		memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN);
pd               2936 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (pd->rx_queue_size)
pd               2937 drivers/net/ethernet/marvell/mv643xx_eth.c 		mp->rx_ring_size = pd->rx_queue_size;
pd               2938 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->rx_desc_sram_addr = pd->rx_sram_addr;
pd               2939 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->rx_desc_sram_size = pd->rx_sram_size;
pd               2941 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->rxq_count = pd->rx_queue_count ? : 1;
pd               2944 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (pd->tx_queue_size)
pd               2945 drivers/net/ethernet/marvell/mv643xx_eth.c 		tx_ring_size = pd->tx_queue_size;
pd               2953 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->tx_desc_sram_addr = pd->tx_sram_addr;
pd               2954 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->tx_desc_sram_size = pd->tx_sram_size;
pd               2956 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->txq_count = pd->tx_queue_count ? : 1;
pd               3080 drivers/net/ethernet/marvell/mv643xx_eth.c 	struct mv643xx_eth_platform_data *pd;
pd               3087 drivers/net/ethernet/marvell/mv643xx_eth.c 	pd = dev_get_platdata(&pdev->dev);
pd               3088 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (pd == NULL) {
pd               3093 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (pd->shared == NULL) {
pd               3106 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->shared = platform_get_drvdata(pd->shared);
pd               3107 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->base = mp->shared->base + 0x0400 + (pd->port_number << 10);
pd               3108 drivers/net/ethernet/marvell/mv643xx_eth.c 	mp->port_num = pd->port_number;
pd               3134 drivers/net/ethernet/marvell/mv643xx_eth.c 	set_params(mp, pd);
pd               3139 drivers/net/ethernet/marvell/mv643xx_eth.c 	if (pd->phy_node) {
pd               3140 drivers/net/ethernet/marvell/mv643xx_eth.c 		phydev = of_phy_connect(mp->dev, pd->phy_node,
pd               3147 drivers/net/ethernet/marvell/mv643xx_eth.c 	} else if (pd->phy_addr != MV643XX_ETH_PHY_NONE) {
pd               3148 drivers/net/ethernet/marvell/mv643xx_eth.c 		phydev = phy_scan(mp, pd->phy_addr);
pd               3153 drivers/net/ethernet/marvell/mv643xx_eth.c 			phy_init(mp, pd->speed, pd->duplex);
pd               3164 drivers/net/ethernet/marvell/mv643xx_eth.c 	init_pscr(mp, pd->speed, pd->duplex);
pd                242 drivers/net/ethernet/marvell/pxa168_eth.c 	struct pxa168_eth_platform_data *pd;
pd               1465 drivers/net/ethernet/marvell/pxa168_eth.c 	pep->pd = dev_get_platdata(&pdev->dev);
pd               1466 drivers/net/ethernet/marvell/pxa168_eth.c 	if (pep->pd) {
pd               1467 drivers/net/ethernet/marvell/pxa168_eth.c 		if (pep->pd->rx_queue_size)
pd               1468 drivers/net/ethernet/marvell/pxa168_eth.c 			pep->rx_ring_size = pep->pd->rx_queue_size;
pd               1470 drivers/net/ethernet/marvell/pxa168_eth.c 		if (pep->pd->tx_queue_size)
pd               1471 drivers/net/ethernet/marvell/pxa168_eth.c 			pep->tx_ring_size = pep->pd->tx_queue_size;
pd               1473 drivers/net/ethernet/marvell/pxa168_eth.c 		pep->port_num = pep->pd->port_number;
pd               1474 drivers/net/ethernet/marvell/pxa168_eth.c 		pep->phy_addr = pep->pd->phy_addr;
pd               1475 drivers/net/ethernet/marvell/pxa168_eth.c 		pep->phy_speed = pep->pd->speed;
pd               1476 drivers/net/ethernet/marvell/pxa168_eth.c 		pep->phy_duplex = pep->pd->duplex;
pd               1477 drivers/net/ethernet/marvell/pxa168_eth.c 		pep->phy_intf = pep->pd->intf;
pd               1479 drivers/net/ethernet/marvell/pxa168_eth.c 		if (pep->pd->init)
pd               1480 drivers/net/ethernet/marvell/pxa168_eth.c 			pep->pd->init();
pd                 49 drivers/net/ethernet/mellanox/mlx4/en_resources.c 	context->pd = cpu_to_be32(mdev->priv_pdn);
pd                374 drivers/net/ethernet/mellanox/mlx4/mlx4.h 	__be32			pd;
pd                371 drivers/net/ethernet/mellanox/mlx4/mr.c 		mmr->pd = be32_to_cpu((*mpt_entry)->pd_flags) & MLX4_MPT_PD_MASK;
pd                417 drivers/net/ethernet/mellanox/mlx4/mr.c static int mlx4_mr_alloc_reserved(struct mlx4_dev *dev, u32 mridx, u32 pd,
pd                423 drivers/net/ethernet/mellanox/mlx4/mr.c 	mr->pd	       = pd;
pd                527 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access,
pd                537 drivers/net/ethernet/mellanox/mlx4/mr.c 	err = mlx4_mr_alloc_reserved(dev, index, pd, iova, size,
pd                648 drivers/net/ethernet/mellanox/mlx4/mr.c 	mpt_entry->pd_flags    = cpu_to_be32(mr->pd | MLX4_MPT_PD_FLAG_EN_INV);
pd                814 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type,
pd                830 drivers/net/ethernet/mellanox/mlx4/mr.c 	mw->pd      = pd;
pd                859 drivers/net/ethernet/mellanox/mlx4/mr.c 	mpt_entry->pd_flags    = cpu_to_be32(mw->pd);
pd               1043 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages,
pd               1064 drivers/net/ethernet/mellanox/mlx4/mr.c 	err = mlx4_mr_alloc(dev, pd, 0, 0, access, max_pages,
pd               2767 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	u32 pd;
pd               2782 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	pd = mr_get_pd(inbox->buf);
pd               2783 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	pd_slave = (pd >> 17) & 0x7f;
pd                198 drivers/net/ethernet/mellanox/mlx4/srq.c 	srq_context->pd			= cpu_to_be32(pdn);
pd                199 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c 	MLX5_SET(mkc, mkc, pd, tracer->buff.pdn);
pd                 81 drivers/net/ethernet/mellanox/mlx5/core/en_common.c 	MLX5_SET(mkc, mkc, pd, pdn);
pd                289 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	MLX5_SET(mkc, mkc, pd, mdev->mlx5e_res.pdn);
pd               2178 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	MLX5_SET(wq, wq, pd,               mdev->mlx5e_res.pdn);
pd               2208 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 	MLX5_SET(wq, wq, pd,            priv->mdev->mlx5e_res.pdn);
pd               3210 drivers/net/ethernet/mellanox/mlx5/core/en_main.c 		MLX5_SET(tisc, tisc, pd, mdev->mlx5e_res.pdn);
pd                240 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	MLX5_SET(mkc, mkc, pd, pdn);
pd                589 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	MLX5_SET(qpc, qpc, pd, fdev->conn_res.pdn);
pd                698 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c 	MLX5_SET(qpc, qpc, pd, conn->fdev->conn_res.pdn);
pd                 50 drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c 	MLX5_SET(encryption_key_obj, obj, pd, mdev->mlx5e_res.pdn);
pd                 83 drivers/net/ethernet/mellanox/mlx5/core/mr.c 	mkey->pd = MLX5_GET(mkc, mkc, pd);
pd                160 drivers/net/ethernet/mellanox/mlx5/core/mr.c 	MLX5_SET(create_psv_in, in, pd, pdn);
pd                 48 drivers/net/ethernet/mellanox/mlx5/core/pd.c 		*pdn = MLX5_GET(alloc_pd_out, out, pd);
pd                 59 drivers/net/ethernet/mellanox/mlx5/core/pd.c 	MLX5_SET(dealloc_pd_in, in, pd, pdn);
pd                 71 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 				 u32 pd, u64 length, u64 start_addr, int mode,
pd                 90 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 	MLX5_SET(mkc, mkc, pd, pd);
pd                164 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	MLX5_SET(qpc, qpc, pd, attr->pdn);
pd                811 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	MLX5_SET(mkc, mkc, pd, pdn);
pd                262 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	const struct mlxfw_mfa2_tlv_package_descriptor *pd;
pd                292 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	pd = mlxfw_mfa2_tlv_package_descriptor_get(mfa2_file, multi_child);
pd                293 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	if (!pd) {
pd                304 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	mfa2_file->dev_count = be16_to_cpu(pd->num_devices);
pd                308 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	mfa2_file->component_count = be16_to_cpu(pd->num_components);
pd                309 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	mfa2_file->cb = fw->data + NLA_ALIGN(be32_to_cpu(pd->cb_offset));
pd                314 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c 	mfa2_file->cb_archive_size = be32_to_cpu(pd->cb_archive_size);
pd               7273 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 pd;
pd               7922 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 pd;
pd               7984 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 pd;
pd               9957 drivers/net/ethernet/qlogic/qed/qed_hsi.h 	__le16 pd;
pd                275 drivers/net/ethernet/qlogic/qed/qed_iwarp.c 	p_ramrod->pd = qp->pd;
pd                954 drivers/net/ethernet/qlogic/qed/qed_rdma.c static int qed_rdma_alloc_pd(void *rdma_cxt, u16 *pd)
pd                968 drivers/net/ethernet/qlogic/qed/qed_rdma.c 	*pd = (u16)returned_id;
pd                974 drivers/net/ethernet/qlogic/qed/qed_rdma.c static void qed_rdma_free_pd(void *rdma_cxt, u16 pd)
pd                978 drivers/net/ethernet/qlogic/qed/qed_rdma.c 	DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "pd = %08x\n", pd);
pd                982 drivers/net/ethernet/qlogic/qed/qed_rdma.c 	qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->pd_map, pd);
pd               1296 drivers/net/ethernet/qlogic/qed/qed_rdma.c 	qp->pd = in_params->pd;
pd               1533 drivers/net/ethernet/qlogic/qed/qed_rdma.c 	p_ramrod->pd = cpu_to_le16(params->pd);
pd                124 drivers/net/ethernet/qlogic/qed/qed_rdma.h 	u16 pd;
pd                327 drivers/net/ethernet/qlogic/qed/qed_roce.c 	p_ramrod->pd = cpu_to_le16(qp->pd);
pd                457 drivers/net/ethernet/qlogic/qed/qed_roce.c 	p_ramrod->pd = cpu_to_le16(qp->pd);
pd               3087 drivers/net/ethernet/renesas/sh_eth.c 			struct sh_eth_plat_data *pd)
pd               3101 drivers/net/ethernet/renesas/sh_eth.c 	bitbang->set_gate = pd->set_mdio_gate;
pd               3116 drivers/net/ethernet/renesas/sh_eth.c 	if (pd->phy_irq > 0)
pd               3117 drivers/net/ethernet/renesas/sh_eth.c 		mdp->mii_bus->irq[pd->phy] = pd->phy_irq;
pd               3242 drivers/net/ethernet/renesas/sh_eth.c 	struct sh_eth_plat_data *pd = dev_get_platdata(&pdev->dev);
pd               3280 drivers/net/ethernet/renesas/sh_eth.c 		pd = sh_eth_parse_dt(&pdev->dev);
pd               3281 drivers/net/ethernet/renesas/sh_eth.c 	if (!pd) {
pd               3288 drivers/net/ethernet/renesas/sh_eth.c 	mdp->phy_id = pd->phy;
pd               3289 drivers/net/ethernet/renesas/sh_eth.c 	mdp->phy_interface = pd->phy_interface;
pd               3290 drivers/net/ethernet/renesas/sh_eth.c 	mdp->no_ether_link = pd->no_ether_link;
pd               3291 drivers/net/ethernet/renesas/sh_eth.c 	mdp->ether_link_active_low = pd->ether_link_active_low;
pd               3332 drivers/net/ethernet/renesas/sh_eth.c 	read_mac_address(ndev, pd->mac_addr);
pd               3385 drivers/net/ethernet/renesas/sh_eth.c 	ret = sh_mdio_init(mdp, pd);
pd                724 drivers/net/ethernet/seeq/sgiseeq.c 	struct sgiseeq_platform_data *pd = dev_get_platdata(&pdev->dev);
pd                725 drivers/net/ethernet/seeq/sgiseeq.c 	struct hpc3_regs *hpcregs = pd->hpc;
pd                727 drivers/net/ethernet/seeq/sgiseeq.c 	unsigned int irq = pd->irq;
pd                759 drivers/net/ethernet/seeq/sgiseeq.c 	memcpy(dev->dev_addr, pd->mac, ETH_ALEN);
pd                255 drivers/net/ethernet/sfc/falcon/tenxpress.c 	struct tenxpress_phy_data *pd = efx->phy_data;
pd                268 drivers/net/ethernet/sfc/falcon/tenxpress.c 			pd->bad_lp_tries++;
pd                272 drivers/net/ethernet/sfc/falcon/tenxpress.c 	if (!pd->bad_lp_tries)
pd                277 drivers/net/ethernet/sfc/falcon/tenxpress.c 	if (!bad_lp || pd->bad_lp_tries == MAX_BAD_LP_TRIES) {
pd                293 drivers/net/ethernet/sfc/falcon/tenxpress.c 		pd->bad_lp_tries = bad_lp;
pd                361 drivers/net/ethernet/sfc/falcon/txc43128_phy.c 	int pd = (1 << TXC_GLCMD_L01PD_LBN) | (1 << TXC_GLCMD_L23PD_LBN);
pd                365 drivers/net/ethernet/sfc/falcon/txc43128_phy.c 		ctl &= ~pd;
pd                367 drivers/net/ethernet/sfc/falcon/txc43128_phy.c 		ctl |= pd;
pd               2076 drivers/net/ethernet/smsc/smc911x.c 		struct smc911x_platdata *pd = dev_get_platdata(&pdev->dev);
pd               2077 drivers/net/ethernet/smsc/smc911x.c 		if (!pd) {
pd               2081 drivers/net/ethernet/smsc/smc911x.c 		memcpy(&lp->cfg, pd, sizeof(lp->cfg));
pd               2232 drivers/net/ethernet/smsc/smc91x.c 	struct smc91x_platdata *pd = dev_get_platdata(&pdev->dev);
pd               2256 drivers/net/ethernet/smsc/smc91x.c 	if (pd) {
pd               2257 drivers/net/ethernet/smsc/smc91x.c 		memcpy(&lp->cfg, pd, sizeof(lp->cfg));
pd               2315 drivers/net/ethernet/smsc/smc91x.c 	if (!pd && !match) {
pd                 86 drivers/net/ethernet/smsc/smsc9420.c static inline u32 smsc9420_reg_read(struct smsc9420_pdata *pd, u32 offset)
pd                 88 drivers/net/ethernet/smsc/smsc9420.c 	return ioread32(pd->ioaddr + offset);
pd                 92 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(struct smsc9420_pdata *pd, u32 offset, u32 value)
pd                 94 drivers/net/ethernet/smsc/smsc9420.c 	iowrite32(value, pd->ioaddr + offset);
pd                 97 drivers/net/ethernet/smsc/smsc9420.c static inline void smsc9420_pci_flush_write(struct smsc9420_pdata *pd)
pd                100 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_read(pd, ID_REV);
pd                105 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = (struct smsc9420_pdata *)bus->priv;
pd                110 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->phy_lock, flags);
pd                113 drivers/net/ethernet/smsc/smsc9420.c 	if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) {
pd                114 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, drv, pd->dev, "MII is busy???\n");
pd                121 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MII_ACCESS, addr);
pd                125 drivers/net/ethernet/smsc/smsc9420.c 		if (!(smsc9420_reg_read(pd, MII_ACCESS) &
pd                127 drivers/net/ethernet/smsc/smsc9420.c 			reg = (u16)smsc9420_reg_read(pd, MII_DATA);
pd                133 drivers/net/ethernet/smsc/smsc9420.c 	netif_warn(pd, drv, pd->dev, "MII busy timeout!\n");
pd                136 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->phy_lock, flags);
pd                143 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = (struct smsc9420_pdata *)bus->priv;
pd                148 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->phy_lock, flags);
pd                151 drivers/net/ethernet/smsc/smsc9420.c 	if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) {
pd                152 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, drv, pd->dev, "MII is busy???\n");
pd                157 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MII_DATA, (u32)val);
pd                162 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MII_ACCESS, addr);
pd                166 drivers/net/ethernet/smsc/smsc9420.c 		if (!(smsc9420_reg_read(pd, MII_ACCESS) &
pd                174 drivers/net/ethernet/smsc/smsc9420.c 	netif_warn(pd, drv, pd->dev, "MII busy timeout!\n");
pd                177 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->phy_lock, flags);
pd                189 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_reload(struct smsc9420_pdata *pd)
pd                193 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd);
pd                195 drivers/net/ethernet/smsc/smsc9420.c 	if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) {
pd                196 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, drv, pd->dev, "%s: Eeprom busy\n", __func__);
pd                200 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, E2P_CMD,
pd                205 drivers/net/ethernet/smsc/smsc9420.c 		if (!(smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_))
pd                209 drivers/net/ethernet/smsc/smsc9420.c 	netif_warn(pd, drv, pd->dev, "%s: Eeprom timed out\n", __func__);
pd                225 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(netdev);
pd                228 drivers/net/ethernet/smsc/smsc9420.c 	strlcpy(drvinfo->bus_info, pci_name(pd->pdev),
pd                235 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(netdev);
pd                236 drivers/net/ethernet/smsc/smsc9420.c 	return pd->msg_enable;
pd                241 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(netdev);
pd                242 drivers/net/ethernet/smsc/smsc9420.c 	pd->msg_enable = data;
pd                255 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                260 drivers/net/ethernet/smsc/smsc9420.c 	regs->version = smsc9420_reg_read(pd, ID_REV);
pd                262 drivers/net/ethernet/smsc/smsc9420.c 		data[j++] = smsc9420_reg_read(pd, i);
pd                273 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_eeprom_enable_access(struct smsc9420_pdata *pd)
pd                275 drivers/net/ethernet/smsc/smsc9420.c 	unsigned int temp = smsc9420_reg_read(pd, GPIO_CFG);
pd                277 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, GPIO_CFG, temp);
pd                281 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_send_cmd(struct smsc9420_pdata *pd, u32 op)
pd                286 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, hw, pd->dev, "op 0x%08x\n", op);
pd                287 drivers/net/ethernet/smsc/smsc9420.c 	if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) {
pd                288 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, hw, pd->dev, "Busy at start\n");
pd                293 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, E2P_CMD, e2cmd);
pd                297 drivers/net/ethernet/smsc/smsc9420.c 		e2cmd = smsc9420_reg_read(pd, E2P_CMD);
pd                301 drivers/net/ethernet/smsc/smsc9420.c 		netif_info(pd, hw, pd->dev, "TIMED OUT\n");
pd                306 drivers/net/ethernet/smsc/smsc9420.c 		netif_info(pd, hw, pd->dev,
pd                314 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_read_location(struct smsc9420_pdata *pd,
pd                320 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, hw, pd->dev, "address 0x%x\n", address);
pd                321 drivers/net/ethernet/smsc/smsc9420.c 	ret = smsc9420_eeprom_send_cmd(pd, op);
pd                324 drivers/net/ethernet/smsc/smsc9420.c 		data[address] = smsc9420_reg_read(pd, E2P_DATA);
pd                329 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_write_location(struct smsc9420_pdata *pd,
pd                335 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, hw, pd->dev, "address 0x%x, data 0x%x\n", address, data);
pd                336 drivers/net/ethernet/smsc/smsc9420.c 	ret = smsc9420_eeprom_send_cmd(pd, op);
pd                340 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, E2P_DATA, (u32)data);
pd                341 drivers/net/ethernet/smsc/smsc9420.c 		ret = smsc9420_eeprom_send_cmd(pd, op);
pd                355 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                359 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_eeprom_enable_access(pd);
pd                363 drivers/net/ethernet/smsc/smsc9420.c 		int ret = smsc9420_eeprom_read_location(pd, i, eeprom_data);
pd                379 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                385 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_eeprom_enable_access(pd);
pd                386 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_eeprom_send_cmd(pd, E2P_CMD_EPC_CMD_EWEN_);
pd                387 drivers/net/ethernet/smsc/smsc9420.c 	ret = smsc9420_eeprom_write_location(pd, eeprom->offset, *data);
pd                388 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_eeprom_send_cmd(pd, E2P_CMD_EPC_CMD_EWDS_);
pd                415 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                421 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, ADDRH, mac_high16);
pd                422 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, ADDRL, mac_low32);
pd                427 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                432 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, probe, pd->dev,
pd                437 drivers/net/ethernet/smsc/smsc9420.c 		u32 mac_high16 = smsc9420_reg_read(pd, ADDRH);
pd                438 drivers/net/ethernet/smsc/smsc9420.c 		u32 mac_low32 = smsc9420_reg_read(pd, ADDRL);
pd                448 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, probe, pd->dev,
pd                454 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, probe, pd->dev,
pd                460 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_stop_tx(struct smsc9420_pdata *pd)
pd                466 drivers/net/ethernet/smsc/smsc9420.c 	dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL);
pd                468 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control);
pd                472 drivers/net/ethernet/smsc/smsc9420.c 		if (smsc9420_reg_read(pd, DMAC_STATUS) & DMAC_STS_TS_)
pd                478 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifdown, pd->dev, "TX DMAC failed to stop\n");
pd                481 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_TXPS_);
pd                484 drivers/net/ethernet/smsc/smsc9420.c 	dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA);
pd                486 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena);
pd                487 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                490 drivers/net/ethernet/smsc/smsc9420.c 	mac_cr = smsc9420_reg_read(pd, MAC_CR) & (~MAC_CR_TXEN_);
pd                491 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MAC_CR, mac_cr);
pd                492 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                495 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_free_tx_ring(struct smsc9420_pdata *pd)
pd                499 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->tx_ring);
pd                501 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->tx_buffers)
pd                505 drivers/net/ethernet/smsc/smsc9420.c 		struct sk_buff *skb = pd->tx_buffers[i].skb;
pd                508 drivers/net/ethernet/smsc/smsc9420.c 			BUG_ON(!pd->tx_buffers[i].mapping);
pd                509 drivers/net/ethernet/smsc/smsc9420.c 			pci_unmap_single(pd->pdev, pd->tx_buffers[i].mapping,
pd                514 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].status = 0;
pd                515 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].length = 0;
pd                516 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].buffer1 = 0;
pd                517 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].buffer2 = 0;
pd                521 drivers/net/ethernet/smsc/smsc9420.c 	kfree(pd->tx_buffers);
pd                522 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_buffers = NULL;
pd                524 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring_head = 0;
pd                525 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring_tail = 0;
pd                528 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_free_rx_ring(struct smsc9420_pdata *pd)
pd                532 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->rx_ring);
pd                534 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->rx_buffers)
pd                538 drivers/net/ethernet/smsc/smsc9420.c 		if (pd->rx_buffers[i].skb)
pd                539 drivers/net/ethernet/smsc/smsc9420.c 			dev_kfree_skb_any(pd->rx_buffers[i].skb);
pd                541 drivers/net/ethernet/smsc/smsc9420.c 		if (pd->rx_buffers[i].mapping)
pd                542 drivers/net/ethernet/smsc/smsc9420.c 			pci_unmap_single(pd->pdev, pd->rx_buffers[i].mapping,
pd                545 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].status = 0;
pd                546 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].length = 0;
pd                547 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].buffer1 = 0;
pd                548 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].buffer2 = 0;
pd                552 drivers/net/ethernet/smsc/smsc9420.c 	kfree(pd->rx_buffers);
pd                553 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers = NULL;
pd                555 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring_head = 0;
pd                556 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring_tail = 0;
pd                559 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_stop_rx(struct smsc9420_pdata *pd)
pd                565 drivers/net/ethernet/smsc/smsc9420.c 	dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA);
pd                567 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena);
pd                568 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                571 drivers/net/ethernet/smsc/smsc9420.c 	mac_cr = smsc9420_reg_read(pd, MAC_CR) & (~MAC_CR_RXEN_);
pd                572 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MAC_CR, mac_cr);
pd                573 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                576 drivers/net/ethernet/smsc/smsc9420.c 	dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL);
pd                578 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control);
pd                579 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                583 drivers/net/ethernet/smsc/smsc9420.c 		if (smsc9420_reg_read(pd, DMAC_STATUS) & DMAC_STS_RS_)
pd                589 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifdown, pd->dev,
pd                593 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_RXPS_);
pd                598 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = dev_id;
pd                603 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd);
pd                604 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->ioaddr);
pd                606 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG);
pd                613 drivers/net/ethernet/smsc/smsc9420.c 	int_sts = smsc9420_reg_read(pd, INT_STAT);
pd                616 drivers/net/ethernet/smsc/smsc9420.c 		u32 status = smsc9420_reg_read(pd, DMAC_STATUS);
pd                621 drivers/net/ethernet/smsc/smsc9420.c 			netif_wake_queue(pd->dev);
pd                626 drivers/net/ethernet/smsc/smsc9420.c 			u32 dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA);
pd                628 drivers/net/ethernet/smsc/smsc9420.c 			smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena);
pd                629 drivers/net/ethernet/smsc/smsc9420.c 			smsc9420_pci_flush_write(pd);
pd                632 drivers/net/ethernet/smsc/smsc9420.c 			napi_schedule(&pd->napi);
pd                636 drivers/net/ethernet/smsc/smsc9420.c 			smsc9420_reg_write(pd, DMAC_STATUS, ints_to_clear);
pd                643 drivers/net/ethernet/smsc/smsc9420.c 		spin_lock_irqsave(&pd->int_lock, flags);
pd                644 drivers/net/ethernet/smsc/smsc9420.c 		int_ctl = smsc9420_reg_read(pd, INT_CTL);
pd                646 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, INT_CTL, int_ctl);
pd                647 drivers/net/ethernet/smsc/smsc9420.c 		spin_unlock_irqrestore(&pd->int_lock, flags);
pd                649 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, INT_STAT, INT_STAT_SW_INT_);
pd                650 drivers/net/ethernet/smsc/smsc9420.c 		pd->software_irq_signal = true;
pd                657 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                665 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                666 drivers/net/ethernet/smsc/smsc9420.c 	const int irq = pd->pdev->irq;
pd                674 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_dmac_soft_reset(struct smsc9420_pdata *pd)
pd                676 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, BUS_MODE, BUS_MODE_SWR_);
pd                677 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_read(pd, BUS_MODE);
pd                679 drivers/net/ethernet/smsc/smsc9420.c 	if (smsc9420_reg_read(pd, BUS_MODE) & BUS_MODE_SWR_)
pd                680 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, drv, pd->dev, "Software reset not cleared\n");
pd                685 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                689 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd);
pd                693 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd                694 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_);
pd                695 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd                696 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd                699 drivers/net/ethernet/smsc/smsc9420.c 	napi_disable(&pd->napi);
pd                701 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_stop_tx(pd);
pd                702 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_free_tx_ring(pd);
pd                704 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_stop_rx(pd);
pd                705 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_free_rx_ring(pd);
pd                707 drivers/net/ethernet/smsc/smsc9420.c 	free_irq(pd->pdev->irq, pd);
pd                709 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_dmac_soft_reset(pd);
pd                714 drivers/net/ethernet/smsc/smsc9420.c 	mdiobus_unregister(pd->mii_bus);
pd                715 drivers/net/ethernet/smsc/smsc9420.c 	mdiobus_free(pd->mii_bus);
pd                744 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_rx_handoff(struct smsc9420_pdata *pd, const int index,
pd                747 drivers/net/ethernet/smsc/smsc9420.c 	struct net_device *dev = pd->dev;
pd                755 drivers/net/ethernet/smsc/smsc9420.c 	if (pd->rx_csum)
pd                761 drivers/net/ethernet/smsc/smsc9420.c 	pci_unmap_single(pd->pdev, pd->rx_buffers[index].mapping,
pd                763 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers[index].mapping = 0;
pd                765 drivers/net/ethernet/smsc/smsc9420.c 	skb = pd->rx_buffers[index].skb;
pd                766 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers[index].skb = NULL;
pd                768 drivers/net/ethernet/smsc/smsc9420.c 	if (pd->rx_csum) {
pd                783 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_rx_buffer(struct smsc9420_pdata *pd, int index)
pd                785 drivers/net/ethernet/smsc/smsc9420.c 	struct sk_buff *skb = netdev_alloc_skb(pd->dev, PKT_BUF_SZ);
pd                788 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->rx_buffers[index].skb);
pd                789 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->rx_buffers[index].mapping);
pd                794 drivers/net/ethernet/smsc/smsc9420.c 	mapping = pci_map_single(pd->pdev, skb_tail_pointer(skb),
pd                796 drivers/net/ethernet/smsc/smsc9420.c 	if (pci_dma_mapping_error(pd->pdev, mapping)) {
pd                798 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, rx_err, pd->dev, "pci_map_single failed!\n");
pd                802 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers[index].skb = skb;
pd                803 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers[index].mapping = mapping;
pd                804 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring[index].buffer1 = mapping + NET_IP_ALIGN;
pd                805 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring[index].status = RDES0_OWN_;
pd                811 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_alloc_new_rx_buffers(struct smsc9420_pdata *pd)
pd                813 drivers/net/ethernet/smsc/smsc9420.c 	while (pd->rx_ring_tail != pd->rx_ring_head) {
pd                814 drivers/net/ethernet/smsc/smsc9420.c 		if (smsc9420_alloc_rx_buffer(pd, pd->rx_ring_tail))
pd                817 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring_tail = (pd->rx_ring_tail + 1) % RX_RING_SIZE;
pd                823 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd =
pd                825 drivers/net/ethernet/smsc/smsc9420.c 	struct net_device *dev = pd->dev;
pd                831 drivers/net/ethernet/smsc/smsc9420.c 		status = pd->rx_ring[pd->rx_ring_head].status;
pd                838 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_rx_handoff(pd, pd->rx_ring_head, status);
pd                839 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring_head = (pd->rx_ring_head + 1) % RX_RING_SIZE;
pd                840 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_alloc_new_rx_buffers(pd);
pd                843 drivers/net/ethernet/smsc/smsc9420.c 	drop_frame_cnt = smsc9420_reg_read(pd, MISS_FRAME_CNTR);
pd                848 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, RX_POLL_DEMAND, 1);
pd                849 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                852 drivers/net/ethernet/smsc/smsc9420.c 		napi_complete_done(&pd->napi, work_done);
pd                855 drivers/net/ethernet/smsc/smsc9420.c 		dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA);
pd                857 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena);
pd                858 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_pci_flush_write(pd);
pd                894 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                896 drivers/net/ethernet/smsc/smsc9420.c 	while (pd->tx_ring_tail != pd->tx_ring_head) {
pd                897 drivers/net/ethernet/smsc/smsc9420.c 		int index = pd->tx_ring_tail;
pd                901 drivers/net/ethernet/smsc/smsc9420.c 		status = pd->tx_ring[index].status;
pd                902 drivers/net/ethernet/smsc/smsc9420.c 		length = pd->tx_ring[index].length;
pd                910 drivers/net/ethernet/smsc/smsc9420.c 		BUG_ON(!pd->tx_buffers[index].skb);
pd                911 drivers/net/ethernet/smsc/smsc9420.c 		BUG_ON(!pd->tx_buffers[index].mapping);
pd                913 drivers/net/ethernet/smsc/smsc9420.c 		pci_unmap_single(pd->pdev, pd->tx_buffers[index].mapping,
pd                914 drivers/net/ethernet/smsc/smsc9420.c 			pd->tx_buffers[index].skb->len, PCI_DMA_TODEVICE);
pd                915 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_buffers[index].mapping = 0;
pd                917 drivers/net/ethernet/smsc/smsc9420.c 		dev_kfree_skb_any(pd->tx_buffers[index].skb);
pd                918 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_buffers[index].skb = NULL;
pd                920 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[index].buffer1 = 0;
pd                923 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring_tail = (pd->tx_ring_tail + 1) % TX_RING_SIZE;
pd                930 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                932 drivers/net/ethernet/smsc/smsc9420.c 	int index = pd->tx_ring_head;
pd                935 drivers/net/ethernet/smsc/smsc9420.c 		(((pd->tx_ring_head + 2) % TX_RING_SIZE) == pd->tx_ring_tail);
pd                940 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->tx_ring[index].status & TDES0_OWN_);
pd                941 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->tx_buffers[index].skb);
pd                942 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->tx_buffers[index].mapping);
pd                944 drivers/net/ethernet/smsc/smsc9420.c 	mapping = pci_map_single(pd->pdev, skb->data,
pd                946 drivers/net/ethernet/smsc/smsc9420.c 	if (pci_dma_mapping_error(pd->pdev, mapping)) {
pd                947 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, tx_err, pd->dev,
pd                952 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_buffers[index].skb = skb;
pd                953 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_buffers[index].mapping = mapping;
pd                958 drivers/net/ethernet/smsc/smsc9420.c 		netif_stop_queue(pd->dev);
pd                965 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring[index].buffer1 = mapping;
pd                966 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring[index].length = tmp_desc1;
pd                970 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring_head = (pd->tx_ring_head + 1) % TX_RING_SIZE;
pd                973 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring[index].status = TDES0_OWN_;
pd                979 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, TX_POLL_DEMAND, 1);
pd                980 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd                987 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                988 drivers/net/ethernet/smsc/smsc9420.c 	u32 counter = smsc9420_reg_read(pd, MISS_FRAME_CNTR);
pd                996 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd                997 drivers/net/ethernet/smsc/smsc9420.c 	u32 mac_cr = smsc9420_reg_read(pd, MAC_CR);
pd               1000 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, hw, pd->dev, "Promiscuous Mode Enabled\n");
pd               1005 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, hw, pd->dev, "Receive all Multicast Enabled\n");
pd               1013 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, hw, pd->dev, "Multicast filter enabled\n");
pd               1024 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, HASHH, hash_hi);
pd               1025 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, HASHL, hash_lo);
pd               1031 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, hw, pd->dev, "Receive own packets only\n");
pd               1032 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, HASHH, 0);
pd               1033 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, HASHL, 0);
pd               1040 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MAC_CR, mac_cr);
pd               1041 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1044 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_phy_update_flowcontrol(struct smsc9420_pdata *pd)
pd               1046 drivers/net/ethernet/smsc/smsc9420.c 	struct net_device *dev = pd->dev;
pd               1060 drivers/net/ethernet/smsc/smsc9420.c 		netif_info(pd, link, pd->dev, "rx pause %s, tx pause %s\n",
pd               1064 drivers/net/ethernet/smsc/smsc9420.c 		netif_info(pd, link, pd->dev, "half duplex\n");
pd               1068 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, FLOW, flow);
pd               1075 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1079 drivers/net/ethernet/smsc/smsc9420.c 	if (phy_dev->duplex != pd->last_duplex) {
pd               1080 drivers/net/ethernet/smsc/smsc9420.c 		u32 mac_cr = smsc9420_reg_read(pd, MAC_CR);
pd               1082 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, link, pd->dev, "full duplex mode\n");
pd               1085 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, link, pd->dev, "half duplex mode\n");
pd               1088 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, MAC_CR, mac_cr);
pd               1090 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_phy_update_flowcontrol(pd);
pd               1091 drivers/net/ethernet/smsc/smsc9420.c 		pd->last_duplex = phy_dev->duplex;
pd               1095 drivers/net/ethernet/smsc/smsc9420.c 	if (carrier != pd->last_carrier) {
pd               1097 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, link, pd->dev, "carrier OK\n");
pd               1099 drivers/net/ethernet/smsc/smsc9420.c 			netif_dbg(pd, link, pd->dev, "no carrier\n");
pd               1100 drivers/net/ethernet/smsc/smsc9420.c 		pd->last_carrier = carrier;
pd               1106 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1112 drivers/net/ethernet/smsc/smsc9420.c 	phydev = mdiobus_get_phy(pd->mii_bus, 1);
pd               1133 drivers/net/ethernet/smsc/smsc9420.c 	pd->last_duplex = -1;
pd               1134 drivers/net/ethernet/smsc/smsc9420.c 	pd->last_carrier = -1;
pd               1141 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1144 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus = mdiobus_alloc();
pd               1145 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->mii_bus) {
pd               1149 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus->name = DRV_MDIONAME;
pd               1150 drivers/net/ethernet/smsc/smsc9420.c 	snprintf(pd->mii_bus->id, MII_BUS_ID_SIZE, "%x",
pd               1151 drivers/net/ethernet/smsc/smsc9420.c 		(pd->pdev->bus->number << 8) | pd->pdev->devfn);
pd               1152 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus->priv = pd;
pd               1153 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus->read = smsc9420_mii_read;
pd               1154 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus->write = smsc9420_mii_write;
pd               1157 drivers/net/ethernet/smsc/smsc9420.c 	pd->mii_bus->phy_mask = ~(1 << 1);
pd               1159 drivers/net/ethernet/smsc/smsc9420.c 	if (mdiobus_register(pd->mii_bus)) {
pd               1160 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, probe, pd->dev, "Error registering mii bus\n");
pd               1165 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, probe, pd->dev, "Error probing mii bus\n");
pd               1172 drivers/net/ethernet/smsc/smsc9420.c 	mdiobus_unregister(pd->mii_bus);
pd               1174 drivers/net/ethernet/smsc/smsc9420.c 	mdiobus_free(pd->mii_bus);
pd               1179 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_tx_ring(struct smsc9420_pdata *pd)
pd               1183 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->tx_ring);
pd               1185 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_buffers = kmalloc_array(TX_RING_SIZE,
pd               1188 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->tx_buffers)
pd               1193 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_buffers[i].skb = NULL;
pd               1194 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_buffers[i].mapping = 0;
pd               1195 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].status = 0;
pd               1196 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].length = 0;
pd               1197 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].buffer1 = 0;
pd               1198 drivers/net/ethernet/smsc/smsc9420.c 		pd->tx_ring[i].buffer2 = 0;
pd               1200 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring[TX_RING_SIZE - 1].length = TDES1_TER_;
pd               1203 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring_head = 0;
pd               1204 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring_tail = 0;
pd               1206 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, TX_BASE_ADDR, pd->tx_dma_addr);
pd               1207 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1212 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_rx_ring(struct smsc9420_pdata *pd)
pd               1216 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->rx_ring);
pd               1218 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_buffers = kmalloc_array(RX_RING_SIZE,
pd               1221 drivers/net/ethernet/smsc/smsc9420.c 	if (pd->rx_buffers == NULL)
pd               1226 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].status = 0;
pd               1227 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].length = PKT_BUF_SZ;
pd               1228 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_ring[i].buffer2 = 0;
pd               1229 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_buffers[i].skb = NULL;
pd               1230 drivers/net/ethernet/smsc/smsc9420.c 		pd->rx_buffers[i].mapping = 0;
pd               1232 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring[RX_RING_SIZE - 1].length = (PKT_BUF_SZ | RDES1_RER_);
pd               1236 drivers/net/ethernet/smsc/smsc9420.c 		if (smsc9420_alloc_rx_buffer(pd, i)) {
pd               1237 drivers/net/ethernet/smsc/smsc9420.c 			netif_warn(pd, ifup, pd->dev,
pd               1243 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring_head = 0;
pd               1244 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring_tail = 0;
pd               1246 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, VLAN1, ETH_P_8021Q);
pd               1247 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, ifup, pd->dev, "VLAN1 = 0x%08x\n",
pd               1248 drivers/net/ethernet/smsc/smsc9420.c 		  smsc9420_reg_read(pd, VLAN1));
pd               1250 drivers/net/ethernet/smsc/smsc9420.c 	if (pd->rx_csum) {
pd               1252 drivers/net/ethernet/smsc/smsc9420.c 		u32 coe = smsc9420_reg_read(pd, COE_CR) | RX_COE_EN;
pd               1253 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_reg_write(pd, COE_CR, coe);
pd               1254 drivers/net/ethernet/smsc/smsc9420.c 		netif_dbg(pd, ifup, pd->dev, "COE_CR = 0x%08x\n", coe);
pd               1257 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, RX_BASE_ADDR, pd->rx_dma_addr);
pd               1258 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1263 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_free_rx_ring(pd);
pd               1270 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1272 drivers/net/ethernet/smsc/smsc9420.c 	const int irq = pd->pdev->irq;
pd               1277 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev,
pd               1286 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd               1287 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_);
pd               1288 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd               1289 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CTL, 0);
pd               1290 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd               1291 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_INTR_ENA, 0);
pd               1292 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_STAT, 0xFFFFFFFF);
pd               1293 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1295 drivers/net/ethernet/smsc/smsc9420.c 	result = request_irq(irq, smsc9420_isr, IRQF_SHARED, DRV_NAME, pd);
pd               1297 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev, "Unable to use IRQ = %d\n", irq);
pd               1302 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_dmac_soft_reset(pd);
pd               1305 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MAC_CR, 0);
pd               1310 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, GPIO_CFG,
pd               1319 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, BUS_MODE, bus_mode);
pd               1321 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1324 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, BUS_CFG, BUS_CFG_RXTXWEIGHT_4_1);
pd               1326 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_CONTROL,
pd               1329 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1332 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, ifup, pd->dev, "Testing ISR using IRQ %d\n", irq);
pd               1333 drivers/net/ethernet/smsc/smsc9420.c 	pd->software_irq_signal = false;
pd               1335 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd               1337 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) | INT_CFG_IRQ_EN_;
pd               1340 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd               1343 drivers/net/ethernet/smsc/smsc9420.c 	int_ctl = smsc9420_reg_read(pd, INT_CTL) | INT_CTL_SW_INT_EN_;
pd               1344 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CTL, int_ctl);
pd               1345 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd               1346 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1350 drivers/net/ethernet/smsc/smsc9420.c 		if (pd->software_irq_signal)
pd               1356 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd               1357 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_);
pd               1358 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd               1359 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd               1361 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->software_irq_signal) {
pd               1362 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev, "ISR failed signaling test\n");
pd               1367 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, ifup, pd->dev, "ISR passed test using IRQ %d\n", irq);
pd               1369 drivers/net/ethernet/smsc/smsc9420.c 	result = smsc9420_alloc_tx_ring(pd);
pd               1371 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev,
pd               1377 drivers/net/ethernet/smsc/smsc9420.c 	result = smsc9420_alloc_rx_ring(pd);
pd               1379 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev,
pd               1387 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev, "Failed to initialize Phy\n");
pd               1395 drivers/net/ethernet/smsc/smsc9420.c 	napi_enable(&pd->napi);
pd               1398 drivers/net/ethernet/smsc/smsc9420.c 	mac_cr = smsc9420_reg_read(pd, MAC_CR) | MAC_CR_TXEN_ | MAC_CR_RXEN_;
pd               1399 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, MAC_CR, mac_cr);
pd               1401 drivers/net/ethernet/smsc/smsc9420.c 	dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL);
pd               1403 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control);
pd               1404 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1406 drivers/net/ethernet/smsc/smsc9420.c 	dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA);
pd               1409 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena);
pd               1410 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_pci_flush_write(pd);
pd               1414 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, RX_POLL_DEMAND, 1);
pd               1417 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd               1418 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) | INT_CFG_IRQ_EN_;
pd               1419 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd               1420 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd               1425 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_free_rx_ring(pd);
pd               1427 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_free_tx_ring(pd);
pd               1429 drivers/net/ethernet/smsc/smsc9420.c 	free_irq(irq, pd);
pd               1439 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1444 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_irqsave(&pd->int_lock, flags);
pd               1445 drivers/net/ethernet/smsc/smsc9420.c 	int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_);
pd               1446 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_reg_write(pd, INT_CFG, int_cfg);
pd               1447 drivers/net/ethernet/smsc/smsc9420.c 	spin_unlock_irqrestore(&pd->int_lock, flags);
pd               1451 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_stop_tx(pd);
pd               1452 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_free_tx_ring(pd);
pd               1454 drivers/net/ethernet/smsc/smsc9420.c 		napi_disable(&pd->napi);
pd               1455 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_stop_rx(pd);
pd               1456 drivers/net/ethernet/smsc/smsc9420.c 		smsc9420_free_rx_ring(pd);
pd               1458 drivers/net/ethernet/smsc/smsc9420.c 		free_irq(pd->pdev->irq, pd);
pd               1474 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd = netdev_priv(dev);
pd               1488 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, ifup, pd->dev, "pci_enable_wake failed: %d\n",
pd               1519 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd;
pd               1535 drivers/net/ethernet/smsc/smsc9420.c 	dev = alloc_etherdev(sizeof(*pd));
pd               1566 drivers/net/ethernet/smsc/smsc9420.c 	pd = netdev_priv(dev);
pd               1569 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_ring = pci_alloc_consistent(pdev,
pd               1572 drivers/net/ethernet/smsc/smsc9420.c 		&pd->rx_dma_addr);
pd               1574 drivers/net/ethernet/smsc/smsc9420.c 	if (!pd->rx_ring)
pd               1578 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_ring = (pd->rx_ring + RX_RING_SIZE);
pd               1579 drivers/net/ethernet/smsc/smsc9420.c 	pd->tx_dma_addr = pd->rx_dma_addr +
pd               1582 drivers/net/ethernet/smsc/smsc9420.c 	pd->pdev = pdev;
pd               1583 drivers/net/ethernet/smsc/smsc9420.c 	pd->dev = dev;
pd               1584 drivers/net/ethernet/smsc/smsc9420.c 	pd->ioaddr = virt_addr;
pd               1585 drivers/net/ethernet/smsc/smsc9420.c 	pd->msg_enable = smsc_debug;
pd               1586 drivers/net/ethernet/smsc/smsc9420.c 	pd->rx_csum = true;
pd               1588 drivers/net/ethernet/smsc/smsc9420.c 	netif_dbg(pd, probe, pd->dev, "lan_base=0x%08lx\n", (ulong)virt_addr);
pd               1590 drivers/net/ethernet/smsc/smsc9420.c 	id_rev = smsc9420_reg_read(pd, ID_REV);
pd               1593 drivers/net/ethernet/smsc/smsc9420.c 		netif_info(pd, probe, pd->dev,
pd               1597 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, probe, pd->dev, "LAN9420 NOT identified\n");
pd               1598 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, probe, pd->dev, "ID_REV=0x%08X\n", id_rev);
pd               1602 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_dmac_soft_reset(pd);
pd               1603 drivers/net/ethernet/smsc/smsc9420.c 	smsc9420_eeprom_reload(pd);
pd               1609 drivers/net/ethernet/smsc/smsc9420.c 	netif_napi_add(dev, &pd->napi, smsc9420_rx_poll, NAPI_WEIGHT);
pd               1613 drivers/net/ethernet/smsc/smsc9420.c 		netif_warn(pd, probe, pd->dev, "error %i registering device\n",
pd               1620 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_init(&pd->int_lock);
pd               1621 drivers/net/ethernet/smsc/smsc9420.c 	spin_lock_init(&pd->phy_lock);
pd               1629 drivers/net/ethernet/smsc/smsc9420.c 		(RX_RING_SIZE + TX_RING_SIZE), pd->rx_ring, pd->rx_dma_addr);
pd               1645 drivers/net/ethernet/smsc/smsc9420.c 	struct smsc9420_pdata *pd;
pd               1651 drivers/net/ethernet/smsc/smsc9420.c 	pd = netdev_priv(dev);
pd               1655 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->tx_buffers);
pd               1656 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(pd->rx_buffers);
pd               1658 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->tx_ring);
pd               1659 drivers/net/ethernet/smsc/smsc9420.c 	BUG_ON(!pd->rx_ring);
pd               1662 drivers/net/ethernet/smsc/smsc9420.c 		(RX_RING_SIZE + TX_RING_SIZE), pd->rx_ring, pd->rx_dma_addr);
pd               1664 drivers/net/ethernet/smsc/smsc9420.c 	iounmap(pd->ioaddr - LAN9420_CPSR_ENDIAN_OFFSET);
pd                708 drivers/net/ethernet/toshiba/tc35815.c 	struct device *pd = bus_find_device(&platform_bus_type, NULL,
pd                710 drivers/net/ethernet/toshiba/tc35815.c 	if (pd) {
pd                711 drivers/net/ethernet/toshiba/tc35815.c 		if (pd->platform_data)
pd                712 drivers/net/ethernet/toshiba/tc35815.c 			memcpy(dev->dev_addr, pd->platform_data, ETH_ALEN);
pd                713 drivers/net/ethernet/toshiba/tc35815.c 		put_device(pd);
pd               1825 drivers/net/fddi/skfp/smt.c 	const  struct smt_pdef	*pd ;
pd               1856 drivers/net/fddi/skfp/smt.c 		for (i = N_SMT_PLEN, pd = smt_pdef; i ; i--,pd++) {
pd               1857 drivers/net/fddi/skfp/smt.c 			if (pd->ptype == type)
pd               1860 drivers/net/fddi/skfp/smt.c 		if (i && pd->pswap) {
pd               1861 drivers/net/fddi/skfp/smt.c 			smt_string_swap(p+PARA_LEN,pd->pswap,len) ;
pd                169 drivers/net/phy/dp83640.c static void dp83640_gpio_defaults(struct ptp_pin_desc *pd)
pd                174 drivers/net/phy/dp83640.c 		snprintf(pd[i].name, sizeof(pd[i].name), "GPIO%d", 1 + i);
pd                175 drivers/net/phy/dp83640.c 		pd[i].index = i;
pd                186 drivers/net/phy/dp83640.c 	pd[index].func = PTP_PF_PHYSYNC;
pd                187 drivers/net/phy/dp83640.c 	pd[index].chan = 0;
pd                190 drivers/net/phy/dp83640.c 	pd[index].func = PTP_PF_PEROUT;
pd                191 drivers/net/phy/dp83640.c 	pd[index].chan = 0;
pd                195 drivers/net/phy/dp83640.c 		pd[index].func = PTP_PF_EXTTS;
pd                196 drivers/net/phy/dp83640.c 		pd[index].chan = i - EXTTS0_GPIO;
pd                698 drivers/net/wireless/ath/ath5k/eeprom.c 			struct ath5k_pdgain_info *pd =
pd                701 drivers/net/wireless/ath/ath5k/eeprom.c 			kfree(pd->pd_step);
pd                702 drivers/net/wireless/ath/ath5k/eeprom.c 			kfree(pd->pd_pwr);
pd                719 drivers/net/wireless/ath/ath5k/eeprom.c 	struct ath5k_pdgain_info *pd;
pd                751 drivers/net/wireless/ath/ath5k/eeprom.c 		pd = &chinfo[pier].pd_curves[idx];
pd                753 drivers/net/wireless/ath/ath5k/eeprom.c 		pd->pd_points = AR5K_EEPROM_N_PWR_POINTS_5111;
pd                756 drivers/net/wireless/ath/ath5k/eeprom.c 		pd->pd_step = kcalloc(AR5K_EEPROM_N_PWR_POINTS_5111,
pd                758 drivers/net/wireless/ath/ath5k/eeprom.c 		if (!pd->pd_step)
pd                761 drivers/net/wireless/ath/ath5k/eeprom.c 		pd->pd_pwr = kcalloc(AR5K_EEPROM_N_PWR_POINTS_5111,
pd                763 drivers/net/wireless/ath/ath5k/eeprom.c 		if (!pd->pd_pwr)
pd                769 drivers/net/wireless/ath/ath5k/eeprom.c 		for (point = 0; point < pd->pd_points; point++) {
pd                772 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_pwr[point] = 2 * pcinfo->pwr[point];
pd                775 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_step[point] = pcinfo->pcdac[point];
pd                779 drivers/net/wireless/ath/ath5k/eeprom.c 		chinfo[pier].min_pwr = pd->pd_pwr[0];
pd                780 drivers/net/wireless/ath/ath5k/eeprom.c 		chinfo[pier].max_pwr = pd->pd_pwr[10];
pd                928 drivers/net/wireless/ath/ath5k/eeprom.c 			struct ath5k_pdgain_info *pd =
pd                934 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_points = AR5K_EEPROM_N_XPD0_POINTS;
pd                937 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_step = kcalloc(pd->pd_points,
pd                940 drivers/net/wireless/ath/ath5k/eeprom.c 				if (!pd->pd_step)
pd                943 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_pwr = kcalloc(pd->pd_points,
pd                946 drivers/net/wireless/ath/ath5k/eeprom.c 				if (!pd->pd_pwr)
pd                951 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_step[0] = pcinfo->pcdac_x0[0];
pd                952 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_pwr[0] = pcinfo->pwr_x0[0];
pd                954 drivers/net/wireless/ath/ath5k/eeprom.c 				for (point = 1; point < pd->pd_points;
pd                957 drivers/net/wireless/ath/ath5k/eeprom.c 					pd->pd_pwr[point] =
pd                961 drivers/net/wireless/ath/ath5k/eeprom.c 					pd->pd_step[point] =
pd                962 drivers/net/wireless/ath/ath5k/eeprom.c 						pd->pd_step[point - 1] +
pd                967 drivers/net/wireless/ath/ath5k/eeprom.c 				chinfo[pier].min_pwr = pd->pd_pwr[0];
pd                972 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_points = AR5K_EEPROM_N_XPD3_POINTS;
pd                975 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_step = kcalloc(pd->pd_points,
pd                978 drivers/net/wireless/ath/ath5k/eeprom.c 				if (!pd->pd_step)
pd                981 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_pwr = kcalloc(pd->pd_points,
pd                984 drivers/net/wireless/ath/ath5k/eeprom.c 				if (!pd->pd_pwr)
pd                989 drivers/net/wireless/ath/ath5k/eeprom.c 				for (point = 0; point < pd->pd_points;
pd                992 drivers/net/wireless/ath/ath5k/eeprom.c 					pd->pd_pwr[point] =
pd                996 drivers/net/wireless/ath/ath5k/eeprom.c 					pd->pd_step[point] =
pd               1002 drivers/net/wireless/ath/ath5k/eeprom.c 				chinfo[pier].min_pwr = pd->pd_pwr[0];
pd               1220 drivers/net/wireless/ath/ath5k/eeprom.c 			struct ath5k_pdgain_info *pd =
pd               1226 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_points = AR5K_EEPROM_N_PD_POINTS;
pd               1228 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_points = AR5K_EEPROM_N_PD_POINTS - 1;
pd               1231 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_step = kcalloc(pd->pd_points,
pd               1234 drivers/net/wireless/ath/ath5k/eeprom.c 			if (!pd->pd_step)
pd               1237 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_pwr = kcalloc(pd->pd_points,
pd               1240 drivers/net/wireless/ath/ath5k/eeprom.c 			if (!pd->pd_pwr)
pd               1246 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_step[0] = pcinfo->pddac_i[pdg];
pd               1247 drivers/net/wireless/ath/ath5k/eeprom.c 			pd->pd_pwr[0] = 4 * pcinfo->pwr_i[pdg];
pd               1249 drivers/net/wireless/ath/ath5k/eeprom.c 			for (point = 1; point < pd->pd_points; point++) {
pd               1251 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_pwr[point] = pd->pd_pwr[point - 1] +
pd               1254 drivers/net/wireless/ath/ath5k/eeprom.c 				pd->pd_step[point] = pd->pd_step[point - 1] +
pd               1261 drivers/net/wireless/ath/ath5k/eeprom.c 				chinfo[pier].min_pwr = pd->pd_pwr[0];
pd               1266 drivers/net/wireless/ath/ath5k/eeprom.c 					pd->pd_pwr[pd->pd_points - 1];
pd                276 drivers/net/wireless/ath/ath9k/dfs.c 	struct dfs_pattern_detector *pd = sc->dfs_detector;
pd                278 drivers/net/wireless/ath/ath9k/dfs.c 	if (pd == NULL)
pd                280 drivers/net/wireless/ath/ath9k/dfs.c 	if (!pd->add_pulse(pd, pe, NULL))
pd                295 drivers/net/wireless/ath/dfs_pattern_detector.c 		struct pri_detector *pd = cd->detectors[i];
pd                296 drivers/net/wireless/ath/dfs_pattern_detector.c 		struct pri_sequence *ps = pd->add_pulse(pd, event);
pd                299 drivers/net/wireless/ath/dfs_pattern_detector.c 				memcpy(rs, pd->rs, sizeof(*rs));
pd                303 drivers/net/wireless/ath/dfs_pattern_detector.c 				event->freq, pd->rs->type_id,
pd                305 drivers/net/wireless/ath/dfs_pattern_detector.c 			pd->reset(pd, dpd->last_pulse_ts);
pd                101 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = drvr->proto->pd;
pd                110 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd;
pd                141 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd;
pd                159 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd;
pd                217 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd;
pd                335 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = drvr->proto->pd;
pd                366 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = bus_if->drvr->proto->pd;
pd                426 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = drvr->proto->pd;
pd                470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	drvr->proto->pd = bcdc;
pd                484 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	struct brcmf_bcdc *bcdc = drvr->proto->pd;
pd                486 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c 	drvr->proto->pd = NULL;
pd                429 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd                490 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd                798 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd                823 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd                832 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd                841 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd               1364 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd               1394 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd               1446 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd               1555 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	drvr->proto->pd = msgbuf;
pd               1634 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 	if (drvr->proto->pd) {
pd               1635 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 		msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd;
pd               1656 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c 		drvr->proto->pd = NULL;
pd                 41 drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h 	void *pd;
pd               3139 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	u32 pd, cs;
pd               3144 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	pd = rtw_read32_mask(rtwdev,
pd               3150 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	pd += pd_diff;
pd               3152 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	if (pd > RTW_CCK_PD_MAX)
pd               3153 drivers/net/wireless/realtek/rtw88/rtw8822c.c 		pd = RTW_CCK_PD_MAX;
pd               3161 drivers/net/wireless/realtek/rtw88/rtw8822c.c 			 pd);
pd                334 drivers/net/wireless/ti/wl1251/acx.c 	struct acx_packet_detection *pd;
pd                339 drivers/net/wireless/ti/wl1251/acx.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                340 drivers/net/wireless/ti/wl1251/acx.c 	if (!pd)
pd                345 drivers/net/wireless/ti/wl1251/acx.c 	ret = wl1251_cmd_configure(wl, ACX_PD_THRESHOLD, pd, sizeof(*pd));
pd                352 drivers/net/wireless/ti/wl1251/acx.c 	kfree(pd);
pd                 39 drivers/nvme/host/rdma.c 	struct ib_pd		*pd;
pd                273 drivers/nvme/host/rdma.c 	ret = rdma_create_qp(queue->cm_id, dev->pd, &init_attr);
pd                339 drivers/nvme/host/rdma.c 	ib_dealloc_pd(ndev->pd);
pd                372 drivers/nvme/host/rdma.c 	ndev->pd = ib_alloc_pd(ndev->dev,
pd                374 drivers/nvme/host/rdma.c 	if (IS_ERR(ndev->pd))
pd                392 drivers/nvme/host/rdma.c 	ib_dealloc_pd(ndev->pd);
pd               1194 drivers/nvme/host/rdma.c 		sge->lkey = queue->device->pd->local_dma_lkey;
pd               1213 drivers/nvme/host/rdma.c 	put_unaligned_le32(queue->device->pd->unsafe_global_rkey, sg->key);
pd               1305 drivers/nvme/host/rdma.c 		if (count == 1 && dev->pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) {
pd               1351 drivers/nvme/host/rdma.c 	sge->lkey   = queue->device->pd->local_dma_lkey;
pd               1382 drivers/nvme/host/rdma.c 	list.lkey   = queue->device->pd->local_dma_lkey;
pd                107 drivers/nvme/target/rdma.c 	struct ib_pd		*pd;
pd                263 drivers/nvme/target/rdma.c 		sge->lkey = ndev->pd->local_dma_lkey;
pd                293 drivers/nvme/target/rdma.c 	c->sge[0].lkey = ndev->pd->local_dma_lkey;
pd                378 drivers/nvme/target/rdma.c 	r->send_sge.lkey = ndev->pd->local_dma_lkey;
pd                865 drivers/nvme/target/rdma.c 	srq = ib_create_srq(ndev->pd, &srq_attr);
pd                909 drivers/nvme/target/rdma.c 	ib_dealloc_pd(ndev->pd);
pd                949 drivers/nvme/target/rdma.c 	ndev->pd = ib_alloc_pd(ndev->device, 0);
pd                950 drivers/nvme/target/rdma.c 	if (IS_ERR(ndev->pd))
pd                966 drivers/nvme/target/rdma.c 	ib_dealloc_pd(ndev->pd);
pd               1023 drivers/nvme/target/rdma.c 	ret = rdma_create_qp(queue->cm_id, ndev->pd, &qp_attr);
pd               1317 drivers/parport/share.c 	struct pardevice *pd;
pd               1354 drivers/parport/share.c 	for (pd = port->waithead; pd; pd = pd->waitnext) {
pd               1355 drivers/parport/share.c 		if (pd->waiting & 2) { /* sleeping in claim_or_block */
pd               1356 drivers/parport/share.c 			parport_claim(pd);
pd               1357 drivers/parport/share.c 			if (waitqueue_active(&pd->wait_q))
pd               1358 drivers/parport/share.c 				wake_up_interruptible(&pd->wait_q);
pd               1360 drivers/parport/share.c 		} else if (pd->wakeup) {
pd               1361 drivers/parport/share.c 			pd->wakeup(pd->private);
pd               1365 drivers/parport/share.c 			printk(KERN_ERR "%s: don't know how to wake %s\n", port->name, pd->name);
pd               1374 drivers/parport/share.c 	for (pd = port->devices; !port->cad && pd; pd = pd->next) {
pd               1375 drivers/parport/share.c 		if (pd->wakeup && pd != dev)
pd               1376 drivers/parport/share.c 			pd->wakeup(pd->private);
pd                240 drivers/pcmcia/bcm63xx_pcmcia.c 	if (gpio_get_value(skt->pd->ready_gpio))
pd                346 drivers/pcmcia/bcm63xx_pcmcia.c 	skt->pd = pdev->dev.platform_data;
pd                347 drivers/pcmcia/bcm63xx_pcmcia.c 	if (!skt->common_res || !skt->attr_res || !irq_res || !skt->pd) {
pd                 26 drivers/pcmcia/bcm63xx_pcmcia.h 	struct bcm63xx_pcmcia_platform_data *pd;
pd                538 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(pd-entry-vld,			0x07),
pd                563 drivers/perf/xgene_pmu.c 	XGENE_PMU_EVENT_ATTR(rank-pd,				0x20),
pd               1502 drivers/pinctrl/intel/pinctrl-intel.c 	const struct pin_desc *pd = pin_desc_get(pctrl->pctldev, pin);
pd               1504 drivers/pinctrl/intel/pinctrl-intel.c 	if (!pd || !intel_pad_usable(pctrl, pin))
pd               1513 drivers/pinctrl/intel/pinctrl-intel.c 	if (pd->mux_owner || pd->gpio_owner ||
pd                457 drivers/pinctrl/nomadik/pinctrl-abx500.c 	bool pd;
pd                485 drivers/pinctrl/nomadik/pinctrl-abx500.c 				gpio_offset, &pd);
pd                489 drivers/pinctrl/nomadik/pinctrl-abx500.c 		seq_printf(s, " %-9s", pull_up_down[pd]);
pd               1724 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c 	u32 ie, oe, pu, pd;
pd               1732 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c 		pd = ioread32(bank->base + NPCM7XX_GP_N_PD) & pinmask;
pd               1734 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c 			rc = (!pu && !pd);
pd               1736 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c 			rc = (pu && !pd);
pd               1738 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c 			rc = (!pu && pd);
pd                783 drivers/pinctrl/pinctrl-amd.c 	const struct pin_desc *pd = pin_desc_get(gpio_dev->pctrl, pin);
pd                785 drivers/pinctrl/pinctrl-amd.c 	if (!pd)
pd                792 drivers/pinctrl/pinctrl-amd.c 	if (pd->mux_owner || pd->gpio_owner ||
pd                432 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	struct pinctrl_pin_desc *pd;
pd                461 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	pd = &iod->pa[pin];
pd                462 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	pd->drv_data = &cfg[pin_index];
pd                656 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	struct pinctrl_pin_desc *pd;
pd                666 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	pd = &iod->pa[pin];
pd                667 drivers/pinctrl/ti/pinctrl-ti-iodelay.c 	cfg = pd->drv_data;
pd                164 drivers/platform/chrome/cros_ec.c 		ec_dev->pd = platform_device_register_data(ec_dev->dev,
pd                168 drivers/platform/chrome/cros_ec.c 		if (IS_ERR(ec_dev->pd)) {
pd                172 drivers/platform/chrome/cros_ec.c 			return PTR_ERR(ec_dev->pd);
pd                179 drivers/platform/chrome/cros_ec.c 			platform_device_unregister(ec_dev->pd);
pd                203 drivers/platform/chrome/cros_ec.c 	if (ec_dev->pd)
pd                204 drivers/platform/chrome/cros_ec.c 		platform_device_unregister(ec_dev->pd);
pd                422 drivers/platform/chrome/cros_ec_debugfs.c static int cros_ec_debugfs_probe(struct platform_device *pd)
pd                424 drivers/platform/chrome/cros_ec_debugfs.c 	struct cros_ec_dev *ec = dev_get_drvdata(pd->dev.parent);
pd                456 drivers/platform/chrome/cros_ec_debugfs.c 	dev_set_drvdata(&pd->dev, ec);
pd                465 drivers/platform/chrome/cros_ec_debugfs.c static int cros_ec_debugfs_remove(struct platform_device *pd)
pd                467 drivers/platform/chrome/cros_ec_debugfs.c 	struct cros_ec_dev *ec = dev_get_drvdata(pd->dev.parent);
pd                556 drivers/platform/chrome/cros_ec_lightbar.c static int cros_ec_lightbar_probe(struct platform_device *pd)
pd                558 drivers/platform/chrome/cros_ec_lightbar.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                560 drivers/platform/chrome/cros_ec_lightbar.c 	struct device *dev = &pd->dev;
pd                589 drivers/platform/chrome/cros_ec_lightbar.c static int cros_ec_lightbar_remove(struct platform_device *pd)
pd                591 drivers/platform/chrome/cros_ec_lightbar.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                344 drivers/platform/chrome/cros_ec_sysfs.c static int cros_ec_sysfs_probe(struct platform_device *pd)
pd                346 drivers/platform/chrome/cros_ec_sysfs.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                347 drivers/platform/chrome/cros_ec_sysfs.c 	struct device *dev = &pd->dev;
pd                357 drivers/platform/chrome/cros_ec_sysfs.c static int cros_ec_sysfs_remove(struct platform_device *pd)
pd                359 drivers/platform/chrome/cros_ec_sysfs.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                110 drivers/platform/chrome/cros_ec_vbc.c static int cros_ec_vbc_probe(struct platform_device *pd)
pd                112 drivers/platform/chrome/cros_ec_vbc.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                113 drivers/platform/chrome/cros_ec_vbc.c 	struct device *dev = &pd->dev;
pd                125 drivers/platform/chrome/cros_ec_vbc.c static int cros_ec_vbc_remove(struct platform_device *pd)
pd                127 drivers/platform/chrome/cros_ec_vbc.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                195 drivers/platform/chrome/cros_usbpd_logger.c static int cros_usbpd_logger_probe(struct platform_device *pd)
pd                197 drivers/platform/chrome/cros_usbpd_logger.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                198 drivers/platform/chrome/cros_usbpd_logger.c 	struct device *dev = &pd->dev;
pd                208 drivers/platform/chrome/cros_usbpd_logger.c 	platform_set_drvdata(pd, logger);
pd                222 drivers/platform/chrome/cros_usbpd_logger.c static int cros_usbpd_logger_remove(struct platform_device *pd)
pd                224 drivers/platform/chrome/cros_usbpd_logger.c 	struct logger_data *logger = platform_get_drvdata(pd);
pd                128 drivers/platform/x86/sony-laptop.c static int sony_nc_kbd_backlight_setup(struct platform_device *pd,
pd                130 drivers/platform/x86/sony-laptop.c static void sony_nc_kbd_backlight_cleanup(struct platform_device *pd,
pd                133 drivers/platform/x86/sony-laptop.c static int sony_nc_battery_care_setup(struct platform_device *pd,
pd                135 drivers/platform/x86/sony-laptop.c static void sony_nc_battery_care_cleanup(struct platform_device *pd);
pd                137 drivers/platform/x86/sony-laptop.c static int sony_nc_thermal_setup(struct platform_device *pd);
pd                138 drivers/platform/x86/sony-laptop.c static void sony_nc_thermal_cleanup(struct platform_device *pd);
pd                140 drivers/platform/x86/sony-laptop.c static int sony_nc_lid_resume_setup(struct platform_device *pd,
pd                142 drivers/platform/x86/sony-laptop.c static void sony_nc_lid_resume_cleanup(struct platform_device *pd);
pd                144 drivers/platform/x86/sony-laptop.c static int sony_nc_gfx_switch_setup(struct platform_device *pd,
pd                146 drivers/platform/x86/sony-laptop.c static void sony_nc_gfx_switch_cleanup(struct platform_device *pd);
pd                149 drivers/platform/x86/sony-laptop.c static int sony_nc_highspeed_charging_setup(struct platform_device *pd);
pd                150 drivers/platform/x86/sony-laptop.c static void sony_nc_highspeed_charging_cleanup(struct platform_device *pd);
pd                152 drivers/platform/x86/sony-laptop.c static int sony_nc_lowbatt_setup(struct platform_device *pd);
pd                153 drivers/platform/x86/sony-laptop.c static void sony_nc_lowbatt_cleanup(struct platform_device *pd);
pd                155 drivers/platform/x86/sony-laptop.c static int sony_nc_fanspeed_setup(struct platform_device *pd);
pd                156 drivers/platform/x86/sony-laptop.c static void sony_nc_fanspeed_cleanup(struct platform_device *pd);
pd                158 drivers/platform/x86/sony-laptop.c static int sony_nc_usb_charge_setup(struct platform_device *pd);
pd                159 drivers/platform/x86/sony-laptop.c static void sony_nc_usb_charge_cleanup(struct platform_device *pd);
pd                161 drivers/platform/x86/sony-laptop.c static int sony_nc_panelid_setup(struct platform_device *pd);
pd                162 drivers/platform/x86/sony-laptop.c static void sony_nc_panelid_cleanup(struct platform_device *pd);
pd                164 drivers/platform/x86/sony-laptop.c static int sony_nc_smart_conn_setup(struct platform_device *pd);
pd                165 drivers/platform/x86/sony-laptop.c static void sony_nc_smart_conn_cleanup(struct platform_device *pd);
pd                167 drivers/platform/x86/sony-laptop.c static int sony_nc_touchpad_setup(struct platform_device *pd,
pd                169 drivers/platform/x86/sony-laptop.c static void sony_nc_touchpad_cleanup(struct platform_device *pd);
pd                838 drivers/platform/x86/sony-laptop.c static int sony_nc_handles_setup(struct platform_device *pd)
pd                864 drivers/platform/x86/sony-laptop.c 		if (device_create_file(&pd->dev, &handles->devattr)) {
pd                874 drivers/platform/x86/sony-laptop.c static int sony_nc_handles_cleanup(struct platform_device *pd)
pd                878 drivers/platform/x86/sony-laptop.c 			device_remove_file(&pd->dev, &handles->devattr);
pd               1431 drivers/platform/x86/sony-laptop.c static void sony_nc_function_cleanup(struct platform_device *pd)
pd               1453 drivers/platform/x86/sony-laptop.c 			sony_nc_touchpad_cleanup(pd);
pd               1458 drivers/platform/x86/sony-laptop.c 			sony_nc_battery_care_cleanup(pd);
pd               1462 drivers/platform/x86/sony-laptop.c 			sony_nc_lid_resume_cleanup(pd);
pd               1465 drivers/platform/x86/sony-laptop.c 			sony_nc_thermal_cleanup(pd);
pd               1470 drivers/platform/x86/sony-laptop.c 			sony_nc_gfx_switch_cleanup(pd);
pd               1473 drivers/platform/x86/sony-laptop.c 			sony_nc_highspeed_charging_cleanup(pd);
pd               1485 drivers/platform/x86/sony-laptop.c 			sony_nc_kbd_backlight_cleanup(pd, handle);
pd               1488 drivers/platform/x86/sony-laptop.c 			sony_nc_lowbatt_cleanup(pd);
pd               1491 drivers/platform/x86/sony-laptop.c 			sony_nc_fanspeed_cleanup(pd);
pd               1494 drivers/platform/x86/sony-laptop.c 			sony_nc_usb_charge_cleanup(pd);
pd               1497 drivers/platform/x86/sony-laptop.c 			sony_nc_panelid_cleanup(pd);
pd               1500 drivers/platform/x86/sony-laptop.c 			sony_nc_smart_conn_cleanup(pd);
pd               1508 drivers/platform/x86/sony-laptop.c 	sony_nc_handles_cleanup(pd);
pd               1870 drivers/platform/x86/sony-laptop.c static int sony_nc_kbd_backlight_setup(struct platform_device *pd,
pd               1929 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &kbdbl_ctl->mode_attr);
pd               1944 drivers/platform/x86/sony-laptop.c 		ret = device_create_file(&pd->dev, &kbdbl_ctl->timeout_attr);
pd               1955 drivers/platform/x86/sony-laptop.c 	device_remove_file(&pd->dev, &kbdbl_ctl->mode_attr);
pd               1962 drivers/platform/x86/sony-laptop.c static void sony_nc_kbd_backlight_cleanup(struct platform_device *pd,
pd               1966 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &kbdbl_ctl->mode_attr);
pd               1968 drivers/platform/x86/sony-laptop.c 			device_remove_file(&pd->dev, &kbdbl_ctl->timeout_attr);
pd               2078 drivers/platform/x86/sony-laptop.c static int sony_nc_battery_care_setup(struct platform_device *pd,
pd               2095 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &bcare_ctl->attrs[0]);
pd               2108 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &bcare_ctl->attrs[1]);
pd               2115 drivers/platform/x86/sony-laptop.c 	device_remove_file(&pd->dev, &bcare_ctl->attrs[0]);
pd               2124 drivers/platform/x86/sony-laptop.c static void sony_nc_battery_care_cleanup(struct platform_device *pd)
pd               2127 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &bcare_ctl->attrs[0]);
pd               2129 drivers/platform/x86/sony-laptop.c 			device_remove_file(&pd->dev, &bcare_ctl->attrs[1]);
pd               2236 drivers/platform/x86/sony-laptop.c static int sony_nc_thermal_setup(struct platform_device *pd)
pd               2267 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &th_handle->profiles_attr);
pd               2271 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &th_handle->mode_attr);
pd               2278 drivers/platform/x86/sony-laptop.c 	device_remove_file(&pd->dev, &th_handle->profiles_attr);
pd               2285 drivers/platform/x86/sony-laptop.c static void sony_nc_thermal_cleanup(struct platform_device *pd)
pd               2288 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &th_handle->profiles_attr);
pd               2289 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &th_handle->mode_attr);
pd               2373 drivers/platform/x86/sony-laptop.c static int sony_nc_lid_resume_setup(struct platform_device *pd,
pd               2410 drivers/platform/x86/sony-laptop.c 		result = device_create_file(&pd->dev, &lid_ctl->attrs[i]);
pd               2419 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &lid_ctl->attrs[i]);
pd               2427 drivers/platform/x86/sony-laptop.c static void sony_nc_lid_resume_cleanup(struct platform_device *pd)
pd               2436 drivers/platform/x86/sony-laptop.c 			device_remove_file(&pd->dev, &lid_ctl->attrs[i]);
pd               2507 drivers/platform/x86/sony-laptop.c static int sony_nc_gfx_switch_setup(struct platform_device *pd,
pd               2523 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, &gfxs_ctl->attr);
pd               2536 drivers/platform/x86/sony-laptop.c static void sony_nc_gfx_switch_cleanup(struct platform_device *pd)
pd               2539 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &gfxs_ctl->attr);
pd               2579 drivers/platform/x86/sony-laptop.c static int sony_nc_highspeed_charging_setup(struct platform_device *pd)
pd               2601 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, hsc_handle);
pd               2611 drivers/platform/x86/sony-laptop.c static void sony_nc_highspeed_charging_cleanup(struct platform_device *pd)
pd               2614 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, hsc_handle);
pd               2653 drivers/platform/x86/sony-laptop.c static int sony_nc_lowbatt_setup(struct platform_device *pd)
pd               2667 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, lowbatt_handle);
pd               2677 drivers/platform/x86/sony-laptop.c static void sony_nc_lowbatt_cleanup(struct platform_device *pd)
pd               2680 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, lowbatt_handle);
pd               2730 drivers/platform/x86/sony-laptop.c static int sony_nc_fanspeed_setup(struct platform_device *pd)
pd               2756 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, fan_handle);
pd               2760 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, hsf_handle);
pd               2767 drivers/platform/x86/sony-laptop.c 	device_remove_file(&pd->dev, fan_handle);
pd               2779 drivers/platform/x86/sony-laptop.c static void sony_nc_fanspeed_cleanup(struct platform_device *pd)
pd               2782 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, fan_handle);
pd               2787 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, hsf_handle);
pd               2826 drivers/platform/x86/sony-laptop.c static int sony_nc_usb_charge_setup(struct platform_device *pd)
pd               2848 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, uc_handle);
pd               2858 drivers/platform/x86/sony-laptop.c static void sony_nc_usb_charge_cleanup(struct platform_device *pd)
pd               2861 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, uc_handle);
pd               2881 drivers/platform/x86/sony-laptop.c static int sony_nc_panelid_setup(struct platform_device *pd)
pd               2895 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, panel_handle);
pd               2905 drivers/platform/x86/sony-laptop.c static void sony_nc_panelid_cleanup(struct platform_device *pd)
pd               2908 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, panel_handle);
pd               2936 drivers/platform/x86/sony-laptop.c static int sony_nc_smart_conn_setup(struct platform_device *pd)
pd               2950 drivers/platform/x86/sony-laptop.c 	result = device_create_file(&pd->dev, sc_handle);
pd               2960 drivers/platform/x86/sony-laptop.c static void sony_nc_smart_conn_cleanup(struct platform_device *pd)
pd               2963 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, sc_handle);
pd               3009 drivers/platform/x86/sony-laptop.c static int sony_nc_touchpad_setup(struct platform_device *pd,
pd               3026 drivers/platform/x86/sony-laptop.c 	ret = device_create_file(&pd->dev, &tp_ctl->attr);
pd               3035 drivers/platform/x86/sony-laptop.c static void sony_nc_touchpad_cleanup(struct platform_device *pd)
pd               3038 drivers/platform/x86/sony-laptop.c 		device_remove_file(&pd->dev, &tp_ctl->attr);
pd                 93 drivers/platform/x86/surface3_button.c 	struct platform_device *pd;
pd                136 drivers/platform/x86/surface3_button.c 	pd = platform_device_alloc("gpio-keys", PLATFORM_DEVID_AUTO);
pd                137 drivers/platform/x86/surface3_button.c 	if (!pd) {
pd                142 drivers/platform/x86/surface3_button.c 	error = platform_device_add_data(pd, gpio_keys_pdata,
pd                147 drivers/platform/x86/surface3_button.c 	error = platform_device_add(pd);
pd                151 drivers/platform/x86/surface3_button.c 	return pd;
pd                154 drivers/platform/x86/surface3_button.c 	platform_device_put(pd);
pd                186 drivers/platform/x86/surface3_button.c 	struct platform_device *pd;
pd                208 drivers/platform/x86/surface3_button.c 		pd = surface3_button_device_create(client,
pd                211 drivers/platform/x86/surface3_button.c 		if (IS_ERR(pd)) {
pd                212 drivers/platform/x86/surface3_button.c 			error = PTR_ERR(pd);
pd                220 drivers/platform/x86/surface3_button.c 		priv->children[i] = pd;
pd                555 drivers/power/supply/cros_usbpd-charger.c static int cros_usbpd_charger_probe(struct platform_device *pd)
pd                557 drivers/power/supply/cros_usbpd-charger.c 	struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent);
pd                560 drivers/power/supply/cros_usbpd-charger.c 	struct device *dev = &pd->dev;
pd                576 drivers/power/supply/cros_usbpd-charger.c 	platform_set_drvdata(pd, charger);
pd                122 drivers/ptp/ptp_chardev.c 	struct ptp_pin_desc pd;
pd                317 drivers/ptp/ptp_chardev.c 		if (copy_from_user(&pd, (void __user *)arg, sizeof(pd))) {
pd                321 drivers/ptp/ptp_chardev.c 		if ((pd.rsv[0] || pd.rsv[1] || pd.rsv[2]
pd                322 drivers/ptp/ptp_chardev.c 				|| pd.rsv[3] || pd.rsv[4])
pd                327 drivers/ptp/ptp_chardev.c 			pd.rsv[0] = 0;
pd                328 drivers/ptp/ptp_chardev.c 			pd.rsv[1] = 0;
pd                329 drivers/ptp/ptp_chardev.c 			pd.rsv[2] = 0;
pd                330 drivers/ptp/ptp_chardev.c 			pd.rsv[3] = 0;
pd                331 drivers/ptp/ptp_chardev.c 			pd.rsv[4] = 0;
pd                333 drivers/ptp/ptp_chardev.c 		pin_index = pd.index;
pd                341 drivers/ptp/ptp_chardev.c 		pd = ops->pin_config[pin_index];
pd                343 drivers/ptp/ptp_chardev.c 		if (!err && copy_to_user((void __user *)arg, &pd, sizeof(pd)))
pd                349 drivers/ptp/ptp_chardev.c 		if (copy_from_user(&pd, (void __user *)arg, sizeof(pd))) {
pd                353 drivers/ptp/ptp_chardev.c 		if ((pd.rsv[0] || pd.rsv[1] || pd.rsv[2]
pd                354 drivers/ptp/ptp_chardev.c 				|| pd.rsv[3] || pd.rsv[4])
pd                359 drivers/ptp/ptp_chardev.c 			pd.rsv[0] = 0;
pd                360 drivers/ptp/ptp_chardev.c 			pd.rsv[1] = 0;
pd                361 drivers/ptp/ptp_chardev.c 			pd.rsv[2] = 0;
pd                362 drivers/ptp/ptp_chardev.c 			pd.rsv[3] = 0;
pd                363 drivers/ptp/ptp_chardev.c 			pd.rsv[4] = 0;
pd                365 drivers/ptp/ptp_chardev.c 		pin_index = pd.index;
pd                373 drivers/ptp/ptp_chardev.c 		err = ptp_set_pinfunc(ptp, pin_index, pd.func, pd.chan);
pd                171 drivers/regulator/max77620-regulator.c 	int pd = rpdata->active_fps_pd_slot;
pd                179 drivers/regulator/max77620-regulator.c 		pd = rpdata->suspend_fps_pd_slot;
pd                189 drivers/regulator/max77620-regulator.c 	if (pd >= 0) {
pd                190 drivers/regulator/max77620-regulator.c 		val |= (pd << MAX77620_FPS_PD_PERIOD_SHIFT);
pd                126 drivers/regulator/max8952.c 	struct max8952_platform_data *pd;
pd                131 drivers/regulator/max8952.c 	pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL);
pd                132 drivers/regulator/max8952.c 	if (!pd)
pd                135 drivers/regulator/max8952.c 	if (of_property_read_u32(np, "max8952,default-mode", &pd->default_mode))
pd                139 drivers/regulator/max8952.c 					pd->dvs_mode, ARRAY_SIZE(pd->dvs_mode));
pd                145 drivers/regulator/max8952.c 	for (i = 0; i < ARRAY_SIZE(pd->dvs_mode); ++i) {
pd                146 drivers/regulator/max8952.c 		if (pd->dvs_mode[i] < 770000 || pd->dvs_mode[i] > 1400000) {
pd                150 drivers/regulator/max8952.c 		pd->dvs_mode[i] = (pd->dvs_mode[i] - 770000) / 10000;
pd                153 drivers/regulator/max8952.c 	if (of_property_read_u32(np, "max8952,sync-freq", &pd->sync_freq))
pd                156 drivers/regulator/max8952.c 	if (of_property_read_u32(np, "max8952,ramp-speed", &pd->ramp_speed))
pd                159 drivers/regulator/max8952.c 	pd->reg_data = of_get_regulator_init_data(dev, np, &regulator);
pd                160 drivers/regulator/max8952.c 	if (!pd->reg_data) {
pd                165 drivers/regulator/max8952.c 	return pd;
pd                 30 drivers/regulator/qcom_rpm-regulator.c 	struct request_member pd;		/* pull down enable */
pd                 72 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 1, 0x00000001,  0 },
pd                 83 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 1, 0x00000001,  0 },
pd                 92 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 0, 0x00000002,  1 },
pd                109 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 0, 0x00800000, 23 },
pd                120 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 0, 0x00800000, 23 },
pd                134 drivers/regulator/qcom_rpm-regulator.c 	.pd             = { 0, 0x00000002,  1 },
pd                700 drivers/regulator/qcom_rpm-regulator.c 		ret = rpm_reg_set(vreg, &vreg->parts->pd, 1);
pd                700 drivers/scsi/aic94xx/aic94xx_scb.c static void set_speed_mask(u8 *speed_mask, struct asd_phy_desc *pd)
pd                706 drivers/scsi/aic94xx/aic94xx_scb.c 	switch (pd->max_sas_lrate) {
pd                718 drivers/scsi/aic94xx/aic94xx_scb.c 	switch (pd->min_sas_lrate) {
pd                730 drivers/scsi/aic94xx/aic94xx_scb.c 	switch (pd->max_sata_lrate) {
pd                739 drivers/scsi/aic94xx/aic94xx_scb.c 	switch (pd->min_sata_lrate) {
pd                889 drivers/scsi/aic94xx/aic94xx_scb.c 	struct asd_phy_desc *pd = asd_ha->phys[phy->id].phy_desc;
pd                901 drivers/scsi/aic94xx/aic94xx_scb.c 			pd->min_sas_lrate = rates->minimum_linkrate;
pd                902 drivers/scsi/aic94xx/aic94xx_scb.c 			pd->min_sata_lrate = rates->minimum_linkrate;
pd                905 drivers/scsi/aic94xx/aic94xx_scb.c 			pd->max_sas_lrate = rates->maximum_linkrate;
pd                906 drivers/scsi/aic94xx/aic94xx_scb.c 			pd->max_sata_lrate = rates->maximum_linkrate;
pd                793 drivers/scsi/aic94xx/aic94xx_sds.c 		struct asd_manuf_phy_desc *pd = &phy_param->phy_desc[i];
pd                794 drivers/scsi/aic94xx/aic94xx_sds.c 		switch (pd->state & 0xF) {
pd                809 drivers/scsi/aic94xx/aic94xx_sds.c 		asd_ha->hw_prof.phy_desc[i].phy_control_0 = pd->phy_control_0;
pd                810 drivers/scsi/aic94xx/aic94xx_sds.c 		asd_ha->hw_prof.phy_desc[i].phy_control_1 = pd->phy_control_1;
pd                811 drivers/scsi/aic94xx/aic94xx_sds.c 		asd_ha->hw_prof.phy_desc[i].phy_control_2 = pd->phy_control_2;
pd                812 drivers/scsi/aic94xx/aic94xx_sds.c 		asd_ha->hw_prof.phy_desc[i].phy_control_3 = pd->phy_control_3;
pd                297 drivers/scsi/be2iscsi/be_cmds.h 	u8 pd[10];		/* dword 1 */
pd                585 drivers/scsi/be2iscsi/be_cmds.h 	u8 pd[10];		/* dword 1 */
pd               2091 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2121 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2140 drivers/scsi/megaraid/megaraid_sas.h 		struct megasas_evtarg_pd pd;
pd               2143 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2149 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2154 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2159 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2164 drivers/scsi/megaraid/megaraid_sas.h 			struct megasas_evtarg_pd pd;
pd               2655 drivers/scsi/megaraid/megaraid_sas.h __le16 MR_PdDevHandleGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map);
pd                121 drivers/scsi/megaraid/megaraid_sas_fp.c 	return le16_to_cpu(map->raidMap.arMapInfo[ar].pd[arm]);
pd                129 drivers/scsi/megaraid/megaraid_sas_fp.c __le16 MR_PdDevHandleGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map)
pd                131 drivers/scsi/megaraid/megaraid_sas_fp.c 	return map->raidMap.devHndlInfo[pd].curDevHdl;
pd                134 drivers/scsi/megaraid/megaraid_sas_fp.c static u8 MR_PdInterfaceTypeGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map)
pd                136 drivers/scsi/megaraid/megaraid_sas_fp.c 	return map->raidMap.devHndlInfo[pd].interfaceType;
pd                694 drivers/scsi/megaraid/megaraid_sas_fp.c 	u32     pd, arRef, r1_alt_pd;
pd                729 drivers/scsi/megaraid/megaraid_sas_fp.c 	pd          = MR_ArPdGet(arRef, physArm, map);
pd                731 drivers/scsi/megaraid/megaraid_sas_fp.c 	if (pd != MR_PD_INVALID) {
pd                732 drivers/scsi/megaraid/megaraid_sas_fp.c 		*pDevHandle = MR_PdDevHandleGet(pd, map);
pd                733 drivers/scsi/megaraid/megaraid_sas_fp.c 		*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
pd                751 drivers/scsi/megaraid/megaraid_sas_fp.c 			pd = MR_ArPdGet(arRef, physArm, map);
pd                752 drivers/scsi/megaraid/megaraid_sas_fp.c 			if (pd != MR_PD_INVALID) {
pd                753 drivers/scsi/megaraid/megaraid_sas_fp.c 				*pDevHandle = MR_PdDevHandleGet(pd, map);
pd                754 drivers/scsi/megaraid/megaraid_sas_fp.c 				*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
pd                770 drivers/scsi/megaraid/megaraid_sas_fp.c 	io_info->pd_after_lb = pd;
pd                797 drivers/scsi/megaraid/megaraid_sas_fp.c 	u32         pd, arRef, r1_alt_pd;
pd                844 drivers/scsi/megaraid/megaraid_sas_fp.c 	pd          = MR_ArPdGet(arRef, physArm, map); /* Get the pd */
pd                846 drivers/scsi/megaraid/megaraid_sas_fp.c 	if (pd != MR_PD_INVALID) {
pd                848 drivers/scsi/megaraid/megaraid_sas_fp.c 		*pDevHandle = MR_PdDevHandleGet(pd, map);
pd                849 drivers/scsi/megaraid/megaraid_sas_fp.c 		*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
pd                868 drivers/scsi/megaraid/megaraid_sas_fp.c 			pd = MR_ArPdGet(arRef, physArm, map);
pd                869 drivers/scsi/megaraid/megaraid_sas_fp.c 			if (pd != MR_PD_INVALID) {
pd                871 drivers/scsi/megaraid/megaraid_sas_fp.c 				*pDevHandle = MR_PdDevHandleGet(pd, map);
pd                872 drivers/scsi/megaraid/megaraid_sas_fp.c 				*pPdInterface = MR_PdInterfaceTypeGet(pd, map);
pd                888 drivers/scsi/megaraid/megaraid_sas_fp.c 	io_info->pd_after_lb = pd;
pd               2995 drivers/scsi/megaraid/megaraid_sas_fusion.c 	u32                         arRef, pd;
pd               3047 drivers/scsi/megaraid/megaraid_sas_fusion.c 		pd = MR_ArPdGet(arRef, physArm, local_map_ptr);
pd               3048 drivers/scsi/megaraid/megaraid_sas_fusion.c 		devHandle = MR_PdDevHandleGet(pd, local_map_ptr);
pd                736 drivers/scsi/megaraid/megaraid_sas_fusion.h 	__le16	pd[MAX_RAIDMAP_ROW_SIZE];
pd               1304 drivers/scsi/qla2xxx/qla_init.c 	struct port_database_24xx *pd;
pd               1331 drivers/scsi/qla2xxx/qla_init.c 	pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma);
pd               1332 drivers/scsi/qla2xxx/qla_init.c 	if (pd == NULL) {
pd               1348 drivers/scsi/qla2xxx/qla_init.c 	mbx->u.mbx.in = (void *)pd;
pd               1363 drivers/scsi/qla2xxx/qla_init.c 	if (pd)
pd               1364 drivers/scsi/qla2xxx/qla_init.c 		dma_pool_free(ha->s_dma_pool, pd, pd_dma);
pd               1408 drivers/scsi/qla2xxx/qla_init.c 	struct port_database_24xx *pd;
pd               1412 drivers/scsi/qla2xxx/qla_init.c 	pd = (struct port_database_24xx *)sp->u.iocb_cmd.u.mbx.in;
pd               1418 drivers/scsi/qla2xxx/qla_init.c 	    fcport->disc_state, pd->current_login_state, fcport->fc4f_nvme,
pd               1425 drivers/scsi/qla2xxx/qla_init.c 		ls = pd->current_login_state >> 4;
pd               1427 drivers/scsi/qla2xxx/qla_init.c 		ls = pd->current_login_state & 0xf;
pd               1444 drivers/scsi/qla2xxx/qla_init.c 		__qla24xx_parse_gpdb(vha, fcport, pd);
pd               1880 drivers/scsi/qla2xxx/qla_mbx.c 	port_database_t *pd;
pd               1889 drivers/scsi/qla2xxx/qla_mbx.c 	pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma);
pd               1890 drivers/scsi/qla2xxx/qla_mbx.c 	if (pd  == NULL) {
pd               1932 drivers/scsi/qla2xxx/qla_mbx.c 		pd24 = (struct port_database_24xx *) pd;
pd               1992 drivers/scsi/qla2xxx/qla_mbx.c 		if (pd->master_state != PD_STATE_PORT_LOGGED_IN &&
pd               1993 drivers/scsi/qla2xxx/qla_mbx.c 		    pd->slave_state != PD_STATE_PORT_LOGGED_IN) {
pd               1996 drivers/scsi/qla2xxx/qla_mbx.c 			    "portid=%02x%02x%02x.\n", pd->master_state,
pd               1997 drivers/scsi/qla2xxx/qla_mbx.c 			    pd->slave_state, fcport->d_id.b.domain,
pd               2005 drivers/scsi/qla2xxx/qla_mbx.c 		     memcmp(fcport->port_name, pd->port_name, 8))) {
pd               2012 drivers/scsi/qla2xxx/qla_mbx.c 		memcpy(fcport->node_name, pd->node_name, WWN_SIZE);
pd               2013 drivers/scsi/qla2xxx/qla_mbx.c 		memcpy(fcport->port_name, pd->port_name, WWN_SIZE);
pd               2016 drivers/scsi/qla2xxx/qla_mbx.c 		fcport->d_id.b.domain = pd->port_id[0];
pd               2017 drivers/scsi/qla2xxx/qla_mbx.c 		fcport->d_id.b.area = pd->port_id[3];
pd               2018 drivers/scsi/qla2xxx/qla_mbx.c 		fcport->d_id.b.al_pa = pd->port_id[2];
pd               2022 drivers/scsi/qla2xxx/qla_mbx.c 		if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0)
pd               2028 drivers/scsi/qla2xxx/qla_mbx.c 		fcport->supported_classes = (pd->options & BIT_4) ?
pd               2033 drivers/scsi/qla2xxx/qla_mbx.c 	dma_pool_free(ha->s_dma_pool, pd, pd_dma);
pd               6311 drivers/scsi/qla2xxx/qla_mbx.c 	struct port_database_24xx *pd;
pd               6318 drivers/scsi/qla2xxx/qla_mbx.c 	pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma);
pd               6319 drivers/scsi/qla2xxx/qla_mbx.c 	if (pd  == NULL) {
pd               6342 drivers/scsi/qla2xxx/qla_mbx.c 	rval = __qla24xx_parse_gpdb(vha, fcport, pd);
pd               6348 drivers/scsi/qla2xxx/qla_mbx.c 	if (pd)
pd               6349 drivers/scsi/qla2xxx/qla_mbx.c 		dma_pool_free(ha->s_dma_pool, pd, pd_dma);
pd               6355 drivers/scsi/qla2xxx/qla_mbx.c     struct port_database_24xx *pd)
pd               6362 drivers/scsi/qla2xxx/qla_mbx.c 		current_login_state = pd->current_login_state >> 4;
pd               6363 drivers/scsi/qla2xxx/qla_mbx.c 		last_login_state = pd->last_login_state >> 4;
pd               6365 drivers/scsi/qla2xxx/qla_mbx.c 		current_login_state = pd->current_login_state & 0xf;
pd               6366 drivers/scsi/qla2xxx/qla_mbx.c 		last_login_state = pd->last_login_state & 0xf;
pd               6380 drivers/scsi/qla2xxx/qla_mbx.c 	     memcmp(fcport->port_name, pd->port_name, 8))) {
pd               6387 drivers/scsi/qla2xxx/qla_mbx.c 	memcpy(fcport->node_name, pd->node_name, WWN_SIZE);
pd               6388 drivers/scsi/qla2xxx/qla_mbx.c 	memcpy(fcport->port_name, pd->port_name, WWN_SIZE);
pd               6391 drivers/scsi/qla2xxx/qla_mbx.c 	fcport->d_id.b.domain = pd->port_id[0];
pd               6392 drivers/scsi/qla2xxx/qla_mbx.c 	fcport->d_id.b.area = pd->port_id[1];
pd               6393 drivers/scsi/qla2xxx/qla_mbx.c 	fcport->d_id.b.al_pa = pd->port_id[2];
pd               6398 drivers/scsi/qla2xxx/qla_mbx.c 		if ((pd->prli_svc_param_word_3[0] & BIT_5) == 0)
pd               6400 drivers/scsi/qla2xxx/qla_mbx.c 		if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0)
pd               6402 drivers/scsi/qla2xxx/qla_mbx.c 		if ((pd->prli_svc_param_word_3[0] & BIT_3) == 0)
pd               6406 drivers/scsi/qla2xxx/qla_mbx.c 		if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0)
pd               6412 drivers/scsi/qla2xxx/qla_mbx.c 	fcport->supported_classes = (pd->flags & PDF_CLASS_2) ?
pd               6415 drivers/scsi/qla2xxx/qla_mbx.c 	if (pd->prli_svc_param_word_3[0] & BIT_7) {
pd                216 drivers/scsi/sgiwd93.c 	struct sgiwd93_platform_data *pd = pdev->dev.platform_data;
pd                217 drivers/scsi/sgiwd93.c 	unsigned char *wdregs = pd->wdregs;
pd                218 drivers/scsi/sgiwd93.c 	struct hpc3_scsiregs *hregs = pd->hregs;
pd                222 drivers/scsi/sgiwd93.c 	unsigned int unit = pd->unit;
pd                223 drivers/scsi/sgiwd93.c 	unsigned int irq = pd->irq;
pd                290 drivers/scsi/sgiwd93.c 	struct sgiwd93_platform_data *pd = pdev->dev.platform_data;
pd                293 drivers/scsi/sgiwd93.c 	free_irq(pd->irq, host);
pd                 47 drivers/soc/actions/owl-sps.c static int owl_sps_set_power(struct owl_sps_domain *pd, bool enable)
pd                 51 drivers/soc/actions/owl-sps.c 	ack_mask = BIT(pd->info->ack_bit);
pd                 52 drivers/soc/actions/owl-sps.c 	pwr_mask = BIT(pd->info->pwr_bit);
pd                 54 drivers/soc/actions/owl-sps.c 	return owl_sps_set_pg(pd->sps->base, pwr_mask, ack_mask, enable);
pd                 59 drivers/soc/actions/owl-sps.c 	struct owl_sps_domain *pd = to_owl_pd(domain);
pd                 61 drivers/soc/actions/owl-sps.c 	dev_dbg(pd->sps->dev, "%s power on", pd->info->name);
pd                 63 drivers/soc/actions/owl-sps.c 	return owl_sps_set_power(pd, true);
pd                 68 drivers/soc/actions/owl-sps.c 	struct owl_sps_domain *pd = to_owl_pd(domain);
pd                 70 drivers/soc/actions/owl-sps.c 	dev_dbg(pd->sps->dev, "%s power off", pd->info->name);
pd                 72 drivers/soc/actions/owl-sps.c 	return owl_sps_set_power(pd, false);
pd                 77 drivers/soc/actions/owl-sps.c 	struct owl_sps_domain *pd;
pd                 79 drivers/soc/actions/owl-sps.c 	pd = devm_kzalloc(sps->dev, sizeof(*pd), GFP_KERNEL);
pd                 80 drivers/soc/actions/owl-sps.c 	if (!pd)
pd                 83 drivers/soc/actions/owl-sps.c 	pd->info = &sps->info->domains[index];
pd                 84 drivers/soc/actions/owl-sps.c 	pd->sps = sps;
pd                 86 drivers/soc/actions/owl-sps.c 	pd->genpd.name = pd->info->name;
pd                 87 drivers/soc/actions/owl-sps.c 	pd->genpd.power_on = owl_sps_power_on;
pd                 88 drivers/soc/actions/owl-sps.c 	pd->genpd.power_off = owl_sps_power_off;
pd                 89 drivers/soc/actions/owl-sps.c 	pd->genpd.flags = pd->info->genpd_flags;
pd                 90 drivers/soc/actions/owl-sps.c 	pm_genpd_init(&pd->genpd, NULL, false);
pd                 92 drivers/soc/actions/owl-sps.c 	sps->genpd_data.domains[index] = &pd->genpd;
pd                 49 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd);
pd                 52 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                 58 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0,
pd                 63 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1,
pd                 68 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0,
pd                 74 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                 79 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	clk_disable_unprepare(pd->vpu_clk);
pd                 80 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	clk_disable_unprepare(pd->vapb_clk);
pd                 87 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd);
pd                 90 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                 96 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0,
pd                101 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1,
pd                106 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG2,
pd                111 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0,
pd                117 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                122 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	clk_disable_unprepare(pd->vpu_clk);
pd                123 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	clk_disable_unprepare(pd->vapb_clk);
pd                128 drivers/soc/amlogic/meson-gx-pwrc-vpu.c static int meson_gx_pwrc_vpu_setup_clk(struct meson_gx_pwrc_vpu *pd)
pd                132 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = clk_prepare_enable(pd->vpu_clk);
pd                136 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = clk_prepare_enable(pd->vapb_clk);
pd                138 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		clk_disable_unprepare(pd->vpu_clk);
pd                145 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd);
pd                149 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                155 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0,
pd                161 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1,
pd                167 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0,
pd                173 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = reset_control_assert(pd->rstc);
pd                177 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                180 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = reset_control_deassert(pd->rstc);
pd                184 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = meson_gx_pwrc_vpu_setup_clk(pd);
pd                193 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd);
pd                197 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                203 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0,
pd                209 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1,
pd                215 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG2,
pd                221 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 		regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0,
pd                227 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = reset_control_assert(pd->rstc);
pd                231 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0,
pd                234 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = reset_control_deassert(pd->rstc);
pd                238 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	ret = meson_gx_pwrc_vpu_setup_clk(pd);
pd                245 drivers/soc/amlogic/meson-gx-pwrc-vpu.c static bool meson_gx_pwrc_vpu_get_power(struct meson_gx_pwrc_vpu *pd)
pd                249 drivers/soc/amlogic/meson-gx-pwrc-vpu.c 	regmap_read(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, &reg);
pd                191 drivers/soc/bcm/bcm2835-power.c static int bcm2835_power_power_off(struct bcm2835_power_domain *pd, u32 pm_reg)
pd                193 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                207 drivers/soc/bcm/bcm2835-power.c static int bcm2835_power_power_on(struct bcm2835_power_domain *pd, u32 pm_reg)
pd                209 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                240 drivers/soc/bcm/bcm2835-power.c 			pd->base.name);
pd                255 drivers/soc/bcm/bcm2835-power.c 				pd->base.name);
pd                273 drivers/soc/bcm/bcm2835-power.c static int bcm2835_asb_power_on(struct bcm2835_power_domain *pd,
pd                279 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                282 drivers/soc/bcm/bcm2835-power.c 	ret = clk_prepare_enable(pd->clk);
pd                285 drivers/soc/bcm/bcm2835-power.c 			pd->base.name);
pd                292 drivers/soc/bcm/bcm2835-power.c 	clk_disable_unprepare(pd->clk);
pd                297 drivers/soc/bcm/bcm2835-power.c 	ret = clk_prepare_enable(pd->clk);
pd                300 drivers/soc/bcm/bcm2835-power.c 			pd->base.name);
pd                307 drivers/soc/bcm/bcm2835-power.c 			pd->base.name);
pd                313 drivers/soc/bcm/bcm2835-power.c 			pd->base.name);
pd                322 drivers/soc/bcm/bcm2835-power.c 	clk_disable_unprepare(pd->clk);
pd                328 drivers/soc/bcm/bcm2835-power.c static int bcm2835_asb_power_off(struct bcm2835_power_domain *pd,
pd                334 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                340 drivers/soc/bcm/bcm2835-power.c 			 pd->base.name);
pd                346 drivers/soc/bcm/bcm2835-power.c 			 pd->base.name);
pd                351 drivers/soc/bcm/bcm2835-power.c 	clk_disable_unprepare(pd->clk);
pd                361 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power_domain *pd =
pd                363 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                365 drivers/soc/bcm/bcm2835-power.c 	switch (pd->domain) {
pd                367 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_power_power_on(pd, PM_GRAFX);
pd                370 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_on(pd, PM_GRAFX,
pd                375 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_power_power_on(pd, PM_IMAGE);
pd                378 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_on(pd, PM_IMAGE,
pd                383 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_on(pd, PM_IMAGE,
pd                388 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_on(pd, PM_IMAGE,
pd                420 drivers/soc/bcm/bcm2835-power.c 		dev_err(power->dev, "Invalid domain %d\n", pd->domain);
pd                427 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power_domain *pd =
pd                429 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power *power = pd->power;
pd                431 drivers/soc/bcm/bcm2835-power.c 	switch (pd->domain) {
pd                433 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_power_power_off(pd, PM_GRAFX);
pd                436 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_off(pd, PM_GRAFX,
pd                441 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_power_power_off(pd, PM_IMAGE);
pd                444 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_off(pd, PM_IMAGE,
pd                449 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_off(pd, PM_IMAGE,
pd                454 drivers/soc/bcm/bcm2835-power.c 		return bcm2835_asb_power_off(pd, PM_IMAGE,
pd                483 drivers/soc/bcm/bcm2835-power.c 		dev_err(power->dev, "Invalid domain %d\n", pd->domain);
pd                536 drivers/soc/bcm/bcm2835-power.c 	struct bcm2835_power_domain *pd;
pd                541 drivers/soc/bcm/bcm2835-power.c 		pd = &power->domains[BCM2835_POWER_DOMAIN_GRAFX_V3D];
pd                544 drivers/soc/bcm/bcm2835-power.c 		pd = &power->domains[BCM2835_POWER_DOMAIN_IMAGE_H264];
pd                547 drivers/soc/bcm/bcm2835-power.c 		pd = &power->domains[BCM2835_POWER_DOMAIN_IMAGE_ISP];
pd                554 drivers/soc/bcm/bcm2835-power.c 	ret = bcm2835_power_pd_power_off(&pd->base);
pd                558 drivers/soc/bcm/bcm2835-power.c 	return bcm2835_power_pd_power_on(&pd->base);
pd                356 drivers/soc/fsl/dpio/dpio-service.c 	struct qbman_pull_desc pd;
pd                359 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_clear(&pd);
pd                360 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1);
pd                361 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_numframes(&pd, (u8)s->max);
pd                362 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_fq(&pd, fqid);
pd                368 drivers/soc/fsl/dpio/dpio-service.c 	err = qbman_swp_pull(d->swp, &pd);
pd                387 drivers/soc/fsl/dpio/dpio-service.c 	struct qbman_pull_desc pd;
pd                390 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_clear(&pd);
pd                391 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1);
pd                392 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_numframes(&pd, (u8)s->max);
pd                393 drivers/soc/fsl/dpio/dpio-service.c 	qbman_pull_desc_set_channel(&pd, channelid, qbman_pull_type_prio);
pd                400 drivers/soc/fsl/dpio/dpio-service.c 	err = qbman_swp_pull(d->swp, &pd);
pd                 61 drivers/soc/imx/gpc.c 	struct imx_pm_domain *pd = to_imx_pm_domain(genpd);
pd                 66 drivers/soc/imx/gpc.c 	regmap_read(pd->regmap, pd->reg_offs + GPC_PGC_PDNSCR_OFFS, &val);
pd                 71 drivers/soc/imx/gpc.c 	regmap_update_bits(pd->regmap, pd->reg_offs + GPC_PGC_CTRL_OFFS,
pd                 75 drivers/soc/imx/gpc.c 	val = BIT(pd->cntr_pdn_bit);
pd                 76 drivers/soc/imx/gpc.c 	regmap_update_bits(pd->regmap, GPC_CNTR, val, val);
pd                 79 drivers/soc/imx/gpc.c 	udelay(DIV_ROUND_UP(iso + iso2sw, pd->ipg_rate_mhz));
pd                 81 drivers/soc/imx/gpc.c 	if (pd->supply)
pd                 82 drivers/soc/imx/gpc.c 		regulator_disable(pd->supply);
pd                 89 drivers/soc/imx/gpc.c 	struct imx_pm_domain *pd = to_imx_pm_domain(genpd);
pd                 93 drivers/soc/imx/gpc.c 	if (pd->supply) {
pd                 94 drivers/soc/imx/gpc.c 		ret = regulator_enable(pd->supply);
pd                103 drivers/soc/imx/gpc.c 	for (i = 0; i < pd->num_clks; i++)
pd                104 drivers/soc/imx/gpc.c 		clk_prepare_enable(pd->clk[i]);
pd                107 drivers/soc/imx/gpc.c 	regmap_update_bits(pd->regmap, pd->reg_offs + GPC_PGC_CTRL_OFFS,
pd                111 drivers/soc/imx/gpc.c 	req = BIT(pd->cntr_pdn_bit + 1);
pd                112 drivers/soc/imx/gpc.c 	regmap_update_bits(pd->regmap, GPC_CNTR, req, req);
pd                115 drivers/soc/imx/gpc.c 	ret = regmap_read_poll_timeout(pd->regmap, GPC_CNTR, val, !(val & req),
pd                124 drivers/soc/imx/gpc.c 	for (i = 0; i < pd->num_clks; i++)
pd                125 drivers/soc/imx/gpc.c 		clk_disable_unprepare(pd->clk[i]);
pd                 90 drivers/soc/qcom/qcom_aoss.c 	struct generic_pm_domain pd;
pd                 93 drivers/soc/qcom/qcom_aoss.c #define to_qmp_pd_resource(res) container_of(res, struct qmp_pd, pd)
pd                319 drivers/soc/qcom/qcom_aoss.c 		 res->pd.name, enable ? "on" : "off");
pd                362 drivers/soc/qcom/qcom_aoss.c 		res[i].pd.name = sdm845_resources[i];
pd                363 drivers/soc/qcom/qcom_aoss.c 		res[i].pd.power_on = qmp_pd_power_on;
pd                364 drivers/soc/qcom/qcom_aoss.c 		res[i].pd.power_off = qmp_pd_power_off;
pd                366 drivers/soc/qcom/qcom_aoss.c 		ret = pm_genpd_init(&res[i].pd, NULL, true);
pd                372 drivers/soc/qcom/qcom_aoss.c 		data->domains[i] = &res[i].pd;
pd                 18 drivers/soc/qcom/rpmhpd.c #define domain_to_rpmhpd(domain) container_of(domain, struct rpmhpd, pd)
pd                 40 drivers/soc/qcom/rpmhpd.c 	struct generic_pm_domain pd;
pd                 63 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "ebi", },
pd                 68 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "lmx", },
pd                 73 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "lcx", },
pd                 78 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "gfx", },
pd                 83 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "mss", },
pd                 89 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "mx", },
pd                 95 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "mx_ao", },
pd                103 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "cx", },
pd                105 drivers/soc/qcom/rpmhpd.c 	.parent = &sdm845_mx.pd,
pd                110 drivers/soc/qcom/rpmhpd.c 	.pd = { .name = "cx_ao", },
pd                113 drivers/soc/qcom/rpmhpd.c 	.parent = &sdm845_mx_ao.pd,
pd                139 drivers/soc/qcom/rpmhpd.c static int rpmhpd_send_corner(struct rpmhpd *pd, int state,
pd                143 drivers/soc/qcom/rpmhpd.c 		.addr = pd->addr,
pd                152 drivers/soc/qcom/rpmhpd.c 		return rpmh_write(pd->dev, state, &cmd, 1);
pd                154 drivers/soc/qcom/rpmhpd.c 		return rpmh_write_async(pd->dev, state, &cmd, 1);
pd                157 drivers/soc/qcom/rpmhpd.c static void to_active_sleep(struct rpmhpd *pd, unsigned int corner,
pd                162 drivers/soc/qcom/rpmhpd.c 	if (pd->active_only)
pd                177 drivers/soc/qcom/rpmhpd.c static int rpmhpd_aggregate_corner(struct rpmhpd *pd, unsigned int corner)
pd                180 drivers/soc/qcom/rpmhpd.c 	struct rpmhpd *peer = pd->peer;
pd                185 drivers/soc/qcom/rpmhpd.c 	to_active_sleep(pd, corner, &this_active_corner, &this_sleep_corner);
pd                193 drivers/soc/qcom/rpmhpd.c 	ret = rpmhpd_send_corner(pd, RPMH_ACTIVE_ONLY_STATE, active_corner,
pd                194 drivers/soc/qcom/rpmhpd.c 				 active_corner > pd->active_corner);
pd                198 drivers/soc/qcom/rpmhpd.c 	pd->active_corner = active_corner;
pd                203 drivers/soc/qcom/rpmhpd.c 		ret = rpmhpd_send_corner(pd, RPMH_WAKE_ONLY_STATE,
pd                210 drivers/soc/qcom/rpmhpd.c 		return rpmhpd_send_corner(pd, RPMH_SLEEP_STATE, sleep_corner,
pd                219 drivers/soc/qcom/rpmhpd.c 	struct rpmhpd *pd = domain_to_rpmhpd(domain);
pd                224 drivers/soc/qcom/rpmhpd.c 	if (pd->corner)
pd                225 drivers/soc/qcom/rpmhpd.c 		ret = rpmhpd_aggregate_corner(pd, pd->corner);
pd                228 drivers/soc/qcom/rpmhpd.c 		pd->enabled = true;
pd                237 drivers/soc/qcom/rpmhpd.c 	struct rpmhpd *pd = domain_to_rpmhpd(domain);
pd                242 drivers/soc/qcom/rpmhpd.c 	ret = rpmhpd_aggregate_corner(pd, pd->level[0]);
pd                245 drivers/soc/qcom/rpmhpd.c 		pd->enabled = false;
pd                255 drivers/soc/qcom/rpmhpd.c 	struct rpmhpd *pd = domain_to_rpmhpd(domain);
pd                260 drivers/soc/qcom/rpmhpd.c 	for (i = 0; i < pd->level_count; i++)
pd                261 drivers/soc/qcom/rpmhpd.c 		if (level <= pd->level[i])
pd                268 drivers/soc/qcom/rpmhpd.c 	if (i == pd->level_count)
pd                271 drivers/soc/qcom/rpmhpd.c 	if (pd->enabled) {
pd                272 drivers/soc/qcom/rpmhpd.c 		ret = rpmhpd_aggregate_corner(pd, i);
pd                277 drivers/soc/qcom/rpmhpd.c 	pd->corner = i;
pd                374 drivers/soc/qcom/rpmhpd.c 		rpmhpds[i]->pd.power_off = rpmhpd_power_off;
pd                375 drivers/soc/qcom/rpmhpd.c 		rpmhpds[i]->pd.power_on = rpmhpd_power_on;
pd                376 drivers/soc/qcom/rpmhpd.c 		rpmhpds[i]->pd.set_performance_state = rpmhpd_set_performance_state;
pd                377 drivers/soc/qcom/rpmhpd.c 		rpmhpds[i]->pd.opp_to_performance_state = rpmhpd_get_performance_state;
pd                378 drivers/soc/qcom/rpmhpd.c 		pm_genpd_init(&rpmhpds[i]->pd, NULL, true);
pd                380 drivers/soc/qcom/rpmhpd.c 		data->domains[i] = &rpmhpds[i]->pd;
pd                389 drivers/soc/qcom/rpmhpd.c 					       &rpmhpds[i]->pd);
pd                 17 drivers/soc/qcom/rpmpd.c #define domain_to_rpmpd(domain) container_of(domain, struct rpmpd, pd)
pd                 43 drivers/soc/qcom/rpmpd.c 		.pd = {	.name = #_name,	},				\
pd                 50 drivers/soc/qcom/rpmpd.c 		.pd = { .name = #_active, },				\
pd                 60 drivers/soc/qcom/rpmpd.c 		.pd = { .name = #_name, },				\
pd                 68 drivers/soc/qcom/rpmpd.c 		.pd = { .name = #_name, },				\
pd                 76 drivers/soc/qcom/rpmpd.c 		.pd = { .name = #_name, },				\
pd                 84 drivers/soc/qcom/rpmpd.c 		.pd = { .name = #_name, },				\
pd                 97 drivers/soc/qcom/rpmpd.c 	struct generic_pm_domain pd;
pd                207 drivers/soc/qcom/rpmpd.c static int rpmpd_send_enable(struct rpmpd *pd, bool enable)
pd                215 drivers/soc/qcom/rpmpd.c 	return qcom_rpm_smd_write(pd->rpm, QCOM_SMD_RPM_ACTIVE_STATE,
pd                216 drivers/soc/qcom/rpmpd.c 				  pd->res_type, pd->res_id, &req, sizeof(req));
pd                219 drivers/soc/qcom/rpmpd.c static int rpmpd_send_corner(struct rpmpd *pd, int state, unsigned int corner)
pd                222 drivers/soc/qcom/rpmpd.c 		.key = pd->key,
pd                227 drivers/soc/qcom/rpmpd.c 	return qcom_rpm_smd_write(pd->rpm, state, pd->res_type, pd->res_id,
pd                231 drivers/soc/qcom/rpmpd.c static void to_active_sleep(struct rpmpd *pd, unsigned int corner,
pd                236 drivers/soc/qcom/rpmpd.c 	if (pd->active_only)
pd                242 drivers/soc/qcom/rpmpd.c static int rpmpd_aggregate_corner(struct rpmpd *pd)
pd                245 drivers/soc/qcom/rpmpd.c 	struct rpmpd *peer = pd->peer;
pd                250 drivers/soc/qcom/rpmpd.c 	to_active_sleep(pd, pd->corner, &this_active_corner, &this_sleep_corner);
pd                258 drivers/soc/qcom/rpmpd.c 	ret = rpmpd_send_corner(pd, QCOM_SMD_RPM_ACTIVE_STATE, active_corner);
pd                264 drivers/soc/qcom/rpmpd.c 	return rpmpd_send_corner(pd, QCOM_SMD_RPM_SLEEP_STATE, sleep_corner);
pd                270 drivers/soc/qcom/rpmpd.c 	struct rpmpd *pd = domain_to_rpmpd(domain);
pd                274 drivers/soc/qcom/rpmpd.c 	ret = rpmpd_send_enable(pd, true);
pd                278 drivers/soc/qcom/rpmpd.c 	pd->enabled = true;
pd                280 drivers/soc/qcom/rpmpd.c 	if (pd->corner)
pd                281 drivers/soc/qcom/rpmpd.c 		ret = rpmpd_aggregate_corner(pd);
pd                292 drivers/soc/qcom/rpmpd.c 	struct rpmpd *pd = domain_to_rpmpd(domain);
pd                296 drivers/soc/qcom/rpmpd.c 	ret = rpmpd_send_enable(pd, false);
pd                298 drivers/soc/qcom/rpmpd.c 		pd->enabled = false;
pd                309 drivers/soc/qcom/rpmpd.c 	struct rpmpd *pd = domain_to_rpmpd(domain);
pd                311 drivers/soc/qcom/rpmpd.c 	if (state > pd->max_state)
pd                312 drivers/soc/qcom/rpmpd.c 		state = pd->max_state;
pd                316 drivers/soc/qcom/rpmpd.c 	pd->corner = state;
pd                319 drivers/soc/qcom/rpmpd.c 	if (!pd->enabled && pd->key != KEY_FLOOR_CORNER &&
pd                320 drivers/soc/qcom/rpmpd.c 	    pd->key != KEY_FLOOR_LEVEL)
pd                323 drivers/soc/qcom/rpmpd.c 	ret = rpmpd_aggregate_corner(pd);
pd                376 drivers/soc/qcom/rpmpd.c 		rpmpds[i]->pd.power_off = rpmpd_power_off;
pd                377 drivers/soc/qcom/rpmpd.c 		rpmpds[i]->pd.power_on = rpmpd_power_on;
pd                378 drivers/soc/qcom/rpmpd.c 		rpmpds[i]->pd.set_performance_state = rpmpd_set_performance;
pd                379 drivers/soc/qcom/rpmpd.c 		rpmpds[i]->pd.opp_to_performance_state = rpmpd_get_performance;
pd                380 drivers/soc/qcom/rpmpd.c 		pm_genpd_init(&rpmpds[i]->pd, NULL, true);
pd                382 drivers/soc/qcom/rpmpd.c 		data->domains[i] = &rpmpds[i]->pd;
pd                183 drivers/soc/renesas/rcar-sysc.c 	struct rcar_sysc_pd *pd = to_rcar_pd(genpd);
pd                186 drivers/soc/renesas/rcar-sysc.c 	return rcar_sysc_power(&pd->ch, false);
pd                191 drivers/soc/renesas/rcar-sysc.c 	struct rcar_sysc_pd *pd = to_rcar_pd(genpd);
pd                194 drivers/soc/renesas/rcar-sysc.c 	return rcar_sysc_power(&pd->ch, true);
pd                199 drivers/soc/renesas/rcar-sysc.c static int __init rcar_sysc_pd_setup(struct rcar_sysc_pd *pd)
pd                201 drivers/soc/renesas/rcar-sysc.c 	struct generic_pm_domain *genpd = &pd->genpd;
pd                202 drivers/soc/renesas/rcar-sysc.c 	const char *name = pd->genpd.name;
pd                205 drivers/soc/renesas/rcar-sysc.c 	if (pd->flags & PD_CPU) {
pd                212 drivers/soc/renesas/rcar-sysc.c 	} else if (pd->flags & PD_SCU) {
pd                220 drivers/soc/renesas/rcar-sysc.c 	} else if (pd->flags & PD_NO_CR) {
pd                227 drivers/soc/renesas/rcar-sysc.c 	if (!(pd->flags & (PD_CPU | PD_SCU))) {
pd                242 drivers/soc/renesas/rcar-sysc.c 	if (pd->flags & (PD_CPU | PD_NO_CR)) {
pd                248 drivers/soc/renesas/rcar-sysc.c 	if (!rcar_sysc_power_is_off(&pd->ch)) {
pd                253 drivers/soc/renesas/rcar-sysc.c 	rcar_sysc_power(&pd->ch, true);
pd                375 drivers/soc/renesas/rcar-sysc.c 		struct rcar_sysc_pd *pd;
pd                382 drivers/soc/renesas/rcar-sysc.c 		pd = kzalloc(sizeof(*pd) + strlen(area->name) + 1, GFP_KERNEL);
pd                383 drivers/soc/renesas/rcar-sysc.c 		if (!pd) {
pd                388 drivers/soc/renesas/rcar-sysc.c 		strcpy(pd->name, area->name);
pd                389 drivers/soc/renesas/rcar-sysc.c 		pd->genpd.name = pd->name;
pd                390 drivers/soc/renesas/rcar-sysc.c 		pd->ch.chan_offs = area->chan_offs;
pd                391 drivers/soc/renesas/rcar-sysc.c 		pd->ch.chan_bit = area->chan_bit;
pd                392 drivers/soc/renesas/rcar-sysc.c 		pd->ch.isr_bit = area->isr_bit;
pd                393 drivers/soc/renesas/rcar-sysc.c 		pd->flags = area->flags;
pd                395 drivers/soc/renesas/rcar-sysc.c 		error = rcar_sysc_pd_setup(pd);
pd                399 drivers/soc/renesas/rcar-sysc.c 		domains->domains[area->isr_bit] = &pd->genpd;
pd                405 drivers/soc/renesas/rcar-sysc.c 					       &pd->genpd);
pd                437 drivers/soc/renesas/rcar-sysc.c 	struct rcar_sysc_pd *pd;
pd                448 drivers/soc/renesas/rcar-sysc.c 		pd = to_rcar_pd(genpd);
pd                449 drivers/soc/renesas/rcar-sysc.c 		if (!(pd->flags & PD_CPU) || pd->ch.chan_bit != idx)
pd                452 drivers/soc/renesas/rcar-sysc.c 		return rcar_sysc_power(&pd->ch, on);
pd                149 drivers/soc/renesas/rmobile-sysc.c 	struct device_node *pd;
pd                166 drivers/soc/renesas/rmobile-sysc.c 	struct device_node *pd;
pd                168 drivers/soc/renesas/rmobile-sysc.c 	pd = of_parse_phandle(np, "power-domains", 0);
pd                169 drivers/soc/renesas/rmobile-sysc.c 	if (!pd)
pd                173 drivers/soc/renesas/rmobile-sysc.c 		if (pd == special_pds[i].pd && type == special_pds[i].type) {
pd                174 drivers/soc/renesas/rmobile-sysc.c 			of_node_put(pd);
pd                180 drivers/soc/renesas/rmobile-sysc.c 		of_node_put(pd);
pd                184 drivers/soc/renesas/rmobile-sysc.c 	pr_debug("Special PM domain %pOFn type %d for %pOF\n", pd, type, np);
pd                186 drivers/soc/renesas/rmobile-sysc.c 	special_pds[num_special_pds].pd = pd;
pd                214 drivers/soc/renesas/rmobile-sysc.c 		of_node_put(special_pds[i].pd);
pd                217 drivers/soc/renesas/rmobile-sysc.c static enum pd_types __init pd_type(const struct device_node *pd)
pd                222 drivers/soc/renesas/rmobile-sysc.c 		if (pd == special_pds[i].pd)
pd                229 drivers/soc/renesas/rmobile-sysc.c 					   struct rmobile_pm_domain *pd)
pd                231 drivers/soc/renesas/rmobile-sysc.c 	const char *name = pd->genpd.name;
pd                240 drivers/soc/renesas/rmobile-sysc.c 		pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON;
pd                245 drivers/soc/renesas/rmobile-sysc.c 		pd->gov = &pm_domain_always_on_gov;
pd                246 drivers/soc/renesas/rmobile-sysc.c 		pd->suspend = rmobile_pd_suspend_console;
pd                256 drivers/soc/renesas/rmobile-sysc.c 		pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON;
pd                265 drivers/soc/renesas/rmobile-sysc.c 		pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON;
pd                269 drivers/soc/renesas/rmobile-sysc.c 		if (pd->bit_shift == ~0) {
pd                272 drivers/soc/renesas/rmobile-sysc.c 			pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON;
pd                277 drivers/soc/renesas/rmobile-sysc.c 	rmobile_init_pm_domain(pd);
pd                287 drivers/soc/renesas/rmobile-sysc.c 		struct rmobile_pm_domain *pd;
pd                294 drivers/soc/renesas/rmobile-sysc.c 		pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                295 drivers/soc/renesas/rmobile-sysc.c 		if (!pd) {
pd                300 drivers/soc/renesas/rmobile-sysc.c 		pd->genpd.name = np->name;
pd                301 drivers/soc/renesas/rmobile-sysc.c 		pd->base = base;
pd                302 drivers/soc/renesas/rmobile-sysc.c 		pd->bit_shift = idx;
pd                304 drivers/soc/renesas/rmobile-sysc.c 		rmobile_setup_pm_domain(np, pd);
pd                306 drivers/soc/renesas/rmobile-sysc.c 			pm_genpd_add_subdomain(genpd_parent, &pd->genpd);
pd                307 drivers/soc/renesas/rmobile-sysc.c 		of_genpd_add_provider_simple(np, &pd->genpd);
pd                309 drivers/soc/renesas/rmobile-sysc.c 		rmobile_add_pm_domains(base, np, &pd->genpd);
pd                134 drivers/soc/rockchip/pm_domains.c static bool rockchip_pmu_domain_is_idle(struct rockchip_pm_domain *pd)
pd                136 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pmu *pmu = pd->pmu;
pd                137 drivers/soc/rockchip/pm_domains.c 	const struct rockchip_domain_info *pd_info = pd->info;
pd                152 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_set_idle_request(struct rockchip_pm_domain *pd,
pd                155 drivers/soc/rockchip/pm_domains.c 	const struct rockchip_domain_info *pd_info = pd->info;
pd                156 drivers/soc/rockchip/pm_domains.c 	struct generic_pm_domain *genpd = &pd->genpd;
pd                157 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pmu *pmu = pd->pmu;
pd                187 drivers/soc/rockchip/pm_domains.c 	ret = readx_poll_timeout_atomic(rockchip_pmu_domain_is_idle, pd,
pd                199 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_save_qos(struct rockchip_pm_domain *pd)
pd                203 drivers/soc/rockchip/pm_domains.c 	for (i = 0; i < pd->num_qos; i++) {
pd                204 drivers/soc/rockchip/pm_domains.c 		regmap_read(pd->qos_regmap[i],
pd                206 drivers/soc/rockchip/pm_domains.c 			    &pd->qos_save_regs[0][i]);
pd                207 drivers/soc/rockchip/pm_domains.c 		regmap_read(pd->qos_regmap[i],
pd                209 drivers/soc/rockchip/pm_domains.c 			    &pd->qos_save_regs[1][i]);
pd                210 drivers/soc/rockchip/pm_domains.c 		regmap_read(pd->qos_regmap[i],
pd                212 drivers/soc/rockchip/pm_domains.c 			    &pd->qos_save_regs[2][i]);
pd                213 drivers/soc/rockchip/pm_domains.c 		regmap_read(pd->qos_regmap[i],
pd                215 drivers/soc/rockchip/pm_domains.c 			    &pd->qos_save_regs[3][i]);
pd                216 drivers/soc/rockchip/pm_domains.c 		regmap_read(pd->qos_regmap[i],
pd                218 drivers/soc/rockchip/pm_domains.c 			    &pd->qos_save_regs[4][i]);
pd                223 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_restore_qos(struct rockchip_pm_domain *pd)
pd                227 drivers/soc/rockchip/pm_domains.c 	for (i = 0; i < pd->num_qos; i++) {
pd                228 drivers/soc/rockchip/pm_domains.c 		regmap_write(pd->qos_regmap[i],
pd                230 drivers/soc/rockchip/pm_domains.c 			     pd->qos_save_regs[0][i]);
pd                231 drivers/soc/rockchip/pm_domains.c 		regmap_write(pd->qos_regmap[i],
pd                233 drivers/soc/rockchip/pm_domains.c 			     pd->qos_save_regs[1][i]);
pd                234 drivers/soc/rockchip/pm_domains.c 		regmap_write(pd->qos_regmap[i],
pd                236 drivers/soc/rockchip/pm_domains.c 			     pd->qos_save_regs[2][i]);
pd                237 drivers/soc/rockchip/pm_domains.c 		regmap_write(pd->qos_regmap[i],
pd                239 drivers/soc/rockchip/pm_domains.c 			     pd->qos_save_regs[3][i]);
pd                240 drivers/soc/rockchip/pm_domains.c 		regmap_write(pd->qos_regmap[i],
pd                242 drivers/soc/rockchip/pm_domains.c 			     pd->qos_save_regs[4][i]);
pd                248 drivers/soc/rockchip/pm_domains.c static bool rockchip_pmu_domain_is_on(struct rockchip_pm_domain *pd)
pd                250 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pmu *pmu = pd->pmu;
pd                254 drivers/soc/rockchip/pm_domains.c 	if (pd->info->status_mask == 0)
pd                255 drivers/soc/rockchip/pm_domains.c 		return !rockchip_pmu_domain_is_idle(pd);
pd                260 drivers/soc/rockchip/pm_domains.c 	return !(val & pd->info->status_mask);
pd                263 drivers/soc/rockchip/pm_domains.c static void rockchip_do_pmu_set_power_domain(struct rockchip_pm_domain *pd,
pd                266 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pmu *pmu = pd->pmu;
pd                267 drivers/soc/rockchip/pm_domains.c 	struct generic_pm_domain *genpd = &pd->genpd;
pd                270 drivers/soc/rockchip/pm_domains.c 	if (pd->info->pwr_mask == 0)
pd                272 drivers/soc/rockchip/pm_domains.c 	else if (pd->info->pwr_w_mask)
pd                274 drivers/soc/rockchip/pm_domains.c 			     on ? pd->info->pwr_w_mask :
pd                275 drivers/soc/rockchip/pm_domains.c 			     (pd->info->pwr_mask | pd->info->pwr_w_mask));
pd                278 drivers/soc/rockchip/pm_domains.c 				   pd->info->pwr_mask, on ? 0 : -1U);
pd                282 drivers/soc/rockchip/pm_domains.c 	if (readx_poll_timeout_atomic(rockchip_pmu_domain_is_on, pd, is_on,
pd                291 drivers/soc/rockchip/pm_domains.c static int rockchip_pd_power(struct rockchip_pm_domain *pd, bool power_on)
pd                293 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pmu *pmu = pd->pmu;
pd                298 drivers/soc/rockchip/pm_domains.c 	if (rockchip_pmu_domain_is_on(pd) != power_on) {
pd                299 drivers/soc/rockchip/pm_domains.c 		ret = clk_bulk_enable(pd->num_clks, pd->clks);
pd                307 drivers/soc/rockchip/pm_domains.c 			rockchip_pmu_save_qos(pd);
pd                310 drivers/soc/rockchip/pm_domains.c 			rockchip_pmu_set_idle_request(pd, true);
pd                313 drivers/soc/rockchip/pm_domains.c 		rockchip_do_pmu_set_power_domain(pd, power_on);
pd                317 drivers/soc/rockchip/pm_domains.c 			rockchip_pmu_set_idle_request(pd, false);
pd                319 drivers/soc/rockchip/pm_domains.c 			rockchip_pmu_restore_qos(pd);
pd                322 drivers/soc/rockchip/pm_domains.c 		clk_bulk_disable(pd->num_clks, pd->clks);
pd                331 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pm_domain *pd = to_rockchip_pd(domain);
pd                333 drivers/soc/rockchip/pm_domains.c 	return rockchip_pd_power(pd, true);
pd                338 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pm_domain *pd = to_rockchip_pd(domain);
pd                340 drivers/soc/rockchip/pm_domains.c 	return rockchip_pd_power(pd, false);
pd                385 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pm_domain *pd;
pd                412 drivers/soc/rockchip/pm_domains.c 	pd = devm_kzalloc(pmu->dev, sizeof(*pd), GFP_KERNEL);
pd                413 drivers/soc/rockchip/pm_domains.c 	if (!pd)
pd                416 drivers/soc/rockchip/pm_domains.c 	pd->info = pd_info;
pd                417 drivers/soc/rockchip/pm_domains.c 	pd->pmu = pmu;
pd                419 drivers/soc/rockchip/pm_domains.c 	pd->num_clks = of_clk_get_parent_count(node);
pd                420 drivers/soc/rockchip/pm_domains.c 	if (pd->num_clks > 0) {
pd                421 drivers/soc/rockchip/pm_domains.c 		pd->clks = devm_kcalloc(pmu->dev, pd->num_clks,
pd                422 drivers/soc/rockchip/pm_domains.c 					sizeof(*pd->clks), GFP_KERNEL);
pd                423 drivers/soc/rockchip/pm_domains.c 		if (!pd->clks)
pd                427 drivers/soc/rockchip/pm_domains.c 			node, pd->num_clks);
pd                428 drivers/soc/rockchip/pm_domains.c 		pd->num_clks = 0;
pd                431 drivers/soc/rockchip/pm_domains.c 	for (i = 0; i < pd->num_clks; i++) {
pd                432 drivers/soc/rockchip/pm_domains.c 		pd->clks[i].clk = of_clk_get(node, i);
pd                433 drivers/soc/rockchip/pm_domains.c 		if (IS_ERR(pd->clks[i].clk)) {
pd                434 drivers/soc/rockchip/pm_domains.c 			error = PTR_ERR(pd->clks[i].clk);
pd                442 drivers/soc/rockchip/pm_domains.c 	error = clk_bulk_prepare(pd->num_clks, pd->clks);
pd                446 drivers/soc/rockchip/pm_domains.c 	pd->num_qos = of_count_phandle_with_args(node, "pm_qos",
pd                449 drivers/soc/rockchip/pm_domains.c 	if (pd->num_qos > 0) {
pd                450 drivers/soc/rockchip/pm_domains.c 		pd->qos_regmap = devm_kcalloc(pmu->dev, pd->num_qos,
pd                451 drivers/soc/rockchip/pm_domains.c 					      sizeof(*pd->qos_regmap),
pd                453 drivers/soc/rockchip/pm_domains.c 		if (!pd->qos_regmap) {
pd                459 drivers/soc/rockchip/pm_domains.c 			pd->qos_save_regs[j] = devm_kcalloc(pmu->dev,
pd                460 drivers/soc/rockchip/pm_domains.c 							    pd->num_qos,
pd                463 drivers/soc/rockchip/pm_domains.c 			if (!pd->qos_save_regs[j]) {
pd                469 drivers/soc/rockchip/pm_domains.c 		for (j = 0; j < pd->num_qos; j++) {
pd                475 drivers/soc/rockchip/pm_domains.c 			pd->qos_regmap[j] = syscon_node_to_regmap(qos_node);
pd                476 drivers/soc/rockchip/pm_domains.c 			if (IS_ERR(pd->qos_regmap[j])) {
pd                485 drivers/soc/rockchip/pm_domains.c 	error = rockchip_pd_power(pd, true);
pd                493 drivers/soc/rockchip/pm_domains.c 	pd->genpd.name = node->name;
pd                494 drivers/soc/rockchip/pm_domains.c 	pd->genpd.power_off = rockchip_pd_power_off;
pd                495 drivers/soc/rockchip/pm_domains.c 	pd->genpd.power_on = rockchip_pd_power_on;
pd                496 drivers/soc/rockchip/pm_domains.c 	pd->genpd.attach_dev = rockchip_pd_attach_dev;
pd                497 drivers/soc/rockchip/pm_domains.c 	pd->genpd.detach_dev = rockchip_pd_detach_dev;
pd                498 drivers/soc/rockchip/pm_domains.c 	pd->genpd.flags = GENPD_FLAG_PM_CLK;
pd                500 drivers/soc/rockchip/pm_domains.c 		pd->genpd.flags |= GENPD_FLAG_ACTIVE_WAKEUP;
pd                501 drivers/soc/rockchip/pm_domains.c 	pm_genpd_init(&pd->genpd, NULL, false);
pd                503 drivers/soc/rockchip/pm_domains.c 	pmu->genpd_data.domains[id] = &pd->genpd;
pd                507 drivers/soc/rockchip/pm_domains.c 	clk_bulk_unprepare(pd->num_clks, pd->clks);
pd                509 drivers/soc/rockchip/pm_domains.c 	clk_bulk_put(pd->num_clks, pd->clks);
pd                513 drivers/soc/rockchip/pm_domains.c static void rockchip_pm_remove_one_domain(struct rockchip_pm_domain *pd)
pd                521 drivers/soc/rockchip/pm_domains.c 	ret = pm_genpd_remove(&pd->genpd);
pd                523 drivers/soc/rockchip/pm_domains.c 		dev_err(pd->pmu->dev, "failed to remove domain '%s' : %d - state may be inconsistent\n",
pd                524 drivers/soc/rockchip/pm_domains.c 			pd->genpd.name, ret);
pd                526 drivers/soc/rockchip/pm_domains.c 	clk_bulk_unprepare(pd->num_clks, pd->clks);
pd                527 drivers/soc/rockchip/pm_domains.c 	clk_bulk_put(pd->num_clks, pd->clks);
pd                530 drivers/soc/rockchip/pm_domains.c 	mutex_lock(&pd->pmu->mutex);
pd                531 drivers/soc/rockchip/pm_domains.c 	pd->num_clks = 0;
pd                532 drivers/soc/rockchip/pm_domains.c 	mutex_unlock(&pd->pmu->mutex);
pd                540 drivers/soc/rockchip/pm_domains.c 	struct rockchip_pm_domain *pd;
pd                546 drivers/soc/rockchip/pm_domains.c 			pd = to_rockchip_pd(genpd);
pd                547 drivers/soc/rockchip/pm_domains.c 			rockchip_pm_remove_one_domain(pd);
pd                 32 drivers/soc/samsung/pm_domains.c 	struct generic_pm_domain pd;
pd                 38 drivers/soc/samsung/pm_domains.c 	struct exynos_pm_domain *pd;
pd                 43 drivers/soc/samsung/pm_domains.c 	pd = container_of(domain, struct exynos_pm_domain, pd);
pd                 44 drivers/soc/samsung/pm_domains.c 	base = pd->base;
pd                 46 drivers/soc/samsung/pm_domains.c 	pwr = power_on ? pd->local_pwr_cfg : 0;
pd                 52 drivers/soc/samsung/pm_domains.c 	while ((readl_relaxed(base + 0x4) & pd->local_pwr_cfg) != pwr) {
pd                111 drivers/soc/samsung/pm_domains.c 		struct exynos_pm_domain *pd;
pd                116 drivers/soc/samsung/pm_domains.c 		pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                117 drivers/soc/samsung/pm_domains.c 		if (!pd) {
pd                121 drivers/soc/samsung/pm_domains.c 		pd->pd.name = exynos_get_domain_name(np);
pd                122 drivers/soc/samsung/pm_domains.c 		if (!pd->pd.name) {
pd                123 drivers/soc/samsung/pm_domains.c 			kfree(pd);
pd                128 drivers/soc/samsung/pm_domains.c 		pd->base = of_iomap(np, 0);
pd                129 drivers/soc/samsung/pm_domains.c 		if (!pd->base) {
pd                131 drivers/soc/samsung/pm_domains.c 			kfree_const(pd->pd.name);
pd                132 drivers/soc/samsung/pm_domains.c 			kfree(pd);
pd                136 drivers/soc/samsung/pm_domains.c 		pd->pd.power_off = exynos_pd_power_off;
pd                137 drivers/soc/samsung/pm_domains.c 		pd->pd.power_on = exynos_pd_power_on;
pd                138 drivers/soc/samsung/pm_domains.c 		pd->local_pwr_cfg = pm_domain_cfg->local_pwr_cfg;
pd                140 drivers/soc/samsung/pm_domains.c 		on = readl_relaxed(pd->base + 0x4) & pd->local_pwr_cfg;
pd                142 drivers/soc/samsung/pm_domains.c 		pm_genpd_init(&pd->pd, NULL, !on);
pd                143 drivers/soc/samsung/pm_domains.c 		of_genpd_add_provider_simple(np, &pd->pd);
pd                 43 drivers/soc/ti/ti_sci_pm_domains.c 	struct generic_pm_domain pd;
pd                 46 drivers/soc/ti/ti_sci_pm_domains.c #define genpd_to_ti_sci_pd(gpd) container_of(gpd, struct ti_sci_pm_domain, pd)
pd                 79 drivers/soc/ti/ti_sci_pm_domains.c 	struct generic_pm_domain *pd = pd_to_genpd(dev->pm_domain);
pd                 80 drivers/soc/ti/ti_sci_pm_domains.c 	struct ti_sci_pm_domain *ti_sci_genpd = genpd_to_ti_sci_pd(pd);
pd                190 drivers/soc/ti/ti_sci_pm_domains.c 	ti_sci_pd->pd.name = "ti_sci_pd";
pd                192 drivers/soc/ti/ti_sci_pm_domains.c 	ti_sci_pd->pd.attach_dev = ti_sci_pd_attach_dev;
pd                193 drivers/soc/ti/ti_sci_pm_domains.c 	ti_sci_pd->pd.detach_dev = ti_sci_pd_detach_dev;
pd                195 drivers/soc/ti/ti_sci_pm_domains.c 	ti_sci_pd->pd.dev_ops.start = ti_sci_dev_start;
pd                196 drivers/soc/ti/ti_sci_pm_domains.c 	ti_sci_pd->pd.dev_ops.stop = ti_sci_dev_stop;
pd                198 drivers/soc/ti/ti_sci_pm_domains.c 	pm_genpd_init(&ti_sci_pd->pd, NULL, true);
pd                200 drivers/soc/ti/ti_sci_pm_domains.c 	ret = of_genpd_add_provider_simple(np, &ti_sci_pd->pd);
pd                 75 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                 80 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = container_of(domain, struct zynqmp_pm_domain, gpd);
pd                 81 drivers/soc/xilinx/zynqmp_pm_domains.c 	ret = eemi_ops->set_requirement(pd->node_id,
pd                 87 drivers/soc/xilinx/zynqmp_pm_domains.c 		       __func__, domain->name, pd->node_id, ret);
pd                108 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                115 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = container_of(domain, struct zynqmp_pm_domain, gpd);
pd                118 drivers/soc/xilinx/zynqmp_pm_domains.c 	if (!(pd->flags & ZYNQMP_PM_DOMAIN_REQUESTED)) {
pd                135 drivers/soc/xilinx/zynqmp_pm_domains.c 	ret = eemi_ops->set_requirement(pd->node_id, capabilities, 0,
pd                143 drivers/soc/xilinx/zynqmp_pm_domains.c 		       __func__, domain->name, pd->node_id, ret);
pd                162 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                167 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = container_of(domain, struct zynqmp_pm_domain, gpd);
pd                173 drivers/soc/xilinx/zynqmp_pm_domains.c 	ret = eemi_ops->request_node(pd->node_id, 0, 0,
pd                178 drivers/soc/xilinx/zynqmp_pm_domains.c 		       __func__, domain->name, pd->node_id, ret);
pd                182 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd->flags |= ZYNQMP_PM_DOMAIN_REQUESTED;
pd                198 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                203 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = container_of(domain, struct zynqmp_pm_domain, gpd);
pd                209 drivers/soc/xilinx/zynqmp_pm_domains.c 	ret = eemi_ops->release_node(pd->node_id);
pd                213 drivers/soc/xilinx/zynqmp_pm_domains.c 		       __func__, domain->name, pd->node_id, ret);
pd                217 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd->flags &= ~ZYNQMP_PM_DOMAIN_REQUESTED;
pd                228 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                230 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = container_of(genpd_data->domains[0], struct zynqmp_pm_domain, gpd);
pd                237 drivers/soc/xilinx/zynqmp_pm_domains.c 		if (pd[i].node_id == idx)
pd                246 drivers/soc/xilinx/zynqmp_pm_domains.c 		if (pd[i].node_id == 0) {
pd                247 drivers/soc/xilinx/zynqmp_pm_domains.c 			pd[i].node_id = idx;
pd                264 drivers/soc/xilinx/zynqmp_pm_domains.c 	struct zynqmp_pm_domain *pd;
pd                271 drivers/soc/xilinx/zynqmp_pm_domains.c 	pd = devm_kcalloc(dev, ZYNQMP_NUM_DOMAINS, sizeof(*pd), GFP_KERNEL);
pd                272 drivers/soc/xilinx/zynqmp_pm_domains.c 	if (!pd)
pd                286 drivers/soc/xilinx/zynqmp_pm_domains.c 	for (i = 0; i < ZYNQMP_NUM_DOMAINS; i++, pd++) {
pd                287 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->node_id = 0;
pd                288 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->gpd.name = kasprintf(GFP_KERNEL, "domain%d", i);
pd                289 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->gpd.power_off = zynqmp_gpd_power_off;
pd                290 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->gpd.power_on = zynqmp_gpd_power_on;
pd                291 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->gpd.attach_dev = zynqmp_gpd_attach_dev;
pd                292 drivers/soc/xilinx/zynqmp_pm_domains.c 		pd->gpd.detach_dev = zynqmp_gpd_detach_dev;
pd                294 drivers/soc/xilinx/zynqmp_pm_domains.c 		domains[i] = &pd->gpd;
pd                297 drivers/soc/xilinx/zynqmp_pm_domains.c 		pm_genpd_init(&pd->gpd, NULL, true);
pd                 54 drivers/spi/spi-butterfly.c 	struct pardevice	*pd;
pd                178 drivers/spi/spi-butterfly.c 	struct pardevice	*pd;
pd                218 drivers/spi/spi-butterfly.c 	pd = parport_register_dev_model(p, "spi_butterfly", &butterfly_cb, 0);
pd                219 drivers/spi/spi-butterfly.c 	if (!pd) {
pd                223 drivers/spi/spi-butterfly.c 	pp->pd = pd;
pd                225 drivers/spi/spi-butterfly.c 	status = parport_claim(pd);
pd                279 drivers/spi/spi-butterfly.c 	parport_release(pp->pd);
pd                281 drivers/spi/spi-butterfly.c 	parport_unregister_device(pd);
pd                308 drivers/spi/spi-butterfly.c 	parport_release(pp->pd);
pd                309 drivers/spi/spi-butterfly.c 	parport_unregister_device(pp->pd);
pd                 73 drivers/spi/spi-lm70llp.c 	struct pardevice	*pd;
pd                190 drivers/spi/spi-lm70llp.c 	struct pardevice	*pd;
pd                227 drivers/spi/spi-lm70llp.c 	pd = parport_register_dev_model(p, DRVNAME, &lm70llp_cb, 0);
pd                229 drivers/spi/spi-lm70llp.c 	if (!pd) {
pd                233 drivers/spi/spi-lm70llp.c 	pp->pd = pd;
pd                235 drivers/spi/spi-lm70llp.c 	status = parport_claim(pd);
pd                244 drivers/spi/spi-lm70llp.c 		dev_warn(&pd->dev, "spi_bitbang_start failed with status %d\n",
pd                272 drivers/spi/spi-lm70llp.c 		dev_warn(&pd->dev, "spi_new_device failed\n");
pd                287 drivers/spi/spi-lm70llp.c 	parport_release(pp->pd);
pd                289 drivers/spi/spi-lm70llp.c 	parport_unregister_device(pd);
pd                309 drivers/spi/spi-lm70llp.c 	parport_release(pp->pd);
pd                310 drivers/spi/spi-lm70llp.c 	parport_unregister_device(pp->pd);
pd               2093 drivers/spi/spi-pl022.c 	struct pl022_ssp_controller *pd;
pd               2101 drivers/spi/spi-pl022.c 	pd = devm_kzalloc(dev, sizeof(struct pl022_ssp_controller), GFP_KERNEL);
pd               2102 drivers/spi/spi-pl022.c 	if (!pd)
pd               2105 drivers/spi/spi-pl022.c 	pd->bus_id = -1;
pd               2106 drivers/spi/spi-pl022.c 	pd->enable_dma = 1;
pd               2108 drivers/spi/spi-pl022.c 	pd->num_chipselect = tmp;
pd               2110 drivers/spi/spi-pl022.c 			     &pd->autosuspend_delay);
pd               2111 drivers/spi/spi-pl022.c 	pd->rt = of_property_read_bool(np, "pl022,rt");
pd               2113 drivers/spi/spi-pl022.c 	return pd;
pd                475 drivers/staging/comedi/drivers/comedi_test.c 				unsigned short *pd;
pd                477 drivers/staging/comedi/drivers/comedi_test.c 				pd = &devpriv->ao_loopbacks[chan];
pd                479 drivers/staging/comedi/drivers/comedi_test.c 				if (!comedi_buf_read_samples(s, pd, 1)) {
pd                552 drivers/staging/fieldbus/anybuss/host.c 	struct area_priv *pd = &t->area_pd;
pd                556 drivers/staging/fieldbus/anybuss/host.c 	if (atomic_read(&cd->ind_ab) & pd->flags) {
pd                567 drivers/staging/fieldbus/anybuss/host.c 	struct area_priv *pd = &t->area_pd;
pd                574 drivers/staging/fieldbus/anybuss/host.c 	if (!(atomic_read(&cd->ind_ab) & pd->flags)) {
pd                584 drivers/staging/fieldbus/anybuss/host.c 	if (pd->is_write)
pd                585 drivers/staging/fieldbus/anybuss/host.c 		regmap_bulk_write(cd->regmap, pd->addr, pd->buf,
pd                586 drivers/staging/fieldbus/anybuss/host.c 				  pd->count);
pd                588 drivers/staging/fieldbus/anybuss/host.c 		regmap_bulk_read(cd->regmap, pd->addr, pd->buf,
pd                589 drivers/staging/fieldbus/anybuss/host.c 				 pd->count);
pd                592 drivers/staging/fieldbus/anybuss/host.c 	ind_ap |= pd->flags;
pd                602 drivers/staging/fieldbus/anybuss/host.c 	struct area_priv *pd = &t->area_pd;
pd                611 drivers/staging/fieldbus/anybuss/host.c 	ind_ap |= pd->flags | IND_AP_ACTION | IND_AP_LOCK;
pd                698 drivers/staging/fieldbus/anybuss/host.c 	struct mbox_priv *pd = &t->mbox_pd;
pd                711 drivers/staging/fieldbus/anybuss/host.c 	regmap_bulk_read(cd->regmap, MBOX_OUT_AREA, &pd->hdr,
pd                712 drivers/staging/fieldbus/anybuss/host.c 			 sizeof(pd->hdr));
pd                713 drivers/staging/fieldbus/anybuss/host.c 	regmap_bulk_read(cd->regmap, MBOX_OUT_AREA + sizeof(pd->hdr),
pd                714 drivers/staging/fieldbus/anybuss/host.c 			 pd->msg, pd->msg_in_sz);
pd                722 drivers/staging/fieldbus/anybuss/host.c 	struct mbox_priv *pd = &t->mbox_pd;
pd                736 drivers/staging/fieldbus/anybuss/host.c 	regmap_bulk_write(cd->regmap, MBOX_IN_AREA, &pd->hdr,
pd                737 drivers/staging/fieldbus/anybuss/host.c 			  sizeof(pd->hdr));
pd                738 drivers/staging/fieldbus/anybuss/host.c 	regmap_bulk_write(cd->regmap, MBOX_IN_AREA + sizeof(pd->hdr),
pd                739 drivers/staging/fieldbus/anybuss/host.c 			  pd->msg, pd->msg_out_sz);
pd                811 drivers/staging/fieldbus/anybuss/host.c 	struct mbox_priv *pd;
pd                824 drivers/staging/fieldbus/anybuss/host.c 	pd = &t->mbox_pd;
pd                825 drivers/staging/fieldbus/anybuss/host.c 	h = &pd->hdr;
pd                841 drivers/staging/fieldbus/anybuss/host.c 	memcpy(pd->msg, msg_out, msg_out_sz);
pd                842 drivers/staging/fieldbus/anybuss/host.c 	pd->msg_out_sz = msg_out_sz;
pd                843 drivers/staging/fieldbus/anybuss/host.c 	pd->msg_in_sz = msg_in_sz;
pd                851 drivers/staging/fieldbus/anybuss/host.c 	err = mbox_cmd_err(cd->dev, pd);
pd                854 drivers/staging/fieldbus/anybuss/host.c 	memcpy(msg_in, pd->msg, msg_in_sz);
pd                 70 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd = card->proclog;
pd                 74 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if (!pd)
pd                 77 drivers/staging/isdn/hysdn/hysdn_proclog.c 	cp = pd->logtmp;
pd                 87 drivers/staging/isdn/hysdn/hysdn_proclog.c 		printk(KERN_INFO "%s", pd->logtmp);
pd                 89 drivers/staging/isdn/hysdn/hysdn_proclog.c 		put_log_buffer(card, pd->logtmp);
pd                103 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd = card->proclog;
pd                106 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if (!pd)
pd                112 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if (pd->if_used <= 0)
pd                119 drivers/staging/isdn/hysdn/hysdn_proclog.c 	ib->proc_ctrl = pd;	/* point to own control structure */
pd                121 drivers/staging/isdn/hysdn/hysdn_proclog.c 	ib->usage_cnt = pd->if_used;
pd                122 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if (!pd->log_head)
pd                123 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->log_head = ib;	/* new head */
pd                125 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->log_tail->next = ib;	/* follows existing messages */
pd                126 drivers/staging/isdn/hysdn/hysdn_proclog.c 	pd->log_tail = ib;	/* new tail */
pd                129 drivers/staging/isdn/hysdn/hysdn_proclog.c 	while (pd->log_head->next) {
pd                130 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if ((pd->log_head->usage_cnt <= 0) &&
pd                131 drivers/staging/isdn/hysdn/hysdn_proclog.c 		    (pd->log_head->next->usage_cnt <= 0)) {
pd                132 drivers/staging/isdn/hysdn/hysdn_proclog.c 			ib = pd->log_head;
pd                133 drivers/staging/isdn/hysdn/hysdn_proclog.c 			pd->log_head = pd->log_head->next;
pd                142 drivers/staging/isdn/hysdn/hysdn_proclog.c 	wake_up_interruptible(&(pd->rd_queue));		/* announce new entry */
pd                177 drivers/staging/isdn/hysdn/hysdn_proclog.c 		struct procdata *pd = card->proclog;
pd                181 drivers/staging/isdn/hysdn/hysdn_proclog.c 		wait_event_interruptible(pd->rd_queue, (inf =
pd                211 drivers/staging/isdn/hysdn/hysdn_proclog.c 		struct procdata *pd = card->proclog;
pd                216 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->if_used++;
pd                217 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if (pd->log_head)
pd                218 drivers/staging/isdn/hysdn/hysdn_proclog.c 			filep->private_data = &pd->log_tail->next;
pd                220 drivers/staging/isdn/hysdn/hysdn_proclog.c 			filep->private_data = &pd->log_head;
pd                241 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd;
pd                254 drivers/staging/isdn/hysdn/hysdn_proclog.c 			pd = (struct procdata *) inf->proc_ctrl;	/* still entries there */
pd                258 drivers/staging/isdn/hysdn/hysdn_proclog.c 			pd = card->proclog;	/* pointer to procfs log */
pd                260 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if (pd)
pd                261 drivers/staging/isdn/hysdn/hysdn_proclog.c 			pd->if_used--;	/* decrement interface usage count by one */
pd                268 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if (pd)
pd                269 drivers/staging/isdn/hysdn/hysdn_proclog.c 			if (pd->if_used <= 0)	/* delete buffers if last file closed */
pd                270 drivers/staging/isdn/hysdn/hysdn_proclog.c 				while (pd->log_head) {
pd                271 drivers/staging/isdn/hysdn/hysdn_proclog.c 					inf = pd->log_head;
pd                272 drivers/staging/isdn/hysdn/hysdn_proclog.c 					pd->log_head = pd->log_head->next;
pd                289 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd = card->proclog;
pd                294 drivers/staging/isdn/hysdn/hysdn_proclog.c 	poll_wait(file, &(pd->rd_queue), wait);
pd                324 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd;
pd                328 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if ((pd = kzalloc(sizeof(struct procdata), GFP_KERNEL)) != NULL) {
pd                329 drivers/staging/isdn/hysdn/hysdn_proclog.c 		sprintf(pd->log_name, "%s%d", PROC_LOG_BASENAME, card->myid);
pd                330 drivers/staging/isdn/hysdn/hysdn_proclog.c 		pd->log = proc_create_data(pd->log_name,
pd                334 drivers/staging/isdn/hysdn/hysdn_proclog.c 		init_waitqueue_head(&(pd->rd_queue));
pd                336 drivers/staging/isdn/hysdn/hysdn_proclog.c 		card->proclog = (void *) pd;	/* remember procfs structure */
pd                349 drivers/staging/isdn/hysdn/hysdn_proclog.c 	struct procdata *pd;
pd                351 drivers/staging/isdn/hysdn/hysdn_proclog.c 	if ((pd = (struct procdata *) card->proclog) != NULL) {
pd                352 drivers/staging/isdn/hysdn/hysdn_proclog.c 		if (pd->log)
pd                353 drivers/staging/isdn/hysdn/hysdn_proclog.c 			remove_proc_entry(pd->log_name, hysdn_proc_entry);
pd                354 drivers/staging/isdn/hysdn/hysdn_proclog.c 		kfree(pd);	/* release memory */
pd                781 drivers/usb/gadget/function/f_uvc.c 	struct uvc_processing_unit_descriptor *pd;
pd                809 drivers/usb/gadget/function/f_uvc.c 	pd = &opts->uvc_processing;
pd                810 drivers/usb/gadget/function/f_uvc.c 	pd->bLength			= UVC_DT_PROCESSING_UNIT_SIZE(2);
pd                811 drivers/usb/gadget/function/f_uvc.c 	pd->bDescriptorType		= USB_DT_CS_INTERFACE;
pd                812 drivers/usb/gadget/function/f_uvc.c 	pd->bDescriptorSubType		= UVC_VC_PROCESSING_UNIT;
pd                813 drivers/usb/gadget/function/f_uvc.c 	pd->bUnitID			= 2;
pd                814 drivers/usb/gadget/function/f_uvc.c 	pd->bSourceID			= 1;
pd                815 drivers/usb/gadget/function/f_uvc.c 	pd->wMaxMultiplier		= cpu_to_le16(16*1024);
pd                816 drivers/usb/gadget/function/f_uvc.c 	pd->bControlSize		= 2;
pd                817 drivers/usb/gadget/function/f_uvc.c 	pd->bmControls[0]		= 1;
pd                818 drivers/usb/gadget/function/f_uvc.c 	pd->bmControls[1]		= 0;
pd                819 drivers/usb/gadget/function/f_uvc.c 	pd->iProcessing			= 0;
pd                843 drivers/usb/gadget/function/f_uvc.c 	ctl_cls[2] = (struct uvc_descriptor_header *)pd;
pd                853 drivers/usb/gadget/function/f_uvc.c 	ctl_cls[2] = (struct uvc_descriptor_header *)pd;
pd                267 drivers/usb/gadget/function/uvc_configfs.c 	struct uvc_processing_unit_descriptor *pd;			\
pd                274 drivers/usb/gadget/function/uvc_configfs.c 	pd = &opts->uvc_processing;					\
pd                277 drivers/usb/gadget/function/uvc_configfs.c 	result = sprintf(page, "%u\n", le##bits##_to_cpu(pd->aname));	\
pd                300 drivers/usb/gadget/function/uvc_configfs.c 	struct uvc_processing_unit_descriptor *pd;
pd                308 drivers/usb/gadget/function/uvc_configfs.c 	pd = &opts->uvc_processing;
pd                311 drivers/usb/gadget/function/uvc_configfs.c 	for (result = 0, i = 0; i < pd->bControlSize; ++i) {
pd                312 drivers/usb/gadget/function/uvc_configfs.c 		result += sprintf(pg, "%u\n", pd->bmControls[i]);
pd                296 drivers/usb/gadget/udc/bcm63xx_udc.c 	struct bcm63xx_usbd_platform_data *pd;
pd                872 drivers/usb/gadget/udc/bcm63xx_udc.c 	u32 val, portmask = BIT(udc->pd->port_no);
pd                912 drivers/usb/gadget/udc/bcm63xx_udc.c 	u32 val, portmask = BIT(udc->pd->port_no);
pd               2283 drivers/usb/gadget/udc/bcm63xx_udc.c 	struct bcm63xx_usbd_platform_data *pd = dev_get_platdata(dev);
pd               2294 drivers/usb/gadget/udc/bcm63xx_udc.c 	udc->pd = pd;
pd               2296 drivers/usb/gadget/udc/bcm63xx_udc.c 	if (!pd) {
pd               2317 drivers/usb/gadget/udc/bcm63xx_udc.c 	if (!pd->use_fullspeed && !use_fullspeed)
pd                144 drivers/usb/gadget/udc/s3c-hsudc.c 	struct s3c24xx_hsudc_platdata *pd;
pd                293 drivers/usb/gadget/udc/s3c-hsudc.c 	for (epnum = 0; epnum < hsudc->pd->epnum; epnum++) {
pd               1036 drivers/usb/gadget/udc/s3c-hsudc.c 	for (epnum = 0; epnum < hsudc->pd->epnum; epnum++)
pd               1113 drivers/usb/gadget/udc/s3c-hsudc.c 			for (ep_idx = 0; ep_idx < hsudc->pd->epnum; ep_idx++) {
pd               1192 drivers/usb/gadget/udc/s3c-hsudc.c 	if (hsudc->pd->gpio_init)
pd               1193 drivers/usb/gadget/udc/s3c-hsudc.c 		hsudc->pd->gpio_init();
pd               1217 drivers/usb/gadget/udc/s3c-hsudc.c 	if (hsudc->pd->gpio_uninit)
pd               1218 drivers/usb/gadget/udc/s3c-hsudc.c 		hsudc->pd->gpio_uninit();
pd               1268 drivers/usb/gadget/udc/s3c-hsudc.c 	struct s3c24xx_hsudc_platdata *pd = dev_get_platdata(&pdev->dev);
pd               1272 drivers/usb/gadget/udc/s3c-hsudc.c 			sizeof(struct s3c_hsudc_ep) * pd->epnum,
pd               1279 drivers/usb/gadget/udc/s3c-hsudc.c 	hsudc->pd = dev_get_platdata(&pdev->dev);
pd                209 drivers/usb/host/ehci-orion.c 	struct orion_ehci_data *pd = dev_get_platdata(&pdev->dev);
pd                290 drivers/usb/host/ehci-orion.c 		phy_version = pd->phy_version;
pd                310 drivers/video/backlight/ams369fg06.c 	struct lcd_platform_data *pd;
pd                313 drivers/video/backlight/ams369fg06.c 	pd = lcd->lcd_pd;
pd                316 drivers/video/backlight/ams369fg06.c 	if (pd->power_on) {
pd                317 drivers/video/backlight/ams369fg06.c 		pd->power_on(lcd->ld, 1);
pd                318 drivers/video/backlight/ams369fg06.c 		msleep(pd->power_on_delay);
pd                321 drivers/video/backlight/ams369fg06.c 	if (!pd->reset) {
pd                326 drivers/video/backlight/ams369fg06.c 	pd->reset(lcd->ld);
pd                327 drivers/video/backlight/ams369fg06.c 	msleep(pd->reset_delay);
pd                354 drivers/video/backlight/ams369fg06.c 	struct lcd_platform_data *pd;
pd                356 drivers/video/backlight/ams369fg06.c 	pd = lcd->lcd_pd;
pd                364 drivers/video/backlight/ams369fg06.c 	msleep(pd->power_off_delay);
pd                366 drivers/video/backlight/ams369fg06.c 	if (pd->power_on)
pd                367 drivers/video/backlight/ams369fg06.c 		pd->power_on(lcd->ld, 0);
pd                224 drivers/video/backlight/lms501kf03.c 	struct lcd_platform_data *pd;
pd                226 drivers/video/backlight/lms501kf03.c 	pd = lcd->lcd_pd;
pd                228 drivers/video/backlight/lms501kf03.c 	if (!pd->power_on) {
pd                233 drivers/video/backlight/lms501kf03.c 	pd->power_on(lcd->ld, 1);
pd                234 drivers/video/backlight/lms501kf03.c 	msleep(pd->power_on_delay);
pd                236 drivers/video/backlight/lms501kf03.c 	if (!pd->reset) {
pd                241 drivers/video/backlight/lms501kf03.c 	pd->reset(lcd->ld);
pd                242 drivers/video/backlight/lms501kf03.c 	msleep(pd->reset_delay);
pd                262 drivers/video/backlight/lms501kf03.c 	struct lcd_platform_data *pd;
pd                264 drivers/video/backlight/lms501kf03.c 	pd = lcd->lcd_pd;
pd                272 drivers/video/backlight/lms501kf03.c 	msleep(pd->power_off_delay);
pd                274 drivers/video/backlight/lms501kf03.c 	pd->power_on(lcd->ld, 0);
pd                171 drivers/video/backlight/lp855x_bl.c 	struct lp855x_platform_data *pd = lp->pdata;
pd                197 drivers/video/backlight/lp855x_bl.c 	val = pd->initial_brightness;
pd                202 drivers/video/backlight/lp855x_bl.c 	val = pd->device_control;
pd                207 drivers/video/backlight/lp855x_bl.c 	if (pd->size_program > 0) {
pd                208 drivers/video/backlight/lp855x_bl.c 		for (i = 0; i < pd->size_program; i++) {
pd                209 drivers/video/backlight/lp855x_bl.c 			addr = pd->rom_data[i].addr;
pd                210 drivers/video/backlight/lp855x_bl.c 			val = pd->rom_data[i].val;
pd                443 drivers/video/console/sticore.c 		if (sti->pd) {
pd                456 drivers/video/console/sticore.c 			newhpa = pci_resource_start (sti->pd, (offs - PCI_BASE_ADDRESS_0) / 4);
pd                819 drivers/video/console/sticore.c 					      struct pci_dev *pd)
pd                857 drivers/video/console/sticore.c 		sti->pd = pd;
pd                905 drivers/video/console/sticore.c 	if (sti->pd) {
pd                907 drivers/video/console/sticore.c 		rom_base = pci_resource_start(sti->pd, PCI_ROM_RESOURCE);	
pd                908 drivers/video/console/sticore.c 		pci_write_config_dword(sti->pd, PCI_ROM_ADDRESS, rom_base & ~PCI_ROM_ADDRESS_ENABLE);
pd                964 drivers/video/console/sticore.c static int sticore_pci_init(struct pci_dev *pd, const struct pci_device_id *ent)
pd                972 drivers/video/console/sticore.c 	err = pci_enable_device(pd);
pd                974 drivers/video/console/sticore.c 		dev_err(&pd->dev, "Cannot enable PCI device\n");
pd                978 drivers/video/console/sticore.c 	fb_base = pci_resource_start(pd, 0);
pd                979 drivers/video/console/sticore.c 	fb_len = pci_resource_len(pd, 0);
pd                980 drivers/video/console/sticore.c 	rom_base = pci_resource_start(pd, PCI_ROM_RESOURCE);
pd                981 drivers/video/console/sticore.c 	rom_len = pci_resource_len(pd, PCI_ROM_RESOURCE);
pd                983 drivers/video/console/sticore.c 		pci_write_config_dword(pd, PCI_ROM_ADDRESS, rom_base | PCI_ROM_ADDRESS_ENABLE);
pd                993 drivers/video/console/sticore.c 	sti = sti_try_rom_generic(rom_base, fb_base, pd);
pd                996 drivers/video/console/sticore.c 		print_pci_hwpath(pd, pa_path);
pd               1002 drivers/video/console/sticore.c 			pci_name(pd));
pd               1011 drivers/video/console/sticore.c static void __exit sticore_pci_remove(struct pci_dev *pd)
pd                149 drivers/video/fbdev/au1200fb.c 	struct au1200fb_platdata *pd;
pd                773 drivers/video/fbdev/au1200fb.c 			    struct au1200fb_platdata *pd)
pd                806 drivers/video/fbdev/au1200fb.c 		if (pd->panel_shutdown)
pd                807 drivers/video/fbdev/au1200fb.c 			pd->panel_shutdown();
pd                866 drivers/video/fbdev/au1200fb.c 	if (pd->panel_init)
pd                867 drivers/video/fbdev/au1200fb.c 		pd->panel_init();
pd               1211 drivers/video/fbdev/au1200fb.c 		au1200_setpanel(panel, fbdev->pd);
pd               1217 drivers/video/fbdev/au1200fb.c 		au1200_setpanel(NULL, fbdev->pd);
pd               1462 drivers/video/fbdev/au1200fb.c 				au1200_setpanel(newpanel, fbdev->pd);
pd               1578 drivers/video/fbdev/au1200fb.c static int au1200fb_setup(struct au1200fb_platdata *pd)
pd               1604 drivers/video/fbdev/au1200fb.c 				panel_idx = pd->panel_index();
pd               1647 drivers/video/fbdev/au1200fb.c 	struct au1200fb_platdata *pd;
pd               1653 drivers/video/fbdev/au1200fb.c 	pd = dev->dev.platform_data;
pd               1654 drivers/video/fbdev/au1200fb.c 	if (!pd)
pd               1658 drivers/video/fbdev/au1200fb.c 	if (au1200fb_setup(pd))
pd               1685 drivers/video/fbdev/au1200fb.c 		fbdev->pd = pd;
pd               1740 drivers/video/fbdev/au1200fb.c 	platform_set_drvdata(dev, pd);
pd               1743 drivers/video/fbdev/au1200fb.c 	au1200_setpanel(panel, pd);
pd               1767 drivers/video/fbdev/au1200fb.c 	struct au1200fb_platdata *pd = platform_get_drvdata(dev);
pd               1772 drivers/video/fbdev/au1200fb.c 	au1200_setpanel(NULL, pd);
pd               1795 drivers/video/fbdev/au1200fb.c 	struct au1200fb_platdata *pd = dev_get_drvdata(dev);
pd               1796 drivers/video/fbdev/au1200fb.c 	au1200_setpanel(NULL, pd);
pd               1806 drivers/video/fbdev/au1200fb.c 	struct au1200fb_platdata *pd = dev_get_drvdata(dev);
pd               1811 drivers/video/fbdev/au1200fb.c 	au1200_setpanel(panel, pd);
pd               1270 drivers/video/fbdev/nvidia/nvidia.c static int nvidiafb_probe(struct pci_dev *pd, const struct pci_device_id *ent)
pd               1278 drivers/video/fbdev/nvidia/nvidia.c 	assert(pd != NULL);
pd               1280 drivers/video/fbdev/nvidia/nvidia.c 	info = framebuffer_alloc(sizeof(struct nvidia_par), &pd->dev);
pd               1286 drivers/video/fbdev/nvidia/nvidia.c 	par->pci_dev = pd;
pd               1292 drivers/video/fbdev/nvidia/nvidia.c 	if (pci_enable_device(pd)) {
pd               1297 drivers/video/fbdev/nvidia/nvidia.c 	if (pci_request_regions(pd, "nvidiafb")) {
pd               1313 drivers/video/fbdev/nvidia/nvidia.c 	pci_read_config_word(pd, PCI_COMMAND, &cmd);
pd               1315 drivers/video/fbdev/nvidia/nvidia.c 	pci_write_config_word(pd, PCI_COMMAND, cmd);
pd               1317 drivers/video/fbdev/nvidia/nvidia.c 	nvidiafb_fix.mmio_start = pci_resource_start(pd, 0);
pd               1318 drivers/video/fbdev/nvidia/nvidia.c 	nvidiafb_fix.smem_start = pci_resource_start(pd, 1);
pd               1319 drivers/video/fbdev/nvidia/nvidia.c 	nvidiafb_fix.mmio_len = pci_resource_len(pd, 0);
pd               1336 drivers/video/fbdev/nvidia/nvidia.c 	sprintf(nvidiafb_fix.id, "NV%x", (pd->device & 0x0ff0) >> 4);
pd               1385 drivers/video/fbdev/nvidia/nvidia.c 	pci_set_drvdata(pd, info);
pd               1412 drivers/video/fbdev/nvidia/nvidia.c 	pci_release_regions(pd);
pd               1421 drivers/video/fbdev/nvidia/nvidia.c static void nvidiafb_remove(struct pci_dev *pd)
pd               1423 drivers/video/fbdev/nvidia/nvidia.c 	struct fb_info *info = pci_get_drvdata(pd);
pd               1436 drivers/video/fbdev/nvidia/nvidia.c 	pci_release_regions(pd);
pd                202 drivers/video/fbdev/omap/lcd_mipid.c 	struct mipid_platform_data *pd = md->spi->dev.platform_data;
pd                204 drivers/video/fbdev/omap/lcd_mipid.c 	if (pd->get_bklight_max == NULL || pd->set_bklight_level == NULL)
pd                206 drivers/video/fbdev/omap/lcd_mipid.c 	if (level > pd->get_bklight_max(pd))
pd                212 drivers/video/fbdev/omap/lcd_mipid.c 	pd->set_bklight_level(pd, level);
pd                220 drivers/video/fbdev/omap/lcd_mipid.c 	struct mipid_platform_data *pd = md->spi->dev.platform_data;
pd                222 drivers/video/fbdev/omap/lcd_mipid.c 	if (pd->get_bklight_level == NULL)
pd                224 drivers/video/fbdev/omap/lcd_mipid.c 	return pd->get_bklight_level(pd);
pd                230 drivers/video/fbdev/omap/lcd_mipid.c 	struct mipid_platform_data *pd = md->spi->dev.platform_data;
pd                232 drivers/video/fbdev/omap/lcd_mipid.c 	if (pd->get_bklight_max == NULL)
pd                235 drivers/video/fbdev/omap/lcd_mipid.c 	return pd->get_bklight_max(pd);
pd               1729 drivers/video/fbdev/riva/fbdev.c static int riva_get_EDID_OF(struct fb_info *info, struct pci_dev *pd)
pd               1740 drivers/video/fbdev/riva/fbdev.c 	dp = pci_device_to_OF_node(pd);
pd               1856 drivers/video/fbdev/riva/fbdev.c static u32 riva_get_arch(struct pci_dev *pd)
pd               1860 drivers/video/fbdev/riva/fbdev.c 	switch (pd->device & 0x0ff0) {
pd               1894 drivers/video/fbdev/riva/fbdev.c static int rivafb_probe(struct pci_dev *pd, const struct pci_device_id *ent)
pd               1901 drivers/video/fbdev/riva/fbdev.c 	assert(pd != NULL);
pd               1903 drivers/video/fbdev/riva/fbdev.c 	info = framebuffer_alloc(sizeof(struct riva_par), &pd->dev);
pd               1909 drivers/video/fbdev/riva/fbdev.c 	default_par->pdev = pd;
pd               1917 drivers/video/fbdev/riva/fbdev.c 	ret = pci_enable_device(pd);
pd               1923 drivers/video/fbdev/riva/fbdev.c 	ret = pci_request_regions(pd, "rivafb");
pd               1930 drivers/video/fbdev/riva/fbdev.c 	default_par->riva.Architecture = riva_get_arch(pd);
pd               1932 drivers/video/fbdev/riva/fbdev.c 	default_par->Chipset = (pd->vendor << 16) | pd->device;
pd               1943 drivers/video/fbdev/riva/fbdev.c 		sprintf(rivafb_fix.id, "NV%x", (pd->device & 0x0ff0) >> 4);
pd               1953 drivers/video/fbdev/riva/fbdev.c 	rivafb_fix.mmio_len = pci_resource_len(pd, 0);
pd               1954 drivers/video/fbdev/riva/fbdev.c 	rivafb_fix.smem_len = pci_resource_len(pd, 1);
pd               1960 drivers/video/fbdev/riva/fbdev.c 		pci_read_config_word(pd, PCI_COMMAND, &cmd);
pd               1962 drivers/video/fbdev/riva/fbdev.c 		pci_write_config_word(pd, PCI_COMMAND, cmd);
pd               1965 drivers/video/fbdev/riva/fbdev.c 	rivafb_fix.mmio_start = pci_resource_start(pd, 0);
pd               1966 drivers/video/fbdev/riva/fbdev.c 	rivafb_fix.smem_start = pci_resource_start(pd, 1);
pd               2023 drivers/video/fbdev/riva/fbdev.c 	riva_get_EDID(info, pd);
pd               2035 drivers/video/fbdev/riva/fbdev.c 	pci_set_drvdata(pd, info);
pd               2068 drivers/video/fbdev/riva/fbdev.c 	pci_release_regions(pd);
pd               2078 drivers/video/fbdev/riva/fbdev.c static void rivafb_remove(struct pci_dev *pd)
pd               2080 drivers/video/fbdev/riva/fbdev.c 	struct fb_info *info = pci_get_drvdata(pd);
pd               2098 drivers/video/fbdev/riva/fbdev.c 	pci_release_regions(pd);
pd               1360 drivers/video/fbdev/s3c-fb.c 	struct s3c_fb_platdata *pd;
pd               1375 drivers/video/fbdev/s3c-fb.c 	pd = dev_get_platdata(&pdev->dev);
pd               1376 drivers/video/fbdev/s3c-fb.c 	if (!pd) {
pd               1388 drivers/video/fbdev/s3c-fb.c 	sfb->pdata = pd;
pd               1442 drivers/video/fbdev/s3c-fb.c 	pd->setup_gpio();
pd               1444 drivers/video/fbdev/s3c-fb.c 	writel(pd->vidcon1, sfb->regs + VIDCON1);
pd               1473 drivers/video/fbdev/s3c-fb.c 		if (!pd->win[win])
pd               1566 drivers/video/fbdev/s3c-fb.c 	struct s3c_fb_platdata *pd = sfb->pdata;
pd               1579 drivers/video/fbdev/s3c-fb.c 	pd->setup_gpio();
pd               1580 drivers/video/fbdev/s3c-fb.c 	writel(pd->vidcon1, sfb->regs + VIDCON1);
pd               1641 drivers/video/fbdev/s3c-fb.c 	struct s3c_fb_platdata *pd = sfb->pdata;
pd               1649 drivers/video/fbdev/s3c-fb.c 	pd->setup_gpio();
pd               1650 drivers/video/fbdev/s3c-fb.c 	writel(pd->vidcon1, sfb->regs + VIDCON1);
pd                 34 drivers/video/fbdev/sh7760fb.c 	struct sh7760fb_platdata *pd;	/* display information */
pd                 73 drivers/video/fbdev/sh7760fb.c 	struct sh7760fb_platdata *pd = par->pd;
pd                 88 drivers/video/fbdev/sh7760fb.c 	if (pd->blank)
pd                 89 drivers/video/fbdev/sh7760fb.c 		pd->blank(blank);
pd                173 drivers/video/fbdev/sh7760fb.c 	ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, NULL);
pd                199 drivers/video/fbdev/sh7760fb.c 	struct fb_videomode *vm = par->pd->def_mode;
pd                206 drivers/video/fbdev/sh7760fb.c 	par->rot = par->pd->rotate;
pd                225 drivers/video/fbdev/sh7760fb.c 	ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, &gray);
pd                234 drivers/video/fbdev/sh7760fb.c 	lddfr = par->pd->lddfr | (1 << 8);
pd                236 drivers/video/fbdev/sh7760fb.c 	lddfr = par->pd->lddfr & ~(1 << 8);
pd                239 drivers/video/fbdev/sh7760fb.c 	ldmtr = par->pd->ldmtr;
pd                249 drivers/video/fbdev/sh7760fb.c 	iowrite16(par->pd->ldickr, par->base + LDICKR);	/* pixclock */
pd                253 drivers/video/fbdev/sh7760fb.c 	iowrite16(par->pd->ldpmmr, par->base + LDPMMR);	/* Power Management */
pd                254 drivers/video/fbdev/sh7760fb.c 	iowrite16(par->pd->ldpspr, par->base + LDPSPR);	/* Power Supply Ctrl */
pd                266 drivers/video/fbdev/sh7760fb.c 	iowrite16(par->pd->ldaclnr, par->base + LDACLNR);
pd                334 drivers/video/fbdev/sh7760fb.c 		(par->pd->ldickr >> 12) & 3, par->pd->ldickr & 0x1f);
pd                335 drivers/video/fbdev/sh7760fb.c 	dev_dbg(info->dev, "ldpmmr: 0x%04x ldpspr: 0x%04x\n", par->pd->ldpmmr,
pd                336 drivers/video/fbdev/sh7760fb.c 		par->pd->ldpspr);
pd                384 drivers/video/fbdev/sh7760fb.c 	ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, NULL);
pd                451 drivers/video/fbdev/sh7760fb.c 	par->pd = pdev->dev.platform_data;
pd                452 drivers/video/fbdev/sh7760fb.c 	if (!par->pd) {
pd                485 drivers/video/fbdev/sh7760fb.c 	fb_videomode_to_var(&info->var, par->pd->def_mode);
pd                148 drivers/video/fbdev/simplefb.c 	struct simplefb_platform_data *pd = dev_get_platdata(&pdev->dev);
pd                151 drivers/video/fbdev/simplefb.c 	params->width = pd->width;
pd                152 drivers/video/fbdev/simplefb.c 	params->height = pd->height;
pd                153 drivers/video/fbdev/simplefb.c 	params->stride = pd->stride;
pd                157 drivers/video/fbdev/simplefb.c 		if (strcmp(pd->format, simplefb_formats[i].name))
pd                716 drivers/video/fbdev/sm501fb.c 	struct sm501_platdata_fbsub *pd = fbi->pdata->fb_pnl;
pd                735 drivers/video/fbdev/sm501fb.c 		if (!(pd->flags & SM501FB_FLAG_PANEL_NO_VBIASEN)) {
pd                736 drivers/video/fbdev/sm501fb.c 			if (pd->flags & SM501FB_FLAG_PANEL_INV_VBIASEN)
pd                746 drivers/video/fbdev/sm501fb.c 		if (!(pd->flags & SM501FB_FLAG_PANEL_NO_FPEN)) {
pd                747 drivers/video/fbdev/sm501fb.c 			if (pd->flags & SM501FB_FLAG_PANEL_INV_FPEN)
pd                758 drivers/video/fbdev/sm501fb.c 		if (!(pd->flags & SM501FB_FLAG_PANEL_NO_FPEN)) {
pd                759 drivers/video/fbdev/sm501fb.c 			if (pd->flags & SM501FB_FLAG_PANEL_INV_FPEN)
pd                769 drivers/video/fbdev/sm501fb.c 		if (!(pd->flags & SM501FB_FLAG_PANEL_NO_VBIASEN)) {
pd                770 drivers/video/fbdev/sm501fb.c 			if (pd->flags & SM501FB_FLAG_PANEL_INV_VBIASEN)
pd               1677 drivers/video/fbdev/sm501fb.c 	struct sm501_platdata_fbsub *pd;
pd               1686 drivers/video/fbdev/sm501fb.c 		pd = info->pdata->fb_crt;
pd               1699 drivers/video/fbdev/sm501fb.c 		pd = info->pdata->fb_pnl;
pd               1705 drivers/video/fbdev/sm501fb.c 		pd = NULL;		/* stop compiler warnings */
pd               1730 drivers/video/fbdev/sm501fb.c 	if ((pd->flags & SM501FB_FLAG_USE_HWCURSOR) == 0)
pd               1772 drivers/video/fbdev/sm501fb.c 	if (enable && (pd->flags & SM501FB_FLAG_USE_INIT_MODE) && 0) {
pd               1775 drivers/video/fbdev/sm501fb.c 		if (pd->def_mode) {
pd               1777 drivers/video/fbdev/sm501fb.c 			fb_videomode_to_var(&fb->var, pd->def_mode);
pd               1779 drivers/video/fbdev/sm501fb.c 			fb->var.bits_per_pixel = pd->def_bpp ? pd->def_bpp : 8;
pd               1863 drivers/video/fbdev/sm501fb.c 	struct sm501_platdata_fbsub *pd;
pd               1867 drivers/video/fbdev/sm501fb.c 	pd = (head == HEAD_CRT) ? info->pdata->fb_crt : info->pdata->fb_pnl;
pd               1870 drivers/video/fbdev/sm501fb.c 	if (pd == NULL) {
pd               1948 drivers/video/fbdev/sm501fb.c 		struct sm501_platdata *pd = dev->parent->platform_data;
pd               1949 drivers/video/fbdev/sm501fb.c 		info->pdata = pd->fb;
pd                367 drivers/video/fbdev/sticore.h 	struct pci_dev *pd;
pd                565 drivers/video/fbdev/xen-fbfront.c 		info->page->pd[i] = vmalloc_to_gfn(&info->gfns[i * epd]);
pd                684 fs/cifs/smbdirect.c 	info->pd = ib_alloc_pd(info->id->device, 0);
pd                685 fs/cifs/smbdirect.c 	if (IS_ERR(info->pd)) {
pd                686 fs/cifs/smbdirect.c 		rc = PTR_ERR(info->pd);
pd                740 fs/cifs/smbdirect.c 	request->sge[0].lkey = info->pd->local_dma_lkey;
pd                905 fs/cifs/smbdirect.c 	request->sge[0].lkey = info->pd->local_dma_lkey;
pd               1007 fs/cifs/smbdirect.c 		request->sge[i+1].lkey = info->pd->local_dma_lkey;
pd               1102 fs/cifs/smbdirect.c 	response->sge.lkey = info->pd->local_dma_lkey;
pd               1466 fs/cifs/smbdirect.c 	ib_dealloc_pd(info->pd);
pd               1689 fs/cifs/smbdirect.c 	rc = rdma_create_qp(info->id, info->pd, &qp_attr);
pd               1810 fs/cifs/smbdirect.c 	ib_dealloc_pd(info->pd);
pd               2287 fs/cifs/smbdirect.c 				info->pd, info->mr_type,
pd               2354 fs/cifs/smbdirect.c 		smbdirect_mr->mr = ib_alloc_mr(info->pd, info->mr_type,
pd                 58 fs/cifs/smbdirect.h 	struct ib_pd *pd;
pd                155 fs/xfs/xfs_quota.h #define xfs_trans_reserve_quota(tp, mp, ud, gd, pd, nb, ni, f) \
pd                156 fs/xfs/xfs_quota.h 	xfs_trans_reserve_quota_bydquots(tp, mp, ud, gd, pd, nb, ni, \
pd                133 include/linux/blk-cgroup.h 	struct blkg_policy_data		*pd[BLKCG_MAX_POLS];
pd                154 include/linux/blk-cgroup.h typedef void (blkcg_pol_init_pd_fn)(struct blkg_policy_data *pd);
pd                155 include/linux/blk-cgroup.h typedef void (blkcg_pol_online_pd_fn)(struct blkg_policy_data *pd);
pd                156 include/linux/blk-cgroup.h typedef void (blkcg_pol_offline_pd_fn)(struct blkg_policy_data *pd);
pd                157 include/linux/blk-cgroup.h typedef void (blkcg_pol_free_pd_fn)(struct blkg_policy_data *pd);
pd                158 include/linux/blk-cgroup.h typedef void (blkcg_pol_reset_pd_stats_fn)(struct blkg_policy_data *pd);
pd                159 include/linux/blk-cgroup.h typedef size_t (blkcg_pol_stat_pd_fn)(struct blkg_policy_data *pd, char *buf,
pd                218 include/linux/blk-cgroup.h u64 __blkg_prfill_u64(struct seq_file *sf, struct blkg_policy_data *pd, u64 v);
pd                219 include/linux/blk-cgroup.h u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd,
pd                221 include/linux/blk-cgroup.h u64 blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd,
pd                410 include/linux/blk-cgroup.h 	return blkg ? blkg->pd[pol->plid] : NULL;
pd                425 include/linux/blk-cgroup.h static inline struct blkcg_gq *pd_to_blkg(struct blkg_policy_data *pd)
pd                427 include/linux/blk-cgroup.h 	return pd ? pd->blkg : NULL;
pd                862 include/linux/blk-cgroup.h static inline struct blkcg_gq *pd_to_blkg(struct blkg_policy_data *pd) { return NULL; }
pd                 79 include/linux/energy_model.h static inline unsigned long em_pd_energy(struct em_perf_domain *pd,
pd                 91 include/linux/energy_model.h 	cpu = cpumask_first(to_cpumask(pd->cpus));
pd                 93 include/linux/energy_model.h 	cs = &pd->table[pd->nr_cap_states - 1];
pd                100 include/linux/energy_model.h 	for (i = 0; i < pd->nr_cap_states; i++) {
pd                101 include/linux/energy_model.h 		cs = &pd->table[i];
pd                157 include/linux/energy_model.h static inline int em_pd_nr_cap_states(struct em_perf_domain *pd)
pd                159 include/linux/energy_model.h 	return pd->nr_cap_states;
pd                176 include/linux/energy_model.h static inline unsigned long em_pd_energy(struct em_perf_domain *pd,
pd                181 include/linux/energy_model.h static inline int em_pd_nr_cap_states(struct em_perf_domain *pd)
pd                 43 include/linux/firmware/trusted_foundations.h void register_trusted_foundations(struct trusted_foundations_platform_data *pd);
pd                 53 include/linux/firmware/trusted_foundations.h 				   struct trusted_foundations_platform_data *pd)
pd                 19 include/linux/hugetlb.h typedef struct { unsigned long pd; } hugepd_t;
pd                693 include/linux/mlx4/device.h 	u32			pd;
pd                705 include/linux/mlx4/device.h 	u32			pd;
pd               1117 include/linux/mlx4/device.h int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access,
pd               1121 include/linux/mlx4/device.h int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type,
pd               1417 include/linux/mlx4/device.h int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages,
pd                194 include/linux/mlx4/qp.h 	__be32			pd;
pd                348 include/linux/mlx5/driver.h 		u32	pd;
pd                376 include/linux/mlx5/driver.h 	u32			pd;
pd               1609 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               2611 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               2856 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               2918 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               3013 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               3266 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               3469 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               7175 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               7525 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               8007 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd               10231 include/linux/mlx5/mlx5_ifc.h 	u8         pd[0x18];
pd                 36 include/linux/padata.h 	struct parallel_data	*pd;
pd                 66 include/linux/padata.h        struct parallel_data *pd;
pd                140 include/linux/padata.h 	struct parallel_data __rcu	*pd;
pd                140 include/linux/pktcdvd.h 	struct pktcdvd_device	*pd;
pd                151 include/linux/pktcdvd.h 	struct pktcdvd_device	*pd;
pd                158 include/linux/pktcdvd.h 	struct pktcdvd_device	*pd;
pd                167 include/linux/platform_data/cros_ec_proto.h 	struct platform_device *pd;
pd                 44 include/linux/platform_data/hwmon-s3c.h extern void __init s3c_hwmon_set_platdata(struct s3c_hwmon_pdata *pd);
pd                 37 include/linux/platform_data/s3c-hsotg.h extern void dwc2_hsotg_set_platdata(struct dwc2_hsotg_plat *pd);
pd                 37 include/linux/platform_data/usb-s3c2410_udc.h extern void __init s3c24xx_hsudc_set_platdata(struct s3c24xx_hsudc_platdata *pd);
pd                148 include/linux/pm_domain.h static inline struct generic_pm_domain *pd_to_genpd(struct dev_pm_domain *pd)
pd                150 include/linux/pm_domain.h 	return container_of(pd, struct generic_pm_domain, domain);
pd                369 include/linux/pm_domain.h void dev_pm_domain_set(struct device *dev, struct dev_pm_domain *pd);
pd                387 include/linux/pm_domain.h 				     struct dev_pm_domain *pd) {}
pd                253 include/linux/qed/qed_rdma_if.h 	u16 pd;
pd                312 include/linux/qed/qed_rdma_if.h 	u16 pd;
pd                612 include/linux/qed/qed_rdma_if.h 	int (*rdma_alloc_pd)(void *rdma_cxt, u16 *pd);
pd                613 include/linux/qed/qed_rdma_if.h 	void (*rdma_dealloc_pd)(void *rdma_cxt, u16 pd);
pd                123 include/net/dsa.h 	struct dsa_platform_data	*pd;
pd                428 include/net/ip_vs.h 	int (*init_netns)(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd);
pd                430 include/net/ip_vs.h 	void (*exit_netns)(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd);
pd                434 include/net/ip_vs.h 			     struct ip_vs_proto_data *pd,
pd                460 include/net/ip_vs.h 				 struct ip_vs_proto_data *pd);
pd                473 include/net/ip_vs.h 	void (*timeout_change)(struct ip_vs_proto_data *pd, int flags);
pd               1391 include/net/ip_vs.h 	       struct ip_vs_proto_data *pd, int *ignored,
pd               1394 include/net/ip_vs.h 		struct ip_vs_proto_data *pd, struct ip_vs_iphdr *iph);
pd                 79 include/rdma/ib_fmr_pool.h struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd             *pd,
pd               1527 include/rdma/ib_verbs.h 	struct ib_pd		*pd;
pd               1566 include/rdma/ib_verbs.h 	struct ib_pd	       *pd;
pd               1615 include/rdma/ib_verbs.h 	struct ib_pd	       *pd;
pd               1713 include/rdma/ib_verbs.h 	struct ib_pd	       *pd;
pd               1762 include/rdma/ib_verbs.h 	struct ib_pd	  *pd;
pd               1785 include/rdma/ib_verbs.h 	struct ib_pd		*pd;
pd               1793 include/rdma/ib_verbs.h 	struct ib_pd		*pd;
pd               2366 include/rdma/ib_verbs.h 	int (*alloc_pd)(struct ib_pd *pd, struct ib_udata *udata);
pd               2367 include/rdma/ib_verbs.h 	void (*dealloc_pd)(struct ib_pd *pd, struct ib_udata *udata);
pd               2381 include/rdma/ib_verbs.h 	struct ib_qp *(*create_qp)(struct ib_pd *pd,
pd               2394 include/rdma/ib_verbs.h 	struct ib_mr *(*get_dma_mr)(struct ib_pd *pd, int mr_access_flags);
pd               2395 include/rdma/ib_verbs.h 	struct ib_mr *(*reg_user_mr)(struct ib_pd *pd, u64 start, u64 length,
pd               2400 include/rdma/ib_verbs.h 			     struct ib_pd *pd, struct ib_udata *udata);
pd               2402 include/rdma/ib_verbs.h 	struct ib_mr *(*alloc_mr)(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               2404 include/rdma/ib_verbs.h 	struct ib_mr *(*alloc_mr_integrity)(struct ib_pd *pd,
pd               2407 include/rdma/ib_verbs.h 	int (*advise_mr)(struct ib_pd *pd,
pd               2415 include/rdma/ib_verbs.h 	struct ib_mw *(*alloc_mw)(struct ib_pd *pd, enum ib_mw_type type,
pd               2418 include/rdma/ib_verbs.h 	struct ib_fmr *(*alloc_fmr)(struct ib_pd *pd, int mr_access_flags,
pd               2452 include/rdma/ib_verbs.h 	struct ib_wq *(*create_wq)(struct ib_pd *pd,
pd               2468 include/rdma/ib_verbs.h 	struct ib_mr *(*reg_dm_mr)(struct ib_pd *pd, struct ib_dm *dm,
pd               3353 include/rdma/ib_verbs.h void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata);
pd               3361 include/rdma/ib_verbs.h static inline void ib_dealloc_pd(struct ib_pd *pd)
pd               3363 include/rdma/ib_verbs.h 	ib_dealloc_pd_user(pd, NULL);
pd               3380 include/rdma/ib_verbs.h struct ib_ah *rdma_create_ah(struct ib_pd *pd, struct rdma_ah_attr *ah_attr,
pd               3395 include/rdma/ib_verbs.h struct ib_ah *rdma_create_user_ah(struct ib_pd *pd,
pd               3450 include/rdma/ib_verbs.h struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc,
pd               3509 include/rdma/ib_verbs.h struct ib_srq *ib_create_srq(struct ib_pd *pd,
pd               3581 include/rdma/ib_verbs.h struct ib_qp *ib_create_qp_user(struct ib_pd *pd,
pd               3596 include/rdma/ib_verbs.h static inline struct ib_qp *ib_create_qp(struct ib_pd *pd,
pd               3599 include/rdma/ib_verbs.h 	return ib_create_qp_user(pd, qp_init_attr, NULL);
pd               4135 include/rdma/ib_verbs.h struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type,
pd               4138 include/rdma/ib_verbs.h static inline struct ib_mr *ib_alloc_mr(struct ib_pd *pd,
pd               4141 include/rdma/ib_verbs.h 	return ib_alloc_mr_user(pd, mr_type, max_num_sg, NULL);
pd               4144 include/rdma/ib_verbs.h struct ib_mr *ib_alloc_mr_integrity(struct ib_pd *pd,
pd               4180 include/rdma/ib_verbs.h struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd,
pd               4323 include/rdma/ib_verbs.h struct ib_wq *ib_create_wq(struct ib_pd *pd,
pd                219 include/rdma/rdma_cm.h int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd,
pd                568 include/rdma/rdma_vt.h int rvt_lkey_ok(struct rvt_lkey_table *rkt, struct rvt_pd *pd,
pd                 74 include/rdma/rdmavt_mr.h 	struct ib_pd *pd;       /* shares refcnt of ibmr.pd */
pd                 14 include/sound/sh_dac_audio.h 	void (*start)(struct dac_audio_pdata *pd);
pd                 15 include/sound/sh_dac_audio.h 	void (*stop)(struct dac_audio_pdata *pd);
pd                130 include/xen/interface/io/fbif.h 	unsigned long pd[256];
pd                 38 kernel/padata.c static void padata_free_pd(struct parallel_data *pd);
pd                 40 kernel/padata.c static int padata_index_to_cpu(struct parallel_data *pd, int cpu_index)
pd                 44 kernel/padata.c 	target_cpu = cpumask_first(pd->cpumask.pcpu);
pd                 46 kernel/padata.c 		target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu);
pd                 51 kernel/padata.c static int padata_cpu_hash(struct parallel_data *pd, unsigned int seq_nr)
pd                 57 kernel/padata.c 	int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu);
pd                 59 kernel/padata.c 	return padata_index_to_cpu(pd, cpu_index);
pd                109 kernel/padata.c 	struct parallel_data *pd;
pd                113 kernel/padata.c 	pd = rcu_dereference_bh(ps->pd);
pd                119 kernel/padata.c 	if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) {
pd                120 kernel/padata.c 		if (!cpumask_weight(pd->cpumask.cbcpu))
pd                124 kernel/padata.c 		cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu);
pd                126 kernel/padata.c 		cpu = cpumask_first(pd->cpumask.cbcpu);
pd                128 kernel/padata.c 			cpu = cpumask_next(cpu, pd->cpumask.cbcpu);
pd                137 kernel/padata.c 	if (atomic_read(&pd->refcnt) >= MAX_OBJ_NUM)
pd                141 kernel/padata.c 	atomic_inc(&pd->refcnt);
pd                142 kernel/padata.c 	padata->pd = pd;
pd                145 kernel/padata.c 	padata->seq_nr = atomic_inc_return(&pd->seq_nr);
pd                146 kernel/padata.c 	target_cpu = padata_cpu_hash(pd, padata->seq_nr);
pd                148 kernel/padata.c 	queue = per_cpu_ptr(pd->pqueue, target_cpu);
pd                175 kernel/padata.c static struct padata_priv *padata_find_next(struct parallel_data *pd,
pd                181 kernel/padata.c 	int cpu = pd->cpu;
pd                183 kernel/padata.c 	next_queue = per_cpu_ptr(pd->pqueue, cpu);
pd                198 kernel/padata.c 	if (padata->seq_nr != pd->processed) {
pd                205 kernel/padata.c 		atomic_dec(&pd->reorder_objects);
pd                206 kernel/padata.c 		++pd->processed;
pd                207 kernel/padata.c 		pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false);
pd                214 kernel/padata.c static void padata_reorder(struct parallel_data *pd)
pd                216 kernel/padata.c 	struct padata_instance *pinst = pd->ps->pinst;
pd                232 kernel/padata.c 	if (!spin_trylock_bh(&pd->lock))
pd                236 kernel/padata.c 		padata = padata_find_next(pd, true);
pd                247 kernel/padata.c 		squeue = per_cpu_ptr(pd->squeue, cb_cpu);
pd                256 kernel/padata.c 	spin_unlock_bh(&pd->lock);
pd                268 kernel/padata.c 	next_queue = per_cpu_ptr(pd->pqueue, pd->cpu);
pd                270 kernel/padata.c 	    padata_find_next(pd, false))
pd                271 kernel/padata.c 		queue_work(pinst->serial_wq, &pd->reorder_work);
pd                276 kernel/padata.c 	struct parallel_data *pd;
pd                279 kernel/padata.c 	pd = container_of(work, struct parallel_data, reorder_work);
pd                280 kernel/padata.c 	padata_reorder(pd);
pd                287 kernel/padata.c 	struct parallel_data *pd;
pd                293 kernel/padata.c 	pd = squeue->pd;
pd                314 kernel/padata.c 	if (atomic_sub_and_test(cnt, &pd->refcnt))
pd                315 kernel/padata.c 		padata_free_pd(pd);
pd                328 kernel/padata.c 	struct parallel_data *pd = padata->pd;
pd                329 kernel/padata.c 	struct padata_parallel_queue *pqueue = per_cpu_ptr(pd->pqueue,
pd                339 kernel/padata.c 	atomic_inc(&pd->reorder_objects);
pd                349 kernel/padata.c 	padata_reorder(pd);
pd                370 kernel/padata.c static int pd_setup_cpumasks(struct parallel_data *pd,
pd                376 kernel/padata.c 	if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL))
pd                378 kernel/padata.c 	if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL))
pd                381 kernel/padata.c 	cpumask_copy(pd->cpumask.pcpu, pcpumask);
pd                382 kernel/padata.c 	cpumask_copy(pd->cpumask.cbcpu, cbcpumask);
pd                387 kernel/padata.c 	free_cpumask_var(pd->cpumask.pcpu);
pd                399 kernel/padata.c static void padata_init_squeues(struct parallel_data *pd)
pd                404 kernel/padata.c 	for_each_cpu(cpu, pd->cpumask.cbcpu) {
pd                405 kernel/padata.c 		squeue = per_cpu_ptr(pd->squeue, cpu);
pd                406 kernel/padata.c 		squeue->pd = pd;
pd                413 kernel/padata.c static void padata_init_pqueues(struct parallel_data *pd)
pd                418 kernel/padata.c 	for_each_cpu(cpu, pd->cpumask.pcpu) {
pd                419 kernel/padata.c 		pqueue = per_cpu_ptr(pd->pqueue, cpu);
pd                434 kernel/padata.c 	struct parallel_data *pd;
pd                439 kernel/padata.c 	pd = kzalloc(sizeof(struct parallel_data), GFP_KERNEL);
pd                440 kernel/padata.c 	if (!pd)
pd                443 kernel/padata.c 	pd->pqueue = alloc_percpu(struct padata_parallel_queue);
pd                444 kernel/padata.c 	if (!pd->pqueue)
pd                447 kernel/padata.c 	pd->squeue = alloc_percpu(struct padata_serial_queue);
pd                448 kernel/padata.c 	if (!pd->squeue)
pd                451 kernel/padata.c 	pd->ps = ps;
pd                452 kernel/padata.c 	if (pd_setup_cpumasks(pd, pcpumask, cbcpumask))
pd                455 kernel/padata.c 	padata_init_pqueues(pd);
pd                456 kernel/padata.c 	padata_init_squeues(pd);
pd                457 kernel/padata.c 	atomic_set(&pd->seq_nr, -1);
pd                458 kernel/padata.c 	atomic_set(&pd->reorder_objects, 0);
pd                459 kernel/padata.c 	atomic_set(&pd->refcnt, 1);
pd                460 kernel/padata.c 	spin_lock_init(&pd->lock);
pd                461 kernel/padata.c 	pd->cpu = cpumask_first(pd->cpumask.pcpu);
pd                462 kernel/padata.c 	INIT_WORK(&pd->reorder_work, invoke_padata_reorder);
pd                464 kernel/padata.c 	return pd;
pd                467 kernel/padata.c 	free_percpu(pd->squeue);
pd                469 kernel/padata.c 	free_percpu(pd->pqueue);
pd                471 kernel/padata.c 	kfree(pd);
pd                476 kernel/padata.c static void padata_free_pd(struct parallel_data *pd)
pd                478 kernel/padata.c 	free_cpumask_var(pd->cpumask.pcpu);
pd                479 kernel/padata.c 	free_cpumask_var(pd->cpumask.cbcpu);
pd                480 kernel/padata.c 	free_percpu(pd->pqueue);
pd                481 kernel/padata.c 	free_percpu(pd->squeue);
pd                482 kernel/padata.c 	kfree(pd);
pd                509 kernel/padata.c 	ps->opd = rcu_dereference_protected(ps->pd, 1);
pd                510 kernel/padata.c 	rcu_assign_pointer(ps->pd, pd_new);
pd               1085 kernel/padata.c 	struct parallel_data *pd;
pd               1095 kernel/padata.c 	pd = padata_alloc_pd(ps);
pd               1098 kernel/padata.c 	if (!pd)
pd               1102 kernel/padata.c 	RCU_INIT_POINTER(ps->pd, pd);
pd               1126 kernel/padata.c 	padata_free_pd(rcu_dereference_protected(ps->pd, 1));
pd                 30 kernel/power/energy_model.c static void em_debug_create_cs(struct em_cap_state *cs, struct dentry *pd)
pd                 38 kernel/power/energy_model.c 	d = debugfs_create_dir(name, pd);
pd                 52 kernel/power/energy_model.c static void em_debug_create_pd(struct em_perf_domain *pd, int cpu)
pd                 63 kernel/power/energy_model.c 	debugfs_create_file("cpus", 0444, d, pd->cpus, &em_debug_cpus_fops);
pd                 66 kernel/power/energy_model.c 	for (i = 0; i < pd->nr_cap_states; i++)
pd                 67 kernel/power/energy_model.c 		em_debug_create_cs(&pd->table[i], d);
pd                 79 kernel/power/energy_model.c static void em_debug_create_pd(struct em_perf_domain *pd, int cpu) {}
pd                 88 kernel/power/energy_model.c 	struct em_perf_domain *pd;
pd                 94 kernel/power/energy_model.c 	pd = kzalloc(sizeof(*pd) + cpumask_size(), GFP_KERNEL);
pd                 95 kernel/power/energy_model.c 	if (!pd)
pd                156 kernel/power/energy_model.c 	pd->table = table;
pd                157 kernel/power/energy_model.c 	pd->nr_cap_states = nr_states;
pd                158 kernel/power/energy_model.c 	cpumask_copy(to_cpumask(pd->cpus), span);
pd                160 kernel/power/energy_model.c 	em_debug_create_pd(pd, cpu);
pd                162 kernel/power/energy_model.c 	return pd;
pd                167 kernel/power/energy_model.c 	kfree(pd);
pd                203 kernel/power/energy_model.c 	struct em_perf_domain *pd;
pd                237 kernel/power/energy_model.c 	pd = em_create_pd(span, nr_states, cb);
pd                238 kernel/power/energy_model.c 	if (!pd) {
pd                249 kernel/power/energy_model.c 		smp_store_release(per_cpu_ptr(&em_data, cpu), pd);
pd               6277 kernel/sched/fair.c compute_energy(struct task_struct *p, int dst_cpu, struct perf_domain *pd)
pd               6279 kernel/sched/fair.c 	struct cpumask *pd_mask = perf_domain_span(pd);
pd               6318 kernel/sched/fair.c 	return em_pd_energy(pd->em_pd, max_util, sum_util);
pd               6367 kernel/sched/fair.c 	struct perf_domain *pd;
pd               6370 kernel/sched/fair.c 	pd = rcu_dereference(rd->pd);
pd               6371 kernel/sched/fair.c 	if (!pd || READ_ONCE(rd->overutilized))
pd               6388 kernel/sched/fair.c 	for (; pd; pd = pd->next) {
pd               6394 kernel/sched/fair.c 		base_energy_pd = compute_energy(p, -1, pd);
pd               6397 kernel/sched/fair.c 		for_each_cpu_and(cpu, perf_domain_span(pd), sched_domain_span(sd)) {
pd               6409 kernel/sched/fair.c 				prev_delta = compute_energy(p, prev_cpu, pd);
pd               6427 kernel/sched/fair.c 			cur_delta = compute_energy(p, max_spare_cap_cpu, pd);
pd               8548 kernel/sched/fair.c 		if (rcu_dereference(rd->pd) && !READ_ONCE(rd->overutilized))
pd                790 kernel/sched/sched.h 	struct perf_domain __rcu *pd;
pd               2460 kernel/sched/sched.h #define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus)))
pd               2471 kernel/sched/sched.h #define perf_domain_span(pd) NULL
pd                235 kernel/sched/topology.c static void free_pd(struct perf_domain *pd)
pd                239 kernel/sched/topology.c 	while (pd) {
pd                240 kernel/sched/topology.c 		tmp = pd->next;
pd                241 kernel/sched/topology.c 		kfree(pd);
pd                242 kernel/sched/topology.c 		pd = tmp;
pd                246 kernel/sched/topology.c static struct perf_domain *find_pd(struct perf_domain *pd, int cpu)
pd                248 kernel/sched/topology.c 	while (pd) {
pd                249 kernel/sched/topology.c 		if (cpumask_test_cpu(cpu, perf_domain_span(pd)))
pd                250 kernel/sched/topology.c 			return pd;
pd                251 kernel/sched/topology.c 		pd = pd->next;
pd                260 kernel/sched/topology.c 	struct perf_domain *pd;
pd                268 kernel/sched/topology.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                269 kernel/sched/topology.c 	if (!pd)
pd                271 kernel/sched/topology.c 	pd->em_pd = obj;
pd                273 kernel/sched/topology.c 	return pd;
pd                277 kernel/sched/topology.c 						struct perf_domain *pd)
pd                279 kernel/sched/topology.c 	if (!sched_debug() || !pd)
pd                284 kernel/sched/topology.c 	while (pd) {
pd                286 kernel/sched/topology.c 				cpumask_first(perf_domain_span(pd)),
pd                287 kernel/sched/topology.c 				cpumask_pr_args(perf_domain_span(pd)),
pd                288 kernel/sched/topology.c 				em_pd_nr_cap_states(pd->em_pd));
pd                289 kernel/sched/topology.c 		pd = pd->next;
pd                297 kernel/sched/topology.c 	struct perf_domain *pd;
pd                299 kernel/sched/topology.c 	pd = container_of(rp, struct perf_domain, rcu);
pd                300 kernel/sched/topology.c 	free_pd(pd);
pd                345 kernel/sched/topology.c 	struct perf_domain *pd = NULL, *tmp;
pd                365 kernel/sched/topology.c 		if (find_pd(pd, i))
pd                375 kernel/sched/topology.c 			if (rd->pd)
pd                385 kernel/sched/topology.c 		tmp->next = pd;
pd                386 kernel/sched/topology.c 		pd = tmp;
pd                393 kernel/sched/topology.c 		nr_cs += em_pd_nr_cap_states(pd->em_pd);
pd                403 kernel/sched/topology.c 	perf_domain_debug(cpu_map, pd);
pd                406 kernel/sched/topology.c 	tmp = rd->pd;
pd                407 kernel/sched/topology.c 	rcu_assign_pointer(rd->pd, pd);
pd                411 kernel/sched/topology.c 	return !!pd;
pd                414 kernel/sched/topology.c 	free_pd(pd);
pd                415 kernel/sched/topology.c 	tmp = rd->pd;
pd                416 kernel/sched/topology.c 	rcu_assign_pointer(rd->pd, NULL);
pd                423 kernel/sched/topology.c static void free_pd(struct perf_domain *pd) { }
pd                436 kernel/sched/topology.c 	free_pd(rd->pd);
pd               2299 kernel/sched/topology.c 			    cpu_rq(cpumask_first(doms_cur[j]))->rd->pd) {
pd                391 lib/bch.c      	unsigned int i, j, tmp, l, pd = 1, d = syn[0];
pd                411 lib/bch.c      			tmp = a_log(bch, d)+n-a_log(bch, pd);
pd                423 lib/bch.c      				pd = d;
pd                 80 net/9p/trans_rdma.c 	struct ib_pd *pd;
pd                370 net/9p/trans_rdma.c 	if (rdma->pd && !IS_ERR(rdma->pd))
pd                371 net/9p/trans_rdma.c 		ib_dealloc_pd(rdma->pd);
pd                399 net/9p/trans_rdma.c 	sge.lkey = rdma->pd->local_dma_lkey;
pd                491 net/9p/trans_rdma.c 	sge.lkey = rdma->pd->local_dma_lkey;
pd                695 net/9p/trans_rdma.c 	rdma->pd = ib_alloc_pd(rdma->cm_id->device, 0);
pd                696 net/9p/trans_rdma.c 	if (IS_ERR(rdma->pd))
pd                711 net/9p/trans_rdma.c 	err = rdma_create_qp(rdma->cm_id, rdma->pd, &qp_attr);
pd                 92 net/ipv4/fou.c 	__be16 *pd = data;
pd                 93 net/ipv4/fou.c 	size_t start = ntohs(pd[0]);
pd                 94 net/ipv4/fou.c 	size_t offset = ntohs(pd[1]);
pd                292 net/ipv4/fou.c 	__be16 *pd = data;
pd                293 net/ipv4/fou.c 	size_t start = ntohs(pd[0]);
pd                294 net/ipv4/fou.c 	size_t offset = ntohs(pd[1]);
pd               1036 net/ipv4/fou.c 			__be16 *pd = data;
pd               1042 net/ipv4/fou.c 			pd[0] = htons(csum_start);
pd               1043 net/ipv4/fou.c 			pd[1] = htons(csum_start + skb->csum_offset);
pd                 43 net/l2tp/l2tp_debugfs.c static void l2tp_dfs_next_tunnel(struct l2tp_dfs_seq_data *pd)
pd                 46 net/l2tp/l2tp_debugfs.c 	if (pd->tunnel)
pd                 47 net/l2tp/l2tp_debugfs.c 		l2tp_tunnel_dec_refcount(pd->tunnel);
pd                 49 net/l2tp/l2tp_debugfs.c 	pd->tunnel = l2tp_tunnel_get_nth(pd->net, pd->tunnel_idx);
pd                 50 net/l2tp/l2tp_debugfs.c 	pd->tunnel_idx++;
pd                 53 net/l2tp/l2tp_debugfs.c static void l2tp_dfs_next_session(struct l2tp_dfs_seq_data *pd)
pd                 56 net/l2tp/l2tp_debugfs.c 	if (pd->session)
pd                 57 net/l2tp/l2tp_debugfs.c 		l2tp_session_dec_refcount(pd->session);
pd                 59 net/l2tp/l2tp_debugfs.c 	pd->session = l2tp_session_get_nth(pd->tunnel, pd->session_idx);
pd                 60 net/l2tp/l2tp_debugfs.c 	pd->session_idx++;
pd                 62 net/l2tp/l2tp_debugfs.c 	if (pd->session == NULL) {
pd                 63 net/l2tp/l2tp_debugfs.c 		pd->session_idx = 0;
pd                 64 net/l2tp/l2tp_debugfs.c 		l2tp_dfs_next_tunnel(pd);
pd                 71 net/l2tp/l2tp_debugfs.c 	struct l2tp_dfs_seq_data *pd = SEQ_START_TOKEN;
pd                 78 net/l2tp/l2tp_debugfs.c 	pd = m->private;
pd                 80 net/l2tp/l2tp_debugfs.c 	if (pd->tunnel == NULL)
pd                 81 net/l2tp/l2tp_debugfs.c 		l2tp_dfs_next_tunnel(pd);
pd                 83 net/l2tp/l2tp_debugfs.c 		l2tp_dfs_next_session(pd);
pd                 86 net/l2tp/l2tp_debugfs.c 	if ((pd->tunnel == NULL) && (pd->session == NULL))
pd                 87 net/l2tp/l2tp_debugfs.c 		pd = NULL;
pd                 90 net/l2tp/l2tp_debugfs.c 	return pd;
pd                102 net/l2tp/l2tp_debugfs.c 	struct l2tp_dfs_seq_data *pd = v;
pd                104 net/l2tp/l2tp_debugfs.c 	if (!pd || pd == SEQ_START_TOKEN)
pd                110 net/l2tp/l2tp_debugfs.c 	if (pd->session) {
pd                111 net/l2tp/l2tp_debugfs.c 		l2tp_session_dec_refcount(pd->session);
pd                112 net/l2tp/l2tp_debugfs.c 		pd->session = NULL;
pd                114 net/l2tp/l2tp_debugfs.c 	if (pd->tunnel) {
pd                115 net/l2tp/l2tp_debugfs.c 		l2tp_tunnel_dec_refcount(pd->tunnel);
pd                116 net/l2tp/l2tp_debugfs.c 		pd->tunnel = NULL;
pd                233 net/l2tp/l2tp_debugfs.c 	struct l2tp_dfs_seq_data *pd = v;
pd                251 net/l2tp/l2tp_debugfs.c 	if (!pd->session)
pd                252 net/l2tp/l2tp_debugfs.c 		l2tp_dfs_seq_tunnel_show(m, pd->tunnel);
pd                254 net/l2tp/l2tp_debugfs.c 		l2tp_dfs_seq_session_show(m, pd->session);
pd                269 net/l2tp/l2tp_debugfs.c 	struct l2tp_dfs_seq_data *pd;
pd                273 net/l2tp/l2tp_debugfs.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                274 net/l2tp/l2tp_debugfs.c 	if (pd == NULL)
pd                280 net/l2tp/l2tp_debugfs.c 	pd->net = get_net_ns_by_pid(current->pid);
pd                281 net/l2tp/l2tp_debugfs.c 	if (IS_ERR(pd->net)) {
pd                282 net/l2tp/l2tp_debugfs.c 		rc = PTR_ERR(pd->net);
pd                291 net/l2tp/l2tp_debugfs.c 	seq->private = pd;
pd                297 net/l2tp/l2tp_debugfs.c 	put_net(pd->net);
pd                299 net/l2tp/l2tp_debugfs.c 	kfree(pd);
pd                305 net/l2tp/l2tp_debugfs.c 	struct l2tp_dfs_seq_data *pd;
pd                309 net/l2tp/l2tp_debugfs.c 	pd = seq->private;
pd                310 net/l2tp/l2tp_debugfs.c 	if (pd->net)
pd                311 net/l2tp/l2tp_debugfs.c 		put_net(pd->net);
pd                312 net/l2tp/l2tp_debugfs.c 	kfree(pd);
pd               1439 net/l2tp/l2tp_ppp.c static void pppol2tp_next_tunnel(struct net *net, struct pppol2tp_seq_data *pd)
pd               1442 net/l2tp/l2tp_ppp.c 	if (pd->tunnel)
pd               1443 net/l2tp/l2tp_ppp.c 		l2tp_tunnel_dec_refcount(pd->tunnel);
pd               1446 net/l2tp/l2tp_ppp.c 		pd->tunnel = l2tp_tunnel_get_nth(net, pd->tunnel_idx);
pd               1447 net/l2tp/l2tp_ppp.c 		pd->tunnel_idx++;
pd               1450 net/l2tp/l2tp_ppp.c 		if (!pd->tunnel || pd->tunnel->version == 2)
pd               1453 net/l2tp/l2tp_ppp.c 		l2tp_tunnel_dec_refcount(pd->tunnel);
pd               1457 net/l2tp/l2tp_ppp.c static void pppol2tp_next_session(struct net *net, struct pppol2tp_seq_data *pd)
pd               1460 net/l2tp/l2tp_ppp.c 	if (pd->session)
pd               1461 net/l2tp/l2tp_ppp.c 		l2tp_session_dec_refcount(pd->session);
pd               1463 net/l2tp/l2tp_ppp.c 	pd->session = l2tp_session_get_nth(pd->tunnel, pd->session_idx);
pd               1464 net/l2tp/l2tp_ppp.c 	pd->session_idx++;
pd               1466 net/l2tp/l2tp_ppp.c 	if (pd->session == NULL) {
pd               1467 net/l2tp/l2tp_ppp.c 		pd->session_idx = 0;
pd               1468 net/l2tp/l2tp_ppp.c 		pppol2tp_next_tunnel(net, pd);
pd               1474 net/l2tp/l2tp_ppp.c 	struct pppol2tp_seq_data *pd = SEQ_START_TOKEN;
pd               1482 net/l2tp/l2tp_ppp.c 	pd = m->private;
pd               1485 net/l2tp/l2tp_ppp.c 	if (pd->tunnel == NULL)
pd               1486 net/l2tp/l2tp_ppp.c 		pppol2tp_next_tunnel(net, pd);
pd               1488 net/l2tp/l2tp_ppp.c 		pppol2tp_next_session(net, pd);
pd               1491 net/l2tp/l2tp_ppp.c 	if ((pd->tunnel == NULL) && (pd->session == NULL))
pd               1492 net/l2tp/l2tp_ppp.c 		pd = NULL;
pd               1495 net/l2tp/l2tp_ppp.c 	return pd;
pd               1506 net/l2tp/l2tp_ppp.c 	struct pppol2tp_seq_data *pd = v;
pd               1508 net/l2tp/l2tp_ppp.c 	if (!pd || pd == SEQ_START_TOKEN)
pd               1514 net/l2tp/l2tp_ppp.c 	if (pd->session) {
pd               1515 net/l2tp/l2tp_ppp.c 		l2tp_session_dec_refcount(pd->session);
pd               1516 net/l2tp/l2tp_ppp.c 		pd->session = NULL;
pd               1518 net/l2tp/l2tp_ppp.c 	if (pd->tunnel) {
pd               1519 net/l2tp/l2tp_ppp.c 		l2tp_tunnel_dec_refcount(pd->tunnel);
pd               1520 net/l2tp/l2tp_ppp.c 		pd->tunnel = NULL;
pd               1600 net/l2tp/l2tp_ppp.c 	struct pppol2tp_seq_data *pd = v;
pd               1614 net/l2tp/l2tp_ppp.c 	if (!pd->session)
pd               1615 net/l2tp/l2tp_ppp.c 		pppol2tp_seq_tunnel_show(m, pd->tunnel);
pd               1617 net/l2tp/l2tp_ppp.c 		pppol2tp_seq_session_show(m, pd->session);
pd                654 net/netfilter/ipvs/ip_vs_conn.c 		struct ip_vs_proto_data *pd;
pd                680 net/netfilter/ipvs/ip_vs_conn.c 		pd = ip_vs_proto_data_get(cp->ipvs, cp->protocol);
pd                681 net/netfilter/ipvs/ip_vs_conn.c 		if (pd && atomic_read(&pd->appcnt))
pd                682 net/netfilter/ipvs/ip_vs_conn.c 			ip_vs_bind_app(cp, pd->pp);
pd                907 net/netfilter/ipvs/ip_vs_conn.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(p->ipvs,
pd                984 net/netfilter/ipvs/ip_vs_conn.c 	if (unlikely(pd && atomic_read(&pd->appcnt)))
pd                985 net/netfilter/ipvs/ip_vs_conn.c 		ip_vs_bind_app(cp, pd->pp);
pd                235 net/netfilter/ipvs/ip_vs_core.c 		struct ip_vs_proto_data *pd)
pd                237 net/netfilter/ipvs/ip_vs_core.c 	if (likely(pd->pp->state_transition))
pd                238 net/netfilter/ipvs/ip_vs_core.c 		pd->pp->state_transition(cp, direction, skb, pd);
pd                454 net/netfilter/ipvs/ip_vs_core.c 	       struct ip_vs_proto_data *pd, int *ignored,
pd                457 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_protocol *pp = pd->pp;
pd                598 net/netfilter/ipvs/ip_vs_core.c 		struct ip_vs_proto_data *pd, struct ip_vs_iphdr *iph)
pd                640 net/netfilter/ipvs/ip_vs_core.c 		ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pd);
pd                643 net/netfilter/ipvs/ip_vs_core.c 		ret = cp->packet_xmit(skb, cp, pd->pp, iph);
pd               1279 net/netfilter/ipvs/ip_vs_core.c handle_response(int af, struct sk_buff *skb, struct ip_vs_proto_data *pd,
pd               1283 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_protocol *pp = pd->pp;
pd               1326 net/netfilter/ipvs/ip_vs_core.c 	ip_vs_set_state(cp, IP_VS_DIR_OUTPUT, skb, pd);
pd               1352 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_proto_data *pd;
pd               1398 net/netfilter/ipvs/ip_vs_core.c 	pd = ip_vs_proto_data_get(ipvs, iph.protocol);
pd               1399 net/netfilter/ipvs/ip_vs_core.c 	if (unlikely(!pd))
pd               1401 net/netfilter/ipvs/ip_vs_core.c 	pp = pd->pp;
pd               1424 net/netfilter/ipvs/ip_vs_core.c 		return handle_response(af, skb, pd, cp, &iph, hooknum);
pd               1437 net/netfilter/ipvs/ip_vs_core.c 				return handle_response(af, skb, pd, cp, &iph,
pd               1546 net/netfilter/ipvs/ip_vs_core.c 		      struct ip_vs_proto_data *pd,
pd               1550 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_protocol *pp = pd->pp;
pd               1558 net/netfilter/ipvs/ip_vs_core.c 		if (!pp->conn_schedule(ipvs, af, skb, pd, verdict, cpp, iph))
pd               1662 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_proto_data *pd;
pd               1756 net/netfilter/ipvs/ip_vs_core.c 	pd = ip_vs_proto_data_get(ipvs, cih->protocol);
pd               1757 net/netfilter/ipvs/ip_vs_core.c 	if (!pd)
pd               1759 net/netfilter/ipvs/ip_vs_core.c 	pp = pd->pp;
pd               1785 net/netfilter/ipvs/ip_vs_core.c 		if (!ip_vs_try_to_schedule(ipvs, AF_INET, skb, pd, &v, &cp, &ciph))
pd               1879 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_proto_data *pd;
pd               1914 net/netfilter/ipvs/ip_vs_core.c 	pd = ip_vs_proto_data_get(ipvs, ciph.protocol);
pd               1915 net/netfilter/ipvs/ip_vs_core.c 	if (!pd)
pd               1917 net/netfilter/ipvs/ip_vs_core.c 	pp = pd->pp;
pd               1938 net/netfilter/ipvs/ip_vs_core.c 		if (!ip_vs_try_to_schedule(ipvs, AF_INET6, skb, pd, &v, &cp, &ciph))
pd               1982 net/netfilter/ipvs/ip_vs_core.c 	struct ip_vs_proto_data *pd;
pd               2044 net/netfilter/ipvs/ip_vs_core.c 	pd = ip_vs_proto_data_get(ipvs, iph.protocol);
pd               2045 net/netfilter/ipvs/ip_vs_core.c 	if (unlikely(!pd)) {
pd               2055 net/netfilter/ipvs/ip_vs_core.c 	pp = pd->pp;
pd               2096 net/netfilter/ipvs/ip_vs_core.c 		if (!ip_vs_try_to_schedule(ipvs, af, skb, pd, &v, &cp, &iph))
pd               2124 net/netfilter/ipvs/ip_vs_core.c 	ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pd);
pd               2297 net/netfilter/ipvs/ip_vs_ctl.c 	struct ip_vs_proto_data *pd;
pd               2319 net/netfilter/ipvs/ip_vs_ctl.c 		pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP);
pd               2320 net/netfilter/ipvs/ip_vs_ctl.c 		pd->timeout_table[IP_VS_TCP_S_ESTABLISHED]
pd               2325 net/netfilter/ipvs/ip_vs_ctl.c 		pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP);
pd               2326 net/netfilter/ipvs/ip_vs_ctl.c 		pd->timeout_table[IP_VS_TCP_S_FIN_WAIT]
pd               2333 net/netfilter/ipvs/ip_vs_ctl.c 		pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP);
pd               2334 net/netfilter/ipvs/ip_vs_ctl.c 		pd->timeout_table[IP_VS_UDP_S_NORMAL]
pd               2688 net/netfilter/ipvs/ip_vs_ctl.c 	struct ip_vs_proto_data *pd;
pd               2694 net/netfilter/ipvs/ip_vs_ctl.c 	pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP);
pd               2695 net/netfilter/ipvs/ip_vs_ctl.c 	u->tcp_timeout = pd->timeout_table[IP_VS_TCP_S_ESTABLISHED] / HZ;
pd               2696 net/netfilter/ipvs/ip_vs_ctl.c 	u->tcp_fin_timeout = pd->timeout_table[IP_VS_TCP_S_FIN_WAIT] / HZ;
pd               2699 net/netfilter/ipvs/ip_vs_ctl.c 	pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP);
pd               2701 net/netfilter/ipvs/ip_vs_ctl.c 			pd->timeout_table[IP_VS_UDP_S_NORMAL] / HZ;
pd                 69 net/netfilter/ipvs/ip_vs_proto.c 	struct ip_vs_proto_data *pd =
pd                 72 net/netfilter/ipvs/ip_vs_proto.c 	if (!pd)
pd                 75 net/netfilter/ipvs/ip_vs_proto.c 	pd->pp = pp;	/* For speed issues */
pd                 76 net/netfilter/ipvs/ip_vs_proto.c 	pd->next = ipvs->proto_data_table[hash];
pd                 77 net/netfilter/ipvs/ip_vs_proto.c 	ipvs->proto_data_table[hash] = pd;
pd                 78 net/netfilter/ipvs/ip_vs_proto.c 	atomic_set(&pd->appcnt, 0);	/* Init app counter */
pd                 81 net/netfilter/ipvs/ip_vs_proto.c 		int ret = pp->init_netns(ipvs, pd);
pd                 84 net/netfilter/ipvs/ip_vs_proto.c 			ipvs->proto_data_table[hash] = pd->next;
pd                 85 net/netfilter/ipvs/ip_vs_proto.c 			kfree(pd);
pd                118 net/netfilter/ipvs/ip_vs_proto.c unregister_ip_vs_proto_netns(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                121 net/netfilter/ipvs/ip_vs_proto.c 	unsigned int hash = IP_VS_PROTO_HASH(pd->pp->protocol);
pd                125 net/netfilter/ipvs/ip_vs_proto.c 		if (*pd_p == pd) {
pd                126 net/netfilter/ipvs/ip_vs_proto.c 			*pd_p = pd->next;
pd                127 net/netfilter/ipvs/ip_vs_proto.c 			if (pd->pp->exit_netns != NULL)
pd                128 net/netfilter/ipvs/ip_vs_proto.c 				pd->pp->exit_netns(ipvs, pd);
pd                129 net/netfilter/ipvs/ip_vs_proto.c 			kfree(pd);
pd                160 net/netfilter/ipvs/ip_vs_proto.c 	struct ip_vs_proto_data *pd;
pd                163 net/netfilter/ipvs/ip_vs_proto.c 	for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) {
pd                164 net/netfilter/ipvs/ip_vs_proto.c 		if (pd->pp->protocol == proto)
pd                165 net/netfilter/ipvs/ip_vs_proto.c 			return pd;
pd                177 net/netfilter/ipvs/ip_vs_proto.c 	struct ip_vs_proto_data *pd;
pd                181 net/netfilter/ipvs/ip_vs_proto.c 		for (pd = ipvs->proto_data_table[i]; pd; pd = pd->next) {
pd                182 net/netfilter/ipvs/ip_vs_proto.c 			if (pd->pp->timeout_change)
pd                183 net/netfilter/ipvs/ip_vs_proto.c 				pd->pp->timeout_change(pd, flags);
pd                331 net/netfilter/ipvs/ip_vs_proto.c 	struct ip_vs_proto_data *pd;
pd                336 net/netfilter/ipvs/ip_vs_proto.c 		while ((pd = ipvs->proto_data_table[i]) != NULL)
pd                337 net/netfilter/ipvs/ip_vs_proto.c 			unregister_ip_vs_proto_netns(ipvs, pd);
pd                104 net/netfilter/ipvs/ip_vs_proto_ah_esp.c 		     struct ip_vs_proto_data *pd,
pd                 17 net/netfilter/ipvs/ip_vs_proto_sctp.c 		   struct ip_vs_proto_data *pd,
pd                 70 net/netfilter/ipvs/ip_vs_proto_sctp.c 		*cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph);
pd                 73 net/netfilter/ipvs/ip_vs_proto_sctp.c 				*verdict = ip_vs_leave(svc, skb, pd, iph);
pd                378 net/netfilter/ipvs/ip_vs_proto_sctp.c set_sctp_state(struct ip_vs_proto_data *pd, struct ip_vs_conn *cp,
pd                441 net/netfilter/ipvs/ip_vs_proto_sctp.c 				pd->pp->name,
pd                467 net/netfilter/ipvs/ip_vs_proto_sctp.c 	if (likely(pd))
pd                468 net/netfilter/ipvs/ip_vs_proto_sctp.c 		cp->timeout = pd->timeout_table[cp->state = next_state];
pd                475 net/netfilter/ipvs/ip_vs_proto_sctp.c 		const struct sk_buff *skb, struct ip_vs_proto_data *pd)
pd                478 net/netfilter/ipvs/ip_vs_proto_sctp.c 	set_sctp_state(pd, cp, direction, skb);
pd                494 net/netfilter/ipvs/ip_vs_proto_sctp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_SCTP);
pd                505 net/netfilter/ipvs/ip_vs_proto_sctp.c 	atomic_inc(&pd->appcnt);
pd                513 net/netfilter/ipvs/ip_vs_proto_sctp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_SCTP);
pd                515 net/netfilter/ipvs/ip_vs_proto_sctp.c 	atomic_dec(&pd->appcnt);
pd                559 net/netfilter/ipvs/ip_vs_proto_sctp.c static int __ip_vs_sctp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                562 net/netfilter/ipvs/ip_vs_proto_sctp.c 	pd->timeout_table = ip_vs_create_timeout_table((int *)sctp_timeouts,
pd                564 net/netfilter/ipvs/ip_vs_proto_sctp.c 	if (!pd->timeout_table)
pd                569 net/netfilter/ipvs/ip_vs_proto_sctp.c static void __ip_vs_sctp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                571 net/netfilter/ipvs/ip_vs_proto_sctp.c 	kfree(pd->timeout_table);
pd                 36 net/netfilter/ipvs/ip_vs_proto_tcp.c 		  struct ip_vs_proto_data *pd,
pd                 90 net/netfilter/ipvs/ip_vs_proto_tcp.c 		*cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph);
pd                 93 net/netfilter/ipvs/ip_vs_proto_tcp.c 				*verdict = ip_vs_leave(svc, skb, pd, iph);
pd                485 net/netfilter/ipvs/ip_vs_proto_tcp.c static void tcp_timeout_change(struct ip_vs_proto_data *pd, int flags)
pd                495 net/netfilter/ipvs/ip_vs_proto_tcp.c 	pd->tcp_state_table = (on ? tcp_states_dos : tcp_states);
pd                512 net/netfilter/ipvs/ip_vs_proto_tcp.c set_tcp_state(struct ip_vs_proto_data *pd, struct ip_vs_conn *cp,
pd                536 net/netfilter/ipvs/ip_vs_proto_tcp.c 		pd->tcp_state_table[state_off+state_idx].next_state[cp->state];
pd                544 net/netfilter/ipvs/ip_vs_proto_tcp.c 			      pd->pp->name,
pd                576 net/netfilter/ipvs/ip_vs_proto_tcp.c 	if (likely(pd))
pd                577 net/netfilter/ipvs/ip_vs_proto_tcp.c 		cp->timeout = pd->timeout_table[cp->state = new_state];
pd                588 net/netfilter/ipvs/ip_vs_proto_tcp.c 		     struct ip_vs_proto_data *pd)
pd                603 net/netfilter/ipvs/ip_vs_proto_tcp.c 	set_tcp_state(pd, cp, direction, th);
pd                620 net/netfilter/ipvs/ip_vs_proto_tcp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP);
pd                631 net/netfilter/ipvs/ip_vs_proto_tcp.c 	atomic_inc(&pd->appcnt);
pd                641 net/netfilter/ipvs/ip_vs_proto_tcp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP);
pd                643 net/netfilter/ipvs/ip_vs_proto_tcp.c 	atomic_dec(&pd->appcnt);
pd                693 net/netfilter/ipvs/ip_vs_proto_tcp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(cp->ipvs, IPPROTO_TCP);
pd                697 net/netfilter/ipvs/ip_vs_proto_tcp.c 	cp->timeout = (pd ? pd->timeout_table[IP_VS_TCP_S_LISTEN]
pd                706 net/netfilter/ipvs/ip_vs_proto_tcp.c static int __ip_vs_tcp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                709 net/netfilter/ipvs/ip_vs_proto_tcp.c 	pd->timeout_table = ip_vs_create_timeout_table((int *)tcp_timeouts,
pd                711 net/netfilter/ipvs/ip_vs_proto_tcp.c 	if (!pd->timeout_table)
pd                713 net/netfilter/ipvs/ip_vs_proto_tcp.c 	pd->tcp_state_table = tcp_states;
pd                717 net/netfilter/ipvs/ip_vs_proto_tcp.c static void __ip_vs_tcp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                719 net/netfilter/ipvs/ip_vs_proto_tcp.c 	kfree(pd->timeout_table);
pd                 32 net/netfilter/ipvs/ip_vs_proto_udp.c 		  struct ip_vs_proto_data *pd,
pd                 78 net/netfilter/ipvs/ip_vs_proto_udp.c 		*cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph);
pd                 81 net/netfilter/ipvs/ip_vs_proto_udp.c 				*verdict = ip_vs_leave(svc, skb, pd, iph);
pd                367 net/netfilter/ipvs/ip_vs_proto_udp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP);
pd                378 net/netfilter/ipvs/ip_vs_proto_udp.c 	atomic_inc(&pd->appcnt);
pd                388 net/netfilter/ipvs/ip_vs_proto_udp.c 	struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP);
pd                390 net/netfilter/ipvs/ip_vs_proto_udp.c 	atomic_dec(&pd->appcnt);
pd                454 net/netfilter/ipvs/ip_vs_proto_udp.c 		     struct ip_vs_proto_data *pd)
pd                456 net/netfilter/ipvs/ip_vs_proto_udp.c 	if (unlikely(!pd)) {
pd                461 net/netfilter/ipvs/ip_vs_proto_udp.c 	cp->timeout = pd->timeout_table[IP_VS_UDP_S_NORMAL];
pd                466 net/netfilter/ipvs/ip_vs_proto_udp.c static int __udp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                469 net/netfilter/ipvs/ip_vs_proto_udp.c 	pd->timeout_table = ip_vs_create_timeout_table((int *)udp_timeouts,
pd                471 net/netfilter/ipvs/ip_vs_proto_udp.c 	if (!pd->timeout_table)
pd                476 net/netfilter/ipvs/ip_vs_proto_udp.c static void __udp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd)
pd                478 net/netfilter/ipvs/ip_vs_proto_udp.c 	kfree(pd->timeout_table);
pd                948 net/netfilter/ipvs/ip_vs_sync.c 		struct ip_vs_proto_data *pd;
pd                950 net/netfilter/ipvs/ip_vs_sync.c 		pd = ip_vs_proto_data_get(ipvs, protocol);
pd                951 net/netfilter/ipvs/ip_vs_sync.c 		if (!(flags & IP_VS_CONN_F_TEMPLATE) && pd && pd->timeout_table)
pd                952 net/netfilter/ipvs/ip_vs_sync.c 			cp->timeout = pd->timeout_table[state];
pd               1056 net/netfilter/ipvs/ip_vs_xmit.c 		__be16 *pd;
pd               1066 net/netfilter/ipvs/ip_vs_xmit.c 		pd = data;
pd               1067 net/netfilter/ipvs/ip_vs_xmit.c 		pd[0] = htons(csum_start);
pd               1068 net/netfilter/ipvs/ip_vs_xmit.c 		pd[1] = htons(csum_start + skb->csum_offset);
pd                108 net/rds/ib.c   	if (rds_ibdev->pd)
pd                109 net/rds/ib.c   		ib_dealloc_pd(rds_ibdev->pd);
pd                180 net/rds/ib.c   	rds_ibdev->pd = ib_alloc_pd(device, 0);
pd                181 net/rds/ib.c   	if (IS_ERR(rds_ibdev->pd)) {
pd                182 net/rds/ib.c   		rds_ibdev->pd = NULL;
pd                248 net/rds/ib.h   	struct ib_pd		*pd;
pd                484 net/rds/ib_cm.c 	ic->i_pd = rds_ibdev->pd;
pd                 70 net/rds/ib_fmr.c 	fmr->fmr = ib_alloc_fmr(rds_ibdev->pd,
pd                 78 net/rds/ib_frmr.c 	frmr->mr = ib_alloc_mr(rds_ibdev->pd, IB_MR_TYPE_MEM_REG,
pd                719 net/sctp/ulpevent.c 	struct sctp_pdapi_event *pd;
pd                728 net/sctp/ulpevent.c 	pd = skb_put(skb, sizeof(struct sctp_pdapi_event));
pd                736 net/sctp/ulpevent.c 	pd->pdapi_type = SCTP_PARTIAL_DELIVERY_EVENT;
pd                737 net/sctp/ulpevent.c 	pd->pdapi_flags = flags;
pd                738 net/sctp/ulpevent.c 	pd->pdapi_stream = sid;
pd                739 net/sctp/ulpevent.c 	pd->pdapi_seq = seq;
pd                747 net/sctp/ulpevent.c 	pd->pdapi_length = sizeof(struct sctp_pdapi_event);
pd                753 net/sctp/ulpevent.c 	pd->pdapi_indication = indication;
pd                760 net/sctp/ulpevent.c 	pd->pdapi_assoc_id = sctp_assoc2id(asoc);
pd                377 net/smc/smc_ib.c int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags,
pd                384 net/smc/smc_ib.c 		ib_alloc_mr(pd, IB_MR_TYPE_MEM_REG, 1 << buf_slot->order);
pd                 72 net/smc/smc_ib.h int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags,
pd                493 security/selinux/ss/conditional.c 	struct policy_data *pd = ptr;
pd                494 security/selinux/ss/conditional.c 	void *fp = pd->fp;
pd               2599 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2600 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               2627 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2628 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               2762 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2763 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               2857 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2858 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               2859 security/selinux/ss/policydb.c 	struct policydb *p = pd->p;
pd               2945 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2946 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               2947 security/selinux/ss/policydb.c 	struct policydb *p = pd->p;
pd               2984 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               2985 security/selinux/ss/policydb.c 	struct policydb *p = pd->p;
pd               2986 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               3025 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               3026 security/selinux/ss/policydb.c 	struct policydb *p = pd->p;
pd               3027 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               3270 security/selinux/ss/policydb.c 	struct policy_data *pd = ptr;
pd               3271 security/selinux/ss/policydb.c 	void *fp = pd->fp;
pd               3272 security/selinux/ss/policydb.c 	struct policydb *p = pd->p;
pd               3297 security/selinux/ss/policydb.c 	struct policy_data pd;
pd               3299 security/selinux/ss/policydb.c 	pd.p = p;
pd               3300 security/selinux/ss/policydb.c 	pd.fp = fp;
pd               3314 security/selinux/ss/policydb.c 	rc = hashtab_map(p->range_tr, range_write_helper, &pd);
pd               3456 security/selinux/ss/policydb.c 		struct policy_data pd;
pd               3458 security/selinux/ss/policydb.c 		pd.fp = fp;
pd               3459 security/selinux/ss/policydb.c 		pd.p = p;
pd               3467 security/selinux/ss/policydb.c 		rc = hashtab_map(p->symtab[i].table, write_f[i], &pd);
pd                161 sound/isa/wavefront/wavefront_fx.c 	unsigned short *pd;
pd                187 sound/isa/wavefront/wavefront_fx.c 			pd = (unsigned short *) &r.data[3];
pd                199 sound/isa/wavefront/wavefront_fx.c 			pd = page_data;
pd                206 sound/isa/wavefront/wavefront_fx.c 			     pd);
pd                870 sound/pci/asihpi/asihpi.c 			char *pd = &s->runtime->dma_area[buf_ofs];
pd                885 sound/pci/asihpi/asihpi.c 						ds->h_stream, pd, xfer1,
pd                889 sound/pci/asihpi/asihpi.c 					pd = s->runtime->dma_area;
pd                896 sound/pci/asihpi/asihpi.c 							ds->h_stream, pd,
pd                906 sound/pci/asihpi/asihpi.c 						pd, xfer1));
pd                908 sound/pci/asihpi/asihpi.c 					pd = s->runtime->dma_area;
pd                914 sound/pci/asihpi/asihpi.c 							pd, xfer2));
pd                 51 sound/sh/aica.c static struct platform_device *pd;
pd                536 sound/sh/aica.c 	err = request_firmware(&fw_entry, "aica_firmware.bin", &pd->dev);
pd                630 sound/sh/aica.c 	pd = platform_device_register_simple(SND_AICA_DRIVER, -1,
pd                632 sound/sh/aica.c 	if (IS_ERR(pd)) {
pd                634 sound/sh/aica.c 		return PTR_ERR(pd);
pd                642 sound/sh/aica.c 	platform_device_unregister(pd);
pd                400 sound/soc/codecs/cros_ec_codec.c static int cros_ec_codec_platform_probe(struct platform_device *pd)
pd                402 sound/soc/codecs/cros_ec_codec.c 	struct device *dev = &pd->dev;
pd                403 sound/soc/codecs/cros_ec_codec.c 	struct cros_ec_device *ec_device = dev_get_drvdata(pd->dev.parent);
pd                414 sound/soc/codecs/cros_ec_codec.c 	platform_set_drvdata(pd, codec_data);
pd                 28 sound/soc/codecs/rl6231.c 	int pd, val;
pd                 39 sound/soc/codecs/rl6231.c 		pd = val + 1;
pd                 42 sound/soc/codecs/rl6231.c 		pd = 6;
pd                 45 sound/soc/codecs/rl6231.c 		pd = 8;
pd                 48 sound/soc/codecs/rl6231.c 		pd = 12;
pd                 51 sound/soc/codecs/rl6231.c 		pd = 16;
pd                 54 sound/soc/codecs/rl6231.c 		pd = -EINVAL;
pd                 58 sound/soc/codecs/rl6231.c 	return pd;
pd                225 sound/soc/codecs/rl6231.c 	static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16};
pd                231 sound/soc/codecs/rl6231.c 	for (i = 0; i < ARRAY_SIZE(pd); i++)
pd                232 sound/soc/codecs/rl6231.c 		if (sclk == rate * pd[i])
pd               1523 sound/soc/codecs/rt1011.c 	static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16};
pd               1529 sound/soc/codecs/rt1011.c 	for (i = 0; i < ARRAY_SIZE(pd); i++)
pd               1530 sound/soc/codecs/rt1011.c 		if (sclk == rate * pd[i])
pd                612 sound/soc/codecs/rt1305.c 	static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16};
pd                618 sound/soc/codecs/rt1305.c 	for (i = 0; i < ARRAY_SIZE(pd); i++)
pd                619 sound/soc/codecs/rt1305.c 		if (sclk == rate * pd[i])
pd                440 sound/soc/codecs/rt1308.c 	static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16};
pd                446 sound/soc/codecs/rt1308.c 	for (i = 0; i < ARRAY_SIZE(pd); i++)
pd                447 sound/soc/codecs/rt1308.c 		if (sclk == rate * pd[i])
pd               1607 sound/soc/codecs/rt5659.c 	int pd, idx = -EINVAL;
pd               1609 sound/soc/codecs/rt5659.c 	pd = rl6231_get_pre_div(rt5659->regmap,
pd               1611 sound/soc/codecs/rt5659.c 	idx = rl6231_calc_dmic_clk(rt5659->sysclk / pd);
pd               1481 sound/soc/codecs/rt5665.c 	int pd, idx;
pd               1483 sound/soc/codecs/rt5665.c 	pd = rl6231_get_pre_div(rt5665->regmap,
pd               1485 sound/soc/codecs/rt5665.c 	idx = rl6231_calc_dmic_clk(rt5665->sysclk / pd);
pd                 39 sound/soc/codecs/uda134x.c 	struct uda134x_platform_data *pd;
pd                 67 sound/soc/codecs/uda134x.c 	struct uda134x_platform_data *pd = context;
pd                 92 sound/soc/codecs/uda134x.c 		ret = l3_write(&pd->l3,
pd                102 sound/soc/codecs/uda134x.c 	ret = l3_write(&pd->l3,
pd                298 sound/soc/codecs/uda134x.c 	struct uda134x_platform_data *pd = uda134x->pd;
pd                306 sound/soc/codecs/uda134x.c 		if (pd->power) {
pd                307 sound/soc/codecs/uda134x.c 			pd->power(1);
pd                315 sound/soc/codecs/uda134x.c 		if (pd->power) {
pd                316 sound/soc/codecs/uda134x.c 			pd->power(0);
pd                450 sound/soc/codecs/uda134x.c 	struct uda134x_platform_data *pd = uda134x->pd;
pd                457 sound/soc/codecs/uda134x.c 	switch (pd->model) {
pd                466 sound/soc/codecs/uda134x.c 			pd->model);
pd                470 sound/soc/codecs/uda134x.c 	if (pd->power)
pd                471 sound/soc/codecs/uda134x.c 		pd->power(1);
pd                475 sound/soc/codecs/uda134x.c 	if (pd->model == UDA134X_UDA1341) {
pd                490 sound/soc/codecs/uda134x.c 	switch (pd->model) {
pd                506 sound/soc/codecs/uda134x.c 			__func__, pd->model);
pd                545 sound/soc/codecs/uda134x.c 	struct uda134x_platform_data *pd = pdev->dev.platform_data;
pd                549 sound/soc/codecs/uda134x.c 	if (!pd) {
pd                558 sound/soc/codecs/uda134x.c 	uda134x->pd = pd;
pd                561 sound/soc/codecs/uda134x.c 	if (pd->l3.use_gpios) {
pd                562 sound/soc/codecs/uda134x.c 		ret = l3_set_gpio_ops(&pdev->dev, &uda134x->pd->l3);
pd                567 sound/soc/codecs/uda134x.c 	uda134x->regmap = devm_regmap_init(&pdev->dev, NULL, pd,
pd                557 sound/soc/dwc/dwc-i2s.c 	dev->play_dma_data.pd.data = pdata->play_dma_data;
pd                558 sound/soc/dwc/dwc-i2s.c 	dev->capture_dma_data.pd.data = pdata->capture_dma_data;
pd                559 sound/soc/dwc/dwc-i2s.c 	dev->play_dma_data.pd.addr = res->start + I2S_TXDMA;
pd                560 sound/soc/dwc/dwc-i2s.c 	dev->capture_dma_data.pd.addr = res->start + I2S_RXDMA;
pd                561 sound/soc/dwc/dwc-i2s.c 	dev->play_dma_data.pd.max_burst = 16;
pd                562 sound/soc/dwc/dwc-i2s.c 	dev->capture_dma_data.pd.max_burst = 16;
pd                563 sound/soc/dwc/dwc-i2s.c 	dev->play_dma_data.pd.addr_width = bus_widths[idx];
pd                564 sound/soc/dwc/dwc-i2s.c 	dev->capture_dma_data.pd.addr_width = bus_widths[idx];
pd                565 sound/soc/dwc/dwc-i2s.c 	dev->play_dma_data.pd.filter = pdata->filter;
pd                566 sound/soc/dwc/dwc-i2s.c 	dev->capture_dma_data.pd.filter = pdata->filter;
pd                 85 sound/soc/dwc/local.h 	struct i2s_dma_data pd;
pd                198 sound/soc/samsung/s3c24xx_simtec.c static int simtec_call_startup(struct s3c24xx_audio_simtec_pdata *pd)
pd                203 sound/soc/samsung/s3c24xx_simtec.c 	if (pd->startup)
pd                204 sound/soc/samsung/s3c24xx_simtec.c 		pd->startup();
pd                222 sound/soc/samsung/s3c24xx_simtec.c 			   struct s3c24xx_audio_simtec_pdata *pd)
pd                228 sound/soc/samsung/s3c24xx_simtec.c 		ret = gpio_request(pd->amp_gain[0], "gpio-amp-gain0");
pd                234 sound/soc/samsung/s3c24xx_simtec.c 		ret = gpio_request(pd->amp_gain[1], "gpio-amp-gain1");
pd                241 sound/soc/samsung/s3c24xx_simtec.c 		gpio_direction_output(pd->amp_gain[0], 0);
pd                242 sound/soc/samsung/s3c24xx_simtec.c 		gpio_direction_output(pd->amp_gain[1], 0);
pd                247 sound/soc/samsung/s3c24xx_simtec.c 		ret = gpio_request(pd->amp_gpio, "gpio-amp");
pd                250 sound/soc/samsung/s3c24xx_simtec.c 				pd->amp_gpio, ret);
pd                261 sound/soc/samsung/s3c24xx_simtec.c 	if (pd->amp_gain[0] > 0) {
pd                262 sound/soc/samsung/s3c24xx_simtec.c 		gpio_free(pd->amp_gain[0]);
pd                263 sound/soc/samsung/s3c24xx_simtec.c 		gpio_free(pd->amp_gain[1]);
pd                269 sound/soc/samsung/s3c24xx_simtec.c static void detach_gpio_amp(struct s3c24xx_audio_simtec_pdata *pd)
pd                271 sound/soc/samsung/s3c24xx_simtec.c 	if (pd->amp_gain[0] > 0) {
pd                272 sound/soc/samsung/s3c24xx_simtec.c 		gpio_free(pd->amp_gain[0]);
pd                273 sound/soc/samsung/s3c24xx_simtec.c 		gpio_free(pd->amp_gain[1]);
pd                276 sound/soc/samsung/s3c24xx_simtec.c 	if (pd->amp_gpio > 0)
pd                277 sound/soc/samsung/s3c24xx_simtec.c 		gpio_free(pd->amp_gpio);
pd                 92 sound/soc/samsung/s3c24xx_simtec_hermes.c static int simtec_audio_hermes_probe(struct platform_device *pd)
pd                 94 sound/soc/samsung/s3c24xx_simtec_hermes.c 	dev_info(&pd->dev, "probing....\n");
pd                 95 sound/soc/samsung/s3c24xx_simtec_hermes.c 	return simtec_audio_core_probe(pd, &snd_soc_machine_simtec_aic33);
pd                 81 sound/soc/samsung/s3c24xx_simtec_tlv320aic23.c static int simtec_audio_tlv320aic23_probe(struct platform_device *pd)
pd                 83 sound/soc/samsung/s3c24xx_simtec_tlv320aic23.c 	return simtec_audio_core_probe(pd, &snd_soc_machine_simtec_aic23);
pd                746 sound/soc/sof/pcm.c 	struct snd_soc_component_driver *pd = &sdev->plat_drv;
pd                752 sound/soc/sof/pcm.c 	pd->name = "sof-audio-component";
pd                753 sound/soc/sof/pcm.c 	pd->probe = sof_pcm_probe;
pd                754 sound/soc/sof/pcm.c 	pd->remove = sof_pcm_remove;
pd                755 sound/soc/sof/pcm.c 	pd->ops	= &sof_pcm_ops;
pd                757 sound/soc/sof/pcm.c 	pd->compr_ops = &sof_compressed_ops;
pd                759 sound/soc/sof/pcm.c 	pd->pcm_new = sof_pcm_new;
pd                760 sound/soc/sof/pcm.c 	pd->ignore_machine = drv_name;
pd                761 sound/soc/sof/pcm.c 	pd->be_hw_params_fixup = sof_pcm_dai_link_fixup;
pd                762 sound/soc/sof/pcm.c 	pd->be_pcm_base = SOF_BE_PCM_BASE;
pd                763 sound/soc/sof/pcm.c 	pd->use_dai_pcm_id = true;
pd                764 sound/soc/sof/pcm.c 	pd->topology_name_prefix = "sof";
pd                767 sound/soc/sof/pcm.c 	pd->module_get_upon_open = 1;
pd                 20 sound/usb/power.c 	struct snd_usb_power_domain *pd;
pd                 23 sound/usb/power.c 	pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd                 24 sound/usb/power.c 	if (!pd)
pd                 38 sound/usb/power.c 				pd->pd_id = pd_desc->bPowerDomainID;
pd                 39 sound/usb/power.c 				pd->pd_d1d0_rec =
pd                 41 sound/usb/power.c 				pd->pd_d2d0_rec =
pd                 43 sound/usb/power.c 				return pd;
pd                 48 sound/usb/power.c 	kfree(pd);
pd                 53 sound/usb/power.c 			     struct snd_usb_power_domain *pd,
pd                 60 sound/usb/power.c 	idx = snd_usb_ctrl_intf(chip) | (pd->pd_id << 8);
pd                 69 sound/usb/power.c 			pd->pd_id);
pd                 75 sound/usb/power.c 			pd->pd_id, state);
pd                 85 sound/usb/power.c 			state, pd->pd_id);
pd                 92 sound/usb/power.c 			udelay(pd->pd_d2d0_rec * 50);
pd                 95 sound/usb/power.c 			udelay(pd->pd_d1d0_rec * 50);
pd                103 sound/usb/power.c 		pd->pd_id, state);
pd                 18 sound/usb/power.h 			     struct snd_usb_power_domain *pd,
pd                 83 sound/usb/stream.c 				   struct snd_usb_power_domain *pd)
pd                108 sound/usb/stream.c 	if (pd) {
pd                109 sound/usb/stream.c 		subs->str_pd = pd;
pd                111 sound/usb/stream.c 		snd_usb_power_domain_set(subs->stream->chip, pd,
pd                479 sound/usb/stream.c 				      struct snd_usb_power_domain *pd)
pd                508 sound/usb/stream.c 		snd_usb_init_substream(as, stream, fp, pd);
pd                536 sound/usb/stream.c 	snd_usb_init_substream(as, stream, fp, pd);
pd                564 sound/usb/stream.c 				       struct snd_usb_power_domain *pd)
pd                566 sound/usb/stream.c 	return __snd_usb_add_audio_stream(chip, stream, fp, pd);
pd                867 sound/usb/stream.c 	struct snd_usb_power_domain *pd;
pd               1049 sound/usb/stream.c 		pd = kzalloc(sizeof(*pd), GFP_KERNEL);
pd               1050 sound/usb/stream.c 		if (!pd) {
pd               1054 sound/usb/stream.c 		pd->pd_id = (stream == SNDRV_PCM_STREAM_PLAYBACK) ?
pd               1056 sound/usb/stream.c 		pd->pd_d1d0_rec = UAC3_BADD_PD_RECOVER_D1D0;
pd               1057 sound/usb/stream.c 		pd->pd_d2d0_rec = UAC3_BADD_PD_RECOVER_D2D0;
pd               1064 sound/usb/stream.c 		pd = snd_usb_find_power_domain(chip->ctrl_intf,
pd               1069 sound/usb/stream.c 			kfree(pd);
pd               1075 sound/usb/stream.c 	if (pd)
pd               1076 sound/usb/stream.c 		*pd_out = pd;
pd               1091 sound/usb/stream.c 	struct snd_usb_power_domain *pd = NULL;
pd               1174 sound/usb/stream.c 			fp = snd_usb_get_audioformat_uac3(chip, alts, &pd,
pd               1188 sound/usb/stream.c 			kfree(pd);
pd               1190 sound/usb/stream.c 			pd = NULL;
pd               1196 sound/usb/stream.c 			err = snd_usb_add_audio_stream_v3(chip, stream, fp, pd);
pd               1202 sound/usb/stream.c 			kfree(pd);
pd                327 tools/perf/pmu-events/jevents.c 	struct perf_entry_data *pd = data;
pd                328 tools/perf/pmu-events/jevents.c 	FILE *outfp = pd->outfp;
pd                329 tools/perf/pmu-events/jevents.c 	char *topic = pd->topic;