pd 1447 arch/alpha/kernel/setup.c struct platform_device *pd; pd 1450 arch/alpha/kernel/setup.c pd = platform_device_alloc("pcspkr", -1); pd 1451 arch/alpha/kernel/setup.c if (!pd) pd 1454 arch/alpha/kernel/setup.c ret = platform_device_add(pd); pd 1456 arch/alpha/kernel/setup.c platform_device_put(pd); pd 805 arch/arm/common/sa1111.c struct sa1111_platform_data *pd = me->platform_data; pd 811 arch/arm/common/sa1111.c if (!pd) pd 831 arch/arm/common/sa1111.c sachip->pdata = pd; pd 867 arch/arm/common/sa1111.c ret = sa1111_setup_irq(sachip, pd->irq_base); pd 909 arch/arm/common/sa1111.c if (pd) pd 910 arch/arm/common/sa1111.c has_devs &= ~pd->disable_devs; pd 91 arch/arm/mach-imx/mach-mx35_3ds.c struct plat_lcd_data *pd, unsigned int power) pd 1070 arch/arm/mach-pxa/devices.c struct platform_device *pd; pd 1072 arch/arm/mach-pxa/devices.c pd = platform_device_alloc("pxa2xx-spi", id); pd 1073 arch/arm/mach-pxa/devices.c if (pd == NULL) { pd 1079 arch/arm/mach-pxa/devices.c pd->dev.platform_data = info; pd 1080 arch/arm/mach-pxa/devices.c platform_device_add(pd); pd 532 arch/arm/mach-pxa/hx4700.c static void hx4700_lcd_set_power(struct plat_lcd_data *pd, unsigned int power) pd 38 arch/arm/mach-rockchip/platsmp.c static int pmu_power_domain_is_on(int pd) pd 47 arch/arm/mach-rockchip/platsmp.c return !(val & BIT(pd)); pd 64 arch/arm/mach-rockchip/platsmp.c static int pmu_set_power_domain(int pd, bool on) pd 66 arch/arm/mach-rockchip/platsmp.c u32 val = (on) ? 0 : BIT(pd); pd 67 arch/arm/mach-rockchip/platsmp.c struct reset_control *rstc = rockchip_get_core_reset(pd); pd 72 arch/arm/mach-rockchip/platsmp.c __func__, pd); pd 85 arch/arm/mach-rockchip/platsmp.c ret = regmap_update_bits(pmu, PMU_PWRDN_CON, BIT(pd), val); pd 94 arch/arm/mach-rockchip/platsmp.c ret = pmu_power_domain_is_on(pd); pd 534 arch/arm/mach-s3c24xx/mach-h1940.c static void h1940_lcd_power_set(struct plat_lcd_data *pd, pd 103 arch/arm/mach-s3c64xx/mach-anw6410.c static void anw6410_lcd_power_set(struct plat_lcd_data *pd, pd 129 arch/arm/mach-s3c64xx/mach-crag6410.c static void crag6410_lcd_power_set(struct plat_lcd_data *pd, unsigned int power) pd 197 arch/arm/mach-s3c64xx/mach-mini6410.c static void mini6410_lcd_power_set(struct plat_lcd_data *pd, pd 233 arch/arm/mach-s3c64xx/mach-smartq.c static void smartq_lcd_power_set(struct plat_lcd_data *pd, unsigned int power) pd 118 arch/arm/mach-s3c64xx/mach-smdk6410.c static void smdk6410_lcd_power_set(struct plat_lcd_data *pd, pd 38 arch/arm/mach-s3c64xx/pm.c struct generic_pm_domain pd; pd 43 arch/arm/mach-s3c64xx/pm.c struct s3c64xx_pm_domain *pd; pd 46 arch/arm/mach-s3c64xx/pm.c pd = container_of(domain, struct s3c64xx_pm_domain, pd); pd 49 arch/arm/mach-s3c64xx/pm.c val &= ~(pd->ena); pd 57 arch/arm/mach-s3c64xx/pm.c struct s3c64xx_pm_domain *pd; pd 61 arch/arm/mach-s3c64xx/pm.c pd = container_of(domain, struct s3c64xx_pm_domain, pd); pd 64 arch/arm/mach-s3c64xx/pm.c val |= pd->ena; pd 68 arch/arm/mach-s3c64xx/pm.c if (pd->pwr_stat) { pd 71 arch/arm/mach-s3c64xx/pm.c if (__raw_readl(S3C64XX_BLK_PWR_STAT) & pd->pwr_stat) pd 76 arch/arm/mach-s3c64xx/pm.c pr_err("Failed to start domain %s\n", pd->name); pd 87 arch/arm/mach-s3c64xx/pm.c .pd = { pd 97 arch/arm/mach-s3c64xx/pm.c .pd = { pd 107 arch/arm/mach-s3c64xx/pm.c .pd = { pd 117 arch/arm/mach-s3c64xx/pm.c .pd = { pd 127 arch/arm/mach-s3c64xx/pm.c .pd = { pd 137 arch/arm/mach-s3c64xx/pm.c .pd = { pd 146 arch/arm/mach-s3c64xx/pm.c .pd = { pd 156 arch/arm/mach-s3c64xx/pm.c .pd = { pd 315 arch/arm/mach-s3c64xx/pm.c pm_genpd_init(&s3c64xx_always_on_pm_domains[i]->pd, pd 319 arch/arm/mach-s3c64xx/pm.c pm_genpd_init(&s3c64xx_pm_domains[i]->pd, NULL, false); pd 323 arch/arm/mach-s3c64xx/pm.c pm_genpd_add_device(&s3c64xx_pm_f.pd, &s3c_device_fb.dev); pd 160 arch/arm/plat-samsung/devs.c void __init s3c_fb_set_platdata(struct s3c_fb_platdata *pd) pd 162 arch/arm/plat-samsung/devs.c s3c_set_platdata(pd, sizeof(struct s3c_fb_platdata), pd 176 arch/arm/plat-samsung/devs.c void __init s3c_hwmon_set_platdata(struct s3c_hwmon_pdata *pd) pd 178 arch/arm/plat-samsung/devs.c s3c_set_platdata(pd, sizeof(struct s3c_hwmon_pdata), pd 209 arch/arm/plat-samsung/devs.c void s3c_sdhci0_set_platdata(struct s3c_sdhci_platdata *pd) pd 211 arch/arm/plat-samsung/devs.c s3c_sdhci_set_platdata(pd, &s3c_hsmmc0_def_platdata); pd 239 arch/arm/plat-samsung/devs.c void s3c_sdhci1_set_platdata(struct s3c_sdhci_platdata *pd) pd 241 arch/arm/plat-samsung/devs.c s3c_sdhci_set_platdata(pd, &s3c_hsmmc1_def_platdata); pd 271 arch/arm/plat-samsung/devs.c void s3c_sdhci2_set_platdata(struct s3c_sdhci_platdata *pd) pd 273 arch/arm/plat-samsung/devs.c s3c_sdhci_set_platdata(pd, &s3c_hsmmc2_def_platdata); pd 301 arch/arm/plat-samsung/devs.c void s3c_sdhci3_set_platdata(struct s3c_sdhci_platdata *pd) pd 303 arch/arm/plat-samsung/devs.c s3c_sdhci_set_platdata(pd, &s3c_hsmmc3_def_platdata); pd 328 arch/arm/plat-samsung/devs.c void __init s3c_i2c0_set_platdata(struct s3c2410_platform_i2c *pd) pd 332 arch/arm/plat-samsung/devs.c if (!pd) { pd 333 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 334 arch/arm/plat-samsung/devs.c pd->bus_num = 0; pd 337 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c0); pd 356 arch/arm/plat-samsung/devs.c void __init s3c_i2c1_set_platdata(struct s3c2410_platform_i2c *pd) pd 360 arch/arm/plat-samsung/devs.c if (!pd) { pd 361 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 362 arch/arm/plat-samsung/devs.c pd->bus_num = 1; pd 365 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c1); pd 385 arch/arm/plat-samsung/devs.c void __init s3c_i2c2_set_platdata(struct s3c2410_platform_i2c *pd) pd 389 arch/arm/plat-samsung/devs.c if (!pd) { pd 390 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 391 arch/arm/plat-samsung/devs.c pd->bus_num = 2; pd 394 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c2); pd 414 arch/arm/plat-samsung/devs.c void __init s3c_i2c3_set_platdata(struct s3c2410_platform_i2c *pd) pd 418 arch/arm/plat-samsung/devs.c if (!pd) { pd 419 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 420 arch/arm/plat-samsung/devs.c pd->bus_num = 3; pd 423 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c3); pd 443 arch/arm/plat-samsung/devs.c void __init s3c_i2c4_set_platdata(struct s3c2410_platform_i2c *pd) pd 447 arch/arm/plat-samsung/devs.c if (!pd) { pd 448 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 449 arch/arm/plat-samsung/devs.c pd->bus_num = 4; pd 452 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c4); pd 472 arch/arm/plat-samsung/devs.c void __init s3c_i2c5_set_platdata(struct s3c2410_platform_i2c *pd) pd 476 arch/arm/plat-samsung/devs.c if (!pd) { pd 477 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 478 arch/arm/plat-samsung/devs.c pd->bus_num = 5; pd 481 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c5); pd 501 arch/arm/plat-samsung/devs.c void __init s3c_i2c6_set_platdata(struct s3c2410_platform_i2c *pd) pd 505 arch/arm/plat-samsung/devs.c if (!pd) { pd 506 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 507 arch/arm/plat-samsung/devs.c pd->bus_num = 6; pd 510 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c6); pd 530 arch/arm/plat-samsung/devs.c void __init s3c_i2c7_set_platdata(struct s3c2410_platform_i2c *pd) pd 534 arch/arm/plat-samsung/devs.c if (!pd) { pd 535 arch/arm/plat-samsung/devs.c pd = &default_i2c_data; pd 536 arch/arm/plat-samsung/devs.c pd->bus_num = 7; pd 539 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_i2c7); pd 601 arch/arm/plat-samsung/devs.c void __init samsung_keypad_set_platdata(struct samsung_keypad_platdata *pd) pd 605 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &samsung_device_keypad); pd 631 arch/arm/plat-samsung/devs.c void __init s3c24xx_fb_set_platdata(struct s3c2410fb_mach_info *pd) pd 635 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_lcd); pd 637 arch/arm/plat-samsung/devs.c npd->displays = kmemdup(pd->displays, pd 795 arch/arm/plat-samsung/devs.c void __init samsung_pwm_set_platdata(struct samsung_pwm_variant *pd) pd 797 arch/arm/plat-samsung/devs.c samsung_device_pwm.dev.platform_data = pd; pd 921 arch/arm/plat-samsung/devs.c void __init s3c64xx_ts_set_platdata(struct s3c2410_ts_mach_info *pd) pd 923 arch/arm/plat-samsung/devs.c if (!pd) pd 924 arch/arm/plat-samsung/devs.c pd = &default_ts_data; pd 926 arch/arm/plat-samsung/devs.c s3c_set_platdata(pd, sizeof(struct s3c2410_ts_mach_info), pd 981 arch/arm/plat-samsung/devs.c void __init s3c24xx_udc_set_platdata(struct s3c2410_udc_mach_info *pd) pd 983 arch/arm/plat-samsung/devs.c s3c_set_platdata(pd, sizeof(*pd), &s3c_device_usbgadget); pd 1006 arch/arm/plat-samsung/devs.c void __init dwc2_hsotg_set_platdata(struct dwc2_hsotg_plat *pd) pd 1010 arch/arm/plat-samsung/devs.c npd = s3c_set_platdata(pd, sizeof(*npd), &s3c_device_usb_hsotg); pd 1038 arch/arm/plat-samsung/devs.c void __init s3c24xx_hsudc_set_platdata(struct s3c24xx_hsudc_platdata *pd) pd 1040 arch/arm/plat-samsung/devs.c s3c_set_platdata(pd, sizeof(*pd), &s3c_device_usb_hsudc); pd 1080 arch/arm/plat-samsung/devs.c struct s3c64xx_spi_info pd; pd 1088 arch/arm/plat-samsung/devs.c pd.num_cs = num_cs; pd 1089 arch/arm/plat-samsung/devs.c pd.src_clk_nr = src_clk_nr; pd 1090 arch/arm/plat-samsung/devs.c pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi0_cfg_gpio; pd 1092 arch/arm/plat-samsung/devs.c s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi0); pd 1116 arch/arm/plat-samsung/devs.c struct s3c64xx_spi_info pd; pd 1124 arch/arm/plat-samsung/devs.c pd.num_cs = num_cs; pd 1125 arch/arm/plat-samsung/devs.c pd.src_clk_nr = src_clk_nr; pd 1126 arch/arm/plat-samsung/devs.c pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi1_cfg_gpio; pd 1128 arch/arm/plat-samsung/devs.c s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi1); pd 1152 arch/arm/plat-samsung/devs.c struct s3c64xx_spi_info pd; pd 1160 arch/arm/plat-samsung/devs.c pd.num_cs = num_cs; pd 1161 arch/arm/plat-samsung/devs.c pd.src_clk_nr = src_clk_nr; pd 1162 arch/arm/plat-samsung/devs.c pd.cfg_gpio = (cfg_gpio) ? cfg_gpio : s3c64xx_spi2_cfg_gpio; pd 1164 arch/arm/plat-samsung/devs.c s3c_set_platdata(&pd, sizeof(pd), &s3c64xx_device_spi2); pd 93 arch/arm/plat-samsung/include/plat/devs.h extern void *s3c_set_platdata(void *pd, size_t pdsize, pd 22 arch/arm/plat-samsung/include/plat/fb.h extern void s3c_fb_set_platdata(struct s3c_fb_platdata *pd); pd 22 arch/arm/plat-samsung/include/plat/keypad.h extern void samsung_keypad_set_platdata(struct samsung_keypad_platdata *pd); pd 14 arch/arm/plat-samsung/include/plat/pwm-core.h extern void samsung_pwm_set_platdata(struct samsung_pwm_variant *pd); pd 16 arch/arm/plat-samsung/include/plat/pwm-core.h static inline void samsung_pwm_set_platdata(struct samsung_pwm_variant *pd) { } pd 24 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci_set_platdata(struct s3c_sdhci_platdata *pd, pd 35 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci0_set_platdata(struct s3c_sdhci_platdata *pd); pd 36 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci1_set_platdata(struct s3c_sdhci_platdata *pd); pd 37 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci2_set_platdata(struct s3c_sdhci_platdata *pd); pd 38 arch/arm/plat-samsung/include/plat/sdhci.h extern void s3c_sdhci3_set_platdata(struct s3c_sdhci_platdata *pd); pd 15 arch/arm/plat-samsung/platformdata.c void __init *s3c_set_platdata(void *pd, size_t pdsize, pd 20 arch/arm/plat-samsung/platformdata.c if (!pd) { pd 26 arch/arm/plat-samsung/platformdata.c npd = kmemdup(pd, pdsize, GFP_KERNEL); pd 34 arch/arm/plat-samsung/platformdata.c void s3c_sdhci_set_platdata(struct s3c_sdhci_platdata *pd, pd 37 arch/arm/plat-samsung/platformdata.c set->cd_type = pd->cd_type; pd 38 arch/arm/plat-samsung/platformdata.c set->ext_cd_init = pd->ext_cd_init; pd 39 arch/arm/plat-samsung/platformdata.c set->ext_cd_cleanup = pd->ext_cd_cleanup; pd 40 arch/arm/plat-samsung/platformdata.c set->ext_cd_gpio = pd->ext_cd_gpio; pd 41 arch/arm/plat-samsung/platformdata.c set->ext_cd_gpio_invert = pd->ext_cd_gpio_invert; pd 43 arch/arm/plat-samsung/platformdata.c if (pd->max_width) pd 44 arch/arm/plat-samsung/platformdata.c set->max_width = pd->max_width; pd 45 arch/arm/plat-samsung/platformdata.c if (pd->cfg_gpio) pd 46 arch/arm/plat-samsung/platformdata.c set->cfg_gpio = pd->cfg_gpio; pd 47 arch/arm/plat-samsung/platformdata.c if (pd->host_caps) pd 48 arch/arm/plat-samsung/platformdata.c set->host_caps |= pd->host_caps; pd 49 arch/arm/plat-samsung/platformdata.c if (pd->host_caps2) pd 50 arch/arm/plat-samsung/platformdata.c set->host_caps2 |= pd->host_caps2; pd 51 arch/arm/plat-samsung/platformdata.c if (pd->pm_caps) pd 52 arch/arm/plat-samsung/platformdata.c set->pm_caps |= pd->pm_caps; pd 137 arch/microblaze/mm/pgtable.c pmd_t *pd; pd 141 arch/microblaze/mm/pgtable.c pd = pmd_offset(pgd_offset_k(va), va); pd 143 arch/microblaze/mm/pgtable.c pg = pte_alloc_kernel(pd, va); /* from powerpc - pgtable.c */ pd 99 arch/mips/alchemy/devboards/platform.c struct platform_device *pd; pd 111 arch/mips/alchemy/devboards/platform.c pd = platform_device_alloc("db1xxx_pcmcia", id); pd 112 arch/mips/alchemy/devboards/platform.c if (!pd) { pd 153 arch/mips/alchemy/devboards/platform.c pd->resource = sr; pd 154 arch/mips/alchemy/devboards/platform.c pd->num_resources = cnt; pd 156 arch/mips/alchemy/devboards/platform.c ret = platform_device_add(pd); pd 160 arch/mips/alchemy/devboards/platform.c platform_device_put(pd); pd 173 arch/mips/alchemy/devboards/platform.c struct platform_device *pd; pd 194 arch/mips/alchemy/devboards/platform.c pd = platform_device_alloc("physmap-flash", 0); pd 195 arch/mips/alchemy/devboards/platform.c if (!pd) pd 248 arch/mips/alchemy/devboards/platform.c pd->dev.platform_data = pfd; pd 249 arch/mips/alchemy/devboards/platform.c pd->resource = res; pd 250 arch/mips/alchemy/devboards/platform.c pd->num_resources = 1; pd 252 arch/mips/alchemy/devboards/platform.c ret = platform_device_add(pd); pd 256 arch/mips/alchemy/devboards/platform.c platform_device_put(pd); pd 211 arch/mips/bcm63xx/dev-enet.c const struct bcm63xx_enet_platform_data *pd) pd 247 arch/mips/bcm63xx/dev-enet.c memcpy(dpd, pd, sizeof(*pd)); pd 291 arch/mips/bcm63xx/dev-enet.c bcm63xx_enetsw_register(const struct bcm63xx_enetsw_platform_data *pd) pd 310 arch/mips/bcm63xx/dev-enet.c memcpy(bcm63xx_enetsw_device.dev.platform_data, pd, sizeof(*pd)); pd 57 arch/mips/bcm63xx/dev-pcmcia.c static struct bcm63xx_pcmcia_platform_data pd; pd 65 arch/mips/bcm63xx/dev-pcmcia.c .platform_data = &pd, pd 114 arch/mips/bcm63xx/dev-pcmcia.c pd.ready_gpio = 22; pd 118 arch/mips/bcm63xx/dev-pcmcia.c pd.ready_gpio = 18; pd 36 arch/mips/bcm63xx/dev-usb-usbd.c int __init bcm63xx_usbd_register(const struct bcm63xx_usbd_platform_data *pd) pd 62 arch/mips/bcm63xx/dev-usb-usbd.c platform_device_add_data(&bcm63xx_usbd_device, pd, sizeof(*pd)); pd 316 arch/mips/cavium-octeon/octeon-platform.c struct platform_device *pd; pd 324 arch/mips/cavium-octeon/octeon-platform.c pd = of_find_device_by_node(ehci_node); pd 326 arch/mips/cavium-octeon/octeon-platform.c if (!pd) pd 329 arch/mips/cavium-octeon/octeon-platform.c pd->dev.platform_data = &octeon_ehci_pdata; pd 330 arch/mips/cavium-octeon/octeon-platform.c octeon_ehci_hw_start(&pd->dev); pd 379 arch/mips/cavium-octeon/octeon-platform.c struct platform_device *pd; pd 387 arch/mips/cavium-octeon/octeon-platform.c pd = of_find_device_by_node(ohci_node); pd 389 arch/mips/cavium-octeon/octeon-platform.c if (!pd) pd 392 arch/mips/cavium-octeon/octeon-platform.c pd->dev.platform_data = &octeon_ohci_pdata; pd 393 arch/mips/cavium-octeon/octeon-platform.c octeon_ohci_hw_start(&pd->dev); pd 404 arch/mips/cavium-octeon/octeon-platform.c struct platform_device *pd; pd 419 arch/mips/cavium-octeon/octeon-platform.c pd = platform_device_alloc("octeon_rng", -1); pd 420 arch/mips/cavium-octeon/octeon-platform.c if (!pd) { pd 425 arch/mips/cavium-octeon/octeon-platform.c ret = platform_device_add_resources(pd, rng_resources, pd 430 arch/mips/cavium-octeon/octeon-platform.c ret = platform_device_add(pd); pd 436 arch/mips/cavium-octeon/octeon-platform.c platform_device_put(pd); pd 20 arch/mips/cobalt/buttons.c struct platform_device *pd; pd 23 arch/mips/cobalt/buttons.c pd = platform_device_alloc("Cobalt buttons", -1); pd 24 arch/mips/cobalt/buttons.c if (!pd) pd 27 arch/mips/cobalt/buttons.c error = platform_device_add_resources(pd, &cobalt_buttons_resource, 1); pd 31 arch/mips/cobalt/buttons.c error = platform_device_add(pd); pd 38 arch/mips/cobalt/buttons.c platform_device_put(pd); pd 103 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_enet.h const struct bcm63xx_enet_platform_data *pd); pd 105 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_enet.h int bcm63xx_enetsw_register(const struct bcm63xx_enetsw_platform_data *pd); pd 16 arch/mips/include/asm/mach-bcm63xx/bcm63xx_dev_usb_usbd.h int bcm63xx_usbd_register(const struct bcm63xx_usbd_platform_data *pd); pd 90 arch/mips/netlogic/common/irq.c struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d); pd 92 arch/mips/netlogic/common/irq.c BUG_ON(!pd); pd 93 arch/mips/netlogic/common/irq.c spin_lock_irqsave(&pd->node->piclock, flags); pd 94 arch/mips/netlogic/common/irq.c nlm_pic_enable_irt(pd->node->picbase, pd->irt); pd 95 arch/mips/netlogic/common/irq.c spin_unlock_irqrestore(&pd->node->piclock, flags); pd 100 arch/mips/netlogic/common/irq.c struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d); pd 103 arch/mips/netlogic/common/irq.c BUG_ON(!pd); pd 104 arch/mips/netlogic/common/irq.c spin_lock_irqsave(&pd->node->piclock, flags); pd 105 arch/mips/netlogic/common/irq.c nlm_pic_disable_irt(pd->node->picbase, pd->irt); pd 106 arch/mips/netlogic/common/irq.c spin_unlock_irqrestore(&pd->node->piclock, flags); pd 111 arch/mips/netlogic/common/irq.c struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d); pd 113 arch/mips/netlogic/common/irq.c clear_c0_eimr(pd->picirq); pd 114 arch/mips/netlogic/common/irq.c ack_c0_eirr(pd->picirq); pd 119 arch/mips/netlogic/common/irq.c struct nlm_pic_irq *pd = irq_data_get_irq_chip_data(d); pd 121 arch/mips/netlogic/common/irq.c BUG_ON(!pd); pd 123 arch/mips/netlogic/common/irq.c if (pd->extra_ack) pd 124 arch/mips/netlogic/common/irq.c pd->extra_ack(d); pd 127 arch/mips/netlogic/common/irq.c set_c0_eimr(pd->picirq); pd 130 arch/mips/netlogic/common/irq.c nlm_pic_ack(pd->node->picbase, pd->irt); pd 364 arch/mips/pci/pci-alchemy.c struct alchemy_pci_platdata *pd = pdev->dev.platform_data; pd 373 arch/mips/pci/pci-alchemy.c if (!pd) { pd 442 arch/mips/pci/pci-alchemy.c if (pd->board_map_irq) pd 443 arch/mips/pci/pci-alchemy.c ctx->board_map_irq = pd->board_map_irq; pd 445 arch/mips/pci/pci-alchemy.c if (pd->board_pci_idsel) pd 446 arch/mips/pci/pci-alchemy.c ctx->board_pci_idsel = pd->board_pci_idsel; pd 474 arch/mips/pci/pci-alchemy.c val &= ~pd->pci_cfg_clr; pd 475 arch/mips/pci/pci-alchemy.c val |= pd->pci_cfg_set; pd 54 arch/mips/sgi-ip32/ip32-platform.c struct platform_device *pd; pd 57 arch/mips/sgi-ip32/ip32-platform.c pd = platform_device_alloc("meth", -1); pd 58 arch/mips/sgi-ip32/ip32-platform.c if (!pd) pd 61 arch/mips/sgi-ip32/ip32-platform.c ret = platform_device_add(pd); pd 63 arch/mips/sgi-ip32/ip32-platform.c platform_device_put(pd); pd 72 arch/mips/sgi-ip32/ip32-platform.c struct platform_device *pd; pd 75 arch/mips/sgi-ip32/ip32-platform.c pd = platform_device_alloc("sgio2audio", -1); pd 76 arch/mips/sgi-ip32/ip32-platform.c if (!pd) pd 79 arch/mips/sgi-ip32/ip32-platform.c ret = platform_device_add(pd); pd 81 arch/mips/sgi-ip32/ip32-platform.c platform_device_put(pd); pd 70 arch/powerpc/include/asm/pgtable-types.h typedef struct { unsigned long pd; } hugepd_t; pd 74 arch/powerpc/include/asm/pgtable-types.h return x.pd; pd 575 arch/powerpc/kernel/setup-common.c struct platform_device *pd; pd 583 arch/powerpc/kernel/setup-common.c pd = platform_device_alloc("pcspkr", -1); pd 584 arch/powerpc/kernel/setup-common.c if (!pd) pd 587 arch/powerpc/kernel/setup-common.c ret = platform_device_add(pd); pd 589 arch/powerpc/kernel/setup-common.c platform_device_put(pd); pd 61 arch/powerpc/mm/pgtable_32.c pmd_t *pd; pd 66 arch/powerpc/mm/pgtable_32.c pd = pmd_offset(pud_offset(pgd_offset_k(va), va), va); pd 69 arch/powerpc/mm/pgtable_32.c pg = pte_alloc_kernel(pd, va); pd 71 arch/powerpc/mm/pgtable_32.c pg = early_pte_alloc_kernel(pd, va); pd 96 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = irq_data_get_irq_chip_data(d); pd 100 arch/powerpc/platforms/ps3/interrupt.c pd->thread_id, d->irq); pd 103 arch/powerpc/platforms/ps3/interrupt.c clear_bit(63 - d->irq, &pd->bmp.mask); pd 104 arch/powerpc/platforms/ps3/interrupt.c lv1_did_update_interrupt_mask(pd->ppe_id, pd->thread_id); pd 117 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = irq_data_get_irq_chip_data(d); pd 121 arch/powerpc/platforms/ps3/interrupt.c pd->thread_id, d->irq); pd 124 arch/powerpc/platforms/ps3/interrupt.c set_bit(63 - d->irq, &pd->bmp.mask); pd 125 arch/powerpc/platforms/ps3/interrupt.c lv1_did_update_interrupt_mask(pd->ppe_id, pd->thread_id); pd 138 arch/powerpc/platforms/ps3/interrupt.c const struct ps3_private *pd = irq_data_get_irq_chip_data(d); pd 142 arch/powerpc/platforms/ps3/interrupt.c if (!test_bit(63 - d->irq, &pd->ipi_mask)) pd 143 arch/powerpc/platforms/ps3/interrupt.c lv1_end_of_interrupt_ext(pd->ppe_id, pd->thread_id, d->irq); pd 172 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd; pd 179 arch/powerpc/platforms/ps3/interrupt.c pd = &per_cpu(ps3_private, cpu); pd 193 arch/powerpc/platforms/ps3/interrupt.c result = irq_set_chip_data(*virq, pd); pd 220 arch/powerpc/platforms/ps3/interrupt.c const struct ps3_private *pd = irq_get_chip_data(virq); pd 223 arch/powerpc/platforms/ps3/interrupt.c __LINE__, pd->ppe_id, pd->thread_id, virq); pd 246 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd; pd 255 arch/powerpc/platforms/ps3/interrupt.c pd = irq_get_chip_data(*virq); pd 259 arch/powerpc/platforms/ps3/interrupt.c result = lv1_connect_irq_plug_ext(pd->ppe_id, pd->thread_id, *virq, pd 290 arch/powerpc/platforms/ps3/interrupt.c const struct ps3_private *pd = irq_get_chip_data(virq); pd 293 arch/powerpc/platforms/ps3/interrupt.c __LINE__, pd->ppe_id, pd->thread_id, virq); pd 297 arch/powerpc/platforms/ps3/interrupt.c result = lv1_disconnect_irq_plug_ext(pd->ppe_id, pd->thread_id, virq); pd 634 arch/powerpc/platforms/ps3/interrupt.c static void _dump_bmp(struct ps3_private* pd, const char* func, int line) pd 638 arch/powerpc/platforms/ps3/interrupt.c spin_lock_irqsave(&pd->bmp_lock, flags); pd 639 arch/powerpc/platforms/ps3/interrupt.c _dump_64_bmp("stat", &pd->bmp.status, pd->thread_id, func, line); pd 640 arch/powerpc/platforms/ps3/interrupt.c _dump_64_bmp("mask", (u64*)&pd->bmp.mask, pd->thread_id, func, line); pd 641 arch/powerpc/platforms/ps3/interrupt.c spin_unlock_irqrestore(&pd->bmp_lock, flags); pd 645 arch/powerpc/platforms/ps3/interrupt.c static void __maybe_unused _dump_mask(struct ps3_private *pd, pd 650 arch/powerpc/platforms/ps3/interrupt.c spin_lock_irqsave(&pd->bmp_lock, flags); pd 651 arch/powerpc/platforms/ps3/interrupt.c _dump_64_bmp("mask", (u64*)&pd->bmp.mask, pd->thread_id, func, line); pd 652 arch/powerpc/platforms/ps3/interrupt.c spin_unlock_irqrestore(&pd->bmp_lock, flags); pd 655 arch/powerpc/platforms/ps3/interrupt.c static void dump_bmp(struct ps3_private* pd) {}; pd 683 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = &per_cpu(ps3_private, cpu); pd 685 arch/powerpc/platforms/ps3/interrupt.c set_bit(63 - virq, &pd->ipi_debug_brk_mask); pd 688 arch/powerpc/platforms/ps3/interrupt.c cpu, virq, pd->ipi_debug_brk_mask); pd 693 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = &per_cpu(ps3_private, cpu); pd 695 arch/powerpc/platforms/ps3/interrupt.c set_bit(63 - virq, &pd->ipi_mask); pd 698 arch/powerpc/platforms/ps3/interrupt.c cpu, virq, pd->ipi_mask); pd 703 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = this_cpu_ptr(&ps3_private); pd 704 arch/powerpc/platforms/ps3/interrupt.c u64 x = (pd->bmp.status & pd->bmp.mask); pd 709 arch/powerpc/platforms/ps3/interrupt.c if (x & pd->ipi_debug_brk_mask) pd 710 arch/powerpc/platforms/ps3/interrupt.c x &= pd->ipi_debug_brk_mask; pd 717 arch/powerpc/platforms/ps3/interrupt.c __LINE__, pd->thread_id); pd 733 arch/powerpc/platforms/ps3/interrupt.c if (test_bit(63 - plug, &pd->ipi_mask)) pd 734 arch/powerpc/platforms/ps3/interrupt.c lv1_end_of_interrupt_ext(pd->ppe_id, pd->thread_id, plug); pd 749 arch/powerpc/platforms/ps3/interrupt.c struct ps3_private *pd = &per_cpu(ps3_private, cpu); pd 751 arch/powerpc/platforms/ps3/interrupt.c lv1_get_logical_ppe_id(&pd->ppe_id); pd 752 arch/powerpc/platforms/ps3/interrupt.c pd->thread_id = get_hard_smp_processor_id(cpu); pd 753 arch/powerpc/platforms/ps3/interrupt.c spin_lock_init(&pd->bmp_lock); pd 756 arch/powerpc/platforms/ps3/interrupt.c __func__, __LINE__, pd->ppe_id, pd->thread_id, pd 757 arch/powerpc/platforms/ps3/interrupt.c ps3_mm_phys_to_lpar(__pa(&pd->bmp))); pd 759 arch/powerpc/platforms/ps3/interrupt.c result = lv1_configure_irq_state_bitmap(pd->ppe_id, pd 760 arch/powerpc/platforms/ps3/interrupt.c pd->thread_id, ps3_mm_phys_to_lpar(__pa(&pd->bmp))); pd 1294 arch/powerpc/sysdev/fsl_pci.c struct mpc85xx_edac_pci_plat_data pd = { pd 1303 arch/powerpc/sysdev/fsl_pci.c &pd, sizeof(pd)); pd 23 arch/powerpc/sysdev/rtc_cmos_setup.c struct platform_device *pd; pd 62 arch/powerpc/sysdev/rtc_cmos_setup.c pd = platform_device_register_simple("rtc_cmos", -1, pd 65 arch/powerpc/sysdev/rtc_cmos_setup.c return PTR_ERR_OR_ZERO(pd); pd 39 arch/s390/include/asm/nmi.h u64 pd : 1; /* 01 instruction-processing damage */ pd 26 arch/s390/include/asm/vx-insn.h \opd = 255 pd 28 arch/s390/include/asm/vx-insn.h \opd = 0 pd 31 arch/s390/include/asm/vx-insn.h \opd = 1 pd 34 arch/s390/include/asm/vx-insn.h \opd = 2 pd 37 arch/s390/include/asm/vx-insn.h \opd = 3 pd 40 arch/s390/include/asm/vx-insn.h \opd = 4 pd 43 arch/s390/include/asm/vx-insn.h \opd = 5 pd 46 arch/s390/include/asm/vx-insn.h \opd = 6 pd 49 arch/s390/include/asm/vx-insn.h \opd = 7 pd 52 arch/s390/include/asm/vx-insn.h \opd = 8 pd 55 arch/s390/include/asm/vx-insn.h \opd = 9 pd 58 arch/s390/include/asm/vx-insn.h \opd = 10 pd 61 arch/s390/include/asm/vx-insn.h \opd = 11 pd 64 arch/s390/include/asm/vx-insn.h \opd = 12 pd 67 arch/s390/include/asm/vx-insn.h \opd = 13 pd 70 arch/s390/include/asm/vx-insn.h \opd = 14 pd 73 arch/s390/include/asm/vx-insn.h \opd = 15 pd 75 arch/s390/include/asm/vx-insn.h .if \opd == 255 pd 76 arch/s390/include/asm/vx-insn.h \opd = \gr pd 90 arch/s390/include/asm/vx-insn.h \opd = 255 pd 92 arch/s390/include/asm/vx-insn.h \opd = 0 pd 95 arch/s390/include/asm/vx-insn.h \opd = 1 pd 98 arch/s390/include/asm/vx-insn.h \opd = 2 pd 101 arch/s390/include/asm/vx-insn.h \opd = 3 pd 104 arch/s390/include/asm/vx-insn.h \opd = 4 pd 107 arch/s390/include/asm/vx-insn.h \opd = 5 pd 110 arch/s390/include/asm/vx-insn.h \opd = 6 pd 113 arch/s390/include/asm/vx-insn.h \opd = 7 pd 116 arch/s390/include/asm/vx-insn.h \opd = 8 pd 119 arch/s390/include/asm/vx-insn.h \opd = 9 pd 122 arch/s390/include/asm/vx-insn.h \opd = 10 pd 125 arch/s390/include/asm/vx-insn.h \opd = 11 pd 128 arch/s390/include/asm/vx-insn.h \opd = 12 pd 131 arch/s390/include/asm/vx-insn.h \opd = 13 pd 134 arch/s390/include/asm/vx-insn.h \opd = 14 pd 137 arch/s390/include/asm/vx-insn.h \opd = 15 pd 140 arch/s390/include/asm/vx-insn.h \opd = 16 pd 143 arch/s390/include/asm/vx-insn.h \opd = 17 pd 146 arch/s390/include/asm/vx-insn.h \opd = 18 pd 149 arch/s390/include/asm/vx-insn.h \opd = 19 pd 152 arch/s390/include/asm/vx-insn.h \opd = 20 pd 155 arch/s390/include/asm/vx-insn.h \opd = 21 pd 158 arch/s390/include/asm/vx-insn.h \opd = 22 pd 161 arch/s390/include/asm/vx-insn.h \opd = 23 pd 164 arch/s390/include/asm/vx-insn.h \opd = 24 pd 167 arch/s390/include/asm/vx-insn.h \opd = 25 pd 170 arch/s390/include/asm/vx-insn.h \opd = 26 pd 173 arch/s390/include/asm/vx-insn.h \opd = 27 pd 176 arch/s390/include/asm/vx-insn.h \opd = 28 pd 179 arch/s390/include/asm/vx-insn.h \opd = 29 pd 182 arch/s390/include/asm/vx-insn.h \opd = 30 pd 185 arch/s390/include/asm/vx-insn.h \opd = 31 pd 187 arch/s390/include/asm/vx-insn.h .if \opd == 255 pd 188 arch/s390/include/asm/vx-insn.h \opd = \vxr pd 361 arch/s390/kernel/nmi.c if (mci.pd && !test_cpu_flag(CIF_MCCK_GUEST)) { pd 1057 arch/sh/boards/mach-ecovec24/setup.c static void __init sh_eth_init(struct sh_eth_plat_data *pd) pd 1068 arch/sh/boards/mach-ecovec24/setup.c for (i = 0; i < sizeof(pd->mac_addr); i++) { pd 1069 arch/sh/boards/mach-ecovec24/setup.c pd->mac_addr[i] = mac_read(a, 0x10 + i); pd 1076 arch/sh/boards/mach-ecovec24/setup.c static void __init sh_eth_init(struct sh_eth_plat_data *pd) pd 19 arch/x86/include/asm/inst.h \opd = REG_NUM_INVALID pd 21 arch/x86/include/asm/inst.h \opd = 0 pd 24 arch/x86/include/asm/inst.h \opd = 1 pd 27 arch/x86/include/asm/inst.h \opd = 2 pd 30 arch/x86/include/asm/inst.h \opd = 3 pd 33 arch/x86/include/asm/inst.h \opd = 4 pd 36 arch/x86/include/asm/inst.h \opd = 5 pd 39 arch/x86/include/asm/inst.h \opd = 6 pd 42 arch/x86/include/asm/inst.h \opd = 7 pd 46 arch/x86/include/asm/inst.h \opd = 8 pd 49 arch/x86/include/asm/inst.h \opd = 9 pd 52 arch/x86/include/asm/inst.h \opd = 10 pd 55 arch/x86/include/asm/inst.h \opd = 11 pd 58 arch/x86/include/asm/inst.h \opd = 12 pd 61 arch/x86/include/asm/inst.h \opd = 13 pd 64 arch/x86/include/asm/inst.h \opd = 14 pd 67 arch/x86/include/asm/inst.h \opd = 15 pd 73 arch/x86/include/asm/inst.h \opd = REG_NUM_INVALID pd 76 arch/x86/include/asm/inst.h \opd = 0 pd 79 arch/x86/include/asm/inst.h \opd = 1 pd 82 arch/x86/include/asm/inst.h \opd = 2 pd 85 arch/x86/include/asm/inst.h \opd = 3 pd 88 arch/x86/include/asm/inst.h \opd = 4 pd 91 arch/x86/include/asm/inst.h \opd = 5 pd 94 arch/x86/include/asm/inst.h \opd = 6 pd 97 arch/x86/include/asm/inst.h \opd = 7 pd 100 arch/x86/include/asm/inst.h \opd = 8 pd 103 arch/x86/include/asm/inst.h \opd = 9 pd 106 arch/x86/include/asm/inst.h \opd = 10 pd 109 arch/x86/include/asm/inst.h \opd = 11 pd 112 arch/x86/include/asm/inst.h \opd = 12 pd 115 arch/x86/include/asm/inst.h \opd = 13 pd 118 arch/x86/include/asm/inst.h \opd = 14 pd 121 arch/x86/include/asm/inst.h \opd = 15 pd 127 arch/x86/include/asm/inst.h \opd = REG_NUM_INVALID pd 129 arch/x86/include/asm/inst.h \opd = 0 pd 132 arch/x86/include/asm/inst.h \opd = 1 pd 135 arch/x86/include/asm/inst.h \opd = 2 pd 138 arch/x86/include/asm/inst.h \opd = 3 pd 141 arch/x86/include/asm/inst.h \opd = 4 pd 144 arch/x86/include/asm/inst.h \opd = 5 pd 147 arch/x86/include/asm/inst.h \opd = 6 pd 150 arch/x86/include/asm/inst.h \opd = 7 pd 153 arch/x86/include/asm/inst.h \opd = 8 pd 156 arch/x86/include/asm/inst.h \opd = 9 pd 159 arch/x86/include/asm/inst.h \opd = 10 pd 162 arch/x86/include/asm/inst.h \opd = 11 pd 165 arch/x86/include/asm/inst.h \opd = 12 pd 168 arch/x86/include/asm/inst.h \opd = 13 pd 171 arch/x86/include/asm/inst.h \opd = 14 pd 174 arch/x86/include/asm/inst.h \opd = 15 pd 8 arch/x86/kernel/pcspeaker.c struct platform_device *pd; pd 10 arch/x86/kernel/pcspeaker.c pd = platform_device_register_simple("pcspkr", -1, NULL, 0); pd 12 arch/x86/kernel/pcspeaker.c return PTR_ERR_OR_ZERO(pd); pd 41 arch/x86/kernel/sysfb.c struct platform_device *pd; pd 64 arch/x86/kernel/sysfb.c pd = platform_device_register_resndata(NULL, name, 0, pd 66 arch/x86/kernel/sysfb.c return PTR_ERR_OR_ZERO(pd); pd 63 arch/x86/kernel/sysfb_simplefb.c struct platform_device *pd; pd 108 arch/x86/kernel/sysfb_simplefb.c pd = platform_device_register_resndata(NULL, "simple-framebuffer", 0, pd 110 arch/x86/kernel/sysfb_simplefb.c return PTR_ERR_OR_ZERO(pd); pd 5892 arch/x86/kvm/x86.c static int kernel_pio(struct kvm_vcpu *vcpu, void *pd) pd 5899 arch/x86/kvm/x86.c vcpu->arch.pio.size, pd); pd 5903 arch/x86/kvm/x86.c pd); pd 5906 arch/x86/kvm/x86.c pd += vcpu->arch.pio.size; pd 90 block/bfq-cgroup.c static u64 blkg_prfill_stat(struct seq_file *sf, struct blkg_policy_data *pd, pd 93 block/bfq-cgroup.c return __blkg_prfill_u64(sf, pd, bfq_stat_read((void *)pd + off)); pd 279 block/bfq-cgroup.c static struct bfq_group *pd_to_bfqg(struct blkg_policy_data *pd) pd 281 block/bfq-cgroup.c return pd ? container_of(pd, struct bfq_group, pd) : NULL; pd 286 block/bfq-cgroup.c return pd_to_blkg(&bfqg->pd); pd 473 block/bfq-cgroup.c return cpd ? container_of(cpd, struct bfq_group_data, pd) : NULL; pd 488 block/bfq-cgroup.c return &bgd->pd; pd 520 block/bfq-cgroup.c return &bfqg->pd; pd 523 block/bfq-cgroup.c static void bfq_pd_init(struct blkg_policy_data *pd) pd 525 block/bfq-cgroup.c struct blkcg_gq *blkg = pd_to_blkg(pd); pd 542 block/bfq-cgroup.c static void bfq_pd_free(struct blkg_policy_data *pd) pd 544 block/bfq-cgroup.c struct bfq_group *bfqg = pd_to_bfqg(pd); pd 550 block/bfq-cgroup.c static void bfq_pd_reset_stats(struct blkg_policy_data *pd) pd 552 block/bfq-cgroup.c struct bfq_group *bfqg = pd_to_bfqg(pd); pd 858 block/bfq-cgroup.c static void bfq_pd_offline(struct blkg_policy_data *pd) pd 861 block/bfq-cgroup.c struct bfq_group *bfqg = pd_to_bfqg(pd); pd 949 block/bfq-cgroup.c struct blkg_policy_data *pd, int off) pd 951 block/bfq-cgroup.c struct bfq_group *bfqg = pd_to_bfqg(pd); pd 955 block/bfq-cgroup.c return __blkg_prfill_u64(sf, pd, bfqg->entity.dev_weight); pd 1102 block/bfq-cgroup.c struct blkg_policy_data *pd, int off) pd 1104 block/bfq-cgroup.c struct blkcg_gq *blkg = pd_to_blkg(pd); pd 1123 block/bfq-cgroup.c return __blkg_prfill_u64(sf, pd, sum); pd 1127 block/bfq-cgroup.c struct blkg_policy_data *pd, int off) pd 1131 block/bfq-cgroup.c blkg_rwstat_recursive_sum(pd_to_blkg(pd), &blkcg_policy_bfq, off, &sum); pd 1132 block/bfq-cgroup.c return __blkg_prfill_rwstat(sf, pd, &sum); pd 1151 block/bfq-cgroup.c static u64 bfqg_prfill_sectors(struct seq_file *sf, struct blkg_policy_data *pd, pd 1154 block/bfq-cgroup.c u64 sum = blkg_rwstat_total(&pd->blkg->stat_bytes); pd 1156 block/bfq-cgroup.c return __blkg_prfill_u64(sf, pd, sum >> 9); pd 1167 block/bfq-cgroup.c struct blkg_policy_data *pd, int off) pd 1171 block/bfq-cgroup.c blkg_rwstat_recursive_sum(pd->blkg, NULL, pd 1174 block/bfq-cgroup.c return __blkg_prfill_u64(sf, pd, pd 1187 block/bfq-cgroup.c struct blkg_policy_data *pd, int off) pd 1189 block/bfq-cgroup.c struct bfq_group *bfqg = pd_to_bfqg(pd); pd 1197 block/bfq-cgroup.c __blkg_prfill_u64(sf, pd, v); pd 853 block/bfq-iosched.h struct blkcg_policy_data pd; pd 892 block/bfq-iosched.h struct blkg_policy_data pd; pd 80 block/blk-cgroup.c if (blkg->pd[i]) pd 81 block/blk-cgroup.c blkcg_policy[i]->pd_free_fn(blkg->pd[i]); pd 172 block/blk-cgroup.c struct blkg_policy_data *pd; pd 178 block/blk-cgroup.c pd = pol->pd_alloc_fn(gfp_mask, q, blkcg); pd 179 block/blk-cgroup.c if (!pd) pd 182 block/blk-cgroup.c blkg->pd[i] = pd; pd 183 block/blk-cgroup.c pd->blkg = blkg; pd 184 block/blk-cgroup.c pd->plid = i; pd 278 block/blk-cgroup.c if (blkg->pd[i] && pol->pd_init_fn) pd 279 block/blk-cgroup.c pol->pd_init_fn(blkg->pd[i]); pd 292 block/blk-cgroup.c if (blkg->pd[i] && pol->pd_online_fn) pd 293 block/blk-cgroup.c pol->pd_online_fn(blkg->pd[i]); pd 409 block/blk-cgroup.c if (blkg->pd[i] && pol->pd_offline_fn) pd 410 block/blk-cgroup.c pol->pd_offline_fn(blkg->pd[i]); pd 484 block/blk-cgroup.c if (blkg->pd[i] && pol->pd_reset_stats_fn) pd 485 block/blk-cgroup.c pol->pd_reset_stats_fn(blkg->pd[i]); pd 533 block/blk-cgroup.c total += prfill(sf, blkg->pd[pol->plid], data); pd 551 block/blk-cgroup.c u64 __blkg_prfill_u64(struct seq_file *sf, struct blkg_policy_data *pd, u64 v) pd 553 block/blk-cgroup.c const char *dname = blkg_dev_name(pd->blkg); pd 571 block/blk-cgroup.c u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd, pd 581 block/blk-cgroup.c const char *dname = blkg_dev_name(pd->blkg); pd 608 block/blk-cgroup.c u64 blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd, pd 613 block/blk-cgroup.c blkg_rwstat_read((void *)pd + off, &rwstat); pd 614 block/blk-cgroup.c return __blkg_prfill_rwstat(sf, pd, &rwstat); pd 619 block/blk-cgroup.c struct blkg_policy_data *pd, int off) pd 623 block/blk-cgroup.c blkg_rwstat_read((void *)pd->blkg + off, &rwstat); pd 624 block/blk-cgroup.c return __blkg_prfill_rwstat(sf, pd, &rwstat); pd 662 block/blk-cgroup.c struct blkg_policy_data *pd, pd 667 block/blk-cgroup.c blkg_rwstat_recursive_sum(pd->blkg, NULL, off, &rwstat); pd 668 block/blk-cgroup.c return __blkg_prfill_rwstat(sf, pd, &rwstat); pd 986 block/blk-cgroup.c if (!blkg->pd[i] || !pol->pd_stat_fn) pd 989 block/blk-cgroup.c written = pol->pd_stat_fn(blkg->pd[i], buf+off, size-off); pd 1381 block/blk-cgroup.c struct blkg_policy_data *pd; pd 1383 block/blk-cgroup.c if (blkg->pd[pol->plid]) pd 1388 block/blk-cgroup.c pd = pd_prealloc; pd 1391 block/blk-cgroup.c pd = pol->pd_alloc_fn(GFP_NOWAIT | __GFP_NOWARN, q, pd 1395 block/blk-cgroup.c if (!pd) { pd 1417 block/blk-cgroup.c blkg->pd[pol->plid] = pd; pd 1418 block/blk-cgroup.c pd->blkg = blkg; pd 1419 block/blk-cgroup.c pd->plid = pol->plid; pd 1425 block/blk-cgroup.c pol->pd_init_fn(blkg->pd[pol->plid]); pd 1444 block/blk-cgroup.c if (blkg->pd[pol->plid]) { pd 1445 block/blk-cgroup.c pol->pd_free_fn(blkg->pd[pol->plid]); pd 1446 block/blk-cgroup.c blkg->pd[pol->plid] = NULL; pd 1479 block/blk-cgroup.c if (blkg->pd[pol->plid]) { pd 1481 block/blk-cgroup.c pol->pd_offline_fn(blkg->pd[pol->plid]); pd 1482 block/blk-cgroup.c pol->pd_free_fn(blkg->pd[pol->plid]); pd 1483 block/blk-cgroup.c blkg->pd[pol->plid] = NULL; pd 431 block/blk-iocost.c struct blkg_policy_data pd; pd 634 block/blk-iocost.c static struct ioc_gq *pd_to_iocg(struct blkg_policy_data *pd) pd 636 block/blk-iocost.c return pd ? container_of(pd, struct ioc_gq, pd) : NULL; pd 646 block/blk-iocost.c return pd_to_blkg(&iocg->pd); pd 2013 block/blk-iocost.c return &iocg->pd; pd 2016 block/blk-iocost.c static void ioc_pd_init(struct blkg_policy_data *pd) pd 2018 block/blk-iocost.c struct ioc_gq *iocg = pd_to_iocg(pd); pd 2019 block/blk-iocost.c struct blkcg_gq *blkg = pd_to_blkg(&iocg->pd); pd 2053 block/blk-iocost.c static void ioc_pd_free(struct blkg_policy_data *pd) pd 2055 block/blk-iocost.c struct ioc_gq *iocg = pd_to_iocg(pd); pd 2072 block/blk-iocost.c static u64 ioc_weight_prfill(struct seq_file *sf, struct blkg_policy_data *pd, pd 2075 block/blk-iocost.c const char *dname = blkg_dev_name(pd->blkg); pd 2076 block/blk-iocost.c struct ioc_gq *iocg = pd_to_iocg(pd); pd 2158 block/blk-iocost.c static u64 ioc_qos_prfill(struct seq_file *sf, struct blkg_policy_data *pd, pd 2161 block/blk-iocost.c const char *dname = blkg_dev_name(pd->blkg); pd 2162 block/blk-iocost.c struct ioc *ioc = pd_to_iocg(pd)->ioc; pd 2330 block/blk-iocost.c struct blkg_policy_data *pd, int off) pd 2332 block/blk-iocost.c const char *dname = blkg_dev_name(pd->blkg); pd 2333 block/blk-iocost.c struct ioc *ioc = pd_to_iocg(pd)->ioc; pd 134 block/blk-iolatency.c struct blkg_policy_data pd; pd 177 block/blk-iolatency.c static inline struct iolatency_grp *pd_to_lat(struct blkg_policy_data *pd) pd 179 block/blk-iolatency.c return pd ? container_of(pd, struct iolatency_grp, pd) : NULL; pd 189 block/blk-iolatency.c return pd_to_blkg(&iolat->pd); pd 868 block/blk-iolatency.c struct blkg_policy_data *pd, int off) pd 870 block/blk-iolatency.c struct iolatency_grp *iolat = pd_to_lat(pd); pd 871 block/blk-iolatency.c const char *dname = blkg_dev_name(pd->blkg); pd 913 block/blk-iolatency.c static size_t iolatency_pd_stat(struct blkg_policy_data *pd, char *buf, pd 916 block/blk-iolatency.c struct iolatency_grp *iolat = pd_to_lat(pd); pd 952 block/blk-iolatency.c return &iolat->pd; pd 955 block/blk-iolatency.c static void iolatency_pd_init(struct blkg_policy_data *pd) pd 957 block/blk-iolatency.c struct iolatency_grp *iolat = pd_to_lat(pd); pd 1000 block/blk-iolatency.c static void iolatency_pd_offline(struct blkg_policy_data *pd) pd 1002 block/blk-iolatency.c struct iolatency_grp *iolat = pd_to_lat(pd); pd 1015 block/blk-iolatency.c static void iolatency_pd_free(struct blkg_policy_data *pd) pd 1017 block/blk-iolatency.c struct iolatency_grp *iolat = pd_to_lat(pd); pd 108 block/blk-throttle.c struct blkg_policy_data pd; pd 227 block/blk-throttle.c static inline struct throtl_grp *pd_to_tg(struct blkg_policy_data *pd) pd 229 block/blk-throttle.c return pd ? container_of(pd, struct throtl_grp, pd) : NULL; pd 239 block/blk-throttle.c return pd_to_blkg(&tg->pd); pd 515 block/blk-throttle.c return &tg->pd; pd 518 block/blk-throttle.c static void throtl_pd_init(struct blkg_policy_data *pd) pd 520 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 562 block/blk-throttle.c static void throtl_pd_online(struct blkg_policy_data *pd) pd 564 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 594 block/blk-throttle.c static void throtl_pd_offline(struct blkg_policy_data *pd) pd 596 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 609 block/blk-throttle.c static void throtl_pd_free(struct blkg_policy_data *pd) pd 611 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 1330 block/blk-throttle.c static u64 tg_prfill_conf_u64(struct seq_file *sf, struct blkg_policy_data *pd, pd 1333 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 1338 block/blk-throttle.c return __blkg_prfill_u64(sf, pd, v); pd 1341 block/blk-throttle.c static u64 tg_prfill_conf_uint(struct seq_file *sf, struct blkg_policy_data *pd, pd 1344 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 1349 block/blk-throttle.c return __blkg_prfill_u64(sf, pd, v); pd 1515 block/blk-throttle.c static u64 tg_prfill_limit(struct seq_file *sf, struct blkg_policy_data *pd, pd 1518 block/blk-throttle.c struct throtl_grp *tg = pd_to_tg(pd); pd 1519 block/blk-throttle.c const char *dname = blkg_dev_name(pd->blkg); pd 36 drivers/ata/pata_pxa.c struct pata_pxa_data *pd = d; pd 39 drivers/ata/pata_pxa.c status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, NULL); pd 41 drivers/ata/pata_pxa.c complete(&pd->dma_done); pd 49 drivers/ata/pata_pxa.c struct pata_pxa_data *pd = qc->ap->private_data; pd 57 drivers/ata/pata_pxa.c tx = dmaengine_prep_slave_sg(pd->dma_chan, qc->sg, qc->n_elem, dir, pd 64 drivers/ata/pata_pxa.c tx->callback_param = pd; pd 65 drivers/ata/pata_pxa.c pd->dma_cookie = dmaengine_submit(tx); pd 82 drivers/ata/pata_pxa.c struct pata_pxa_data *pd = qc->ap->private_data; pd 83 drivers/ata/pata_pxa.c init_completion(&pd->dma_done); pd 84 drivers/ata/pata_pxa.c dma_async_issue_pending(pd->dma_chan); pd 92 drivers/ata/pata_pxa.c struct pata_pxa_data *pd = qc->ap->private_data; pd 95 drivers/ata/pata_pxa.c status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, NULL); pd 97 drivers/ata/pata_pxa.c wait_for_completion_timeout(&pd->dma_done, HZ)) pd 100 drivers/ata/pata_pxa.c dmaengine_terminate_all(pd->dma_chan); pd 109 drivers/ata/pata_pxa.c struct pata_pxa_data *pd = ap->private_data; pd 114 drivers/ata/pata_pxa.c status = dmaengine_tx_status(pd->dma_chan, pd->dma_cookie, &state); pd 1427 drivers/base/platform.c struct platform_device *pd; pd 1429 drivers/base/platform.c list_for_each_entry(pd, &early_platform_device_list, dev.devres_head) pd 1430 drivers/base/platform.c if (platform_match(&pd->dev, &epdrv->pdrv->driver)) pd 1431 drivers/base/platform.c if (pd->id == id) pd 1432 drivers/base/platform.c return pd; pd 1445 drivers/base/platform.c struct platform_device *pd; pd 1447 drivers/base/platform.c list_for_each_entry(pd, &early_platform_device_list, dev.devres_head) pd 1448 drivers/base/platform.c if (platform_match(&pd->dev, &epdrv->pdrv->driver)) pd 1449 drivers/base/platform.c if (pd->id >= id) pd 1582 drivers/base/platform.c struct platform_device *pd, *pd2; pd 1585 drivers/base/platform.c list_for_each_entry_safe(pd, pd2, &early_platform_device_list, pd 1587 drivers/base/platform.c list_del(&pd->dev.devres_head); pd 1588 drivers/base/platform.c memset(&pd->dev.devres_head, 0, sizeof(pd->dev.devres_head)); pd 200 drivers/base/power/common.c void dev_pm_domain_set(struct device *dev, struct dev_pm_domain *pd) pd 202 drivers/base/power/common.c if (dev->pm_domain == pd) pd 205 drivers/base/power/common.c WARN(pd && device_is_bound(dev), pd 207 drivers/base/power/common.c dev->pm_domain = pd; pd 2360 drivers/base/power/domain.c struct generic_pm_domain *pd; pd 2364 drivers/base/power/domain.c pd = dev_to_genpd(dev); pd 2365 drivers/base/power/domain.c if (IS_ERR(pd)) pd 2368 drivers/base/power/domain.c dev_dbg(dev, "removing from PM domain %s\n", pd->name); pd 2371 drivers/base/power/domain.c ret = genpd_remove_device(pd, dev); pd 2381 drivers/base/power/domain.c pd->name, ret); pd 2386 drivers/base/power/domain.c genpd_queue_power_off_work(pd); pd 2395 drivers/base/power/domain.c struct generic_pm_domain *pd; pd 2397 drivers/base/power/domain.c pd = dev_to_genpd(dev); pd 2398 drivers/base/power/domain.c if (IS_ERR(pd)) pd 2401 drivers/base/power/domain.c genpd_queue_power_off_work(pd); pd 2408 drivers/base/power/domain.c struct generic_pm_domain *pd; pd 2417 drivers/base/power/domain.c pd = genpd_get_from_provider(&pd_args); pd 2419 drivers/base/power/domain.c if (IS_ERR(pd)) { pd 2422 drivers/base/power/domain.c __func__, PTR_ERR(pd)); pd 2426 drivers/base/power/domain.c dev_dbg(dev, "adding to PM domain %s\n", pd->name); pd 2428 drivers/base/power/domain.c ret = genpd_add_device(pd, dev, base_dev); pd 2434 drivers/base/power/domain.c pd->name, ret); pd 2442 drivers/base/power/domain.c genpd_lock(pd); pd 2443 drivers/base/power/domain.c ret = genpd_power_on(pd, 0); pd 2444 drivers/base/power/domain.c genpd_unlock(pd); pd 2448 drivers/base/power/domain.c genpd_remove_device(pd, dev); pd 120 drivers/base/power/domain_governor.c static bool __default_power_down_ok(struct dev_pm_domain *pd, pd 123 drivers/base/power/domain_governor.c struct generic_pm_domain *genpd = pd_to_genpd(pd); pd 209 drivers/base/power/domain_governor.c static bool default_power_down_ok(struct dev_pm_domain *pd) pd 211 drivers/base/power/domain_governor.c struct generic_pm_domain *genpd = pd_to_genpd(pd); pd 234 drivers/base/power/domain_governor.c while (!__default_power_down_ok(pd, genpd->state_idx)) { pd 252 drivers/base/power/domain_governor.c static bool cpu_power_down_ok(struct dev_pm_domain *pd) pd 254 drivers/base/power/domain_governor.c struct generic_pm_domain *genpd = pd_to_genpd(pd); pd 261 drivers/base/power/domain_governor.c if (!default_power_down_ok(pd)) pd 1383 drivers/block/drbd/drbd_worker.c static int pd_send_unplug_remote(struct drbd_peer_device *pd) pd 1385 drivers/block/drbd/drbd_worker.c struct drbd_socket *sock = &pd->connection->data; pd 1386 drivers/block/drbd/drbd_worker.c if (!drbd_prepare_command(pd, sock)) pd 1388 drivers/block/drbd/drbd_worker.c return drbd_send_command(pd, sock, P_UNPLUG_REMOTE, 0, NULL, 0); pd 243 drivers/block/paride/pd.c static struct pd_unit pd[PD_UNITS]; pd 404 drivers/block/paride/pd.c disk = pd[pd_queue].gd; pd 899 drivers/block/paride/pd.c p->first_minor = (disk - pd) << PD_BITS; pd 944 drivers/block/paride/pd.c struct pd_unit *disk = pd + unit; pd 965 drivers/block/paride/pd.c disk = pd; pd 974 drivers/block/paride/pd.c for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) { pd 987 drivers/block/paride/pd.c for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) { pd 1027 drivers/block/paride/pd.c for (unit = 0, disk = pd; unit < PD_UNITS; unit++, disk++) { pd 75 drivers/block/pktcdvd.c #define pkt_err(pd, fmt, ...) \ pd 76 drivers/block/pktcdvd.c pr_err("%s: " fmt, pd->name, ##__VA_ARGS__) pd 77 drivers/block/pktcdvd.c #define pkt_notice(pd, fmt, ...) \ pd 78 drivers/block/pktcdvd.c pr_notice("%s: " fmt, pd->name, ##__VA_ARGS__) pd 79 drivers/block/pktcdvd.c #define pkt_info(pd, fmt, ...) \ pd 80 drivers/block/pktcdvd.c pr_info("%s: " fmt, pd->name, ##__VA_ARGS__) pd 82 drivers/block/pktcdvd.c #define pkt_dbg(level, pd, fmt, ...) \ pd 86 drivers/block/pktcdvd.c pd->name, __func__, ##__VA_ARGS__); \ pd 88 drivers/block/pktcdvd.c pr_notice("%s: " fmt, pd->name, ##__VA_ARGS__); \ pd 111 drivers/block/pktcdvd.c static sector_t get_zone(sector_t sector, struct pktcdvd_device *pd) pd 113 drivers/block/pktcdvd.c return (sector + pd->offset) & ~(sector_t)(pd->settings.size - 1); pd 119 drivers/block/pktcdvd.c static struct pktcdvd_kobj* pkt_kobj_create(struct pktcdvd_device *pd, pd 130 drivers/block/pktcdvd.c p->pd = pd; pd 210 drivers/block/pktcdvd.c struct pktcdvd_device *pd = to_pktcdvdkobj(kobj)->pd; pd 214 drivers/block/pktcdvd.c n = sprintf(data, "%lu\n", pd->stats.pkt_started); pd 217 drivers/block/pktcdvd.c n = sprintf(data, "%lu\n", pd->stats.pkt_ended); pd 220 drivers/block/pktcdvd.c n = sprintf(data, "%lu\n", pd->stats.secs_w >> 1); pd 223 drivers/block/pktcdvd.c n = sprintf(data, "%lu\n", pd->stats.secs_r >> 1); pd 226 drivers/block/pktcdvd.c n = sprintf(data, "%lu\n", pd->stats.secs_rg >> 1); pd 229 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 230 drivers/block/pktcdvd.c v = pd->bio_queue_size; pd 231 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 235 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 236 drivers/block/pktcdvd.c v = pd->write_congestion_off; pd 237 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 241 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 242 drivers/block/pktcdvd.c v = pd->write_congestion_on; pd 243 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 270 drivers/block/pktcdvd.c struct pktcdvd_device *pd = to_pktcdvdkobj(kobj)->pd; pd 274 drivers/block/pktcdvd.c pd->stats.pkt_started = 0; pd 275 drivers/block/pktcdvd.c pd->stats.pkt_ended = 0; pd 276 drivers/block/pktcdvd.c pd->stats.secs_w = 0; pd 277 drivers/block/pktcdvd.c pd->stats.secs_rg = 0; pd 278 drivers/block/pktcdvd.c pd->stats.secs_r = 0; pd 282 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 283 drivers/block/pktcdvd.c pd->write_congestion_off = val; pd 284 drivers/block/pktcdvd.c init_write_congestion_marks(&pd->write_congestion_off, pd 285 drivers/block/pktcdvd.c &pd->write_congestion_on); pd 286 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 290 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 291 drivers/block/pktcdvd.c pd->write_congestion_on = val; pd 292 drivers/block/pktcdvd.c init_write_congestion_marks(&pd->write_congestion_off, pd 293 drivers/block/pktcdvd.c &pd->write_congestion_on); pd 294 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 314 drivers/block/pktcdvd.c static void pkt_sysfs_dev_new(struct pktcdvd_device *pd) pd 317 drivers/block/pktcdvd.c pd->dev = device_create(class_pktcdvd, NULL, MKDEV(0, 0), NULL, pd 318 drivers/block/pktcdvd.c "%s", pd->name); pd 319 drivers/block/pktcdvd.c if (IS_ERR(pd->dev)) pd 320 drivers/block/pktcdvd.c pd->dev = NULL; pd 322 drivers/block/pktcdvd.c if (pd->dev) { pd 323 drivers/block/pktcdvd.c pd->kobj_stat = pkt_kobj_create(pd, "stat", pd 324 drivers/block/pktcdvd.c &pd->dev->kobj, pd 326 drivers/block/pktcdvd.c pd->kobj_wqueue = pkt_kobj_create(pd, "write_queue", pd 327 drivers/block/pktcdvd.c &pd->dev->kobj, pd 332 drivers/block/pktcdvd.c static void pkt_sysfs_dev_remove(struct pktcdvd_device *pd) pd 334 drivers/block/pktcdvd.c pkt_kobj_remove(pd->kobj_stat); pd 335 drivers/block/pktcdvd.c pkt_kobj_remove(pd->kobj_wqueue); pd 337 drivers/block/pktcdvd.c device_unregister(pd->dev); pd 360 drivers/block/pktcdvd.c struct pktcdvd_device *pd = pkt_devs[idx]; pd 361 drivers/block/pktcdvd.c if (!pd) pd 364 drivers/block/pktcdvd.c pd->name, pd 365 drivers/block/pktcdvd.c MAJOR(pd->pkt_dev), MINOR(pd->pkt_dev), pd 366 drivers/block/pktcdvd.c MAJOR(pd->bdev->bd_dev), pd 367 drivers/block/pktcdvd.c MINOR(pd->bdev->bd_dev)); pd 473 drivers/block/pktcdvd.c static void pkt_debugfs_dev_new(struct pktcdvd_device *pd) pd 477 drivers/block/pktcdvd.c pd->dfs_d_root = debugfs_create_dir(pd->name, pkt_debugfs_root); pd 478 drivers/block/pktcdvd.c if (!pd->dfs_d_root) pd 481 drivers/block/pktcdvd.c pd->dfs_f_info = debugfs_create_file("info", 0444, pd 482 drivers/block/pktcdvd.c pd->dfs_d_root, pd, &debug_fops); pd 485 drivers/block/pktcdvd.c static void pkt_debugfs_dev_remove(struct pktcdvd_device *pd) pd 489 drivers/block/pktcdvd.c debugfs_remove(pd->dfs_f_info); pd 490 drivers/block/pktcdvd.c debugfs_remove(pd->dfs_d_root); pd 491 drivers/block/pktcdvd.c pd->dfs_f_info = NULL; pd 492 drivers/block/pktcdvd.c pd->dfs_d_root = NULL; pd 509 drivers/block/pktcdvd.c static void pkt_bio_finished(struct pktcdvd_device *pd) pd 511 drivers/block/pktcdvd.c BUG_ON(atomic_read(&pd->cdrw.pending_bios) <= 0); pd 512 drivers/block/pktcdvd.c if (atomic_dec_and_test(&pd->cdrw.pending_bios)) { pd 513 drivers/block/pktcdvd.c pkt_dbg(2, pd, "queue empty\n"); pd 514 drivers/block/pktcdvd.c atomic_set(&pd->iosched.attention, 1); pd 515 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 591 drivers/block/pktcdvd.c static void pkt_shrink_pktlist(struct pktcdvd_device *pd) pd 595 drivers/block/pktcdvd.c BUG_ON(!list_empty(&pd->cdrw.pkt_active_list)); pd 597 drivers/block/pktcdvd.c list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_free_list, list) { pd 600 drivers/block/pktcdvd.c INIT_LIST_HEAD(&pd->cdrw.pkt_free_list); pd 603 drivers/block/pktcdvd.c static int pkt_grow_pktlist(struct pktcdvd_device *pd, int nr_packets) pd 607 drivers/block/pktcdvd.c BUG_ON(!list_empty(&pd->cdrw.pkt_free_list)); pd 610 drivers/block/pktcdvd.c pkt = pkt_alloc_packet_data(pd->settings.size >> 2); pd 612 drivers/block/pktcdvd.c pkt_shrink_pktlist(pd); pd 616 drivers/block/pktcdvd.c pkt->pd = pd; pd 617 drivers/block/pktcdvd.c list_add(&pkt->list, &pd->cdrw.pkt_free_list); pd 631 drivers/block/pktcdvd.c static void pkt_rbtree_erase(struct pktcdvd_device *pd, struct pkt_rb_node *node) pd 633 drivers/block/pktcdvd.c rb_erase(&node->rb_node, &pd->bio_queue); pd 634 drivers/block/pktcdvd.c mempool_free(node, &pd->rb_pool); pd 635 drivers/block/pktcdvd.c pd->bio_queue_size--; pd 636 drivers/block/pktcdvd.c BUG_ON(pd->bio_queue_size < 0); pd 642 drivers/block/pktcdvd.c static struct pkt_rb_node *pkt_rbtree_find(struct pktcdvd_device *pd, sector_t s) pd 644 drivers/block/pktcdvd.c struct rb_node *n = pd->bio_queue.rb_node; pd 649 drivers/block/pktcdvd.c BUG_ON(pd->bio_queue_size > 0); pd 676 drivers/block/pktcdvd.c static void pkt_rbtree_insert(struct pktcdvd_device *pd, struct pkt_rb_node *node) pd 678 drivers/block/pktcdvd.c struct rb_node **p = &pd->bio_queue.rb_node; pd 692 drivers/block/pktcdvd.c rb_insert_color(&node->rb_node, &pd->bio_queue); pd 693 drivers/block/pktcdvd.c pd->bio_queue_size++; pd 700 drivers/block/pktcdvd.c static int pkt_generic_packet(struct pktcdvd_device *pd, struct packet_command *cgc) pd 702 drivers/block/pktcdvd.c struct request_queue *q = bdev_get_queue(pd->bdev); pd 725 drivers/block/pktcdvd.c blk_execute_rq(rq->q, pd->bdev->bd_disk, rq, 0); pd 748 drivers/block/pktcdvd.c static void pkt_dump_sense(struct pktcdvd_device *pd, pd 754 drivers/block/pktcdvd.c pkt_err(pd, "%*ph - sense %02x.%02x.%02x (%s)\n", pd 759 drivers/block/pktcdvd.c pkt_err(pd, "%*ph - no sense\n", CDROM_PACKET_SIZE, cgc->cmd); pd 765 drivers/block/pktcdvd.c static int pkt_flush_cache(struct pktcdvd_device *pd) pd 780 drivers/block/pktcdvd.c return pkt_generic_packet(pd, &cgc); pd 786 drivers/block/pktcdvd.c static noinline_for_stack int pkt_set_speed(struct pktcdvd_device *pd, pd 801 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 803 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 812 drivers/block/pktcdvd.c static void pkt_queue_bio(struct pktcdvd_device *pd, struct bio *bio) pd 814 drivers/block/pktcdvd.c spin_lock(&pd->iosched.lock); pd 816 drivers/block/pktcdvd.c bio_list_add(&pd->iosched.read_queue, bio); pd 818 drivers/block/pktcdvd.c bio_list_add(&pd->iosched.write_queue, bio); pd 819 drivers/block/pktcdvd.c spin_unlock(&pd->iosched.lock); pd 821 drivers/block/pktcdvd.c atomic_set(&pd->iosched.attention, 1); pd 822 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 841 drivers/block/pktcdvd.c static void pkt_iosched_process_queue(struct pktcdvd_device *pd) pd 844 drivers/block/pktcdvd.c if (atomic_read(&pd->iosched.attention) == 0) pd 846 drivers/block/pktcdvd.c atomic_set(&pd->iosched.attention, 0); pd 852 drivers/block/pktcdvd.c spin_lock(&pd->iosched.lock); pd 853 drivers/block/pktcdvd.c reads_queued = !bio_list_empty(&pd->iosched.read_queue); pd 854 drivers/block/pktcdvd.c writes_queued = !bio_list_empty(&pd->iosched.write_queue); pd 855 drivers/block/pktcdvd.c spin_unlock(&pd->iosched.lock); pd 860 drivers/block/pktcdvd.c if (pd->iosched.writing) { pd 862 drivers/block/pktcdvd.c spin_lock(&pd->iosched.lock); pd 863 drivers/block/pktcdvd.c bio = bio_list_peek(&pd->iosched.write_queue); pd 864 drivers/block/pktcdvd.c spin_unlock(&pd->iosched.lock); pd 866 drivers/block/pktcdvd.c pd->iosched.last_write)) pd 869 drivers/block/pktcdvd.c if (atomic_read(&pd->cdrw.pending_bios) > 0) { pd 870 drivers/block/pktcdvd.c pkt_dbg(2, pd, "write, waiting\n"); pd 873 drivers/block/pktcdvd.c pkt_flush_cache(pd); pd 874 drivers/block/pktcdvd.c pd->iosched.writing = 0; pd 878 drivers/block/pktcdvd.c if (atomic_read(&pd->cdrw.pending_bios) > 0) { pd 879 drivers/block/pktcdvd.c pkt_dbg(2, pd, "read, waiting\n"); pd 882 drivers/block/pktcdvd.c pd->iosched.writing = 1; pd 886 drivers/block/pktcdvd.c spin_lock(&pd->iosched.lock); pd 887 drivers/block/pktcdvd.c if (pd->iosched.writing) pd 888 drivers/block/pktcdvd.c bio = bio_list_pop(&pd->iosched.write_queue); pd 890 drivers/block/pktcdvd.c bio = bio_list_pop(&pd->iosched.read_queue); pd 891 drivers/block/pktcdvd.c spin_unlock(&pd->iosched.lock); pd 897 drivers/block/pktcdvd.c pd->iosched.successive_reads += pd 900 drivers/block/pktcdvd.c pd->iosched.successive_reads = 0; pd 901 drivers/block/pktcdvd.c pd->iosched.last_write = bio_end_sector(bio); pd 903 drivers/block/pktcdvd.c if (pd->iosched.successive_reads >= HI_SPEED_SWITCH) { pd 904 drivers/block/pktcdvd.c if (pd->read_speed == pd->write_speed) { pd 905 drivers/block/pktcdvd.c pd->read_speed = MAX_SPEED; pd 906 drivers/block/pktcdvd.c pkt_set_speed(pd, pd->write_speed, pd->read_speed); pd 909 drivers/block/pktcdvd.c if (pd->read_speed != pd->write_speed) { pd 910 drivers/block/pktcdvd.c pd->read_speed = pd->write_speed; pd 911 drivers/block/pktcdvd.c pkt_set_speed(pd, pd->write_speed, pd->read_speed); pd 915 drivers/block/pktcdvd.c atomic_inc(&pd->cdrw.pending_bios); pd 924 drivers/block/pktcdvd.c static int pkt_set_segment_merging(struct pktcdvd_device *pd, struct request_queue *q) pd 926 drivers/block/pktcdvd.c if ((pd->settings.size << 9) / CD_FRAMESIZE pd 931 drivers/block/pktcdvd.c clear_bit(PACKET_MERGE_SEGS, &pd->flags); pd 933 drivers/block/pktcdvd.c } else if ((pd->settings.size << 9) / PAGE_SIZE pd 939 drivers/block/pktcdvd.c set_bit(PACKET_MERGE_SEGS, &pd->flags); pd 942 drivers/block/pktcdvd.c pkt_err(pd, "cdrom max_phys_segments too small\n"); pd 950 drivers/block/pktcdvd.c struct pktcdvd_device *pd = pkt->pd; pd 951 drivers/block/pktcdvd.c BUG_ON(!pd); pd 953 drivers/block/pktcdvd.c pkt_dbg(2, pd, "bio=%p sec0=%llx sec=%llx err=%d\n", pd 961 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 963 drivers/block/pktcdvd.c pkt_bio_finished(pd); pd 969 drivers/block/pktcdvd.c struct pktcdvd_device *pd = pkt->pd; pd 970 drivers/block/pktcdvd.c BUG_ON(!pd); pd 972 drivers/block/pktcdvd.c pkt_dbg(2, pd, "id=%d, err=%d\n", pkt->id, bio->bi_status); pd 974 drivers/block/pktcdvd.c pd->stats.pkt_ended++; pd 976 drivers/block/pktcdvd.c pkt_bio_finished(pd); pd 979 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 985 drivers/block/pktcdvd.c static void pkt_gather_data(struct pktcdvd_device *pd, struct packet_data *pkt) pd 1006 drivers/block/pktcdvd.c pd->stats.secs_w += num_frames * (CD_FRAMESIZE >> 9); pd 1015 drivers/block/pktcdvd.c pkt_dbg(2, pd, "zone %llx cached\n", pd 1032 drivers/block/pktcdvd.c bio_set_dev(bio, pd->bdev); pd 1038 drivers/block/pktcdvd.c pkt_dbg(2, pd, "Adding frame %d, page:%p offs:%d\n", pd 1045 drivers/block/pktcdvd.c pkt_queue_bio(pd, bio); pd 1050 drivers/block/pktcdvd.c pkt_dbg(2, pd, "need %d frames for zone %llx\n", pd 1052 drivers/block/pktcdvd.c pd->stats.pkt_started++; pd 1053 drivers/block/pktcdvd.c pd->stats.secs_rg += frames_read * (CD_FRAMESIZE >> 9); pd 1060 drivers/block/pktcdvd.c static struct packet_data *pkt_get_packet_data(struct pktcdvd_device *pd, int zone) pd 1064 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_free_list, list) { pd 1065 drivers/block/pktcdvd.c if (pkt->sector == zone || pkt->list.next == &pd->cdrw.pkt_free_list) { pd 1076 drivers/block/pktcdvd.c static void pkt_put_packet_data(struct pktcdvd_device *pd, struct packet_data *pkt) pd 1079 drivers/block/pktcdvd.c list_add(&pkt->list, &pd->cdrw.pkt_free_list); pd 1081 drivers/block/pktcdvd.c list_add_tail(&pkt->list, &pd->cdrw.pkt_free_list); pd 1100 drivers/block/pktcdvd.c struct pktcdvd_device *pd = rq->rq_disk->private_data; pd 1106 drivers/block/pktcdvd.c pkt_bdev = bdget(kdev_t_to_nr(pd->pkt_dev)); pd 1126 drivers/block/pktcdvd.c bio_set_dev(pkt->bio, pd->bdev); pd 1151 drivers/block/pktcdvd.c pkt_dbg(2, pd, "pkt %2d : s=%6llx %s -> %s\n", pd 1162 drivers/block/pktcdvd.c static int pkt_handle_queue(struct pktcdvd_device *pd) pd 1171 drivers/block/pktcdvd.c atomic_set(&pd->scan_queue, 0); pd 1173 drivers/block/pktcdvd.c if (list_empty(&pd->cdrw.pkt_free_list)) { pd 1174 drivers/block/pktcdvd.c pkt_dbg(2, pd, "no pkt\n"); pd 1181 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 1182 drivers/block/pktcdvd.c first_node = pkt_rbtree_find(pd, pd->current_sector); pd 1184 drivers/block/pktcdvd.c n = rb_first(&pd->bio_queue); pd 1191 drivers/block/pktcdvd.c zone = get_zone(bio->bi_iter.bi_sector, pd); pd 1192 drivers/block/pktcdvd.c list_for_each_entry(p, &pd->cdrw.pkt_active_list, list) { pd 1202 drivers/block/pktcdvd.c n = rb_first(&pd->bio_queue); pd 1209 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 1211 drivers/block/pktcdvd.c pkt_dbg(2, pd, "no bio\n"); pd 1215 drivers/block/pktcdvd.c pkt = pkt_get_packet_data(pd, zone); pd 1217 drivers/block/pktcdvd.c pd->current_sector = zone + pd->settings.size; pd 1219 drivers/block/pktcdvd.c BUG_ON(pkt->frames != pd->settings.size >> 2); pd 1226 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 1227 drivers/block/pktcdvd.c pkt_dbg(2, pd, "looking for zone %llx\n", (unsigned long long)zone); pd 1228 drivers/block/pktcdvd.c while ((node = pkt_rbtree_find(pd, zone)) != NULL) { pd 1230 drivers/block/pktcdvd.c pkt_dbg(2, pd, "found zone=%llx\n", (unsigned long long) pd 1231 drivers/block/pktcdvd.c get_zone(bio->bi_iter.bi_sector, pd)); pd 1232 drivers/block/pktcdvd.c if (get_zone(bio->bi_iter.bi_sector, pd) != zone) pd 1234 drivers/block/pktcdvd.c pkt_rbtree_erase(pd, node); pd 1242 drivers/block/pktcdvd.c wakeup = (pd->write_congestion_on > 0 pd 1243 drivers/block/pktcdvd.c && pd->bio_queue_size <= pd->write_congestion_off); pd 1244 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 1246 drivers/block/pktcdvd.c clear_bdi_congested(pd->disk->queue->backing_dev_info, pd 1254 drivers/block/pktcdvd.c spin_lock(&pd->cdrw.active_list_lock); pd 1255 drivers/block/pktcdvd.c list_add(&pkt->list, &pd->cdrw.pkt_active_list); pd 1256 drivers/block/pktcdvd.c spin_unlock(&pd->cdrw.active_list_lock); pd 1265 drivers/block/pktcdvd.c static void pkt_start_write(struct pktcdvd_device *pd, struct packet_data *pkt) pd 1271 drivers/block/pktcdvd.c bio_set_dev(pkt->w_bio, pd->bdev); pd 1283 drivers/block/pktcdvd.c pkt_dbg(2, pd, "vcnt=%d\n", pkt->w_bio->bi_vcnt); pd 1294 drivers/block/pktcdvd.c pkt_dbg(2, pd, "Writing %d frames for zone %llx\n", pd 1297 drivers/block/pktcdvd.c if (test_bit(PACKET_MERGE_SEGS, &pd->flags) || (pkt->write_size < pkt->frames)) pd 1305 drivers/block/pktcdvd.c pkt_queue_bio(pd, pkt->w_bio); pd 1322 drivers/block/pktcdvd.c static void pkt_run_state_machine(struct pktcdvd_device *pd, struct packet_data *pkt) pd 1324 drivers/block/pktcdvd.c pkt_dbg(2, pd, "pkt %d\n", pkt->id); pd 1333 drivers/block/pktcdvd.c pkt_gather_data(pd, pkt); pd 1344 drivers/block/pktcdvd.c pkt_start_write(pd, pkt); pd 1361 drivers/block/pktcdvd.c pkt_start_write(pd, pkt); pd 1363 drivers/block/pktcdvd.c pkt_dbg(2, pd, "No recovery possible\n"); pd 1379 drivers/block/pktcdvd.c static void pkt_handle_packets(struct pktcdvd_device *pd) pd 1386 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 1389 drivers/block/pktcdvd.c pkt_run_state_machine(pd, pkt); pd 1396 drivers/block/pktcdvd.c spin_lock(&pd->cdrw.active_list_lock); pd 1397 drivers/block/pktcdvd.c list_for_each_entry_safe(pkt, next, &pd->cdrw.pkt_active_list, list) { pd 1400 drivers/block/pktcdvd.c pkt_put_packet_data(pd, pkt); pd 1402 drivers/block/pktcdvd.c atomic_set(&pd->scan_queue, 1); pd 1405 drivers/block/pktcdvd.c spin_unlock(&pd->cdrw.active_list_lock); pd 1408 drivers/block/pktcdvd.c static void pkt_count_states(struct pktcdvd_device *pd, int *states) pd 1416 drivers/block/pktcdvd.c spin_lock(&pd->cdrw.active_list_lock); pd 1417 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 1420 drivers/block/pktcdvd.c spin_unlock(&pd->cdrw.active_list_lock); pd 1429 drivers/block/pktcdvd.c struct pktcdvd_device *pd = foobar; pd 1442 drivers/block/pktcdvd.c add_wait_queue(&pd->wqueue, &wait); pd 1447 drivers/block/pktcdvd.c if (atomic_read(&pd->scan_queue) > 0) pd 1451 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 1457 drivers/block/pktcdvd.c if (atomic_read(&pd->iosched.attention) != 0) pd 1463 drivers/block/pktcdvd.c pkt_count_states(pd, states); pd 1464 drivers/block/pktcdvd.c pkt_dbg(2, pd, "i:%d ow:%d rw:%d ww:%d rec:%d fin:%d\n", pd 1470 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 1475 drivers/block/pktcdvd.c pkt_dbg(2, pd, "sleeping\n"); pd 1477 drivers/block/pktcdvd.c pkt_dbg(2, pd, "wake up\n"); pd 1482 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 1497 drivers/block/pktcdvd.c remove_wait_queue(&pd->wqueue, &wait); pd 1506 drivers/block/pktcdvd.c while (pkt_handle_queue(pd)) pd 1512 drivers/block/pktcdvd.c pkt_handle_packets(pd); pd 1517 drivers/block/pktcdvd.c pkt_iosched_process_queue(pd); pd 1523 drivers/block/pktcdvd.c static void pkt_print_settings(struct pktcdvd_device *pd) pd 1525 drivers/block/pktcdvd.c pkt_info(pd, "%s packets, %u blocks, Mode-%c disc\n", pd 1526 drivers/block/pktcdvd.c pd->settings.fp ? "Fixed" : "Variable", pd 1527 drivers/block/pktcdvd.c pd->settings.size >> 2, pd 1528 drivers/block/pktcdvd.c pd->settings.block_mode == 8 ? '1' : '2'); pd 1531 drivers/block/pktcdvd.c static int pkt_mode_sense(struct pktcdvd_device *pd, struct packet_command *cgc, int page_code, int page_control) pd 1540 drivers/block/pktcdvd.c return pkt_generic_packet(pd, cgc); pd 1543 drivers/block/pktcdvd.c static int pkt_mode_select(struct pktcdvd_device *pd, struct packet_command *cgc) pd 1552 drivers/block/pktcdvd.c return pkt_generic_packet(pd, cgc); pd 1555 drivers/block/pktcdvd.c static int pkt_get_disc_info(struct pktcdvd_device *pd, disc_information *di) pd 1566 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 1580 drivers/block/pktcdvd.c return pkt_generic_packet(pd, &cgc); pd 1583 drivers/block/pktcdvd.c static int pkt_get_track_info(struct pktcdvd_device *pd, __u16 track, __u8 type, track_information *ti) pd 1596 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 1607 drivers/block/pktcdvd.c return pkt_generic_packet(pd, &cgc); pd 1610 drivers/block/pktcdvd.c static noinline_for_stack int pkt_get_last_written(struct pktcdvd_device *pd, pd 1618 drivers/block/pktcdvd.c ret = pkt_get_disc_info(pd, &di); pd 1623 drivers/block/pktcdvd.c ret = pkt_get_track_info(pd, last_track, 1, &ti); pd 1630 drivers/block/pktcdvd.c ret = pkt_get_track_info(pd, last_track, 1, &ti); pd 1651 drivers/block/pktcdvd.c static noinline_for_stack int pkt_set_write_settings(struct pktcdvd_device *pd) pd 1660 drivers/block/pktcdvd.c if ((pd->mmc3_profile == 0x1a) || (pd->mmc3_profile == 0x12)) pd 1666 drivers/block/pktcdvd.c ret = pkt_mode_sense(pd, &cgc, GPMODE_WRITE_PARMS_PAGE, 0); pd 1668 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 1673 drivers/block/pktcdvd.c pd->mode_offset = (buffer[6] << 8) | (buffer[7] & 0xff); pd 1682 drivers/block/pktcdvd.c ret = pkt_mode_sense(pd, &cgc, GPMODE_WRITE_PARMS_PAGE, 0); pd 1684 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 1691 drivers/block/pktcdvd.c wp = (write_param_page *) &buffer[sizeof(struct mode_page_header) + pd->mode_offset]; pd 1693 drivers/block/pktcdvd.c wp->fp = pd->settings.fp; pd 1694 drivers/block/pktcdvd.c wp->track_mode = pd->settings.track_mode; pd 1695 drivers/block/pktcdvd.c wp->write_type = pd->settings.write_type; pd 1696 drivers/block/pktcdvd.c wp->data_block_type = pd->settings.block_mode; pd 1719 drivers/block/pktcdvd.c pkt_err(pd, "write mode wrong %d\n", wp->data_block_type); pd 1722 drivers/block/pktcdvd.c wp->packet_size = cpu_to_be32(pd->settings.size >> 2); pd 1725 drivers/block/pktcdvd.c ret = pkt_mode_select(pd, &cgc); pd 1727 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 1731 drivers/block/pktcdvd.c pkt_print_settings(pd); pd 1738 drivers/block/pktcdvd.c static int pkt_writable_track(struct pktcdvd_device *pd, track_information *ti) pd 1740 drivers/block/pktcdvd.c switch (pd->mmc3_profile) { pd 1764 drivers/block/pktcdvd.c pkt_err(pd, "bad state %d-%d-%d\n", ti->rt, ti->blank, ti->packet); pd 1771 drivers/block/pktcdvd.c static int pkt_writable_disc(struct pktcdvd_device *pd, disc_information *di) pd 1773 drivers/block/pktcdvd.c switch (pd->mmc3_profile) { pd 1782 drivers/block/pktcdvd.c pkt_dbg(2, pd, "Wrong disc profile (%x)\n", pd 1783 drivers/block/pktcdvd.c pd->mmc3_profile); pd 1792 drivers/block/pktcdvd.c pkt_notice(pd, "unknown disc - no track?\n"); pd 1797 drivers/block/pktcdvd.c pkt_err(pd, "wrong disc type (%x)\n", di->disc_type); pd 1802 drivers/block/pktcdvd.c pkt_notice(pd, "disc not erasable\n"); pd 1807 drivers/block/pktcdvd.c pkt_err(pd, "can't write to last track (reserved)\n"); pd 1814 drivers/block/pktcdvd.c static noinline_for_stack int pkt_probe_settings(struct pktcdvd_device *pd) pd 1825 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 1826 drivers/block/pktcdvd.c pd->mmc3_profile = ret ? 0xffff : buf[6] << 8 | buf[7]; pd 1831 drivers/block/pktcdvd.c ret = pkt_get_disc_info(pd, &di); pd 1833 drivers/block/pktcdvd.c pkt_err(pd, "failed get_disc\n"); pd 1837 drivers/block/pktcdvd.c if (!pkt_writable_disc(pd, &di)) pd 1840 drivers/block/pktcdvd.c pd->type = di.erasable ? PACKET_CDRW : PACKET_CDR; pd 1843 drivers/block/pktcdvd.c ret = pkt_get_track_info(pd, track, 1, &ti); pd 1845 drivers/block/pktcdvd.c pkt_err(pd, "failed get_track\n"); pd 1849 drivers/block/pktcdvd.c if (!pkt_writable_track(pd, &ti)) { pd 1850 drivers/block/pktcdvd.c pkt_err(pd, "can't write to this track\n"); pd 1858 drivers/block/pktcdvd.c pd->settings.size = be32_to_cpu(ti.fixed_packet_size) << 2; pd 1859 drivers/block/pktcdvd.c if (pd->settings.size == 0) { pd 1860 drivers/block/pktcdvd.c pkt_notice(pd, "detected zero packet size!\n"); pd 1863 drivers/block/pktcdvd.c if (pd->settings.size > PACKET_MAX_SECTORS) { pd 1864 drivers/block/pktcdvd.c pkt_err(pd, "packet size is too big\n"); pd 1867 drivers/block/pktcdvd.c pd->settings.fp = ti.fp; pd 1868 drivers/block/pktcdvd.c pd->offset = (be32_to_cpu(ti.track_start) << 2) & (pd->settings.size - 1); pd 1871 drivers/block/pktcdvd.c pd->nwa = be32_to_cpu(ti.next_writable); pd 1872 drivers/block/pktcdvd.c set_bit(PACKET_NWA_VALID, &pd->flags); pd 1881 drivers/block/pktcdvd.c pd->lra = be32_to_cpu(ti.last_rec_address); pd 1882 drivers/block/pktcdvd.c set_bit(PACKET_LRA_VALID, &pd->flags); pd 1884 drivers/block/pktcdvd.c pd->lra = 0xffffffff; pd 1885 drivers/block/pktcdvd.c set_bit(PACKET_LRA_VALID, &pd->flags); pd 1891 drivers/block/pktcdvd.c pd->settings.link_loss = 7; pd 1892 drivers/block/pktcdvd.c pd->settings.write_type = 0; /* packet */ pd 1893 drivers/block/pktcdvd.c pd->settings.track_mode = ti.track_mode; pd 1900 drivers/block/pktcdvd.c pd->settings.block_mode = PACKET_BLOCK_MODE1; pd 1903 drivers/block/pktcdvd.c pd->settings.block_mode = PACKET_BLOCK_MODE2; pd 1906 drivers/block/pktcdvd.c pkt_err(pd, "unknown data mode\n"); pd 1915 drivers/block/pktcdvd.c static noinline_for_stack int pkt_write_caching(struct pktcdvd_device *pd, pd 1925 drivers/block/pktcdvd.c cgc.buflen = pd->mode_offset + 12; pd 1932 drivers/block/pktcdvd.c ret = pkt_mode_sense(pd, &cgc, GPMODE_WCACHING_PAGE, 0); pd 1936 drivers/block/pktcdvd.c buf[pd->mode_offset + 10] |= (!!set << 2); pd 1939 drivers/block/pktcdvd.c ret = pkt_mode_select(pd, &cgc); pd 1941 drivers/block/pktcdvd.c pkt_err(pd, "write caching control failed\n"); pd 1942 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 1944 drivers/block/pktcdvd.c pkt_notice(pd, "enabled write caching\n"); pd 1948 drivers/block/pktcdvd.c static int pkt_lock_door(struct pktcdvd_device *pd, int lockflag) pd 1955 drivers/block/pktcdvd.c return pkt_generic_packet(pd, &cgc); pd 1961 drivers/block/pktcdvd.c static noinline_for_stack int pkt_get_max_speed(struct pktcdvd_device *pd, pd 1970 drivers/block/pktcdvd.c cap_buf = &buf[sizeof(struct mode_page_header) + pd->mode_offset]; pd 1974 drivers/block/pktcdvd.c ret = pkt_mode_sense(pd, &cgc, GPMODE_CAPABILITIES_PAGE, 0); pd 1976 drivers/block/pktcdvd.c cgc.buflen = pd->mode_offset + cap_buf[1] + 2 + pd 1978 drivers/block/pktcdvd.c ret = pkt_mode_sense(pd, &cgc, GPMODE_CAPABILITIES_PAGE, 0); pd 1980 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 2022 drivers/block/pktcdvd.c static noinline_for_stack int pkt_media_speed(struct pktcdvd_device *pd, pd 2037 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 2039 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 2052 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 2054 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 2059 drivers/block/pktcdvd.c pkt_notice(pd, "disc type is not CD-RW\n"); pd 2063 drivers/block/pktcdvd.c pkt_notice(pd, "A1 values on media are not valid, maybe not CDRW?\n"); pd 2083 drivers/block/pktcdvd.c pkt_notice(pd, "unknown disc sub-type %d\n", st); pd 2087 drivers/block/pktcdvd.c pkt_info(pd, "maximum media speed: %d\n", *speed); pd 2090 drivers/block/pktcdvd.c pkt_notice(pd, "unknown speed %d for sub-type %d\n", sp, st); pd 2095 drivers/block/pktcdvd.c static noinline_for_stack int pkt_perform_opc(struct pktcdvd_device *pd) pd 2101 drivers/block/pktcdvd.c pkt_dbg(2, pd, "Performing OPC\n"); pd 2108 drivers/block/pktcdvd.c ret = pkt_generic_packet(pd, &cgc); pd 2110 drivers/block/pktcdvd.c pkt_dump_sense(pd, &cgc); pd 2114 drivers/block/pktcdvd.c static int pkt_open_write(struct pktcdvd_device *pd) pd 2119 drivers/block/pktcdvd.c ret = pkt_probe_settings(pd); pd 2121 drivers/block/pktcdvd.c pkt_dbg(2, pd, "failed probe\n"); pd 2125 drivers/block/pktcdvd.c ret = pkt_set_write_settings(pd); pd 2127 drivers/block/pktcdvd.c pkt_dbg(1, pd, "failed saving write settings\n"); pd 2131 drivers/block/pktcdvd.c pkt_write_caching(pd, USE_WCACHING); pd 2133 drivers/block/pktcdvd.c ret = pkt_get_max_speed(pd, &write_speed); pd 2136 drivers/block/pktcdvd.c switch (pd->mmc3_profile) { pd 2140 drivers/block/pktcdvd.c pkt_dbg(1, pd, "write speed %ukB/s\n", write_speed); pd 2143 drivers/block/pktcdvd.c ret = pkt_media_speed(pd, &media_write_speed); pd 2147 drivers/block/pktcdvd.c pkt_dbg(1, pd, "write speed %ux\n", write_speed / 176); pd 2152 drivers/block/pktcdvd.c ret = pkt_set_speed(pd, write_speed, read_speed); pd 2154 drivers/block/pktcdvd.c pkt_dbg(1, pd, "couldn't set write speed\n"); pd 2157 drivers/block/pktcdvd.c pd->write_speed = write_speed; pd 2158 drivers/block/pktcdvd.c pd->read_speed = read_speed; pd 2160 drivers/block/pktcdvd.c ret = pkt_perform_opc(pd); pd 2162 drivers/block/pktcdvd.c pkt_dbg(1, pd, "Optimum Power Calibration failed\n"); pd 2171 drivers/block/pktcdvd.c static int pkt_open_dev(struct pktcdvd_device *pd, fmode_t write) pd 2182 drivers/block/pktcdvd.c bdget(pd->bdev->bd_dev); pd 2183 drivers/block/pktcdvd.c ret = blkdev_get(pd->bdev, FMODE_READ | FMODE_EXCL, pd); pd 2187 drivers/block/pktcdvd.c ret = pkt_get_last_written(pd, &lba); pd 2189 drivers/block/pktcdvd.c pkt_err(pd, "pkt_get_last_written failed\n"); pd 2193 drivers/block/pktcdvd.c set_capacity(pd->disk, lba << 2); pd 2194 drivers/block/pktcdvd.c set_capacity(pd->bdev->bd_disk, lba << 2); pd 2195 drivers/block/pktcdvd.c bd_set_size(pd->bdev, (loff_t)lba << 11); pd 2197 drivers/block/pktcdvd.c q = bdev_get_queue(pd->bdev); pd 2199 drivers/block/pktcdvd.c ret = pkt_open_write(pd); pd 2206 drivers/block/pktcdvd.c blk_queue_max_hw_sectors(q, pd->settings.size); pd 2207 drivers/block/pktcdvd.c set_bit(PACKET_WRITABLE, &pd->flags); pd 2209 drivers/block/pktcdvd.c pkt_set_speed(pd, MAX_SPEED, MAX_SPEED); pd 2210 drivers/block/pktcdvd.c clear_bit(PACKET_WRITABLE, &pd->flags); pd 2213 drivers/block/pktcdvd.c ret = pkt_set_segment_merging(pd, q); pd 2218 drivers/block/pktcdvd.c if (!pkt_grow_pktlist(pd, CONFIG_CDROM_PKTCDVD_BUFFERS)) { pd 2219 drivers/block/pktcdvd.c pkt_err(pd, "not enough memory for buffers\n"); pd 2223 drivers/block/pktcdvd.c pkt_info(pd, "%lukB available on disc\n", lba << 1); pd 2229 drivers/block/pktcdvd.c blkdev_put(pd->bdev, FMODE_READ | FMODE_EXCL); pd 2238 drivers/block/pktcdvd.c static void pkt_release_dev(struct pktcdvd_device *pd, int flush) pd 2240 drivers/block/pktcdvd.c if (flush && pkt_flush_cache(pd)) pd 2241 drivers/block/pktcdvd.c pkt_dbg(1, pd, "not flushing cache\n"); pd 2243 drivers/block/pktcdvd.c pkt_lock_door(pd, 0); pd 2245 drivers/block/pktcdvd.c pkt_set_speed(pd, MAX_SPEED, MAX_SPEED); pd 2246 drivers/block/pktcdvd.c blkdev_put(pd->bdev, FMODE_READ | FMODE_EXCL); pd 2248 drivers/block/pktcdvd.c pkt_shrink_pktlist(pd); pd 2262 drivers/block/pktcdvd.c struct pktcdvd_device *pd = NULL; pd 2267 drivers/block/pktcdvd.c pd = pkt_find_dev_from_minor(MINOR(bdev->bd_dev)); pd 2268 drivers/block/pktcdvd.c if (!pd) { pd 2272 drivers/block/pktcdvd.c BUG_ON(pd->refcnt < 0); pd 2274 drivers/block/pktcdvd.c pd->refcnt++; pd 2275 drivers/block/pktcdvd.c if (pd->refcnt > 1) { pd 2277 drivers/block/pktcdvd.c !test_bit(PACKET_WRITABLE, &pd->flags)) { pd 2282 drivers/block/pktcdvd.c ret = pkt_open_dev(pd, mode & FMODE_WRITE); pd 2297 drivers/block/pktcdvd.c pd->refcnt--; pd 2306 drivers/block/pktcdvd.c struct pktcdvd_device *pd = disk->private_data; pd 2310 drivers/block/pktcdvd.c pd->refcnt--; pd 2311 drivers/block/pktcdvd.c BUG_ON(pd->refcnt < 0); pd 2312 drivers/block/pktcdvd.c if (pd->refcnt == 0) { pd 2313 drivers/block/pktcdvd.c int flush = test_bit(PACKET_WRITABLE, &pd->flags); pd 2314 drivers/block/pktcdvd.c pkt_release_dev(pd, flush); pd 2324 drivers/block/pktcdvd.c struct pktcdvd_device *pd = psd->pd; pd 2330 drivers/block/pktcdvd.c pkt_bio_finished(pd); pd 2333 drivers/block/pktcdvd.c static void pkt_make_request_read(struct pktcdvd_device *pd, struct bio *bio) pd 2338 drivers/block/pktcdvd.c psd->pd = pd; pd 2340 drivers/block/pktcdvd.c bio_set_dev(cloned_bio, pd->bdev); pd 2343 drivers/block/pktcdvd.c pd->stats.secs_r += bio_sectors(bio); pd 2344 drivers/block/pktcdvd.c pkt_queue_bio(pd, cloned_bio); pd 2349 drivers/block/pktcdvd.c struct pktcdvd_device *pd = q->queuedata; pd 2355 drivers/block/pktcdvd.c zone = get_zone(bio->bi_iter.bi_sector, pd); pd 2361 drivers/block/pktcdvd.c spin_lock(&pd->cdrw.active_list_lock); pd 2363 drivers/block/pktcdvd.c list_for_each_entry(pkt, &pd->cdrw.pkt_active_list, list) { pd 2374 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 2377 drivers/block/pktcdvd.c spin_unlock(&pd->cdrw.active_list_lock); pd 2385 drivers/block/pktcdvd.c spin_unlock(&pd->cdrw.active_list_lock); pd 2392 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 2393 drivers/block/pktcdvd.c if (pd->write_congestion_on > 0 pd 2394 drivers/block/pktcdvd.c && pd->bio_queue_size >= pd->write_congestion_on) { pd 2397 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 2399 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 2400 drivers/block/pktcdvd.c } while(pd->bio_queue_size > pd->write_congestion_off); pd 2402 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 2407 drivers/block/pktcdvd.c node = mempool_alloc(&pd->rb_pool, GFP_NOIO); pd 2409 drivers/block/pktcdvd.c spin_lock(&pd->lock); pd 2410 drivers/block/pktcdvd.c BUG_ON(pd->bio_queue_size < 0); pd 2411 drivers/block/pktcdvd.c was_empty = (pd->bio_queue_size == 0); pd 2412 drivers/block/pktcdvd.c pkt_rbtree_insert(pd, node); pd 2413 drivers/block/pktcdvd.c spin_unlock(&pd->lock); pd 2418 drivers/block/pktcdvd.c atomic_set(&pd->scan_queue, 1); pd 2421 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 2422 drivers/block/pktcdvd.c } else if (!list_empty(&pd->cdrw.pkt_free_list) && !blocked_bio) { pd 2427 drivers/block/pktcdvd.c wake_up(&pd->wqueue); pd 2433 drivers/block/pktcdvd.c struct pktcdvd_device *pd; pd 2439 drivers/block/pktcdvd.c pd = q->queuedata; pd 2440 drivers/block/pktcdvd.c if (!pd) { pd 2445 drivers/block/pktcdvd.c pkt_dbg(2, pd, "start = %6llx stop = %6llx\n", pd 2453 drivers/block/pktcdvd.c pkt_make_request_read(pd, bio); pd 2457 drivers/block/pktcdvd.c if (!test_bit(PACKET_WRITABLE, &pd->flags)) { pd 2458 drivers/block/pktcdvd.c pkt_notice(pd, "WRITE for ro device (%llu)\n", pd 2464 drivers/block/pktcdvd.c pkt_err(pd, "wrong bio size\n"); pd 2469 drivers/block/pktcdvd.c sector_t zone = get_zone(bio->bi_iter.bi_sector, pd); pd 2470 drivers/block/pktcdvd.c sector_t last_zone = get_zone(bio_end_sector(bio) - 1, pd); pd 2473 drivers/block/pktcdvd.c BUG_ON(last_zone != zone + pd->settings.size); pd 2492 drivers/block/pktcdvd.c static void pkt_init_queue(struct pktcdvd_device *pd) pd 2494 drivers/block/pktcdvd.c struct request_queue *q = pd->disk->queue; pd 2499 drivers/block/pktcdvd.c q->queuedata = pd; pd 2504 drivers/block/pktcdvd.c struct pktcdvd_device *pd = m->private; pd 2509 drivers/block/pktcdvd.c seq_printf(m, "Writer %s mapped to %s:\n", pd->name, pd 2510 drivers/block/pktcdvd.c bdevname(pd->bdev, bdev_buf)); pd 2513 drivers/block/pktcdvd.c seq_printf(m, "\tpacket size:\t\t%dkB\n", pd->settings.size / 2); pd 2515 drivers/block/pktcdvd.c if (pd->settings.write_type == 0) pd 2521 drivers/block/pktcdvd.c seq_printf(m, "\tpacket type:\t\t%s\n", pd->settings.fp ? "Fixed" : "Variable"); pd 2522 drivers/block/pktcdvd.c seq_printf(m, "\tlink loss:\t\t%d\n", pd->settings.link_loss); pd 2524 drivers/block/pktcdvd.c seq_printf(m, "\ttrack mode:\t\t%d\n", pd->settings.track_mode); pd 2526 drivers/block/pktcdvd.c if (pd->settings.block_mode == PACKET_BLOCK_MODE1) pd 2528 drivers/block/pktcdvd.c else if (pd->settings.block_mode == PACKET_BLOCK_MODE2) pd 2535 drivers/block/pktcdvd.c seq_printf(m, "\tpackets started:\t%lu\n", pd->stats.pkt_started); pd 2536 drivers/block/pktcdvd.c seq_printf(m, "\tpackets ended:\t\t%lu\n", pd->stats.pkt_ended); pd 2537 drivers/block/pktcdvd.c seq_printf(m, "\twritten:\t\t%lukB\n", pd->stats.secs_w >> 1); pd 2538 drivers/block/pktcdvd.c seq_printf(m, "\tread gather:\t\t%lukB\n", pd->stats.secs_rg >> 1); pd 2539 drivers/block/pktcdvd.c seq_printf(m, "\tread:\t\t\t%lukB\n", pd->stats.secs_r >> 1); pd 2542 drivers/block/pktcdvd.c seq_printf(m, "\treference count:\t%d\n", pd->refcnt); pd 2543 drivers/block/pktcdvd.c seq_printf(m, "\tflags:\t\t\t0x%lx\n", pd->flags); pd 2544 drivers/block/pktcdvd.c seq_printf(m, "\tread speed:\t\t%ukB/s\n", pd->read_speed); pd 2545 drivers/block/pktcdvd.c seq_printf(m, "\twrite speed:\t\t%ukB/s\n", pd->write_speed); pd 2546 drivers/block/pktcdvd.c seq_printf(m, "\tstart offset:\t\t%lu\n", pd->offset); pd 2547 drivers/block/pktcdvd.c seq_printf(m, "\tmode page offset:\t%u\n", pd->mode_offset); pd 2550 drivers/block/pktcdvd.c seq_printf(m, "\tbios queued:\t\t%d\n", pd->bio_queue_size); pd 2551 drivers/block/pktcdvd.c seq_printf(m, "\tbios pending:\t\t%d\n", atomic_read(&pd->cdrw.pending_bios)); pd 2552 drivers/block/pktcdvd.c seq_printf(m, "\tcurrent sector:\t\t0x%llx\n", (unsigned long long)pd->current_sector); pd 2554 drivers/block/pktcdvd.c pkt_count_states(pd, states); pd 2559 drivers/block/pktcdvd.c pd->write_congestion_off, pd 2560 drivers/block/pktcdvd.c pd->write_congestion_on); pd 2564 drivers/block/pktcdvd.c static int pkt_new_dev(struct pktcdvd_device *pd, dev_t dev) pd 2571 drivers/block/pktcdvd.c if (pd->pkt_dev == dev) { pd 2572 drivers/block/pktcdvd.c pkt_err(pd, "recursive setup not allowed\n"); pd 2580 drivers/block/pktcdvd.c pkt_err(pd, "%s already setup\n", pd 2585 drivers/block/pktcdvd.c pkt_err(pd, "can't chain pktcdvd devices\n"); pd 2604 drivers/block/pktcdvd.c pd->bdev = bdev; pd 2607 drivers/block/pktcdvd.c pkt_init_queue(pd); pd 2609 drivers/block/pktcdvd.c atomic_set(&pd->cdrw.pending_bios, 0); pd 2610 drivers/block/pktcdvd.c pd->cdrw.thread = kthread_run(kcdrwd, pd, "%s", pd->name); pd 2611 drivers/block/pktcdvd.c if (IS_ERR(pd->cdrw.thread)) { pd 2612 drivers/block/pktcdvd.c pkt_err(pd, "can't start kernel thread\n"); pd 2617 drivers/block/pktcdvd.c proc_create_single_data(pd->name, 0, pkt_proc, pkt_seq_show, pd); pd 2618 drivers/block/pktcdvd.c pkt_dbg(1, pd, "writer mapped to %s\n", bdevname(bdev, b)); pd 2630 drivers/block/pktcdvd.c struct pktcdvd_device *pd = bdev->bd_disk->private_data; pd 2633 drivers/block/pktcdvd.c pkt_dbg(2, pd, "cmd %x, dev %d:%d\n", pd 2643 drivers/block/pktcdvd.c if (pd->refcnt == 1) pd 2644 drivers/block/pktcdvd.c pkt_lock_door(pd, 0); pd 2654 drivers/block/pktcdvd.c ret = __blkdev_driver_ioctl(pd->bdev, mode, cmd, arg); pd 2658 drivers/block/pktcdvd.c pkt_dbg(2, pd, "Unknown ioctl (%x)\n", cmd); pd 2669 drivers/block/pktcdvd.c struct pktcdvd_device *pd = disk->private_data; pd 2672 drivers/block/pktcdvd.c if (!pd) pd 2674 drivers/block/pktcdvd.c if (!pd->bdev) pd 2676 drivers/block/pktcdvd.c attached_disk = pd->bdev->bd_disk; pd 2702 drivers/block/pktcdvd.c struct pktcdvd_device *pd; pd 2716 drivers/block/pktcdvd.c pd = kzalloc(sizeof(struct pktcdvd_device), GFP_KERNEL); pd 2717 drivers/block/pktcdvd.c if (!pd) pd 2720 drivers/block/pktcdvd.c ret = mempool_init_kmalloc_pool(&pd->rb_pool, PKT_RB_POOL_SIZE, pd 2725 drivers/block/pktcdvd.c INIT_LIST_HEAD(&pd->cdrw.pkt_free_list); pd 2726 drivers/block/pktcdvd.c INIT_LIST_HEAD(&pd->cdrw.pkt_active_list); pd 2727 drivers/block/pktcdvd.c spin_lock_init(&pd->cdrw.active_list_lock); pd 2729 drivers/block/pktcdvd.c spin_lock_init(&pd->lock); pd 2730 drivers/block/pktcdvd.c spin_lock_init(&pd->iosched.lock); pd 2731 drivers/block/pktcdvd.c bio_list_init(&pd->iosched.read_queue); pd 2732 drivers/block/pktcdvd.c bio_list_init(&pd->iosched.write_queue); pd 2733 drivers/block/pktcdvd.c sprintf(pd->name, DRIVER_NAME"%d", idx); pd 2734 drivers/block/pktcdvd.c init_waitqueue_head(&pd->wqueue); pd 2735 drivers/block/pktcdvd.c pd->bio_queue = RB_ROOT; pd 2737 drivers/block/pktcdvd.c pd->write_congestion_on = write_congestion_on; pd 2738 drivers/block/pktcdvd.c pd->write_congestion_off = write_congestion_off; pd 2744 drivers/block/pktcdvd.c pd->disk = disk; pd 2749 drivers/block/pktcdvd.c strcpy(disk->disk_name, pd->name); pd 2751 drivers/block/pktcdvd.c disk->private_data = pd; pd 2756 drivers/block/pktcdvd.c pd->pkt_dev = MKDEV(pktdev_major, idx); pd 2757 drivers/block/pktcdvd.c ret = pkt_new_dev(pd, dev); pd 2762 drivers/block/pktcdvd.c disk->events = pd->bdev->bd_disk->events; pd 2766 drivers/block/pktcdvd.c pkt_sysfs_dev_new(pd); pd 2767 drivers/block/pktcdvd.c pkt_debugfs_dev_new(pd); pd 2769 drivers/block/pktcdvd.c pkt_devs[idx] = pd; pd 2771 drivers/block/pktcdvd.c *pkt_dev = pd->pkt_dev; pd 2779 drivers/block/pktcdvd.c mempool_exit(&pd->rb_pool); pd 2780 drivers/block/pktcdvd.c kfree(pd); pd 2792 drivers/block/pktcdvd.c struct pktcdvd_device *pd; pd 2799 drivers/block/pktcdvd.c pd = pkt_devs[idx]; pd 2800 drivers/block/pktcdvd.c if (pd && (pd->pkt_dev == pkt_dev)) pd 2809 drivers/block/pktcdvd.c if (pd->refcnt > 0) { pd 2813 drivers/block/pktcdvd.c if (!IS_ERR(pd->cdrw.thread)) pd 2814 drivers/block/pktcdvd.c kthread_stop(pd->cdrw.thread); pd 2818 drivers/block/pktcdvd.c pkt_debugfs_dev_remove(pd); pd 2819 drivers/block/pktcdvd.c pkt_sysfs_dev_remove(pd); pd 2821 drivers/block/pktcdvd.c blkdev_put(pd->bdev, FMODE_READ | FMODE_NDELAY); pd 2823 drivers/block/pktcdvd.c remove_proc_entry(pd->name, pkt_proc); pd 2824 drivers/block/pktcdvd.c pkt_dbg(1, pd, "writer unmapped\n"); pd 2826 drivers/block/pktcdvd.c del_gendisk(pd->disk); pd 2827 drivers/block/pktcdvd.c blk_cleanup_queue(pd->disk->queue); pd 2828 drivers/block/pktcdvd.c put_disk(pd->disk); pd 2830 drivers/block/pktcdvd.c mempool_exit(&pd->rb_pool); pd 2831 drivers/block/pktcdvd.c kfree(pd); pd 2843 drivers/block/pktcdvd.c struct pktcdvd_device *pd; pd 2847 drivers/block/pktcdvd.c pd = pkt_find_dev_from_minor(ctrl_cmd->dev_index); pd 2848 drivers/block/pktcdvd.c if (pd) { pd 2849 drivers/block/pktcdvd.c ctrl_cmd->dev = new_encode_dev(pd->bdev->bd_dev); pd 2850 drivers/block/pktcdvd.c ctrl_cmd->pkt_dev = new_encode_dev(pd->pkt_dev); pd 85 drivers/cdrom/gdrom.c static struct platform_device *pd; pd 850 drivers/cdrom/gdrom.c pd = platform_device_register_simple(GDROM_DEV_NAME, -1, NULL, 0); pd 851 drivers/cdrom/gdrom.c if (IS_ERR(pd)) { pd 853 drivers/cdrom/gdrom.c return PTR_ERR(pd); pd 860 drivers/cdrom/gdrom.c platform_device_unregister(pd); pd 83 drivers/clk/davinci/psc.c u32 pd; pd 121 drivers/clk/davinci/psc.c regmap_read(lpsc->regmap, PDSTAT(lpsc->pd), &pdstat); pd 123 drivers/clk/davinci/psc.c regmap_write_bits(lpsc->regmap, PDCTL(lpsc->pd), PDCTL_NEXT, pd 126 drivers/clk/davinci/psc.c regmap_write(lpsc->regmap, PTCMD, BIT(lpsc->pd)); pd 129 drivers/clk/davinci/psc.c epcpr & BIT(lpsc->pd), 0, 0); pd 131 drivers/clk/davinci/psc.c regmap_write_bits(lpsc->regmap, PDCTL(lpsc->pd), PDCTL_EPCGOOD, pd 134 drivers/clk/davinci/psc.c regmap_write(lpsc->regmap, PTCMD, BIT(lpsc->pd)); pd 138 drivers/clk/davinci/psc.c !(ptstat & BIT(lpsc->pd)), 0, 0); pd 236 drivers/clk/davinci/psc.c u32 md, u32 pd, u32 flags) pd 263 drivers/clk/davinci/psc.c lpsc->pd = pd; pd 410 drivers/clk/davinci/psc.c regmap, info->md, info->pd, pd 65 drivers/clk/davinci/psc.h u32 pd; pd 75 drivers/clk/davinci/psc.h .pd = (d), \ pd 58 drivers/clk/imx/clk-pllv1.c unsigned int mfi, mfn, mfd, pd; pd 77 drivers/clk/imx/clk-pllv1.c pd = (reg >> 26) & 0xf; pd 96 drivers/clk/imx/clk-pllv1.c rate /= pd + 1; pd 1526 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 1535 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 1544 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 1553 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 1562 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 1571 drivers/clk/qcom/camcc-sdm845.c .pd = { pd 572 drivers/clk/qcom/dispcc-sdm845.c .pd = { pd 3250 drivers/clk/qcom/gcc-apq8084.c .pd = { pd 3258 drivers/clk/qcom/gcc-apq8084.c .pd = { pd 3266 drivers/clk/qcom/gcc-apq8084.c .pd = { pd 3274 drivers/clk/qcom/gcc-apq8084.c .pd = { pd 3052 drivers/clk/qcom/gcc-msm8916.c .pd = { pd 3060 drivers/clk/qcom/gcc-msm8916.c .pd = { pd 3068 drivers/clk/qcom/gcc-msm8916.c .pd = { pd 3076 drivers/clk/qcom/gcc-msm8916.c .pd = { pd 3084 drivers/clk/qcom/gcc-msm8916.c .pd = { pd 2427 drivers/clk/qcom/gcc-msm8974.c .pd = { pd 3207 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3216 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3225 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3234 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3243 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3251 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3259 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3267 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 3275 drivers/clk/qcom/gcc-msm8996.c .pd = { pd 2644 drivers/clk/qcom/gcc-msm8998.c .pd = { pd 2654 drivers/clk/qcom/gcc-msm8998.c .pd = { pd 2664 drivers/clk/qcom/gcc-msm8998.c .pd = { pd 2228 drivers/clk/qcom/gcc-sdm660.c .pd = { pd 2238 drivers/clk/qcom/gcc-sdm660.c .pd = { pd 2248 drivers/clk/qcom/gcc-sdm660.c .pd = { pd 3200 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3209 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3218 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3227 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3236 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3245 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3254 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3263 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3272 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3281 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3290 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3299 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 3308 drivers/clk/qcom/gcc-sdm845.c .pd = { pd 43 drivers/clk/qcom/gdsc.c #define domain_to_gdsc(domain) container_of(domain, struct gdsc, pd) pd 145 drivers/clk/qcom/gdsc.c WARN(ret, "%s status stuck at 'o%s'", sc->pd.name, status ? "ff" : "n"); pd 332 drivers/clk/qcom/gdsc.c gdsc_enable(&sc->pd); pd 337 drivers/clk/qcom/gdsc.c gdsc_enable(&sc->pd); pd 339 drivers/clk/qcom/gdsc.c sc->pd.flags |= GENPD_FLAG_ALWAYS_ON; pd 347 drivers/clk/qcom/gdsc.c if (!sc->pd.power_off) pd 348 drivers/clk/qcom/gdsc.c sc->pd.power_off = gdsc_disable; pd 349 drivers/clk/qcom/gdsc.c if (!sc->pd.power_on) pd 350 drivers/clk/qcom/gdsc.c sc->pd.power_on = gdsc_enable; pd 351 drivers/clk/qcom/gdsc.c pm_genpd_init(&sc->pd, NULL, !on); pd 383 drivers/clk/qcom/gdsc.c data->domains[i] = &scs[i]->pd; pd 391 drivers/clk/qcom/gdsc.c pm_genpd_add_subdomain(scs[i]->parent, &scs[i]->pd); pd 409 drivers/clk/qcom/gdsc.c pm_genpd_remove_subdomain(scs[i]->parent, &scs[i]->pd); pd 29 drivers/clk/qcom/gdsc.h struct generic_pm_domain pd; pd 127 drivers/clk/qcom/gpucc-sdm845.c .pd = { pd 162 drivers/clk/qcom/gpucc-sdm845.c .pd = { pd 3059 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3067 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3075 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3085 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3093 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3103 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3113 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 3121 drivers/clk/qcom/mmcc-apq8084.c .pd = { pd 2342 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2352 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2362 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2372 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2382 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2390 drivers/clk/qcom/mmcc-msm8974.c .pd = { pd 2393 drivers/clk/qcom/mmcc-msm8974.c .parent = &oxili_gdsc.pd, pd 2899 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2909 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2919 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2929 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2940 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2943 drivers/clk/qcom/mmcc-msm8996.c .parent = &mmagic_video_gdsc.pd, pd 2951 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2954 drivers/clk/qcom/mmcc-msm8996.c .parent = &venus_gdsc.pd, pd 2963 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2966 drivers/clk/qcom/mmcc-msm8996.c .parent = &venus_gdsc.pd, pd 2975 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2978 drivers/clk/qcom/mmcc-msm8996.c .parent = &mmagic_camss_gdsc.pd, pd 2986 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 2989 drivers/clk/qcom/mmcc-msm8996.c .parent = &camss_gdsc.pd, pd 2997 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3000 drivers/clk/qcom/mmcc-msm8996.c .parent = &camss_gdsc.pd, pd 3008 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3011 drivers/clk/qcom/mmcc-msm8996.c .parent = &camss_gdsc.pd, pd 3019 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3022 drivers/clk/qcom/mmcc-msm8996.c .parent = &camss_gdsc.pd, pd 3030 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3033 drivers/clk/qcom/mmcc-msm8996.c .parent = &camss_gdsc.pd, pd 3041 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3044 drivers/clk/qcom/mmcc-msm8996.c .parent = &mmagic_mdss_gdsc.pd, pd 3051 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 3063 drivers/clk/qcom/mmcc-msm8996.c .pd = { pd 248 drivers/clk/qcom/videocc-sdm845.c .pd = { pd 259 drivers/clk/qcom/videocc-sdm845.c .pd = { pd 270 drivers/clk/qcom/videocc-sdm845.c .pd = { pd 324 drivers/clk/renesas/clk-mstp.c struct generic_pm_domain *pd; pd 332 drivers/clk/renesas/clk-mstp.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 333 drivers/clk/renesas/clk-mstp.c if (!pd) pd 336 drivers/clk/renesas/clk-mstp.c pd->name = np->name; pd 337 drivers/clk/renesas/clk-mstp.c pd->flags = GENPD_FLAG_PM_CLK | GENPD_FLAG_ALWAYS_ON | pd 339 drivers/clk/renesas/clk-mstp.c pd->attach_dev = cpg_mstp_attach_dev; pd 340 drivers/clk/renesas/clk-mstp.c pd->detach_dev = cpg_mstp_detach_dev; pd 341 drivers/clk/renesas/clk-mstp.c pm_genpd_init(pd, &pm_domain_always_on_gov, false); pd 343 drivers/clk/renesas/clk-mstp.c of_genpd_add_provider_simple(np, pd); pd 380 drivers/clk/renesas/r9a06g032-clocks.c static int r9a06g032_attach_dev(struct generic_pm_domain *pd, pd 391 drivers/clk/renesas/r9a06g032-clocks.c if (clkspec.np != pd->dev.of_node) pd 417 drivers/clk/renesas/r9a06g032-clocks.c struct generic_pm_domain *pd; pd 419 drivers/clk/renesas/r9a06g032-clocks.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 420 drivers/clk/renesas/r9a06g032-clocks.c if (!pd) pd 423 drivers/clk/renesas/r9a06g032-clocks.c pd->name = np->name; pd 424 drivers/clk/renesas/r9a06g032-clocks.c pd->flags = GENPD_FLAG_PM_CLK | GENPD_FLAG_ALWAYS_ON | pd 426 drivers/clk/renesas/r9a06g032-clocks.c pd->attach_dev = r9a06g032_attach_dev; pd 427 drivers/clk/renesas/r9a06g032-clocks.c pd->detach_dev = r9a06g032_detach_dev; pd 428 drivers/clk/renesas/r9a06g032-clocks.c pm_genpd_init(pd, &pm_domain_always_on_gov, false); pd 430 drivers/clk/renesas/r9a06g032-clocks.c of_genpd_add_provider_simple(np, pd); pd 459 drivers/clk/renesas/renesas-cpg-mssr.c struct cpg_mssr_clk_domain *pd) pd 463 drivers/clk/renesas/renesas-cpg-mssr.c if (clkspec->np != pd->genpd.dev.of_node || clkspec->args_count != 2) pd 468 drivers/clk/renesas/renesas-cpg-mssr.c for (i = 0; i < pd->num_core_pm_clks; i++) pd 469 drivers/clk/renesas/renesas-cpg-mssr.c if (clkspec->args[1] == pd->core_pm_clks[i]) pd 483 drivers/clk/renesas/renesas-cpg-mssr.c struct cpg_mssr_clk_domain *pd = cpg_mssr_clk_domain; pd 490 drivers/clk/renesas/renesas-cpg-mssr.c if (!pd) { pd 497 drivers/clk/renesas/renesas-cpg-mssr.c if (cpg_mssr_is_pm_clk(&clkspec, pd)) pd 542 drivers/clk/renesas/renesas-cpg-mssr.c struct cpg_mssr_clk_domain *pd; pd 545 drivers/clk/renesas/renesas-cpg-mssr.c pd = devm_kzalloc(dev, sizeof(*pd) + pm_size, GFP_KERNEL); pd 546 drivers/clk/renesas/renesas-cpg-mssr.c if (!pd) pd 549 drivers/clk/renesas/renesas-cpg-mssr.c pd->num_core_pm_clks = num_core_pm_clks; pd 550 drivers/clk/renesas/renesas-cpg-mssr.c memcpy(pd->core_pm_clks, core_pm_clks, pm_size); pd 552 drivers/clk/renesas/renesas-cpg-mssr.c genpd = &pd->genpd; pd 559 drivers/clk/renesas/renesas-cpg-mssr.c cpg_mssr_clk_domain = pd; pd 163 drivers/clk/sifive/fu540-prci.c void (*enable_bypass)(struct __prci_data *pd); pd 164 drivers/clk/sifive/fu540-prci.c void (*disable_bypass)(struct __prci_data *pd); pd 186 drivers/clk/sifive/fu540-prci.c struct __prci_data *pd; pd 208 drivers/clk/sifive/fu540-prci.c static u32 __prci_readl(struct __prci_data *pd, u32 offs) pd 210 drivers/clk/sifive/fu540-prci.c return readl_relaxed(pd->va + offs); pd 213 drivers/clk/sifive/fu540-prci.c static void __prci_writel(u32 v, u32 offs, struct __prci_data *pd) pd 215 drivers/clk/sifive/fu540-prci.c writel_relaxed(v, pd->va + offs); pd 303 drivers/clk/sifive/fu540-prci.c static void __prci_wrpll_read_cfg(struct __prci_data *pd, pd 306 drivers/clk/sifive/fu540-prci.c __prci_wrpll_unpack(&pwd->c, __prci_readl(pd, pwd->cfg0_offs)); pd 323 drivers/clk/sifive/fu540-prci.c static void __prci_wrpll_write_cfg(struct __prci_data *pd, pd 327 drivers/clk/sifive/fu540-prci.c __prci_writel(__prci_wrpll_pack(c), pwd->cfg0_offs, pd); pd 343 drivers/clk/sifive/fu540-prci.c static void __prci_coreclksel_use_hfclk(struct __prci_data *pd) pd 347 drivers/clk/sifive/fu540-prci.c r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); pd 349 drivers/clk/sifive/fu540-prci.c __prci_writel(r, PRCI_CORECLKSEL_OFFSET, pd); pd 351 drivers/clk/sifive/fu540-prci.c r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); /* barrier */ pd 363 drivers/clk/sifive/fu540-prci.c static void __prci_coreclksel_use_corepll(struct __prci_data *pd) pd 367 drivers/clk/sifive/fu540-prci.c r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); pd 369 drivers/clk/sifive/fu540-prci.c __prci_writel(r, PRCI_CORECLKSEL_OFFSET, pd); pd 371 drivers/clk/sifive/fu540-prci.c r = __prci_readl(pd, PRCI_CORECLKSEL_OFFSET); /* barrier */ pd 411 drivers/clk/sifive/fu540-prci.c struct __prci_data *pd = pc->pd; pd 419 drivers/clk/sifive/fu540-prci.c pwd->enable_bypass(pd); pd 421 drivers/clk/sifive/fu540-prci.c __prci_wrpll_write_cfg(pd, pwd, &pwd->c); pd 426 drivers/clk/sifive/fu540-prci.c pwd->disable_bypass(pd); pd 447 drivers/clk/sifive/fu540-prci.c struct __prci_data *pd = pc->pd; pd 451 drivers/clk/sifive/fu540-prci.c v = __prci_readl(pd, PRCI_CLKMUXSTATUSREG_OFFSET); pd 519 drivers/clk/sifive/fu540-prci.c static int __prci_register_clocks(struct device *dev, struct __prci_data *pd) pd 542 drivers/clk/sifive/fu540-prci.c pic->pd = pd; pd 545 drivers/clk/sifive/fu540-prci.c __prci_wrpll_read_cfg(pd, pic->pwd); pd 561 drivers/clk/sifive/fu540-prci.c pd->hw_clks.hws[i] = &pic->hw; pd 564 drivers/clk/sifive/fu540-prci.c pd->hw_clks.num = i; pd 567 drivers/clk/sifive/fu540-prci.c &pd->hw_clks); pd 586 drivers/clk/sifive/fu540-prci.c struct __prci_data *pd; pd 589 drivers/clk/sifive/fu540-prci.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 590 drivers/clk/sifive/fu540-prci.c if (!pd) pd 594 drivers/clk/sifive/fu540-prci.c pd->va = devm_ioremap_resource(dev, res); pd 595 drivers/clk/sifive/fu540-prci.c if (IS_ERR(pd->va)) pd 596 drivers/clk/sifive/fu540-prci.c return PTR_ERR(pd->va); pd 598 drivers/clk/sifive/fu540-prci.c r = __prci_register_clocks(dev, pd); pd 64 drivers/clocksource/timer-stm32.c struct stm32_timer_private *pd = to->private_data; pd 66 drivers/clocksource/timer-stm32.c pd->bits = bits; pd 80 drivers/clocksource/timer-stm32.c struct stm32_timer_private *pd = to->private_data; pd 82 drivers/clocksource/timer-stm32.c return pd->bits; pd 198 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd = &dev->pdr[i]; pd 201 drivers/crypto/amcc/crypto4xx_core.c pd->sa = dev->shadow_sa_pool_pa + pd 446 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd, pd 520 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd) pd 529 drivers/crypto/amcc/crypto4xx_core.c crypto4xx_copy_pkt_to_dst(dev, pd, pd_uinfo, pd 572 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd) pd 583 drivers/crypto/amcc/crypto4xx_core.c crypto4xx_copy_pkt_to_dst(dev, pd, pd_uinfo, pd 584 drivers/crypto/amcc/crypto4xx_core.c pd->pd_ctl_len.bf.pkt_len, pd 587 drivers/crypto/amcc/crypto4xx_core.c dma_unmap_page(dev->core_dev->device, pd->dest, dst->length, pd 612 drivers/crypto/amcc/crypto4xx_core.c if (pd->pd_ctl.bf.status & 0xff) { pd 614 drivers/crypto/amcc/crypto4xx_core.c if (pd->pd_ctl.bf.status & 2) pd 616 drivers/crypto/amcc/crypto4xx_core.c if (pd->pd_ctl.bf.status & 4) pd 618 drivers/crypto/amcc/crypto4xx_core.c if (pd->pd_ctl.bf.status & 8) pd 621 drivers/crypto/amcc/crypto4xx_core.c pd->pd_ctl.bf.status & 0xff); pd 623 drivers/crypto/amcc/crypto4xx_core.c pd->pd_ctl.bf.pd_pad_ctl); pd 636 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd = &dev->pdr[idx]; pd 641 drivers/crypto/amcc/crypto4xx_core.c crypto4xx_cipher_done(dev, pd_uinfo, pd); pd 644 drivers/crypto/amcc/crypto4xx_core.c crypto4xx_aead_done(dev, pd_uinfo, pd); pd 692 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd; pd 808 drivers/crypto/amcc/crypto4xx_core.c pd = &dev->pdr[pd_entry]; pd 809 drivers/crypto/amcc/crypto4xx_core.c pd->sa_len = sa_len; pd 835 drivers/crypto/amcc/crypto4xx_core.c pd->src = gd_dma; pd 859 drivers/crypto/amcc/crypto4xx_core.c pd->src = (u32)dma_map_page(dev->core_dev->device, sg_page(src), pd 878 drivers/crypto/amcc/crypto4xx_core.c pd->dest = (u32)dma_map_page(dev->core_dev->device, pd 891 drivers/crypto/amcc/crypto4xx_core.c pd->dest = sd_dma; pd 918 drivers/crypto/amcc/crypto4xx_core.c pd->pd_ctl.w = PD_CTL_HOST_READY | pd 922 drivers/crypto/amcc/crypto4xx_core.c pd->pd_ctl_len.w = 0x00400000 | (assoclen + datalen); pd 1079 drivers/crypto/amcc/crypto4xx_core.c struct ce_pd *pd; pd 1085 drivers/crypto/amcc/crypto4xx_core.c pd = &core_dev->dev->pdr[tail]; pd 1087 drivers/crypto/amcc/crypto4xx_core.c ((READ_ONCE(pd->pd_ctl.w) & pd 285 drivers/dma/amba-pl08x.c struct pl08x_platform_data *pd; pd 335 drivers/dma/amba-pl08x.c const struct pl08x_platform_data *pd = plchan->host->pd; pd 338 drivers/dma/amba-pl08x.c if (plchan->mux_use++ == 0 && pd->get_xfer_signal) { pd 339 drivers/dma/amba-pl08x.c ret = pd->get_xfer_signal(plchan->cd); pd 352 drivers/dma/amba-pl08x.c const struct pl08x_platform_data *pd = plchan->host->pd; pd 357 drivers/dma/amba-pl08x.c if (--plchan->mux_use == 0 && pd->put_xfer_signal) { pd 358 drivers/dma/amba-pl08x.c pd->put_xfer_signal(plchan->cd, plchan->signal); pd 463 drivers/dma/amba-pl08x.c switch (pl08x->pd->memcpy_burst_size) { pd 500 drivers/dma/amba-pl08x.c if (pl08x->pd->memcpy_prot_buff) pd 502 drivers/dma/amba-pl08x.c if (pl08x->pd->memcpy_prot_cache) pd 1766 drivers/dma/amba-pl08x.c switch (pl08x->pd->memcpy_burst_size) { pd 1805 drivers/dma/amba-pl08x.c switch (pl08x->pd->memcpy_bus_width) { pd 1825 drivers/dma/amba-pl08x.c if (pl08x->pd->memcpy_prot_buff) pd 1827 drivers/dma/amba-pl08x.c if (pl08x->pd->memcpy_prot_cache) pd 1849 drivers/dma/amba-pl08x.c switch (pl08x->pd->memcpy_bus_width) { pd 2392 drivers/dma/amba-pl08x.c chan->cd = &pl08x->pd->slave_channels[i]; pd 2407 drivers/dma/amba-pl08x.c chan->cd->periph_buses = pl08x->pd->mem_buses; pd 2576 drivers/dma/amba-pl08x.c struct pl08x_platform_data *pd; pd 2582 drivers/dma/amba-pl08x.c pd = devm_kzalloc(&adev->dev, sizeof(*pd), GFP_KERNEL); pd 2583 drivers/dma/amba-pl08x.c if (!pd) pd 2588 drivers/dma/amba-pl08x.c pd->lli_buses |= PL08X_AHB1; pd 2590 drivers/dma/amba-pl08x.c pd->lli_buses |= PL08X_AHB2; pd 2591 drivers/dma/amba-pl08x.c if (!pd->lli_buses) { pd 2593 drivers/dma/amba-pl08x.c pd->lli_buses |= PL08X_AHB1 | PL08X_AHB2; pd 2598 drivers/dma/amba-pl08x.c pd->mem_buses |= PL08X_AHB1; pd 2600 drivers/dma/amba-pl08x.c pd->mem_buses |= PL08X_AHB2; pd 2601 drivers/dma/amba-pl08x.c if (!pd->mem_buses) { pd 2603 drivers/dma/amba-pl08x.c pd->mem_buses |= PL08X_AHB1 | PL08X_AHB2; pd 2617 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_1; pd 2620 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_4; pd 2623 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_8; pd 2626 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_16; pd 2629 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_32; pd 2632 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_64; pd 2635 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_128; pd 2638 drivers/dma/amba-pl08x.c pd->memcpy_burst_size = PL08X_BURST_SZ_256; pd 2652 drivers/dma/amba-pl08x.c pd->memcpy_bus_width = PL08X_BUS_WIDTH_8_BITS; pd 2655 drivers/dma/amba-pl08x.c pd->memcpy_bus_width = PL08X_BUS_WIDTH_16_BITS; pd 2658 drivers/dma/amba-pl08x.c pd->memcpy_bus_width = PL08X_BUS_WIDTH_32_BITS; pd 2676 drivers/dma/amba-pl08x.c pd->slave_channels = chanp; pd 2684 drivers/dma/amba-pl08x.c pd->num_slave_channels = pl08x->vd->signals; pd 2687 drivers/dma/amba-pl08x.c pl08x->pd = pd; pd 2810 drivers/dma/amba-pl08x.c pl08x->pd = dev_get_platdata(&adev->dev); pd 2811 drivers/dma/amba-pl08x.c if (!pl08x->pd) { pd 2822 drivers/dma/amba-pl08x.c pl08x->slave.filter.map = pl08x->pd->slave_map; pd 2823 drivers/dma/amba-pl08x.c pl08x->slave.filter.mapcnt = pl08x->pd->slave_map_len; pd 2831 drivers/dma/amba-pl08x.c pl08x->lli_buses = pl08x->pd->lli_buses; pd 2832 drivers/dma/amba-pl08x.c pl08x->mem_buses = pl08x->pd->mem_buses; pd 2934 drivers/dma/amba-pl08x.c pl08x->pd->num_slave_channels, true); pd 284 drivers/dma/imx-sdma.c u32 pd; pd 132 drivers/dma/pch_dma.c #define dma_readl(pd, name) \ pd 133 drivers/dma/pch_dma.c readl((pd)->membase + PCH_DMA_##name) pd 134 drivers/dma/pch_dma.c #define dma_writel(pd, name, val) \ pd 135 drivers/dma/pch_dma.c writel((val), (pd)->membase + PCH_DMA_##name) pd 179 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(chan->device); pd 188 drivers/dma/pch_dma.c val = dma_readl(pd, CTL2); pd 195 drivers/dma/pch_dma.c dma_writel(pd, CTL2, val); pd 204 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(chan->device); pd 210 drivers/dma/pch_dma.c val = dma_readl(pd, CTL0); pd 225 drivers/dma/pch_dma.c dma_writel(pd, CTL0, val); pd 228 drivers/dma/pch_dma.c val = dma_readl(pd, CTL3); pd 242 drivers/dma/pch_dma.c dma_writel(pd, CTL3, val); pd 251 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(chan->device); pd 261 drivers/dma/pch_dma.c val = dma_readl(pd, CTL0); pd 265 drivers/dma/pch_dma.c dma_writel(pd, CTL0, val); pd 272 drivers/dma/pch_dma.c val = dma_readl(pd, CTL3); pd 276 drivers/dma/pch_dma.c dma_writel(pd, CTL3, val); pd 285 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(pd_chan->chan.device); pd 288 drivers/dma/pch_dma.c val = dma_readl(pd, STS0); pd 295 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(pd_chan->chan.device); pd 298 drivers/dma/pch_dma.c val = dma_readl(pd, STS2); pd 429 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(chan->device); pd 432 drivers/dma/pch_dma.c desc = dma_pool_zalloc(pd->pool, flags, &addr); pd 530 drivers/dma/pch_dma.c struct pch_dma *pd = to_pd(chan->device); pd 544 drivers/dma/pch_dma.c dma_pool_free(pd->pool, desc, desc->txd.phys); pd 694 drivers/dma/pch_dma.c struct pch_dma *pd = (struct pch_dma *)devid; pd 702 drivers/dma/pch_dma.c sts0 = dma_readl(pd, STS0); pd 703 drivers/dma/pch_dma.c sts2 = dma_readl(pd, STS2); pd 705 drivers/dma/pch_dma.c dev_dbg(pd->dma.dev, "pd_irq sts0: %x\n", sts0); pd 707 drivers/dma/pch_dma.c for (i = 0; i < pd->dma.chancnt; i++) { pd 708 drivers/dma/pch_dma.c pd_chan = &pd->channels[i]; pd 731 drivers/dma/pch_dma.c dma_writel(pd, STS0, sts0); pd 733 drivers/dma/pch_dma.c dma_writel(pd, STS2, sts2); pd 739 drivers/dma/pch_dma.c static void pch_dma_save_regs(struct pch_dma *pd) pd 745 drivers/dma/pch_dma.c pd->regs.dma_ctl0 = dma_readl(pd, CTL0); pd 746 drivers/dma/pch_dma.c pd->regs.dma_ctl1 = dma_readl(pd, CTL1); pd 747 drivers/dma/pch_dma.c pd->regs.dma_ctl2 = dma_readl(pd, CTL2); pd 748 drivers/dma/pch_dma.c pd->regs.dma_ctl3 = dma_readl(pd, CTL3); pd 750 drivers/dma/pch_dma.c list_for_each_entry_safe(chan, _c, &pd->dma.channels, device_node) { pd 753 drivers/dma/pch_dma.c pd->ch_regs[i].dev_addr = channel_readl(pd_chan, DEV_ADDR); pd 754 drivers/dma/pch_dma.c pd->ch_regs[i].mem_addr = channel_readl(pd_chan, MEM_ADDR); pd 755 drivers/dma/pch_dma.c pd->ch_regs[i].size = channel_readl(pd_chan, SIZE); pd 756 drivers/dma/pch_dma.c pd->ch_regs[i].next = channel_readl(pd_chan, NEXT); pd 762 drivers/dma/pch_dma.c static void pch_dma_restore_regs(struct pch_dma *pd) pd 768 drivers/dma/pch_dma.c dma_writel(pd, CTL0, pd->regs.dma_ctl0); pd 769 drivers/dma/pch_dma.c dma_writel(pd, CTL1, pd->regs.dma_ctl1); pd 770 drivers/dma/pch_dma.c dma_writel(pd, CTL2, pd->regs.dma_ctl2); pd 771 drivers/dma/pch_dma.c dma_writel(pd, CTL3, pd->regs.dma_ctl3); pd 773 drivers/dma/pch_dma.c list_for_each_entry_safe(chan, _c, &pd->dma.channels, device_node) { pd 776 drivers/dma/pch_dma.c channel_writel(pd_chan, DEV_ADDR, pd->ch_regs[i].dev_addr); pd 777 drivers/dma/pch_dma.c channel_writel(pd_chan, MEM_ADDR, pd->ch_regs[i].mem_addr); pd 778 drivers/dma/pch_dma.c channel_writel(pd_chan, SIZE, pd->ch_regs[i].size); pd 779 drivers/dma/pch_dma.c channel_writel(pd_chan, NEXT, pd->ch_regs[i].next); pd 787 drivers/dma/pch_dma.c struct pch_dma *pd = pci_get_drvdata(pdev); pd 789 drivers/dma/pch_dma.c if (pd) pd 790 drivers/dma/pch_dma.c pch_dma_save_regs(pd); pd 801 drivers/dma/pch_dma.c struct pch_dma *pd = pci_get_drvdata(pdev); pd 813 drivers/dma/pch_dma.c if (pd) pd 814 drivers/dma/pch_dma.c pch_dma_restore_regs(pd); pd 823 drivers/dma/pch_dma.c struct pch_dma *pd; pd 830 drivers/dma/pch_dma.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 831 drivers/dma/pch_dma.c if (!pd) pd 834 drivers/dma/pch_dma.c pci_set_drvdata(pdev, pd); pd 860 drivers/dma/pch_dma.c regs = pd->membase = pci_iomap(pdev, 1, 0); pd 861 drivers/dma/pch_dma.c if (!pd->membase) { pd 868 drivers/dma/pch_dma.c pd->dma.dev = &pdev->dev; pd 870 drivers/dma/pch_dma.c err = request_irq(pdev->irq, pd_irq, IRQF_SHARED, DRV_NAME, pd); pd 876 drivers/dma/pch_dma.c pd->pool = dma_pool_create("pch_dma_desc_pool", &pdev->dev, pd 878 drivers/dma/pch_dma.c if (!pd->pool) { pd 885 drivers/dma/pch_dma.c INIT_LIST_HEAD(&pd->dma.channels); pd 888 drivers/dma/pch_dma.c struct pch_dma_chan *pd_chan = &pd->channels[i]; pd 890 drivers/dma/pch_dma.c pd_chan->chan.device = &pd->dma; pd 903 drivers/dma/pch_dma.c list_add_tail(&pd_chan->chan.device_node, &pd->dma.channels); pd 906 drivers/dma/pch_dma.c dma_cap_zero(pd->dma.cap_mask); pd 907 drivers/dma/pch_dma.c dma_cap_set(DMA_PRIVATE, pd->dma.cap_mask); pd 908 drivers/dma/pch_dma.c dma_cap_set(DMA_SLAVE, pd->dma.cap_mask); pd 910 drivers/dma/pch_dma.c pd->dma.device_alloc_chan_resources = pd_alloc_chan_resources; pd 911 drivers/dma/pch_dma.c pd->dma.device_free_chan_resources = pd_free_chan_resources; pd 912 drivers/dma/pch_dma.c pd->dma.device_tx_status = pd_tx_status; pd 913 drivers/dma/pch_dma.c pd->dma.device_issue_pending = pd_issue_pending; pd 914 drivers/dma/pch_dma.c pd->dma.device_prep_slave_sg = pd_prep_slave_sg; pd 915 drivers/dma/pch_dma.c pd->dma.device_terminate_all = pd_device_terminate_all; pd 917 drivers/dma/pch_dma.c err = dma_async_device_register(&pd->dma); pd 926 drivers/dma/pch_dma.c dma_pool_destroy(pd->pool); pd 928 drivers/dma/pch_dma.c free_irq(pdev->irq, pd); pd 930 drivers/dma/pch_dma.c pci_iounmap(pdev, pd->membase); pd 936 drivers/dma/pch_dma.c kfree(pd); pd 942 drivers/dma/pch_dma.c struct pch_dma *pd = pci_get_drvdata(pdev); pd 946 drivers/dma/pch_dma.c if (pd) { pd 947 drivers/dma/pch_dma.c dma_async_device_unregister(&pd->dma); pd 949 drivers/dma/pch_dma.c free_irq(pdev->irq, pd); pd 951 drivers/dma/pch_dma.c list_for_each_entry_safe(chan, _c, &pd->dma.channels, pd 958 drivers/dma/pch_dma.c dma_pool_destroy(pd->pool); pd 959 drivers/dma/pch_dma.c pci_iounmap(pdev, pd->membase); pd 962 drivers/dma/pch_dma.c kfree(pd); pd 3000 drivers/dma/pl330.c struct dma_device *pd; pd 3015 drivers/dma/pl330.c pd = &pl330->ddma; pd 3016 drivers/dma/pl330.c pd->dev = &adev->dev; pd 3086 drivers/dma/pl330.c INIT_LIST_HEAD(&pd->channels); pd 3108 drivers/dma/pl330.c pch->chan.device = pd; pd 3113 drivers/dma/pl330.c list_add_tail(&pch->chan.device_node, &pd->channels); pd 3116 drivers/dma/pl330.c dma_cap_set(DMA_MEMCPY, pd->cap_mask); pd 3118 drivers/dma/pl330.c dma_cap_set(DMA_SLAVE, pd->cap_mask); pd 3119 drivers/dma/pl330.c dma_cap_set(DMA_CYCLIC, pd->cap_mask); pd 3120 drivers/dma/pl330.c dma_cap_set(DMA_PRIVATE, pd->cap_mask); pd 3123 drivers/dma/pl330.c pd->device_alloc_chan_resources = pl330_alloc_chan_resources; pd 3124 drivers/dma/pl330.c pd->device_free_chan_resources = pl330_free_chan_resources; pd 3125 drivers/dma/pl330.c pd->device_prep_dma_memcpy = pl330_prep_dma_memcpy; pd 3126 drivers/dma/pl330.c pd->device_prep_dma_cyclic = pl330_prep_dma_cyclic; pd 3127 drivers/dma/pl330.c pd->device_tx_status = pl330_tx_status; pd 3128 drivers/dma/pl330.c pd->device_prep_slave_sg = pl330_prep_slave_sg; pd 3129 drivers/dma/pl330.c pd->device_config = pl330_config; pd 3130 drivers/dma/pl330.c pd->device_pause = pl330_pause; pd 3131 drivers/dma/pl330.c pd->device_terminate_all = pl330_terminate_all; pd 3132 drivers/dma/pl330.c pd->device_issue_pending = pl330_issue_pending; pd 3133 drivers/dma/pl330.c pd->src_addr_widths = PL330_DMA_BUSWIDTHS; pd 3134 drivers/dma/pl330.c pd->dst_addr_widths = PL330_DMA_BUSWIDTHS; pd 3135 drivers/dma/pl330.c pd->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); pd 3136 drivers/dma/pl330.c pd->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; pd 3137 drivers/dma/pl330.c pd->max_burst = ((pl330->quirks & PL330_QUIRK_BROKEN_NO_FLUSHP) ? pd 3140 drivers/dma/pl330.c ret = dma_async_device_register(pd); pd 284 drivers/firewire/net.c static bool fwnet_frag_overlap(struct fwnet_partial_datagram *pd, pd 290 drivers/firewire/net.c list_for_each_entry(fi, &pd->fi_list, fi_link) pd 299 drivers/firewire/net.c struct fwnet_partial_datagram *pd, unsigned offset, unsigned len) pd 304 drivers/firewire/net.c list = &pd->fi_list; pd 305 drivers/firewire/net.c list_for_each_entry(fi, &pd->fi_list, fi_link) { pd 401 drivers/firewire/net.c struct fwnet_partial_datagram *pd; pd 403 drivers/firewire/net.c list_for_each_entry(pd, &peer->pd_list, pd_link) pd 404 drivers/firewire/net.c if (pd->datagram_label == datagram_label) pd 405 drivers/firewire/net.c return pd; pd 424 drivers/firewire/net.c struct fwnet_partial_datagram *pd, void *frag_buf, pd 427 drivers/firewire/net.c if (fwnet_frag_new(pd, frag_off, frag_len) == NULL) pd 430 drivers/firewire/net.c memcpy(pd->pbuf + frag_off, frag_buf, frag_len); pd 436 drivers/firewire/net.c list_move_tail(&pd->pd_link, &peer->pd_list); pd 441 drivers/firewire/net.c static bool fwnet_pd_is_complete(struct fwnet_partial_datagram *pd) pd 445 drivers/firewire/net.c fi = list_entry(pd->fi_list.next, struct fwnet_fragment_info, fi_link); pd 447 drivers/firewire/net.c return fi->len == pd->datagram_size; pd 579 drivers/firewire/net.c struct fwnet_partial_datagram *pd; pd 643 drivers/firewire/net.c pd = fwnet_pd_find(peer, datagram_label); pd 644 drivers/firewire/net.c if (pd == NULL) { pd 651 drivers/firewire/net.c pd = fwnet_pd_new(net, peer, datagram_label, pd 653 drivers/firewire/net.c if (pd == NULL) { pd 659 drivers/firewire/net.c if (fwnet_frag_overlap(pd, fg_off, len) || pd 660 drivers/firewire/net.c pd->datagram_size != dg_size) { pd 665 drivers/firewire/net.c fwnet_pd_delete(pd); pd 666 drivers/firewire/net.c pd = fwnet_pd_new(net, peer, datagram_label, pd 668 drivers/firewire/net.c if (pd == NULL) { pd 674 drivers/firewire/net.c if (!fwnet_pd_update(peer, pd, buf, fg_off, len)) { pd 680 drivers/firewire/net.c fwnet_pd_delete(pd); pd 689 drivers/firewire/net.c pd->ether_type = ether_type; pd 691 drivers/firewire/net.c if (fwnet_pd_is_complete(pd)) { pd 692 drivers/firewire/net.c ether_type = pd->ether_type; pd 694 drivers/firewire/net.c skb = skb_get(pd->skb); pd 695 drivers/firewire/net.c fwnet_pd_delete(pd); pd 1549 drivers/firewire/net.c struct fwnet_partial_datagram *pd, *pd_next; pd 1557 drivers/firewire/net.c list_for_each_entry_safe(pd, pd_next, &peer->pd_list, pd_link) pd 1558 drivers/firewire/net.c fwnet_pd_delete(pd); pd 2767 drivers/firewire/ohci.c struct descriptor *pd; pd 2770 drivers/firewire/ohci.c for (pd = d; pd <= last; pd++) pd 2771 drivers/firewire/ohci.c if (pd->transfer_status) pd 2773 drivers/firewire/ohci.c if (pd > last) pd 2848 drivers/firewire/ohci.c struct descriptor *pd) pd 2854 drivers/firewire/ohci.c if (pd->control & cpu_to_le16(DESCRIPTOR_BRANCH_ALWAYS)) pd 2858 drivers/firewire/ohci.c pd += 2; pd 2865 drivers/firewire/ohci.c if ((le32_to_cpu(pd->data_address) & PAGE_MASK) == pd 2867 drivers/firewire/ohci.c if (pd->control & cpu_to_le16(DESCRIPTOR_BRANCH_ALWAYS)) pd 2869 drivers/firewire/ohci.c pd++; pd 2873 drivers/firewire/ohci.c buffer_dma = le32_to_cpu(pd->data_address); pd 2877 drivers/firewire/ohci.c le16_to_cpu(pd->req_count), pd 2879 drivers/firewire/ohci.c control = pd->control; pd 2880 drivers/firewire/ohci.c pd++; pd 2890 drivers/firewire/ohci.c struct descriptor *pd; pd 2893 drivers/firewire/ohci.c for (pd = d; pd <= last; pd++) pd 2894 drivers/firewire/ohci.c if (pd->transfer_status) pd 2896 drivers/firewire/ohci.c if (pd > last) pd 2911 drivers/firewire/ohci.c *ctx_hdr = cpu_to_be32((le16_to_cpu(pd->transfer_status) << 16) | pd 2912 drivers/firewire/ohci.c le16_to_cpu(pd->res_count)); pd 3197 drivers/firewire/ohci.c struct descriptor *d, *last, *pd; pd 3260 drivers/firewire/ohci.c pd = d + z - payload_z; pd 3268 drivers/firewire/ohci.c pd[i].req_count = cpu_to_le16(length); pd 3271 drivers/firewire/ohci.c pd[i].data_address = cpu_to_le32(page_bus + offset); pd 3302 drivers/firewire/ohci.c struct descriptor *d, *pd; pd 3339 drivers/firewire/ohci.c pd = d; pd 3341 drivers/firewire/ohci.c pd++; pd 3342 drivers/firewire/ohci.c pd->control = cpu_to_le16(DESCRIPTOR_STATUS | pd 3349 drivers/firewire/ohci.c pd->req_count = cpu_to_le16(length); pd 3350 drivers/firewire/ohci.c pd->res_count = pd->req_count; pd 3351 drivers/firewire/ohci.c pd->transfer_status = 0; pd 3354 drivers/firewire/ohci.c pd->data_address = cpu_to_le32(page_bus + offset); pd 3365 drivers/firewire/ohci.c pd->control = cpu_to_le16(DESCRIPTOR_STATUS | pd 3369 drivers/firewire/ohci.c pd->control |= cpu_to_le16(DESCRIPTOR_IRQ_ALWAYS); pd 27 drivers/firmware/arm_scmi/scmi_pm_domain.c struct scmi_pm_domain *pd = to_scmi_pd(domain); pd 28 drivers/firmware/arm_scmi/scmi_pm_domain.c const struct scmi_power_ops *ops = pd->handle->power_ops; pd 35 drivers/firmware/arm_scmi/scmi_pm_domain.c ret = ops->state_set(pd->handle, pd->domain, state); pd 37 drivers/firmware/arm_scmi/scmi_pm_domain.c ret = ops->state_get(pd->handle, pd->domain, &ret_state); pd 272 drivers/firmware/efi/arm-init.c void *pd; pd 277 drivers/firmware/efi/arm-init.c pd = platform_device_register_data(NULL, "efi-framebuffer", 0, pd 279 drivers/firmware/efi/arm-init.c return PTR_ERR_OR_ZERO(pd); pd 68 drivers/firmware/imx/scu-pd.c struct generic_pm_domain pd; pd 173 drivers/firmware/imx/scu-pd.c return container_of(genpd, struct imx_sc_pm_domain, pd); pd 180 drivers/firmware/imx/scu-pd.c struct imx_sc_pm_domain *pd; pd 183 drivers/firmware/imx/scu-pd.c pd = to_imx_sc_pd(domain); pd 190 drivers/firmware/imx/scu-pd.c msg.resource = pd->rsrc; pd 196 drivers/firmware/imx/scu-pd.c power_on ? "up" : "off", pd->rsrc, ret); pd 223 drivers/firmware/imx/scu-pd.c domain = &sc_pd->pd; pd 243 drivers/firmware/imx/scu-pd.c sc_pd->pd.power_off = imx_sc_pd_power_off; pd 244 drivers/firmware/imx/scu-pd.c sc_pd->pd.power_on = imx_sc_pd_power_on; pd 253 drivers/firmware/imx/scu-pd.c sc_pd->pd.name = sc_pd->name; pd 263 drivers/firmware/imx/scu-pd.c ret = pm_genpd_init(&sc_pd->pd, NULL, true); pd 302 drivers/firmware/imx/scu-pd.c domains[count++] = &sc_pd->pd; pd 303 drivers/firmware/imx/scu-pd.c dev_dbg(dev, "added power domain %s\n", sc_pd->pd.name); pd 34 drivers/firmware/scpi_pm_domain.c static int scpi_pd_power(struct scpi_pm_domain *pd, bool power_on) pd 44 drivers/firmware/scpi_pm_domain.c ret = pd->ops->device_set_power_state(pd->domain, state); pd 48 drivers/firmware/scpi_pm_domain.c return !(state == pd->ops->device_get_power_state(pd->domain)); pd 53 drivers/firmware/scpi_pm_domain.c struct scpi_pm_domain *pd = to_scpi_pd(domain); pd 55 drivers/firmware/scpi_pm_domain.c return scpi_pd_power(pd, true); pd 60 drivers/firmware/scpi_pm_domain.c struct scpi_pm_domain *pd = to_scpi_pd(domain); pd 62 drivers/firmware/scpi_pm_domain.c return scpi_pd_power(pd, false); pd 134 drivers/firmware/trusted_foundations.c void register_trusted_foundations(struct trusted_foundations_platform_data *pd) pd 312 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo *pd = vm->root.base.bo; pd 313 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); pd 327 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_amdkfd_validate(¶m, pd); pd 336 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c ret = amdgpu_bo_kmap(pd, NULL); pd 348 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo *pd = vm->root.base.bo; pd 349 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); pd 812 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo *pd = peer_vm->root.base.bo; pd 815 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c sync, pd->tbo.base.resv, pd 995 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo *pd = vm->root.base.bo; pd 1001 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_bo_reserve(pd, false); pd 1002 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_bo_fence(pd, NULL, false); pd 1003 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c amdgpu_bo_unreserve(pd); pd 1062 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_bo *pd = avm->root.base.bo; pd 1063 drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); pd 71 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c struct amdgpu_bo_list_entry pd; pd 81 drivers/gpu/drm/amd/amdgpu/amdgpu_csa.c amdgpu_vm_get_pd_bo(vm, &list, &pd); pd 1048 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c struct amdgpu_bo *pd; pd 1073 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c pd = amdgpu_bo_ref(fpriv->vm.root.base.bo); pd 1079 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c amdgpu_pasid_free_delayed(pd->tbo.base.resv, pasid); pd 1080 drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c amdgpu_bo_unref(&pd); pd 128 drivers/gpu/drm/gma500/mmu.c void psb_mmu_set_pd_context(struct psb_mmu_pd *pd, int hw_context) pd 130 drivers/gpu/drm/gma500/mmu.c struct drm_device *dev = pd->driver->dev; pd 135 drivers/gpu/drm/gma500/mmu.c down_write(&pd->driver->sem); pd 136 drivers/gpu/drm/gma500/mmu.c PSB_WSGX32(page_to_pfn(pd->p) << PAGE_SHIFT, offset); pd 138 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush_pd_locked(pd->driver, 1); pd 139 drivers/gpu/drm/gma500/mmu.c pd->hw_context = hw_context; pd 140 drivers/gpu/drm/gma500/mmu.c up_write(&pd->driver->sem); pd 168 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pd *pd = kmalloc(sizeof(*pd), GFP_KERNEL); pd 172 drivers/gpu/drm/gma500/mmu.c if (!pd) pd 175 drivers/gpu/drm/gma500/mmu.c pd->p = alloc_page(GFP_DMA32); pd 176 drivers/gpu/drm/gma500/mmu.c if (!pd->p) pd 178 drivers/gpu/drm/gma500/mmu.c pd->dummy_pt = alloc_page(GFP_DMA32); pd 179 drivers/gpu/drm/gma500/mmu.c if (!pd->dummy_pt) pd 181 drivers/gpu/drm/gma500/mmu.c pd->dummy_page = alloc_page(GFP_DMA32); pd 182 drivers/gpu/drm/gma500/mmu.c if (!pd->dummy_page) pd 186 drivers/gpu/drm/gma500/mmu.c pd->invalid_pde = psb_mmu_mask_pte(page_to_pfn(pd->dummy_pt), pd 188 drivers/gpu/drm/gma500/mmu.c pd->invalid_pte = psb_mmu_mask_pte(page_to_pfn(pd->dummy_page), pd 191 drivers/gpu/drm/gma500/mmu.c pd->invalid_pde = 0; pd 192 drivers/gpu/drm/gma500/mmu.c pd->invalid_pte = 0; pd 195 drivers/gpu/drm/gma500/mmu.c v = kmap(pd->dummy_pt); pd 197 drivers/gpu/drm/gma500/mmu.c v[i] = pd->invalid_pte; pd 199 drivers/gpu/drm/gma500/mmu.c kunmap(pd->dummy_pt); pd 201 drivers/gpu/drm/gma500/mmu.c v = kmap(pd->p); pd 203 drivers/gpu/drm/gma500/mmu.c v[i] = pd->invalid_pde; pd 205 drivers/gpu/drm/gma500/mmu.c kunmap(pd->p); pd 207 drivers/gpu/drm/gma500/mmu.c clear_page(kmap(pd->dummy_page)); pd 208 drivers/gpu/drm/gma500/mmu.c kunmap(pd->dummy_page); pd 210 drivers/gpu/drm/gma500/mmu.c pd->tables = vmalloc_user(sizeof(struct psb_mmu_pt *) * 1024); pd 211 drivers/gpu/drm/gma500/mmu.c if (!pd->tables) pd 214 drivers/gpu/drm/gma500/mmu.c pd->hw_context = -1; pd 215 drivers/gpu/drm/gma500/mmu.c pd->pd_mask = PSB_PTE_VALID; pd 216 drivers/gpu/drm/gma500/mmu.c pd->driver = driver; pd 218 drivers/gpu/drm/gma500/mmu.c return pd; pd 221 drivers/gpu/drm/gma500/mmu.c __free_page(pd->dummy_page); pd 223 drivers/gpu/drm/gma500/mmu.c __free_page(pd->dummy_pt); pd 225 drivers/gpu/drm/gma500/mmu.c __free_page(pd->p); pd 227 drivers/gpu/drm/gma500/mmu.c kfree(pd); pd 237 drivers/gpu/drm/gma500/mmu.c void psb_mmu_free_pagedir(struct psb_mmu_pd *pd) pd 239 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_driver *driver = pd->driver; pd 246 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) { pd 247 drivers/gpu/drm/gma500/mmu.c PSB_WSGX32(0, PSB_CR_BIF_DIR_LIST_BASE0 + pd->hw_context * 4); pd 255 drivers/gpu/drm/gma500/mmu.c pt = pd->tables[i]; pd 260 drivers/gpu/drm/gma500/mmu.c vfree(pd->tables); pd 261 drivers/gpu/drm/gma500/mmu.c __free_page(pd->dummy_page); pd 262 drivers/gpu/drm/gma500/mmu.c __free_page(pd->dummy_pt); pd 263 drivers/gpu/drm/gma500/mmu.c __free_page(pd->p); pd 264 drivers/gpu/drm/gma500/mmu.c kfree(pd); pd 268 drivers/gpu/drm/gma500/mmu.c static struct psb_mmu_pt *psb_mmu_alloc_pt(struct psb_mmu_pd *pd) pd 272 drivers/gpu/drm/gma500/mmu.c uint32_t clflush_add = pd->driver->clflush_add >> PAGE_SHIFT; pd 274 drivers/gpu/drm/gma500/mmu.c spinlock_t *lock = &pd->driver->lock; pd 294 drivers/gpu/drm/gma500/mmu.c *ptes++ = pd->invalid_pte; pd 297 drivers/gpu/drm/gma500/mmu.c if (pd->driver->has_clflush && pd->hw_context != -1) { pd 310 drivers/gpu/drm/gma500/mmu.c pt->pd = pd; pd 316 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pt *psb_mmu_pt_alloc_map_lock(struct psb_mmu_pd *pd, pd 322 drivers/gpu/drm/gma500/mmu.c spinlock_t *lock = &pd->driver->lock; pd 325 drivers/gpu/drm/gma500/mmu.c pt = pd->tables[index]; pd 328 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_alloc_pt(pd); pd 333 drivers/gpu/drm/gma500/mmu.c if (pd->tables[index]) { pd 337 drivers/gpu/drm/gma500/mmu.c pt = pd->tables[index]; pd 341 drivers/gpu/drm/gma500/mmu.c v = kmap_atomic(pd->p); pd 342 drivers/gpu/drm/gma500/mmu.c pd->tables[index] = pt; pd 343 drivers/gpu/drm/gma500/mmu.c v[index] = (page_to_pfn(pt->p) << 12) | pd->pd_mask; pd 347 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) { pd 348 drivers/gpu/drm/gma500/mmu.c psb_mmu_clflush(pd->driver, (void *)&v[index]); pd 349 drivers/gpu/drm/gma500/mmu.c atomic_set(&pd->driver->needs_tlbflush, 1); pd 356 drivers/gpu/drm/gma500/mmu.c static struct psb_mmu_pt *psb_mmu_pt_map_lock(struct psb_mmu_pd *pd, pd 361 drivers/gpu/drm/gma500/mmu.c spinlock_t *lock = &pd->driver->lock; pd 364 drivers/gpu/drm/gma500/mmu.c pt = pd->tables[index]; pd 375 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pd *pd = pt->pd; pd 380 drivers/gpu/drm/gma500/mmu.c v = kmap_atomic(pd->p); pd 381 drivers/gpu/drm/gma500/mmu.c v[pt->index] = pd->invalid_pde; pd 382 drivers/gpu/drm/gma500/mmu.c pd->tables[pt->index] = NULL; pd 384 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) { pd 385 drivers/gpu/drm/gma500/mmu.c psb_mmu_clflush(pd->driver, (void *)&v[pt->index]); pd 386 drivers/gpu/drm/gma500/mmu.c atomic_set(&pd->driver->needs_tlbflush, 1); pd 389 drivers/gpu/drm/gma500/mmu.c spin_unlock(&pd->driver->lock); pd 393 drivers/gpu/drm/gma500/mmu.c spin_unlock(&pd->driver->lock); pd 405 drivers/gpu/drm/gma500/mmu.c pt->v[psb_mmu_pt_index(addr)] = pt->pd->invalid_pte; pd 410 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pd *pd; pd 413 drivers/gpu/drm/gma500/mmu.c pd = driver->default_pd; pd 416 drivers/gpu/drm/gma500/mmu.c return pd; pd 422 drivers/gpu/drm/gma500/mmu.c struct psb_mmu_pd *pd; pd 424 drivers/gpu/drm/gma500/mmu.c pd = psb_mmu_get_default_pd(driver); pd 425 drivers/gpu/drm/gma500/mmu.c return page_to_pfn(pd->p) << PAGE_SHIFT; pd 499 drivers/gpu/drm/gma500/mmu.c static void psb_mmu_flush_ptes(struct psb_mmu_pd *pd, unsigned long address, pd 511 drivers/gpu/drm/gma500/mmu.c unsigned long clflush_add = pd->driver->clflush_add; pd 512 drivers/gpu/drm/gma500/mmu.c unsigned long clflush_mask = pd->driver->clflush_mask; pd 514 drivers/gpu/drm/gma500/mmu.c if (!pd->driver->has_clflush) pd 532 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_map_lock(pd, addr); pd 547 drivers/gpu/drm/gma500/mmu.c static void psb_mmu_flush_ptes(struct psb_mmu_pd *pd, unsigned long address, pd 555 drivers/gpu/drm/gma500/mmu.c void psb_mmu_remove_pfn_sequence(struct psb_mmu_pd *pd, pd 564 drivers/gpu/drm/gma500/mmu.c down_read(&pd->driver->sem); pd 571 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_alloc_map_lock(pd, addr); pd 583 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 584 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); pd 586 drivers/gpu/drm/gma500/mmu.c up_read(&pd->driver->sem); pd 588 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 589 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush(pd->driver); pd 594 drivers/gpu/drm/gma500/mmu.c void psb_mmu_remove_pages(struct psb_mmu_pd *pd, unsigned long address, pd 616 drivers/gpu/drm/gma500/mmu.c down_read(&pd->driver->sem); pd 627 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_map_lock(pd, addr); pd 640 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 641 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush_ptes(pd, f_address, num_pages, pd 644 drivers/gpu/drm/gma500/mmu.c up_read(&pd->driver->sem); pd 646 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 647 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush(pd->driver); pd 650 drivers/gpu/drm/gma500/mmu.c int psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd, uint32_t start_pfn, pd 662 drivers/gpu/drm/gma500/mmu.c down_read(&pd->driver->sem); pd 669 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_alloc_map_lock(pd, addr); pd 685 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 686 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); pd 688 drivers/gpu/drm/gma500/mmu.c up_read(&pd->driver->sem); pd 690 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 691 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush(pd->driver); pd 696 drivers/gpu/drm/gma500/mmu.c int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages, pd 724 drivers/gpu/drm/gma500/mmu.c down_read(&pd->driver->sem); pd 733 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_alloc_map_lock(pd, addr); pd 751 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 752 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush_ptes(pd, f_address, num_pages, pd 755 drivers/gpu/drm/gma500/mmu.c up_read(&pd->driver->sem); pd 757 drivers/gpu/drm/gma500/mmu.c if (pd->hw_context != -1) pd 758 drivers/gpu/drm/gma500/mmu.c psb_mmu_flush(pd->driver); pd 763 drivers/gpu/drm/gma500/mmu.c int psb_mmu_virtual_to_pfn(struct psb_mmu_pd *pd, uint32_t virtual, pd 769 drivers/gpu/drm/gma500/mmu.c spinlock_t *lock = &pd->driver->lock; pd 771 drivers/gpu/drm/gma500/mmu.c down_read(&pd->driver->sem); pd 772 drivers/gpu/drm/gma500/mmu.c pt = psb_mmu_pt_map_lock(pd, virtual); pd 777 drivers/gpu/drm/gma500/mmu.c v = kmap_atomic(pd->p); pd 782 drivers/gpu/drm/gma500/mmu.c if (tmp != pd->invalid_pde || !(tmp & PSB_PTE_VALID) || pd 783 drivers/gpu/drm/gma500/mmu.c !(pd->invalid_pte & PSB_PTE_VALID)) { pd 788 drivers/gpu/drm/gma500/mmu.c *pfn = pd->invalid_pte >> PAGE_SHIFT; pd 800 drivers/gpu/drm/gma500/mmu.c up_read(&pd->driver->sem); pd 36 drivers/gpu/drm/gma500/mmu.h struct psb_mmu_pd *pd; pd 65 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_free_pagedir(struct psb_mmu_pd *pd); pd 67 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_remove_pfn_sequence(struct psb_mmu_pd *pd, pd 70 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_insert_pfn_sequence(struct psb_mmu_pd *pd, pd 74 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_virtual_to_pfn(struct psb_mmu_pd *pd, uint32_t virtual, pd 76 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_set_pd_context(struct psb_mmu_pd *pd, int hw_context); pd 77 drivers/gpu/drm/gma500/mmu.h extern int psb_mmu_insert_pages(struct psb_mmu_pd *pd, struct page **pages, pd 81 drivers/gpu/drm/gma500/mmu.h extern void psb_mmu_remove_pages(struct psb_mmu_pd *pd, pd 994 drivers/gpu/drm/i915/gem/i915_gem_context.c const dma_addr_t pd_daddr = px_dma(ppgtt->pd); pd 58 drivers/gpu/drm/i915/gt/intel_lrc_reg.h const u64 addr__ = px_dma(ppgtt->pd); \ pd 1545 drivers/gpu/drm/i915/gt/intel_ringbuffer.c *cs++ = px_base(ppgtt->pd)->ggtt_offset << 10; pd 372 drivers/gpu/drm/i915/gvt/scheduler.c px_dma(ppgtt->pd) = mm->ppgtt_mm.shadow_pdps[0]; pd 375 drivers/gpu/drm/i915/gvt/scheduler.c struct i915_page_directory * const pd = pd 376 drivers/gpu/drm/i915/gvt/scheduler.c i915_pd_entry(ppgtt->pd, i); pd 380 drivers/gpu/drm/i915/gvt/scheduler.c if (!pd) pd 382 drivers/gpu/drm/i915/gvt/scheduler.c px_dma(pd) = mm->ppgtt_mm.shadow_pdps[i]; pd 1152 drivers/gpu/drm/i915/gvt/scheduler.c px_dma(ppgtt->pd) = s->i915_context_pml4; pd 1155 drivers/gpu/drm/i915/gvt/scheduler.c struct i915_page_directory * const pd = pd 1156 drivers/gpu/drm/i915/gvt/scheduler.c i915_pd_entry(ppgtt->pd, i); pd 1158 drivers/gpu/drm/i915/gvt/scheduler.c px_dma(pd) = s->i915_context_pdps[i]; pd 1213 drivers/gpu/drm/i915/gvt/scheduler.c s->i915_context_pml4 = px_dma(ppgtt->pd); pd 1216 drivers/gpu/drm/i915/gvt/scheduler.c struct i915_page_directory * const pd = pd 1217 drivers/gpu/drm/i915/gvt/scheduler.c i915_pd_entry(ppgtt->pd, i); pd 1219 drivers/gpu/drm/i915/gvt/scheduler.c s->i915_context_pdps[i] = px_dma(pd); pd 724 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd; pd 726 drivers/gpu/drm/i915/i915_gem_gtt.c pd = kzalloc(sz, I915_GFP_ALLOW_FAIL); pd 727 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(!pd)) pd 730 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock_init(&pd->lock); pd 731 drivers/gpu/drm/i915/i915_gem_gtt.c return pd; pd 736 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd; pd 738 drivers/gpu/drm/i915/i915_gem_gtt.c pd = __alloc_pd(sizeof(*pd)); pd 739 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(!pd)) pd 742 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(setup_page_dma(vm, px_base(pd)))) { pd 743 drivers/gpu/drm/i915/i915_gem_gtt.c kfree(pd); pd 747 drivers/gpu/drm/i915/i915_gem_gtt.c return pd; pd 750 drivers/gpu/drm/i915/i915_gem_gtt.c static void free_pd(struct i915_address_space *vm, struct i915_page_dma *pd) pd 752 drivers/gpu/drm/i915/i915_gem_gtt.c cleanup_page_dma(vm, pd); pd 753 drivers/gpu/drm/i915/i915_gem_gtt.c kfree(pd); pd 770 drivers/gpu/drm/i915/i915_gem_gtt.c __set_pd_entry(struct i915_page_directory * const pd, pd 776 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(px_used(pd)) > 2 * ARRAY_SIZE(pd->entry)); pd 778 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_inc(px_used(pd)); pd 779 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = to; pd 780 drivers/gpu/drm/i915/i915_gem_gtt.c write_dma_entry(px_base(pd), idx, encode(to->daddr, I915_CACHE_LLC)); pd 783 drivers/gpu/drm/i915/i915_gem_gtt.c #define set_pd_entry(pd, idx, to) \ pd 784 drivers/gpu/drm/i915/i915_gem_gtt.c __set_pd_entry((pd), (idx), px_base(to), gen8_pde_encode) pd 787 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(struct i915_page_directory * const pd, pd 791 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(atomic_read(px_used(pd)) == 0); pd 793 drivers/gpu/drm/i915/i915_gem_gtt.c write_dma_entry(px_base(pd), idx, scratch->encode); pd 794 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[idx] = NULL; pd 795 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_dec(px_used(pd)); pd 799 drivers/gpu/drm/i915/i915_gem_gtt.c release_pd_entry(struct i915_page_directory * const pd, pd 809 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 811 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(pd, idx, scratch); pd 814 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 837 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_inc(px_used(ppgtt->pd)); /* never remove */ pd 839 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_dec(px_used(ppgtt->pd)); pd 844 drivers/gpu/drm/i915/i915_gem_gtt.c const u64 daddr = px_dma(ppgtt->pd); pd 923 drivers/gpu/drm/i915/i915_gem_gtt.c return ppgtt->pd; pd 925 drivers/gpu/drm/i915/i915_gem_gtt.c return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm->top)); pd 935 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd, pd 939 drivers/gpu/drm/i915/i915_gem_gtt.c void **pde = pd->entry; pd 949 drivers/gpu/drm/i915/i915_gem_gtt.c free_px(vm, pd); pd 959 drivers/gpu/drm/i915/i915_gem_gtt.c __gen8_ppgtt_cleanup(vm, ppgtt->pd, gen8_pd_top_count(vm), vm->top); pd 964 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd, pd 975 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); pd 976 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!len || len >= atomic_read(px_used(pd))); pd 979 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; pd 985 drivers/gpu/drm/i915/i915_gem_gtt.c clear_pd_entry(pd, idx, scratch); pd 1015 drivers/gpu/drm/i915/i915_gem_gtt.c if (release_pd_entry(pd, idx, pt, scratch)) pd 1033 drivers/gpu/drm/i915/i915_gem_gtt.c __gen8_ppgtt_clear(vm, i915_vm_to_ppgtt(vm)->pd, pd 1038 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd, pd 1051 drivers/gpu/drm/i915/i915_gem_gtt.c idx, len, atomic_read(px_used(pd))); pd 1054 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 1055 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(!atomic_read(px_used(pd))); /* Must be pinned! */ pd 1057 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_table *pt = pd->entry[idx]; pd 1060 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 1090 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 1091 drivers/gpu/drm/i915/i915_gem_gtt.c if (likely(!pd->entry[idx])) pd 1092 drivers/gpu/drm/i915/i915_gem_gtt.c set_pd_entry(pd, idx, pt); pd 1094 drivers/gpu/drm/i915/i915_gem_gtt.c alloc = pt, pt = pd->entry[idx]; pd 1099 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 1104 drivers/gpu/drm/i915/i915_gem_gtt.c if (release_pd_entry(pd, idx, pt, scratch)) pd 1109 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 1126 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 1148 drivers/gpu/drm/i915/i915_gem_gtt.c err = __gen8_ppgtt_alloc(vm, i915_vm_to_ppgtt(vm)->pd, pd 1151 drivers/gpu/drm/i915/i915_gem_gtt.c __gen8_ppgtt_clear(vm, i915_vm_to_ppgtt(vm)->pd, pd 1174 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd; pd 1178 drivers/gpu/drm/i915/i915_gem_gtt.c pd = i915_pd_entry(pdp, gen8_pd_index(idx, 2)); pd 1179 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1))); pd 1202 drivers/gpu/drm/i915/i915_gem_gtt.c pd = pdp->entry[gen8_pd_index(idx, 2)]; pd 1206 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, gen8_pd_index(idx, 1))); pd 1228 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = pd 1244 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(pd); pd 1247 drivers/gpu/drm/i915/i915_gem_gtt.c i915_pt_entry(pd, __gen8_pte_index(start, 1)); pd 1303 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(pd); pd 1321 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, maybe_64K)); pd 1409 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd = ppgtt->pd; pd 1423 drivers/gpu/drm/i915/i915_gem_gtt.c set_pd_entry(pd, idx, pde); pd 1451 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pd; pd 1453 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(count > ARRAY_SIZE(pd->entry)); pd 1455 drivers/gpu/drm/i915/i915_gem_gtt.c pd = __alloc_pd(offsetof(typeof(*pd), entry[count])); pd 1456 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(!pd)) pd 1459 drivers/gpu/drm/i915/i915_gem_gtt.c if (unlikely(setup_page_dma(vm, px_base(pd)))) { pd 1460 drivers/gpu/drm/i915/i915_gem_gtt.c kfree(pd); pd 1464 drivers/gpu/drm/i915/i915_gem_gtt.c fill_page_dma(px_base(pd), vm->scratch[vm->top].encode, count); pd 1465 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_inc(px_used(pd)); /* mark as pinned */ pd 1466 drivers/gpu/drm/i915/i915_gem_gtt.c return pd; pd 1506 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt->pd = gen8_alloc_top_pd(&ppgtt->vm); pd 1507 drivers/gpu/drm/i915/i915_gem_gtt.c if (IS_ERR(ppgtt->pd)) { pd 1508 drivers/gpu/drm/i915/i915_gem_gtt.c err = PTR_ERR(ppgtt->pd); pd 1532 drivers/gpu/drm/i915/i915_gem_gtt.c __gen8_ppgtt_cleanup(&ppgtt->vm, ppgtt->pd, pd 1616 drivers/gpu/drm/i915/i915_gem_gtt.c i915_pt_entry(ppgtt->base.pd, pde++); pd 1649 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = ppgtt->pd; pd 1657 drivers/gpu/drm/i915/i915_gem_gtt.c GEM_BUG_ON(pd->entry[act_pt] == &vm->scratch[1]); pd 1659 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, act_pt)); pd 1676 drivers/gpu/drm/i915/i915_gem_gtt.c vaddr = kmap_atomic_px(i915_pt_entry(pd, ++act_pt)); pd 1689 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = ppgtt->base.pd; pd 1699 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 1700 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_each_pde(pt, pd, start, length, pde) { pd 1704 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 1716 drivers/gpu/drm/i915/i915_gem_gtt.c spin_lock(&pd->lock); pd 1717 drivers/gpu/drm/i915/i915_gem_gtt.c if (pd->entry[pde] == &vm->scratch[1]) { pd 1718 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = pt; pd 1726 drivers/gpu/drm/i915/i915_gem_gtt.c pt = pd->entry[pde]; pd 1732 drivers/gpu/drm/i915/i915_gem_gtt.c spin_unlock(&pd->lock); pd 1753 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = ppgtt->base.pd; pd 1770 drivers/gpu/drm/i915/i915_gem_gtt.c memset_p(pd->entry, &vm->scratch[1], I915_PDES); pd 1777 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = ppgtt->base.pd; pd 1783 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, pd, pde) pd 1800 drivers/gpu/drm/i915/i915_gem_gtt.c kfree(ppgtt->base.pd); pd 1826 drivers/gpu/drm/i915/i915_gem_gtt.c px_base(ppgtt->base.pd)->ggtt_offset = ggtt_offset * sizeof(gen6_pte_t); pd 1829 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, ppgtt->base.pd, pde) pd 1841 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory * const pd = ppgtt->base.pd; pd 1851 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, ppgtt->base.pd, pde) { pd 1856 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = scratch; pd 1979 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt->base.pd = __alloc_pd(sizeof(*ppgtt->base.pd)); pd 1980 drivers/gpu/drm/i915/i915_gem_gtt.c if (!ppgtt->base.pd) { pd 2000 drivers/gpu/drm/i915/i915_gem_gtt.c kfree(ppgtt->base.pd); pd 426 drivers/gpu/drm/i915/i915_gem_gtt.h struct i915_page_directory *pd; pd 455 drivers/gpu/drm/i915/i915_gem_gtt.h #define gen6_for_each_pde(pt, pd, start, length, iter) \ pd 458 drivers/gpu/drm/i915/i915_gem_gtt.h (pt = i915_pt_entry(pd, iter), true); \ pd 463 drivers/gpu/drm/i915/i915_gem_gtt.h #define gen6_for_all_pdes(pt, pd, iter) \ pd 466 drivers/gpu/drm/i915/i915_gem_gtt.h (pt = i915_pt_entry(pd, iter), true); \ pd 517 drivers/gpu/drm/i915/i915_gem_gtt.h i915_pt_entry(const struct i915_page_directory * const pd, pd 520 drivers/gpu/drm/i915/i915_gem_gtt.h return pd->entry[n]; pd 533 drivers/gpu/drm/i915/i915_gem_gtt.h struct i915_page_dma *pt = ppgtt->pd->entry[n]; pd 92 drivers/gpu/drm/lima/lima_mmu.c mmu_write(LIMA_MMU_DTE_ADDR, dev->empty_vm->pd.dma); pd 113 drivers/gpu/drm/lima/lima_mmu.c mmu_write(LIMA_MMU_DTE_ADDR, vm->pd.dma); pd 137 drivers/gpu/drm/lima/lima_mmu.c mmu_write(LIMA_MMU_DTE_ADDR, dev->empty_vm->pd.dma); pd 59 drivers/gpu/drm/lima/lima_vm.c u32 *pd; pd 72 drivers/gpu/drm/lima/lima_vm.c pd = vm->pd.cpu + (pbe << LIMA_VM_NUM_PT_PER_BT_SHIFT); pd 74 drivers/gpu/drm/lima/lima_vm.c pd[j] = pts | LIMA_VM_FLAG_PRESENT; pd 212 drivers/gpu/drm/lima/lima_vm.c vm->pd.cpu = dma_alloc_wc(dev->dev, LIMA_PAGE_SIZE, &vm->pd.dma, pd 214 drivers/gpu/drm/lima/lima_vm.c if (!vm->pd.cpu) pd 230 drivers/gpu/drm/lima/lima_vm.c dma_free_wc(dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma); pd 249 drivers/gpu/drm/lima/lima_vm.c if (vm->pd.cpu) pd 250 drivers/gpu/drm/lima/lima_vm.c dma_free_wc(vm->dev->dev, LIMA_PAGE_SIZE, vm->pd.cpu, vm->pd.dma); pd 258 drivers/gpu/drm/lima/lima_vm.c u32 *pd, *pt; pd 260 drivers/gpu/drm/lima/lima_vm.c if (!vm->pd.cpu) pd 263 drivers/gpu/drm/lima/lima_vm.c pd = vm->pd.cpu; pd 272 drivers/gpu/drm/lima/lima_vm.c printk(KERN_INFO "lima vm pd %03x:%08x\n", idx, pd[idx]); pd 37 drivers/gpu/drm/lima/lima_vm.h struct lima_vm_page pd; pd 157 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c s32 hb_en, hb_en_ckln, pd_ckln, pd; pd 171 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c pd = timing->hs_prep_dly; pd 173 drivers/gpu/drm/msm/dsi/phy/dsi_phy.c val = (hb_en << 2) + (pd << 1); pd 166 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c static int pll_get_post_div(struct hdmi_8996_post_divider *pd, u64 bclk) pd 207 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c pd->vco_freq = vco_optimal; pd 208 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c pd->tx_band_sel = tx_band_sel[vco_optimal_index % 4]; pd 209 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c pd->vco_ratio = ratio[vco_optimal_index / 4]; pd 210 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c pd->hsclk_divsel = hs_divsel[vco_optimal_index / 4]; pd 221 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c struct hdmi_8996_post_divider pd; pd 244 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c ret = pll_get_post_div(&pd, bclk); pd 248 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c dec_start = pd.vco_freq; pd 252 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c frac_start = pd.vco_freq * (1 << 20); pd 265 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c fdata = pd.vco_freq; pd 266 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c do_div(fdata, pd.vco_ratio); pd 270 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c DBG("VCO freq: %llu", pd.vco_freq); pd 274 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c DBG("HSCLK_SEL: %d", pd.hsclk_divsel); pd 281 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c DBG("TX_BAND: %d", pd.tx_band_sel); pd 290 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c cfg->com_hsclk_sel = (0x20 | pd.hsclk_divsel); pd 318 drivers/gpu/drm/msm/hdmi/hdmi_phy_8996.c cfg->tx_lx_tx_band[i] = pd.tx_band_sel + 4; pd 36 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h struct nvkm_vmm_pt *pd; pd 53 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv04.c device->mmu->vmm->pd->pt[0]->memory; pd 37 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c int pd, lf, xd, vh, vr, vo, l3; pd 45 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c pd = ram->next->bios.ramcfg_11_01_80; pd 104 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c pd = 1; /* binary driver does this.. bug? */ pd 107 drivers/gpu/drm/nouveau/nvkm/subdev/fb/gddr5.c ram->mr[6] |= (pd & 0x01) << 0; pd 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv41.c nvkm_wr32(device, 0x100800, 0x00000002 | mmu->vmm->pd->pt[0]->addr); pd 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/nv44.c struct nvkm_memory *pt = mmu->vmm->pd->pt[0]->memory; pd 523 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pt[it.max] = vmm->pd; pd 1012 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vmm->pd) { pd 1013 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_mmu_ptc_put(vmm->mmu, true, &vmm->pd->pt[0]); pd 1014 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c nvkm_vmm_pt_del(&vmm->pd); pd 1072 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vmm->pd = nvkm_vmm_pt_new(desc, false, NULL); pd 1073 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!vmm->pd) pd 1075 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vmm->pd->refs[0] = 1; pd 1083 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vmm->pd->pt[0] = nvkm_mmu_ptc_get(mmu, size, desc->align, true); pd 1084 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!vmm->pd->pt[0]) pd 109 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c struct nvkm_mmu_pt *pd = pgd->pt[0]; pd 141 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c nvkm_kmap(pd->memory); pd 142 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c VMM_WO064(pd, vmm, pdei * 8, data); pd 143 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c nvkm_done(pd->memory); pd 192 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c struct nvkm_mmu_pt *pd = vmm->pd->pt[0]; pd 205 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c switch (nvkm_memory_target(pd->memory)) { pd 213 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c addr |= (vmm->pd->pt[0]->addr >> 12) << 4; pd 343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c struct nvkm_mmu_pt *pd = vmm->pd->pt[0]; pd 345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c switch (nvkm_memory_target(pd->memory)) { pd 355 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c base |= pd->addr; pd 233 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct nvkm_mmu_pt *pd = pgd->pt[0]; pd 241 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c nvkm_kmap(pd->memory); pd 242 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c VMM_WO128(pd, vmm, pdei * 0x10, data[0], data[1]); pd 243 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c nvkm_done(pd->memory); pd 273 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct nvkm_mmu_pt *pd = pgd->pt[0]; pd 279 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c nvkm_kmap(pd->memory); pd 280 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c VMM_WO064(pd, vmm, pdei * 8, data); pd 281 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c nvkm_done(pd->memory); pd 135 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv04.c mem = vmm->pd->pt[0]->memory; pd 354 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c if (!nv50_vmm_pde(vmm, vmm->pd->pde[pdei], &data)) { pd 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c nvkm_wr32(device, 0xb830a0, vmm->pd->pt[0]->addr >> 8); pd 644 drivers/gpu/drm/radeon/radeon_vm.c struct radeon_bo *pd = vm->page_directory; pd 645 drivers/gpu/drm/radeon/radeon_vm.c uint64_t pd_addr = radeon_bo_gpu_offset(pd); pd 705 drivers/gpu/drm/radeon/radeon_vm.c radeon_sync_resv(rdev, &ib.sync, pd->tbo.base.resv, true); pd 713 drivers/gpu/drm/radeon/radeon_vm.c radeon_bo_fence(pd, ib.fence, false); pd 259 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c static u16 sun6i_dsi_crc_repeat(u8 pd, u8 *buffer, size_t len) pd 261 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c memset(buffer, pd, len); pd 284 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c static u32 sun6i_dsi_build_blk1_pkt(u16 pd, u8 *buffer, size_t len) pd 286 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c u32 val = SUN6I_DSI_BLK_PD(pd); pd 288 drivers/gpu/drm/sun4i/sun6i_mipi_dsi.c return val | SUN6I_DSI_BLK_PF(sun6i_dsi_crc_repeat(pd, buffer, len)); pd 326 drivers/hsi/controllers/omap_ssi_core.c static int ssi_get_iomem(struct platform_device *pd, pd 331 drivers/hsi/controllers/omap_ssi_core.c struct hsi_controller *ssi = platform_get_drvdata(pd); pd 333 drivers/hsi/controllers/omap_ssi_core.c mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name); pd 347 drivers/hsi/controllers/omap_ssi_core.c struct platform_device *pd) pd 362 drivers/hsi/controllers/omap_ssi_core.c ssi->device.parent = &pd->dev; pd 366 drivers/hsi/controllers/omap_ssi_core.c err = ssi_get_iomem(pd, "sys", &omap_ssi->sys, NULL); pd 369 drivers/hsi/controllers/omap_ssi_core.c err = ssi_get_iomem(pd, "gdd", &omap_ssi->gdd, NULL); pd 372 drivers/hsi/controllers/omap_ssi_core.c err = platform_get_irq_byname(pd, "gdd_mpu"); pd 395 drivers/hsi/controllers/omap_ssi_core.c dev_err(&pd->dev, "Could not acquire clock \"ssi_ssr_fck\": %li\n", pd 480 drivers/hsi/controllers/omap_ssi_core.c static int ssi_probe(struct platform_device *pd) pd 483 drivers/hsi/controllers/omap_ssi_core.c struct device_node *np = pd->dev.of_node; pd 490 drivers/hsi/controllers/omap_ssi_core.c dev_err(&pd->dev, "missing device tree data\n"); pd 498 drivers/hsi/controllers/omap_ssi_core.c dev_err(&pd->dev, "No memory for controller\n"); pd 502 drivers/hsi/controllers/omap_ssi_core.c platform_set_drvdata(pd, ssi); pd 504 drivers/hsi/controllers/omap_ssi_core.c err = ssi_add_controller(ssi, pd); pd 508 drivers/hsi/controllers/omap_ssi_core.c pm_runtime_enable(&pd->dev); pd 523 drivers/hsi/controllers/omap_ssi_core.c childpdev = of_platform_device_create(child, NULL, &pd->dev); pd 526 drivers/hsi/controllers/omap_ssi_core.c dev_err(&pd->dev, "failed to create ssi controller port\n"); pd 531 drivers/hsi/controllers/omap_ssi_core.c dev_info(&pd->dev, "ssi controller %d initialized (%d ports)!\n", pd 535 drivers/hsi/controllers/omap_ssi_core.c device_for_each_child(&pd->dev, NULL, ssi_remove_ports); pd 539 drivers/hsi/controllers/omap_ssi_core.c platform_set_drvdata(pd, NULL); pd 540 drivers/hsi/controllers/omap_ssi_core.c pm_runtime_disable(&pd->dev); pd 545 drivers/hsi/controllers/omap_ssi_core.c static int ssi_remove(struct platform_device *pd) pd 547 drivers/hsi/controllers/omap_ssi_core.c struct hsi_controller *ssi = platform_get_drvdata(pd); pd 550 drivers/hsi/controllers/omap_ssi_core.c device_for_each_child(&pd->dev, NULL, ssi_remove_ports); pd 556 drivers/hsi/controllers/omap_ssi_core.c platform_set_drvdata(pd, NULL); pd 558 drivers/hsi/controllers/omap_ssi_core.c pm_runtime_disable(&pd->dev); pd 1035 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_irq(struct hsi_port *port, struct platform_device *pd) pd 1040 drivers/hsi/controllers/omap_ssi_port.c err = platform_get_irq(pd, 0); pd 1052 drivers/hsi/controllers/omap_ssi_port.c static int ssi_wake_irq(struct hsi_port *port, struct platform_device *pd) pd 1092 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_get_iomem(struct platform_device *pd, pd 1095 drivers/hsi/controllers/omap_ssi_port.c struct hsi_port *port = platform_get_drvdata(pd); pd 1100 drivers/hsi/controllers/omap_ssi_port.c mem = platform_get_resource_byname(pd, IORESOURCE_MEM, name); pd 1102 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "IO memory region missing (%s)\n", name); pd 1106 drivers/hsi/controllers/omap_ssi_port.c resource_size(mem), dev_name(&pd->dev)); pd 1108 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "%s IO memory region request failed\n", pd 1114 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "%s IO remap failed\n", mem->name); pd 1125 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_probe(struct platform_device *pd) pd 1127 drivers/hsi/controllers/omap_ssi_port.c struct device_node *np = pd->dev.of_node; pd 1130 drivers/hsi/controllers/omap_ssi_port.c struct hsi_controller *ssi = dev_get_drvdata(pd->dev.parent); pd 1136 drivers/hsi/controllers/omap_ssi_port.c dev_dbg(&pd->dev, "init ssi port...\n"); pd 1139 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "ssi controller not initialized!\n"); pd 1150 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "port id out of range!\n"); pd 1158 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "missing device tree data\n"); pd 1163 drivers/hsi/controllers/omap_ssi_port.c cawake_gpio = devm_gpiod_get(&pd->dev, "ti,ssi-cawake", GPIOD_IN); pd 1166 drivers/hsi/controllers/omap_ssi_port.c dev_err(&pd->dev, "couldn't get cawake gpio (err=%d)!\n", err); pd 1176 drivers/hsi/controllers/omap_ssi_port.c omap_port->pdev = &pd->dev; pd 1192 drivers/hsi/controllers/omap_ssi_port.c platform_set_drvdata(pd, port); pd 1194 drivers/hsi/controllers/omap_ssi_port.c err = ssi_port_get_iomem(pd, "tx", &omap_port->sst_base, pd 1198 drivers/hsi/controllers/omap_ssi_port.c err = ssi_port_get_iomem(pd, "rx", &omap_port->ssr_base, pd 1203 drivers/hsi/controllers/omap_ssi_port.c err = ssi_port_irq(port, pd); pd 1206 drivers/hsi/controllers/omap_ssi_port.c err = ssi_wake_irq(port, pd); pd 1229 drivers/hsi/controllers/omap_ssi_port.c dev_info(&pd->dev, "ssi port %u successfully initialized\n", port_id); pd 1237 drivers/hsi/controllers/omap_ssi_port.c static int ssi_port_remove(struct platform_device *pd) pd 1239 drivers/hsi/controllers/omap_ssi_port.c struct hsi_port *port = platform_get_drvdata(pd); pd 1260 drivers/hsi/controllers/omap_ssi_port.c platform_set_drvdata(pd, NULL); pd 1262 drivers/hsi/controllers/omap_ssi_port.c pm_runtime_dont_use_autosuspend(&pd->dev); pd 1263 drivers/hsi/controllers/omap_ssi_port.c pm_runtime_disable(&pd->dev); pd 648 drivers/hwmon/coretemp.c struct platform_data *pd; pd 668 drivers/hwmon/coretemp.c pd = platform_get_drvdata(pdev); pd 669 drivers/hwmon/coretemp.c tdata = pd->core_data[indx]; pd 671 drivers/hwmon/coretemp.c cpumask_clear_cpu(cpu, &pd->cpumask); pd 678 drivers/hwmon/coretemp.c target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu)); pd 680 drivers/hwmon/coretemp.c coretemp_remove_core(pd, indx); pd 692 drivers/hwmon/coretemp.c if (cpumask_empty(&pd->cpumask)) { pd 702 drivers/hwmon/coretemp.c tdata = pd->core_data[PKG_SYSFS_ATTR_NO]; pd 704 drivers/hwmon/coretemp.c target = cpumask_first(&pd->cpumask); pd 876 drivers/i2c/busses/i2c-mv64xxx.c mv64xxx_i2c_probe(struct platform_device *pd) pd 879 drivers/i2c/busses/i2c-mv64xxx.c struct mv64xxx_i2c_pdata *pdata = dev_get_platdata(&pd->dev); pd 883 drivers/i2c/busses/i2c-mv64xxx.c if ((!pdata && !pd->dev.of_node)) pd 886 drivers/i2c/busses/i2c-mv64xxx.c drv_data = devm_kzalloc(&pd->dev, sizeof(struct mv64xxx_i2c_data), pd 891 drivers/i2c/busses/i2c-mv64xxx.c r = platform_get_resource(pd, IORESOURCE_MEM, 0); pd 892 drivers/i2c/busses/i2c-mv64xxx.c drv_data->reg_base = devm_ioremap_resource(&pd->dev, r); pd 903 drivers/i2c/busses/i2c-mv64xxx.c drv_data->clk = devm_clk_get(&pd->dev, NULL); pd 909 drivers/i2c/busses/i2c-mv64xxx.c drv_data->reg_clk = devm_clk_get(&pd->dev, "reg"); pd 916 drivers/i2c/busses/i2c-mv64xxx.c drv_data->irq = platform_get_irq(pd, 0); pd 924 drivers/i2c/busses/i2c-mv64xxx.c } else if (pd->dev.of_node) { pd 925 drivers/i2c/busses/i2c-mv64xxx.c rc = mv64xxx_of_config(drv_data, &pd->dev); pd 934 drivers/i2c/busses/i2c-mv64xxx.c drv_data->adapter.dev.parent = &pd->dev; pd 938 drivers/i2c/busses/i2c-mv64xxx.c drv_data->adapter.nr = pd->id; pd 939 drivers/i2c/busses/i2c-mv64xxx.c drv_data->adapter.dev.of_node = pd->dev.of_node; pd 940 drivers/i2c/busses/i2c-mv64xxx.c platform_set_drvdata(pd, drv_data); pd 37 drivers/i2c/busses/i2c-pca-isa.c static void pca_isa_writebyte(void *pd, int reg, int val) pd 47 drivers/i2c/busses/i2c-pca-isa.c static int pca_isa_readbyte(void *pd, int reg) pd 59 drivers/i2c/busses/i2c-pca-isa.c static int pca_isa_waitforcompletion(void *pd) pd 66 drivers/i2c/busses/i2c-pca-isa.c pca_isa_readbyte(pd, I2C_PCA_CON) pd 73 drivers/i2c/busses/i2c-pca-isa.c if (pca_isa_readbyte(pd, I2C_PCA_CON) pd 83 drivers/i2c/busses/i2c-pca-isa.c static void pca_isa_resetchip(void *pd) pd 42 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte8(void *pd, int reg) pd 44 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 48 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte16(void *pd, int reg) pd 50 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 54 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_readbyte32(void *pd, int reg) pd 56 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 60 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte8(void *pd, int reg, int val) pd 62 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 66 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte16(void *pd, int reg, int val) pd 68 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 72 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_writebyte32(void *pd, int reg, int val) pd 74 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 79 drivers/i2c/busses/i2c-pca-platform.c static int i2c_pca_pf_waitforcompletion(void *pd) pd 81 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 104 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_dummyreset(void *pd) pd 106 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 111 drivers/i2c/busses/i2c-pca-platform.c static void i2c_pca_pf_resetchip(void *pd) pd 113 drivers/i2c/busses/i2c-pca-platform.c struct i2c_pca_pf_data *i2c = pd; pd 434 drivers/i2c/busses/i2c-sh7760.c struct sh7760_i2c_platdata *pd; pd 439 drivers/i2c/busses/i2c-sh7760.c pd = dev_get_platdata(&pdev->dev); pd 440 drivers/i2c/busses/i2c-sh7760.c if (!pd) { pd 497 drivers/i2c/busses/i2c-sh7760.c ret = calc_CCR(pd->speed_khz * 1000); pd 500 drivers/i2c/busses/i2c-sh7760.c pd->speed_khz); pd 519 drivers/i2c/busses/i2c-sh7760.c pd->speed_khz, res->start, id->irq); pd 143 drivers/i2c/busses/i2c-sh_mobile.c int (*setup)(struct sh_mobile_i2c_data *pd); pd 187 drivers/i2c/busses/i2c-sh_mobile.c static void iic_wr(struct sh_mobile_i2c_data *pd, int offs, unsigned char data) pd 190 drivers/i2c/busses/i2c-sh_mobile.c data |= pd->icic; pd 192 drivers/i2c/busses/i2c-sh_mobile.c iowrite8(data, pd->reg + offs); pd 195 drivers/i2c/busses/i2c-sh_mobile.c static unsigned char iic_rd(struct sh_mobile_i2c_data *pd, int offs) pd 197 drivers/i2c/busses/i2c-sh_mobile.c return ioread8(pd->reg + offs); pd 200 drivers/i2c/busses/i2c-sh_mobile.c static void iic_set_clr(struct sh_mobile_i2c_data *pd, int offs, pd 203 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, offs, (iic_rd(pd, offs) | set) & ~clr); pd 240 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_check_timing(struct sh_mobile_i2c_data *pd) pd 242 drivers/i2c/busses/i2c-sh_mobile.c u16 max_val = pd->flags & IIC_FLAG_HAS_ICIC67 ? 0x1ff : 0xff; pd 244 drivers/i2c/busses/i2c-sh_mobile.c if (pd->iccl > max_val || pd->icch > max_val) { pd 245 drivers/i2c/busses/i2c-sh_mobile.c dev_err(pd->dev, "timing values out of range: L/H=0x%x/0x%x\n", pd 246 drivers/i2c/busses/i2c-sh_mobile.c pd->iccl, pd->icch); pd 251 drivers/i2c/busses/i2c-sh_mobile.c if (pd->iccl & 0x100) pd 252 drivers/i2c/busses/i2c-sh_mobile.c pd->icic |= ICIC_ICCLB8; pd 254 drivers/i2c/busses/i2c-sh_mobile.c pd->icic &= ~ICIC_ICCLB8; pd 257 drivers/i2c/busses/i2c-sh_mobile.c if (pd->icch & 0x100) pd 258 drivers/i2c/busses/i2c-sh_mobile.c pd->icic |= ICIC_ICCHB8; pd 260 drivers/i2c/busses/i2c-sh_mobile.c pd->icic &= ~ICIC_ICCHB8; pd 262 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "timing values: L/H=0x%x/0x%x\n", pd->iccl, pd->icch); pd 266 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_init(struct sh_mobile_i2c_data *pd) pd 271 drivers/i2c/busses/i2c-sh_mobile.c i2c_clk_khz = clk_get_rate(pd->clk) / 1000 / pd->clks_per_count; pd 273 drivers/i2c/busses/i2c-sh_mobile.c if (pd->bus_speed == STANDARD_MODE) { pd 277 drivers/i2c/busses/i2c-sh_mobile.c } else if (pd->bus_speed == FAST_MODE) { pd 282 drivers/i2c/busses/i2c-sh_mobile.c dev_err(pd->dev, "unrecognized bus speed %lu Hz\n", pd 283 drivers/i2c/busses/i2c-sh_mobile.c pd->bus_speed); pd 287 drivers/i2c/busses/i2c-sh_mobile.c pd->iccl = sh_mobile_i2c_iccl(i2c_clk_khz, tLOW, tf); pd 288 drivers/i2c/busses/i2c-sh_mobile.c pd->icch = sh_mobile_i2c_icch(i2c_clk_khz, tHIGH, tf); pd 290 drivers/i2c/busses/i2c-sh_mobile.c return sh_mobile_i2c_check_timing(pd); pd 293 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_v2_init(struct sh_mobile_i2c_data *pd) pd 298 drivers/i2c/busses/i2c-sh_mobile.c clks_per_cycle = clk_get_rate(pd->clk) / pd->bus_speed; pd 299 drivers/i2c/busses/i2c-sh_mobile.c pd->iccl = DIV_ROUND_UP(clks_per_cycle * 5 / 9 - 1, pd->clks_per_count); pd 300 drivers/i2c/busses/i2c-sh_mobile.c pd->icch = DIV_ROUND_UP(clks_per_cycle * 4 / 9 - 5, pd->clks_per_count); pd 302 drivers/i2c/busses/i2c-sh_mobile.c return sh_mobile_i2c_check_timing(pd); pd 305 drivers/i2c/busses/i2c-sh_mobile.c static unsigned char i2c_op(struct sh_mobile_i2c_data *pd, enum sh_mobile_i2c_op op) pd 310 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "op %d\n", op); pd 312 drivers/i2c/busses/i2c-sh_mobile.c spin_lock_irqsave(&pd->lock, flags); pd 316 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_ICE | ICCR_TRS | ICCR_BBSY); pd 319 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICIC, ICIC_WAITE | ICIC_ALE | ICIC_TACKE); pd 320 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICDR, i2c_8bit_addr_from_msg(pd->msg)); pd 323 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICDR, pd->msg->buf[pd->pos]); pd 326 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, pd->send_stop ? ICCR_ICE | ICCR_TRS pd 330 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP); pd 333 drivers/i2c/busses/i2c-sh_mobile.c ret = iic_rd(pd, ICDR); pd 336 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICIC, pd 338 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK); pd 341 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICIC, pd 343 drivers/i2c/busses/i2c-sh_mobile.c ret = iic_rd(pd, ICDR); pd 344 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK); pd 348 drivers/i2c/busses/i2c-sh_mobile.c spin_unlock_irqrestore(&pd->lock, flags); pd 350 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "op %d, data out 0x%02x\n", op, ret); pd 354 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_isr_tx(struct sh_mobile_i2c_data *pd) pd 356 drivers/i2c/busses/i2c-sh_mobile.c if (pd->pos == pd->msg->len) { pd 357 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_TX_STOP); pd 361 drivers/i2c/busses/i2c-sh_mobile.c if (pd->pos == -1) pd 362 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_TX_FIRST); pd 364 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_TX); pd 366 drivers/i2c/busses/i2c-sh_mobile.c pd->pos++; pd 370 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_isr_rx(struct sh_mobile_i2c_data *pd) pd 376 drivers/i2c/busses/i2c-sh_mobile.c real_pos = pd->pos - 2; pd 378 drivers/i2c/busses/i2c-sh_mobile.c if (pd->pos == -1) { pd 379 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_TX_FIRST); pd 380 drivers/i2c/busses/i2c-sh_mobile.c } else if (pd->pos == 0) { pd 381 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_TX_TO_RX); pd 382 drivers/i2c/busses/i2c-sh_mobile.c } else if (pd->pos == pd->msg->len) { pd 383 drivers/i2c/busses/i2c-sh_mobile.c if (pd->stop_after_dma) { pd 385 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_RX_STOP); pd 386 drivers/i2c/busses/i2c-sh_mobile.c pd->pos++; pd 391 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_RX_STOP); pd 393 drivers/i2c/busses/i2c-sh_mobile.c data = i2c_op(pd, OP_RX_STOP_DATA); pd 395 drivers/i2c/busses/i2c-sh_mobile.c data = i2c_op(pd, OP_RX); pd 399 drivers/i2c/busses/i2c-sh_mobile.c pd->msg->buf[real_pos] = data; pd 401 drivers/i2c/busses/i2c-sh_mobile.c pd->pos++; pd 402 drivers/i2c/busses/i2c-sh_mobile.c return pd->pos == (pd->msg->len + 2); pd 407 drivers/i2c/busses/i2c-sh_mobile.c struct sh_mobile_i2c_data *pd = dev_id; pd 411 drivers/i2c/busses/i2c-sh_mobile.c sr = iic_rd(pd, ICSR); pd 412 drivers/i2c/busses/i2c-sh_mobile.c pd->sr |= sr; /* remember state */ pd 414 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "i2c_isr 0x%02x 0x%02x %s %d %d!\n", sr, pd->sr, pd 415 drivers/i2c/busses/i2c-sh_mobile.c (pd->msg->flags & I2C_M_RD) ? "read" : "write", pd 416 drivers/i2c/busses/i2c-sh_mobile.c pd->pos, pd->msg->len); pd 419 drivers/i2c/busses/i2c-sh_mobile.c if (pd->dma_direction == DMA_TO_DEVICE && pd->pos == 0) pd 420 drivers/i2c/busses/i2c-sh_mobile.c iic_set_clr(pd, ICIC, ICIC_TDMAE, 0); pd 423 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICSR, sr & ~(ICSR_AL | ICSR_TACK)); pd 424 drivers/i2c/busses/i2c-sh_mobile.c else if (pd->msg->flags & I2C_M_RD) pd 425 drivers/i2c/busses/i2c-sh_mobile.c wakeup = sh_mobile_i2c_isr_rx(pd); pd 427 drivers/i2c/busses/i2c-sh_mobile.c wakeup = sh_mobile_i2c_isr_tx(pd); pd 430 drivers/i2c/busses/i2c-sh_mobile.c if (pd->dma_direction == DMA_FROM_DEVICE && pd->pos == 1) pd 431 drivers/i2c/busses/i2c-sh_mobile.c iic_set_clr(pd, ICIC, ICIC_RDMAE, 0); pd 434 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICSR, sr & ~ICSR_WAIT); pd 437 drivers/i2c/busses/i2c-sh_mobile.c pd->sr |= SW_DONE; pd 438 drivers/i2c/busses/i2c-sh_mobile.c wake_up(&pd->wait); pd 442 drivers/i2c/busses/i2c-sh_mobile.c iic_rd(pd, ICSR); pd 447 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_dma_unmap(struct sh_mobile_i2c_data *pd) pd 449 drivers/i2c/busses/i2c-sh_mobile.c struct dma_chan *chan = pd->dma_direction == DMA_FROM_DEVICE pd 450 drivers/i2c/busses/i2c-sh_mobile.c ? pd->dma_rx : pd->dma_tx; pd 452 drivers/i2c/busses/i2c-sh_mobile.c dma_unmap_single(chan->device->dev, sg_dma_address(&pd->sg), pd 453 drivers/i2c/busses/i2c-sh_mobile.c pd->msg->len, pd->dma_direction); pd 455 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_direction = DMA_NONE; pd 458 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_cleanup_dma(struct sh_mobile_i2c_data *pd) pd 460 drivers/i2c/busses/i2c-sh_mobile.c if (pd->dma_direction == DMA_NONE) pd 462 drivers/i2c/busses/i2c-sh_mobile.c else if (pd->dma_direction == DMA_FROM_DEVICE) pd 463 drivers/i2c/busses/i2c-sh_mobile.c dmaengine_terminate_all(pd->dma_rx); pd 464 drivers/i2c/busses/i2c-sh_mobile.c else if (pd->dma_direction == DMA_TO_DEVICE) pd 465 drivers/i2c/busses/i2c-sh_mobile.c dmaengine_terminate_all(pd->dma_tx); pd 467 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_dma_unmap(pd); pd 472 drivers/i2c/busses/i2c-sh_mobile.c struct sh_mobile_i2c_data *pd = data; pd 474 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_dma_unmap(pd); pd 475 drivers/i2c/busses/i2c-sh_mobile.c pd->pos = pd->msg->len; pd 476 drivers/i2c/busses/i2c-sh_mobile.c pd->stop_after_dma = true; pd 478 drivers/i2c/busses/i2c-sh_mobile.c iic_set_clr(pd, ICIC, 0, ICIC_TDMAE | ICIC_RDMAE); pd 517 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_xfer_dma(struct sh_mobile_i2c_data *pd) pd 519 drivers/i2c/busses/i2c-sh_mobile.c bool read = pd->msg->flags & I2C_M_RD; pd 521 drivers/i2c/busses/i2c-sh_mobile.c struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; pd 528 drivers/i2c/busses/i2c-sh_mobile.c chan = pd->dma_rx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_DEV_TO_MEM, pd 529 drivers/i2c/busses/i2c-sh_mobile.c pd->res->start + ICDR); pd 531 drivers/i2c/busses/i2c-sh_mobile.c chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV, pd 532 drivers/i2c/busses/i2c-sh_mobile.c pd->res->start + ICDR); pd 538 drivers/i2c/busses/i2c-sh_mobile.c dma_addr = dma_map_single(chan->device->dev, pd->dma_buf, pd->msg->len, dir); pd 540 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "dma map failed, using PIO\n"); pd 544 drivers/i2c/busses/i2c-sh_mobile.c sg_dma_len(&pd->sg) = pd->msg->len; pd 545 drivers/i2c/busses/i2c-sh_mobile.c sg_dma_address(&pd->sg) = dma_addr; pd 547 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_direction = dir; pd 549 drivers/i2c/busses/i2c-sh_mobile.c txdesc = dmaengine_prep_slave_sg(chan, &pd->sg, 1, pd 553 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "dma prep slave sg failed, using PIO\n"); pd 554 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_cleanup_dma(pd); pd 559 drivers/i2c/busses/i2c-sh_mobile.c txdesc->callback_param = pd; pd 563 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "submitting dma failed, using PIO\n"); pd 564 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_cleanup_dma(pd); pd 571 drivers/i2c/busses/i2c-sh_mobile.c static void start_ch(struct sh_mobile_i2c_data *pd, struct i2c_msg *usr_msg, pd 576 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_SCP); pd 579 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP); pd 582 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCL, pd->iccl & 0xff); pd 583 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCH, pd->icch & 0xff); pd 586 drivers/i2c/busses/i2c-sh_mobile.c pd->msg = usr_msg; pd 587 drivers/i2c/busses/i2c-sh_mobile.c pd->pos = -1; pd 588 drivers/i2c/busses/i2c-sh_mobile.c pd->sr = 0; pd 590 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_buf = i2c_get_dma_safe_msg_buf(pd->msg, 8); pd 591 drivers/i2c/busses/i2c-sh_mobile.c if (pd->dma_buf) pd 592 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_xfer_dma(pd); pd 595 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICIC, ICIC_DTEE | ICIC_WAITE | ICIC_ALE | ICIC_TACKE); pd 598 drivers/i2c/busses/i2c-sh_mobile.c static int poll_dte(struct sh_mobile_i2c_data *pd) pd 603 drivers/i2c/busses/i2c-sh_mobile.c u_int8_t val = iic_rd(pd, ICSR); pd 617 drivers/i2c/busses/i2c-sh_mobile.c static int poll_busy(struct sh_mobile_i2c_data *pd) pd 622 drivers/i2c/busses/i2c-sh_mobile.c u_int8_t val = iic_rd(pd, ICSR); pd 624 drivers/i2c/busses/i2c-sh_mobile.c dev_dbg(pd->dev, "val 0x%02x pd->sr 0x%02x\n", val, pd->sr); pd 632 drivers/i2c/busses/i2c-sh_mobile.c val |= pd->sr; pd 650 drivers/i2c/busses/i2c-sh_mobile.c struct sh_mobile_i2c_data *pd = i2c_get_adapdata(adapter); pd 657 drivers/i2c/busses/i2c-sh_mobile.c pm_runtime_get_sync(pd->dev); pd 661 drivers/i2c/busses/i2c-sh_mobile.c bool do_start = pd->send_stop || !i; pd 663 drivers/i2c/busses/i2c-sh_mobile.c pd->send_stop = i == num - 1 || msg->flags & I2C_M_STOP; pd 664 drivers/i2c/busses/i2c-sh_mobile.c pd->stop_after_dma = false; pd 666 drivers/i2c/busses/i2c-sh_mobile.c start_ch(pd, msg, do_start); pd 669 drivers/i2c/busses/i2c-sh_mobile.c i2c_op(pd, OP_START); pd 672 drivers/i2c/busses/i2c-sh_mobile.c timeout = wait_event_timeout(pd->wait, pd 673 drivers/i2c/busses/i2c-sh_mobile.c pd->sr & (ICSR_TACK | SW_DONE), pd 677 drivers/i2c/busses/i2c-sh_mobile.c i2c_put_dma_safe_msg_buf(pd->dma_buf, pd->msg, pd->stop_after_dma); pd 680 drivers/i2c/busses/i2c-sh_mobile.c dev_err(pd->dev, "Transfer request timed out\n"); pd 681 drivers/i2c/busses/i2c-sh_mobile.c if (pd->dma_direction != DMA_NONE) pd 682 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_cleanup_dma(pd); pd 688 drivers/i2c/busses/i2c-sh_mobile.c if (pd->send_stop) pd 689 drivers/i2c/busses/i2c-sh_mobile.c err = poll_busy(pd); pd 691 drivers/i2c/busses/i2c-sh_mobile.c err = poll_dte(pd); pd 697 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_SCP); pd 700 drivers/i2c/busses/i2c-sh_mobile.c pm_runtime_put_sync(pd->dev); pd 722 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_r8a7740_workaround(struct sh_mobile_i2c_data *pd) pd 724 drivers/i2c/busses/i2c-sh_mobile.c iic_set_clr(pd, ICCR, ICCR_ICE, 0); pd 725 drivers/i2c/busses/i2c-sh_mobile.c iic_rd(pd, ICCR); /* dummy read */ pd 727 drivers/i2c/busses/i2c-sh_mobile.c iic_set_clr(pd, ICSTART, ICSTART_ICSTART, 0); pd 728 drivers/i2c/busses/i2c-sh_mobile.c iic_rd(pd, ICSTART); /* dummy read */ pd 732 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_SCP); pd 733 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICSTART, 0); pd 737 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_TRS); pd 739 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, 0); pd 741 drivers/i2c/busses/i2c-sh_mobile.c iic_wr(pd, ICCR, ICCR_TRS); pd 744 drivers/i2c/busses/i2c-sh_mobile.c return sh_mobile_i2c_init(pd); pd 786 drivers/i2c/busses/i2c-sh_mobile.c static void sh_mobile_i2c_release_dma(struct sh_mobile_i2c_data *pd) pd 788 drivers/i2c/busses/i2c-sh_mobile.c if (!IS_ERR(pd->dma_tx)) { pd 789 drivers/i2c/busses/i2c-sh_mobile.c dma_release_channel(pd->dma_tx); pd 790 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_tx = ERR_PTR(-EPROBE_DEFER); pd 793 drivers/i2c/busses/i2c-sh_mobile.c if (!IS_ERR(pd->dma_rx)) { pd 794 drivers/i2c/busses/i2c-sh_mobile.c dma_release_channel(pd->dma_rx); pd 795 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_rx = ERR_PTR(-EPROBE_DEFER); pd 799 drivers/i2c/busses/i2c-sh_mobile.c static int sh_mobile_i2c_hook_irqs(struct platform_device *dev, struct sh_mobile_i2c_data *pd) pd 808 drivers/i2c/busses/i2c-sh_mobile.c 0, dev_name(&dev->dev), pd); pd 822 drivers/i2c/busses/i2c-sh_mobile.c struct sh_mobile_i2c_data *pd; pd 829 drivers/i2c/busses/i2c-sh_mobile.c pd = devm_kzalloc(&dev->dev, sizeof(struct sh_mobile_i2c_data), GFP_KERNEL); pd 830 drivers/i2c/busses/i2c-sh_mobile.c if (!pd) pd 833 drivers/i2c/busses/i2c-sh_mobile.c pd->clk = devm_clk_get(&dev->dev, NULL); pd 834 drivers/i2c/busses/i2c-sh_mobile.c if (IS_ERR(pd->clk)) { pd 836 drivers/i2c/busses/i2c-sh_mobile.c return PTR_ERR(pd->clk); pd 839 drivers/i2c/busses/i2c-sh_mobile.c ret = sh_mobile_i2c_hook_irqs(dev, pd); pd 843 drivers/i2c/busses/i2c-sh_mobile.c pd->dev = &dev->dev; pd 844 drivers/i2c/busses/i2c-sh_mobile.c platform_set_drvdata(dev, pd); pd 848 drivers/i2c/busses/i2c-sh_mobile.c pd->res = res; pd 849 drivers/i2c/busses/i2c-sh_mobile.c pd->reg = devm_ioremap_resource(&dev->dev, res); pd 850 drivers/i2c/busses/i2c-sh_mobile.c if (IS_ERR(pd->reg)) pd 851 drivers/i2c/busses/i2c-sh_mobile.c return PTR_ERR(pd->reg); pd 854 drivers/i2c/busses/i2c-sh_mobile.c pd->bus_speed = (ret || !bus_speed) ? STANDARD_MODE : bus_speed; pd 855 drivers/i2c/busses/i2c-sh_mobile.c pd->clks_per_count = 1; pd 859 drivers/i2c/busses/i2c-sh_mobile.c pd->flags |= IIC_FLAG_HAS_ICIC67; pd 866 drivers/i2c/busses/i2c-sh_mobile.c pd->clks_per_count = config->clks_per_count; pd 867 drivers/i2c/busses/i2c-sh_mobile.c ret = config->setup(pd); pd 869 drivers/i2c/busses/i2c-sh_mobile.c ret = sh_mobile_i2c_init(pd); pd 877 drivers/i2c/busses/i2c-sh_mobile.c sg_init_table(&pd->sg, 1); pd 878 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_direction = DMA_NONE; pd 879 drivers/i2c/busses/i2c-sh_mobile.c pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); pd 882 drivers/i2c/busses/i2c-sh_mobile.c adap = &pd->adap; pd 883 drivers/i2c/busses/i2c-sh_mobile.c i2c_set_adapdata(adap, pd); pd 895 drivers/i2c/busses/i2c-sh_mobile.c spin_lock_init(&pd->lock); pd 896 drivers/i2c/busses/i2c-sh_mobile.c init_waitqueue_head(&pd->wait); pd 900 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_release_dma(pd); pd 904 drivers/i2c/busses/i2c-sh_mobile.c dev_info(&dev->dev, "I2C adapter %d, bus speed %lu Hz\n", adap->nr, pd->bus_speed); pd 911 drivers/i2c/busses/i2c-sh_mobile.c struct sh_mobile_i2c_data *pd = platform_get_drvdata(dev); pd 913 drivers/i2c/busses/i2c-sh_mobile.c i2c_del_adapter(&pd->adap); pd 914 drivers/i2c/busses/i2c-sh_mobile.c sh_mobile_i2c_release_dma(pd); pd 36 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd = pw; pd 37 drivers/i2c/busses/i2c-simtec.c writeb(CMD_SET_SDA | (state ? STATE_SDA : 0), pd->reg); pd 42 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd = pw; pd 43 drivers/i2c/busses/i2c-simtec.c writeb(CMD_SET_SCL | (state ? STATE_SCL : 0), pd->reg); pd 48 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd = pw; pd 49 drivers/i2c/busses/i2c-simtec.c return readb(pd->reg) & STATE_SDA ? 1 : 0; pd 54 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd = pw; pd 55 drivers/i2c/busses/i2c-simtec.c return readb(pd->reg) & STATE_SCL ? 1 : 0; pd 62 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd; pd 67 drivers/i2c/busses/i2c-simtec.c pd = kzalloc(sizeof(struct simtec_i2c_data), GFP_KERNEL); pd 68 drivers/i2c/busses/i2c-simtec.c if (pd == NULL) pd 71 drivers/i2c/busses/i2c-simtec.c platform_set_drvdata(dev, pd); pd 82 drivers/i2c/busses/i2c-simtec.c pd->ioarea = request_mem_region(res->start, size, dev->name); pd 83 drivers/i2c/busses/i2c-simtec.c if (pd->ioarea == NULL) { pd 89 drivers/i2c/busses/i2c-simtec.c pd->reg = ioremap(res->start, size); pd 90 drivers/i2c/busses/i2c-simtec.c if (pd->reg == NULL) { pd 98 drivers/i2c/busses/i2c-simtec.c pd->adap.owner = THIS_MODULE; pd 99 drivers/i2c/busses/i2c-simtec.c pd->adap.algo_data = &pd->bit; pd 100 drivers/i2c/busses/i2c-simtec.c pd->adap.dev.parent = &dev->dev; pd 102 drivers/i2c/busses/i2c-simtec.c strlcpy(pd->adap.name, "Simtec I2C", sizeof(pd->adap.name)); pd 104 drivers/i2c/busses/i2c-simtec.c pd->bit.data = pd; pd 105 drivers/i2c/busses/i2c-simtec.c pd->bit.setsda = simtec_i2c_setsda; pd 106 drivers/i2c/busses/i2c-simtec.c pd->bit.setscl = simtec_i2c_setscl; pd 107 drivers/i2c/busses/i2c-simtec.c pd->bit.getsda = simtec_i2c_getsda; pd 108 drivers/i2c/busses/i2c-simtec.c pd->bit.getscl = simtec_i2c_getscl; pd 109 drivers/i2c/busses/i2c-simtec.c pd->bit.timeout = HZ; pd 110 drivers/i2c/busses/i2c-simtec.c pd->bit.udelay = 20; pd 112 drivers/i2c/busses/i2c-simtec.c ret = i2c_bit_add_bus(&pd->adap); pd 119 drivers/i2c/busses/i2c-simtec.c iounmap(pd->reg); pd 122 drivers/i2c/busses/i2c-simtec.c release_mem_region(pd->ioarea->start, size); pd 125 drivers/i2c/busses/i2c-simtec.c kfree(pd); pd 131 drivers/i2c/busses/i2c-simtec.c struct simtec_i2c_data *pd = platform_get_drvdata(dev); pd 133 drivers/i2c/busses/i2c-simtec.c i2c_del_adapter(&pd->adap); pd 135 drivers/i2c/busses/i2c-simtec.c iounmap(pd->reg); pd 136 drivers/i2c/busses/i2c-simtec.c release_mem_region(pd->ioarea->start, resource_size(pd->ioarea)); pd 137 drivers/i2c/busses/i2c-simtec.c kfree(pd); pd 391 drivers/iio/dac/mcp4725.c u8 pd; pd 470 drivers/iio/dac/mcp4725.c pd = (inbuf[0] >> 1) & 0x3; pd 471 drivers/iio/dac/mcp4725.c data->powerdown = pd > 0; pd 472 drivers/iio/dac/mcp4725.c data->powerdown_mode = pd ? pd - 1 : 2; /* largest resistor to gnd */ pd 102 drivers/infiniband/core/agent.c ah = ib_create_ah_from_wc(agent->qp->pd, wc, grh, port_num); pd 347 drivers/infiniband/core/cm.c ah = rdma_create_ah(mad_agent->qp->pd, &av->ah_attr, 0); pd 392 drivers/infiniband/core/cm.c ah = ib_create_ah_from_wc(port->mad_agent->qp->pd, mad_recv_wc->wc, pd 950 drivers/infiniband/core/cma.c int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd, pd 958 drivers/infiniband/core/cma.c if (id->device != pd->device) pd 962 drivers/infiniband/core/cma.c qp = ib_create_qp(pd, qp_init_attr); pd 321 drivers/infiniband/core/core_priv.h struct ib_pd *pd, pd 333 drivers/infiniband/core/core_priv.h qp = dev->ops.create_qp(pd, attr, udata); pd 338 drivers/infiniband/core/core_priv.h qp->pd = pd; pd 194 drivers/infiniband/core/fmr_pool.c struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd, pd 206 drivers/infiniband/core/fmr_pool.c device = pd->device; pd 284 drivers/infiniband/core/fmr_pool.c fmr->fmr = ib_alloc_fmr(pd, params->access, &fmr_attr); pd 1097 drivers/infiniband/core/mad.c mad_send_wr->sg_list[0].lkey = mad_agent->qp->pd->local_dma_lkey; pd 1106 drivers/infiniband/core/mad.c mad_send_wr->sg_list[1].lkey = mad_agent->qp->pd->local_dma_lkey; pd 2936 drivers/infiniband/core/mad.c sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey; pd 3165 drivers/infiniband/core/mad.c qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr); pd 3227 drivers/infiniband/core/mad.c port_priv->pd = ib_alloc_pd(device, 0); pd 3228 drivers/infiniband/core/mad.c if (IS_ERR(port_priv->pd)) { pd 3230 drivers/infiniband/core/mad.c ret = PTR_ERR(port_priv->pd); pd 3285 drivers/infiniband/core/mad.c ib_dealloc_pd(port_priv->pd); pd 3316 drivers/infiniband/core/mad.c ib_dealloc_pd(port_priv->pd); pd 204 drivers/infiniband/core/mad_priv.h struct ib_pd *pd; pd 163 drivers/infiniband/core/mad_rmpp.c ah = ib_create_ah_from_wc(agent->qp->pd, recv_wc->wc, pd 295 drivers/infiniband/core/mad_rmpp.c rmpp_recv->ah = ib_create_ah_from_wc(agent->agent.qp->pd, pd 45 drivers/infiniband/core/mr_pool.c mr = ib_alloc_mr_integrity(qp->pd, max_num_sg, pd 48 drivers/infiniband/core/mr_pool.c mr = ib_alloc_mr(qp->pd, type, max_num_sg); pd 473 drivers/infiniband/core/nldev.c nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PDN, qp->pd->res.id)) pd 582 drivers/infiniband/core/nldev.c struct ib_device *dev = mr->pd->device; pd 599 drivers/infiniband/core/nldev.c nla_put_u32(msg, RDMA_NLDEV_ATTR_RES_PDN, mr->pd->res.id)) pd 616 drivers/infiniband/core/nldev.c struct ib_pd *pd = container_of(res, struct ib_pd, res); pd 617 drivers/infiniband/core/nldev.c struct ib_device *dev = pd->device; pd 621 drivers/infiniband/core/nldev.c pd->local_dma_lkey)) pd 623 drivers/infiniband/core/nldev.c if ((pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) && pd 625 drivers/infiniband/core/nldev.c pd->unsafe_global_rkey)) pd 629 drivers/infiniband/core/nldev.c atomic_read(&pd->usecnt), RDMA_NLDEV_ATTR_PAD)) pd 637 drivers/infiniband/core/nldev.c pd->uobject->context->res.id)) pd 1417 drivers/infiniband/core/nldev.c RES_GET_FUNCS(pd, RDMA_RESTRACK_PD); pd 131 drivers/infiniband/core/restrack.c struct ib_pd *pd; pd 135 drivers/infiniband/core/restrack.c pd = container_of(res, struct ib_qp, res)->pd; pd 136 drivers/infiniband/core/restrack.c if (!pd) { pd 143 drivers/infiniband/core/restrack.c pd = container_of(res, struct ib_mr, res)->pd; pd 147 drivers/infiniband/core/restrack.c pd = NULL; pd 151 drivers/infiniband/core/restrack.c if (pd) pd 152 drivers/infiniband/core/restrack.c res->kern_name = pd->res.kern_name; pd 89 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pd 123 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pd 224 drivers/infiniband/core/rw.c sge->lkey = qp->pd->local_dma_lkey; pd 253 drivers/infiniband/core/rw.c ctx->single.sge.lkey = qp->pd->local_dma_lkey; pd 307 drivers/infiniband/core/rw.c struct ib_device *dev = qp->pd->device; pd 377 drivers/infiniband/core/rw.c struct ib_device *dev = qp->pd->device; pd 378 drivers/infiniband/core/rw.c u32 pages_per_mr = rdma_rw_fr_page_list_len(qp->pd->device, pd 599 drivers/infiniband/core/rw.c rdma_rw_unmap_sg(qp->pd->device, sg, sg_cnt, dir); pd 626 drivers/infiniband/core/rw.c ib_dma_unmap_sg(qp->pd->device, sg, sg_cnt, dir); pd 628 drivers/infiniband/core/rw.c ib_dma_unmap_sg(qp->pd->device, prot_sg, prot_sg_cnt, dir); pd 691 drivers/infiniband/core/rw.c struct ib_device *dev = qp->pd->device; pd 2271 drivers/infiniband/core/sa_query.c new_ah->ah = rdma_create_ah(port->agent->qp->pd, &ah_attr, pd 547 drivers/infiniband/core/user_mad.c ah = rdma_create_user_ah(agent->qp->pd, &ah_attr, NULL); pd 410 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 422 drivers/infiniband/core/uverbs_cmd.c pd = rdma_zalloc_drv_obj(ib_dev, ib_pd); pd 423 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 428 drivers/infiniband/core/uverbs_cmd.c pd->device = ib_dev; pd 429 drivers/infiniband/core/uverbs_cmd.c pd->uobject = uobj; pd 430 drivers/infiniband/core/uverbs_cmd.c pd->__internal_mr = NULL; pd 431 drivers/infiniband/core/uverbs_cmd.c atomic_set(&pd->usecnt, 0); pd 432 drivers/infiniband/core/uverbs_cmd.c pd->res.type = RDMA_RESTRACK_PD; pd 434 drivers/infiniband/core/uverbs_cmd.c ret = ib_dev->ops.alloc_pd(pd, &attrs->driver_udata); pd 438 drivers/infiniband/core/uverbs_cmd.c uobj->object = pd; pd 441 drivers/infiniband/core/uverbs_cmd.c rdma_restrack_uadd(&pd->res); pd 450 drivers/infiniband/core/uverbs_cmd.c ib_dealloc_pd_user(pd, uverbs_get_cleared_udata(attrs)); pd 451 drivers/infiniband/core/uverbs_cmd.c pd = NULL; pd 453 drivers/infiniband/core/uverbs_cmd.c kfree(pd); pd 709 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 729 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs); pd 730 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 736 drivers/infiniband/core/uverbs_cmd.c if (!(pd->device->attrs.device_cap_flags & pd 744 drivers/infiniband/core/uverbs_cmd.c mr = pd->device->ops.reg_user_mr(pd, cmd.start, cmd.length, cmd.hca_va, pd 752 drivers/infiniband/core/uverbs_cmd.c mr->device = pd->device; pd 753 drivers/infiniband/core/uverbs_cmd.c mr->pd = pd; pd 758 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 773 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 781 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 792 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd = NULL; pd 828 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, pd 830 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 836 drivers/infiniband/core/uverbs_cmd.c old_pd = mr->pd; pd 839 drivers/infiniband/core/uverbs_cmd.c cmd.access_flags, pd, pd 845 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 846 drivers/infiniband/core/uverbs_cmd.c mr->pd = pd; pd 858 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 883 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 896 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs); pd 897 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 907 drivers/infiniband/core/uverbs_cmd.c mw = pd->device->ops.alloc_mw(pd, cmd.mw_type, &attrs->driver_udata); pd 913 drivers/infiniband/core/uverbs_cmd.c mw->device = pd->device; pd 914 drivers/infiniband/core/uverbs_cmd.c mw->pd = pd; pd 916 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 928 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 934 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 1274 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd = NULL; pd 1365 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle, pd 1367 drivers/infiniband/core/uverbs_cmd.c if (!pd || (!scq && has_sq)) { pd 1372 drivers/infiniband/core/uverbs_cmd.c device = pd->device; pd 1419 drivers/infiniband/core/uverbs_cmd.c qp = ib_create_qp(pd, &attr); pd 1421 drivers/infiniband/core/uverbs_cmd.c qp = _ib_create_qp(device, pd, &attr, &attrs->driver_udata, pd 1434 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 1471 drivers/infiniband/core/uverbs_cmd.c if (pd) pd 1472 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 1489 drivers/infiniband/core/uverbs_cmd.c if (pd) pd 1490 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 2396 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 2415 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs); pd 2416 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 2439 drivers/infiniband/core/uverbs_cmd.c ah = rdma_create_user_ah(pd, &attr, &attrs->driver_udata); pd 2455 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 2463 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 2897 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 2914 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd.pd_handle, attrs); pd 2915 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 2936 drivers/infiniband/core/uverbs_cmd.c wq = pd->device->ops.create_wq(pd, &wq_init_attr, &attrs->driver_udata); pd 2946 drivers/infiniband/core/uverbs_cmd.c wq->pd = pd; pd 2947 drivers/infiniband/core/uverbs_cmd.c wq->device = pd->device; pd 2950 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 2965 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 2974 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 3358 drivers/infiniband/core/uverbs_cmd.c struct ib_pd *pd; pd 3400 drivers/infiniband/core/uverbs_cmd.c pd = uobj_get_obj_read(pd, UVERBS_OBJECT_PD, cmd->pd_handle, attrs); pd 3401 drivers/infiniband/core/uverbs_cmd.c if (!pd) { pd 3422 drivers/infiniband/core/uverbs_cmd.c srq->device = pd->device; pd 3423 drivers/infiniband/core/uverbs_cmd.c srq->pd = pd; pd 3429 drivers/infiniband/core/uverbs_cmd.c ret = pd->device->ops.create_srq(srq, &attr, udata); pd 3443 drivers/infiniband/core/uverbs_cmd.c atomic_inc(&pd->usecnt); pd 3466 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 3476 drivers/infiniband/core/uverbs_cmd.c uobj_put_obj_read(pd); pd 107 drivers/infiniband/core/uverbs_main.c struct ib_pd *pd = mw->pd; pd 112 drivers/infiniband/core/uverbs_main.c atomic_dec(&pd->usecnt); pd 194 drivers/infiniband/core/uverbs_std_types.c struct ib_pd *pd = uobject->object; pd 197 drivers/infiniband/core/uverbs_std_types.c ret = ib_destroy_usecnt(&pd->usecnt, why, uobject); pd 201 drivers/infiniband/core/uverbs_std_types.c ib_dealloc_pd_user(pd, &attrs->driver_udata); pd 48 drivers/infiniband/core/uverbs_std_types_mr.c struct ib_pd *pd = pd 51 drivers/infiniband/core/uverbs_std_types_mr.c struct ib_device *ib_dev = pd->device; pd 77 drivers/infiniband/core/uverbs_std_types_mr.c return ib_dev->ops.advise_mr(pd, advice, flags, sg_list, num_sge, pd 89 drivers/infiniband/core/uverbs_std_types_mr.c struct ib_pd *pd = pd 91 drivers/infiniband/core/uverbs_std_types_mr.c struct ib_device *ib_dev = pd->device; pd 125 drivers/infiniband/core/uverbs_std_types_mr.c mr = pd->device->ops.reg_dm_mr(pd, dm, &attr, attrs); pd 129 drivers/infiniband/core/uverbs_std_types_mr.c mr->device = pd->device; pd 130 drivers/infiniband/core/uverbs_std_types_mr.c mr->pd = pd; pd 134 drivers/infiniband/core/uverbs_std_types_mr.c atomic_inc(&pd->usecnt); pd 257 drivers/infiniband/core/verbs.c struct ib_pd *pd; pd 261 drivers/infiniband/core/verbs.c pd = rdma_zalloc_drv_obj(device, ib_pd); pd 262 drivers/infiniband/core/verbs.c if (!pd) pd 265 drivers/infiniband/core/verbs.c pd->device = device; pd 266 drivers/infiniband/core/verbs.c pd->uobject = NULL; pd 267 drivers/infiniband/core/verbs.c pd->__internal_mr = NULL; pd 268 drivers/infiniband/core/verbs.c atomic_set(&pd->usecnt, 0); pd 269 drivers/infiniband/core/verbs.c pd->flags = flags; pd 271 drivers/infiniband/core/verbs.c pd->res.type = RDMA_RESTRACK_PD; pd 272 drivers/infiniband/core/verbs.c rdma_restrack_set_task(&pd->res, caller); pd 274 drivers/infiniband/core/verbs.c ret = device->ops.alloc_pd(pd, NULL); pd 276 drivers/infiniband/core/verbs.c kfree(pd); pd 279 drivers/infiniband/core/verbs.c rdma_restrack_kadd(&pd->res); pd 282 drivers/infiniband/core/verbs.c pd->local_dma_lkey = device->local_dma_lkey; pd 294 drivers/infiniband/core/verbs.c mr = pd->device->ops.get_dma_mr(pd, mr_access_flags); pd 296 drivers/infiniband/core/verbs.c ib_dealloc_pd(pd); pd 300 drivers/infiniband/core/verbs.c mr->device = pd->device; pd 301 drivers/infiniband/core/verbs.c mr->pd = pd; pd 306 drivers/infiniband/core/verbs.c pd->__internal_mr = mr; pd 309 drivers/infiniband/core/verbs.c pd->local_dma_lkey = pd->__internal_mr->lkey; pd 312 drivers/infiniband/core/verbs.c pd->unsafe_global_rkey = pd->__internal_mr->rkey; pd 315 drivers/infiniband/core/verbs.c return pd; pd 328 drivers/infiniband/core/verbs.c void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata) pd 332 drivers/infiniband/core/verbs.c if (pd->__internal_mr) { pd 333 drivers/infiniband/core/verbs.c ret = pd->device->ops.dereg_mr(pd->__internal_mr, NULL); pd 335 drivers/infiniband/core/verbs.c pd->__internal_mr = NULL; pd 340 drivers/infiniband/core/verbs.c WARN_ON(atomic_read(&pd->usecnt)); pd 342 drivers/infiniband/core/verbs.c rdma_restrack_del(&pd->res); pd 343 drivers/infiniband/core/verbs.c pd->device->ops.dealloc_pd(pd, udata); pd 344 drivers/infiniband/core/verbs.c kfree(pd); pd 497 drivers/infiniband/core/verbs.c static struct ib_ah *_rdma_create_ah(struct ib_pd *pd, pd 502 drivers/infiniband/core/verbs.c struct ib_device *device = pd->device; pd 518 drivers/infiniband/core/verbs.c ah->pd = pd; pd 528 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 543 drivers/infiniband/core/verbs.c struct ib_ah *rdma_create_ah(struct ib_pd *pd, struct rdma_ah_attr *ah_attr, pd 550 drivers/infiniband/core/verbs.c ret = rdma_fill_sgid_attr(pd->device, ah_attr, &old_sgid_attr); pd 554 drivers/infiniband/core/verbs.c ah = _rdma_create_ah(pd, ah_attr, flags, NULL); pd 574 drivers/infiniband/core/verbs.c struct ib_ah *rdma_create_user_ah(struct ib_pd *pd, pd 582 drivers/infiniband/core/verbs.c err = rdma_fill_sgid_attr(pd->device, ah_attr, &old_sgid_attr); pd 587 drivers/infiniband/core/verbs.c err = ib_resolve_eth_dmac(pd->device, ah_attr); pd 594 drivers/infiniband/core/verbs.c ah = _rdma_create_ah(pd, ah_attr, RDMA_CREATE_AH_SLEEPABLE, udata); pd 896 drivers/infiniband/core/verbs.c struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc, pd 903 drivers/infiniband/core/verbs.c ret = ib_init_ah_attr_from_wc(pd->device, port_num, wc, grh, &ah_attr); pd 907 drivers/infiniband/core/verbs.c ah = rdma_create_ah(pd, &ah_attr, RDMA_CREATE_AH_SLEEPABLE); pd 949 drivers/infiniband/core/verbs.c struct ib_pd *pd; pd 953 drivers/infiniband/core/verbs.c pd = ah->pd; pd 956 drivers/infiniband/core/verbs.c atomic_dec(&pd->usecnt); pd 967 drivers/infiniband/core/verbs.c struct ib_srq *ib_create_srq(struct ib_pd *pd, pd 973 drivers/infiniband/core/verbs.c if (!pd->device->ops.create_srq) pd 976 drivers/infiniband/core/verbs.c srq = rdma_zalloc_drv_obj(pd->device, ib_srq); pd 980 drivers/infiniband/core/verbs.c srq->device = pd->device; pd 981 drivers/infiniband/core/verbs.c srq->pd = pd; pd 994 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 996 drivers/infiniband/core/verbs.c ret = pd->device->ops.create_srq(srq, srq_init_attr, NULL); pd 998 drivers/infiniband/core/verbs.c atomic_dec(&srq->pd->usecnt); pd 1036 drivers/infiniband/core/verbs.c atomic_dec(&srq->pd->usecnt); pd 1132 drivers/infiniband/core/verbs.c qp->pd = NULL; pd 1148 drivers/infiniband/core/verbs.c struct ib_qp *ib_create_qp_user(struct ib_pd *pd, pd 1152 drivers/infiniband/core/verbs.c struct ib_device *device = pd ? pd->device : qp_init_attr->xrcd->device; pd 1175 drivers/infiniband/core/verbs.c qp = _ib_create_qp(device, pd, qp_init_attr, NULL, NULL); pd 1211 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 1860 drivers/infiniband/core/verbs.c struct ib_pd *pd; pd 1875 drivers/infiniband/core/verbs.c pd = qp->pd; pd 1895 drivers/infiniband/core/verbs.c if (pd) pd 1896 drivers/infiniband/core/verbs.c atomic_dec(&pd->usecnt); pd 1983 drivers/infiniband/core/verbs.c struct ib_pd *pd = mr->pd; pd 1991 drivers/infiniband/core/verbs.c atomic_dec(&pd->usecnt); pd 2014 drivers/infiniband/core/verbs.c struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type, pd 2019 drivers/infiniband/core/verbs.c if (!pd->device->ops.alloc_mr) pd 2025 drivers/infiniband/core/verbs.c mr = pd->device->ops.alloc_mr(pd, mr_type, max_num_sg, udata); pd 2027 drivers/infiniband/core/verbs.c mr->device = pd->device; pd 2028 drivers/infiniband/core/verbs.c mr->pd = pd; pd 2031 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 2055 drivers/infiniband/core/verbs.c struct ib_mr *ib_alloc_mr_integrity(struct ib_pd *pd, pd 2062 drivers/infiniband/core/verbs.c if (!pd->device->ops.alloc_mr_integrity || pd 2063 drivers/infiniband/core/verbs.c !pd->device->ops.map_mr_sg_pi) pd 2073 drivers/infiniband/core/verbs.c mr = pd->device->ops.alloc_mr_integrity(pd, max_num_data_sg, pd 2080 drivers/infiniband/core/verbs.c mr->device = pd->device; pd 2081 drivers/infiniband/core/verbs.c mr->pd = pd; pd 2084 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 2097 drivers/infiniband/core/verbs.c struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd, pd 2103 drivers/infiniband/core/verbs.c if (!pd->device->ops.alloc_fmr) pd 2106 drivers/infiniband/core/verbs.c fmr = pd->device->ops.alloc_fmr(pd, mr_access_flags, fmr_attr); pd 2108 drivers/infiniband/core/verbs.c fmr->device = pd->device; pd 2109 drivers/infiniband/core/verbs.c fmr->pd = pd; pd 2110 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 2131 drivers/infiniband/core/verbs.c struct ib_pd *pd; pd 2134 drivers/infiniband/core/verbs.c pd = fmr->pd; pd 2137 drivers/infiniband/core/verbs.c atomic_dec(&pd->usecnt); pd 2273 drivers/infiniband/core/verbs.c struct ib_wq *ib_create_wq(struct ib_pd *pd, pd 2278 drivers/infiniband/core/verbs.c if (!pd->device->ops.create_wq) pd 2281 drivers/infiniband/core/verbs.c wq = pd->device->ops.create_wq(pd, wq_attr, NULL); pd 2287 drivers/infiniband/core/verbs.c wq->device = pd->device; pd 2288 drivers/infiniband/core/verbs.c wq->pd = pd; pd 2290 drivers/infiniband/core/verbs.c atomic_inc(&pd->usecnt); pd 2306 drivers/infiniband/core/verbs.c struct ib_pd *pd = wq->pd; pd 2312 drivers/infiniband/core/verbs.c atomic_dec(&pd->usecnt); pd 405 drivers/infiniband/hw/bnxt_re/ib_verbs.c static void bnxt_re_create_fence_wqe(struct bnxt_re_pd *pd) pd 407 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_fence_data *fence = &pd->fence; pd 433 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct ib_pd *ib_pd = qp->ib_qp.pd; pd 434 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 435 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_fence_data *fence = &pd->fence; pd 446 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe.bind.r_key, qp->qplib_qp.id, pd); pd 457 drivers/infiniband/hw/bnxt_re/ib_verbs.c static void bnxt_re_destroy_fence_mr(struct bnxt_re_pd *pd) pd 459 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_fence_data *fence = &pd->fence; pd 460 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 484 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_create_fence_mr(struct bnxt_re_pd *pd) pd 487 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_fence_data *fence = &pd->fence; pd 488 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 515 drivers/infiniband/hw/bnxt_re/ib_verbs.c mr->qplib_mr.pd = &pd->qplib_pd; pd 538 drivers/infiniband/hw/bnxt_re/ib_verbs.c mw = bnxt_re_alloc_mw(&pd->ib_pd, IB_MW_TYPE_1, NULL); pd 541 drivers/infiniband/hw/bnxt_re/ib_verbs.c "Failed to create fence-MW for PD: %p\n", pd); pd 547 drivers/infiniband/hw/bnxt_re/ib_verbs.c bnxt_re_create_fence_wqe(pd); pd 551 drivers/infiniband/hw/bnxt_re/ib_verbs.c bnxt_re_destroy_fence_mr(pd); pd 558 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 559 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 561 drivers/infiniband/hw/bnxt_re/ib_verbs.c bnxt_re_destroy_fence_mr(pd); pd 563 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (pd->qplib_pd.id) pd 565 drivers/infiniband/hw/bnxt_re/ib_verbs.c &pd->qplib_pd); pd 574 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ibpd, struct bnxt_re_pd, ib_pd); pd 577 drivers/infiniband/hw/bnxt_re/ib_verbs.c pd->rdev = rdev; pd 578 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (bnxt_qplib_alloc_pd(&rdev->qplib_res.pd_tbl, &pd->qplib_pd)) { pd 599 drivers/infiniband/hw/bnxt_re/ib_verbs.c resp.pdid = pd->qplib_pd.id; pd 613 drivers/infiniband/hw/bnxt_re/ib_verbs.c if (bnxt_re_create_fence_mr(pd)) pd 619 drivers/infiniband/hw/bnxt_re/ib_verbs.c &pd->qplib_pd); pd 655 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct ib_pd *ib_pd = ib_ah->pd; pd 656 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 658 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 670 drivers/infiniband/hw/bnxt_re/ib_verbs.c ah->qplib_ah.pd = &pd->qplib_pd; pd 836 drivers/infiniband/hw/bnxt_re/ib_verbs.c static int bnxt_re_init_user_qp(struct bnxt_re_dev *rdev, struct bnxt_re_pd *pd, pd 892 drivers/infiniband/hw/bnxt_re/ib_verbs.c (struct bnxt_re_pd *pd, pd 896 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 906 drivers/infiniband/hw/bnxt_re/ib_verbs.c ah->qplib_ah.pd = &pd->qplib_pd; pd 939 drivers/infiniband/hw/bnxt_re/ib_verbs.c (struct bnxt_re_pd *pd, pd 943 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 956 drivers/infiniband/hw/bnxt_re/ib_verbs.c qp->qplib_qp.pd = &pd->qplib_pd; pd 1005 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 1006 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 1026 drivers/infiniband/hw/bnxt_re/ib_verbs.c qp->qplib_qp.pd = &pd->qplib_pd; pd 1127 drivers/infiniband/hw/bnxt_re/ib_verbs.c rdev->qp1_sqp = bnxt_re_create_shadow_qp(pd, &rdev->qplib_res, pd 1135 drivers/infiniband/hw/bnxt_re/ib_verbs.c rdev->sqp_ah = bnxt_re_create_shadow_qp_ah(pd, &rdev->qplib_res, pd 1166 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_init_user_qp(rdev, pd, qp, udata); pd 1309 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd, pd 1343 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct ib_pd *ib_pd = ib_srq->pd; pd 1344 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 1345 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 1364 drivers/infiniband/hw/bnxt_re/ib_verbs.c srq->qplib_srq.pd = &pd->qplib_pd; pd 1381 drivers/infiniband/hw/bnxt_re/ib_verbs.c rc = bnxt_re_init_user_srq(rdev, pd, srq, udata); pd 3277 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 3278 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 3288 drivers/infiniband/hw/bnxt_re/ib_verbs.c mr->qplib_mr.pd = &pd->qplib_pd; pd 3368 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 3369 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 3385 drivers/infiniband/hw/bnxt_re/ib_verbs.c mr->qplib_mr.pd = &pd->qplib_pd; pd 3424 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 3425 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 3433 drivers/infiniband/hw/bnxt_re/ib_verbs.c mw->qplib_mw.pd = &pd->qplib_pd; pd 3505 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_pd *pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd); pd 3506 drivers/infiniband/hw/bnxt_re/ib_verbs.c struct bnxt_re_dev *rdev = pd->rdev; pd 3523 drivers/infiniband/hw/bnxt_re/ib_verbs.c mr->qplib_mr.pd = &pd->qplib_pd; pd 164 drivers/infiniband/hw/bnxt_re/ib_verbs.h int bnxt_re_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 165 drivers/infiniband/hw/bnxt_re/ib_verbs.h void bnxt_re_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 181 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_qp *bnxt_re_create_qp(struct ib_pd *pd, pd 198 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_mr *bnxt_re_get_dma_mr(struct ib_pd *pd, int mr_access_flags); pd 208 drivers/infiniband/hw/bnxt_re/ib_verbs.h struct ib_mr *bnxt_re_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 582 drivers/infiniband/hw/bnxt_re/qplib_fp.c req.pd_id = cpu_to_le32(srq->pd->id); pd 834 drivers/infiniband/hw/bnxt_re/qplib_fp.c req.pd_id = cpu_to_le32(qp->pd->id); pd 1058 drivers/infiniband/hw/bnxt_re/qplib_fp.c req.pd_id = cpu_to_le32(qp->pd->id); pd 43 drivers/infiniband/hw/bnxt_re/qplib_fp.h struct bnxt_qplib_pd *pd; pd 254 drivers/infiniband/hw/bnxt_re/qplib_fp.h struct bnxt_qplib_pd *pd; pd 575 drivers/infiniband/hw/bnxt_re/qplib_res.c int bnxt_qplib_alloc_pd(struct bnxt_qplib_pd_tbl *pdt, struct bnxt_qplib_pd *pd) pd 585 drivers/infiniband/hw/bnxt_re/qplib_res.c pd->id = bit_num; pd 591 drivers/infiniband/hw/bnxt_re/qplib_res.c struct bnxt_qplib_pd *pd) pd 593 drivers/infiniband/hw/bnxt_re/qplib_res.c if (test_and_set_bit(pd->id, pdt->tbl)) { pd 595 drivers/infiniband/hw/bnxt_re/qplib_res.c pd->id); pd 598 drivers/infiniband/hw/bnxt_re/qplib_res.c pd->id = 0; pd 245 drivers/infiniband/hw/bnxt_re/qplib_res.h struct bnxt_qplib_pd *pd); pd 248 drivers/infiniband/hw/bnxt_re/qplib_res.h struct bnxt_qplib_pd *pd); pd 521 drivers/infiniband/hw/bnxt_re/qplib_sp.c req.pd_id = cpu_to_le32(ah->pd->id); pd 604 drivers/infiniband/hw/bnxt_re/qplib_sp.c req.pd_id = cpu_to_le32(mrw->pd->id); pd 94 drivers/infiniband/hw/bnxt_re/qplib_sp.h struct bnxt_qplib_pd *pd; pd 109 drivers/infiniband/hw/bnxt_re/qplib_sp.h struct bnxt_qplib_pd *pd; pd 293 drivers/infiniband/hw/cxgb3/iwch_provider.c static void iwch_deallocate_pd(struct ib_pd *pd, struct ib_udata *udata) pd 298 drivers/infiniband/hw/cxgb3/iwch_provider.c php = to_iwch_pd(pd); pd 300 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ibpd %p pdid 0x%x\n", __func__, pd, php->pdid); pd 304 drivers/infiniband/hw/cxgb3/iwch_provider.c static int iwch_allocate_pd(struct ib_pd *pd, struct ib_udata *udata) pd 306 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_pd *php = to_iwch_pd(pd); pd 307 drivers/infiniband/hw/cxgb3/iwch_provider.c struct ib_device *ibdev = pd->device; pd 355 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_get_dma_mr(struct ib_pd *pd, int acc) pd 359 drivers/infiniband/hw/cxgb3/iwch_provider.c struct iwch_pd *php = to_iwch_pd(pd); pd 365 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ib_pd %p\n", __func__, pd); pd 433 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 444 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ib_pd %p\n", __func__, pd); pd 446 drivers/infiniband/hw/cxgb3/iwch_provider.c php = to_iwch_pd(pd); pd 531 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mw *iwch_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 544 drivers/infiniband/hw/cxgb3/iwch_provider.c php = to_iwch_pd(pd); pd 585 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_mr *iwch_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 599 drivers/infiniband/hw/cxgb3/iwch_provider.c php = to_iwch_pd(pd); pd 695 drivers/infiniband/hw/cxgb3/iwch_provider.c static struct ib_qp *iwch_create_qp(struct ib_pd *pd, pd 708 drivers/infiniband/hw/cxgb3/iwch_provider.c pr_debug("%s ib_pd %p\n", __func__, pd); pd 711 drivers/infiniband/hw/cxgb3/iwch_provider.c php = to_iwch_pd(pd); pd 769 drivers/infiniband/hw/cxgb3/iwch_provider.c qhp->attr.pd = php->pdid; pd 150 drivers/infiniband/hw/cxgb3/iwch_provider.h u32 pd; /* IN */ pd 824 drivers/infiniband/hw/cxgb3/iwch_qp.c init_attr.pdid = qhp->attr.pd; pd 476 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pd.total, dev->rdev.stats.pd.cur, pd 477 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pd.max, dev->rdev.stats.pd.fail); pd 524 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pd.max = 0; pd 525 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pd.fail = 0; pd 842 drivers/infiniband/hw/cxgb4/device.c rdev->stats.pd.total = T4_MAX_NUM_PD; pd 130 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct c4iw_stat pd; pd 470 drivers/infiniband/hw/cxgb4/iw_cxgb4.h u32 pd; pd 982 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 988 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 990 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start, pd 993 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc); pd 1006 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct ib_qp *c4iw_create_qp(struct ib_pd *pd, pd 447 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_get_dma_mr(struct ib_pd *pd, int acc) pd 455 drivers/infiniband/hw/cxgb4/mem.c pr_debug("ib_pd %p\n", pd); pd 456 drivers/infiniband/hw/cxgb4/mem.c php = to_c4iw_pd(pd); pd 508 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 519 drivers/infiniband/hw/cxgb4/mem.c pr_debug("ib_pd %p\n", pd); pd 527 drivers/infiniband/hw/cxgb4/mem.c php = to_c4iw_pd(pd); pd 615 drivers/infiniband/hw/cxgb4/mem.c struct ib_mw *c4iw_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 628 drivers/infiniband/hw/cxgb4/mem.c php = to_c4iw_pd(pd); pd 693 drivers/infiniband/hw/cxgb4/mem.c struct ib_mr *c4iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 704 drivers/infiniband/hw/cxgb4/mem.c php = to_c4iw_pd(pd); pd 193 drivers/infiniband/hw/cxgb4/provider.c static void c4iw_deallocate_pd(struct ib_pd *pd, struct ib_udata *udata) pd 198 drivers/infiniband/hw/cxgb4/provider.c php = to_c4iw_pd(pd); pd 200 drivers/infiniband/hw/cxgb4/provider.c pr_debug("ibpd %p pdid 0x%x\n", pd, php->pdid); pd 203 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.cur--; pd 207 drivers/infiniband/hw/cxgb4/provider.c static int c4iw_allocate_pd(struct ib_pd *pd, struct ib_udata *udata) pd 209 drivers/infiniband/hw/cxgb4/provider.c struct c4iw_pd *php = to_c4iw_pd(pd); pd 210 drivers/infiniband/hw/cxgb4/provider.c struct ib_device *ibdev = pd->device; pd 231 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.cur++; pd 232 drivers/infiniband/hw/cxgb4/provider.c if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max) pd 233 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.max = rhp->rdev.stats.pd.cur; pd 1799 drivers/infiniband/hw/cxgb4/qp.c wqe->u.init.pdid = cpu_to_be32(qhp->attr.pd); pd 2111 drivers/infiniband/hw/cxgb4/qp.c struct ib_qp *c4iw_create_qp(struct ib_pd *pd, struct ib_qp_init_attr *attrs, pd 2127 drivers/infiniband/hw/cxgb4/qp.c pr_debug("ib_pd %p\n", pd); pd 2132 drivers/infiniband/hw/cxgb4/qp.c php = to_c4iw_pd(pd); pd 2196 drivers/infiniband/hw/cxgb4/qp.c qhp->attr.pd = php->pdid; pd 2672 drivers/infiniband/hw/cxgb4/qp.c struct ib_pd *pd = ib_srq->pd; pd 2683 drivers/infiniband/hw/cxgb4/qp.c pr_debug("%s ib_pd %p\n", __func__, pd); pd 2685 drivers/infiniband/hw/cxgb4/qp.c php = to_c4iw_pd(pd); pd 105 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 287 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 310 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 329 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 701 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 711 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u16 pd; pd 26 drivers/infiniband/hw/efa/efa_com_cmd.c create_qp_cmd.pd = params->pd; pd 228 drivers/infiniband/hw/efa/efa_com_cmd.c mr_cmd.pd = params->pd; pd 308 drivers/infiniband/hw/efa/efa_com_cmd.c ah_cmd.pd = params->pdn; pd 337 drivers/infiniband/hw/efa/efa_com_cmd.c ah_cmd.pd = params->pdn; pd 347 drivers/infiniband/hw/efa/efa_com_cmd.c ah_cmd.ah, ah_cmd.pd, err); pd 648 drivers/infiniband/hw/efa/efa_com_cmd.c result->pdn = resp.pd; pd 662 drivers/infiniband/hw/efa/efa_com_cmd.c cmd.pd = params->pdn; pd 672 drivers/infiniband/hw/efa/efa_com_cmd.c cmd.pd, err); pd 27 drivers/infiniband/hw/efa/efa_com_cmd.h u16 pd; pd 177 drivers/infiniband/hw/efa/efa_com_cmd.h u16 pd; pd 433 drivers/infiniband/hw/efa/efa_verbs.c struct efa_pd *pd = to_epd(ibpd); pd 448 drivers/infiniband/hw/efa/efa_verbs.c pd->pdn = result.pdn; pd 461 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Allocated pd[%d]\n", pd->pdn); pd 475 drivers/infiniband/hw/efa/efa_verbs.c struct efa_pd *pd = to_epd(ibpd); pd 477 drivers/infiniband/hw/efa/efa_verbs.c ibdev_dbg(&dev->ibdev, "Dealloc pd[%d]\n", pd->pdn); pd 478 drivers/infiniband/hw/efa/efa_verbs.c efa_pd_dealloc(dev, pd->pdn); pd 490 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibqp->pd->device); pd 691 drivers/infiniband/hw/efa/efa_verbs.c create_qp_params.pd = to_epd(ibpd)->pdn; pd 1434 drivers/infiniband/hw/efa/efa_verbs.c params.pd = to_epd(ibpd)->pdn; pd 1675 drivers/infiniband/hw/efa/efa_verbs.c .pdn = to_epd(ah->ibah.pd)->pdn, pd 1709 drivers/infiniband/hw/efa/efa_verbs.c params.pdn = to_epd(ibah->pd)->pdn; pd 1741 drivers/infiniband/hw/efa/efa_verbs.c struct efa_dev *dev = to_edev(ibah->pd->device); pd 308 drivers/infiniband/hw/hfi1/mad.c ah = rdma_create_ah(qp0->ibqp.pd, &attr, 0); pd 1782 drivers/infiniband/hw/hfi1/mad.c u8 *pd = data; pd 1789 drivers/infiniband/hw/hfi1/mad.c if ((pd[i] & 0x1f) == 0xf) pd 1790 drivers/infiniband/hw/hfi1/mad.c pd[i] = ILLEGAL_VL; pd 1792 drivers/infiniband/hw/hfi1/mad.c if ((pd[i] & 0x1f) == ILLEGAL_VL) pd 1793 drivers/infiniband/hw/hfi1/mad.c pd[i] = 0xf; pd 390 drivers/infiniband/hw/hns/hns_roce_device.h u32 pd; /* PD num of MR */ pd 1185 drivers/infiniband/hw/hns/hns_roce_device.h int hns_roce_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 1186 drivers/infiniband/hw/hns/hns_roce_device.h void hns_roce_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 1188 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_get_dma_mr(struct ib_pd *pd, int acc); pd 1189 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 1193 drivers/infiniband/hw/hns/hns_roce_device.h u64 virt_addr, int mr_access_flags, struct ib_pd *pd, pd 1195 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mr *hns_roce_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1205 drivers/infiniband/hw/hns/hns_roce_device.h struct ib_mw *hns_roce_alloc_mw(struct ib_pd *pd, enum ib_mw_type, pd 677 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct ib_pd *pd) pd 689 drivers/infiniband/hw/hns/hns_roce_hw_v1.c qp = hns_roce_create_qp(pd, &init_attr, NULL); pd 709 drivers/infiniband/hw/hns/hns_roce_hw_v1.c struct ib_pd *pd; pd 745 drivers/infiniband/hw/hns/hns_roce_hw_v1.c pd = rdma_zalloc_drv_obj(ibdev, ib_pd); pd 746 drivers/infiniband/hw/hns/hns_roce_hw_v1.c if (!pd) { pd 751 drivers/infiniband/hw/hns/hns_roce_hw_v1.c pd->device = ibdev; pd 752 drivers/infiniband/hw/hns/hns_roce_hw_v1.c ret = hns_roce_alloc_pd(pd, NULL); pd 756 drivers/infiniband/hw/hns/hns_roce_hw_v1.c free_mr->mr_free_pd = to_hr_pd(pd); pd 796 drivers/infiniband/hw/hns/hns_roce_hw_v1.c free_mr->mr_free_qp[i] = hns_roce_v1_create_lp_qp(hr_dev, pd); pd 810 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hr_qp->ibqp.pd = pd; pd 862 drivers/infiniband/hw/hns/hns_roce_hw_v1.c hns_roce_dealloc_pd(pd, NULL); pd 865 drivers/infiniband/hw/hns/hns_roce_hw_v1.c kfree(pd); pd 1875 drivers/infiniband/hw/hns/hns_roce_hw_v1.c MPT_BYTE_28_PD_S, mr->pd); pd 2605 drivers/infiniband/hw/hns/hns_roce_hw_v1.c QP1C_BYTES_4_PD_S, to_hr_pd(ibqp->pd)->pdn); pd 2791 drivers/infiniband/hw/hns/hns_roce_hw_v1.c to_hr_pd(ibqp->pd)->pdn); pd 2857 drivers/infiniband/hw/hns/hns_roce_hw_v1.c to_hr_pd(ibqp->pd)->pdn); pd 375 drivers/infiniband/hw/hns/hns_roce_hw_v2.c to_hr_pd(ibqp->pd)->pdn); pd 2287 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_MPT_BYTE_4_PD_S, mr->pd); pd 2336 drivers/infiniband/hw/hns/hns_roce_hw_v2.c mr->pd = pdn; pd 2385 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_MPT_BYTE_4_PD_S, mr->pd); pd 3282 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_QPC_BYTE_16_PD_S, to_hr_pd(ibqp->pd)->pdn); pd 3602 drivers/infiniband/hw/hns/hns_roce_hw_v2.c V2_QPC_BYTE_16_PD_S, to_hr_pd(ibqp->pd)->pdn); pd 571 drivers/infiniband/hw/hns/hns_roce_mr.c static int hns_roce_mr_alloc(struct hns_roce_dev *hr_dev, u32 pd, u64 iova, pd 586 drivers/infiniband/hw/hns/hns_roce_mr.c mr->pd = pd; /* MR num */ pd 985 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_get_dma_mr(struct ib_pd *pd, int acc) pd 997 drivers/infiniband/hw/hns/hns_roce_mr.c ret = hns_roce_mr_alloc(to_hr_dev(pd->device), to_hr_pd(pd)->pdn, 0, pd 1002 drivers/infiniband/hw/hns/hns_roce_mr.c ret = hns_roce_mr_enable(to_hr_dev(pd->device), mr); pd 1012 drivers/infiniband/hw/hns/hns_roce_mr.c hns_roce_mr_free(to_hr_dev(pd->device), mr); pd 1130 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 1134 drivers/infiniband/hw/hns/hns_roce_mr.c struct hns_roce_dev *hr_dev = to_hr_dev(pd->device); pd 1180 drivers/infiniband/hw/hns/hns_roce_mr.c ret = hns_roce_mr_alloc(hr_dev, to_hr_pd(pd)->pdn, virt_addr, length, pd 1286 drivers/infiniband/hw/hns/hns_roce_mr.c u64 virt_addr, int mr_access_flags, struct ib_pd *pd, pd 1318 drivers/infiniband/hw/hns/hns_roce_mr.c pdn = to_hr_pd(pd)->pdn; pd 1374 drivers/infiniband/hw/hns/hns_roce_mr.c struct ib_mr *hns_roce_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1377 drivers/infiniband/hw/hns/hns_roce_mr.c struct hns_roce_dev *hr_dev = to_hr_dev(pd->device); pd 1403 drivers/infiniband/hw/hns/hns_roce_mr.c ret = hns_roce_mr_alloc(hr_dev, to_hr_pd(pd)->pdn, 0, length, pd 1418 drivers/infiniband/hw/hns/hns_roce_mr.c hns_roce_mr_free(to_hr_dev(pd->device), mr); pd 65 drivers/infiniband/hw/hns/hns_roce_pd.c struct hns_roce_pd *pd = to_hr_pd(ibpd); pd 68 drivers/infiniband/hw/hns/hns_roce_pd.c ret = hns_roce_pd_alloc(to_hr_dev(ib_dev), &pd->pdn); pd 75 drivers/infiniband/hw/hns/hns_roce_pd.c struct hns_roce_ib_alloc_pd_resp uresp = {.pdn = pd->pdn}; pd 78 drivers/infiniband/hw/hns/hns_roce_pd.c hns_roce_pd_free(to_hr_dev(ib_dev), pd->pdn); pd 87 drivers/infiniband/hw/hns/hns_roce_pd.c void hns_roce_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 89 drivers/infiniband/hw/hns/hns_roce_pd.c hns_roce_pd_free(to_hr_dev(pd->device), to_hr_pd(pd)->pdn); pd 1016 drivers/infiniband/hw/hns/hns_roce_qp.c struct ib_qp *hns_roce_create_qp(struct ib_pd *pd, pd 1020 drivers/infiniband/hw/hns/hns_roce_qp.c struct hns_roce_dev *hr_dev = to_hr_dev(pd->device); pd 1032 drivers/infiniband/hw/hns/hns_roce_qp.c ret = hns_roce_create_qp_common(hr_dev, pd, init_attr, udata, 0, pd 1067 drivers/infiniband/hw/hns/hns_roce_qp.c ret = hns_roce_create_qp_common(hr_dev, pd, init_attr, udata, pd 252 drivers/infiniband/hw/hns/hns_roce_srq.c static int hns_roce_create_idx_que(struct ib_pd *pd, struct hns_roce_srq *srq, pd 255 drivers/infiniband/hw/hns/hns_roce_srq.c struct hns_roce_dev *hr_dev = to_hr_dev(pd->device); pd 296 drivers/infiniband/hw/hns/hns_roce_srq.c ret = hns_roce_create_idx_que(srq->ibsrq.pd, srq, page_shift); pd 413 drivers/infiniband/hw/hns/hns_roce_srq.c ret = hns_roce_srq_alloc(hr_dev, to_hr_pd(ib_srq->pd)->pdn, cqn, 0, pd 3720 drivers/infiniband/hw/i40iw/i40iw_cm.c ibmr->pd = &iwpd->ibpd; pd 418 drivers/infiniband/hw/i40iw/i40iw_ctrl.c struct i40iw_sc_pd *pd, pd 422 drivers/infiniband/hw/i40iw/i40iw_ctrl.c pd->size = sizeof(*pd); pd 423 drivers/infiniband/hw/i40iw/i40iw_ctrl.c pd->pd_id = pd_id; pd 424 drivers/infiniband/hw/i40iw/i40iw_ctrl.c pd->abi_ver = abi_ver; pd 425 drivers/infiniband/hw/i40iw/i40iw_ctrl.c pd->dev = dev; pd 2364 drivers/infiniband/hw/i40iw/i40iw_ctrl.c qp->dev = info->pd->dev; pd 2373 drivers/infiniband/hw/i40iw/i40iw_ctrl.c qp->pd = info->pd; pd 2375 drivers/infiniband/hw/i40iw/i40iw_ctrl.c offset = (qp->pd->dev->is_pf) ? I40E_PFPE_WQEALLOC : I40E_VFPE_WQEALLOC1; pd 2376 drivers/infiniband/hw/i40iw/i40iw_ctrl.c if (i40iw_get_hw_addr(qp->pd->dev)) pd 2377 drivers/infiniband/hw/i40iw/i40iw_ctrl.c wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) + pd 2381 drivers/infiniband/hw/i40iw/i40iw_ctrl.c info->qp_uk_init_info.abi_ver = qp->pd->abi_ver; pd 2387 drivers/infiniband/hw/i40iw/i40iw_ctrl.c pble_obj_cnt = info->pd->dev->hmc_info->hmc_obj[I40IW_HMC_IW_PBLE].cnt; pd 2401 drivers/infiniband/hw/i40iw/i40iw_ctrl.c switch (qp->pd->abi_ver) { pd 2452 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cqp = qp->pd->dev->cqp; pd 2500 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cqp = qp->pd->dev->cqp; pd 2567 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cqp = qp->pd->dev->cqp; pd 2620 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cqp = qp->pd->dev->cqp; pd 2677 drivers/infiniband/hw/i40iw/i40iw_ctrl.c cqp = qp->pd->dev->cqp; pd 577 drivers/infiniband/hw/i40iw/i40iw_puda.c qp->pd = &rsrc->sc_pd; pd 603 drivers/infiniband/hw/i40iw/i40iw_puda.c if (qp->pd->dev->is_pf) pd 604 drivers/infiniband/hw/i40iw/i40iw_puda.c ukqp->wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) + pd 607 drivers/infiniband/hw/i40iw/i40iw_puda.c ukqp->wqe_alloc_reg = (u32 __iomem *)(i40iw_get_hw_addr(qp->pd->dev) + pd 376 drivers/infiniband/hw/i40iw/i40iw_type.h struct i40iw_sc_pd *pd; pd 844 drivers/infiniband/hw/i40iw/i40iw_type.h struct i40iw_sc_pd *pd; pd 285 drivers/infiniband/hw/i40iw/i40iw_verbs.c static int i40iw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 287 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(pd); pd 288 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(pd->device); pd 617 drivers/infiniband/hw/i40iw/i40iw_verbs.c init_info.pd = &iwpd->sc_pd; pd 1512 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(iwmr->ibmr.pd); pd 1550 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1553 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(pd); pd 1554 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(pd->device); pd 1575 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.pd = pd; pd 1576 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.device = pd->device; pd 1678 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(iwmr->ibmr.pd); pd 1735 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_reg_user_mr(struct ib_pd *pd, pd 1742 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(pd); pd 1743 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(pd->device); pd 1784 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.pd = pd; pd 1785 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.device = pd->device; pd 1885 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_mr *i40iw_reg_phys_mr(struct ib_pd *pd, pd 1891 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_pd *iwpd = to_iwpd(pd); pd 1892 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct i40iw_device *iwdev = to_iwdev(pd->device); pd 1903 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.pd = pd; pd 1904 drivers/infiniband/hw/i40iw/i40iw_verbs.c iwmr->ibmr.device = pd->device; pd 1940 drivers/infiniband/hw/i40iw/i40iw_verbs.c static struct ib_mr *i40iw_get_dma_mr(struct ib_pd *pd, int acc) pd 1944 drivers/infiniband/hw/i40iw/i40iw_verbs.c return i40iw_reg_phys_mr(pd, 0, 0, acc, &kva); pd 1986 drivers/infiniband/hw/i40iw/i40iw_verbs.c struct ib_pd *ibpd = ib_mr->pd; pd 48 drivers/infiniband/hw/mlx4/ah.c ah->av.ib.port_pd = cpu_to_be32(to_mpd(ib_ah->pd)->pdn | pd 118 drivers/infiniband/hw/mlx4/ah.c ah->av.eth.port_pd = cpu_to_be32(to_mpd(ib_ah->pd)->pdn | pd 204 drivers/infiniband/hw/mlx4/mad.c new_ah = rdma_create_ah(dev->send_agent[port_num - 1][0]->qp->pd, pd 570 drivers/infiniband/hw/mlx4/mad.c ah = rdma_create_ah(tun_ctx->pd, &attr, 0); pd 638 drivers/infiniband/hw/mlx4/mad.c list.lkey = tun_ctx->pd->local_dma_lkey; pd 1324 drivers/infiniband/hw/mlx4/mad.c sg_list.lkey = ctx->pd->local_dma_lkey; pd 1396 drivers/infiniband/hw/mlx4/mad.c ah = rdma_zalloc_drv_obj(sqp_ctx->pd->device, ib_ah); pd 1400 drivers/infiniband/hw/mlx4/mad.c ah->device = sqp_ctx->pd->device; pd 1401 drivers/infiniband/hw/mlx4/mad.c ah->pd = sqp_ctx->pd; pd 1437 drivers/infiniband/hw/mlx4/mad.c list.lkey = sqp_ctx->pd->local_dma_lkey; pd 1834 drivers/infiniband/hw/mlx4/mad.c tun_qp->qp = ib_create_qp(ctx->pd, &qp_init_attr.init_attr); pd 2020 drivers/infiniband/hw/mlx4/mad.c ctx->pd = ib_alloc_pd(ctx->ib_dev, 0); pd 2021 drivers/infiniband/hw/mlx4/mad.c if (IS_ERR(ctx->pd)) { pd 2022 drivers/infiniband/hw/mlx4/mad.c ret = PTR_ERR(ctx->pd); pd 2070 drivers/infiniband/hw/mlx4/mad.c ib_dealloc_pd(ctx->pd); pd 2071 drivers/infiniband/hw/mlx4/mad.c ctx->pd = NULL; pd 2105 drivers/infiniband/hw/mlx4/mad.c ib_dealloc_pd(ctx->pd); pd 2106 drivers/infiniband/hw/mlx4/mad.c ctx->pd = NULL; pd 2237 drivers/infiniband/hw/mlx4/mad.c ib_dealloc_pd(sqp_ctx->pd); pd 2238 drivers/infiniband/hw/mlx4/mad.c sqp_ctx->pd = NULL; pd 1196 drivers/infiniband/hw/mlx4/main.c struct mlx4_ib_pd *pd = to_mpd(ibpd); pd 1200 drivers/infiniband/hw/mlx4/main.c err = mlx4_pd_alloc(to_mdev(ibdev)->dev, &pd->pdn); pd 1204 drivers/infiniband/hw/mlx4/main.c if (udata && ib_copy_to_udata(udata, &pd->pdn, sizeof(__u32))) { pd 1205 drivers/infiniband/hw/mlx4/main.c mlx4_pd_free(to_mdev(ibdev)->dev, pd->pdn); pd 1211 drivers/infiniband/hw/mlx4/main.c static void mlx4_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 1213 drivers/infiniband/hw/mlx4/main.c mlx4_pd_free(to_mdev(pd->device)->dev, to_mpd(pd)->pdn); pd 1234 drivers/infiniband/hw/mlx4/main.c xrcd->pd = ib_alloc_pd(ibdev, 0); pd 1235 drivers/infiniband/hw/mlx4/main.c if (IS_ERR(xrcd->pd)) { pd 1236 drivers/infiniband/hw/mlx4/main.c err = PTR_ERR(xrcd->pd); pd 1250 drivers/infiniband/hw/mlx4/main.c ib_dealloc_pd(xrcd->pd); pd 1261 drivers/infiniband/hw/mlx4/main.c ib_dealloc_pd(to_mxrcd(xrcd)->pd); pd 100 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_pd *pd; pd 459 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_pd *pd; pd 730 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc); pd 733 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 737 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 740 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 771 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd, pd 797 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int mr_access_flags, pd 896 drivers/infiniband/hw/mlx4/mlx4_ib.h int mr_access_flags, struct ib_pd *pd, pd 906 drivers/infiniband/hw/mlx4/mlx4_ib.h struct ib_wq *mlx4_ib_create_wq(struct ib_pd *pd, pd 58 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_get_dma_mr(struct ib_pd *pd, int acc) pd 67 drivers/infiniband/hw/mlx4/mr.c err = mlx4_mr_alloc(to_mdev(pd->device)->dev, to_mpd(pd)->pdn, 0, pd 72 drivers/infiniband/hw/mlx4/mr.c err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); pd 82 drivers/infiniband/hw/mlx4/mr.c (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); pd 404 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 408 drivers/infiniband/hw/mlx4/mr.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 427 drivers/infiniband/hw/mlx4/mr.c err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, virt_addr, length, pd 448 drivers/infiniband/hw/mlx4/mr.c (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); pd 461 drivers/infiniband/hw/mlx4/mr.c int mr_access_flags, struct ib_pd *pd, pd 481 drivers/infiniband/hw/mlx4/mr.c to_mpd(pd)->pdn); pd 614 drivers/infiniband/hw/mlx4/mr.c struct ib_mw *mlx4_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 617 drivers/infiniband/hw/mlx4/mr.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 625 drivers/infiniband/hw/mlx4/mr.c err = mlx4_mw_alloc(dev->dev, to_mpd(pd)->pdn, pd 657 drivers/infiniband/hw/mlx4/mr.c struct ib_mr *mlx4_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 660 drivers/infiniband/hw/mlx4/mr.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 672 drivers/infiniband/hw/mlx4/mr.c err = mlx4_mr_alloc(dev->dev, to_mpd(pd)->pdn, 0, 0, 0, pd 677 drivers/infiniband/hw/mlx4/mr.c err = mlx4_alloc_priv_pages(pd->device, mr, max_num_sg); pd 692 drivers/infiniband/hw/mlx4/mr.c mr->ibmr.device = pd->device; pd 701 drivers/infiniband/hw/mlx4/mr.c struct ib_fmr *mlx4_ib_fmr_alloc(struct ib_pd *pd, int acc, pd 704 drivers/infiniband/hw/mlx4/mr.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 712 drivers/infiniband/hw/mlx4/mr.c err = mlx4_fmr_alloc(dev->dev, to_mpd(pd)->pdn, convert_access(acc), pd 718 drivers/infiniband/hw/mlx4/mr.c err = mlx4_fmr_enable(to_mdev(pd->device)->dev, &fmr->mfmr); pd 727 drivers/infiniband/hw/mlx4/mr.c (void) mlx4_mr_free(to_mdev(pd->device)->dev, &fmr->mfmr.mr); pd 699 drivers/infiniband/hw/mlx4/qp.c static struct ib_qp *_mlx4_ib_create_qp_rss(struct ib_pd *pd, pd 764 drivers/infiniband/hw/mlx4/qp.c err = create_qp_rss(to_mdev(pd->device), init_attr, &ucmd, qp); pd 858 drivers/infiniband/hw/mlx4/qp.c static int create_rq(struct ib_pd *pd, struct ib_qp_init_attr *init_attr, pd 861 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 990 drivers/infiniband/hw/mlx4/qp.c static int create_qp_common(struct ib_pd *pd, struct ib_qp_init_attr *init_attr, pd 994 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_dev *dev = to_mdev(pd->device); pd 1197 drivers/infiniband/hw/mlx4/qp.c if (alloc_proxy_bufs(pd->device, qp)) { pd 1273 drivers/infiniband/hw/mlx4/qp.c free_proxy_bufs(pd->device, qp); pd 1360 drivers/infiniband/hw/mlx4/qp.c return to_mpd(to_mxrcd(qp->ibqp.xrcd)->pd); pd 1362 drivers/infiniband/hw/mlx4/qp.c return to_mpd(qp->ibqp.pd); pd 1532 drivers/infiniband/hw/mlx4/qp.c static struct ib_qp *_mlx4_ib_create_qp(struct ib_pd *pd, pd 1542 drivers/infiniband/hw/mlx4/qp.c return _mlx4_ib_create_qp_rss(pd, init_attr, udata); pd 1578 drivers/infiniband/hw/mlx4/qp.c pd = to_mxrcd(init_attr->xrcd)->pd; pd 1583 drivers/infiniband/hw/mlx4/qp.c if (!(to_mdev(pd->device)->dev->caps.flags & MLX4_DEV_CAP_FLAG_XRC)) pd 1598 drivers/infiniband/hw/mlx4/qp.c err = create_qp_common(pd, init_attr, udata, 0, &qp); pd 1618 drivers/infiniband/hw/mlx4/qp.c int res = mlx4_qp_reserve_range(to_mdev(pd->device)->dev, pd 1625 drivers/infiniband/hw/mlx4/qp.c sqpn = get_sqp_num(to_mdev(pd->device), init_attr); pd 1628 drivers/infiniband/hw/mlx4/qp.c err = create_qp_common(pd, init_attr, udata, sqpn, &qp); pd 1645 drivers/infiniband/hw/mlx4/qp.c struct ib_qp *mlx4_ib_create_qp(struct ib_pd *pd, pd 1648 drivers/infiniband/hw/mlx4/qp.c struct ib_device *device = pd ? pd->device : init_attr->xrcd->device; pd 1652 drivers/infiniband/hw/mlx4/qp.c ibqp = _mlx4_ib_create_qp(pd, init_attr, udata); pd 1663 drivers/infiniband/hw/mlx4/qp.c sqp->roce_v2_gsi = ib_create_qp(pd, init_attr); pd 2143 drivers/infiniband/hw/mlx4/qp.c struct mlx4_ib_pd *pd; pd 2163 drivers/infiniband/hw/mlx4/qp.c pd = to_mpd(ibwq->pd); pd 2173 drivers/infiniband/hw/mlx4/qp.c pd = get_pd(qp); pd 2375 drivers/infiniband/hw/mlx4/qp.c context->pd = cpu_to_be32(pd->pdn); pd 4129 drivers/infiniband/hw/mlx4/qp.c struct ib_wq *mlx4_ib_create_wq(struct ib_pd *pd, pd 4133 drivers/infiniband/hw/mlx4/qp.c struct mlx4_dev *dev = to_mdev(pd->device)->dev; pd 4188 drivers/infiniband/hw/mlx4/qp.c err = create_rq(pd, &ib_qp_init_attr, udata, qp); pd 179 drivers/infiniband/hw/mlx4/srq.c err = mlx4_srq_alloc(dev->dev, to_mpd(ib_srq->pd)->pdn, cqn, xrcdn, pd 272 drivers/infiniband/hw/mlx5/cmd.c MLX5_SET(dealloc_pd_in, in, pd, pdn); pd 1265 drivers/infiniband/hw/mlx5/devx.c mkey->pd = MLX5_GET(mkc, mkc, pd); pd 115 drivers/infiniband/hw/mlx5/gsi.c struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd, pd 118 drivers/infiniband/hw/mlx5/gsi.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 122 drivers/infiniband/hw/mlx5/gsi.c const int num_pkeys = pd->device->attrs.max_pkeys; pd 171 drivers/infiniband/hw/mlx5/gsi.c gsi->cq = ib_alloc_cq(pd->device, gsi, init_attr->cap.max_send_wr, 0, pd 187 drivers/infiniband/hw/mlx5/gsi.c gsi->rx_qp = ib_create_qp(pd, &hw_init_attr); pd 253 drivers/infiniband/hw/mlx5/gsi.c struct ib_pd *pd = gsi->rx_qp->pd; pd 269 drivers/infiniband/hw/mlx5/gsi.c return ib_create_qp(pd, &init_attr); pd 2469 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_pd *pd = to_mpd(ibpd); pd 2487 drivers/infiniband/hw/mlx5/main.c pd->pdn = MLX5_GET(alloc_pd_out, out, pd); pd 2488 drivers/infiniband/hw/mlx5/main.c pd->uid = uid; pd 2490 drivers/infiniband/hw/mlx5/main.c resp.pdn = pd->pdn; pd 2492 drivers/infiniband/hw/mlx5/main.c mlx5_cmd_dealloc_pd(to_mdev(ibdev)->mdev, pd->pdn, uid); pd 2500 drivers/infiniband/hw/mlx5/main.c static void mlx5_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 2502 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_dev *mdev = to_mdev(pd->device); pd 2503 drivers/infiniband/hw/mlx5/main.c struct mlx5_ib_pd *mpd = to_mpd(pd); pd 4347 drivers/infiniband/hw/mlx5/main.c uid = ibqp->pd ? pd 4348 drivers/infiniband/hw/mlx5/main.c to_mpd(ibqp->pd)->uid : 0; pd 4369 drivers/infiniband/hw/mlx5/main.c uid = ibqp->pd ? pd 4370 drivers/infiniband/hw/mlx5/main.c to_mpd(ibqp->pd)->uid : 0; pd 4796 drivers/infiniband/hw/mlx5/main.c if (dev->umrc.pd) pd 4797 drivers/infiniband/hw/mlx5/main.c ib_dealloc_pd(dev->umrc.pd); pd 4808 drivers/infiniband/hw/mlx5/main.c struct ib_pd *pd; pd 4820 drivers/infiniband/hw/mlx5/main.c pd = ib_alloc_pd(&dev->ib_dev, 0); pd 4821 drivers/infiniband/hw/mlx5/main.c if (IS_ERR(pd)) { pd 4823 drivers/infiniband/hw/mlx5/main.c ret = PTR_ERR(pd); pd 4841 drivers/infiniband/hw/mlx5/main.c qp = mlx5_ib_create_qp(pd, init_attr, NULL); pd 4883 drivers/infiniband/hw/mlx5/main.c dev->umrc.pd = pd; pd 4906 drivers/infiniband/hw/mlx5/main.c ib_dealloc_pd(pd); pd 4907 drivers/infiniband/hw/mlx5/main.c dev->umrc.pd = NULL; pd 5002 drivers/infiniband/hw/mlx5/main.c devr->s0->pd = devr->p0; pd 5026 drivers/infiniband/hw/mlx5/main.c devr->s1->pd = devr->p0; pd 480 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_pd *pd; pd 652 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_pd *pd; pd 1120 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd, pd 1147 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc); pd 1148 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 1151 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_advise_mr(struct ib_pd *pd, pd 1157 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 1162 drivers/infiniband/hw/mlx5/mlx5_ib.h struct mlx5_ib_mr *mlx5_ib_alloc_implicit_mr(struct mlx5_ib_pd *pd, pd 1168 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_pd *pd, struct ib_udata *udata); pd 1170 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1172 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_alloc_mr_integrity(struct ib_pd *pd, pd 1229 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd, pd 1245 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_mr *mlx5_ib_reg_dm_mr(struct ib_pd *pd, struct ib_dm *dm, pd 1261 drivers/infiniband/hw/mlx5/mlx5_ib.h int mlx5_ib_advise_mr_prefetch(struct ib_pd *pd, pd 1280 drivers/infiniband/hw/mlx5/mlx5_ib.h mlx5_ib_advise_mr_prefetch(struct ib_pd *pd, pd 1314 drivers/infiniband/hw/mlx5/mlx5_ib.h struct ib_qp *mlx5_ib_gsi_create_qp(struct ib_pd *pd, pd 682 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc) pd 684 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 712 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); pd 850 drivers/infiniband/hw/mlx5/mr.c struct ib_pd *pd, struct ib_umem *umem, pd 854 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 874 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.pd = pd; pd 880 drivers/infiniband/hw/mlx5/mr.c mr->mmkey.pd = to_mpd(pd)->pdn; pd 985 drivers/infiniband/hw/mlx5/mr.c sg.lkey = dev->umrc.pd->local_dma_lkey; pd 995 drivers/infiniband/hw/mlx5/mr.c wr.pd = mr->ibmr.pd; pd 1050 drivers/infiniband/hw/mlx5/mr.c static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd, pd 1056 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1069 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.pd = pd; pd 1101 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); pd 1147 drivers/infiniband/hw/mlx5/mr.c static struct ib_mr *mlx5_ib_get_dm_mr(struct ib_pd *pd, u64 start_addr, pd 1150 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1179 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); pd 1203 drivers/infiniband/hw/mlx5/mr.c int mlx5_ib_advise_mr(struct ib_pd *pd, pd 1214 drivers/infiniband/hw/mlx5/mr.c return mlx5_ib_advise_mr_prefetch(pd, advice, flags, pd 1218 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_reg_dm_mr(struct ib_pd *pd, struct ib_dm *dm, pd 1246 drivers/infiniband/hw/mlx5/mr.c return mlx5_ib_get_dm_mr(pd, start_addr, attr->length, pd 1250 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 1254 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1276 drivers/infiniband/hw/mlx5/mr.c mr = mlx5_ib_alloc_implicit_mr(to_mpd(pd), udata, access_flags); pd 1291 drivers/infiniband/hw/mlx5/mr.c mr = alloc_mr_from_cache(pd, umem, virt_addr, length, ncont, pd 1308 drivers/infiniband/hw/mlx5/mr.c mr = reg_create(NULL, pd, virt_addr, length, umem, ncont, pd 1362 drivers/infiniband/hw/mlx5/mr.c umrwr.pd = dev->umrc.pd; pd 1369 drivers/infiniband/hw/mlx5/mr.c static int rereg_umr(struct ib_pd *pd, struct mlx5_ib_mr *mr, pd 1372 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1382 drivers/infiniband/hw/mlx5/mr.c umrwr.pd = pd; pd 1398 drivers/infiniband/hw/mlx5/mr.c struct ib_pd *pd = (flags & IB_MR_REREG_PD) ? new_pd : ib_mr->pd; pd 1456 drivers/infiniband/hw/mlx5/mr.c mr = reg_create(ib_mr, pd, addr, len, mr->umem, ncont, pd 1470 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.pd = pd; pd 1474 drivers/infiniband/hw/mlx5/mr.c mr->mmkey.pd = to_mpd(pd)->pdn; pd 1485 drivers/infiniband/hw/mlx5/mr.c err = rereg_umr(pd, mr, access_flags, flags); pd 1642 drivers/infiniband/hw/mlx5/mr.c static void mlx5_set_umr_free_mkey(struct ib_pd *pd, u32 *in, int ndescs, pd 1651 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); pd 1659 drivers/infiniband/hw/mlx5/mr.c static int _mlx5_alloc_mkey_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, pd 1663 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1670 drivers/infiniband/hw/mlx5/mr.c err = mlx5_alloc_priv_descs(pd->device, mr, ndescs, desc_size); pd 1674 drivers/infiniband/hw/mlx5/mr.c mlx5_set_umr_free_mkey(pd, in, ndescs, access_mode, page_shift); pd 1691 drivers/infiniband/hw/mlx5/mr.c static struct mlx5_ib_mr *mlx5_ib_alloc_pi_mr(struct ib_pd *pd, pd 1706 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.pd = pd; pd 1707 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.device = pd->device; pd 1718 drivers/infiniband/hw/mlx5/mr.c err = _mlx5_alloc_mkey_descs(pd, mr, ndescs, desc_size, page_shift, pd 1735 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_mem_reg_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, pd 1738 drivers/infiniband/hw/mlx5/mr.c return _mlx5_alloc_mkey_descs(pd, mr, ndescs, sizeof(struct mlx5_mtt), pd 1743 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_sg_gaps_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, pd 1746 drivers/infiniband/hw/mlx5/mr.c return _mlx5_alloc_mkey_descs(pd, mr, ndescs, sizeof(struct mlx5_klm), pd 1750 drivers/infiniband/hw/mlx5/mr.c static int mlx5_alloc_integrity_descs(struct ib_pd *pd, struct mlx5_ib_mr *mr, pd 1754 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1764 drivers/infiniband/hw/mlx5/mr.c err = mlx5_core_create_psv(dev->mdev, to_mpd(pd)->pdn, 2, psv_index); pd 1775 drivers/infiniband/hw/mlx5/mr.c mr->klm_mr = mlx5_ib_alloc_pi_mr(pd, max_num_sg, max_num_meta_sg, pd 1782 drivers/infiniband/hw/mlx5/mr.c mr->mtt_mr = mlx5_ib_alloc_pi_mr(pd, max_num_sg, max_num_meta_sg, pd 1795 drivers/infiniband/hw/mlx5/mr.c err = _mlx5_alloc_mkey_descs(pd, mr, 4, sizeof(struct mlx5_klm), 0, pd 1821 drivers/infiniband/hw/mlx5/mr.c static struct ib_mr *__mlx5_ib_alloc_mr(struct ib_pd *pd, pd 1825 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1842 drivers/infiniband/hw/mlx5/mr.c mr->ibmr.device = pd->device; pd 1847 drivers/infiniband/hw/mlx5/mr.c err = mlx5_alloc_mem_reg_descs(pd, mr, ndescs, in, inlen); pd 1850 drivers/infiniband/hw/mlx5/mr.c err = mlx5_alloc_sg_gaps_descs(pd, mr, ndescs, in, inlen); pd 1853 drivers/infiniband/hw/mlx5/mr.c err = mlx5_alloc_integrity_descs(pd, mr, max_num_sg, pd 1875 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1878 drivers/infiniband/hw/mlx5/mr.c return __mlx5_ib_alloc_mr(pd, mr_type, max_num_sg, 0); pd 1881 drivers/infiniband/hw/mlx5/mr.c struct ib_mr *mlx5_ib_alloc_mr_integrity(struct ib_pd *pd, pd 1884 drivers/infiniband/hw/mlx5/mr.c return __mlx5_ib_alloc_mr(pd, IB_MR_TYPE_INTEGRITY, max_num_sg, pd 1888 drivers/infiniband/hw/mlx5/mr.c struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, pd 1891 drivers/infiniband/hw/mlx5/mr.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1929 drivers/infiniband/hw/mlx5/mr.c MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); pd 2073 drivers/infiniband/hw/mlx5/mr.c u32 lkey = mr->ibmr.pd->local_dma_lkey; pd 166 drivers/infiniband/hw/mlx5/odp.c struct ib_pd *pd = mr->ibmr.pd; pd 167 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 267 drivers/infiniband/hw/mlx5/odp.c if (!mr || !mr->ibmr.pd) pd 419 drivers/infiniband/hw/mlx5/odp.c static struct mlx5_ib_mr *implicit_mr_alloc(struct ib_pd *pd, pd 423 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 433 drivers/infiniband/hw/mlx5/odp.c mr->ibmr.pd = pd; pd 478 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(mr->ibmr.pd->device); pd 502 drivers/infiniband/hw/mlx5/odp.c mtt = implicit_mr_alloc(mr->ibmr.pd, odp, 0, pd 549 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_mr *mlx5_ib_alloc_implicit_mr(struct mlx5_ib_pd *pd, pd 560 drivers/infiniband/hw/mlx5/odp.c imr = implicit_mr_alloc(&pd->ibpd, umem_odp, 1, access_flags); pd 775 drivers/infiniband/hw/mlx5/odp.c struct ib_pd *pd, u32 key, pd 812 drivers/infiniband/hw/mlx5/odp.c if (!smp_load_acquire(&mr->live) || !mr->ibmr.pd) { pd 820 drivers/infiniband/hw/mlx5/odp.c mr->ibmr.pd != pd) { pd 1627 drivers/infiniband/hw/mlx5/odp.c struct ib_pd *pd; pd 1655 drivers/infiniband/hw/mlx5/odp.c static bool num_pending_prefetch_inc(struct ib_pd *pd, pd 1658 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1685 drivers/infiniband/hw/mlx5/odp.c if (mr->ibmr.pd != pd) { pd 1699 drivers/infiniband/hw/mlx5/odp.c static int mlx5_ib_prefetch_sg_list(struct ib_pd *pd, u32 pf_flags, pd 1704 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1710 drivers/infiniband/hw/mlx5/odp.c ret = pagefault_single_data_segment(dev, pd, sg->lkey, sg->addr, pd 1726 drivers/infiniband/hw/mlx5/odp.c if (ib_device_try_get(w->pd->device)) { pd 1727 drivers/infiniband/hw/mlx5/odp.c mlx5_ib_prefetch_sg_list(w->pd, w->pf_flags, w->sg_list, pd 1729 drivers/infiniband/hw/mlx5/odp.c ib_device_put(w->pd->device); pd 1732 drivers/infiniband/hw/mlx5/odp.c num_pending_prefetch_dec(to_mdev(w->pd->device), w->sg_list, pd 1737 drivers/infiniband/hw/mlx5/odp.c int mlx5_ib_advise_mr_prefetch(struct ib_pd *pd, pd 1741 drivers/infiniband/hw/mlx5/odp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 1751 drivers/infiniband/hw/mlx5/odp.c return mlx5_ib_prefetch_sg_list(pd, pf_flags, sg_list, pd 1764 drivers/infiniband/hw/mlx5/odp.c work->pd = pd; pd 1772 drivers/infiniband/hw/mlx5/odp.c valid_req = num_pending_prefetch_inc(pd, sg_list, num_sge); pd 778 drivers/infiniband/hw/mlx5/qp.c static void destroy_user_rq(struct mlx5_ib_dev *dev, struct ib_pd *pd, pd 794 drivers/infiniband/hw/mlx5/qp.c static int create_user_rq(struct mlx5_ib_dev *dev, struct ib_pd *pd, pd 855 drivers/infiniband/hw/mlx5/qp.c static int create_user_qp(struct mlx5_ib_dev *dev, struct ib_pd *pd, pd 937 drivers/infiniband/hw/mlx5/qp.c attr->qp_type != IB_QPT_XRC_INI) ? to_mpd(pd)->uid : 0; pd 985 drivers/infiniband/hw/mlx5/qp.c static void destroy_qp_user(struct mlx5_ib_dev *dev, struct ib_pd *pd, pd 1193 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd) pd 1198 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_tis_in, in, uid, to_mpd(pd)->uid); pd 1207 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_sq *sq, struct ib_pd *pd) pd 1209 drivers/infiniband/hw/mlx5/qp.c mlx5_cmd_destroy_tis(dev->mdev, sq->tisn, to_mpd(pd)->uid); pd 1222 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd) pd 1250 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_sq_in, in, uid, to_mpd(pd)->uid); pd 1266 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd)); pd 1318 drivers/infiniband/hw/mlx5/qp.c size_t qpinlen, struct ib_pd *pd) pd 1339 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid); pd 1357 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(wq, wq, pd, MLX5_GET(qpc, qpc, pd)); pd 1390 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd) pd 1395 drivers/infiniband/hw/mlx5/qp.c mlx5_cmd_destroy_tir(dev->mdev, rq->tirn, to_mpd(pd)->uid); pd 1401 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd, pd 1415 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid); pd 1443 drivers/infiniband/hw/mlx5/qp.c destroy_raw_packet_qp_tir(dev, rq, 0, pd); pd 1452 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd, pd 1463 drivers/infiniband/hw/mlx5/qp.c u16 uid = to_mpd(pd)->uid; pd 1467 drivers/infiniband/hw/mlx5/qp.c err = create_raw_packet_qp_tis(dev, qp, sq, tdn, pd); pd 1471 drivers/infiniband/hw/mlx5/qp.c err = create_raw_packet_qp_sq(dev, udata, sq, in, pd); pd 1493 drivers/infiniband/hw/mlx5/qp.c err = create_raw_packet_qp_rq(dev, rq, in, inlen, pd); pd 1498 drivers/infiniband/hw/mlx5/qp.c dev, rq, tdn, &qp->flags_en, pd, out, pd 1534 drivers/infiniband/hw/mlx5/qp.c destroy_raw_packet_qp_tir(dev, rq, qp->flags_en, pd); pd 1542 drivers/infiniband/hw/mlx5/qp.c destroy_raw_packet_qp_tis(dev, sq, pd); pd 1555 drivers/infiniband/hw/mlx5/qp.c destroy_raw_packet_qp_tir(dev, rq, qp->flags_en, qp->ibqp.pd); pd 1561 drivers/infiniband/hw/mlx5/qp.c destroy_raw_packet_qp_tis(dev, sq, qp->ibqp.pd); pd 1583 drivers/infiniband/hw/mlx5/qp.c to_mpd(qp->ibqp.pd)->uid); pd 1587 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd, pd 1684 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_tir_in, in, uid, to_mpd(pd)->uid); pd 1802 drivers/infiniband/hw/mlx5/qp.c to_mpd(pd)->uid); pd 1945 drivers/infiniband/hw/mlx5/qp.c static int create_qp_common(struct mlx5_ib_dev *dev, struct ib_pd *pd, pd 1978 drivers/infiniband/hw/mlx5/qp.c err = create_rss_raw_qp_tir(dev, qp, pd, init_attr, udata); pd 2127 drivers/infiniband/hw/mlx5/qp.c if (pd) { pd 2147 drivers/infiniband/hw/mlx5/qp.c err = create_user_qp(dev, pd, qp, udata, init_attr, &in, pd 2177 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(qpc, qpc, pd, to_mpd(pd ? pd : devr->p0)->pdn); pd 2284 drivers/infiniband/hw/mlx5/qp.c err = create_raw_packet_qp(dev, qp, in, inlen, pd, udata, pd 2321 drivers/infiniband/hw/mlx5/qp.c destroy_qp_user(dev, pd, qp, base, udata); pd 2390 drivers/infiniband/hw/mlx5/qp.c return to_mpd(qp->ibqp.pd); pd 2540 drivers/infiniband/hw/mlx5/qp.c static struct ib_qp *mlx5_ib_create_dct(struct ib_pd *pd, pd 2569 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_dct_in, qp->dct.in, uid, to_mpd(pd)->uid); pd 2572 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(dctc, dctc, pd, to_mpd(pd)->pdn); pd 2627 drivers/infiniband/hw/mlx5/qp.c struct ib_qp *mlx5_ib_create_qp(struct ib_pd *pd, pd 2640 drivers/infiniband/hw/mlx5/qp.c if (pd) { pd 2641 drivers/infiniband/hw/mlx5/qp.c dev = to_mdev(pd->device); pd 2679 drivers/infiniband/hw/mlx5/qp.c return mlx5_ib_create_dct(pd, init_attr, &ucmd, udata); pd 2709 drivers/infiniband/hw/mlx5/qp.c err = create_qp_common(dev, pd, init_attr, udata, qp); pd 2733 drivers/infiniband/hw/mlx5/qp.c return mlx5_ib_gsi_create_qp(pd, init_attr); pd 2855 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd) pd 2868 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid); pd 2882 drivers/infiniband/hw/mlx5/qp.c struct ib_pd *pd) pd 2895 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_tis_in, in, uid, to_mpd(pd)->uid); pd 2980 drivers/infiniband/hw/mlx5/qp.c sl & 0xf, qp->ibqp.pd); pd 3151 drivers/infiniband/hw/mlx5/qp.c const struct mlx5_modify_raw_qp_param *raw_qp_param, struct ib_pd *pd) pd 3164 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_rq_in, in, uid, to_mpd(pd)->uid); pd 3193 drivers/infiniband/hw/mlx5/qp.c const struct mlx5_modify_raw_qp_param *raw_qp_param, struct ib_pd *pd) pd 3210 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_sq_in, in, uid, to_mpd(pd)->uid); pd 3316 drivers/infiniband/hw/mlx5/qp.c qp->ibqp.pd); pd 3327 drivers/infiniband/hw/mlx5/qp.c qp->ibqp.pd); pd 3338 drivers/infiniband/hw/mlx5/qp.c raw_qp_param, qp->ibqp.pd); pd 3357 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_pd *pd, pd 3457 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_pd *pd; pd 3475 drivers/infiniband/hw/mlx5/qp.c pd = get_pd(qp); pd 3504 drivers/infiniband/hw/mlx5/qp.c tx_affinity = get_tx_affinity(dev, pd, base, p, pd 3566 drivers/infiniband/hw/mlx5/qp.c context->flags_pd = cpu_to_be32(pd ? pd->pdn : to_mpd(dev->devr.p0)->pdn); pd 4379 drivers/infiniband/hw/mlx5/qp.c if (umrwr->pd) pd 4380 drivers/infiniband/hw/mlx5/qp.c seg->flags_pd = cpu_to_be32(to_mpd(umrwr->pd)->pdn); pd 4394 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_pd *pd) pd 4400 drivers/infiniband/hw/mlx5/qp.c dseg->lkey = cpu_to_be32(pd->ibpd.local_dma_lkey); pd 4814 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_pd *pd = to_mpd(qp->ibqp.pd); pd 4815 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_dev *dev = to_mdev(pd->ibpd.device); pd 4857 drivers/infiniband/hw/mlx5/qp.c set_reg_data_seg(*seg, mr, pd); pd 5128 drivers/infiniband/hw/mlx5/qp.c mr->ibmr.pd->local_dma_lkey; pd 5918 drivers/infiniband/hw/mlx5/qp.c static int create_rq(struct mlx5_ib_rwq *rwq, struct ib_pd *pd, pd 5930 drivers/infiniband/hw/mlx5/qp.c dev = to_mdev(pd->device); pd 5937 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(create_rq_in, in, uid, to_mpd(pd)->uid); pd 5968 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(wq, wq, pd, to_mpd(pd)->pdn); pd 6040 drivers/infiniband/hw/mlx5/qp.c static int prepare_user_rq(struct ib_pd *pd, pd 6045 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 6110 drivers/infiniband/hw/mlx5/qp.c err = create_user_rq(dev, pd, udata, rwq, &ucmd); pd 6120 drivers/infiniband/hw/mlx5/qp.c struct ib_wq *mlx5_ib_create_wq(struct ib_pd *pd, pd 6141 drivers/infiniband/hw/mlx5/qp.c dev = to_mdev(pd->device); pd 6147 drivers/infiniband/hw/mlx5/qp.c err = prepare_user_rq(pd, init_attr, udata, rwq); pd 6150 drivers/infiniband/hw/mlx5/qp.c err = create_rq(rwq, pd, init_attr); pd 6177 drivers/infiniband/hw/mlx5/qp.c destroy_user_rq(dev, pd, rwq, udata); pd 6189 drivers/infiniband/hw/mlx5/qp.c destroy_user_rq(dev, wq->pd, rwq, udata); pd 6244 drivers/infiniband/hw/mlx5/qp.c rwq_ind_tbl->uid = to_mpd(init_attr->ind_tbl[0]->pd)->uid; pd 6327 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(modify_rq_in, in, uid, to_mpd(wq->pd)->uid); pd 44 drivers/infiniband/hw/mlx5/srq.c static int create_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq, pd 48 drivers/infiniband/hw/mlx5/srq.c struct mlx5_ib_dev *dev = to_mdev(pd->device); pd 115 drivers/infiniband/hw/mlx5/srq.c in->uid = (in->type != IB_SRQT_XRC) ? to_mpd(pd)->uid : 0; pd 197 drivers/infiniband/hw/mlx5/srq.c static void destroy_srq_user(struct ib_pd *pd, struct mlx5_ib_srq *srq, pd 262 drivers/infiniband/hw/mlx5/srq.c err = create_srq_user(ib_srq->pd, srq, &in, udata, buf_size); pd 299 drivers/infiniband/hw/mlx5/srq.c in.pd = to_mpd(ib_srq->pd)->pdn; pd 329 drivers/infiniband/hw/mlx5/srq.c destroy_srq_user(ib_srq->pd, srq, udata); pd 26 drivers/infiniband/hw/mlx5/srq.h u32 pd; pd 36 drivers/infiniband/hw/mlx5/srq_cmd.c MLX5_SET(wq, wq, pd, in->pd); pd 49 drivers/infiniband/hw/mlx5/srq_cmd.c MLX5_SET(srqc, srqc, pd, in->pd); pd 64 drivers/infiniband/hw/mlx5/srq_cmd.c in->pd = MLX5_GET(wq, wq, pd); pd 77 drivers/infiniband/hw/mlx5/srq_cmd.c in->pd = MLX5_GET(srqc, srqc, pd); pd 196 drivers/infiniband/hw/mthca/mthca_allocator.c union mthca_buf *buf, int *is_direct, struct mthca_pd *pd, pd 262 drivers/infiniband/hw/mthca/mthca_allocator.c err = mthca_mr_alloc_phys(dev, pd->pd_num, pd 154 drivers/infiniband/hw/mthca/mthca_av.c struct mthca_pd *pd, pd 170 drivers/infiniband/hw/mthca/mthca_av.c } else if (!atomic_read(&pd->sqp_count) && pd 197 drivers/infiniband/hw/mthca/mthca_av.c ah->key = pd->ntmr.ibmr.lkey; pd 199 drivers/infiniband/hw/mthca/mthca_av.c av->port_pd = cpu_to_be32(pd->pd_num | pd 70 drivers/infiniband/hw/mthca/mthca_cq.c __be32 pd; pd 833 drivers/infiniband/hw/mthca/mthca_cq.c cq_context->pd = cpu_to_be32(pdn); pd 425 drivers/infiniband/hw/mthca/mthca_dev.h union mthca_buf *buf, int *is_direct, struct mthca_pd *pd, pd 462 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd); pd 463 drivers/infiniband/hw/mthca/mthca_dev.h void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd); pd 471 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift, pd 473 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd, pd 475 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, pd 481 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd, pd 512 drivers/infiniband/hw/mthca/mthca_dev.h int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd, pd 545 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_pd *pd, pd 554 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_pd *pd, pd 565 drivers/infiniband/hw/mthca/mthca_dev.h struct mthca_pd *pd, pd 91 drivers/infiniband/hw/mthca/mthca_mad.c new_ah = rdma_create_ah(dev->send_agent[port_num - 1][0]->qp->pd, pd 54 drivers/infiniband/hw/mthca/mthca_mr.c __be32 pd; pd 429 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc(struct mthca_dev *dev, u32 pd, int buffer_size_shift, pd 468 drivers/infiniband/hw/mthca/mthca_mr.c mpt_entry->pd = cpu_to_be32(pd); pd 512 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd, pd 516 drivers/infiniband/hw/mthca/mthca_mr.c return mthca_mr_alloc(dev, pd, 12, 0, ~0ULL, access, mr); pd 519 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, pd 536 drivers/infiniband/hw/mthca/mthca_mr.c err = mthca_mr_alloc(dev, pd, buffer_size_shift, iova, pd 567 drivers/infiniband/hw/mthca/mthca_mr.c int mthca_fmr_alloc(struct mthca_dev *dev, u32 pd, pd 638 drivers/infiniband/hw/mthca/mthca_mr.c mpt_entry->pd = cpu_to_be32(pd); pd 39 drivers/infiniband/hw/mthca/mthca_pd.c int mthca_pd_alloc(struct mthca_dev *dev, int privileged, struct mthca_pd *pd) pd 43 drivers/infiniband/hw/mthca/mthca_pd.c pd->privileged = privileged; pd 45 drivers/infiniband/hw/mthca/mthca_pd.c atomic_set(&pd->sqp_count, 0); pd 46 drivers/infiniband/hw/mthca/mthca_pd.c pd->pd_num = mthca_alloc(&dev->pd_table.alloc); pd 47 drivers/infiniband/hw/mthca/mthca_pd.c if (pd->pd_num == -1) pd 51 drivers/infiniband/hw/mthca/mthca_pd.c err = mthca_mr_alloc_notrans(dev, pd->pd_num, pd 54 drivers/infiniband/hw/mthca/mthca_pd.c &pd->ntmr); pd 56 drivers/infiniband/hw/mthca/mthca_pd.c mthca_free(&dev->pd_table.alloc, pd->pd_num); pd 62 drivers/infiniband/hw/mthca/mthca_pd.c void mthca_pd_free(struct mthca_dev *dev, struct mthca_pd *pd) pd 64 drivers/infiniband/hw/mthca/mthca_pd.c if (pd->privileged) pd 65 drivers/infiniband/hw/mthca/mthca_pd.c mthca_free_mr(dev, &pd->ntmr); pd 66 drivers/infiniband/hw/mthca/mthca_pd.c mthca_free(&dev->pd_table.alloc, pd->pd_num); pd 369 drivers/infiniband/hw/mthca/mthca_provider.c struct mthca_pd *pd = to_mpd(ibpd); pd 372 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd); pd 377 drivers/infiniband/hw/mthca/mthca_provider.c if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) { pd 378 drivers/infiniband/hw/mthca/mthca_provider.c mthca_pd_free(to_mdev(ibdev), pd); pd 386 drivers/infiniband/hw/mthca/mthca_provider.c static void mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 388 drivers/infiniband/hw/mthca/mthca_provider.c mthca_pd_free(to_mdev(pd->device), to_mpd(pd)); pd 397 drivers/infiniband/hw/mthca/mthca_provider.c return mthca_create_ah(to_mdev(ibah->device), to_mpd(ibah->pd), ah_attr, pd 434 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_alloc_srq(to_mdev(ibsrq->device), to_mpd(ibsrq->pd), pd 468 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_qp *mthca_create_qp(struct ib_pd *pd, pd 496 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(pd->device), &context->uar, pd 504 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_map_user_db(to_mdev(pd->device), &context->uar, pd 508 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(pd->device), pd 521 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_alloc_qp(to_mdev(pd->device), to_mpd(pd), pd 528 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(pd->device), pd 532 drivers/infiniband/hw/mthca/mthca_provider.c mthca_unmap_user_db(to_mdev(pd->device), pd 554 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_alloc_sqp(to_mdev(pd->device), to_mpd(pd), pd 831 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_mr *mthca_get_dma_mr(struct ib_pd *pd, int acc) pd 840 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_mr_alloc_notrans(to_mdev(pd->device), pd 841 drivers/infiniband/hw/mthca/mthca_provider.c to_mpd(pd)->pd_num, pd 854 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_mr *mthca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 857 drivers/infiniband/hw/mthca/mthca_provider.c struct mthca_dev *dev = to_mdev(pd->device); pd 932 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_mr_alloc(dev, to_mpd(pd)->pd_num, PAGE_SHIFT, virt, length, pd 962 drivers/infiniband/hw/mthca/mthca_provider.c static struct ib_fmr *mthca_alloc_fmr(struct ib_pd *pd, int mr_access_flags, pd 973 drivers/infiniband/hw/mthca/mthca_provider.c err = mthca_fmr_alloc(to_mdev(pd->device), to_mpd(pd)->pd_num, pd 134 drivers/infiniband/hw/mthca/mthca_qp.c __be32 pd; pd 709 drivers/infiniband/hw/mthca/mthca_qp.c qp_context->pd = cpu_to_be32(to_mpd(ibqp->pd)->pd_num); pd 956 drivers/infiniband/hw/mthca/mthca_qp.c static inline int mthca_max_inline_data(struct mthca_pd *pd, int max_data_size) pd 959 drivers/infiniband/hw/mthca/mthca_qp.c return pd->ibpd.uobject ? max_data_size - MTHCA_INLINE_HEADER_SIZE : 0; pd 963 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, pd 970 drivers/infiniband/hw/mthca/mthca_qp.c qp->max_inline_data = mthca_max_inline_data(pd, max_data_size); pd 988 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, pd 1069 drivers/infiniband/hw/mthca/mthca_qp.c &qp->queue, &qp->is_direct, pd, 0, &qp->mr); pd 1160 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, pd 1188 drivers/infiniband/hw/mthca/mthca_qp.c ret = mthca_alloc_wqe_buf(dev, pd, qp, udata); pd 1194 drivers/infiniband/hw/mthca/mthca_qp.c mthca_adjust_qp_caps(dev, pd, qp); pd 1250 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, struct mthca_qp *qp) pd 1259 drivers/infiniband/hw/mthca/mthca_qp.c cap->max_inline_data > mthca_max_inline_data(pd, max_data_size)) pd 1289 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, pd 1307 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_set_qp_size(dev, cap, pd, qp); pd 1318 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, pd 1364 drivers/infiniband/hw/mthca/mthca_qp.c struct mthca_pd *pd, pd 1378 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_set_qp_size(dev, cap, pd, &sqp->qp); pd 1402 drivers/infiniband/hw/mthca/mthca_qp.c err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, pd 1407 drivers/infiniband/hw/mthca/mthca_qp.c atomic_inc(&pd->sqp_count); pd 1489 drivers/infiniband/hw/mthca/mthca_qp.c atomic_dec(&(to_mpd(qp->ibqp.pd)->sqp_count)); pd 1559 drivers/infiniband/hw/mthca/mthca_qp.c data->lkey = cpu_to_be32(to_mpd(sqp->qp.ibqp.pd)->ntmr.ibmr.lkey); pd 98 drivers/infiniband/hw/mthca/mthca_srq.c struct mthca_pd *pd, pd 109 drivers/infiniband/hw/mthca/mthca_srq.c context->state_pd = cpu_to_be32(pd->pd_num); pd 119 drivers/infiniband/hw/mthca/mthca_srq.c struct mthca_pd *pd, pd 144 drivers/infiniband/hw/mthca/mthca_srq.c context->eq_pd = cpu_to_be32(MTHCA_EQ_ASYNC << 24 | pd->pd_num); pd 154 drivers/infiniband/hw/mthca/mthca_srq.c static int mthca_alloc_srq_buf(struct mthca_dev *dev, struct mthca_pd *pd, pd 171 drivers/infiniband/hw/mthca/mthca_srq.c &srq->queue, &srq->is_direct, pd, 1, &srq->mr); pd 206 drivers/infiniband/hw/mthca/mthca_srq.c int mthca_alloc_srq(struct mthca_dev *dev, struct mthca_pd *pd, pd 262 drivers/infiniband/hw/mthca/mthca_srq.c err = mthca_alloc_srq_buf(dev, pd, srq, udata); pd 272 drivers/infiniband/hw/mthca/mthca_srq.c mthca_arbel_init_srq_context(dev, pd, srq, mailbox->buf, udata); pd 274 drivers/infiniband/hw/mthca/mthca_srq.c mthca_tavor_init_srq_context(dev, pd, srq, mailbox->buf, udata); pd 386 drivers/infiniband/hw/ocrdma/ocrdma.h struct ocrdma_pd *pd; pd 423 drivers/infiniband/hw/ocrdma/ocrdma.h struct ocrdma_pd *pd; pd 167 drivers/infiniband/hw/ocrdma/ocrdma_ah.c struct ocrdma_pd *pd = get_ocrdma_pd(ibah->pd); pd 189 drivers/infiniband/hw/ocrdma/ocrdma_ah.c status = set_av_attr(dev, ah, attr, &sgid_attr->gid, pd->id, pd 195 drivers/infiniband/hw/ocrdma/ocrdma_ah.c if ((pd->uctx) && (pd->uctx->ah_tbl.va)) { pd 196 drivers/infiniband/hw/ocrdma/ocrdma_ah.c ahid_addr = pd->uctx->ah_tbl.va + rdma_ah_get_dlid(attr); pd 1463 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_mbx_alloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd) pd 1472 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if (pd->dpp_enabled) pd 1478 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->id = rsp->dpp_page_pdid & OCRDMA_ALLOC_PD_RSP_PDID_MASK; pd 1480 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->dpp_enabled = true; pd 1481 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->dpp_page = rsp->dpp_page_pdid >> pd 1484 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->dpp_enabled = false; pd 1485 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->num_dpp_qp = 0; pd 1492 drivers/infiniband/hw/ocrdma/ocrdma_hw.c int ocrdma_mbx_dealloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd) pd 1500 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->id = pd->id; pd 2188 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd = qp->pd; pd 2189 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 2243 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd = qp->pd; pd 2244 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 2284 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd, pd 2288 drivers/infiniband/hw/ocrdma/ocrdma_hw.c pd->num_dpp_qp--; pd 2302 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd = qp->pd; pd 2303 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 2373 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd = qp->pd; pd 2374 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 2419 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->type_pgsz_pdn |= (pd->id << OCRDMA_CREATE_QP_REQ_PD_ID_SHIFT) & pd 2440 drivers/infiniband/hw/ocrdma/ocrdma_hw.c if (pd->dpp_enabled && attrs->cap.max_inline_data && pd->num_dpp_qp && pd 2442 drivers/infiniband/hw/ocrdma/ocrdma_hw.c ocrdma_set_create_qp_dpp_cmd(cmd, pd, qp, enable_dpp_cq, pd 2739 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp->pd->num_dpp_qp++; pd 2745 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd) pd 2760 drivers/infiniband/hw/ocrdma/ocrdma_hw.c cmd->pgsz_pdid = pd->id & OCRDMA_CREATE_SRQ_PD_ID_MASK; pd 2822 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_pd *pd = srq->pd; pd 2823 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 321 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int ocrdma_get_pd_num(struct ocrdma_dev *dev, struct ocrdma_pd *pd) pd 327 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->dpp_enabled) { pd 331 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->id = dev->pd_mgr->pd_dpp_start + pd_idx; pd 332 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->dpp_page = dev->pd_mgr->dpp_page_index + pd_idx; pd 336 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->id = dev->pd_mgr->pd_norm_start + pd_idx; pd 337 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->dpp_enabled = false; pd 344 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->id = dev->pd_mgr->pd_norm_start + pd_idx; pd 363 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int _ocrdma_alloc_pd(struct ocrdma_dev *dev, struct ocrdma_pd *pd, pd 370 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->dpp_enabled = pd 372 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->num_dpp_qp = pd 373 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->dpp_enabled ? (dev->nic_info.db_page_size / pd 378 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c return ocrdma_get_pd_num(dev, pd); pd 381 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_mbx_alloc_pd(dev, pd); pd 383 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->dpp_enabled) { pd 384 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->dpp_enabled = false; pd 385 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->num_dpp_qp = 0; pd 395 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd) pd 397 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c return (uctx->cntxt_pd == pd); pd 401 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd) pd 404 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_put_pd_num(dev, pd->id, pd->dpp_enabled); pd 406 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_mbx_dealloc_pd(dev, pd); pd 414 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ib_pd *pd; pd 417 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = rdma_zalloc_drv_obj(ibdev, ib_pd); pd 418 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (!pd) pd 421 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->device = ibdev; pd 422 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c uctx->cntxt_pd = get_ocrdma_pd(pd); pd 438 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = uctx->cntxt_pd; pd 439 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 443 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c __func__, dev->id, pd->id); pd 447 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c _ocrdma_dealloc_pd(dev, pd); pd 452 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = NULL; pd 457 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = uctx->cntxt_pd; pd 461 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c return pd; pd 588 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c static int ocrdma_copy_pd_uresp(struct ocrdma_dev *dev, struct ocrdma_pd *pd, pd 600 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c rsp.id = pd->id; pd 601 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c rsp.dpp_enabled = pd->dpp_enabled; pd 602 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c db_page_addr = ocrdma_get_db_addr(dev, pd->id); pd 609 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->dpp_enabled) { pd 611 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * PAGE_SIZE); pd 624 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd->uctx = uctx; pd 628 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->dpp_enabled) pd 629 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, dpp_page_addr, PAGE_SIZE); pd 631 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, db_page_addr, db_page_size); pd 639 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd; pd 646 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = ocrdma_get_ucontext_pd(uctx); pd 647 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd) { pd 653 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = get_ocrdma_pd(ibpd); pd 654 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = _ocrdma_alloc_pd(dev, pd, uctx, udata); pd 660 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_copy_pd_uresp(dev, pd, udata); pd 670 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c _ocrdma_dealloc_pd(dev, pd); pd 677 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = get_ocrdma_pd(ibpd); pd 682 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c uctx = pd->uctx; pd 685 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * PAGE_SIZE); pd 686 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->dpp_enabled) pd 687 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, dpp_db, PAGE_SIZE); pd 688 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c usr_db = ocrdma_get_db_addr(dev, pd->id); pd 689 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, usr_db, dev->nic_info.db_page_size); pd 691 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (is_ucontext_pd(uctx, pd)) { pd 696 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c _ocrdma_dealloc_pd(dev, pd); pd 727 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = get_ocrdma_pd(ibpd); pd 739 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_alloc_lkey(dev, mr, pd->id, acc, 0, pd 867 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd; pd 870 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = get_ocrdma_pd(ibpd); pd 901 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_reg_mr(dev, &mr->hwmr, pd->id, acc); pd 1191 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = qp->pd; pd 1192 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_dev *dev = get_ocrdma_dev(pd->ibpd.device); pd 1196 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * dev->nic_info.db_page_size); pd 1225 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_add_mmap(pd->uctx, uresp.sq_page_addr[0], pd 1231 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_add_mmap(pd->uctx, uresp.rq_page_addr[0], pd 1238 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, uresp.sq_page_addr[0], uresp.sq_page_size); pd 1244 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd) pd 1248 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * dev->nic_info.db_page_size) + pd 1251 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * dev->nic_info.db_page_size) + pd 1255 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * dev->nic_info.db_page_size) + pd 1258 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (pd->id * dev->nic_info.db_page_size) + pd 1279 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd, pd 1282 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c qp->pd = pd; pd 1311 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = get_ocrdma_pd(ibpd); pd 1331 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_set_qp_init_params(qp, pd, attrs); pd 1353 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_set_qp_db(dev, qp, pd); pd 1686 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd; pd 1696 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c pd = qp->pd; pd 1725 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (!pd->uctx) { pd 1731 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (pd->uctx) { pd 1732 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, (u64) qp->sq.pa, pd 1735 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(pd->uctx, (u64) qp->rq.pa, pd 1759 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c (srq->pd->id * dev->nic_info.db_page_size); pd 1773 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_add_mmap(srq->pd->uctx, uresp.rq_page_addr[0], pd 1784 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = get_ocrdma_pd(ibsrq->pd); pd 1794 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c srq->pd = pd; pd 1795 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c srq->db = dev->nic_info.db + (pd->id * dev->nic_info.db_page_size); pd 1796 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_mbx_create_srq(dev, srq, init_attr, pd); pd 1877 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c if (srq->pd->uctx) pd 1878 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c ocrdma_del_mmap(srq->pd->uctx, (u64) srq->rq.pa, pd 2916 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c struct ocrdma_pd *pd = get_ocrdma_pd(ibpd); pd 2947 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c status = ocrdma_reg_mr(dev, &mr->hwmr, pd->id, 0); pd 71 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h int ocrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 72 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h void ocrdma_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 105 drivers/infiniband/hw/ocrdma/ocrdma_verbs.h struct ib_mr *ocrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 232 drivers/infiniband/hw/qedr/qedr.h struct qedr_pd *pd; pd 407 drivers/infiniband/hw/qedr/qedr.h struct qedr_pd *pd; pd 438 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 454 drivers/infiniband/hw/qedr/verbs.c pd->pd_id = pd_id; pd 470 drivers/infiniband/hw/qedr/verbs.c pd->uctx = context; pd 471 drivers/infiniband/hw/qedr/verbs.c pd->uctx->pd = pd; pd 480 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 482 drivers/infiniband/hw/qedr/verbs.c DP_DEBUG(dev, QEDR_MSG_INIT, "Deallocating PD %d\n", pd->pd_id); pd 483 drivers/infiniband/hw/qedr/verbs.c dev->ops->rdma_dealloc_pd(dev->rdma_ctx, pd->pd_id); pd 1193 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd, pd 1201 drivers/infiniband/hw/qedr/verbs.c qp->pd = pd; pd 1222 drivers/infiniband/hw/qedr/verbs.c pd->pd_id, qp->qp_type, qp->max_inline_data, pd 1364 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibsrq->pd); pd 1374 drivers/infiniband/hw/qedr/verbs.c (udata) ? "User lib" : "kernel", pd); pd 1416 drivers/infiniband/hw/qedr/verbs.c in_params.pd_id = pd->pd_id; pd 1515 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd, pd 1527 drivers/infiniband/hw/qedr/verbs.c params->pd = pd->pd_id; pd 1528 drivers/infiniband/hw/qedr/verbs.c params->dpi = pd->uctx ? pd->uctx->dpi : dev->dpi; pd 1602 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 1630 drivers/infiniband/hw/qedr/verbs.c qedr_init_common_qp_in_params(dev, pd, qp, attrs, false, &in_params); pd 1814 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 1863 drivers/infiniband/hw/qedr/verbs.c qedr_init_common_qp_in_params(dev, pd, qp, attrs, true, &in_params); pd 1889 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 1895 drivers/infiniband/hw/qedr/verbs.c udata ? "user library" : "kernel", pd); pd 1903 drivers/infiniband/hw/qedr/verbs.c udata ? "user library" : "kernel", attrs->event_handler, pd, pd 1915 drivers/infiniband/hw/qedr/verbs.c qedr_set_common_qp_params(dev, qp, pd, attrs); pd 2607 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd; pd 2610 drivers/infiniband/hw/qedr/verbs.c pd = get_qedr_pd(ibpd); pd 2613 drivers/infiniband/hw/qedr/verbs.c pd->pd_id, start, len, usr_addr, acc); pd 2646 drivers/infiniband/hw/qedr/verbs.c mr->hw_mr.pd = pd->pd_id; pd 2714 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 2720 drivers/infiniband/hw/qedr/verbs.c "qedr_alloc_frmr pd = %d max_page_list_len= %d\n", pd->pd_id, pd 2743 drivers/infiniband/hw/qedr/verbs.c mr->hw_mr.pd = pd->pd_id; pd 2856 drivers/infiniband/hw/qedr/verbs.c struct qedr_pd *pd = get_qedr_pd(ibpd); pd 2874 drivers/infiniband/hw/qedr/verbs.c mr->hw_mr.pd = pd->pd_id; pd 50 drivers/infiniband/hw/qedr/verbs.h int qedr_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 51 drivers/infiniband/hw/qedr/verbs.h void qedr_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 87 drivers/infiniband/hw/qedr/verbs.h struct ib_mr *qedr_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 40 drivers/infiniband/hw/qib/qib_user_sdma.h int qib_user_sdma_writev(struct qib_ctxtdata *pd, pd 1355 drivers/infiniband/hw/qib/qib_verbs.c ah = rdma_create_ah(qp0->ibqp.pd, &attr, 0); pd 96 drivers/infiniband/hw/usnic/usnic_ib.h struct usnic_ib_pd *pd; pd 587 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c struct usnic_ib_pd *pd, pd 597 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c err = usnic_uiom_attach_dev_to_pd(pd->umem_pd, &pdev->dev); pd 603 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c vf->pd = pd; pd 607 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c WARN_ON(vf->pd != pd); pd 616 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c struct usnic_ib_pd *pd; pd 620 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c pd = qp_grp->vf->pd; pd 623 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c qp_grp->vf->pd = NULL; pd 624 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c usnic_uiom_detach_dev_from_pd(pd->umem_pd, &pdev->dev); pd 670 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c struct usnic_ib_pd *pd, pd 702 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c err = qp_grp_and_vf_bind(vf, pd, qp_grp); pd 94 drivers/infiniband/hw/usnic/usnic_ib_qp_grp.h struct usnic_ib_pd *pd, pd 173 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_pd *pd, pd 192 drivers/infiniband/hw/usnic/usnic_ib_verbs.c dev_list = usnic_uiom_get_dev_list(pd->umem_pd); pd 206 drivers/infiniband/hw/usnic/usnic_ib_verbs.c vf, pd, pd 226 drivers/infiniband/hw/usnic/usnic_ib_verbs.c pd, res_spec, pd 453 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct usnic_ib_pd *pd = to_upd(ibpd); pd 456 drivers/infiniband/hw/usnic/usnic_ib_verbs.c umem_pd = pd->umem_pd = usnic_uiom_alloc_pd(); pd 464 drivers/infiniband/hw/usnic/usnic_ib_verbs.c void usnic_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 466 drivers/infiniband/hw/usnic/usnic_ib_verbs.c usnic_uiom_dealloc_pd((to_upd(pd))->umem_pd); pd 469 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct ib_qp *usnic_ib_create_qp(struct ib_pd *pd, pd 485 drivers/infiniband/hw/usnic/usnic_ib_verbs.c us_ibdev = to_usdev(pd->device); pd 515 drivers/infiniband/hw/usnic/usnic_ib_verbs.c qp_grp = find_free_vf_and_create_qp_grp(us_ibdev, to_upd(pd), pd 605 drivers/infiniband/hw/usnic/usnic_ib_verbs.c struct ib_mr *usnic_ib_reg_mr(struct ib_pd *pd, u64 start, u64 length, pd 619 drivers/infiniband/hw/usnic/usnic_ib_verbs.c mr->umem = usnic_uiom_reg_get(to_upd(pd)->umem_pd, start, length, pd 54 drivers/infiniband/hw/usnic/usnic_ib_verbs.h void usnic_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 55 drivers/infiniband/hw/usnic/usnic_ib_verbs.h struct ib_qp *usnic_ib_create_qp(struct ib_pd *pd, pd 64 drivers/infiniband/hw/usnic/usnic_ib_verbs.h struct ib_mr *usnic_ib_reg_mr(struct ib_pd *pd, u64 start, u64 length, pd 196 drivers/infiniband/hw/usnic/usnic_uiom.c struct usnic_uiom_pd *pd) pd 207 drivers/infiniband/hw/usnic/usnic_uiom.c iommu_unmap(pd->domain, va, PAGE_SIZE); pd 214 drivers/infiniband/hw/usnic/usnic_uiom.c static void __usnic_uiom_reg_release(struct usnic_uiom_pd *pd, pd 228 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock(&pd->lock); pd 229 drivers/infiniband/hw/usnic/usnic_uiom.c usnic_uiom_remove_interval(&pd->root, vpn_start, pd 231 drivers/infiniband/hw/usnic/usnic_uiom.c usnic_uiom_unmap_sorted_intervals(&rm_intervals, pd); pd 241 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 255 drivers/infiniband/hw/usnic/usnic_uiom.c struct usnic_uiom_pd *pd = uiomr->pd; pd 284 drivers/infiniband/hw/usnic/usnic_uiom.c err = iommu_map(pd->domain, va_start, pa_start, pd 301 drivers/infiniband/hw/usnic/usnic_uiom.c err = iommu_map(pd->domain, va_start, pa_start, pd 330 drivers/infiniband/hw/usnic/usnic_uiom.c usnic_uiom_unmap_sorted_intervals(intervals, pd); pd 334 drivers/infiniband/hw/usnic/usnic_uiom.c struct usnic_uiom_reg *usnic_uiom_reg_get(struct usnic_uiom_pd *pd, pd 367 drivers/infiniband/hw/usnic/usnic_uiom.c uiomr->pd = pd; pd 377 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock(&pd->lock); pd 381 drivers/infiniband/hw/usnic/usnic_uiom.c &pd->root, pd 397 drivers/infiniband/hw/usnic/usnic_uiom.c err = usnic_uiom_insert_interval(&pd->root, vpn_start, vpn_last, pd 406 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 411 drivers/infiniband/hw/usnic/usnic_uiom.c usnic_uiom_unmap_sorted_intervals(&sorted_diff_intervals, pd); pd 416 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 436 drivers/infiniband/hw/usnic/usnic_uiom.c __usnic_uiom_reg_release(uiomr->pd, uiomr, 1); pd 444 drivers/infiniband/hw/usnic/usnic_uiom.c struct usnic_uiom_pd *pd; pd 447 drivers/infiniband/hw/usnic/usnic_uiom.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 448 drivers/infiniband/hw/usnic/usnic_uiom.c if (!pd) pd 451 drivers/infiniband/hw/usnic/usnic_uiom.c pd->domain = domain = iommu_domain_alloc(&pci_bus_type); pd 454 drivers/infiniband/hw/usnic/usnic_uiom.c kfree(pd); pd 458 drivers/infiniband/hw/usnic/usnic_uiom.c iommu_set_fault_handler(pd->domain, usnic_uiom_dma_fault, NULL); pd 460 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock_init(&pd->lock); pd 461 drivers/infiniband/hw/usnic/usnic_uiom.c INIT_LIST_HEAD(&pd->devs); pd 463 drivers/infiniband/hw/usnic/usnic_uiom.c return pd; pd 466 drivers/infiniband/hw/usnic/usnic_uiom.c void usnic_uiom_dealloc_pd(struct usnic_uiom_pd *pd) pd 468 drivers/infiniband/hw/usnic/usnic_uiom.c iommu_domain_free(pd->domain); pd 469 drivers/infiniband/hw/usnic/usnic_uiom.c kfree(pd); pd 472 drivers/infiniband/hw/usnic/usnic_uiom.c int usnic_uiom_attach_dev_to_pd(struct usnic_uiom_pd *pd, struct device *dev) pd 482 drivers/infiniband/hw/usnic/usnic_uiom.c err = iommu_attach_device(pd->domain, dev); pd 493 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock(&pd->lock); pd 494 drivers/infiniband/hw/usnic/usnic_uiom.c list_add_tail(&uiom_dev->link, &pd->devs); pd 495 drivers/infiniband/hw/usnic/usnic_uiom.c pd->dev_cnt++; pd 496 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 501 drivers/infiniband/hw/usnic/usnic_uiom.c iommu_detach_device(pd->domain, dev); pd 507 drivers/infiniband/hw/usnic/usnic_uiom.c void usnic_uiom_detach_dev_from_pd(struct usnic_uiom_pd *pd, struct device *dev) pd 512 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock(&pd->lock); pd 513 drivers/infiniband/hw/usnic/usnic_uiom.c list_for_each_entry(uiom_dev, &pd->devs, link) { pd 523 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 528 drivers/infiniband/hw/usnic/usnic_uiom.c pd->dev_cnt--; pd 529 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 531 drivers/infiniband/hw/usnic/usnic_uiom.c return iommu_detach_device(pd->domain, dev); pd 534 drivers/infiniband/hw/usnic/usnic_uiom.c struct device **usnic_uiom_get_dev_list(struct usnic_uiom_pd *pd) pd 540 drivers/infiniband/hw/usnic/usnic_uiom.c spin_lock(&pd->lock); pd 541 drivers/infiniband/hw/usnic/usnic_uiom.c devs = kcalloc(pd->dev_cnt + 1, sizeof(*devs), GFP_ATOMIC); pd 547 drivers/infiniband/hw/usnic/usnic_uiom.c list_for_each_entry(uiom_dev, &pd->devs, link) { pd 551 drivers/infiniband/hw/usnic/usnic_uiom.c spin_unlock(&pd->lock); pd 66 drivers/infiniband/hw/usnic/usnic_uiom.h struct usnic_uiom_pd *pd; pd 84 drivers/infiniband/hw/usnic/usnic_uiom.h void usnic_uiom_dealloc_pd(struct usnic_uiom_pd *pd); pd 85 drivers/infiniband/hw/usnic/usnic_uiom.h int usnic_uiom_attach_dev_to_pd(struct usnic_uiom_pd *pd, struct device *dev); pd 86 drivers/infiniband/hw/usnic/usnic_uiom.h void usnic_uiom_detach_dev_from_pd(struct usnic_uiom_pd *pd, pd 88 drivers/infiniband/hw/usnic/usnic_uiom.h struct device **usnic_uiom_get_dev_list(struct usnic_uiom_pd *pd); pd 90 drivers/infiniband/hw/usnic/usnic_uiom.h struct usnic_uiom_reg *usnic_uiom_reg_get(struct usnic_uiom_pd *pd, pd 58 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_get_dma_mr(struct ib_pd *pd, int acc) pd 60 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct pvrdma_dev *dev = to_vdev(pd->device); pd 81 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c cmd->pd_handle = to_vpd(pd)->pd_handle; pd 111 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 115 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct pvrdma_dev *dev = to_vdev(pd->device); pd 169 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c cmd->pd_handle = to_vpd(pd)->pd_handle; pd 204 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct ib_mr *pvrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 207 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c struct pvrdma_dev *dev = to_vdev(pd->device); pd 240 drivers/infiniband/hw/vmw_pvrdma/pvrdma_mr.c cmd->pd_handle = to_vpd(pd)->pd_handle; pd 188 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct ib_qp *pvrdma_create_qp(struct ib_pd *pd, pd 193 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c struct pvrdma_dev *dev = to_vdev(pd->device); pd 230 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c init_attr->port_num > pd->device->phys_port_cnt || pd 293 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c ret = pvrdma_set_sq_size(to_vdev(pd->device), pd 298 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c ret = pvrdma_set_rq_size(to_vdev(pd->device), pd 349 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c cmd->pd_handle = to_vpd(pd)->pd_handle; pd 177 drivers/infiniband/hw/vmw_pvrdma/pvrdma_srq.c cmd->pd_handle = to_vpd(ibsrq->pd)->pd_handle; pd 430 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_pd *pd = to_vpd(ibpd); pd 455 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pd->privileged = !udata; pd 456 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pd->pd_handle = resp->pd_handle; pd 457 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pd->pdn = resp->pd_handle; pd 464 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c pvrdma_dealloc_pd(&pd->ibpd, udata); pd 484 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c void pvrdma_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 486 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c struct pvrdma_dev *dev = to_vdev(pd->device); pd 492 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c cmd->pd_handle = to_vpd(pd)->pd_handle; pd 531 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.c ah->av.port_pd = to_vpd(ibah->pd)->pd_handle | (port_num << 24); pd 401 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h int pvrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 403 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_get_dma_mr(struct ib_pd *pd, int acc); pd 404 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 408 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_mr *pvrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 428 drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h struct ib_qp *pvrdma_create_qp(struct ib_pd *pd, pd 136 drivers/infiniband/sw/rdmavt/mr.c static int rvt_init_mregion(struct rvt_mregion *mr, struct ib_pd *pd, pd 140 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *dev = ib_to_rvt(pd->device); pd 158 drivers/infiniband/sw/rdmavt/mr.c mr->pd = pd; pd 184 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *dev = ib_to_rvt(mr->pd->device); pd 253 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *dev = ib_to_rvt(mr->pd->device); pd 280 drivers/infiniband/sw/rdmavt/mr.c static struct rvt_mr *__rvt_alloc_mr(int count, struct ib_pd *pd) pd 292 drivers/infiniband/sw/rdmavt/mr.c rval = rvt_init_mregion(&mr->mr, pd, count, 0); pd 331 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_get_dma_mr(struct ib_pd *pd, int acc) pd 337 drivers/infiniband/sw/rdmavt/mr.c if (ibpd_to_rvtpd(pd)->user) pd 346 drivers/infiniband/sw/rdmavt/mr.c rval = rvt_init_mregion(&mr->mr, pd, 0, 0); pd 380 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 399 drivers/infiniband/sw/rdmavt/mr.c mr = __rvt_alloc_mr(n, pd); pd 456 drivers/infiniband/sw/rdmavt/mr.c if (mr->pd != qp->ibqp.pd) pd 470 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *rdi = ib_to_rvt(mr->pd->device); pd 489 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *rdi = ib_to_rvt(mr->pd->device); pd 502 drivers/infiniband/sw/rdmavt/mr.c t, mr, mr->pd, mr->lkey, pd 579 drivers/infiniband/sw/rdmavt/mr.c struct ib_mr *rvt_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 587 drivers/infiniband/sw/rdmavt/mr.c mr = __rvt_alloc_mr(max_num_sg, pd); pd 663 drivers/infiniband/sw/rdmavt/mr.c if (qp->ibqp.pd != mr->mr.pd) pd 703 drivers/infiniband/sw/rdmavt/mr.c if (unlikely(!mr || mr->lkey != rkey || qp->ibqp.pd != mr->pd)) pd 724 drivers/infiniband/sw/rdmavt/mr.c struct ib_fmr *rvt_alloc_fmr(struct ib_pd *pd, int mr_access_flags, pd 738 drivers/infiniband/sw/rdmavt/mr.c rval = rvt_init_mregion(&fmr->mr, pd, fmr_attr->max_pages, pd 914 drivers/infiniband/sw/rdmavt/mr.c int rvt_lkey_ok(struct rvt_lkey_table *rkt, struct rvt_pd *pd, pd 927 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *dev = ib_to_rvt(pd->ibpd.device); pd 929 drivers/infiniband/sw/rdmavt/mr.c if (pd->user) pd 959 drivers/infiniband/sw/rdmavt/mr.c mr->lkey != sge->lkey || mr->pd != &pd->ibpd)) pd 1039 drivers/infiniband/sw/rdmavt/mr.c struct rvt_pd *pd = ibpd_to_rvtpd(qp->ibqp.pd); pd 1040 drivers/infiniband/sw/rdmavt/mr.c struct rvt_dev_info *rdi = ib_to_rvt(pd->ibpd.device); pd 1042 drivers/infiniband/sw/rdmavt/mr.c if (pd->user) pd 1067 drivers/infiniband/sw/rdmavt/mr.c mr->lkey != rkey || qp->ibqp.pd != mr->pd)) pd 77 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_get_dma_mr(struct ib_pd *pd, int acc); pd 78 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, pd 82 drivers/infiniband/sw/rdmavt/mr.h struct ib_mr *rvt_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 86 drivers/infiniband/sw/rdmavt/mr.h struct ib_fmr *rvt_alloc_fmr(struct ib_pd *pd, int mr_access_flags, pd 64 drivers/infiniband/sw/rdmavt/pd.c struct rvt_pd *pd = ibpd_to_rvtpd(ibpd); pd 85 drivers/infiniband/sw/rdmavt/pd.c pd->user = !!udata; pd 53 drivers/infiniband/sw/rdmavt/pd.h int rvt_alloc_pd(struct ib_pd *pd, struct ib_udata *udata); pd 1914 drivers/infiniband/sw/rdmavt/qp.c ibpd_to_rvtpd(qp->ibqp.pd)->user) pd 1928 drivers/infiniband/sw/rdmavt/qp.c if (qp->ibqp.pd != ud_wr(wr)->ah->pd) pd 2009 drivers/infiniband/sw/rdmavt/qp.c struct rvt_pd *pd; pd 2076 drivers/infiniband/sw/rdmavt/qp.c pd = ibpd_to_rvtpd(qp->ibqp.pd); pd 2094 drivers/infiniband/sw/rdmavt/qp.c ret = rvt_lkey_ok(rkt, pd, &wqe->sg_list[j], last_sge, pd 2311 drivers/infiniband/sw/rdmavt/qp.c struct rvt_pd *pd; pd 2316 drivers/infiniband/sw/rdmavt/qp.c pd = ibpd_to_rvtpd(qp->ibqp.srq ? qp->ibqp.srq->pd : qp->ibqp.pd); pd 2324 drivers/infiniband/sw/rdmavt/qp.c ret = rvt_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge, pd 66 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ENTRY(ib_to_rvt(mr->pd->device)) pd 79 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ASSIGN(ib_to_rvt(mr->pd->device)); pd 127 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ENTRY(ib_to_rvt(sge->mr->pd->device)) pd 142 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ASSIGN(ib_to_rvt(sge->mr->pd->device)); pd 154 drivers/infiniband/sw/rdmavt/trace_mr.h __entry->user = ibpd_to_rvtpd(sge->mr->pd)->user; pd 189 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ENTRY(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) pd 198 drivers/infiniband/sw/rdmavt/trace_mr.h RDI_DEV_ASSIGN(ib_to_rvt(to_imr(ibmr)->mr.pd->device)) pd 373 drivers/infiniband/sw/rxe/rxe_comp.c ret = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE, pd 393 drivers/infiniband/sw/rxe/rxe_comp.c ret = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE, pd 106 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_dma(struct rxe_pd *pd, pd 109 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_user(struct rxe_pd *pd, u64 start, pd 113 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_mem_init_fast(struct rxe_pd *pd, pd 119 drivers/infiniband/sw/rxe/rxe_loc.h int copy_data(struct rxe_pd *pd, int access, pd 130 drivers/infiniband/sw/rxe/rxe_loc.h struct rxe_mem *lookup_mem(struct rxe_pd *pd, int access, u32 key, pd 157 drivers/infiniband/sw/rxe/rxe_loc.h int rxe_qp_from_init(struct rxe_dev *rxe, struct rxe_qp *qp, struct rxe_pd *pd, pd 147 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_dma(struct rxe_pd *pd, pd 152 drivers/infiniband/sw/rxe/rxe_mr.c mem->pd = pd; pd 160 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_user(struct rxe_pd *pd, u64 start, pd 222 drivers/infiniband/sw/rxe/rxe_mr.c mem->pd = pd; pd 238 drivers/infiniband/sw/rxe/rxe_mr.c int rxe_mem_init_fast(struct rxe_pd *pd, pd 252 drivers/infiniband/sw/rxe/rxe_mr.c mem->pd = pd; pd 372 drivers/infiniband/sw/rxe/rxe_mr.c *crcp = rxe_crc32(to_rdev(mem->pd->ibpd.device), pd 406 drivers/infiniband/sw/rxe/rxe_mr.c crc = rxe_crc32(to_rdev(mem->pd->ibpd.device), pd 436 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_pd *pd, pd 461 drivers/infiniband/sw/rxe/rxe_mr.c mem = lookup_mem(pd, access, sge->lkey, lookup_local); pd 486 drivers/infiniband/sw/rxe/rxe_mr.c mem = lookup_mem(pd, access, sge->lkey, pd 568 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_mem *lookup_mem(struct rxe_pd *pd, int access, u32 key, pd 572 drivers/infiniband/sw/rxe/rxe_mr.c struct rxe_dev *rxe = to_rdev(pd->ibpd.device); pd 581 drivers/infiniband/sw/rxe/rxe_mr.c mem->pd != pd || pd 336 drivers/infiniband/sw/rxe/rxe_qp.c int rxe_qp_from_init(struct rxe_dev *rxe, struct rxe_qp *qp, struct rxe_pd *pd, pd 347 drivers/infiniband/sw/rxe/rxe_qp.c rxe_add_ref(pd); pd 353 drivers/infiniband/sw/rxe/rxe_qp.c qp->pd = pd; pd 380 drivers/infiniband/sw/rxe/rxe_qp.c rxe_drop_ref(pd); pd 824 drivers/infiniband/sw/rxe/rxe_qp.c if (qp->pd) pd 825 drivers/infiniband/sw/rxe/rxe_qp.c rxe_drop_ref(qp->pd); pd 496 drivers/infiniband/sw/rxe/rxe_req.c err = copy_data(qp->pd, 0, &wqe->dma, pd 459 drivers/infiniband/sw/rxe/rxe_resp.c mem = lookup_mem(qp->pd, access, rkey, lookup_remote); pd 512 drivers/infiniband/sw/rxe/rxe_resp.c err = copy_data(qp->pd, IB_ACCESS_LOCAL_WRITE, &qp->resp.wqe->dma, pd 182 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 184 drivers/infiniband/sw/rxe/rxe_verbs.c return rxe_add_to_pool(&rxe->pd_pool, &pd->pelem); pd 189 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 191 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(pd); pd 297 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibsrq->pd); pd 315 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_add_ref(pd); pd 316 drivers/infiniband/sw/rxe/rxe_verbs.c srq->pd = pd; pd 325 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(pd); pd 383 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(srq->pd); pd 417 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 447 drivers/infiniband/sw/rxe/rxe_verbs.c err = rxe_qp_from_init(rxe, qp, pd, init, uresp, ibpd, udata); pd 897 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 909 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_add_ref(pd); pd 911 drivers/infiniband/sw/rxe/rxe_verbs.c err = rxe_mem_init_dma(pd, access, mr); pd 918 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(pd); pd 933 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 944 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_add_ref(pd); pd 946 drivers/infiniband/sw/rxe/rxe_verbs.c err = rxe_mem_init_user(pd, start, length, iova, pd 954 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(pd); pd 966 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(mr->pd); pd 976 drivers/infiniband/sw/rxe/rxe_verbs.c struct rxe_pd *pd = to_rpd(ibpd); pd 991 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_add_ref(pd); pd 993 drivers/infiniband/sw/rxe/rxe_verbs.c err = rxe_mem_init_fast(pd, max_num_sg, mr); pd 1000 drivers/infiniband/sw/rxe/rxe_verbs.c rxe_drop_ref(pd); pd 76 drivers/infiniband/sw/rxe/rxe_verbs.h struct rxe_pd *pd; pd 125 drivers/infiniband/sw/rxe/rxe_verbs.h struct rxe_pd *pd; pd 243 drivers/infiniband/sw/rxe/rxe_verbs.h struct rxe_pd *pd; pd 325 drivers/infiniband/sw/rxe/rxe_verbs.h struct rxe_pd *pd; pd 433 drivers/infiniband/sw/rxe/rxe_verbs.h static inline struct rxe_pd *to_rpd(struct ib_pd *pd) pd 435 drivers/infiniband/sw/rxe/rxe_verbs.h return pd ? container_of(pd, struct rxe_pd, ibpd) : NULL; pd 173 drivers/infiniband/sw/siw/siw.h struct ib_pd *pd; pd 443 drivers/infiniband/sw/siw/siw.h struct ib_pd *pd; pd 728 drivers/infiniband/sw/siw/siw.h #define siw_dbg_pd(pd, fmt, ...) \ pd 729 drivers/infiniband/sw/siw/siw.h ibdev_dbg(pd->device, "PD[%u] %s: " fmt, pd->res.id, __func__, \ pd 89 drivers/infiniband/sw/siw/siw_mem.c int siw_mr_add_mem(struct siw_mr *mr, struct ib_pd *pd, void *mem_obj, pd 92 drivers/infiniband/sw/siw/siw_mem.c struct siw_device *sdev = to_siw_dev(pd->device); pd 105 drivers/infiniband/sw/siw/siw_mem.c mem->pd = pd; pd 168 drivers/infiniband/sw/siw/siw_mem.c int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr, pd 172 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag 0x%08x invalid\n", mem->stag); pd 175 drivers/infiniband/sw/siw/siw_mem.c if (mem->pd != pd) { pd 176 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag 0x%08x: PD mismatch\n", mem->stag); pd 183 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "permissions 0x%08x < 0x%08x\n", pd 191 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "MEM interval len %d\n", len); pd 192 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "[0x%pK, 0x%pK] out of bounds\n", pd 195 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "[0x%pK, 0x%pK] STag=0x%08x\n", pd 222 drivers/infiniband/sw/siw/siw_mem.c int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[], pd 225 drivers/infiniband/sw/siw/siw_mem.c struct siw_device *sdev = to_siw_dev(pd->device); pd 236 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey); pd 248 drivers/infiniband/sw/siw/siw_mem.c rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len); pd 292 drivers/infiniband/sw/siw/siw_mem.c int siw_invalidate_stag(struct ib_pd *pd, u32 stag) pd 294 drivers/infiniband/sw/siw/siw_mem.c struct siw_device *sdev = to_siw_dev(pd->device); pd 299 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag 0x%08x unknown\n", stag); pd 302 drivers/infiniband/sw/siw/siw_mem.c if (unlikely(mem->pd != pd)) { pd 303 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "PD mismatch for STag 0x%08x\n", stag); pd 313 drivers/infiniband/sw/siw/siw_mem.c siw_dbg_pd(pd, "STag 0x%08x now invalid\n", stag); pd 15 drivers/infiniband/sw/siw/siw_mem.h int siw_invalidate_stag(struct ib_pd *pd, u32 stag); pd 16 drivers/infiniband/sw/siw/siw_mem.h int siw_check_mem(struct ib_pd *pd, struct siw_mem *mem, u64 addr, pd 18 drivers/infiniband/sw/siw/siw_mem.h int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, pd 22 drivers/infiniband/sw/siw/siw_mem.h int siw_mr_add_mem(struct siw_mr *mr, struct ib_pd *pd, void *mem_obj, pd 454 drivers/infiniband/sw/siw/siw_qp_rx.c struct ib_pd *pd; pd 474 drivers/infiniband/sw/siw/siw_qp_rx.c pd = qp->srq == NULL ? qp->pd : qp->srq->base_srq.pd; pd 476 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE, pd 588 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_check_mem(qp->pd, mem, srx->ddp_to + srx->fpdu_part_rcvd, pd 825 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0, pd 1234 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_invalidate_stag(qp->pd, srx->inval_stag); pd 1272 drivers/infiniband/sw/siw/siw_qp_rx.c rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey); pd 751 drivers/infiniband/sw/siw/siw_qp_tx.c static int siw_check_sgl_tx(struct ib_pd *pd, struct siw_wqe *wqe, pd 765 drivers/infiniband/sw/siw/siw_qp_tx.c int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0, pd 806 drivers/infiniband/sw/siw/siw_qp_tx.c rv = siw_check_sgl_tx(qp->pd, wqe, 0); pd 919 drivers/infiniband/sw/siw/siw_qp_tx.c static int siw_fastreg_mr(struct ib_pd *pd, struct siw_sqe *sqe) pd 922 drivers/infiniband/sw/siw/siw_qp_tx.c struct siw_device *sdev = to_siw_dev(pd->device); pd 926 drivers/infiniband/sw/siw/siw_qp_tx.c siw_dbg_pd(pd, "STag 0x%08x\n", sqe->rkey); pd 944 drivers/infiniband/sw/siw/siw_qp_tx.c if (unlikely(mem->pd != pd)) { pd 972 drivers/infiniband/sw/siw/siw_qp_tx.c rv = siw_fastreg_mr(qp->pd, &wqe->sqe); pd 976 drivers/infiniband/sw/siw/siw_qp_tx.c rv = siw_invalidate_stag(qp->pd, wqe->sqe.rkey); pd 265 drivers/infiniband/sw/siw/siw_verbs.c int siw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 267 drivers/infiniband/sw/siw/siw_verbs.c struct siw_device *sdev = to_siw_dev(pd->device); pd 273 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "now %d PD's(s)\n", atomic_read(&sdev->num_pd)); pd 278 drivers/infiniband/sw/siw/siw_verbs.c void siw_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) pd 280 drivers/infiniband/sw/siw/siw_verbs.c struct siw_device *sdev = to_siw_dev(pd->device); pd 282 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "free PD\n"); pd 306 drivers/infiniband/sw/siw/siw_verbs.c struct ib_qp *siw_create_qp(struct ib_pd *pd, pd 312 drivers/infiniband/sw/siw/siw_verbs.c struct ib_device *base_dev = pd->device; pd 416 drivers/infiniband/sw/siw/siw_verbs.c qp->pd = pd; pd 1295 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_reg_user_mr(struct ib_pd *pd, u64 start, u64 len, pd 1301 drivers/infiniband/sw/siw/siw_verbs.c struct siw_device *sdev = to_siw_dev(pd->device); pd 1306 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "start: 0x%pK, va: 0x%pK, len: %llu\n", pd 1311 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "too many mr's\n"); pd 1325 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "pages req %lu, max %lu, lock %lu\n", pd 1335 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "getting user memory failed: %d\n", rv); pd 1344 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_mr_add_mem(mr, pd, umem, start, len, rights); pd 1390 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, pd 1393 drivers/infiniband/sw/siw/siw_verbs.c struct siw_device *sdev = to_siw_dev(pd->device); pd 1399 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "too many mr's\n"); pd 1404 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "mr type %d unsupported\n", mr_type); pd 1409 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "too many sge's: %d\n", max_sge); pd 1416 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "pbl allocation failed: %d\n", rv); pd 1425 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0); pd 1431 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag); pd 1445 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "failed: %d\n", rv); pd 1524 drivers/infiniband/sw/siw/siw_verbs.c struct ib_mr *siw_get_dma_mr(struct ib_pd *pd, int rights) pd 1526 drivers/infiniband/sw/siw/siw_verbs.c struct siw_device *sdev = to_siw_dev(pd->device); pd 1531 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "too many mr's\n"); pd 1540 drivers/infiniband/sw/siw/siw_verbs.c rv = siw_mr_add_mem(mr, pd, NULL, 0, ULONG_MAX, rights); pd 1546 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(pd, "[MEM %u]: success\n", mr->mem->stag); pd 1581 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, "too many SRQ's\n"); pd 1632 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, "[SRQ]: success\n"); pd 1752 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, pd 1769 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, "SRQ full\n"); pd 1774 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, pd 1794 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(base_srq->pd, "[SRQ]: error %d\n", rv); pd 1847 drivers/infiniband/sw/siw/siw_verbs.c siw_dbg_pd(srq->base_srq.pd, pd 364 drivers/infiniband/ulp/ipoib/ipoib.h struct ib_pd *pd; pd 487 drivers/infiniband/ulp/ipoib/ipoib.h struct ib_pd *pd, struct rdma_ah_attr *attr); pd 270 drivers/infiniband/ulp/ipoib/ipoib_cm.c return ib_create_qp(priv->pd, &attr); pd 336 drivers/infiniband/ulp/ipoib/ipoib_cm.c sge[i].lkey = priv->pd->local_dma_lkey; pd 1077 drivers/infiniband/ulp/ipoib/ipoib_cm.c tx_qp = ib_create_qp(priv->pd, &attr); pd 1567 drivers/infiniband/ulp/ipoib/ipoib_cm.c priv->cm.srq = ib_create_srq(priv->pd, &srq_init_attr); pd 56 drivers/infiniband/ulp/ipoib/ipoib_ib.c struct ib_pd *pd, struct rdma_ah_attr *attr) pd 69 drivers/infiniband/ulp/ipoib/ipoib_ib.c vah = rdma_create_ah(pd, attr, RDMA_CREATE_AH_SLEEPABLE); pd 1316 drivers/infiniband/ulp/ipoib/ipoib_ib.c if (priv->pd) { pd 1317 drivers/infiniband/ulp/ipoib/ipoib_ib.c ib_dealloc_pd(priv->pd); pd 1318 drivers/infiniband/ulp/ipoib/ipoib_ib.c priv->pd = NULL; pd 770 drivers/infiniband/ulp/ipoib/ipoib_main.c ah = ipoib_create_ah(dev, priv->pd, &av); pd 1765 drivers/infiniband/ulp/ipoib/ipoib_main.c priv->pd = ib_alloc_pd(priv->ca, 0); pd 1766 drivers/infiniband/ulp/ipoib/ipoib_main.c if (IS_ERR(priv->pd)) { pd 1800 drivers/infiniband/ulp/ipoib/ipoib_main.c if (priv->pd) { pd 1801 drivers/infiniband/ulp/ipoib/ipoib_main.c ib_dealloc_pd(priv->pd); pd 1802 drivers/infiniband/ulp/ipoib/ipoib_main.c priv->pd = NULL; pd 289 drivers/infiniband/ulp/ipoib/ipoib_multicast.c ah = ipoib_create_ah(dev, priv->pd, &av); pd 209 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->qp = ib_create_qp(priv->pd, &init_attr); pd 219 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->tx_sge[i].lkey = priv->pd->local_dma_lkey; pd 225 drivers/infiniband/ulp/ipoib/ipoib_verbs.c priv->rx_sge[0].lkey = priv->pd->local_dma_lkey; pd 214 drivers/infiniband/ulp/iser/iscsi_iser.c tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey; pd 361 drivers/infiniband/ulp/iser/iscsi_iser.h struct ib_pd *pd; pd 280 drivers/infiniband/ulp/iser/iser_initiator.c rx_sg->lkey = device->pd->local_dma_lkey; pd 534 drivers/infiniband/ulp/iser/iser_initiator.c tx_dsg->lkey = device->pd->local_dma_lkey; pd 196 drivers/infiniband/ulp/iser/iser_memory.c reg->sge.lkey = device->pd->local_dma_lkey; pd 202 drivers/infiniband/ulp/iser/iser_memory.c if (device->pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) pd 203 drivers/infiniband/ulp/iser/iser_memory.c reg->rkey = device->pd->unsafe_global_rkey; pd 91 drivers/infiniband/ulp/iser/iser_verbs.c device->pd = ib_alloc_pd(ib_dev, pd 93 drivers/infiniband/ulp/iser/iser_verbs.c if (IS_ERR(device->pd)) pd 119 drivers/infiniband/ulp/iser/iser_verbs.c ib_dealloc_pd(device->pd); pd 143 drivers/infiniband/ulp/iser/iser_verbs.c ib_dealloc_pd(device->pd); pd 147 drivers/infiniband/ulp/iser/iser_verbs.c device->pd = NULL; pd 195 drivers/infiniband/ulp/iser/iser_verbs.c fmr_pool = ib_create_fmr_pool(device->pd, ¶ms); pd 238 drivers/infiniband/ulp/iser/iser_verbs.c struct ib_pd *pd, pd 256 drivers/infiniband/ulp/iser/iser_verbs.c desc->rsc.mr = ib_alloc_mr(pd, mr_type, size); pd 264 drivers/infiniband/ulp/iser/iser_verbs.c desc->rsc.sig_mr = ib_alloc_mr_integrity(pd, size, size); pd 314 drivers/infiniband/ulp/iser/iser_verbs.c desc = iser_create_fastreg_desc(device, device->pd, pd 420 drivers/infiniband/ulp/iser/iser_verbs.c ret = rdma_create_qp(ib_conn->cma_id, device->pd, &init_attr); pd 981 drivers/infiniband/ulp/iser/iser_verbs.c desc->sge.lkey = ib_conn->device->pd->local_dma_lkey; pd 138 drivers/infiniband/ulp/isert/ib_isert.c ret = rdma_create_qp(cma_id, device->pd, &attr); pd 195 drivers/infiniband/ulp/isert/ib_isert.c rx_sg->lkey = device->pd->local_dma_lkey; pd 302 drivers/infiniband/ulp/isert/ib_isert.c device->pd = ib_alloc_pd(ib_dev, 0); pd 303 drivers/infiniband/ulp/isert/ib_isert.c if (IS_ERR(device->pd)) { pd 304 drivers/infiniband/ulp/isert/ib_isert.c ret = PTR_ERR(device->pd); pd 329 drivers/infiniband/ulp/isert/ib_isert.c ib_dealloc_pd(device->pd); pd 891 drivers/infiniband/ulp/isert/ib_isert.c if (tx_desc->tx_sg[0].lkey != device->pd->local_dma_lkey) { pd 892 drivers/infiniband/ulp/isert/ib_isert.c tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey; pd 929 drivers/infiniband/ulp/isert/ib_isert.c tx_desc->tx_sg[0].lkey = device->pd->local_dma_lkey; pd 969 drivers/infiniband/ulp/isert/ib_isert.c sge.lkey = isert_conn->device->pd->local_dma_lkey; pd 1018 drivers/infiniband/ulp/isert/ib_isert.c tx_dsg->lkey = isert_conn->device->pd->local_dma_lkey; pd 1880 drivers/infiniband/ulp/isert/ib_isert.c tx_dsg->lkey = device->pd->local_dma_lkey; pd 2008 drivers/infiniband/ulp/isert/ib_isert.c tx_dsg->lkey = device->pd->local_dma_lkey; pd 2051 drivers/infiniband/ulp/isert/ib_isert.c tx_dsg->lkey = device->pd->local_dma_lkey; pd 188 drivers/infiniband/ulp/isert/ib_isert.h struct ib_pd *pd; pd 636 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc, pd 780 drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c ah = rdma_create_ah(port->mad_agent->qp->pd, &ah_attr, 0); pd 406 drivers/infiniband/ulp/srp/ib_srp.c return ib_create_fmr_pool(dev->pd, &fmr_param); pd 436 drivers/infiniband/ulp/srp/ib_srp.c struct ib_pd *pd, int pool_size, pd 462 drivers/infiniband/ulp/srp/ib_srp.c mr = ib_alloc_mr(pd, mr_type, max_page_list_len); pd 529 drivers/infiniband/ulp/srp/ib_srp.c return srp_create_fr_pool(dev->dev, dev->pd, target->mr_pool_size, pd 593 drivers/infiniband/ulp/srp/ib_srp.c ret = rdma_create_qp(ch->rdma_cm.cm_id, dev->pd, init_attr); pd 596 drivers/infiniband/ulp/srp/ib_srp.c qp = ib_create_qp(dev->pd, init_attr); pd 3804 drivers/infiniband/ulp/srp/ib_srp.c target->lkey = host->srp_dev->pd->local_dma_lkey; pd 4176 drivers/infiniband/ulp/srp/ib_srp.c srp_dev->pd = ib_alloc_pd(device, flags); pd 4177 drivers/infiniband/ulp/srp/ib_srp.c if (IS_ERR(srp_dev->pd)) pd 4181 drivers/infiniband/ulp/srp/ib_srp.c srp_dev->global_rkey = srp_dev->pd->unsafe_global_rkey; pd 4233 drivers/infiniband/ulp/srp/ib_srp.c ib_dealloc_pd(srp_dev->pd); pd 104 drivers/infiniband/ulp/srp/ib_srp.h struct ib_pd *pd; pd 483 drivers/infiniband/ulp/srpt/ib_srpt.c ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc, pd 1833 drivers/infiniband/ulp/srpt/ib_srpt.c ret = rdma_create_qp(ch->rdma_cm.cm_id, sdev->pd, qp_init); pd 1836 drivers/infiniband/ulp/srpt/ib_srpt.c ch->qp = ib_create_qp(sdev->pd, qp_init); pd 3050 drivers/infiniband/ulp/srpt/ib_srpt.c srq = ib_create_srq(sdev->pd, &srq_attr); pd 3125 drivers/infiniband/ulp/srpt/ib_srpt.c sdev->pd = ib_alloc_pd(device, 0); pd 3126 drivers/infiniband/ulp/srpt/ib_srpt.c if (IS_ERR(sdev->pd)) pd 3129 drivers/infiniband/ulp/srpt/ib_srpt.c sdev->lkey = sdev->pd->local_dma_lkey; pd 3207 drivers/infiniband/ulp/srpt/ib_srpt.c ib_dealloc_pd(sdev->pd); pd 3259 drivers/infiniband/ulp/srpt/ib_srpt.c ib_dealloc_pd(sdev->pd); pd 424 drivers/infiniband/ulp/srpt/ib_srpt.h struct ib_pd *pd; pd 90 drivers/input/joystick/db9.c struct pardevice *pd; pd 358 drivers/input/joystick/db9.c struct parport *port = db9->pd->port; pd 510 drivers/input/joystick/db9.c struct parport *port = db9->pd->port; pd 518 drivers/input/joystick/db9.c parport_claim(db9->pd); pd 534 drivers/input/joystick/db9.c struct parport *port = db9->pd->port; pd 541 drivers/input/joystick/db9.c parport_release(db9->pd); pd 550 drivers/input/joystick/db9.c struct pardevice *pd; pd 587 drivers/input/joystick/db9.c pd = parport_register_dev_model(pp, "db9", &db9_parport_cb, port_idx); pd 588 drivers/input/joystick/db9.c if (!pd) { pd 598 drivers/input/joystick/db9.c db9->pd = pd; pd 612 drivers/input/joystick/db9.c "%s/input%d", db9->pd->port->name, i); pd 650 drivers/input/joystick/db9.c parport_unregister_device(pd); pd 671 drivers/input/joystick/db9.c parport_unregister_device(db9->pd); pd 74 drivers/input/joystick/gamecon.c struct pardevice *pd; pd 136 drivers/input/joystick/gamecon.c struct parport *port = gc->pd->port; pd 149 drivers/input/joystick/gamecon.c struct parport *port = gc->pd->port; pd 191 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_N64_POWER_R); pd 193 drivers/input/joystick/gamecon.c data[i] = parport_read_status(gc->pd->port); pd 194 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_N64_POWER_R | GC_N64_CLOCK); pd 345 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK | GC_NES_LATCH); pd 347 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK); pd 351 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_NES_POWER); pd 352 drivers/input/joystick/gamecon.c data[i] = parport_read_status(gc->pd->port) ^ 0x7f; pd 354 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_NES_POWER | GC_NES_CLOCK); pd 465 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, ~(1 << i)); pd 466 drivers/input/joystick/gamecon.c data[i] = parport_read_status(gc->pd->port) ^ 0x7f; pd 552 drivers/input/joystick/gamecon.c struct parport *port = gc->pd->port; pd 571 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, cmd | GC_PSX_CLOCK | GC_PSX_POWER); pd 590 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_SELECT | GC_PSX_POWER); pd 593 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_POWER); pd 622 drivers/input/joystick/gamecon.c parport_write_data(gc->pd->port, GC_PSX_CLOCK | GC_PSX_SELECT | GC_PSX_POWER); pd 778 drivers/input/joystick/gamecon.c parport_claim(gc->pd); pd 779 drivers/input/joystick/gamecon.c parport_write_control(gc->pd->port, 0x04); pd 794 drivers/input/joystick/gamecon.c parport_write_control(gc->pd->port, 0x00); pd 795 drivers/input/joystick/gamecon.c parport_release(gc->pd); pd 821 drivers/input/joystick/gamecon.c "%s/input%d", gc->pd->port->name, idx); pd 924 drivers/input/joystick/gamecon.c struct pardevice *pd; pd 948 drivers/input/joystick/gamecon.c pd = parport_register_dev_model(pp, "gamecon", &gc_parport_cb, pd 950 drivers/input/joystick/gamecon.c if (!pd) { pd 962 drivers/input/joystick/gamecon.c gc->pd = pd; pd 991 drivers/input/joystick/gamecon.c parport_unregister_device(pd); pd 1013 drivers/input/joystick/gamecon.c parport_unregister_device(gc->pd); pd 61 drivers/input/joystick/turbografx.c struct pardevice *pd; pd 87 drivers/input/joystick/turbografx.c parport_write_data(tgfx->pd->port, ~(1 << i)); pd 88 drivers/input/joystick/turbografx.c data1 = parport_read_status(tgfx->pd->port) ^ 0x7f; pd 89 drivers/input/joystick/turbografx.c data2 = parport_read_control(tgfx->pd->port) ^ 0x04; /* CAVEAT parport */ pd 116 drivers/input/joystick/turbografx.c parport_claim(tgfx->pd); pd 117 drivers/input/joystick/turbografx.c parport_write_control(tgfx->pd->port, 0x04); pd 132 drivers/input/joystick/turbografx.c parport_write_control(tgfx->pd->port, 0x00); pd 133 drivers/input/joystick/turbografx.c parport_release(tgfx->pd); pd 148 drivers/input/joystick/turbografx.c struct pardevice *pd; pd 171 drivers/input/joystick/turbografx.c pd = parport_register_dev_model(pp, "turbografx", &tgfx_parport_cb, pd 173 drivers/input/joystick/turbografx.c if (!pd) { pd 185 drivers/input/joystick/turbografx.c tgfx->pd = pd; pd 208 drivers/input/joystick/turbografx.c "%s/input%d", tgfx->pd->port->name, i); pd 250 drivers/input/joystick/turbografx.c parport_unregister_device(pd); pd 272 drivers/input/joystick/turbografx.c parport_unregister_device(tgfx->pd); pd 86 drivers/input/keyboard/snvs_pwrkey.c struct pwrkey_drv_data *pd = pdata; pd 88 drivers/input/keyboard/snvs_pwrkey.c del_timer_sync(&pd->check_timer); pd 69 drivers/input/misc/soc_button_array.c struct platform_device *pd; pd 130 drivers/input/misc/soc_button_array.c pd = platform_device_register_resndata(&pdev->dev, "gpio-keys", pd 134 drivers/input/misc/soc_button_array.c error = PTR_ERR_OR_ZERO(pd); pd 141 drivers/input/misc/soc_button_array.c return pd; pd 326 drivers/input/misc/soc_button_array.c struct platform_device *pd; pd 358 drivers/input/misc/soc_button_array.c pd = soc_button_device_create(pdev, button_info, i == 0); pd 359 drivers/input/misc/soc_button_array.c if (IS_ERR(pd)) { pd 360 drivers/input/misc/soc_button_array.c error = PTR_ERR(pd); pd 368 drivers/input/misc/soc_button_array.c priv->children[i] = pd; pd 54 drivers/iommu/tegra-smmu.c struct page *pd; pd 290 drivers/iommu/tegra-smmu.c as->pd = alloc_page(GFP_KERNEL | __GFP_DMA | __GFP_ZERO); pd 291 drivers/iommu/tegra-smmu.c if (!as->pd) { pd 298 drivers/iommu/tegra-smmu.c __free_page(as->pd); pd 306 drivers/iommu/tegra-smmu.c __free_page(as->pd); pd 414 drivers/iommu/tegra-smmu.c as->pd_dma = dma_map_page(smmu->dev, as->pd, 0, SMMU_SIZE_PD, pd 525 drivers/iommu/tegra-smmu.c u32 *pd = page_address(as->pd); pd 526 drivers/iommu/tegra-smmu.c unsigned long offset = pd_index * sizeof(*pd); pd 529 drivers/iommu/tegra-smmu.c pd[pd_index] = value; pd 533 drivers/iommu/tegra-smmu.c sizeof(*pd), DMA_TO_DEVICE); pd 554 drivers/iommu/tegra-smmu.c u32 *pd; pd 560 drivers/iommu/tegra-smmu.c pd = page_address(as->pd); pd 561 drivers/iommu/tegra-smmu.c *dmap = smmu_pde_to_dma(smmu, pd[pd_index]); pd 601 drivers/iommu/tegra-smmu.c u32 *pd = page_address(as->pd); pd 603 drivers/iommu/tegra-smmu.c *dmap = smmu_pde_to_dma(smmu, pd[pde]); pd 627 drivers/iommu/tegra-smmu.c u32 *pd = page_address(as->pd); pd 628 drivers/iommu/tegra-smmu.c dma_addr_t pte_dma = smmu_pde_to_dma(smmu, pd[pde]); pd 38 drivers/leds/leds-s3c24xx.c struct s3c24xx_led_platdata *pd = led->pdata; pd 39 drivers/leds/leds-s3c24xx.c int state = (value ? 1 : 0) ^ (pd->flags & S3C24XX_LEDF_ACTLOW); pd 44 drivers/leds/leds-s3c24xx.c gpio_set_value(pd->gpio, state); pd 46 drivers/leds/leds-s3c24xx.c if (pd->flags & S3C24XX_LEDF_TRISTATE) { pd 48 drivers/leds/leds-s3c24xx.c gpio_direction_output(pd->gpio, state); pd 50 drivers/leds/leds-s3c24xx.c gpio_direction_input(pd->gpio); pd 415 drivers/media/i2c/adp1653.c struct adp1653_platform_data *pd; pd 418 drivers/media/i2c/adp1653.c pd = devm_kzalloc(&client->dev, sizeof(*pd), GFP_KERNEL); pd 419 drivers/media/i2c/adp1653.c if (!pd) pd 421 drivers/media/i2c/adp1653.c flash->platform_data = pd; pd 428 drivers/media/i2c/adp1653.c &pd->max_flash_timeout)) pd 432 drivers/media/i2c/adp1653.c &pd->max_flash_intensity)) pd 435 drivers/media/i2c/adp1653.c pd->max_flash_intensity /= 1000; pd 438 drivers/media/i2c/adp1653.c &pd->max_torch_intensity)) pd 441 drivers/media/i2c/adp1653.c pd->max_torch_intensity /= 1000; pd 449 drivers/media/i2c/adp1653.c &pd->max_indicator_intensity)) pd 454 drivers/media/i2c/adp1653.c pd->enable_gpio = devm_gpiod_get(&client->dev, "enable", GPIOD_OUT_LOW); pd 455 drivers/media/i2c/adp1653.c if (IS_ERR(pd->enable_gpio)) { pd 457 drivers/media/i2c/adp1653.c return PTR_ERR(pd->enable_gpio); pd 128 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = vb2_get_drv_priv(vq); pd 129 drivers/media/pci/dt3155/dt3155.c unsigned size = pd->width * pd->height; pd 142 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = vb2_get_drv_priv(vb->vb2_queue); pd 144 drivers/media/pci/dt3155/dt3155.c vb2_set_plane_payload(vb, 0, pd->width * pd->height); pd 150 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = vb2_get_drv_priv(q); pd 151 drivers/media/pci/dt3155/dt3155.c struct vb2_buffer *vb = &pd->curr_buf->vb2_buf; pd 154 drivers/media/pci/dt3155/dt3155.c pd->sequence = 0; pd 156 drivers/media/pci/dt3155/dt3155.c iowrite32(dma_addr, pd->regs + EVEN_DMA_START); pd 157 drivers/media/pci/dt3155/dt3155.c iowrite32(dma_addr + pd->width, pd->regs + ODD_DMA_START); pd 158 drivers/media/pci/dt3155/dt3155.c iowrite32(pd->width, pd->regs + EVEN_DMA_STRIDE); pd 159 drivers/media/pci/dt3155/dt3155.c iowrite32(pd->width, pd->regs + ODD_DMA_STRIDE); pd 162 drivers/media/pci/dt3155/dt3155.c FLD_END_EVEN | FLD_END_ODD, pd->regs + INT_CSR); pd 165 drivers/media/pci/dt3155/dt3155.c pd->regs + CSR1); pd 166 drivers/media/pci/dt3155/dt3155.c wait_i2c_reg(pd->regs); pd 167 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, CONFIG, pd->config); pd 168 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, EVEN_CSR, CSR_ERROR | CSR_DONE); pd 169 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, ODD_CSR, CSR_ERROR | CSR_DONE); pd 172 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, CSR2, pd->csr2 | BUSY_EVEN | BUSY_ODD); pd 178 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = vb2_get_drv_priv(q); pd 181 drivers/media/pci/dt3155/dt3155.c spin_lock_irq(&pd->lock); pd 183 drivers/media/pci/dt3155/dt3155.c write_i2c_reg_nowait(pd->regs, CSR2, pd->csr2); pd 185 drivers/media/pci/dt3155/dt3155.c FLD_DN_ODD | FLD_DN_EVEN, pd->regs + CSR1); pd 187 drivers/media/pci/dt3155/dt3155.c iowrite32(FLD_START | FLD_END_EVEN | FLD_END_ODD, pd->regs + INT_CSR); pd 188 drivers/media/pci/dt3155/dt3155.c spin_unlock_irq(&pd->lock); pd 197 drivers/media/pci/dt3155/dt3155.c spin_lock_irq(&pd->lock); pd 198 drivers/media/pci/dt3155/dt3155.c if (pd->curr_buf) { pd 199 drivers/media/pci/dt3155/dt3155.c vb2_buffer_done(&pd->curr_buf->vb2_buf, VB2_BUF_STATE_ERROR); pd 200 drivers/media/pci/dt3155/dt3155.c pd->curr_buf = NULL; pd 203 drivers/media/pci/dt3155/dt3155.c while (!list_empty(&pd->dmaq)) { pd 204 drivers/media/pci/dt3155/dt3155.c vb = list_first_entry(&pd->dmaq, typeof(*vb), done_entry); pd 208 drivers/media/pci/dt3155/dt3155.c spin_unlock_irq(&pd->lock); pd 214 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = vb2_get_drv_priv(vb->vb2_queue); pd 217 drivers/media/pci/dt3155/dt3155.c spin_lock_irq(&pd->lock); pd 218 drivers/media/pci/dt3155/dt3155.c if (pd->curr_buf) pd 219 drivers/media/pci/dt3155/dt3155.c list_add_tail(&vb->done_entry, &pd->dmaq); pd 221 drivers/media/pci/dt3155/dt3155.c pd->curr_buf = vbuf; pd 222 drivers/media/pci/dt3155/dt3155.c spin_unlock_irq(&pd->lock); pd 295 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 299 drivers/media/pci/dt3155/dt3155.c sprintf(cap->bus_info, "PCI:%s", pci_name(pd->pdev)); pd 314 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 316 drivers/media/pci/dt3155/dt3155.c f->fmt.pix.width = pd->width; pd 317 drivers/media/pci/dt3155/dt3155.c f->fmt.pix.height = pd->height; pd 328 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 330 drivers/media/pci/dt3155/dt3155.c *norm = pd->std; pd 336 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 338 drivers/media/pci/dt3155/dt3155.c if (pd->std == norm) pd 340 drivers/media/pci/dt3155/dt3155.c if (vb2_is_busy(&pd->vidq)) pd 342 drivers/media/pci/dt3155/dt3155.c pd->std = norm; pd 343 drivers/media/pci/dt3155/dt3155.c if (pd->std & V4L2_STD_525_60) { pd 344 drivers/media/pci/dt3155/dt3155.c pd->csr2 = VT_60HZ; pd 345 drivers/media/pci/dt3155/dt3155.c pd->width = 640; pd 346 drivers/media/pci/dt3155/dt3155.c pd->height = 480; pd 348 drivers/media/pci/dt3155/dt3155.c pd->csr2 = VT_50HZ; pd 349 drivers/media/pci/dt3155/dt3155.c pd->width = 768; pd 350 drivers/media/pci/dt3155/dt3155.c pd->height = 576; pd 373 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 375 drivers/media/pci/dt3155/dt3155.c *i = pd->input; pd 381 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = video_drvdata(filp); pd 385 drivers/media/pci/dt3155/dt3155.c pd->input = i; pd 386 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG); pd 387 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_CMD, (i << 6) | (i << 4) | SYNC_LVL_3); pd 412 drivers/media/pci/dt3155/dt3155.c static int dt3155_init_board(struct dt3155_priv *pd) pd 414 drivers/media/pci/dt3155/dt3155.c struct pci_dev *pdev = pd->pdev; pd 422 drivers/media/pci/dt3155/dt3155.c FLD_DN_ODD | FLD_DN_EVEN, pd->regs + CSR1); pd 426 drivers/media/pci/dt3155/dt3155.c iowrite32(FIFO_EN | SRST, pd->regs + CSR1); pd 427 drivers/media/pci/dt3155/dt3155.c iowrite32(0xEEEEEE01, pd->regs + EVEN_PIXEL_FMT); pd 428 drivers/media/pci/dt3155/dt3155.c iowrite32(0xEEEEEE01, pd->regs + ODD_PIXEL_FMT); pd 429 drivers/media/pci/dt3155/dt3155.c iowrite32(0x00000020, pd->regs + FIFO_TRIGER); pd 430 drivers/media/pci/dt3155/dt3155.c iowrite32(0x00000103, pd->regs + XFER_MODE); pd 431 drivers/media/pci/dt3155/dt3155.c iowrite32(0, pd->regs + RETRY_WAIT_CNT); pd 432 drivers/media/pci/dt3155/dt3155.c iowrite32(0, pd->regs + INT_CSR); pd 433 drivers/media/pci/dt3155/dt3155.c iowrite32(1, pd->regs + EVEN_FLD_MASK); pd 434 drivers/media/pci/dt3155/dt3155.c iowrite32(1, pd->regs + ODD_FLD_MASK); pd 435 drivers/media/pci/dt3155/dt3155.c iowrite32(0, pd->regs + MASK_LENGTH); pd 436 drivers/media/pci/dt3155/dt3155.c iowrite32(0x0005007C, pd->regs + FIFO_FLAG_CNT); pd 437 drivers/media/pci/dt3155/dt3155.c iowrite32(0x01010101, pd->regs + IIC_CLK_DUR); pd 440 drivers/media/pci/dt3155/dt3155.c read_i2c_reg(pd->regs, DT_ID, &tmp); pd 445 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, 0); pd 447 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_LUT, i); pd 451 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG); pd 452 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_CMD, VIDEO_CNL_1 | SYNC_CNL_1 | SYNC_LVL_3); pd 453 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, AD_POS_REF); pd 454 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_CMD, 34); pd 455 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, AD_NEG_REF); pd 456 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_CMD, 0); pd 459 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, CONFIG, pd->config | PM_LUT_PGM); pd 461 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, PM_LUT_ADDR, i); pd 462 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, PM_LUT_DATA, i); pd 464 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, CONFIG, pd->config | PM_LUT_PGM | PM_LUT_SEL); pd 466 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, PM_LUT_ADDR, i); pd 467 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, PM_LUT_DATA, i); pd 469 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, CONFIG, pd->config); /* ACQ_MODE_EVEN */ pd 472 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_ADDR, AD_CMD_REG); pd 473 drivers/media/pci/dt3155/dt3155.c write_i2c_reg(pd->regs, AD_CMD, VIDEO_CNL_1 | SYNC_CNL_1 | SYNC_LVL_3); pd 477 drivers/media/pci/dt3155/dt3155.c pd->regs + INT_CSR); pd 496 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd; pd 501 drivers/media/pci/dt3155/dt3155.c pd = devm_kzalloc(&pdev->dev, sizeof(*pd), GFP_KERNEL); pd 502 drivers/media/pci/dt3155/dt3155.c if (!pd) pd 505 drivers/media/pci/dt3155/dt3155.c err = v4l2_device_register(&pdev->dev, &pd->v4l2_dev); pd 508 drivers/media/pci/dt3155/dt3155.c pd->vdev = dt3155_vdev; pd 509 drivers/media/pci/dt3155/dt3155.c pd->vdev.v4l2_dev = &pd->v4l2_dev; pd 510 drivers/media/pci/dt3155/dt3155.c video_set_drvdata(&pd->vdev, pd); /* for use in video_fops */ pd 511 drivers/media/pci/dt3155/dt3155.c pd->pdev = pdev; pd 512 drivers/media/pci/dt3155/dt3155.c pd->std = V4L2_STD_625_50; pd 513 drivers/media/pci/dt3155/dt3155.c pd->csr2 = VT_50HZ; pd 514 drivers/media/pci/dt3155/dt3155.c pd->width = 768; pd 515 drivers/media/pci/dt3155/dt3155.c pd->height = 576; pd 516 drivers/media/pci/dt3155/dt3155.c INIT_LIST_HEAD(&pd->dmaq); pd 517 drivers/media/pci/dt3155/dt3155.c mutex_init(&pd->mux); pd 518 drivers/media/pci/dt3155/dt3155.c pd->vdev.lock = &pd->mux; /* for locking v4l2_file_operations */ pd 519 drivers/media/pci/dt3155/dt3155.c pd->vidq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; pd 520 drivers/media/pci/dt3155/dt3155.c pd->vidq.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; pd 521 drivers/media/pci/dt3155/dt3155.c pd->vidq.io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ; pd 522 drivers/media/pci/dt3155/dt3155.c pd->vidq.ops = &q_ops; pd 523 drivers/media/pci/dt3155/dt3155.c pd->vidq.mem_ops = &vb2_dma_contig_memops; pd 524 drivers/media/pci/dt3155/dt3155.c pd->vidq.drv_priv = pd; pd 525 drivers/media/pci/dt3155/dt3155.c pd->vidq.min_buffers_needed = 2; pd 526 drivers/media/pci/dt3155/dt3155.c pd->vidq.gfp_flags = GFP_DMA32; pd 527 drivers/media/pci/dt3155/dt3155.c pd->vidq.lock = &pd->mux; /* for locking v4l2_file_operations */ pd 528 drivers/media/pci/dt3155/dt3155.c pd->vidq.dev = &pdev->dev; pd 529 drivers/media/pci/dt3155/dt3155.c pd->vdev.queue = &pd->vidq; pd 530 drivers/media/pci/dt3155/dt3155.c err = vb2_queue_init(&pd->vidq); pd 533 drivers/media/pci/dt3155/dt3155.c spin_lock_init(&pd->lock); pd 534 drivers/media/pci/dt3155/dt3155.c pd->config = ACQ_MODE_EVEN; pd 541 drivers/media/pci/dt3155/dt3155.c pd->regs = pci_iomap(pdev, 0, pci_resource_len(pd->pdev, 0)); pd 542 drivers/media/pci/dt3155/dt3155.c if (!pd->regs) { pd 546 drivers/media/pci/dt3155/dt3155.c err = dt3155_init_board(pd); pd 549 drivers/media/pci/dt3155/dt3155.c err = request_irq(pd->pdev->irq, dt3155_irq_handler_even, pd 550 drivers/media/pci/dt3155/dt3155.c IRQF_SHARED, DT3155_NAME, pd); pd 553 drivers/media/pci/dt3155/dt3155.c err = video_register_device(&pd->vdev, VFL_TYPE_GRABBER, -1); pd 556 drivers/media/pci/dt3155/dt3155.c dev_info(&pdev->dev, "/dev/video%i is ready\n", pd->vdev.minor); pd 560 drivers/media/pci/dt3155/dt3155.c free_irq(pd->pdev->irq, pd); pd 562 drivers/media/pci/dt3155/dt3155.c pci_iounmap(pdev, pd->regs); pd 568 drivers/media/pci/dt3155/dt3155.c v4l2_device_unregister(&pd->v4l2_dev); pd 575 drivers/media/pci/dt3155/dt3155.c struct dt3155_priv *pd = container_of(v4l2_dev, struct dt3155_priv, pd 578 drivers/media/pci/dt3155/dt3155.c video_unregister_device(&pd->vdev); pd 579 drivers/media/pci/dt3155/dt3155.c free_irq(pd->pdev->irq, pd); pd 580 drivers/media/pci/dt3155/dt3155.c vb2_queue_release(&pd->vidq); pd 581 drivers/media/pci/dt3155/dt3155.c v4l2_device_unregister(&pd->v4l2_dev); pd 582 drivers/media/pci/dt3155/dt3155.c pci_iounmap(pdev, pd->regs); pd 387 drivers/media/platform/exynos4-is/media-dev.c struct fimc_source_info *pd = &fmd->sensor[index].pdata; pd 408 drivers/media/platform/exynos4-is/media-dev.c pd->mux_id = (endpoint.base.port - 1) & 0x1; pd 420 drivers/media/platform/exynos4-is/media-dev.c pd->sensor_bus_type = FIMC_BUS_TYPE_ITU_601; pd 422 drivers/media/platform/exynos4-is/media-dev.c pd->sensor_bus_type = FIMC_BUS_TYPE_ITU_656; pd 423 drivers/media/platform/exynos4-is/media-dev.c pd->flags = endpoint.bus.parallel.flags; pd 429 drivers/media/platform/exynos4-is/media-dev.c pd->sensor_bus_type = FIMC_BUS_TYPE_MIPI_CSI2; pd 445 drivers/media/platform/exynos4-is/media-dev.c pd->fimc_bus_type = FIMC_BUS_TYPE_ISP_WRITEBACK; pd 447 drivers/media/platform/exynos4-is/media-dev.c pd->fimc_bus_type = pd->sensor_bus_type; pd 1612 drivers/media/radio/si4713/si4713.c sdev->pd = si4713_pdev; pd 1631 drivers/media/radio/si4713/si4713.c platform_device_unregister(sdev->pd); pd 242 drivers/media/radio/si4713/si4713.h struct platform_device *pd; pd 555 drivers/media/tuners/tda18271-common.c u8 d, pd; pd 558 drivers/media/tuners/tda18271-common.c int ret = tda18271_lookup_pll_map(fe, MAIN_PLL, &freq, &pd, &d); pd 562 drivers/media/tuners/tda18271-common.c regs[R_MPD] = (0x7f & pd); pd 578 drivers/media/tuners/tda18271-common.c u8 d, pd; pd 581 drivers/media/tuners/tda18271-common.c int ret = tda18271_lookup_pll_map(fe, CAL_PLL, &freq, &pd, &d); pd 585 drivers/media/tuners/tda18271-common.c regs[R_CPD] = pd; pd 13 drivers/media/tuners/tda18271-maps.c u8 pd; /* post div */ pd 25 drivers/media/tuners/tda18271-maps.c { .lomax = 32000, .pd = 0x5f, .d = 0xf0 }, pd 26 drivers/media/tuners/tda18271-maps.c { .lomax = 35000, .pd = 0x5e, .d = 0xe0 }, pd 27 drivers/media/tuners/tda18271-maps.c { .lomax = 37000, .pd = 0x5d, .d = 0xd0 }, pd 28 drivers/media/tuners/tda18271-maps.c { .lomax = 41000, .pd = 0x5c, .d = 0xc0 }, pd 29 drivers/media/tuners/tda18271-maps.c { .lomax = 44000, .pd = 0x5b, .d = 0xb0 }, pd 30 drivers/media/tuners/tda18271-maps.c { .lomax = 49000, .pd = 0x5a, .d = 0xa0 }, pd 31 drivers/media/tuners/tda18271-maps.c { .lomax = 54000, .pd = 0x59, .d = 0x90 }, pd 32 drivers/media/tuners/tda18271-maps.c { .lomax = 61000, .pd = 0x58, .d = 0x80 }, pd 33 drivers/media/tuners/tda18271-maps.c { .lomax = 65000, .pd = 0x4f, .d = 0x78 }, pd 34 drivers/media/tuners/tda18271-maps.c { .lomax = 70000, .pd = 0x4e, .d = 0x70 }, pd 35 drivers/media/tuners/tda18271-maps.c { .lomax = 75000, .pd = 0x4d, .d = 0x68 }, pd 36 drivers/media/tuners/tda18271-maps.c { .lomax = 82000, .pd = 0x4c, .d = 0x60 }, pd 37 drivers/media/tuners/tda18271-maps.c { .lomax = 89000, .pd = 0x4b, .d = 0x58 }, pd 38 drivers/media/tuners/tda18271-maps.c { .lomax = 98000, .pd = 0x4a, .d = 0x50 }, pd 39 drivers/media/tuners/tda18271-maps.c { .lomax = 109000, .pd = 0x49, .d = 0x48 }, pd 40 drivers/media/tuners/tda18271-maps.c { .lomax = 123000, .pd = 0x48, .d = 0x40 }, pd 41 drivers/media/tuners/tda18271-maps.c { .lomax = 131000, .pd = 0x3f, .d = 0x3c }, pd 42 drivers/media/tuners/tda18271-maps.c { .lomax = 141000, .pd = 0x3e, .d = 0x38 }, pd 43 drivers/media/tuners/tda18271-maps.c { .lomax = 151000, .pd = 0x3d, .d = 0x34 }, pd 44 drivers/media/tuners/tda18271-maps.c { .lomax = 164000, .pd = 0x3c, .d = 0x30 }, pd 45 drivers/media/tuners/tda18271-maps.c { .lomax = 179000, .pd = 0x3b, .d = 0x2c }, pd 46 drivers/media/tuners/tda18271-maps.c { .lomax = 197000, .pd = 0x3a, .d = 0x28 }, pd 47 drivers/media/tuners/tda18271-maps.c { .lomax = 219000, .pd = 0x39, .d = 0x24 }, pd 48 drivers/media/tuners/tda18271-maps.c { .lomax = 246000, .pd = 0x38, .d = 0x20 }, pd 49 drivers/media/tuners/tda18271-maps.c { .lomax = 263000, .pd = 0x2f, .d = 0x1e }, pd 50 drivers/media/tuners/tda18271-maps.c { .lomax = 282000, .pd = 0x2e, .d = 0x1c }, pd 51 drivers/media/tuners/tda18271-maps.c { .lomax = 303000, .pd = 0x2d, .d = 0x1a }, pd 52 drivers/media/tuners/tda18271-maps.c { .lomax = 329000, .pd = 0x2c, .d = 0x18 }, pd 53 drivers/media/tuners/tda18271-maps.c { .lomax = 359000, .pd = 0x2b, .d = 0x16 }, pd 54 drivers/media/tuners/tda18271-maps.c { .lomax = 395000, .pd = 0x2a, .d = 0x14 }, pd 55 drivers/media/tuners/tda18271-maps.c { .lomax = 438000, .pd = 0x29, .d = 0x12 }, pd 56 drivers/media/tuners/tda18271-maps.c { .lomax = 493000, .pd = 0x28, .d = 0x10 }, pd 57 drivers/media/tuners/tda18271-maps.c { .lomax = 526000, .pd = 0x1f, .d = 0x0f }, pd 58 drivers/media/tuners/tda18271-maps.c { .lomax = 564000, .pd = 0x1e, .d = 0x0e }, pd 59 drivers/media/tuners/tda18271-maps.c { .lomax = 607000, .pd = 0x1d, .d = 0x0d }, pd 60 drivers/media/tuners/tda18271-maps.c { .lomax = 658000, .pd = 0x1c, .d = 0x0c }, pd 61 drivers/media/tuners/tda18271-maps.c { .lomax = 718000, .pd = 0x1b, .d = 0x0b }, pd 62 drivers/media/tuners/tda18271-maps.c { .lomax = 790000, .pd = 0x1a, .d = 0x0a }, pd 63 drivers/media/tuners/tda18271-maps.c { .lomax = 877000, .pd = 0x19, .d = 0x09 }, pd 64 drivers/media/tuners/tda18271-maps.c { .lomax = 987000, .pd = 0x18, .d = 0x08 }, pd 65 drivers/media/tuners/tda18271-maps.c { .lomax = 0, .pd = 0x00, .d = 0x00 }, /* end */ pd 69 drivers/media/tuners/tda18271-maps.c { .lomax = 33125, .pd = 0x57, .d = 0xf0 }, pd 70 drivers/media/tuners/tda18271-maps.c { .lomax = 35500, .pd = 0x56, .d = 0xe0 }, pd 71 drivers/media/tuners/tda18271-maps.c { .lomax = 38188, .pd = 0x55, .d = 0xd0 }, pd 72 drivers/media/tuners/tda18271-maps.c { .lomax = 41375, .pd = 0x54, .d = 0xc0 }, pd 73 drivers/media/tuners/tda18271-maps.c { .lomax = 45125, .pd = 0x53, .d = 0xb0 }, pd 74 drivers/media/tuners/tda18271-maps.c { .lomax = 49688, .pd = 0x52, .d = 0xa0 }, pd 75 drivers/media/tuners/tda18271-maps.c { .lomax = 55188, .pd = 0x51, .d = 0x90 }, pd 76 drivers/media/tuners/tda18271-maps.c { .lomax = 62125, .pd = 0x50, .d = 0x80 }, pd 77 drivers/media/tuners/tda18271-maps.c { .lomax = 66250, .pd = 0x47, .d = 0x78 }, pd 78 drivers/media/tuners/tda18271-maps.c { .lomax = 71000, .pd = 0x46, .d = 0x70 }, pd 79 drivers/media/tuners/tda18271-maps.c { .lomax = 76375, .pd = 0x45, .d = 0x68 }, pd 80 drivers/media/tuners/tda18271-maps.c { .lomax = 82750, .pd = 0x44, .d = 0x60 }, pd 81 drivers/media/tuners/tda18271-maps.c { .lomax = 90250, .pd = 0x43, .d = 0x58 }, pd 82 drivers/media/tuners/tda18271-maps.c { .lomax = 99375, .pd = 0x42, .d = 0x50 }, pd 83 drivers/media/tuners/tda18271-maps.c { .lomax = 110375, .pd = 0x41, .d = 0x48 }, pd 84 drivers/media/tuners/tda18271-maps.c { .lomax = 124250, .pd = 0x40, .d = 0x40 }, pd 85 drivers/media/tuners/tda18271-maps.c { .lomax = 132500, .pd = 0x37, .d = 0x3c }, pd 86 drivers/media/tuners/tda18271-maps.c { .lomax = 142000, .pd = 0x36, .d = 0x38 }, pd 87 drivers/media/tuners/tda18271-maps.c { .lomax = 152750, .pd = 0x35, .d = 0x34 }, pd 88 drivers/media/tuners/tda18271-maps.c { .lomax = 165500, .pd = 0x34, .d = 0x30 }, pd 89 drivers/media/tuners/tda18271-maps.c { .lomax = 180500, .pd = 0x33, .d = 0x2c }, pd 90 drivers/media/tuners/tda18271-maps.c { .lomax = 198750, .pd = 0x32, .d = 0x28 }, pd 91 drivers/media/tuners/tda18271-maps.c { .lomax = 220750, .pd = 0x31, .d = 0x24 }, pd 92 drivers/media/tuners/tda18271-maps.c { .lomax = 248500, .pd = 0x30, .d = 0x20 }, pd 93 drivers/media/tuners/tda18271-maps.c { .lomax = 265000, .pd = 0x27, .d = 0x1e }, pd 94 drivers/media/tuners/tda18271-maps.c { .lomax = 284000, .pd = 0x26, .d = 0x1c }, pd 95 drivers/media/tuners/tda18271-maps.c { .lomax = 305500, .pd = 0x25, .d = 0x1a }, pd 96 drivers/media/tuners/tda18271-maps.c { .lomax = 331000, .pd = 0x24, .d = 0x18 }, pd 97 drivers/media/tuners/tda18271-maps.c { .lomax = 361000, .pd = 0x23, .d = 0x16 }, pd 98 drivers/media/tuners/tda18271-maps.c { .lomax = 397500, .pd = 0x22, .d = 0x14 }, pd 99 drivers/media/tuners/tda18271-maps.c { .lomax = 441500, .pd = 0x21, .d = 0x12 }, pd 100 drivers/media/tuners/tda18271-maps.c { .lomax = 497000, .pd = 0x20, .d = 0x10 }, pd 101 drivers/media/tuners/tda18271-maps.c { .lomax = 530000, .pd = 0x17, .d = 0x0f }, pd 102 drivers/media/tuners/tda18271-maps.c { .lomax = 568000, .pd = 0x16, .d = 0x0e }, pd 103 drivers/media/tuners/tda18271-maps.c { .lomax = 611000, .pd = 0x15, .d = 0x0d }, pd 104 drivers/media/tuners/tda18271-maps.c { .lomax = 662000, .pd = 0x14, .d = 0x0c }, pd 105 drivers/media/tuners/tda18271-maps.c { .lomax = 722000, .pd = 0x13, .d = 0x0b }, pd 106 drivers/media/tuners/tda18271-maps.c { .lomax = 795000, .pd = 0x12, .d = 0x0a }, pd 107 drivers/media/tuners/tda18271-maps.c { .lomax = 883000, .pd = 0x11, .d = 0x09 }, pd 108 drivers/media/tuners/tda18271-maps.c { .lomax = 994000, .pd = 0x10, .d = 0x08 }, pd 109 drivers/media/tuners/tda18271-maps.c { .lomax = 0, .pd = 0x00, .d = 0x00 }, /* end */ pd 113 drivers/media/tuners/tda18271-maps.c { .lomax = 33000, .pd = 0xdd, .d = 0xd0 }, pd 114 drivers/media/tuners/tda18271-maps.c { .lomax = 36000, .pd = 0xdc, .d = 0xc0 }, pd 115 drivers/media/tuners/tda18271-maps.c { .lomax = 40000, .pd = 0xdb, .d = 0xb0 }, pd 116 drivers/media/tuners/tda18271-maps.c { .lomax = 44000, .pd = 0xda, .d = 0xa0 }, pd 117 drivers/media/tuners/tda18271-maps.c { .lomax = 49000, .pd = 0xd9, .d = 0x90 }, pd 118 drivers/media/tuners/tda18271-maps.c { .lomax = 55000, .pd = 0xd8, .d = 0x80 }, pd 119 drivers/media/tuners/tda18271-maps.c { .lomax = 63000, .pd = 0xd3, .d = 0x70 }, pd 120 drivers/media/tuners/tda18271-maps.c { .lomax = 67000, .pd = 0xcd, .d = 0x68 }, pd 121 drivers/media/tuners/tda18271-maps.c { .lomax = 73000, .pd = 0xcc, .d = 0x60 }, pd 122 drivers/media/tuners/tda18271-maps.c { .lomax = 80000, .pd = 0xcb, .d = 0x58 }, pd 123 drivers/media/tuners/tda18271-maps.c { .lomax = 88000, .pd = 0xca, .d = 0x50 }, pd 124 drivers/media/tuners/tda18271-maps.c { .lomax = 98000, .pd = 0xc9, .d = 0x48 }, pd 125 drivers/media/tuners/tda18271-maps.c { .lomax = 110000, .pd = 0xc8, .d = 0x40 }, pd 126 drivers/media/tuners/tda18271-maps.c { .lomax = 126000, .pd = 0xc3, .d = 0x38 }, pd 127 drivers/media/tuners/tda18271-maps.c { .lomax = 135000, .pd = 0xbd, .d = 0x34 }, pd 128 drivers/media/tuners/tda18271-maps.c { .lomax = 147000, .pd = 0xbc, .d = 0x30 }, pd 129 drivers/media/tuners/tda18271-maps.c { .lomax = 160000, .pd = 0xbb, .d = 0x2c }, pd 130 drivers/media/tuners/tda18271-maps.c { .lomax = 176000, .pd = 0xba, .d = 0x28 }, pd 131 drivers/media/tuners/tda18271-maps.c { .lomax = 196000, .pd = 0xb9, .d = 0x24 }, pd 132 drivers/media/tuners/tda18271-maps.c { .lomax = 220000, .pd = 0xb8, .d = 0x20 }, pd 133 drivers/media/tuners/tda18271-maps.c { .lomax = 252000, .pd = 0xb3, .d = 0x1c }, pd 134 drivers/media/tuners/tda18271-maps.c { .lomax = 271000, .pd = 0xad, .d = 0x1a }, pd 135 drivers/media/tuners/tda18271-maps.c { .lomax = 294000, .pd = 0xac, .d = 0x18 }, pd 136 drivers/media/tuners/tda18271-maps.c { .lomax = 321000, .pd = 0xab, .d = 0x16 }, pd 137 drivers/media/tuners/tda18271-maps.c { .lomax = 353000, .pd = 0xaa, .d = 0x14 }, pd 138 drivers/media/tuners/tda18271-maps.c { .lomax = 392000, .pd = 0xa9, .d = 0x12 }, pd 139 drivers/media/tuners/tda18271-maps.c { .lomax = 441000, .pd = 0xa8, .d = 0x10 }, pd 140 drivers/media/tuners/tda18271-maps.c { .lomax = 505000, .pd = 0xa3, .d = 0x0e }, pd 141 drivers/media/tuners/tda18271-maps.c { .lomax = 543000, .pd = 0x9d, .d = 0x0d }, pd 142 drivers/media/tuners/tda18271-maps.c { .lomax = 589000, .pd = 0x9c, .d = 0x0c }, pd 143 drivers/media/tuners/tda18271-maps.c { .lomax = 642000, .pd = 0x9b, .d = 0x0b }, pd 144 drivers/media/tuners/tda18271-maps.c { .lomax = 707000, .pd = 0x9a, .d = 0x0a }, pd 145 drivers/media/tuners/tda18271-maps.c { .lomax = 785000, .pd = 0x99, .d = 0x09 }, pd 146 drivers/media/tuners/tda18271-maps.c { .lomax = 883000, .pd = 0x98, .d = 0x08 }, pd 147 drivers/media/tuners/tda18271-maps.c { .lomax = 1010000, .pd = 0x93, .d = 0x07 }, pd 148 drivers/media/tuners/tda18271-maps.c { .lomax = 0, .pd = 0x00, .d = 0x00 }, /* end */ pd 152 drivers/media/tuners/tda18271-maps.c { .lomax = 33813, .pd = 0xdd, .d = 0xd0 }, pd 153 drivers/media/tuners/tda18271-maps.c { .lomax = 36625, .pd = 0xdc, .d = 0xc0 }, pd 154 drivers/media/tuners/tda18271-maps.c { .lomax = 39938, .pd = 0xdb, .d = 0xb0 }, pd 155 drivers/media/tuners/tda18271-maps.c { .lomax = 43938, .pd = 0xda, .d = 0xa0 }, pd 156 drivers/media/tuners/tda18271-maps.c { .lomax = 48813, .pd = 0xd9, .d = 0x90 }, pd 157 drivers/media/tuners/tda18271-maps.c { .lomax = 54938, .pd = 0xd8, .d = 0x80 }, pd 158 drivers/media/tuners/tda18271-maps.c { .lomax = 62813, .pd = 0xd3, .d = 0x70 }, pd 159 drivers/media/tuners/tda18271-maps.c { .lomax = 67625, .pd = 0xcd, .d = 0x68 }, pd 160 drivers/media/tuners/tda18271-maps.c { .lomax = 73250, .pd = 0xcc, .d = 0x60 }, pd 161 drivers/media/tuners/tda18271-maps.c { .lomax = 79875, .pd = 0xcb, .d = 0x58 }, pd 162 drivers/media/tuners/tda18271-maps.c { .lomax = 87875, .pd = 0xca, .d = 0x50 }, pd 163 drivers/media/tuners/tda18271-maps.c { .lomax = 97625, .pd = 0xc9, .d = 0x48 }, pd 164 drivers/media/tuners/tda18271-maps.c { .lomax = 109875, .pd = 0xc8, .d = 0x40 }, pd 165 drivers/media/tuners/tda18271-maps.c { .lomax = 125625, .pd = 0xc3, .d = 0x38 }, pd 166 drivers/media/tuners/tda18271-maps.c { .lomax = 135250, .pd = 0xbd, .d = 0x34 }, pd 167 drivers/media/tuners/tda18271-maps.c { .lomax = 146500, .pd = 0xbc, .d = 0x30 }, pd 168 drivers/media/tuners/tda18271-maps.c { .lomax = 159750, .pd = 0xbb, .d = 0x2c }, pd 169 drivers/media/tuners/tda18271-maps.c { .lomax = 175750, .pd = 0xba, .d = 0x28 }, pd 170 drivers/media/tuners/tda18271-maps.c { .lomax = 195250, .pd = 0xb9, .d = 0x24 }, pd 171 drivers/media/tuners/tda18271-maps.c { .lomax = 219750, .pd = 0xb8, .d = 0x20 }, pd 172 drivers/media/tuners/tda18271-maps.c { .lomax = 251250, .pd = 0xb3, .d = 0x1c }, pd 173 drivers/media/tuners/tda18271-maps.c { .lomax = 270500, .pd = 0xad, .d = 0x1a }, pd 174 drivers/media/tuners/tda18271-maps.c { .lomax = 293000, .pd = 0xac, .d = 0x18 }, pd 175 drivers/media/tuners/tda18271-maps.c { .lomax = 319500, .pd = 0xab, .d = 0x16 }, pd 176 drivers/media/tuners/tda18271-maps.c { .lomax = 351500, .pd = 0xaa, .d = 0x14 }, pd 177 drivers/media/tuners/tda18271-maps.c { .lomax = 390500, .pd = 0xa9, .d = 0x12 }, pd 178 drivers/media/tuners/tda18271-maps.c { .lomax = 439500, .pd = 0xa8, .d = 0x10 }, pd 179 drivers/media/tuners/tda18271-maps.c { .lomax = 502500, .pd = 0xa3, .d = 0x0e }, pd 180 drivers/media/tuners/tda18271-maps.c { .lomax = 541000, .pd = 0x9d, .d = 0x0d }, pd 181 drivers/media/tuners/tda18271-maps.c { .lomax = 586000, .pd = 0x9c, .d = 0x0c }, pd 182 drivers/media/tuners/tda18271-maps.c { .lomax = 639000, .pd = 0x9b, .d = 0x0b }, pd 183 drivers/media/tuners/tda18271-maps.c { .lomax = 703000, .pd = 0x9a, .d = 0x0a }, pd 184 drivers/media/tuners/tda18271-maps.c { .lomax = 781000, .pd = 0x99, .d = 0x09 }, pd 185 drivers/media/tuners/tda18271-maps.c { .lomax = 879000, .pd = 0x98, .d = 0x08 }, pd 186 drivers/media/tuners/tda18271-maps.c { .lomax = 0, .pd = 0x00, .d = 0x00 }, /* end */ pd 1093 drivers/media/tuners/tda18271-maps.c *post_div = map[i].pd; pd 1201 drivers/memory/emif.c struct emif_platform_data *pd = emif->plat_data; pd 1203 drivers/memory/emif.c pd->timings = lpddr2_jedec_timings; pd 1204 drivers/memory/emif.c pd->timings_arr_size = ARRAY_SIZE(lpddr2_jedec_timings); pd 1347 drivers/memory/emif.c struct emif_platform_data *pd = NULL; pd 1355 drivers/memory/emif.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 1358 drivers/memory/emif.c if (!emif || !pd || !dev_info) { pd 1364 drivers/memory/emif.c emif->plat_data = pd; pd 1365 drivers/memory/emif.c pd->device_info = dev_info; pd 1375 drivers/memory/emif.c of_property_read_u32(np_emif, "phy-type", &pd->phy_type); pd 1378 drivers/memory/emif.c pd->hw_caps |= EMIF_HW_CAPS_LL_INTERFACE; pd 1381 drivers/memory/emif.c if (!is_dev_data_valid(pd->device_info->type, pd->device_info->density, pd 1382 drivers/memory/emif.c pd->device_info->io_width, pd->phy_type, pd->ip_rev, pd 1431 drivers/memory/emif.c struct emif_platform_data *pd; pd 1435 drivers/memory/emif.c pd = pdev->dev.platform_data; pd 1438 drivers/memory/emif.c if (!(pd && pd->device_info && is_dev_data_valid(pd->device_info->type, pd 1439 drivers/memory/emif.c pd->device_info->density, pd->device_info->io_width, pd 1440 drivers/memory/emif.c pd->phy_type, pd->ip_rev, dev))) { pd 1446 drivers/memory/emif.c temp = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 1449 drivers/memory/emif.c if (!emif || !pd || !dev_info) { pd 1454 drivers/memory/emif.c memcpy(temp, pd, sizeof(*pd)); pd 1455 drivers/memory/emif.c pd = temp; pd 1456 drivers/memory/emif.c memcpy(dev_info, pd->device_info, sizeof(*dev_info)); pd 1458 drivers/memory/emif.c pd->device_info = dev_info; pd 1459 drivers/memory/emif.c emif->plat_data = pd; pd 1474 drivers/memory/emif.c pd->timings = NULL; pd 1475 drivers/memory/emif.c pd->min_tck = NULL; pd 1486 drivers/memory/emif.c cust_cfgs = pd->custom_configs; pd 1494 drivers/memory/emif.c pd->custom_configs = temp; pd 1501 drivers/memory/emif.c size = sizeof(struct lpddr2_timings) * pd->timings_arr_size; pd 1502 drivers/memory/emif.c if (pd->timings) { pd 1505 drivers/memory/emif.c memcpy(temp, pd->timings, size); pd 1506 drivers/memory/emif.c pd->timings = temp; pd 1516 drivers/memory/emif.c if (pd->min_tck) { pd 1517 drivers/memory/emif.c temp = devm_kzalloc(dev, sizeof(*pd->min_tck), GFP_KERNEL); pd 1519 drivers/memory/emif.c memcpy(temp, pd->min_tck, sizeof(*pd->min_tck)); pd 1520 drivers/memory/emif.c pd->min_tck = temp; pd 1524 drivers/memory/emif.c pd->min_tck = &lpddr2_jedec_min_tck; pd 1527 drivers/memory/emif.c pd->min_tck = &lpddr2_jedec_min_tck; pd 134 drivers/mfd/max8997.c struct max8997_platform_data *pd; pd 136 drivers/mfd/max8997.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 137 drivers/mfd/max8997.c if (!pd) pd 140 drivers/mfd/max8997.c pd->ono = irq_of_parse_and_map(dev->of_node, 1); pd 142 drivers/mfd/max8997.c return pd; pd 139 drivers/mfd/max8998.c struct max8998_platform_data *pd; pd 141 drivers/mfd/max8998.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 142 drivers/mfd/max8998.c if (!pd) pd 145 drivers/mfd/max8998.c pd->ono = irq_of_parse_and_map(dev->of_node, 1); pd 152 drivers/mfd/max8998.c return pd; pd 297 drivers/mfd/sec-core.c struct sec_platform_data *pd; pd 299 drivers/mfd/sec-core.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 300 drivers/mfd/sec-core.c if (!pd) pd 309 drivers/mfd/sec-core.c pd->manual_poweroff = of_property_read_bool(dev->of_node, pd 311 drivers/mfd/sec-core.c pd->disable_wrstbi = of_property_read_bool(dev->of_node, pd 313 drivers/mfd/sec-core.c return pd; pd 1445 drivers/mfd/sm501.c struct sm501_platdata *pd = sm->platdata; pd 1447 drivers/mfd/sm501.c if (!pd) pd 1450 drivers/mfd/sm501.c if (pd->get_power) { pd 1451 drivers/mfd/sm501.c if (pd->get_power(sm->dev) == on) { pd 1457 drivers/mfd/sm501.c if (pd->set_power) { pd 1460 drivers/mfd/sm501.c pd->set_power(sm->dev, on); pd 278 drivers/misc/eeprom/eeprom_93xx46.c struct eeprom_93xx46_platform_data *pd = edev->pdata; pd 319 drivers/misc/eeprom/eeprom_93xx46.c if (pd->finish) pd 320 drivers/misc/eeprom/eeprom_93xx46.c pd->finish(edev); pd 375 drivers/misc/eeprom/eeprom_93xx46.c struct eeprom_93xx46_platform_data *pd; pd 379 drivers/misc/eeprom/eeprom_93xx46.c pd = devm_kzalloc(&spi->dev, sizeof(*pd), GFP_KERNEL); pd 380 drivers/misc/eeprom/eeprom_93xx46.c if (!pd) pd 390 drivers/misc/eeprom/eeprom_93xx46.c pd->flags |= EE_ADDR8; pd 392 drivers/misc/eeprom/eeprom_93xx46.c pd->flags |= EE_ADDR16; pd 399 drivers/misc/eeprom/eeprom_93xx46.c pd->flags |= EE_READONLY; pd 401 drivers/misc/eeprom/eeprom_93xx46.c pd->select = devm_gpiod_get_optional(&spi->dev, "select", pd 403 drivers/misc/eeprom/eeprom_93xx46.c if (IS_ERR(pd->select)) pd 404 drivers/misc/eeprom/eeprom_93xx46.c return PTR_ERR(pd->select); pd 406 drivers/misc/eeprom/eeprom_93xx46.c pd->prepare = select_assert; pd 407 drivers/misc/eeprom/eeprom_93xx46.c pd->finish = select_deassert; pd 408 drivers/misc/eeprom/eeprom_93xx46.c gpiod_direction_output(pd->select, 0); pd 413 drivers/misc/eeprom/eeprom_93xx46.c pd->quirks = data->quirks; pd 416 drivers/misc/eeprom/eeprom_93xx46.c spi->dev.platform_data = pd; pd 423 drivers/misc/eeprom/eeprom_93xx46.c struct eeprom_93xx46_platform_data *pd; pd 433 drivers/misc/eeprom/eeprom_93xx46.c pd = spi->dev.platform_data; pd 434 drivers/misc/eeprom/eeprom_93xx46.c if (!pd) { pd 443 drivers/misc/eeprom/eeprom_93xx46.c if (pd->flags & EE_ADDR8) pd 445 drivers/misc/eeprom/eeprom_93xx46.c else if (pd->flags & EE_ADDR16) pd 455 drivers/misc/eeprom/eeprom_93xx46.c edev->pdata = pd; pd 460 drivers/misc/eeprom/eeprom_93xx46.c edev->nvmem_config.read_only = pd->flags & EE_READONLY; pd 477 drivers/misc/eeprom/eeprom_93xx46.c (pd->flags & EE_ADDR8) ? 8 : 16, pd 478 drivers/misc/eeprom/eeprom_93xx46.c (pd->flags & EE_READONLY) ? "(readonly)" : ""); pd 480 drivers/misc/eeprom/eeprom_93xx46.c if (!(pd->flags & EE_READONLY)) { pd 201 drivers/misc/fastrpc.c int pd; pd 897 drivers/misc/fastrpc.c msg->ctx = ctx->ctxid | fl->pd; pd 1009 drivers/misc/fastrpc.c fl->pd = 1; pd 1252 drivers/misc/fastrpc.c fl->pd = 0; pd 223 drivers/mmc/host/sh_mmcif.c struct platform_device *pd; pd 258 drivers/mmc/host/sh_mmcif.c #define sh_mmcif_host_to_dev(host) (&host->pd->dev) pd 407 drivers/mmc/host/sh_mmcif.c res = platform_get_resource(host->pd, IORESOURCE_MEM, 0); pd 1370 drivers/mmc/host/sh_mmcif.c struct sh_mmcif_plat_data *pd = dev->platform_data; pd 1375 drivers/mmc/host/sh_mmcif.c if (!pd) pd 1379 drivers/mmc/host/sh_mmcif.c mmc->ocr_avail = pd->ocr; pd 1380 drivers/mmc/host/sh_mmcif.c else if (pd->ocr) pd 1390 drivers/mmc/host/sh_mmcif.c struct sh_mmcif_plat_data *pd = dev->platform_data; pd 1420 drivers/mmc/host/sh_mmcif.c host->pd = pdev; pd 1431 drivers/mmc/host/sh_mmcif.c if (pd && pd->caps) pd 1432 drivers/mmc/host/sh_mmcif.c mmc->caps |= pd->caps; pd 378 drivers/mtd/nand/raw/au1550nd.c struct au1550nd_platdata *pd; pd 385 drivers/mtd/nand/raw/au1550nd.c pd = dev_get_platdata(&pdev->dev); pd 386 drivers/mtd/nand/raw/au1550nd.c if (!pd) { pd 436 drivers/mtd/nand/raw/au1550nd.c if (pd->devwidth) pd 439 drivers/mtd/nand/raw/au1550nd.c this->legacy.read_byte = (pd->devwidth) ? au_read_byte16 : au_read_byte; pd 440 drivers/mtd/nand/raw/au1550nd.c ctx->write_byte = (pd->devwidth) ? au_write_byte16 : au_write_byte; pd 441 drivers/mtd/nand/raw/au1550nd.c this->legacy.write_buf = (pd->devwidth) ? au_write_buf16 : au_write_buf; pd 442 drivers/mtd/nand/raw/au1550nd.c this->legacy.read_buf = (pd->devwidth) ? au_read_buf16 : au_read_buf; pd 450 drivers/mtd/nand/raw/au1550nd.c mtd_device_register(mtd, pd->parts, pd->num_parts); pd 1470 drivers/net/can/grcan.c static void grcan_sanitize_##name(struct platform_device *pd) \ pd 1475 drivers/net/can/grcan.c dev_err(&pd->dev, \ pd 41 drivers/net/dsa/sja1105/sja1105_clocking.c u64 pd; pd 57 drivers/net/dsa/sja1105/sja1105_clocking.c u64 pd; pd 87 drivers/net/dsa/sja1105/sja1105_clocking.c u64 pd; pd 98 drivers/net/dsa/sja1105/sja1105_clocking.c sja1105_packing(buf, &idiv->pd, 0, 0, size, op); pd 118 drivers/net/dsa/sja1105/sja1105_clocking.c idiv.pd = enabled ? 0 : 1; /* Power down? */ pd 134 drivers/net/dsa/sja1105/sja1105_clocking.c sja1105_packing(buf, &cmd->pd, 0, 0, size, op); pd 167 drivers/net/dsa/sja1105/sja1105_clocking.c mii_tx_clk.pd = 0; /* Power Down off => enabled */ pd 192 drivers/net/dsa/sja1105/sja1105_clocking.c mii_rx_clk.pd = 0; /* Power Down off => enabled */ pd 217 drivers/net/dsa/sja1105/sja1105_clocking.c mii_ext_tx_clk.pd = 0; /* Power Down off => enabled */ pd 242 drivers/net/dsa/sja1105/sja1105_clocking.c mii_ext_rx_clk.pd = 0; /* Power Down off => enabled */ pd 313 drivers/net/dsa/sja1105/sja1105_clocking.c sja1105_packing(buf, &cmd->pd, 0, 0, size, op); pd 337 drivers/net/dsa/sja1105/sja1105_clocking.c txc.pd = 0; pd 547 drivers/net/dsa/sja1105/sja1105_clocking.c ref_clk.pd = 0; /* Power Down off => enabled */ pd 565 drivers/net/dsa/sja1105/sja1105_clocking.c ext_tx_clk.pd = 0; /* Power Down off => enabled */ pd 595 drivers/net/dsa/sja1105/sja1105_clocking.c pll.pd = 0x1; pd 606 drivers/net/dsa/sja1105/sja1105_clocking.c pll.pd = 0x0; pd 1081 drivers/net/ethernet/amd/au1000_eth.c struct au1000_eth_platform_data *pd; pd 1193 drivers/net/ethernet/amd/au1000_eth.c pd = dev_get_platdata(&pdev->dev); pd 1194 drivers/net/ethernet/amd/au1000_eth.c if (!pd) { pd 1199 drivers/net/ethernet/amd/au1000_eth.c if (is_valid_ether_addr(pd->mac)) { pd 1200 drivers/net/ethernet/amd/au1000_eth.c memcpy(dev->dev_addr, pd->mac, ETH_ALEN); pd 1206 drivers/net/ethernet/amd/au1000_eth.c aup->phy_static_config = pd->phy_static_config; pd 1207 drivers/net/ethernet/amd/au1000_eth.c aup->phy_search_highest_addr = pd->phy_search_highest_addr; pd 1208 drivers/net/ethernet/amd/au1000_eth.c aup->phy1_search_mac0 = pd->phy1_search_mac0; pd 1209 drivers/net/ethernet/amd/au1000_eth.c aup->phy_addr = pd->phy_addr; pd 1210 drivers/net/ethernet/amd/au1000_eth.c aup->phy_busid = pd->phy_busid; pd 1211 drivers/net/ethernet/amd/au1000_eth.c aup->phy_irq = pd->phy_irq; pd 1695 drivers/net/ethernet/broadcom/bcm63xx_enet.c struct bcm63xx_enet_platform_data *pd; pd 1745 drivers/net/ethernet/broadcom/bcm63xx_enet.c pd = dev_get_platdata(&pdev->dev); pd 1746 drivers/net/ethernet/broadcom/bcm63xx_enet.c if (pd) { pd 1747 drivers/net/ethernet/broadcom/bcm63xx_enet.c memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN); pd 1748 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->has_phy = pd->has_phy; pd 1749 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->phy_id = pd->phy_id; pd 1750 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->has_phy_interrupt = pd->has_phy_interrupt; pd 1751 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->phy_interrupt = pd->phy_interrupt; pd 1752 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->use_external_mii = !pd->use_internal_phy; pd 1753 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->pause_auto = pd->pause_auto; pd 1754 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->pause_rx = pd->pause_rx; pd 1755 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->pause_tx = pd->pause_tx; pd 1756 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->force_duplex_full = pd->force_duplex_full; pd 1757 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->force_speed_100 = pd->force_speed_100; pd 1758 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_en_mask = pd->dma_chan_en_mask; pd 1759 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_int_mask = pd->dma_chan_int_mask; pd 1760 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_width = pd->dma_chan_width; pd 1761 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_has_sram = pd->dma_has_sram; pd 1762 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_desc_shift = pd->dma_desc_shift; pd 1763 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->rx_chan = pd->rx_chan; pd 1764 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->tx_chan = pd->tx_chan; pd 1816 drivers/net/ethernet/broadcom/bcm63xx_enet.c if (pd && pd->mii_config && pd 1817 drivers/net/ethernet/broadcom/bcm63xx_enet.c pd->mii_config(dev, 1, bcm_enet_mdio_read_mii, pd 1899 drivers/net/ethernet/broadcom/bcm63xx_enet.c struct bcm63xx_enet_platform_data *pd; pd 1901 drivers/net/ethernet/broadcom/bcm63xx_enet.c pd = dev_get_platdata(&pdev->dev); pd 1902 drivers/net/ethernet/broadcom/bcm63xx_enet.c if (pd && pd->mii_config) pd 1903 drivers/net/ethernet/broadcom/bcm63xx_enet.c pd->mii_config(dev, 0, bcm_enet_mdio_read_mii, pd 2643 drivers/net/ethernet/broadcom/bcm63xx_enet.c struct bcm63xx_enetsw_platform_data *pd; pd 2670 drivers/net/ethernet/broadcom/bcm63xx_enet.c pd = dev_get_platdata(&pdev->dev); pd 2671 drivers/net/ethernet/broadcom/bcm63xx_enet.c if (pd) { pd 2672 drivers/net/ethernet/broadcom/bcm63xx_enet.c memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN); pd 2673 drivers/net/ethernet/broadcom/bcm63xx_enet.c memcpy(priv->used_ports, pd->used_ports, pd 2674 drivers/net/ethernet/broadcom/bcm63xx_enet.c sizeof(pd->used_ports)); pd 2675 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->num_ports = pd->num_ports; pd 2676 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_has_sram = pd->dma_has_sram; pd 2677 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_en_mask = pd->dma_chan_en_mask; pd 2678 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_int_mask = pd->dma_chan_int_mask; pd 2679 drivers/net/ethernet/broadcom/bcm63xx_enet.c priv->dma_chan_width = pd->dma_chan_width; pd 5736 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c u16 pd, gp_status4; pd 5749 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c &pd); pd 5750 drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c if (pd & (1<<15)) pd 3439 drivers/net/ethernet/broadcom/genet/bcmgenet.c struct bcmgenet_platform_data *pd = pdev->dev.platform_data; pd 3481 drivers/net/ethernet/broadcom/genet/bcmgenet.c macaddr = pd->mac_address; pd 3524 drivers/net/ethernet/broadcom/genet/bcmgenet.c priv->version = pd->genet_version; pd 394 drivers/net/ethernet/broadcom/genet/bcmmii.c struct bcmgenet_platform_data *pd = kdev->platform_data; pd 396 drivers/net/ethernet/broadcom/genet/bcmmii.c if (pd->phy_interface != PHY_INTERFACE_MODE_MOCA && pd->mdio_enabled) { pd 400 drivers/net/ethernet/broadcom/genet/bcmmii.c if (pd->phy_address >= 0 && pd->phy_address < PHY_MAX_ADDR) pd 401 drivers/net/ethernet/broadcom/genet/bcmmii.c ppd->phy_mask = 1 << pd->phy_address; pd 530 drivers/net/ethernet/broadcom/genet/bcmmii.c struct bcmgenet_platform_data *pd = kdev->platform_data; pd 538 drivers/net/ethernet/broadcom/genet/bcmmii.c if (pd->phy_interface != PHY_INTERFACE_MODE_MOCA && pd->mdio_enabled) { pd 540 drivers/net/ethernet/broadcom/genet/bcmmii.c mdio_bus_id, pd->phy_address); pd 545 drivers/net/ethernet/broadcom/genet/bcmmii.c phydev = phy_attach(priv->dev, phy_name, pd->phy_interface); pd 557 drivers/net/ethernet/broadcom/genet/bcmmii.c .speed = pd->phy_speed, pd 558 drivers/net/ethernet/broadcom/genet/bcmmii.c .duplex = pd->phy_duplex, pd 574 drivers/net/ethernet/broadcom/genet/bcmmii.c priv->phy_interface = pd->phy_interface; pd 12003 drivers/net/ethernet/broadcom/tg3.c u8 *pd; pd 12047 drivers/net/ethernet/broadcom/tg3.c pd = &data[eeprom->len]; pd 12056 drivers/net/ethernet/broadcom/tg3.c memcpy(pd + i, &val, 4); pd 12070 drivers/net/ethernet/broadcom/tg3.c pd = &data[eeprom->len]; pd 12076 drivers/net/ethernet/broadcom/tg3.c memcpy(pd, &val, b_count); pd 1475 drivers/net/ethernet/cavium/thunder/thunder_bgx.c struct phy_device *pd; pd 1501 drivers/net/ethernet/cavium/thunder/thunder_bgx.c pd = of_phy_find_device(phy_np); pd 1502 drivers/net/ethernet/cavium/thunder/thunder_bgx.c if (!pd) pd 1504 drivers/net/ethernet/cavium/thunder/thunder_bgx.c bgx->lmac[lmac].phydev = pd; pd 372 drivers/net/ethernet/emulex/benet/be_cmds.h u8 pd[10]; /* dword 1*/ pd 464 drivers/net/ethernet/emulex/benet/be_cmds.h u8 pd[10]; /* dword 1*/ pd 385 drivers/net/ethernet/ibm/ehea/ehea.h u32 pd; /* protection domain */ pd 1512 drivers/net/ethernet/ibm/ehea/ehea_main.c pr->qp = ehea_create_qp(adapter, adapter->pd, init_attr); pd 3399 drivers/net/ethernet/ibm/ehea/ehea_main.c adapter->pd = EHEA_PD_ID; pd 199 drivers/net/ethernet/ibm/ehea/ehea_phyp.c struct ehea_qp_init_attr *init_attr, const u32 pd, pd 216 drivers/net/ethernet/ibm/ehea/ehea_phyp.c u64 r9_reg = EHEA_BMASK_SET(H_ALL_RES_QP_PD, pd) pd 434 drivers/net/ethernet/ibm/ehea/ehea_phyp.c const u64 vaddr_in, const u32 access_ctrl, const u32 pd, pd 446 drivers/net/ethernet/ibm/ehea/ehea_phyp.c pd, /* R8 */ pd 479 drivers/net/ethernet/ibm/ehea/ehea_phyp.c const u32 pd, u64 *mr_handle, u32 *lkey) pd 491 drivers/net/ethernet/ibm/ehea/ehea_phyp.c pd, /* R9 */ pd 368 drivers/net/ethernet/ibm/ehea/ehea_phyp.h const u32 pd, pd 394 drivers/net/ethernet/ibm/ehea/ehea_phyp.h const u32 pd, u64 *mr_handle, u32 *lkey); pd 401 drivers/net/ethernet/ibm/ehea/ehea_phyp.h const u64 vaddr_in, const u32 access_ctrl, const u32 pd, pd 398 drivers/net/ethernet/ibm/ehea/ehea_qmr.c u32 pd, struct ehea_qp_init_attr *init_attr) pd 413 drivers/net/ethernet/ibm/ehea/ehea_qmr.c hret = ehea_h_alloc_resource_qp(adapter->handle, init_attr, pd, pd 877 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ehea_mr_len, acc_ctrl, adapter->pd, pd 941 drivers/net/ethernet/ibm/ehea/ehea_qmr.c adapter->pd, shared_mr); pd 369 drivers/net/ethernet/ibm/ehea/ehea_qmr.h struct ehea_qp *ehea_create_qp(struct ehea_adapter *adapter, u32 pd, pd 1484 drivers/net/ethernet/ibm/emac/core.c u32 pd, int len, int last, u16 base_ctrl) pd 1499 drivers/net/ethernet/ibm/emac/core.c dev->tx_desc[slot].data_ptr = pd; pd 1507 drivers/net/ethernet/ibm/emac/core.c pd += chunk; pd 1521 drivers/net/ethernet/ibm/emac/core.c u32 pd; pd 1543 drivers/net/ethernet/ibm/emac/core.c dev->tx_desc[slot].data_ptr = pd = pd 1548 drivers/net/ethernet/ibm/emac/core.c slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags, pd 1558 drivers/net/ethernet/ibm/emac/core.c pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len, pd 1561 drivers/net/ethernet/ibm/emac/core.c slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1, pd 2786 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_shared_platform_data *pd; pd 2794 drivers/net/ethernet/marvell/mv643xx_eth.c pd = devm_kzalloc(&pdev->dev, sizeof(*pd), GFP_KERNEL); pd 2795 drivers/net/ethernet/marvell/mv643xx_eth.c if (!pd) pd 2797 drivers/net/ethernet/marvell/mv643xx_eth.c pdev->dev.platform_data = pd; pd 2799 drivers/net/ethernet/marvell/mv643xx_eth.c mv643xx_eth_property(np, "tx-checksum-limit", pd->tx_csum_limit); pd 2834 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_shared_platform_data *pd; pd 2871 drivers/net/ethernet/marvell/mv643xx_eth.c pd = dev_get_platdata(&pdev->dev); pd 2873 drivers/net/ethernet/marvell/mv643xx_eth.c msp->tx_csum_limit = (pd != NULL && pd->tx_csum_limit) ? pd 2874 drivers/net/ethernet/marvell/mv643xx_eth.c pd->tx_csum_limit : 9 * 1024; pd 2925 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_platform_data *pd) pd 2930 drivers/net/ethernet/marvell/mv643xx_eth.c if (is_valid_ether_addr(pd->mac_addr)) pd 2931 drivers/net/ethernet/marvell/mv643xx_eth.c memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN); pd 2936 drivers/net/ethernet/marvell/mv643xx_eth.c if (pd->rx_queue_size) pd 2937 drivers/net/ethernet/marvell/mv643xx_eth.c mp->rx_ring_size = pd->rx_queue_size; pd 2938 drivers/net/ethernet/marvell/mv643xx_eth.c mp->rx_desc_sram_addr = pd->rx_sram_addr; pd 2939 drivers/net/ethernet/marvell/mv643xx_eth.c mp->rx_desc_sram_size = pd->rx_sram_size; pd 2941 drivers/net/ethernet/marvell/mv643xx_eth.c mp->rxq_count = pd->rx_queue_count ? : 1; pd 2944 drivers/net/ethernet/marvell/mv643xx_eth.c if (pd->tx_queue_size) pd 2945 drivers/net/ethernet/marvell/mv643xx_eth.c tx_ring_size = pd->tx_queue_size; pd 2953 drivers/net/ethernet/marvell/mv643xx_eth.c mp->tx_desc_sram_addr = pd->tx_sram_addr; pd 2954 drivers/net/ethernet/marvell/mv643xx_eth.c mp->tx_desc_sram_size = pd->tx_sram_size; pd 2956 drivers/net/ethernet/marvell/mv643xx_eth.c mp->txq_count = pd->tx_queue_count ? : 1; pd 3080 drivers/net/ethernet/marvell/mv643xx_eth.c struct mv643xx_eth_platform_data *pd; pd 3087 drivers/net/ethernet/marvell/mv643xx_eth.c pd = dev_get_platdata(&pdev->dev); pd 3088 drivers/net/ethernet/marvell/mv643xx_eth.c if (pd == NULL) { pd 3093 drivers/net/ethernet/marvell/mv643xx_eth.c if (pd->shared == NULL) { pd 3106 drivers/net/ethernet/marvell/mv643xx_eth.c mp->shared = platform_get_drvdata(pd->shared); pd 3107 drivers/net/ethernet/marvell/mv643xx_eth.c mp->base = mp->shared->base + 0x0400 + (pd->port_number << 10); pd 3108 drivers/net/ethernet/marvell/mv643xx_eth.c mp->port_num = pd->port_number; pd 3134 drivers/net/ethernet/marvell/mv643xx_eth.c set_params(mp, pd); pd 3139 drivers/net/ethernet/marvell/mv643xx_eth.c if (pd->phy_node) { pd 3140 drivers/net/ethernet/marvell/mv643xx_eth.c phydev = of_phy_connect(mp->dev, pd->phy_node, pd 3147 drivers/net/ethernet/marvell/mv643xx_eth.c } else if (pd->phy_addr != MV643XX_ETH_PHY_NONE) { pd 3148 drivers/net/ethernet/marvell/mv643xx_eth.c phydev = phy_scan(mp, pd->phy_addr); pd 3153 drivers/net/ethernet/marvell/mv643xx_eth.c phy_init(mp, pd->speed, pd->duplex); pd 3164 drivers/net/ethernet/marvell/mv643xx_eth.c init_pscr(mp, pd->speed, pd->duplex); pd 242 drivers/net/ethernet/marvell/pxa168_eth.c struct pxa168_eth_platform_data *pd; pd 1465 drivers/net/ethernet/marvell/pxa168_eth.c pep->pd = dev_get_platdata(&pdev->dev); pd 1466 drivers/net/ethernet/marvell/pxa168_eth.c if (pep->pd) { pd 1467 drivers/net/ethernet/marvell/pxa168_eth.c if (pep->pd->rx_queue_size) pd 1468 drivers/net/ethernet/marvell/pxa168_eth.c pep->rx_ring_size = pep->pd->rx_queue_size; pd 1470 drivers/net/ethernet/marvell/pxa168_eth.c if (pep->pd->tx_queue_size) pd 1471 drivers/net/ethernet/marvell/pxa168_eth.c pep->tx_ring_size = pep->pd->tx_queue_size; pd 1473 drivers/net/ethernet/marvell/pxa168_eth.c pep->port_num = pep->pd->port_number; pd 1474 drivers/net/ethernet/marvell/pxa168_eth.c pep->phy_addr = pep->pd->phy_addr; pd 1475 drivers/net/ethernet/marvell/pxa168_eth.c pep->phy_speed = pep->pd->speed; pd 1476 drivers/net/ethernet/marvell/pxa168_eth.c pep->phy_duplex = pep->pd->duplex; pd 1477 drivers/net/ethernet/marvell/pxa168_eth.c pep->phy_intf = pep->pd->intf; pd 1479 drivers/net/ethernet/marvell/pxa168_eth.c if (pep->pd->init) pd 1480 drivers/net/ethernet/marvell/pxa168_eth.c pep->pd->init(); pd 49 drivers/net/ethernet/mellanox/mlx4/en_resources.c context->pd = cpu_to_be32(mdev->priv_pdn); pd 374 drivers/net/ethernet/mellanox/mlx4/mlx4.h __be32 pd; pd 371 drivers/net/ethernet/mellanox/mlx4/mr.c mmr->pd = be32_to_cpu((*mpt_entry)->pd_flags) & MLX4_MPT_PD_MASK; pd 417 drivers/net/ethernet/mellanox/mlx4/mr.c static int mlx4_mr_alloc_reserved(struct mlx4_dev *dev, u32 mridx, u32 pd, pd 423 drivers/net/ethernet/mellanox/mlx4/mr.c mr->pd = pd; pd 527 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access, pd 537 drivers/net/ethernet/mellanox/mlx4/mr.c err = mlx4_mr_alloc_reserved(dev, index, pd, iova, size, pd 648 drivers/net/ethernet/mellanox/mlx4/mr.c mpt_entry->pd_flags = cpu_to_be32(mr->pd | MLX4_MPT_PD_FLAG_EN_INV); pd 814 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type, pd 830 drivers/net/ethernet/mellanox/mlx4/mr.c mw->pd = pd; pd 859 drivers/net/ethernet/mellanox/mlx4/mr.c mpt_entry->pd_flags = cpu_to_be32(mw->pd); pd 1043 drivers/net/ethernet/mellanox/mlx4/mr.c int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages, pd 1064 drivers/net/ethernet/mellanox/mlx4/mr.c err = mlx4_mr_alloc(dev, pd, 0, 0, access, max_pages, pd 2767 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u32 pd; pd 2782 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c pd = mr_get_pd(inbox->buf); pd 2783 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c pd_slave = (pd >> 17) & 0x7f; pd 198 drivers/net/ethernet/mellanox/mlx4/srq.c srq_context->pd = cpu_to_be32(pdn); pd 199 drivers/net/ethernet/mellanox/mlx5/core/diag/fw_tracer.c MLX5_SET(mkc, mkc, pd, tracer->buff.pdn); pd 81 drivers/net/ethernet/mellanox/mlx5/core/en_common.c MLX5_SET(mkc, mkc, pd, pdn); pd 289 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(mkc, mkc, pd, mdev->mlx5e_res.pdn); pd 2178 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(wq, wq, pd, mdev->mlx5e_res.pdn); pd 2208 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(wq, wq, pd, priv->mdev->mlx5e_res.pdn); pd 3210 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(tisc, tisc, pd, mdev->mlx5e_res.pdn); pd 240 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(mkc, mkc, pd, pdn); pd 589 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(qpc, qpc, pd, fdev->conn_res.pdn); pd 698 drivers/net/ethernet/mellanox/mlx5/core/fpga/conn.c MLX5_SET(qpc, qpc, pd, conn->fdev->conn_res.pdn); pd 50 drivers/net/ethernet/mellanox/mlx5/core/lib/crypto.c MLX5_SET(encryption_key_obj, obj, pd, mdev->mlx5e_res.pdn); pd 83 drivers/net/ethernet/mellanox/mlx5/core/mr.c mkey->pd = MLX5_GET(mkc, mkc, pd); pd 160 drivers/net/ethernet/mellanox/mlx5/core/mr.c MLX5_SET(create_psv_in, in, pd, pdn); pd 48 drivers/net/ethernet/mellanox/mlx5/core/pd.c *pdn = MLX5_GET(alloc_pd_out, out, pd); pd 59 drivers/net/ethernet/mellanox/mlx5/core/pd.c MLX5_SET(dealloc_pd_in, in, pd, pdn); pd 71 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c u32 pd, u64 length, u64 start_addr, int mode, pd 90 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c MLX5_SET(mkc, mkc, pd, pd); pd 164 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c MLX5_SET(qpc, qpc, pd, attr->pdn); pd 811 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c MLX5_SET(mkc, mkc, pd, pdn); pd 262 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c const struct mlxfw_mfa2_tlv_package_descriptor *pd; pd 292 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c pd = mlxfw_mfa2_tlv_package_descriptor_get(mfa2_file, multi_child); pd 293 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c if (!pd) { pd 304 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mfa2_file->dev_count = be16_to_cpu(pd->num_devices); pd 308 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mfa2_file->component_count = be16_to_cpu(pd->num_components); pd 309 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mfa2_file->cb = fw->data + NLA_ALIGN(be32_to_cpu(pd->cb_offset)); pd 314 drivers/net/ethernet/mellanox/mlxfw/mlxfw_mfa2.c mfa2_file->cb_archive_size = be32_to_cpu(pd->cb_archive_size); pd 7273 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 pd; pd 7922 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 pd; pd 7984 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 pd; pd 9957 drivers/net/ethernet/qlogic/qed/qed_hsi.h __le16 pd; pd 275 drivers/net/ethernet/qlogic/qed/qed_iwarp.c p_ramrod->pd = qp->pd; pd 954 drivers/net/ethernet/qlogic/qed/qed_rdma.c static int qed_rdma_alloc_pd(void *rdma_cxt, u16 *pd) pd 968 drivers/net/ethernet/qlogic/qed/qed_rdma.c *pd = (u16)returned_id; pd 974 drivers/net/ethernet/qlogic/qed/qed_rdma.c static void qed_rdma_free_pd(void *rdma_cxt, u16 pd) pd 978 drivers/net/ethernet/qlogic/qed/qed_rdma.c DP_VERBOSE(p_hwfn, QED_MSG_RDMA, "pd = %08x\n", pd); pd 982 drivers/net/ethernet/qlogic/qed/qed_rdma.c qed_bmap_release_id(p_hwfn, &p_hwfn->p_rdma_info->pd_map, pd); pd 1296 drivers/net/ethernet/qlogic/qed/qed_rdma.c qp->pd = in_params->pd; pd 1533 drivers/net/ethernet/qlogic/qed/qed_rdma.c p_ramrod->pd = cpu_to_le16(params->pd); pd 124 drivers/net/ethernet/qlogic/qed/qed_rdma.h u16 pd; pd 327 drivers/net/ethernet/qlogic/qed/qed_roce.c p_ramrod->pd = cpu_to_le16(qp->pd); pd 457 drivers/net/ethernet/qlogic/qed/qed_roce.c p_ramrod->pd = cpu_to_le16(qp->pd); pd 3087 drivers/net/ethernet/renesas/sh_eth.c struct sh_eth_plat_data *pd) pd 3101 drivers/net/ethernet/renesas/sh_eth.c bitbang->set_gate = pd->set_mdio_gate; pd 3116 drivers/net/ethernet/renesas/sh_eth.c if (pd->phy_irq > 0) pd 3117 drivers/net/ethernet/renesas/sh_eth.c mdp->mii_bus->irq[pd->phy] = pd->phy_irq; pd 3242 drivers/net/ethernet/renesas/sh_eth.c struct sh_eth_plat_data *pd = dev_get_platdata(&pdev->dev); pd 3280 drivers/net/ethernet/renesas/sh_eth.c pd = sh_eth_parse_dt(&pdev->dev); pd 3281 drivers/net/ethernet/renesas/sh_eth.c if (!pd) { pd 3288 drivers/net/ethernet/renesas/sh_eth.c mdp->phy_id = pd->phy; pd 3289 drivers/net/ethernet/renesas/sh_eth.c mdp->phy_interface = pd->phy_interface; pd 3290 drivers/net/ethernet/renesas/sh_eth.c mdp->no_ether_link = pd->no_ether_link; pd 3291 drivers/net/ethernet/renesas/sh_eth.c mdp->ether_link_active_low = pd->ether_link_active_low; pd 3332 drivers/net/ethernet/renesas/sh_eth.c read_mac_address(ndev, pd->mac_addr); pd 3385 drivers/net/ethernet/renesas/sh_eth.c ret = sh_mdio_init(mdp, pd); pd 724 drivers/net/ethernet/seeq/sgiseeq.c struct sgiseeq_platform_data *pd = dev_get_platdata(&pdev->dev); pd 725 drivers/net/ethernet/seeq/sgiseeq.c struct hpc3_regs *hpcregs = pd->hpc; pd 727 drivers/net/ethernet/seeq/sgiseeq.c unsigned int irq = pd->irq; pd 759 drivers/net/ethernet/seeq/sgiseeq.c memcpy(dev->dev_addr, pd->mac, ETH_ALEN); pd 255 drivers/net/ethernet/sfc/falcon/tenxpress.c struct tenxpress_phy_data *pd = efx->phy_data; pd 268 drivers/net/ethernet/sfc/falcon/tenxpress.c pd->bad_lp_tries++; pd 272 drivers/net/ethernet/sfc/falcon/tenxpress.c if (!pd->bad_lp_tries) pd 277 drivers/net/ethernet/sfc/falcon/tenxpress.c if (!bad_lp || pd->bad_lp_tries == MAX_BAD_LP_TRIES) { pd 293 drivers/net/ethernet/sfc/falcon/tenxpress.c pd->bad_lp_tries = bad_lp; pd 361 drivers/net/ethernet/sfc/falcon/txc43128_phy.c int pd = (1 << TXC_GLCMD_L01PD_LBN) | (1 << TXC_GLCMD_L23PD_LBN); pd 365 drivers/net/ethernet/sfc/falcon/txc43128_phy.c ctl &= ~pd; pd 367 drivers/net/ethernet/sfc/falcon/txc43128_phy.c ctl |= pd; pd 2076 drivers/net/ethernet/smsc/smc911x.c struct smc911x_platdata *pd = dev_get_platdata(&pdev->dev); pd 2077 drivers/net/ethernet/smsc/smc911x.c if (!pd) { pd 2081 drivers/net/ethernet/smsc/smc911x.c memcpy(&lp->cfg, pd, sizeof(lp->cfg)); pd 2232 drivers/net/ethernet/smsc/smc91x.c struct smc91x_platdata *pd = dev_get_platdata(&pdev->dev); pd 2256 drivers/net/ethernet/smsc/smc91x.c if (pd) { pd 2257 drivers/net/ethernet/smsc/smc91x.c memcpy(&lp->cfg, pd, sizeof(lp->cfg)); pd 2315 drivers/net/ethernet/smsc/smc91x.c if (!pd && !match) { pd 86 drivers/net/ethernet/smsc/smsc9420.c static inline u32 smsc9420_reg_read(struct smsc9420_pdata *pd, u32 offset) pd 88 drivers/net/ethernet/smsc/smsc9420.c return ioread32(pd->ioaddr + offset); pd 92 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(struct smsc9420_pdata *pd, u32 offset, u32 value) pd 94 drivers/net/ethernet/smsc/smsc9420.c iowrite32(value, pd->ioaddr + offset); pd 97 drivers/net/ethernet/smsc/smsc9420.c static inline void smsc9420_pci_flush_write(struct smsc9420_pdata *pd) pd 100 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_read(pd, ID_REV); pd 105 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = (struct smsc9420_pdata *)bus->priv; pd 110 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->phy_lock, flags); pd 113 drivers/net/ethernet/smsc/smsc9420.c if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) { pd 114 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "MII is busy???\n"); pd 121 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MII_ACCESS, addr); pd 125 drivers/net/ethernet/smsc/smsc9420.c if (!(smsc9420_reg_read(pd, MII_ACCESS) & pd 127 drivers/net/ethernet/smsc/smsc9420.c reg = (u16)smsc9420_reg_read(pd, MII_DATA); pd 133 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "MII busy timeout!\n"); pd 136 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->phy_lock, flags); pd 143 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = (struct smsc9420_pdata *)bus->priv; pd 148 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->phy_lock, flags); pd 151 drivers/net/ethernet/smsc/smsc9420.c if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) { pd 152 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "MII is busy???\n"); pd 157 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MII_DATA, (u32)val); pd 162 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MII_ACCESS, addr); pd 166 drivers/net/ethernet/smsc/smsc9420.c if (!(smsc9420_reg_read(pd, MII_ACCESS) & pd 174 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "MII busy timeout!\n"); pd 177 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->phy_lock, flags); pd 189 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_reload(struct smsc9420_pdata *pd) pd 193 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd); pd 195 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) { pd 196 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, drv, pd->dev, "%s: Eeprom busy\n", __func__); pd 200 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, E2P_CMD, pd 205 drivers/net/ethernet/smsc/smsc9420.c if (!(smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_)) pd 209 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "%s: Eeprom timed out\n", __func__); pd 225 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(netdev); pd 228 drivers/net/ethernet/smsc/smsc9420.c strlcpy(drvinfo->bus_info, pci_name(pd->pdev), pd 235 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(netdev); pd 236 drivers/net/ethernet/smsc/smsc9420.c return pd->msg_enable; pd 241 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(netdev); pd 242 drivers/net/ethernet/smsc/smsc9420.c pd->msg_enable = data; pd 255 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 260 drivers/net/ethernet/smsc/smsc9420.c regs->version = smsc9420_reg_read(pd, ID_REV); pd 262 drivers/net/ethernet/smsc/smsc9420.c data[j++] = smsc9420_reg_read(pd, i); pd 273 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_eeprom_enable_access(struct smsc9420_pdata *pd) pd 275 drivers/net/ethernet/smsc/smsc9420.c unsigned int temp = smsc9420_reg_read(pd, GPIO_CFG); pd 277 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, GPIO_CFG, temp); pd 281 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_send_cmd(struct smsc9420_pdata *pd, u32 op) pd 286 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "op 0x%08x\n", op); pd 287 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) { pd 288 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, hw, pd->dev, "Busy at start\n"); pd 293 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, E2P_CMD, e2cmd); pd 297 drivers/net/ethernet/smsc/smsc9420.c e2cmd = smsc9420_reg_read(pd, E2P_CMD); pd 301 drivers/net/ethernet/smsc/smsc9420.c netif_info(pd, hw, pd->dev, "TIMED OUT\n"); pd 306 drivers/net/ethernet/smsc/smsc9420.c netif_info(pd, hw, pd->dev, pd 314 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_read_location(struct smsc9420_pdata *pd, pd 320 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "address 0x%x\n", address); pd 321 drivers/net/ethernet/smsc/smsc9420.c ret = smsc9420_eeprom_send_cmd(pd, op); pd 324 drivers/net/ethernet/smsc/smsc9420.c data[address] = smsc9420_reg_read(pd, E2P_DATA); pd 329 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_eeprom_write_location(struct smsc9420_pdata *pd, pd 335 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "address 0x%x, data 0x%x\n", address, data); pd 336 drivers/net/ethernet/smsc/smsc9420.c ret = smsc9420_eeprom_send_cmd(pd, op); pd 340 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, E2P_DATA, (u32)data); pd 341 drivers/net/ethernet/smsc/smsc9420.c ret = smsc9420_eeprom_send_cmd(pd, op); pd 355 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 359 drivers/net/ethernet/smsc/smsc9420.c smsc9420_eeprom_enable_access(pd); pd 363 drivers/net/ethernet/smsc/smsc9420.c int ret = smsc9420_eeprom_read_location(pd, i, eeprom_data); pd 379 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 385 drivers/net/ethernet/smsc/smsc9420.c smsc9420_eeprom_enable_access(pd); pd 386 drivers/net/ethernet/smsc/smsc9420.c smsc9420_eeprom_send_cmd(pd, E2P_CMD_EPC_CMD_EWEN_); pd 387 drivers/net/ethernet/smsc/smsc9420.c ret = smsc9420_eeprom_write_location(pd, eeprom->offset, *data); pd 388 drivers/net/ethernet/smsc/smsc9420.c smsc9420_eeprom_send_cmd(pd, E2P_CMD_EPC_CMD_EWDS_); pd 415 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 421 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, ADDRH, mac_high16); pd 422 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, ADDRL, mac_low32); pd 427 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 432 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, probe, pd->dev, pd 437 drivers/net/ethernet/smsc/smsc9420.c u32 mac_high16 = smsc9420_reg_read(pd, ADDRH); pd 438 drivers/net/ethernet/smsc/smsc9420.c u32 mac_low32 = smsc9420_reg_read(pd, ADDRL); pd 448 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, probe, pd->dev, pd 454 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, probe, pd->dev, pd 460 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_stop_tx(struct smsc9420_pdata *pd) pd 466 drivers/net/ethernet/smsc/smsc9420.c dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL); pd 468 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control); pd 472 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_reg_read(pd, DMAC_STATUS) & DMAC_STS_TS_) pd 478 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifdown, pd->dev, "TX DMAC failed to stop\n"); pd 481 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_TXPS_); pd 484 drivers/net/ethernet/smsc/smsc9420.c dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA); pd 486 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena); pd 487 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 490 drivers/net/ethernet/smsc/smsc9420.c mac_cr = smsc9420_reg_read(pd, MAC_CR) & (~MAC_CR_TXEN_); pd 491 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, mac_cr); pd 492 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 495 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_free_tx_ring(struct smsc9420_pdata *pd) pd 499 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_ring); pd 501 drivers/net/ethernet/smsc/smsc9420.c if (!pd->tx_buffers) pd 505 drivers/net/ethernet/smsc/smsc9420.c struct sk_buff *skb = pd->tx_buffers[i].skb; pd 508 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_buffers[i].mapping); pd 509 drivers/net/ethernet/smsc/smsc9420.c pci_unmap_single(pd->pdev, pd->tx_buffers[i].mapping, pd 514 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].status = 0; pd 515 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].length = 0; pd 516 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].buffer1 = 0; pd 517 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].buffer2 = 0; pd 521 drivers/net/ethernet/smsc/smsc9420.c kfree(pd->tx_buffers); pd 522 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers = NULL; pd 524 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_head = 0; pd 525 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_tail = 0; pd 528 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_free_rx_ring(struct smsc9420_pdata *pd) pd 532 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->rx_ring); pd 534 drivers/net/ethernet/smsc/smsc9420.c if (!pd->rx_buffers) pd 538 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_buffers[i].skb) pd 539 drivers/net/ethernet/smsc/smsc9420.c dev_kfree_skb_any(pd->rx_buffers[i].skb); pd 541 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_buffers[i].mapping) pd 542 drivers/net/ethernet/smsc/smsc9420.c pci_unmap_single(pd->pdev, pd->rx_buffers[i].mapping, pd 545 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].status = 0; pd 546 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].length = 0; pd 547 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].buffer1 = 0; pd 548 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].buffer2 = 0; pd 552 drivers/net/ethernet/smsc/smsc9420.c kfree(pd->rx_buffers); pd 553 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers = NULL; pd 555 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_head = 0; pd 556 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_tail = 0; pd 559 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_stop_rx(struct smsc9420_pdata *pd) pd 565 drivers/net/ethernet/smsc/smsc9420.c dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA); pd 567 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena); pd 568 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 571 drivers/net/ethernet/smsc/smsc9420.c mac_cr = smsc9420_reg_read(pd, MAC_CR) & (~MAC_CR_RXEN_); pd 572 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, mac_cr); pd 573 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 576 drivers/net/ethernet/smsc/smsc9420.c dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL); pd 578 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control); pd 579 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 583 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_reg_read(pd, DMAC_STATUS) & DMAC_STS_RS_) pd 589 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifdown, pd->dev, pd 593 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_RXPS_); pd 598 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = dev_id; pd 603 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd); pd 604 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->ioaddr); pd 606 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG); pd 613 drivers/net/ethernet/smsc/smsc9420.c int_sts = smsc9420_reg_read(pd, INT_STAT); pd 616 drivers/net/ethernet/smsc/smsc9420.c u32 status = smsc9420_reg_read(pd, DMAC_STATUS); pd 621 drivers/net/ethernet/smsc/smsc9420.c netif_wake_queue(pd->dev); pd 626 drivers/net/ethernet/smsc/smsc9420.c u32 dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA); pd 628 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena); pd 629 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 632 drivers/net/ethernet/smsc/smsc9420.c napi_schedule(&pd->napi); pd 636 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_STATUS, ints_to_clear); pd 643 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 644 drivers/net/ethernet/smsc/smsc9420.c int_ctl = smsc9420_reg_read(pd, INT_CTL); pd 646 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CTL, int_ctl); pd 647 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 649 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_STAT, INT_STAT_SW_INT_); pd 650 drivers/net/ethernet/smsc/smsc9420.c pd->software_irq_signal = true; pd 657 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 665 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 666 drivers/net/ethernet/smsc/smsc9420.c const int irq = pd->pdev->irq; pd 674 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_dmac_soft_reset(struct smsc9420_pdata *pd) pd 676 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, BUS_MODE, BUS_MODE_SWR_); pd 677 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_read(pd, BUS_MODE); pd 679 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_reg_read(pd, BUS_MODE) & BUS_MODE_SWR_) pd 680 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, drv, pd->dev, "Software reset not cleared\n"); pd 685 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 689 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd); pd 693 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 694 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_); pd 695 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 696 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 699 drivers/net/ethernet/smsc/smsc9420.c napi_disable(&pd->napi); pd 701 drivers/net/ethernet/smsc/smsc9420.c smsc9420_stop_tx(pd); pd 702 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_tx_ring(pd); pd 704 drivers/net/ethernet/smsc/smsc9420.c smsc9420_stop_rx(pd); pd 705 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_rx_ring(pd); pd 707 drivers/net/ethernet/smsc/smsc9420.c free_irq(pd->pdev->irq, pd); pd 709 drivers/net/ethernet/smsc/smsc9420.c smsc9420_dmac_soft_reset(pd); pd 714 drivers/net/ethernet/smsc/smsc9420.c mdiobus_unregister(pd->mii_bus); pd 715 drivers/net/ethernet/smsc/smsc9420.c mdiobus_free(pd->mii_bus); pd 744 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_rx_handoff(struct smsc9420_pdata *pd, const int index, pd 747 drivers/net/ethernet/smsc/smsc9420.c struct net_device *dev = pd->dev; pd 755 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_csum) pd 761 drivers/net/ethernet/smsc/smsc9420.c pci_unmap_single(pd->pdev, pd->rx_buffers[index].mapping, pd 763 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[index].mapping = 0; pd 765 drivers/net/ethernet/smsc/smsc9420.c skb = pd->rx_buffers[index].skb; pd 766 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[index].skb = NULL; pd 768 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_csum) { pd 783 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_rx_buffer(struct smsc9420_pdata *pd, int index) pd 785 drivers/net/ethernet/smsc/smsc9420.c struct sk_buff *skb = netdev_alloc_skb(pd->dev, PKT_BUF_SZ); pd 788 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->rx_buffers[index].skb); pd 789 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->rx_buffers[index].mapping); pd 794 drivers/net/ethernet/smsc/smsc9420.c mapping = pci_map_single(pd->pdev, skb_tail_pointer(skb), pd 796 drivers/net/ethernet/smsc/smsc9420.c if (pci_dma_mapping_error(pd->pdev, mapping)) { pd 798 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, rx_err, pd->dev, "pci_map_single failed!\n"); pd 802 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[index].skb = skb; pd 803 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[index].mapping = mapping; pd 804 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[index].buffer1 = mapping + NET_IP_ALIGN; pd 805 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[index].status = RDES0_OWN_; pd 811 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_alloc_new_rx_buffers(struct smsc9420_pdata *pd) pd 813 drivers/net/ethernet/smsc/smsc9420.c while (pd->rx_ring_tail != pd->rx_ring_head) { pd 814 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_alloc_rx_buffer(pd, pd->rx_ring_tail)) pd 817 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_tail = (pd->rx_ring_tail + 1) % RX_RING_SIZE; pd 823 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = pd 825 drivers/net/ethernet/smsc/smsc9420.c struct net_device *dev = pd->dev; pd 831 drivers/net/ethernet/smsc/smsc9420.c status = pd->rx_ring[pd->rx_ring_head].status; pd 838 drivers/net/ethernet/smsc/smsc9420.c smsc9420_rx_handoff(pd, pd->rx_ring_head, status); pd 839 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_head = (pd->rx_ring_head + 1) % RX_RING_SIZE; pd 840 drivers/net/ethernet/smsc/smsc9420.c smsc9420_alloc_new_rx_buffers(pd); pd 843 drivers/net/ethernet/smsc/smsc9420.c drop_frame_cnt = smsc9420_reg_read(pd, MISS_FRAME_CNTR); pd 848 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, RX_POLL_DEMAND, 1); pd 849 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 852 drivers/net/ethernet/smsc/smsc9420.c napi_complete_done(&pd->napi, work_done); pd 855 drivers/net/ethernet/smsc/smsc9420.c dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA); pd 857 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena); pd 858 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 894 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 896 drivers/net/ethernet/smsc/smsc9420.c while (pd->tx_ring_tail != pd->tx_ring_head) { pd 897 drivers/net/ethernet/smsc/smsc9420.c int index = pd->tx_ring_tail; pd 901 drivers/net/ethernet/smsc/smsc9420.c status = pd->tx_ring[index].status; pd 902 drivers/net/ethernet/smsc/smsc9420.c length = pd->tx_ring[index].length; pd 910 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_buffers[index].skb); pd 911 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_buffers[index].mapping); pd 913 drivers/net/ethernet/smsc/smsc9420.c pci_unmap_single(pd->pdev, pd->tx_buffers[index].mapping, pd 914 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[index].skb->len, PCI_DMA_TODEVICE); pd 915 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[index].mapping = 0; pd 917 drivers/net/ethernet/smsc/smsc9420.c dev_kfree_skb_any(pd->tx_buffers[index].skb); pd 918 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[index].skb = NULL; pd 920 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[index].buffer1 = 0; pd 923 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_tail = (pd->tx_ring_tail + 1) % TX_RING_SIZE; pd 930 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 932 drivers/net/ethernet/smsc/smsc9420.c int index = pd->tx_ring_head; pd 935 drivers/net/ethernet/smsc/smsc9420.c (((pd->tx_ring_head + 2) % TX_RING_SIZE) == pd->tx_ring_tail); pd 940 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->tx_ring[index].status & TDES0_OWN_); pd 941 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->tx_buffers[index].skb); pd 942 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->tx_buffers[index].mapping); pd 944 drivers/net/ethernet/smsc/smsc9420.c mapping = pci_map_single(pd->pdev, skb->data, pd 946 drivers/net/ethernet/smsc/smsc9420.c if (pci_dma_mapping_error(pd->pdev, mapping)) { pd 947 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, tx_err, pd->dev, pd 952 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[index].skb = skb; pd 953 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[index].mapping = mapping; pd 958 drivers/net/ethernet/smsc/smsc9420.c netif_stop_queue(pd->dev); pd 965 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[index].buffer1 = mapping; pd 966 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[index].length = tmp_desc1; pd 970 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_head = (pd->tx_ring_head + 1) % TX_RING_SIZE; pd 973 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[index].status = TDES0_OWN_; pd 979 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, TX_POLL_DEMAND, 1); pd 980 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 987 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 988 drivers/net/ethernet/smsc/smsc9420.c u32 counter = smsc9420_reg_read(pd, MISS_FRAME_CNTR); pd 996 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 997 drivers/net/ethernet/smsc/smsc9420.c u32 mac_cr = smsc9420_reg_read(pd, MAC_CR); pd 1000 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "Promiscuous Mode Enabled\n"); pd 1005 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "Receive all Multicast Enabled\n"); pd 1013 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "Multicast filter enabled\n"); pd 1024 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, HASHH, hash_hi); pd 1025 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, HASHL, hash_lo); pd 1031 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, hw, pd->dev, "Receive own packets only\n"); pd 1032 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, HASHH, 0); pd 1033 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, HASHL, 0); pd 1040 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, mac_cr); pd 1041 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1044 drivers/net/ethernet/smsc/smsc9420.c static void smsc9420_phy_update_flowcontrol(struct smsc9420_pdata *pd) pd 1046 drivers/net/ethernet/smsc/smsc9420.c struct net_device *dev = pd->dev; pd 1060 drivers/net/ethernet/smsc/smsc9420.c netif_info(pd, link, pd->dev, "rx pause %s, tx pause %s\n", pd 1064 drivers/net/ethernet/smsc/smsc9420.c netif_info(pd, link, pd->dev, "half duplex\n"); pd 1068 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, FLOW, flow); pd 1075 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1079 drivers/net/ethernet/smsc/smsc9420.c if (phy_dev->duplex != pd->last_duplex) { pd 1080 drivers/net/ethernet/smsc/smsc9420.c u32 mac_cr = smsc9420_reg_read(pd, MAC_CR); pd 1082 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, link, pd->dev, "full duplex mode\n"); pd 1085 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, link, pd->dev, "half duplex mode\n"); pd 1088 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, mac_cr); pd 1090 drivers/net/ethernet/smsc/smsc9420.c smsc9420_phy_update_flowcontrol(pd); pd 1091 drivers/net/ethernet/smsc/smsc9420.c pd->last_duplex = phy_dev->duplex; pd 1095 drivers/net/ethernet/smsc/smsc9420.c if (carrier != pd->last_carrier) { pd 1097 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, link, pd->dev, "carrier OK\n"); pd 1099 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, link, pd->dev, "no carrier\n"); pd 1100 drivers/net/ethernet/smsc/smsc9420.c pd->last_carrier = carrier; pd 1106 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1112 drivers/net/ethernet/smsc/smsc9420.c phydev = mdiobus_get_phy(pd->mii_bus, 1); pd 1133 drivers/net/ethernet/smsc/smsc9420.c pd->last_duplex = -1; pd 1134 drivers/net/ethernet/smsc/smsc9420.c pd->last_carrier = -1; pd 1141 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1144 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus = mdiobus_alloc(); pd 1145 drivers/net/ethernet/smsc/smsc9420.c if (!pd->mii_bus) { pd 1149 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus->name = DRV_MDIONAME; pd 1150 drivers/net/ethernet/smsc/smsc9420.c snprintf(pd->mii_bus->id, MII_BUS_ID_SIZE, "%x", pd 1151 drivers/net/ethernet/smsc/smsc9420.c (pd->pdev->bus->number << 8) | pd->pdev->devfn); pd 1152 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus->priv = pd; pd 1153 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus->read = smsc9420_mii_read; pd 1154 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus->write = smsc9420_mii_write; pd 1157 drivers/net/ethernet/smsc/smsc9420.c pd->mii_bus->phy_mask = ~(1 << 1); pd 1159 drivers/net/ethernet/smsc/smsc9420.c if (mdiobus_register(pd->mii_bus)) { pd 1160 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, probe, pd->dev, "Error registering mii bus\n"); pd 1165 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, probe, pd->dev, "Error probing mii bus\n"); pd 1172 drivers/net/ethernet/smsc/smsc9420.c mdiobus_unregister(pd->mii_bus); pd 1174 drivers/net/ethernet/smsc/smsc9420.c mdiobus_free(pd->mii_bus); pd 1179 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_tx_ring(struct smsc9420_pdata *pd) pd 1183 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_ring); pd 1185 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers = kmalloc_array(TX_RING_SIZE, pd 1188 drivers/net/ethernet/smsc/smsc9420.c if (!pd->tx_buffers) pd 1193 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[i].skb = NULL; pd 1194 drivers/net/ethernet/smsc/smsc9420.c pd->tx_buffers[i].mapping = 0; pd 1195 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].status = 0; pd 1196 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].length = 0; pd 1197 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].buffer1 = 0; pd 1198 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[i].buffer2 = 0; pd 1200 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring[TX_RING_SIZE - 1].length = TDES1_TER_; pd 1203 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_head = 0; pd 1204 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring_tail = 0; pd 1206 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, TX_BASE_ADDR, pd->tx_dma_addr); pd 1207 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1212 drivers/net/ethernet/smsc/smsc9420.c static int smsc9420_alloc_rx_ring(struct smsc9420_pdata *pd) pd 1216 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->rx_ring); pd 1218 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers = kmalloc_array(RX_RING_SIZE, pd 1221 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_buffers == NULL) pd 1226 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].status = 0; pd 1227 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].length = PKT_BUF_SZ; pd 1228 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[i].buffer2 = 0; pd 1229 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[i].skb = NULL; pd 1230 drivers/net/ethernet/smsc/smsc9420.c pd->rx_buffers[i].mapping = 0; pd 1232 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring[RX_RING_SIZE - 1].length = (PKT_BUF_SZ | RDES1_RER_); pd 1236 drivers/net/ethernet/smsc/smsc9420.c if (smsc9420_alloc_rx_buffer(pd, i)) { pd 1237 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, pd 1243 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_head = 0; pd 1244 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring_tail = 0; pd 1246 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, VLAN1, ETH_P_8021Q); pd 1247 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, ifup, pd->dev, "VLAN1 = 0x%08x\n", pd 1248 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_read(pd, VLAN1)); pd 1250 drivers/net/ethernet/smsc/smsc9420.c if (pd->rx_csum) { pd 1252 drivers/net/ethernet/smsc/smsc9420.c u32 coe = smsc9420_reg_read(pd, COE_CR) | RX_COE_EN; pd 1253 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, COE_CR, coe); pd 1254 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, ifup, pd->dev, "COE_CR = 0x%08x\n", coe); pd 1257 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, RX_BASE_ADDR, pd->rx_dma_addr); pd 1258 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1263 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_rx_ring(pd); pd 1270 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1272 drivers/net/ethernet/smsc/smsc9420.c const int irq = pd->pdev->irq; pd 1277 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, pd 1286 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 1287 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_); pd 1288 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 1289 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CTL, 0); pd 1290 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 1291 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, 0); pd 1292 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_STAT, 0xFFFFFFFF); pd 1293 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1295 drivers/net/ethernet/smsc/smsc9420.c result = request_irq(irq, smsc9420_isr, IRQF_SHARED, DRV_NAME, pd); pd 1297 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, "Unable to use IRQ = %d\n", irq); pd 1302 drivers/net/ethernet/smsc/smsc9420.c smsc9420_dmac_soft_reset(pd); pd 1305 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, 0); pd 1310 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, GPIO_CFG, pd 1319 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, BUS_MODE, bus_mode); pd 1321 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1324 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, BUS_CFG, BUS_CFG_RXTXWEIGHT_4_1); pd 1326 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_CONTROL, pd 1329 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1332 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, ifup, pd->dev, "Testing ISR using IRQ %d\n", irq); pd 1333 drivers/net/ethernet/smsc/smsc9420.c pd->software_irq_signal = false; pd 1335 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 1337 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) | INT_CFG_IRQ_EN_; pd 1340 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 1343 drivers/net/ethernet/smsc/smsc9420.c int_ctl = smsc9420_reg_read(pd, INT_CTL) | INT_CTL_SW_INT_EN_; pd 1344 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CTL, int_ctl); pd 1345 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 1346 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1350 drivers/net/ethernet/smsc/smsc9420.c if (pd->software_irq_signal) pd 1356 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 1357 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_); pd 1358 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 1359 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 1361 drivers/net/ethernet/smsc/smsc9420.c if (!pd->software_irq_signal) { pd 1362 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, "ISR failed signaling test\n"); pd 1367 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, ifup, pd->dev, "ISR passed test using IRQ %d\n", irq); pd 1369 drivers/net/ethernet/smsc/smsc9420.c result = smsc9420_alloc_tx_ring(pd); pd 1371 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, pd 1377 drivers/net/ethernet/smsc/smsc9420.c result = smsc9420_alloc_rx_ring(pd); pd 1379 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, pd 1387 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, "Failed to initialize Phy\n"); pd 1395 drivers/net/ethernet/smsc/smsc9420.c napi_enable(&pd->napi); pd 1398 drivers/net/ethernet/smsc/smsc9420.c mac_cr = smsc9420_reg_read(pd, MAC_CR) | MAC_CR_TXEN_ | MAC_CR_RXEN_; pd 1399 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, MAC_CR, mac_cr); pd 1401 drivers/net/ethernet/smsc/smsc9420.c dmac_control = smsc9420_reg_read(pd, DMAC_CONTROL); pd 1403 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_CONTROL, dmac_control); pd 1404 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1406 drivers/net/ethernet/smsc/smsc9420.c dma_intr_ena = smsc9420_reg_read(pd, DMAC_INTR_ENA); pd 1409 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, DMAC_INTR_ENA, dma_intr_ena); pd 1410 drivers/net/ethernet/smsc/smsc9420.c smsc9420_pci_flush_write(pd); pd 1414 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, RX_POLL_DEMAND, 1); pd 1417 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 1418 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) | INT_CFG_IRQ_EN_; pd 1419 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 1420 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 1425 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_rx_ring(pd); pd 1427 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_tx_ring(pd); pd 1429 drivers/net/ethernet/smsc/smsc9420.c free_irq(irq, pd); pd 1439 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1444 drivers/net/ethernet/smsc/smsc9420.c spin_lock_irqsave(&pd->int_lock, flags); pd 1445 drivers/net/ethernet/smsc/smsc9420.c int_cfg = smsc9420_reg_read(pd, INT_CFG) & (~INT_CFG_IRQ_EN_); pd 1446 drivers/net/ethernet/smsc/smsc9420.c smsc9420_reg_write(pd, INT_CFG, int_cfg); pd 1447 drivers/net/ethernet/smsc/smsc9420.c spin_unlock_irqrestore(&pd->int_lock, flags); pd 1451 drivers/net/ethernet/smsc/smsc9420.c smsc9420_stop_tx(pd); pd 1452 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_tx_ring(pd); pd 1454 drivers/net/ethernet/smsc/smsc9420.c napi_disable(&pd->napi); pd 1455 drivers/net/ethernet/smsc/smsc9420.c smsc9420_stop_rx(pd); pd 1456 drivers/net/ethernet/smsc/smsc9420.c smsc9420_free_rx_ring(pd); pd 1458 drivers/net/ethernet/smsc/smsc9420.c free_irq(pd->pdev->irq, pd); pd 1474 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd = netdev_priv(dev); pd 1488 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, ifup, pd->dev, "pci_enable_wake failed: %d\n", pd 1519 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd; pd 1535 drivers/net/ethernet/smsc/smsc9420.c dev = alloc_etherdev(sizeof(*pd)); pd 1566 drivers/net/ethernet/smsc/smsc9420.c pd = netdev_priv(dev); pd 1569 drivers/net/ethernet/smsc/smsc9420.c pd->rx_ring = pci_alloc_consistent(pdev, pd 1572 drivers/net/ethernet/smsc/smsc9420.c &pd->rx_dma_addr); pd 1574 drivers/net/ethernet/smsc/smsc9420.c if (!pd->rx_ring) pd 1578 drivers/net/ethernet/smsc/smsc9420.c pd->tx_ring = (pd->rx_ring + RX_RING_SIZE); pd 1579 drivers/net/ethernet/smsc/smsc9420.c pd->tx_dma_addr = pd->rx_dma_addr + pd 1582 drivers/net/ethernet/smsc/smsc9420.c pd->pdev = pdev; pd 1583 drivers/net/ethernet/smsc/smsc9420.c pd->dev = dev; pd 1584 drivers/net/ethernet/smsc/smsc9420.c pd->ioaddr = virt_addr; pd 1585 drivers/net/ethernet/smsc/smsc9420.c pd->msg_enable = smsc_debug; pd 1586 drivers/net/ethernet/smsc/smsc9420.c pd->rx_csum = true; pd 1588 drivers/net/ethernet/smsc/smsc9420.c netif_dbg(pd, probe, pd->dev, "lan_base=0x%08lx\n", (ulong)virt_addr); pd 1590 drivers/net/ethernet/smsc/smsc9420.c id_rev = smsc9420_reg_read(pd, ID_REV); pd 1593 drivers/net/ethernet/smsc/smsc9420.c netif_info(pd, probe, pd->dev, pd 1597 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, probe, pd->dev, "LAN9420 NOT identified\n"); pd 1598 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, probe, pd->dev, "ID_REV=0x%08X\n", id_rev); pd 1602 drivers/net/ethernet/smsc/smsc9420.c smsc9420_dmac_soft_reset(pd); pd 1603 drivers/net/ethernet/smsc/smsc9420.c smsc9420_eeprom_reload(pd); pd 1609 drivers/net/ethernet/smsc/smsc9420.c netif_napi_add(dev, &pd->napi, smsc9420_rx_poll, NAPI_WEIGHT); pd 1613 drivers/net/ethernet/smsc/smsc9420.c netif_warn(pd, probe, pd->dev, "error %i registering device\n", pd 1620 drivers/net/ethernet/smsc/smsc9420.c spin_lock_init(&pd->int_lock); pd 1621 drivers/net/ethernet/smsc/smsc9420.c spin_lock_init(&pd->phy_lock); pd 1629 drivers/net/ethernet/smsc/smsc9420.c (RX_RING_SIZE + TX_RING_SIZE), pd->rx_ring, pd->rx_dma_addr); pd 1645 drivers/net/ethernet/smsc/smsc9420.c struct smsc9420_pdata *pd; pd 1651 drivers/net/ethernet/smsc/smsc9420.c pd = netdev_priv(dev); pd 1655 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->tx_buffers); pd 1656 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(pd->rx_buffers); pd 1658 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->tx_ring); pd 1659 drivers/net/ethernet/smsc/smsc9420.c BUG_ON(!pd->rx_ring); pd 1662 drivers/net/ethernet/smsc/smsc9420.c (RX_RING_SIZE + TX_RING_SIZE), pd->rx_ring, pd->rx_dma_addr); pd 1664 drivers/net/ethernet/smsc/smsc9420.c iounmap(pd->ioaddr - LAN9420_CPSR_ENDIAN_OFFSET); pd 708 drivers/net/ethernet/toshiba/tc35815.c struct device *pd = bus_find_device(&platform_bus_type, NULL, pd 710 drivers/net/ethernet/toshiba/tc35815.c if (pd) { pd 711 drivers/net/ethernet/toshiba/tc35815.c if (pd->platform_data) pd 712 drivers/net/ethernet/toshiba/tc35815.c memcpy(dev->dev_addr, pd->platform_data, ETH_ALEN); pd 713 drivers/net/ethernet/toshiba/tc35815.c put_device(pd); pd 1825 drivers/net/fddi/skfp/smt.c const struct smt_pdef *pd ; pd 1856 drivers/net/fddi/skfp/smt.c for (i = N_SMT_PLEN, pd = smt_pdef; i ; i--,pd++) { pd 1857 drivers/net/fddi/skfp/smt.c if (pd->ptype == type) pd 1860 drivers/net/fddi/skfp/smt.c if (i && pd->pswap) { pd 1861 drivers/net/fddi/skfp/smt.c smt_string_swap(p+PARA_LEN,pd->pswap,len) ; pd 169 drivers/net/phy/dp83640.c static void dp83640_gpio_defaults(struct ptp_pin_desc *pd) pd 174 drivers/net/phy/dp83640.c snprintf(pd[i].name, sizeof(pd[i].name), "GPIO%d", 1 + i); pd 175 drivers/net/phy/dp83640.c pd[i].index = i; pd 186 drivers/net/phy/dp83640.c pd[index].func = PTP_PF_PHYSYNC; pd 187 drivers/net/phy/dp83640.c pd[index].chan = 0; pd 190 drivers/net/phy/dp83640.c pd[index].func = PTP_PF_PEROUT; pd 191 drivers/net/phy/dp83640.c pd[index].chan = 0; pd 195 drivers/net/phy/dp83640.c pd[index].func = PTP_PF_EXTTS; pd 196 drivers/net/phy/dp83640.c pd[index].chan = i - EXTTS0_GPIO; pd 698 drivers/net/wireless/ath/ath5k/eeprom.c struct ath5k_pdgain_info *pd = pd 701 drivers/net/wireless/ath/ath5k/eeprom.c kfree(pd->pd_step); pd 702 drivers/net/wireless/ath/ath5k/eeprom.c kfree(pd->pd_pwr); pd 719 drivers/net/wireless/ath/ath5k/eeprom.c struct ath5k_pdgain_info *pd; pd 751 drivers/net/wireless/ath/ath5k/eeprom.c pd = &chinfo[pier].pd_curves[idx]; pd 753 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_points = AR5K_EEPROM_N_PWR_POINTS_5111; pd 756 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step = kcalloc(AR5K_EEPROM_N_PWR_POINTS_5111, pd 758 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_step) pd 761 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr = kcalloc(AR5K_EEPROM_N_PWR_POINTS_5111, pd 763 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_pwr) pd 769 drivers/net/wireless/ath/ath5k/eeprom.c for (point = 0; point < pd->pd_points; point++) { pd 772 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[point] = 2 * pcinfo->pwr[point]; pd 775 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[point] = pcinfo->pcdac[point]; pd 779 drivers/net/wireless/ath/ath5k/eeprom.c chinfo[pier].min_pwr = pd->pd_pwr[0]; pd 780 drivers/net/wireless/ath/ath5k/eeprom.c chinfo[pier].max_pwr = pd->pd_pwr[10]; pd 928 drivers/net/wireless/ath/ath5k/eeprom.c struct ath5k_pdgain_info *pd = pd 934 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_points = AR5K_EEPROM_N_XPD0_POINTS; pd 937 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step = kcalloc(pd->pd_points, pd 940 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_step) pd 943 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr = kcalloc(pd->pd_points, pd 946 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_pwr) pd 951 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[0] = pcinfo->pcdac_x0[0]; pd 952 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[0] = pcinfo->pwr_x0[0]; pd 954 drivers/net/wireless/ath/ath5k/eeprom.c for (point = 1; point < pd->pd_points; pd 957 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[point] = pd 961 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[point] = pd 962 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[point - 1] + pd 967 drivers/net/wireless/ath/ath5k/eeprom.c chinfo[pier].min_pwr = pd->pd_pwr[0]; pd 972 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_points = AR5K_EEPROM_N_XPD3_POINTS; pd 975 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step = kcalloc(pd->pd_points, pd 978 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_step) pd 981 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr = kcalloc(pd->pd_points, pd 984 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_pwr) pd 989 drivers/net/wireless/ath/ath5k/eeprom.c for (point = 0; point < pd->pd_points; pd 992 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[point] = pd 996 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[point] = pd 1002 drivers/net/wireless/ath/ath5k/eeprom.c chinfo[pier].min_pwr = pd->pd_pwr[0]; pd 1220 drivers/net/wireless/ath/ath5k/eeprom.c struct ath5k_pdgain_info *pd = pd 1226 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_points = AR5K_EEPROM_N_PD_POINTS; pd 1228 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_points = AR5K_EEPROM_N_PD_POINTS - 1; pd 1231 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step = kcalloc(pd->pd_points, pd 1234 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_step) pd 1237 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr = kcalloc(pd->pd_points, pd 1240 drivers/net/wireless/ath/ath5k/eeprom.c if (!pd->pd_pwr) pd 1246 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[0] = pcinfo->pddac_i[pdg]; pd 1247 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[0] = 4 * pcinfo->pwr_i[pdg]; pd 1249 drivers/net/wireless/ath/ath5k/eeprom.c for (point = 1; point < pd->pd_points; point++) { pd 1251 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[point] = pd->pd_pwr[point - 1] + pd 1254 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_step[point] = pd->pd_step[point - 1] + pd 1261 drivers/net/wireless/ath/ath5k/eeprom.c chinfo[pier].min_pwr = pd->pd_pwr[0]; pd 1266 drivers/net/wireless/ath/ath5k/eeprom.c pd->pd_pwr[pd->pd_points - 1]; pd 276 drivers/net/wireless/ath/ath9k/dfs.c struct dfs_pattern_detector *pd = sc->dfs_detector; pd 278 drivers/net/wireless/ath/ath9k/dfs.c if (pd == NULL) pd 280 drivers/net/wireless/ath/ath9k/dfs.c if (!pd->add_pulse(pd, pe, NULL)) pd 295 drivers/net/wireless/ath/dfs_pattern_detector.c struct pri_detector *pd = cd->detectors[i]; pd 296 drivers/net/wireless/ath/dfs_pattern_detector.c struct pri_sequence *ps = pd->add_pulse(pd, event); pd 299 drivers/net/wireless/ath/dfs_pattern_detector.c memcpy(rs, pd->rs, sizeof(*rs)); pd 303 drivers/net/wireless/ath/dfs_pattern_detector.c event->freq, pd->rs->type_id, pd 305 drivers/net/wireless/ath/dfs_pattern_detector.c pd->reset(pd, dpd->last_pulse_ts); pd 101 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = drvr->proto->pd; pd 110 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd; pd 141 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd; pd 159 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd; pd 217 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = (struct brcmf_bcdc *)drvr->proto->pd; pd 335 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = drvr->proto->pd; pd 366 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = bus_if->drvr->proto->pd; pd 426 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = drvr->proto->pd; pd 470 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c drvr->proto->pd = bcdc; pd 484 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c struct brcmf_bcdc *bcdc = drvr->proto->pd; pd 486 drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcdc.c drvr->proto->pd = NULL; pd 429 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 490 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 798 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 823 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 832 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 841 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 1364 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 1394 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 1446 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c struct brcmf_msgbuf *msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 1555 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c drvr->proto->pd = msgbuf; pd 1634 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c if (drvr->proto->pd) { pd 1635 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c msgbuf = (struct brcmf_msgbuf *)drvr->proto->pd; pd 1656 drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c drvr->proto->pd = NULL; pd 41 drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h void *pd; pd 3139 drivers/net/wireless/realtek/rtw88/rtw8822c.c u32 pd, cs; pd 3144 drivers/net/wireless/realtek/rtw88/rtw8822c.c pd = rtw_read32_mask(rtwdev, pd 3150 drivers/net/wireless/realtek/rtw88/rtw8822c.c pd += pd_diff; pd 3152 drivers/net/wireless/realtek/rtw88/rtw8822c.c if (pd > RTW_CCK_PD_MAX) pd 3153 drivers/net/wireless/realtek/rtw88/rtw8822c.c pd = RTW_CCK_PD_MAX; pd 3161 drivers/net/wireless/realtek/rtw88/rtw8822c.c pd); pd 334 drivers/net/wireless/ti/wl1251/acx.c struct acx_packet_detection *pd; pd 339 drivers/net/wireless/ti/wl1251/acx.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 340 drivers/net/wireless/ti/wl1251/acx.c if (!pd) pd 345 drivers/net/wireless/ti/wl1251/acx.c ret = wl1251_cmd_configure(wl, ACX_PD_THRESHOLD, pd, sizeof(*pd)); pd 352 drivers/net/wireless/ti/wl1251/acx.c kfree(pd); pd 39 drivers/nvme/host/rdma.c struct ib_pd *pd; pd 273 drivers/nvme/host/rdma.c ret = rdma_create_qp(queue->cm_id, dev->pd, &init_attr); pd 339 drivers/nvme/host/rdma.c ib_dealloc_pd(ndev->pd); pd 372 drivers/nvme/host/rdma.c ndev->pd = ib_alloc_pd(ndev->dev, pd 374 drivers/nvme/host/rdma.c if (IS_ERR(ndev->pd)) pd 392 drivers/nvme/host/rdma.c ib_dealloc_pd(ndev->pd); pd 1194 drivers/nvme/host/rdma.c sge->lkey = queue->device->pd->local_dma_lkey; pd 1213 drivers/nvme/host/rdma.c put_unaligned_le32(queue->device->pd->unsafe_global_rkey, sg->key); pd 1305 drivers/nvme/host/rdma.c if (count == 1 && dev->pd->flags & IB_PD_UNSAFE_GLOBAL_RKEY) { pd 1351 drivers/nvme/host/rdma.c sge->lkey = queue->device->pd->local_dma_lkey; pd 1382 drivers/nvme/host/rdma.c list.lkey = queue->device->pd->local_dma_lkey; pd 107 drivers/nvme/target/rdma.c struct ib_pd *pd; pd 263 drivers/nvme/target/rdma.c sge->lkey = ndev->pd->local_dma_lkey; pd 293 drivers/nvme/target/rdma.c c->sge[0].lkey = ndev->pd->local_dma_lkey; pd 378 drivers/nvme/target/rdma.c r->send_sge.lkey = ndev->pd->local_dma_lkey; pd 865 drivers/nvme/target/rdma.c srq = ib_create_srq(ndev->pd, &srq_attr); pd 909 drivers/nvme/target/rdma.c ib_dealloc_pd(ndev->pd); pd 949 drivers/nvme/target/rdma.c ndev->pd = ib_alloc_pd(ndev->device, 0); pd 950 drivers/nvme/target/rdma.c if (IS_ERR(ndev->pd)) pd 966 drivers/nvme/target/rdma.c ib_dealloc_pd(ndev->pd); pd 1023 drivers/nvme/target/rdma.c ret = rdma_create_qp(queue->cm_id, ndev->pd, &qp_attr); pd 1317 drivers/parport/share.c struct pardevice *pd; pd 1354 drivers/parport/share.c for (pd = port->waithead; pd; pd = pd->waitnext) { pd 1355 drivers/parport/share.c if (pd->waiting & 2) { /* sleeping in claim_or_block */ pd 1356 drivers/parport/share.c parport_claim(pd); pd 1357 drivers/parport/share.c if (waitqueue_active(&pd->wait_q)) pd 1358 drivers/parport/share.c wake_up_interruptible(&pd->wait_q); pd 1360 drivers/parport/share.c } else if (pd->wakeup) { pd 1361 drivers/parport/share.c pd->wakeup(pd->private); pd 1365 drivers/parport/share.c printk(KERN_ERR "%s: don't know how to wake %s\n", port->name, pd->name); pd 1374 drivers/parport/share.c for (pd = port->devices; !port->cad && pd; pd = pd->next) { pd 1375 drivers/parport/share.c if (pd->wakeup && pd != dev) pd 1376 drivers/parport/share.c pd->wakeup(pd->private); pd 240 drivers/pcmcia/bcm63xx_pcmcia.c if (gpio_get_value(skt->pd->ready_gpio)) pd 346 drivers/pcmcia/bcm63xx_pcmcia.c skt->pd = pdev->dev.platform_data; pd 347 drivers/pcmcia/bcm63xx_pcmcia.c if (!skt->common_res || !skt->attr_res || !irq_res || !skt->pd) { pd 26 drivers/pcmcia/bcm63xx_pcmcia.h struct bcm63xx_pcmcia_platform_data *pd; pd 538 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pd-entry-vld, 0x07), pd 563 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(rank-pd, 0x20), pd 1502 drivers/pinctrl/intel/pinctrl-intel.c const struct pin_desc *pd = pin_desc_get(pctrl->pctldev, pin); pd 1504 drivers/pinctrl/intel/pinctrl-intel.c if (!pd || !intel_pad_usable(pctrl, pin)) pd 1513 drivers/pinctrl/intel/pinctrl-intel.c if (pd->mux_owner || pd->gpio_owner || pd 457 drivers/pinctrl/nomadik/pinctrl-abx500.c bool pd; pd 485 drivers/pinctrl/nomadik/pinctrl-abx500.c gpio_offset, &pd); pd 489 drivers/pinctrl/nomadik/pinctrl-abx500.c seq_printf(s, " %-9s", pull_up_down[pd]); pd 1724 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c u32 ie, oe, pu, pd; pd 1732 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c pd = ioread32(bank->base + NPCM7XX_GP_N_PD) & pinmask; pd 1734 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c rc = (!pu && !pd); pd 1736 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c rc = (pu && !pd); pd 1738 drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c rc = (!pu && pd); pd 783 drivers/pinctrl/pinctrl-amd.c const struct pin_desc *pd = pin_desc_get(gpio_dev->pctrl, pin); pd 785 drivers/pinctrl/pinctrl-amd.c if (!pd) pd 792 drivers/pinctrl/pinctrl-amd.c if (pd->mux_owner || pd->gpio_owner || pd 432 drivers/pinctrl/ti/pinctrl-ti-iodelay.c struct pinctrl_pin_desc *pd; pd 461 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pd = &iod->pa[pin]; pd 462 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pd->drv_data = &cfg[pin_index]; pd 656 drivers/pinctrl/ti/pinctrl-ti-iodelay.c struct pinctrl_pin_desc *pd; pd 666 drivers/pinctrl/ti/pinctrl-ti-iodelay.c pd = &iod->pa[pin]; pd 667 drivers/pinctrl/ti/pinctrl-ti-iodelay.c cfg = pd->drv_data; pd 164 drivers/platform/chrome/cros_ec.c ec_dev->pd = platform_device_register_data(ec_dev->dev, pd 168 drivers/platform/chrome/cros_ec.c if (IS_ERR(ec_dev->pd)) { pd 172 drivers/platform/chrome/cros_ec.c return PTR_ERR(ec_dev->pd); pd 179 drivers/platform/chrome/cros_ec.c platform_device_unregister(ec_dev->pd); pd 203 drivers/platform/chrome/cros_ec.c if (ec_dev->pd) pd 204 drivers/platform/chrome/cros_ec.c platform_device_unregister(ec_dev->pd); pd 422 drivers/platform/chrome/cros_ec_debugfs.c static int cros_ec_debugfs_probe(struct platform_device *pd) pd 424 drivers/platform/chrome/cros_ec_debugfs.c struct cros_ec_dev *ec = dev_get_drvdata(pd->dev.parent); pd 456 drivers/platform/chrome/cros_ec_debugfs.c dev_set_drvdata(&pd->dev, ec); pd 465 drivers/platform/chrome/cros_ec_debugfs.c static int cros_ec_debugfs_remove(struct platform_device *pd) pd 467 drivers/platform/chrome/cros_ec_debugfs.c struct cros_ec_dev *ec = dev_get_drvdata(pd->dev.parent); pd 556 drivers/platform/chrome/cros_ec_lightbar.c static int cros_ec_lightbar_probe(struct platform_device *pd) pd 558 drivers/platform/chrome/cros_ec_lightbar.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 560 drivers/platform/chrome/cros_ec_lightbar.c struct device *dev = &pd->dev; pd 589 drivers/platform/chrome/cros_ec_lightbar.c static int cros_ec_lightbar_remove(struct platform_device *pd) pd 591 drivers/platform/chrome/cros_ec_lightbar.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 344 drivers/platform/chrome/cros_ec_sysfs.c static int cros_ec_sysfs_probe(struct platform_device *pd) pd 346 drivers/platform/chrome/cros_ec_sysfs.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 347 drivers/platform/chrome/cros_ec_sysfs.c struct device *dev = &pd->dev; pd 357 drivers/platform/chrome/cros_ec_sysfs.c static int cros_ec_sysfs_remove(struct platform_device *pd) pd 359 drivers/platform/chrome/cros_ec_sysfs.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 110 drivers/platform/chrome/cros_ec_vbc.c static int cros_ec_vbc_probe(struct platform_device *pd) pd 112 drivers/platform/chrome/cros_ec_vbc.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 113 drivers/platform/chrome/cros_ec_vbc.c struct device *dev = &pd->dev; pd 125 drivers/platform/chrome/cros_ec_vbc.c static int cros_ec_vbc_remove(struct platform_device *pd) pd 127 drivers/platform/chrome/cros_ec_vbc.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 195 drivers/platform/chrome/cros_usbpd_logger.c static int cros_usbpd_logger_probe(struct platform_device *pd) pd 197 drivers/platform/chrome/cros_usbpd_logger.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 198 drivers/platform/chrome/cros_usbpd_logger.c struct device *dev = &pd->dev; pd 208 drivers/platform/chrome/cros_usbpd_logger.c platform_set_drvdata(pd, logger); pd 222 drivers/platform/chrome/cros_usbpd_logger.c static int cros_usbpd_logger_remove(struct platform_device *pd) pd 224 drivers/platform/chrome/cros_usbpd_logger.c struct logger_data *logger = platform_get_drvdata(pd); pd 128 drivers/platform/x86/sony-laptop.c static int sony_nc_kbd_backlight_setup(struct platform_device *pd, pd 130 drivers/platform/x86/sony-laptop.c static void sony_nc_kbd_backlight_cleanup(struct platform_device *pd, pd 133 drivers/platform/x86/sony-laptop.c static int sony_nc_battery_care_setup(struct platform_device *pd, pd 135 drivers/platform/x86/sony-laptop.c static void sony_nc_battery_care_cleanup(struct platform_device *pd); pd 137 drivers/platform/x86/sony-laptop.c static int sony_nc_thermal_setup(struct platform_device *pd); pd 138 drivers/platform/x86/sony-laptop.c static void sony_nc_thermal_cleanup(struct platform_device *pd); pd 140 drivers/platform/x86/sony-laptop.c static int sony_nc_lid_resume_setup(struct platform_device *pd, pd 142 drivers/platform/x86/sony-laptop.c static void sony_nc_lid_resume_cleanup(struct platform_device *pd); pd 144 drivers/platform/x86/sony-laptop.c static int sony_nc_gfx_switch_setup(struct platform_device *pd, pd 146 drivers/platform/x86/sony-laptop.c static void sony_nc_gfx_switch_cleanup(struct platform_device *pd); pd 149 drivers/platform/x86/sony-laptop.c static int sony_nc_highspeed_charging_setup(struct platform_device *pd); pd 150 drivers/platform/x86/sony-laptop.c static void sony_nc_highspeed_charging_cleanup(struct platform_device *pd); pd 152 drivers/platform/x86/sony-laptop.c static int sony_nc_lowbatt_setup(struct platform_device *pd); pd 153 drivers/platform/x86/sony-laptop.c static void sony_nc_lowbatt_cleanup(struct platform_device *pd); pd 155 drivers/platform/x86/sony-laptop.c static int sony_nc_fanspeed_setup(struct platform_device *pd); pd 156 drivers/platform/x86/sony-laptop.c static void sony_nc_fanspeed_cleanup(struct platform_device *pd); pd 158 drivers/platform/x86/sony-laptop.c static int sony_nc_usb_charge_setup(struct platform_device *pd); pd 159 drivers/platform/x86/sony-laptop.c static void sony_nc_usb_charge_cleanup(struct platform_device *pd); pd 161 drivers/platform/x86/sony-laptop.c static int sony_nc_panelid_setup(struct platform_device *pd); pd 162 drivers/platform/x86/sony-laptop.c static void sony_nc_panelid_cleanup(struct platform_device *pd); pd 164 drivers/platform/x86/sony-laptop.c static int sony_nc_smart_conn_setup(struct platform_device *pd); pd 165 drivers/platform/x86/sony-laptop.c static void sony_nc_smart_conn_cleanup(struct platform_device *pd); pd 167 drivers/platform/x86/sony-laptop.c static int sony_nc_touchpad_setup(struct platform_device *pd, pd 169 drivers/platform/x86/sony-laptop.c static void sony_nc_touchpad_cleanup(struct platform_device *pd); pd 838 drivers/platform/x86/sony-laptop.c static int sony_nc_handles_setup(struct platform_device *pd) pd 864 drivers/platform/x86/sony-laptop.c if (device_create_file(&pd->dev, &handles->devattr)) { pd 874 drivers/platform/x86/sony-laptop.c static int sony_nc_handles_cleanup(struct platform_device *pd) pd 878 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &handles->devattr); pd 1431 drivers/platform/x86/sony-laptop.c static void sony_nc_function_cleanup(struct platform_device *pd) pd 1453 drivers/platform/x86/sony-laptop.c sony_nc_touchpad_cleanup(pd); pd 1458 drivers/platform/x86/sony-laptop.c sony_nc_battery_care_cleanup(pd); pd 1462 drivers/platform/x86/sony-laptop.c sony_nc_lid_resume_cleanup(pd); pd 1465 drivers/platform/x86/sony-laptop.c sony_nc_thermal_cleanup(pd); pd 1470 drivers/platform/x86/sony-laptop.c sony_nc_gfx_switch_cleanup(pd); pd 1473 drivers/platform/x86/sony-laptop.c sony_nc_highspeed_charging_cleanup(pd); pd 1485 drivers/platform/x86/sony-laptop.c sony_nc_kbd_backlight_cleanup(pd, handle); pd 1488 drivers/platform/x86/sony-laptop.c sony_nc_lowbatt_cleanup(pd); pd 1491 drivers/platform/x86/sony-laptop.c sony_nc_fanspeed_cleanup(pd); pd 1494 drivers/platform/x86/sony-laptop.c sony_nc_usb_charge_cleanup(pd); pd 1497 drivers/platform/x86/sony-laptop.c sony_nc_panelid_cleanup(pd); pd 1500 drivers/platform/x86/sony-laptop.c sony_nc_smart_conn_cleanup(pd); pd 1508 drivers/platform/x86/sony-laptop.c sony_nc_handles_cleanup(pd); pd 1870 drivers/platform/x86/sony-laptop.c static int sony_nc_kbd_backlight_setup(struct platform_device *pd, pd 1929 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &kbdbl_ctl->mode_attr); pd 1944 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &kbdbl_ctl->timeout_attr); pd 1955 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &kbdbl_ctl->mode_attr); pd 1962 drivers/platform/x86/sony-laptop.c static void sony_nc_kbd_backlight_cleanup(struct platform_device *pd, pd 1966 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &kbdbl_ctl->mode_attr); pd 1968 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &kbdbl_ctl->timeout_attr); pd 2078 drivers/platform/x86/sony-laptop.c static int sony_nc_battery_care_setup(struct platform_device *pd, pd 2095 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &bcare_ctl->attrs[0]); pd 2108 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &bcare_ctl->attrs[1]); pd 2115 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &bcare_ctl->attrs[0]); pd 2124 drivers/platform/x86/sony-laptop.c static void sony_nc_battery_care_cleanup(struct platform_device *pd) pd 2127 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &bcare_ctl->attrs[0]); pd 2129 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &bcare_ctl->attrs[1]); pd 2236 drivers/platform/x86/sony-laptop.c static int sony_nc_thermal_setup(struct platform_device *pd) pd 2267 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &th_handle->profiles_attr); pd 2271 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &th_handle->mode_attr); pd 2278 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &th_handle->profiles_attr); pd 2285 drivers/platform/x86/sony-laptop.c static void sony_nc_thermal_cleanup(struct platform_device *pd) pd 2288 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &th_handle->profiles_attr); pd 2289 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &th_handle->mode_attr); pd 2373 drivers/platform/x86/sony-laptop.c static int sony_nc_lid_resume_setup(struct platform_device *pd, pd 2410 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, &lid_ctl->attrs[i]); pd 2419 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &lid_ctl->attrs[i]); pd 2427 drivers/platform/x86/sony-laptop.c static void sony_nc_lid_resume_cleanup(struct platform_device *pd) pd 2436 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &lid_ctl->attrs[i]); pd 2507 drivers/platform/x86/sony-laptop.c static int sony_nc_gfx_switch_setup(struct platform_device *pd, pd 2523 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, &gfxs_ctl->attr); pd 2536 drivers/platform/x86/sony-laptop.c static void sony_nc_gfx_switch_cleanup(struct platform_device *pd) pd 2539 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &gfxs_ctl->attr); pd 2579 drivers/platform/x86/sony-laptop.c static int sony_nc_highspeed_charging_setup(struct platform_device *pd) pd 2601 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, hsc_handle); pd 2611 drivers/platform/x86/sony-laptop.c static void sony_nc_highspeed_charging_cleanup(struct platform_device *pd) pd 2614 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, hsc_handle); pd 2653 drivers/platform/x86/sony-laptop.c static int sony_nc_lowbatt_setup(struct platform_device *pd) pd 2667 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, lowbatt_handle); pd 2677 drivers/platform/x86/sony-laptop.c static void sony_nc_lowbatt_cleanup(struct platform_device *pd) pd 2680 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, lowbatt_handle); pd 2730 drivers/platform/x86/sony-laptop.c static int sony_nc_fanspeed_setup(struct platform_device *pd) pd 2756 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, fan_handle); pd 2760 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, hsf_handle); pd 2767 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, fan_handle); pd 2779 drivers/platform/x86/sony-laptop.c static void sony_nc_fanspeed_cleanup(struct platform_device *pd) pd 2782 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, fan_handle); pd 2787 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, hsf_handle); pd 2826 drivers/platform/x86/sony-laptop.c static int sony_nc_usb_charge_setup(struct platform_device *pd) pd 2848 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, uc_handle); pd 2858 drivers/platform/x86/sony-laptop.c static void sony_nc_usb_charge_cleanup(struct platform_device *pd) pd 2861 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, uc_handle); pd 2881 drivers/platform/x86/sony-laptop.c static int sony_nc_panelid_setup(struct platform_device *pd) pd 2895 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, panel_handle); pd 2905 drivers/platform/x86/sony-laptop.c static void sony_nc_panelid_cleanup(struct platform_device *pd) pd 2908 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, panel_handle); pd 2936 drivers/platform/x86/sony-laptop.c static int sony_nc_smart_conn_setup(struct platform_device *pd) pd 2950 drivers/platform/x86/sony-laptop.c result = device_create_file(&pd->dev, sc_handle); pd 2960 drivers/platform/x86/sony-laptop.c static void sony_nc_smart_conn_cleanup(struct platform_device *pd) pd 2963 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, sc_handle); pd 3009 drivers/platform/x86/sony-laptop.c static int sony_nc_touchpad_setup(struct platform_device *pd, pd 3026 drivers/platform/x86/sony-laptop.c ret = device_create_file(&pd->dev, &tp_ctl->attr); pd 3035 drivers/platform/x86/sony-laptop.c static void sony_nc_touchpad_cleanup(struct platform_device *pd) pd 3038 drivers/platform/x86/sony-laptop.c device_remove_file(&pd->dev, &tp_ctl->attr); pd 93 drivers/platform/x86/surface3_button.c struct platform_device *pd; pd 136 drivers/platform/x86/surface3_button.c pd = platform_device_alloc("gpio-keys", PLATFORM_DEVID_AUTO); pd 137 drivers/platform/x86/surface3_button.c if (!pd) { pd 142 drivers/platform/x86/surface3_button.c error = platform_device_add_data(pd, gpio_keys_pdata, pd 147 drivers/platform/x86/surface3_button.c error = platform_device_add(pd); pd 151 drivers/platform/x86/surface3_button.c return pd; pd 154 drivers/platform/x86/surface3_button.c platform_device_put(pd); pd 186 drivers/platform/x86/surface3_button.c struct platform_device *pd; pd 208 drivers/platform/x86/surface3_button.c pd = surface3_button_device_create(client, pd 211 drivers/platform/x86/surface3_button.c if (IS_ERR(pd)) { pd 212 drivers/platform/x86/surface3_button.c error = PTR_ERR(pd); pd 220 drivers/platform/x86/surface3_button.c priv->children[i] = pd; pd 555 drivers/power/supply/cros_usbpd-charger.c static int cros_usbpd_charger_probe(struct platform_device *pd) pd 557 drivers/power/supply/cros_usbpd-charger.c struct cros_ec_dev *ec_dev = dev_get_drvdata(pd->dev.parent); pd 560 drivers/power/supply/cros_usbpd-charger.c struct device *dev = &pd->dev; pd 576 drivers/power/supply/cros_usbpd-charger.c platform_set_drvdata(pd, charger); pd 122 drivers/ptp/ptp_chardev.c struct ptp_pin_desc pd; pd 317 drivers/ptp/ptp_chardev.c if (copy_from_user(&pd, (void __user *)arg, sizeof(pd))) { pd 321 drivers/ptp/ptp_chardev.c if ((pd.rsv[0] || pd.rsv[1] || pd.rsv[2] pd 322 drivers/ptp/ptp_chardev.c || pd.rsv[3] || pd.rsv[4]) pd 327 drivers/ptp/ptp_chardev.c pd.rsv[0] = 0; pd 328 drivers/ptp/ptp_chardev.c pd.rsv[1] = 0; pd 329 drivers/ptp/ptp_chardev.c pd.rsv[2] = 0; pd 330 drivers/ptp/ptp_chardev.c pd.rsv[3] = 0; pd 331 drivers/ptp/ptp_chardev.c pd.rsv[4] = 0; pd 333 drivers/ptp/ptp_chardev.c pin_index = pd.index; pd 341 drivers/ptp/ptp_chardev.c pd = ops->pin_config[pin_index]; pd 343 drivers/ptp/ptp_chardev.c if (!err && copy_to_user((void __user *)arg, &pd, sizeof(pd))) pd 349 drivers/ptp/ptp_chardev.c if (copy_from_user(&pd, (void __user *)arg, sizeof(pd))) { pd 353 drivers/ptp/ptp_chardev.c if ((pd.rsv[0] || pd.rsv[1] || pd.rsv[2] pd 354 drivers/ptp/ptp_chardev.c || pd.rsv[3] || pd.rsv[4]) pd 359 drivers/ptp/ptp_chardev.c pd.rsv[0] = 0; pd 360 drivers/ptp/ptp_chardev.c pd.rsv[1] = 0; pd 361 drivers/ptp/ptp_chardev.c pd.rsv[2] = 0; pd 362 drivers/ptp/ptp_chardev.c pd.rsv[3] = 0; pd 363 drivers/ptp/ptp_chardev.c pd.rsv[4] = 0; pd 365 drivers/ptp/ptp_chardev.c pin_index = pd.index; pd 373 drivers/ptp/ptp_chardev.c err = ptp_set_pinfunc(ptp, pin_index, pd.func, pd.chan); pd 171 drivers/regulator/max77620-regulator.c int pd = rpdata->active_fps_pd_slot; pd 179 drivers/regulator/max77620-regulator.c pd = rpdata->suspend_fps_pd_slot; pd 189 drivers/regulator/max77620-regulator.c if (pd >= 0) { pd 190 drivers/regulator/max77620-regulator.c val |= (pd << MAX77620_FPS_PD_PERIOD_SHIFT); pd 126 drivers/regulator/max8952.c struct max8952_platform_data *pd; pd 131 drivers/regulator/max8952.c pd = devm_kzalloc(dev, sizeof(*pd), GFP_KERNEL); pd 132 drivers/regulator/max8952.c if (!pd) pd 135 drivers/regulator/max8952.c if (of_property_read_u32(np, "max8952,default-mode", &pd->default_mode)) pd 139 drivers/regulator/max8952.c pd->dvs_mode, ARRAY_SIZE(pd->dvs_mode)); pd 145 drivers/regulator/max8952.c for (i = 0; i < ARRAY_SIZE(pd->dvs_mode); ++i) { pd 146 drivers/regulator/max8952.c if (pd->dvs_mode[i] < 770000 || pd->dvs_mode[i] > 1400000) { pd 150 drivers/regulator/max8952.c pd->dvs_mode[i] = (pd->dvs_mode[i] - 770000) / 10000; pd 153 drivers/regulator/max8952.c if (of_property_read_u32(np, "max8952,sync-freq", &pd->sync_freq)) pd 156 drivers/regulator/max8952.c if (of_property_read_u32(np, "max8952,ramp-speed", &pd->ramp_speed)) pd 159 drivers/regulator/max8952.c pd->reg_data = of_get_regulator_init_data(dev, np, ®ulator); pd 160 drivers/regulator/max8952.c if (!pd->reg_data) { pd 165 drivers/regulator/max8952.c return pd; pd 30 drivers/regulator/qcom_rpm-regulator.c struct request_member pd; /* pull down enable */ pd 72 drivers/regulator/qcom_rpm-regulator.c .pd = { 1, 0x00000001, 0 }, pd 83 drivers/regulator/qcom_rpm-regulator.c .pd = { 1, 0x00000001, 0 }, pd 92 drivers/regulator/qcom_rpm-regulator.c .pd = { 0, 0x00000002, 1 }, pd 109 drivers/regulator/qcom_rpm-regulator.c .pd = { 0, 0x00800000, 23 }, pd 120 drivers/regulator/qcom_rpm-regulator.c .pd = { 0, 0x00800000, 23 }, pd 134 drivers/regulator/qcom_rpm-regulator.c .pd = { 0, 0x00000002, 1 }, pd 700 drivers/regulator/qcom_rpm-regulator.c ret = rpm_reg_set(vreg, &vreg->parts->pd, 1); pd 700 drivers/scsi/aic94xx/aic94xx_scb.c static void set_speed_mask(u8 *speed_mask, struct asd_phy_desc *pd) pd 706 drivers/scsi/aic94xx/aic94xx_scb.c switch (pd->max_sas_lrate) { pd 718 drivers/scsi/aic94xx/aic94xx_scb.c switch (pd->min_sas_lrate) { pd 730 drivers/scsi/aic94xx/aic94xx_scb.c switch (pd->max_sata_lrate) { pd 739 drivers/scsi/aic94xx/aic94xx_scb.c switch (pd->min_sata_lrate) { pd 889 drivers/scsi/aic94xx/aic94xx_scb.c struct asd_phy_desc *pd = asd_ha->phys[phy->id].phy_desc; pd 901 drivers/scsi/aic94xx/aic94xx_scb.c pd->min_sas_lrate = rates->minimum_linkrate; pd 902 drivers/scsi/aic94xx/aic94xx_scb.c pd->min_sata_lrate = rates->minimum_linkrate; pd 905 drivers/scsi/aic94xx/aic94xx_scb.c pd->max_sas_lrate = rates->maximum_linkrate; pd 906 drivers/scsi/aic94xx/aic94xx_scb.c pd->max_sata_lrate = rates->maximum_linkrate; pd 793 drivers/scsi/aic94xx/aic94xx_sds.c struct asd_manuf_phy_desc *pd = &phy_param->phy_desc[i]; pd 794 drivers/scsi/aic94xx/aic94xx_sds.c switch (pd->state & 0xF) { pd 809 drivers/scsi/aic94xx/aic94xx_sds.c asd_ha->hw_prof.phy_desc[i].phy_control_0 = pd->phy_control_0; pd 810 drivers/scsi/aic94xx/aic94xx_sds.c asd_ha->hw_prof.phy_desc[i].phy_control_1 = pd->phy_control_1; pd 811 drivers/scsi/aic94xx/aic94xx_sds.c asd_ha->hw_prof.phy_desc[i].phy_control_2 = pd->phy_control_2; pd 812 drivers/scsi/aic94xx/aic94xx_sds.c asd_ha->hw_prof.phy_desc[i].phy_control_3 = pd->phy_control_3; pd 297 drivers/scsi/be2iscsi/be_cmds.h u8 pd[10]; /* dword 1 */ pd 585 drivers/scsi/be2iscsi/be_cmds.h u8 pd[10]; /* dword 1 */ pd 2091 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2121 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2140 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2143 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2149 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2154 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2159 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2164 drivers/scsi/megaraid/megaraid_sas.h struct megasas_evtarg_pd pd; pd 2655 drivers/scsi/megaraid/megaraid_sas.h __le16 MR_PdDevHandleGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map); pd 121 drivers/scsi/megaraid/megaraid_sas_fp.c return le16_to_cpu(map->raidMap.arMapInfo[ar].pd[arm]); pd 129 drivers/scsi/megaraid/megaraid_sas_fp.c __le16 MR_PdDevHandleGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map) pd 131 drivers/scsi/megaraid/megaraid_sas_fp.c return map->raidMap.devHndlInfo[pd].curDevHdl; pd 134 drivers/scsi/megaraid/megaraid_sas_fp.c static u8 MR_PdInterfaceTypeGet(u32 pd, struct MR_DRV_RAID_MAP_ALL *map) pd 136 drivers/scsi/megaraid/megaraid_sas_fp.c return map->raidMap.devHndlInfo[pd].interfaceType; pd 694 drivers/scsi/megaraid/megaraid_sas_fp.c u32 pd, arRef, r1_alt_pd; pd 729 drivers/scsi/megaraid/megaraid_sas_fp.c pd = MR_ArPdGet(arRef, physArm, map); pd 731 drivers/scsi/megaraid/megaraid_sas_fp.c if (pd != MR_PD_INVALID) { pd 732 drivers/scsi/megaraid/megaraid_sas_fp.c *pDevHandle = MR_PdDevHandleGet(pd, map); pd 733 drivers/scsi/megaraid/megaraid_sas_fp.c *pPdInterface = MR_PdInterfaceTypeGet(pd, map); pd 751 drivers/scsi/megaraid/megaraid_sas_fp.c pd = MR_ArPdGet(arRef, physArm, map); pd 752 drivers/scsi/megaraid/megaraid_sas_fp.c if (pd != MR_PD_INVALID) { pd 753 drivers/scsi/megaraid/megaraid_sas_fp.c *pDevHandle = MR_PdDevHandleGet(pd, map); pd 754 drivers/scsi/megaraid/megaraid_sas_fp.c *pPdInterface = MR_PdInterfaceTypeGet(pd, map); pd 770 drivers/scsi/megaraid/megaraid_sas_fp.c io_info->pd_after_lb = pd; pd 797 drivers/scsi/megaraid/megaraid_sas_fp.c u32 pd, arRef, r1_alt_pd; pd 844 drivers/scsi/megaraid/megaraid_sas_fp.c pd = MR_ArPdGet(arRef, physArm, map); /* Get the pd */ pd 846 drivers/scsi/megaraid/megaraid_sas_fp.c if (pd != MR_PD_INVALID) { pd 848 drivers/scsi/megaraid/megaraid_sas_fp.c *pDevHandle = MR_PdDevHandleGet(pd, map); pd 849 drivers/scsi/megaraid/megaraid_sas_fp.c *pPdInterface = MR_PdInterfaceTypeGet(pd, map); pd 868 drivers/scsi/megaraid/megaraid_sas_fp.c pd = MR_ArPdGet(arRef, physArm, map); pd 869 drivers/scsi/megaraid/megaraid_sas_fp.c if (pd != MR_PD_INVALID) { pd 871 drivers/scsi/megaraid/megaraid_sas_fp.c *pDevHandle = MR_PdDevHandleGet(pd, map); pd 872 drivers/scsi/megaraid/megaraid_sas_fp.c *pPdInterface = MR_PdInterfaceTypeGet(pd, map); pd 888 drivers/scsi/megaraid/megaraid_sas_fp.c io_info->pd_after_lb = pd; pd 2995 drivers/scsi/megaraid/megaraid_sas_fusion.c u32 arRef, pd; pd 3047 drivers/scsi/megaraid/megaraid_sas_fusion.c pd = MR_ArPdGet(arRef, physArm, local_map_ptr); pd 3048 drivers/scsi/megaraid/megaraid_sas_fusion.c devHandle = MR_PdDevHandleGet(pd, local_map_ptr); pd 736 drivers/scsi/megaraid/megaraid_sas_fusion.h __le16 pd[MAX_RAIDMAP_ROW_SIZE]; pd 1304 drivers/scsi/qla2xxx/qla_init.c struct port_database_24xx *pd; pd 1331 drivers/scsi/qla2xxx/qla_init.c pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma); pd 1332 drivers/scsi/qla2xxx/qla_init.c if (pd == NULL) { pd 1348 drivers/scsi/qla2xxx/qla_init.c mbx->u.mbx.in = (void *)pd; pd 1363 drivers/scsi/qla2xxx/qla_init.c if (pd) pd 1364 drivers/scsi/qla2xxx/qla_init.c dma_pool_free(ha->s_dma_pool, pd, pd_dma); pd 1408 drivers/scsi/qla2xxx/qla_init.c struct port_database_24xx *pd; pd 1412 drivers/scsi/qla2xxx/qla_init.c pd = (struct port_database_24xx *)sp->u.iocb_cmd.u.mbx.in; pd 1418 drivers/scsi/qla2xxx/qla_init.c fcport->disc_state, pd->current_login_state, fcport->fc4f_nvme, pd 1425 drivers/scsi/qla2xxx/qla_init.c ls = pd->current_login_state >> 4; pd 1427 drivers/scsi/qla2xxx/qla_init.c ls = pd->current_login_state & 0xf; pd 1444 drivers/scsi/qla2xxx/qla_init.c __qla24xx_parse_gpdb(vha, fcport, pd); pd 1880 drivers/scsi/qla2xxx/qla_mbx.c port_database_t *pd; pd 1889 drivers/scsi/qla2xxx/qla_mbx.c pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma); pd 1890 drivers/scsi/qla2xxx/qla_mbx.c if (pd == NULL) { pd 1932 drivers/scsi/qla2xxx/qla_mbx.c pd24 = (struct port_database_24xx *) pd; pd 1992 drivers/scsi/qla2xxx/qla_mbx.c if (pd->master_state != PD_STATE_PORT_LOGGED_IN && pd 1993 drivers/scsi/qla2xxx/qla_mbx.c pd->slave_state != PD_STATE_PORT_LOGGED_IN) { pd 1996 drivers/scsi/qla2xxx/qla_mbx.c "portid=%02x%02x%02x.\n", pd->master_state, pd 1997 drivers/scsi/qla2xxx/qla_mbx.c pd->slave_state, fcport->d_id.b.domain, pd 2005 drivers/scsi/qla2xxx/qla_mbx.c memcmp(fcport->port_name, pd->port_name, 8))) { pd 2012 drivers/scsi/qla2xxx/qla_mbx.c memcpy(fcport->node_name, pd->node_name, WWN_SIZE); pd 2013 drivers/scsi/qla2xxx/qla_mbx.c memcpy(fcport->port_name, pd->port_name, WWN_SIZE); pd 2016 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = pd->port_id[0]; pd 2017 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = pd->port_id[3]; pd 2018 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = pd->port_id[2]; pd 2022 drivers/scsi/qla2xxx/qla_mbx.c if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0) pd 2028 drivers/scsi/qla2xxx/qla_mbx.c fcport->supported_classes = (pd->options & BIT_4) ? pd 2033 drivers/scsi/qla2xxx/qla_mbx.c dma_pool_free(ha->s_dma_pool, pd, pd_dma); pd 6311 drivers/scsi/qla2xxx/qla_mbx.c struct port_database_24xx *pd; pd 6318 drivers/scsi/qla2xxx/qla_mbx.c pd = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &pd_dma); pd 6319 drivers/scsi/qla2xxx/qla_mbx.c if (pd == NULL) { pd 6342 drivers/scsi/qla2xxx/qla_mbx.c rval = __qla24xx_parse_gpdb(vha, fcport, pd); pd 6348 drivers/scsi/qla2xxx/qla_mbx.c if (pd) pd 6349 drivers/scsi/qla2xxx/qla_mbx.c dma_pool_free(ha->s_dma_pool, pd, pd_dma); pd 6355 drivers/scsi/qla2xxx/qla_mbx.c struct port_database_24xx *pd) pd 6362 drivers/scsi/qla2xxx/qla_mbx.c current_login_state = pd->current_login_state >> 4; pd 6363 drivers/scsi/qla2xxx/qla_mbx.c last_login_state = pd->last_login_state >> 4; pd 6365 drivers/scsi/qla2xxx/qla_mbx.c current_login_state = pd->current_login_state & 0xf; pd 6366 drivers/scsi/qla2xxx/qla_mbx.c last_login_state = pd->last_login_state & 0xf; pd 6380 drivers/scsi/qla2xxx/qla_mbx.c memcmp(fcport->port_name, pd->port_name, 8))) { pd 6387 drivers/scsi/qla2xxx/qla_mbx.c memcpy(fcport->node_name, pd->node_name, WWN_SIZE); pd 6388 drivers/scsi/qla2xxx/qla_mbx.c memcpy(fcport->port_name, pd->port_name, WWN_SIZE); pd 6391 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.domain = pd->port_id[0]; pd 6392 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.area = pd->port_id[1]; pd 6393 drivers/scsi/qla2xxx/qla_mbx.c fcport->d_id.b.al_pa = pd->port_id[2]; pd 6398 drivers/scsi/qla2xxx/qla_mbx.c if ((pd->prli_svc_param_word_3[0] & BIT_5) == 0) pd 6400 drivers/scsi/qla2xxx/qla_mbx.c if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0) pd 6402 drivers/scsi/qla2xxx/qla_mbx.c if ((pd->prli_svc_param_word_3[0] & BIT_3) == 0) pd 6406 drivers/scsi/qla2xxx/qla_mbx.c if ((pd->prli_svc_param_word_3[0] & BIT_4) == 0) pd 6412 drivers/scsi/qla2xxx/qla_mbx.c fcport->supported_classes = (pd->flags & PDF_CLASS_2) ? pd 6415 drivers/scsi/qla2xxx/qla_mbx.c if (pd->prli_svc_param_word_3[0] & BIT_7) { pd 216 drivers/scsi/sgiwd93.c struct sgiwd93_platform_data *pd = pdev->dev.platform_data; pd 217 drivers/scsi/sgiwd93.c unsigned char *wdregs = pd->wdregs; pd 218 drivers/scsi/sgiwd93.c struct hpc3_scsiregs *hregs = pd->hregs; pd 222 drivers/scsi/sgiwd93.c unsigned int unit = pd->unit; pd 223 drivers/scsi/sgiwd93.c unsigned int irq = pd->irq; pd 290 drivers/scsi/sgiwd93.c struct sgiwd93_platform_data *pd = pdev->dev.platform_data; pd 293 drivers/scsi/sgiwd93.c free_irq(pd->irq, host); pd 47 drivers/soc/actions/owl-sps.c static int owl_sps_set_power(struct owl_sps_domain *pd, bool enable) pd 51 drivers/soc/actions/owl-sps.c ack_mask = BIT(pd->info->ack_bit); pd 52 drivers/soc/actions/owl-sps.c pwr_mask = BIT(pd->info->pwr_bit); pd 54 drivers/soc/actions/owl-sps.c return owl_sps_set_pg(pd->sps->base, pwr_mask, ack_mask, enable); pd 59 drivers/soc/actions/owl-sps.c struct owl_sps_domain *pd = to_owl_pd(domain); pd 61 drivers/soc/actions/owl-sps.c dev_dbg(pd->sps->dev, "%s power on", pd->info->name); pd 63 drivers/soc/actions/owl-sps.c return owl_sps_set_power(pd, true); pd 68 drivers/soc/actions/owl-sps.c struct owl_sps_domain *pd = to_owl_pd(domain); pd 70 drivers/soc/actions/owl-sps.c dev_dbg(pd->sps->dev, "%s power off", pd->info->name); pd 72 drivers/soc/actions/owl-sps.c return owl_sps_set_power(pd, false); pd 77 drivers/soc/actions/owl-sps.c struct owl_sps_domain *pd; pd 79 drivers/soc/actions/owl-sps.c pd = devm_kzalloc(sps->dev, sizeof(*pd), GFP_KERNEL); pd 80 drivers/soc/actions/owl-sps.c if (!pd) pd 83 drivers/soc/actions/owl-sps.c pd->info = &sps->info->domains[index]; pd 84 drivers/soc/actions/owl-sps.c pd->sps = sps; pd 86 drivers/soc/actions/owl-sps.c pd->genpd.name = pd->info->name; pd 87 drivers/soc/actions/owl-sps.c pd->genpd.power_on = owl_sps_power_on; pd 88 drivers/soc/actions/owl-sps.c pd->genpd.power_off = owl_sps_power_off; pd 89 drivers/soc/actions/owl-sps.c pd->genpd.flags = pd->info->genpd_flags; pd 90 drivers/soc/actions/owl-sps.c pm_genpd_init(&pd->genpd, NULL, false); pd 92 drivers/soc/actions/owl-sps.c sps->genpd_data.domains[index] = &pd->genpd; pd 49 drivers/soc/amlogic/meson-gx-pwrc-vpu.c struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd); pd 52 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 58 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0, pd 63 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1, pd 68 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0, pd 74 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 79 drivers/soc/amlogic/meson-gx-pwrc-vpu.c clk_disable_unprepare(pd->vpu_clk); pd 80 drivers/soc/amlogic/meson-gx-pwrc-vpu.c clk_disable_unprepare(pd->vapb_clk); pd 87 drivers/soc/amlogic/meson-gx-pwrc-vpu.c struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd); pd 90 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 96 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0, pd 101 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1, pd 106 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG2, pd 111 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0, pd 117 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 122 drivers/soc/amlogic/meson-gx-pwrc-vpu.c clk_disable_unprepare(pd->vpu_clk); pd 123 drivers/soc/amlogic/meson-gx-pwrc-vpu.c clk_disable_unprepare(pd->vapb_clk); pd 128 drivers/soc/amlogic/meson-gx-pwrc-vpu.c static int meson_gx_pwrc_vpu_setup_clk(struct meson_gx_pwrc_vpu *pd) pd 132 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = clk_prepare_enable(pd->vpu_clk); pd 136 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = clk_prepare_enable(pd->vapb_clk); pd 138 drivers/soc/amlogic/meson-gx-pwrc-vpu.c clk_disable_unprepare(pd->vpu_clk); pd 145 drivers/soc/amlogic/meson-gx-pwrc-vpu.c struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd); pd 149 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 155 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0, pd 161 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1, pd 167 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0, pd 173 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = reset_control_assert(pd->rstc); pd 177 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 180 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = reset_control_deassert(pd->rstc); pd 184 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = meson_gx_pwrc_vpu_setup_clk(pd); pd 193 drivers/soc/amlogic/meson-gx-pwrc-vpu.c struct meson_gx_pwrc_vpu *pd = genpd_to_pd(genpd); pd 197 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 203 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG0, pd 209 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG1, pd 215 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_VPU_MEM_PD_REG2, pd 221 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_hhi, HHI_MEM_PD_REG0, pd 227 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = reset_control_assert(pd->rstc); pd 231 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_update_bits(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, pd 234 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = reset_control_deassert(pd->rstc); pd 238 drivers/soc/amlogic/meson-gx-pwrc-vpu.c ret = meson_gx_pwrc_vpu_setup_clk(pd); pd 245 drivers/soc/amlogic/meson-gx-pwrc-vpu.c static bool meson_gx_pwrc_vpu_get_power(struct meson_gx_pwrc_vpu *pd) pd 249 drivers/soc/amlogic/meson-gx-pwrc-vpu.c regmap_read(pd->regmap_ao, AO_RTI_GEN_PWR_SLEEP0, ®); pd 191 drivers/soc/bcm/bcm2835-power.c static int bcm2835_power_power_off(struct bcm2835_power_domain *pd, u32 pm_reg) pd 193 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 207 drivers/soc/bcm/bcm2835-power.c static int bcm2835_power_power_on(struct bcm2835_power_domain *pd, u32 pm_reg) pd 209 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 240 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 255 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 273 drivers/soc/bcm/bcm2835-power.c static int bcm2835_asb_power_on(struct bcm2835_power_domain *pd, pd 279 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 282 drivers/soc/bcm/bcm2835-power.c ret = clk_prepare_enable(pd->clk); pd 285 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 292 drivers/soc/bcm/bcm2835-power.c clk_disable_unprepare(pd->clk); pd 297 drivers/soc/bcm/bcm2835-power.c ret = clk_prepare_enable(pd->clk); pd 300 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 307 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 313 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 322 drivers/soc/bcm/bcm2835-power.c clk_disable_unprepare(pd->clk); pd 328 drivers/soc/bcm/bcm2835-power.c static int bcm2835_asb_power_off(struct bcm2835_power_domain *pd, pd 334 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 340 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 346 drivers/soc/bcm/bcm2835-power.c pd->base.name); pd 351 drivers/soc/bcm/bcm2835-power.c clk_disable_unprepare(pd->clk); pd 361 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power_domain *pd = pd 363 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 365 drivers/soc/bcm/bcm2835-power.c switch (pd->domain) { pd 367 drivers/soc/bcm/bcm2835-power.c return bcm2835_power_power_on(pd, PM_GRAFX); pd 370 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_on(pd, PM_GRAFX, pd 375 drivers/soc/bcm/bcm2835-power.c return bcm2835_power_power_on(pd, PM_IMAGE); pd 378 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_on(pd, PM_IMAGE, pd 383 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_on(pd, PM_IMAGE, pd 388 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_on(pd, PM_IMAGE, pd 420 drivers/soc/bcm/bcm2835-power.c dev_err(power->dev, "Invalid domain %d\n", pd->domain); pd 427 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power_domain *pd = pd 429 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power *power = pd->power; pd 431 drivers/soc/bcm/bcm2835-power.c switch (pd->domain) { pd 433 drivers/soc/bcm/bcm2835-power.c return bcm2835_power_power_off(pd, PM_GRAFX); pd 436 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_off(pd, PM_GRAFX, pd 441 drivers/soc/bcm/bcm2835-power.c return bcm2835_power_power_off(pd, PM_IMAGE); pd 444 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_off(pd, PM_IMAGE, pd 449 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_off(pd, PM_IMAGE, pd 454 drivers/soc/bcm/bcm2835-power.c return bcm2835_asb_power_off(pd, PM_IMAGE, pd 483 drivers/soc/bcm/bcm2835-power.c dev_err(power->dev, "Invalid domain %d\n", pd->domain); pd 536 drivers/soc/bcm/bcm2835-power.c struct bcm2835_power_domain *pd; pd 541 drivers/soc/bcm/bcm2835-power.c pd = &power->domains[BCM2835_POWER_DOMAIN_GRAFX_V3D]; pd 544 drivers/soc/bcm/bcm2835-power.c pd = &power->domains[BCM2835_POWER_DOMAIN_IMAGE_H264]; pd 547 drivers/soc/bcm/bcm2835-power.c pd = &power->domains[BCM2835_POWER_DOMAIN_IMAGE_ISP]; pd 554 drivers/soc/bcm/bcm2835-power.c ret = bcm2835_power_pd_power_off(&pd->base); pd 558 drivers/soc/bcm/bcm2835-power.c return bcm2835_power_pd_power_on(&pd->base); pd 356 drivers/soc/fsl/dpio/dpio-service.c struct qbman_pull_desc pd; pd 359 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_clear(&pd); pd 360 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1); pd 361 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_numframes(&pd, (u8)s->max); pd 362 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_fq(&pd, fqid); pd 368 drivers/soc/fsl/dpio/dpio-service.c err = qbman_swp_pull(d->swp, &pd); pd 387 drivers/soc/fsl/dpio/dpio-service.c struct qbman_pull_desc pd; pd 390 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_clear(&pd); pd 391 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_storage(&pd, s->vaddr, s->paddr, 1); pd 392 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_numframes(&pd, (u8)s->max); pd 393 drivers/soc/fsl/dpio/dpio-service.c qbman_pull_desc_set_channel(&pd, channelid, qbman_pull_type_prio); pd 400 drivers/soc/fsl/dpio/dpio-service.c err = qbman_swp_pull(d->swp, &pd); pd 61 drivers/soc/imx/gpc.c struct imx_pm_domain *pd = to_imx_pm_domain(genpd); pd 66 drivers/soc/imx/gpc.c regmap_read(pd->regmap, pd->reg_offs + GPC_PGC_PDNSCR_OFFS, &val); pd 71 drivers/soc/imx/gpc.c regmap_update_bits(pd->regmap, pd->reg_offs + GPC_PGC_CTRL_OFFS, pd 75 drivers/soc/imx/gpc.c val = BIT(pd->cntr_pdn_bit); pd 76 drivers/soc/imx/gpc.c regmap_update_bits(pd->regmap, GPC_CNTR, val, val); pd 79 drivers/soc/imx/gpc.c udelay(DIV_ROUND_UP(iso + iso2sw, pd->ipg_rate_mhz)); pd 81 drivers/soc/imx/gpc.c if (pd->supply) pd 82 drivers/soc/imx/gpc.c regulator_disable(pd->supply); pd 89 drivers/soc/imx/gpc.c struct imx_pm_domain *pd = to_imx_pm_domain(genpd); pd 93 drivers/soc/imx/gpc.c if (pd->supply) { pd 94 drivers/soc/imx/gpc.c ret = regulator_enable(pd->supply); pd 103 drivers/soc/imx/gpc.c for (i = 0; i < pd->num_clks; i++) pd 104 drivers/soc/imx/gpc.c clk_prepare_enable(pd->clk[i]); pd 107 drivers/soc/imx/gpc.c regmap_update_bits(pd->regmap, pd->reg_offs + GPC_PGC_CTRL_OFFS, pd 111 drivers/soc/imx/gpc.c req = BIT(pd->cntr_pdn_bit + 1); pd 112 drivers/soc/imx/gpc.c regmap_update_bits(pd->regmap, GPC_CNTR, req, req); pd 115 drivers/soc/imx/gpc.c ret = regmap_read_poll_timeout(pd->regmap, GPC_CNTR, val, !(val & req), pd 124 drivers/soc/imx/gpc.c for (i = 0; i < pd->num_clks; i++) pd 125 drivers/soc/imx/gpc.c clk_disable_unprepare(pd->clk[i]); pd 90 drivers/soc/qcom/qcom_aoss.c struct generic_pm_domain pd; pd 93 drivers/soc/qcom/qcom_aoss.c #define to_qmp_pd_resource(res) container_of(res, struct qmp_pd, pd) pd 319 drivers/soc/qcom/qcom_aoss.c res->pd.name, enable ? "on" : "off"); pd 362 drivers/soc/qcom/qcom_aoss.c res[i].pd.name = sdm845_resources[i]; pd 363 drivers/soc/qcom/qcom_aoss.c res[i].pd.power_on = qmp_pd_power_on; pd 364 drivers/soc/qcom/qcom_aoss.c res[i].pd.power_off = qmp_pd_power_off; pd 366 drivers/soc/qcom/qcom_aoss.c ret = pm_genpd_init(&res[i].pd, NULL, true); pd 372 drivers/soc/qcom/qcom_aoss.c data->domains[i] = &res[i].pd; pd 18 drivers/soc/qcom/rpmhpd.c #define domain_to_rpmhpd(domain) container_of(domain, struct rpmhpd, pd) pd 40 drivers/soc/qcom/rpmhpd.c struct generic_pm_domain pd; pd 63 drivers/soc/qcom/rpmhpd.c .pd = { .name = "ebi", }, pd 68 drivers/soc/qcom/rpmhpd.c .pd = { .name = "lmx", }, pd 73 drivers/soc/qcom/rpmhpd.c .pd = { .name = "lcx", }, pd 78 drivers/soc/qcom/rpmhpd.c .pd = { .name = "gfx", }, pd 83 drivers/soc/qcom/rpmhpd.c .pd = { .name = "mss", }, pd 89 drivers/soc/qcom/rpmhpd.c .pd = { .name = "mx", }, pd 95 drivers/soc/qcom/rpmhpd.c .pd = { .name = "mx_ao", }, pd 103 drivers/soc/qcom/rpmhpd.c .pd = { .name = "cx", }, pd 105 drivers/soc/qcom/rpmhpd.c .parent = &sdm845_mx.pd, pd 110 drivers/soc/qcom/rpmhpd.c .pd = { .name = "cx_ao", }, pd 113 drivers/soc/qcom/rpmhpd.c .parent = &sdm845_mx_ao.pd, pd 139 drivers/soc/qcom/rpmhpd.c static int rpmhpd_send_corner(struct rpmhpd *pd, int state, pd 143 drivers/soc/qcom/rpmhpd.c .addr = pd->addr, pd 152 drivers/soc/qcom/rpmhpd.c return rpmh_write(pd->dev, state, &cmd, 1); pd 154 drivers/soc/qcom/rpmhpd.c return rpmh_write_async(pd->dev, state, &cmd, 1); pd 157 drivers/soc/qcom/rpmhpd.c static void to_active_sleep(struct rpmhpd *pd, unsigned int corner, pd 162 drivers/soc/qcom/rpmhpd.c if (pd->active_only) pd 177 drivers/soc/qcom/rpmhpd.c static int rpmhpd_aggregate_corner(struct rpmhpd *pd, unsigned int corner) pd 180 drivers/soc/qcom/rpmhpd.c struct rpmhpd *peer = pd->peer; pd 185 drivers/soc/qcom/rpmhpd.c to_active_sleep(pd, corner, &this_active_corner, &this_sleep_corner); pd 193 drivers/soc/qcom/rpmhpd.c ret = rpmhpd_send_corner(pd, RPMH_ACTIVE_ONLY_STATE, active_corner, pd 194 drivers/soc/qcom/rpmhpd.c active_corner > pd->active_corner); pd 198 drivers/soc/qcom/rpmhpd.c pd->active_corner = active_corner; pd 203 drivers/soc/qcom/rpmhpd.c ret = rpmhpd_send_corner(pd, RPMH_WAKE_ONLY_STATE, pd 210 drivers/soc/qcom/rpmhpd.c return rpmhpd_send_corner(pd, RPMH_SLEEP_STATE, sleep_corner, pd 219 drivers/soc/qcom/rpmhpd.c struct rpmhpd *pd = domain_to_rpmhpd(domain); pd 224 drivers/soc/qcom/rpmhpd.c if (pd->corner) pd 225 drivers/soc/qcom/rpmhpd.c ret = rpmhpd_aggregate_corner(pd, pd->corner); pd 228 drivers/soc/qcom/rpmhpd.c pd->enabled = true; pd 237 drivers/soc/qcom/rpmhpd.c struct rpmhpd *pd = domain_to_rpmhpd(domain); pd 242 drivers/soc/qcom/rpmhpd.c ret = rpmhpd_aggregate_corner(pd, pd->level[0]); pd 245 drivers/soc/qcom/rpmhpd.c pd->enabled = false; pd 255 drivers/soc/qcom/rpmhpd.c struct rpmhpd *pd = domain_to_rpmhpd(domain); pd 260 drivers/soc/qcom/rpmhpd.c for (i = 0; i < pd->level_count; i++) pd 261 drivers/soc/qcom/rpmhpd.c if (level <= pd->level[i]) pd 268 drivers/soc/qcom/rpmhpd.c if (i == pd->level_count) pd 271 drivers/soc/qcom/rpmhpd.c if (pd->enabled) { pd 272 drivers/soc/qcom/rpmhpd.c ret = rpmhpd_aggregate_corner(pd, i); pd 277 drivers/soc/qcom/rpmhpd.c pd->corner = i; pd 374 drivers/soc/qcom/rpmhpd.c rpmhpds[i]->pd.power_off = rpmhpd_power_off; pd 375 drivers/soc/qcom/rpmhpd.c rpmhpds[i]->pd.power_on = rpmhpd_power_on; pd 376 drivers/soc/qcom/rpmhpd.c rpmhpds[i]->pd.set_performance_state = rpmhpd_set_performance_state; pd 377 drivers/soc/qcom/rpmhpd.c rpmhpds[i]->pd.opp_to_performance_state = rpmhpd_get_performance_state; pd 378 drivers/soc/qcom/rpmhpd.c pm_genpd_init(&rpmhpds[i]->pd, NULL, true); pd 380 drivers/soc/qcom/rpmhpd.c data->domains[i] = &rpmhpds[i]->pd; pd 389 drivers/soc/qcom/rpmhpd.c &rpmhpds[i]->pd); pd 17 drivers/soc/qcom/rpmpd.c #define domain_to_rpmpd(domain) container_of(domain, struct rpmpd, pd) pd 43 drivers/soc/qcom/rpmpd.c .pd = { .name = #_name, }, \ pd 50 drivers/soc/qcom/rpmpd.c .pd = { .name = #_active, }, \ pd 60 drivers/soc/qcom/rpmpd.c .pd = { .name = #_name, }, \ pd 68 drivers/soc/qcom/rpmpd.c .pd = { .name = #_name, }, \ pd 76 drivers/soc/qcom/rpmpd.c .pd = { .name = #_name, }, \ pd 84 drivers/soc/qcom/rpmpd.c .pd = { .name = #_name, }, \ pd 97 drivers/soc/qcom/rpmpd.c struct generic_pm_domain pd; pd 207 drivers/soc/qcom/rpmpd.c static int rpmpd_send_enable(struct rpmpd *pd, bool enable) pd 215 drivers/soc/qcom/rpmpd.c return qcom_rpm_smd_write(pd->rpm, QCOM_SMD_RPM_ACTIVE_STATE, pd 216 drivers/soc/qcom/rpmpd.c pd->res_type, pd->res_id, &req, sizeof(req)); pd 219 drivers/soc/qcom/rpmpd.c static int rpmpd_send_corner(struct rpmpd *pd, int state, unsigned int corner) pd 222 drivers/soc/qcom/rpmpd.c .key = pd->key, pd 227 drivers/soc/qcom/rpmpd.c return qcom_rpm_smd_write(pd->rpm, state, pd->res_type, pd->res_id, pd 231 drivers/soc/qcom/rpmpd.c static void to_active_sleep(struct rpmpd *pd, unsigned int corner, pd 236 drivers/soc/qcom/rpmpd.c if (pd->active_only) pd 242 drivers/soc/qcom/rpmpd.c static int rpmpd_aggregate_corner(struct rpmpd *pd) pd 245 drivers/soc/qcom/rpmpd.c struct rpmpd *peer = pd->peer; pd 250 drivers/soc/qcom/rpmpd.c to_active_sleep(pd, pd->corner, &this_active_corner, &this_sleep_corner); pd 258 drivers/soc/qcom/rpmpd.c ret = rpmpd_send_corner(pd, QCOM_SMD_RPM_ACTIVE_STATE, active_corner); pd 264 drivers/soc/qcom/rpmpd.c return rpmpd_send_corner(pd, QCOM_SMD_RPM_SLEEP_STATE, sleep_corner); pd 270 drivers/soc/qcom/rpmpd.c struct rpmpd *pd = domain_to_rpmpd(domain); pd 274 drivers/soc/qcom/rpmpd.c ret = rpmpd_send_enable(pd, true); pd 278 drivers/soc/qcom/rpmpd.c pd->enabled = true; pd 280 drivers/soc/qcom/rpmpd.c if (pd->corner) pd 281 drivers/soc/qcom/rpmpd.c ret = rpmpd_aggregate_corner(pd); pd 292 drivers/soc/qcom/rpmpd.c struct rpmpd *pd = domain_to_rpmpd(domain); pd 296 drivers/soc/qcom/rpmpd.c ret = rpmpd_send_enable(pd, false); pd 298 drivers/soc/qcom/rpmpd.c pd->enabled = false; pd 309 drivers/soc/qcom/rpmpd.c struct rpmpd *pd = domain_to_rpmpd(domain); pd 311 drivers/soc/qcom/rpmpd.c if (state > pd->max_state) pd 312 drivers/soc/qcom/rpmpd.c state = pd->max_state; pd 316 drivers/soc/qcom/rpmpd.c pd->corner = state; pd 319 drivers/soc/qcom/rpmpd.c if (!pd->enabled && pd->key != KEY_FLOOR_CORNER && pd 320 drivers/soc/qcom/rpmpd.c pd->key != KEY_FLOOR_LEVEL) pd 323 drivers/soc/qcom/rpmpd.c ret = rpmpd_aggregate_corner(pd); pd 376 drivers/soc/qcom/rpmpd.c rpmpds[i]->pd.power_off = rpmpd_power_off; pd 377 drivers/soc/qcom/rpmpd.c rpmpds[i]->pd.power_on = rpmpd_power_on; pd 378 drivers/soc/qcom/rpmpd.c rpmpds[i]->pd.set_performance_state = rpmpd_set_performance; pd 379 drivers/soc/qcom/rpmpd.c rpmpds[i]->pd.opp_to_performance_state = rpmpd_get_performance; pd 380 drivers/soc/qcom/rpmpd.c pm_genpd_init(&rpmpds[i]->pd, NULL, true); pd 382 drivers/soc/qcom/rpmpd.c data->domains[i] = &rpmpds[i]->pd; pd 183 drivers/soc/renesas/rcar-sysc.c struct rcar_sysc_pd *pd = to_rcar_pd(genpd); pd 186 drivers/soc/renesas/rcar-sysc.c return rcar_sysc_power(&pd->ch, false); pd 191 drivers/soc/renesas/rcar-sysc.c struct rcar_sysc_pd *pd = to_rcar_pd(genpd); pd 194 drivers/soc/renesas/rcar-sysc.c return rcar_sysc_power(&pd->ch, true); pd 199 drivers/soc/renesas/rcar-sysc.c static int __init rcar_sysc_pd_setup(struct rcar_sysc_pd *pd) pd 201 drivers/soc/renesas/rcar-sysc.c struct generic_pm_domain *genpd = &pd->genpd; pd 202 drivers/soc/renesas/rcar-sysc.c const char *name = pd->genpd.name; pd 205 drivers/soc/renesas/rcar-sysc.c if (pd->flags & PD_CPU) { pd 212 drivers/soc/renesas/rcar-sysc.c } else if (pd->flags & PD_SCU) { pd 220 drivers/soc/renesas/rcar-sysc.c } else if (pd->flags & PD_NO_CR) { pd 227 drivers/soc/renesas/rcar-sysc.c if (!(pd->flags & (PD_CPU | PD_SCU))) { pd 242 drivers/soc/renesas/rcar-sysc.c if (pd->flags & (PD_CPU | PD_NO_CR)) { pd 248 drivers/soc/renesas/rcar-sysc.c if (!rcar_sysc_power_is_off(&pd->ch)) { pd 253 drivers/soc/renesas/rcar-sysc.c rcar_sysc_power(&pd->ch, true); pd 375 drivers/soc/renesas/rcar-sysc.c struct rcar_sysc_pd *pd; pd 382 drivers/soc/renesas/rcar-sysc.c pd = kzalloc(sizeof(*pd) + strlen(area->name) + 1, GFP_KERNEL); pd 383 drivers/soc/renesas/rcar-sysc.c if (!pd) { pd 388 drivers/soc/renesas/rcar-sysc.c strcpy(pd->name, area->name); pd 389 drivers/soc/renesas/rcar-sysc.c pd->genpd.name = pd->name; pd 390 drivers/soc/renesas/rcar-sysc.c pd->ch.chan_offs = area->chan_offs; pd 391 drivers/soc/renesas/rcar-sysc.c pd->ch.chan_bit = area->chan_bit; pd 392 drivers/soc/renesas/rcar-sysc.c pd->ch.isr_bit = area->isr_bit; pd 393 drivers/soc/renesas/rcar-sysc.c pd->flags = area->flags; pd 395 drivers/soc/renesas/rcar-sysc.c error = rcar_sysc_pd_setup(pd); pd 399 drivers/soc/renesas/rcar-sysc.c domains->domains[area->isr_bit] = &pd->genpd; pd 405 drivers/soc/renesas/rcar-sysc.c &pd->genpd); pd 437 drivers/soc/renesas/rcar-sysc.c struct rcar_sysc_pd *pd; pd 448 drivers/soc/renesas/rcar-sysc.c pd = to_rcar_pd(genpd); pd 449 drivers/soc/renesas/rcar-sysc.c if (!(pd->flags & PD_CPU) || pd->ch.chan_bit != idx) pd 452 drivers/soc/renesas/rcar-sysc.c return rcar_sysc_power(&pd->ch, on); pd 149 drivers/soc/renesas/rmobile-sysc.c struct device_node *pd; pd 166 drivers/soc/renesas/rmobile-sysc.c struct device_node *pd; pd 168 drivers/soc/renesas/rmobile-sysc.c pd = of_parse_phandle(np, "power-domains", 0); pd 169 drivers/soc/renesas/rmobile-sysc.c if (!pd) pd 173 drivers/soc/renesas/rmobile-sysc.c if (pd == special_pds[i].pd && type == special_pds[i].type) { pd 174 drivers/soc/renesas/rmobile-sysc.c of_node_put(pd); pd 180 drivers/soc/renesas/rmobile-sysc.c of_node_put(pd); pd 184 drivers/soc/renesas/rmobile-sysc.c pr_debug("Special PM domain %pOFn type %d for %pOF\n", pd, type, np); pd 186 drivers/soc/renesas/rmobile-sysc.c special_pds[num_special_pds].pd = pd; pd 214 drivers/soc/renesas/rmobile-sysc.c of_node_put(special_pds[i].pd); pd 217 drivers/soc/renesas/rmobile-sysc.c static enum pd_types __init pd_type(const struct device_node *pd) pd 222 drivers/soc/renesas/rmobile-sysc.c if (pd == special_pds[i].pd) pd 229 drivers/soc/renesas/rmobile-sysc.c struct rmobile_pm_domain *pd) pd 231 drivers/soc/renesas/rmobile-sysc.c const char *name = pd->genpd.name; pd 240 drivers/soc/renesas/rmobile-sysc.c pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON; pd 245 drivers/soc/renesas/rmobile-sysc.c pd->gov = &pm_domain_always_on_gov; pd 246 drivers/soc/renesas/rmobile-sysc.c pd->suspend = rmobile_pd_suspend_console; pd 256 drivers/soc/renesas/rmobile-sysc.c pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON; pd 265 drivers/soc/renesas/rmobile-sysc.c pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON; pd 269 drivers/soc/renesas/rmobile-sysc.c if (pd->bit_shift == ~0) { pd 272 drivers/soc/renesas/rmobile-sysc.c pd->genpd.flags |= GENPD_FLAG_ALWAYS_ON; pd 277 drivers/soc/renesas/rmobile-sysc.c rmobile_init_pm_domain(pd); pd 287 drivers/soc/renesas/rmobile-sysc.c struct rmobile_pm_domain *pd; pd 294 drivers/soc/renesas/rmobile-sysc.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 295 drivers/soc/renesas/rmobile-sysc.c if (!pd) { pd 300 drivers/soc/renesas/rmobile-sysc.c pd->genpd.name = np->name; pd 301 drivers/soc/renesas/rmobile-sysc.c pd->base = base; pd 302 drivers/soc/renesas/rmobile-sysc.c pd->bit_shift = idx; pd 304 drivers/soc/renesas/rmobile-sysc.c rmobile_setup_pm_domain(np, pd); pd 306 drivers/soc/renesas/rmobile-sysc.c pm_genpd_add_subdomain(genpd_parent, &pd->genpd); pd 307 drivers/soc/renesas/rmobile-sysc.c of_genpd_add_provider_simple(np, &pd->genpd); pd 309 drivers/soc/renesas/rmobile-sysc.c rmobile_add_pm_domains(base, np, &pd->genpd); pd 134 drivers/soc/rockchip/pm_domains.c static bool rockchip_pmu_domain_is_idle(struct rockchip_pm_domain *pd) pd 136 drivers/soc/rockchip/pm_domains.c struct rockchip_pmu *pmu = pd->pmu; pd 137 drivers/soc/rockchip/pm_domains.c const struct rockchip_domain_info *pd_info = pd->info; pd 152 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_set_idle_request(struct rockchip_pm_domain *pd, pd 155 drivers/soc/rockchip/pm_domains.c const struct rockchip_domain_info *pd_info = pd->info; pd 156 drivers/soc/rockchip/pm_domains.c struct generic_pm_domain *genpd = &pd->genpd; pd 157 drivers/soc/rockchip/pm_domains.c struct rockchip_pmu *pmu = pd->pmu; pd 187 drivers/soc/rockchip/pm_domains.c ret = readx_poll_timeout_atomic(rockchip_pmu_domain_is_idle, pd, pd 199 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_save_qos(struct rockchip_pm_domain *pd) pd 203 drivers/soc/rockchip/pm_domains.c for (i = 0; i < pd->num_qos; i++) { pd 204 drivers/soc/rockchip/pm_domains.c regmap_read(pd->qos_regmap[i], pd 206 drivers/soc/rockchip/pm_domains.c &pd->qos_save_regs[0][i]); pd 207 drivers/soc/rockchip/pm_domains.c regmap_read(pd->qos_regmap[i], pd 209 drivers/soc/rockchip/pm_domains.c &pd->qos_save_regs[1][i]); pd 210 drivers/soc/rockchip/pm_domains.c regmap_read(pd->qos_regmap[i], pd 212 drivers/soc/rockchip/pm_domains.c &pd->qos_save_regs[2][i]); pd 213 drivers/soc/rockchip/pm_domains.c regmap_read(pd->qos_regmap[i], pd 215 drivers/soc/rockchip/pm_domains.c &pd->qos_save_regs[3][i]); pd 216 drivers/soc/rockchip/pm_domains.c regmap_read(pd->qos_regmap[i], pd 218 drivers/soc/rockchip/pm_domains.c &pd->qos_save_regs[4][i]); pd 223 drivers/soc/rockchip/pm_domains.c static int rockchip_pmu_restore_qos(struct rockchip_pm_domain *pd) pd 227 drivers/soc/rockchip/pm_domains.c for (i = 0; i < pd->num_qos; i++) { pd 228 drivers/soc/rockchip/pm_domains.c regmap_write(pd->qos_regmap[i], pd 230 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[0][i]); pd 231 drivers/soc/rockchip/pm_domains.c regmap_write(pd->qos_regmap[i], pd 233 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[1][i]); pd 234 drivers/soc/rockchip/pm_domains.c regmap_write(pd->qos_regmap[i], pd 236 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[2][i]); pd 237 drivers/soc/rockchip/pm_domains.c regmap_write(pd->qos_regmap[i], pd 239 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[3][i]); pd 240 drivers/soc/rockchip/pm_domains.c regmap_write(pd->qos_regmap[i], pd 242 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[4][i]); pd 248 drivers/soc/rockchip/pm_domains.c static bool rockchip_pmu_domain_is_on(struct rockchip_pm_domain *pd) pd 250 drivers/soc/rockchip/pm_domains.c struct rockchip_pmu *pmu = pd->pmu; pd 254 drivers/soc/rockchip/pm_domains.c if (pd->info->status_mask == 0) pd 255 drivers/soc/rockchip/pm_domains.c return !rockchip_pmu_domain_is_idle(pd); pd 260 drivers/soc/rockchip/pm_domains.c return !(val & pd->info->status_mask); pd 263 drivers/soc/rockchip/pm_domains.c static void rockchip_do_pmu_set_power_domain(struct rockchip_pm_domain *pd, pd 266 drivers/soc/rockchip/pm_domains.c struct rockchip_pmu *pmu = pd->pmu; pd 267 drivers/soc/rockchip/pm_domains.c struct generic_pm_domain *genpd = &pd->genpd; pd 270 drivers/soc/rockchip/pm_domains.c if (pd->info->pwr_mask == 0) pd 272 drivers/soc/rockchip/pm_domains.c else if (pd->info->pwr_w_mask) pd 274 drivers/soc/rockchip/pm_domains.c on ? pd->info->pwr_w_mask : pd 275 drivers/soc/rockchip/pm_domains.c (pd->info->pwr_mask | pd->info->pwr_w_mask)); pd 278 drivers/soc/rockchip/pm_domains.c pd->info->pwr_mask, on ? 0 : -1U); pd 282 drivers/soc/rockchip/pm_domains.c if (readx_poll_timeout_atomic(rockchip_pmu_domain_is_on, pd, is_on, pd 291 drivers/soc/rockchip/pm_domains.c static int rockchip_pd_power(struct rockchip_pm_domain *pd, bool power_on) pd 293 drivers/soc/rockchip/pm_domains.c struct rockchip_pmu *pmu = pd->pmu; pd 298 drivers/soc/rockchip/pm_domains.c if (rockchip_pmu_domain_is_on(pd) != power_on) { pd 299 drivers/soc/rockchip/pm_domains.c ret = clk_bulk_enable(pd->num_clks, pd->clks); pd 307 drivers/soc/rockchip/pm_domains.c rockchip_pmu_save_qos(pd); pd 310 drivers/soc/rockchip/pm_domains.c rockchip_pmu_set_idle_request(pd, true); pd 313 drivers/soc/rockchip/pm_domains.c rockchip_do_pmu_set_power_domain(pd, power_on); pd 317 drivers/soc/rockchip/pm_domains.c rockchip_pmu_set_idle_request(pd, false); pd 319 drivers/soc/rockchip/pm_domains.c rockchip_pmu_restore_qos(pd); pd 322 drivers/soc/rockchip/pm_domains.c clk_bulk_disable(pd->num_clks, pd->clks); pd 331 drivers/soc/rockchip/pm_domains.c struct rockchip_pm_domain *pd = to_rockchip_pd(domain); pd 333 drivers/soc/rockchip/pm_domains.c return rockchip_pd_power(pd, true); pd 338 drivers/soc/rockchip/pm_domains.c struct rockchip_pm_domain *pd = to_rockchip_pd(domain); pd 340 drivers/soc/rockchip/pm_domains.c return rockchip_pd_power(pd, false); pd 385 drivers/soc/rockchip/pm_domains.c struct rockchip_pm_domain *pd; pd 412 drivers/soc/rockchip/pm_domains.c pd = devm_kzalloc(pmu->dev, sizeof(*pd), GFP_KERNEL); pd 413 drivers/soc/rockchip/pm_domains.c if (!pd) pd 416 drivers/soc/rockchip/pm_domains.c pd->info = pd_info; pd 417 drivers/soc/rockchip/pm_domains.c pd->pmu = pmu; pd 419 drivers/soc/rockchip/pm_domains.c pd->num_clks = of_clk_get_parent_count(node); pd 420 drivers/soc/rockchip/pm_domains.c if (pd->num_clks > 0) { pd 421 drivers/soc/rockchip/pm_domains.c pd->clks = devm_kcalloc(pmu->dev, pd->num_clks, pd 422 drivers/soc/rockchip/pm_domains.c sizeof(*pd->clks), GFP_KERNEL); pd 423 drivers/soc/rockchip/pm_domains.c if (!pd->clks) pd 427 drivers/soc/rockchip/pm_domains.c node, pd->num_clks); pd 428 drivers/soc/rockchip/pm_domains.c pd->num_clks = 0; pd 431 drivers/soc/rockchip/pm_domains.c for (i = 0; i < pd->num_clks; i++) { pd 432 drivers/soc/rockchip/pm_domains.c pd->clks[i].clk = of_clk_get(node, i); pd 433 drivers/soc/rockchip/pm_domains.c if (IS_ERR(pd->clks[i].clk)) { pd 434 drivers/soc/rockchip/pm_domains.c error = PTR_ERR(pd->clks[i].clk); pd 442 drivers/soc/rockchip/pm_domains.c error = clk_bulk_prepare(pd->num_clks, pd->clks); pd 446 drivers/soc/rockchip/pm_domains.c pd->num_qos = of_count_phandle_with_args(node, "pm_qos", pd 449 drivers/soc/rockchip/pm_domains.c if (pd->num_qos > 0) { pd 450 drivers/soc/rockchip/pm_domains.c pd->qos_regmap = devm_kcalloc(pmu->dev, pd->num_qos, pd 451 drivers/soc/rockchip/pm_domains.c sizeof(*pd->qos_regmap), pd 453 drivers/soc/rockchip/pm_domains.c if (!pd->qos_regmap) { pd 459 drivers/soc/rockchip/pm_domains.c pd->qos_save_regs[j] = devm_kcalloc(pmu->dev, pd 460 drivers/soc/rockchip/pm_domains.c pd->num_qos, pd 463 drivers/soc/rockchip/pm_domains.c if (!pd->qos_save_regs[j]) { pd 469 drivers/soc/rockchip/pm_domains.c for (j = 0; j < pd->num_qos; j++) { pd 475 drivers/soc/rockchip/pm_domains.c pd->qos_regmap[j] = syscon_node_to_regmap(qos_node); pd 476 drivers/soc/rockchip/pm_domains.c if (IS_ERR(pd->qos_regmap[j])) { pd 485 drivers/soc/rockchip/pm_domains.c error = rockchip_pd_power(pd, true); pd 493 drivers/soc/rockchip/pm_domains.c pd->genpd.name = node->name; pd 494 drivers/soc/rockchip/pm_domains.c pd->genpd.power_off = rockchip_pd_power_off; pd 495 drivers/soc/rockchip/pm_domains.c pd->genpd.power_on = rockchip_pd_power_on; pd 496 drivers/soc/rockchip/pm_domains.c pd->genpd.attach_dev = rockchip_pd_attach_dev; pd 497 drivers/soc/rockchip/pm_domains.c pd->genpd.detach_dev = rockchip_pd_detach_dev; pd 498 drivers/soc/rockchip/pm_domains.c pd->genpd.flags = GENPD_FLAG_PM_CLK; pd 500 drivers/soc/rockchip/pm_domains.c pd->genpd.flags |= GENPD_FLAG_ACTIVE_WAKEUP; pd 501 drivers/soc/rockchip/pm_domains.c pm_genpd_init(&pd->genpd, NULL, false); pd 503 drivers/soc/rockchip/pm_domains.c pmu->genpd_data.domains[id] = &pd->genpd; pd 507 drivers/soc/rockchip/pm_domains.c clk_bulk_unprepare(pd->num_clks, pd->clks); pd 509 drivers/soc/rockchip/pm_domains.c clk_bulk_put(pd->num_clks, pd->clks); pd 513 drivers/soc/rockchip/pm_domains.c static void rockchip_pm_remove_one_domain(struct rockchip_pm_domain *pd) pd 521 drivers/soc/rockchip/pm_domains.c ret = pm_genpd_remove(&pd->genpd); pd 523 drivers/soc/rockchip/pm_domains.c dev_err(pd->pmu->dev, "failed to remove domain '%s' : %d - state may be inconsistent\n", pd 524 drivers/soc/rockchip/pm_domains.c pd->genpd.name, ret); pd 526 drivers/soc/rockchip/pm_domains.c clk_bulk_unprepare(pd->num_clks, pd->clks); pd 527 drivers/soc/rockchip/pm_domains.c clk_bulk_put(pd->num_clks, pd->clks); pd 530 drivers/soc/rockchip/pm_domains.c mutex_lock(&pd->pmu->mutex); pd 531 drivers/soc/rockchip/pm_domains.c pd->num_clks = 0; pd 532 drivers/soc/rockchip/pm_domains.c mutex_unlock(&pd->pmu->mutex); pd 540 drivers/soc/rockchip/pm_domains.c struct rockchip_pm_domain *pd; pd 546 drivers/soc/rockchip/pm_domains.c pd = to_rockchip_pd(genpd); pd 547 drivers/soc/rockchip/pm_domains.c rockchip_pm_remove_one_domain(pd); pd 32 drivers/soc/samsung/pm_domains.c struct generic_pm_domain pd; pd 38 drivers/soc/samsung/pm_domains.c struct exynos_pm_domain *pd; pd 43 drivers/soc/samsung/pm_domains.c pd = container_of(domain, struct exynos_pm_domain, pd); pd 44 drivers/soc/samsung/pm_domains.c base = pd->base; pd 46 drivers/soc/samsung/pm_domains.c pwr = power_on ? pd->local_pwr_cfg : 0; pd 52 drivers/soc/samsung/pm_domains.c while ((readl_relaxed(base + 0x4) & pd->local_pwr_cfg) != pwr) { pd 111 drivers/soc/samsung/pm_domains.c struct exynos_pm_domain *pd; pd 116 drivers/soc/samsung/pm_domains.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 117 drivers/soc/samsung/pm_domains.c if (!pd) { pd 121 drivers/soc/samsung/pm_domains.c pd->pd.name = exynos_get_domain_name(np); pd 122 drivers/soc/samsung/pm_domains.c if (!pd->pd.name) { pd 123 drivers/soc/samsung/pm_domains.c kfree(pd); pd 128 drivers/soc/samsung/pm_domains.c pd->base = of_iomap(np, 0); pd 129 drivers/soc/samsung/pm_domains.c if (!pd->base) { pd 131 drivers/soc/samsung/pm_domains.c kfree_const(pd->pd.name); pd 132 drivers/soc/samsung/pm_domains.c kfree(pd); pd 136 drivers/soc/samsung/pm_domains.c pd->pd.power_off = exynos_pd_power_off; pd 137 drivers/soc/samsung/pm_domains.c pd->pd.power_on = exynos_pd_power_on; pd 138 drivers/soc/samsung/pm_domains.c pd->local_pwr_cfg = pm_domain_cfg->local_pwr_cfg; pd 140 drivers/soc/samsung/pm_domains.c on = readl_relaxed(pd->base + 0x4) & pd->local_pwr_cfg; pd 142 drivers/soc/samsung/pm_domains.c pm_genpd_init(&pd->pd, NULL, !on); pd 143 drivers/soc/samsung/pm_domains.c of_genpd_add_provider_simple(np, &pd->pd); pd 43 drivers/soc/ti/ti_sci_pm_domains.c struct generic_pm_domain pd; pd 46 drivers/soc/ti/ti_sci_pm_domains.c #define genpd_to_ti_sci_pd(gpd) container_of(gpd, struct ti_sci_pm_domain, pd) pd 79 drivers/soc/ti/ti_sci_pm_domains.c struct generic_pm_domain *pd = pd_to_genpd(dev->pm_domain); pd 80 drivers/soc/ti/ti_sci_pm_domains.c struct ti_sci_pm_domain *ti_sci_genpd = genpd_to_ti_sci_pd(pd); pd 190 drivers/soc/ti/ti_sci_pm_domains.c ti_sci_pd->pd.name = "ti_sci_pd"; pd 192 drivers/soc/ti/ti_sci_pm_domains.c ti_sci_pd->pd.attach_dev = ti_sci_pd_attach_dev; pd 193 drivers/soc/ti/ti_sci_pm_domains.c ti_sci_pd->pd.detach_dev = ti_sci_pd_detach_dev; pd 195 drivers/soc/ti/ti_sci_pm_domains.c ti_sci_pd->pd.dev_ops.start = ti_sci_dev_start; pd 196 drivers/soc/ti/ti_sci_pm_domains.c ti_sci_pd->pd.dev_ops.stop = ti_sci_dev_stop; pd 198 drivers/soc/ti/ti_sci_pm_domains.c pm_genpd_init(&ti_sci_pd->pd, NULL, true); pd 200 drivers/soc/ti/ti_sci_pm_domains.c ret = of_genpd_add_provider_simple(np, &ti_sci_pd->pd); pd 75 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 80 drivers/soc/xilinx/zynqmp_pm_domains.c pd = container_of(domain, struct zynqmp_pm_domain, gpd); pd 81 drivers/soc/xilinx/zynqmp_pm_domains.c ret = eemi_ops->set_requirement(pd->node_id, pd 87 drivers/soc/xilinx/zynqmp_pm_domains.c __func__, domain->name, pd->node_id, ret); pd 108 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 115 drivers/soc/xilinx/zynqmp_pm_domains.c pd = container_of(domain, struct zynqmp_pm_domain, gpd); pd 118 drivers/soc/xilinx/zynqmp_pm_domains.c if (!(pd->flags & ZYNQMP_PM_DOMAIN_REQUESTED)) { pd 135 drivers/soc/xilinx/zynqmp_pm_domains.c ret = eemi_ops->set_requirement(pd->node_id, capabilities, 0, pd 143 drivers/soc/xilinx/zynqmp_pm_domains.c __func__, domain->name, pd->node_id, ret); pd 162 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 167 drivers/soc/xilinx/zynqmp_pm_domains.c pd = container_of(domain, struct zynqmp_pm_domain, gpd); pd 173 drivers/soc/xilinx/zynqmp_pm_domains.c ret = eemi_ops->request_node(pd->node_id, 0, 0, pd 178 drivers/soc/xilinx/zynqmp_pm_domains.c __func__, domain->name, pd->node_id, ret); pd 182 drivers/soc/xilinx/zynqmp_pm_domains.c pd->flags |= ZYNQMP_PM_DOMAIN_REQUESTED; pd 198 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 203 drivers/soc/xilinx/zynqmp_pm_domains.c pd = container_of(domain, struct zynqmp_pm_domain, gpd); pd 209 drivers/soc/xilinx/zynqmp_pm_domains.c ret = eemi_ops->release_node(pd->node_id); pd 213 drivers/soc/xilinx/zynqmp_pm_domains.c __func__, domain->name, pd->node_id, ret); pd 217 drivers/soc/xilinx/zynqmp_pm_domains.c pd->flags &= ~ZYNQMP_PM_DOMAIN_REQUESTED; pd 228 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 230 drivers/soc/xilinx/zynqmp_pm_domains.c pd = container_of(genpd_data->domains[0], struct zynqmp_pm_domain, gpd); pd 237 drivers/soc/xilinx/zynqmp_pm_domains.c if (pd[i].node_id == idx) pd 246 drivers/soc/xilinx/zynqmp_pm_domains.c if (pd[i].node_id == 0) { pd 247 drivers/soc/xilinx/zynqmp_pm_domains.c pd[i].node_id = idx; pd 264 drivers/soc/xilinx/zynqmp_pm_domains.c struct zynqmp_pm_domain *pd; pd 271 drivers/soc/xilinx/zynqmp_pm_domains.c pd = devm_kcalloc(dev, ZYNQMP_NUM_DOMAINS, sizeof(*pd), GFP_KERNEL); pd 272 drivers/soc/xilinx/zynqmp_pm_domains.c if (!pd) pd 286 drivers/soc/xilinx/zynqmp_pm_domains.c for (i = 0; i < ZYNQMP_NUM_DOMAINS; i++, pd++) { pd 287 drivers/soc/xilinx/zynqmp_pm_domains.c pd->node_id = 0; pd 288 drivers/soc/xilinx/zynqmp_pm_domains.c pd->gpd.name = kasprintf(GFP_KERNEL, "domain%d", i); pd 289 drivers/soc/xilinx/zynqmp_pm_domains.c pd->gpd.power_off = zynqmp_gpd_power_off; pd 290 drivers/soc/xilinx/zynqmp_pm_domains.c pd->gpd.power_on = zynqmp_gpd_power_on; pd 291 drivers/soc/xilinx/zynqmp_pm_domains.c pd->gpd.attach_dev = zynqmp_gpd_attach_dev; pd 292 drivers/soc/xilinx/zynqmp_pm_domains.c pd->gpd.detach_dev = zynqmp_gpd_detach_dev; pd 294 drivers/soc/xilinx/zynqmp_pm_domains.c domains[i] = &pd->gpd; pd 297 drivers/soc/xilinx/zynqmp_pm_domains.c pm_genpd_init(&pd->gpd, NULL, true); pd 54 drivers/spi/spi-butterfly.c struct pardevice *pd; pd 178 drivers/spi/spi-butterfly.c struct pardevice *pd; pd 218 drivers/spi/spi-butterfly.c pd = parport_register_dev_model(p, "spi_butterfly", &butterfly_cb, 0); pd 219 drivers/spi/spi-butterfly.c if (!pd) { pd 223 drivers/spi/spi-butterfly.c pp->pd = pd; pd 225 drivers/spi/spi-butterfly.c status = parport_claim(pd); pd 279 drivers/spi/spi-butterfly.c parport_release(pp->pd); pd 281 drivers/spi/spi-butterfly.c parport_unregister_device(pd); pd 308 drivers/spi/spi-butterfly.c parport_release(pp->pd); pd 309 drivers/spi/spi-butterfly.c parport_unregister_device(pp->pd); pd 73 drivers/spi/spi-lm70llp.c struct pardevice *pd; pd 190 drivers/spi/spi-lm70llp.c struct pardevice *pd; pd 227 drivers/spi/spi-lm70llp.c pd = parport_register_dev_model(p, DRVNAME, &lm70llp_cb, 0); pd 229 drivers/spi/spi-lm70llp.c if (!pd) { pd 233 drivers/spi/spi-lm70llp.c pp->pd = pd; pd 235 drivers/spi/spi-lm70llp.c status = parport_claim(pd); pd 244 drivers/spi/spi-lm70llp.c dev_warn(&pd->dev, "spi_bitbang_start failed with status %d\n", pd 272 drivers/spi/spi-lm70llp.c dev_warn(&pd->dev, "spi_new_device failed\n"); pd 287 drivers/spi/spi-lm70llp.c parport_release(pp->pd); pd 289 drivers/spi/spi-lm70llp.c parport_unregister_device(pd); pd 309 drivers/spi/spi-lm70llp.c parport_release(pp->pd); pd 310 drivers/spi/spi-lm70llp.c parport_unregister_device(pp->pd); pd 2093 drivers/spi/spi-pl022.c struct pl022_ssp_controller *pd; pd 2101 drivers/spi/spi-pl022.c pd = devm_kzalloc(dev, sizeof(struct pl022_ssp_controller), GFP_KERNEL); pd 2102 drivers/spi/spi-pl022.c if (!pd) pd 2105 drivers/spi/spi-pl022.c pd->bus_id = -1; pd 2106 drivers/spi/spi-pl022.c pd->enable_dma = 1; pd 2108 drivers/spi/spi-pl022.c pd->num_chipselect = tmp; pd 2110 drivers/spi/spi-pl022.c &pd->autosuspend_delay); pd 2111 drivers/spi/spi-pl022.c pd->rt = of_property_read_bool(np, "pl022,rt"); pd 2113 drivers/spi/spi-pl022.c return pd; pd 475 drivers/staging/comedi/drivers/comedi_test.c unsigned short *pd; pd 477 drivers/staging/comedi/drivers/comedi_test.c pd = &devpriv->ao_loopbacks[chan]; pd 479 drivers/staging/comedi/drivers/comedi_test.c if (!comedi_buf_read_samples(s, pd, 1)) { pd 552 drivers/staging/fieldbus/anybuss/host.c struct area_priv *pd = &t->area_pd; pd 556 drivers/staging/fieldbus/anybuss/host.c if (atomic_read(&cd->ind_ab) & pd->flags) { pd 567 drivers/staging/fieldbus/anybuss/host.c struct area_priv *pd = &t->area_pd; pd 574 drivers/staging/fieldbus/anybuss/host.c if (!(atomic_read(&cd->ind_ab) & pd->flags)) { pd 584 drivers/staging/fieldbus/anybuss/host.c if (pd->is_write) pd 585 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_write(cd->regmap, pd->addr, pd->buf, pd 586 drivers/staging/fieldbus/anybuss/host.c pd->count); pd 588 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_read(cd->regmap, pd->addr, pd->buf, pd 589 drivers/staging/fieldbus/anybuss/host.c pd->count); pd 592 drivers/staging/fieldbus/anybuss/host.c ind_ap |= pd->flags; pd 602 drivers/staging/fieldbus/anybuss/host.c struct area_priv *pd = &t->area_pd; pd 611 drivers/staging/fieldbus/anybuss/host.c ind_ap |= pd->flags | IND_AP_ACTION | IND_AP_LOCK; pd 698 drivers/staging/fieldbus/anybuss/host.c struct mbox_priv *pd = &t->mbox_pd; pd 711 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_read(cd->regmap, MBOX_OUT_AREA, &pd->hdr, pd 712 drivers/staging/fieldbus/anybuss/host.c sizeof(pd->hdr)); pd 713 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_read(cd->regmap, MBOX_OUT_AREA + sizeof(pd->hdr), pd 714 drivers/staging/fieldbus/anybuss/host.c pd->msg, pd->msg_in_sz); pd 722 drivers/staging/fieldbus/anybuss/host.c struct mbox_priv *pd = &t->mbox_pd; pd 736 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_write(cd->regmap, MBOX_IN_AREA, &pd->hdr, pd 737 drivers/staging/fieldbus/anybuss/host.c sizeof(pd->hdr)); pd 738 drivers/staging/fieldbus/anybuss/host.c regmap_bulk_write(cd->regmap, MBOX_IN_AREA + sizeof(pd->hdr), pd 739 drivers/staging/fieldbus/anybuss/host.c pd->msg, pd->msg_out_sz); pd 811 drivers/staging/fieldbus/anybuss/host.c struct mbox_priv *pd; pd 824 drivers/staging/fieldbus/anybuss/host.c pd = &t->mbox_pd; pd 825 drivers/staging/fieldbus/anybuss/host.c h = &pd->hdr; pd 841 drivers/staging/fieldbus/anybuss/host.c memcpy(pd->msg, msg_out, msg_out_sz); pd 842 drivers/staging/fieldbus/anybuss/host.c pd->msg_out_sz = msg_out_sz; pd 843 drivers/staging/fieldbus/anybuss/host.c pd->msg_in_sz = msg_in_sz; pd 851 drivers/staging/fieldbus/anybuss/host.c err = mbox_cmd_err(cd->dev, pd); pd 854 drivers/staging/fieldbus/anybuss/host.c memcpy(msg_in, pd->msg, msg_in_sz); pd 70 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd = card->proclog; pd 74 drivers/staging/isdn/hysdn/hysdn_proclog.c if (!pd) pd 77 drivers/staging/isdn/hysdn/hysdn_proclog.c cp = pd->logtmp; pd 87 drivers/staging/isdn/hysdn/hysdn_proclog.c printk(KERN_INFO "%s", pd->logtmp); pd 89 drivers/staging/isdn/hysdn/hysdn_proclog.c put_log_buffer(card, pd->logtmp); pd 103 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd = card->proclog; pd 106 drivers/staging/isdn/hysdn/hysdn_proclog.c if (!pd) pd 112 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd->if_used <= 0) pd 119 drivers/staging/isdn/hysdn/hysdn_proclog.c ib->proc_ctrl = pd; /* point to own control structure */ pd 121 drivers/staging/isdn/hysdn/hysdn_proclog.c ib->usage_cnt = pd->if_used; pd 122 drivers/staging/isdn/hysdn/hysdn_proclog.c if (!pd->log_head) pd 123 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_head = ib; /* new head */ pd 125 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_tail->next = ib; /* follows existing messages */ pd 126 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_tail = ib; /* new tail */ pd 129 drivers/staging/isdn/hysdn/hysdn_proclog.c while (pd->log_head->next) { pd 130 drivers/staging/isdn/hysdn/hysdn_proclog.c if ((pd->log_head->usage_cnt <= 0) && pd 131 drivers/staging/isdn/hysdn/hysdn_proclog.c (pd->log_head->next->usage_cnt <= 0)) { pd 132 drivers/staging/isdn/hysdn/hysdn_proclog.c ib = pd->log_head; pd 133 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_head = pd->log_head->next; pd 142 drivers/staging/isdn/hysdn/hysdn_proclog.c wake_up_interruptible(&(pd->rd_queue)); /* announce new entry */ pd 177 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd = card->proclog; pd 181 drivers/staging/isdn/hysdn/hysdn_proclog.c wait_event_interruptible(pd->rd_queue, (inf = pd 211 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd = card->proclog; pd 216 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->if_used++; pd 217 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd->log_head) pd 218 drivers/staging/isdn/hysdn/hysdn_proclog.c filep->private_data = &pd->log_tail->next; pd 220 drivers/staging/isdn/hysdn/hysdn_proclog.c filep->private_data = &pd->log_head; pd 241 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd; pd 254 drivers/staging/isdn/hysdn/hysdn_proclog.c pd = (struct procdata *) inf->proc_ctrl; /* still entries there */ pd 258 drivers/staging/isdn/hysdn/hysdn_proclog.c pd = card->proclog; /* pointer to procfs log */ pd 260 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd) pd 261 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->if_used--; /* decrement interface usage count by one */ pd 268 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd) pd 269 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd->if_used <= 0) /* delete buffers if last file closed */ pd 270 drivers/staging/isdn/hysdn/hysdn_proclog.c while (pd->log_head) { pd 271 drivers/staging/isdn/hysdn/hysdn_proclog.c inf = pd->log_head; pd 272 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log_head = pd->log_head->next; pd 289 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd = card->proclog; pd 294 drivers/staging/isdn/hysdn/hysdn_proclog.c poll_wait(file, &(pd->rd_queue), wait); pd 324 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd; pd 328 drivers/staging/isdn/hysdn/hysdn_proclog.c if ((pd = kzalloc(sizeof(struct procdata), GFP_KERNEL)) != NULL) { pd 329 drivers/staging/isdn/hysdn/hysdn_proclog.c sprintf(pd->log_name, "%s%d", PROC_LOG_BASENAME, card->myid); pd 330 drivers/staging/isdn/hysdn/hysdn_proclog.c pd->log = proc_create_data(pd->log_name, pd 334 drivers/staging/isdn/hysdn/hysdn_proclog.c init_waitqueue_head(&(pd->rd_queue)); pd 336 drivers/staging/isdn/hysdn/hysdn_proclog.c card->proclog = (void *) pd; /* remember procfs structure */ pd 349 drivers/staging/isdn/hysdn/hysdn_proclog.c struct procdata *pd; pd 351 drivers/staging/isdn/hysdn/hysdn_proclog.c if ((pd = (struct procdata *) card->proclog) != NULL) { pd 352 drivers/staging/isdn/hysdn/hysdn_proclog.c if (pd->log) pd 353 drivers/staging/isdn/hysdn/hysdn_proclog.c remove_proc_entry(pd->log_name, hysdn_proc_entry); pd 354 drivers/staging/isdn/hysdn/hysdn_proclog.c kfree(pd); /* release memory */ pd 781 drivers/usb/gadget/function/f_uvc.c struct uvc_processing_unit_descriptor *pd; pd 809 drivers/usb/gadget/function/f_uvc.c pd = &opts->uvc_processing; pd 810 drivers/usb/gadget/function/f_uvc.c pd->bLength = UVC_DT_PROCESSING_UNIT_SIZE(2); pd 811 drivers/usb/gadget/function/f_uvc.c pd->bDescriptorType = USB_DT_CS_INTERFACE; pd 812 drivers/usb/gadget/function/f_uvc.c pd->bDescriptorSubType = UVC_VC_PROCESSING_UNIT; pd 813 drivers/usb/gadget/function/f_uvc.c pd->bUnitID = 2; pd 814 drivers/usb/gadget/function/f_uvc.c pd->bSourceID = 1; pd 815 drivers/usb/gadget/function/f_uvc.c pd->wMaxMultiplier = cpu_to_le16(16*1024); pd 816 drivers/usb/gadget/function/f_uvc.c pd->bControlSize = 2; pd 817 drivers/usb/gadget/function/f_uvc.c pd->bmControls[0] = 1; pd 818 drivers/usb/gadget/function/f_uvc.c pd->bmControls[1] = 0; pd 819 drivers/usb/gadget/function/f_uvc.c pd->iProcessing = 0; pd 843 drivers/usb/gadget/function/f_uvc.c ctl_cls[2] = (struct uvc_descriptor_header *)pd; pd 853 drivers/usb/gadget/function/f_uvc.c ctl_cls[2] = (struct uvc_descriptor_header *)pd; pd 267 drivers/usb/gadget/function/uvc_configfs.c struct uvc_processing_unit_descriptor *pd; \ pd 274 drivers/usb/gadget/function/uvc_configfs.c pd = &opts->uvc_processing; \ pd 277 drivers/usb/gadget/function/uvc_configfs.c result = sprintf(page, "%u\n", le##bits##_to_cpu(pd->aname)); \ pd 300 drivers/usb/gadget/function/uvc_configfs.c struct uvc_processing_unit_descriptor *pd; pd 308 drivers/usb/gadget/function/uvc_configfs.c pd = &opts->uvc_processing; pd 311 drivers/usb/gadget/function/uvc_configfs.c for (result = 0, i = 0; i < pd->bControlSize; ++i) { pd 312 drivers/usb/gadget/function/uvc_configfs.c result += sprintf(pg, "%u\n", pd->bmControls[i]); pd 296 drivers/usb/gadget/udc/bcm63xx_udc.c struct bcm63xx_usbd_platform_data *pd; pd 872 drivers/usb/gadget/udc/bcm63xx_udc.c u32 val, portmask = BIT(udc->pd->port_no); pd 912 drivers/usb/gadget/udc/bcm63xx_udc.c u32 val, portmask = BIT(udc->pd->port_no); pd 2283 drivers/usb/gadget/udc/bcm63xx_udc.c struct bcm63xx_usbd_platform_data *pd = dev_get_platdata(dev); pd 2294 drivers/usb/gadget/udc/bcm63xx_udc.c udc->pd = pd; pd 2296 drivers/usb/gadget/udc/bcm63xx_udc.c if (!pd) { pd 2317 drivers/usb/gadget/udc/bcm63xx_udc.c if (!pd->use_fullspeed && !use_fullspeed) pd 144 drivers/usb/gadget/udc/s3c-hsudc.c struct s3c24xx_hsudc_platdata *pd; pd 293 drivers/usb/gadget/udc/s3c-hsudc.c for (epnum = 0; epnum < hsudc->pd->epnum; epnum++) { pd 1036 drivers/usb/gadget/udc/s3c-hsudc.c for (epnum = 0; epnum < hsudc->pd->epnum; epnum++) pd 1113 drivers/usb/gadget/udc/s3c-hsudc.c for (ep_idx = 0; ep_idx < hsudc->pd->epnum; ep_idx++) { pd 1192 drivers/usb/gadget/udc/s3c-hsudc.c if (hsudc->pd->gpio_init) pd 1193 drivers/usb/gadget/udc/s3c-hsudc.c hsudc->pd->gpio_init(); pd 1217 drivers/usb/gadget/udc/s3c-hsudc.c if (hsudc->pd->gpio_uninit) pd 1218 drivers/usb/gadget/udc/s3c-hsudc.c hsudc->pd->gpio_uninit(); pd 1268 drivers/usb/gadget/udc/s3c-hsudc.c struct s3c24xx_hsudc_platdata *pd = dev_get_platdata(&pdev->dev); pd 1272 drivers/usb/gadget/udc/s3c-hsudc.c sizeof(struct s3c_hsudc_ep) * pd->epnum, pd 1279 drivers/usb/gadget/udc/s3c-hsudc.c hsudc->pd = dev_get_platdata(&pdev->dev); pd 209 drivers/usb/host/ehci-orion.c struct orion_ehci_data *pd = dev_get_platdata(&pdev->dev); pd 290 drivers/usb/host/ehci-orion.c phy_version = pd->phy_version; pd 310 drivers/video/backlight/ams369fg06.c struct lcd_platform_data *pd; pd 313 drivers/video/backlight/ams369fg06.c pd = lcd->lcd_pd; pd 316 drivers/video/backlight/ams369fg06.c if (pd->power_on) { pd 317 drivers/video/backlight/ams369fg06.c pd->power_on(lcd->ld, 1); pd 318 drivers/video/backlight/ams369fg06.c msleep(pd->power_on_delay); pd 321 drivers/video/backlight/ams369fg06.c if (!pd->reset) { pd 326 drivers/video/backlight/ams369fg06.c pd->reset(lcd->ld); pd 327 drivers/video/backlight/ams369fg06.c msleep(pd->reset_delay); pd 354 drivers/video/backlight/ams369fg06.c struct lcd_platform_data *pd; pd 356 drivers/video/backlight/ams369fg06.c pd = lcd->lcd_pd; pd 364 drivers/video/backlight/ams369fg06.c msleep(pd->power_off_delay); pd 366 drivers/video/backlight/ams369fg06.c if (pd->power_on) pd 367 drivers/video/backlight/ams369fg06.c pd->power_on(lcd->ld, 0); pd 224 drivers/video/backlight/lms501kf03.c struct lcd_platform_data *pd; pd 226 drivers/video/backlight/lms501kf03.c pd = lcd->lcd_pd; pd 228 drivers/video/backlight/lms501kf03.c if (!pd->power_on) { pd 233 drivers/video/backlight/lms501kf03.c pd->power_on(lcd->ld, 1); pd 234 drivers/video/backlight/lms501kf03.c msleep(pd->power_on_delay); pd 236 drivers/video/backlight/lms501kf03.c if (!pd->reset) { pd 241 drivers/video/backlight/lms501kf03.c pd->reset(lcd->ld); pd 242 drivers/video/backlight/lms501kf03.c msleep(pd->reset_delay); pd 262 drivers/video/backlight/lms501kf03.c struct lcd_platform_data *pd; pd 264 drivers/video/backlight/lms501kf03.c pd = lcd->lcd_pd; pd 272 drivers/video/backlight/lms501kf03.c msleep(pd->power_off_delay); pd 274 drivers/video/backlight/lms501kf03.c pd->power_on(lcd->ld, 0); pd 171 drivers/video/backlight/lp855x_bl.c struct lp855x_platform_data *pd = lp->pdata; pd 197 drivers/video/backlight/lp855x_bl.c val = pd->initial_brightness; pd 202 drivers/video/backlight/lp855x_bl.c val = pd->device_control; pd 207 drivers/video/backlight/lp855x_bl.c if (pd->size_program > 0) { pd 208 drivers/video/backlight/lp855x_bl.c for (i = 0; i < pd->size_program; i++) { pd 209 drivers/video/backlight/lp855x_bl.c addr = pd->rom_data[i].addr; pd 210 drivers/video/backlight/lp855x_bl.c val = pd->rom_data[i].val; pd 443 drivers/video/console/sticore.c if (sti->pd) { pd 456 drivers/video/console/sticore.c newhpa = pci_resource_start (sti->pd, (offs - PCI_BASE_ADDRESS_0) / 4); pd 819 drivers/video/console/sticore.c struct pci_dev *pd) pd 857 drivers/video/console/sticore.c sti->pd = pd; pd 905 drivers/video/console/sticore.c if (sti->pd) { pd 907 drivers/video/console/sticore.c rom_base = pci_resource_start(sti->pd, PCI_ROM_RESOURCE); pd 908 drivers/video/console/sticore.c pci_write_config_dword(sti->pd, PCI_ROM_ADDRESS, rom_base & ~PCI_ROM_ADDRESS_ENABLE); pd 964 drivers/video/console/sticore.c static int sticore_pci_init(struct pci_dev *pd, const struct pci_device_id *ent) pd 972 drivers/video/console/sticore.c err = pci_enable_device(pd); pd 974 drivers/video/console/sticore.c dev_err(&pd->dev, "Cannot enable PCI device\n"); pd 978 drivers/video/console/sticore.c fb_base = pci_resource_start(pd, 0); pd 979 drivers/video/console/sticore.c fb_len = pci_resource_len(pd, 0); pd 980 drivers/video/console/sticore.c rom_base = pci_resource_start(pd, PCI_ROM_RESOURCE); pd 981 drivers/video/console/sticore.c rom_len = pci_resource_len(pd, PCI_ROM_RESOURCE); pd 983 drivers/video/console/sticore.c pci_write_config_dword(pd, PCI_ROM_ADDRESS, rom_base | PCI_ROM_ADDRESS_ENABLE); pd 993 drivers/video/console/sticore.c sti = sti_try_rom_generic(rom_base, fb_base, pd); pd 996 drivers/video/console/sticore.c print_pci_hwpath(pd, pa_path); pd 1002 drivers/video/console/sticore.c pci_name(pd)); pd 1011 drivers/video/console/sticore.c static void __exit sticore_pci_remove(struct pci_dev *pd) pd 149 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd; pd 773 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd) pd 806 drivers/video/fbdev/au1200fb.c if (pd->panel_shutdown) pd 807 drivers/video/fbdev/au1200fb.c pd->panel_shutdown(); pd 866 drivers/video/fbdev/au1200fb.c if (pd->panel_init) pd 867 drivers/video/fbdev/au1200fb.c pd->panel_init(); pd 1211 drivers/video/fbdev/au1200fb.c au1200_setpanel(panel, fbdev->pd); pd 1217 drivers/video/fbdev/au1200fb.c au1200_setpanel(NULL, fbdev->pd); pd 1462 drivers/video/fbdev/au1200fb.c au1200_setpanel(newpanel, fbdev->pd); pd 1578 drivers/video/fbdev/au1200fb.c static int au1200fb_setup(struct au1200fb_platdata *pd) pd 1604 drivers/video/fbdev/au1200fb.c panel_idx = pd->panel_index(); pd 1647 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd; pd 1653 drivers/video/fbdev/au1200fb.c pd = dev->dev.platform_data; pd 1654 drivers/video/fbdev/au1200fb.c if (!pd) pd 1658 drivers/video/fbdev/au1200fb.c if (au1200fb_setup(pd)) pd 1685 drivers/video/fbdev/au1200fb.c fbdev->pd = pd; pd 1740 drivers/video/fbdev/au1200fb.c platform_set_drvdata(dev, pd); pd 1743 drivers/video/fbdev/au1200fb.c au1200_setpanel(panel, pd); pd 1767 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd = platform_get_drvdata(dev); pd 1772 drivers/video/fbdev/au1200fb.c au1200_setpanel(NULL, pd); pd 1795 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd = dev_get_drvdata(dev); pd 1796 drivers/video/fbdev/au1200fb.c au1200_setpanel(NULL, pd); pd 1806 drivers/video/fbdev/au1200fb.c struct au1200fb_platdata *pd = dev_get_drvdata(dev); pd 1811 drivers/video/fbdev/au1200fb.c au1200_setpanel(panel, pd); pd 1270 drivers/video/fbdev/nvidia/nvidia.c static int nvidiafb_probe(struct pci_dev *pd, const struct pci_device_id *ent) pd 1278 drivers/video/fbdev/nvidia/nvidia.c assert(pd != NULL); pd 1280 drivers/video/fbdev/nvidia/nvidia.c info = framebuffer_alloc(sizeof(struct nvidia_par), &pd->dev); pd 1286 drivers/video/fbdev/nvidia/nvidia.c par->pci_dev = pd; pd 1292 drivers/video/fbdev/nvidia/nvidia.c if (pci_enable_device(pd)) { pd 1297 drivers/video/fbdev/nvidia/nvidia.c if (pci_request_regions(pd, "nvidiafb")) { pd 1313 drivers/video/fbdev/nvidia/nvidia.c pci_read_config_word(pd, PCI_COMMAND, &cmd); pd 1315 drivers/video/fbdev/nvidia/nvidia.c pci_write_config_word(pd, PCI_COMMAND, cmd); pd 1317 drivers/video/fbdev/nvidia/nvidia.c nvidiafb_fix.mmio_start = pci_resource_start(pd, 0); pd 1318 drivers/video/fbdev/nvidia/nvidia.c nvidiafb_fix.smem_start = pci_resource_start(pd, 1); pd 1319 drivers/video/fbdev/nvidia/nvidia.c nvidiafb_fix.mmio_len = pci_resource_len(pd, 0); pd 1336 drivers/video/fbdev/nvidia/nvidia.c sprintf(nvidiafb_fix.id, "NV%x", (pd->device & 0x0ff0) >> 4); pd 1385 drivers/video/fbdev/nvidia/nvidia.c pci_set_drvdata(pd, info); pd 1412 drivers/video/fbdev/nvidia/nvidia.c pci_release_regions(pd); pd 1421 drivers/video/fbdev/nvidia/nvidia.c static void nvidiafb_remove(struct pci_dev *pd) pd 1423 drivers/video/fbdev/nvidia/nvidia.c struct fb_info *info = pci_get_drvdata(pd); pd 1436 drivers/video/fbdev/nvidia/nvidia.c pci_release_regions(pd); pd 202 drivers/video/fbdev/omap/lcd_mipid.c struct mipid_platform_data *pd = md->spi->dev.platform_data; pd 204 drivers/video/fbdev/omap/lcd_mipid.c if (pd->get_bklight_max == NULL || pd->set_bklight_level == NULL) pd 206 drivers/video/fbdev/omap/lcd_mipid.c if (level > pd->get_bklight_max(pd)) pd 212 drivers/video/fbdev/omap/lcd_mipid.c pd->set_bklight_level(pd, level); pd 220 drivers/video/fbdev/omap/lcd_mipid.c struct mipid_platform_data *pd = md->spi->dev.platform_data; pd 222 drivers/video/fbdev/omap/lcd_mipid.c if (pd->get_bklight_level == NULL) pd 224 drivers/video/fbdev/omap/lcd_mipid.c return pd->get_bklight_level(pd); pd 230 drivers/video/fbdev/omap/lcd_mipid.c struct mipid_platform_data *pd = md->spi->dev.platform_data; pd 232 drivers/video/fbdev/omap/lcd_mipid.c if (pd->get_bklight_max == NULL) pd 235 drivers/video/fbdev/omap/lcd_mipid.c return pd->get_bklight_max(pd); pd 1729 drivers/video/fbdev/riva/fbdev.c static int riva_get_EDID_OF(struct fb_info *info, struct pci_dev *pd) pd 1740 drivers/video/fbdev/riva/fbdev.c dp = pci_device_to_OF_node(pd); pd 1856 drivers/video/fbdev/riva/fbdev.c static u32 riva_get_arch(struct pci_dev *pd) pd 1860 drivers/video/fbdev/riva/fbdev.c switch (pd->device & 0x0ff0) { pd 1894 drivers/video/fbdev/riva/fbdev.c static int rivafb_probe(struct pci_dev *pd, const struct pci_device_id *ent) pd 1901 drivers/video/fbdev/riva/fbdev.c assert(pd != NULL); pd 1903 drivers/video/fbdev/riva/fbdev.c info = framebuffer_alloc(sizeof(struct riva_par), &pd->dev); pd 1909 drivers/video/fbdev/riva/fbdev.c default_par->pdev = pd; pd 1917 drivers/video/fbdev/riva/fbdev.c ret = pci_enable_device(pd); pd 1923 drivers/video/fbdev/riva/fbdev.c ret = pci_request_regions(pd, "rivafb"); pd 1930 drivers/video/fbdev/riva/fbdev.c default_par->riva.Architecture = riva_get_arch(pd); pd 1932 drivers/video/fbdev/riva/fbdev.c default_par->Chipset = (pd->vendor << 16) | pd->device; pd 1943 drivers/video/fbdev/riva/fbdev.c sprintf(rivafb_fix.id, "NV%x", (pd->device & 0x0ff0) >> 4); pd 1953 drivers/video/fbdev/riva/fbdev.c rivafb_fix.mmio_len = pci_resource_len(pd, 0); pd 1954 drivers/video/fbdev/riva/fbdev.c rivafb_fix.smem_len = pci_resource_len(pd, 1); pd 1960 drivers/video/fbdev/riva/fbdev.c pci_read_config_word(pd, PCI_COMMAND, &cmd); pd 1962 drivers/video/fbdev/riva/fbdev.c pci_write_config_word(pd, PCI_COMMAND, cmd); pd 1965 drivers/video/fbdev/riva/fbdev.c rivafb_fix.mmio_start = pci_resource_start(pd, 0); pd 1966 drivers/video/fbdev/riva/fbdev.c rivafb_fix.smem_start = pci_resource_start(pd, 1); pd 2023 drivers/video/fbdev/riva/fbdev.c riva_get_EDID(info, pd); pd 2035 drivers/video/fbdev/riva/fbdev.c pci_set_drvdata(pd, info); pd 2068 drivers/video/fbdev/riva/fbdev.c pci_release_regions(pd); pd 2078 drivers/video/fbdev/riva/fbdev.c static void rivafb_remove(struct pci_dev *pd) pd 2080 drivers/video/fbdev/riva/fbdev.c struct fb_info *info = pci_get_drvdata(pd); pd 2098 drivers/video/fbdev/riva/fbdev.c pci_release_regions(pd); pd 1360 drivers/video/fbdev/s3c-fb.c struct s3c_fb_platdata *pd; pd 1375 drivers/video/fbdev/s3c-fb.c pd = dev_get_platdata(&pdev->dev); pd 1376 drivers/video/fbdev/s3c-fb.c if (!pd) { pd 1388 drivers/video/fbdev/s3c-fb.c sfb->pdata = pd; pd 1442 drivers/video/fbdev/s3c-fb.c pd->setup_gpio(); pd 1444 drivers/video/fbdev/s3c-fb.c writel(pd->vidcon1, sfb->regs + VIDCON1); pd 1473 drivers/video/fbdev/s3c-fb.c if (!pd->win[win]) pd 1566 drivers/video/fbdev/s3c-fb.c struct s3c_fb_platdata *pd = sfb->pdata; pd 1579 drivers/video/fbdev/s3c-fb.c pd->setup_gpio(); pd 1580 drivers/video/fbdev/s3c-fb.c writel(pd->vidcon1, sfb->regs + VIDCON1); pd 1641 drivers/video/fbdev/s3c-fb.c struct s3c_fb_platdata *pd = sfb->pdata; pd 1649 drivers/video/fbdev/s3c-fb.c pd->setup_gpio(); pd 1650 drivers/video/fbdev/s3c-fb.c writel(pd->vidcon1, sfb->regs + VIDCON1); pd 34 drivers/video/fbdev/sh7760fb.c struct sh7760fb_platdata *pd; /* display information */ pd 73 drivers/video/fbdev/sh7760fb.c struct sh7760fb_platdata *pd = par->pd; pd 88 drivers/video/fbdev/sh7760fb.c if (pd->blank) pd 89 drivers/video/fbdev/sh7760fb.c pd->blank(blank); pd 173 drivers/video/fbdev/sh7760fb.c ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, NULL); pd 199 drivers/video/fbdev/sh7760fb.c struct fb_videomode *vm = par->pd->def_mode; pd 206 drivers/video/fbdev/sh7760fb.c par->rot = par->pd->rotate; pd 225 drivers/video/fbdev/sh7760fb.c ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, &gray); pd 234 drivers/video/fbdev/sh7760fb.c lddfr = par->pd->lddfr | (1 << 8); pd 236 drivers/video/fbdev/sh7760fb.c lddfr = par->pd->lddfr & ~(1 << 8); pd 239 drivers/video/fbdev/sh7760fb.c ldmtr = par->pd->ldmtr; pd 249 drivers/video/fbdev/sh7760fb.c iowrite16(par->pd->ldickr, par->base + LDICKR); /* pixclock */ pd 253 drivers/video/fbdev/sh7760fb.c iowrite16(par->pd->ldpmmr, par->base + LDPMMR); /* Power Management */ pd 254 drivers/video/fbdev/sh7760fb.c iowrite16(par->pd->ldpspr, par->base + LDPSPR); /* Power Supply Ctrl */ pd 266 drivers/video/fbdev/sh7760fb.c iowrite16(par->pd->ldaclnr, par->base + LDACLNR); pd 334 drivers/video/fbdev/sh7760fb.c (par->pd->ldickr >> 12) & 3, par->pd->ldickr & 0x1f); pd 335 drivers/video/fbdev/sh7760fb.c dev_dbg(info->dev, "ldpmmr: 0x%04x ldpspr: 0x%04x\n", par->pd->ldpmmr, pd 336 drivers/video/fbdev/sh7760fb.c par->pd->ldpspr); pd 384 drivers/video/fbdev/sh7760fb.c ret = sh7760fb_get_color_info(info->dev, par->pd->lddfr, &bpp, NULL); pd 451 drivers/video/fbdev/sh7760fb.c par->pd = pdev->dev.platform_data; pd 452 drivers/video/fbdev/sh7760fb.c if (!par->pd) { pd 485 drivers/video/fbdev/sh7760fb.c fb_videomode_to_var(&info->var, par->pd->def_mode); pd 148 drivers/video/fbdev/simplefb.c struct simplefb_platform_data *pd = dev_get_platdata(&pdev->dev); pd 151 drivers/video/fbdev/simplefb.c params->width = pd->width; pd 152 drivers/video/fbdev/simplefb.c params->height = pd->height; pd 153 drivers/video/fbdev/simplefb.c params->stride = pd->stride; pd 157 drivers/video/fbdev/simplefb.c if (strcmp(pd->format, simplefb_formats[i].name)) pd 716 drivers/video/fbdev/sm501fb.c struct sm501_platdata_fbsub *pd = fbi->pdata->fb_pnl; pd 735 drivers/video/fbdev/sm501fb.c if (!(pd->flags & SM501FB_FLAG_PANEL_NO_VBIASEN)) { pd 736 drivers/video/fbdev/sm501fb.c if (pd->flags & SM501FB_FLAG_PANEL_INV_VBIASEN) pd 746 drivers/video/fbdev/sm501fb.c if (!(pd->flags & SM501FB_FLAG_PANEL_NO_FPEN)) { pd 747 drivers/video/fbdev/sm501fb.c if (pd->flags & SM501FB_FLAG_PANEL_INV_FPEN) pd 758 drivers/video/fbdev/sm501fb.c if (!(pd->flags & SM501FB_FLAG_PANEL_NO_FPEN)) { pd 759 drivers/video/fbdev/sm501fb.c if (pd->flags & SM501FB_FLAG_PANEL_INV_FPEN) pd 769 drivers/video/fbdev/sm501fb.c if (!(pd->flags & SM501FB_FLAG_PANEL_NO_VBIASEN)) { pd 770 drivers/video/fbdev/sm501fb.c if (pd->flags & SM501FB_FLAG_PANEL_INV_VBIASEN) pd 1677 drivers/video/fbdev/sm501fb.c struct sm501_platdata_fbsub *pd; pd 1686 drivers/video/fbdev/sm501fb.c pd = info->pdata->fb_crt; pd 1699 drivers/video/fbdev/sm501fb.c pd = info->pdata->fb_pnl; pd 1705 drivers/video/fbdev/sm501fb.c pd = NULL; /* stop compiler warnings */ pd 1730 drivers/video/fbdev/sm501fb.c if ((pd->flags & SM501FB_FLAG_USE_HWCURSOR) == 0) pd 1772 drivers/video/fbdev/sm501fb.c if (enable && (pd->flags & SM501FB_FLAG_USE_INIT_MODE) && 0) { pd 1775 drivers/video/fbdev/sm501fb.c if (pd->def_mode) { pd 1777 drivers/video/fbdev/sm501fb.c fb_videomode_to_var(&fb->var, pd->def_mode); pd 1779 drivers/video/fbdev/sm501fb.c fb->var.bits_per_pixel = pd->def_bpp ? pd->def_bpp : 8; pd 1863 drivers/video/fbdev/sm501fb.c struct sm501_platdata_fbsub *pd; pd 1867 drivers/video/fbdev/sm501fb.c pd = (head == HEAD_CRT) ? info->pdata->fb_crt : info->pdata->fb_pnl; pd 1870 drivers/video/fbdev/sm501fb.c if (pd == NULL) { pd 1948 drivers/video/fbdev/sm501fb.c struct sm501_platdata *pd = dev->parent->platform_data; pd 1949 drivers/video/fbdev/sm501fb.c info->pdata = pd->fb; pd 367 drivers/video/fbdev/sticore.h struct pci_dev *pd; pd 565 drivers/video/fbdev/xen-fbfront.c info->page->pd[i] = vmalloc_to_gfn(&info->gfns[i * epd]); pd 684 fs/cifs/smbdirect.c info->pd = ib_alloc_pd(info->id->device, 0); pd 685 fs/cifs/smbdirect.c if (IS_ERR(info->pd)) { pd 686 fs/cifs/smbdirect.c rc = PTR_ERR(info->pd); pd 740 fs/cifs/smbdirect.c request->sge[0].lkey = info->pd->local_dma_lkey; pd 905 fs/cifs/smbdirect.c request->sge[0].lkey = info->pd->local_dma_lkey; pd 1007 fs/cifs/smbdirect.c request->sge[i+1].lkey = info->pd->local_dma_lkey; pd 1102 fs/cifs/smbdirect.c response->sge.lkey = info->pd->local_dma_lkey; pd 1466 fs/cifs/smbdirect.c ib_dealloc_pd(info->pd); pd 1689 fs/cifs/smbdirect.c rc = rdma_create_qp(info->id, info->pd, &qp_attr); pd 1810 fs/cifs/smbdirect.c ib_dealloc_pd(info->pd); pd 2287 fs/cifs/smbdirect.c info->pd, info->mr_type, pd 2354 fs/cifs/smbdirect.c smbdirect_mr->mr = ib_alloc_mr(info->pd, info->mr_type, pd 58 fs/cifs/smbdirect.h struct ib_pd *pd; pd 155 fs/xfs/xfs_quota.h #define xfs_trans_reserve_quota(tp, mp, ud, gd, pd, nb, ni, f) \ pd 156 fs/xfs/xfs_quota.h xfs_trans_reserve_quota_bydquots(tp, mp, ud, gd, pd, nb, ni, \ pd 133 include/linux/blk-cgroup.h struct blkg_policy_data *pd[BLKCG_MAX_POLS]; pd 154 include/linux/blk-cgroup.h typedef void (blkcg_pol_init_pd_fn)(struct blkg_policy_data *pd); pd 155 include/linux/blk-cgroup.h typedef void (blkcg_pol_online_pd_fn)(struct blkg_policy_data *pd); pd 156 include/linux/blk-cgroup.h typedef void (blkcg_pol_offline_pd_fn)(struct blkg_policy_data *pd); pd 157 include/linux/blk-cgroup.h typedef void (blkcg_pol_free_pd_fn)(struct blkg_policy_data *pd); pd 158 include/linux/blk-cgroup.h typedef void (blkcg_pol_reset_pd_stats_fn)(struct blkg_policy_data *pd); pd 159 include/linux/blk-cgroup.h typedef size_t (blkcg_pol_stat_pd_fn)(struct blkg_policy_data *pd, char *buf, pd 218 include/linux/blk-cgroup.h u64 __blkg_prfill_u64(struct seq_file *sf, struct blkg_policy_data *pd, u64 v); pd 219 include/linux/blk-cgroup.h u64 __blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd, pd 221 include/linux/blk-cgroup.h u64 blkg_prfill_rwstat(struct seq_file *sf, struct blkg_policy_data *pd, pd 410 include/linux/blk-cgroup.h return blkg ? blkg->pd[pol->plid] : NULL; pd 425 include/linux/blk-cgroup.h static inline struct blkcg_gq *pd_to_blkg(struct blkg_policy_data *pd) pd 427 include/linux/blk-cgroup.h return pd ? pd->blkg : NULL; pd 862 include/linux/blk-cgroup.h static inline struct blkcg_gq *pd_to_blkg(struct blkg_policy_data *pd) { return NULL; } pd 79 include/linux/energy_model.h static inline unsigned long em_pd_energy(struct em_perf_domain *pd, pd 91 include/linux/energy_model.h cpu = cpumask_first(to_cpumask(pd->cpus)); pd 93 include/linux/energy_model.h cs = &pd->table[pd->nr_cap_states - 1]; pd 100 include/linux/energy_model.h for (i = 0; i < pd->nr_cap_states; i++) { pd 101 include/linux/energy_model.h cs = &pd->table[i]; pd 157 include/linux/energy_model.h static inline int em_pd_nr_cap_states(struct em_perf_domain *pd) pd 159 include/linux/energy_model.h return pd->nr_cap_states; pd 176 include/linux/energy_model.h static inline unsigned long em_pd_energy(struct em_perf_domain *pd, pd 181 include/linux/energy_model.h static inline int em_pd_nr_cap_states(struct em_perf_domain *pd) pd 43 include/linux/firmware/trusted_foundations.h void register_trusted_foundations(struct trusted_foundations_platform_data *pd); pd 53 include/linux/firmware/trusted_foundations.h struct trusted_foundations_platform_data *pd) pd 19 include/linux/hugetlb.h typedef struct { unsigned long pd; } hugepd_t; pd 693 include/linux/mlx4/device.h u32 pd; pd 705 include/linux/mlx4/device.h u32 pd; pd 1117 include/linux/mlx4/device.h int mlx4_mr_alloc(struct mlx4_dev *dev, u32 pd, u64 iova, u64 size, u32 access, pd 1121 include/linux/mlx4/device.h int mlx4_mw_alloc(struct mlx4_dev *dev, u32 pd, enum mlx4_mw_type type, pd 1417 include/linux/mlx4/device.h int mlx4_fmr_alloc(struct mlx4_dev *dev, u32 pd, u32 access, int max_pages, pd 194 include/linux/mlx4/qp.h __be32 pd; pd 348 include/linux/mlx5/driver.h u32 pd; pd 376 include/linux/mlx5/driver.h u32 pd; pd 1609 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 2611 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 2856 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 2918 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 3013 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 3266 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 3469 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 7175 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 7525 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 8007 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 10231 include/linux/mlx5/mlx5_ifc.h u8 pd[0x18]; pd 36 include/linux/padata.h struct parallel_data *pd; pd 66 include/linux/padata.h struct parallel_data *pd; pd 140 include/linux/padata.h struct parallel_data __rcu *pd; pd 140 include/linux/pktcdvd.h struct pktcdvd_device *pd; pd 151 include/linux/pktcdvd.h struct pktcdvd_device *pd; pd 158 include/linux/pktcdvd.h struct pktcdvd_device *pd; pd 167 include/linux/platform_data/cros_ec_proto.h struct platform_device *pd; pd 44 include/linux/platform_data/hwmon-s3c.h extern void __init s3c_hwmon_set_platdata(struct s3c_hwmon_pdata *pd); pd 37 include/linux/platform_data/s3c-hsotg.h extern void dwc2_hsotg_set_platdata(struct dwc2_hsotg_plat *pd); pd 37 include/linux/platform_data/usb-s3c2410_udc.h extern void __init s3c24xx_hsudc_set_platdata(struct s3c24xx_hsudc_platdata *pd); pd 148 include/linux/pm_domain.h static inline struct generic_pm_domain *pd_to_genpd(struct dev_pm_domain *pd) pd 150 include/linux/pm_domain.h return container_of(pd, struct generic_pm_domain, domain); pd 369 include/linux/pm_domain.h void dev_pm_domain_set(struct device *dev, struct dev_pm_domain *pd); pd 387 include/linux/pm_domain.h struct dev_pm_domain *pd) {} pd 253 include/linux/qed/qed_rdma_if.h u16 pd; pd 312 include/linux/qed/qed_rdma_if.h u16 pd; pd 612 include/linux/qed/qed_rdma_if.h int (*rdma_alloc_pd)(void *rdma_cxt, u16 *pd); pd 613 include/linux/qed/qed_rdma_if.h void (*rdma_dealloc_pd)(void *rdma_cxt, u16 pd); pd 123 include/net/dsa.h struct dsa_platform_data *pd; pd 428 include/net/ip_vs.h int (*init_netns)(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd); pd 430 include/net/ip_vs.h void (*exit_netns)(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd); pd 434 include/net/ip_vs.h struct ip_vs_proto_data *pd, pd 460 include/net/ip_vs.h struct ip_vs_proto_data *pd); pd 473 include/net/ip_vs.h void (*timeout_change)(struct ip_vs_proto_data *pd, int flags); pd 1391 include/net/ip_vs.h struct ip_vs_proto_data *pd, int *ignored, pd 1394 include/net/ip_vs.h struct ip_vs_proto_data *pd, struct ip_vs_iphdr *iph); pd 79 include/rdma/ib_fmr_pool.h struct ib_fmr_pool *ib_create_fmr_pool(struct ib_pd *pd, pd 1527 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1566 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1615 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1713 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1762 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1785 include/rdma/ib_verbs.h struct ib_pd *pd; pd 1793 include/rdma/ib_verbs.h struct ib_pd *pd; pd 2366 include/rdma/ib_verbs.h int (*alloc_pd)(struct ib_pd *pd, struct ib_udata *udata); pd 2367 include/rdma/ib_verbs.h void (*dealloc_pd)(struct ib_pd *pd, struct ib_udata *udata); pd 2381 include/rdma/ib_verbs.h struct ib_qp *(*create_qp)(struct ib_pd *pd, pd 2394 include/rdma/ib_verbs.h struct ib_mr *(*get_dma_mr)(struct ib_pd *pd, int mr_access_flags); pd 2395 include/rdma/ib_verbs.h struct ib_mr *(*reg_user_mr)(struct ib_pd *pd, u64 start, u64 length, pd 2400 include/rdma/ib_verbs.h struct ib_pd *pd, struct ib_udata *udata); pd 2402 include/rdma/ib_verbs.h struct ib_mr *(*alloc_mr)(struct ib_pd *pd, enum ib_mr_type mr_type, pd 2404 include/rdma/ib_verbs.h struct ib_mr *(*alloc_mr_integrity)(struct ib_pd *pd, pd 2407 include/rdma/ib_verbs.h int (*advise_mr)(struct ib_pd *pd, pd 2415 include/rdma/ib_verbs.h struct ib_mw *(*alloc_mw)(struct ib_pd *pd, enum ib_mw_type type, pd 2418 include/rdma/ib_verbs.h struct ib_fmr *(*alloc_fmr)(struct ib_pd *pd, int mr_access_flags, pd 2452 include/rdma/ib_verbs.h struct ib_wq *(*create_wq)(struct ib_pd *pd, pd 2468 include/rdma/ib_verbs.h struct ib_mr *(*reg_dm_mr)(struct ib_pd *pd, struct ib_dm *dm, pd 3353 include/rdma/ib_verbs.h void ib_dealloc_pd_user(struct ib_pd *pd, struct ib_udata *udata); pd 3361 include/rdma/ib_verbs.h static inline void ib_dealloc_pd(struct ib_pd *pd) pd 3363 include/rdma/ib_verbs.h ib_dealloc_pd_user(pd, NULL); pd 3380 include/rdma/ib_verbs.h struct ib_ah *rdma_create_ah(struct ib_pd *pd, struct rdma_ah_attr *ah_attr, pd 3395 include/rdma/ib_verbs.h struct ib_ah *rdma_create_user_ah(struct ib_pd *pd, pd 3450 include/rdma/ib_verbs.h struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, const struct ib_wc *wc, pd 3509 include/rdma/ib_verbs.h struct ib_srq *ib_create_srq(struct ib_pd *pd, pd 3581 include/rdma/ib_verbs.h struct ib_qp *ib_create_qp_user(struct ib_pd *pd, pd 3596 include/rdma/ib_verbs.h static inline struct ib_qp *ib_create_qp(struct ib_pd *pd, pd 3599 include/rdma/ib_verbs.h return ib_create_qp_user(pd, qp_init_attr, NULL); pd 4135 include/rdma/ib_verbs.h struct ib_mr *ib_alloc_mr_user(struct ib_pd *pd, enum ib_mr_type mr_type, pd 4138 include/rdma/ib_verbs.h static inline struct ib_mr *ib_alloc_mr(struct ib_pd *pd, pd 4141 include/rdma/ib_verbs.h return ib_alloc_mr_user(pd, mr_type, max_num_sg, NULL); pd 4144 include/rdma/ib_verbs.h struct ib_mr *ib_alloc_mr_integrity(struct ib_pd *pd, pd 4180 include/rdma/ib_verbs.h struct ib_fmr *ib_alloc_fmr(struct ib_pd *pd, pd 4323 include/rdma/ib_verbs.h struct ib_wq *ib_create_wq(struct ib_pd *pd, pd 219 include/rdma/rdma_cm.h int rdma_create_qp(struct rdma_cm_id *id, struct ib_pd *pd, pd 568 include/rdma/rdma_vt.h int rvt_lkey_ok(struct rvt_lkey_table *rkt, struct rvt_pd *pd, pd 74 include/rdma/rdmavt_mr.h struct ib_pd *pd; /* shares refcnt of ibmr.pd */ pd 14 include/sound/sh_dac_audio.h void (*start)(struct dac_audio_pdata *pd); pd 15 include/sound/sh_dac_audio.h void (*stop)(struct dac_audio_pdata *pd); pd 130 include/xen/interface/io/fbif.h unsigned long pd[256]; pd 38 kernel/padata.c static void padata_free_pd(struct parallel_data *pd); pd 40 kernel/padata.c static int padata_index_to_cpu(struct parallel_data *pd, int cpu_index) pd 44 kernel/padata.c target_cpu = cpumask_first(pd->cpumask.pcpu); pd 46 kernel/padata.c target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); pd 51 kernel/padata.c static int padata_cpu_hash(struct parallel_data *pd, unsigned int seq_nr) pd 57 kernel/padata.c int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); pd 59 kernel/padata.c return padata_index_to_cpu(pd, cpu_index); pd 109 kernel/padata.c struct parallel_data *pd; pd 113 kernel/padata.c pd = rcu_dereference_bh(ps->pd); pd 119 kernel/padata.c if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) { pd 120 kernel/padata.c if (!cpumask_weight(pd->cpumask.cbcpu)) pd 124 kernel/padata.c cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu); pd 126 kernel/padata.c cpu = cpumask_first(pd->cpumask.cbcpu); pd 128 kernel/padata.c cpu = cpumask_next(cpu, pd->cpumask.cbcpu); pd 137 kernel/padata.c if (atomic_read(&pd->refcnt) >= MAX_OBJ_NUM) pd 141 kernel/padata.c atomic_inc(&pd->refcnt); pd 142 kernel/padata.c padata->pd = pd; pd 145 kernel/padata.c padata->seq_nr = atomic_inc_return(&pd->seq_nr); pd 146 kernel/padata.c target_cpu = padata_cpu_hash(pd, padata->seq_nr); pd 148 kernel/padata.c queue = per_cpu_ptr(pd->pqueue, target_cpu); pd 175 kernel/padata.c static struct padata_priv *padata_find_next(struct parallel_data *pd, pd 181 kernel/padata.c int cpu = pd->cpu; pd 183 kernel/padata.c next_queue = per_cpu_ptr(pd->pqueue, cpu); pd 198 kernel/padata.c if (padata->seq_nr != pd->processed) { pd 205 kernel/padata.c atomic_dec(&pd->reorder_objects); pd 206 kernel/padata.c ++pd->processed; pd 207 kernel/padata.c pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false); pd 214 kernel/padata.c static void padata_reorder(struct parallel_data *pd) pd 216 kernel/padata.c struct padata_instance *pinst = pd->ps->pinst; pd 232 kernel/padata.c if (!spin_trylock_bh(&pd->lock)) pd 236 kernel/padata.c padata = padata_find_next(pd, true); pd 247 kernel/padata.c squeue = per_cpu_ptr(pd->squeue, cb_cpu); pd 256 kernel/padata.c spin_unlock_bh(&pd->lock); pd 268 kernel/padata.c next_queue = per_cpu_ptr(pd->pqueue, pd->cpu); pd 270 kernel/padata.c padata_find_next(pd, false)) pd 271 kernel/padata.c queue_work(pinst->serial_wq, &pd->reorder_work); pd 276 kernel/padata.c struct parallel_data *pd; pd 279 kernel/padata.c pd = container_of(work, struct parallel_data, reorder_work); pd 280 kernel/padata.c padata_reorder(pd); pd 287 kernel/padata.c struct parallel_data *pd; pd 293 kernel/padata.c pd = squeue->pd; pd 314 kernel/padata.c if (atomic_sub_and_test(cnt, &pd->refcnt)) pd 315 kernel/padata.c padata_free_pd(pd); pd 328 kernel/padata.c struct parallel_data *pd = padata->pd; pd 329 kernel/padata.c struct padata_parallel_queue *pqueue = per_cpu_ptr(pd->pqueue, pd 339 kernel/padata.c atomic_inc(&pd->reorder_objects); pd 349 kernel/padata.c padata_reorder(pd); pd 370 kernel/padata.c static int pd_setup_cpumasks(struct parallel_data *pd, pd 376 kernel/padata.c if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL)) pd 378 kernel/padata.c if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL)) pd 381 kernel/padata.c cpumask_copy(pd->cpumask.pcpu, pcpumask); pd 382 kernel/padata.c cpumask_copy(pd->cpumask.cbcpu, cbcpumask); pd 387 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); pd 399 kernel/padata.c static void padata_init_squeues(struct parallel_data *pd) pd 404 kernel/padata.c for_each_cpu(cpu, pd->cpumask.cbcpu) { pd 405 kernel/padata.c squeue = per_cpu_ptr(pd->squeue, cpu); pd 406 kernel/padata.c squeue->pd = pd; pd 413 kernel/padata.c static void padata_init_pqueues(struct parallel_data *pd) pd 418 kernel/padata.c for_each_cpu(cpu, pd->cpumask.pcpu) { pd 419 kernel/padata.c pqueue = per_cpu_ptr(pd->pqueue, cpu); pd 434 kernel/padata.c struct parallel_data *pd; pd 439 kernel/padata.c pd = kzalloc(sizeof(struct parallel_data), GFP_KERNEL); pd 440 kernel/padata.c if (!pd) pd 443 kernel/padata.c pd->pqueue = alloc_percpu(struct padata_parallel_queue); pd 444 kernel/padata.c if (!pd->pqueue) pd 447 kernel/padata.c pd->squeue = alloc_percpu(struct padata_serial_queue); pd 448 kernel/padata.c if (!pd->squeue) pd 451 kernel/padata.c pd->ps = ps; pd 452 kernel/padata.c if (pd_setup_cpumasks(pd, pcpumask, cbcpumask)) pd 455 kernel/padata.c padata_init_pqueues(pd); pd 456 kernel/padata.c padata_init_squeues(pd); pd 457 kernel/padata.c atomic_set(&pd->seq_nr, -1); pd 458 kernel/padata.c atomic_set(&pd->reorder_objects, 0); pd 459 kernel/padata.c atomic_set(&pd->refcnt, 1); pd 460 kernel/padata.c spin_lock_init(&pd->lock); pd 461 kernel/padata.c pd->cpu = cpumask_first(pd->cpumask.pcpu); pd 462 kernel/padata.c INIT_WORK(&pd->reorder_work, invoke_padata_reorder); pd 464 kernel/padata.c return pd; pd 467 kernel/padata.c free_percpu(pd->squeue); pd 469 kernel/padata.c free_percpu(pd->pqueue); pd 471 kernel/padata.c kfree(pd); pd 476 kernel/padata.c static void padata_free_pd(struct parallel_data *pd) pd 478 kernel/padata.c free_cpumask_var(pd->cpumask.pcpu); pd 479 kernel/padata.c free_cpumask_var(pd->cpumask.cbcpu); pd 480 kernel/padata.c free_percpu(pd->pqueue); pd 481 kernel/padata.c free_percpu(pd->squeue); pd 482 kernel/padata.c kfree(pd); pd 509 kernel/padata.c ps->opd = rcu_dereference_protected(ps->pd, 1); pd 510 kernel/padata.c rcu_assign_pointer(ps->pd, pd_new); pd 1085 kernel/padata.c struct parallel_data *pd; pd 1095 kernel/padata.c pd = padata_alloc_pd(ps); pd 1098 kernel/padata.c if (!pd) pd 1102 kernel/padata.c RCU_INIT_POINTER(ps->pd, pd); pd 1126 kernel/padata.c padata_free_pd(rcu_dereference_protected(ps->pd, 1)); pd 30 kernel/power/energy_model.c static void em_debug_create_cs(struct em_cap_state *cs, struct dentry *pd) pd 38 kernel/power/energy_model.c d = debugfs_create_dir(name, pd); pd 52 kernel/power/energy_model.c static void em_debug_create_pd(struct em_perf_domain *pd, int cpu) pd 63 kernel/power/energy_model.c debugfs_create_file("cpus", 0444, d, pd->cpus, &em_debug_cpus_fops); pd 66 kernel/power/energy_model.c for (i = 0; i < pd->nr_cap_states; i++) pd 67 kernel/power/energy_model.c em_debug_create_cs(&pd->table[i], d); pd 79 kernel/power/energy_model.c static void em_debug_create_pd(struct em_perf_domain *pd, int cpu) {} pd 88 kernel/power/energy_model.c struct em_perf_domain *pd; pd 94 kernel/power/energy_model.c pd = kzalloc(sizeof(*pd) + cpumask_size(), GFP_KERNEL); pd 95 kernel/power/energy_model.c if (!pd) pd 156 kernel/power/energy_model.c pd->table = table; pd 157 kernel/power/energy_model.c pd->nr_cap_states = nr_states; pd 158 kernel/power/energy_model.c cpumask_copy(to_cpumask(pd->cpus), span); pd 160 kernel/power/energy_model.c em_debug_create_pd(pd, cpu); pd 162 kernel/power/energy_model.c return pd; pd 167 kernel/power/energy_model.c kfree(pd); pd 203 kernel/power/energy_model.c struct em_perf_domain *pd; pd 237 kernel/power/energy_model.c pd = em_create_pd(span, nr_states, cb); pd 238 kernel/power/energy_model.c if (!pd) { pd 249 kernel/power/energy_model.c smp_store_release(per_cpu_ptr(&em_data, cpu), pd); pd 6277 kernel/sched/fair.c compute_energy(struct task_struct *p, int dst_cpu, struct perf_domain *pd) pd 6279 kernel/sched/fair.c struct cpumask *pd_mask = perf_domain_span(pd); pd 6318 kernel/sched/fair.c return em_pd_energy(pd->em_pd, max_util, sum_util); pd 6367 kernel/sched/fair.c struct perf_domain *pd; pd 6370 kernel/sched/fair.c pd = rcu_dereference(rd->pd); pd 6371 kernel/sched/fair.c if (!pd || READ_ONCE(rd->overutilized)) pd 6388 kernel/sched/fair.c for (; pd; pd = pd->next) { pd 6394 kernel/sched/fair.c base_energy_pd = compute_energy(p, -1, pd); pd 6397 kernel/sched/fair.c for_each_cpu_and(cpu, perf_domain_span(pd), sched_domain_span(sd)) { pd 6409 kernel/sched/fair.c prev_delta = compute_energy(p, prev_cpu, pd); pd 6427 kernel/sched/fair.c cur_delta = compute_energy(p, max_spare_cap_cpu, pd); pd 8548 kernel/sched/fair.c if (rcu_dereference(rd->pd) && !READ_ONCE(rd->overutilized)) pd 790 kernel/sched/sched.h struct perf_domain __rcu *pd; pd 2460 kernel/sched/sched.h #define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus))) pd 2471 kernel/sched/sched.h #define perf_domain_span(pd) NULL pd 235 kernel/sched/topology.c static void free_pd(struct perf_domain *pd) pd 239 kernel/sched/topology.c while (pd) { pd 240 kernel/sched/topology.c tmp = pd->next; pd 241 kernel/sched/topology.c kfree(pd); pd 242 kernel/sched/topology.c pd = tmp; pd 246 kernel/sched/topology.c static struct perf_domain *find_pd(struct perf_domain *pd, int cpu) pd 248 kernel/sched/topology.c while (pd) { pd 249 kernel/sched/topology.c if (cpumask_test_cpu(cpu, perf_domain_span(pd))) pd 250 kernel/sched/topology.c return pd; pd 251 kernel/sched/topology.c pd = pd->next; pd 260 kernel/sched/topology.c struct perf_domain *pd; pd 268 kernel/sched/topology.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 269 kernel/sched/topology.c if (!pd) pd 271 kernel/sched/topology.c pd->em_pd = obj; pd 273 kernel/sched/topology.c return pd; pd 277 kernel/sched/topology.c struct perf_domain *pd) pd 279 kernel/sched/topology.c if (!sched_debug() || !pd) pd 284 kernel/sched/topology.c while (pd) { pd 286 kernel/sched/topology.c cpumask_first(perf_domain_span(pd)), pd 287 kernel/sched/topology.c cpumask_pr_args(perf_domain_span(pd)), pd 288 kernel/sched/topology.c em_pd_nr_cap_states(pd->em_pd)); pd 289 kernel/sched/topology.c pd = pd->next; pd 297 kernel/sched/topology.c struct perf_domain *pd; pd 299 kernel/sched/topology.c pd = container_of(rp, struct perf_domain, rcu); pd 300 kernel/sched/topology.c free_pd(pd); pd 345 kernel/sched/topology.c struct perf_domain *pd = NULL, *tmp; pd 365 kernel/sched/topology.c if (find_pd(pd, i)) pd 375 kernel/sched/topology.c if (rd->pd) pd 385 kernel/sched/topology.c tmp->next = pd; pd 386 kernel/sched/topology.c pd = tmp; pd 393 kernel/sched/topology.c nr_cs += em_pd_nr_cap_states(pd->em_pd); pd 403 kernel/sched/topology.c perf_domain_debug(cpu_map, pd); pd 406 kernel/sched/topology.c tmp = rd->pd; pd 407 kernel/sched/topology.c rcu_assign_pointer(rd->pd, pd); pd 411 kernel/sched/topology.c return !!pd; pd 414 kernel/sched/topology.c free_pd(pd); pd 415 kernel/sched/topology.c tmp = rd->pd; pd 416 kernel/sched/topology.c rcu_assign_pointer(rd->pd, NULL); pd 423 kernel/sched/topology.c static void free_pd(struct perf_domain *pd) { } pd 436 kernel/sched/topology.c free_pd(rd->pd); pd 2299 kernel/sched/topology.c cpu_rq(cpumask_first(doms_cur[j]))->rd->pd) { pd 391 lib/bch.c unsigned int i, j, tmp, l, pd = 1, d = syn[0]; pd 411 lib/bch.c tmp = a_log(bch, d)+n-a_log(bch, pd); pd 423 lib/bch.c pd = d; pd 80 net/9p/trans_rdma.c struct ib_pd *pd; pd 370 net/9p/trans_rdma.c if (rdma->pd && !IS_ERR(rdma->pd)) pd 371 net/9p/trans_rdma.c ib_dealloc_pd(rdma->pd); pd 399 net/9p/trans_rdma.c sge.lkey = rdma->pd->local_dma_lkey; pd 491 net/9p/trans_rdma.c sge.lkey = rdma->pd->local_dma_lkey; pd 695 net/9p/trans_rdma.c rdma->pd = ib_alloc_pd(rdma->cm_id->device, 0); pd 696 net/9p/trans_rdma.c if (IS_ERR(rdma->pd)) pd 711 net/9p/trans_rdma.c err = rdma_create_qp(rdma->cm_id, rdma->pd, &qp_attr); pd 92 net/ipv4/fou.c __be16 *pd = data; pd 93 net/ipv4/fou.c size_t start = ntohs(pd[0]); pd 94 net/ipv4/fou.c size_t offset = ntohs(pd[1]); pd 292 net/ipv4/fou.c __be16 *pd = data; pd 293 net/ipv4/fou.c size_t start = ntohs(pd[0]); pd 294 net/ipv4/fou.c size_t offset = ntohs(pd[1]); pd 1036 net/ipv4/fou.c __be16 *pd = data; pd 1042 net/ipv4/fou.c pd[0] = htons(csum_start); pd 1043 net/ipv4/fou.c pd[1] = htons(csum_start + skb->csum_offset); pd 43 net/l2tp/l2tp_debugfs.c static void l2tp_dfs_next_tunnel(struct l2tp_dfs_seq_data *pd) pd 46 net/l2tp/l2tp_debugfs.c if (pd->tunnel) pd 47 net/l2tp/l2tp_debugfs.c l2tp_tunnel_dec_refcount(pd->tunnel); pd 49 net/l2tp/l2tp_debugfs.c pd->tunnel = l2tp_tunnel_get_nth(pd->net, pd->tunnel_idx); pd 50 net/l2tp/l2tp_debugfs.c pd->tunnel_idx++; pd 53 net/l2tp/l2tp_debugfs.c static void l2tp_dfs_next_session(struct l2tp_dfs_seq_data *pd) pd 56 net/l2tp/l2tp_debugfs.c if (pd->session) pd 57 net/l2tp/l2tp_debugfs.c l2tp_session_dec_refcount(pd->session); pd 59 net/l2tp/l2tp_debugfs.c pd->session = l2tp_session_get_nth(pd->tunnel, pd->session_idx); pd 60 net/l2tp/l2tp_debugfs.c pd->session_idx++; pd 62 net/l2tp/l2tp_debugfs.c if (pd->session == NULL) { pd 63 net/l2tp/l2tp_debugfs.c pd->session_idx = 0; pd 64 net/l2tp/l2tp_debugfs.c l2tp_dfs_next_tunnel(pd); pd 71 net/l2tp/l2tp_debugfs.c struct l2tp_dfs_seq_data *pd = SEQ_START_TOKEN; pd 78 net/l2tp/l2tp_debugfs.c pd = m->private; pd 80 net/l2tp/l2tp_debugfs.c if (pd->tunnel == NULL) pd 81 net/l2tp/l2tp_debugfs.c l2tp_dfs_next_tunnel(pd); pd 83 net/l2tp/l2tp_debugfs.c l2tp_dfs_next_session(pd); pd 86 net/l2tp/l2tp_debugfs.c if ((pd->tunnel == NULL) && (pd->session == NULL)) pd 87 net/l2tp/l2tp_debugfs.c pd = NULL; pd 90 net/l2tp/l2tp_debugfs.c return pd; pd 102 net/l2tp/l2tp_debugfs.c struct l2tp_dfs_seq_data *pd = v; pd 104 net/l2tp/l2tp_debugfs.c if (!pd || pd == SEQ_START_TOKEN) pd 110 net/l2tp/l2tp_debugfs.c if (pd->session) { pd 111 net/l2tp/l2tp_debugfs.c l2tp_session_dec_refcount(pd->session); pd 112 net/l2tp/l2tp_debugfs.c pd->session = NULL; pd 114 net/l2tp/l2tp_debugfs.c if (pd->tunnel) { pd 115 net/l2tp/l2tp_debugfs.c l2tp_tunnel_dec_refcount(pd->tunnel); pd 116 net/l2tp/l2tp_debugfs.c pd->tunnel = NULL; pd 233 net/l2tp/l2tp_debugfs.c struct l2tp_dfs_seq_data *pd = v; pd 251 net/l2tp/l2tp_debugfs.c if (!pd->session) pd 252 net/l2tp/l2tp_debugfs.c l2tp_dfs_seq_tunnel_show(m, pd->tunnel); pd 254 net/l2tp/l2tp_debugfs.c l2tp_dfs_seq_session_show(m, pd->session); pd 269 net/l2tp/l2tp_debugfs.c struct l2tp_dfs_seq_data *pd; pd 273 net/l2tp/l2tp_debugfs.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 274 net/l2tp/l2tp_debugfs.c if (pd == NULL) pd 280 net/l2tp/l2tp_debugfs.c pd->net = get_net_ns_by_pid(current->pid); pd 281 net/l2tp/l2tp_debugfs.c if (IS_ERR(pd->net)) { pd 282 net/l2tp/l2tp_debugfs.c rc = PTR_ERR(pd->net); pd 291 net/l2tp/l2tp_debugfs.c seq->private = pd; pd 297 net/l2tp/l2tp_debugfs.c put_net(pd->net); pd 299 net/l2tp/l2tp_debugfs.c kfree(pd); pd 305 net/l2tp/l2tp_debugfs.c struct l2tp_dfs_seq_data *pd; pd 309 net/l2tp/l2tp_debugfs.c pd = seq->private; pd 310 net/l2tp/l2tp_debugfs.c if (pd->net) pd 311 net/l2tp/l2tp_debugfs.c put_net(pd->net); pd 312 net/l2tp/l2tp_debugfs.c kfree(pd); pd 1439 net/l2tp/l2tp_ppp.c static void pppol2tp_next_tunnel(struct net *net, struct pppol2tp_seq_data *pd) pd 1442 net/l2tp/l2tp_ppp.c if (pd->tunnel) pd 1443 net/l2tp/l2tp_ppp.c l2tp_tunnel_dec_refcount(pd->tunnel); pd 1446 net/l2tp/l2tp_ppp.c pd->tunnel = l2tp_tunnel_get_nth(net, pd->tunnel_idx); pd 1447 net/l2tp/l2tp_ppp.c pd->tunnel_idx++; pd 1450 net/l2tp/l2tp_ppp.c if (!pd->tunnel || pd->tunnel->version == 2) pd 1453 net/l2tp/l2tp_ppp.c l2tp_tunnel_dec_refcount(pd->tunnel); pd 1457 net/l2tp/l2tp_ppp.c static void pppol2tp_next_session(struct net *net, struct pppol2tp_seq_data *pd) pd 1460 net/l2tp/l2tp_ppp.c if (pd->session) pd 1461 net/l2tp/l2tp_ppp.c l2tp_session_dec_refcount(pd->session); pd 1463 net/l2tp/l2tp_ppp.c pd->session = l2tp_session_get_nth(pd->tunnel, pd->session_idx); pd 1464 net/l2tp/l2tp_ppp.c pd->session_idx++; pd 1466 net/l2tp/l2tp_ppp.c if (pd->session == NULL) { pd 1467 net/l2tp/l2tp_ppp.c pd->session_idx = 0; pd 1468 net/l2tp/l2tp_ppp.c pppol2tp_next_tunnel(net, pd); pd 1474 net/l2tp/l2tp_ppp.c struct pppol2tp_seq_data *pd = SEQ_START_TOKEN; pd 1482 net/l2tp/l2tp_ppp.c pd = m->private; pd 1485 net/l2tp/l2tp_ppp.c if (pd->tunnel == NULL) pd 1486 net/l2tp/l2tp_ppp.c pppol2tp_next_tunnel(net, pd); pd 1488 net/l2tp/l2tp_ppp.c pppol2tp_next_session(net, pd); pd 1491 net/l2tp/l2tp_ppp.c if ((pd->tunnel == NULL) && (pd->session == NULL)) pd 1492 net/l2tp/l2tp_ppp.c pd = NULL; pd 1495 net/l2tp/l2tp_ppp.c return pd; pd 1506 net/l2tp/l2tp_ppp.c struct pppol2tp_seq_data *pd = v; pd 1508 net/l2tp/l2tp_ppp.c if (!pd || pd == SEQ_START_TOKEN) pd 1514 net/l2tp/l2tp_ppp.c if (pd->session) { pd 1515 net/l2tp/l2tp_ppp.c l2tp_session_dec_refcount(pd->session); pd 1516 net/l2tp/l2tp_ppp.c pd->session = NULL; pd 1518 net/l2tp/l2tp_ppp.c if (pd->tunnel) { pd 1519 net/l2tp/l2tp_ppp.c l2tp_tunnel_dec_refcount(pd->tunnel); pd 1520 net/l2tp/l2tp_ppp.c pd->tunnel = NULL; pd 1600 net/l2tp/l2tp_ppp.c struct pppol2tp_seq_data *pd = v; pd 1614 net/l2tp/l2tp_ppp.c if (!pd->session) pd 1615 net/l2tp/l2tp_ppp.c pppol2tp_seq_tunnel_show(m, pd->tunnel); pd 1617 net/l2tp/l2tp_ppp.c pppol2tp_seq_session_show(m, pd->session); pd 654 net/netfilter/ipvs/ip_vs_conn.c struct ip_vs_proto_data *pd; pd 680 net/netfilter/ipvs/ip_vs_conn.c pd = ip_vs_proto_data_get(cp->ipvs, cp->protocol); pd 681 net/netfilter/ipvs/ip_vs_conn.c if (pd && atomic_read(&pd->appcnt)) pd 682 net/netfilter/ipvs/ip_vs_conn.c ip_vs_bind_app(cp, pd->pp); pd 907 net/netfilter/ipvs/ip_vs_conn.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(p->ipvs, pd 984 net/netfilter/ipvs/ip_vs_conn.c if (unlikely(pd && atomic_read(&pd->appcnt))) pd 985 net/netfilter/ipvs/ip_vs_conn.c ip_vs_bind_app(cp, pd->pp); pd 235 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd) pd 237 net/netfilter/ipvs/ip_vs_core.c if (likely(pd->pp->state_transition)) pd 238 net/netfilter/ipvs/ip_vs_core.c pd->pp->state_transition(cp, direction, skb, pd); pd 454 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd, int *ignored, pd 457 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_protocol *pp = pd->pp; pd 598 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd, struct ip_vs_iphdr *iph) pd 640 net/netfilter/ipvs/ip_vs_core.c ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pd); pd 643 net/netfilter/ipvs/ip_vs_core.c ret = cp->packet_xmit(skb, cp, pd->pp, iph); pd 1279 net/netfilter/ipvs/ip_vs_core.c handle_response(int af, struct sk_buff *skb, struct ip_vs_proto_data *pd, pd 1283 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_protocol *pp = pd->pp; pd 1326 net/netfilter/ipvs/ip_vs_core.c ip_vs_set_state(cp, IP_VS_DIR_OUTPUT, skb, pd); pd 1352 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd; pd 1398 net/netfilter/ipvs/ip_vs_core.c pd = ip_vs_proto_data_get(ipvs, iph.protocol); pd 1399 net/netfilter/ipvs/ip_vs_core.c if (unlikely(!pd)) pd 1401 net/netfilter/ipvs/ip_vs_core.c pp = pd->pp; pd 1424 net/netfilter/ipvs/ip_vs_core.c return handle_response(af, skb, pd, cp, &iph, hooknum); pd 1437 net/netfilter/ipvs/ip_vs_core.c return handle_response(af, skb, pd, cp, &iph, pd 1546 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd, pd 1550 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_protocol *pp = pd->pp; pd 1558 net/netfilter/ipvs/ip_vs_core.c if (!pp->conn_schedule(ipvs, af, skb, pd, verdict, cpp, iph)) pd 1662 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd; pd 1756 net/netfilter/ipvs/ip_vs_core.c pd = ip_vs_proto_data_get(ipvs, cih->protocol); pd 1757 net/netfilter/ipvs/ip_vs_core.c if (!pd) pd 1759 net/netfilter/ipvs/ip_vs_core.c pp = pd->pp; pd 1785 net/netfilter/ipvs/ip_vs_core.c if (!ip_vs_try_to_schedule(ipvs, AF_INET, skb, pd, &v, &cp, &ciph)) pd 1879 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd; pd 1914 net/netfilter/ipvs/ip_vs_core.c pd = ip_vs_proto_data_get(ipvs, ciph.protocol); pd 1915 net/netfilter/ipvs/ip_vs_core.c if (!pd) pd 1917 net/netfilter/ipvs/ip_vs_core.c pp = pd->pp; pd 1938 net/netfilter/ipvs/ip_vs_core.c if (!ip_vs_try_to_schedule(ipvs, AF_INET6, skb, pd, &v, &cp, &ciph)) pd 1982 net/netfilter/ipvs/ip_vs_core.c struct ip_vs_proto_data *pd; pd 2044 net/netfilter/ipvs/ip_vs_core.c pd = ip_vs_proto_data_get(ipvs, iph.protocol); pd 2045 net/netfilter/ipvs/ip_vs_core.c if (unlikely(!pd)) { pd 2055 net/netfilter/ipvs/ip_vs_core.c pp = pd->pp; pd 2096 net/netfilter/ipvs/ip_vs_core.c if (!ip_vs_try_to_schedule(ipvs, af, skb, pd, &v, &cp, &iph)) pd 2124 net/netfilter/ipvs/ip_vs_core.c ip_vs_set_state(cp, IP_VS_DIR_INPUT, skb, pd); pd 2297 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_proto_data *pd; pd 2319 net/netfilter/ipvs/ip_vs_ctl.c pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP); pd 2320 net/netfilter/ipvs/ip_vs_ctl.c pd->timeout_table[IP_VS_TCP_S_ESTABLISHED] pd 2325 net/netfilter/ipvs/ip_vs_ctl.c pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP); pd 2326 net/netfilter/ipvs/ip_vs_ctl.c pd->timeout_table[IP_VS_TCP_S_FIN_WAIT] pd 2333 net/netfilter/ipvs/ip_vs_ctl.c pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP); pd 2334 net/netfilter/ipvs/ip_vs_ctl.c pd->timeout_table[IP_VS_UDP_S_NORMAL] pd 2688 net/netfilter/ipvs/ip_vs_ctl.c struct ip_vs_proto_data *pd; pd 2694 net/netfilter/ipvs/ip_vs_ctl.c pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP); pd 2695 net/netfilter/ipvs/ip_vs_ctl.c u->tcp_timeout = pd->timeout_table[IP_VS_TCP_S_ESTABLISHED] / HZ; pd 2696 net/netfilter/ipvs/ip_vs_ctl.c u->tcp_fin_timeout = pd->timeout_table[IP_VS_TCP_S_FIN_WAIT] / HZ; pd 2699 net/netfilter/ipvs/ip_vs_ctl.c pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP); pd 2701 net/netfilter/ipvs/ip_vs_ctl.c pd->timeout_table[IP_VS_UDP_S_NORMAL] / HZ; pd 69 net/netfilter/ipvs/ip_vs_proto.c struct ip_vs_proto_data *pd = pd 72 net/netfilter/ipvs/ip_vs_proto.c if (!pd) pd 75 net/netfilter/ipvs/ip_vs_proto.c pd->pp = pp; /* For speed issues */ pd 76 net/netfilter/ipvs/ip_vs_proto.c pd->next = ipvs->proto_data_table[hash]; pd 77 net/netfilter/ipvs/ip_vs_proto.c ipvs->proto_data_table[hash] = pd; pd 78 net/netfilter/ipvs/ip_vs_proto.c atomic_set(&pd->appcnt, 0); /* Init app counter */ pd 81 net/netfilter/ipvs/ip_vs_proto.c int ret = pp->init_netns(ipvs, pd); pd 84 net/netfilter/ipvs/ip_vs_proto.c ipvs->proto_data_table[hash] = pd->next; pd 85 net/netfilter/ipvs/ip_vs_proto.c kfree(pd); pd 118 net/netfilter/ipvs/ip_vs_proto.c unregister_ip_vs_proto_netns(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 121 net/netfilter/ipvs/ip_vs_proto.c unsigned int hash = IP_VS_PROTO_HASH(pd->pp->protocol); pd 125 net/netfilter/ipvs/ip_vs_proto.c if (*pd_p == pd) { pd 126 net/netfilter/ipvs/ip_vs_proto.c *pd_p = pd->next; pd 127 net/netfilter/ipvs/ip_vs_proto.c if (pd->pp->exit_netns != NULL) pd 128 net/netfilter/ipvs/ip_vs_proto.c pd->pp->exit_netns(ipvs, pd); pd 129 net/netfilter/ipvs/ip_vs_proto.c kfree(pd); pd 160 net/netfilter/ipvs/ip_vs_proto.c struct ip_vs_proto_data *pd; pd 163 net/netfilter/ipvs/ip_vs_proto.c for (pd = ipvs->proto_data_table[hash]; pd; pd = pd->next) { pd 164 net/netfilter/ipvs/ip_vs_proto.c if (pd->pp->protocol == proto) pd 165 net/netfilter/ipvs/ip_vs_proto.c return pd; pd 177 net/netfilter/ipvs/ip_vs_proto.c struct ip_vs_proto_data *pd; pd 181 net/netfilter/ipvs/ip_vs_proto.c for (pd = ipvs->proto_data_table[i]; pd; pd = pd->next) { pd 182 net/netfilter/ipvs/ip_vs_proto.c if (pd->pp->timeout_change) pd 183 net/netfilter/ipvs/ip_vs_proto.c pd->pp->timeout_change(pd, flags); pd 331 net/netfilter/ipvs/ip_vs_proto.c struct ip_vs_proto_data *pd; pd 336 net/netfilter/ipvs/ip_vs_proto.c while ((pd = ipvs->proto_data_table[i]) != NULL) pd 337 net/netfilter/ipvs/ip_vs_proto.c unregister_ip_vs_proto_netns(ipvs, pd); pd 104 net/netfilter/ipvs/ip_vs_proto_ah_esp.c struct ip_vs_proto_data *pd, pd 17 net/netfilter/ipvs/ip_vs_proto_sctp.c struct ip_vs_proto_data *pd, pd 70 net/netfilter/ipvs/ip_vs_proto_sctp.c *cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph); pd 73 net/netfilter/ipvs/ip_vs_proto_sctp.c *verdict = ip_vs_leave(svc, skb, pd, iph); pd 378 net/netfilter/ipvs/ip_vs_proto_sctp.c set_sctp_state(struct ip_vs_proto_data *pd, struct ip_vs_conn *cp, pd 441 net/netfilter/ipvs/ip_vs_proto_sctp.c pd->pp->name, pd 467 net/netfilter/ipvs/ip_vs_proto_sctp.c if (likely(pd)) pd 468 net/netfilter/ipvs/ip_vs_proto_sctp.c cp->timeout = pd->timeout_table[cp->state = next_state]; pd 475 net/netfilter/ipvs/ip_vs_proto_sctp.c const struct sk_buff *skb, struct ip_vs_proto_data *pd) pd 478 net/netfilter/ipvs/ip_vs_proto_sctp.c set_sctp_state(pd, cp, direction, skb); pd 494 net/netfilter/ipvs/ip_vs_proto_sctp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_SCTP); pd 505 net/netfilter/ipvs/ip_vs_proto_sctp.c atomic_inc(&pd->appcnt); pd 513 net/netfilter/ipvs/ip_vs_proto_sctp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_SCTP); pd 515 net/netfilter/ipvs/ip_vs_proto_sctp.c atomic_dec(&pd->appcnt); pd 559 net/netfilter/ipvs/ip_vs_proto_sctp.c static int __ip_vs_sctp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 562 net/netfilter/ipvs/ip_vs_proto_sctp.c pd->timeout_table = ip_vs_create_timeout_table((int *)sctp_timeouts, pd 564 net/netfilter/ipvs/ip_vs_proto_sctp.c if (!pd->timeout_table) pd 569 net/netfilter/ipvs/ip_vs_proto_sctp.c static void __ip_vs_sctp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 571 net/netfilter/ipvs/ip_vs_proto_sctp.c kfree(pd->timeout_table); pd 36 net/netfilter/ipvs/ip_vs_proto_tcp.c struct ip_vs_proto_data *pd, pd 90 net/netfilter/ipvs/ip_vs_proto_tcp.c *cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph); pd 93 net/netfilter/ipvs/ip_vs_proto_tcp.c *verdict = ip_vs_leave(svc, skb, pd, iph); pd 485 net/netfilter/ipvs/ip_vs_proto_tcp.c static void tcp_timeout_change(struct ip_vs_proto_data *pd, int flags) pd 495 net/netfilter/ipvs/ip_vs_proto_tcp.c pd->tcp_state_table = (on ? tcp_states_dos : tcp_states); pd 512 net/netfilter/ipvs/ip_vs_proto_tcp.c set_tcp_state(struct ip_vs_proto_data *pd, struct ip_vs_conn *cp, pd 536 net/netfilter/ipvs/ip_vs_proto_tcp.c pd->tcp_state_table[state_off+state_idx].next_state[cp->state]; pd 544 net/netfilter/ipvs/ip_vs_proto_tcp.c pd->pp->name, pd 576 net/netfilter/ipvs/ip_vs_proto_tcp.c if (likely(pd)) pd 577 net/netfilter/ipvs/ip_vs_proto_tcp.c cp->timeout = pd->timeout_table[cp->state = new_state]; pd 588 net/netfilter/ipvs/ip_vs_proto_tcp.c struct ip_vs_proto_data *pd) pd 603 net/netfilter/ipvs/ip_vs_proto_tcp.c set_tcp_state(pd, cp, direction, th); pd 620 net/netfilter/ipvs/ip_vs_proto_tcp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP); pd 631 net/netfilter/ipvs/ip_vs_proto_tcp.c atomic_inc(&pd->appcnt); pd 641 net/netfilter/ipvs/ip_vs_proto_tcp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_TCP); pd 643 net/netfilter/ipvs/ip_vs_proto_tcp.c atomic_dec(&pd->appcnt); pd 693 net/netfilter/ipvs/ip_vs_proto_tcp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(cp->ipvs, IPPROTO_TCP); pd 697 net/netfilter/ipvs/ip_vs_proto_tcp.c cp->timeout = (pd ? pd->timeout_table[IP_VS_TCP_S_LISTEN] pd 706 net/netfilter/ipvs/ip_vs_proto_tcp.c static int __ip_vs_tcp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 709 net/netfilter/ipvs/ip_vs_proto_tcp.c pd->timeout_table = ip_vs_create_timeout_table((int *)tcp_timeouts, pd 711 net/netfilter/ipvs/ip_vs_proto_tcp.c if (!pd->timeout_table) pd 713 net/netfilter/ipvs/ip_vs_proto_tcp.c pd->tcp_state_table = tcp_states; pd 717 net/netfilter/ipvs/ip_vs_proto_tcp.c static void __ip_vs_tcp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 719 net/netfilter/ipvs/ip_vs_proto_tcp.c kfree(pd->timeout_table); pd 32 net/netfilter/ipvs/ip_vs_proto_udp.c struct ip_vs_proto_data *pd, pd 78 net/netfilter/ipvs/ip_vs_proto_udp.c *cpp = ip_vs_schedule(svc, skb, pd, &ignored, iph); pd 81 net/netfilter/ipvs/ip_vs_proto_udp.c *verdict = ip_vs_leave(svc, skb, pd, iph); pd 367 net/netfilter/ipvs/ip_vs_proto_udp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP); pd 378 net/netfilter/ipvs/ip_vs_proto_udp.c atomic_inc(&pd->appcnt); pd 388 net/netfilter/ipvs/ip_vs_proto_udp.c struct ip_vs_proto_data *pd = ip_vs_proto_data_get(ipvs, IPPROTO_UDP); pd 390 net/netfilter/ipvs/ip_vs_proto_udp.c atomic_dec(&pd->appcnt); pd 454 net/netfilter/ipvs/ip_vs_proto_udp.c struct ip_vs_proto_data *pd) pd 456 net/netfilter/ipvs/ip_vs_proto_udp.c if (unlikely(!pd)) { pd 461 net/netfilter/ipvs/ip_vs_proto_udp.c cp->timeout = pd->timeout_table[IP_VS_UDP_S_NORMAL]; pd 466 net/netfilter/ipvs/ip_vs_proto_udp.c static int __udp_init(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 469 net/netfilter/ipvs/ip_vs_proto_udp.c pd->timeout_table = ip_vs_create_timeout_table((int *)udp_timeouts, pd 471 net/netfilter/ipvs/ip_vs_proto_udp.c if (!pd->timeout_table) pd 476 net/netfilter/ipvs/ip_vs_proto_udp.c static void __udp_exit(struct netns_ipvs *ipvs, struct ip_vs_proto_data *pd) pd 478 net/netfilter/ipvs/ip_vs_proto_udp.c kfree(pd->timeout_table); pd 948 net/netfilter/ipvs/ip_vs_sync.c struct ip_vs_proto_data *pd; pd 950 net/netfilter/ipvs/ip_vs_sync.c pd = ip_vs_proto_data_get(ipvs, protocol); pd 951 net/netfilter/ipvs/ip_vs_sync.c if (!(flags & IP_VS_CONN_F_TEMPLATE) && pd && pd->timeout_table) pd 952 net/netfilter/ipvs/ip_vs_sync.c cp->timeout = pd->timeout_table[state]; pd 1056 net/netfilter/ipvs/ip_vs_xmit.c __be16 *pd; pd 1066 net/netfilter/ipvs/ip_vs_xmit.c pd = data; pd 1067 net/netfilter/ipvs/ip_vs_xmit.c pd[0] = htons(csum_start); pd 1068 net/netfilter/ipvs/ip_vs_xmit.c pd[1] = htons(csum_start + skb->csum_offset); pd 108 net/rds/ib.c if (rds_ibdev->pd) pd 109 net/rds/ib.c ib_dealloc_pd(rds_ibdev->pd); pd 180 net/rds/ib.c rds_ibdev->pd = ib_alloc_pd(device, 0); pd 181 net/rds/ib.c if (IS_ERR(rds_ibdev->pd)) { pd 182 net/rds/ib.c rds_ibdev->pd = NULL; pd 248 net/rds/ib.h struct ib_pd *pd; pd 484 net/rds/ib_cm.c ic->i_pd = rds_ibdev->pd; pd 70 net/rds/ib_fmr.c fmr->fmr = ib_alloc_fmr(rds_ibdev->pd, pd 78 net/rds/ib_frmr.c frmr->mr = ib_alloc_mr(rds_ibdev->pd, IB_MR_TYPE_MEM_REG, pd 719 net/sctp/ulpevent.c struct sctp_pdapi_event *pd; pd 728 net/sctp/ulpevent.c pd = skb_put(skb, sizeof(struct sctp_pdapi_event)); pd 736 net/sctp/ulpevent.c pd->pdapi_type = SCTP_PARTIAL_DELIVERY_EVENT; pd 737 net/sctp/ulpevent.c pd->pdapi_flags = flags; pd 738 net/sctp/ulpevent.c pd->pdapi_stream = sid; pd 739 net/sctp/ulpevent.c pd->pdapi_seq = seq; pd 747 net/sctp/ulpevent.c pd->pdapi_length = sizeof(struct sctp_pdapi_event); pd 753 net/sctp/ulpevent.c pd->pdapi_indication = indication; pd 760 net/sctp/ulpevent.c pd->pdapi_assoc_id = sctp_assoc2id(asoc); pd 377 net/smc/smc_ib.c int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags, pd 384 net/smc/smc_ib.c ib_alloc_mr(pd, IB_MR_TYPE_MEM_REG, 1 << buf_slot->order); pd 72 net/smc/smc_ib.h int smc_ib_get_memory_region(struct ib_pd *pd, int access_flags, pd 493 security/selinux/ss/conditional.c struct policy_data *pd = ptr; pd 494 security/selinux/ss/conditional.c void *fp = pd->fp; pd 2599 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2600 security/selinux/ss/policydb.c void *fp = pd->fp; pd 2627 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2628 security/selinux/ss/policydb.c void *fp = pd->fp; pd 2762 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2763 security/selinux/ss/policydb.c void *fp = pd->fp; pd 2857 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2858 security/selinux/ss/policydb.c void *fp = pd->fp; pd 2859 security/selinux/ss/policydb.c struct policydb *p = pd->p; pd 2945 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2946 security/selinux/ss/policydb.c void *fp = pd->fp; pd 2947 security/selinux/ss/policydb.c struct policydb *p = pd->p; pd 2984 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 2985 security/selinux/ss/policydb.c struct policydb *p = pd->p; pd 2986 security/selinux/ss/policydb.c void *fp = pd->fp; pd 3025 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 3026 security/selinux/ss/policydb.c struct policydb *p = pd->p; pd 3027 security/selinux/ss/policydb.c void *fp = pd->fp; pd 3270 security/selinux/ss/policydb.c struct policy_data *pd = ptr; pd 3271 security/selinux/ss/policydb.c void *fp = pd->fp; pd 3272 security/selinux/ss/policydb.c struct policydb *p = pd->p; pd 3297 security/selinux/ss/policydb.c struct policy_data pd; pd 3299 security/selinux/ss/policydb.c pd.p = p; pd 3300 security/selinux/ss/policydb.c pd.fp = fp; pd 3314 security/selinux/ss/policydb.c rc = hashtab_map(p->range_tr, range_write_helper, &pd); pd 3456 security/selinux/ss/policydb.c struct policy_data pd; pd 3458 security/selinux/ss/policydb.c pd.fp = fp; pd 3459 security/selinux/ss/policydb.c pd.p = p; pd 3467 security/selinux/ss/policydb.c rc = hashtab_map(p->symtab[i].table, write_f[i], &pd); pd 161 sound/isa/wavefront/wavefront_fx.c unsigned short *pd; pd 187 sound/isa/wavefront/wavefront_fx.c pd = (unsigned short *) &r.data[3]; pd 199 sound/isa/wavefront/wavefront_fx.c pd = page_data; pd 206 sound/isa/wavefront/wavefront_fx.c pd); pd 870 sound/pci/asihpi/asihpi.c char *pd = &s->runtime->dma_area[buf_ofs]; pd 885 sound/pci/asihpi/asihpi.c ds->h_stream, pd, xfer1, pd 889 sound/pci/asihpi/asihpi.c pd = s->runtime->dma_area; pd 896 sound/pci/asihpi/asihpi.c ds->h_stream, pd, pd 906 sound/pci/asihpi/asihpi.c pd, xfer1)); pd 908 sound/pci/asihpi/asihpi.c pd = s->runtime->dma_area; pd 914 sound/pci/asihpi/asihpi.c pd, xfer2)); pd 51 sound/sh/aica.c static struct platform_device *pd; pd 536 sound/sh/aica.c err = request_firmware(&fw_entry, "aica_firmware.bin", &pd->dev); pd 630 sound/sh/aica.c pd = platform_device_register_simple(SND_AICA_DRIVER, -1, pd 632 sound/sh/aica.c if (IS_ERR(pd)) { pd 634 sound/sh/aica.c return PTR_ERR(pd); pd 642 sound/sh/aica.c platform_device_unregister(pd); pd 400 sound/soc/codecs/cros_ec_codec.c static int cros_ec_codec_platform_probe(struct platform_device *pd) pd 402 sound/soc/codecs/cros_ec_codec.c struct device *dev = &pd->dev; pd 403 sound/soc/codecs/cros_ec_codec.c struct cros_ec_device *ec_device = dev_get_drvdata(pd->dev.parent); pd 414 sound/soc/codecs/cros_ec_codec.c platform_set_drvdata(pd, codec_data); pd 28 sound/soc/codecs/rl6231.c int pd, val; pd 39 sound/soc/codecs/rl6231.c pd = val + 1; pd 42 sound/soc/codecs/rl6231.c pd = 6; pd 45 sound/soc/codecs/rl6231.c pd = 8; pd 48 sound/soc/codecs/rl6231.c pd = 12; pd 51 sound/soc/codecs/rl6231.c pd = 16; pd 54 sound/soc/codecs/rl6231.c pd = -EINVAL; pd 58 sound/soc/codecs/rl6231.c return pd; pd 225 sound/soc/codecs/rl6231.c static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16}; pd 231 sound/soc/codecs/rl6231.c for (i = 0; i < ARRAY_SIZE(pd); i++) pd 232 sound/soc/codecs/rl6231.c if (sclk == rate * pd[i]) pd 1523 sound/soc/codecs/rt1011.c static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16}; pd 1529 sound/soc/codecs/rt1011.c for (i = 0; i < ARRAY_SIZE(pd); i++) pd 1530 sound/soc/codecs/rt1011.c if (sclk == rate * pd[i]) pd 612 sound/soc/codecs/rt1305.c static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16}; pd 618 sound/soc/codecs/rt1305.c for (i = 0; i < ARRAY_SIZE(pd); i++) pd 619 sound/soc/codecs/rt1305.c if (sclk == rate * pd[i]) pd 440 sound/soc/codecs/rt1308.c static const int pd[] = {1, 2, 3, 4, 6, 8, 12, 16}; pd 446 sound/soc/codecs/rt1308.c for (i = 0; i < ARRAY_SIZE(pd); i++) pd 447 sound/soc/codecs/rt1308.c if (sclk == rate * pd[i]) pd 1607 sound/soc/codecs/rt5659.c int pd, idx = -EINVAL; pd 1609 sound/soc/codecs/rt5659.c pd = rl6231_get_pre_div(rt5659->regmap, pd 1611 sound/soc/codecs/rt5659.c idx = rl6231_calc_dmic_clk(rt5659->sysclk / pd); pd 1481 sound/soc/codecs/rt5665.c int pd, idx; pd 1483 sound/soc/codecs/rt5665.c pd = rl6231_get_pre_div(rt5665->regmap, pd 1485 sound/soc/codecs/rt5665.c idx = rl6231_calc_dmic_clk(rt5665->sysclk / pd); pd 39 sound/soc/codecs/uda134x.c struct uda134x_platform_data *pd; pd 67 sound/soc/codecs/uda134x.c struct uda134x_platform_data *pd = context; pd 92 sound/soc/codecs/uda134x.c ret = l3_write(&pd->l3, pd 102 sound/soc/codecs/uda134x.c ret = l3_write(&pd->l3, pd 298 sound/soc/codecs/uda134x.c struct uda134x_platform_data *pd = uda134x->pd; pd 306 sound/soc/codecs/uda134x.c if (pd->power) { pd 307 sound/soc/codecs/uda134x.c pd->power(1); pd 315 sound/soc/codecs/uda134x.c if (pd->power) { pd 316 sound/soc/codecs/uda134x.c pd->power(0); pd 450 sound/soc/codecs/uda134x.c struct uda134x_platform_data *pd = uda134x->pd; pd 457 sound/soc/codecs/uda134x.c switch (pd->model) { pd 466 sound/soc/codecs/uda134x.c pd->model); pd 470 sound/soc/codecs/uda134x.c if (pd->power) pd 471 sound/soc/codecs/uda134x.c pd->power(1); pd 475 sound/soc/codecs/uda134x.c if (pd->model == UDA134X_UDA1341) { pd 490 sound/soc/codecs/uda134x.c switch (pd->model) { pd 506 sound/soc/codecs/uda134x.c __func__, pd->model); pd 545 sound/soc/codecs/uda134x.c struct uda134x_platform_data *pd = pdev->dev.platform_data; pd 549 sound/soc/codecs/uda134x.c if (!pd) { pd 558 sound/soc/codecs/uda134x.c uda134x->pd = pd; pd 561 sound/soc/codecs/uda134x.c if (pd->l3.use_gpios) { pd 562 sound/soc/codecs/uda134x.c ret = l3_set_gpio_ops(&pdev->dev, &uda134x->pd->l3); pd 567 sound/soc/codecs/uda134x.c uda134x->regmap = devm_regmap_init(&pdev->dev, NULL, pd, pd 557 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.data = pdata->play_dma_data; pd 558 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.data = pdata->capture_dma_data; pd 559 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.addr = res->start + I2S_TXDMA; pd 560 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.addr = res->start + I2S_RXDMA; pd 561 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.max_burst = 16; pd 562 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.max_burst = 16; pd 563 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.addr_width = bus_widths[idx]; pd 564 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.addr_width = bus_widths[idx]; pd 565 sound/soc/dwc/dwc-i2s.c dev->play_dma_data.pd.filter = pdata->filter; pd 566 sound/soc/dwc/dwc-i2s.c dev->capture_dma_data.pd.filter = pdata->filter; pd 85 sound/soc/dwc/local.h struct i2s_dma_data pd; pd 198 sound/soc/samsung/s3c24xx_simtec.c static int simtec_call_startup(struct s3c24xx_audio_simtec_pdata *pd) pd 203 sound/soc/samsung/s3c24xx_simtec.c if (pd->startup) pd 204 sound/soc/samsung/s3c24xx_simtec.c pd->startup(); pd 222 sound/soc/samsung/s3c24xx_simtec.c struct s3c24xx_audio_simtec_pdata *pd) pd 228 sound/soc/samsung/s3c24xx_simtec.c ret = gpio_request(pd->amp_gain[0], "gpio-amp-gain0"); pd 234 sound/soc/samsung/s3c24xx_simtec.c ret = gpio_request(pd->amp_gain[1], "gpio-amp-gain1"); pd 241 sound/soc/samsung/s3c24xx_simtec.c gpio_direction_output(pd->amp_gain[0], 0); pd 242 sound/soc/samsung/s3c24xx_simtec.c gpio_direction_output(pd->amp_gain[1], 0); pd 247 sound/soc/samsung/s3c24xx_simtec.c ret = gpio_request(pd->amp_gpio, "gpio-amp"); pd 250 sound/soc/samsung/s3c24xx_simtec.c pd->amp_gpio, ret); pd 261 sound/soc/samsung/s3c24xx_simtec.c if (pd->amp_gain[0] > 0) { pd 262 sound/soc/samsung/s3c24xx_simtec.c gpio_free(pd->amp_gain[0]); pd 263 sound/soc/samsung/s3c24xx_simtec.c gpio_free(pd->amp_gain[1]); pd 269 sound/soc/samsung/s3c24xx_simtec.c static void detach_gpio_amp(struct s3c24xx_audio_simtec_pdata *pd) pd 271 sound/soc/samsung/s3c24xx_simtec.c if (pd->amp_gain[0] > 0) { pd 272 sound/soc/samsung/s3c24xx_simtec.c gpio_free(pd->amp_gain[0]); pd 273 sound/soc/samsung/s3c24xx_simtec.c gpio_free(pd->amp_gain[1]); pd 276 sound/soc/samsung/s3c24xx_simtec.c if (pd->amp_gpio > 0) pd 277 sound/soc/samsung/s3c24xx_simtec.c gpio_free(pd->amp_gpio); pd 92 sound/soc/samsung/s3c24xx_simtec_hermes.c static int simtec_audio_hermes_probe(struct platform_device *pd) pd 94 sound/soc/samsung/s3c24xx_simtec_hermes.c dev_info(&pd->dev, "probing....\n"); pd 95 sound/soc/samsung/s3c24xx_simtec_hermes.c return simtec_audio_core_probe(pd, &snd_soc_machine_simtec_aic33); pd 81 sound/soc/samsung/s3c24xx_simtec_tlv320aic23.c static int simtec_audio_tlv320aic23_probe(struct platform_device *pd) pd 83 sound/soc/samsung/s3c24xx_simtec_tlv320aic23.c return simtec_audio_core_probe(pd, &snd_soc_machine_simtec_aic23); pd 746 sound/soc/sof/pcm.c struct snd_soc_component_driver *pd = &sdev->plat_drv; pd 752 sound/soc/sof/pcm.c pd->name = "sof-audio-component"; pd 753 sound/soc/sof/pcm.c pd->probe = sof_pcm_probe; pd 754 sound/soc/sof/pcm.c pd->remove = sof_pcm_remove; pd 755 sound/soc/sof/pcm.c pd->ops = &sof_pcm_ops; pd 757 sound/soc/sof/pcm.c pd->compr_ops = &sof_compressed_ops; pd 759 sound/soc/sof/pcm.c pd->pcm_new = sof_pcm_new; pd 760 sound/soc/sof/pcm.c pd->ignore_machine = drv_name; pd 761 sound/soc/sof/pcm.c pd->be_hw_params_fixup = sof_pcm_dai_link_fixup; pd 762 sound/soc/sof/pcm.c pd->be_pcm_base = SOF_BE_PCM_BASE; pd 763 sound/soc/sof/pcm.c pd->use_dai_pcm_id = true; pd 764 sound/soc/sof/pcm.c pd->topology_name_prefix = "sof"; pd 767 sound/soc/sof/pcm.c pd->module_get_upon_open = 1; pd 20 sound/usb/power.c struct snd_usb_power_domain *pd; pd 23 sound/usb/power.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 24 sound/usb/power.c if (!pd) pd 38 sound/usb/power.c pd->pd_id = pd_desc->bPowerDomainID; pd 39 sound/usb/power.c pd->pd_d1d0_rec = pd 41 sound/usb/power.c pd->pd_d2d0_rec = pd 43 sound/usb/power.c return pd; pd 48 sound/usb/power.c kfree(pd); pd 53 sound/usb/power.c struct snd_usb_power_domain *pd, pd 60 sound/usb/power.c idx = snd_usb_ctrl_intf(chip) | (pd->pd_id << 8); pd 69 sound/usb/power.c pd->pd_id); pd 75 sound/usb/power.c pd->pd_id, state); pd 85 sound/usb/power.c state, pd->pd_id); pd 92 sound/usb/power.c udelay(pd->pd_d2d0_rec * 50); pd 95 sound/usb/power.c udelay(pd->pd_d1d0_rec * 50); pd 103 sound/usb/power.c pd->pd_id, state); pd 18 sound/usb/power.h struct snd_usb_power_domain *pd, pd 83 sound/usb/stream.c struct snd_usb_power_domain *pd) pd 108 sound/usb/stream.c if (pd) { pd 109 sound/usb/stream.c subs->str_pd = pd; pd 111 sound/usb/stream.c snd_usb_power_domain_set(subs->stream->chip, pd, pd 479 sound/usb/stream.c struct snd_usb_power_domain *pd) pd 508 sound/usb/stream.c snd_usb_init_substream(as, stream, fp, pd); pd 536 sound/usb/stream.c snd_usb_init_substream(as, stream, fp, pd); pd 564 sound/usb/stream.c struct snd_usb_power_domain *pd) pd 566 sound/usb/stream.c return __snd_usb_add_audio_stream(chip, stream, fp, pd); pd 867 sound/usb/stream.c struct snd_usb_power_domain *pd; pd 1049 sound/usb/stream.c pd = kzalloc(sizeof(*pd), GFP_KERNEL); pd 1050 sound/usb/stream.c if (!pd) { pd 1054 sound/usb/stream.c pd->pd_id = (stream == SNDRV_PCM_STREAM_PLAYBACK) ? pd 1056 sound/usb/stream.c pd->pd_d1d0_rec = UAC3_BADD_PD_RECOVER_D1D0; pd 1057 sound/usb/stream.c pd->pd_d2d0_rec = UAC3_BADD_PD_RECOVER_D2D0; pd 1064 sound/usb/stream.c pd = snd_usb_find_power_domain(chip->ctrl_intf, pd 1069 sound/usb/stream.c kfree(pd); pd 1075 sound/usb/stream.c if (pd) pd 1076 sound/usb/stream.c *pd_out = pd; pd 1091 sound/usb/stream.c struct snd_usb_power_domain *pd = NULL; pd 1174 sound/usb/stream.c fp = snd_usb_get_audioformat_uac3(chip, alts, &pd, pd 1188 sound/usb/stream.c kfree(pd); pd 1190 sound/usb/stream.c pd = NULL; pd 1196 sound/usb/stream.c err = snd_usb_add_audio_stream_v3(chip, stream, fp, pd); pd 1202 sound/usb/stream.c kfree(pd); pd 327 tools/perf/pmu-events/jevents.c struct perf_entry_data *pd = data; pd 328 tools/perf/pmu-events/jevents.c FILE *outfp = pd->outfp; pd 329 tools/perf/pmu-events/jevents.c char *topic = pd->topic;