psl 804 arch/x86/kernel/cpu/resctrl/rdtgroup.c int i, hwb, swb, excl, psl; psl 850 arch/x86/kernel/cpu/resctrl/rdtgroup.c psl = test_bit(i, &pseudo_locked); psl 859 arch/x86/kernel/cpu/resctrl/rdtgroup.c else if (psl) psl 393 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c ATOM_PPLIB_PhaseSheddingLimits_Table *psl = psl 400 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c kcalloc(psl->ucNumEntries, psl 408 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c entry = &psl->entries[0]; psl 409 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c for (i = 0; i < psl->ucNumEntries; i++) { psl 420 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c psl->ucNumEntries; psl 986 drivers/gpu/drm/radeon/r600_dpm.c ATOM_PPLIB_PhaseSheddingLimits_Table *psl = psl 993 drivers/gpu/drm/radeon/r600_dpm.c kcalloc(psl->ucNumEntries, psl 1001 drivers/gpu/drm/radeon/r600_dpm.c entry = &psl->entries[0]; psl 1002 drivers/gpu/drm/radeon/r600_dpm.c for (i = 0; i < psl->ucNumEntries; i++) { psl 1013 drivers/gpu/drm/radeon/r600_dpm.c psl->ucNumEntries; psl 2299 net/ipv4/igmp.c struct ip_sf_socklist *psl; psl 2344 net/ipv4/igmp.c psl = rtnl_dereference(pmc->sflist); psl 2346 net/ipv4/igmp.c if (!psl) psl 2349 net/ipv4/igmp.c for (i = 0; i < psl->sl_count; i++) { psl 2350 net/ipv4/igmp.c rv = memcmp(&psl->sl_addr[i], &mreqs->imr_sourceaddr, psl 2359 net/ipv4/igmp.c if (psl->sl_count == 1 && omode == MCAST_INCLUDE) { psl 2368 net/ipv4/igmp.c for (j = i+1; j < psl->sl_count; j++) psl 2369 net/ipv4/igmp.c psl->sl_addr[j-1] = psl->sl_addr[j]; psl 2370 net/ipv4/igmp.c psl->sl_count--; psl 2376 net/ipv4/igmp.c if (psl && psl->sl_count >= net->ipv4.sysctl_igmp_max_msf) { psl 2380 net/ipv4/igmp.c if (!psl || psl->sl_count == psl->sl_max) { psl 2384 net/ipv4/igmp.c if (psl) psl 2385 net/ipv4/igmp.c count += psl->sl_max; psl 2393 net/ipv4/igmp.c if (psl) { psl 2394 net/ipv4/igmp.c for (i = 0; i < psl->sl_count; i++) psl 2395 net/ipv4/igmp.c newpsl->sl_addr[i] = psl->sl_addr[i]; psl 2397 net/ipv4/igmp.c atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc); psl 2398 net/ipv4/igmp.c kfree_rcu(psl, rcu); psl 2401 net/ipv4/igmp.c psl = newpsl; psl 2404 net/ipv4/igmp.c for (i = 0; i < psl->sl_count; i++) { psl 2405 net/ipv4/igmp.c rv = memcmp(&psl->sl_addr[i], &mreqs->imr_sourceaddr, psl 2412 net/ipv4/igmp.c for (j = psl->sl_count-1; j >= i; j--) psl 2413 net/ipv4/igmp.c psl->sl_addr[j+1] = psl->sl_addr[j]; psl 2414 net/ipv4/igmp.c psl->sl_addr[i] = mreqs->imr_sourceaddr; psl 2415 net/ipv4/igmp.c psl->sl_count++; psl 2434 net/ipv4/igmp.c struct ip_sf_socklist *newpsl, *psl; psl 2492 net/ipv4/igmp.c psl = rtnl_dereference(pmc->sflist); psl 2493 net/ipv4/igmp.c if (psl) { psl 2495 net/ipv4/igmp.c psl->sl_count, psl->sl_addr, 0); psl 2497 net/ipv4/igmp.c atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc); psl 2498 net/ipv4/igmp.c kfree_rcu(psl, rcu); psl 2520 net/ipv4/igmp.c struct ip_sf_socklist *psl; psl 2547 net/ipv4/igmp.c psl = rtnl_dereference(pmc->sflist); psl 2548 net/ipv4/igmp.c if (!psl) { psl 2552 net/ipv4/igmp.c count = psl->sl_count; psl 2555 net/ipv4/igmp.c len = copycount * sizeof(psl->sl_addr[0]); psl 2562 net/ipv4/igmp.c copy_to_user(&optval->imsf_slist[0], psl->sl_addr, len)) psl 2577 net/ipv4/igmp.c struct ip_sf_socklist *psl; psl 2598 net/ipv4/igmp.c psl = rtnl_dereference(pmc->sflist); psl 2599 net/ipv4/igmp.c count = psl ? psl->sl_count : 0; psl 2612 net/ipv4/igmp.c psin->sin_addr.s_addr = psl->sl_addr[i]; psl 2629 net/ipv4/igmp.c struct ip_sf_socklist *psl; psl 2647 net/ipv4/igmp.c psl = rcu_dereference(pmc->sflist); psl 2649 net/ipv4/igmp.c if (!psl) psl 2652 net/ipv4/igmp.c for (i = 0; i < psl->sl_count; i++) { psl 2653 net/ipv4/igmp.c if (psl->sl_addr[i] == rmt_addr) psl 2657 net/ipv4/igmp.c if (pmc->sfmode == MCAST_INCLUDE && i >= psl->sl_count) psl 2659 net/ipv4/igmp.c if (pmc->sfmode == MCAST_EXCLUDE && i < psl->sl_count) psl 335 net/ipv6/mcast.c struct ip6_sf_socklist *psl; psl 383 net/ipv6/mcast.c psl = pmc->sflist; psl 385 net/ipv6/mcast.c if (!psl) psl 388 net/ipv6/mcast.c for (i = 0; i < psl->sl_count; i++) { psl 389 net/ipv6/mcast.c rv = !ipv6_addr_equal(&psl->sl_addr[i], source); psl 397 net/ipv6/mcast.c if (psl->sl_count == 1 && omode == MCAST_INCLUDE) { psl 405 net/ipv6/mcast.c for (j = i+1; j < psl->sl_count; j++) psl 406 net/ipv6/mcast.c psl->sl_addr[j-1] = psl->sl_addr[j]; psl 407 net/ipv6/mcast.c psl->sl_count--; psl 413 net/ipv6/mcast.c if (psl && psl->sl_count >= sysctl_mld_max_msf) { psl 417 net/ipv6/mcast.c if (!psl || psl->sl_count == psl->sl_max) { psl 421 net/ipv6/mcast.c if (psl) psl 422 net/ipv6/mcast.c count += psl->sl_max; psl 430 net/ipv6/mcast.c if (psl) { psl 431 net/ipv6/mcast.c for (i = 0; i < psl->sl_count; i++) psl 432 net/ipv6/mcast.c newpsl->sl_addr[i] = psl->sl_addr[i]; psl 433 net/ipv6/mcast.c sock_kfree_s(sk, psl, IP6_SFLSIZE(psl->sl_max)); psl 435 net/ipv6/mcast.c pmc->sflist = psl = newpsl; psl 438 net/ipv6/mcast.c for (i = 0; i < psl->sl_count; i++) { psl 439 net/ipv6/mcast.c rv = !ipv6_addr_equal(&psl->sl_addr[i], source); psl 443 net/ipv6/mcast.c for (j = psl->sl_count-1; j >= i; j--) psl 444 net/ipv6/mcast.c psl->sl_addr[j+1] = psl->sl_addr[j]; psl 445 net/ipv6/mcast.c psl->sl_addr[i] = *source; psl 446 net/ipv6/mcast.c psl->sl_count++; psl 466 net/ipv6/mcast.c struct ip6_sf_socklist *newpsl, *psl; psl 530 net/ipv6/mcast.c psl = pmc->sflist; psl 531 net/ipv6/mcast.c if (psl) { psl 533 net/ipv6/mcast.c psl->sl_count, psl->sl_addr, 0); psl 534 net/ipv6/mcast.c sock_kfree_s(sk, psl, IP6_SFLSIZE(psl->sl_max)); psl 557 net/ipv6/mcast.c struct ip6_sf_socklist *psl; psl 588 net/ipv6/mcast.c psl = pmc->sflist; psl 589 net/ipv6/mcast.c count = psl ? psl->sl_count : 0; psl 609 net/ipv6/mcast.c psin6->sin6_addr = psl->sl_addr[i]; psl 625 net/ipv6/mcast.c struct ip6_sf_socklist *psl; psl 638 net/ipv6/mcast.c psl = mc->sflist; psl 639 net/ipv6/mcast.c if (!psl) { psl 644 net/ipv6/mcast.c for (i = 0; i < psl->sl_count; i++) { psl 645 net/ipv6/mcast.c if (ipv6_addr_equal(&psl->sl_addr[i], src_addr)) psl 648 net/ipv6/mcast.c if (mc->sfmode == MCAST_INCLUDE && i >= psl->sl_count) psl 650 net/ipv6/mcast.c if (mc->sfmode == MCAST_EXCLUDE && i < psl->sl_count)