psl               804 arch/x86/kernel/cpu/resctrl/rdtgroup.c 	int i, hwb, swb, excl, psl;
psl               850 arch/x86/kernel/cpu/resctrl/rdtgroup.c 			psl = test_bit(i, &pseudo_locked);
psl               859 arch/x86/kernel/cpu/resctrl/rdtgroup.c 			else if (psl)
psl               393 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			ATOM_PPLIB_PhaseSheddingLimits_Table *psl =
psl               400 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				kcalloc(psl->ucNumEntries,
psl               408 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			entry = &psl->entries[0];
psl               409 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 			for (i = 0; i < psl->ucNumEntries; i++) {
psl               420 drivers/gpu/drm/amd/amdgpu/amdgpu_dpm.c 				psl->ucNumEntries;
psl               986 drivers/gpu/drm/radeon/r600_dpm.c 			ATOM_PPLIB_PhaseSheddingLimits_Table *psl =
psl               993 drivers/gpu/drm/radeon/r600_dpm.c 				kcalloc(psl->ucNumEntries,
psl              1001 drivers/gpu/drm/radeon/r600_dpm.c 			entry = &psl->entries[0];
psl              1002 drivers/gpu/drm/radeon/r600_dpm.c 			for (i = 0; i < psl->ucNumEntries; i++) {
psl              1013 drivers/gpu/drm/radeon/r600_dpm.c 				psl->ucNumEntries;
psl              2299 net/ipv4/igmp.c 	struct ip_sf_socklist *psl;
psl              2344 net/ipv4/igmp.c 	psl = rtnl_dereference(pmc->sflist);
psl              2346 net/ipv4/igmp.c 		if (!psl)
psl              2349 net/ipv4/igmp.c 		for (i = 0; i < psl->sl_count; i++) {
psl              2350 net/ipv4/igmp.c 			rv = memcmp(&psl->sl_addr[i], &mreqs->imr_sourceaddr,
psl              2359 net/ipv4/igmp.c 		if (psl->sl_count == 1 && omode == MCAST_INCLUDE) {
psl              2368 net/ipv4/igmp.c 		for (j = i+1; j < psl->sl_count; j++)
psl              2369 net/ipv4/igmp.c 			psl->sl_addr[j-1] = psl->sl_addr[j];
psl              2370 net/ipv4/igmp.c 		psl->sl_count--;
psl              2376 net/ipv4/igmp.c 	if (psl && psl->sl_count >= net->ipv4.sysctl_igmp_max_msf) {
psl              2380 net/ipv4/igmp.c 	if (!psl || psl->sl_count == psl->sl_max) {
psl              2384 net/ipv4/igmp.c 		if (psl)
psl              2385 net/ipv4/igmp.c 			count += psl->sl_max;
psl              2393 net/ipv4/igmp.c 		if (psl) {
psl              2394 net/ipv4/igmp.c 			for (i = 0; i < psl->sl_count; i++)
psl              2395 net/ipv4/igmp.c 				newpsl->sl_addr[i] = psl->sl_addr[i];
psl              2397 net/ipv4/igmp.c 			atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc);
psl              2398 net/ipv4/igmp.c 			kfree_rcu(psl, rcu);
psl              2401 net/ipv4/igmp.c 		psl = newpsl;
psl              2404 net/ipv4/igmp.c 	for (i = 0; i < psl->sl_count; i++) {
psl              2405 net/ipv4/igmp.c 		rv = memcmp(&psl->sl_addr[i], &mreqs->imr_sourceaddr,
psl              2412 net/ipv4/igmp.c 	for (j = psl->sl_count-1; j >= i; j--)
psl              2413 net/ipv4/igmp.c 		psl->sl_addr[j+1] = psl->sl_addr[j];
psl              2414 net/ipv4/igmp.c 	psl->sl_addr[i] = mreqs->imr_sourceaddr;
psl              2415 net/ipv4/igmp.c 	psl->sl_count++;
psl              2434 net/ipv4/igmp.c 	struct ip_sf_socklist *newpsl, *psl;
psl              2492 net/ipv4/igmp.c 	psl = rtnl_dereference(pmc->sflist);
psl              2493 net/ipv4/igmp.c 	if (psl) {
psl              2495 net/ipv4/igmp.c 			psl->sl_count, psl->sl_addr, 0);
psl              2497 net/ipv4/igmp.c 		atomic_sub(IP_SFLSIZE(psl->sl_max), &sk->sk_omem_alloc);
psl              2498 net/ipv4/igmp.c 		kfree_rcu(psl, rcu);
psl              2520 net/ipv4/igmp.c 	struct ip_sf_socklist *psl;
psl              2547 net/ipv4/igmp.c 	psl = rtnl_dereference(pmc->sflist);
psl              2548 net/ipv4/igmp.c 	if (!psl) {
psl              2552 net/ipv4/igmp.c 		count = psl->sl_count;
psl              2555 net/ipv4/igmp.c 	len = copycount * sizeof(psl->sl_addr[0]);
psl              2562 net/ipv4/igmp.c 	    copy_to_user(&optval->imsf_slist[0], psl->sl_addr, len))
psl              2577 net/ipv4/igmp.c 	struct ip_sf_socklist *psl;
psl              2598 net/ipv4/igmp.c 	psl = rtnl_dereference(pmc->sflist);
psl              2599 net/ipv4/igmp.c 	count = psl ? psl->sl_count : 0;
psl              2612 net/ipv4/igmp.c 		psin->sin_addr.s_addr = psl->sl_addr[i];
psl              2629 net/ipv4/igmp.c 	struct ip_sf_socklist *psl;
psl              2647 net/ipv4/igmp.c 	psl = rcu_dereference(pmc->sflist);
psl              2649 net/ipv4/igmp.c 	if (!psl)
psl              2652 net/ipv4/igmp.c 	for (i = 0; i < psl->sl_count; i++) {
psl              2653 net/ipv4/igmp.c 		if (psl->sl_addr[i] == rmt_addr)
psl              2657 net/ipv4/igmp.c 	if (pmc->sfmode == MCAST_INCLUDE && i >= psl->sl_count)
psl              2659 net/ipv4/igmp.c 	if (pmc->sfmode == MCAST_EXCLUDE && i < psl->sl_count)
psl               335 net/ipv6/mcast.c 	struct ip6_sf_socklist *psl;
psl               383 net/ipv6/mcast.c 	psl = pmc->sflist;
psl               385 net/ipv6/mcast.c 		if (!psl)
psl               388 net/ipv6/mcast.c 		for (i = 0; i < psl->sl_count; i++) {
psl               389 net/ipv6/mcast.c 			rv = !ipv6_addr_equal(&psl->sl_addr[i], source);
psl               397 net/ipv6/mcast.c 		if (psl->sl_count == 1 && omode == MCAST_INCLUDE) {
psl               405 net/ipv6/mcast.c 		for (j = i+1; j < psl->sl_count; j++)
psl               406 net/ipv6/mcast.c 			psl->sl_addr[j-1] = psl->sl_addr[j];
psl               407 net/ipv6/mcast.c 		psl->sl_count--;
psl               413 net/ipv6/mcast.c 	if (psl && psl->sl_count >= sysctl_mld_max_msf) {
psl               417 net/ipv6/mcast.c 	if (!psl || psl->sl_count == psl->sl_max) {
psl               421 net/ipv6/mcast.c 		if (psl)
psl               422 net/ipv6/mcast.c 			count += psl->sl_max;
psl               430 net/ipv6/mcast.c 		if (psl) {
psl               431 net/ipv6/mcast.c 			for (i = 0; i < psl->sl_count; i++)
psl               432 net/ipv6/mcast.c 				newpsl->sl_addr[i] = psl->sl_addr[i];
psl               433 net/ipv6/mcast.c 			sock_kfree_s(sk, psl, IP6_SFLSIZE(psl->sl_max));
psl               435 net/ipv6/mcast.c 		pmc->sflist = psl = newpsl;
psl               438 net/ipv6/mcast.c 	for (i = 0; i < psl->sl_count; i++) {
psl               439 net/ipv6/mcast.c 		rv = !ipv6_addr_equal(&psl->sl_addr[i], source);
psl               443 net/ipv6/mcast.c 	for (j = psl->sl_count-1; j >= i; j--)
psl               444 net/ipv6/mcast.c 		psl->sl_addr[j+1] = psl->sl_addr[j];
psl               445 net/ipv6/mcast.c 	psl->sl_addr[i] = *source;
psl               446 net/ipv6/mcast.c 	psl->sl_count++;
psl               466 net/ipv6/mcast.c 	struct ip6_sf_socklist *newpsl, *psl;
psl               530 net/ipv6/mcast.c 	psl = pmc->sflist;
psl               531 net/ipv6/mcast.c 	if (psl) {
psl               533 net/ipv6/mcast.c 			psl->sl_count, psl->sl_addr, 0);
psl               534 net/ipv6/mcast.c 		sock_kfree_s(sk, psl, IP6_SFLSIZE(psl->sl_max));
psl               557 net/ipv6/mcast.c 	struct ip6_sf_socklist *psl;
psl               588 net/ipv6/mcast.c 	psl = pmc->sflist;
psl               589 net/ipv6/mcast.c 	count = psl ? psl->sl_count : 0;
psl               609 net/ipv6/mcast.c 		psin6->sin6_addr = psl->sl_addr[i];
psl               625 net/ipv6/mcast.c 	struct ip6_sf_socklist *psl;
psl               638 net/ipv6/mcast.c 	psl = mc->sflist;
psl               639 net/ipv6/mcast.c 	if (!psl) {
psl               644 net/ipv6/mcast.c 		for (i = 0; i < psl->sl_count; i++) {
psl               645 net/ipv6/mcast.c 			if (ipv6_addr_equal(&psl->sl_addr[i], src_addr))
psl               648 net/ipv6/mcast.c 		if (mc->sfmode == MCAST_INCLUDE && i >= psl->sl_count)
psl               650 net/ipv6/mcast.c 		if (mc->sfmode == MCAST_EXCLUDE && i < psl->sl_count)