new_rl 3197 drivers/infiniband/hw/mlx5/qp.c struct mlx5_rate_limit new_rl = old_rl; new_rl 3221 drivers/infiniband/hw/mlx5/qp.c new_rl = raw_qp_param->rl; new_rl 3224 drivers/infiniband/hw/mlx5/qp.c if (!mlx5_rl_are_equal(&old_rl, &new_rl)) { new_rl 3225 drivers/infiniband/hw/mlx5/qp.c if (new_rl.rate) { new_rl 3226 drivers/infiniband/hw/mlx5/qp.c err = mlx5_rl_add_rate(dev, &rl_index, &new_rl); new_rl 3230 drivers/infiniband/hw/mlx5/qp.c err, new_rl.rate, new_rl.max_burst_sz, new_rl 3231 drivers/infiniband/hw/mlx5/qp.c new_rl.typical_pkt_sz); new_rl 3247 drivers/infiniband/hw/mlx5/qp.c mlx5_rl_remove_rate(dev, &new_rl); new_rl 3252 drivers/infiniband/hw/mlx5/qp.c if ((old_rl.rate && !mlx5_rl_are_equal(&old_rl, &new_rl)) || new_rl 3256 drivers/infiniband/hw/mlx5/qp.c memset(&new_rl, 0, sizeof(new_rl)); new_rl 3259 drivers/infiniband/hw/mlx5/qp.c ibqp->rl = new_rl; new_rl 63 fs/ntfs/runlist.c runlist_element *new_rl; new_rl 70 fs/ntfs/runlist.c new_rl = ntfs_malloc_nofs(new_size); new_rl 71 fs/ntfs/runlist.c if (unlikely(!new_rl)) new_rl 77 fs/ntfs/runlist.c memcpy(new_rl, rl, old_size); new_rl 80 fs/ntfs/runlist.c return new_rl; new_rl 109 fs/ntfs/runlist.c runlist_element *new_rl; new_rl 116 fs/ntfs/runlist.c new_rl = ntfs_malloc_nofs_nofail(new_size); new_rl 117 fs/ntfs/runlist.c BUG_ON(!new_rl); new_rl 122 fs/ntfs/runlist.c memcpy(new_rl, rl, old_size); new_rl 125 fs/ntfs/runlist.c return new_rl; new_rl 1457 fs/ocfs2/refcounttree.c struct ocfs2_refcount_list *new_rl = &new_rb->rf_records; new_rl 1489 fs/ocfs2/refcounttree.c memcpy(new_rl->rl_recs, &rl->rl_recs[split_index], new_rl 1498 fs/ocfs2/refcounttree.c new_rl->rl_used = cpu_to_le16(num_moved); new_rl 1504 fs/ocfs2/refcounttree.c sort(&new_rl->rl_recs, le16_to_cpu(new_rl->rl_used),