mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-12-29 17:25:38 +00:00
all: replace find_next{,_zero}_bit with find_first{,_zero}_bit where appropriate
find_first{,_zero}_bit is a more effective analogue of 'next' version if start == 0. This patch replaces 'next' with 'first' where things look trivial. Signed-off-by: Yury Norov <yury.norov@gmail.com> Tested-by: Wolfram Sang <wsa+renesas@sang-engineering.com>
This commit is contained in:
parent
93ba139ba8
commit
b5c7e7ec7d
@ -375,7 +375,7 @@ int pasemi_dma_alloc_flag(void)
|
||||
int bit;
|
||||
|
||||
retry:
|
||||
bit = find_next_bit(flags_free, MAX_FLAGS, 0);
|
||||
bit = find_first_bit(flags_free, MAX_FLAGS);
|
||||
if (bit >= MAX_FLAGS)
|
||||
return -ENOSPC;
|
||||
if (!test_and_clear_bit(bit, flags_free))
|
||||
@ -440,7 +440,7 @@ int pasemi_dma_alloc_fun(void)
|
||||
int bit;
|
||||
|
||||
retry:
|
||||
bit = find_next_bit(fun_free, MAX_FLAGS, 0);
|
||||
bit = find_first_bit(fun_free, MAX_FLAGS);
|
||||
if (bit >= MAX_FLAGS)
|
||||
return -ENOSPC;
|
||||
if (!test_and_clear_bit(bit, fun_free))
|
||||
|
@ -2021,7 +2021,7 @@ static unsigned long kvm_s390_next_dirty_cmma(struct kvm_memslots *slots,
|
||||
while ((slotidx > 0) && (ofs >= ms->npages)) {
|
||||
slotidx--;
|
||||
ms = slots->memslots + slotidx;
|
||||
ofs = find_next_bit(kvm_second_dirty_bitmap(ms), ms->npages, 0);
|
||||
ofs = find_first_bit(kvm_second_dirty_bitmap(ms), ms->npages);
|
||||
}
|
||||
return ms->base_gfn + ofs;
|
||||
}
|
||||
|
@ -196,7 +196,7 @@ rnbd_get_cpu_qlist(struct rnbd_clt_session *sess, int cpu)
|
||||
return per_cpu_ptr(sess->cpu_queues, bit);
|
||||
} else if (cpu != 0) {
|
||||
/* Search from 0 to cpu */
|
||||
bit = find_next_bit(sess->cpu_queues_bm, cpu, 0);
|
||||
bit = find_first_bit(sess->cpu_queues_bm, cpu);
|
||||
if (bit < cpu)
|
||||
return per_cpu_ptr(sess->cpu_queues, bit);
|
||||
}
|
||||
|
@ -1681,7 +1681,7 @@ static irqreturn_t dma_ccerr_handler(int irq, void *data)
|
||||
|
||||
dev_dbg(ecc->dev, "EMR%d 0x%08x\n", j, val);
|
||||
emr = val;
|
||||
for (i = find_next_bit(&emr, 32, 0); i < 32;
|
||||
for (i = find_first_bit(&emr, 32); i < 32;
|
||||
i = find_next_bit(&emr, 32, i + 1)) {
|
||||
int k = (j << 5) + i;
|
||||
|
||||
|
@ -347,7 +347,7 @@ static int ad7124_find_free_config_slot(struct ad7124_state *st)
|
||||
{
|
||||
unsigned int free_cfg_slot;
|
||||
|
||||
free_cfg_slot = find_next_zero_bit(&st->cfg_slots_status, AD7124_MAX_CONFIGS, 0);
|
||||
free_cfg_slot = find_first_zero_bit(&st->cfg_slots_status, AD7124_MAX_CONFIGS);
|
||||
if (free_cfg_slot == AD7124_MAX_CONFIGS)
|
||||
return -1;
|
||||
|
||||
|
@ -1709,14 +1709,14 @@ static enum irdma_status_code irdma_setup_init_state(struct irdma_pci_f *rf)
|
||||
*/
|
||||
static void irdma_get_used_rsrc(struct irdma_device *iwdev)
|
||||
{
|
||||
iwdev->rf->used_pds = find_next_zero_bit(iwdev->rf->allocated_pds,
|
||||
iwdev->rf->max_pd, 0);
|
||||
iwdev->rf->used_qps = find_next_zero_bit(iwdev->rf->allocated_qps,
|
||||
iwdev->rf->max_qp, 0);
|
||||
iwdev->rf->used_cqs = find_next_zero_bit(iwdev->rf->allocated_cqs,
|
||||
iwdev->rf->max_cq, 0);
|
||||
iwdev->rf->used_mrs = find_next_zero_bit(iwdev->rf->allocated_mrs,
|
||||
iwdev->rf->max_mr, 0);
|
||||
iwdev->rf->used_pds = find_first_zero_bit(iwdev->rf->allocated_pds,
|
||||
iwdev->rf->max_pd);
|
||||
iwdev->rf->used_qps = find_first_zero_bit(iwdev->rf->allocated_qps,
|
||||
iwdev->rf->max_qp);
|
||||
iwdev->rf->used_cqs = find_first_zero_bit(iwdev->rf->allocated_cqs,
|
||||
iwdev->rf->max_cq);
|
||||
iwdev->rf->used_mrs = find_first_zero_bit(iwdev->rf->allocated_mrs,
|
||||
iwdev->rf->max_mr);
|
||||
}
|
||||
|
||||
void irdma_ctrl_deinit_hw(struct irdma_pci_f *rf)
|
||||
|
@ -106,7 +106,7 @@ static int __must_check cec_devnode_register(struct cec_devnode *devnode,
|
||||
|
||||
/* Part 1: Find a free minor number */
|
||||
mutex_lock(&cec_devnode_lock);
|
||||
minor = find_next_zero_bit(cec_devnode_nums, CEC_NUM_DEVICES, 0);
|
||||
minor = find_first_zero_bit(cec_devnode_nums, CEC_NUM_DEVICES);
|
||||
if (minor == CEC_NUM_DEVICES) {
|
||||
mutex_unlock(&cec_devnode_lock);
|
||||
pr_err("could not get a free minor\n");
|
||||
|
@ -217,7 +217,7 @@ int __must_check media_devnode_register(struct media_device *mdev,
|
||||
|
||||
/* Part 1: Find a free minor number */
|
||||
mutex_lock(&media_devnode_lock);
|
||||
minor = find_next_zero_bit(media_devnode_nums, MEDIA_NUM_DEVICES, 0);
|
||||
minor = find_first_zero_bit(media_devnode_nums, MEDIA_NUM_DEVICES);
|
||||
if (minor == MEDIA_NUM_DEVICES) {
|
||||
mutex_unlock(&media_devnode_lock);
|
||||
pr_err("could not get a free minor\n");
|
||||
|
@ -213,7 +213,7 @@ static int dra7xx_pcie_handle_msi(struct pcie_port *pp, int index)
|
||||
if (!val)
|
||||
return 0;
|
||||
|
||||
pos = find_next_bit(&val, MAX_MSI_IRQS_PER_CTRL, 0);
|
||||
pos = find_first_bit(&val, MAX_MSI_IRQS_PER_CTRL);
|
||||
while (pos != MAX_MSI_IRQS_PER_CTRL) {
|
||||
generic_handle_domain_irq(pp->irq_domain,
|
||||
(index * MAX_MSI_IRQS_PER_CTRL) + pos);
|
||||
|
@ -17990,8 +17990,8 @@ lpfc_sli4_alloc_xri(struct lpfc_hba *phba)
|
||||
* the driver starts at 0 each time.
|
||||
*/
|
||||
spin_lock_irq(&phba->hbalock);
|
||||
xri = find_next_zero_bit(phba->sli4_hba.xri_bmask,
|
||||
phba->sli4_hba.max_cfg_param.max_xri, 0);
|
||||
xri = find_first_zero_bit(phba->sli4_hba.xri_bmask,
|
||||
phba->sli4_hba.max_cfg_param.max_xri);
|
||||
if (xri >= phba->sli4_hba.max_cfg_param.max_xri) {
|
||||
spin_unlock_irq(&phba->hbalock);
|
||||
return NO_XRI;
|
||||
@ -19668,7 +19668,7 @@ lpfc_sli4_alloc_rpi(struct lpfc_hba *phba)
|
||||
max_rpi = phba->sli4_hba.max_cfg_param.max_rpi;
|
||||
rpi_limit = phba->sli4_hba.next_rpi;
|
||||
|
||||
rpi = find_next_zero_bit(phba->sli4_hba.rpi_bmask, rpi_limit, 0);
|
||||
rpi = find_first_zero_bit(phba->sli4_hba.rpi_bmask, rpi_limit);
|
||||
if (rpi >= rpi_limit)
|
||||
rpi = LPFC_RPI_ALLOC_ERROR;
|
||||
else {
|
||||
@ -20311,8 +20311,8 @@ lpfc_sli4_fcf_rr_next_index_get(struct lpfc_hba *phba)
|
||||
* have been tested so that we can detect when we should
|
||||
* change the priority level.
|
||||
*/
|
||||
next_fcf_index = find_next_bit(phba->fcf.fcf_rr_bmask,
|
||||
LPFC_SLI4_FCF_TBL_INDX_MAX, 0);
|
||||
next_fcf_index = find_first_bit(phba->fcf.fcf_rr_bmask,
|
||||
LPFC_SLI4_FCF_TBL_INDX_MAX);
|
||||
}
|
||||
|
||||
|
||||
|
@ -358,8 +358,8 @@ struct k3_ring *k3_ringacc_request_ring(struct k3_ringacc *ringacc,
|
||||
goto out;
|
||||
|
||||
if (flags & K3_RINGACC_RING_USE_PROXY) {
|
||||
proxy_id = find_next_zero_bit(ringacc->proxy_inuse,
|
||||
ringacc->num_proxies, 0);
|
||||
proxy_id = find_first_zero_bit(ringacc->proxy_inuse,
|
||||
ringacc->num_proxies);
|
||||
if (proxy_id == ringacc->num_proxies)
|
||||
goto error;
|
||||
}
|
||||
|
@ -1975,7 +1975,7 @@ static bool canon_copy_from_read_buf(struct tty_struct *tty,
|
||||
more = n - (size - tail);
|
||||
if (eol == N_TTY_BUF_SIZE && more) {
|
||||
/* scan wrapped without finding set bit */
|
||||
eol = find_next_bit(ldata->read_flags, more, 0);
|
||||
eol = find_first_bit(ldata->read_flags, more);
|
||||
found = eol != more;
|
||||
} else
|
||||
found = eol != size;
|
||||
|
@ -246,8 +246,7 @@ void acrn_ioreq_request_clear(struct acrn_vm *vm)
|
||||
spin_lock_bh(&vm->ioreq_clients_lock);
|
||||
client = vm->default_client;
|
||||
if (client) {
|
||||
vcpu = find_next_bit(client->ioreqs_map,
|
||||
ACRN_IO_REQUEST_MAX, 0);
|
||||
vcpu = find_first_bit(client->ioreqs_map, ACRN_IO_REQUEST_MAX);
|
||||
while (vcpu < ACRN_IO_REQUEST_MAX) {
|
||||
acrn_ioreq_complete_request(client, vcpu, NULL);
|
||||
vcpu = find_next_bit(client->ioreqs_map,
|
||||
|
@ -2558,8 +2558,8 @@ static void get_new_segment(struct f2fs_sb_info *sbi,
|
||||
secno = find_next_zero_bit(free_i->free_secmap, MAIN_SECS(sbi), hint);
|
||||
if (secno >= MAIN_SECS(sbi)) {
|
||||
if (dir == ALLOC_RIGHT) {
|
||||
secno = find_next_zero_bit(free_i->free_secmap,
|
||||
MAIN_SECS(sbi), 0);
|
||||
secno = find_first_zero_bit(free_i->free_secmap,
|
||||
MAIN_SECS(sbi));
|
||||
f2fs_bug_on(sbi, secno >= MAIN_SECS(sbi));
|
||||
} else {
|
||||
go_left = 1;
|
||||
@ -2574,8 +2574,8 @@ static void get_new_segment(struct f2fs_sb_info *sbi,
|
||||
left_start--;
|
||||
continue;
|
||||
}
|
||||
left_start = find_next_zero_bit(free_i->free_secmap,
|
||||
MAIN_SECS(sbi), 0);
|
||||
left_start = find_first_zero_bit(free_i->free_secmap,
|
||||
MAIN_SECS(sbi));
|
||||
f2fs_bug_on(sbi, left_start >= MAIN_SECS(sbi));
|
||||
break;
|
||||
}
|
||||
|
@ -379,7 +379,7 @@ static void o2hb_nego_timeout(struct work_struct *work)
|
||||
|
||||
o2hb_fill_node_map(live_node_bitmap, sizeof(live_node_bitmap));
|
||||
/* lowest node as master node to make negotiate decision. */
|
||||
master_node = find_next_bit(live_node_bitmap, O2NM_MAX_NODES, 0);
|
||||
master_node = find_first_bit(live_node_bitmap, O2NM_MAX_NODES);
|
||||
|
||||
if (master_node == o2nm_this_node()) {
|
||||
if (!test_bit(master_node, reg->hr_nego_node_bitmap)) {
|
||||
|
@ -1045,7 +1045,7 @@ static int dlm_send_regions(struct dlm_ctxt *dlm, unsigned long *node_map)
|
||||
int status, ret = 0, i;
|
||||
char *p;
|
||||
|
||||
if (find_next_bit(node_map, O2NM_MAX_NODES, 0) >= O2NM_MAX_NODES)
|
||||
if (find_first_bit(node_map, O2NM_MAX_NODES) >= O2NM_MAX_NODES)
|
||||
goto bail;
|
||||
|
||||
qr = kzalloc(sizeof(struct dlm_query_region), GFP_KERNEL);
|
||||
@ -1217,7 +1217,7 @@ static int dlm_send_nodeinfo(struct dlm_ctxt *dlm, unsigned long *node_map)
|
||||
struct o2nm_node *node;
|
||||
int ret = 0, status, count, i;
|
||||
|
||||
if (find_next_bit(node_map, O2NM_MAX_NODES, 0) >= O2NM_MAX_NODES)
|
||||
if (find_first_bit(node_map, O2NM_MAX_NODES) >= O2NM_MAX_NODES)
|
||||
goto bail;
|
||||
|
||||
qn = kzalloc(sizeof(struct dlm_query_nodeinfo), GFP_KERNEL);
|
||||
|
@ -861,7 +861,7 @@ struct dlm_lock_resource * dlm_get_lock_resource(struct dlm_ctxt *dlm,
|
||||
* to see if there are any nodes that still need to be
|
||||
* considered. these will not appear in the mle nodemap
|
||||
* but they might own this lockres. wait on them. */
|
||||
bit = find_next_bit(dlm->recovery_map, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(dlm->recovery_map, O2NM_MAX_NODES);
|
||||
if (bit < O2NM_MAX_NODES) {
|
||||
mlog(0, "%s: res %.*s, At least one node (%d) "
|
||||
"to recover before lock mastery can begin\n",
|
||||
@ -912,7 +912,7 @@ struct dlm_lock_resource * dlm_get_lock_resource(struct dlm_ctxt *dlm,
|
||||
dlm_wait_for_recovery(dlm);
|
||||
|
||||
spin_lock(&dlm->spinlock);
|
||||
bit = find_next_bit(dlm->recovery_map, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(dlm->recovery_map, O2NM_MAX_NODES);
|
||||
if (bit < O2NM_MAX_NODES) {
|
||||
mlog(0, "%s: res %.*s, At least one node (%d) "
|
||||
"to recover before lock mastery can begin\n",
|
||||
@ -1079,7 +1079,7 @@ static int dlm_wait_for_lock_mastery(struct dlm_ctxt *dlm,
|
||||
sleep = 1;
|
||||
/* have all nodes responded? */
|
||||
if (voting_done && !*blocked) {
|
||||
bit = find_next_bit(mle->maybe_map, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(mle->maybe_map, O2NM_MAX_NODES);
|
||||
if (dlm->node_num <= bit) {
|
||||
/* my node number is lowest.
|
||||
* now tell other nodes that I am
|
||||
@ -1234,8 +1234,8 @@ static int dlm_restart_lock_mastery(struct dlm_ctxt *dlm,
|
||||
} else {
|
||||
mlog(ML_ERROR, "node down! %d\n", node);
|
||||
if (blocked) {
|
||||
int lowest = find_next_bit(mle->maybe_map,
|
||||
O2NM_MAX_NODES, 0);
|
||||
int lowest = find_first_bit(mle->maybe_map,
|
||||
O2NM_MAX_NODES);
|
||||
|
||||
/* act like it was never there */
|
||||
clear_bit(node, mle->maybe_map);
|
||||
@ -1795,7 +1795,7 @@ int dlm_assert_master_handler(struct o2net_msg *msg, u32 len, void *data,
|
||||
"MLE for it! (%.*s)\n", assert->node_idx,
|
||||
namelen, name);
|
||||
} else {
|
||||
int bit = find_next_bit (mle->maybe_map, O2NM_MAX_NODES, 0);
|
||||
int bit = find_first_bit(mle->maybe_map, O2NM_MAX_NODES);
|
||||
if (bit >= O2NM_MAX_NODES) {
|
||||
/* not necessarily an error, though less likely.
|
||||
* could be master just re-asserting. */
|
||||
@ -2521,7 +2521,7 @@ static int dlm_is_lockres_migratable(struct dlm_ctxt *dlm,
|
||||
}
|
||||
|
||||
if (!nonlocal) {
|
||||
node_ref = find_next_bit(res->refmap, O2NM_MAX_NODES, 0);
|
||||
node_ref = find_first_bit(res->refmap, O2NM_MAX_NODES);
|
||||
if (node_ref >= O2NM_MAX_NODES)
|
||||
return 0;
|
||||
}
|
||||
@ -3303,7 +3303,7 @@ static void dlm_clean_block_mle(struct dlm_ctxt *dlm,
|
||||
BUG_ON(mle->type != DLM_MLE_BLOCK);
|
||||
|
||||
spin_lock(&mle->spinlock);
|
||||
bit = find_next_bit(mle->maybe_map, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(mle->maybe_map, O2NM_MAX_NODES);
|
||||
if (bit != dead_node) {
|
||||
mlog(0, "mle found, but dead node %u would not have been "
|
||||
"master\n", dead_node);
|
||||
@ -3542,7 +3542,7 @@ void dlm_force_free_mles(struct dlm_ctxt *dlm)
|
||||
spin_lock(&dlm->master_lock);
|
||||
|
||||
BUG_ON(dlm->dlm_state != DLM_CTXT_LEAVING);
|
||||
BUG_ON((find_next_bit(dlm->domain_map, O2NM_MAX_NODES, 0) < O2NM_MAX_NODES));
|
||||
BUG_ON((find_first_bit(dlm->domain_map, O2NM_MAX_NODES) < O2NM_MAX_NODES));
|
||||
|
||||
for (i = 0; i < DLM_HASH_BUCKETS; i++) {
|
||||
bucket = dlm_master_hash(dlm, i);
|
||||
|
@ -451,7 +451,7 @@ static int dlm_do_recovery(struct dlm_ctxt *dlm)
|
||||
if (dlm->reco.dead_node == O2NM_INVALID_NODE_NUM) {
|
||||
int bit;
|
||||
|
||||
bit = find_next_bit (dlm->recovery_map, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(dlm->recovery_map, O2NM_MAX_NODES);
|
||||
if (bit >= O2NM_MAX_NODES || bit < 0)
|
||||
dlm_set_reco_dead_node(dlm, O2NM_INVALID_NODE_NUM);
|
||||
else
|
||||
|
@ -92,7 +92,7 @@ int __dlm_lockres_unused(struct dlm_lock_resource *res)
|
||||
return 0;
|
||||
|
||||
/* Another node has this resource with this node as the master */
|
||||
bit = find_next_bit(res->refmap, O2NM_MAX_NODES, 0);
|
||||
bit = find_first_bit(res->refmap, O2NM_MAX_NODES);
|
||||
if (bit < O2NM_MAX_NODES)
|
||||
return 0;
|
||||
|
||||
|
@ -251,7 +251,7 @@ void gen_pool_destroy(struct gen_pool *pool)
|
||||
list_del(&chunk->next_chunk);
|
||||
|
||||
end_bit = chunk_size(chunk) >> order;
|
||||
bit = find_next_bit(chunk->bits, end_bit, 0);
|
||||
bit = find_first_bit(chunk->bits, end_bit);
|
||||
BUG_ON(bit < end_bit);
|
||||
|
||||
vfree(chunk);
|
||||
|
@ -608,7 +608,7 @@ static int clear_one_vid(struct ncsi_dev_priv *ndp, struct ncsi_channel *nc,
|
||||
bitmap = &ncf->bitmap;
|
||||
|
||||
spin_lock_irqsave(&nc->lock, flags);
|
||||
index = find_next_bit(bitmap, ncf->n_vids, 0);
|
||||
index = find_first_bit(bitmap, ncf->n_vids);
|
||||
if (index >= ncf->n_vids) {
|
||||
spin_unlock_irqrestore(&nc->lock, flags);
|
||||
return -1;
|
||||
@ -667,7 +667,7 @@ static int set_one_vid(struct ncsi_dev_priv *ndp, struct ncsi_channel *nc,
|
||||
return -1;
|
||||
}
|
||||
|
||||
index = find_next_zero_bit(bitmap, ncf->n_vids, 0);
|
||||
index = find_first_zero_bit(bitmap, ncf->n_vids);
|
||||
if (index < 0 || index >= ncf->n_vids) {
|
||||
netdev_err(ndp->ndev.dev,
|
||||
"Channel %u already has all VLAN filters set\n",
|
||||
|
Loading…
Reference in New Issue
Block a user