index() will give the first occurrence of an item in a list. Is there a neat trick which returns all indices in a list for an element?
22 Answers
You can use a list comprehension with enumerate:
indices = [i for i, x in enumerate(my_list) if x == "whatever"]
The iterator enumerate(my_list) yields pairs (index, item) for each item in the list. Using i, x as loop variable target unpacks these pairs into the index i and the list item x. We filter down to all x that match our criterion, and select the indices i of these elements.
Comments
While not a solution for lists directly, numpy really shines for this sort of thing:
import numpy as np
values = np.array([1,2,3,1,2,4,5,6,3,2,1])
searchval = 3
ii = np.where(values == searchval)[0]
returns:
ii ==>array([2, 8])
This can be significantly faster for lists (arrays) with a large number of elements vs some of the other solutions.
4 Comments
values could be a NumPy array or a Python list.np.where([7, 8, 9, 8] == 8)[0] and np.where(np.array([7, 8, 9, 8]) == 8)[0]; only the latter works as intended.np.where isn't really recommended anymore (though not deprecated). Instead, use np.nonzero: ii = np.nonzero(values == searchval)[0]A solution using list.index:
def indices(lst, element):
result = []
offset = -1
while True:
try:
offset = lst.index(element, offset+1)
except ValueError:
return result
result.append(offset)
It's much faster than the list comprehension with enumerate, for large lists. It is also much slower than the numpy solution if you already have the array, otherwise the cost of converting outweighs the speed gain (tested on integer lists with 100, 1000 and 10000 elements).
NOTE: A note of caution based on Chris_Rands' comment: this solution is faster than the list comprehension if the results are sufficiently sparse, but if the list has many instances of the element that is being searched (more than ~15% of the list, on a test with a list of 1000 integers), the list comprehension is faster.
2 Comments
timeit.timeit with randomly generated lists. That's an important point though, and I suppose that may be why you ask. At the time it didn't occur to me, but the speed gains are only true if the results are sufficiently sparse. I just tested with a list full of the element to search for, and it's much slower than the list comprehension.more_itertools.locate finds indices for all items that satisfy a condition.
from more_itertools import locate
list(locate([0, 1, 1, 0, 1, 0, 0]))
# [1, 2, 4]
list(locate(['a', 'b', 'c', 'b'], lambda x: x == 'b'))
# [1, 3]
more_itertools is a third-party library > pip install more_itertools.
Comments
- The
while-loopin this answer is the fastest implementation tested.- It is 26% faster than the accepted answer,
test2()below.
- It is 26% faster than the accepted answer,
- There’s an answer using
np.whereto find the indices of a single value, which is not faster than a list-comprehension, if the time to convert a list to an array is included - The overhead of importing
numpyand converting alistto anumpy.arrayprobably makes usingnumpya less efficient option for most circumstances. A careful timing analysis would be necessary.- In cases where multiple functions/operations will need to be performed on the
list, converting thelistto anarray, and then usingnumpyfunctions will likely be a faster option.
- In cases where multiple functions/operations will need to be performed on the
- This solution uses
np.whereandnp.uniqueto find the indices of all unique elements in a list.- Using
np.whereon an array (including the time to convert the list to an array) is slightly slower than a list-comprehension on a list, for finding all indices of all unique elements. - This has been tested on an 2M element list with 4 unique values, and the size of the list/array and number of unique elements will have an impact.
- Using
- Other solutions using
numpyon an array can be found in Get a list of all indices of repeated elements in a numpy array - Tested in
[python 3.10.4, numpy 1.23.1]and[python 3.11.0, numpy 1.23.4]
import numpy as np
import random # to create test list
# create sample list
random.seed(365)
l = [random.choice(['s1', 's2', 's3', 's4']) for _ in range(20)]
# convert the list to an array for use with these numpy methods
a = np.array(l)
# create a dict of each unique entry and the associated indices
idx = {v: np.where(a == v)[0].tolist() for v in np.unique(a)}
# print(idx)
{'s1': [7, 9, 10, 11, 17],
's2': [1, 3, 6, 8, 14, 18, 19],
's3': [0, 2, 13, 16],
's4': [4, 5, 12, 15]}
%timeit on a 2M element list with 4 unique str elements
# create 2M element list
random.seed(365)
l = [random.choice(['s1', 's2', 's3', 's4']) for _ in range(2000000)]
Functions
def test1():
# np.where: convert list to array and find indices of a single element
a = np.array(l)
return np.where(a == 's1')
def test2():
# list-comprehension: on list l and find indices of a single element
return [i for i, x in enumerate(l) if x == "s1"]
def test3():
# filter: on list l and find indices of a single element
return list(filter(lambda i: l[i]=="s1", range(len(l))))
def test4():
# use np.where and np.unique to find indices of all unique elements: convert list to array
a = np.array(l)
return {v: np.where(a == v)[0].tolist() for v in np.unique(a)}
def test5():
# list comprehension inside dict comprehension: on list l and find indices of all unique elements
return {req_word: [idx for idx, word in enumerate(l) if word == req_word] for req_word in set(l)}
def get_indices1(x: list, value: int) -> list:
indices = list()
for i in range(len(x)):
if x[i] == value:
indices.append(i)
return indices
def get_indices2(x: list, value: int) -> list:
indices = list()
i = 0
while True:
try:
# find an occurrence of value and update i to that index
i = x.index(value, i)
# add i to the list
indices.append(i)
# advance i by 1
i += 1
except ValueError as e:
break
return indices
Function Call
%timeit test1() # list of indices for specified value
%timeit test2() # list of indices for specified value
%timeit test3() # list of indices for specified value
%timeit test4() # dict of indices of all values
%timeit test5() # dict of indices of all values
%timeit get_indices1(l, 's1') # list of indices for specified value
%timeit get_indices2(l, 's1') # list of indices for specified value
Result python 3.12.0
209 ms ± 2.93 ms per loop (mean ± std. dev. of 7 runs, 1 loop each)
78.5 ms ± 733 µs per loop (mean ± std. dev. of 7 runs, 10 loops each)
125 ms ± 757 µs per loop (mean ± std. dev. of 7 runs, 10 loops each)
340 ms ± 8.16 ms per loop (mean ± std. dev. of 7 runs, 1 loop each)
319 ms ± 2.97 ms per loop (mean ± std. dev. of 7 runs, 1 loop each)
74.9 ms ± 1.99 ms per loop (mean ± std. dev. of 7 runs, 10 loops each)
58.2 ms ± 1.03 ms per loop (mean ± std. dev. of 7 runs, 10 loops each)
Comments
Getting all the occurrences and the position of one or more (identical) items in a list
With enumerate(alist) you can store the first element (n) that is the index of the list when the element x is equal to what you look for.
>>> alist = ['foo', 'spam', 'egg', 'foo']
>>> foo_indexes = [n for n,x in enumerate(alist) if x=='foo']
>>> foo_indexes
[0, 3]
>>>
Let's make our function findindex
This function takes the item and the list as arguments and return the position of the item in the list, like we saw before.
def indexlist(item2find, list_or_string):
"Returns all indexes of an item in a list or a string"
return [n for n,item in enumerate(list_or_string) if item==item2find]
print(indexlist("1", "010101010"))
Output
[1, 3, 5, 7]
Simple
for n, i in enumerate([1, 2, 3, 4, 1]):
if i == 1:
print(n)
Output:
0
4
Comments
A dynamic list comprehension based solution incase we do not know in advance which element:
lst = ['to', 'be', 'or', 'not', 'to', 'be']
{req_word: [idx for idx, word in enumerate(lst) if word == req_word] for req_word in set(lst)}
results in:
{'be': [1, 5], 'or': [2], 'to': [0, 4], 'not': [3]}
You can think of all other ways along the same lines as well but with index() you can find only one index although you can set occurrence number yourself.
Comments
Using a for-loop:
- Answers with
enumerateand a list comprehension are more pythonic, but not necessarily faster. However, this answer is aimed at students who may not be allowed to use some of those built-in functions. - create an empty list,
indices - create the loop with
for i in range(len(x)):, which essentially iterates through a list of index locations[0, 1, 2, 3, ..., len(x)-1] - in the loop, add any
i, wherex[i]is a match tovalue, toindices
def get_indices(x: list, value: int) -> list:
indices = list()
for i in range(len(x)):
if x[i] == value:
indices.append(i)
return indices
n = [1, 2, 3, -50, -60, 0, 6, 9, -60, -60]
print(get_indices(n, -60))
>>> [4, 8, 9]
- The functions,
get_indices, are implemented with type hints. In this case, the list,n, is a bunch ofints, therefore we search forvalue, also defined as anint.
Using a while-loop and .index:
- This is the fastest implementation tested in this answer.
- It is 26% faster than the accepted answer.
- With
.index, usetry-exceptfor error handling, because aValueErrorwill occur ifvalueis not in thelist.
def get_indices(x: list, value: int) -> list:
indices = list()
i = 0
while True:
try:
# find an occurrence of value and update i to that index
i = x.index(value, i)
# add i to the list
indices.append(i)
# advance i by 1
i += 1
except ValueError as e:
break
return indices
print(get_indices(n, -60))
>>> [4, 8, 9]
1 Comment
get_indeices is a bit faster(~15%) than normal list comprehension. I am trying to figure it out.Here are different methods to find all occurrences of "bar" in a list, I have ranked by execution speed and usecase:
lst = ["foo", "bar", "baz", "bar", "qux", "bar"]
item = "bar"
itertools.compress (Fastest: ~0.032s) Memory Efficient
from itertools import compress
indices = list(compress(range(len(lst)), [val == item for val in lst]))
List Comprehension (~0.020s) readability & simplicity
indices = [i for i, v in enumerate(lst) if v == item]
collections.deque (~0.063s) Faster Append
from collections import deque
indices = deque()
for i, val in enumerate(lst):
if val == item:
indices.append(i)
indices = list(indices)
filter + lambda (~0.042s) Functional Approach
indices = list(filter(lambda i: lst[i] == item, range(len(lst))))
NumPy (np.where) (~0.194s) Faster for Large Lists
import numpy as np
indices = np.where(np.array(lst) == item)[0].tolist()
Pandas (pd.Series.index) (~9.874s) Best for DataFrames & Large Datasets
import pandas as pd
indices = pd.Series(lst).index[pd.Series(lst) == item].tolist()
Output:
[1, 3, 5]
Comments
If you are using Python 2, you can achieve the same functionality with this:
def f(my_list, value):
return filter(lambda x: my_list[x] == value, range(len(my_list)))
Where my_list is the list you want to get the indexes of, and value is the value searched. Usage:
f(some_list, some_element)
Comments
Create a generator
Generators are fast and use a tiny memory footprint. They give you flexibility in how you use the result.
def indices(iter, val):
"""Generator: Returns all indices of val in iter
Raises a ValueError if no val does not occur in iter
Passes on the AttributeError if iter does not have an index method (e.g. is a set)
"""
i = -1
NotFound = False
while not NotFound:
try:
i = iter.index(val, i+1)
except ValueError:
NotFound = True
else:
yield i
if i == -1:
raise ValueError("No occurrences of {v} in {i}".format(v = val, i = iter))
The above code can be use to create a list of the indices: list(indices(input,value)); use them as dictionary keys: dict(indices(input,value)); sum them: sum(indices(input,value)); in a for loop for index_ in indices(input,value):; ...etc... without creating an interim list/tuple or similar.
In a for loop you will get your next index back when you call for it, without waiting for all the others to be calculated first. That means: if you break out of the loop for some reason you save the time needed to find indices you never needed.
How it works
- Call
.indexon the inputiterto find the next occurrence ofval - Use the second parameter to
.indexto start at the point after the last found occurrence - Yield the index
- Repeat until
indexraises aValueError
Alternative versions
I tried four different versions for flow control; two EAFP (using try - except) and two TBYL (with a logical test in the while statement):
- "WhileTrueBreak":
while True:...except ValueError: break. Surprisingly, this was usually a touch slower than option 2 and (IMV) less readable - "WhileErrFalse": Using a bool variable
errto identify when aValueErroris raised. This is generally the fastest and more readable than 1 - "RemainingSlice": Check whether val is in the remaining part of the input using slicing:
while val in iter[i:]. Unsurprisingly, this does not scale well - "LastOccurrence": Check first where the last occurrence is, keep going
while i < last
The overall performance differences between 1,2 and 4 are negligible, so it comes down to personal style and preference. Given that .index uses ValueError to let you know it didn't find anything, rather than e.g. returning None, an EAFP-approach seems fitting to me.
Here are the 4 code variants and results from timeit (in milliseconds) for different lengths of input and sparsity of matches
@version("WhileTrueBreak", versions)
def indices2(iter, val):
i = -1
while True:
try:
i = iter.index(val, i+1)
except ValueError:
break
else:
yield i
@version("WhileErrFalse", versions)
def indices5(iter, val):
i = -1
err = False
while not err:
try:
i = iter.index(val, i+1)
except ValueError:
err = True
else:
yield i
@version("RemainingSlice", versions)
def indices1(iter, val):
i = 0
while val in iter[i:]:
i = iter.index(val, i)
yield i
i += 1
@version("LastOccurrence", versions)
def indices4(iter,val):
i = 0
last = len(iter) - tuple(reversed(iter)).index(val)
while i < last:
i = iter.index(val, i)
yield i
i += 1
Length: 100, Ocurrences: 4.0%
{'WhileTrueBreak': 0.0074799987487494946, 'WhileErrFalse': 0.006440002471208572, 'RemainingSlice': 0.01221001148223877, 'LastOccurrence': 0.00801000278443098}
Length: 1000, Ocurrences: 1.2%
{'WhileTrueBreak': 0.03101000329479575, 'WhileErrFalse': 0.0278000021353364, 'RemainingSlice': 0.08278000168502331, 'LastOccurrence': 0.03986000083386898}
Length: 10000, Ocurrences: 2.05%
{'WhileTrueBreak': 0.18062000162899494, 'WhileErrFalse': 0.1810499932616949, 'RemainingSlice': 2.9145700042136014, 'LastOccurrence': 0.2049500006251037}
Length: 100000, Ocurrences: 1.977%
{'WhileTrueBreak': 1.9361200043931603, 'WhileErrFalse': 1.7280600033700466, 'RemainingSlice': 254.4725100044161, 'LastOccurrence': 1.9101499929092824}
Length: 100000, Ocurrences: 9.873%
{'WhileTrueBreak': 2.832529996521771, 'WhileErrFalse': 2.9984100023284554, 'RemainingSlice': 1132.4922299943864, 'LastOccurrence': 2.6660699979402125}
Length: 100000, Ocurrences: 25.058%
{'WhileTrueBreak': 5.119729996658862, 'WhileErrFalse': 5.2082200068980455, 'RemainingSlice': 2443.0577100021765, 'LastOccurrence': 4.75954000139609}
Length: 100000, Ocurrences: 49.698%
{'WhileTrueBreak': 9.372120001353323, 'WhileErrFalse': 8.447749994229525, 'RemainingSlice': 5042.717969999649, 'LastOccurrence': 8.050809998530895}
Comments
A very old question but I don't see this wrinkle in the answers, altough it is alluded to.
Using the accepted answer's enumerate but with the in option, allows a string search of a list, with the added bonus of partial hits and case insensitivity, should you so choose.
Providing either index positions or the found string itself.
Given a list of iso country names, the following gives an idea of its use
>>> query = "Tre".casefold()
>>>
>>> print([i for i, x in enumerate(country_list) if query in x.casefold()])
[280, 303, 352, 489]
>>>
>>> print([x for i, x in enumerate(country_list) if query in x.casefold()])
['Eritrea', 'Spain Extremadura', 'France Centre-Val de Loire', 'Italy Trentino-South Tyrol']
>>>
>>> country_list[280]
'Eritrea'
>>>
>>> country_list[489]
'Italy Trentino-South Tyrol'
>>> query = "out".casefold()
>>> print([i for i, x in enumerate(country_list) if query in x.casefold()])
[48, 54, 253, 421, 489, 651, 1075, 1099, 1147, 1240, 1242, 1248, 1306]
>>> print([x for i, x in enumerate(country_list) if query in x.casefold()])
['Australia New South Wales', 'Australia South Australia', 'Djibouti', 'South Georgia and Sandwich Islands', 'Italy Trentino-South Tyrol', 'South Korea', 'South Sudan', 'French Southern Territories', 'United States Outlying Islands', 'United States of America South Carolina', 'United States of America South Dakota', 'United States of America Outlying Islands', 'South Africa']
It's a simple case insensitive, partial string finder.
I suspect it's terribly inefficient on massive searches but for small ones it's easy and clean.
Comments
If you need to perform multiple lookups on a list, it may be more efficient to compile a dictionary with all of the positions of every distinct element. At that point, any single lookup is a constant-time operation.
To do this we can:
- Map an index onto each element.
enumerateprovides this quite readily. - Sort based on the element, ignoring the index. It's important to sort first so that we can...
- Use
itertools.groupbyto group those matching elements together. - Iterate over that
groupbyobject to build a dictionary.
from itertools import groupby
from operator import itemgetter
def build_table(lst):
snd = itemgetter(1)
e = enumerate(lst)
s = sorted(e, key=snd)
g = groupby(s, key=snd)
return {
k: [x[0] for x in v]
for k, v in g
}
Comments
Here is a time performance comparison between using np.where vs list_comprehension. Seems like np.where is faster on average.
# np.where
start_times = []
end_times = []
for i in range(10000):
start = time.time()
start_times.append(start)
temp_list = np.array([1,2,3,3,5])
ixs = np.where(temp_list==3)[0].tolist()
end = time.time()
end_times.append(end)
print("Took on average {} seconds".format(
np.mean(end_times)-np.mean(start_times)))
Took on average 3.81469726562e-06 seconds
# list_comprehension
start_times = []
end_times = []
for i in range(10000):
start = time.time()
start_times.append(start)
temp_list = np.array([1,2,3,3,5])
ixs = [i for i in range(len(temp_list)) if temp_list[i]==3]
end = time.time()
end_times.append(end)
print("Took on average {} seconds".format(
np.mean(end_times)-np.mean(start_times)))
Took on average 4.05311584473e-06 seconds
Comments
The simplest and far fastest I have found is to recreate as a dict that you can then loop over i.e.
l_full_list_w_duplicates=[...]
l_full_list_wo_duplicates=list(set(l_full_list_w_duplicates))
dict_location={}
for i in l_full_list_wo_duplicates:
dict_location[i]=[]
i_counter=0
for iin l_full_list_w_duplicates:
dict_location[i]+=[i_counter]
i_counter+=1
This runs in linear time and is very scalable