I'm referring to this: http://docs.python.org/tutorial/datastructures.html
What would be the running time of list.index(x) function in terms of Big O notation?
I'm referring to this: http://docs.python.org/tutorial/datastructures.html
What would be the running time of list.index(x) function in terms of Big O notation?
It's O(n), also check out: http://wiki.python.org/moin/TimeComplexity
This page documents the time-complexity (aka "Big O" or "Big Oh") of various operations in current CPython. Other Python implementations (or older or still-under development versions of CPython) may have slightly different performance characteristics. However, it is generally safe to assume that they are not slower by more than a factor of O(log n)...
list or other data structures, it is implement as linear search hence O(n).binary=True or sorted=True were an argument one could provideAccording to said documentation:
list.index(x)Return the index in the list of the first item whose value is x. It is an error if there is no such item.
Which implies searching. You're effectively doing x in s but rather than returning True or False you're returning the index of x. As such, I'd go with the listed time complexity of O(n).
Any list implementation is going to have an O(n) complexity for a linear search (e.g., list.index). Although maybe there are some wacky implementations out there that do worse...
You can improve lookup complexity by using different data structures, such as ordered lists or sets. These are usually implemented with binary trees. However, these data structures put constraints on the elements they contain. In the case of a binary tree, the elements need to be orderable, but the lookup cost goes down to O(log n).
As mentioned previously, look here for run time costs of standard Python data structures: http://wiki.python.org/moin/TimeComplexity
The documentation provided above did not cover list.index()
from my understanding, list.index is O(1) operation. Here is a link if you want to know more. https://www.ics.uci.edu/~pattis/ICS-33/lectures/complexitypython.txt
Use the following code to check the timing. Its complexity is O(n).
import time
class TimeChecker:
def __init__(self, name):
self.name = name
def __enter__(self):
self.start = self.get_time_in_sec()
return self
def __exit__(self, exc_type, exc_val, exc_tb):
now = self.get_time_in_sec()
time_taken = now - self.start # in seconds
print("Time Taken by " + self.name + ": " + str(time_taken))
def get_time_in_sec(self):
return int(round(time.time() * 1000))
def test_list_index_func(range_num):
lis = [1,2,3,4,5]
with TimeChecker('Process 1') as tim:
for i in range(range_num):
lis.index(4)
test_list_index_func(1000)
test_list_index_func(10000)
test_list_index_func(100000)
test_list_index_func(1000000)
print("Time: O(n)")
list.index operates in linear time. It does not compare how long list.index takes to run on varying input sizes, but it simply runs list.index multiple times. Even if you were computing 1+1, if you compute 1+1 a thousand times it will take 1000x longer than computing it once. To ensure this is true, I tested your code with binary search, which should be O(log n), and with accessing an element of the list, which should be O(1). Both of them, naturally, took 10x longer with each call of test_list_index_func, which is linear growth, which is incorrect.list.index never change, it's always the same list and the same search value. When you call test_list_index_func with increasing values, it isn't changing the input size to list.index, it's just changing the number of times that the exact same operation is performed. Doing that will always show a linear pattern, for any deterministic function called with the same inputs (which, again, is what you're doing here).
O(n).%%timeitsaid 2.2ns whereas fetching an attribute via an ORM (warm queryset) was 80ns.