Recently, I noticed some people mentioning that std::list::size()
has a linear complexity.
According to some sources, this is in fact implementation dependent as the standard doesn't say what the complexity has to be.
The comment in this blog entry says:
Actually, it depends on which STL you are using. Microsoft Visual Studio V6 implements size() as {return (_Size); } whereas gcc (at least in versions 3.3.2 and 4.1.0) do it as { return std::distance(begin(), end()); } The first has constant speed, the second has o(N) speed
size()
has constant complexity as Dinkumware
probably won't have changed that fact since VC6. Am I right there?gcc
? If it is really O(n), why did the
developers choose to do so?You are correct that the standard does not state what the complexity of list::size()
must be - however, it does recommend that it "should have constant complexity" (Note A in Table 65).
Here's an interesting article by Howard Hinnant that explains why some people think list::size()
should have O(N) complexity (basically because they believe that O(1) list::size()
makes list::splice()
have O(N) complexity) and why an O(1) list::size()
is be a good idea (in the author's opinion):
I think the main points in the paper are:
list::size()
can be O(1) causes the splice operation to become linear size()
(such as his one example where list::size()
is called while holding a lock).size()
be O(N), in the interest of 'least surprise', the standard should require any container that implements size()
to implement it in an O(1) fashion. If a container cannot do this, it should not implement size()
at all. In this case, the user of the container will be made aware that size()
is unavailable, and if they still want or need to get the number of elements in the container they can still use container::distance( begin(), end())
to get that value - but they will be completely aware that it's an O(N) operation.I think I tend to agree with most of his reasoning. However, I do not like his proposed addition to the splice()
overloads. Having to pass in an n
that must be equal to distance( first, last)
to get correct behavior seems like a recipe for hard to diagnose bugs.
I'm not sure what should or could be done moving forward, as any change would have a significant impact on existing code. But as it stands, I think that existing code is already impacted - behavior might be rather significantly different from one implementation to another for something that should have been well-defined. Maybe onebyone's comment about having the size 'cached' and marked known/unknown might work well - you get amortized O(1) behavior - the only time you get O(N) behavior is when the list is modified by some splice() operations. The nice thing about this is that it can be done by implementors today without a change to the standard (unless I'm missing something).
As far as I know, C++0x is not changing anything in this area.