Reputation: 13321
I have a pandas.DataFrame
with a column called name
containing strings.
I would like to get a list of the names which occur more than once in the column. How do I do that?
I tried:
funcs_groups = funcs.groupby(funcs.name)
funcs_groups[(funcs_groups.count().name>1)]
But it doesn't filter out the singleton names.
Upvotes: 20
Views: 47034
Reputation: 4467
Most of the responses given demonstrate how to remove the duplicates, not find them.
The following will select each row in the data frame with a duplicate 'name'
field. Note that this will find each instance, not just duplicates after the first occurrence. The keep
argument accepts additional values that can exclude either the first or last occurrence.
df[df.duplicated(['name'], keep=False)]
The pandas reference for duplicated()
can be found here.
Upvotes: 3
Reputation: 6312
value_counts will give you the number of duplicates as well.
names = df.name.value_counts()
names[names > 1]
Upvotes: 8
Reputation: 6428
A one liner can be:
x.set_index('name').index.get_duplicates()
the index contains a method for finding duplicates, columns does not seem to have a similar method..
Upvotes: 11
Reputation: 14953
I had a similar problem and came across this answer.
I guess this also works:
counts = df.groupby('name').size()
df2 = pd.DataFrame(counts, columns = ['size'])
df2 = df2[df2.size>1]
and df2.index
will give you a list of names with duplicates
Upvotes: 1
Reputation: 93804
If you want to find the rows with duplicated name (except the first time we see that), you can try this
In [16]: import pandas as pd
In [17]: p1 = {'name': 'willy', 'age': 10}
In [18]: p2 = {'name': 'willy', 'age': 11}
In [19]: p3 = {'name': 'zoe', 'age': 10}
In [20]: df = pd.DataFrame([p1, p2, p3])
In [21]: df
Out[21]:
age name
0 10 willy
1 11 willy
2 10 zoe
In [22]: df.duplicated('name')
Out[22]:
0 False
1 True
2 False
Upvotes: 39