Skip to content

MAINT: Remove assertIn from testing #16101

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Apr 23, 2017
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion pandas/tests/frame/test_alter_axes.py
Original file line number Diff line number Diff line change
Expand Up @@ -138,7 +138,7 @@ def test_set_index_nonuniq(self):
'E': np.random.randn(5)})
with assertRaisesRegexp(ValueError, 'Index has duplicate keys'):
df.set_index('A', verify_integrity=True, inplace=True)
self.assertIn('A', df)
assert 'A' in df

def test_set_index_bug(self):
# GH1590
Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/frame/test_axis_select_reindex.py
Original file line number Diff line number Diff line change
Expand Up @@ -734,7 +734,7 @@ def test_filter_regex_search(self):
# regex
filtered = fcopy.filter(regex='[A]+')
self.assertEqual(len(filtered.columns), 2)
self.assertIn('AA', filtered)
assert 'AA' in filtered

# doesn't have to be at beginning
df = DataFrame({'aBBa': [1, 2],
Expand Down
6 changes: 3 additions & 3 deletions pandas/tests/frame/test_convert_to.py
Original file line number Diff line number Diff line change
Expand Up @@ -156,16 +156,16 @@ def test_to_records_index_name(self):
df = DataFrame(np.random.randn(3, 3))
df.index.name = 'X'
rs = df.to_records()
self.assertIn('X', rs.dtype.fields)
assert 'X' in rs.dtype.fields

df = DataFrame(np.random.randn(3, 3))
rs = df.to_records()
self.assertIn('index', rs.dtype.fields)
assert 'index' in rs.dtype.fields

df.index = MultiIndex.from_tuples([('a', 'x'), ('a', 'y'), ('b', 'z')])
df.index.names = ['A', None]
rs = df.to_records()
self.assertIn('level_0', rs.dtype.fields)
assert 'level_0' in rs.dtype.fields

def test_to_records_with_unicode_index(self):
# GH13172
Expand Down
4 changes: 2 additions & 2 deletions pandas/tests/frame/test_indexing.py
Original file line number Diff line number Diff line change
Expand Up @@ -422,7 +422,7 @@ def test_setitem(self):
# not sure what else to do here
series = self.frame['A'][::2]
self.frame['col5'] = series
self.assertIn('col5', self.frame)
assert 'col5' in self.frame

self.assertEqual(len(series), 15)
self.assertEqual(len(self.frame), 30)
Expand Down Expand Up @@ -600,7 +600,7 @@ def test_setitem_corner(self):
index=np.arange(3))
del df['B']
df['B'] = [1., 2., 3.]
self.assertIn('B', df)
assert 'B' in df
self.assertEqual(len(df.columns), 2)

df['A'] = 'beginning'
Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/frame/test_operators.py
Original file line number Diff line number Diff line change
Expand Up @@ -831,7 +831,7 @@ def test_combineSeries(self):

for key, s in compat.iteritems(self.frame):
assert_series_equal(larger_added[key], s + series[key])
self.assertIn('E', larger_added)
assert 'E' in larger_added
self.assertTrue(np.isnan(larger_added['E']).all())

# vs mix (upcast) as needed
Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/frame/test_repr_info.py
Original file line number Diff line number Diff line change
Expand Up @@ -171,7 +171,7 @@ def test_repr_column_name_unicode_truncation_bug(self):
' the File through the code..')})

result = repr(df)
self.assertIn('StringCol', result)
assert 'StringCol' in result

def test_latex_repr(self):
result = r"""\begin{tabular}{llll}
Expand Down
4 changes: 2 additions & 2 deletions pandas/tests/frame/test_to_csv.py
Original file line number Diff line number Diff line change
Expand Up @@ -909,7 +909,7 @@ def test_to_csv_compression_gzip(self):
text = f.read().decode('utf8')
f.close()
for col in df.columns:
self.assertIn(col, text)
assert col in text

def test_to_csv_compression_bz2(self):
# GH7615
Expand All @@ -932,7 +932,7 @@ def test_to_csv_compression_bz2(self):
text = f.read().decode('utf8')
f.close()
for col in df.columns:
self.assertIn(col, text)
assert col in text

def test_to_csv_compression_xz(self):
# GH11852
Expand Down
32 changes: 16 additions & 16 deletions pandas/tests/groupby/test_groupby.py
Original file line number Diff line number Diff line change
Expand Up @@ -2483,14 +2483,14 @@ def test_groupby_series_with_name(self):
result = self.df.groupby(self.df['A']).mean()
result2 = self.df.groupby(self.df['A'], as_index=False).mean()
self.assertEqual(result.index.name, 'A')
self.assertIn('A', result2)
assert 'A' in result2

result = self.df.groupby([self.df['A'], self.df['B']]).mean()
result2 = self.df.groupby([self.df['A'], self.df['B']],
as_index=False).mean()
self.assertEqual(result.index.names, ('A', 'B'))
self.assertIn('A', result2)
self.assertIn('B', result2)
assert 'A' in result2
assert 'B' in result2

def test_seriesgroupby_name_attr(self):
# GH 6265
Expand Down Expand Up @@ -3357,10 +3357,10 @@ def test_groupby_with_small_elem(self):
'change': [1234, 5678]},
index=pd.DatetimeIndex(['2014-09-10', '2013-10-10']))
grouped = df.groupby([pd.TimeGrouper(freq='M'), 'event'])
self.assertEqual(len(grouped.groups), 2)
self.assertEqual(grouped.ngroups, 2)
self.assertIn((pd.Timestamp('2014-09-30'), 'start'), grouped.groups)
self.assertIn((pd.Timestamp('2013-10-31'), 'start'), grouped.groups)
assert len(grouped.groups) == 2
assert grouped.ngroups == 2
assert (pd.Timestamp('2014-09-30'), 'start') in grouped.groups
assert (pd.Timestamp('2013-10-31'), 'start') in grouped.groups

res = grouped.get_group((pd.Timestamp('2014-09-30'), 'start'))
tm.assert_frame_equal(res, df.iloc[[0], :])
Expand All @@ -3372,10 +3372,10 @@ def test_groupby_with_small_elem(self):
index=pd.DatetimeIndex(['2014-09-10', '2013-10-10',
'2014-09-15']))
grouped = df.groupby([pd.TimeGrouper(freq='M'), 'event'])
self.assertEqual(len(grouped.groups), 2)
self.assertEqual(grouped.ngroups, 2)
self.assertIn((pd.Timestamp('2014-09-30'), 'start'), grouped.groups)
self.assertIn((pd.Timestamp('2013-10-31'), 'start'), grouped.groups)
assert len(grouped.groups) == 2
assert grouped.ngroups == 2
assert (pd.Timestamp('2014-09-30'), 'start') in grouped.groups
assert (pd.Timestamp('2013-10-31'), 'start') in grouped.groups

res = grouped.get_group((pd.Timestamp('2014-09-30'), 'start'))
tm.assert_frame_equal(res, df.iloc[[0, 2], :])
Expand All @@ -3388,11 +3388,11 @@ def test_groupby_with_small_elem(self):
index=pd.DatetimeIndex(['2014-09-10', '2013-10-10',
'2014-08-05']))
grouped = df.groupby([pd.TimeGrouper(freq='M'), 'event'])
self.assertEqual(len(grouped.groups), 3)
self.assertEqual(grouped.ngroups, 3)
self.assertIn((pd.Timestamp('2014-09-30'), 'start'), grouped.groups)
self.assertIn((pd.Timestamp('2013-10-31'), 'start'), grouped.groups)
self.assertIn((pd.Timestamp('2014-08-31'), 'start'), grouped.groups)
assert len(grouped.groups) == 3
assert grouped.ngroups == 3
assert (pd.Timestamp('2014-09-30'), 'start') in grouped.groups
assert (pd.Timestamp('2013-10-31'), 'start') in grouped.groups
assert (pd.Timestamp('2014-08-31'), 'start') in grouped.groups

res = grouped.get_group((pd.Timestamp('2014-09-30'), 'start'))
tm.assert_frame_equal(res, df.iloc[[0], :])
Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/indexes/datetimes/test_datetime.py
Original file line number Diff line number Diff line change
Expand Up @@ -101,7 +101,7 @@ def test_reasonable_keyerror(self):
try:
index.get_loc('1/1/2000')
except KeyError as e:
self.assertIn('2000', str(e))
assert '2000' in str(e)

def test_roundtrip_pickle_with_tz(self):

Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/indexes/datetimes/test_ops.py
Original file line number Diff line number Diff line change
Expand Up @@ -632,7 +632,7 @@ def test_nonunique_contains(self):
for idx in map(DatetimeIndex,
([0, 1, 0], [0, 0, -1], [0, -1, -1],
['2015', '2015', '2016'], ['2015', '2015', '2014'])):
tm.assertIn(idx[0], idx)
assert idx[0] in idx

def test_order(self):
# with freq
Expand Down
8 changes: 4 additions & 4 deletions pandas/tests/indexes/test_base.py
Original file line number Diff line number Diff line change
Expand Up @@ -857,10 +857,10 @@ def test_add_string(self):
def test_iadd_string(self):
index = pd.Index(['a', 'b', 'c'])
# doesn't fail test unless there is a check before `+=`
self.assertIn('a', index)
assert 'a' in index

index += '_x'
self.assertIn('a_x', index)
assert 'a_x' in index

def test_difference(self):

Expand Down Expand Up @@ -963,8 +963,8 @@ def test_summary(self):
ind = Index(['{other}%s', "~:{range}:0"], name='A')
result = ind.summary()
# shouldn't be formatted accidentally.
self.assertIn('~:{range}:0', result)
self.assertIn('{other}%s', result)
assert '~:{range}:0' in result
assert '{other}%s' in result

def test_format(self):
self._check_method_works(Index.format)
Expand Down
4 changes: 2 additions & 2 deletions pandas/tests/indexes/test_multi.py
Original file line number Diff line number Diff line change
Expand Up @@ -1597,8 +1597,8 @@ def test_union(self):
# other = Index(['A', 'B', 'C'])

# result = other.union(self.index)
# self.assertIn(('foo', 'one'), result)
# self.assertIn('B', result)
# assert ('foo', 'one') in result
# assert 'B' in result

# result2 = self.index.union(other)
# self.assertTrue(result.equals(result2))
Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/indexes/timedeltas/test_ops.py
Original file line number Diff line number Diff line change
Expand Up @@ -561,7 +561,7 @@ def test_nonunique_contains(self):
for idx in map(TimedeltaIndex, ([0, 1, 0], [0, 0, -1], [0, -1, -1],
['00:01:00', '00:01:00', '00:02:00'],
['00:01:00', '00:01:00', '00:00:01'])):
tm.assertIn(idx[0], idx)
assert idx[0] in idx

def test_unknown_attribute(self):
# see gh-9680
Expand Down
8 changes: 4 additions & 4 deletions pandas/tests/indexing/test_chaining_and_caching.py
Original file line number Diff line number Diff line change
Expand Up @@ -373,15 +373,15 @@ def test_cache_updating(self):
df['A'] # cache series
with catch_warnings(record=True):
df.ix["Hello Friend"] = df.ix[0]
self.assertIn("Hello Friend", df['A'].index)
self.assertIn("Hello Friend", df['B'].index)
assert "Hello Friend" in df['A'].index
assert "Hello Friend" in df['B'].index

with catch_warnings(record=True):
panel = tm.makePanel()
panel.ix[0] # get first item into cache
panel.ix[:, :, 'A+1'] = panel.ix[:, :, 'A'] + 1
self.assertIn("A+1", panel.ix[0].columns)
self.assertIn("A+1", panel.ix[1].columns)
assert "A+1" in panel.ix[0].columns
assert "A+1" in panel.ix[1].columns

# 5216
# make sure that we don't try to set a dead cache
Expand Down
8 changes: 4 additions & 4 deletions pandas/tests/io/formats/test_format.py
Original file line number Diff line number Diff line change
Expand Up @@ -959,7 +959,7 @@ def test_wide_repr_named(self):
self.assertTrue(len(wider_repr) < len(wide_repr))

for line in wide_repr.splitlines()[1::13]:
self.assertIn('DataFrame Index', line)
assert 'DataFrame Index' in line

reset_option('display.expand_frame_repr')

Expand All @@ -981,7 +981,7 @@ def test_wide_repr_multiindex(self):
self.assertTrue(len(wider_repr) < len(wide_repr))

for line in wide_repr.splitlines()[1::13]:
self.assertIn('Level 0 Level 1', line)
assert 'Level 0 Level 1' in line

reset_option('display.expand_frame_repr')

Expand Down Expand Up @@ -1875,9 +1875,9 @@ def test_float_trim_zeros(self):
if line.startswith('dtype:'):
continue
if _three_digit_exp():
self.assertIn('+010', line)
assert '+010' in line
else:
self.assertIn('+10', line)
assert '+10' in line

def test_datetimeindex(self):

Expand Down
2 changes: 1 addition & 1 deletion pandas/tests/io/parser/c_parser_only.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ def test_buffer_overflow(self):
try:
self.read_table(StringIO(malf))
except Exception as err:
self.assertIn(cperr, str(err))
assert cperr in str(err)

def test_buffer_rd_bytes(self):
# see gh-12098: src->buffer in the C parser can be freed twice leading
Expand Down
12 changes: 6 additions & 6 deletions pandas/tests/io/parser/parse_dates.py
Original file line number Diff line number Diff line change
Expand Up @@ -135,7 +135,7 @@ def test_multiple_date_cols_int_cast(self):
# it works!
df = self.read_csv(StringIO(data), header=None, parse_dates=date_spec,
date_parser=conv.parse_date_time)
self.assertIn('nominal', df)
assert 'nominal' in df

def test_multiple_date_col_timestamp_parse(self):
data = """05/31/2012,15:30:00.029,1306.25,1,E,0,,1306.25
Expand Down Expand Up @@ -530,7 +530,7 @@ def test_parse_date_time(self):
df = self.read_csv(StringIO(data), sep=',', header=0,
parse_dates=datecols,
date_parser=conv.parse_date_time)
self.assertIn('date_time', df)
assert 'date_time' in df
self.assertEqual(df.date_time.loc[0], datetime(2001, 1, 5, 10, 0, 0))

data = ("KORD,19990127, 19:00:00, 18:56:00, 0.8100\n"
Expand Down Expand Up @@ -558,7 +558,7 @@ def test_parse_date_fields(self):
df = self.read_csv(StringIO(data), sep=',', header=0,
parse_dates=datecols,
date_parser=conv.parse_date_fields)
self.assertIn('ymd', df)
assert 'ymd' in df
self.assertEqual(df.ymd.loc[0], datetime(2001, 1, 10))

def test_datetime_six_col(self):
Expand All @@ -585,7 +585,7 @@ def test_datetime_six_col(self):
df = self.read_csv(StringIO(data), sep=',', header=0,
parse_dates=datecols,
date_parser=conv.parse_all_fields)
self.assertIn('ymdHMS', df)
assert 'ymdHMS' in df
self.assertEqual(df.ymdHMS.loc[0], datetime(2001, 1, 5, 10, 0, 0))

def test_datetime_fractional_seconds(self):
Expand All @@ -598,7 +598,7 @@ def test_datetime_fractional_seconds(self):
df = self.read_csv(StringIO(data), sep=',', header=0,
parse_dates=datecols,
date_parser=conv.parse_all_fields)
self.assertIn('ymdHMS', df)
assert 'ymdHMS' in df
self.assertEqual(df.ymdHMS.loc[0], datetime(2001, 1, 5, 10, 0, 0,
microsecond=123456))
self.assertEqual(df.ymdHMS.loc[1], datetime(2001, 1, 5, 10, 0, 0,
Expand All @@ -611,7 +611,7 @@ def test_generic(self):
df = self.read_csv(StringIO(data), sep=',', header=0,
parse_dates=datecols,
date_parser=dateconverter)
self.assertIn('ym', df)
assert 'ym' in df
self.assertEqual(df.ym.loc[0], date(2001, 1, 1))

def test_dateparser_resolution_if_not_ns(self):
Expand Down
4 changes: 2 additions & 2 deletions pandas/tests/io/test_html.py
Original file line number Diff line number Diff line change
Expand Up @@ -566,10 +566,10 @@ def test_gold_canyon(self):
with open(self.banklist_data, 'r') as f:
raw_text = f.read()

self.assertIn(gc, raw_text)
assert gc in raw_text
df = self.read_html(self.banklist_data, 'Gold Canyon',
attrs={'id': 'table'})[0]
self.assertIn(gc, df.to_string())
assert gc in df.to_string()

def test_different_number_of_rows(self):
expected = """<table border="1" class="dataframe">
Expand Down
10 changes: 5 additions & 5 deletions pandas/tests/reshape/test_join.py
Original file line number Diff line number Diff line change
Expand Up @@ -153,15 +153,15 @@ def test_handle_overlap(self):
joined = merge(self.df, self.df2, on='key2',
suffixes=['.foo', '.bar'])

self.assertIn('key1.foo', joined)
self.assertIn('key1.bar', joined)
assert 'key1.foo' in joined
assert 'key1.bar' in joined

def test_handle_overlap_arbitrary_key(self):
joined = merge(self.df, self.df2,
left_on='key2', right_on='key1',
suffixes=['.foo', '.bar'])
self.assertIn('key1.foo', joined)
self.assertIn('key2.bar', joined)
assert 'key1.foo' in joined
assert 'key2.bar' in joined

def test_join_on(self):
target = self.target
Expand Down Expand Up @@ -251,7 +251,7 @@ def test_join_with_len0(self):
# nothing to merge
merged = self.target.join(self.source.reindex([]), on='C')
for col in self.source:
self.assertIn(col, merged)
assert col in merged
self.assertTrue(merged[col].isnull().all())

merged2 = self.target.join(self.source.reindex([]), on='C',
Expand Down
4 changes: 2 additions & 2 deletions pandas/tests/reshape/test_merge.py
Original file line number Diff line number Diff line change
Expand Up @@ -128,8 +128,8 @@ def test_merge_overlap(self):
merged = merge(self.left, self.left, on='key')
exp_len = (self.left['key'].value_counts() ** 2).sum()
self.assertEqual(len(merged), exp_len)
self.assertIn('v1_x', merged)
self.assertIn('v1_y', merged)
assert 'v1_x' in merged
assert 'v1_y' in merged

def test_merge_different_column_key_names(self):
left = DataFrame({'lkey': ['foo', 'bar', 'baz', 'foo'],
Expand Down
Loading