@@ -125,20 +125,20 @@ def test_resample_basic(series, closed, expected):
125
125
def test_resample_integerarray ():
126
126
# GH 25580, resample on IntegerArray
127
127
ts = Series (
128
- range (9 ), index = pd . date_range ("1/1/2000" , periods = 9 , freq = "T" ), dtype = "Int64"
128
+ range (9 ), index = date_range ("1/1/2000" , periods = 9 , freq = "T" ), dtype = "Int64"
129
129
)
130
130
result = ts .resample ("3T" ).sum ()
131
131
expected = Series (
132
132
[3 , 12 , 21 ],
133
- index = pd . date_range ("1/1/2000" , periods = 3 , freq = "3T" ),
133
+ index = date_range ("1/1/2000" , periods = 3 , freq = "3T" ),
134
134
dtype = "Int64" ,
135
135
)
136
136
tm .assert_series_equal (result , expected )
137
137
138
138
result = ts .resample ("3T" ).mean ()
139
139
expected = Series (
140
140
[1 , 4 , 7 ],
141
- index = pd . date_range ("1/1/2000" , periods = 3 , freq = "3T" ),
141
+ index = date_range ("1/1/2000" , periods = 3 , freq = "3T" ),
142
142
dtype = "Float64" ,
143
143
)
144
144
tm .assert_series_equal (result , expected )
@@ -529,8 +529,8 @@ def test_resample_ohlc(series):
529
529
def test_resample_ohlc_result ():
530
530
531
531
# GH 12332
532
- index = pd . date_range ("1-1-2000" , "2-15-2000" , freq = "h" )
533
- index = index .union (pd . date_range ("4-15-2000" , "5-15-2000" , freq = "h" ))
532
+ index = date_range ("1-1-2000" , "2-15-2000" , freq = "h" )
533
+ index = index .union (date_range ("4-15-2000" , "5-15-2000" , freq = "h" ))
534
534
s = Series (range (len (index )), index = index )
535
535
536
536
a = s .loc [:"4-15-2000" ].resample ("30T" ).ohlc ()
@@ -805,7 +805,7 @@ def test_resample_bad_offset(offset):
805
805
def test_resample_origin_prime_freq ():
806
806
# GH 31809
807
807
start , end = "2000-10-01 23:30:00" , "2000-10-02 00:30:00"
808
- rng = pd . date_range (start , end , freq = "7min" )
808
+ rng = date_range (start , end , freq = "7min" )
809
809
ts = Series (np .random .randn (len (rng )), index = rng )
810
810
811
811
exp_rng = date_range ("2000-10-01 23:14:00" , "2000-10-02 00:22:00" , freq = "17min" )
@@ -863,7 +863,7 @@ def test_resample_origin_with_tz():
863
863
def test_resample_origin_epoch_with_tz_day_vs_24h ():
864
864
# GH 34474
865
865
start , end = "2000-10-01 23:30:00+0500" , "2000-12-02 00:30:00+0500"
866
- rng = pd . date_range (start , end , freq = "7min" )
866
+ rng = date_range (start , end , freq = "7min" )
867
867
random_values = np .random .randn (len (rng ))
868
868
ts_1 = Series (random_values , index = rng )
869
869
@@ -880,7 +880,7 @@ def test_resample_origin_epoch_with_tz_day_vs_24h():
880
880
881
881
# check that we have the similar results with two different timezones (+2H and +5H)
882
882
start , end = "2000-10-01 23:30:00+0200" , "2000-12-02 00:30:00+0200"
883
- rng = pd . date_range (start , end , freq = "7min" )
883
+ rng = date_range (start , end , freq = "7min" )
884
884
ts_2 = Series (random_values , index = rng )
885
885
result_5 = ts_2 .resample ("D" , origin = "epoch" ).mean ()
886
886
result_6 = ts_2 .resample ("24H" , origin = "epoch" ).mean ()
@@ -903,7 +903,7 @@ def _create_series(values, timestamps, freq="D"):
903
903
# test classical behavior of origin in a DST context
904
904
start = Timestamp ("2013-11-02" , tz = tz )
905
905
end = Timestamp ("2013-11-03 23:59" , tz = tz )
906
- rng = pd . date_range (start , end , freq = "1h" )
906
+ rng = date_range (start , end , freq = "1h" )
907
907
ts = Series (np .ones (len (rng )), index = rng )
908
908
909
909
expected = _create_series ([24.0 , 25.0 ], ["2013-11-02" , "2013-11-03" ])
@@ -914,7 +914,7 @@ def _create_series(values, timestamps, freq="D"):
914
914
# test complex behavior of origin/offset in a DST context
915
915
start = Timestamp ("2013-11-03" , tz = tz )
916
916
end = Timestamp ("2013-11-03 23:59" , tz = tz )
917
- rng = pd . date_range (start , end , freq = "1h" )
917
+ rng = date_range (start , end , freq = "1h" )
918
918
ts = Series (np .ones (len (rng )), index = rng )
919
919
920
920
expected_ts = ["2013-11-02 22:00-05:00" , "2013-11-03 22:00-06:00" ]
@@ -969,7 +969,7 @@ def test_period_with_agg():
969
969
# aggregate a period resampler with a lambda
970
970
s2 = Series (
971
971
np .random .randint (0 , 5 , 50 ),
972
- index = pd . period_range ("2012-01-01" , freq = "H" , periods = 50 ),
972
+ index = period_range ("2012-01-01" , freq = "H" , periods = 50 ),
973
973
dtype = "float64" ,
974
974
)
975
975
@@ -1003,7 +1003,7 @@ def test_resample_dtype_preservation():
1003
1003
1004
1004
df = DataFrame (
1005
1005
{
1006
- "date" : pd . date_range (start = "2016-01-01" , periods = 4 , freq = "W" ),
1006
+ "date" : date_range (start = "2016-01-01" , periods = 4 , freq = "W" ),
1007
1007
"group" : [1 , 1 , 2 , 2 ],
1008
1008
"val" : Series ([5 , 6 , 7 , 8 ], dtype = "int32" ),
1009
1009
}
@@ -1022,7 +1022,7 @@ def test_resample_dtype_coercion():
1022
1022
1023
1023
# GH 16361
1024
1024
df = {"a" : [1 , 3 , 1 , 4 ]}
1025
- df = DataFrame (df , index = pd . date_range ("2017-01-01" , "2017-01-04" ))
1025
+ df = DataFrame (df , index = date_range ("2017-01-01" , "2017-01-04" ))
1026
1026
1027
1027
expected = df .astype ("float64" ).resample ("H" ).mean ()["a" ].interpolate ("cubic" )
1028
1028
@@ -1056,12 +1056,12 @@ def test_nanosecond_resample_error():
1056
1056
# Resampling using pd.tseries.offsets.Nano as period
1057
1057
start = 1443707890427
1058
1058
exp_start = 1443707890400
1059
- indx = pd . date_range (start = pd .to_datetime (start ), periods = 10 , freq = "100n" )
1059
+ indx = date_range (start = pd .to_datetime (start ), periods = 10 , freq = "100n" )
1060
1060
ts = Series (range (len (indx )), index = indx )
1061
1061
r = ts .resample (pd .tseries .offsets .Nano (100 ))
1062
1062
result = r .agg ("mean" )
1063
1063
1064
- exp_indx = pd . date_range (start = pd .to_datetime (exp_start ), periods = 10 , freq = "100n" )
1064
+ exp_indx = date_range (start = pd .to_datetime (exp_start ), periods = 10 , freq = "100n" )
1065
1065
exp = Series (range (len (exp_indx )), index = exp_indx )
1066
1066
1067
1067
tm .assert_series_equal (result , exp )
@@ -1128,8 +1128,8 @@ def test_resample_anchored_multiday():
1128
1128
#
1129
1129
# See: https://github.com/pandas-dev/pandas/issues/8683
1130
1130
1131
- index1 = pd . date_range ("2014-10-14 23:06:23.206" , periods = 3 , freq = "400L" )
1132
- index2 = pd . date_range ("2014-10-15 23:00:00" , periods = 2 , freq = "2200L" )
1131
+ index1 = date_range ("2014-10-14 23:06:23.206" , periods = 3 , freq = "400L" )
1132
+ index2 = date_range ("2014-10-15 23:00:00" , periods = 2 , freq = "2200L" )
1133
1133
index = index1 .union (index2 )
1134
1134
1135
1135
s = Series (np .random .randn (5 ), index = index )
@@ -1174,7 +1174,7 @@ def test_anchored_lowercase_buglet():
1174
1174
1175
1175
def test_upsample_apply_functions ():
1176
1176
# #1596
1177
- rng = pd . date_range ("2012-06-12" , periods = 4 , freq = "h" )
1177
+ rng = date_range ("2012-06-12" , periods = 4 , freq = "h" )
1178
1178
1179
1179
ts = Series (np .random .randn (len (rng )), index = rng )
1180
1180
@@ -1183,7 +1183,7 @@ def test_upsample_apply_functions():
1183
1183
1184
1184
1185
1185
def test_resample_not_monotonic ():
1186
- rng = pd . date_range ("2012-06-12" , periods = 200 , freq = "h" )
1186
+ rng = date_range ("2012-06-12" , periods = 200 , freq = "h" )
1187
1187
ts = Series (np .random .randn (len (rng )), index = rng )
1188
1188
1189
1189
ts = ts .take (np .random .permutation (len (ts )))
@@ -1255,12 +1255,12 @@ def test_resample_consistency():
1255
1255
# GH 6418
1256
1256
# resample with bfill / limit / reindex consistency
1257
1257
1258
- i30 = pd . date_range ("2002-02-02" , periods = 4 , freq = "30T" )
1258
+ i30 = date_range ("2002-02-02" , periods = 4 , freq = "30T" )
1259
1259
s = Series (np .arange (4.0 ), index = i30 )
1260
1260
s [2 ] = np .NaN
1261
1261
1262
1262
# Upsample by factor 3 with reindex() and resample() methods:
1263
- i10 = pd . date_range (i30 [0 ], i30 [- 1 ], freq = "10T" )
1263
+ i10 = date_range (i30 [0 ], i30 [- 1 ], freq = "10T" )
1264
1264
1265
1265
s10 = s .reindex (index = i10 , method = "bfill" )
1266
1266
s10_2 = s .reindex (index = i10 , method = "bfill" , limit = 2 )
@@ -1364,8 +1364,8 @@ def test_resample_nunique_preserves_column_level_names():
1364
1364
1365
1365
def test_resample_nunique_with_date_gap ():
1366
1366
# GH 13453
1367
- index = pd . date_range ("1-1-2000" , "2-15-2000" , freq = "h" )
1368
- index2 = pd . date_range ("4-15-2000" , "5-15-2000" , freq = "h" )
1367
+ index = date_range ("1-1-2000" , "2-15-2000" , freq = "h" )
1368
+ index2 = date_range ("4-15-2000" , "5-15-2000" , freq = "h" )
1369
1369
index3 = index .append (index2 )
1370
1370
s = Series (range (len (index3 )), index = index3 , dtype = "int64" )
1371
1371
r = s .resample ("M" )
@@ -1462,7 +1462,7 @@ def test_groupby_with_dst_time_change():
1462
1462
1463
1463
df = DataFrame ([1 , 2 ], index = index )
1464
1464
result = df .groupby (Grouper (freq = "1d" )).last ()
1465
- expected_index_values = pd . date_range (
1465
+ expected_index_values = date_range (
1466
1466
"2016-11-02" , "2016-11-24" , freq = "d" , tz = "America/Chicago"
1467
1467
)
1468
1468
@@ -1587,11 +1587,11 @@ def test_downsample_across_dst_weekly():
1587
1587
)
1588
1588
tm .assert_frame_equal (result , expected )
1589
1589
1590
- idx = pd . date_range ("2013-04-01" , "2013-05-01" , tz = "Europe/London" , freq = "H" )
1590
+ idx = date_range ("2013-04-01" , "2013-05-01" , tz = "Europe/London" , freq = "H" )
1591
1591
s = Series (index = idx , dtype = np .float64 )
1592
1592
result = s .resample ("W" ).mean ()
1593
1593
expected = Series (
1594
- index = pd . date_range ("2013-04-07" , freq = "W" , periods = 5 , tz = "Europe/London" ),
1594
+ index = date_range ("2013-04-07" , freq = "W" , periods = 5 , tz = "Europe/London" ),
1595
1595
dtype = np .float64 ,
1596
1596
)
1597
1597
tm .assert_series_equal (result , expected )
@@ -1601,7 +1601,7 @@ def test_downsample_dst_at_midnight():
1601
1601
# GH 25758
1602
1602
start = datetime (2018 , 11 , 3 , 12 )
1603
1603
end = datetime (2018 , 11 , 5 , 12 )
1604
- index = pd . date_range (start , end , freq = "1H" )
1604
+ index = date_range (start , end , freq = "1H" )
1605
1605
index = index .tz_localize ("UTC" ).tz_convert ("America/Havana" )
1606
1606
data = list (range (len (index )))
1607
1607
dataframe = DataFrame (data , index = index )
@@ -1681,7 +1681,7 @@ def f(data, add_arg):
1681
1681
tm .assert_series_equal (result , expected )
1682
1682
1683
1683
# Testing dataframe
1684
- df = DataFrame ({"A" : 1 , "B" : 2 }, index = pd . date_range ("2017" , periods = 10 ))
1684
+ df = DataFrame ({"A" : 1 , "B" : 2 }, index = date_range ("2017" , periods = 10 ))
1685
1685
result = df .groupby ("A" ).resample ("D" ).agg (f , multiplier )
1686
1686
expected = df .groupby ("A" ).resample ("D" ).mean ().multiply (multiplier )
1687
1687
tm .assert_frame_equal (result , expected )
@@ -1707,7 +1707,7 @@ def test_resample_equivalent_offsets(n1, freq1, n2, freq2, k):
1707
1707
# GH 24127
1708
1708
n1_ = n1 * k
1709
1709
n2_ = n2 * k
1710
- s = Series (0 , index = pd . date_range ("19910905 13:00" , "19911005 07:00" , freq = freq1 ))
1710
+ s = Series (0 , index = date_range ("19910905 13:00" , "19911005 07:00" , freq = freq1 ))
1711
1711
s = s + range (len (s ))
1712
1712
1713
1713
result1 = s .resample (str (n1_ ) + freq1 ).mean ()
@@ -1797,10 +1797,10 @@ def test_resample_calendar_day_with_dst(
1797
1797
first : str , last : str , freq_in : str , freq_out : str , exp_last : str
1798
1798
):
1799
1799
# GH 35219
1800
- ts = Series (1.0 , pd . date_range (first , last , freq = freq_in , tz = "Europe/Amsterdam" ))
1800
+ ts = Series (1.0 , date_range (first , last , freq = freq_in , tz = "Europe/Amsterdam" ))
1801
1801
result = ts .resample (freq_out ).pad ()
1802
1802
expected = Series (
1803
- 1.0 , pd . date_range (first , exp_last , freq = freq_out , tz = "Europe/Amsterdam" )
1803
+ 1.0 , date_range (first , exp_last , freq = freq_out , tz = "Europe/Amsterdam" )
1804
1804
)
1805
1805
tm .assert_series_equal (result , expected )
1806
1806
0 commit comments