Skip to content

Commit b76031d

Browse files
authored
DOC: normalize usage of word "pandas" (#36845)
1 parent 6f6310b commit b76031d

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

44 files changed

+134
-134
lines changed

doc/source/development/code_style.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ pandas code style guide
99
.. contents:: Table of contents:
1010
:local:
1111

12-
*pandas* follows the `PEP8 <https://www.python.org/dev/peps/pep-0008/>`_
12+
pandas follows the `PEP8 <https://www.python.org/dev/peps/pep-0008/>`_
1313
standard and uses `Black <https://black.readthedocs.io/en/stable/>`_
1414
and `Flake8 <https://flake8.pycqa.org/en/latest/>`_ to ensure a
1515
consistent code format throughout the project. For details see the

doc/source/development/contributing.rst

+3-3
Original file line numberDiff line numberDiff line change
@@ -155,7 +155,7 @@ Using a Docker container
155155

156156
Instead of manually setting up a development environment, you can use `Docker
157157
<https://docs.docker.com/get-docker/>`_ to automatically create the environment with just several
158-
commands. Pandas provides a ``DockerFile`` in the root directory to build a Docker image
158+
commands. pandas provides a ``DockerFile`` in the root directory to build a Docker image
159159
with a full pandas development environment.
160160

161161
**Docker Commands**
@@ -190,7 +190,7 @@ Note that you might need to rebuild the C extensions if/when you merge with upst
190190
Installing a C compiler
191191
~~~~~~~~~~~~~~~~~~~~~~~
192192

193-
Pandas uses C extensions (mostly written using Cython) to speed up certain
193+
pandas uses C extensions (mostly written using Cython) to speed up certain
194194
operations. To install pandas from source, you need to compile these C
195195
extensions, which means you need a C compiler. This process depends on which
196196
platform you're using.
@@ -1219,7 +1219,7 @@ This test shows off several useful features of Hypothesis, as well as
12191219
demonstrating a good use-case: checking properties that should hold over
12201220
a large or complicated domain of inputs.
12211221

1222-
To keep the Pandas test suite running quickly, parametrized tests are
1222+
To keep the pandas test suite running quickly, parametrized tests are
12231223
preferred if the inputs or logic are simple, with Hypothesis tests reserved
12241224
for cases with complex logic or where there are too many combinations of
12251225
options or subtle interactions to test (or think of!) all of them.

doc/source/development/maintaining.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -207,7 +207,7 @@ Only core team members can merge pull requests. We have a few guidelines.
207207
1. You should typically not self-merge your own pull requests. Exceptions include
208208
things like small changes to fix CI (e.g. pinning a package version).
209209
2. You should not merge pull requests that have an active discussion, or pull
210-
requests that has any ``-1`` votes from a core maintainer. Pandas operates
210+
requests that has any ``-1`` votes from a core maintainer. pandas operates
211211
by consensus.
212212
3. For larger changes, it's good to have a +1 from at least two core team members.
213213

doc/source/ecosystem.rst

+11-11
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,7 @@ With Altair, you can spend more time understanding your data and its
9898
meaning. Altair's API is simple, friendly and consistent and built on
9999
top of the powerful Vega-Lite JSON specification. This elegant
100100
simplicity produces beautiful and effective visualizations with a
101-
minimal amount of code. Altair works with Pandas DataFrames.
101+
minimal amount of code. Altair works with pandas DataFrames.
102102

103103

104104
`Bokeh <https://bokeh.pydata.org>`__
@@ -110,7 +110,7 @@ graphics in the style of Protovis/D3, while delivering high-performance interact
110110
large data to thin clients.
111111

112112
`Pandas-Bokeh <https://github.com/PatrikHlobil/Pandas-Bokeh>`__ provides a high level API
113-
for Bokeh that can be loaded as a native Pandas plotting backend via
113+
for Bokeh that can be loaded as a native pandas plotting backend via
114114

115115
.. code:: python
116116
@@ -187,7 +187,7 @@ IDE
187187
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
188188

189189
IPython is an interactive command shell and distributed computing
190-
environment. IPython tab completion works with Pandas methods and also
190+
environment. IPython tab completion works with pandas methods and also
191191
attributes like DataFrame columns.
192192

193193
`Jupyter Notebook / Jupyter Lab <https://jupyter.org>`__
@@ -201,7 +201,7 @@ Jupyter notebooks can be converted to a number of open standard output formats
201201
Python) through 'Download As' in the web interface and ``jupyter convert``
202202
in a shell.
203203

204-
Pandas DataFrames implement ``_repr_html_``and ``_repr_latex`` methods
204+
pandas DataFrames implement ``_repr_html_``and ``_repr_latex`` methods
205205
which are utilized by Jupyter Notebook for displaying
206206
(abbreviated) HTML or LaTeX tables. LaTeX output is properly escaped.
207207
(Note: HTML tables may or may not be
@@ -229,7 +229,7 @@ Its `Variable Explorer <https://docs.spyder-ide.org/variableexplorer.html>`__
229229
allows users to view, manipulate and edit pandas ``Index``, ``Series``,
230230
and ``DataFrame`` objects like a "spreadsheet", including copying and modifying
231231
values, sorting, displaying a "heatmap", converting data types and more.
232-
Pandas objects can also be renamed, duplicated, new columns added,
232+
pandas objects can also be renamed, duplicated, new columns added,
233233
copyed/pasted to/from the clipboard (as TSV), and saved/loaded to/from a file.
234234
Spyder can also import data from a variety of plain text and binary files
235235
or the clipboard into a new pandas DataFrame via a sophisticated import wizard.
@@ -276,13 +276,13 @@ The following data feeds are available:
276276
`Quandl/Python <https://github.com/quandl/Python>`__
277277
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
278278
Quandl API for Python wraps the Quandl REST API to return
279-
Pandas DataFrames with timeseries indexes.
279+
pandas DataFrames with timeseries indexes.
280280

281281
`Pydatastream <https://github.com/vfilimonov/pydatastream>`__
282282
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
283283
PyDatastream is a Python interface to the
284284
`Refinitiv Datastream (DWS) <https://www.refinitiv.com/en/products/datastream-macroeconomic-analysis>`__
285-
REST API to return indexed Pandas DataFrames with financial data.
285+
REST API to return indexed pandas DataFrames with financial data.
286286
This package requires valid credentials for this API (non free).
287287

288288
`pandaSDMX <https://pandasdmx.readthedocs.io>`__
@@ -357,7 +357,7 @@ Out-of-core
357357
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
358358

359359
Blaze provides a standard API for doing computations with various
360-
in-memory and on-disk backends: NumPy, Pandas, SQLAlchemy, MongoDB, PyTables,
360+
in-memory and on-disk backends: NumPy, pandas, SQLAlchemy, MongoDB, PyTables,
361361
PySpark.
362362

363363
`Dask <https://dask.readthedocs.io/en/latest/>`__
@@ -403,7 +403,7 @@ If also displays progress bars.
403403
`Ray <https://ray.readthedocs.io/en/latest/pandas_on_ray.html>`__
404404
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
405405

406-
Pandas on Ray is an early stage DataFrame library that wraps Pandas and transparently distributes the data and computation. The user does not need to know how many cores their system has, nor do they need to specify how to distribute the data. In fact, users can continue using their previous Pandas notebooks while experiencing a considerable speedup from Pandas on Ray, even on a single machine. Only a modification of the import statement is needed, as we demonstrate below. Once you’ve changed your import statement, you’re ready to use Pandas on Ray just like you would Pandas.
406+
pandas on Ray is an early stage DataFrame library that wraps pandas and transparently distributes the data and computation. The user does not need to know how many cores their system has, nor do they need to specify how to distribute the data. In fact, users can continue using their previous pandas notebooks while experiencing a considerable speedup from pandas on Ray, even on a single machine. Only a modification of the import statement is needed, as we demonstrate below. Once you’ve changed your import statement, you’re ready to use pandas on Ray just like you would pandas.
407407

408408
.. code:: python
409409
@@ -414,7 +414,7 @@ Pandas on Ray is an early stage DataFrame library that wraps Pandas and transpar
414414
`Vaex <https://docs.vaex.io/>`__
415415
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
416416

417-
Increasingly, packages are being built on top of pandas to address specific needs in data preparation, analysis and visualization. Vaex is a python library for Out-of-Core DataFrames (similar to Pandas), to visualize and explore big tabular datasets. It can calculate statistics such as mean, sum, count, standard deviation etc, on an N-dimensional grid up to a billion (10\ :sup:`9`) objects/rows per second. Visualization is done using histograms, density plots and 3d volume rendering, allowing interactive exploration of big data. Vaex uses memory mapping, zero memory copy policy and lazy computations for best performance (no memory wasted).
417+
Increasingly, packages are being built on top of pandas to address specific needs in data preparation, analysis and visualization. Vaex is a python library for Out-of-Core DataFrames (similar to pandas), to visualize and explore big tabular datasets. It can calculate statistics such as mean, sum, count, standard deviation etc, on an N-dimensional grid up to a billion (10\ :sup:`9`) objects/rows per second. Visualization is done using histograms, density plots and 3d volume rendering, allowing interactive exploration of big data. Vaex uses memory mapping, zero memory copy policy and lazy computations for best performance (no memory wasted).
418418

419419
* vaex.from_pandas
420420
* vaex.to_pandas_df
@@ -424,7 +424,7 @@ Increasingly, packages are being built on top of pandas to address specific need
424424
Extension data types
425425
--------------------
426426

427-
Pandas provides an interface for defining
427+
pandas provides an interface for defining
428428
:ref:`extension types <extending.extension-types>` to extend NumPy's type
429429
system. The following libraries implement that interface to provide types not
430430
found in NumPy or pandas, which work well with pandas' data containers.

doc/source/getting_started/comparison/comparison_with_r.rst

+8-8
Original file line numberDiff line numberDiff line change
@@ -5,11 +5,11 @@
55
Comparison with R / R libraries
66
*******************************
77

8-
Since ``pandas`` aims to provide a lot of the data manipulation and analysis
8+
Since pandas aims to provide a lot of the data manipulation and analysis
99
functionality that people use `R <https://www.r-project.org/>`__ for, this page
1010
was started to provide a more detailed look at the `R language
1111
<https://en.wikipedia.org/wiki/R_(programming_language)>`__ and its many third
12-
party libraries as they relate to ``pandas``. In comparisons with R and CRAN
12+
party libraries as they relate to pandas. In comparisons with R and CRAN
1313
libraries, we care about the following things:
1414

1515
* **Functionality / flexibility**: what can/cannot be done with each tool
@@ -21,7 +21,7 @@ libraries, we care about the following things:
2121
This page is also here to offer a bit of a translation guide for users of these
2222
R packages.
2323

24-
For transfer of ``DataFrame`` objects from ``pandas`` to R, one option is to
24+
For transfer of ``DataFrame`` objects from pandas to R, one option is to
2525
use HDF5 files, see :ref:`io.external_compatibility` for an
2626
example.
2727

@@ -118,7 +118,7 @@ or by integer location
118118
df <- data.frame(matrix(rnorm(1000), ncol=100))
119119
df[, c(1:10, 25:30, 40, 50:100)]
120120
121-
Selecting multiple columns by name in ``pandas`` is straightforward
121+
Selecting multiple columns by name in pandas is straightforward
122122

123123
.. ipython:: python
124124
@@ -235,7 +235,7 @@ since the subclass sizes are possibly irregular. Using a data.frame called
235235
tapply(baseball$batting.average, baseball.example$team,
236236
max)
237237
238-
In ``pandas`` we may use :meth:`~pandas.pivot_table` method to handle this:
238+
In pandas we may use :meth:`~pandas.pivot_table` method to handle this:
239239

240240
.. ipython:: python
241241
@@ -268,7 +268,7 @@ column's values are less than another column's values:
268268
subset(df, a <= b)
269269
df[df$a <= df$b,] # note the comma
270270
271-
In ``pandas``, there are a few ways to perform subsetting. You can use
271+
In pandas, there are a few ways to perform subsetting. You can use
272272
:meth:`~pandas.DataFrame.query` or pass an expression as if it were an
273273
index/slice as well as standard boolean indexing:
274274

@@ -295,7 +295,7 @@ An expression using a data.frame called ``df`` in R with the columns ``a`` and
295295
with(df, a + b)
296296
df$a + df$b # same as the previous expression
297297
298-
In ``pandas`` the equivalent expression, using the
298+
In pandas the equivalent expression, using the
299299
:meth:`~pandas.DataFrame.eval` method, would be:
300300

301301
.. ipython:: python
@@ -347,7 +347,7 @@ summarize ``x`` by ``month``:
347347
mean = round(mean(x), 2),
348348
sd = round(sd(x), 2))
349349
350-
In ``pandas`` the equivalent expression, using the
350+
In pandas the equivalent expression, using the
351351
:meth:`~pandas.DataFrame.groupby` method, would be:
352352

353353
.. ipython:: python

doc/source/getting_started/comparison/comparison_with_stata.rst

+4-4
Original file line numberDiff line numberDiff line change
@@ -146,7 +146,7 @@ the pandas command would be:
146146
# alternatively, read_table is an alias to read_csv with tab delimiter
147147
tips = pd.read_table("tips.csv", header=None)
148148
149-
Pandas can also read Stata data sets in ``.dta`` format with the :func:`read_stata` function.
149+
pandas can also read Stata data sets in ``.dta`` format with the :func:`read_stata` function.
150150

151151
.. code-block:: python
152152
@@ -172,7 +172,7 @@ Similarly in pandas, the opposite of ``read_csv`` is :meth:`DataFrame.to_csv`.
172172
173173
tips.to_csv("tips2.csv")
174174
175-
Pandas can also export to Stata file format with the :meth:`DataFrame.to_stata` method.
175+
pandas can also export to Stata file format with the :meth:`DataFrame.to_stata` method.
176176

177177
.. code-block:: python
178178
@@ -583,7 +583,7 @@ should be used for comparisons.
583583
outer_join[pd.isna(outer_join["value_x"])]
584584
outer_join[pd.notna(outer_join["value_x"])]
585585
586-
Pandas also provides a variety of methods to work with missing data -- some of
586+
pandas also provides a variety of methods to work with missing data -- some of
587587
which would be challenging to express in Stata. For example, there are methods to
588588
drop all rows with any missing values, replacing missing values with a specified
589589
value, like the mean, or forward filling from previous rows. See the
@@ -674,7 +674,7 @@ Other considerations
674674
Disk vs memory
675675
~~~~~~~~~~~~~~
676676

677-
Pandas and Stata both operate exclusively in memory. This means that the size of
677+
pandas and Stata both operate exclusively in memory. This means that the size of
678678
data able to be loaded in pandas is limited by your machine's memory.
679679
If out of core processing is needed, one possibility is the
680680
`dask.dataframe <https://dask.pydata.org/en/latest/dataframe.html>`_

doc/source/getting_started/install.rst

+2-2
Original file line numberDiff line numberDiff line change
@@ -184,7 +184,7 @@ You can find simple installation instructions for pandas in this document: ``ins
184184
Installing from source
185185
~~~~~~~~~~~~~~~~~~~~~~
186186

187-
See the :ref:`contributing guide <contributing>` for complete instructions on building from the git source tree. Further, see :ref:`creating a development environment <contributing.dev_env>` if you wish to create a *pandas* development environment.
187+
See the :ref:`contributing guide <contributing>` for complete instructions on building from the git source tree. Further, see :ref:`creating a development environment <contributing.dev_env>` if you wish to create a pandas development environment.
188188

189189
Running the test suite
190190
----------------------
@@ -249,7 +249,7 @@ Recommended dependencies
249249
Optional dependencies
250250
~~~~~~~~~~~~~~~~~~~~~
251251

252-
Pandas has many optional dependencies that are only used for specific methods.
252+
pandas has many optional dependencies that are only used for specific methods.
253253
For example, :func:`pandas.read_hdf` requires the ``pytables`` package, while
254254
:meth:`DataFrame.to_markdown` requires the ``tabulate`` package. If the
255255
optional dependency is not installed, pandas will raise an ``ImportError`` when

doc/source/getting_started/overview.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66
Package overview
77
****************
88

9-
**pandas** is a `Python <https://www.python.org>`__ package providing fast,
9+
pandas is a `Python <https://www.python.org>`__ package providing fast,
1010
flexible, and expressive data structures designed to make working with
1111
"relational" or "labeled" data both easy and intuitive. It aims to be the
1212
fundamental high-level building block for doing practical, **real-world** data

doc/source/reference/arrays.rst

+7-7
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ For some data types, pandas extends NumPy's type system. String aliases for thes
1616
can be found at :ref:`basics.dtypes`.
1717

1818
=================== ========================= ================== =============================
19-
Kind of Data Pandas Data Type Scalar Array
19+
Kind of Data pandas Data Type Scalar Array
2020
=================== ========================= ================== =============================
2121
TZ-aware datetime :class:`DatetimeTZDtype` :class:`Timestamp` :ref:`api.arrays.datetime`
2222
Timedeltas (none) :class:`Timedelta` :ref:`api.arrays.timedelta`
@@ -29,7 +29,7 @@ Strings :class:`StringDtype` :class:`str` :ref:`api.array
2929
Boolean (with NA) :class:`BooleanDtype` :class:`bool` :ref:`api.arrays.bool`
3030
=================== ========================= ================== =============================
3131

32-
Pandas and third-party libraries can extend NumPy's type system (see :ref:`extending.extension-types`).
32+
pandas and third-party libraries can extend NumPy's type system (see :ref:`extending.extension-types`).
3333
The top-level :meth:`array` method can be used to create a new array, which may be
3434
stored in a :class:`Series`, :class:`Index`, or as a column in a :class:`DataFrame`.
3535

@@ -43,7 +43,7 @@ stored in a :class:`Series`, :class:`Index`, or as a column in a :class:`DataFra
4343
Datetime data
4444
-------------
4545

46-
NumPy cannot natively represent timezone-aware datetimes. Pandas supports this
46+
NumPy cannot natively represent timezone-aware datetimes. pandas supports this
4747
with the :class:`arrays.DatetimeArray` extension array, which can hold timezone-naive
4848
or timezone-aware values.
4949

@@ -162,7 +162,7 @@ If the data are tz-aware, then every value in the array must have the same timez
162162
Timedelta data
163163
--------------
164164

165-
NumPy can natively represent timedeltas. Pandas provides :class:`Timedelta`
165+
NumPy can natively represent timedeltas. pandas provides :class:`Timedelta`
166166
for symmetry with :class:`Timestamp`.
167167

168168
.. autosummary::
@@ -217,7 +217,7 @@ A collection of timedeltas may be stored in a :class:`TimedeltaArray`.
217217
Timespan data
218218
-------------
219219

220-
Pandas represents spans of times as :class:`Period` objects.
220+
pandas represents spans of times as :class:`Period` objects.
221221

222222
Period
223223
------
@@ -352,7 +352,7 @@ Nullable integer
352352
----------------
353353

354354
:class:`numpy.ndarray` cannot natively represent integer-data with missing values.
355-
Pandas provides this through :class:`arrays.IntegerArray`.
355+
pandas provides this through :class:`arrays.IntegerArray`.
356356

357357
.. autosummary::
358358
:toctree: api/
@@ -378,7 +378,7 @@ Pandas provides this through :class:`arrays.IntegerArray`.
378378
Categorical data
379379
----------------
380380

381-
Pandas defines a custom data type for representing data that can take only a
381+
pandas defines a custom data type for representing data that can take only a
382382
limited, fixed set of values. The dtype of a ``Categorical`` can be described by
383383
a :class:`pandas.api.types.CategoricalDtype`.
384384

doc/source/reference/series.rst

+1-1
Original file line numberDiff line numberDiff line change
@@ -280,7 +280,7 @@ Time Series-related
280280
Accessors
281281
---------
282282

283-
Pandas provides dtype-specific methods under various accessors.
283+
pandas provides dtype-specific methods under various accessors.
284284
These are separate namespaces within :class:`Series` that only apply
285285
to specific data types.
286286

0 commit comments

Comments
 (0)