Browse Source

DOC: pydata/pandas -> pandas-dev/pandas (#14409)

pull/14424/head
Jeff Reback 6 years ago committed by Joris Van den Bossche
parent
commit
7d40f18718
  1. 2
      asv_bench/benchmarks/attrs_caching.py
  2. 2
      asv_bench/benchmarks/ctors.py
  3. 2
      asv_bench/benchmarks/frame_ctor.py
  4. 2
      asv_bench/benchmarks/hdfstore_bench.py
  5. 2
      asv_bench/benchmarks/index_object.py
  6. 2
      asv_bench/benchmarks/io_sql.py
  7. 2
      asv_bench/benchmarks/panel_ctor.py
  8. 2
      asv_bench/benchmarks/panel_methods.py
  9. 2
      asv_bench/benchmarks/replace.py
  10. 2
      asv_bench/benchmarks/reshape.py
  11. 2
      asv_bench/benchmarks/stat_ops.py
  12. 2
      asv_bench/benchmarks/strings.py
  13. 4
      doc/README.rst
  14. 2
      doc/_templates/autosummary/accessor_attribute.rst
  15. 2
      doc/_templates/autosummary/accessor_method.rst
  16. 2
      doc/source/categorical.rst
  17. 4
      doc/source/comparison_with_sas.rst
  18. 2
      doc/source/comparison_with_sql.rst
  19. 8
      doc/source/conf.py
  20. 30
      doc/source/contributing.rst
  21. 14
      doc/source/cookbook.rst
  22. 2
      doc/source/ecosystem.rst
  23. 2
      doc/source/gotchas.rst
  24. 2
      doc/source/install.rst
  25. 2
      doc/source/io.rst
  26. 4
      doc/source/overview.rst
  27. 2
      doc/source/r_interface.rst
  28. 4
      doc/source/release.rst
  29. 2
      doc/source/remote_data.rst
  30. 4
      doc/source/visualization.rst
  31. 2
      pandas/compat/__init__.py
  32. 4
      pandas/computation/tests/test_eval.py
  33. 4
      pandas/core/categorical.py
  34. 2
      pandas/io/data.py
  35. 2
      pandas/io/gbq.py
  36. 2
      pandas/io/tests/json/test_pandas.py
  37. 2
      pandas/io/tests/parser/common.py
  38. 2
      pandas/io/tests/parser/test_network.py
  39. 2
      pandas/io/tests/test_excel.py
  40. 4
      pandas/io/tests/test_gbq.py
  41. 2
      pandas/io/tests/test_packers.py
  42. 2
      pandas/io/tests/test_sql.py
  43. 2
      pandas/io/wb.py
  44. 8
      pandas/tests/formats/test_style.py
  45. 2
      pandas/tests/indexing/test_categorical.py
  46. 2
      pandas/tests/plotting/test_boxplot_method.py
  47. 18
      pandas/tests/plotting/test_frame.py
  48. 2
      pandas/tests/series/test_datetime_values.py
  49. 2
      pandas/tests/series/test_operators.py
  50. 2
      pandas/tests/test_algos.py
  51. 16
      pandas/tests/test_categorical.py
  52. 2
      pandas/tests/test_config.py
  53. 2
      pandas/tests/test_groupby.py
  54. 4
      pandas/tests/test_strings.py
  55. 4
      pandas/tools/plotting.py
  56. 2
      pandas/tseries/resample.py
  57. 2
      pandas/tseries/tests/test_offsets.py
  58. 2
      pandas/tseries/tests/test_resample.py
  59. 6
      pandas/tseries/tests/test_timezones.py
  60. 2
      scripts/find_undoc_args.py
  61. 2
      scripts/gen_release_notes.py
  62. 2
      scripts/touchup_gh_issues.py
  63. 2
      vb_suite/perf_HEAD.py
  64. 4
      vb_suite/suite.py

2
asv_bench/benchmarks/attrs_caching.py

@ -20,4 +20,4 @@ class setattr_dataframe_index(object):
self.cur_index = self.df.index
def time_setattr_dataframe_index(self):
self.df.index = self.cur_index
self.df.index = self.cur_index

2
asv_bench/benchmarks/ctors.py

@ -49,4 +49,4 @@ class index_from_series_ctor(object):
self.s = Series(([Timestamp('20110101'), Timestamp('20120101'), Timestamp('20130101')] * 1000))
def time_index_from_series_ctor(self):
Index(self.s)
Index(self.s)

2
asv_bench/benchmarks/frame_ctor.py

@ -1703,4 +1703,4 @@ class series_ctor_from_dict(object):
self.dict_list = [dict(zip(self.columns, row)) for row in self.frame.values]
def time_series_ctor_from_dict(self):
Series(self.some_dict)
Series(self.some_dict)

2
asv_bench/benchmarks/hdfstore_bench.py

@ -348,4 +348,4 @@ class write_store_table_wide(object):
try:
os.remove(self.f)
except:
pass
pass

2
asv_bench/benchmarks/index_object.py

@ -344,4 +344,4 @@ class multiindex_with_datetime_level_sliced(object):
self.mi = MultiIndex.from_product([self.level1, self.level2])
def time_multiindex_with_datetime_level_sliced(self):
self.mi[:10].values
self.mi[:10].values

2
asv_bench/benchmarks/io_sql.py

@ -212,4 +212,4 @@ class sql_write_sqlalchemy(object):
self.df = DataFrame({'float1': randn(10000), 'float2': randn(10000), 'string1': (['foo'] * 10000), 'bool1': ([True] * 10000), 'int1': np.random.randint(0, 100000, size=10000), }, index=self.index)
def time_sql_write_sqlalchemy(self):
self.df.to_sql('test1', self.engine, if_exists='replace')
self.df.to_sql('test1', self.engine, if_exists='replace')

2
asv_bench/benchmarks/panel_ctor.py

@ -61,4 +61,4 @@ class panel_from_dict_two_different_indexes(object):
self.data_frames[x] = self.df
def time_panel_from_dict_two_different_indexes(self):
Panel.from_dict(self.data_frames)
Panel.from_dict(self.data_frames)

2
asv_bench/benchmarks/panel_methods.py

@ -53,4 +53,4 @@ class panel_shift_minor(object):
self.panel = Panel(np.random.randn(100, len(self.index), 1000))
def time_panel_shift_minor(self):
self.panel.shift(1, axis='minor')
self.panel.shift(1, axis='minor')

2
asv_bench/benchmarks/replace.py

@ -45,4 +45,4 @@ class replace_replacena(object):
self.ts = Series(np.random.randn(self.N), index=self.rng)
def time_replace_replacena(self):
self.ts.replace(np.nan, 0.0, inplace=True)
self.ts.replace(np.nan, 0.0, inplace=True)

2
asv_bench/benchmarks/reshape.py

@ -73,4 +73,4 @@ class unstack_sparse_keyspace(object):
break
def time_unstack_sparse_keyspace(self):
self.idf.unstack()
self.idf.unstack()

2
asv_bench/benchmarks/stat_ops.py

@ -258,4 +258,4 @@ class stats_rolling_mean(object):
rolling_skew(self.arr, self.win)
def time_rolling_kurt(self):
rolling_kurt(self.arr, self.win)
rolling_kurt(self.arr, self.win)

2
asv_bench/benchmarks/strings.py

@ -390,4 +390,4 @@ class strings_upper(object):
self.many.str.upper()
def make_series(self, letters, strlen, size):
return Series([str(x) for x in np.fromiter(IT.cycle(letters), count=(size * strlen), dtype='|S1').view('|S{}'.format(strlen))])
return Series([str(x) for x in np.fromiter(IT.cycle(letters), count=(size * strlen), dtype='|S1').view('|S{}'.format(strlen))])

4
doc/README.rst

@ -155,9 +155,9 @@ Where to start?
---------------
There are a number of issues listed under `Docs
<https://github.com/pydata/pandas/issues?labels=Docs&sort=updated&state=open>`_
<https://github.com/pandas-dev/pandas/issues?labels=Docs&sort=updated&state=open>`_
and `Good as first PR
<https://github.com/pydata/pandas/issues?labels=Good+as+first+PR&sort=updated&state=open>`_
<https://github.com/pandas-dev/pandas/issues?labels=Good+as+first+PR&sort=updated&state=open>`_
where you could start out.
Or maybe you have an idea of your own, by using pandas, looking for something

2
doc/_templates/autosummary/accessor_attribute.rst vendored

@ -3,4 +3,4 @@
.. currentmodule:: {{ module.split('.')[0] }}
.. autoaccessorattribute:: {{ [module.split('.')[1], objname]|join('.') }}
.. autoaccessorattribute:: {{ [module.split('.')[1], objname]|join('.') }}

2
doc/_templates/autosummary/accessor_method.rst vendored

@ -3,4 +3,4 @@
.. currentmodule:: {{ module.split('.')[0] }}
.. autoaccessormethod:: {{ [module.split('.')[1], objname]|join('.') }}
.. autoaccessormethod:: {{ [module.split('.')[1], objname]|join('.') }}

2
doc/source/categorical.rst

@ -973,7 +973,7 @@ are not numeric data (even in the case that ``.categories`` is numeric).
print("TypeError: " + str(e))
.. note::
If such a function works, please file a bug at https://github.com/pydata/pandas!
If such a function works, please file a bug at https://github.com/pandas-dev/pandas!
dtype in apply
~~~~~~~~~~~~~~

4
doc/source/comparison_with_sas.rst

@ -116,7 +116,7 @@ Reading External Data
Like SAS, pandas provides utilities for reading in data from
many formats. The ``tips`` dataset, found within the pandas
tests (`csv <https://raw.github.com/pydata/pandas/master/pandas/tests/data/tips.csv>`_)
tests (`csv <https://raw.github.com/pandas-dev/pandas/master/pandas/tests/data/tips.csv>`_)
will be used in many of the following examples.
SAS provides ``PROC IMPORT`` to read csv data into a data set.
@ -131,7 +131,7 @@ The pandas method is :func:`read_csv`, which works similarly.
.. ipython:: python
url = 'https://raw.github.com/pydata/pandas/master/pandas/tests/data/tips.csv'
url = 'https://raw.github.com/pandas-dev/pandas/master/pandas/tests/data/tips.csv'
tips = pd.read_csv(url)
tips.head()

2
doc/source/comparison_with_sql.rst

@ -23,7 +23,7 @@ structure.
.. ipython:: python
url = 'https://raw.github.com/pydata/pandas/master/pandas/tests/data/tips.csv'
url = 'https://raw.github.com/pandas-dev/pandas/master/pandas/tests/data/tips.csv'
tips = pd.read_csv(url)
tips.head()

8
doc/source/conf.py

@ -301,9 +301,9 @@ import glob
autosummary_generate = glob.glob("*.rst")
# extlinks alias
extlinks = {'issue': ('https://github.com/pydata/pandas/issues/%s',
extlinks = {'issue': ('https://github.com/pandas-dev/pandas/issues/%s',
'GH'),
'wiki': ('https://github.com/pydata/pandas/wiki/%s',
'wiki': ('https://github.com/pandas-dev/pandas/wiki/%s',
'wiki ')}
ipython_exec_lines = [
@ -468,10 +468,10 @@ def linkcode_resolve(domain, info):
fn = os.path.relpath(fn, start=os.path.dirname(pandas.__file__))
if '+' in pandas.__version__:
return "http://github.com/pydata/pandas/blob/master/pandas/%s%s" % (
return "http://github.com/pandas-dev/pandas/blob/master/pandas/%s%s" % (
fn, linespec)
else:
return "http://github.com/pydata/pandas/blob/v%s/pandas/%s%s" % (
return "http://github.com/pandas-dev/pandas/blob/v%s/pandas/%s%s" % (
pandas.__version__, fn, linespec)

30
doc/source/contributing.rst

@ -14,11 +14,11 @@ All contributions, bug reports, bug fixes, documentation improvements,
enhancements and ideas are welcome.
If you are simply looking to start working with the *pandas* codebase, navigate to the
`GitHub "issues" tab <https://github.com/pydata/pandas/issues>`_ and start looking through
`GitHub "issues" tab <https://github.com/pandas-dev/pandas/issues>`_ and start looking through
interesting issues. There are a number of issues listed under `Docs
<https://github.com/pydata/pandas/issues?labels=Docs&sort=updated&state=open>`_
<https://github.com/pandas-dev/pandas/issues?labels=Docs&sort=updated&state=open>`_
and `Difficulty Novice
<https://github.com/pydata/pandas/issues?q=is%3Aopen+is%3Aissue+label%3A%22Difficulty+Novice%22>`_
<https://github.com/pandas-dev/pandas/issues?q=is%3Aopen+is%3Aissue+label%3A%22Difficulty+Novice%22>`_
where you could start out.
Or maybe through using *pandas* you have an idea of your own or are looking for something
@ -27,7 +27,7 @@ about it!
Feel free to ask questions on the `mailing list
<https://groups.google.com/forum/?fromgroups#!forum/pydata>`_ or on `Gitter
<https://gitter.im/pydata/pandas>`_.
<https://gitter.im/pandas-dev/pandas>`_.
Bug reports and enhancement requests
====================================
@ -79,7 +79,7 @@ It can very quickly become overwhelming, but sticking to the guidelines below wi
straightforward and mostly trouble free. As always, if you are having difficulties please
feel free to ask for help.
The code is hosted on `GitHub <https://www.github.com/pydata/pandas>`_. To
The code is hosted on `GitHub <https://www.github.com/pandas-dev/pandas>`_. To
contribute you will need to sign up for a `free GitHub account
<https://github.com/signup/free>`_. We use `Git <http://git-scm.com/>`_ for
version control to allow many people to work together on the project.
@ -103,12 +103,12 @@ Forking
-------
You will need your own fork to work on the code. Go to the `pandas project
page <https://github.com/pydata/pandas>`_ and hit the ``Fork`` button. You will
page <https://github.com/pandas-dev/pandas>`_ and hit the ``Fork`` button. You will
want to clone your fork to your machine::
git clone git@github.com:your-user-name/pandas.git pandas-yourname
cd pandas-yourname
git remote add upstream git://github.com/pydata/pandas.git
git remote add upstream git://github.com/pandas-dev/pandas.git
This creates the directory `pandas-yourname` and connects your repository to
the upstream (main project) *pandas* repository.
@ -467,7 +467,7 @@ and make these changes with::
pep8radius master --diff --in-place
Additional standards are outlined on the `code style wiki
page <https://github.com/pydata/pandas/wiki/Code-Style-and-Conventions>`_.
page <https://github.com/pandas-dev/pandas/wiki/Code-Style-and-Conventions>`_.
Please try to maintain backward compatibility. *pandas* has lots of users with lots of
existing code, so don't break it if at all possible. If you think breakage is required,
@ -501,7 +501,7 @@ All tests should go into the ``tests`` subdirectory of the specific package.
This folder contains many current examples of tests, and we suggest looking to these for
inspiration. If your test requires working with files or
network connectivity, there is more information on the `testing page
<https://github.com/pydata/pandas/wiki/Testing>`_ of the wiki.
<https://github.com/pandas-dev/pandas/wiki/Testing>`_ of the wiki.
The ``pandas.util.testing`` module has many special ``assert`` functions that
make it easier to make statements about whether Series or DataFrame objects are
@ -639,7 +639,7 @@ on Travis-CI. The first step is to create a `service account
Integration tests for ``pandas.io.gbq`` are skipped in pull requests because
the credentials that are required for running Google BigQuery integration
tests are `encrypted <https://docs.travis-ci.com/user/encrypting-files/>`__
on Travis-CI and are only accessible from the pydata/pandas repository. The
on Travis-CI and are only accessible from the pandas-dev/pandas repository. The
credentials won't be available on forks of pandas. Here are the steps to run
gbq integration tests on a forked repository:
@ -688,7 +688,7 @@ performance regressions.
You can run specific benchmarks using the ``-r`` flag, which takes a regular expression.
See the `performance testing wiki <https://github.com/pydata/pandas/wiki/Performance-Testing>`_ for information
See the `performance testing wiki <https://github.com/pandas-dev/pandas/wiki/Performance-Testing>`_ for information
on how to write a benchmark.
Documenting your code
@ -712,8 +712,8 @@ directive is used. The sphinx syntax for that is:
This will put the text *New in version 0.17.0* wherever you put the sphinx
directive. This should also be put in the docstring when adding a new function
or method (`example <https://github.com/pydata/pandas/blob/v0.16.2/pandas/core/generic.py#L1959>`__)
or a new keyword argument (`example <https://github.com/pydata/pandas/blob/v0.16.2/pandas/core/frame.py#L1171>`__).
or method (`example <https://github.com/pandas-dev/pandas/blob/v0.16.2/pandas/core/generic.py#L1959>`__)
or a new keyword argument (`example <https://github.com/pandas-dev/pandas/blob/v0.16.2/pandas/core/frame.py#L1171>`__).
Contributing your changes to *pandas*
=====================================
@ -806,8 +806,8 @@ like::
origin git@github.com:yourname/pandas.git (fetch)
origin git@github.com:yourname/pandas.git (push)
upstream git://github.com/pydata/pandas.git (fetch)
upstream git://github.com/pydata/pandas.git (push)
upstream git://github.com/pandas-dev/pandas.git (fetch)
upstream git://github.com/pandas-dev/pandas.git (push)
Now your code is on GitHub, but it is not yet a part of the *pandas* project. For that to
happen, a pull request needs to be submitted on GitHub.

14
doc/source/cookbook.rst

@ -200,7 +200,7 @@ The :ref:`indexing <indexing>` docs.
df[(df.AAA <= 6) & (df.index.isin([0,2,4]))]
`Use loc for label-oriented slicing and iloc positional slicing
<https://github.com/pydata/pandas/issues/2904>`__
<https://github.com/pandas-dev/pandas/issues/2904>`__
.. ipython:: python
@ -410,7 +410,7 @@ Sorting
df.sort_values(by=('Labs', 'II'), ascending=False)
`Partial Selection, the need for sortedness;
<https://github.com/pydata/pandas/issues/2995>`__
<https://github.com/pandas-dev/pandas/issues/2995>`__
Levels
******
@ -787,7 +787,7 @@ The :ref:`Resample <timeseries.resampling>` docs.
<http://stackoverflow.com/questions/14569223/timegrouper-pandas>`__
`Using TimeGrouper and another grouping to create subgroups, then apply a custom function
<https://github.com/pydata/pandas/issues/3791>`__
<https://github.com/pandas-dev/pandas/issues/3791>`__
`Resampling with custom periods
<http://stackoverflow.com/questions/15408156/resampling-with-custom-periods>`__
@ -823,7 +823,7 @@ ignore_index is needed in pandas < v0.13, and depending on df construction
df = df1.append(df2,ignore_index=True); df
`Self Join of a DataFrame
<https://github.com/pydata/pandas/issues/2996>`__
<https://github.com/pandas-dev/pandas/issues/2996>`__
.. ipython:: python
@ -936,7 +936,7 @@ using that handle to read.
<http://stackoverflow.com/questions/15555005/get-inferred-dataframe-types-iteratively-using-chunksize>`__
`Dealing with bad lines
<http://github.com/pydata/pandas/issues/2886>`__
<http://github.com/pandas-dev/pandas/issues/2886>`__
`Dealing with bad lines II
<http://nipunbatra.github.io/2013/06/reading-unclean-data-csv-using-pandas/>`__
@ -1075,7 +1075,7 @@ The :ref:`HDFStores <io.hdf5>` docs
<http://stackoverflow.com/questions/13926089/selecting-columns-from-pandas-hdfstore-table>`__
`Managing heterogeneous data using a linked multiple table hierarchy
<http://github.com/pydata/pandas/issues/3032>`__
<http://github.com/pandas-dev/pandas/issues/3032>`__
`Merging on-disk tables with millions of rows
<http://stackoverflow.com/questions/14614512/merging-two-tables-with-millions-of-rows-in-python/14617925#14617925>`__
@ -1216,7 +1216,7 @@ Timedeltas
The :ref:`Timedeltas <timedeltas.timedeltas>` docs.
`Using timedeltas
<http://github.com/pydata/pandas/pull/2899>`__
<http://github.com/pandas-dev/pandas/pull/2899>`__
.. ipython:: python

2
doc/source/ecosystem.rst

@ -143,7 +143,7 @@ both "column wise min/max and global min/max coloring."
API
-----
`pandas-datareader <https://github.com/pydata/pandas-datareader>`__
`pandas-datareader <https://github.com/pandas-dev/pandas-datareader>`__
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
``pandas-datareader`` is a remote data access library for pandas. ``pandas.io`` from pandas < 0.17.0 is now refactored/split-off to and importable from ``pandas_datareader`` (PyPI:``pandas-datareader``). Many/most of the supported APIs have at least a documentation paragraph in the `pandas-datareader docs <https://pandas-datareader.readthedocs.io/en/latest/>`_:

2
doc/source/gotchas.rst

@ -391,7 +391,7 @@ This is because ``reindex_like`` silently inserts ``NaNs`` and the ``dtype``
changes accordingly. This can cause some issues when using ``numpy`` ``ufuncs``
such as ``numpy.logical_and``.
See the `this old issue <https://github.com/pydata/pandas/issues/2388>`__ for a more
See the `this old issue <https://github.com/pandas-dev/pandas/issues/2388>`__ for a more
detailed discussion.
Parsing Dates from Text Files

2
doc/source/install.rst

@ -13,7 +13,7 @@ This is the recommended installation method for most users.
Instructions for installing from source,
`PyPI <http://pypi.python.org/pypi/pandas>`__, various Linux distributions, or a
`development version <http://github.com/pydata/pandas>`__ are also provided.
`development version <http://github.com/pandas-dev/pandas>`__ are also provided.
Python version support
----------------------

2
doc/source/io.rst

@ -2035,7 +2035,7 @@ You can even pass in an instance of ``StringIO`` if you so desire
that having so many network-accessing functions slows down the documentation
build. If you spot an error or an example that doesn't run, please do not
hesitate to report it over on `pandas GitHub issues page
<http://www.github.com/pydata/pandas/issues>`__.
<http://www.github.com/pandas-dev/pandas/issues>`__.
Read a URL and match a table that contains specific text

4
doc/source/overview.rst

@ -81,7 +81,7 @@ Getting Support
---------------
The first stop for pandas issues and ideas is the `Github Issue Tracker
<https://github.com/pydata/pandas/issues>`__. If you have a general question,
<https://github.com/pandas-dev/pandas/issues>`__. If you have a general question,
pandas community experts can answer through `Stack Overflow
<http://stackoverflow.com/questions/tagged/pandas>`__.
@ -103,7 +103,7 @@ training, and consulting for pandas.
pandas is only made possible by a group of people around the world like you
who have contributed new code, bug reports, fixes, comments and ideas. A
complete list can be found `on Github <http://www.github.com/pydata/pandas/contributors>`__.
complete list can be found `on Github <http://www.github.com/pandas-dev/pandas/contributors>`__.
Development Team
----------------

2
doc/source/r_interface.rst

@ -71,7 +71,7 @@ The ``convert_to_r_matrix`` function can be replaced by the normal
Not all conversion functions in rpy2 are working exactly the same as the
current methods in pandas. If you experience problems or limitations in
comparison to the ones in pandas, please report this at the
`issue tracker <https://github.com/pydata/pandas/issues>`_.
`issue tracker <https://github.com/pandas-dev/pandas/issues>`_.
See also the documentation of the `rpy2 <http://rpy2.bitbucket.org/>`__ project.

4
doc/source/release.rst

@ -20,7 +20,7 @@ Release Notes
*************
This is the list of changes to pandas between each release. For full details,
see the commit logs at http://github.com/pydata/pandas
see the commit logs at http://github.com/pandas-dev/pandas
**What is it**
@ -33,7 +33,7 @@ analysis / manipulation tool available in any language.
**Where to get it**
* Source code: http://github.com/pydata/pandas
* Source code: http://github.com/pandas-dev/pandas
* Binary installers on PyPI: http://pypi.python.org/pypi/pandas
* Documentation: http://pandas.pydata.org

2
doc/source/remote_data.rst

@ -13,7 +13,7 @@ DataReader
The sub-package ``pandas.io.data`` is removed in favor of a separately
installable `pandas-datareader package
<https://github.com/pydata/pandas-datareader>`_. This will allow the data
<https://github.com/pandas-dev/pandas-datareader>`_. This will allow the data
modules to be independently updated to your pandas installation. The API for
``pandas-datareader v0.1.1`` is the same as in ``pandas v0.16.1``.
(:issue:`8961`)

4
doc/source/visualization.rst

@ -892,7 +892,7 @@ for Fourier series. By coloring these curves differently for each class
it is possible to visualize data clustering. Curves belonging to samples
of the same class will usually be closer together and form larger structures.
**Note**: The "Iris" dataset is available `here <https://raw.github.com/pydata/pandas/master/pandas/tests/data/iris.csv>`__.
**Note**: The "Iris" dataset is available `here <https://raw.github.com/pandas-dev/pandas/master/pandas/tests/data/iris.csv>`__.
.. ipython:: python
@ -1044,7 +1044,7 @@ forces acting on our sample are at an equilibrium) is where a dot representing
our sample will be drawn. Depending on which class that sample belongs it will
be colored differently.
**Note**: The "Iris" dataset is available `here <https://raw.github.com/pydata/pandas/master/pandas/tests/data/iris.csv>`__.
**Note**: The "Iris" dataset is available `here <https://raw.github.com/pandas-dev/pandas/master/pandas/tests/data/iris.csv>`__.
.. ipython:: python

2
pandas/compat/__init__.py

@ -392,7 +392,7 @@ class OrderedDefaultdict(OrderedDict):
return type(self), args, None, None, list(self.items())
# https://github.com/pydata/pandas/pull/9123
# https://github.com/pandas-dev/pandas/pull/9123
def is_platform_little_endian():
""" am I little endian """
return sys.byteorder == 'little'

4
pandas/computation/tests/test_eval.py

@ -1693,11 +1693,11 @@ class TestMathPythonPython(tm.TestCase):
self.check_result_type(np.float64, np.float64)
def test_result_types2(self):
# xref https://github.com/pydata/pandas/issues/12293
# xref https://github.com/pandas-dev/pandas/issues/12293
raise nose.SkipTest("unreliable tests on complex128")
# Did not test complex64 because DataFrame is converting it to
# complex128. Due to https://github.com/pydata/pandas/issues/10952
# complex128. Due to https://github.com/pandas-dev/pandas/issues/10952
self.check_result_type(np.complex128, np.complex128)
def test_undefined_func(self):

4
pandas/core/categorical.py

@ -1681,7 +1681,7 @@ class Categorical(PandasObject):
else:
# There is a bug in numpy, which does not accept a Series as a
# indexer
# https://github.com/pydata/pandas/issues/6168
# https://github.com/pandas-dev/pandas/issues/6168
# https://github.com/numpy/numpy/issues/4240 -> fixed in numpy 1.9
# FIXME: remove when numpy 1.9 is the lowest numpy version pandas
# accepts...
@ -1690,7 +1690,7 @@ class Categorical(PandasObject):
lindexer = self.categories.get_indexer(rvalue)
# FIXME: the following can be removed after GH7820 is fixed:
# https://github.com/pydata/pandas/issues/7820
# https://github.com/pandas-dev/pandas/issues/7820
# float categories do currently return -1 for np.nan, even if np.nan is
# included in the index -> "repair" this here
if isnull(rvalue).any() and isnull(self.categories).any():

2
pandas/io/data.py

@ -1,6 +1,6 @@
raise ImportError(
"The pandas.io.data module is moved to a separate package "
"(pandas-datareader). After installing the pandas-datareader package "
"(https://github.com/pydata/pandas-datareader), you can change "
"(https://github.com/pandas-dev/pandas-datareader), you can change "
"the import ``from pandas.io import data, wb`` to "
"``from pandas_datareader import data, wb``.")

2
pandas/io/gbq.py

@ -236,7 +236,7 @@ class GbqConnector(object):
return credentials
def get_service_account_credentials(self):
# Bug fix for https://github.com/pydata/pandas/issues/12572
# Bug fix for https://github.com/pandas-dev/pandas/issues/12572
# We need to know that a supported version of oauth2client is installed
# Test that either of the following is installed:
# - SignedJwtAssertionCredentials from oauth2client.client

2
pandas/io/tests/json/test_pandas.py

@ -767,7 +767,7 @@ DataFrame\\.index values are different \\(100\\.0 %\\)
@network
def test_url(self):
url = 'https://api.github.com/repos/pydata/pandas/issues?per_page=5'
url = 'https://api.github.com/repos/pandas-dev/pandas/issues?per_page=5' # noqa
result = read_json(url, convert_dates=True)
for c in ['created_at', 'closed_at', 'updated_at']:
self.assertEqual(result[c].dtype, 'datetime64[ns]')

2
pandas/io/tests/parser/common.py

@ -629,7 +629,7 @@ bar"""
@tm.network
def test_url(self):
# HTTP(S)
url = ('https://raw.github.com/pydata/pandas/master/'
url = ('https://raw.github.com/pandas-dev/pandas/master/'
'pandas/io/tests/parser/data/salary.table.csv')
url_table = self.read_table(url)
dirpath = tm.get_data_path()

2
pandas/io/tests/parser/test_network.py

@ -23,7 +23,7 @@ class TestUrlGz(tm.TestCase):
@tm.network
def test_url_gz(self):
url = ('https://raw.github.com/pydata/pandas/'
url = ('https://raw.github.com/pandas-dev/pandas/'
'master/pandas/io/tests/parser/data/salary.table.gz')
url_table = read_table(url, compression="gzip", engine="python")
tm.assert_frame_equal(url_table, self.local_table)

2
pandas/io/tests/test_excel.py

@ -543,7 +543,7 @@ class XlrdTests(ReadingTestsBase):
@tm.network
def test_read_from_http_url(self):
url = ('https://raw.github.com/pydata/pandas/master/'
url = ('https://raw.github.com/pandas-dev/pandas/master/'
'pandas/io/tests/data/test1' + self.ext)
url_table = read_excel(url)
local_table = self.get_exceldf('test1')

4
pandas/io/tests/test_gbq.py

@ -150,7 +150,7 @@ def _test_imports():
raise ImportError(
"pandas requires httplib2 for Google BigQuery support")
# Bug fix for https://github.com/pydata/pandas/issues/12572
# Bug fix for https://github.com/pandas-dev/pandas/issues/12572
# We need to know that a supported version of oauth2client is installed
# Test that either of the following is installed:
# - SignedJwtAssertionCredentials from oauth2client.client
@ -651,7 +651,7 @@ class TestReadGBQIntegration(tm.TestCase):
self.assertEqual(len(df.drop_duplicates()), test_size)
def test_zero_rows(self):
# Bug fix for https://github.com/pydata/pandas/issues/10273
# Bug fix for https://github.com/pandas-dev/pandas/issues/10273
df = gbq.read_gbq("SELECT title, id "
"FROM [publicdata:samples.wikipedia] "
"WHERE timestamp=-9999999",

2
pandas/io/tests/test_packers.py

@ -544,7 +544,7 @@ class TestSparse(TestPackers):
class TestCompression(TestPackers):
"""See https://github.com/pydata/pandas/pull/9783
"""See https://github.com/pandas-dev/pandas/pull/9783
"""
def setUp(self):

2
pandas/io/tests/test_sql.py

@ -1610,7 +1610,7 @@ class _TestSQLAlchemy(SQLAlchemyMixIn, PandasSQLTest):
def test_connectable_issue_example(self):
# This tests the example raised in issue
# https://github.com/pydata/pandas/issues/10104
# https://github.com/pandas-dev/pandas/issues/10104
def foo(connection):
query = 'SELECT test_foo_data FROM test_foo_data'

2
pandas/io/wb.py

@ -1,6 +1,6 @@
raise ImportError(
"The pandas.io.wb module is moved to a separate package "
"(pandas-datareader). After installing the pandas-datareader package "
"(https://github.com/pydata/pandas-datareader), you can change "
"(https://github.com/pandas-dev/pandas-datareader), you can change "
"the import ``from pandas.io import data, wb`` to "
"``from pandas_datareader import data, wb``.")

8
pandas/tests/formats/test_style.py

@ -144,7 +144,7 @@ class TestStyler(TestCase):
self.assertEqual(result, expected)
def test_empty_index_name_doesnt_display(self):
# https://github.com/pydata/pandas/pull/12090#issuecomment-180695902
# https://github.com/pandas-dev/pandas/pull/12090#issuecomment-180695902
df = pd.DataFrame({'A': [1, 2], 'B': [3, 4], 'C': [5, 6]})
result = df.style._translate()
@ -175,7 +175,7 @@ class TestStyler(TestCase):
self.assertEqual(result['head'], expected)
def test_index_name(self):
# https://github.com/pydata/pandas/issues/11655
# https://github.com/pandas-dev/pandas/issues/11655
df = pd.DataFrame({'A': [1, 2], 'B': [3, 4], 'C': [5, 6]})
result = df.set_index('A').style._translate()
@ -195,7 +195,7 @@ class TestStyler(TestCase):
self.assertEqual(result['head'], expected)
def test_multiindex_name(self):
# https://github.com/pydata/pandas/issues/11655
# https://github.com/pandas-dev/pandas/issues/11655
df = pd.DataFrame({'A': [1, 2], 'B': [3, 4], 'C': [5, 6]})
result = df.set_index(['A', 'B']).style._translate()
@ -217,7 +217,7 @@ class TestStyler(TestCase):
self.assertEqual(result['head'], expected)
def test_numeric_columns(self):
# https://github.com/pydata/pandas/issues/12125
# https://github.com/pandas-dev/pandas/issues/12125
# smoke test for _translate
df = pd.DataFrame({0: [1, 2, 3]})
df.style._translate()

2
pandas/tests/indexing/test_categorical.py

@ -392,7 +392,7 @@ class TestCategoricalIndex(tm.TestCase):
def test_indexing_with_category(self):
# https://github.com/pydata/pandas/issues/12564
# https://github.com/pandas-dev/pandas/issues/12564
# consistent result if comparing as Dataframe
cat = DataFrame({'A': ['foo', 'bar', 'baz']})

2
pandas/tests/plotting/test_boxplot_method.py

@ -100,7 +100,7 @@ class TestDataFramePlots(TestPlotBase):
@slow
def test_boxplot_return_type_legacy(self):
# API change in https://github.com/pydata/pandas/pull/7096
# API change in https://github.com/pandas-dev/pandas/pull/7096
import matplotlib as mpl # noqa
df = DataFrame(randn(6, 4),

18
pandas/tests/plotting/test_frame.py

@ -84,7 +84,7 @@ class TestDataFramePlots(TestPlotBase):
# We have to redo it here because _check_plot_works does two plots,
# once without an ax kwarg and once with an ax kwarg and the new sharex
# behaviour does not remove the visibility of the latter axis (as ax is
# present). see: https://github.com/pydata/pandas/issues/9737
# present). see: https://github.com/pandas-dev/pandas/issues/9737
axes = df.plot(subplots=True, title='blah')
self._check_axes_shape(axes, axes_num=3, layout=(3, 1))
@ -927,7 +927,7 @@ class TestDataFramePlots(TestPlotBase):
# Ensure that we can pass an np.array straight through to matplotlib,
# this functionality was accidentally removed previously.
# See https://github.com/pydata/pandas/issues/8852 for bug report
# See https://github.com/pandas-dev/pandas/issues/8852 for bug report
#
# Exercise colormap path and non-colormap path as they are independent
#
@ -2115,7 +2115,7 @@ class TestDataFramePlots(TestPlotBase):
self.assertEqual(result, expected)
# legend labels
# NaN's not included in legend with subplots
# see https://github.com/pydata/pandas/issues/8390
# see https://github.com/pandas-dev/pandas/issues/8390
self.assertEqual([x.get_text() for x in
ax.get_legend().get_texts()],
base_expected[:i] + base_expected[i + 1:])
@ -2336,9 +2336,9 @@ class TestDataFramePlots(TestPlotBase):
@slow
def test_sharex_and_ax(self):
# https://github.com/pydata/pandas/issues/9737 using gridspec, the axis
# in fig.get_axis() are sorted differently than pandas expected them,
# so make sure that only the right ones are removed
# https://github.com/pandas-dev/pandas/issues/9737 using gridspec,
# the axis in fig.get_axis() are sorted differently than pandas
# expected them, so make sure that only the right ones are removed
import matplotlib.pyplot as plt
plt.close('all')
gs, axes = _generate_4_axes_via_gridspec()
@ -2388,9 +2388,9 @@ class TestDataFramePlots(TestPlotBase):
@slow
def test_sharey_and_ax(self):
# https://github.com/pydata/pandas/issues/9737 using gridspec, the axis
# in fig.get_axis() are sorted differently than pandas expected them,
# so make sure that only the right ones are removed
# https://github.com/pandas-dev/pandas/issues/9737 using gridspec,
# the axis in fig.get_axis() are sorted differently than pandas
# expected them, so make sure that only the right ones are removed
import matplotlib.pyplot as plt
gs, axes = _generate_4_axes_via_gridspec()

2
pandas/tests/series/test_datetime_values.py

@ -273,7 +273,7 @@ class TestSeriesDatetimeValues(TestData, tm.TestCase):
self.assertRaises(com.SettingWithCopyError, f)
def test_dt_accessor_no_new_attributes(self):
# https://github.com/pydata/pandas/issues/10673
# https://github.com/pandas-dev/pandas/issues/10673
s = Series(date_range('20130101', periods=5, freq='D'))
with tm.assertRaisesRegexp(AttributeError,
"You cannot add any new attribute"):

2
pandas/tests/series/test_operators.py

@ -1412,7 +1412,7 @@ class TestSeriesOperators(TestData, tm.TestCase):
# NotImplemented
# this is an alignment issue; these are equivalent
# https://github.com/pydata/pandas/issues/5284
# https://github.com/pandas-dev/pandas/issues/5284
self.assertRaises(ValueError, lambda: d.__and__(s, axis='columns'))
self.assertRaises(ValueError, tester, s, d)

2
pandas/tests/test_algos.py

@ -595,7 +595,7 @@ class TestValueCounts(tm.TestCase):
tm.assert_series_equal(result, expected, check_index_type=True)
def test_dropna(self):
# https://github.com/pydata/pandas/issues/9443#issuecomment-73719328
# https://github.com/pandas-dev/pandas/issues/9443#issuecomment-73719328
tm.assert_series_equal(
pd.Series([True, True, False]).value_counts(dropna=True),

16
pandas/tests/test_categorical.py

@ -191,7 +191,7 @@ class TestCategorical(tm.TestCase):
cat = pd.Categorical([1, 2, 3, np.nan], categories=[1, 2, 3])
self.assertTrue(is_integer_dtype(cat.categories))
# https://github.com/pydata/pandas/issues/3678
# https://github.com/pandas-dev/pandas/issues/3678
cat = pd.Categorical([np.nan, 1, 2, 3])
self.assertTrue(is_integer_dtype(cat.categories))
@ -618,7 +618,7 @@ class TestCategorical(tm.TestCase):
index=exp_index)
tm.assert_frame_equal(desc, expected)
# https://github.com/pydata/pandas/issues/3678
# https://github.com/pandas-dev/pandas/issues/3678
# describe should work with NaN
cat = pd.Categorical([np.nan, 1, 2, 2])
desc = cat.describe()
@ -1547,7 +1547,7 @@ Categories (3, object): [ああああ, いいいいい, ううううううう]""
self.assertTrue(abs(diff) < 100)
def test_searchsorted(self):
# https://github.com/pydata/pandas/issues/8420
# https://github.com/pandas-dev/pandas/issues/8420
s1 = pd.Series(['apple', 'bread', 'bread', 'cheese', 'milk'])
s2 = pd.Series(['apple', 'bread', 'bread', 'cheese', 'milk', 'donuts'])
c1 = pd.Categorical(s1, ordered=True)
@ -1633,7 +1633,7 @@ Categories (3, object): [ああああ, いいいいい, ううううううう]""
np.array([False, True, True]))
def test_comparison_with_unknown_scalars(self):
# https://github.com/pydata/pandas/issues/9836#issuecomment-92123057
# https://github.com/pandas-dev/pandas/issues/9836#issuecomment-92123057
# and following comparisons with scalars not in categories should raise
# for unequal comps, but not for equal/not equal
cat = pd.Categorical([1, 2, 3], ordered=True)
@ -3829,7 +3829,7 @@ Categories (10, timedelta64[ns]): [0 days 01:00:00 < 1 days 01:00:00 < 2 days 01
self.assertRaises(TypeError, f)
# https://github.com/pydata/pandas/issues/9836#issuecomment-92123057
# https://github.com/pandas-dev/pandas/issues/9836#issuecomment-92123057
# and following comparisons with scalars not in categories should raise
# for unequal comps, but not for equal/not equal
cat = Series(Categorical(list("abc"), ordered=True))
@ -4303,14 +4303,14 @@ Categories (10, timedelta64[ns]): [0 days 01:00:00 < 1 days 01:00:00 < 2 days 01
self.assertFalse(hasattr(invalid, 'cat'))
def test_cat_accessor_no_new_attributes(self):
# https://github.com/pydata/pandas/issues/10673
# https://github.com/pandas-dev/pandas/issues/10673
c = Series(list('aabbcde')).astype('category')
with tm.assertRaisesRegexp(AttributeError,
"You cannot add any new attribute"):
c.cat.xlabel = "a"
def test_str_accessor_api_for_categorical(self):
# https://github.com/pydata/pandas/issues/10661
# https://github.com/pandas-dev/pandas/issues/10661
from pandas.core.strings import StringMethods
s = Series(list('aabb'))
s = s + " " + s
@ -4385,7 +4385,7 @@ Categories (10, timedelta64[ns]): [0 days 01:00:00 < 1 days 01:00:00 < 2 days 01
self.assertFalse(hasattr(invalid, 'str'))
def test_dt_accessor_api_for_categorical(self):
# https://github.com/pydata/pandas/issues/10661
# https://github.com/pandas-dev/pandas/issues/10661
from pandas.tseries.common import Properties
from pandas.tseries.index import date_range, DatetimeIndex
from pandas.tseries.period import period_range, PeriodIndex

2
pandas/tests/test_config.py

@ -427,7 +427,7 @@ class TestConfig(unittest.TestCase):
def test_option_context_scope(self):
# Ensure that creating a context does not affect the existing
# environment as it is supposed to be used with the `with` statement.
# See https://github.com/pydata/pandas/issues/8514
# See https://github.com/pandas-dev/pandas/issues/8514
original_value = 60
context_value = 10

2
pandas/tests/test_groupby.py

@ -6443,7 +6443,7 @@ class TestGroupBy(tm.TestCase):
def test_groupby_categorical_two_columns(self):
# https://github.com/pydata/pandas/issues/8138
# https://github.com/pandas-dev/pandas/issues/8138
d = {'cat':
pd.Categorical(["a", "b", "a", "b"], categories=["a", "b", "c"],
ordered=True),

4
pandas/tests/test_strings.py

@ -2604,7 +2604,7 @@ class TestStringMethods(tm.TestCase):
self.assertEqual(str_multiple.loc[1], '2011 2 2')
def test_str_cat_raises_intuitive_error(self):
# https://github.com/pydata/pandas/issues/11334
# https://github.com/pandas-dev/pandas/issues/11334
s = Series(['a', 'b', 'c', 'd'])
message = "Did you mean to supply a `sep` keyword?"
with tm.assertRaisesRegexp(ValueError, message):
@ -2661,7 +2661,7 @@ class TestStringMethods(tm.TestCase):
idx.str
def test_str_accessor_no_new_attributes(self):
# https://github.com/pydata/pandas/issues/10673
# https://github.com/pandas-dev/pandas/issues/10673
s = Series(list('aabbcde'))
with tm.assertRaisesRegexp(AttributeError,
"You cannot add any new attribute"):

4
pandas/tools/plotting.py

@ -722,7 +722,7 @@ def parallel_coordinates(frame, class_column, cols=None, ax=None, color=None,
>>> from pandas import read_csv
>>> from pandas.tools.plotting import parallel_coordinates
>>> from matplotlib import pyplot as plt
>>> df = read_csv('https://raw.github.com/pydata/pandas/master'
>>> df = read_csv('https://raw.github.com/pandas-dev/pandas/master'
'/pandas/tests/data/iris.csv')
>>> parallel_coordinates(df, 'Name', color=('#556270',
'#4ECDC4', '#C7F464'))
@ -2773,7 +2773,7 @@ def boxplot(data, column=None, by=None, ax=None, fontsize=None,
if by is not None:
# Prefer array return type for 2-D plots to match the subplot layout
# https://github.com/pydata/pandas/pull/12216#issuecomment-241175580
# https://github.com/pandas-dev/pandas/pull/12216#issuecomment-241175580
result = _grouped_plot_by_column(plot_group, data, columns=columns,
by=by, grid=grid, figsize=figsize,
ax=ax, layout=layout,

2
pandas/tseries/resample.py

@ -1281,7 +1281,7 @@ def _adjust_dates_anchored(first, last, offset, closed='right', base=0):
# error cause by resampling across multiple days when a one day period is
# not a multiple of the frequency.
#
# See https://github.com/pydata/pandas/issues/8683
# See https://github.com/pandas-dev/pandas/issues/8683
first_tzinfo = first.tzinfo
first = first.tz_localize(None)

2
pandas/tseries/tests/test_offsets.py

@ -4606,7 +4606,7 @@ class TestParseTimeString(tm.TestCase):
self.assertEqual(reso, reso_lower)
def test_parse_time_quarter_w_dash(self):
# https://github.com/pydata/pandas/issue/9688
# https://github.com/pandas-dev/pandas/issue/9688
pairs = [('1988-Q2', '1988Q2'), ('2Q-1988', '2Q1988'), ]
for dashed, normal in pairs:

2
pandas/tseries/tests/test_resample.py

@ -1678,7 +1678,7 @@ class TestDatetimeIndex(Base, tm.TestCase):
# start date gets used to determine the offset. Fixes issue where
# a one day period is not a multiple of the frequency.
#
# See: https://github.com/pydata/pandas/issues/8683
# See: https://github.com/pandas-dev/pandas/issues/8683
index = pd.date_range(
'2014-10-14 23:06:23.206', periods=3, freq='400L'

6
pandas/tseries/tests/test_timezones.py

@ -942,7 +942,7 @@ class TestTimeZoneSupportDateutil(TestTimeZoneSupportPytz):
def test_tz_convert_hour_overflow_dst(self):
# Regression test for:
# https://github.com/pydata/pandas/issues/13306
# https://github.com/pandas-dev/pandas/issues/13306
# sorted case US/Eastern -> UTC
ts = ['2008-05-12 09:50:00',
@ -982,7 +982,7 @@ class TestTimeZoneSupportDateutil(TestTimeZoneSupportPytz):
def test_tz_convert_hour_overflow_dst_timestamps(self):
# Regression test for:
# https://github.com/pydata/pandas/issues/13306
# https://github.com/pandas-dev/pandas/issues/13306
tz = self.tzstr('US/Eastern')
@ -1024,7 +1024,7 @@ class TestTimeZoneSupportDateutil(TestTimeZoneSupportPytz):
def test_tslib_tz_convert_trans_pos_plus_1__bug(self):
# Regression test for tslib.tz_convert(vals, tz1, tz2).
# See https://github.com/pydata/pandas/issues/4496 for details.
# See https://github.com/pandas-dev/pandas/issues/4496 for details.
for freq, n in [('H', 1), ('T', 60), ('S', 3600)]:
idx = date_range(datetime(2011, 3, 26, 23),
datetime(2011, 3, 27, 1), freq=freq)

2
scripts/find_undoc_args.py

@ -19,7 +19,7 @@ parser.add_argument('-p', '--path', metavar='PATH', type=str, required=False,
parser.add_argument('-m', '--module', metavar='MODULE', type=str,required=True,
help='name of package to import and examine',action='store')
parser.add_argument('-G', '--github_repo', metavar='REPO', type=str,required=False,
help='github project where the the code lives, e.g. "pydata/pandas"',
help='github project where the the code lives, e.g. "pandas-dev/pandas"',
default=None,action='store')
args = parser.parse_args()

2
scripts/gen_release_notes.py

@ -46,7 +46,7 @@ def get_issues():
def _get_page(page_number):
gh_url = ('https://api.github.com/repos/pydata/pandas/issues?'
gh_url = ('https://api.github.com/repos/pandas-dev/pandas/issues?'
'milestone=*&state=closed&assignee=*&page=%d') % page_number
with urlopen(gh_url) as resp:
rs = resp.readlines()[0]

2
scripts/touchup_gh_issues.py

@ -14,7 +14,7 @@ section header
pat = "((?:\s*GH\s*)?)#(\d{3,4})([^_]|$)?"
rep_pat = r"\1GH\2_\3"
anchor_pat = ".. _GH{id}: https://github.com/pydata/pandas/issues/{id}"
anchor_pat = ".. _GH{id}: https://github.com/pandas-dev/pandas/issues/{id}"
section_pat = "^pandas\s[\d\.]+\s*$"

2
vb_suite/perf_HEAD.py

@ -192,7 +192,7 @@ def get_build_results(build):
return convert_json_to_df(r_url)
def get_all_results(repo_id=53976): # travis pydata/pandas id
def get_all_results(repo_id=53976): # travis pandas-dev/pandas id
"""Fetches the VBENCH results for all travis builds, and returns a list of result df
unsuccesful individual vbenches are dropped.

4
vb_suite/suite.py

@ -67,7 +67,7 @@ try:
TMP_DIR = config.get('setup', 'tmp_dir')
except:
REPO_PATH = os.path.abspath(os.path.join(os.path.dirname(__file__), "../"))
REPO_URL = 'git@github.com:pydata/pandas.git'
REPO_URL = 'git@github.com:pandas-dev/pandas.git'
DB_PATH = os.path.join(REPO_PATH, 'vb_suite/benchmarks.db')
TMP_DIR = os.path.join(HOME, 'tmp/vb_pandas')
@ -138,7 +138,7 @@ These historical benchmark graphs were produced with `vbench
The ``.pandas_vb_common`` setup script can be found here_
.. _here: https://github.com/pydata/pandas/tree/master/vb_suite
.. _here: https://github.com/pandas-dev/pandas/tree/master/vb_suite
Produced on a machine with