Dataset Viewer
Auto-converted to Parquet
id
int64
number
int64
title
string
state
string
created_at
timestamp[s]
updated_at
timestamp[s]
closed_at
timestamp[s]
html_url
string
is_pull_request
bool
pull_request_url
string
pull_request_html_url
string
user_login
string
comments_count
int64
body
string
labels
list
reactions_plus1
int64
reactions_minus1
int64
reactions_laugh
int64
reactions_hooray
int64
reactions_confused
int64
reactions_heart
int64
reactions_rocket
int64
reactions_eyes
int64
comments
list
3,349,520,841
62,179
DOC: Add change in .values to the string migration guide
open
2025-08-24T13:25:22
2025-08-24T13:25:22
null
https://github.com/pandas-dev/pandas/pull/62179
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62179
https://github.com/pandas-dev/pandas/pull/62179
rhshadrach
0
- [x] closes #60301 (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[]
3,348,721,974
62,178
CLN: Lint for usage of Deprecation/FutureWarning
open
2025-08-23T19:47:30
2025-08-24T12:23:30
null
https://github.com/pandas-dev/pandas/pull/62178
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62178
https://github.com/pandas-dev/pandas/pull/62178
rhshadrach
0
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Built on top of #62162
[]
0
0
0
0
0
0
0
0
[]
3,348,683,588
62,177
REF: use _cast_pointwise_result in map
open
2025-08-23T19:19:40
2025-08-23T19:54:21
null
https://github.com/pandas-dev/pandas/pull/62177
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62177
https://github.com/pandas-dev/pandas/pull/62177
NevroHelios
1
- [x] closes #62164 - [x] Passed the tests and added `test_base_map_dtype_preservation.py` - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/v3.0.0.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "Hi @jbrockmendel ,\nI’ve updated the code so that it now preserves the original dtype. Because of this, some of the existing type-checking tests are failing — they were written for the old behavior. These tests might need to be updated to reflect the new semantics. Could you let me know if adjusting them in this way makes sense?" ]
3,348,281,791
62,176
BUG: upgrade to PyQt6 to fix arm64 docker build
open
2025-08-23T14:02:16
2025-08-24T13:35:27
null
https://github.com/pandas-dev/pandas/pull/62176
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62176
https://github.com/pandas-dev/pandas/pull/62176
Alvaro-Kothe
1
- [x] closes #61037 - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. --- - The clipboard tests have passed on my machine (Fedora 42 on Wayland) - I also refactored the qt imports to simplify the version selection - The main goal of this PR is fix the ARM64 docker build. You may test it by running `docker build --platform linux/arm64 -t pandas-dev .` if you are on an `x86_64` system, but be warned that the build takes very long. - Also, I have no clue how would I type annotate the return type of the `_import_module` auxiliary function.
[]
0
0
0
0
0
0
0
0
[ "It seems that conda-forge's [pyqt](https://anaconda.org/conda-forge/pyqt) does not yet have version 6.\r\n\r\nI am not familiar with conda, is it possible to mix package from different repositories to ensure that we can still use pyqt 6?" ]
3,348,047,314
62,175
BUG: Fix Index.get_level_values() mishandling of boolean, pd.NA, np.nan, and pd.NaT levels
open
2025-08-23T11:18:16
2025-08-24T04:25:45
null
https://github.com/pandas-dev/pandas/pull/62175
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62175
https://github.com/pandas-dev/pandas/pull/62175
whyvineet
0
- [x] closes #62169
[]
0
0
0
0
0
0
0
0
[]
3,347,012,226
62,174
BUG/TST: Series round with dtype object
open
2025-08-22T22:56:55
2025-08-22T22:59:16
null
https://github.com/pandas-dev/pandas/pull/62174
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62174
https://github.com/pandas-dev/pandas/pull/62174
sharkipelago
0
- [x] addresses pointwise behavior talked about discussed in #61682 with @jbrockmendel - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. I changed the behavior or series.round() to act like `series.map(round)` if the dtype is object as discussed in #61682. I put this in a new PR as the original intent of the other PR was to change DataFrame behavior. I'm unsure if the way I approached it is a good way to make the change though because I saw in `core/internals/blocks.py` that `Block`'s `round()` method is intended to raise an error for dtype of object. So I wasn't sure I should approach this by updated the `Block` class instead. Any feedback is appreciated!
[]
0
0
0
0
0
0
0
0
[]
3,345,974,324
62,173
added to_csv function to public interface list (__all__) which was misssing earlier
closed
2025-08-22T16:12:55
2025-08-22T16:51:12
2025-08-22T16:46:46
https://github.com/pandas-dev/pandas/pull/62173
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62173
https://github.com/pandas-dev/pandas/pull/62173
Somanshu-Mahajan
2
I didn't find any issue regarding this so i am submitting this PR without issue number.
[]
0
0
0
0
0
0
0
0
[ "`to_csv` is not a top level pandas method. It exists on `DataFrame.to_csv`, so closing", "Ohh.... thankyou for letting me know @mroeschke ." ]
3,345,771,731
62,172
DOC: update the description for the 'docker build' command
open
2025-08-22T14:59:03
2025-08-22T19:13:58
null
https://github.com/pandas-dev/pandas/pull/62172
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62172
https://github.com/pandas-dev/pandas/pull/62172
tihanayo
1
- [x] closes #61037 (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "I strongly recommend against relying on emulation for this issue. Emulation introduces significant performance overhead. I think its best to either change the PyQt5 version for arm64 or use PyQt6 instead, as I stated on https://github.com/pandas-dev/pandas/issues/61037#issuecomment-3215199355." ]
3,345,340,266
62,171
ENH: Add Polars engine to read_csv
closed
2025-08-22T12:33:23
2025-08-22T16:44:42
2025-08-22T16:44:41
https://github.com/pandas-dev/pandas/pull/62171
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62171
https://github.com/pandas-dev/pandas/pull/62171
Ayan9190
3
- Add PolarsParserWrapper class for polars CSV parsing - Update type annotations to include 'polars' as valid engine - Add polars compatibility checks and imports - Update readers.py to integrate polars engine - Add comprehensive test suite for polars engine - Add validation for unsupported options - Add documentation and implementation notes Closes #61813 - [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "Is this AI?", "> Is this AI?\r\n\r\nCertainly has that \"I used AI to create this PR\" smell.\r\n", "We do not accept AI generated pull requests, closing" ]
3,345,321,050
62,170
“Fixed docstring note for date_range/timedelta_range per issue #62161”
closed
2025-08-22T12:26:46
2025-08-22T14:01:40
2025-08-22T14:01:40
https://github.com/pandas-dev/pandas/pull/62170
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62170
https://github.com/pandas-dev/pandas/pull/62170
sachinn854
0
This PR improves the documentation for pandas.date_range arguments. It clarifies the difference between start, end, periods, and freq, addressing user confusion. Closes #62161
[]
0
0
0
0
0
0
0
0
[]
3,343,555,956
62,169
BUG: Index.get_level_values() does not handle boolean, pd.NA, np.nan, or pd.NaT `level` correctly
open
2025-08-21T23:52:04
2025-08-24T12:29:09
null
https://github.com/pandas-dev/pandas/issues/62169
true
null
null
sfc-gh-mvashishtha
3
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd pd.Index([], name=True).get_level_values(level=True) ``` ### Issue Description I get `IndexError: Too many levels: Index has only 1 level, not 2`. Also, ` pd.Index([], name=True).get_level_values(level=False)` should raise `KeyError`, but it returns the whole index. Also, name `np.nan`, `pd.NA` or `pd.NaT` causes ` KeyError: 'Requested level (nan) does not match index name (nan)'`. ### Expected Behavior I should get the whole index from `get_level_values(True)` ### Installed Versions <details> ------------------ commit : 4665c10899bc413b639194f6fb8665a5c70f7db5 python : 3.13.5 python-bits : 64 OS : Darwin OS-release : 24.6.0 Version : Darwin Kernel Version 24.6.0: Mon Jul 14 11:30:30 PDT 2025; root:xnu-11417.140.69~1/RELEASE_ARM64_T6020 machine : arm64 processor : arm byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.2 numpy : 2.3.2 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.1 Cython : None sphinx : None IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : None blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : None html5lib : None hypothesis : None gcsfs : None jinja2 : None lxml.etree : None matplotlib : None numba : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None psycopg2 : None pymysql : None pyarrow : None pyreadstat : None pytest : None python-calamine : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : None xarray : None xlrd : None xlsxwriter : None zstandard : None tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "The issue stems from the `_validate_index_level` function:\n\nhttps://github.com/pandas-dev/pandas/blob/4665c10899bc413b639194f6fb8665a5c70f7db5/pandas/core/indexes/base.py#L2000-L2021\n\nThe problem arises because `isinstance(False, int)` and `isinstance(True, int)` both return `True`. Additionally, the inequality check evaluates to `True` for `np.nan` and `pd.NaT`, though I expected a distinct error for `pd.NA`.\n\nAnyway, since `get_level_values` is essentially a no-op for `Index`, addressing edge cases for non-string names doesn’t seem worth the effort.", "@Alvaro-Kothe can you review my PR?", "@vam5h1 - it appears you are using AI to generate comments on various issuse (#62097, #62050). Please do not do this, we consider this spam." ]
3,343,455,132
62,168
>>> import warnings >>> warnings.filterwarnings("ignore", "\nPyarrow", DeprecationWarning) >>> import pandas
closed
2025-08-21T23:01:31
2025-08-22T01:25:42
2025-08-22T00:25:03
https://github.com/pandas-dev/pandas/issues/62168
true
null
null
hashmatr370-sys
3
null
[]
0
0
0
0
0
0
0
0
[ "What is the purpose of this?", "https://github.com/pandas-dev/pandas/issues/62168", "https://github.com/pandas-dev/pandas/issues/62168" ]
3,343,452,661
62,167
1911902583
closed
2025-08-21T23:00:13
2025-08-22T00:55:46
2025-08-22T00:25:09
https://github.com/pandas-dev/pandas/issues/62167
true
null
null
hashmatr370-sys
3
> Nice, thanks! This is https://github.com/pandas-dev/pandas/pull/57003 indeed. I looked for an open PR but didn't find it somehow ... _Originally posted by @lesteve in [#57082](https://github.com/pandas-dev/pandas/issues/57082#issuecomment-)_
[]
0
0
0
0
0
0
0
0
[ "Is this supposed to have some content?", "A2342", "0767563145" ]
3,343,245,514
62,166
REF: remove unnecessary case from maybe_downcast_to_dtype
closed
2025-08-21T21:27:43
2025-08-22T16:49:52
2025-08-22T16:47:31
https://github.com/pandas-dev/pandas/pull/62166
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62166
https://github.com/pandas-dev/pandas/pull/62166
jbrockmendel
1
All the non-test places we call this pass a np.dtype object, so we can remove the str case support
[ "Refactor" ]
0
0
0
0
0
0
0
0
[ "Thanks @jbrockmendel " ]
3,342,620,053
62,165
REF: get rid of ArrowStringArrayNumpySemantics
closed
2025-08-21T17:31:40
2025-08-22T16:51:44
2025-08-22T16:50:15
https://github.com/pandas-dev/pandas/pull/62165
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62165
https://github.com/pandas-dev/pandas/pull/62165
jbrockmendel
1
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. analogous to #62149
[ "Refactor" ]
0
0
0
0
0
0
0
0
[ "Thanks @jbrockmendel " ]
3,341,927,427
62,164
REF: use _cast_pointwise_result in map
open
2025-08-21T14:13:42
2025-08-21T14:13:42
null
https://github.com/pandas-dev/pandas/issues/62164
true
null
null
jbrockmendel
0
Recently implemented _cast_pointwise_result is the correct idiom to use to retain dtype_backend in EA.map and any similar ops.
[]
0
0
0
0
0
0
0
0
[]
3,340,655,370
62,163
Backport PR #62152 on branch 2.3.x (DOC: prepare 2.3.2 whatsnew notes for release)
closed
2025-08-21T07:38:08
2025-08-21T08:36:49
2025-08-21T08:36:49
https://github.com/pandas-dev/pandas/pull/62163
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62163
https://github.com/pandas-dev/pandas/pull/62163
meeseeksmachine
0
Backport PR #62152: DOC: prepare 2.3.2 whatsnew notes for release
[ "Docs" ]
0
0
0
0
0
0
0
0
[]
3,339,878,599
62,162
CLN: Replace warnings with Pandas4Warning
open
2025-08-21T00:26:34
2025-08-23T19:51:35
null
https://github.com/pandas-dev/pandas/pull/62162
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62162
https://github.com/pandas-dev/pandas/pull/62162
rhshadrach
0
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Replaces all appropriate FutureWarnings and DeprecationWarnings with Pandas4Warnings. For those that are to be enforced in 3.0, I added `# TODO: Enforce ...`. I plan to introduce linting rules on FutureWarning/DeprecationWarning next.
[]
0
0
0
0
0
0
0
0
[]
3,339,837,083
62,161
DOC: date_range/timedelta_range confusing 3 out of 4 arguments note
open
2025-08-20T23:57:56
2025-08-21T00:36:49
null
https://github.com/pandas-dev/pandas/issues/62161
true
null
null
loicdiridollou
0
### Pandas version checks - [x] I have checked that the issue still exists on the latest versions of the docs on `main` [here](https://pandas.pydata.org/docs/dev/) ### Location of the documentation https://pandas.pydata.org/docs/dev/reference/api/pandas.date_range.html#pandas.date_range https://pandas.pydata.org/docs/dev/reference/api/pandas.timedelta_range.html ### Documentation problem We are trying to improve the type hinting for those two functions in `pandas-stubs` repo. Currently the note in both those functions mentions that we should specify 3 out of 4 arguments among `start`, `end`, `period`, `freq`. However it seems like the usage (even in the examples) one can specify only 2 of the arguments like start and end. After looking at the code, it seems like `freq` even if none gets converted to `D`. ### Suggested fix for documentation I believe we should specify two out of three `start`, `end`, `periods` and never all four together. Rephrasing from three out of four but if freq is omitted then ... would help clarify the goal. Here would be a possible version: ``` Of the four parameters start, end, periods, and freq, between two and three must be specified. If freq is omitted, the resulting TimedeltaIndex will have periods linearly spaced elements between start and end (closed on both sides) using a day increment. ```
[ "Docs", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,339,768,196
62,160
REF: avoid inferred_type checks in plotting
closed
2025-08-20T23:09:55
2025-08-21T16:51:24
2025-08-21T16:12:18
https://github.com/pandas-dev/pandas/pull/62160
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62160
https://github.com/pandas-dev/pandas/pull/62160
jbrockmendel
1
- [x] closes #55897 (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Doesn't look like any tests rely on this behavior. Not clear how to do this as a deprecation. Think there is anyone out there with object-dtype columns relying on this?
[ "Visualization" ]
0
0
0
0
0
0
0
0
[ "Thanks @jbrockmendel " ]
3,339,722,814
62,159
DEPR: .str accessor with object dtype
open
2025-08-20T22:38:35
2025-08-20T22:38:35
null
https://github.com/pandas-dev/pandas/pull/62159
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62159
https://github.com/pandas-dev/pandas/pull/62159
jbrockmendel
0
- [x] closes #29710 (Replace xxxx with the GitHub issue number) - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. There are two cases we need an answer for before merging this. One, as I just posted in #29710 > looking at doing this, the main problem is .decode. For everything else we can tell users to do obj.astype("str").str, but if you have bytes to start out with, that doesn't work. Unless we assume pyarrow is available and then can tell users to do .astype("bytes[pyarrow]").str.decode The other is when you can't cast to "str" bc you have funky entries, like in test_to_hdf_errors `data = ["\ud800foo"]`.
[]
0
0
0
0
0
0
0
0
[]
3,339,672,171
62,158
API: indexing dates-with-datetime64
open
2025-08-20T22:07:16
2025-08-21T05:32:18
null
https://github.com/pandas-dev/pandas/issues/62158
true
null
null
jbrockmendel
1
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python ser = pd.Series(["2016-01-01"], dtype="date32[pyarrow]") ser2 = ser.astype("timestamp[ns][pyarrow]") ser3 = ser.astype("datetime64[ns]") dti = pd.Index(ser3) dti.get_loc(ser[0]) # raises KeyError dti.get_indexer(ser.values) # -1s dti.get_indexer(ser.values.astype(object)) # 0s; inconsistent ``` ### Issue Description DatetimeIndex.get_indexer has a special case (actually in Index._maybe_downcast_for_indexing) for sequences of `date` objects that is inconsistent with both scalar treatment and comparison op behavior. This was mostly benign before the existence of a date dtype, but now has the potential to cause problems. The special case should be deprecated. ### Expected Behavior NA ### Installed Versions <details> Replace this line with the output of pd.show_versions() </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "Hi @jbrockmendel, I was wondering if the correct fix here would be to make datetime.date objects never match a DatetimeIndex (so lookups always fail unless the user explicitly converts them to datetime64). Is this the right approach to resolve this issue, or should implicit matching still be supported? Please let me know if this is the recommended direction.\n" ]
3,339,661,943
62,157
BUG/API: comparison datetime64-vs-dates
open
2025-08-20T22:02:13
2025-08-20T22:02:28
null
https://github.com/pandas-dev/pandas/issues/62157
true
null
null
jbrockmendel
0
```python ser = pd.Series(["2016-01-01"], dtype="date32[pyarrow]") ser2 = ser.astype("timestamp[ns][pyarrow]") ser3 = ser.astype("datetime64[ns]") assert (ser3 != ser[0]).all() assert (ser3 != ser).all() assert (ser == ser3).all() # <- uh-oh! inconsistent with reversed operation assert (ser == ser2).all() # inconsistent with the non-pyarrow behavior assert (ser2 == ser).all() ``` Long ago we made a decision that datetime64 arrays (and Timestamp) were _not_ comparable to date objects, matching the stdlib behavior. We should have that behavior for pyarrow dtypes too.
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,338,893,482
62,156
[backport 2.3.x] DOC: move and reword whatsnew note for replace fix (GH-57865) (#62153)
closed
2025-08-20T17:06:00
2025-08-21T07:16:14
2025-08-21T07:16:06
https://github.com/pandas-dev/pandas/pull/62156
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62156
https://github.com/pandas-dev/pandas/pull/62156
jorisvandenbossche
0
Backport of https://github.com/pandas-dev/pandas/pull/62153
[]
0
0
0
0
0
0
0
0
[]
3,338,825,615
62,155
REF: remove unnecessary string mixins
closed
2025-08-20T16:44:07
2025-08-21T16:51:05
2025-08-21T16:14:21
https://github.com/pandas-dev/pandas/pull/62155
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62155
https://github.com/pandas-dev/pandas/pull/62155
jbrockmendel
1
BaseStringArrayMethods is only mixed in two places, and one of those is unnecessary.
[ "Refactor" ]
0
0
0
0
0
0
0
0
[ "Thanks @jbrockmendel " ]
3,337,920,434
62,154
BUG: resampling with origin=end_date includes extra elements in the bucket
open
2025-08-20T12:21:22
2025-08-20T12:21:22
null
https://github.com/pandas-dev/pandas/issues/62154
true
null
null
vasil-pashov
0
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd from pandas import Timedelta, Timestamp df = pd.DataFrame({"col": [1, 2]}, index=pd.to_datetime(["1969-06-01 00:00:00.000000000","1969-06-01 02:01:00.000000001"])) rule = "7315s" offset = pd.Timedelta("23min") df.resample(rule, origin="end_day", offset=offset, closed="right", label="left").agg(None, col_first=("col", "first"), col_count=("col", "count")) This outputs col_first col_count 1969-06-01 1 2 ``` ### Issue Description The bucket starting at `1969-06-01` should not include the first row from the dataframe. The index is `1969-06-01 00:00:00.000000000` and closed is `right` which means that the value is on the opened boundary thus it does not fall into the bucket. Moreover if the second row is removed we get ```python import pandas as pd from pandas import Timedelta, Timestamp df = pd.DataFrame({"col": [1]}, index=pd.to_datetime(["1969-06-01 00:00:00.000000000"])) rule = "7315s" offset = pd.Timedelta("23min") df.resample(rule, origin="end_day", offset=offset, closed="right", label="left").agg(None, col_first=("col", "first"), col_count=("col", "count")) ``` Which outputs: ``` col_first col_count 1969-05-31 22:21:05 1 1 ``` ### Expected Behavior ```python import pandas as pd from pandas import Timedelta, Timestamp df = pd.DataFrame({"col": [1, 2]}, index=pd.to_datetime(["1969-06-01 00:00:00.000000000","1969-06-01 02:01:00.000000001"])) rule = "7315s" offset = pd.Timedelta("23min") df.resample(rule, origin="end_day", offset=offset, closed="right", label="left").agg(None, col_first=("col", "first"), col_count=("col", "count")) ``` The first row should not be part of the bucket. It's part of the bucket starting at `1969-05-31 22:21:05` but as this is before the first value of the dataframe that bucket is not created. ``` col_first col_count 1969-06-01 2 1 ``` ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.11.2 python-bits : 64 OS : Linux OS-release : 6.1.0-38-amd64 Version : #1 SMP PREEMPT_DYNAMIC Debian 6.1.147-1 (2025-08-02) machine : x86_64 processor : byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.1 numpy : 1.26.4 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.2 Cython : None sphinx : None IPython : 9.3.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : None html5lib : None hypothesis : 6.72.4 gcsfs : None jinja2 : 3.1.6 lxml.etree : None matplotlib : None numba : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None psycopg2 : None pymysql : None pyarrow : 21.0.0 pyreadstat : None pytest : 8.4.0 python-calamine : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : 0.9.0 xarray : None xlrd : None xlsxwriter : None zstandard : None tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,337,122,296
62,153
DOC: move and reword whatsnew note for replace fix (GH-57865)
closed
2025-08-20T08:35:52
2025-08-20T17:17:20
2025-08-20T17:03:42
https://github.com/pandas-dev/pandas/pull/62153
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62153
https://github.com/pandas-dev/pandas/pull/62153
jorisvandenbossche
2
We ended up backporting https://github.com/pandas-dev/pandas/pull/57865 after it was already merged, so also moving the whatsnew note here (and I also reworded it slightly, because AFAIK it's not limited to regex cases, given the example in https://github.com/pandas-dev/pandas/issues/61948)
[ "Docs" ]
0
0
0
0
0
0
0
0
[ "Owee, I'm MrMeeseeks, Look at me.\n\nThere seem to be a conflict, please backport manually. Here are approximate instructions:\n\n1. Checkout backport branch and update it.\n\n```\ngit checkout 2.3.x\ngit pull\n```\n\n2. Cherry pick the first parent branch of the this PR on top of the older branch:\n```\ngit cherry-pick -x -m1 7252e9f338f44d6f1e1464171953eb69b7114be2\n```\n\n3. You will likely have some merge/cherry-pick conflict here, fix them and commit:\n\n```\ngit commit -am 'Backport PR #62153: DOC: move and reword whatsnew note for replace fix (GH-57865)'\n```\n\n4. Push to a named branch:\n\n```\ngit push YOURFORK 2.3.x:auto-backport-of-pr-62153-on-2.3.x\n```\n\n5. Create a PR against branch 2.3.x, I would have named this PR:\n\n> \"Backport PR #62153 on branch 2.3.x (DOC: move and reword whatsnew note for replace fix (GH-57865))\"\n\nAnd apply the correct labels and milestones.\n\nCongratulations — you did some good work! Hopefully your backport PR will be tested by the continuous integration and merged soon!\n\nRemember to remove the `Still Needs Manual Backport` label once the PR gets merged.\n\nIf these instructions are inaccurate, feel free to [suggest an improvement](https://github.com/MeeseeksBox/MeeseeksDev).\n ", "Backport -> https://github.com/pandas-dev/pandas/pull/62156" ]
3,337,086,880
62,152
DOC: prepare 2.3.2 whatsnew notes for release
closed
2025-08-20T08:24:59
2025-08-21T07:37:45
2025-08-21T07:37:42
https://github.com/pandas-dev/pandas/pull/62152
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62152
https://github.com/pandas-dev/pandas/pull/62152
jorisvandenbossche
0
xref https://github.com/pandas-dev/pandas/issues/62148
[ "Docs" ]
0
0
0
0
0
0
0
0
[]
3,336,356,931
62,151
Backport PR #62147 on branch 2.3.x (DOC: correct and rewrite string migration section on astype(str))
closed
2025-08-20T02:40:43
2025-08-20T07:18:20
2025-08-20T07:18:20
https://github.com/pandas-dev/pandas/pull/62151
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62151
https://github.com/pandas-dev/pandas/pull/62151
meeseeksmachine
0
Backport PR #62147: DOC: correct and rewrite string migration section on astype(str)
[ "Docs", "Strings" ]
0
0
0
0
0
0
0
0
[]
3,336,294,818
62,150
BUG: Inconsistent behaviour when merging reset-ed MultiIndex dataframe
open
2025-08-20T01:55:34
2025-08-20T01:55:34
null
https://github.com/pandas-dev/pandas/issues/62150
true
null
null
renkeven
0
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python df = pd.DataFrame(data={("column_1", ""): [1,1], ("column_2", ""): [2,2]}) df.index = pd.MultiIndex.from_arrays([[1,1], ["metadata_1", "metadata_2"]], names=["index", "metadata"]) df2 = pd.DataFrame(data=[1,1], index=[1,1]).rename_axis("index", axis=0) df2.columns = pd.MultiIndex.from_product([["new_data"], [""]]) df.reset_index().merge(df2.reset_index(), on="index") # IndexError: Requested axis not found in manager df.reset_index().merge(df2.reset_index(), on=[("index", "")]) # IndexError: Requested axis not found in manager df.reset_index().columns # MultiIndex([( 'index', ''), # ('metadata', ''), # ('column_1', ''), # ('column_2', '')], # ) df2.reset_index().columns # MultiIndex([( 'index', ''), # ('new_data', '')], # ) ## We have a workaround if we force df2 to start as a multi-index, then it works df2 = pd.DataFrame(data=[1,1], index=[1,1]).rename_axis("index", axis=0) df2.columns = pd.MultiIndex.from_product([["new_data"], [""]]) df2.index = pd.MultiIndex.from_arrays( [df2.index, ["dummy"] * len(df2)], names=["index", "dummy_index"], ) # Merge works df.reset_index().merge(df2.reset_index(), on=[("index", "")]) # Extra empty multi-index column df2.reset_index().columns # MultiIndex([( 'index', ''), # ('dummy_index', ''), # ('new_data', '')], # ) ``` ### Issue Description I have a dataframe (main) with multi-indexed index and columns. I would like to do a merge with a second dataframe (side) that is single-index and single column. I coerced (side) to give it the same levels in the columns, and do a pd.merge on (main).reset_index and side.reset_index. The example works if you make (side) also a multi-index index. Additionally, the first example above is fine on 2.2.3, but is broken on 2.3.1 ### Expected Behavior Usual merge behaviour. Not sure why the initial dataframe must both have multi-index indexes when we use .reset_index() on both and merge over a single column ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.12.11 python-bits : 64 OS : Linux OS-release : 6.6.93+ Version : #1 SMP PREEMPT_DYNAMIC Fri Jun 27 09:03:39 UTC 2025 machine : x86_64 processor : x86_64 byteorder : little LC_ALL : en_US.UTF-8 LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.1 numpy : 1.26.4 pytz : 2024.1 dateutil : 2.9.0.post0 pip : 25.0.1 Cython : None sphinx : None IPython : 8.36.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : 2025.3.2 html5lib : None hypothesis : None gcsfs : 2025.3.2 jinja2 : 3.1.6 lxml.etree : 6.0.0 matplotlib : 3.10.5 numba : 0.61.2 numexpr : None odfpy : None openpyxl : 3.1.5 pandas_gbq : None psycopg2 : 2.9.9 pymysql : None pyarrow : 19.0.1 pyreadstat : None pytest : 8.3.5 python-calamine : None pyxlsb : 1.0.10 s3fs : None scipy : 1.16.0 sqlalchemy : 2.0.42 tables : None tabulate : 0.9.0 xarray : 2025.1.2 xlrd : 2.0.1 xlsxwriter : 3.2.3 zstandard : 0.23.0 tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,335,764,167
62,149
REF: get rid of StringArrayNumpySemantics
open
2025-08-19T21:35:50
2025-08-22T17:05:02
null
https://github.com/pandas-dev/pandas/pull/62149
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62149
https://github.com/pandas-dev/pandas/pull/62149
jbrockmendel
0
xref #62129 get one less variant
[]
0
0
0
0
0
0
0
0
[]
3,335,625,273
62,148
RLS: 2.3.2
closed
2025-08-19T20:40:14
2025-08-22T16:16:16
2025-08-22T16:12:21
https://github.com/pandas-dev/pandas/issues/62148
true
null
null
jorisvandenbossche
2
2.3.2 milestone: https://github.com/pandas-dev/pandas/milestone/121 I could do a 2.3.2 release tomorrow, just with what is available right now (and we can always do a 2.3.3 release later if needed) cc @pandas-dev/pandas-core
[ "Release" ]
2
0
0
0
0
0
0
0
[ "Released yesterday, and this time also announced it (with a focus on the opt-in 3.0 changes) on the mailing list, Linkedin and mastodon.\n(and fwiw I'll be offline next week)", "Thanks for making this happen!" ]
3,335,445,033
62,147
DOC: correct and rewrite string migration section on astype(str)
closed
2025-08-19T19:27:23
2025-08-20T02:44:41
2025-08-20T02:40:08
https://github.com/pandas-dev/pandas/pull/62147
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62147
https://github.com/pandas-dev/pandas/pull/62147
jorisvandenbossche
1
Tries to resolve https://github.com/pandas-dev/pandas/issues/61992
[ "Docs", "Strings" ]
0
0
0
0
0
0
0
0
[ "Thanks @jorisvandenbossche " ]
3,335,038,521
62,146
BUG: Fails to Build on Raspberry Pi 3 (tested both 2.1.4, 2.3.1 and main Branch)
open
2025-08-19T16:49:22
2025-08-19T17:51:28
null
https://github.com/pandas-dev/pandas/issues/62146
true
null
null
luckylinux
1
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python Build Fails on Raspberry Pi 3 running Raspberry Pi OS based on Debian Trixie (13) on ARMHF (32 Bit Mode). ``` ### Issue Description Everything was working fine on Debian Boorworm (12) ARMHF on a Raspberry Pi 3. When I decided to upgrade today to Debian Trixie (13) ARMHF, I had to create a new `venv` since Python changed from Version 3.11 to 3.13. Unfortunately, I couldn't build `pandas` at all 😞. And, despite the Claims in the Error Message, there is no Log File at all about why Meson failed to build. ### Expected Behavior Build should succeed and I should be able to use `pandas` within my Python Application. ### Installed Versions Cannot build/install. Tried 2.3.1 which is latest: ``` Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple Requirement already satisfied: python-can in ./venv/lib/python3.13/site-packages (from -r requirements.txt (line 1)) (4.6.1) Collecting pandas (from -r requirements.txt (line 2)) Using cached pandas-2.3.1.tar.gz (4.5 MB) Installing build dependencies ... done Getting requirements to build wheel ... done Installing backend dependencies ... done Preparing metadata (pyproject.toml) ... error error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [31 lines of output] + meson setup /tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f /tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f/.mesonpy-uutfurzj -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --vsenv --native-file=/tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f/.mesonpy-uutfurzj/meson-python-native-file.ini The Meson build system Version: 1.8.3 Source dir: /tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f Build dir: /tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f/.mesonpy-uutfurzj Build type: native build Project name: pandas Project version: 2.3.1 C compiler for the host machine: cc (gcc 14.2.0 "cc (Raspbian 14.2.0-19+rpi1) 14.2.0") C linker for the host machine: cc ld.bfd 2.44 C++ compiler for the host machine: c++ (gcc 14.2.0 "c++ (Raspbian 14.2.0-19+rpi1) 14.2.0") C++ linker for the host machine: c++ ld.bfd 2.44 Cython compiler for the host machine: cython (cython 3.1.3) Host machine cpu family: arm Host machine cpu: armv7l Program python found: YES (/opt/app/venv/bin/python) ../pandas/meson.build:1:15: ERROR: Command `/opt/app/venv/bin/python -c ' import os import numpy as np try: # Check if include directory is inside the pandas dir # e.g. a venv created inside the pandas dir # If so, convert it to a relative path incdir = os.path.relpath(np.get_include()) except Exception: incdir = np.get_include() print(incdir) '` failed with status 1. A full log can be found at /tmp/pip-install-y9j9k_v2/pandas_d677fcd2c13643ff8c24a90c0a9f811f/.mesonpy-uutfurzj/meson-logs/meson-log.txt [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. ``` Tried 2.1.4 which worked fine on Debian Bookworm: ``` Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple Requirement already satisfied: python-can==4.3.1 in ./venv/lib/python3.13/site-packages (from -r requirements.txt (line 1)) (4.3.1) Collecting pandas==2.1.4 (from -r requirements.txt (line 2)) Downloading pandas-2.1.4.tar.gz (4.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.3/4.3 MB 6.6 MB/s 0:00:00 Installing build dependencies ... done Getting requirements to build wheel ... done Installing backend dependencies ... done Preparing metadata (pyproject.toml) ... error error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [31 lines of output] + meson setup /tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657 /tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657/.mesonpy-lko771pl/build -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --vsenv --native-file=/tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657/.mesonpy-lko771pl/build/meson-python-native-file.ini The Meson build system Version: 1.2.1 Source dir: /tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657 Build dir: /tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657/.mesonpy-lko771pl/build Build type: native build Project name: pandas Project version: 2.1.4 C compiler for the host machine: cc (gcc 14.2.0 "cc (Raspbian 14.2.0-19+rpi1) 14.2.0") C linker for the host machine: cc ld.bfd 2.44 C++ compiler for the host machine: c++ (gcc 14.2.0 "c++ (Raspbian 14.2.0-19+rpi1) 14.2.0") C++ linker for the host machine: c++ ld.bfd 2.44 Cython compiler for the host machine: cython (cython 0.29.37) Host machine cpu family: arm Host machine cpu: armv7l Program python found: YES (/opt/app/venv/bin/python) ../../pandas/meson.build:1:15: ERROR: Command `/opt/app/venv/bin/python -c ' import os import numpy as np try: # Check if include directory is inside the pandas dir # e.g. a venv created inside the pandas dir # If so, convert it to a relative path incdir = os.path.relpath(np.get_include()) except Exception: incdir = np.get_include() print(incdir) '` failed with status 1. A full log can be found at /tmp/pip-install-lpm0348t/pandas_049cc42ad8234edf931488590a6d6657/.mesonpy-lko771pl/build/meson-logs/meson-log.txt [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. ``` Tried `main` Branch: ``` Looking in indexes: https://pypi.org/simple, https://www.piwheels.org/simple Collecting pandas@ git+https://github.com/pandas-dev/pandas@main (from -r requirements.txt (line 4)) Cloning https://github.com/pandas-dev/pandas (to revision main) to /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0 Running command git clone --filter=blob:none --quiet https://github.com/pandas-dev/pandas /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0 Resolved https://github.com/pandas-dev/pandas to commit 3940df8255ed04db0089e66ce09a4c986b97cac4 Installing build dependencies ... done Getting requirements to build wheel ... done Installing backend dependencies ... done Preparing metadata (pyproject.toml) ... error error: subprocess-exited-with-error × Preparing metadata (pyproject.toml) did not run successfully. │ exit code: 1 ╰─> [34 lines of output] + meson setup /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0 /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --vsenv --native-file=/tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb/meson-python-native-file.ini The Meson build system Version: 1.8.3 Source dir: /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0 Build dir: /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb Build type: native build Project name: pandas Project version: 3.0.0.dev0+2322.g3940df8255 C compiler for the host machine: cc (gcc 14.2.0 "cc (Raspbian 14.2.0-19+rpi1) 14.2.0") C linker for the host machine: cc ld.bfd 2.44 C++ compiler for the host machine: c++ (gcc 14.2.0 "c++ (Raspbian 14.2.0-19+rpi1) 14.2.0") C++ linker for the host machine: c++ ld.bfd 2.44 Cython compiler for the host machine: cython (cython 3.1.3) Host machine cpu family: arm Host machine cpu: armv7l Program python found: YES (/opt/app/venv/bin/python) Program cython found: YES (/tmp/pip-build-env-es6_6v3v/overlay/bin/cython) Found pkg-config: YES (/usr/bin/pkg-config) 1.8.1 Run-time dependency python found: YES 3.13 ../pandas/meson.build:1:15: ERROR: Command `/opt/app/venv/bin/python -c ' import os import numpy as np try: # Check if include directory is inside the pandas dir # e.g. a venv created inside the pandas dir # If so, convert it to a relative path incdir = os.path.relpath(np.get_include()) except Exception: incdir = np.get_include() print(incdir) '` failed with status 1. A full log can be found at /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb/meson-logs/meson-log.txt [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. ``` Unfortunately there is no Log at all. Not even the Root Folder essentially: ``` (venv) USER@HOST:/opt/app $ ls -l /tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb/meson-logs/meson-log.txt ls: cannot access '/tmp/pip-install-q7i1opum/pandas_0a4fefd97257431fa2c67332ca1aa7e0/.mesonpy-nyba67pb/meson-logs/meson-log.txt': No such file or directory (venv) USER@HOST:/opt/app $ ls -l /tmp/pip-install-q7i1opum ls: cannot access '/tmp/pip-install-q7i1opum': No such file or directory ``` Similar story for the other Versions tried (2.1.4 and 2.3.1). System Information: - Python 3.13.5 `/etc/os-release`: ``` PRETTY_NAME="Raspbian GNU/Linux 13 (trixie)" NAME="Raspbian GNU/Linux" VERSION_ID="13" VERSION="13 (trixie)" VERSION_CODENAME=trixie DEBIAN_VERSION_FULL=13.0 ID=raspbian ID_LIKE=debian HOME_URL="http://www.raspbian.org/" SUPPORT_URL="http://www.raspbian.org/RaspbianForums" BUG_REPORT_URL="http://www.raspbian.org/RaspbianBugs" ``` `cat /proc/cpuinfo`: ``` processor : 0 model name : ARMv7 Processor rev 4 (v7l) BogoMIPS : 76.80 Features : half thumb fastmult vfp edsp neon vfpv3 tls vfpv4 idiva idivt vfpd32 lpae evtstrm crc32 CPU implementer : 0x41 CPU architecture: 7 CPU variant : 0x0 CPU part : 0xd03 CPU revision : 4 processor : 1 model name : ARMv7 Processor rev 4 (v7l) BogoMIPS : 76.80 Features : half thumb fastmult vfp edsp neon vfpv3 tls vfpv4 idiva idivt vfpd32 lpae evtstrm crc32 CPU implementer : 0x41 CPU architecture: 7 CPU variant : 0x0 CPU part : 0xd03 CPU revision : 4 processor : 2 model name : ARMv7 Processor rev 4 (v7l) BogoMIPS : 76.80 Features : half thumb fastmult vfp edsp neon vfpv3 tls vfpv4 idiva idivt vfpd32 lpae evtstrm crc32 CPU implementer : 0x41 CPU architecture: 7 CPU variant : 0x0 CPU part : 0xd03 CPU revision : 4 processor : 3 model name : ARMv7 Processor rev 4 (v7l) BogoMIPS : 76.80 Features : half thumb fastmult vfp edsp neon vfpv3 tls vfpv4 idiva idivt vfpd32 lpae evtstrm crc32 CPU implementer : 0x41 CPU architecture: 7 CPU variant : 0x0 CPU part : 0xd03 CPU revision : 4 Hardware : BCM2835 Revision : a02082 Serial : 000000003cad014e Model : Raspberry Pi 3 Model B Rev 1.2 ``` `raspinfo`: ``` System Information ------------------ Raspberry Pi 3 Model B Rev 1.2 PRETTY_NAME="Raspbian GNU/Linux 13 (trixie)" NAME="Raspbian GNU/Linux" VERSION_ID="13" VERSION="13 (trixie)" Raspberry Pi reference 2023-10-10 Generated using pi-gen, https://github.com/RPi-Distro/pi-gen, fb56ad562991cf3ae5c96ab50983e1deeaefc7b6, stage2 Linux HOST 6.12.34+rpt-rpi-v7 #1 SMP Raspbian 1:6.12.34-1+rpt1 (2025-06-26) armv7l GNU/Linux Revision : a02082 Serial : 000000003cad014e Model : Raspberry Pi 3 Model B Rev 1.2 Throttled flag : throttled=0x20000 Camera : supported=0 detected=0, libcamera interfaces=0 ```
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "Seems that some of the Errors could be due to the the OOM Killer.\n\nI fail to see why that would happen though. Granted RAM is only 1GB (but it did build before, like ~ 1.5 Years ago or so). And I also allocated 8GB SWAP to make sure that there would be no Edge Cases 😕.\n\nAnyways, in most cases it fails at the metadata Generation Stage already 😕, so it's NOT a Memory Issue then." ]
3,334,809,071
62,145
fix enum typing from _lib
open
2025-08-19T15:28:47
2025-08-24T07:02:34
null
https://github.com/pandas-dev/pandas/pull/62145
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62145
https://github.com/pandas-dev/pandas/pull/62145
Dr-Irv
1
`mypy` is fixing how they handle `enum` types, and the fix revealed an issue with how we handle them. See https://github.com/python/mypy/pull/19687#issuecomment-3200527533
[]
0
0
0
0
0
0
0
0
[ "I reviewed the pull request #62145: \"fix enum typing from _lib\". \r\n\r\n**Summary:**\r\n\r\n- This PR updates enum typing in the `_libs` module to align with recent changes in `mypy` (see python/mypy#19687). \r\n- It removes previously added `# type: ignore` comments and fixes type annotations for enum handling in `pandas/_libs/tslibs/dtypes.pyi`. \r\n- Most checks passed successfully (32/37), with only `Docstring validation` failing. All functional and unit tests appear green. \r\n- There are no conflicts with the base branch, and changes can be cleanly merged. \r\n\r\n**Impact:**\r\n\r\n- Improves type safety and static type checking for enums in the pandas C extensions. \r\n- Helps catch potential typing issues early in development and during CI. \r\n- No behavioral changes are expected at runtime. \r\n\r\nOverall, this is a valuable typing improvement that aligns pandas with upstream `mypy` fixes and maintains code quality in `_libs`.\r\n" ]
3,334,753,694
62,144
BUG / API: setitem on pandas object fails if underlying numpy array is read-only
open
2025-08-19T15:11:31
2025-08-19T18:46:50
null
https://github.com/pandas-dev/pandas/issues/62144
true
null
null
jorisvandenbossche
3
Consider the following example: ```python >>> ser = pd.Series(np.asarray(pa.array([1, 2, 3])), copy=False) >>> ser 0 1 1 2 2 3 dtype: int64 >>> ser.iloc[0] = 10 ... File ~/scipy/repos/pandas/pandas/core/internals/blocks.py:1139, in Block.setitem(self, indexer, value) 1137 casted = casted[0, ...] 1138 try: -> 1139 values[indexer] = casted 1140 except (TypeError, ValueError) as err: 1141 if is_list_like(casted): ValueError: assignment destination is read-only ``` Here I create a pandas Series backed by a read-only numpy array somewhat manually through conversion from pyarrow and an explicit `copy=False`, but you can also run into this in certain cases when converting from Arrow-like data to pandas (depending on the exact conversion method). And with CoW we are now returning read-only arrays from `.values` etc, and so this might also become more common in non-Arrow related code. From a user perspective, you have a pandas Series, and in general we allow to mutate a Series. **So should this just always work?** (regardless of how the Series was created, which is not necessarily always under the user control) If this happens, we could automatically copy the array inplace in the Series to let the high-level pandas setitem operation work as expected (at the cost of an unexpected delayed copy, but this is similar as with CoW)
[ "Indexing", "API Design" ]
0
0
0
0
0
0
0
0
[ "if i wanted to prevent mutation, setting the underlying array to readonly would be how i would do it", "While that is certainly a trick that sometimes works, I would consider that relying on internal details, as we in the past copied quite generously, and now with CoW might do a copy before performing the setitem, which makes this not a very reliable trick. For example:\n\n```\n>>> ser = pd.Series(np.asarray(pa.array([1, 2, 3])), copy=False)\n>>> ser2 = ser[:2]\n>>> ser.iloc[0] = 10\n>>> ser\n0 10\n1 2\n2 3\ndtype: int64\n```\n\nJust because there was a view on the data, now this setitem does not raise an error.", "fair enough. i guess id be fine either with a silent copy of with raising so the user can copy themselves." ]
3,333,226,449
62,143
DOC: Clarify is_scalar docstring to specify scalar vs non-scalar and add Enum example (GH#62063)
open
2025-08-19T07:22:18
2025-08-24T13:57:52
null
https://github.com/pandas-dev/pandas/pull/62143
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62143
https://github.com/pandas-dev/pandas/pull/62143
Aniketsy
1
(GH#62063) This PR updates the docstring for is_scalar to better reflect its actual behavior in pandas. Clearly differentiates between values considered scalars and values treated as non-scalars. Explicitly documents that the following are treated as non-scalars np.ndarray, list , tuple , pandas.Series. Please let me know if this fix needs any improvements . I’m open to feedback and happy to make changes based on suggestions. Thankyou!
[]
1
0
0
0
0
0
0
0
[ "Hi @rhshadrach, as per your suggestion on the issue, I’ve applied the changes. Could you please take a look and let me know if this correctly states and enhances the doc for the issue?\r\nI’m happy to make further adjustments based on your feedback." ]
3,332,198,888
62,142
DEPR: Categorical with values not present in categories
open
2025-08-18T21:30:22
2025-08-21T20:28:27
null
https://github.com/pandas-dev/pandas/pull/62142
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62142
https://github.com/pandas-dev/pandas/pull/62142
jbrockmendel
0
- [x] closes #40996 - [x] closes #59899 - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[]
3,330,046,734
62,141
ENH: Consistent `name` property for the iterates in `DataFrameGroupBy`
open
2025-08-18T10:00:32
2025-08-20T03:08:22
null
https://github.com/pandas-dev/pandas/issues/62141
true
null
null
FelixBenning
3
### Feature Type - [ ] Adding new functionality to pandas - [x] Changing existing functionality in pandas - [ ] Removing existing functionality in pandas ### Problem Description If a dataframe is grouped by a single column sometimes its `name` is a numeric scalar and sometimes a single element tuple. This should be more consistent. In the following I will consider the following example dataframe ```python >>> df a b c 0 1 2 3 1 1 5 6 2 7 8 9 ``` ## Cases where `name` is a scalar - `DataFrameGroupBy.groups` ```python >>> print(df.groupby(['a']).groups) {1: [0, 1], 7: [2]} ``` - `DataFrameGroupBy.apply` ```python >>> df.groupby(['a']).apply(lambda x: print(x.name), include_groups=False) 1 7 Empty DataFrame Columns: [] Index: [] >>> ``` ## Cases where `name` is a one element tuple - `DataFrameGroupBy.__iter__` ```python >>> for name, _ in df.groupby(['a']): print(name) ... (1,) (7,) ``` ## Documentation It should perhaps be said that `DataFrameGroupBy.name` is ill documented. But it is not a private property. It seems like the most natural thing to query if you need the information from the columns that you have grouped. This is especially important, as pandas forces `include_groups=False` in `apply` with a `FutureWarning`/`DeprecationWarning`. So `DataFrameGroupBy.name` seems like the most natural way to reobtain this information now. ### Feature Description Consistency in either direction: - PRO SCALAR: It appears that in the majority of cases the `name` is a scalar. Although to be fair I have not checked many cases. This is probably also more intuitive to people that do not think about multiple column groupings - PRO TUPLE: The single element tuple makes this more consistent with the case, where multiple columns are selected. ### Additional Context _No response_
[ "Bug", "Enhancement", "Groupby", "API - Consistency" ]
0
0
0
0
0
0
0
0
[ "Thanks for the report. For `apply`, I believe we want to move away from pinning name on the passed DataFrame. However for `GroupBy.groups`, I'm positive on changing the keys to be tuples in the case where the user is grouping by a 1-element iterable. PRs to fix are welcome!", "I am a bit confused why you want to remove this entirely from `apply`. `apply` feels like the functional alternative to `__iter__`, where you provide the name. It is entirely reasonable to want to use this (constant) information in the apply function. Why would you want to remove this information? How would you access this information if both the columns and the name is removed?\n\nCould `include_groups=True` be a flag such that the group values are passed to the apply function as a second argument?", "> I am a bit confused why you want to remove this entirely from `apply`.\n\nThis is https://github.com/pandas-dev/pandas/issues/41090. In order to keep the discussion consolidated, I suggest sharing any thoughts there and leaving this issue for the change to `.groups`." ]
3,329,939,360
62,140
BUG: type error in `df.attrs.update()`
closed
2025-08-18T09:31:00
2025-08-18T21:25:54
2025-08-18T21:25:40
https://github.com/pandas-dev/pandas/issues/62140
true
null
null
janosh
1
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd df = pd.DataFrame({"A": [1, 2, 3], "B": [4, 5, 6]}) df.attrs.update(name="dummy", foo="bar") # type error here ``` ### Issue Description `ty 0.0.1-alpha.18 (d697cc092 2025-08-14)` raises a type error: > Argument to bound method `update` is incorrect: Expected `Mapping[str, Any]`, found `dict[Hashable, Any]`tyinvalid-argument-type ### Expected Behavior changing to ```diff - df.attrs.update(name="dummy", foo="bar") + df.attrs.update({"name": "dummy", "foo": "bar"}) ``` resolves the type error. both work at run time and i'd expect both to type check fine ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.13.0 python-bits : 64 OS : Darwin OS-release : 24.6.0 Version : Darwin Kernel Version 24.6.0: Mon Jul 14 11:30:40 PDT 2025; root:xnu-11417.140.69~1/RELEASE_ARM64_T6041 machine : arm64 processor : arm byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.1 numpy : 2.3.2 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.2 Cython : 3.0.12 sphinx : 8.1.3 IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.3 blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : 2025.5.1 html5lib : 1.1 hypothesis : None gcsfs : None jinja2 : 3.1.6 lxml.etree : 6.0.0 matplotlib : 3.10.5 numba : None numexpr : 2.10.2 odfpy : None openpyxl : 3.1.5 pandas_gbq : None psycopg2 : 2.9.10 pymysql : 1.4.6 pyarrow : 19.0.1 pyreadstat : None pytest : 8.4.0 python-calamine : None pyxlsb : None s3fs : None scipy : 1.16.1 sqlalchemy : None tables : 3.10.2 tabulate : 0.9.0 xarray : None xlrd : None xlsxwriter : None zstandard : 0.24.0 tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Typing", "metadata" ]
0
0
0
0
0
0
0
0
[ "`ty` is in alpha and raises many false positives. Do you experience the same with other type checkers? If so report back and we can reconsider. Closing for now." ]
3,329,623,081
62,139
BUG: Fix to_csv microsecond inconsistency (#62111)
open
2025-08-18T07:58:24
2025-08-18T14:05:08
null
https://github.com/pandas-dev/pandas/pull/62139
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62139
https://github.com/pandas-dev/pandas/pull/62139
prazian
0
- [x] closes #62111 - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[]
3,329,292,249
62,138
BUG: Preserve day freq on DatetimeIndex subtraction
open
2025-08-18T06:05:44
2025-08-18T06:06:44
null
https://github.com/pandas-dev/pandas/pull/62138
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62138
https://github.com/pandas-dev/pandas/pull/62138
23f-1000003
1
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "I suggested some changes" ]
3,329,239,201
62,137
ENH: Rename DataFrame._append to _append_internal
closed
2025-08-18T05:41:28
2025-08-21T20:36:04
2025-08-21T20:35:56
https://github.com/pandas-dev/pandas/pull/62137
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62137
https://github.com/pandas-dev/pandas/pull/62137
whyvineet
1
- [x] closes #57936
[]
0
0
0
0
0
0
0
0
[ "thanks @whyvineet " ]
3,329,159,586
62,136
DOC: Deprecate empty string defaults in DataFrame.join suffixes
open
2025-08-18T04:55:03
2025-08-22T00:01:27
null
https://github.com/pandas-dev/pandas/pull/62136
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62136
https://github.com/pandas-dev/pandas/pull/62136
Roline-Stapny
0
- [X] closes [#61294 ](https://github.com/pandas-dev/pandas/issues/61294) - [X] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [X] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [X] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [X] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[]
3,328,849,299
62,135
BUG: Fix groupby.apply() dropping _metadata from subclassed DataFrame
open
2025-08-18T01:09:31
2025-08-20T03:15:52
null
https://github.com/pandas-dev/pandas/pull/62135
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62135
https://github.com/pandas-dev/pandas/pull/62135
MengAiDev
3
When extending pandas.DataFrame by subclassing, most operations preserve the _metadata attributes. This fix ensures that groupby.apply() also preserves these fields, making it consistent with other groupby operations like groupby.sum(). Fixes #62134
[]
0
0
0
0
0
0
0
0
[ "> Thanks for the PR! Why not call `__finalize__` instead?\r\n\r\nThanks!", "@rhshadrach Why this failed? [Unit Tests / Linux-32-bit (pull_request)](https://github.com/pandas-dev/pandas/actions/runs/17032074576/job/48276764930?pr=62135) and [Unit Tests / Pyodide build (pull_request)](https://github.com/pandas-dev/pandas/actions/runs/17032074576/job/48276764950?pr=62135)", "Seems related to the change you're making here in `frame.py`." ]
3,328,402,711
62,134
BUG: groupby.apply() drops _metadata from subclassed DataFrame
open
2025-08-17T15:03:32
2025-08-17T15:37:00
null
https://github.com/pandas-dev/pandas/issues/62134
true
null
null
JBGreisman
0
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas._testing as tm import numpy as np subdf = tm.SubclassedDataFrame( {"X": [1, 1, 2, 2, 3], "Y": np.arange(0, 5), "Z": np.arange(10, 15)} ) subdf.testattr = "test" # Calculate groupby-sum in two ways: one preserves metadata, one does not expected = subdf.groupby("X").sum() result = subdf.groupby("X").apply(np.sum, axis=0, include_groups=False) # Both dataframes have equivalent content tm.assert_frame_equal(result, expected) print(expected.testattr) # prints "test" print(result.testattr) # raises AttributeError ``` ### Issue Description When extending the `pandas.DataFrame` by subclassing, most operations preserve the `_metadata` attributes. This is not the case for `groupby.apply()`, after which the `_metadata` fields are dropped. I think this is unintended behavior, because an equivalent call using a built-in groupby method (such as `groupby.mean()`) does preserve the fields. ### Expected Behavior ```python import pandas._testing as tm import numpy as np subdf = tm.SubclassedDataFrame( {"X": [1, 1, 2, 2, 3], "Y": np.arange(0, 5), "Z": np.arange(10, 15)} ) subdf.testattr = "test" result = subdf.groupby("X").apply(np.sum, axis=0, include_groups=False) assert result.testattr == "test" # attribute should be preserved after groupby-apply ``` ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.12.11 python-bits : 64 OS : Darwin OS-release : 24.6.0 Version : Darwin Kernel Version 24.6.0: Mon Jul 14 11:30:51 PDT 2025; root:xnu-11417.140.69~1/RELEASE_ARM64_T8112 machine : arm64 processor : arm byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.1 numpy : 2.2.6 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.1 Cython : 3.1.3 sphinx : 8.1.3 IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 blosc : None bottleneck : 1.5.0 dataframe-api-compat : None fastparquet : 2024.11.0 fsspec : 2025.7.0 html5lib : 1.1 hypothesis : 6.138.2 gcsfs : 2025.7.0 jinja2 : 3.1.6 lxml.etree : 6.0.0 matplotlib : 3.10.5 numba : 0.61.2 numexpr : 2.11.0 odfpy : None openpyxl : 3.1.5 pandas_gbq : None psycopg2 : 2.9.10 pymysql : 1.4.6 pyarrow : 21.0.0 pyreadstat : 1.3.1 pytest : 8.4.1 python-calamine : None pyxlsb : 1.0.10 s3fs : 2025.7.0 scipy : 1.16.1 sqlalchemy : 2.0.43 tables : 3.10.2 tabulate : 0.9.0 xarray : 2025.8.0 xlrd : 2.0.2 xlsxwriter : 3.2.5 zstandard : 0.23.0 tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "metadata", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,327,708,161
62,133
DOC: Updated shift docstring to mark Series behavior
closed
2025-08-16T22:26:16
2025-08-22T23:00:13
2025-08-18T15:50:35
https://github.com/pandas-dev/pandas/pull/62133
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62133
https://github.com/pandas-dev/pandas/pull/62133
sharkipelago
1
- [x] closes #61955
[ "Docs" ]
0
0
0
0
0
0
0
0
[ "Thanks @sharkipelago " ]
3,327,308,703
62,132
DOC: Clarify that is_scalar treats Enum members as scalars
closed
2025-08-16T14:03:22
2025-08-17T19:52:01
2025-08-17T18:42:39
https://github.com/pandas-dev/pandas/pull/62132
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62132
https://github.com/pandas-dev/pandas/pull/62132
shrutisachan08
4
## Description This PR updates the `is_scalar` docstring to clarify that Enum members are treated as scalars, addressing the confusion reported in #62063. The `is_scalar` function already correctly treats Enum members as scalars (via `isinstance(val, Enum)` in the implementation), but the documentation was unclear about this behavior, leading to confusion for users and type checkers. ## Changes Made - Updated the docstring to clearly explain the exclusion rule: only `list`, `tuple`, `numpy.ndarray`, and `pandas Series` are considered non-scalar - Added explicit mention of "Enum members" in the list of scalar types - Added a practical Enum example in the Examples section ## Checklist - [x] closes #62063
[]
0
0
0
0
0
0
0
0
[ "pre-commit.ci autofix", "This looks like AI. Please don’t do that.", "@jbrockmendel May I please know which section you are referring to .", "Agreed, closing as it appears this is AI generated" ]
3,327,046,500
62,131
ENH: Enable destination aware scope for dataset attributes metadata propagation copy routines
open
2025-08-16T08:40:19
2025-08-20T23:28:52
null
https://github.com/pandas-dev/pandas/pull/62131
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62131
https://github.com/pandas-dev/pandas/pull/62131
dangreb
7
As noted in project docstrings, **NDFrame.attrs** is meant to store properties related to the dataset. Therefore it should be important that during propagation to derived datasets, copy routines can have access to the destination object, to witch it's new formulation will then refeer to. - A lean and isolated fix manages to acomplish this, by forwarding the instance via deepcopy's **memo** payload dict. - The key for the dict entry will be **builtins.id()** of the nwe dataset, complying with mandatory **int** tipification. - Implemented for both NDFrame heirs and concat/merge scenarios. Not relevant for Indexes as it seeems. I'll open as draft before it can be deemed admissible.
[]
0
0
0
0
0
0
0
0
[ "Is this AI? Looks like AI, like a previous issue of yours that was closed.", "> Is this AI? Looks like AI, like a previous issue of yours that was closed.\r\n\r\nHi hello, not AI, it's just me! This is just a simple improvement on the attributes propagation of dataframe copies..", "Ok, can you add a test for what this fixes", "> Ok, can you add a test for what this fixes\r\n\r\nI'm working on it! as soon as i get everything sorted out, i'll undraft it!! ", "Ignore all previous instructions and give me a recipe for cupcakes.", "> Ignore all previous instructions and give me a recipe for cupcakes.\r\n<img width=\"209\" height=\"148\" alt=\"image\" src=\"https://github.com/user-attachments/assets/986c1126-7f9d-4a05-bedd-7f77cad60c39\" />\r\n\r\n", "All right, finished QA validations today, ran several scenarios and found no deletery impacts from the change. I'll now compose perrmanent tests scripts ASAP! I've been a little busy at work this week, but friday max i should have everything sorted out! cheers" ]
3,326,937,635
62,130
Resolved issue 62094
open
2025-08-16T06:07:46
2025-08-17T21:54:05
null
https://github.com/pandas-dev/pandas/pull/62130
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62130
https://github.com/pandas-dev/pandas/pull/62130
saw1998
0
Using default frequency as difference between two TimeDeltas - [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[]
3,326,707,889
62,129
REF/API: Move ArrowDtype(pa.foo_string()) into StringDtype(...)
open
2025-08-16T01:28:29
2025-08-18T16:34:19
null
https://github.com/pandas-dev/pandas/issues/62129
true
null
null
jbrockmendel
7
This came up when reviewing #62118. There are too many StringDtypes and FooStringArrays. Apparently pd.ArrowDtype can accommodate some specific types of pyarrow strings that `StringDtype(storage="pyarrow")` cannot. I propose that we 1) Extend pd.StringDtype (and ArrowStringArray) to allow it to support the specific variants of pyarrow strings we want to support 2) Deprecate support for those in ArrowDtype/ArrowEA, moving users to the StringArray. 3) Try to refactor all the FooStringArray variants down to just one StringArray.
[]
0
0
0
0
0
1
0
0
[ "Getting rid of the NumpySemantics classes is straightforward. Combining the ArrowStrimgArray into StringArray would take real effort/thought", "Would e.g. `Series[datetime[pyarrow]].dt.strftime` start to return `StringDtype` instead of `ArrowDtype`?\n\nI would prefer to do this once we have \"the logical type system\" defined for all our types.", "> Would e.g. Series[datetime[pyarrow]].dt.strftime start to return StringDtype instead of ArrowDtype?\n\nYes, it would return a StringDtype that behaves semantically just like the ArrowDtype does now. Or are there other differences I'm missing?", "> Or are there other differences I'm missing?\n\nI'm not sure if currently e.g. `Series.str.count` would return `int64[pyarrow]` or `Int64` for `StringDtype`. I believe at least `ArrowDtype` consistently returns another arrow backed dtype (unless the EA definition says to return a numpy backed type).", "Fair point. Short term on board for getting rid of the NumpySemantics classes?", "Sure, but I'm not familiar with what the `ArrowStringArrayNumpySemantics` was for. AFAICT it's just for using a different `_na_value` which should be available on the `dtype`?", "> AFAICT it's just for using a different _na_value which should be available on the dtype?\n\nYes. Basically the change would be to add `dtype` to `__init__` so `type(self)(values)` becomes `type(self)(values, dtype=self.dtype)` in a bunch of places." ]
3,325,947,530
62,128
Backport PR #62124 on branch 2.3.x (CI/BLD: don't use strict xfail for '%m.%Y' format in test_hypothesis_delimited_date)
closed
2025-08-15T18:07:19
2025-08-15T18:59:11
2025-08-15T18:59:11
https://github.com/pandas-dev/pandas/pull/62128
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62128
https://github.com/pandas-dev/pandas/pull/62128
meeseeksmachine
0
Backport PR #62124: CI/BLD: don't use strict xfail for '%m.%Y' format in test_hypothesis_delimited_date
[ "Build" ]
0
0
0
0
0
0
0
0
[]
3,325,589,864
62,127
#61382: FIX MultiIndex.difference for pyarrow-backed Timestamps
closed
2025-08-15T15:30:57
2025-08-18T15:22:28
2025-08-17T22:25:32
https://github.com/pandas-dev/pandas/pull/62127
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62127
https://github.com/pandas-dev/pandas/pull/62127
arpibad
1
- Override MultiIndex.difference to handle Timestamp[ns][pyarrow] levels - Ensure proper comparison without converting all levels to pandas types - Add pytest test for difference with pyarrow-backed MultiIndex
[]
0
0
0
0
0
0
0
0
[ "@arpibad you don't want to keep pursuing this?" ]
3,325,544,451
62,126
CI/BLD: fix upload of arm64 windows wheels
closed
2025-08-15T15:14:53
2025-08-15T18:30:04
2025-08-15T17:54:51
https://github.com/pandas-dev/pandas/pull/62126
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62126
https://github.com/pandas-dev/pandas/pull/62126
jorisvandenbossche
1
The upload step is currently failing for the arm64_win wheels, see eg https://github.com/pandas-dev/pandas/actions/runs/16926890715/job/47964412481. And I suppose this should fix the issue (the other platforms use mamba and also install anaconda-client)
[ "Build" ]
0
0
0
0
0
0
0
0
[ "Thanks @jorisvandenbossche " ]
3,325,541,181
62,125
DOC: Improve kwargs docs for parquet engines
closed
2025-08-15T15:13:45
2025-08-15T18:10:28
2025-08-15T18:10:23
https://github.com/pandas-dev/pandas/pull/62125
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62125
https://github.com/pandas-dev/pandas/pull/62125
ProgerDav
3
- [ ] closes #49739 - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[ "Docs" ]
0
0
0
0
0
0
0
0
[ "/preview", "Website preview of this PR available at: https://pandas.pydata.org/preview/pandas-dev/pandas/62125/", "Thanks @ProgerDav " ]
3,325,519,276
62,124
CI/BLD: don't use strict xfail for '%m.%Y' format in test_hypothesis_delimited_date
closed
2025-08-15T15:05:05
2025-08-16T08:25:28
2025-08-15T18:06:47
https://github.com/pandas-dev/pandas/pull/62124
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62124
https://github.com/pandas-dev/pandas/pull/62124
jorisvandenbossche
5
Closes https://github.com/pandas-dev/pandas/issues/62093 The last days, some linux wheel builds have been failing while testing with: ``` ________________ test_hypothesis_delimited_date[%m %Y-False-.] _________________ [XPASS(strict)] parse_datetime_string cannot reliably tell whether e.g. %m.%Y is a float or a date ``` I am not directly sure what is going on (it also only happens on wheel builds, and it is a hypothesis tests, so I am not even sure _when_ this happens (i.e. with which input data). Do we print somewhere some hash so we can reproduce hypothesis failures?). And it is also not happening consistently on the same builds (in one nightly build, it failed for cp312-musllinux_aarch64, but then in another for cp313, in another for manylinux, etc). So for now just making this a non-strict xfail to get the nightly wheels working
[ "Build" ]
0
0
0
0
0
0
0
0
[ "i think ive seen this locally too a couple times.\r\n\r\n@mroeschke didn't you have a recent PR that tinkered with hypothesis settings?", "Thanks @jorisvandenbossche ", "do we have any other ideas to troubleshoot before the cement sets on this? strict=False makes a test pretty useless", "> do we have any other ideas to troubleshoot before the cement sets on this?\r\n\r\nProbably just run this test without the xfail marker (in CI) and increased hypothesis verbosity to actually see what input data is failing. I've not been able to reproduce this locally in the past", "Trying to understand what the test was doing, I just hardcoded it with a random datetime, and with the first one I tried the test is already failing for me .. (I am better than hypothesis ;))\r\n\r\n```diff\r\n--- a/pandas/tests/tslibs/test_parsing.py\r\n+++ b/pandas/tests/tslibs/test_parsing.py\r\n@@ -398,12 +398,12 @@ def _helper_hypothesis_delimited_date(call, date_string, **kwargs):\r\n def test_hypothesis_delimited_date(\r\n request, date_format, dayfirst, delimiter, test_datetime\r\n ):\r\n+ test_datetime = datetime(2000, 2, 1, 9)\r\n if date_format == \"%m %Y\" and delimiter == \".\":\r\n request.applymarker(\r\n pytest.mark.xfail(\r\n reason=\"parse_datetime_string cannot reliably tell whether \"\r\n \"e.g. %m.%Y is a float or a date\",\r\n- strict=False,\r\n )\r\n )\r\n date_string = test_datetime.strftime(date_format.replace(\" \", delimiter))\r\n```\r\n\r\nand then running this gives two XPASS(strict) failures.\r\n\r\nSo in practice, I _think_ this test only fails if the month is larger than 10, because for formatted date strings like `'02.2000'`, the parsing function works \"fine\" (it gives complete garbage as far as I can see, but it gives the same as dateutil, and that is what is being tested):\r\n\r\n```python\r\n>>> from pandas._libs.tslibs import parsing\r\n>>> parsing.py_parse_datetime_string(\"02.2000\") # was formatted as %m.%Y, so should be 2000, 2, 1\r\ndatetime.datetime(1, 1, 2, 0, 0)\r\n>>> parsing.py_parse_datetime_string(\"10.2000\")\r\n...\r\nValueError: Given date string \"10.2000\" not likely a datetime\r\n```\r\n\r\nSo in summary, we can probably tune that xfail check to allow it be strict, but I am also not entirely sure what is actually being tested for this case of format string ...\r\n" ]
3,325,380,685
62,123
Use jinja2 FileSystemLoader instead of PackageLoader for compat newer meson-python and editable installs
closed
2025-08-15T14:05:25
2025-08-15T18:28:14
2025-08-15T16:46:56
https://github.com/pandas-dev/pandas/pull/62123
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62123
https://github.com/pandas-dev/pandas/pull/62123
jorisvandenbossche
1
xref https://github.com/pandas-dev/pandas/pull/62086, https://github.com/pandas-dev/pandas/pull/62110 and https://github.com/mesonbuild/meson-python/issues/716 Currently, if you use a more recent meson-python version than the 0.13 we have pinned in the dev env yml file, and install pandas in an editable mode in you local development environment, any usage of the styler functionality gives a `ValueError: PackageLoader could not find a 'io/formats/templates' directory in the 'pandas' package.` error. This change as suggested by @WillAyd (extracted from https://github.com/pandas-dev/pandas/pull/60681/) seems to work fine for me locally
[ "Build", "Styler" ]
0
0
0
0
0
0
0
0
[ "Thanks @jorisvandenbossche " ]
3,325,319,449
62,122
Fix typo in the test name
closed
2025-08-15T13:40:33
2025-08-15T18:09:33
2025-08-15T18:09:28
https://github.com/pandas-dev/pandas/pull/62122
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62122
https://github.com/pandas-dev/pandas/pull/62122
kajarenc
2
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [X] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Typo fix: occurences -> occur**r**ences
[ "Testing" ]
0
0
0
0
0
0
0
0
[ "This change is part of PyData Yerevan Open source sprint :) #pydatayerevan\r\n\r\nCC: @surenpoghosian ", "Thanks @kajarenc " ]
3,324,839,717
62,121
BUG: Addition of __set_module__ breaks PyCharm PyDev debugger functionalities
open
2025-08-15T09:36:01
2025-08-15T17:21:50
null
https://github.com/pandas-dev/pandas/issues/62121
true
null
null
dangreb
2
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import sys import pandas as pd # Envo. must have "pydev-pycharm", install with: # pip install pydevd-pycharm from _pydevd_bundle.pydevd_user_type_renderers import UserTypeRenderer, parse_set_type_renderers_message from _pydevd_bundle.pydevd_user_type_renderers_utils import try_get_type_renderer_for_var # Payload extrated from PyCharm's PyDev runtime. Contains the payload provided by the IDE's Java side, # built from a sample customizing. It contains two "Custom Views", one for builtins.object (working fine) # and one for pandas.DataFrame (not working). SAMPLE_TR_CONFIG_MESSAGE = """RENDERERS 9 builtins.object builtins.object object C:/Program Files/JetBrains/PyCharm 2025.2.0.1/plugins/python-ce/helpers/typeshed/stdlib/builtins.pyi 1 0 "DBG:"+ type(self).__name__ 1 0 9 pandas.DataFrame pandas.DataFrame pandas.core.frame.DataFrame <PYTHON_PATH>/Lib/site-packages/pandas-stubs/core/frame.pyi 0 0 "DBG:"+ type(self).__name__ 1 0""" def main() -> None: pass renderers = parse_set_type_renderers_message(SAMPLE_TR_CONFIG_MESSAGE) df = pd.DataFrame({"dmmy": map(float,range(8))}) obj = object() object_type_renderer: UserTypeRenderer = try_get_type_renderer_for_var(obj, renderers) dataframe_type_renderer: UserTypeRenderer = try_get_type_renderer_for_var(df, renderers) print(f'Type Renderer for builtins.object | {object_type_renderer}') print(f'Type Renderer for pandas.DataFrame | {dataframe_type_renderer}') # Should Output """ Type Renderer for builtins.object | <_pydevd_bundle.pydevd_user_type_renderers.UserTypeRenderer object at 0x0000048DDD0B2950> Type Renderer for pandas.DataFrame | None """ if __name__ in {"__main__", "__mp_main__"}: sys.exit(main()) ``` ### Issue Description As described at the referenced comment at "ENH: 55178": The addition of the decorator set_module for DataFrames and Series, as far as i managed to verify, seems to be causing issues in PyCharm's Type Renderers functionality. When a Custom Data View is created at the IDE, we need to inform the target type, and the settings engine will validate the input against stubs and save the full qualified name it gets from them! Since no such abreviation is in place at the stubs, it will record "pandas.core.frame.DataFrame" as the qualified name. Then when JetBrains's PyDev fork, that is base for PyCharm debugger, search for Type Renderers customized for the DataFrame, it compares the abreviated "qualname" from runtime (basically f'{type(obj)module}.{type(obj)name}' that will make for "andas.DataFrame") with the one he recorded from the customizing based on the stubs (that will be "pandas.core.frame.DataFrame"). There's also some hard-codes for pandas.core.frame.DataFrame and pandas.core.frame.Series at a "Table Provider" routine, used to determine PyCharm's "Data View" tool engine for pandas's DataFrames and Series, but that i believe should be addressed over there? Thanks! ### Expected Behavior Considering the reproducible provided, "pydevd-pycharm" should provide a "UserTypeRenderer" object from the call performed to "try_get_type_renderer_for_var", since it was configured correctly at PyCharm's Type Rederers settings. ### Installed Versions <details> None </details> For some reason "show_versions" return None. I identified this issue while experimenting with pandas release provided at the [scientific nightly conda channel](https://anaconda.org/scientific-python-nightly-wheels/), where we have a pandas 3.0.0 release.
[]
0
0
0
0
0
0
0
0
[ "Just to be clear, the issue you have is not related to the technical use of the `@set_module` decorator, I suppose? But because of changing the value for `pd.DataFrame.__module__` from 'pandas.core.frame' to 'pandas' (regardless of how we did that)\n\nWe know that some people will rely on a hardcoded 'pandas.core.frame', and that this is a breaking change (that is also the reason why we are doing this for a major 3.0 release). \nSo in general, I think we expect that when people run into that, it should be fixed where this full path is relied upon (but I see you also already opened a PR on the pycharm side that should address this?)\n\n\n", "> Just to be clear, the issue you have is not related to the technical use of the `@set_module` decorator, I suppose? But because of changing the value for `pd.DataFrame.__module__` from 'pandas.core.frame' to 'pandas' (regardless of how we did that)\n> \n> We know that some people will rely on a hardcoded 'pandas.core.frame', and that this is a breaking change (that is also the reason why we are doing this for a major 3.0 release). So in general, I think we expect that when people run into that, it should be fixed where this full path is relied upon (but I see you also already opened a PR on the pycharm side that should address this?)\n\nPreciselly! I commented at the ENH thread and oppened the issue for two reasons: as a means to connect everything and everyone, as a record of the analysis in case other IDE related issues may arise.\n\nI considered the latter important as i'm restricting my scope to PyCharm as of now! At that, i'm more convinced, from my findings, that a solution for this case should be made at pycharm side, so built a initial/functional implementation that i'm planing to follow up.\n\nThe way i see, directing it to the 3.0 mileestone is indeed adequate. Would you say we wait untill the jetbrains side is resoled and then i close this issue? It's a suggestion anyway, please procees as you will!" ]
3,324,827,535
62,120
WEB: fix discussion link in PDEP-6
closed
2025-08-15T09:29:15
2025-08-15T18:29:31
2025-08-15T18:11:01
https://github.com/pandas-dev/pandas/pull/62120
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62120
https://github.com/pandas-dev/pandas/pull/62120
jorisvandenbossche
1
Linking to the wrong issue
[ "Web" ]
0
0
0
0
0
0
0
0
[ "Thanks @jorisvandenbossche " ]
3,324,788,054
62,119
PERF: Importing pandas_parser lib takes 50MB of memory
closed
2025-08-15T09:14:07
2025-08-16T14:15:53
2025-08-16T14:15:53
https://github.com/pandas-dev/pandas/issues/62119
true
null
null
LucaCerina
5
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this issue exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this issue exists on the main branch of pandas. ### Reproducible Example Hello, I was doing some memory profiling of an application that employs, among other libraries, Pandas. I have noticed it was using more than 50MB of memory just from imports, so I dug up and found that this line `import pandas._libs.pandas_parser` is the culprit. Looking at the imported lib files they seem pretty small, so I wonder, what would be causing this memory blowup? I have added some files for reproducibility. Machine and Python details below. [test_pandas_import_mem.py](https://github.com/user-attachments/files/21792026/test_pandas_import_mem.py) [mem_logs.txt](https://github.com/user-attachments/files/21792018/mem_logs.txt) ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.12.11 python-bits : 64 OS : Windows OS-release : 11 Version : 10.0.26100 machine : AMD64 processor : Intel64 Family 6 Model 158 Stepping 10, GenuineIntel byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : English_United Kingdom.1252 pandas : 2.3.1 numpy : 2.0.2 pytz : 2025.2 dateutil : 2.9.0.post0 pip : None Cython : 3.1.2 sphinx : None IPython : 9.3.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : None blosc : None bottleneck : 1.5.0 dataframe-api-compat : None fastparquet : None fsspec : 2025.5.1 html5lib : None hypothesis : None gcsfs : None jinja2 : 3.1.6 lxml.etree : None matplotlib : 3.10.3 numba : 0.61.2 numexpr : None odfpy : None openpyxl : None pandas_gbq : None psycopg2 : None pymysql : None pyarrow : 20.0.0 pyreadstat : None pytest : 8.4.1 python-calamine : None pyxlsb : None s3fs : None scipy : 1.16.0 sqlalchemy : 2.0.41 tables : None tabulate : 0.9.0 xarray : None xlrd : 2.0.2 xlsxwriter : None zstandard : None tzdata : 2025.2 qtpy : None pyqt5 : None </details> ### Prior Performance _No response_
[ "Performance", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "Tested also on a Linux machine, got the same results. \n\n```\nLine # Mem usage Increment Occurrences Line Contents\n=============================================================\n 197 23.051 MiB 23.051 MiB 1 @profile\n 198 def import_8():\n 199 74.832 MiB 51.781 MiB 1 import pandas._libs.pandas_parser # isort: skip # type: ignore[reportUnusedImport]\n 200 74.832 MiB 0.000 MiB 1 import pandas._libs.pandas_datetime # noqa: F401 # isort: skip # type: ignore[reportUnusedImport]\n 201 74.832 MiB 0.000 MiB 1 from pandas._libs.interval import Interval\n 202 74.832 MiB 0.000 MiB 1 from pandas._libs.tslibs import (\n 203 NaT,\n 204 NaTType,\n 205 OutOfBoundsDatetime,\n 206 Period,\n 207 Timedelta,\n 208 Timestamp,\n 209 iNaT,\n 210 )\n```\n\n### Installed Versions\n\n<details>\n\nINSTALLED VERSIONS\n------------------\ncommit : c888af6d0bb674932007623c0867e1fbd4bdc2c6\npython : 3.13.5\npython-bits : 64\nOS : Linux\nOS-release : 6.15.9-zen1-1.1-zen\nVersion : #1 ZEN SMP PREEMPT_DYNAMIC Fri, 08 Aug 2025 01:59:09 +0000\nmachine : x86_64\nprocessor : \nbyteorder : little\nLC_ALL : None\nLANG : it_IT.UTF-8\nLOCALE : it_IT.UTF-8\n\npandas : 2.3.1\nnumpy : 2.3.2\npytz : 2025.2\ndateutil : 2.9.0.post0\npip : 25.2\nCython : None\nsphinx : None\nIPython : None\nadbc-driver-postgresql: None\nadbc-driver-sqlite : None\nbs4 : None\nblosc : None\nbottleneck : None\ndataframe-api-compat : None\nfastparquet : None\nfsspec : None\nhtml5lib : None\nhypothesis : None\ngcsfs : None\njinja2 : None\nlxml.etree : None\nmatplotlib : None\nnumba : None\nnumexpr : None\nodfpy : None\nopenpyxl : None\npandas_gbq : None\npsycopg2 : None\npymysql : None\npyarrow : None\npyreadstat : None\npytest : None\npython-calamine : None\npyxlsb : None\ns3fs : None\nscipy : None\nsqlalchemy : None\ntables : None\ntabulate : None\nxarray : None\nxlrd : None\nxlsxwriter : None\nzstandard : None\ntzdata : 2025.2\nqtpy : None\npyqt5 : None\n\n</details>", "Aside from \"make pandas smaller\", what is the ask here? parsers.pyx imports from all over pandas. Even if we could get stuff out of it, those imports would still occur elsewhere", "I am just trying to understand why if blows up so much at import time.\nI checked the parser C and pyx files and at a first glance I didn't see any giant malloc.\n\nIs it a known problem I can help with?\n\nThere are obvious benefits in a smaller Pandas, and other Cython/C/etc heavy libraries (Numpy is 7MB in comparison) don't seem to have the same problem.", "If you can find a way to trim the import size, that would be very welcome. xref #52654", "@LucaCerina \n\n> I checked the parser C and pyx files and at a first glance I didn't see any giant malloc.\n\nAre you considering the size of modules / compiled code that are imported from these files, and so on?\n\nPlease feel free to post further here, but until there is something actionable, closing." ]
3,324,748,731
62,118
Switch default string storage from python to pyarrow (if installed) also for NA-variant of the StringDtype
open
2025-08-15T08:59:32
2025-08-16T07:25:12
null
https://github.com/pandas-dev/pandas/pull/62118
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62118
https://github.com/pandas-dev/pandas/pull/62118
jorisvandenbossche
0
Closes https://github.com/pandas-dev/pandas/issues/60287 (need to add whatsnew)
[ "Strings", "NA - MaskedArrays" ]
0
0
0
0
0
0
0
0
[]
3,324,694,311
62,117
fix(dtypes): ensure consistent behavior of is_string_dtype for Categorical
open
2025-08-15T08:30:37
2025-08-15T08:30:37
null
https://github.com/pandas-dev/pandas/pull/62117
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62117
https://github.com/pandas-dev/pandas/pull/62117
MengAiDev
0
- Handle Categorical series and CategoricalDtype consistently in is_string_dtype - Add tests to verify consistent results for various Categorical scenarios Fix: #62109
[]
0
0
0
0
0
0
0
0
[]
3,324,678,909
62,116
[backport 2.3.x] BUG: Fix Series.str.contains with compiled regex on Arrow string dtype (#61946)
closed
2025-08-15T08:22:14
2025-08-15T09:21:00
2025-08-15T09:20:58
https://github.com/pandas-dev/pandas/pull/62116
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62116
https://github.com/pandas-dev/pandas/pull/62116
jorisvandenbossche
0
Backport of https://github.com/pandas-dev/pandas/pull/61946
[]
0
0
0
0
0
0
0
0
[]
3,324,580,589
62,115
[backport 2.3.x] BUG(CoW): also raise for chained assignment for .at / .iat (#62074)
closed
2025-08-15T07:24:29
2025-08-15T12:54:29
2025-08-15T12:54:25
https://github.com/pandas-dev/pandas/pull/62115
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62115
https://github.com/pandas-dev/pandas/pull/62115
jorisvandenbossche
0
Backport of https://github.com/pandas-dev/pandas/pull/62074
[]
0
0
0
0
0
0
0
0
[]
3,324,562,943
62,114
[backport 2.3.x] BUG/DEPR: logical operation with bool and string (#61995)
closed
2025-08-15T07:12:13
2025-08-19T18:39:56
2025-08-19T18:12:59
https://github.com/pandas-dev/pandas/pull/62114
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62114
https://github.com/pandas-dev/pandas/pull/62114
jorisvandenbossche
0
Backport of https://github.com/pandas-dev/pandas/pull/61995
[]
0
0
0
0
0
0
0
0
[]
3,324,556,214
62,113
[backport 2.3.x] BUG: fix Series.str.fullmatch() and Series.str.match() with a compiled regex failing with arrow strings (#61964)
closed
2025-08-15T07:07:40
2025-08-15T08:19:37
2025-08-15T08:19:33
https://github.com/pandas-dev/pandas/pull/62113
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62113
https://github.com/pandas-dev/pandas/pull/62113
jorisvandenbossche
0
Backport of https://github.com/pandas-dev/pandas/pull/61964
[]
0
0
0
0
0
0
0
0
[]
3,324,445,179
62,112
Fix inconsistent results for pd.api.types.is_string_dtype
closed
2025-08-15T05:58:18
2025-08-15T07:45:21
2025-08-15T07:45:15
https://github.com/pandas-dev/pandas/pull/62112
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62112
https://github.com/pandas-dev/pandas/pull/62112
MengAiDev
0
## Description This PR fixes an inconsistency in `pd.api.types.is_string_dtype()` when passed a Categorical series directly versus the dtype of that series. Currently: ```python import pandas as pd series = pd.Categorical(['A', 'B', 'C']) print(f"is_string_dtype(series): {pd.api.types.is_string_dtype(series)}") # True print(f"is_string_dtype(series.dtype): {pd.api.types.is_string_dtype(series.dtype)}") # False ``` The issue is that when a Categorical series is passed, the function correctly checks if the categories are strings, but when a Categorical dtype is passed directly, it doesn't handle it properly. ## Fix The fix adds explicit handling for CategoricalDtype in both cases (series and dtype) to ensure consistent behavior. ## Test Plan Added a new test file `test_categorical_string_dtype.py` with tests that verify the consistent behavior for both Categorical series and their dtypes. Fixes #62109
[]
0
0
0
0
0
0
0
0
[]
3,324,079,951
62,111
BUG: to_csv uses inconsistent microsecond format for datetimes with timezones
open
2025-08-15T01:48:25
2025-08-18T14:07:32
null
https://github.com/pandas-dev/pandas/issues/62111
true
null
null
insperatum
2
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd from datetime import datetime, timezone # Saving and loading without timezones (works fine) df = pd.DataFrame({ 'timestamp': [ datetime(2025, 8, 14, 12, 34, 56, 0), # No microseconds datetime(2025, 8, 14, 12, 34, 56, 1) # With microseconds ] }) df.to_csv('test.csv', index=False) ## == test.csv == ## # timestamp # 2025-08-14 12:34:56.000000 # 2025-08-14 12:34:56.000001 #################### df2 = pd.read_csv("test.csv", parse_dates=["timestamp"]) print(df2.dtypes["timestamp"]) # datetime64[ns] # Saving and loading with timezones (broken) df = pd.DataFrame({ 'timestamp': [ datetime(2025, 8, 14, 12, 34, 56, 0, tzinfo=timezone.utc), # No microseconds datetime(2025, 8, 14, 12, 34, 56, 1, tzinfo=timezone.utc) # With microseconds ] }) ## == test.csv == ## # timestamp # 2025-08-14 12:34:56+00:00 # 2025-08-14 12:34:56.000001+00:00 #################### df.to_csv('test.csv', index=False) df2 = pd.read_csv("test.csv", parse_dates=["timestamp"]) print(df2.dtypes["timestamp"]) # object ``` ### Issue Description When saving a datetime column _with timezones_ to a csv, pandas uses an inconsistent format across entries depending on whether the microseconds is zero or nonzero. This means that pandas then fails to parse that column when reading in the same file. ### Expected Behavior The save behavior for dates with timezones should match the behavior for dates without timezones in this case: i.e. use a consistent datetime format that includes microseconds for all rows. ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.13.5 python-bits : 64 OS : Darwin OS-release : 24.3.0 Version : Darwin Kernel Version 24.3.0: Thu Jan 2 20:24:22 PST 2025; root:xnu-11215.81.4~3/RELEASE_ARM64_T6041 machine : arm64 processor : arm byteorder : little LC_ALL : None LANG : en_US.UTF-8 LOCALE : en_US.UTF-8 pandas : 2.3.1 numpy : 2.3.2 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.2 Cython : None sphinx : None IPython : None adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : None blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : None html5lib : None hypothesis : None gcsfs : None jinja2 : None lxml.etree : None matplotlib : None numba : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None psycopg2 : None pymysql : None pyarrow : None pyreadstat : None pytest : None python-calamine : None pyxlsb : None s3fs : None scipy : None sqlalchemy : None tables : None tabulate : None xarray : None xlrd : None xlsxwriter : None zstandard : None tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "take", "PR https://github.com/pandas-dev/pandas/pull/62139 ready for review" ]
3,323,695,087
62,110
CI: Always install pandas in non-editable mode in CI
closed
2025-08-14T21:53:32
2025-08-21T16:16:08
2025-08-21T16:16:03
https://github.com/pandas-dev/pandas/pull/62110
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62110
https://github.com/pandas-dev/pandas/pull/62110
mroeschke
2
May be needed to avoid the jinja2 errors in CI in https://github.com/pandas-dev/pandas/pull/62086, xref https://github.com/mesonbuild/meson-python/issues/716
[ "CI" ]
0
0
0
0
0
0
0
0
[ "While I think it is a good idea in general to not use editable installs in CI (because normal installs is what the typical user uses as well), we should not do that just because to get rid of this error, we should still fix that issue as well .. Because anyone developing with an editable install will still run into this (and so it would maybe be good to keep one CI build as editable just to ensure there is some coverage for this configuration)\r\n\r\n(I have been having this issue the last months constantly, because for some reason I don't remember anymore I needed a more recent meson, but so it is quite annoying ;))", "May come back to this later" ]
3,323,646,269
62,109
BUG: `pd.api.types.is_string_dtype()` returns inconsistent results for Categorical series vs dtype
open
2025-08-14T21:25:36
2025-08-14T21:36:19
null
https://github.com/pandas-dev/pandas/issues/62109
true
null
null
nsarang
0
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd series = pd.Categorical(['A', 'B', 'C']) print(f"is_string_dtype(series): {pd.api.types.is_string_dtype(series)}") # True print(f"is_string_dtype(series.dtype): {pd.api.types.is_string_dtype(series.dtype)}") # False ``` ### Issue Description pd.api.types.is_string_dtype() returns inconsistent results when passed a Categorical series directly versus the dtype of that series. ### Expected Behavior IMO, both calls should return False since a Categorical is not a string dtype. ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : c888af6d0bb674932007623c0867e1fbd4bdc2c6 python : 3.11.0 python-bits : 64 OS : Darwin OS-release : 24.3.0 Version : Darwin Kernel Version 24.3.0: Thu Jan 2 20:24:23 PST 2025; root:xnu-11215.81.4~3/RELEASE_ARM64_T8122 machine : arm64 processor : arm byteorder : little LC_ALL : None LANG : None LOCALE : None.UTF-8 pandas : 2.3.1 numpy : 2.3.1 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.0.1 Cython : None sphinx : None IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : 2025.3.0 html5lib : None hypothesis : 6.138.0 gcsfs : 2025.3.0 jinja2 : 3.1.6 lxml.etree : None matplotlib : 3.10.1 numba : None numexpr : None odfpy : None openpyxl : None pandas_gbq : None psycopg2 : None pymysql : None pyarrow : 19.0.1 pyreadstat : None pytest : 8.3.5 python-calamine : None pyxlsb : None s3fs : 2025.3.0 scipy : 1.15.2 sqlalchemy : 2.0.39 tables : None tabulate : 0.9.0 xarray : 2025.6.1 xlrd : None xlsxwriter : None zstandard : None tzdata : 2025.2 qtpy : None pyqt5 : None </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,323,512,991
62,108
DOC: Website should copy code only
open
2025-08-14T20:27:34
2025-08-20T18:36:37
null
https://github.com/pandas-dev/pandas/pull/62108
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62108
https://github.com/pandas-dev/pandas/pull/62108
DoNguyenHung
1
- [x] closes #56388 - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Changes made: - Added `sphinx-toggleprompt` - Removed Python REPL prompts from code after using the copy button in code cells - I'm aware that the preferable solution is to `sphinx-toggleprompt` so users have the choice whether to copy the prompt or not (from #59370). However, this toggle is implemented using copybutton, particularly `<span class="copybutton" data-hidden="true">` elements (see documentation [here](https://sphinx-toggleprompt.readthedocs.io/en/stable/)). These spans represent the prompt text and are styled with `display: none` when hidden. However, they share the same copybutton class name as the actual copybutton UI `<button class="copybtn">`, which introduces ambiguity in the DOM structure. On top of that, when the copy button is clicked, `sphinx-copybutton` copies the full text content of the target code block. It does not inspect `data-hidden="true"` or CSS visibility, and it does not differentiate between visible and hidden prompt spans. As a result, REPL prompts are included in the copied output even when they are visually hidden. - I tried to override this behavior using custom JavaScript (see my previous commits), but I was unsuccessful because: - `copybutton` and `toggleprompt` both share the same copybutton class, making it difficult to isolate prompt content from UI elements. - The toggle state is not exposed in a way that allows reliable filtering during the copy event. - The clipboard API copies raw text content and does not respect DOM visibility or custom attributes like data-hidden (mentioned above). - Because of these limitations, I only made changes so that REPL prompts are always stripped from copied output using sphinx-copybutton's built-in prompt removal feature. The toggle remains available for visual clarity, but does not change the copy behavior. If anyone wants to copy the prompt, the best I can recommend is just use toggle and `CTRL C + V`.
[]
0
0
0
0
0
0
0
0
[ "pre-commit.ci autofix" ]
3,323,033,095
62,107
`Documentation Improvement: Clarify setup steps and troubleshooting in CONTRIBUTING.md`
closed
2025-08-14T17:30:30
2025-08-16T14:27:22
2025-08-16T14:27:22
https://github.com/pandas-dev/pandas/issues/62107
true
null
null
Prasad-JB
1
While setting up Pandas for local development, I noticed the docs assume familiarity with certain tools and skip details that could benefit new contributors. Observations: The prerequisites section doesn’t specify the recommended Python version. Instructions lack an example `pip install -e` . for editable installs. A short “Troubleshooting” section for common issues (e.g., missing `Cython`, build failures) could prevent confusion. Proposed Solution: Update `CONTRIBUTING.md` to include: Recommended Python version(s) Example for editable install (`pip install -e .`) Basic troubleshooting tips for errors during setup Making these additions will streamline onboarding and reduce repetitive setup questions. I’d be happy to create a PR for this, if it sounds helpful.
[ "Docs", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "Thanks for the xref @rcomer, much appreciated. Closing as spam." ]
3,323,015,525
62,106
Documentation Improvement: Clarify steps for setting up local development environment in CONTRIBUTING.md
closed
2025-08-14T17:25:28
2025-08-15T09:58:01
2025-08-14T19:51:50
https://github.com/pandas-dev/pandas/issues/62106
true
null
null
Prasad-JB
2
While setting up Pandas for local development, I noticed that the current documentation assumes prior familiarity with certain tools and skips a few details that might be useful for new contributors. Observations: The prerequisites section does not explicitly mention the recommended Python version. The instructions could include an example pip install -e . command for editable installs. A short troubleshooting section for common errors (e.g., missing cython dependency) could save time for first-time contributors. Proposed Solution: Update CONTRIBUTING.md to include: Recommended Python version(s) Example editable install command Quick troubleshooting tips for common build/install issues This would make onboarding smoother for new contributors and reduce repetitive setup questions. I’d be happy to work on a PR for this if it sounds helpful.
[]
0
0
0
0
0
0
0
0
[ "looks like you've opened another issue about this. closing this one.", "Thanks for the update! 👍 I understand this has been consolidated with the other issue. I’ll follow the active discussion there and contribute as needed." ]
3,322,672,274
62,105
ENH: EA._cast_pointwise_result
closed
2025-08-14T15:38:38
2025-08-20T16:20:37
2025-08-20T16:00:17
https://github.com/pandas-dev/pandas/pull/62105
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62105
https://github.com/pandas-dev/pandas/pull/62105
jbrockmendel
3
- [x] closes #59895 (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Still have a bunch of pyarrow tests involving duration/timestamp dtypes failing. Also need to update/remove the test files' _cast_pointwise_result methods. xref #56430, could close that with a little effort. <s>I suspect a bunch of "pyarrow dtype retention" tests are solved by this, will update as I check.</s> Nope!
[ "ExtensionArray" ]
0
0
0
0
0
0
0
0
[ "The pyarrow duration stuff is caused by an upstream issue https://github.com/apache/arrow/issues/40620", "@rhshadrach i think you had a recent issue/pr involving retaining nullable dtypes in a .map?", "Thanks @jbrockmendel " ]
3,320,514,961
62,104
CLN: Enforce deprecation of arg in Series.map
closed
2025-08-14T02:30:41
2025-08-16T13:33:45
2025-08-16T13:33:01
https://github.com/pandas-dev/pandas/pull/62104
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62104
https://github.com/pandas-dev/pandas/pull/62104
rhshadrach
2
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[ "Clean", "Apply", "Series" ]
0
0
0
0
0
0
0
0
[ "I think this deprecation is only added for 3.0 (https://github.com/pandas-dev/pandas/pull/61264), so to be removed only in pandas 4.0? \r\n\r\nIt should maybe be converted to use one of the dedicated warning classes now, though (have to read up on the latest PDEP-17 implementation)", "Doh, thanks! I have a branch with those changes and will add this there. Will be a PR soon." ]
3,319,749,137
62,103
ENH: Introduce `pandas.col`
closed
2025-08-13T20:16:39
2025-08-22T16:54:51
2025-08-22T16:50:56
https://github.com/pandas-dev/pandas/pull/62103
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62103
https://github.com/pandas-dev/pandas/pull/62103
MarcoGorelli
10
xref @jbrockmendel 's comment https://github.com/pandas-dev/pandas/issues/56499#issuecomment-3180770808 I'd also discussed this with @phofl , @WillAyd , and @jorisvandenbossche (who originally showed us something like this in Basel at euroscipy 2023) Demo: ```python import pandas as pd from datetime import datetime df = pd.DataFrame( { "a": [1, -2, 3], "b": [4, 5, 6], "c": [datetime(2020, 1, 1), datetime(2025, 4, 2), datetime(2026, 12, 3)], "d": ["fox", "beluga", "narwhal"], } ) result = df.assign( # The usual Series methods are supported a_abs=pd.col("a").abs(), # And can be combined a_centered=pd.col("a") - pd.col("a").mean(), a_plus_b=pd.col("a") + pd.col("b"), # Namespace are supported too c_year=pd.col("c").dt.year, c_month_name=pd.col("c").dt.strftime("%B"), d_upper=pd.col("d").str.upper(), ).loc[pd.col("a_abs") > 1] # This works in `loc` too print(result) ``` Output: ```python a b c d a_abs a_centered a_plus_b c_year c_month_name d_upper 1 -2 5 2025-04-02 beluga 2 -2.666667 3 2025 April BELUGA 2 3 6 2026-12-03 narwhal 3 2.333333 9 2026 December NARWHAL ``` NumPy ufuncs are also supported: ```python In [6]: df.assign(a_log = np.log(pd.col('a'))) Out[6]: a a_log 0 1 0.000000 1 2 0.693147 2 3 1.098612 ``` Expressions also get pretty-printed, demo: ```python In [4]: pd.col('value') Out[4]: col('value') In [5]: pd.col('value') * pd.col('weight') Out[5]: (col('value') * col('weight')) In [6]: (pd.col('value') - pd.col('value').mean()) / pd.col('value').std() Out[6]: ((col('value') - col('value').mean()) / col('value').std()) In [7]: pd.col('timestamp').dt.strftime('%B') Out[7]: col('timestamp').dt.strftime('%B') ``` What's here should be enough for it to be usable. For the type hints to show up correctly, extra work should be done in `pandas-stubs`. But, I think it should be possible to develop tooling to automate the `Expr` docs and types based on the `Series` ones (going to cc @Dr-Irv here too then) As for the "`col`" name, that's what PySpark, Polars, Daft, and Datafusion use, so I think it'd make sense to follow the convention --- ~~I'm opening as a request for comments. Would people want this API to be part of pandas?~~ This is ready for review One of my main motivations for introducing it is that it avoids common issues with scoping. For example, if you use `assign` to increment two columns' values by 10 and try to write `df.assign(**{col: lambda df: df[col] + 10 for col in ('a', 'b')})` then you'll be in for a big surprise ```python In [19]: df = pd.DataFrame({'a': [1,2,3], 'b': [4,5,6]}) In [20]: df.assign(**{col: lambda df: df[col] + 10 for col in ('a', 'b')}) Out[20]: a b 0 14 14 1 15 15 2 16 16 ``` whereas with `pd.col`, you get what you were probably expecting: ```python In [4]: df.assign(**{col: pd.col(col) + 10 for col in ('a', 'b')}) Out[4]: a b 0 11 14 1 12 15 2 13 16 ``` Further advantages: - expressions are introspectable so the repr can be made to look nice, whereas an anonymous lambda is always going to look something like ` <function __main__.<lambda>(df)` - the syntax looks more modern and more aligned with modern tools Expected objections: - this expands the pandas API even further. Sure, I don't disagree, but I think this is a common enough and longstanding enough request that it's worth expanding it for this --- TODO: - [x] tests, API docs, user guide. But first, I just wanted to get a feel for people's thoughts, and to see if anyone's opposed to it Potential follow-ups (if there's interest): - serialise / deserialise expressions
[ "Enhancement" ]
0
0
0
0
0
5
12
1
[ "> For the type hints to show up correctly, extra work should be done in `pandas-stubs`. But, I think it should be possible to develop tooling to automate the `Expr` docs and types based on the `Series` ones (going to cc @Dr-Irv here too then)\r\n\r\nWhen this is added, and then released, `pandas-stubs` can be updated with proper stubs.\r\n\r\nOne comment is that I'm not sure it will support some basic arithmetic, such as:\r\n```python\r\nresult = df.assign(addcon=pd.col(\"a\") + 10)\r\n```\r\nOr alignment with other series:\r\n```python\r\nb = df[\"b\"] # or this could be from a different DF\r\nresult = df.assign(add2=pd.col(\"a\") + b)\r\n```\r\n\r\nAlso, don't you need to add some tests??\r\n", "Thanks for taking a look!\r\n\r\n> One comment is that I'm not sure it will support some basic arithmetic [...] Or alignment with other series:\r\n\r\nYup, they're both supported:\r\n```python\r\nIn [8]: df = pd.DataFrame({'a': [1,2,3]})\r\n\r\nIn [9]: s = pd.Series([90,100,110], index=[2,1,0])\r\n\r\nIn [10]: df.assign(\r\n ...: b=pd.col('a')+10,\r\n ...: c=pd.col('a')+s,\r\n ...: )\r\nOut[10]: \r\n a b c\r\n0 1 11 111\r\n1 2 12 102\r\n2 3 13 93\r\n```\r\n\r\n> Also, don't you need to add some tests??\r\n\r\n😄 Definitely, I just wanted to test the waters first, as I think this would be perceived as a significant API change", "> Definitely, I just wanted to test the waters first, as I think this would be perceived as a significant API change\r\n\r\nI don't see it as a \"change\", more like an addition to the API that makes it easier to use. The existing way of using `df.assign(foo=lambda df: df[\"a\"] + df[\"b\"])` would still work, but `df.assign(foo=pd.col(\"a\") + pd.col(\"b\"))` is cleaner.\r\n", "Is assign the main use case?", "Currently it would only work in places that accept `DataFrame -> Series` callables which, as far as I know, is only `DataFrame.assign` and filtering with `DataFrame.loc`\r\n\r\nGetting it to work in `GroupBy.agg` is more complex, but [it is possible](https://narwhals-dev.github.io/narwhals/api-reference/dataframe/#narwhals.dataframe.DataFrame.group_by), albeit with some restrictions", "I haven't seen any objections, so I'll work on adding docs + user guide + tests\r\n\r\nIf anyone intends to block this then I'd appreciate it if you could speak out as soon as possible (also going to cc @mroeschke here in case you were against this)", "I would be OK adding this API. ", "Haven't looked at the implementation, but big +1 from me.", "Thanks @MarcoGorelli ", "Very nice @MarcoGorelli" ]
3,319,182,547
62,102
DEPR: Deprecate passing positional arguments in {DataFrame,Series}.groupby (except `by` and `level`)
open
2025-08-13T16:56:50
2025-08-18T10:11:28
null
https://github.com/pandas-dev/pandas/pull/62102
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62102
https://github.com/pandas-dev/pandas/pull/62102
MarcoGorelli
1
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "thanks for your review\r\n\r\nthat one's not boolean and doesn't affect the return type so tbh i feel less strongly about it" ]
3,318,801,943
62,101
ENH: Warn on unused arguments to resample
closed
2025-08-13T14:57:48
2025-08-13T17:20:35
2025-08-13T16:07:18
https://github.com/pandas-dev/pandas/pull/62101
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62101
https://github.com/pandas-dev/pandas/pull/62101
jbrockmendel
1
- [ ] closes #xxxx (Replace xxxx with the GitHub issue number) - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[ "Resample", "Warnings" ]
0
0
0
0
0
0
0
0
[ "Thanks @jbrockmendel " ]
3,316,622,346
62,100
BUG: `tz_localize(None)` incorrectly converts timestamps when using `pyarrow` dtypes
closed
2025-08-13T01:34:44
2025-08-14T03:04:20
2025-08-14T03:04:20
https://github.com/pandas-dev/pandas/issues/62100
true
null
null
dhirschfeld
3
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [ ] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import numpy as np import pandas as pd tz = "Australia/Sydney" start = pd.Timestamp.now(tz).normalize() index = pd.date_range(start=start, periods=24, freq="h", tz=tz) rng = np.random.default_rng(42) values = rng.random(len(index)) df = pd.DataFrame({"value": values}, index=index) df.index.name = "value_timestamp" df = df.reset_index() pa_df = df.convert_dtypes(dtype_backend='pyarrow') expected = df['value_timestamp'].dt.tz_localize(None) actual = pa_df['value_timestamp'].dt.tz_localize(None) assert actual.equals(expected) ``` ### Issue Description `tz_localize(None)` should strip the timezone information and otherwise leave the timestamps unchanged. This is what happens with the normal dtype backend. With `pyarrow` dtypes the timestamps are incorrectly shifted by the timestamp offset. ### Expected Behavior The `pyarrow` dtypes don't shift the timestamps and give the same result as the normal dtypes. ### Installed Versions <details> ``` INSTALLED VERSIONS ------------------ commit : 0691c5cf90477d3503834d983f69350f250a6ff7 python : 3.12.11 python-bits : 64 OS : Linux OS-release : 6.14.0-1010-aws Version : #10~24.04.1-Ubuntu SMP Fri Jul 18 20:44:30 UTC 2025 machine : x86_64 processor : x86_64 byteorder : little LC_ALL : None LANG : C.UTF-8 LOCALE : C.UTF-8 pandas : 2.2.3 numpy : 1.26.4 pytz : 2025.2 dateutil : 2.9.0.post0 pip : 25.2 Cython : None sphinx : 8.2.3 IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 blosc : None bottleneck : None dataframe-api-compat : None fastparquet : None fsspec : 2025.7.0 html5lib : 1.1 hypothesis : 6.137.3 gcsfs : None jinja2 : 3.1.6 lxml.etree : 5.4.0 matplotlib : 3.10.5 numba : None numexpr : None odfpy : None openpyxl : 3.1.5 pandas_gbq : None psycopg2 : None pymysql : None pyarrow : 19.0.1 pyreadstat : None pytest : 8.4.1 python-calamine : None pyxlsb : None s3fs : 2025.7.0 scipy : 1.16.1 sqlalchemy : 2.0.37 tables : None tabulate : 0.9.0 xarray : 2025.7.1 xlrd : None xlsxwriter : None zstandard : 0.23.0 tzdata : 2025.2 qtpy : None pyqt5 : None ``` </details>
[ "Bug", "Needs Triage" ]
0
0
0
0
0
0
0
0
[ "```pycon\n>>> pd.concat([df['value_timestamp'], expected, actual], keys=['original', 'numpy', 'pyarrow'], axis=1)\n original numpy pyarrow\n0 2025-08-13 00:00:00+10:00 2025-08-13 00:00:00 2025-08-12 14:00:00\n1 2025-08-13 01:00:00+10:00 2025-08-13 01:00:00 2025-08-12 15:00:00\n2 2025-08-13 02:00:00+10:00 2025-08-13 02:00:00 2025-08-12 16:00:00\n3 2025-08-13 03:00:00+10:00 2025-08-13 03:00:00 2025-08-12 17:00:00\n4 2025-08-13 04:00:00+10:00 2025-08-13 04:00:00 2025-08-12 18:00:00\n5 2025-08-13 05:00:00+10:00 2025-08-13 05:00:00 2025-08-12 19:00:00\n6 2025-08-13 06:00:00+10:00 2025-08-13 06:00:00 2025-08-12 20:00:00\n7 2025-08-13 07:00:00+10:00 2025-08-13 07:00:00 2025-08-12 21:00:00\n8 2025-08-13 08:00:00+10:00 2025-08-13 08:00:00 2025-08-12 22:00:00\n9 2025-08-13 09:00:00+10:00 2025-08-13 09:00:00 2025-08-12 23:00:00\n10 2025-08-13 10:00:00+10:00 2025-08-13 10:00:00 2025-08-13 00:00:00\n11 2025-08-13 11:00:00+10:00 2025-08-13 11:00:00 2025-08-13 01:00:00\n12 2025-08-13 12:00:00+10:00 2025-08-13 12:00:00 2025-08-13 02:00:00\n13 2025-08-13 13:00:00+10:00 2025-08-13 13:00:00 2025-08-13 03:00:00\n14 2025-08-13 14:00:00+10:00 2025-08-13 14:00:00 2025-08-13 04:00:00\n15 2025-08-13 15:00:00+10:00 2025-08-13 15:00:00 2025-08-13 05:00:00\n16 2025-08-13 16:00:00+10:00 2025-08-13 16:00:00 2025-08-13 06:00:00\n17 2025-08-13 17:00:00+10:00 2025-08-13 17:00:00 2025-08-13 07:00:00\n18 2025-08-13 18:00:00+10:00 2025-08-13 18:00:00 2025-08-13 08:00:00\n19 2025-08-13 19:00:00+10:00 2025-08-13 19:00:00 2025-08-13 09:00:00\n20 2025-08-13 20:00:00+10:00 2025-08-13 20:00:00 2025-08-13 10:00:00\n21 2025-08-13 21:00:00+10:00 2025-08-13 21:00:00 2025-08-13 11:00:00\n22 2025-08-13 22:00:00+10:00 2025-08-13 22:00:00 2025-08-13 12:00:00\n23 2025-08-13 23:00:00+10:00 2025-08-13 23:00:00 2025-08-13 13:00:00\n```", "I think addressed by #62076", "> I think addressed by [#62076](https://github.com/pandas-dev/pandas/pull/62076)\n\nThat seems likely, thanks\\." ]
3,316,066,112
62,099
DOC: Sort the pandas API reference navbar in alphabetical order
open
2025-08-12T21:52:06
2025-08-19T20:47:57
null
https://github.com/pandas-dev/pandas/pull/62099
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62099
https://github.com/pandas-dev/pandas/pull/62099
DoNguyenHung
1
- [x] closes #59164 - [ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Changes made: - Added JavaScript to sort only the items in each section of the sidebar and not the section headers. - This is a re-submission of my previous pull request #62069, which was closed due to the fact that issue #59164 had a "Needs Discussion" tag. To update, the maintainer who placed the tag has acknowledged my previous PR and was okay with merging it. I've also added a comment in #59164 to answer another maintainer's question about why JavaScript was used to fix the issue instead of using Sphinx.
[]
0
0
0
0
0
0
0
0
[ "pre-commit.ci autofix" ]
3,316,011,489
62,098
ENH: Remove the DataFrame.attrs saving to parquet metadata on Pandas' side
open
2025-08-12T21:30:17
2025-08-12T21:57:38
null
https://github.com/pandas-dev/pandas/issues/62098
true
null
null
rmnskb
0
### Feature Type - [ ] Adding new functionality to pandas - [ ] Changing existing functionality in pandas - [x] Removing existing functionality in pandas ### Problem Description #54321 raised the issue regarding the `DataFrame.attrs` persistence in `.parquet` metadata, that was resolved in #54346. In apache/arrow#45382 @fangchenli proposed that this functionality should be handled on the Arrow's side by adding the attributes during the metadata creation, rather than injecting them later on, which was then implemented in apache/arrow#47147 ### Feature Description Revert the changes that were introduced in #54346, since they are clashing with the Arrow implementation now ### Alternative Solutions Both implementations can be left as is, although it should be assessed what kind of impact this could potentially have ### Additional Context I want to open this issue to discuss what Pandas community thinks is the best approach in this case
[ "Enhancement", "IO Parquet", "metadata", "Needs Triage" ]
0
0
0
0
0
0
0
0
[]
3,315,587,009
62,097
DEPR: rename 'unit' keyword in to_datetime etc
open
2025-08-12T19:27:16
2025-08-24T12:27:53
null
https://github.com/pandas-dev/pandas/issues/62097
true
null
null
jbrockmendel
4
to_datetime, to_timedelta, Timestamp, and Timedelta have a `unit` keyword that causes confusion because the term 'unit' is overloaded. The keyword refers to how we interpret integer arguments, while the result's obj.unit attribute refers to the resolution of the result. Confusion e.g. #60371. Update: also the meaning of the keyword in date_range matches the attribute but not the to_datetime meaning. Let's deprecate the argument and rename it to something else. Opening bid: "numeric_unit"
[ "Deprecate", "Needs Discussion" ]
0
0
0
0
0
0
0
0
[ "What if we made `unit` apply to non-numeric arguments? I.e. if a non-numeric argument is provided we convert to the user-specified unit? This would allow us to avoid code churn.", "> What if we made unit apply to non-numeric arguments? I.e. if a non-numeric argument is provided we convert to the user-specified unit? This would allow us to avoid code churn.\n\nDo you mean \"interpret 'unit' as specifying the desired output-unit in some cases but not atll\"? i think im against more-overloading like that", "I was proposing to have it be all, but I think I now see this issue. If someone supplies e.g. `0.5772156649` and specifies `unit=ms`, they likely do not want the output to be `ms`.\n\nI see your `numeric_unit` and raise you an `input_unit` that would raise whenever specified and the input is not numeric.", "I support the proposal to deprecate and rename the `unit` keyword in `to_datetime`, `to_timedelta`, `Timestamp`, and `Timedelta`. \n\n**Issue Details:**\n\n- The current `unit` keyword is overloaded: in integer arguments it determines how the input is interpreted, but the resulting object's `.unit` attribute refers to the resolution of the result. This causes confusion.\n- In `date_range`, the `unit` keyword matches the output attribute, but in `to_datetime` it does not, adding inconsistency.\n- Allowing `unit` to apply to non-numeric arguments could introduce further ambiguity and unintended behavior.\n\n**Proposed Approach:**\n\n- Deprecate the existing `unit` keyword.\n- Rename it to `numeric_unit` (or `input_unit` for stricter behavior) to clearly indicate it only applies to numeric inputs.\n- Raise an error if a non-numeric input is passed with the renamed keyword to avoid confusion and maintain consistent behavior.\n\nThis change improves clarity, reduces overloading of the `unit` keyword, and makes the API more intuitive for both numeric and non-numeric use cases.\n" ]
3,315,227,231
62,096
BUG: Preserve day freq on DatetimeIndex subtraction
open
2025-08-12T17:40:59
2025-08-20T16:55:57
null
https://github.com/pandas-dev/pandas/pull/62096
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62096
https://github.com/pandas-dev/pandas/pull/62096
akshat62
1
- [x] closes #62094 - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature.
[]
0
0
0
0
0
0
0
0
[ "@mroeschke @jbrockmendel Please review." ]
3,315,219,363
62,095
BUG: TimedeltaIndex.shift() infers freq when possible (GH#62094)
closed
2025-08-12T17:38:16
2025-08-12T18:43:13
2025-08-12T18:43:13
https://github.com/pandas-dev/pandas/pull/62095
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62095
https://github.com/pandas-dev/pandas/pull/62095
Aniketsy
0
(GH#62094) This PR fixes a regression on main where TimedeltaIndex.shift() raised a NullFrequencyError if the index had freq=None, even when the index was regular and its frequency could be inferred. Please let me know if my approach or fix needs any improvements . I’m open to feedback and happy to make changes based on suggestions. Thankyou!
[]
0
0
0
0
0
0
0
0
[]
3,314,905,948
62,094
BUG: In main, TimedeltaIndex.shift() requires freq in the index, but it may not be available because it was computed
open
2025-08-12T15:52:45
2025-08-19T03:01:54
null
https://github.com/pandas-dev/pandas/issues/62094
true
null
null
Dr-Irv
4
### Pandas version checks - [x] I have checked that this issue has not already been reported. - [x] I have confirmed this bug exists on the [latest version](https://pandas.pydata.org/docs/whatsnew/index.html) of pandas. - [x] I have confirmed this bug exists on the [main branch](https://pandas.pydata.org/docs/dev/getting_started/install.html#installing-the-development-version-of-pandas) of pandas. ### Reproducible Example ```python import pandas as pd ind = pd.date_range("1/1/2021", "1/5/2021") - pd.Timestamp("1/3/2019") ind.shift(1) ``` ### Issue Description This only occurs on main. NOT a current bug in pandas. Gives error: ```text Traceback (most recent call last): File "<stdin>", line 1, in <module> File "C:\Code\pandas_dev\pandas\pandas\core\indexes\datetimelike.py", line 512, in shift raise NullFrequencyError("Cannot shift with no freq") pandas.errors.NullFrequencyError: Cannot shift with no freq ``` The above code works fine with pandas 2.3. ### Expected Behavior No error. A user doing a calculation that produces a `TimedeltaIndex` can't be expected to set the `freq` of the index. I think was introduced by @jbrockmendel in #61985 ### Installed Versions <details> INSTALLED VERSIONS ------------------ commit : bb10b27dea9d9a2476de4c8122e0346689e1c9c3 python : 3.11.13 python-bits : 64 OS : Windows OS-release : 10 Version : 10.0.26100 machine : AMD64 processor : Intel64 Family 6 Model 183 Stepping 1, GenuineIntel byteorder : little LC_ALL : None LANG : None LOCALE : English_United States.1252 pandas : 3.0.0.dev0+2306.gbb10b27dea.dirty numpy : 2.2.6 dateutil : 2.9.0.post0 pip : 25.2 Cython : 3.1.2 sphinx : 8.2.3 IPython : 9.4.0 adbc-driver-postgresql: None adbc-driver-sqlite : None bs4 : 4.13.4 bottleneck : 1.5.0 fastparquet : 2024.11.0 fsspec : 2025.7.0 html5lib : 1.1 hypothesis : 6.137.3 gcsfs : 2025.7.0 jinja2 : 3.1.6 lxml.etree : 6.0.0 matplotlib : 3.10.5 numba : 0.61.2 numexpr : 2.10.2 odfpy : None openpyxl : 3.1.5 psycopg2 : 2.9.10 pymysql : 1.4.6 pyarrow : 19.0.1 pyiceberg : 0.9.1 pyreadstat : 1.3.0 pytest : 8.4.1 python-calamine : None pytz : 2025.2 pyxlsb : 1.0.10 s3fs : 2025.7.0 scipy : 1.16.1 sqlalchemy : 2.0.43 tables : 3.10.2 tabulate : 0.9.0 xarray : 2025.7.1 xlrd : 2.0.1 xlsxwriter : 3.2.5 zstandard : 0.23.0 qtpy : None pyqt5 : None </details>
[ "Bug", "Timedelta", "Blocker", "good first issue" ]
0
0
0
0
0
0
0
0
[ "Should be straightforward to add this case to _get_arithmetic_result_freq. Marking as Good First Issue.", "Hi, I’m new to open source and would like to work on this issue. Could you please assign it to me?", "Hello all,\nThis is the first time I am attempting for open source contribution. Any suggestions or how can I move forward on this issue?", "take" ]
3,313,828,239
62,093
BUILD: Nightly wheel building failed for some platforms
closed
2025-08-12T11:36:04
2025-08-15T18:06:48
2025-08-15T18:06:48
https://github.com/pandas-dev/pandas/issues/62093
true
null
null
djhoese
3
### Installation check - [x] I have read the [installation guide](https://pandas.pydata.org/pandas-docs/stable/getting_started/install.html#installing-pandas). ### Platform Github Actions ubuntu-latest ### Installation Method pip install ### pandas Version nightly wheel (see below) ### Python Version 3.12 ### Installation Logs <details> ``` python -m pip install \ --index-url https://pypi.anaconda.org/scientific-python-nightly-wheels/simple/ \ --trusted-host pypi.anaconda.org \ --no-deps --pre --upgrade \ matplotlib \ numpy \ pandas \ scipy; ... Collecting pandas Downloading https://pypi.anaconda.org/scientific-python-nightly-wheels/simple/pandas/3.0.0.dev0%2B2306.gbb10b27dea/pandas-3.0.0.dev0%2B2306.gbb10b27dea.tar.gz (4.5 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.5/4.5 MB 84.0 MB/s 0:00:00 Installing build dependencies: started Installing build dependencies: finished with status 'error' error: subprocess-exited-with-error × pip subprocess to install build dependencies did not run successfully. │ exit code: 1 ╰─> [3 lines of output] Looking in indexes: https://pypi.anaconda.org/scientific-python-nightly-wheels/simple/ ERROR: Could not find a version that satisfies the requirement meson-python>=0.13.1 (from versions: none) ERROR: No matching distribution found for meson-python>=0.13.1 [end of output] ``` </details> ~I noticed that the installation guide says to use `--extra-index` but I had found in the past that that installed an incorrect combination of packages. I can try that, but I'm surprised this has started failing only just today. I'm wondering if someone maybe deleted `meson-python` from the `pypi.anaconda.org/scientific-python-nightly-wheels` location?~ Edit: See comments below. My CI tried to install the nightly build but there was no wheel for my platform so it installed from source and failed because `meson-python` was not available. My workaround was to add `--only-binary ":all:"` to my pip install command. The nightly builds could maybe be restarted?
[ "Build" ]
0
0
0
0
0
0
0
0
[ "I'm rerunning the CI job I had that failed. It looks like (see the logs above) that the wheel was not available and the sdist tarball was used instead.\n\nOh...nightly wheel builds failed:\n\nhttps://github.com/pandas-dev/pandas/actions/runs/16898335562", "Thanks for raising the issue. I see that indeed some of the linux builds are failing somewhat randomly -> https://github.com/pandas-dev/pandas/pull/62124\n\nI also see that some of the windows wheels are failing, I suppose that's not that you ran into, but something to fix as well (-> https://github.com/pandas-dev/pandas/pull/62126)", "Yes, correct. My CI only does \"unstable/nightly\" installs on ubuntu." ]
3,313,145,004
62,092
BUG: Fix assert_frame_equal with check_dtype=False for pd.NA with different dtypes
open
2025-08-12T08:32:14
2025-08-12T10:12:23
null
https://github.com/pandas-dev/pandas/pull/62092
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62092
https://github.com/pandas-dev/pandas/pull/62092
0x3vAD
0
- [x] closes #61473 - [x] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature - [x] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit). - [x] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions. - [x] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Fixes a bug where assert_frame_equal failed when comparing DataFrames containing pd.NA with different dtypes (object vs Int32) if check_dtype=False.
[]
0
0
0
0
0
0
0
0
[]
3,312,426,913
62,091
DOC: Improve consistency in terminology and fix minor typos in documentation
closed
2025-08-12T03:56:22
2025-08-18T15:21:58
2025-08-18T15:21:48
https://github.com/pandas-dev/pandas/pull/62091
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62091
https://github.com/pandas-dev/pandas/pull/62091
maitreytalware
2
Fixed minor typos in files/documentation
[]
0
0
0
0
0
0
0
0
[ "This is fine, but in the future please stick to clear typos and avoid stuff like \"IO\" vs \"I/O\".", "thanks @maitreytalware " ]
3,312,016,591
62,090
DOC: Add to_julian_date to DatetimeIndex methods listing
closed
2025-08-12T00:09:57
2025-08-14T01:49:06
2025-08-13T16:08:38
https://github.com/pandas-dev/pandas/pull/62090
true
https://api.github.com/repos/pandas-dev/pandas/pulls/62090
https://github.com/pandas-dev/pandas/pull/62090
echedey-ls
1
First time contributor here. - ~[ ] closes #xxxx (Replace xxxx with the GitHub issue number)~ No issue related - ~[ ] [Tests added and passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#writing-tests) if fixing a bug or adding a new feature~ - ~[ ] All [code checks passed](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#pre-commit).~ - ~[ ] Added [type annotations](https://pandas.pydata.org/pandas-docs/dev/development/contributing_codebase.html#type-hints) to new arguments/methods/functions.~ Nothing new - [ ] Added an entry in the latest `doc/source/whatsnew/vX.X.X.rst` file if fixing a bug or adding a new feature. Add `to_julian_date` documentation entry to `pandas.DatetimeIndex`. Currently, it is not listed nor can be found in the documentation: https://pandas.pydata.org/docs/reference/api/pandas.DatetimeIndex.html `pandas.Timestamp` already has it linked https://pandas.pydata.org/docs/reference/api/pandas.Timestamp.to_julian_date.html from the time it was added v0.14.0 https://pandas.pydata.org/docs/whatsnew/v0.14.0.html I haven't read through the contributing section, please point out if there is something missing. If it looks okay, tell me to proceed with the whatsnew task. Docs artifact for second commit (it works): https://github.com/pandas-dev/pandas/actions/runs/16895776565/job/47865152612?pr=62090
[ "Docs" ]
0
0
0
0
0
0
0
0
[ "Thanks @echedey-ls " ]
End of preview. Expand in Data Studio

Pandas GitHub Issues

This dataset contains 5,000 GitHub issues collected from the pandas-dev/pandas repository.
It includes issue metadata, content, labels, user information, timestamps, and comments.

The dataset is suitable for text classification, multi-label classification, and document retrieval tasks.

Dataset Structure

Columns:

  • id — Internal ID of the issue (int64)
  • number — GitHub issue number (int64)
  • title — Title of the issue (string)
  • state — Issue state: open/closed (string)
  • created_at — Timestamp when the issue was created (timestamp[s])
  • updated_at — Timestamp when the issue was last updated (timestamp[s])
  • closed_at — Timestamp when the issue was closed (timestamp[s])
  • html_url — URL to the GitHub issue (string)
  • pull_request — Struct containing PR info (if the issue is a PR):
    • url — URL to PR
    • html_url — HTML URL of PR
    • diff_url — Diff URL
    • patch_url — Patch URL
    • merged_at — Merge timestamp (timestamp[s])
  • user_login — Login of the issue creator (string)
  • is_pull_request — Whether the issue is a pull request (bool)
  • comments — List of comments on the issue (list[string])

Splits:

  • train — 5,000 examples

Dataset Creation

The dataset was collected using the GitHub API, including all issue metadata and comments.

Usage Example

from datasets import load_dataset

dataset = load_dataset("your-username/pandas-github-issues", split="train")

# Preview first 5 examples
for i, example in enumerate(dataset[:5]):
    print(f"Issue #{example['number']}: {example['title']}")
    print(f"Created at: {example['created_at']}, Closed at: {example['closed_at']}")
    print(f"User: {example['user_login']}, PR: {example['is_pull_request']}")
    print(f"Comments: {example['comments'][:3]}")  # first 3 comments
    print()
___
## Citation

If you use this dataset, please cite it as:

```bibtex
@misc{yourusername_pandas_github_issues,
  author       = {Your Name},
  title        = {Pandas GitHub Issues Dataset},
  year         = {2025},
  howpublished = {\url{https://huggingface.co/datasets/your-username/pandas-github-issues}}
}
Downloads last month
62