-
Notifications
You must be signed in to change notification settings - Fork 117
Closed
Labels
area: variablesIssues related to Variables category.Issues related to Variables category.lang: pythonperformance
Milestone
Description
Positron Version: git main
Steps to reproduce the issue:
- Load data frame with over 100,000 rows
- Transpose data frame into new variable (
df2 = df.T
)
For example
import numpy as np
import pandas as pd
df = pd.DataFrame({'a': np.arange(1000000)})
df
df2 = df.T
(without Positron in the loop, df.T
is virtually instantaneous)
In [4]: %timeit df.T
20.6 µs ± 76.6 ns per loop (mean ± std. dev. of 7 runs, 10,000 loops each)
The variables pane takes over 10 seconds for the update to come through.
We should look at the variables logic around data frames to make sure that we don't do excess computations for data frames with > 10,000 columns. For some users, this case is less unusual than one might think, and so we don't want the UI to get blocked in this scenario.
Metadata
Metadata
Assignees
Labels
area: variablesIssues related to Variables category.Issues related to Variables category.lang: pythonperformance