Category "pandas"

Pandas implement an any check

How do I check a Pandas column for "any" row that matches a condition? (in my case, I want to test for type string). Background: I was using the df.columnName.d

Count all NaNs in a pandas DataFrame

I'm trying to count NaN element (data type class 'numpy.float64')in pandas series to know how many are there which data type is class 'pandas.core.series.Seri

Adding new multiindex level

I have a DataFrame: d = pd.DataFrame({'i1': ['A', 'B', 'C', 'D', 'E'], 'i2': ['I', 'II', 'III', 'IV', 'V'], 'val': ["lol1",

Changing column order in python pandas using pivot table pd

I would love to change the order of my pivot table using pd (pandas) Current layout Code: pd.pivot_table(df.loc[filt], index=['region','brand'], values

Python code to return element value in dataframe based on another dataframe

I have a dataset similar to this generated from a file with yearly data d1 = pd.DataFrame({'category': ['A', 'B', 'C', 'D', 'E', 'F'], 'col

How to add new edges to the stellargraph dataset?

I need to add some extra edges to Cora dataset using stellargraph. Is there ane way to add edges to the current dataset in stellargraph library? import stellarg

Convert Sharepoint List into Pandas Dataframe

I have a table in Sharepoint that I'm wanting to convert into a Pandas Dataframe. I've largely used this question to try and frame a solution Get SharePoint Lis

How to calculate bed mates and length of time spent together

I am working with some patient data in Pandas that looks roughly like the following patient_id room_id start_date end_date 1 1001 01-01-2022 04:30:26 05-01-202

How to hide text labels for rangeslider in Plotly?

I'm trying to plot a time-series plot with a range slider using Plotly in Python. import plotly.express as px import pandas as pd data = pd.read_csv('https://r

How to give space before and after on a pandas dataframe object?

I have a dataframe_ df date 13MAY2022 13MAY2022 13MAY2022 13MAY2022 How can we give space between these objects? My expectations like this_ df date

pandas check if there are duplicates of repeated values between the two columns and not inside one column

I have 2 columns and I want to check if there are duplicates of repeated values between the two columns and not inside one column. The length of the datasets is

Basic Trading strategy with python

I am trying to code a basic trading strategy call sell the winner and buy the loser. I have listed the three conditions I want to code in python. I am a newbie

Limit writing of pandas to_excel to 1 million rows per sheet

I have a dataFrame with around 28 millions rows (5 columns) and I'm struggling to write that to an excel, which is limited to 1,048,576 rows, I can't have that

Placeholder for DataFrame in pd.query

I use pd.query and pd.eval a lot. However, sometimes I find myself in situations where I would like to filter an unnamed DataFrame with pd.query and it would be

Issue when selecting desired elements from dict_keys in Python

I have a list of disctionary keys ['A_report1', 'A_report2', ..., 'A_report10','B_report1', 'B_report2', ..., 'B_report10',]. I want to extract all 'report1' fr

ValueError: 18 columns passed, passed data had 1 columnss

#pip install sqlalchemy #from Google import Create_Service # link to source code is in the description import pyodbc as odbc # pip install pypyodbc

Python SUMIF with one condition across dataframes

I'm working with two dataframes MRP: Material Description Septiembre 1208181 ADSV,NA,MX,ADH HOTMET 814433PM 630.2888856 1206500 SHWP,NA,MX,WRAP M-WRAP 18' 459.

How can I get last value of STOCHRSI with Ta-Lib?

I implemented it but it prints all. print(ta.STOCHRSI(df["close"], 14, 5, 3, 0)[-1]) 2022-04-20 17:00:00 NaN 2022-04-20 18:00:00 NaN 2022-04-

How to sample a python df on daily rate when it is greater than 500 yrs

I need to sample a dataframe that has a date range of 100 years at a daily rate because I want to get yearly totals (so I thought resample at daily rate then su

pandas df.to_parquet write to multiple smaller files

Is it possible to use Pandas' DataFrame.to_parquet functionality to split writing into multiple files of some approximate desired size? I have a very large Data