pdlog provides logging for pandas dataframes, to better enable you to monitor and debug your data pipelines.
For example:
>>> import pdlog
>>> df = df.log.dropna()
2020-05-26 20:55:30,049 INFO <pdlog> dropna: dropped 1 row (17%), 5 rows remainingThe above assumes that the logging module has been configured and that data has been loaded into a pandas DataFrame. Let's walk through those steps with a simple example.
-
Configure
logging:>>> import logging >>> fmt = "{asctime} {levelname} <{name}> {message}" >>> logging.basicConfig(format=fmt, style="{", level=logging.INFO)
-
Load data into a
pandas.DataFrame:>>> import pandas as pd >>> df = pd.DataFrame([0, 1, 2, None, 4]) >>> df.head() 0 0 0.0 1 1.0 2 2.0 3 NaN 4 4.0
-
Importing
pdlogand call a method under thelogaccessor:>>> import pdlog >>> df = df.log.dropna() 2020-05-26 20:55:30,049 INFO <pdlog> dropna: dropped 1 row (17%), 5 rows remaining
pdlog currently supports the following pandas.DataFrame methods:
- Filter rows and select columns:
drop_duplicatesdropdropnaheadquerysampletail
- (Re-)set indexes:
reset_indexset_index
- Rename indexes:
rename
- Reshape:
meltpivot
- Impute:
bfillffillfillna
pandas-log is aimed at interactive usage. Its messages are friendlier and more verbose than pdlog aims to be.
Ideally, each pdlog message should be a single line of dense information to help you understand whether your production code is doing what you think it is, while not overflowing your logs.
These don't tend to make particularly friendly messages.
pdlog can be considered a port of tidylog (R package) to pandas.
Their goals align with ours, and we think they've done a great job at reaching those goals.