diff --git a/pyproject.toml b/pyproject.toml index cff54a2..8ad89c1 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -41,7 +41,7 @@ cmd = "nvim" [tool.poe.tasks.extract] help = "Extract the csv data from raw yaml files" shell = """ -python src/prep_data.py > 02-data/processed/extracted.csv +python src/matrix.py > 02-data/processed/extracted.csv """ [tool.poe.tasks.milestone] help = "Extract, render, commit and version a finished artifact" diff --git a/scoping_review.qmd b/scoping_review.qmd index e49e2ac..71dbb3f 100644 --- a/scoping_review.qmd +++ b/scoping_review.qmd @@ -613,9 +613,9 @@ to better identify areas of strong analytical lenses or areas of more limited an ```{python} #| label: fig-validity -from src import prep_data +from src.model import validity -validities = prep_data.calculate_validities(by_intervention) +validities = validity.calculate(by_intervention) validities["identifier"] = validities["author"].str.replace(r',.*$', '', regex=True) + " (" + validities["year"].astype(str) + ")" g = sns.PairGrid(validities[["internal_validity", "external_validity", "identifier"]].drop_duplicates(subset="identifier"), diff --git a/src/matrix.py b/src/matrix.py new file mode 100644 index 0000000..d4072d5 --- /dev/null +++ b/src/matrix.py @@ -0,0 +1,38 @@ +from io import StringIO +from pathlib import Path + +from pandas import DataFrame + +try: + from src.model import validity # for quarto document scripts +except ModuleNotFoundError: + from model import validity # for directly running the package + + +def extract(df: DataFrame, file: Path | StringIO) -> None: + ( + validity.calculate(df) + .drop(labels=["observation"], axis="columns") + .to_csv(file, index=False, encoding="utf-8") + ) + + +if __name__ == "__main__": + import os + import sys + + import load_data + + if len(sys.argv) == 2: + df = load_data.from_yml(Path(sys.argv[1])) + else: + df = load_data.from_yml() + + output = StringIO() + extract(df, output) + output.seek(0) + try: + print(output.read()) + except BrokenPipeError: + devnull = os.open(os.devnull, os.O_WRONLY) + os.dup2(devnull, sys.stdout.fileno()) diff --git a/src/model/validity.py b/src/model/validity.py index 946e05f..d493fb5 100644 --- a/src/model/validity.py +++ b/src/model/validity.py @@ -1,9 +1,8 @@ import math +from typing import cast from pandas import DataFrame -from src import prep_data - def _binned_strength(strength: float) -> str: if strength < 3.0: @@ -28,15 +27,74 @@ def _combined_validities( return r"\-" +def calculate( + df: DataFrame, repr_col: str = "representativeness", method_col: str = "method" +) -> DataFrame: + EXT_COL_NAME: str = "external_validity" + INT_COL_NAME: str = "internal_validity" + cols = {EXT_COL_NAME: 0.0, INT_COL_NAME: 0.0} + + vd = df[ + (df["design"] == "quasi-experimental") | (df["design"] == "experimental") + ].copy() + vd.assign(**cols) + vd = cast(DataFrame, vd) + + vd[repr_col] = vd[repr_col].fillna("") + vd[method_col] = vd[method_col].fillna("") + # needs to check national before subnational and census, subnational before local + vd.loc[ + vd[repr_col].str.contains("|".join(["national", "regional"])), EXT_COL_NAME + ] = 4.0 + vd.loc[vd[repr_col].str.contains("census"), EXT_COL_NAME] = 5.0 + vd.loc[vd[repr_col].str.contains("subnational"), EXT_COL_NAME] = 3.0 + vd.loc[vd[repr_col].str.contains("local"), EXT_COL_NAME] = 2.0 + + # needs to go lowest to highest in case of multiple mentioned approaches + vd.loc[ + vd[method_col].str.contains( + "|".join(["OLS", "ordinary.least.square", "logistic.regression"]) + ), + INT_COL_NAME, + ] = 2.0 + vd.loc[ + vd[method_col].str.contains("|".join(["DM", "discontinuity.matching"])), + INT_COL_NAME, + ] = 3.0 + vd.loc[ + vd[method_col].str.contains( + "|".join(["DID", "difference.in.diff", "diff.in.diff", "triple.diff"]) + ), + INT_COL_NAME, + ] = 3.0 + vd.loc[ + vd[method_col].str.contains( + "|".join(["PSM", "propensity.score.matching", "score.matching"]) + ), + INT_COL_NAME, + ] = 3.5 + vd.loc[ + vd[method_col].str.contains("|".join(["IV", "instrumental.variable"])), + INT_COL_NAME, + ] = 4.0 + vd.loc[ + vd[method_col].str.contains("|".join(["RD", "regression.discontinuity"])), + INT_COL_NAME, + ] = 4.5 + vd.loc[vd[method_col].str.contains("RCT"), INT_COL_NAME] = 5.0 + + return vd + + def add_to_findings( findings_df: DataFrame, studies_by_intervention: DataFrame ) -> DataFrame: valid_subset = ( - prep_data.calculate_validities(studies_by_intervention)[ + calculate(studies_by_intervention)[ ["internal_validity", "external_validity", "citation"] ] .fillna(1.0) - .drop_duplicates(subset=["citation"]) # type: ignore + .drop_duplicates(subset=["citation"]) # type: ignore .sort_values("internal_validity") ) diff --git a/src/prep_data.py b/src/prep_data.py deleted file mode 100644 index 21d9ec4..0000000 --- a/src/prep_data.py +++ /dev/null @@ -1,82 +0,0 @@ -from typing import cast - -from pandas import DataFrame - - -def calculate_validities( - df: DataFrame, repr_col: str = "representativeness", method_col: str = "method" -) -> DataFrame: - EXT_COL_NAME: str = "external_validity" - INT_COL_NAME: str = "internal_validity" - cols = {EXT_COL_NAME: 0.0, INT_COL_NAME: 0.0} - - vd = df[ - (df["design"] == "quasi-experimental") | (df["design"] == "experimental") - ].copy() - vd.assign(**cols) - vd = cast(DataFrame, vd) - - vd[repr_col] = vd[repr_col].fillna("") - vd[method_col] = vd[method_col].fillna("") - # needs to check national before subnational and census, subnational before local - vd.loc[vd[repr_col].str.contains("|".join(["national", "regional"])), EXT_COL_NAME] = 4.0 - vd.loc[vd[repr_col].str.contains("census"), EXT_COL_NAME] = 5.0 - vd.loc[vd[repr_col].str.contains("subnational"), EXT_COL_NAME] = 3.0 - vd.loc[vd[repr_col].str.contains("local"), EXT_COL_NAME] = 2.0 - - # needs to go lowest to highest in case of multiple mentioned approaches - vd.loc[ - vd[method_col].str.contains("|".join(["OLS", "ordinary.least.square", "logistic.regression"])), - INT_COL_NAME, - ] = 2.0 - vd.loc[ - vd[method_col].str.contains("|".join(["DM", "discontinuity.matching"])), - INT_COL_NAME, - ] = 3.0 - vd.loc[ - vd[method_col].str.contains( - "|".join(["DID", "difference.in.diff", "diff.in.diff", "triple.diff"]) - ), - INT_COL_NAME, - ] = 3.0 - vd.loc[ - vd[method_col].str.contains( - "|".join(["PSM", "propensity.score.matching", "score.matching"]) - ), - INT_COL_NAME, - ] = 3.5 - vd.loc[ - vd[method_col].str.contains("|".join(["IV", "instrumental.variable"])), - INT_COL_NAME, - ] = 4.0 - vd.loc[ - vd[method_col].str.contains("|".join(["RD", "regression.discontinuity"])), - INT_COL_NAME, - ] = 4.5 - vd.loc[vd[method_col].str.contains("RCT"), INT_COL_NAME] = 5.0 - - return vd - - -if __name__ == "__main__": - import os - import sys - from io import StringIO - from pathlib import Path - - import load_data - - if len(sys.argv) == 2: - df = load_data.from_yml(Path(sys.argv[1])) - else: - df = load_data.from_yml() - - df = calculate_validities(df) - output = StringIO() - df.to_csv(output) - output.seek(0) - try: - print(output.read()) - except BrokenPipeError: - devnull = os.open(os.devnull, os.O_WRONLY) - os.dup2(devnull, sys.stdout.fileno())