Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 12 additions & 4 deletions pyprophet/io/export/osw.py
Original file line number Diff line number Diff line change
Expand Up @@ -127,20 +127,21 @@ def _read_sqlite(self, con):
)

if self._is_unscored_file(con):
logger.info("Reading unscored data from Parquet file.")
logger.info("Reading unscored data from OSW file.")
return self._read_unscored_data(con)

ipf_present = self._check_ipf_presence(con, cfg)

if ipf_present and cfg.ipf == "peptidoform":
logger.info("Reading peptidoform IPF data from Parquet file.")
logger.info("Reading peptidoform IPF data from OSW file.")
data = self._read_peptidoform_data(con, cfg)
elif ipf_present and cfg.ipf == "augmented":
logger.info("Reading augmented data with IPF from Parquet file.")
logger.info("Reading augmented data with IPF from OSW file.")
data = self._read_augmented_data(con, cfg)
else:
logger.info("Reading standard OpenSWATH data from Parquet file.")
logger.info("Reading standard OpenSWATH data from OSW file.")
data = self._read_standard_data(con, cfg)
logger.debug(f"Base data read with {len(data)} rows")

# Apply common augmentations to all scored data types
return self._augment_data(data, con, cfg)
Expand Down Expand Up @@ -493,6 +494,9 @@ def _read_standard_data(self, con, cfg):
ORDER BY transition_group_id, peak_group_rank;
"""
data = pd.read_sql_query(query, con)
logger.trace(
f"Initial standard data read with {len(data)} rows and columns: {data.columns.tolist()}"
)

# Ensure id column is Int64 to preserve precision for large feature IDs
if "id" in data.columns:
Expand Down Expand Up @@ -656,6 +660,10 @@ def _read_standard_data(self, con, cfg):
f"Assigned alignment_group_id to {mask.sum()} reference features"
)

logger.trace(
f"Data after merging alignment features has {len(data)} rows and columns: {data.columns.tolist()}"
)

return data

def _augment_data(self, data, con, cfg):
Expand Down
Loading
Loading