Skip to content

Commit

Permalink
expanded data shape to accommodate the biggest datasets
Browse files Browse the repository at this point in the history
  • Loading branch information
Dawith committed Oct 19, 2025
1 parent c7cff3c commit 52dde05
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion pipe/etl.py
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ def split_sets(data: DataFrame, split=[0.99, 0.005, 0.005]) -> tuple:
def trim(dataframe, column):

ndarray = np.array(dataframe.select(column).collect()) \
.reshape(-1, 32, 133)
.reshape(-1, 34, 133)

return ndarray

Expand Down
2 changes: 1 addition & 1 deletion pipe/extract.py
Original file line number Diff line number Diff line change
Expand Up @@ -105,7 +105,7 @@ def metadata_read(self, metapath: Path, labels:list,

def read_matfiles(self, specpath: Path,
datakinds: List[DataKind],
default_size: tuple = (32, 130),
default_size: tuple = (34, 130),
pad_value: float = 0.) -> DataFrame:
"""
Loads data for each stack iteration from a set of mat files,
Expand Down

0 comments on commit 52dde05

Please sign in to comment.