Skip to content

Commit

Permalink
changing the yaml configs with the new caching logic
Browse files Browse the repository at this point in the history
  • Loading branch information
DomInvivo committed Aug 10, 2023
1 parent cc27c8f commit bdf14d1
Show file tree
Hide file tree
Showing 33 changed files with 13 additions and 43 deletions.
1 change: 0 additions & 1 deletion expts/configs/config_gps_10M_pcqm4m.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -112,7 +112,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 0 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
1 change: 0 additions & 1 deletion expts/configs/config_gps_10M_pcqm4m_mod.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -81,7 +81,6 @@ datamodule:
# Data handling-related
batch_size_training: 64
batch_size_inference: 16
# cache_data_path: .
num_workers: 0 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
2 changes: 1 addition & 1 deletion expts/configs/config_mpnn_10M_b3lyp.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -93,6 +93,7 @@ datamodule:
featurization_progress: True
featurization_backend: "loky"
processed_graph_data_path: "../datacache/b3lyp/"
dataloading_from: ram
featurization:
# OGB: ['atomic_num', 'degree', 'possible_formal_charge', 'possible_numH' (total-valence),
# 'possible_number_radical_e', 'possible_is_aromatic', 'possible_is_in_ring',
Expand Down Expand Up @@ -123,7 +124,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 0 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
3 changes: 1 addition & 2 deletions expts/configs/config_mpnn_pcqm4m.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,8 @@ datamodule:
featurization_n_jobs: 20
featurization_progress: True
featurization_backend: "loky"
cache_data_path: "./datacache"
processed_graph_data_path: "graphium/data/PCQM4Mv2/"
dataloading_from: ram
featurization:
# OGB: ['atomic_num', 'degree', 'possible_formal_charge', 'possible_numH' (total-valence),
# 'possible_number_radical_e', 'possible_is_aromatic', 'possible_is_in_ring',
Expand All @@ -58,7 +58,6 @@ datamodule:
# Data handling-related
batch_size_training: 64
batch_size_inference: 16
# cache_data_path: .
num_workers: 40 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
1 change: 0 additions & 1 deletion expts/neurips2023_configs/base_config/large.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -168,7 +168,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 32 # -1 to use all
persistent_workers: True # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
1 change: 0 additions & 1 deletion expts/neurips2023_configs/base_config/small.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -132,7 +132,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -131,7 +131,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -111,7 +111,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 5 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
1 change: 0 additions & 1 deletion expts/neurips2023_configs/config_luis_jama.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 4 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
1 change: 0 additions & 1 deletion expts/neurips2023_configs/debug/config_debug.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -105,7 +105,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 0 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
3 changes: 1 addition & 2 deletions expts/neurips2023_configs/debug/config_large_gcn_debug.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down Expand Up @@ -327,7 +326,7 @@ predictor:
l1000_mcf7: []
pcba_1328: []
pcqm4m_g25: []
pcqm4m_n4: []
pcqm4m_n4: []
loss_fun:
l1000_vcap:
name: hybrid_ce_ipu
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -103,7 +103,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -118,7 +118,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -103,7 +103,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -118,7 +118,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: 30 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ datamodule:
# Data handling-related
batch_size_training: 16
batch_size_inference: 16
# cache_data_path: null

architecture: # The parameters for the full graph network are taken from `config_micro_ZINC.yaml`
model_type: FullGraphMultiTaskNetwork
Expand Down Expand Up @@ -111,7 +110,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
graph:
pooling: [sum, max]
out_dim: 1
Expand All @@ -122,7 +121,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
edge:
out_dim: 16
hidden_dims: 32
Expand All @@ -132,7 +131,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
nodepair:
out_dim: 16
hidden_dims: 32
Expand All @@ -142,7 +141,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none

task_heads: # Set as null to avoid task heads. Recall that the arguments for the TaskHeads is a List of TaskHeadParams
task_1:
Expand Down
9 changes: 4 additions & 5 deletions graphium/config/fake_multilevel_multitask_pyg.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ datamodule:
# Data handling-related
batch_size_training: 16
batch_size_inference: 16
# cache_data_path: null

architecture: # The parameters for the full graph network are taken from `config_micro_ZINC.yaml`
model_type: FullGraphMultiTaskNetwork
Expand Down Expand Up @@ -111,7 +110,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
graph:
pooling: [sum, max]
out_dim: 1
Expand All @@ -122,7 +121,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
edge:
out_dim: 16
hidden_dims: 32
Expand All @@ -132,7 +131,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none
nodepair:
out_dim: 16
hidden_dims: 32
Expand All @@ -142,7 +141,7 @@ architecture: # The parameters for the full graph network are taken from `co
dropout: *dropout
normalization: *normalization
last_normalization: "none"
residual_type: none
residual_type: none

task_heads: # Set as null to avoid task heads. Recall that the arguments for the TaskHeads is a List of TaskHeadParams
task_1:
Expand Down
1 change: 0 additions & 1 deletion graphium/config/zinc_default_multitask_pyg.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ datamodule:
# Data handling-related
batch_size_training: 16
batch_size_inference: 16
# cache_data_path: null

architecture: # The parameters for the full graph network are taken from `config_micro_ZINC.yaml`
model_type: FullGraphMultiTaskNetwork
Expand Down
2 changes: 1 addition & 1 deletion profiling/configs_profiling.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ datamodule:
module_type: "DGLFromSmilesDataModule"
args:
df_path: https://storage.googleapis.com/graphium-public/datasets/graphium-zinc-bench-gnn/smiles_score.csv.gz
cache_data_path: null # graphium/data/cache/ZINC_bench_gnn/smiles_score.cache
processed_graph_data_path: null
label_cols: ['score']
smiles_col: SMILES

Expand Down
1 change: 0 additions & 1 deletion tests/config_test_ipu_dataloader_multitask.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -130,7 +130,6 @@ datamodule:
pos_type: rw_return_probs
ksteps: 16

# cache_data_path: .
num_workers: -1 # -1 to use all
persistent_workers: False # if use persistent worker at the start of each epoch.
# Using persistent_workers false might make the start of each epoch very long.
Expand Down
2 changes: 1 addition & 1 deletion tests/data/config_micro_ZINC.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ datamodule:
module_type: "DGLFromSmilesDataModule"
args:
df_path: graphium/data/micro_ZINC/micro_ZINC.csv
cache_data_path: graphium/data/cache/micro_ZINC/full.cache
processed_graph_data_path: graphium/data/cache/micro_ZINC/
label_cols: ['score']
smiles_col: SMILES

Expand Down

0 comments on commit bdf14d1

Please sign in to comment.