Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Binary file added artifacts/deepsdf_2d.pth
Binary file not shown.
Binary file added artifacts/deepsdf_2d_unit_init.pth
Binary file not shown.
48 changes: 0 additions & 48 deletions artifacts/experiment_20250717_115709/data_processing_results.txt

This file was deleted.

10 changes: 0 additions & 10 deletions artifacts/experiment_20250717_115709/model_architecture.txt

This file was deleted.

20 changes: 0 additions & 20 deletions artifacts/experiment_20250717_115709/pipeline_summary.txt

This file was deleted.

Binary file not shown.

This file was deleted.

138 changes: 0 additions & 138 deletions artifacts/experiment_20250717_115709/training_results.txt

This file was deleted.

Binary file not shown.
Binary file added artifacts/lip-10.pth
Binary file not shown.
Binary file added artifacts/lip-5_2d.pth
Binary file not shown.
Binary file added artifacts/lip-6.pth
Binary file not shown.
Binary file added artifacts/lip-6_2d.pth
Binary file not shown.
Binary file added artifacts/lip-8.pth
Binary file not shown.
25 changes: 13 additions & 12 deletions config/config_examples.yaml
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
home: &home_path /Users/marina.levay/Documents/GitHub/topology-control # all other paths are relative to home
home: &home_path /homes/dnogina/code/topology-control # all other paths are relative to home

artifacts_config:
save_artifacts_to: artifacts
Expand All @@ -9,15 +9,15 @@ processor_config:
raw: &raw_data_path data/raw
processed: &processed_data_path data/processed

train_val_split: 0.8
train_val_split: .99

# Updated point cloud params to match your notebook
point_cloud_params:
radius: 0.02
sigma: 0.02 # Match your notebook's sigma = radius
mu: 0.0
n_gaussian: 5 # N = 10 from your notebook
n_uniform: 10000 # NUM_SAMPLES = 10000 from your notebook
n_uniform: 50000 # NUM_SAMPLES = 10000 from your notebook

# Volume processor params for SDF generation
volume_processor_params:
Expand All @@ -26,28 +26,29 @@ processor_config:

model_config:
skip_building: false
model_name: deepsdf
model_name: deepsdf #lipschitz_deepsdf #deepsdf
# DeepSDF specific parameters matching your notebook
z_dim: 16 # Z_DIM = 16 from your notebook
layer_size: 64 # HIDDEN_DIM = 64 from your notebook
z_dim: 1 #16 # Z_DIM = 16 from your notebook
layer_size: 256 # HIDDEN_DIM = 64 from your notebook
coord_dim: 3 # 3D coordinates
dropout_p: 0.2 # Match your notebook's dropout_p = 0.2

trainer_config:
skip_training: false
processed_data_path: *processed_data_path
dataset_type: sdf # Important: specify SDF dataset type
dataset_type: sdf #lipschitz_sdf #sdf
lipschitz_lambda: 0.00001
loss_function: deepsdf # Use custom DeepSDF loss
optimizer: adam
network_learning_rate: 0.001 # lr=1e-3 from your notebook
latent_learning_rate: 0.01
latent_learning_rate: 0 #0.01
batch_size: 8 # BATCH_SIZE = 16 from your notebook
num_epochs: 1000 # max_epochs=20 from your notebook
max_points: 15000 # Allow more points for better SDF learning
max_points: 300000 # Allow more points for better SDF learning

# DeepSDF specific training parameters
sdf_delta: 1.0 # DELTA = 1.0 from your notebook
latent_sd: 0.01 # LATENT_SD = 0.01 from your notebook
latent_sd: .01
latent_mean: 0.0 # LATENT_MEAN = 0 from your notebook

# SDF Dataset configuration - ensure dataset_info is available
Expand All @@ -61,10 +62,10 @@ trainer_config:
# Model saving
save_model: true
save_best_only: true
save_frequency: 5
save_frequency: 20

evaluator_config:
skip_evaluation: false
skip_evaluation: true
# Evaluation specific configs for mesh extraction
batch_size: 8
resolution: 50 # Match training resolution for mesh extraction
Expand Down
Binary file added data/_processed/train/bimba_sampled_points.npy
Binary file not shown.
Binary file added data/_processed/train/bimba_signed_distances.npy
Binary file not shown.
Binary file added data/_processed/train/bunny_sampled_points.npy
Binary file not shown.
Binary file added data/_processed/train/bunny_signed_distances.npy
Binary file not shown.
Binary file added data/_processed/train/cup_sampled_points.npy
Binary file not shown.
Binary file added data/_processed/train/cup_signed_distances.npy
Binary file not shown.
Binary file added data/_processed/train/torus_sampled_points.npy
Binary file not shown.
Binary file added data/_processed/train/torus_signed_distances.npy
Binary file not shown.
File renamed without changes.
File renamed without changes.
Binary file removed data/processed/train/bimba_sampled_points.npy
Binary file not shown.
Binary file removed data/processed/train/bimba_signed_distances.npy
Binary file not shown.
Binary file removed data/processed/train/bunny_sampled_points.npy
Binary file not shown.
Binary file removed data/processed/train/bunny_signed_distances.npy
Binary file not shown.
Binary file added data/processed/train/cup_sampled_points.npy
Binary file not shown.
Binary file added data/processed/train/cup_signed_distances.npy
Binary file not shown.
Binary file added data/processed/train/torus_sampled_points.npy
Binary file not shown.
Binary file added data/processed/train/torus_signed_distances.npy
Binary file not shown.
Binary file added data/processed/val/cup_sampled_points.npy
Binary file not shown.
Binary file added data/processed/val/cup_signed_distances.npy
Binary file not shown.
Binary file removed data/processed/val/torus_sampled_points.npy
Binary file not shown.
Binary file removed data/processed/val/torus_signed_distances.npy
Binary file not shown.
Loading