Skip to content

Commit

Permalink
chore: deletes print() in unit tests (#306)
Browse files Browse the repository at this point in the history
  • Loading branch information
anthonyduong9 authored Sep 29, 2024
1 parent e88c15e commit 7d9fe10
Show file tree
Hide file tree
Showing 2 changed files with 0 additions and 11 deletions.
2 changes: 0 additions & 2 deletions tests/unit/test_evals.py
Original file line number Diff line number Diff line change
Expand Up @@ -155,7 +155,6 @@ def test_run_evals_training_sae(
eval_config=get_eval_everything_config(),
)

print(eval_metrics)
for key in all_expected_keys:
assert key in eval_metrics

Expand All @@ -178,7 +177,6 @@ def test_run_evals_training_sae_ignore_bos(
}, # type: ignore
)

print(eval_metrics)
for key in all_expected_keys:
assert key in eval_metrics

Expand Down
9 changes: 0 additions & 9 deletions tests/unit/training/test_cache_activations_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -25,17 +25,12 @@ def test_cache_activations_runner(tmp_path: Path):

# total_training_steps = 20_000
context_size = 1024
print(f"n tokens per context: {context_size}")
n_batches_in_buffer = 32
print(f"n batches in buffer: {n_batches_in_buffer}")
store_batch_size = 1
print(f"store_batch_size: {store_batch_size}")
n_buffers = 3
print(f"n_buffers: {n_buffers}")

tokens_in_buffer = n_batches_in_buffer * store_batch_size * context_size
total_training_tokens = n_buffers * tokens_in_buffer
print(f"Total Training Tokens: {total_training_tokens}")

# better if we can look at the files (change tmp_path to a real path to look at the files)
# tmp_path = os.path.join(os.path.dirname(__file__), "tmp")
Expand Down Expand Up @@ -103,13 +98,9 @@ def test_load_cached_activations():

# total_training_steps = 20_000
context_size = 256
print(f"n tokens per context: {context_size}")
n_batches_in_buffer = 32
print(f"n batches in buffer: {n_batches_in_buffer}")
store_batch_size = 1
print(f"store_batch_size: {store_batch_size}")
n_buffers = 3
print(f"n_buffers: {n_buffers}")

tokens_in_buffer = n_batches_in_buffer * store_batch_size * context_size
total_training_tokens = n_buffers * tokens_in_buffer
Expand Down

0 comments on commit 7d9fe10

Please sign in to comment.