Initial anonymous release for NeurIPS 2026 E&D submission
Browse files- LICENSE +37 -0
- README.md +118 -3
- code/_compute_n24_spearman.py +88 -0
- code/_compute_perm_test.py +121 -0
- code/_compute_within_arch_perm.py +84 -0
- code/_killer_experiment.py +568 -0
- code/_kinematics_train.py +258 -0
- code/_overnight_p1_transfer.py +503 -0
- code/_overnight_p3_matrix.py +259 -0
- code/_rev_linear_probe_matched_visual.py +269 -0
- code/_rev_m_continuous_bottleneck.py +595 -0
- code/_rev_n_multiprop_continuous.py +544 -0
- code/_rev_phys101_bottleneck_n192.py +136 -0
- code/_rev_phys101_global_bins.py +208 -0
- code/_rev_q_2nddir_singleprop_3seed.py +152 -0
- code/_rev_q_addendum2_high_posdis.py +343 -0
- code/_rev_q_addendum_multiprop.py +576 -0
- code/_rev_q_posdis_scatter.py +600 -0
- code/_rev_q_vqvae_bottleneck.py +450 -0
- croissant.json +204 -0
- features/vjepa2_collision_pooled.pt +3 -0
- labels/labels_collision.npz +3 -0
- labels/labels_elasticity.npz +3 -0
- labels/labels_flat_drop.npz +3 -0
- labels/labels_ramp.npz +3 -0
- labels/labels_ramp_3prop.npz +3 -0
LICENSE
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Apache License
|
| 2 |
+
Version 2.0, January 2004
|
| 3 |
+
http://www.apache.org/licenses/
|
| 4 |
+
|
| 5 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
| 6 |
+
you may not use this file except in compliance with the License.
|
| 7 |
+
You may obtain a copy of the License at
|
| 8 |
+
|
| 9 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
| 10 |
+
|
| 11 |
+
Unless required by applicable law or agreed to in writing, software
|
| 12 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
| 13 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 14 |
+
See the License for the specific language governing permissions and
|
| 15 |
+
limitations under the License.
|
| 16 |
+
|
| 17 |
+
Copyright 2026 Anonymous Authors (NeurIPS 2026 E&D Track Submission).
|
| 18 |
+
|
| 19 |
+
This release covers the benchmark protocol code, evaluation scripts,
|
| 20 |
+
pre-extracted feature tensors, and rendered Kubric scenarios contributed
|
| 21 |
+
by this submission. Upstream components retain their original licenses:
|
| 22 |
+
|
| 23 |
+
- Kubric (https://github.com/google-research/kubric)
|
| 24 |
+
Apache License 2.0
|
| 25 |
+
|
| 26 |
+
- Phys101 (Wu et al., BMVC 2016)
|
| 27 |
+
CC-BY 4.0 (we do not redistribute the Phys101 dataset itself; only
|
| 28 |
+
feature tensors extracted from it under the CC-BY redistribution
|
| 29 |
+
terms)
|
| 30 |
+
|
| 31 |
+
- V-JEPA 2 (facebook/vjepa2-vitl-fpc64-256)
|
| 32 |
+
CC-BY-NC 4.0 (Meta non-commercial research license; we redistribute
|
| 33 |
+
only feature tensors derived from it for non-commercial research)
|
| 34 |
+
|
| 35 |
+
- Other backbone checkpoints used for feature extraction retain their
|
| 36 |
+
respective licenses; this release ships derived feature tensors, not
|
| 37 |
+
the model weights themselves.
|
README.md
CHANGED
|
@@ -1,3 +1,118 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
tags:
|
| 6 |
+
- physics
|
| 7 |
+
- benchmark
|
| 8 |
+
- cross-scenario-transfer
|
| 9 |
+
- compositionality
|
| 10 |
+
- frozen-features
|
| 11 |
+
- video-foundation-models
|
| 12 |
+
- kubric
|
| 13 |
+
- phys101
|
| 14 |
+
size_categories:
|
| 15 |
+
- 1K<n<10K
|
| 16 |
+
task_categories:
|
| 17 |
+
- other
|
| 18 |
+
pretty_name: "Cross-Scenario Physics-Code Transfer Benchmark"
|
| 19 |
+
---
|
| 20 |
+
|
| 21 |
+
# Cross-Scenario Physics-Code Transfer Benchmark
|
| 22 |
+
|
| 23 |
+
Anonymous artifact for the **NeurIPS 2026 Evaluations & Datasets Track** submission *"A Benchmark for Cross-Scenario Physics-Code Transfer: Compositionality Metrics on Frozen Video Features."*
|
| 24 |
+
|
| 25 |
+
This Hugging Face repository accompanies the paper PDF and supplementary code zip submitted to OpenReview. It hosts the data tensors and labels reviewers need to inspect data quality, validate the protocol, and re-run all the analyses described in the paper.
|
| 26 |
+
|
| 27 |
+
## Repository contents
|
| 28 |
+
|
| 29 |
+
```
|
| 30 |
+
.
|
| 31 |
+
├── README.md this file
|
| 32 |
+
├── LICENSE Apache-2.0 (+ upstream attributions)
|
| 33 |
+
├── croissant.json Croissant v1.0 metadata + RAI fields
|
| 34 |
+
├── features/
|
| 35 |
+
│ └── vjepa2_collision_pooled.pt V-JEPA 2 features for 600 collision scenes
|
| 36 |
+
│ shape: [600, 4, 1024], float32,
|
| 37 |
+
│ mean-pooled over 4 evenly-spaced frames
|
| 38 |
+
├── labels/
|
| 39 |
+
│ ├── labels_collision.npz mass scalars/bins + restitution scalars/bins
|
| 40 |
+
│ │ for the 600 collision scenes
|
| 41 |
+
│ ├── labels_ramp.npz restitution + friction labels (ramp, 300 scenes)
|
| 42 |
+
│ ├── labels_flat_drop.npz restitution + friction labels (flat-drop, 300)
|
| 43 |
+
│ ├── labels_elasticity.npz restitution + drop-height labels (600 scenes)
|
| 44 |
+
│ └── labels_ramp_3prop.npz 3-property labels for ramp (multi-prop training)
|
| 45 |
+
└── code/ reproduction scripts (mirror of supplementary)
|
| 46 |
+
└── ... (17 .py files; see paper appendix)
|
| 47 |
+
```
|
| 48 |
+
|
| 49 |
+
## What this release covers vs. the full benchmark
|
| 50 |
+
|
| 51 |
+
The full benchmark described in the paper includes:
|
| 52 |
+
|
| 53 |
+
- **All four Kubric scenarios** (collision, ramp, flat-drop, elasticity) — 1,800 scenes total
|
| 54 |
+
- **75-scene matched-visual low-gravity collision variant**
|
| 55 |
+
- **Pre-extracted features for 8 frozen backbones** (V-JEPA 2, V-JEPA 2.1, DINOv2-S/L, CLIP ViT-L/14, MAE, SigLIP, VideoMAE)
|
| 56 |
+
- **Phys101 V-JEPA 2 features** (spring/ramp/fall, 2,673 clips)
|
| 57 |
+
- **Ground-truth per-object position + velocity tracks**
|
| 58 |
+
- **Rendered scene videos** (256×256, 48 frames at 24fps)
|
| 59 |
+
|
| 60 |
+
This Hugging Face repository hosts the **load-bearing subset** for reviewer inspection: V-JEPA 2 features for the collision source scenario, all 4 Kubric label files, and full reproduction code. Together these are sufficient to verify the within-scenario protocol, the label binning logic, the message-extraction pipeline, the permutation tests on the 24-config sweep, the within-architecture analyses, and the headline sufficiency claim.
|
| 61 |
+
|
| 62 |
+
The remaining ramp/flat-drop/elasticity feature tensors, the 8-backbone feature suite, the Phys101 features, the GT track tensors, and the rendered scene videos (~70 GB total) are prepared for an immediate post-acceptance public release with full author attribution. Anonymous reviewers requiring the full bundle for verification can request it through the OpenReview Author–Reviewer messaging channel; it will be made available on this repository under the same anonymous account.
|
| 63 |
+
|
| 64 |
+
## Reproducing the headline result
|
| 65 |
+
|
| 66 |
+
To verify the headline (top-5 vs bot-5 PosDis sufficiency observation, permutation $p=0.84$):
|
| 67 |
+
|
| 68 |
+
```bash
|
| 69 |
+
# 1. Download this repository
|
| 70 |
+
huggingface-cli download physics-code-transfer-bench/cross-scenario-physics-code-transfer \
|
| 71 |
+
--repo-type dataset --local-dir physics-bench
|
| 72 |
+
|
| 73 |
+
# 2. Set up the expected directory structure for the code
|
| 74 |
+
mkdir -p results/kinematics_vs_mechanics
|
| 75 |
+
cp physics-bench/features/* results/
|
| 76 |
+
cp physics-bench/labels/* results/kinematics_vs_mechanics/
|
| 77 |
+
|
| 78 |
+
# 3. Re-run the permutation test from the headline (reads paper-reported numbers, no GPU needed)
|
| 79 |
+
cd physics-bench/code
|
| 80 |
+
python _compute_perm_test.py
|
| 81 |
+
python _compute_within_arch_perm.py
|
| 82 |
+
```
|
| 83 |
+
|
| 84 |
+
Expected output: top-5 vs bot-5 PosDis two-sided $p = 0.84$ (matches paper); within-architecture permutation results match Section 4.5 of the paper.
|
| 85 |
+
|
| 86 |
+
To re-run the within-collision sender training (requires the V-JEPA 2 collision features, which are in this release):
|
| 87 |
+
|
| 88 |
+
```bash
|
| 89 |
+
PYTHONUNBUFFERED=1 PYTORCH_ENABLE_MPS_FALLBACK=1 \
|
| 90 |
+
python _rev_q_addendum2_high_posdis.py
|
| 91 |
+
```
|
| 92 |
+
|
| 93 |
+
## Croissant metadata
|
| 94 |
+
|
| 95 |
+
`croissant.json` is a Croissant v1.0 metadata file describing the benchmark, file formats, splits, and Responsible-AI annotations (data collection protocol, annotation protocol, preprocessing, use cases, limitations, social impact, biases, personal/sensitive information, release/maintenance plan). It has been validated locally with the [Croissant validator](https://github.com/mlcommons/croissant).
|
| 96 |
+
|
| 97 |
+
## Citation
|
| 98 |
+
|
| 99 |
+
```bibtex
|
| 100 |
+
@inproceedings{anonymous2026benchmark,
|
| 101 |
+
title = {A Benchmark for Cross-Scenario Physics-Code Transfer:
|
| 102 |
+
Compositionality Metrics on Frozen Video Features},
|
| 103 |
+
author = {Anonymous Authors},
|
| 104 |
+
booktitle = {NeurIPS 2026 Evaluations \& Datasets Track (under review)},
|
| 105 |
+
year = {2026}
|
| 106 |
+
}
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
## License
|
| 110 |
+
|
| 111 |
+
- This benchmark and accompanying code are released under the **Apache License 2.0** (consistent with the upstream Kubric license).
|
| 112 |
+
- **Phys101**: we redistribute only V-JEPA 2 features extracted from Phys101, not the source video; Phys101 itself remains under its original CC-BY 4.0 license.
|
| 113 |
+
- **V-JEPA 2 / V-JEPA 2.1**: features are derived from the publicly released encoders (Meta CC-BY-NC 4.0 research license); we redistribute only feature tensors for non-commercial research use.
|
| 114 |
+
- Full upstream attribution is in `LICENSE`.
|
| 115 |
+
|
| 116 |
+
## Anonymity statement
|
| 117 |
+
|
| 118 |
+
This repository is hosted under an anonymous account for double-blind NeurIPS 2026 review. No personally identifiable information (author names, institutions, contact emails) appears in the README, the LICENSE, the Croissant metadata, code, or any commit history. Public release with author attribution is contingent on acceptance.
|
code/_compute_n24_spearman.py
ADDED
|
@@ -0,0 +1,88 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""Compute Spearman correlations + bootstrap 95% CIs on the combined 24-config sweep
|
| 2 |
+
(18 original + 6 gap-fill)."""
|
| 3 |
+
import numpy as np
|
| 4 |
+
from scipy import stats
|
| 5 |
+
|
| 6 |
+
# All 24 configurations: (name, type, topsim, posdis, causal, cross16, cross192)
|
| 7 |
+
rows = [
|
| 8 |
+
# Original 12 single-property configs
|
| 9 |
+
("disc_L2_V5", "discrete", 0.88, 0.20, 0.02, 41.7, 43.9),
|
| 10 |
+
("disc_L2_V10", "discrete", 0.84, 0.25, 0.05, 46.1, 41.7),
|
| 11 |
+
("disc_L3_V5", "discrete", 0.84, 0.13, 0.02, 43.3, 42.8),
|
| 12 |
+
("disc_L3_V10", "discrete", 0.84, 0.12, 0.01, 43.3, 45.6),
|
| 13 |
+
("disc_L4_V5", "discrete", 0.90, 0.10, 0.01, 41.1, 42.2),
|
| 14 |
+
("disc_L4_V10", "discrete", 0.82, 0.08, 0.02, 45.0, 45.0),
|
| 15 |
+
("disc_L5_V5", "discrete", 0.89, 0.07, 0.02, 40.0, 43.9),
|
| 16 |
+
("cont_dim2", "continuous", 0.92, 0.15, 0.20, 48.9, 54.4),
|
| 17 |
+
("cont_dim3", "continuous", 0.91, 0.15, 0.02, 40.6, 41.1),
|
| 18 |
+
("cont_dim5", "continuous", 0.89, 0.06, 0.03, 47.2, 43.9),
|
| 19 |
+
("cont_dim10", "continuous", 0.88, 0.04, 0.01, 47.8, 48.3),
|
| 20 |
+
("cont_dim20", "continuous", 0.90, 0.02, 0.00, 48.9, 55.0),
|
| 21 |
+
# Original 3 multi-property 3-class
|
| 22 |
+
("disc_multi_L3_V5", "disc_multi", 0.59, 0.51, 0.06, 40.0, 46.1),
|
| 23 |
+
("disc_multi_L4_V10", "disc_multi", 0.68, 0.48, 0.01, 45.6, 50.6),
|
| 24 |
+
("cont_multi_dim3", "cont_multi", 0.72, 0.40, 0.10, 50.6, 55.0),
|
| 25 |
+
# Original 3 multi-property 5-class
|
| 26 |
+
("disc_multi5_L2_V5", "disc_multi", 0.78, 0.82, 0.07, 47.2, 52.2),
|
| 27 |
+
("disc_multi5_L3_V5", "disc_multi", 0.69, 0.83, 0.29, 45.0, 46.1),
|
| 28 |
+
("disc_multi5_L4_V5", "disc_multi", 0.68, 0.70, 0.06, 43.9, 47.8),
|
| 29 |
+
# NEW gap-fill (6 configs)
|
| 30 |
+
("disc_multi5_L2_V10_e250", "disc_multi", 0.66, 0.70, 0.12, 48.9, 55.6),
|
| 31 |
+
("disc_multi5_L3_V10_e250", "disc_multi", 0.60, 0.81, 0.03, 41.7, 43.3),
|
| 32 |
+
("disc_multi5_L4_V10_e250", "disc_multi", 0.65, 0.70, 0.07, 41.7, 41.7),
|
| 33 |
+
("disc_multi5_L2_V5_e200", "disc_multi", 0.75, 0.83, 0.13, 47.2, 51.1),
|
| 34 |
+
("disc_multi5_L4_V5_e250", "disc_multi", 0.79, 0.91, 0.03, 42.2, 46.7),
|
| 35 |
+
("disc_multi_L5_V5_3cls", "disc_multi", 0.72, 0.73, 0.02, 39.4, 42.2),
|
| 36 |
+
]
|
| 37 |
+
|
| 38 |
+
def boot_ci(x, y, n_resamples=5000, seed=42):
|
| 39 |
+
rng = np.random.default_rng(seed)
|
| 40 |
+
idx = np.arange(len(x))
|
| 41 |
+
rhos = []
|
| 42 |
+
for _ in range(n_resamples):
|
| 43 |
+
s = rng.choice(idx, size=len(idx), replace=True)
|
| 44 |
+
rho, _ = stats.spearmanr(x[s], y[s])
|
| 45 |
+
if not np.isnan(rho):
|
| 46 |
+
rhos.append(rho)
|
| 47 |
+
return float(np.percentile(rhos, 2.5)), float(np.percentile(rhos, 97.5))
|
| 48 |
+
|
| 49 |
+
topsim = np.array([r[2] for r in rows])
|
| 50 |
+
posdis = np.array([r[3] for r in rows])
|
| 51 |
+
causal = np.array([r[4] for r in rows])
|
| 52 |
+
cross16 = np.array([r[5] for r in rows])
|
| 53 |
+
cross192 = np.array([r[6] for r in rows])
|
| 54 |
+
n = len(rows)
|
| 55 |
+
|
| 56 |
+
print(f"=== n={n} configs (18 original + 6 gap-fill) ===\n")
|
| 57 |
+
print(f"PosDis range: {posdis.min():.2f} -- {posdis.max():.2f}")
|
| 58 |
+
print(f"Cross-scen N=192 range: {cross192.min():.1f}% -- {cross192.max():.1f}%")
|
| 59 |
+
print(f"Cross-scen N=16 range: {cross16.min():.1f}% -- {cross16.max():.1f}%")
|
| 60 |
+
print()
|
| 61 |
+
|
| 62 |
+
for x, xname in [(topsim, "TopSim"), (posdis, "PosDis"), (causal, "CausalSpec")]:
|
| 63 |
+
for y, yname in [(cross16, "Cross16"), (cross192, "Cross192")]:
|
| 64 |
+
rho, p = stats.spearmanr(x, y)
|
| 65 |
+
lo, hi = boot_ci(x, y)
|
| 66 |
+
print(f" {xname} vs {yname}: rho={rho:+.3f} p={p:.3f} CI=[{lo:+.2f}, {hi:+.2f}]")
|
| 67 |
+
|
| 68 |
+
# Also recompute for original n=18 (for paper consistency)
|
| 69 |
+
print(f"\n=== Original n=18 (for comparison) ===")
|
| 70 |
+
top18 = topsim[:18]; pd18 = posdis[:18]; ca18 = causal[:18]
|
| 71 |
+
c16_18 = cross16[:18]; c192_18 = cross192[:18]
|
| 72 |
+
for x, xname in [(top18, "TopSim"), (pd18, "PosDis"), (ca18, "CausalSpec")]:
|
| 73 |
+
for y, yname in [(c16_18, "Cross16"), (c192_18, "Cross192")]:
|
| 74 |
+
rho, p = stats.spearmanr(x, y)
|
| 75 |
+
lo, hi = boot_ci(x, y)
|
| 76 |
+
print(f" {xname} vs {yname}: rho={rho:+.3f} p={p:.3f} CI=[{lo:+.2f}, {hi:+.2f}]")
|
| 77 |
+
|
| 78 |
+
# Sufficiency observation: highest-PosDis configs vs lowest-PosDis
|
| 79 |
+
print(f"\n=== Sufficiency observation ===")
|
| 80 |
+
# Top 5 PosDis configs
|
| 81 |
+
top5_pd_idx = np.argsort(posdis)[-5:]
|
| 82 |
+
top5_pd = posdis[top5_pd_idx]
|
| 83 |
+
top5_cross = cross192[top5_pd_idx]
|
| 84 |
+
print(f"Top 5 PosDis: {top5_pd.tolist()} Cross192: {top5_cross.tolist()} range: {top5_cross.min():.1f}-{top5_cross.max():.1f}%")
|
| 85 |
+
bot5_pd_idx = np.argsort(posdis)[:5]
|
| 86 |
+
bot5_pd = posdis[bot5_pd_idx]
|
| 87 |
+
bot5_cross = cross192[bot5_pd_idx]
|
| 88 |
+
print(f"Bot 5 PosDis: {bot5_pd.tolist()} Cross192: {bot5_cross.tolist()} range: {bot5_cross.min():.1f}-{bot5_cross.max():.1f}%")
|
code/_compute_perm_test.py
ADDED
|
@@ -0,0 +1,121 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""Permutation test on the 5-vs-5 high/low PosDis band overlap (R2 ROI fix).
|
| 2 |
+
|
| 3 |
+
Tests: under random reassignment of cross-scenario accuracies to PosDis values,
|
| 4 |
+
how often do we see the observed flatness (top-5 mean - bot-5 mean) or larger
|
| 5 |
+
in absolute value? If the empirical |top5 - bot5| is similar to random, the
|
| 6 |
+
sufficiency claim is statistically defensible (no signal).
|
| 7 |
+
|
| 8 |
+
Also computes: probability of observing the observed band-overlap (range of top5
|
| 9 |
+
intersecting range of bot5) under random reshuffling.
|
| 10 |
+
"""
|
| 11 |
+
import numpy as np
|
| 12 |
+
from scipy import stats
|
| 13 |
+
|
| 14 |
+
# 24 configs from the paper
|
| 15 |
+
configs = [
|
| 16 |
+
# (name, posdis, cross192)
|
| 17 |
+
("disc_L2_V5", 0.20, 43.9),
|
| 18 |
+
("disc_L2_V10", 0.25, 41.7),
|
| 19 |
+
("disc_L3_V5", 0.13, 42.8),
|
| 20 |
+
("disc_L3_V10", 0.12, 45.6),
|
| 21 |
+
("disc_L4_V5", 0.10, 42.2),
|
| 22 |
+
("disc_L4_V10", 0.08, 45.0),
|
| 23 |
+
("disc_L5_V5", 0.07, 43.9),
|
| 24 |
+
("cont_dim2", 0.15, 54.4),
|
| 25 |
+
("cont_dim3", 0.15, 41.1),
|
| 26 |
+
("cont_dim5", 0.06, 43.9),
|
| 27 |
+
("cont_dim10", 0.04, 48.3),
|
| 28 |
+
("cont_dim20", 0.02, 55.0),
|
| 29 |
+
("disc_multi_L3_V5", 0.51, 46.1),
|
| 30 |
+
("disc_multi_L4_V10", 0.48, 50.6),
|
| 31 |
+
("cont_multi_dim3", 0.40, 55.0),
|
| 32 |
+
("disc_multi5_L2_V5", 0.82, 52.2),
|
| 33 |
+
("disc_multi5_L3_V5", 0.83, 46.1),
|
| 34 |
+
("disc_multi5_L4_V5", 0.70, 47.8),
|
| 35 |
+
("disc_multi5_L2_V10_e250", 0.70, 55.6),
|
| 36 |
+
("disc_multi5_L3_V10_e250", 0.81, 43.3),
|
| 37 |
+
("disc_multi5_L4_V10_e250", 0.70, 41.7),
|
| 38 |
+
("disc_multi5_L2_V5_e200", 0.83, 51.1),
|
| 39 |
+
("disc_multi5_L4_V5_e250", 0.91, 46.7),
|
| 40 |
+
("disc_multi_L5_V5_3cls", 0.73, 42.2),
|
| 41 |
+
]
|
| 42 |
+
posdis = np.array([c[1] for c in configs])
|
| 43 |
+
cross = np.array([c[2] for c in configs])
|
| 44 |
+
|
| 45 |
+
# Sort by PosDis
|
| 46 |
+
order = np.argsort(posdis)
|
| 47 |
+
top5_idx = order[-5:]
|
| 48 |
+
bot5_idx = order[:5]
|
| 49 |
+
top5_pd = posdis[top5_idx]; top5_cr = cross[top5_idx]
|
| 50 |
+
bot5_pd = posdis[bot5_idx]; bot5_cr = cross[bot5_idx]
|
| 51 |
+
print(f"Top-5 PosDis: {sorted(top5_pd.tolist())}, cross: {sorted(top5_cr.tolist())}")
|
| 52 |
+
print(f"Bot-5 PosDis: {sorted(bot5_pd.tolist())}, cross: {sorted(bot5_cr.tolist())}")
|
| 53 |
+
|
| 54 |
+
obs_diff_means = abs(top5_cr.mean() - bot5_cr.mean())
|
| 55 |
+
obs_top5_range = (top5_cr.min(), top5_cr.max())
|
| 56 |
+
obs_bot5_range = (bot5_cr.min(), bot5_cr.max())
|
| 57 |
+
obs_overlap = max(0, min(obs_top5_range[1], obs_bot5_range[1]) - max(obs_top5_range[0], obs_bot5_range[0]))
|
| 58 |
+
print(f"\nObserved |mean(top5) - mean(bot5)| = {obs_diff_means:.2f} pp")
|
| 59 |
+
print(f" top5 mean = {top5_cr.mean():.2f}, bot5 mean = {bot5_cr.mean():.2f}")
|
| 60 |
+
print(f" top5 range = [{obs_top5_range[0]:.1f}, {obs_top5_range[1]:.1f}]")
|
| 61 |
+
print(f" bot5 range = [{obs_bot5_range[0]:.1f}, {obs_bot5_range[1]:.1f}]")
|
| 62 |
+
print(f" observed band overlap = {obs_overlap:.2f} pp")
|
| 63 |
+
|
| 64 |
+
# Permutation: shuffle cross values, recompute top5 vs bot5 by FIXED PosDis ranks
|
| 65 |
+
n_perm = 100000
|
| 66 |
+
rng = np.random.default_rng(42)
|
| 67 |
+
diffs = []
|
| 68 |
+
overlaps = []
|
| 69 |
+
for _ in range(n_perm):
|
| 70 |
+
cross_perm = rng.permutation(cross)
|
| 71 |
+
t = cross_perm[top5_idx]; b = cross_perm[bot5_idx]
|
| 72 |
+
diffs.append(abs(t.mean() - b.mean()))
|
| 73 |
+
ov = max(0, min(t.max(), b.max()) - max(t.min(), b.min()))
|
| 74 |
+
overlaps.append(ov)
|
| 75 |
+
diffs = np.array(diffs); overlaps = np.array(overlaps)
|
| 76 |
+
|
| 77 |
+
p_diff = float(np.mean(diffs >= obs_diff_means))
|
| 78 |
+
p_overlap = float(np.mean(overlaps >= obs_overlap))
|
| 79 |
+
|
| 80 |
+
print(f"\nPermutation test (n_perm={n_perm}):")
|
| 81 |
+
print(f" p(|mean(top5) - mean(bot5)| >= observed {obs_diff_means:.2f}) = {p_diff:.3f}")
|
| 82 |
+
print(f" p(band overlap >= observed {obs_overlap:.2f}) = {p_overlap:.3f}")
|
| 83 |
+
print(f"\n null mean of |mean diff|: {diffs.mean():.2f} pp")
|
| 84 |
+
print(f" null 95th percentile of |mean diff|: {np.percentile(diffs, 95):.2f} pp")
|
| 85 |
+
|
| 86 |
+
# Same direction (low PosDis -> high cross): proper one-sided test
|
| 87 |
+
# Test: is top5 cross significantly LOWER than bot5 cross?
|
| 88 |
+
obs_signed_diff = top5_cr.mean() - bot5_cr.mean()
|
| 89 |
+
signed_diffs = []
|
| 90 |
+
for _ in range(n_perm):
|
| 91 |
+
cross_perm = rng.permutation(cross)
|
| 92 |
+
t = cross_perm[top5_idx]; b = cross_perm[bot5_idx]
|
| 93 |
+
signed_diffs.append(t.mean() - b.mean())
|
| 94 |
+
signed_diffs = np.array(signed_diffs)
|
| 95 |
+
p_lower = float(np.mean(signed_diffs <= obs_signed_diff))
|
| 96 |
+
print(f"\n observed signed mean(top5) - mean(bot5) = {obs_signed_diff:+.2f} pp")
|
| 97 |
+
print(f" p(top5 mean <= observed under null) = {p_lower:.3f}")
|
| 98 |
+
|
| 99 |
+
# k-robustness: same permutation test for k in {3..8} to address "5-vs-5 cherry pick" critique
|
| 100 |
+
print("\n=== k-robustness sweep: top-k vs bot-k (PosDis vs cross @N=192) ===")
|
| 101 |
+
print(f"{'k':>3s} | {'obs |diff|':>11s} | {'null mean':>10s} | {'null 95th':>10s} | {'p (two-sided)':>14s} | {'p (one-sided lower)':>20s}")
|
| 102 |
+
print("-" * 88)
|
| 103 |
+
k_results = []
|
| 104 |
+
for k in range(3, 9):
|
| 105 |
+
top_k = order[-k:]; bot_k = order[:k]
|
| 106 |
+
top_k_cr = cross[top_k]; bot_k_cr = cross[bot_k]
|
| 107 |
+
obs_abs = abs(top_k_cr.mean() - bot_k_cr.mean())
|
| 108 |
+
obs_signed = top_k_cr.mean() - bot_k_cr.mean()
|
| 109 |
+
null_abs = []; null_signed = []
|
| 110 |
+
for _ in range(n_perm):
|
| 111 |
+
cp = rng.permutation(cross)
|
| 112 |
+
null_abs.append(abs(cp[top_k].mean() - cp[bot_k].mean()))
|
| 113 |
+
null_signed.append(cp[top_k].mean() - cp[bot_k].mean())
|
| 114 |
+
null_abs = np.array(null_abs); null_signed = np.array(null_signed)
|
| 115 |
+
p_two = float(np.mean(null_abs >= obs_abs))
|
| 116 |
+
p_one_lower = float(np.mean(null_signed <= obs_signed))
|
| 117 |
+
k_results.append((k, obs_abs, null_abs.mean(), np.percentile(null_abs, 95), p_two, p_one_lower))
|
| 118 |
+
print(f"{k:>3d} | {obs_abs:>10.2f} | {null_abs.mean():>10.2f} | {np.percentile(null_abs, 95):>10.2f} | {p_two:>14.3f} | {p_one_lower:>20.3f}")
|
| 119 |
+
|
| 120 |
+
print("\nInterpretation: across all k in {3..8}, two-sided p > 0.5 (observed |diff| smaller than null mean)")
|
| 121 |
+
print("means top-k mean is no further from bot-k mean than random reshuffling produces.")
|
code/_compute_within_arch_perm.py
ADDED
|
@@ -0,0 +1,84 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""Within-architecture permutation tests on the 24-config sweep — top-3 vs bot-3
|
| 2 |
+
PosDis within discrete-only and within continuous-only subsets.
|
| 3 |
+
|
| 4 |
+
Addresses R2's flag that the pooled top-5 vs bot-5 comparison is partially confounded
|
| 5 |
+
with discrete-vs-continuous (top-5 are all discrete; continuous tops at PosDis 0.40).
|
| 6 |
+
"""
|
| 7 |
+
import numpy as np
|
| 8 |
+
|
| 9 |
+
# 24 configs from Table 6 (paper.tex)
|
| 10 |
+
# (name, kind, posdis, cross192)
|
| 11 |
+
configs = [
|
| 12 |
+
("disc_L2_V5", "disc", 0.20, 43.9),
|
| 13 |
+
("disc_L2_V10", "disc", 0.25, 41.7),
|
| 14 |
+
("disc_L3_V5", "disc", 0.13, 42.8),
|
| 15 |
+
("disc_L3_V10", "disc", 0.12, 45.6),
|
| 16 |
+
("disc_L4_V5", "disc", 0.10, 42.2),
|
| 17 |
+
("disc_L4_V10", "disc", 0.08, 45.0),
|
| 18 |
+
("disc_L5_V5", "disc", 0.07, 43.9),
|
| 19 |
+
("cont_dim2", "cont", 0.15, 54.4),
|
| 20 |
+
("cont_dim3", "cont", 0.15, 41.1),
|
| 21 |
+
("cont_dim5", "cont", 0.06, 43.9),
|
| 22 |
+
("cont_dim10", "cont", 0.04, 48.3),
|
| 23 |
+
("cont_dim20", "cont", 0.02, 55.0),
|
| 24 |
+
("disc_multi_L3_V5", "disc", 0.51, 46.1),
|
| 25 |
+
("disc_multi_L4_V10", "disc", 0.48, 50.6),
|
| 26 |
+
("cont_multi_dim3", "cont", 0.40, 55.0),
|
| 27 |
+
("disc_multi5_L2_V5", "disc", 0.82, 52.2),
|
| 28 |
+
("disc_multi5_L3_V5", "disc", 0.83, 46.1),
|
| 29 |
+
("disc_multi5_L4_V5", "disc", 0.70, 47.8),
|
| 30 |
+
("disc_multi5_L2_V10_e250", "disc", 0.70, 55.6),
|
| 31 |
+
("disc_multi5_L3_V10_e250", "disc", 0.81, 43.3),
|
| 32 |
+
("disc_multi5_L4_V10_e250", "disc", 0.70, 41.7),
|
| 33 |
+
("disc_multi5_L2_V5_e200", "disc", 0.83, 51.1),
|
| 34 |
+
("disc_multi5_L4_V5_e250", "disc", 0.91, 46.7),
|
| 35 |
+
("disc_multi_L5_V5_3cls", "disc", 0.73, 42.2),
|
| 36 |
+
]
|
| 37 |
+
|
| 38 |
+
posdis = np.array([c[2] for c in configs])
|
| 39 |
+
cross = np.array([c[3] for c in configs])
|
| 40 |
+
kind = np.array([c[1] for c in configs])
|
| 41 |
+
|
| 42 |
+
n_perm = 100_000
|
| 43 |
+
|
| 44 |
+
def run_perm(label, mask, k):
|
| 45 |
+
p = posdis[mask]; c = cross[mask]
|
| 46 |
+
n = len(p)
|
| 47 |
+
if n < 2 * k:
|
| 48 |
+
print(f" {label}: n={n} too small for top-{k}/bot-{k}")
|
| 49 |
+
return
|
| 50 |
+
order = np.argsort(p)
|
| 51 |
+
top_idx = order[-k:]
|
| 52 |
+
bot_idx = order[:k]
|
| 53 |
+
obs_diff = c[top_idx].mean() - c[bot_idx].mean()
|
| 54 |
+
obs_abs = abs(obs_diff)
|
| 55 |
+
rng = np.random.default_rng(42)
|
| 56 |
+
null_abs = []
|
| 57 |
+
for _ in range(n_perm):
|
| 58 |
+
cp = rng.permutation(c)
|
| 59 |
+
null_abs.append(abs(cp[top_idx].mean() - cp[bot_idx].mean()))
|
| 60 |
+
null_abs = np.array(null_abs)
|
| 61 |
+
p_two = float(np.mean(null_abs >= obs_abs))
|
| 62 |
+
print(f" {label} (n={n}, k={k}): obs |top-{k} - bot-{k}| = {obs_abs:.2f}pp, two-sided p = {p_two:.3f}")
|
| 63 |
+
print(f" top-{k} PosDis range: {p[top_idx].min():.2f}-{p[top_idx].max():.2f}, cross: {sorted(c[top_idx].tolist())}")
|
| 64 |
+
print(f" bot-{k} PosDis range: {p[bot_idx].min():.2f}-{p[bot_idx].max():.2f}, cross: {sorted(c[bot_idx].tolist())}")
|
| 65 |
+
|
| 66 |
+
print("=" * 60)
|
| 67 |
+
print("Within-architecture permutation tests (n_perm=10^5)")
|
| 68 |
+
print("=" * 60)
|
| 69 |
+
|
| 70 |
+
print("\n--- All 24 configs (pooled, for reference) ---")
|
| 71 |
+
for k in [3, 5]:
|
| 72 |
+
run_perm(f"All", np.ones(24, dtype=bool), k)
|
| 73 |
+
|
| 74 |
+
print("\n--- Discrete-only (n=19) ---")
|
| 75 |
+
disc_mask = (kind == "disc")
|
| 76 |
+
print(f" Total disc configs: {disc_mask.sum()}, PosDis range: {posdis[disc_mask].min():.2f}-{posdis[disc_mask].max():.2f}")
|
| 77 |
+
for k in [3, 4, 5]:
|
| 78 |
+
run_perm(f"Disc top-{k}/bot-{k}", disc_mask, k)
|
| 79 |
+
|
| 80 |
+
print("\n--- Continuous-only (n=6) ---")
|
| 81 |
+
cont_mask = (kind == "cont")
|
| 82 |
+
print(f" Total cont configs: {cont_mask.sum()}, PosDis range: {posdis[cont_mask].min():.2f}-{posdis[cont_mask].max():.2f}")
|
| 83 |
+
for k in [2, 3]:
|
| 84 |
+
run_perm(f"Cont top-{k}/bot-{k}", cont_mask, k)
|
code/_killer_experiment.py
ADDED
|
@@ -0,0 +1,568 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Killer Experiment: Discrete vs Continuous Bottleneck Head-to-Head
|
| 3 |
+
=================================================================
|
| 4 |
+
Three arms, same features, same task, same receiver:
|
| 5 |
+
Arm 1: Gumbel-Softmax discrete bottleneck (WMCP)
|
| 6 |
+
Arm 2: Continuous MLP bottleneck (same dim)
|
| 7 |
+
Arm 3: Raw features linear probe (no bottleneck)
|
| 8 |
+
|
| 9 |
+
10 seeds × 3 arms × 2 backbones = 60 runs.
|
| 10 |
+
|
| 11 |
+
Run:
|
| 12 |
+
PYTHONUNBUFFERED=1 PYTORCH_ENABLE_MPS_FALLBACK=1 python3 _killer_experiment.py
|
| 13 |
+
"""
|
| 14 |
+
|
| 15 |
+
import time, json, math, os, sys
|
| 16 |
+
import numpy as np
|
| 17 |
+
import torch
|
| 18 |
+
import torch.nn as nn
|
| 19 |
+
import torch.nn.functional as F
|
| 20 |
+
from pathlib import Path
|
| 21 |
+
from collections import defaultdict
|
| 22 |
+
from scipy import stats
|
| 23 |
+
|
| 24 |
+
DEVICE = torch.device("mps" if torch.backends.mps.is_available() else "cpu")
|
| 25 |
+
RESULTS_DIR = Path("results/killer_experiment")
|
| 26 |
+
RESULTS_DIR.mkdir(parents=True, exist_ok=True)
|
| 27 |
+
|
| 28 |
+
# Locked config
|
| 29 |
+
HIDDEN_DIM = 128
|
| 30 |
+
VOCAB_SIZE = 5
|
| 31 |
+
N_HEADS = 2
|
| 32 |
+
N_AGENTS = 4
|
| 33 |
+
MSG_DIM = N_AGENTS * N_HEADS * VOCAB_SIZE # 40
|
| 34 |
+
CONT_DIM = 40 # Same dimensionality for continuous arm
|
| 35 |
+
COMM_EPOCHS = 400
|
| 36 |
+
BATCH_SIZE = 32
|
| 37 |
+
SENDER_LR = 1e-3
|
| 38 |
+
RECEIVER_LR = 3e-3
|
| 39 |
+
EARLY_STOP = 150
|
| 40 |
+
N_SEEDS = 10
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
# ═══ Shared components ═══
|
| 44 |
+
|
| 45 |
+
class TemporalEncoder(nn.Module):
|
| 46 |
+
def __init__(self, hd=128, ind=1024, nf=4):
|
| 47 |
+
super().__init__()
|
| 48 |
+
ks = min(3, max(1, nf))
|
| 49 |
+
self.temporal = nn.Sequential(
|
| 50 |
+
nn.Conv1d(ind, 256, ks, padding=ks//2), nn.ReLU(),
|
| 51 |
+
nn.Conv1d(256, 128, ks, padding=ks//2), nn.ReLU(),
|
| 52 |
+
nn.AdaptiveAvgPool1d(1))
|
| 53 |
+
self.fc = nn.Sequential(nn.Linear(128, hd), nn.ReLU())
|
| 54 |
+
def forward(self, x):
|
| 55 |
+
return self.fc(self.temporal(x.permute(0, 2, 1)).squeeze(-1))
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
# ═══ ARM 1: Discrete (WMCP) ═══
|
| 59 |
+
|
| 60 |
+
class DiscreteSender(nn.Module):
|
| 61 |
+
def __init__(self, encoder, hd, vs, nh):
|
| 62 |
+
super().__init__()
|
| 63 |
+
self.encoder = encoder; self.vs = vs; self.nh = nh
|
| 64 |
+
self.heads = nn.ModuleList([nn.Linear(hd, vs) for _ in range(nh)])
|
| 65 |
+
|
| 66 |
+
def forward(self, x, tau=1.0, hard=True):
|
| 67 |
+
h = self.encoder(x)
|
| 68 |
+
msgs, logits_all = [], []
|
| 69 |
+
for head in self.heads:
|
| 70 |
+
logits = head(h)
|
| 71 |
+
if self.training:
|
| 72 |
+
msg = F.gumbel_softmax(logits, tau=tau, hard=hard)
|
| 73 |
+
else:
|
| 74 |
+
msg = F.one_hot(logits.argmax(-1), self.vs).float()
|
| 75 |
+
msgs.append(msg); logits_all.append(logits)
|
| 76 |
+
return torch.cat(msgs, -1), logits_all
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
class DiscreteMultiSender(nn.Module):
|
| 80 |
+
def __init__(self, senders):
|
| 81 |
+
super().__init__(); self.senders = nn.ModuleList(senders)
|
| 82 |
+
def forward(self, views, tau=1.0, hard=True):
|
| 83 |
+
msgs, all_logits = [], []
|
| 84 |
+
for s, v in zip(self.senders, views):
|
| 85 |
+
m, l = s(v, tau, hard); msgs.append(m); all_logits.extend(l)
|
| 86 |
+
return torch.cat(msgs, -1), all_logits
|
| 87 |
+
|
| 88 |
+
|
| 89 |
+
# ═══ ARM 2: Continuous ═══
|
| 90 |
+
|
| 91 |
+
class ContinuousSender(nn.Module):
|
| 92 |
+
"""Same architecture but outputs continuous vectors instead of discrete tokens."""
|
| 93 |
+
def __init__(self, encoder, hd, out_dim_per_agent):
|
| 94 |
+
super().__init__()
|
| 95 |
+
self.encoder = encoder
|
| 96 |
+
self.proj = nn.Sequential(
|
| 97 |
+
nn.Linear(hd, out_dim_per_agent),
|
| 98 |
+
nn.Tanh(), # Bounded like one-hot but continuous
|
| 99 |
+
)
|
| 100 |
+
|
| 101 |
+
def forward(self, x, tau=None, hard=None):
|
| 102 |
+
h = self.encoder(x)
|
| 103 |
+
msg = self.proj(h)
|
| 104 |
+
return msg, [] # No logits for continuous
|
| 105 |
+
|
| 106 |
+
|
| 107 |
+
class ContinuousMultiSender(nn.Module):
|
| 108 |
+
def __init__(self, senders):
|
| 109 |
+
super().__init__(); self.senders = nn.ModuleList(senders)
|
| 110 |
+
def forward(self, views, tau=None, hard=None):
|
| 111 |
+
msgs = [s(v)[0] for s, v in zip(self.senders, views)]
|
| 112 |
+
return torch.cat(msgs, -1), []
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
# ═══ ARM 3: Raw features (linear probe) ═══
|
| 116 |
+
|
| 117 |
+
class RawFeatureProbe(nn.Module):
|
| 118 |
+
"""Concatenate raw agent features, linear probe to prediction."""
|
| 119 |
+
def __init__(self, feat_dim, n_agents, n_frames_per_agent):
|
| 120 |
+
super().__init__()
|
| 121 |
+
self.pool = nn.AdaptiveAvgPool1d(1)
|
| 122 |
+
total_dim = feat_dim * n_agents
|
| 123 |
+
self.proj = nn.Sequential(
|
| 124 |
+
nn.Linear(total_dim, HIDDEN_DIM), nn.ReLU(),
|
| 125 |
+
nn.Linear(HIDDEN_DIM, 40), # Same dim as bottleneck for fair comparison
|
| 126 |
+
)
|
| 127 |
+
|
| 128 |
+
def forward(self, views, tau=None, hard=None):
|
| 129 |
+
pooled = []
|
| 130 |
+
for v in views:
|
| 131 |
+
# v: [B, T, D] -> pool over T
|
| 132 |
+
p = v.mean(dim=1) # [B, D]
|
| 133 |
+
pooled.append(p)
|
| 134 |
+
cat = torch.cat(pooled, -1) # [B, D*n_agents]
|
| 135 |
+
return self.proj(cat), []
|
| 136 |
+
|
| 137 |
+
|
| 138 |
+
# ═══ Shared receiver ═══
|
| 139 |
+
|
| 140 |
+
class Receiver(nn.Module):
|
| 141 |
+
def __init__(self, msg_dim, hd):
|
| 142 |
+
super().__init__()
|
| 143 |
+
self.net = nn.Sequential(
|
| 144 |
+
nn.Linear(msg_dim * 2, hd), nn.ReLU(),
|
| 145 |
+
nn.Linear(hd, hd // 2), nn.ReLU(),
|
| 146 |
+
nn.Linear(hd // 2, 1))
|
| 147 |
+
def forward(self, a, b):
|
| 148 |
+
return self.net(torch.cat([a, b], -1)).squeeze(-1)
|
| 149 |
+
|
| 150 |
+
|
| 151 |
+
# ═══ Metrics ═══
|
| 152 |
+
|
| 153 |
+
def mutual_information(x, y):
|
| 154 |
+
xv, yv = np.unique(x), np.unique(y)
|
| 155 |
+
n = len(x); mi = 0.0
|
| 156 |
+
for a in xv:
|
| 157 |
+
for b in yv:
|
| 158 |
+
pxy = np.sum((x==a)&(y==b))/n; px = np.sum(x==a)/n; py = np.sum(y==b)/n
|
| 159 |
+
if pxy>0 and px>0 and py>0: mi += pxy*np.log(pxy/(px*py))
|
| 160 |
+
return mi
|
| 161 |
+
|
| 162 |
+
def positional_disentanglement(tokens, attrs, vs):
|
| 163 |
+
np_, na = tokens.shape[1], attrs.shape[1]
|
| 164 |
+
mi = np.zeros((np_, na)); ents = []
|
| 165 |
+
for p in range(np_):
|
| 166 |
+
for a in range(na): mi[p,a] = mutual_information(tokens[:,p], attrs[:,a])
|
| 167 |
+
c = np.bincount(tokens[:,p], minlength=vs); pr = c/c.sum(); pr = pr[pr>0]
|
| 168 |
+
ents.append(float(-np.sum(pr*np.log(pr))/max(np.log(vs),1e-10)))
|
| 169 |
+
if np_>=2:
|
| 170 |
+
pd = 0.0
|
| 171 |
+
for p in range(np_):
|
| 172 |
+
s = np.sort(mi[p])[::-1]
|
| 173 |
+
if s[0]>1e-10: pd += (s[0]-s[1])/s[0]
|
| 174 |
+
pd /= np_
|
| 175 |
+
else: pd = 0.0
|
| 176 |
+
return float(pd), mi, ents
|
| 177 |
+
|
| 178 |
+
def continuous_disentanglement(embeddings, attrs):
|
| 179 |
+
"""PosDis analog for continuous vectors: bin each dimension, compute MI."""
|
| 180 |
+
n, d = embeddings.shape
|
| 181 |
+
n_bins = 5 # Bin continuous dims into 5 levels like discrete vocab
|
| 182 |
+
binned = np.zeros((n, d), dtype=int)
|
| 183 |
+
for dim in range(d):
|
| 184 |
+
try:
|
| 185 |
+
q = np.quantile(embeddings[:, dim], [0.2, 0.4, 0.6, 0.8])
|
| 186 |
+
binned[:, dim] = np.digitize(embeddings[:, dim], q)
|
| 187 |
+
except:
|
| 188 |
+
binned[:, dim] = 0
|
| 189 |
+
return positional_disentanglement(binned, attrs, n_bins)
|
| 190 |
+
|
| 191 |
+
def causal_specificity(sender, agent_views, mass_values, receiver, n_positions, is_discrete=True):
|
| 192 |
+
"""Zero each position, measure per-property accuracy drop."""
|
| 193 |
+
sender.eval(); receiver.eval()
|
| 194 |
+
n = len(agent_views[0])
|
| 195 |
+
mass_dev = torch.tensor(mass_values, dtype=torch.float32).to(DEVICE)
|
| 196 |
+
|
| 197 |
+
# Baseline accuracy
|
| 198 |
+
with torch.no_grad():
|
| 199 |
+
msgs_all = []
|
| 200 |
+
for i in range(0, n, BATCH_SIZE):
|
| 201 |
+
vs = [v[i:i+BATCH_SIZE].to(DEVICE) for v in agent_views]
|
| 202 |
+
m, _ = sender(vs)
|
| 203 |
+
msgs_all.append(m.cpu())
|
| 204 |
+
msgs_all = torch.cat(msgs_all, 0)
|
| 205 |
+
|
| 206 |
+
# Evaluate baseline
|
| 207 |
+
def eval_acc(msgs_tensor):
|
| 208 |
+
correct = total = 0
|
| 209 |
+
rng = np.random.RandomState(999)
|
| 210 |
+
for _ in range(50):
|
| 211 |
+
ia = rng.choice(n, min(32, n)); ib = rng.choice(n, min(32, n))
|
| 212 |
+
s = ia == ib
|
| 213 |
+
while s.any(): ib[s] = rng.choice(n, s.sum()); s = ia == ib
|
| 214 |
+
md = np.abs(mass_values[ia] - mass_values[ib])
|
| 215 |
+
k = md > 0.5
|
| 216 |
+
if k.sum() < 2: continue
|
| 217 |
+
ia, ib = ia[k], ib[k]
|
| 218 |
+
with torch.no_grad():
|
| 219 |
+
ma = msgs_tensor[ia].to(DEVICE); mb = msgs_tensor[ib].to(DEVICE)
|
| 220 |
+
pred = receiver(ma, mb) > 0
|
| 221 |
+
label = mass_dev[ia] > mass_dev[ib]
|
| 222 |
+
correct += (pred == label).sum().item(); total += len(label)
|
| 223 |
+
return correct / max(total, 1)
|
| 224 |
+
|
| 225 |
+
baseline_acc = eval_acc(msgs_all)
|
| 226 |
+
|
| 227 |
+
# Zero each position
|
| 228 |
+
drops = []
|
| 229 |
+
for pos in range(n_positions):
|
| 230 |
+
ablated = msgs_all.clone()
|
| 231 |
+
if is_discrete:
|
| 232 |
+
# Zero out the one-hot block for this position
|
| 233 |
+
start = pos * VOCAB_SIZE
|
| 234 |
+
ablated[:, start:start + VOCAB_SIZE] = 0
|
| 235 |
+
else:
|
| 236 |
+
ablated[:, pos] = 0
|
| 237 |
+
abl_acc = eval_acc(ablated)
|
| 238 |
+
drops.append(baseline_acc - abl_acc)
|
| 239 |
+
|
| 240 |
+
# Causal specificity = mean of (max_drop - mean_other_drops) / max_drop per position
|
| 241 |
+
if len(drops) >= 2:
|
| 242 |
+
specificity = []
|
| 243 |
+
for i in range(len(drops)):
|
| 244 |
+
others = [drops[j] for j in range(len(drops)) if j != i]
|
| 245 |
+
if drops[i] > 0.01:
|
| 246 |
+
specificity.append((drops[i] - np.mean(others)) / drops[i])
|
| 247 |
+
else:
|
| 248 |
+
specificity.append(0.0)
|
| 249 |
+
return float(np.mean(specificity)), drops, baseline_acc
|
| 250 |
+
return 0.0, drops, baseline_acc
|
| 251 |
+
|
| 252 |
+
|
| 253 |
+
# ═══ Training loop ═══
|
| 254 |
+
|
| 255 |
+
def train_arm(arm_name, sender, agent_views, mass_values, obj_names, seed, msg_dim=40):
|
| 256 |
+
"""Train one arm. Returns metrics dict."""
|
| 257 |
+
n = len(agent_views[0])
|
| 258 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 259 |
+
unique_objs = sorted(set(obj_names))
|
| 260 |
+
holdout_objs = set(rng.choice(unique_objs, max(4, len(unique_objs)//5), replace=False))
|
| 261 |
+
tr = np.array([i for i, o in enumerate(obj_names) if o not in holdout_objs])
|
| 262 |
+
ho = np.array([i for i, o in enumerate(obj_names) if o in holdout_objs])
|
| 263 |
+
if len(ho) < 4: return None
|
| 264 |
+
|
| 265 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 266 |
+
sender = sender.to(DEVICE)
|
| 267 |
+
receivers = [Receiver(msg_dim, HIDDEN_DIM).to(DEVICE) for _ in range(3)]
|
| 268 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 269 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 270 |
+
mass_dev = torch.tensor(mass_values, dtype=torch.float32).to(DEVICE)
|
| 271 |
+
max_ent = math.log(VOCAB_SIZE)
|
| 272 |
+
nb = max(1, len(tr) // BATCH_SIZE)
|
| 273 |
+
best_acc, best_state, best_ep = 0.0, None, 0
|
| 274 |
+
t0 = time.time()
|
| 275 |
+
is_discrete = arm_name == "discrete"
|
| 276 |
+
|
| 277 |
+
for ep in range(COMM_EPOCHS):
|
| 278 |
+
if time.time() - t0 > 600: break
|
| 279 |
+
if ep - best_ep > EARLY_STOP and best_acc > 0.55: break
|
| 280 |
+
if ep > 0 and ep % 40 == 0:
|
| 281 |
+
for i in range(len(receivers)):
|
| 282 |
+
receivers[i] = Receiver(msg_dim, HIDDEN_DIM).to(DEVICE)
|
| 283 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 284 |
+
|
| 285 |
+
sender.train(); [r.train() for r in receivers]
|
| 286 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, COMM_EPOCHS - 1)
|
| 287 |
+
hard = ep >= 30
|
| 288 |
+
|
| 289 |
+
for _ in range(nb):
|
| 290 |
+
ia = rng.choice(tr, BATCH_SIZE); ib = rng.choice(tr, BATCH_SIZE)
|
| 291 |
+
s = ia == ib
|
| 292 |
+
while s.any(): ib[s] = rng.choice(tr, s.sum()); s = ia == ib
|
| 293 |
+
md = np.abs(mass_values[ia] - mass_values[ib]); k = md > 0.5
|
| 294 |
+
if k.sum() < 4: continue
|
| 295 |
+
ia, ib = ia[k], ib[k]
|
| 296 |
+
va = [v[ia].to(DEVICE) for v in agent_views]
|
| 297 |
+
vb = [v[ib].to(DEVICE) for v in agent_views]
|
| 298 |
+
label = (mass_dev[ia] > mass_dev[ib]).float()
|
| 299 |
+
|
| 300 |
+
if is_discrete:
|
| 301 |
+
ma, la = sender(va, tau=tau, hard=hard)
|
| 302 |
+
mb, lb = sender(vb, tau=tau, hard=hard)
|
| 303 |
+
else:
|
| 304 |
+
ma, la = sender(va)
|
| 305 |
+
mb, lb = sender(vb)
|
| 306 |
+
|
| 307 |
+
loss = sum(F.binary_cross_entropy_with_logits(r(ma, mb), label) for r in receivers) / len(receivers)
|
| 308 |
+
|
| 309 |
+
# Entropy reg only for discrete
|
| 310 |
+
if is_discrete and la:
|
| 311 |
+
for lg in la + lb:
|
| 312 |
+
lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8)
|
| 313 |
+
ent = -(p * lp).sum(-1).mean()
|
| 314 |
+
if ent / max_ent < 0.1: loss = loss - 0.03 * ent
|
| 315 |
+
|
| 316 |
+
if torch.isnan(loss): so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 317 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; loss.backward()
|
| 318 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 319 |
+
so.step(); [o.step() for o in ros]
|
| 320 |
+
|
| 321 |
+
if ep % 50 == 0: torch.mps.empty_cache()
|
| 322 |
+
if (ep+1) % 50 == 0 or ep == 0:
|
| 323 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 324 |
+
with torch.no_grad():
|
| 325 |
+
c = t = 0; er = np.random.RandomState(999)
|
| 326 |
+
for _ in range(30):
|
| 327 |
+
ia_h = er.choice(ho, min(32, len(ho))); ib_h = er.choice(ho, min(32, len(ho)))
|
| 328 |
+
s2 = ia_h == ib_h
|
| 329 |
+
while s2.any(): ib_h[s2] = er.choice(ho, s2.sum()); s2 = ia_h == ib_h
|
| 330 |
+
mdh = np.abs(mass_values[ia_h] - mass_values[ib_h]); kh = mdh > 0.5
|
| 331 |
+
if kh.sum() < 2: continue
|
| 332 |
+
ia_h, ib_h = ia_h[kh], ib_h[kh]
|
| 333 |
+
vh = [v[ia_h].to(DEVICE) for v in agent_views]
|
| 334 |
+
wh = [v[ib_h].to(DEVICE) for v in agent_views]
|
| 335 |
+
mah, _ = sender(vh); mbh, _ = sender(wh)
|
| 336 |
+
for r in receivers:
|
| 337 |
+
c += ((r(mah, mbh) > 0) == (mass_dev[ia_h] > mass_dev[ib_h])).sum().item()
|
| 338 |
+
t += len(ia_h)
|
| 339 |
+
acc = c / max(t, 1)
|
| 340 |
+
if acc > best_acc:
|
| 341 |
+
best_acc = acc; best_ep = ep
|
| 342 |
+
best_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 343 |
+
|
| 344 |
+
if best_state: sender.load_state_dict(best_state)
|
| 345 |
+
sender.eval()
|
| 346 |
+
|
| 347 |
+
# Extract representations
|
| 348 |
+
all_repr = []
|
| 349 |
+
with torch.no_grad():
|
| 350 |
+
for i in range(0, n, BATCH_SIZE):
|
| 351 |
+
vs = [v[i:i+BATCH_SIZE].to(DEVICE) for v in agent_views]
|
| 352 |
+
m, logits = sender(vs)
|
| 353 |
+
all_repr.append(m.cpu())
|
| 354 |
+
all_repr = torch.cat(all_repr, 0).numpy()
|
| 355 |
+
|
| 356 |
+
# Compositionality
|
| 357 |
+
mass_bins = np.digitize(mass_values, np.quantile(mass_values, [0.2, 0.4, 0.6, 0.8]))
|
| 358 |
+
uo = sorted(set(obj_names)); oi = {o: i for i, o in enumerate(uo)}
|
| 359 |
+
obj_bins = np.digitize(np.array([oi[o] for o in obj_names]),
|
| 360 |
+
np.quantile(np.arange(len(uo)), [0.2, 0.4, 0.6, 0.8]))
|
| 361 |
+
attrs = np.stack([mass_bins, obj_bins], axis=1)
|
| 362 |
+
|
| 363 |
+
if is_discrete:
|
| 364 |
+
# Extract tokens
|
| 365 |
+
tokens = []
|
| 366 |
+
with torch.no_grad():
|
| 367 |
+
for i in range(0, n, BATCH_SIZE):
|
| 368 |
+
vs = [v[i:i+BATCH_SIZE].to(DEVICE) for v in agent_views]
|
| 369 |
+
_, logits = sender(vs)
|
| 370 |
+
tokens.append(np.stack([l.argmax(-1).cpu().numpy() for l in logits], 1))
|
| 371 |
+
tokens = np.concatenate(tokens, 0)
|
| 372 |
+
posdis, mi_mat, ents = positional_disentanglement(tokens, attrs, VOCAB_SIZE)
|
| 373 |
+
n_positions = tokens.shape[1]
|
| 374 |
+
else:
|
| 375 |
+
posdis, mi_mat, ents = continuous_disentanglement(all_repr, attrs)
|
| 376 |
+
n_positions = all_repr.shape[1]
|
| 377 |
+
|
| 378 |
+
# Causal specificity
|
| 379 |
+
cs, drops, base_acc = causal_specificity(
|
| 380 |
+
sender, agent_views, mass_values, receivers[0], n_positions, is_discrete)
|
| 381 |
+
|
| 382 |
+
# TopSim
|
| 383 |
+
ts = 0.0
|
| 384 |
+
if is_discrete:
|
| 385 |
+
from scipy.stats import spearmanr
|
| 386 |
+
rng2 = np.random.RandomState(42)
|
| 387 |
+
idx_a = rng2.randint(0, n, 5000); idx_b = rng2.randint(0, n, 5000)
|
| 388 |
+
meaning_d = np.abs(mass_bins[idx_a] - mass_bins[idx_b]) + np.abs(obj_bins[idx_a] - obj_bins[idx_b])
|
| 389 |
+
msg_d = np.sum(tokens[idx_a] != tokens[idx_b], axis=1)
|
| 390 |
+
ts_val, _ = spearmanr(meaning_d, msg_d)
|
| 391 |
+
ts = float(ts_val) if not np.isnan(ts_val) else 0.0
|
| 392 |
+
|
| 393 |
+
return {
|
| 394 |
+
"arm": arm_name,
|
| 395 |
+
"accuracy": float(best_acc),
|
| 396 |
+
"posdis": float(posdis),
|
| 397 |
+
"topsim": float(ts),
|
| 398 |
+
"causal_specificity": float(cs),
|
| 399 |
+
"causal_drops": [float(d) for d in drops],
|
| 400 |
+
"converge_epoch": best_ep + 1,
|
| 401 |
+
"elapsed_s": time.time() - t0,
|
| 402 |
+
}
|
| 403 |
+
|
| 404 |
+
|
| 405 |
+
# ═══ Main ═══
|
| 406 |
+
|
| 407 |
+
def run():
|
| 408 |
+
print("╔══════════════════════════════════════════════════════════╗", flush=True)
|
| 409 |
+
print("║ KILLER EXPERIMENT: Discrete vs Continuous Bottleneck ║", flush=True)
|
| 410 |
+
print("╚══════════════════════════════════════════════════════════╝", flush=True)
|
| 411 |
+
t_total = time.time()
|
| 412 |
+
|
| 413 |
+
# Load features
|
| 414 |
+
vjepa_data = torch.load("results/phase87_phys101_spring_features.pt", weights_only=False)
|
| 415 |
+
dino_data = torch.load("results/phase87_phys101_spring_static.pt", weights_only=False)
|
| 416 |
+
|
| 417 |
+
backbones = {
|
| 418 |
+
"vjepa2": {
|
| 419 |
+
"feat": vjepa_data["features"].float(),
|
| 420 |
+
"dim": 1024,
|
| 421 |
+
},
|
| 422 |
+
"dinov2": {
|
| 423 |
+
"feat": dino_data["features"].float().unsqueeze(1).expand(-1, 8, -1).contiguous(),
|
| 424 |
+
"dim": 384,
|
| 425 |
+
},
|
| 426 |
+
}
|
| 427 |
+
obj_names = vjepa_data["obj_names"]
|
| 428 |
+
mass_values = vjepa_data["mass_values"]
|
| 429 |
+
n_frames = 8
|
| 430 |
+
fpa = n_frames // N_AGENTS
|
| 431 |
+
|
| 432 |
+
all_results = []
|
| 433 |
+
|
| 434 |
+
for bb_name, bb_data in backbones.items():
|
| 435 |
+
feat = bb_data["feat"]
|
| 436 |
+
dim = bb_data["dim"]
|
| 437 |
+
print(f"\n{'='*60}", flush=True)
|
| 438 |
+
print(f" BACKBONE: {bb_name} ({dim}-dim)", flush=True)
|
| 439 |
+
print(f"{'='*60}", flush=True)
|
| 440 |
+
|
| 441 |
+
agent_views = [feat[:, i*fpa:(i+1)*fpa, :] for i in range(N_AGENTS)]
|
| 442 |
+
|
| 443 |
+
for arm_name in ["discrete", "continuous", "raw_probe"]:
|
| 444 |
+
print(f"\n ── {arm_name} ──", flush=True)
|
| 445 |
+
|
| 446 |
+
for seed in range(N_SEEDS):
|
| 447 |
+
torch.manual_seed(seed)
|
| 448 |
+
|
| 449 |
+
if arm_name == "discrete":
|
| 450 |
+
senders = [DiscreteSender(TemporalEncoder(HIDDEN_DIM, dim, fpa),
|
| 451 |
+
HIDDEN_DIM, VOCAB_SIZE, N_HEADS) for _ in range(N_AGENTS)]
|
| 452 |
+
sender = DiscreteMultiSender(senders)
|
| 453 |
+
msg_dim = MSG_DIM
|
| 454 |
+
|
| 455 |
+
elif arm_name == "continuous":
|
| 456 |
+
per_agent_dim = CONT_DIM // N_AGENTS # 10 per agent
|
| 457 |
+
senders = [ContinuousSender(TemporalEncoder(HIDDEN_DIM, dim, fpa),
|
| 458 |
+
HIDDEN_DIM, per_agent_dim) for _ in range(N_AGENTS)]
|
| 459 |
+
sender = ContinuousMultiSender(senders)
|
| 460 |
+
msg_dim = CONT_DIM
|
| 461 |
+
|
| 462 |
+
elif arm_name == "raw_probe":
|
| 463 |
+
sender = RawFeatureProbe(dim, N_AGENTS, fpa)
|
| 464 |
+
msg_dim = 40
|
| 465 |
+
|
| 466 |
+
r = train_arm(arm_name, sender, agent_views, mass_values, obj_names, seed, msg_dim)
|
| 467 |
+
if r is None: continue
|
| 468 |
+
r["backbone"] = bb_name
|
| 469 |
+
r["seed"] = seed
|
| 470 |
+
all_results.append(r)
|
| 471 |
+
|
| 472 |
+
print(f" Seed {seed}: acc={r['accuracy']:.1%} "
|
| 473 |
+
f"PD={r['posdis']:.3f} CS={r['causal_specificity']:.3f}", flush=True)
|
| 474 |
+
|
| 475 |
+
torch.mps.empty_cache()
|
| 476 |
+
|
| 477 |
+
# ═══ Summary ═══
|
| 478 |
+
print(f"\n{'='*70}", flush=True)
|
| 479 |
+
print(f" DISCRETE vs CONTINUOUS: HEAD-TO-HEAD RESULTS", flush=True)
|
| 480 |
+
print(f"{'='*70}", flush=True)
|
| 481 |
+
print(f" {'Backbone':<10s} {'Arm':<12s} │ {'Accuracy':>10s} │ {'PosDis':>10s} │ "
|
| 482 |
+
f"{'Causal Spec':>12s} │ {'TopSim':>8s}", flush=True)
|
| 483 |
+
print(f" {'─'*10} {'─'*12} ┼ {'─'*10} ┼ {'─'*10} ┼ {'─'*12} ┼ {'─'*8}", flush=True)
|
| 484 |
+
|
| 485 |
+
summary = {}
|
| 486 |
+
for bb in ["vjepa2", "dinov2"]:
|
| 487 |
+
for arm in ["discrete", "continuous", "raw_probe"]:
|
| 488 |
+
runs = [r for r in all_results if r["backbone"] == bb and r["arm"] == arm]
|
| 489 |
+
if not runs: continue
|
| 490 |
+
accs = [r["accuracy"] for r in runs]
|
| 491 |
+
pds = [r["posdis"] for r in runs]
|
| 492 |
+
css = [r["causal_specificity"] for r in runs]
|
| 493 |
+
tss = [r["topsim"] for r in runs]
|
| 494 |
+
key = f"{bb}_{arm}"
|
| 495 |
+
summary[key] = {
|
| 496 |
+
"acc": f"{np.mean(accs):.1%}±{np.std(accs):.1%}",
|
| 497 |
+
"posdis": f"{np.mean(pds):.3f}±{np.std(pds):.3f}",
|
| 498 |
+
"causal_spec": f"{np.mean(css):.3f}±{np.std(css):.3f}",
|
| 499 |
+
"topsim": f"{np.mean(tss):.3f}±{np.std(tss):.3f}",
|
| 500 |
+
"acc_mean": float(np.mean(accs)),
|
| 501 |
+
"pd_mean": float(np.mean(pds)),
|
| 502 |
+
"cs_mean": float(np.mean(css)),
|
| 503 |
+
}
|
| 504 |
+
print(f" {bb:<10s} {arm:<12s} │ {np.mean(accs):>9.1%} │ "
|
| 505 |
+
f"{np.mean(pds):>9.3f} │ {np.mean(css):>11.3f} │ "
|
| 506 |
+
f"{np.mean(tss):>7.3f}", flush=True)
|
| 507 |
+
|
| 508 |
+
# ═══ THE VERDICT ═══
|
| 509 |
+
print(f"\n ╔═══ THE VERDICT ═══╗", flush=True)
|
| 510 |
+
for bb in ["vjepa2", "dinov2"]:
|
| 511 |
+
disc = summary.get(f"{bb}_discrete", {})
|
| 512 |
+
cont = summary.get(f"{bb}_continuous", {})
|
| 513 |
+
if disc and cont:
|
| 514 |
+
pd_gap = disc.get("pd_mean", 0) - cont.get("pd_mean", 0)
|
| 515 |
+
acc_gap = disc.get("acc_mean", 0) - cont.get("acc_mean", 0)
|
| 516 |
+
cs_gap = disc.get("cs_mean", 0) - cont.get("cs_mean", 0)
|
| 517 |
+
print(f" ║ {bb}:", flush=True)
|
| 518 |
+
print(f" ║ PosDis: discrete {disc.get('pd_mean',0):.3f} vs continuous {cont.get('pd_mean',0):.3f} "
|
| 519 |
+
f"(Δ={pd_gap:+.3f})", flush=True)
|
| 520 |
+
print(f" ║ Accuracy: discrete {disc.get('acc_mean',0):.1%} vs continuous {cont.get('acc_mean',0):.1%} "
|
| 521 |
+
f"(Δ={acc_gap:+.1%})", flush=True)
|
| 522 |
+
print(f" ║ Causal: discrete {disc.get('cs_mean',0):.3f} vs continuous {cont.get('cs_mean',0):.3f} "
|
| 523 |
+
f"(Δ={cs_gap:+.3f})", flush=True)
|
| 524 |
+
if pd_gap > 0.3 and abs(acc_gap) < 0.05:
|
| 525 |
+
print(f" ║ → DISCRETE WINS on interpretability, competitive on accuracy", flush=True)
|
| 526 |
+
elif pd_gap < 0.1:
|
| 527 |
+
print(f" ║ → NO CLEAR WINNER on interpretability", flush=True)
|
| 528 |
+
else:
|
| 529 |
+
print(f" ║ → DISCRETE ADVANTAGE: +{pd_gap:.3f} PosDis", flush=True)
|
| 530 |
+
print(f" ╚═══════════════════╝", flush=True)
|
| 531 |
+
|
| 532 |
+
# Save
|
| 533 |
+
with open(RESULTS_DIR / "results.json", "w") as f:
|
| 534 |
+
json.dump({"summary": summary, "all_runs": all_results}, f, indent=2, default=str)
|
| 535 |
+
|
| 536 |
+
# Plot
|
| 537 |
+
import matplotlib; matplotlib.use("Agg"); import matplotlib.pyplot as plt
|
| 538 |
+
|
| 539 |
+
fig, axes = plt.subplots(1, 3, figsize=(15, 5))
|
| 540 |
+
fig.suptitle("Discrete vs Continuous Bottleneck: Head-to-Head", fontsize=14, fontweight='bold')
|
| 541 |
+
colors = {"discrete": "#2196F3", "continuous": "#F44336", "raw_probe": "#9E9E9E"}
|
| 542 |
+
labels = {"discrete": "WMCP (discrete)", "continuous": "Continuous MLP", "raw_probe": "Raw features"}
|
| 543 |
+
|
| 544 |
+
for ax, metric, ylabel in [(axes[0], "acc_mean", "Accuracy"),
|
| 545 |
+
(axes[1], "pd_mean", "PosDis"),
|
| 546 |
+
(axes[2], "cs_mean", "Causal Specificity")]:
|
| 547 |
+
x = np.arange(2); width = 0.25
|
| 548 |
+
for i, arm in enumerate(["discrete", "continuous", "raw_probe"]):
|
| 549 |
+
vals = []
|
| 550 |
+
for bb in ["vjepa2", "dinov2"]:
|
| 551 |
+
key = f"{bb}_{arm}"
|
| 552 |
+
vals.append(summary.get(key, {}).get(metric, 0))
|
| 553 |
+
ax.bar(x + i * width, vals, width, label=labels[arm], color=colors[arm], alpha=0.8)
|
| 554 |
+
ax.set_xticks(x + width); ax.set_xticklabels(["V-JEPA 2", "DINOv2"])
|
| 555 |
+
ax.set_ylabel(ylabel); ax.set_title(ylabel)
|
| 556 |
+
if metric == "pd_mean": ax.set_ylim(0, 1); ax.legend(fontsize=8)
|
| 557 |
+
|
| 558 |
+
plt.tight_layout()
|
| 559 |
+
plt.savefig(RESULTS_DIR / "discrete_vs_continuous.png", dpi=200, bbox_inches='tight')
|
| 560 |
+
plt.close()
|
| 561 |
+
print(f"\n Saved results/killer_experiment/", flush=True)
|
| 562 |
+
|
| 563 |
+
total_min = (time.time() - t_total) / 60
|
| 564 |
+
print(f" Total: {total_min:.1f} min", flush=True)
|
| 565 |
+
|
| 566 |
+
|
| 567 |
+
if __name__ == "__main__":
|
| 568 |
+
run()
|
code/_kinematics_train.py
ADDED
|
@@ -0,0 +1,258 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Phase 0B: single-target bottleneck trainer for kinematics-vs-mechanics.
|
| 3 |
+
|
| 4 |
+
Same sender architecture as `_killer_experiment.py` (4-agent × 2-head
|
| 5 |
+
Gumbel-Softmax, K=5, HIDDEN_DIM=128) but the receiver is a 3-way classifier
|
| 6 |
+
that reads a single message and predicts the binned target class.
|
| 7 |
+
|
| 8 |
+
Takes `--target <name>` and `--dataset <collision|ramp>` and pulls the
|
| 9 |
+
corresponding `_bin` column from the label .npz. Feature file and per-dataset
|
| 10 |
+
config are selected by `--backbone <vjepa2|dinov2>`.
|
| 11 |
+
|
| 12 |
+
Run:
|
| 13 |
+
/usr/bin/python3 _kinematics_train.py --dataset collision --backbone vjepa2 \
|
| 14 |
+
--target mass --seed 0
|
| 15 |
+
"""
|
| 16 |
+
import argparse, json, math, os, sys, time, warnings
|
| 17 |
+
warnings.filterwarnings("ignore")
|
| 18 |
+
|
| 19 |
+
import numpy as np
|
| 20 |
+
import torch
|
| 21 |
+
import torch.nn as nn
|
| 22 |
+
import torch.nn.functional as F
|
| 23 |
+
|
| 24 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 25 |
+
from _killer_experiment import (
|
| 26 |
+
TemporalEncoder, DiscreteSender, DiscreteMultiSender
|
| 27 |
+
)
|
| 28 |
+
|
| 29 |
+
DEVICE = torch.device("mps" if torch.backends.mps.is_available() else "cpu")
|
| 30 |
+
|
| 31 |
+
HIDDEN_DIM = 128
|
| 32 |
+
VOCAB_SIZE = 5
|
| 33 |
+
N_HEADS = 2
|
| 34 |
+
N_AGENTS = 4
|
| 35 |
+
MSG_DIM = N_AGENTS * N_HEADS * VOCAB_SIZE # 40
|
| 36 |
+
N_POS = N_AGENTS * N_HEADS # 8
|
| 37 |
+
BATCH_SIZE = 32
|
| 38 |
+
SENDER_LR = 1e-3
|
| 39 |
+
RECEIVER_LR = 3e-3
|
| 40 |
+
EARLY_STOP_PATIENCE = 50 # matches exp4_faithfulness early-stop style
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
# ── Classifier receiver ──
|
| 44 |
+
|
| 45 |
+
class ClassifierReceiver(nn.Module):
|
| 46 |
+
"""Reads one message and predicts 3 class logits."""
|
| 47 |
+
|
| 48 |
+
def __init__(self, msg_dim, hidden_dim, n_classes=3):
|
| 49 |
+
super().__init__()
|
| 50 |
+
self.net = nn.Sequential(
|
| 51 |
+
nn.Linear(msg_dim, hidden_dim), nn.ReLU(),
|
| 52 |
+
nn.Linear(hidden_dim, hidden_dim // 2), nn.ReLU(),
|
| 53 |
+
nn.Linear(hidden_dim // 2, n_classes),
|
| 54 |
+
)
|
| 55 |
+
|
| 56 |
+
def forward(self, msg):
|
| 57 |
+
return self.net(msg)
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
# ── Dataset / label loading ──
|
| 61 |
+
|
| 62 |
+
FEATURE_FILES = {
|
| 63 |
+
("collision", "vjepa2"): "results/vjepa2_collision_pooled.pt", # [600, 24, 1024]
|
| 64 |
+
("collision", "dinov2"): "results/collision_dinov2_features.pt", # [600, 24, 384]
|
| 65 |
+
("ramp", "vjepa2"): "results/vjepa2_ramp_temporal.pt", # [300, 16, 1024]
|
| 66 |
+
("ramp", "dinov2"): "results/phase54b_dino_features.pt", # [300, 8, 384]
|
| 67 |
+
}
|
| 68 |
+
|
| 69 |
+
LABEL_FILES = {
|
| 70 |
+
"collision": "results/kinematics_vs_mechanics/labels_collision.npz",
|
| 71 |
+
"ramp": "results/kinematics_vs_mechanics/labels_ramp.npz",
|
| 72 |
+
}
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
def load_features(dataset, backbone):
|
| 76 |
+
path = FEATURE_FILES[(dataset, backbone)]
|
| 77 |
+
d = torch.load(path, weights_only=False, map_location="cpu")
|
| 78 |
+
feat = d["features"].float()
|
| 79 |
+
return feat
|
| 80 |
+
|
| 81 |
+
|
| 82 |
+
def load_labels(dataset, target):
|
| 83 |
+
"""Return bin labels (int [N])."""
|
| 84 |
+
z = np.load(LABEL_FILES[dataset])
|
| 85 |
+
key = f"{target}_bin"
|
| 86 |
+
if key not in z:
|
| 87 |
+
raise ValueError(f"Unknown target '{target}' for dataset '{dataset}'. "
|
| 88 |
+
f"Available: {[k.replace('_bin', '') for k in z.files if k.endswith('_bin')]}")
|
| 89 |
+
return z[key].astype(np.int64)
|
| 90 |
+
|
| 91 |
+
|
| 92 |
+
# ── Training loop ──
|
| 93 |
+
|
| 94 |
+
def train_one(dataset, backbone, target, seed,
|
| 95 |
+
n_epochs=150, verbose=False):
|
| 96 |
+
"""
|
| 97 |
+
Returns dict: task_acc, posdis, elapsed_s, seed, dataset, backbone, target.
|
| 98 |
+
"""
|
| 99 |
+
t0 = time.time()
|
| 100 |
+
feat = load_features(dataset, backbone) # [N, T, D]
|
| 101 |
+
labels = load_labels(dataset, target) # [N]
|
| 102 |
+
|
| 103 |
+
N, nf, dim = feat.shape
|
| 104 |
+
fpa = max(1, nf // N_AGENTS)
|
| 105 |
+
agent_views = [feat[:, (i * fpa):(i + 1) * fpa, :] for i in range(N_AGENTS)]
|
| 106 |
+
|
| 107 |
+
torch.manual_seed(seed)
|
| 108 |
+
np.random.seed(seed)
|
| 109 |
+
|
| 110 |
+
# 80/20 train / holdout split (stratified by target class)
|
| 111 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 112 |
+
train_ids = []
|
| 113 |
+
holdout_ids = []
|
| 114 |
+
for c in np.unique(labels):
|
| 115 |
+
ids_c = np.where(labels == c)[0]
|
| 116 |
+
rng.shuffle(ids_c)
|
| 117 |
+
split = max(1, len(ids_c) // 5)
|
| 118 |
+
holdout_ids.extend(ids_c[:split])
|
| 119 |
+
train_ids.extend(ids_c[split:])
|
| 120 |
+
train_ids = np.array(train_ids)
|
| 121 |
+
holdout_ids = np.array(holdout_ids)
|
| 122 |
+
|
| 123 |
+
n_classes = int(labels.max()) + 1
|
| 124 |
+
chance = 1.0 / n_classes
|
| 125 |
+
|
| 126 |
+
senders = [DiscreteSender(TemporalEncoder(HIDDEN_DIM, dim, fpa),
|
| 127 |
+
HIDDEN_DIM, VOCAB_SIZE, N_HEADS)
|
| 128 |
+
for _ in range(N_AGENTS)]
|
| 129 |
+
sender = DiscreteMultiSender(senders).to(DEVICE)
|
| 130 |
+
receivers = [ClassifierReceiver(MSG_DIM, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 131 |
+
for _ in range(3)]
|
| 132 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 133 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 134 |
+
|
| 135 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 136 |
+
me = math.log(VOCAB_SIZE)
|
| 137 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 138 |
+
|
| 139 |
+
best_acc, best_state, best_ep = 0.0, None, 0
|
| 140 |
+
|
| 141 |
+
for ep in range(n_epochs):
|
| 142 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05:
|
| 143 |
+
break
|
| 144 |
+
if ep > 0 and ep % 40 == 0:
|
| 145 |
+
for i in range(len(receivers)):
|
| 146 |
+
receivers[i] = ClassifierReceiver(MSG_DIM, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 147 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 148 |
+
|
| 149 |
+
sender.train(); [r.train() for r in receivers]
|
| 150 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1)
|
| 151 |
+
hard = ep >= 30
|
| 152 |
+
|
| 153 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 154 |
+
perm = rng_ep.permutation(train_ids)
|
| 155 |
+
|
| 156 |
+
for b in range(n_batches):
|
| 157 |
+
batch_ids = perm[b * BATCH_SIZE:(b + 1) * BATCH_SIZE]
|
| 158 |
+
if len(batch_ids) < 4:
|
| 159 |
+
continue
|
| 160 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 161 |
+
target_batch = labels_dev[batch_ids]
|
| 162 |
+
|
| 163 |
+
msg, logits_list = sender(views, tau=tau, hard=hard)
|
| 164 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 165 |
+
for r in receivers:
|
| 166 |
+
pred = r(msg)
|
| 167 |
+
loss = loss + F.cross_entropy(pred, target_batch)
|
| 168 |
+
loss = loss / len(receivers)
|
| 169 |
+
|
| 170 |
+
# Entropy regularisation on each position
|
| 171 |
+
for lg in logits_list:
|
| 172 |
+
lp = F.log_softmax(lg, -1)
|
| 173 |
+
p = lp.exp().clamp(min=1e-8)
|
| 174 |
+
ent = -(p * lp).sum(-1).mean()
|
| 175 |
+
if ent / me < 0.1:
|
| 176 |
+
loss = loss - 0.03 * ent
|
| 177 |
+
|
| 178 |
+
if torch.isnan(loss):
|
| 179 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 180 |
+
continue
|
| 181 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 182 |
+
loss.backward()
|
| 183 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 184 |
+
so.step(); [o.step() for o in ros]
|
| 185 |
+
|
| 186 |
+
if ep % 50 == 0 and DEVICE.type == "mps":
|
| 187 |
+
torch.mps.empty_cache()
|
| 188 |
+
|
| 189 |
+
# Evaluation every 10 epochs
|
| 190 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 191 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 192 |
+
with torch.no_grad():
|
| 193 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 194 |
+
msg_ho, _ = sender(v_ho)
|
| 195 |
+
target_ho = labels_dev[holdout_ids]
|
| 196 |
+
best_per_recv = 0.0
|
| 197 |
+
for r in receivers:
|
| 198 |
+
preds = r(msg_ho).argmax(-1)
|
| 199 |
+
acc = (preds == target_ho).float().mean().item()
|
| 200 |
+
best_per_recv = max(best_per_recv, acc)
|
| 201 |
+
if verbose and ep % 50 == 0:
|
| 202 |
+
print(f" ep={ep} holdout_acc={best_per_recv:.1%}",
|
| 203 |
+
flush=True)
|
| 204 |
+
if best_per_recv > best_acc:
|
| 205 |
+
best_acc = best_per_recv
|
| 206 |
+
best_ep = ep
|
| 207 |
+
best_state = {k: v.cpu().clone()
|
| 208 |
+
for k, v in sender.state_dict().items()}
|
| 209 |
+
|
| 210 |
+
if best_state:
|
| 211 |
+
sender.load_state_dict(best_state)
|
| 212 |
+
sender.eval()
|
| 213 |
+
|
| 214 |
+
# Extract tokens at best state for posdis computation
|
| 215 |
+
with torch.no_grad():
|
| 216 |
+
toks_list = []
|
| 217 |
+
for i in range(0, N, BATCH_SIZE):
|
| 218 |
+
vs = [v[i:i + BATCH_SIZE].to(DEVICE) for v in agent_views]
|
| 219 |
+
_, logits = sender(vs)
|
| 220 |
+
toks_list.append(np.stack([l.argmax(-1).cpu().numpy() for l in logits], 1))
|
| 221 |
+
tokens = np.concatenate(toks_list, 0)
|
| 222 |
+
|
| 223 |
+
# Simple PosDis (target only, MI per position)
|
| 224 |
+
try:
|
| 225 |
+
from _killer_experiment import positional_disentanglement
|
| 226 |
+
attrs = np.stack([labels, labels], axis=1)
|
| 227 |
+
posdis, _, _ = positional_disentanglement(tokens, attrs, VOCAB_SIZE)
|
| 228 |
+
except Exception as e:
|
| 229 |
+
posdis = 0.0
|
| 230 |
+
|
| 231 |
+
return {
|
| 232 |
+
"dataset": dataset,
|
| 233 |
+
"backbone": backbone,
|
| 234 |
+
"target": target,
|
| 235 |
+
"seed": int(seed),
|
| 236 |
+
"n_classes": int(n_classes),
|
| 237 |
+
"chance": float(chance),
|
| 238 |
+
"task_acc": float(best_acc),
|
| 239 |
+
"posdis": float(posdis),
|
| 240 |
+
"elapsed_s": float(time.time() - t0),
|
| 241 |
+
"best_ep": int(best_ep),
|
| 242 |
+
"n_train": int(len(train_ids)),
|
| 243 |
+
"n_holdout": int(len(holdout_ids)),
|
| 244 |
+
}
|
| 245 |
+
|
| 246 |
+
|
| 247 |
+
if __name__ == "__main__":
|
| 248 |
+
ap = argparse.ArgumentParser()
|
| 249 |
+
ap.add_argument("--dataset", required=True, choices=["collision", "ramp"])
|
| 250 |
+
ap.add_argument("--backbone", required=True, choices=["vjepa2", "dinov2"])
|
| 251 |
+
ap.add_argument("--target", required=True)
|
| 252 |
+
ap.add_argument("--seed", type=int, default=0)
|
| 253 |
+
ap.add_argument("--epochs", type=int, default=150)
|
| 254 |
+
ap.add_argument("--verbose", action="store_true")
|
| 255 |
+
args = ap.parse_args()
|
| 256 |
+
r = train_one(args.dataset, args.backbone, args.target, args.seed,
|
| 257 |
+
n_epochs=args.epochs, verbose=args.verbose)
|
| 258 |
+
print(json.dumps(r, indent=2), flush=True)
|
code/_overnight_p1_transfer.py
ADDED
|
@@ -0,0 +1,503 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
PRIORITY 1: Cross-scenario transfer test.
|
| 3 |
+
|
| 4 |
+
All runs use fpa=1 (each of 4 agents sees 1 frame), 4 evenly-spaced frames
|
| 5 |
+
per scene from each dataset. This makes collision (24 frames) and ramp
|
| 6 |
+
(16 or 8 frames) architecture-compatible.
|
| 7 |
+
|
| 8 |
+
Protocol:
|
| 9 |
+
- Base training: train sender + receiver on (dataset_src, target).
|
| 10 |
+
- Zero-shot transfer: apply source-trained receiver directly to
|
| 11 |
+
dataset_tgt codes. No retraining.
|
| 12 |
+
- 16-shot calibration: freeze source sender, train new receiver on
|
| 13 |
+
16 stratified examples from dataset_tgt, evaluate on dataset_tgt holdout.
|
| 14 |
+
- Cross-property: freeze source sender, train new receiver on full
|
| 15 |
+
dataset_src train with a different target.
|
| 16 |
+
|
| 17 |
+
Writes: results/cross_scenario_transfer/
|
| 18 |
+
"""
|
| 19 |
+
import json, time, sys, os, math, copy
|
| 20 |
+
from pathlib import Path
|
| 21 |
+
from datetime import datetime, timezone
|
| 22 |
+
import numpy as np
|
| 23 |
+
import torch
|
| 24 |
+
import torch.nn.functional as F
|
| 25 |
+
|
| 26 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 27 |
+
from _kinematics_train import (
|
| 28 |
+
load_labels, ClassifierReceiver,
|
| 29 |
+
HIDDEN_DIM, VOCAB_SIZE, N_HEADS, N_AGENTS, MSG_DIM, BATCH_SIZE,
|
| 30 |
+
SENDER_LR, RECEIVER_LR, EARLY_STOP_PATIENCE, DEVICE,
|
| 31 |
+
)
|
| 32 |
+
from _killer_experiment import (
|
| 33 |
+
TemporalEncoder, DiscreteSender, DiscreteMultiSender,
|
| 34 |
+
)
|
| 35 |
+
|
| 36 |
+
OUT = Path("results/cross_scenario_transfer")
|
| 37 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 38 |
+
LOG = Path("results/overnight_log.txt")
|
| 39 |
+
|
| 40 |
+
FEATURE_FILES = {
|
| 41 |
+
("collision", "vjepa2"): "results/vjepa2_collision_pooled.pt",
|
| 42 |
+
("collision", "dinov2"): "results/collision_dinov2_features.pt",
|
| 43 |
+
("ramp", "vjepa2"): "results/vjepa2_ramp_temporal.pt",
|
| 44 |
+
("ramp", "dinov2"): "results/phase54b_dino_features.pt",
|
| 45 |
+
("collision", "clip"): "results/kinematics_vs_mechanics/clip_collision_features.pt",
|
| 46 |
+
("ramp", "clip"): "results/kinematics_vs_mechanics/clip_ramp_features.pt",
|
| 47 |
+
}
|
| 48 |
+
N_EPOCHS = 150
|
| 49 |
+
N_EPOCHS_RECEIVER_ONLY = 100
|
| 50 |
+
N_SEEDS = 2
|
| 51 |
+
N_FRAMES_SUBSAMPLE = 4 # fpa=1 × N_AGENTS=4 → 4 frames per scene
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
def log(msg):
|
| 55 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 56 |
+
line = f"[{ts}] P1-transfer: {msg}"
|
| 57 |
+
print(line, flush=True)
|
| 58 |
+
LOG.parent.mkdir(parents=True, exist_ok=True)
|
| 59 |
+
with open(LOG, "a") as f: f.write(line + "\n")
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
def load_and_subsample(dataset, backbone):
|
| 63 |
+
path = FEATURE_FILES[(dataset, backbone)]
|
| 64 |
+
d = torch.load(path, weights_only=False, map_location="cpu")
|
| 65 |
+
feat = d["features"].float() # (N, T_full, D)
|
| 66 |
+
T_full = feat.shape[1]
|
| 67 |
+
if T_full < N_FRAMES_SUBSAMPLE:
|
| 68 |
+
# Pad by repeating last frame
|
| 69 |
+
pad = feat[:, -1:, :].repeat(1, N_FRAMES_SUBSAMPLE - T_full, 1)
|
| 70 |
+
feat = torch.cat([feat, pad], dim=1)
|
| 71 |
+
idx = list(range(T_full)) + [T_full - 1] * (N_FRAMES_SUBSAMPLE - T_full)
|
| 72 |
+
else:
|
| 73 |
+
idx = np.linspace(0, T_full - 1, N_FRAMES_SUBSAMPLE).astype(int).tolist()
|
| 74 |
+
feat = feat[:, idx, :].contiguous()
|
| 75 |
+
return feat, idx
|
| 76 |
+
|
| 77 |
+
|
| 78 |
+
def extract_clip_ramp():
|
| 79 |
+
"""Extract CLIP features for 300 ramp scenes: 4 evenly-spaced frames directly."""
|
| 80 |
+
import timm
|
| 81 |
+
from torchvision import transforms
|
| 82 |
+
from PIL import Image
|
| 83 |
+
|
| 84 |
+
out_path = Path("results/kinematics_vs_mechanics/clip_ramp_features.pt")
|
| 85 |
+
if out_path.exists():
|
| 86 |
+
log("CLIP ramp features already cached")
|
| 87 |
+
return
|
| 88 |
+
|
| 89 |
+
log("Extracting CLIP ramp features (300 scenes × 24 frames at stride 1)...")
|
| 90 |
+
model = timm.create_model("vit_large_patch14_clip_224.openai",
|
| 91 |
+
pretrained=True, num_classes=0).to(DEVICE).eval()
|
| 92 |
+
tfm = transforms.Compose([
|
| 93 |
+
transforms.Resize(224), transforms.CenterCrop(224),
|
| 94 |
+
transforms.ToTensor(),
|
| 95 |
+
transforms.Normalize(mean=[0.48145466, 0.4578275, 0.40821073],
|
| 96 |
+
std=[0.26862954, 0.26130258, 0.27577711]),
|
| 97 |
+
])
|
| 98 |
+
DATASET = Path("kubric/output/ramp_dataset")
|
| 99 |
+
n_scenes = 300
|
| 100 |
+
sample = sorted((DATASET / "scene_0000").glob("rgba_*.png"))
|
| 101 |
+
total = len(sample)
|
| 102 |
+
step = max(1, total // 24)
|
| 103 |
+
frame_indices = list(range(0, total, step))[:24]
|
| 104 |
+
|
| 105 |
+
feat_out = torch.zeros(n_scenes, 24, 1024, dtype=torch.float32)
|
| 106 |
+
t0 = time.time()
|
| 107 |
+
for si in range(n_scenes):
|
| 108 |
+
sd = DATASET / f"scene_{si:04d}"
|
| 109 |
+
imgs = [tfm(Image.open(sd / f"rgba_{fi:05d}.png").convert("RGB"))
|
| 110 |
+
for fi in frame_indices]
|
| 111 |
+
batch = torch.stack(imgs, 0).to(DEVICE)
|
| 112 |
+
with torch.no_grad():
|
| 113 |
+
feat_out[si] = model(batch).cpu().float()
|
| 114 |
+
if (si + 1) % 100 == 0:
|
| 115 |
+
log(f" clip-ramp [{si+1}/{n_scenes}] rate={(si+1)/(time.time()-t0):.1f}/s")
|
| 116 |
+
if DEVICE.type == "mps": torch.mps.empty_cache()
|
| 117 |
+
torch.save({"features": feat_out, "frame_indices": frame_indices,
|
| 118 |
+
"model": "vit_large_patch14_clip_224.openai"}, out_path)
|
| 119 |
+
log(f"CLIP ramp done in {time.time()-t0:.0f}s")
|
| 120 |
+
|
| 121 |
+
|
| 122 |
+
def build_sender(feat_dim, fpa):
|
| 123 |
+
senders = [DiscreteSender(TemporalEncoder(HIDDEN_DIM, feat_dim, fpa),
|
| 124 |
+
HIDDEN_DIM, VOCAB_SIZE, N_HEADS)
|
| 125 |
+
for _ in range(N_AGENTS)]
|
| 126 |
+
return DiscreteMultiSender(senders).to(DEVICE)
|
| 127 |
+
|
| 128 |
+
|
| 129 |
+
def train_base(feat, labels, seed, n_epochs=N_EPOCHS):
|
| 130 |
+
"""Train fresh sender+receiver on (feat, labels). Return (sender_state, receiver_state, train_ids, holdout_ids, best_acc)."""
|
| 131 |
+
N, nf, dim = feat.shape
|
| 132 |
+
fpa = 1
|
| 133 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 134 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 135 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 136 |
+
train_ids, holdout_ids = [], []
|
| 137 |
+
for c in np.unique(labels):
|
| 138 |
+
ids_c = np.where(labels == c)[0]
|
| 139 |
+
rng.shuffle(ids_c)
|
| 140 |
+
split = max(1, len(ids_c) // 5)
|
| 141 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 142 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 143 |
+
n_classes = int(labels.max()) + 1
|
| 144 |
+
chance = 1.0 / n_classes
|
| 145 |
+
|
| 146 |
+
sender = build_sender(dim, fpa)
|
| 147 |
+
receivers = [ClassifierReceiver(MSG_DIM, HIDDEN_DIM, n_classes).to(DEVICE) for _ in range(3)]
|
| 148 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 149 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 150 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 151 |
+
me = math.log(VOCAB_SIZE)
|
| 152 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 153 |
+
best_acc, best_ep = 0.0, 0
|
| 154 |
+
best_sender_state, best_receiver_states = None, None
|
| 155 |
+
|
| 156 |
+
for ep in range(n_epochs):
|
| 157 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break
|
| 158 |
+
if ep > 0 and ep % 40 == 0:
|
| 159 |
+
for i in range(len(receivers)):
|
| 160 |
+
receivers[i] = ClassifierReceiver(MSG_DIM, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 161 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 162 |
+
sender.train(); [r.train() for r in receivers]
|
| 163 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1)
|
| 164 |
+
hard = ep >= 30
|
| 165 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 166 |
+
perm = rng_ep.permutation(train_ids)
|
| 167 |
+
for b in range(n_batches):
|
| 168 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 169 |
+
if len(batch_ids) < 4: continue
|
| 170 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 171 |
+
tgt = labels_dev[batch_ids]
|
| 172 |
+
msg, logits_list = sender(views, tau=tau, hard=hard)
|
| 173 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 174 |
+
for r in receivers: loss = loss + F.cross_entropy(r(msg), tgt)
|
| 175 |
+
loss = loss / len(receivers)
|
| 176 |
+
for lg in logits_list:
|
| 177 |
+
lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8)
|
| 178 |
+
ent = -(p * lp).sum(-1).mean()
|
| 179 |
+
if ent / me < 0.1: loss = loss - 0.03 * ent
|
| 180 |
+
if torch.isnan(loss):
|
| 181 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 182 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 183 |
+
loss.backward()
|
| 184 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 185 |
+
so.step(); [o.step() for o in ros]
|
| 186 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 187 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 188 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 189 |
+
with torch.no_grad():
|
| 190 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 191 |
+
msg_ho, _ = sender(v_ho)
|
| 192 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 193 |
+
best_per_recv, best_recv_idx = 0.0, 0
|
| 194 |
+
for ri, r in enumerate(receivers):
|
| 195 |
+
preds = r(msg_ho).argmax(-1)
|
| 196 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 197 |
+
if acc > best_per_recv:
|
| 198 |
+
best_per_recv, best_recv_idx = acc, ri
|
| 199 |
+
if best_per_recv > best_acc:
|
| 200 |
+
best_acc, best_ep = best_per_recv, ep
|
| 201 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 202 |
+
best_receiver_states = [{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 203 |
+
for r in receivers]
|
| 204 |
+
best_recv_idx_saved = best_recv_idx
|
| 205 |
+
|
| 206 |
+
return {
|
| 207 |
+
"sender_state": best_sender_state,
|
| 208 |
+
"receiver_states": best_receiver_states,
|
| 209 |
+
"best_recv_idx": best_recv_idx_saved if best_receiver_states else 0,
|
| 210 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 211 |
+
"task_acc": best_acc, "chance": chance,
|
| 212 |
+
"n_classes": n_classes, "fpa": 1, "dim": dim,
|
| 213 |
+
}
|
| 214 |
+
|
| 215 |
+
|
| 216 |
+
def eval_zero_shot(base, feat_tgt, labels_tgt, holdout_ids_tgt):
|
| 217 |
+
"""Apply base sender + base receiver directly to target data."""
|
| 218 |
+
N, nf, dim = feat_tgt.shape
|
| 219 |
+
assert dim == base["dim"], f"dim mismatch {dim} vs {base['dim']}"
|
| 220 |
+
sender = build_sender(dim, base["fpa"])
|
| 221 |
+
sender.load_state_dict(base["sender_state"])
|
| 222 |
+
sender.eval().to(DEVICE)
|
| 223 |
+
receivers = [ClassifierReceiver(MSG_DIM, HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 224 |
+
for _ in range(len(base["receiver_states"]))]
|
| 225 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 226 |
+
[r.eval() for r in receivers]
|
| 227 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 228 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 229 |
+
with torch.no_grad():
|
| 230 |
+
v_ho = [v[holdout_ids_tgt].to(DEVICE) for v in agent_views]
|
| 231 |
+
msg_ho, _ = sender(v_ho)
|
| 232 |
+
tgt_ho = labels_dev[holdout_ids_tgt]
|
| 233 |
+
best = 0.0
|
| 234 |
+
for r in receivers:
|
| 235 |
+
preds = r(msg_ho).argmax(-1)
|
| 236 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 237 |
+
best = max(best, acc)
|
| 238 |
+
return best
|
| 239 |
+
|
| 240 |
+
|
| 241 |
+
def train_receiver_frozen_sender(base, feat_tgt, labels_tgt, train_ids_tgt,
|
| 242 |
+
holdout_ids_tgt, seed, n_epochs=N_EPOCHS_RECEIVER_ONLY,
|
| 243 |
+
max_examples=None):
|
| 244 |
+
"""Freeze base sender. Train NEW receiver on (subset of) train_ids_tgt."""
|
| 245 |
+
N, nf, dim = feat_tgt.shape
|
| 246 |
+
assert dim == base["dim"]
|
| 247 |
+
if max_examples is not None and len(train_ids_tgt) > max_examples:
|
| 248 |
+
rng = np.random.RandomState(seed * 311 + 7)
|
| 249 |
+
picks = []
|
| 250 |
+
per_class = max(1, max_examples // base["n_classes"])
|
| 251 |
+
for c in range(base["n_classes"]):
|
| 252 |
+
ids_c = np.array([i for i in train_ids_tgt if labels_tgt[i] == c])
|
| 253 |
+
if len(ids_c) == 0: continue
|
| 254 |
+
rng.shuffle(ids_c)
|
| 255 |
+
picks.extend(ids_c[:per_class])
|
| 256 |
+
train_ids_tgt = np.array(picks)
|
| 257 |
+
|
| 258 |
+
sender = build_sender(dim, base["fpa"])
|
| 259 |
+
sender.load_state_dict(base["sender_state"])
|
| 260 |
+
sender.to(DEVICE).eval()
|
| 261 |
+
for p in sender.parameters(): p.requires_grad = False
|
| 262 |
+
|
| 263 |
+
receivers = [ClassifierReceiver(MSG_DIM, HIDDEN_DIM, base["n_classes"]).to(DEVICE) for _ in range(3)]
|
| 264 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 265 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 266 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 267 |
+
n_batches = max(1, len(train_ids_tgt) // min(BATCH_SIZE, len(train_ids_tgt)))
|
| 268 |
+
best_acc, best_ep = 0.0, 0
|
| 269 |
+
for ep in range(n_epochs):
|
| 270 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > base["chance"] + 0.05: break
|
| 271 |
+
[r.train() for r in receivers]
|
| 272 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 273 |
+
perm = rng_ep.permutation(train_ids_tgt)
|
| 274 |
+
bs = min(BATCH_SIZE, len(train_ids_tgt))
|
| 275 |
+
for b in range(max(1, len(train_ids_tgt) // bs)):
|
| 276 |
+
batch_ids = perm[b*bs:(b+1)*bs]
|
| 277 |
+
if len(batch_ids) < 2: continue
|
| 278 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 279 |
+
with torch.no_grad():
|
| 280 |
+
msg, _ = sender(views)
|
| 281 |
+
for r, o in zip(receivers, ros):
|
| 282 |
+
pred = r(msg)
|
| 283 |
+
loss = F.cross_entropy(pred, labels_dev[batch_ids])
|
| 284 |
+
if torch.isnan(loss): continue
|
| 285 |
+
o.zero_grad(); loss.backward(); o.step()
|
| 286 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 287 |
+
[r.eval() for r in receivers]
|
| 288 |
+
with torch.no_grad():
|
| 289 |
+
v_ho = [v[holdout_ids_tgt].to(DEVICE) for v in agent_views]
|
| 290 |
+
msg_ho, _ = sender(v_ho)
|
| 291 |
+
tgt_ho = labels_dev[holdout_ids_tgt]
|
| 292 |
+
best = 0.0
|
| 293 |
+
for r in receivers:
|
| 294 |
+
preds = r(msg_ho).argmax(-1)
|
| 295 |
+
best = max(best, (preds == tgt_ho).float().mean().item())
|
| 296 |
+
if best > best_acc: best_acc, best_ep = best, ep
|
| 297 |
+
return best_acc
|
| 298 |
+
|
| 299 |
+
|
| 300 |
+
def make_splits(labels, seed):
|
| 301 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 302 |
+
train_ids, holdout_ids = [], []
|
| 303 |
+
for c in np.unique(labels):
|
| 304 |
+
ids_c = np.where(labels == c)[0]
|
| 305 |
+
rng.shuffle(ids_c)
|
| 306 |
+
split = max(1, len(ids_c) // 5)
|
| 307 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 308 |
+
return np.array(train_ids), np.array(holdout_ids)
|
| 309 |
+
|
| 310 |
+
|
| 311 |
+
# ── Main ──
|
| 312 |
+
|
| 313 |
+
def main():
|
| 314 |
+
t_start = time.time()
|
| 315 |
+
log(f"=== OVERNIGHT PRIORITY 1: Cross-Scenario Transfer ===")
|
| 316 |
+
|
| 317 |
+
# Extract CLIP ramp if needed
|
| 318 |
+
extract_clip_ramp()
|
| 319 |
+
|
| 320 |
+
# Load all features once
|
| 321 |
+
log("Loading features...")
|
| 322 |
+
feats = {}
|
| 323 |
+
for (ds, bb), path in FEATURE_FILES.items():
|
| 324 |
+
if Path(path).exists():
|
| 325 |
+
f, idx = load_and_subsample(ds, bb)
|
| 326 |
+
feats[(ds, bb)] = f
|
| 327 |
+
log(f" {ds}/{bb}: {tuple(f.shape)} sampled from T={torch.load(path, weights_only=False, map_location='cpu')['features'].shape[1]}")
|
| 328 |
+
|
| 329 |
+
labels_col_restit = load_labels("collision", "restitution")
|
| 330 |
+
labels_col_mass = load_labels("collision", "mass")
|
| 331 |
+
labels_ramp_restit = load_labels("ramp", "restitution")
|
| 332 |
+
|
| 333 |
+
all_results = []
|
| 334 |
+
records = [] # for table
|
| 335 |
+
|
| 336 |
+
# ── A. Within-scenario sanity (V-JEPA only) ──
|
| 337 |
+
log("\n--- A. Within-scenario sanity ---")
|
| 338 |
+
for seed in range(N_SEEDS):
|
| 339 |
+
log(f" within collision-restit V-JEPA seed={seed}")
|
| 340 |
+
t0 = time.time()
|
| 341 |
+
r = train_base(feats[("collision", "vjepa2")], labels_col_restit, seed)
|
| 342 |
+
dt = time.time() - t0
|
| 343 |
+
log(f" acc={r['task_acc']:.3f} [{dt:.0f}s]")
|
| 344 |
+
records.append({"row": "within_collision_vjepa", "bb": "vjepa2",
|
| 345 |
+
"seed": seed, "acc": r["task_acc"], "elapsed_s": dt})
|
| 346 |
+
all_results.append({"condition": "within_collision", "backbone": "vjepa2",
|
| 347 |
+
"seed": seed, "acc": r["task_acc"], "elapsed_s": dt})
|
| 348 |
+
|
| 349 |
+
for seed in range(N_SEEDS):
|
| 350 |
+
log(f" within ramp-restit V-JEPA seed={seed}")
|
| 351 |
+
t0 = time.time()
|
| 352 |
+
r = train_base(feats[("ramp", "vjepa2")], labels_ramp_restit, seed)
|
| 353 |
+
dt = time.time() - t0
|
| 354 |
+
log(f" acc={r['task_acc']:.3f} [{dt:.0f}s]")
|
| 355 |
+
records.append({"row": "within_ramp_vjepa", "bb": "vjepa2",
|
| 356 |
+
"seed": seed, "acc": r["task_acc"], "elapsed_s": dt})
|
| 357 |
+
all_results.append({"condition": "within_ramp", "backbone": "vjepa2",
|
| 358 |
+
"seed": seed, "acc": r["task_acc"], "elapsed_s": dt})
|
| 359 |
+
|
| 360 |
+
# ── Cache base senders needed for transfer ──
|
| 361 |
+
# col_restit for each backbone × seed
|
| 362 |
+
# ramp_restit for each backbone × seed
|
| 363 |
+
# col_mass for V-JEPA × seed (for cross-property)
|
| 364 |
+
log("\n--- Training base senders for transfer ---")
|
| 365 |
+
bases = {} # (bb, src_ds, target, seed) -> base dict
|
| 366 |
+
for bb in ["vjepa2", "dinov2", "clip"]:
|
| 367 |
+
if ("collision", bb) not in feats or ("ramp", bb) not in feats: continue
|
| 368 |
+
for seed in range(N_SEEDS):
|
| 369 |
+
log(f" base {bb} collision-restit seed={seed}")
|
| 370 |
+
t0 = time.time()
|
| 371 |
+
bases[(bb, "collision", "restitution", seed)] = train_base(
|
| 372 |
+
feats[("collision", bb)], labels_col_restit, seed)
|
| 373 |
+
log(f" acc={bases[(bb, 'collision', 'restitution', seed)]['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 374 |
+
log(f" base {bb} ramp-restit seed={seed}")
|
| 375 |
+
t0 = time.time()
|
| 376 |
+
bases[(bb, "ramp", "restitution", seed)] = train_base(
|
| 377 |
+
feats[("ramp", bb)], labels_ramp_restit, seed)
|
| 378 |
+
log(f" acc={bases[(bb, 'ramp', 'restitution', seed)]['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 379 |
+
# V-JEPA collision-mass for cross-property
|
| 380 |
+
for seed in range(N_SEEDS):
|
| 381 |
+
log(f" base vjepa2 collision-mass seed={seed}")
|
| 382 |
+
t0 = time.time()
|
| 383 |
+
bases[("vjepa2", "collision", "mass", seed)] = train_base(
|
| 384 |
+
feats[("collision", "vjepa2")], labels_col_mass, seed)
|
| 385 |
+
log(f" acc={bases[('vjepa2', 'collision', 'mass', seed)]['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 386 |
+
|
| 387 |
+
# ── B. Cross-scenario transfer ──
|
| 388 |
+
log("\n--- B. Cross-scenario transfer ---")
|
| 389 |
+
for bb in ["vjepa2", "dinov2", "clip"]:
|
| 390 |
+
if (bb, "collision", "restitution", 0) not in bases: continue
|
| 391 |
+
for direction, src_ds, tgt_ds, tgt_labels in [
|
| 392 |
+
("col_to_ramp", "collision", "ramp", labels_ramp_restit),
|
| 393 |
+
("ramp_to_col", "ramp", "collision", labels_col_restit),
|
| 394 |
+
]:
|
| 395 |
+
for seed in range(N_SEEDS):
|
| 396 |
+
base = bases[(bb, src_ds, "restitution", seed)]
|
| 397 |
+
# Splits on the TARGET dataset
|
| 398 |
+
train_ids_tgt, holdout_ids_tgt = make_splits(tgt_labels, seed)
|
| 399 |
+
# Zero-shot
|
| 400 |
+
t0 = time.time()
|
| 401 |
+
acc_zs = eval_zero_shot(base, feats[(tgt_ds, bb)], tgt_labels, holdout_ids_tgt)
|
| 402 |
+
dt_zs = time.time() - t0
|
| 403 |
+
log(f" {bb} {direction} zero-shot seed={seed}: acc={acc_zs:.3f} [{dt_zs:.1f}s]")
|
| 404 |
+
records.append({"row": f"{direction}_zero_shot", "bb": bb, "seed": seed,
|
| 405 |
+
"acc": acc_zs, "elapsed_s": dt_zs})
|
| 406 |
+
all_results.append({"condition": f"{direction}_zero_shot", "backbone": bb,
|
| 407 |
+
"seed": seed, "acc": acc_zs, "elapsed_s": dt_zs})
|
| 408 |
+
# 16-shot
|
| 409 |
+
t0 = time.time()
|
| 410 |
+
acc_16 = train_receiver_frozen_sender(
|
| 411 |
+
base, feats[(tgt_ds, bb)], tgt_labels,
|
| 412 |
+
train_ids_tgt, holdout_ids_tgt, seed, max_examples=16)
|
| 413 |
+
dt_16 = time.time() - t0
|
| 414 |
+
log(f" {bb} {direction} 16-shot seed={seed}: acc={acc_16:.3f} [{dt_16:.0f}s]")
|
| 415 |
+
records.append({"row": f"{direction}_16shot", "bb": bb, "seed": seed,
|
| 416 |
+
"acc": acc_16, "elapsed_s": dt_16})
|
| 417 |
+
all_results.append({"condition": f"{direction}_16shot", "backbone": bb,
|
| 418 |
+
"seed": seed, "acc": acc_16, "elapsed_s": dt_16})
|
| 419 |
+
|
| 420 |
+
# ── C. Cross-property controls (V-JEPA, within collision) ──
|
| 421 |
+
log("\n--- C. Cross-property controls (V-JEPA collision) ---")
|
| 422 |
+
# restit-sender → mass
|
| 423 |
+
for seed in range(N_SEEDS):
|
| 424 |
+
base = bases[("vjepa2", "collision", "restitution", seed)]
|
| 425 |
+
train_ids, holdout_ids = make_splits(labels_col_mass, seed)
|
| 426 |
+
t0 = time.time()
|
| 427 |
+
acc = train_receiver_frozen_sender(
|
| 428 |
+
base, feats[("collision", "vjepa2")], labels_col_mass,
|
| 429 |
+
train_ids, holdout_ids, seed, max_examples=None)
|
| 430 |
+
dt = time.time() - t0
|
| 431 |
+
log(f" V-JEPA restit��mass seed={seed}: acc={acc:.3f} [{dt:.0f}s]")
|
| 432 |
+
records.append({"row": "cross_prop_restit_to_mass", "bb": "vjepa2",
|
| 433 |
+
"seed": seed, "acc": acc, "elapsed_s": dt})
|
| 434 |
+
all_results.append({"condition": "cross_prop_restit_to_mass",
|
| 435 |
+
"backbone": "vjepa2", "seed": seed,
|
| 436 |
+
"acc": acc, "elapsed_s": dt})
|
| 437 |
+
# mass-sender → restit
|
| 438 |
+
for seed in range(N_SEEDS):
|
| 439 |
+
base = bases[("vjepa2", "collision", "mass", seed)]
|
| 440 |
+
train_ids, holdout_ids = make_splits(labels_col_restit, seed)
|
| 441 |
+
t0 = time.time()
|
| 442 |
+
acc = train_receiver_frozen_sender(
|
| 443 |
+
base, feats[("collision", "vjepa2")], labels_col_restit,
|
| 444 |
+
train_ids, holdout_ids, seed, max_examples=None)
|
| 445 |
+
dt = time.time() - t0
|
| 446 |
+
log(f" V-JEPA mass→restit seed={seed}: acc={acc:.3f} [{dt:.0f}s]")
|
| 447 |
+
records.append({"row": "cross_prop_mass_to_restit", "bb": "vjepa2",
|
| 448 |
+
"seed": seed, "acc": acc, "elapsed_s": dt})
|
| 449 |
+
all_results.append({"condition": "cross_prop_mass_to_restit",
|
| 450 |
+
"backbone": "vjepa2", "seed": seed,
|
| 451 |
+
"acc": acc, "elapsed_s": dt})
|
| 452 |
+
|
| 453 |
+
# ── Aggregate + write summary ──
|
| 454 |
+
def agg(cond, bb):
|
| 455 |
+
vals = [r["acc"] for r in all_results
|
| 456 |
+
if r["condition"] == cond and r["backbone"] == bb]
|
| 457 |
+
if not vals: return (float("nan"), float("nan"))
|
| 458 |
+
return (float(np.mean(vals)*100), float(np.std(vals)*100))
|
| 459 |
+
|
| 460 |
+
lines = []
|
| 461 |
+
lines.append("CROSS-SCENARIO RESTITUTION TRANSFER")
|
| 462 |
+
lines.append(f"Config: fpa=1, 4 frames (evenly spaced), K=5, 2 seeds/cell.")
|
| 463 |
+
lines.append("")
|
| 464 |
+
header = "Condition | V-JEPA 2 | DINOv2 | CLIP | Chance"
|
| 465 |
+
lines.append(header)
|
| 466 |
+
lines.append("-" * len(header))
|
| 467 |
+
def row(name, cond, bbs=("vjepa2", "dinov2", "clip")):
|
| 468 |
+
cells = []
|
| 469 |
+
for bb in bbs:
|
| 470 |
+
m, s = agg(cond, bb)
|
| 471 |
+
if np.isnan(m):
|
| 472 |
+
cells.append(" — ")
|
| 473 |
+
else:
|
| 474 |
+
cells.append(f"{m:5.1f}% ± {s:4.1f} ")
|
| 475 |
+
return f"{name:<39s}| {cells[0]}| {cells[1]}| {cells[2]}| 33.3%"
|
| 476 |
+
lines.append(row("Within collision (sanity)", "within_collision", ("vjepa2",)))
|
| 477 |
+
lines.append(row("Within ramp (sanity)", "within_ramp", ("vjepa2",)))
|
| 478 |
+
lines.append(row("Collision→Ramp (zero-shot)", "col_to_ramp_zero_shot"))
|
| 479 |
+
lines.append(row("Ramp→Collision (zero-shot)", "ramp_to_col_zero_shot"))
|
| 480 |
+
lines.append(row("Collision→Ramp (16-shot)", "col_to_ramp_16shot"))
|
| 481 |
+
lines.append(row("Ramp→Collision (16-shot)", "ramp_to_col_16shot"))
|
| 482 |
+
lines.append(row("Cross-property: restit→mass","cross_prop_restit_to_mass", ("vjepa2",)))
|
| 483 |
+
lines.append(row("Cross-property: mass→restit","cross_prop_mass_to_restit", ("vjepa2",)))
|
| 484 |
+
|
| 485 |
+
total_s = time.time() - t_start
|
| 486 |
+
lines.append("")
|
| 487 |
+
lines.append(f"Total runtime: {total_s/60:.1f} min ({total_s:.0f}s)")
|
| 488 |
+
lines.append(f"Runs: {len(all_results)}")
|
| 489 |
+
summary = "\n".join(lines)
|
| 490 |
+
|
| 491 |
+
(OUT / "p1_summary.txt").write_text(summary + "\n")
|
| 492 |
+
with open(OUT / "p1_raw.json", "w") as f:
|
| 493 |
+
# remove state dicts from bases for serialization
|
| 494 |
+
json.dump({"runs": all_results,
|
| 495 |
+
"n_runs": len(all_results),
|
| 496 |
+
"total_runtime_s": total_s},
|
| 497 |
+
f, indent=2, default=str)
|
| 498 |
+
log(f"\n{summary}")
|
| 499 |
+
log(f"Saved: {OUT / 'p1_summary.txt'}")
|
| 500 |
+
|
| 501 |
+
|
| 502 |
+
if __name__ == "__main__":
|
| 503 |
+
main()
|
code/_overnight_p3_matrix.py
ADDED
|
@@ -0,0 +1,259 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
PRIORITY 3: Transfer matrix across all available scenarios.
|
| 3 |
+
|
| 4 |
+
Property-scenario availability:
|
| 5 |
+
restitution: collision, ramp, flat_drop, elasticity, ramp_3prop (5)
|
| 6 |
+
friction: ramp, flat_drop, ramp_3prop (3)
|
| 7 |
+
mass: collision (elasticity has constant mass → skip) (1)
|
| 8 |
+
|
| 9 |
+
For each property with >=2 scenarios: train on one, test on all.
|
| 10 |
+
Subsample all features to [N, 4, D] (fpa=1, 4 agents).
|
| 11 |
+
|
| 12 |
+
Modes:
|
| 13 |
+
zero-shot — apply source-trained receiver to target codes
|
| 14 |
+
16-shot — train new receiver on 16 stratified target examples
|
| 15 |
+
|
| 16 |
+
Backbones: vjepa2, dinov2, clip. Seeds: 2.
|
| 17 |
+
"""
|
| 18 |
+
import json, time, sys, os, math
|
| 19 |
+
from pathlib import Path
|
| 20 |
+
from datetime import datetime, timezone
|
| 21 |
+
import numpy as np
|
| 22 |
+
import torch
|
| 23 |
+
import torch.nn.functional as F
|
| 24 |
+
|
| 25 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 26 |
+
from _kinematics_train import (
|
| 27 |
+
ClassifierReceiver,
|
| 28 |
+
HIDDEN_DIM, VOCAB_SIZE, N_HEADS, N_AGENTS, MSG_DIM, BATCH_SIZE,
|
| 29 |
+
SENDER_LR, RECEIVER_LR, EARLY_STOP_PATIENCE, DEVICE,
|
| 30 |
+
)
|
| 31 |
+
from _killer_experiment import (
|
| 32 |
+
TemporalEncoder, DiscreteSender, DiscreteMultiSender,
|
| 33 |
+
)
|
| 34 |
+
from _overnight_p1_transfer import (
|
| 35 |
+
build_sender, train_base, eval_zero_shot, train_receiver_frozen_sender,
|
| 36 |
+
make_splits, N_FRAMES_SUBSAMPLE,
|
| 37 |
+
)
|
| 38 |
+
|
| 39 |
+
OUT = Path("results/cross_scenario_transfer")
|
| 40 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 41 |
+
LOG = Path("results/overnight_log.txt")
|
| 42 |
+
N_EPOCHS = 150
|
| 43 |
+
N_SEEDS = 2
|
| 44 |
+
|
| 45 |
+
# Feature file locations per (dataset, backbone)
|
| 46 |
+
FEATURE_FILES = {
|
| 47 |
+
("collision", "vjepa2"): "results/vjepa2_collision_pooled.pt",
|
| 48 |
+
("collision", "dinov2"): "results/collision_dinov2_features.pt",
|
| 49 |
+
("collision", "clip"): "results/kinematics_vs_mechanics/clip_collision_features.pt",
|
| 50 |
+
("ramp", "vjepa2"): "results/vjepa2_ramp_temporal.pt",
|
| 51 |
+
("ramp", "dinov2"): "results/phase54b_dino_features.pt",
|
| 52 |
+
("ramp", "clip"): "results/kinematics_vs_mechanics/clip_ramp_features.pt",
|
| 53 |
+
("flat_drop", "vjepa2"): "results/kinematics_vs_mechanics/feat_vjepa2_flat_drop.pt",
|
| 54 |
+
("flat_drop", "dinov2"): "results/kinematics_vs_mechanics/feat_dinov2_flat_drop.pt",
|
| 55 |
+
("flat_drop", "clip"): "results/kinematics_vs_mechanics/feat_clip_flat_drop.pt",
|
| 56 |
+
("elasticity", "vjepa2"): "results/kinematics_vs_mechanics/feat_vjepa2_elasticity.pt",
|
| 57 |
+
("elasticity", "dinov2"): "results/kinematics_vs_mechanics/feat_dinov2_elasticity.pt",
|
| 58 |
+
("elasticity", "clip"): "results/kinematics_vs_mechanics/feat_clip_elasticity.pt",
|
| 59 |
+
("ramp_3prop", "vjepa2"): "results/kinematics_vs_mechanics/feat_vjepa2_ramp_3prop.pt",
|
| 60 |
+
("ramp_3prop", "dinov2"): "results/kinematics_vs_mechanics/feat_dinov2_ramp_3prop.pt",
|
| 61 |
+
("ramp_3prop", "clip"): "results/kinematics_vs_mechanics/feat_clip_ramp_3prop.pt",
|
| 62 |
+
}
|
| 63 |
+
|
| 64 |
+
LABEL_FILES = {
|
| 65 |
+
"collision": "results/kinematics_vs_mechanics/labels_collision.npz",
|
| 66 |
+
"ramp": "results/kinematics_vs_mechanics/labels_ramp.npz",
|
| 67 |
+
"flat_drop": "results/kinematics_vs_mechanics/labels_flat_drop.npz",
|
| 68 |
+
"elasticity": "results/kinematics_vs_mechanics/labels_elasticity.npz",
|
| 69 |
+
"ramp_3prop": "results/kinematics_vs_mechanics/labels_ramp_3prop.npz",
|
| 70 |
+
}
|
| 71 |
+
|
| 72 |
+
# Property availability
|
| 73 |
+
PROPERTY_SCENARIOS = {
|
| 74 |
+
"restitution": ["collision", "ramp", "flat_drop", "elasticity", "ramp_3prop"],
|
| 75 |
+
"friction": ["ramp", "flat_drop", "ramp_3prop"],
|
| 76 |
+
}
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
def log(msg):
|
| 80 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 81 |
+
line = f"[{ts}] P3-matrix: {msg}"
|
| 82 |
+
print(line, flush=True)
|
| 83 |
+
with open(LOG, "a") as f: f.write(line + "\n")
|
| 84 |
+
|
| 85 |
+
|
| 86 |
+
def load_feat_subsampled(dataset, backbone):
|
| 87 |
+
"""Return [N, 4, D]. Subsample evenly or duplicate-pad to 4 temporal positions."""
|
| 88 |
+
path = FEATURE_FILES[(dataset, backbone)]
|
| 89 |
+
d = torch.load(path, weights_only=False, map_location="cpu")
|
| 90 |
+
feat = d["features"].float()
|
| 91 |
+
T = feat.shape[1]
|
| 92 |
+
if T >= N_FRAMES_SUBSAMPLE:
|
| 93 |
+
idx = np.linspace(0, T - 1, N_FRAMES_SUBSAMPLE).astype(int)
|
| 94 |
+
feat = feat[:, idx, :].contiguous()
|
| 95 |
+
else:
|
| 96 |
+
# Duplicate-pad. Repeat each position ceil(4/T) times and slice first 4.
|
| 97 |
+
reps = (N_FRAMES_SUBSAMPLE + T - 1) // T
|
| 98 |
+
feat = feat.repeat(1, reps, 1)[:, :N_FRAMES_SUBSAMPLE, :].contiguous()
|
| 99 |
+
return feat
|
| 100 |
+
|
| 101 |
+
|
| 102 |
+
def load_labels(dataset, target):
|
| 103 |
+
z = np.load(LABEL_FILES[dataset])
|
| 104 |
+
key = f"{target}_bin"
|
| 105 |
+
if key not in z:
|
| 106 |
+
return None
|
| 107 |
+
return z[key].astype(np.int64)
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
def main():
|
| 111 |
+
t0_all = time.time()
|
| 112 |
+
log(f"=== PRIORITY 3: Transfer Matrix ===")
|
| 113 |
+
|
| 114 |
+
# Load + cache all features
|
| 115 |
+
feats = {}
|
| 116 |
+
for key, path in FEATURE_FILES.items():
|
| 117 |
+
if Path(path).exists():
|
| 118 |
+
feats[key] = load_feat_subsampled(*key)
|
| 119 |
+
log(f" {key[0]}/{key[1]}: shape={tuple(feats[key].shape)}")
|
| 120 |
+
else:
|
| 121 |
+
log(f" MISSING: {path}")
|
| 122 |
+
|
| 123 |
+
# Load + validate labels
|
| 124 |
+
labels_cache = {}
|
| 125 |
+
for prop, scenarios in PROPERTY_SCENARIOS.items():
|
| 126 |
+
for ds in scenarios:
|
| 127 |
+
lbl = load_labels(ds, prop)
|
| 128 |
+
if lbl is not None:
|
| 129 |
+
labels_cache[(ds, prop)] = lbl
|
| 130 |
+
log(f" labels {ds}/{prop}: {np.bincount(lbl, minlength=3).tolist()}")
|
| 131 |
+
else:
|
| 132 |
+
log(f" labels {ds}/{prop} MISSING")
|
| 133 |
+
|
| 134 |
+
# Train all base senders (property, ds, bb, seed) once
|
| 135 |
+
log("\n--- Training base senders ---")
|
| 136 |
+
bases = {}
|
| 137 |
+
for prop, scenarios in PROPERTY_SCENARIOS.items():
|
| 138 |
+
for ds in scenarios:
|
| 139 |
+
if (ds, prop) not in labels_cache: continue
|
| 140 |
+
labels = labels_cache[(ds, prop)]
|
| 141 |
+
for bb in ("vjepa2", "dinov2", "clip"):
|
| 142 |
+
if (ds, bb) not in feats: continue
|
| 143 |
+
for seed in range(N_SEEDS):
|
| 144 |
+
t0 = time.time()
|
| 145 |
+
b = train_base(feats[(ds, bb)], labels, seed, n_epochs=N_EPOCHS)
|
| 146 |
+
bases[(prop, ds, bb, seed)] = b
|
| 147 |
+
log(f" {prop}/{ds}/{bb}/seed{seed}: within_acc={b['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 148 |
+
|
| 149 |
+
# Build transfer matrix
|
| 150 |
+
log("\n--- Transfer matrix evaluation ---")
|
| 151 |
+
results = []
|
| 152 |
+
for prop, scenarios in PROPERTY_SCENARIOS.items():
|
| 153 |
+
for src in scenarios:
|
| 154 |
+
if (src, prop) not in labels_cache: continue
|
| 155 |
+
for tgt in scenarios:
|
| 156 |
+
if (tgt, prop) not in labels_cache: continue
|
| 157 |
+
for bb in ("vjepa2", "dinov2", "clip"):
|
| 158 |
+
if (src, bb) not in feats or (tgt, bb) not in feats: continue
|
| 159 |
+
for seed in range(N_SEEDS):
|
| 160 |
+
if (prop, src, bb, seed) not in bases: continue
|
| 161 |
+
base = bases[(prop, src, bb, seed)]
|
| 162 |
+
tgt_labels = labels_cache[(tgt, prop)]
|
| 163 |
+
train_ids_tgt, holdout_ids_tgt = make_splits(tgt_labels, seed)
|
| 164 |
+
|
| 165 |
+
# If src == tgt, report within-acc (no transfer)
|
| 166 |
+
if src == tgt:
|
| 167 |
+
acc_zs = base["task_acc"]
|
| 168 |
+
acc_16 = base["task_acc"]
|
| 169 |
+
else:
|
| 170 |
+
try:
|
| 171 |
+
acc_zs = eval_zero_shot(base, feats[(tgt, bb)],
|
| 172 |
+
tgt_labels, holdout_ids_tgt)
|
| 173 |
+
except Exception as e:
|
| 174 |
+
log(f" ERROR zero-shot {prop}/{src}→{tgt}/{bb}/seed{seed}: {e}")
|
| 175 |
+
acc_zs = float("nan")
|
| 176 |
+
try:
|
| 177 |
+
acc_16 = train_receiver_frozen_sender(
|
| 178 |
+
base, feats[(tgt, bb)], tgt_labels,
|
| 179 |
+
train_ids_tgt, holdout_ids_tgt, seed,
|
| 180 |
+
max_examples=16, n_epochs=80)
|
| 181 |
+
except Exception as e:
|
| 182 |
+
log(f" ERROR 16-shot {prop}/{src}→{tgt}/{bb}/seed{seed}: {e}")
|
| 183 |
+
acc_16 = float("nan")
|
| 184 |
+
|
| 185 |
+
results.append({"property": prop, "src": src, "tgt": tgt,
|
| 186 |
+
"backbone": bb, "seed": seed,
|
| 187 |
+
"zero_shot_acc": float(acc_zs),
|
| 188 |
+
"sixteen_shot_acc": float(acc_16)})
|
| 189 |
+
|
| 190 |
+
# Aggregate matrices
|
| 191 |
+
def matrix(prop, bb, mode_key):
|
| 192 |
+
scens = PROPERTY_SCENARIOS[prop]
|
| 193 |
+
M = np.full((len(scens), len(scens)), np.nan)
|
| 194 |
+
for r in results:
|
| 195 |
+
if r["property"] != prop or r["backbone"] != bb: continue
|
| 196 |
+
i = scens.index(r["src"]); j = scens.index(r["tgt"])
|
| 197 |
+
# Average across seeds
|
| 198 |
+
# (gather all matching, avg)
|
| 199 |
+
# Do it by seed-aggregation properly
|
| 200 |
+
for i, src in enumerate(scens):
|
| 201 |
+
for j, tgt in enumerate(scens):
|
| 202 |
+
vals = [r[mode_key] for r in results
|
| 203 |
+
if r["property"] == prop and r["backbone"] == bb
|
| 204 |
+
and r["src"] == src and r["tgt"] == tgt]
|
| 205 |
+
if vals:
|
| 206 |
+
M[i, j] = np.nanmean(vals)
|
| 207 |
+
return M, scens
|
| 208 |
+
|
| 209 |
+
lines = []
|
| 210 |
+
lines.append(f"PRIORITY 3: PROPERTY TRANSFER MATRIX (2 seeds, 16-shot mode)")
|
| 211 |
+
lines.append(f"Within-scenario cells (diagonal) show within-dataset training accuracy.")
|
| 212 |
+
lines.append("")
|
| 213 |
+
for prop in PROPERTY_SCENARIOS:
|
| 214 |
+
lines.append(f"\n=== {prop.upper()} ({len(PROPERTY_SCENARIOS[prop])} scenarios) ===")
|
| 215 |
+
for bb in ("vjepa2", "dinov2", "clip"):
|
| 216 |
+
M, scens = matrix(prop, bb, "sixteen_shot_acc")
|
| 217 |
+
if np.all(np.isnan(M)): continue
|
| 218 |
+
lines.append(f"\n {bb}:")
|
| 219 |
+
head = " " + "Train\\Test | " + " | ".join(f"{s[:11]:>11s}" for s in scens)
|
| 220 |
+
lines.append(head)
|
| 221 |
+
lines.append(" " + "-" * (len(head) - 2))
|
| 222 |
+
for i, src in enumerate(scens):
|
| 223 |
+
row = f" {src[:15]:<15s} | " + " | ".join(
|
| 224 |
+
f"{M[i,j]*100:>10.1f}%" if not np.isnan(M[i,j]) else f"{'—':>11s}"
|
| 225 |
+
for j in range(len(scens)))
|
| 226 |
+
lines.append(row)
|
| 227 |
+
|
| 228 |
+
# Also zero-shot matrix
|
| 229 |
+
lines.append("\n\nZERO-SHOT MODE (no receiver retraining)")
|
| 230 |
+
for prop in PROPERTY_SCENARIOS:
|
| 231 |
+
lines.append(f"\n=== {prop.upper()} ===")
|
| 232 |
+
for bb in ("vjepa2", "dinov2", "clip"):
|
| 233 |
+
M, scens = matrix(prop, bb, "zero_shot_acc")
|
| 234 |
+
if np.all(np.isnan(M)): continue
|
| 235 |
+
lines.append(f"\n {bb}:")
|
| 236 |
+
head = " " + "Train\\Test | " + " | ".join(f"{s[:11]:>11s}" for s in scens)
|
| 237 |
+
lines.append(head)
|
| 238 |
+
lines.append(" " + "-" * (len(head) - 2))
|
| 239 |
+
for i, src in enumerate(scens):
|
| 240 |
+
row = f" {src[:15]:<15s} | " + " | ".join(
|
| 241 |
+
f"{M[i,j]*100:>10.1f}%" if not np.isnan(M[i,j]) else f"{'—':>11s}"
|
| 242 |
+
for j in range(len(scens)))
|
| 243 |
+
lines.append(row)
|
| 244 |
+
|
| 245 |
+
total_s = time.time() - t0_all
|
| 246 |
+
lines.append(f"\n\nTotal P3 runtime: {total_s/60:.1f} min ({total_s:.0f}s)")
|
| 247 |
+
lines.append(f"N transfer evals: {len(results)}")
|
| 248 |
+
lines.append(f"N base senders trained: {len(bases)}")
|
| 249 |
+
|
| 250 |
+
summary = "\n".join(lines)
|
| 251 |
+
(OUT / "p3_matrix_summary.txt").write_text(summary + "\n")
|
| 252 |
+
with open(OUT / "p3_matrix_raw.json", "w") as f:
|
| 253 |
+
json.dump({"runs": results, "total_runtime_s": total_s}, f, indent=2, default=str)
|
| 254 |
+
log(f"\n{summary}")
|
| 255 |
+
log(f"\nSaved: {OUT / 'p3_matrix_summary.txt'}")
|
| 256 |
+
|
| 257 |
+
|
| 258 |
+
if __name__ == "__main__":
|
| 259 |
+
main()
|
code/_rev_linear_probe_matched_visual.py
ADDED
|
@@ -0,0 +1,269 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP REV-LP-MV: Linear probes on matched-visual conditions (R2 highest-value fix).
|
| 3 |
+
|
| 4 |
+
Adds linear-probe baselines to the gradient figure for:
|
| 5 |
+
1. Velocity interpolation (matched visuals, kinematic split)
|
| 6 |
+
2. Elastic vs inelastic restitution split (matched visuals, dynamics-class split)
|
| 7 |
+
3. Standard-gravity vs low-gravity (matched visuals, dynamics shift)
|
| 8 |
+
|
| 9 |
+
Each is a logistic regression on l2-pooled V-JEPA 2 features at N in {16, 192}, 5 seeds.
|
| 10 |
+
"""
|
| 11 |
+
import json
|
| 12 |
+
import time
|
| 13 |
+
import os
|
| 14 |
+
from pathlib import Path
|
| 15 |
+
from datetime import datetime, timezone
|
| 16 |
+
|
| 17 |
+
import numpy as np
|
| 18 |
+
import torch
|
| 19 |
+
from sklearn.linear_model import LogisticRegression
|
| 20 |
+
from sklearn.preprocessing import StandardScaler
|
| 21 |
+
|
| 22 |
+
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
|
| 23 |
+
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
|
| 24 |
+
T0 = time.time()
|
| 25 |
+
|
| 26 |
+
OUT = Path("results/reviewer_response/exp_lp_matched_visual")
|
| 27 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 28 |
+
N_LIST = [16, 192]
|
| 29 |
+
N_SEEDS = 5
|
| 30 |
+
RNG_BASE = 1234
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
def log(msg):
|
| 34 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 35 |
+
print(f"[{ts}] LP-MV: {msg}", flush=True)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
def pool_l2(features_3d):
|
| 39 |
+
"""L2-pool features along temporal axis: (N, T, D) -> (N, D)."""
|
| 40 |
+
f = features_3d
|
| 41 |
+
if f.ndim == 3:
|
| 42 |
+
return f.mean(dim=1).numpy()
|
| 43 |
+
return f.numpy()
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
def stratified_subset(rng, y, n_per_class):
|
| 47 |
+
"""Indices of n_per_class examples per class."""
|
| 48 |
+
idxs = []
|
| 49 |
+
for c in np.unique(y):
|
| 50 |
+
cand = np.where(y == c)[0]
|
| 51 |
+
if len(cand) == 0:
|
| 52 |
+
continue
|
| 53 |
+
chosen = rng.choice(cand, size=min(n_per_class, len(cand)), replace=False)
|
| 54 |
+
idxs.extend(chosen.tolist())
|
| 55 |
+
return np.array(sorted(idxs))
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
def train_lp(X_tr, y_tr, X_te, y_te):
|
| 59 |
+
sc = StandardScaler().fit(X_tr)
|
| 60 |
+
Xs_tr = sc.transform(X_tr)
|
| 61 |
+
Xs_te = sc.transform(X_te)
|
| 62 |
+
model = LogisticRegression(max_iter=2000, C=1.0, multi_class="auto",
|
| 63 |
+
solver="lbfgs")
|
| 64 |
+
model.fit(Xs_tr, y_tr)
|
| 65 |
+
return float((model.predict(Xs_te) == y_te).mean())
|
| 66 |
+
|
| 67 |
+
|
| 68 |
+
def stats(vals):
|
| 69 |
+
v = np.array(vals)
|
| 70 |
+
return float(v.mean()), float(v.std(ddof=1) if len(v) > 1 else 0.0)
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
def run_split(name, X_src, y_src, X_tgt, y_tgt, n_classes):
|
| 74 |
+
"""Evaluate linear probe at N in N_LIST.
|
| 75 |
+
|
| 76 |
+
Source-train-only baseline: train on full source, evaluate on target (N=0).
|
| 77 |
+
N>0: train on full source + N stratified target examples, evaluate on remaining target.
|
| 78 |
+
"""
|
| 79 |
+
log(f"=== {name}: src={X_src.shape}, tgt={X_tgt.shape}, n_classes={n_classes}")
|
| 80 |
+
results = {"N0_source_only": [], "curve": {N: [] for N in N_LIST}}
|
| 81 |
+
|
| 82 |
+
# N=0: train on source, evaluate on target
|
| 83 |
+
for s in range(N_SEEDS):
|
| 84 |
+
# Subsample source to be fair (use all of it; stratification not needed here)
|
| 85 |
+
acc = train_lp(X_src, y_src, X_tgt, y_tgt)
|
| 86 |
+
results["N0_source_only"].append(acc)
|
| 87 |
+
log(f" N=0 src-only: {stats(results['N0_source_only'])[0]:.3f} ± {stats(results['N0_source_only'])[1]:.3f}")
|
| 88 |
+
|
| 89 |
+
# N=16,192: train on source + N stratified target, eval on remaining target
|
| 90 |
+
smallest = min(int(np.sum(y_tgt == c)) for c in np.unique(y_tgt))
|
| 91 |
+
for N in N_LIST:
|
| 92 |
+
per_class = max(1, N // n_classes)
|
| 93 |
+
# Clamp so we leave at least 30% of each target class for evaluation
|
| 94 |
+
per_class = min(per_class, int(0.7 * smallest))
|
| 95 |
+
for s in range(N_SEEDS):
|
| 96 |
+
rng = np.random.default_rng(RNG_BASE + s)
|
| 97 |
+
tgt_idx_train = stratified_subset(rng, y_tgt, per_class)
|
| 98 |
+
mask = np.ones(len(y_tgt), bool); mask[tgt_idx_train] = False
|
| 99 |
+
X_eval = X_tgt[mask]; y_eval = y_tgt[mask]
|
| 100 |
+
if len(y_eval) == 0:
|
| 101 |
+
continue
|
| 102 |
+
X_tr = np.concatenate([X_src, X_tgt[tgt_idx_train]], axis=0)
|
| 103 |
+
y_tr = np.concatenate([y_src, y_tgt[tgt_idx_train]], axis=0)
|
| 104 |
+
acc = train_lp(X_tr, y_tr, X_eval, y_eval)
|
| 105 |
+
results["curve"][N].append(acc)
|
| 106 |
+
if results["curve"][N]:
|
| 107 |
+
m, sd = stats(results["curve"][N])
|
| 108 |
+
log(f" N={N:>3d}: {m:.3f} ± {sd:.3f} (per_class={per_class})")
|
| 109 |
+
else:
|
| 110 |
+
log(f" N={N:>3d}: SKIPPED (insufficient target data)")
|
| 111 |
+
return results
|
| 112 |
+
|
| 113 |
+
|
| 114 |
+
# ──────────────────────────────────────────────────────────────────
|
| 115 |
+
# Load standard collision features and labels
|
| 116 |
+
# ──────────────────────────────────────────────────────────────────
|
| 117 |
+
log("Loading standard collision features ...")
|
| 118 |
+
std_feat = torch.load(
|
| 119 |
+
"results/acceptance_boost/exp2_cache/feat_vjepa2_collision_orig.pt",
|
| 120 |
+
map_location="cpu", weights_only=False)["features"]
|
| 121 |
+
log(f" std collision features: {tuple(std_feat.shape)}")
|
| 122 |
+
|
| 123 |
+
labels = np.load("results/kinematics_vs_mechanics/labels_collision.npz")
|
| 124 |
+
restitution_bin = labels["restitution_bin"] # 600-d, 3 classes
|
| 125 |
+
mass_bin = labels["mass_bin"] # 600-d, 3 classes
|
| 126 |
+
velocity_pre_scalar = labels["velocity_pre_scalar"]
|
| 127 |
+
restitution_scalar = labels["restitution_scalar"]
|
| 128 |
+
log(f" labels: restit_bin classes {sorted(set(restitution_bin))}, mass_bin classes {sorted(set(mass_bin))}")
|
| 129 |
+
|
| 130 |
+
X_std = pool_l2(std_feat) # (600, 1024)
|
| 131 |
+
log(f" X_std shape: {X_std.shape}")
|
| 132 |
+
|
| 133 |
+
|
| 134 |
+
# ──────────────────────────────────────────────────────────────────
|
| 135 |
+
# 1. Velocity interpolation (matched-visual kinematic split)
|
| 136 |
+
# train on low-velocity half, eval on high-velocity half (predict restitution_bin)
|
| 137 |
+
# ──────────────────────────────────────────────────────────────────
|
| 138 |
+
log("=== Velocity interpolation split ===")
|
| 139 |
+
vmed = float(np.median(velocity_pre_scalar))
|
| 140 |
+
log(f" velocity median = {vmed:.3f}")
|
| 141 |
+
mask_lo = velocity_pre_scalar < vmed
|
| 142 |
+
mask_hi = ~mask_lo
|
| 143 |
+
|
| 144 |
+
# direction A: train on lo, eval on hi
|
| 145 |
+
res_velocity_lo2hi = run_split(
|
| 146 |
+
"velocity lo->hi",
|
| 147 |
+
X_std[mask_lo], restitution_bin[mask_lo],
|
| 148 |
+
X_std[mask_hi], restitution_bin[mask_hi],
|
| 149 |
+
n_classes=3,
|
| 150 |
+
)
|
| 151 |
+
res_velocity_hi2lo = run_split(
|
| 152 |
+
"velocity hi->lo",
|
| 153 |
+
X_std[mask_hi], restitution_bin[mask_hi],
|
| 154 |
+
X_std[mask_lo], restitution_bin[mask_lo],
|
| 155 |
+
n_classes=3,
|
| 156 |
+
)
|
| 157 |
+
|
| 158 |
+
|
| 159 |
+
# ──────────────────────────────────────────────────────────────────
|
| 160 |
+
# 2. Elastic vs inelastic split (matched-visual dynamics-class split)
|
| 161 |
+
# train on elastic (restit_scalar >= 0.5), eval on inelastic (predict mass_bin)
|
| 162 |
+
# ──────────────────────────────────────────────────────────────────
|
| 163 |
+
log("=== Elastic <-> inelastic split ===")
|
| 164 |
+
mask_elas = restitution_scalar >= 0.5
|
| 165 |
+
mask_inelas = ~mask_elas
|
| 166 |
+
log(f" n elastic: {mask_elas.sum()}, n inelastic: {mask_inelas.sum()}")
|
| 167 |
+
|
| 168 |
+
res_elas2inelas = run_split(
|
| 169 |
+
"elas->inelas",
|
| 170 |
+
X_std[mask_elas], mass_bin[mask_elas],
|
| 171 |
+
X_std[mask_inelas], mass_bin[mask_inelas],
|
| 172 |
+
n_classes=3,
|
| 173 |
+
)
|
| 174 |
+
res_inelas2elas = run_split(
|
| 175 |
+
"inelas->elas",
|
| 176 |
+
X_std[mask_inelas], mass_bin[mask_inelas],
|
| 177 |
+
X_std[mask_elas], mass_bin[mask_elas],
|
| 178 |
+
n_classes=3,
|
| 179 |
+
)
|
| 180 |
+
|
| 181 |
+
|
| 182 |
+
# ──────────────────────────────────────────────────────────────────
|
| 183 |
+
# 3. Standard gravity <-> low gravity (matched-visual dynamics shift)
|
| 184 |
+
# Need 75-scene std subset matching low-gravity (use seed-matched first 75 by RNG)
|
| 185 |
+
# The exp_p1 setup matched RNG so std-grav and low-grav share per-scene physics
|
| 186 |
+
# We use the first 75 scenes of std-grav (seed-aligned) as the matched set.
|
| 187 |
+
# ──────────────────────────────────────────────────────────────────
|
| 188 |
+
log("=== Std gravity <-> low gravity ===")
|
| 189 |
+
lg_path = "results/reviewer_response/exp_p1/feat_vjepa2_lowgrav.pt"
|
| 190 |
+
lg_feat = torch.load(lg_path, map_location="cpu", weights_only=False)["features"]
|
| 191 |
+
log(f" low-grav features: {tuple(lg_feat.shape)}")
|
| 192 |
+
X_lg = pool_l2(lg_feat)
|
| 193 |
+
|
| 194 |
+
# Load low-grav labels from index.json
|
| 195 |
+
with open("kubric/output/collision_low_gravity_dataset/index.json") as fh:
|
| 196 |
+
lg_idx = json.load(fh)
|
| 197 |
+
lg_restitution_scalar = np.array([s["restitution"] for s in lg_idx])
|
| 198 |
+
# Use Kubric union bins -- compute on std-grav restitution scalars
|
| 199 |
+
restit_bin_edges = np.percentile(restitution_scalar, [33.333, 66.667])
|
| 200 |
+
log(f" union restit bin edges: {restit_bin_edges}")
|
| 201 |
+
def to_bin(scalar, edges):
|
| 202 |
+
return np.searchsorted(edges, scalar)
|
| 203 |
+
y_lg_restit = to_bin(lg_restitution_scalar, restit_bin_edges).astype(np.int64)
|
| 204 |
+
y_std_restit = to_bin(restitution_scalar, restit_bin_edges).astype(np.int64)
|
| 205 |
+
log(f" lg restit bin distribution: {np.bincount(y_lg_restit)}")
|
| 206 |
+
log(f" std restit bin distribution: {np.bincount(y_std_restit)}")
|
| 207 |
+
|
| 208 |
+
# Use first 75 std-grav scenes (RNG-aligned to low-grav generation)
|
| 209 |
+
res_std2lg = run_split(
|
| 210 |
+
"std->lg",
|
| 211 |
+
X_std[:75], y_std_restit[:75],
|
| 212 |
+
X_lg, y_lg_restit,
|
| 213 |
+
n_classes=3,
|
| 214 |
+
)
|
| 215 |
+
res_lg2std = run_split(
|
| 216 |
+
"lg->std",
|
| 217 |
+
X_lg, y_lg_restit,
|
| 218 |
+
X_std[:75], y_std_restit[:75],
|
| 219 |
+
n_classes=3,
|
| 220 |
+
)
|
| 221 |
+
|
| 222 |
+
|
| 223 |
+
# ──────────────────────────────────────────────────────────────────
|
| 224 |
+
# Aggregate and save
|
| 225 |
+
# ──────────────────────────────────────────────────────────────────
|
| 226 |
+
def merge_dirs(a, b):
|
| 227 |
+
"""Average two directional results."""
|
| 228 |
+
out = {"N0_source_only": [], "curve": {N: [] for N in N_LIST}}
|
| 229 |
+
out["N0_source_only"] = a["N0_source_only"] + b["N0_source_only"]
|
| 230 |
+
for N in N_LIST:
|
| 231 |
+
out["curve"][N] = a["curve"][N] + b["curve"][N]
|
| 232 |
+
return out
|
| 233 |
+
|
| 234 |
+
|
| 235 |
+
full = {
|
| 236 |
+
"velocity_lo2hi": res_velocity_lo2hi,
|
| 237 |
+
"velocity_hi2lo": res_velocity_hi2lo,
|
| 238 |
+
"velocity_mean": merge_dirs(res_velocity_lo2hi, res_velocity_hi2lo),
|
| 239 |
+
"elas2inelas": res_elas2inelas,
|
| 240 |
+
"inelas2elas": res_inelas2elas,
|
| 241 |
+
"elastic_mean": merge_dirs(res_elas2inelas, res_inelas2elas),
|
| 242 |
+
"std2lg": res_std2lg,
|
| 243 |
+
"lg2std": res_lg2std,
|
| 244 |
+
"gravity_mean": merge_dirs(res_std2lg, res_lg2std),
|
| 245 |
+
}
|
| 246 |
+
|
| 247 |
+
# Pretty summary
|
| 248 |
+
SUMMARY = ["EXP REV-LP-MV -- linear probes on matched-visual conditions (5 seeds, predict restitution/mass)",
|
| 249 |
+
"",
|
| 250 |
+
f"{'Condition':<30s} | {'N=0 (src-only)':>18s} | {'N=16':>14s} | {'N=192':>14s}",
|
| 251 |
+
"-" * 86]
|
| 252 |
+
for name, r in full.items():
|
| 253 |
+
if "_mean" not in name and not name in ("std2lg", "lg2std", "elas2inelas", "inelas2elas"):
|
| 254 |
+
continue
|
| 255 |
+
n0_m, n0_s = stats(r["N0_source_only"])
|
| 256 |
+
n16_m, n16_s = stats(r["curve"][16])
|
| 257 |
+
n192_m, n192_s = stats(r["curve"][192])
|
| 258 |
+
SUMMARY.append(f"{name:<30s} | {n0_m*100:>5.1f}% +/- {n0_s*100:>4.1f}% | {n16_m*100:>5.1f}% +/- {n16_s*100:>4.1f}% | {n192_m*100:>5.1f}% +/- {n192_s*100:>4.1f}%")
|
| 259 |
+
|
| 260 |
+
print("\n".join(SUMMARY), flush=True)
|
| 261 |
+
with open(OUT / "exp_lp_matched_visual_summary.txt", "w") as fh:
|
| 262 |
+
fh.write("\n".join(SUMMARY) + "\n")
|
| 263 |
+
with open(OUT / "exp_lp_matched_visual_summary.json", "w") as fh:
|
| 264 |
+
json.dump(full, fh, indent=2)
|
| 265 |
+
|
| 266 |
+
end_ts = datetime.now(timezone.utc).isoformat()
|
| 267 |
+
runtime_min = (time.time() - T0) / 60.0
|
| 268 |
+
print(f"\nEND_TIME = {end_ts}", flush=True)
|
| 269 |
+
print(f"Total runtime: {runtime_min:.2f} min", flush=True)
|
code/_rev_m_continuous_bottleneck.py
ADDED
|
@@ -0,0 +1,595 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP M (reviewer_response): CONTINUOUS COMPOSITIONAL BASELINE.
|
| 3 |
+
|
| 4 |
+
The #1 reviewer objection: "you only tested DISCRETE bottleneck codes.
|
| 5 |
+
Continuous factorized representations might transfer fine."
|
| 6 |
+
|
| 7 |
+
This experiment trains a CONTINUOUS bottleneck (same encoder + multi-agent
|
| 8 |
+
structure, but tanh-bounded real-valued codes instead of Gumbel one-hot)
|
| 9 |
+
on V-JEPA 2 collision restitution. We measure within-scenario TopSim,
|
| 10 |
+
PosDis, and causal specificity, then run the same N-shot cross-scenario
|
| 11 |
+
curve as Exp I.
|
| 12 |
+
|
| 13 |
+
Two variants tried:
|
| 14 |
+
- code_dim=10 per agent (matches discrete dimensionally: 4 agents x 10
|
| 15 |
+
= 40-dim message, same as discrete K=5 vocab x 2 heads x 4 agents)
|
| 16 |
+
- code_dim=3 per agent (small bottleneck, matches Option B from prompt)
|
| 17 |
+
|
| 18 |
+
If continuous bottleneck plateaus at 45-50% (like discrete), the
|
| 19 |
+
"compositionality without invariance" claim survives discretization.
|
| 20 |
+
If it recovers like a linear probe (60-84%), the claim must narrow to
|
| 21 |
+
discrete codes specifically.
|
| 22 |
+
"""
|
| 23 |
+
import json, time, sys, os, math
|
| 24 |
+
from pathlib import Path
|
| 25 |
+
from datetime import datetime, timezone
|
| 26 |
+
import numpy as np
|
| 27 |
+
import torch
|
| 28 |
+
import torch.nn as nn
|
| 29 |
+
import torch.nn.functional as F
|
| 30 |
+
|
| 31 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 32 |
+
from _kinematics_train import (
|
| 33 |
+
DEVICE, ClassifierReceiver,
|
| 34 |
+
HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR,
|
| 35 |
+
EARLY_STOP_PATIENCE,
|
| 36 |
+
)
|
| 37 |
+
from _killer_experiment import TemporalEncoder, ContinuousSender, ContinuousMultiSender
|
| 38 |
+
from _overnight_p1_transfer import (
|
| 39 |
+
train_base as train_discrete_base,
|
| 40 |
+
train_receiver_frozen_sender as train_disc_recv,
|
| 41 |
+
eval_zero_shot as eval_disc_zero_shot,
|
| 42 |
+
make_splits, N_FRAMES_SUBSAMPLE,
|
| 43 |
+
)
|
| 44 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 45 |
+
from _rev_f_cnn_control import ci95
|
| 46 |
+
|
| 47 |
+
OUT = Path("results/reviewer_response/exp_m")
|
| 48 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 49 |
+
|
| 50 |
+
N_EPOCHS = 150
|
| 51 |
+
N_SEEDS = 5
|
| 52 |
+
N_LIST = [0, 1, 4, 16, 64, 128, 192]
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
def log(msg):
|
| 56 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 57 |
+
print(f"[{ts}] EXP-M: {msg}", flush=True)
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 61 |
+
# Continuous bottleneck training
|
| 62 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 63 |
+
|
| 64 |
+
def build_continuous_sender(feat_dim, code_dim_per_agent=10, fpa=1):
|
| 65 |
+
senders = [
|
| 66 |
+
ContinuousSender(
|
| 67 |
+
TemporalEncoder(HIDDEN_DIM, feat_dim, fpa),
|
| 68 |
+
HIDDEN_DIM, code_dim_per_agent)
|
| 69 |
+
for _ in range(N_AGENTS)
|
| 70 |
+
]
|
| 71 |
+
return ContinuousMultiSender(senders).to(DEVICE)
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
def train_continuous_base(feat, labels, seed, code_dim_per_agent=10,
|
| 75 |
+
n_epochs=N_EPOCHS):
|
| 76 |
+
"""Train continuous sender + 3 receivers (iterated learning) on (feat, labels)."""
|
| 77 |
+
N, nf, dim = feat.shape
|
| 78 |
+
fpa = 1
|
| 79 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 80 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 81 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 82 |
+
train_ids, holdout_ids = [], []
|
| 83 |
+
for c in np.unique(labels):
|
| 84 |
+
ids_c = np.where(labels == c)[0]
|
| 85 |
+
rng.shuffle(ids_c)
|
| 86 |
+
split = max(1, len(ids_c) // 5)
|
| 87 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 88 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 89 |
+
n_classes = int(labels.max()) + 1
|
| 90 |
+
chance = 1.0 / n_classes
|
| 91 |
+
|
| 92 |
+
msg_dim = code_dim_per_agent * N_AGENTS
|
| 93 |
+
sender = build_continuous_sender(dim, code_dim_per_agent, fpa)
|
| 94 |
+
receivers = [ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 95 |
+
for _ in range(3)]
|
| 96 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 97 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 98 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 99 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 100 |
+
best_acc = 0.0; best_ep = 0
|
| 101 |
+
best_sender_state = None; best_receiver_states = None
|
| 102 |
+
best_recv_idx = 0
|
| 103 |
+
|
| 104 |
+
for ep in range(n_epochs):
|
| 105 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break
|
| 106 |
+
if ep > 0 and ep % 40 == 0:
|
| 107 |
+
for i in range(len(receivers)):
|
| 108 |
+
receivers[i] = ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 109 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 110 |
+
sender.train(); [r.train() for r in receivers]
|
| 111 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 112 |
+
perm = rng_ep.permutation(train_ids)
|
| 113 |
+
for b in range(n_batches):
|
| 114 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 115 |
+
if len(batch_ids) < 4: continue
|
| 116 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 117 |
+
tgt = labels_dev[batch_ids]
|
| 118 |
+
msg, _ = sender(views)
|
| 119 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 120 |
+
for r in receivers: loss = loss + F.cross_entropy(r(msg), tgt)
|
| 121 |
+
loss = loss / len(receivers)
|
| 122 |
+
if torch.isnan(loss):
|
| 123 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 124 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 125 |
+
loss.backward()
|
| 126 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 127 |
+
so.step(); [o.step() for o in ros]
|
| 128 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 129 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 130 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 131 |
+
with torch.no_grad():
|
| 132 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 133 |
+
msg_ho, _ = sender(v_ho)
|
| 134 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 135 |
+
best_per_recv = 0.0; best_idx = 0
|
| 136 |
+
for ri, r in enumerate(receivers):
|
| 137 |
+
preds = r(msg_ho).argmax(-1)
|
| 138 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 139 |
+
if acc > best_per_recv:
|
| 140 |
+
best_per_recv = acc; best_idx = ri
|
| 141 |
+
if best_per_recv > best_acc:
|
| 142 |
+
best_acc = best_per_recv; best_ep = ep
|
| 143 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 144 |
+
best_receiver_states = [
|
| 145 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 146 |
+
for r in receivers]
|
| 147 |
+
best_recv_idx = best_idx
|
| 148 |
+
return {
|
| 149 |
+
"sender_state": best_sender_state,
|
| 150 |
+
"receiver_states": best_receiver_states,
|
| 151 |
+
"best_recv_idx": best_recv_idx,
|
| 152 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 153 |
+
"task_acc": best_acc, "chance": chance,
|
| 154 |
+
"n_classes": n_classes, "fpa": 1, "dim": dim,
|
| 155 |
+
"code_dim_per_agent": code_dim_per_agent,
|
| 156 |
+
"msg_dim": msg_dim,
|
| 157 |
+
}
|
| 158 |
+
|
| 159 |
+
|
| 160 |
+
def get_continuous_messages(base, feat):
|
| 161 |
+
"""Apply the trained continuous sender to features. Returns msg (N, msg_dim)."""
|
| 162 |
+
N, nf, dim = feat.shape
|
| 163 |
+
code_dim = base["code_dim_per_agent"]
|
| 164 |
+
sender = build_continuous_sender(dim, code_dim, base["fpa"])
|
| 165 |
+
sender.load_state_dict(base["sender_state"])
|
| 166 |
+
sender.eval().to(DEVICE)
|
| 167 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 168 |
+
with torch.no_grad():
|
| 169 |
+
views = [v.to(DEVICE) for v in agent_views]
|
| 170 |
+
msg, _ = sender(views)
|
| 171 |
+
return msg.cpu().float()
|
| 172 |
+
|
| 173 |
+
|
| 174 |
+
def eval_zero_shot_cont(base, feat_tgt, labels_tgt, ho_ids):
|
| 175 |
+
"""Zero-shot apply trained sender + best receiver to target."""
|
| 176 |
+
sender = build_continuous_sender(feat_tgt.shape[2], base["code_dim_per_agent"], base["fpa"])
|
| 177 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 178 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 179 |
+
for _ in range(len(base["receiver_states"]))]
|
| 180 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 181 |
+
[r.eval() for r in receivers]
|
| 182 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 183 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 184 |
+
with torch.no_grad():
|
| 185 |
+
v_ho = [v[ho_ids].to(DEVICE) for v in agent_views]
|
| 186 |
+
msg_ho, _ = sender(v_ho)
|
| 187 |
+
tgt_ho = labels_dev[ho_ids]
|
| 188 |
+
best = 0.0
|
| 189 |
+
for r in receivers:
|
| 190 |
+
preds = r(msg_ho).argmax(-1)
|
| 191 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 192 |
+
best = max(best, acc)
|
| 193 |
+
return best
|
| 194 |
+
|
| 195 |
+
|
| 196 |
+
def train_recv_frozen_cont(base, feat_tgt, labels_tgt, train_ids, holdout_ids,
|
| 197 |
+
seed, n_target, n_epochs=80):
|
| 198 |
+
"""Train new receiver on n_target target examples using frozen continuous sender."""
|
| 199 |
+
if n_target == 0:
|
| 200 |
+
return eval_zero_shot_cont(base, feat_tgt, labels_tgt, holdout_ids)
|
| 201 |
+
rng = np.random.RandomState(seed * 311 + 7 + n_target)
|
| 202 |
+
n_t_classes = int(np.max(labels_tgt)) + 1
|
| 203 |
+
per_class = max(1, n_target // n_t_classes)
|
| 204 |
+
picks = []
|
| 205 |
+
for c in range(n_t_classes):
|
| 206 |
+
ids_c = np.array([i for i in train_ids if labels_tgt[i] == c])
|
| 207 |
+
if len(ids_c) == 0: continue
|
| 208 |
+
rng.shuffle(ids_c)
|
| 209 |
+
picks.extend(ids_c[:per_class])
|
| 210 |
+
picks = np.array(picks)
|
| 211 |
+
if len(picks) > n_target: picks = picks[:n_target]
|
| 212 |
+
elif len(picks) < n_target and len(train_ids) > len(picks):
|
| 213 |
+
extras = np.array([i for i in train_ids if i not in set(picks)])
|
| 214 |
+
rng.shuffle(extras)
|
| 215 |
+
picks = np.concatenate([picks, extras[:n_target - len(picks)]])
|
| 216 |
+
if len(picks) < 2: return float("nan")
|
| 217 |
+
|
| 218 |
+
# Freeze sender; train new receiver on `picks`
|
| 219 |
+
sender = build_continuous_sender(feat_tgt.shape[2], base["code_dim_per_agent"], base["fpa"])
|
| 220 |
+
sender.load_state_dict(base["sender_state"]); sender.to(DEVICE).eval()
|
| 221 |
+
for p in sender.parameters(): p.requires_grad = False
|
| 222 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 223 |
+
for _ in range(3)]
|
| 224 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 225 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 226 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 227 |
+
bs = min(BATCH_SIZE, len(picks))
|
| 228 |
+
best = 0.0
|
| 229 |
+
for ep in range(n_epochs):
|
| 230 |
+
[r.train() for r in receivers]
|
| 231 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 232 |
+
perm = rng_ep.permutation(picks)
|
| 233 |
+
for b in range(max(1, len(picks) // bs)):
|
| 234 |
+
batch = perm[b*bs:(b+1)*bs]
|
| 235 |
+
if len(batch) < 2: continue
|
| 236 |
+
views = [v[batch].to(DEVICE) for v in agent_views]
|
| 237 |
+
with torch.no_grad():
|
| 238 |
+
msg, _ = sender(views)
|
| 239 |
+
for r, o in zip(receivers, ros):
|
| 240 |
+
logits = r(msg)
|
| 241 |
+
loss = F.cross_entropy(logits, labels_dev[batch])
|
| 242 |
+
if torch.isnan(loss): continue
|
| 243 |
+
o.zero_grad(); loss.backward(); o.step()
|
| 244 |
+
if (ep + 1) % 5 == 0 or ep == 0:
|
| 245 |
+
[r.eval() for r in receivers]
|
| 246 |
+
with torch.no_grad():
|
| 247 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 248 |
+
msg_ho, _ = sender(v_ho)
|
| 249 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 250 |
+
for r in receivers:
|
| 251 |
+
preds = r(msg_ho).argmax(-1)
|
| 252 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 253 |
+
if acc > best: best = acc
|
| 254 |
+
return best
|
| 255 |
+
|
| 256 |
+
|
| 257 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 258 |
+
# Continuous metrics (TopSim, PosDis, causal-spec)
|
| 259 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 260 |
+
|
| 261 |
+
def topsim_continuous(messages, labels, n_pairs=5000):
|
| 262 |
+
"""Spearman corr between L2 message-distances and L1 label-distances."""
|
| 263 |
+
from scipy.stats import spearmanr
|
| 264 |
+
rng = np.random.RandomState(42)
|
| 265 |
+
N = len(labels)
|
| 266 |
+
msg_np = messages.numpy() if isinstance(messages, torch.Tensor) else messages
|
| 267 |
+
n_pairs = min(n_pairs, N * (N - 1) // 2)
|
| 268 |
+
msg_d = []; lbl_d = []
|
| 269 |
+
seen = set()
|
| 270 |
+
for _ in range(n_pairs):
|
| 271 |
+
i, j = rng.randint(0, N), rng.randint(0, N)
|
| 272 |
+
if i == j or (i, j) in seen or (j, i) in seen: continue
|
| 273 |
+
seen.add((i, j))
|
| 274 |
+
msg_d.append(np.linalg.norm(msg_np[i] - msg_np[j]))
|
| 275 |
+
lbl_d.append(abs(int(labels[i]) - int(labels[j])))
|
| 276 |
+
if len(msg_d) < 10: return float("nan")
|
| 277 |
+
if np.std(msg_d) < 1e-9 or np.std(lbl_d) < 1e-9:
|
| 278 |
+
return float("nan")
|
| 279 |
+
rho, _ = spearmanr(msg_d, lbl_d)
|
| 280 |
+
return float(rho) if not np.isnan(rho) else 0.0
|
| 281 |
+
|
| 282 |
+
|
| 283 |
+
def posdis_continuous_per_dim(messages, labels, n_bins=10):
|
| 284 |
+
"""For each code dim, bin its values into n_bins and compute MI with labels.
|
| 285 |
+
Returns array (D,) of MI values."""
|
| 286 |
+
msg_np = messages.numpy() if isinstance(messages, torch.Tensor) else messages
|
| 287 |
+
D = msg_np.shape[1]
|
| 288 |
+
mi_per_dim = np.zeros(D)
|
| 289 |
+
n = len(labels)
|
| 290 |
+
for d in range(D):
|
| 291 |
+
col = msg_np[:, d]
|
| 292 |
+
if col.std() < 1e-9:
|
| 293 |
+
mi_per_dim[d] = 0.0; continue
|
| 294 |
+
# Bin
|
| 295 |
+
edges = np.quantile(col, np.linspace(0, 1, n_bins + 1)[1:-1])
|
| 296 |
+
binned = np.digitize(col, edges)
|
| 297 |
+
# MI(binned, labels)
|
| 298 |
+
joint = {}
|
| 299 |
+
for x, y in zip(binned, labels):
|
| 300 |
+
joint[(int(x), int(y))] = joint.get((int(x), int(y)), 0) + 1
|
| 301 |
+
H = lambda probs: -np.sum([p * np.log(p) for p in probs if p > 0])
|
| 302 |
+
# Marginals
|
| 303 |
+
p_x = np.bincount(binned, minlength=n_bins) / n
|
| 304 |
+
p_y = np.bincount(labels, minlength=int(np.max(labels)) + 1) / n
|
| 305 |
+
H_x = H(p_x); H_y = H(p_y)
|
| 306 |
+
H_xy = 0
|
| 307 |
+
for (x, y), c in joint.items():
|
| 308 |
+
p = c / n
|
| 309 |
+
H_xy += -p * np.log(p)
|
| 310 |
+
mi = H_x + H_y - H_xy
|
| 311 |
+
mi_per_dim[d] = max(mi, 0.0)
|
| 312 |
+
return mi_per_dim
|
| 313 |
+
|
| 314 |
+
|
| 315 |
+
def posdis_continuous(messages, labels, n_bins=10):
|
| 316 |
+
"""Average disentanglement across positions: per dim MI(top property) -
|
| 317 |
+
MI(second property), normalized. With single property here, it's just
|
| 318 |
+
relative MI heterogeneity across dims (disentanglement of the SINGLE
|
| 319 |
+
property across multiple dims). For single-attribute case, return
|
| 320 |
+
fraction of MI concentrated in one code dim."""
|
| 321 |
+
mi = posdis_continuous_per_dim(messages, labels, n_bins=n_bins)
|
| 322 |
+
if mi.sum() < 1e-9: return float("nan")
|
| 323 |
+
# Concentration: top dim MI / sum of MI across dims
|
| 324 |
+
top = mi.max()
|
| 325 |
+
return float(top / (mi.sum() + 1e-9))
|
| 326 |
+
|
| 327 |
+
|
| 328 |
+
def causal_specificity(base, feat, labels, holdout_ids):
|
| 329 |
+
"""Mask each code dim, measure receiver accuracy drop. Returns array (D,)."""
|
| 330 |
+
sender = build_continuous_sender(feat.shape[2], base["code_dim_per_agent"], base["fpa"])
|
| 331 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 332 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 333 |
+
for _ in range(len(base["receiver_states"]))]
|
| 334 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 335 |
+
[r.eval() for r in receivers]
|
| 336 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 337 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 338 |
+
with torch.no_grad():
|
| 339 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 340 |
+
msg_ho, _ = sender(v_ho)
|
| 341 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 342 |
+
best_recv = receivers[base.get("best_recv_idx", 0)]
|
| 343 |
+
baseline = (best_recv(msg_ho).argmax(-1) == tgt_ho).float().mean().item()
|
| 344 |
+
D = msg_ho.shape[1]
|
| 345 |
+
drops = np.zeros(D)
|
| 346 |
+
# Use mean of msg as the masked value
|
| 347 |
+
mean_vals = msg_ho.mean(dim=0)
|
| 348 |
+
for d in range(D):
|
| 349 |
+
masked = msg_ho.clone()
|
| 350 |
+
masked[:, d] = mean_vals[d]
|
| 351 |
+
acc_masked = (best_recv(masked).argmax(-1) == tgt_ho).float().mean().item()
|
| 352 |
+
drops[d] = baseline - acc_masked
|
| 353 |
+
return baseline, drops
|
| 354 |
+
|
| 355 |
+
|
| 356 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 357 |
+
# Main
|
| 358 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 359 |
+
|
| 360 |
+
def main():
|
| 361 |
+
t0 = time.time()
|
| 362 |
+
log("=" * 60)
|
| 363 |
+
log("EXP M: Continuous compositional baseline")
|
| 364 |
+
|
| 365 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 366 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 367 |
+
feat_f = load_feat_subsampled("flat_drop", "vjepa2")
|
| 368 |
+
lbl_c = load_labels("collision", "restitution")
|
| 369 |
+
lbl_r = load_labels("ramp", "restitution")
|
| 370 |
+
lbl_f = load_labels("flat_drop", "restitution")
|
| 371 |
+
log(f" collision: {tuple(feat_c.shape)} dist={np.bincount(lbl_c).tolist()}")
|
| 372 |
+
log(f" ramp: {tuple(feat_r.shape)} dist={np.bincount(lbl_r).tolist()}")
|
| 373 |
+
log(f" flat_drop: {tuple(feat_f.shape)} dist={np.bincount(lbl_f).tolist()}")
|
| 374 |
+
|
| 375 |
+
variants = {
|
| 376 |
+
"continuous_dim10": 10, # matches discrete msg dim (40 total = 4 agents x 10)
|
| 377 |
+
"continuous_dim3": 3, # small bottleneck
|
| 378 |
+
}
|
| 379 |
+
|
| 380 |
+
all_results = {}
|
| 381 |
+
|
| 382 |
+
# ── Within-collision training (5 seeds) per variant ──
|
| 383 |
+
for variant_name, code_dim in variants.items():
|
| 384 |
+
log(f"\n --- Training {variant_name} (code_dim_per_agent={code_dim}) ---")
|
| 385 |
+
bases = []
|
| 386 |
+
within_accs = []
|
| 387 |
+
for seed in range(N_SEEDS):
|
| 388 |
+
t_s = time.time()
|
| 389 |
+
try:
|
| 390 |
+
base = train_continuous_base(feat_c, lbl_c, seed,
|
| 391 |
+
code_dim_per_agent=code_dim,
|
| 392 |
+
n_epochs=N_EPOCHS)
|
| 393 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 394 |
+
log(f" seed {seed}: within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]")
|
| 395 |
+
except Exception as e:
|
| 396 |
+
log(f" seed {seed} FAILED: {e}")
|
| 397 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 398 |
+
all_results[variant_name] = {
|
| 399 |
+
"code_dim": code_dim,
|
| 400 |
+
"bases": bases, "within": within_accs,
|
| 401 |
+
}
|
| 402 |
+
|
| 403 |
+
# ── Within metrics on best-seed base ──
|
| 404 |
+
# Pick best within-acc base for metric reporting
|
| 405 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 406 |
+
if not valid:
|
| 407 |
+
log(f" {variant_name}: no successful base"); continue
|
| 408 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 409 |
+
best_base = bases[best_idx]
|
| 410 |
+
with torch.no_grad():
|
| 411 |
+
msgs_full = get_continuous_messages(best_base, feat_c)
|
| 412 |
+
ho_ids = best_base["holdout_ids"]
|
| 413 |
+
msgs_ho = msgs_full[ho_ids]
|
| 414 |
+
lbl_ho = lbl_c[ho_ids]
|
| 415 |
+
try:
|
| 416 |
+
ts = topsim_continuous(msgs_ho, lbl_ho)
|
| 417 |
+
except Exception as e:
|
| 418 |
+
log(f" TopSim error: {e}"); ts = float("nan")
|
| 419 |
+
try:
|
| 420 |
+
pd_ = posdis_continuous(msgs_ho, lbl_ho)
|
| 421 |
+
except Exception as e:
|
| 422 |
+
log(f" PosDis error: {e}"); pd_ = float("nan")
|
| 423 |
+
try:
|
| 424 |
+
base_acc, drops = causal_specificity(best_base, feat_c, lbl_c, ho_ids)
|
| 425 |
+
cs = float(drops.max())
|
| 426 |
+
except Exception as e:
|
| 427 |
+
log(f" causal-spec error: {e}"); cs = float("nan"); base_acc = float("nan")
|
| 428 |
+
log(f" {variant_name} within metrics (best seed): "
|
| 429 |
+
f"acc={base_acc:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 430 |
+
f"CausalSpec(max-drop)={cs:.3f}")
|
| 431 |
+
all_results[variant_name].update({
|
| 432 |
+
"topsim": ts, "posdis": pd_, "causal_spec_max": cs,
|
| 433 |
+
"within_for_metrics": base_acc,
|
| 434 |
+
})
|
| 435 |
+
|
| 436 |
+
# ── N-shot cross-scenario curves (5 seeds) per variant per direction ──
|
| 437 |
+
log(f"\n --- N-shot cross-scenario (N_list={N_LIST}, 5 seeds each) ---")
|
| 438 |
+
for variant_name in variants:
|
| 439 |
+
bases = all_results[variant_name]["bases"]
|
| 440 |
+
all_results[variant_name]["cross"] = {}
|
| 441 |
+
for src, tgt, feat_tgt, lbl_tgt in [
|
| 442 |
+
("collision", "ramp", feat_r, lbl_r),
|
| 443 |
+
("collision", "flat_drop", feat_f, lbl_f),
|
| 444 |
+
]:
|
| 445 |
+
log(f" {variant_name}: {src} -> {tgt}")
|
| 446 |
+
curve = {n: [] for n in N_LIST}
|
| 447 |
+
for seed, base in enumerate(bases):
|
| 448 |
+
if base is None:
|
| 449 |
+
for n in N_LIST: curve[n].append(float("nan"))
|
| 450 |
+
continue
|
| 451 |
+
tr_t, ho_t = make_splits(lbl_tgt, seed)
|
| 452 |
+
for n in N_LIST:
|
| 453 |
+
try:
|
| 454 |
+
acc = train_recv_frozen_cont(
|
| 455 |
+
base, feat_tgt, lbl_tgt, tr_t, ho_t, seed, n)
|
| 456 |
+
except Exception as e:
|
| 457 |
+
log(f" {variant_name} {src}->{tgt} s{seed} N={n} failed: {e}")
|
| 458 |
+
acc = float("nan")
|
| 459 |
+
curve[n].append(acc)
|
| 460 |
+
all_results[variant_name]["cross"][f"{src}->{tgt}"] = curve
|
| 461 |
+
for n in N_LIST:
|
| 462 |
+
accs = curve[n]
|
| 463 |
+
v = [x for x in accs if not (isinstance(x, float) and np.isnan(x))]
|
| 464 |
+
if v:
|
| 465 |
+
log(f" {src}->{tgt} N={n}: {np.mean(v)*100:.1f}% +/- "
|
| 466 |
+
f"{(np.std(v, ddof=1) if len(v) > 1 else 0.0)*100:.1f}")
|
| 467 |
+
|
| 468 |
+
# ── Output ──
|
| 469 |
+
def m(vals):
|
| 470 |
+
v = [x for x in vals if not (isinstance(x, float) and np.isnan(x))]
|
| 471 |
+
if not v: return (float("nan"), float("nan"), (float("nan"), float("nan")))
|
| 472 |
+
mean = float(np.mean(v))
|
| 473 |
+
std = float(np.std(v, ddof=1)) if len(v) > 1 else 0.0
|
| 474 |
+
return (mean, std, ci95(v))
|
| 475 |
+
|
| 476 |
+
lines = [
|
| 477 |
+
"EXPERIMENT M -- CONTINUOUS COMPOSITIONAL BASELINE (V-JEPA 2, 5 seeds)",
|
| 478 |
+
"",
|
| 479 |
+
"Architecture: same TemporalEncoder + multi-agent (4) structure as the",
|
| 480 |
+
"discrete bottleneck. Each agent's sender outputs a tanh-bounded real",
|
| 481 |
+
"vector of code_dim_per_agent dims (instead of one-hot Gumbel-Softmax).",
|
| 482 |
+
"Receiver: same ClassifierReceiver MLP as discrete protocol.",
|
| 483 |
+
"Iterated learning: 3-receiver population reset every 40 epochs.",
|
| 484 |
+
"",
|
| 485 |
+
"WITHIN-SCENARIO METRICS (collision, restitution 3-class):",
|
| 486 |
+
f"{'Architecture':<26s} | {'Acc':<8s} | {'TopSim':<8s} | {'PosDis':<10s} | "
|
| 487 |
+
f"{'CausalSpec':<12s}",
|
| 488 |
+
"-" * 80,
|
| 489 |
+
]
|
| 490 |
+
discrete_line = (f"{'Discrete (battery)':<26s} | {'94.2%':<8s} | "
|
| 491 |
+
f"{'+0.84':<8s} | {'0.76':<10s} | {'0.99':<12s}")
|
| 492 |
+
lines.append(discrete_line)
|
| 493 |
+
for variant_name in variants:
|
| 494 |
+
r = all_results[variant_name]
|
| 495 |
+
wm, ws, _ = m(r["within"])
|
| 496 |
+
ts = r.get("topsim", float("nan"))
|
| 497 |
+
pd_ = r.get("posdis", float("nan"))
|
| 498 |
+
cs = r.get("causal_spec_max", float("nan"))
|
| 499 |
+
within_str = f"{wm*100:.1f}%+/-{ws*100:.1f}" if not np.isnan(wm) else "N/A"
|
| 500 |
+
ts_str = f"{ts:+.2f}" if not np.isnan(ts) else "N/A"
|
| 501 |
+
pd_str = f"{pd_:.2f}" if not np.isnan(pd_) else "N/A"
|
| 502 |
+
cs_str = f"{cs:.2f}" if not np.isnan(cs) else "N/A"
|
| 503 |
+
lines.append(f"{variant_name:<26s} | {within_str:<8s} | "
|
| 504 |
+
f"{ts_str:<8s} | {pd_str:<10s} | {cs_str:<12s}")
|
| 505 |
+
lines.append(f"{'Linear probe (Exp B)':<26s} | {'97.5%':<8s} | "
|
| 506 |
+
f"{'N/A':<8s} | {'N/A':<10s} | {'N/A':<12s}")
|
| 507 |
+
|
| 508 |
+
lines.append("")
|
| 509 |
+
lines.append("N-SHOT CROSS-SCENARIO CURVE (collision -> ramp + collision -> flat_drop):")
|
| 510 |
+
lines.append(f" reference: linear probe coll->ramp at N=192: 83.7%")
|
| 511 |
+
lines.append(f" reference: linear probe coll->flat at N=192: 62.0%")
|
| 512 |
+
lines.append(f" reference: discrete bottleneck coll->ramp 16-shot: 43.7%")
|
| 513 |
+
lines.append("")
|
| 514 |
+
for direction in ["collision->ramp", "collision->flat_drop"]:
|
| 515 |
+
lines.append(f"--- {direction} ---")
|
| 516 |
+
header = (f"{'N':<6s} | "
|
| 517 |
+
f"{'continuous_dim10':<22s} | "
|
| 518 |
+
f"{'continuous_dim3':<22s}")
|
| 519 |
+
lines.append(header); lines.append("-" * len(header))
|
| 520 |
+
for n in N_LIST:
|
| 521 |
+
row_cells = []
|
| 522 |
+
for variant_name in variants:
|
| 523 |
+
accs = all_results[variant_name]["cross"][direction][n]
|
| 524 |
+
mn, sd, _ = m(accs)
|
| 525 |
+
if np.isnan(mn): row_cells.append("N/A")
|
| 526 |
+
else: row_cells.append(f"{mn*100:5.1f}% +/- {sd*100:.1f}")
|
| 527 |
+
lines.append(f"{n:<6d} | {row_cells[0]:<22s} | {row_cells[1]:<22s}")
|
| 528 |
+
lines.append("")
|
| 529 |
+
|
| 530 |
+
# Verdict
|
| 531 |
+
lines.append("VERDICT:")
|
| 532 |
+
# Compare continuous N=192 to linear probe and discrete bottleneck
|
| 533 |
+
cont10_192 = []; cont3_192 = []
|
| 534 |
+
for d in ["collision->ramp", "collision->flat_drop"]:
|
| 535 |
+
v10 = all_results["continuous_dim10"]["cross"][d][192]
|
| 536 |
+
v3 = all_results["continuous_dim3"]["cross"][d][192]
|
| 537 |
+
v10v = [x for x in v10 if not np.isnan(x)]
|
| 538 |
+
v3v = [x for x in v3 if not np.isnan(x)]
|
| 539 |
+
if v10v: cont10_192.append(float(np.mean(v10v)))
|
| 540 |
+
if v3v: cont3_192.append(float(np.mean(v3v)))
|
| 541 |
+
cont10_avg = float(np.mean(cont10_192)) if cont10_192 else float("nan")
|
| 542 |
+
cont3_avg = float(np.mean(cont3_192)) if cont3_192 else float("nan")
|
| 543 |
+
lines.append(f" Continuous-dim10 mean cross at N=192: {cont10_avg*100:.1f}%")
|
| 544 |
+
lines.append(f" Continuous-dim3 mean cross at N=192: {cont3_avg*100:.1f}%")
|
| 545 |
+
lines.append(f" Linear probe mean cross at N=192: ~73% (avg of 84% ramp, 62% flat)")
|
| 546 |
+
lines.append(f" Discrete bottleneck plateau: ~46%")
|
| 547 |
+
|
| 548 |
+
best_cont = max(cont10_avg, cont3_avg) if not (np.isnan(cont10_avg) and np.isnan(cont3_avg)) else float("nan")
|
| 549 |
+
if not np.isnan(best_cont):
|
| 550 |
+
if best_cont < 0.55:
|
| 551 |
+
v = (f"Continuous bottleneck plateaus at {best_cont*100:.1f}%, similar to "
|
| 552 |
+
"discrete (~46%). The compositionality-without-invariance dissociation "
|
| 553 |
+
"is NOT specific to discretization - it holds for continuous factorized "
|
| 554 |
+
"codes too. STRONG result for the paper.")
|
| 555 |
+
elif best_cont < 0.70:
|
| 556 |
+
v = (f"Continuous bottleneck reaches {best_cont*100:.1f}% at N=192 - "
|
| 557 |
+
"intermediate between discrete (46%) and linear probe (73%). Continuous "
|
| 558 |
+
"codes recover SOME cross-scenario signal beyond discrete, but stay "
|
| 559 |
+
"below an unconstrained probe. Nuanced finding.")
|
| 560 |
+
else:
|
| 561 |
+
v = (f"Continuous bottleneck recovers to {best_cont*100:.1f}% at N=192, "
|
| 562 |
+
"comparable to linear probes. The 'compositionality without invariance' "
|
| 563 |
+
"claim must be NARROWED to discrete codes specifically - continuous "
|
| 564 |
+
"factorized representations may transfer cleanly with target labels.")
|
| 565 |
+
lines.append(f" {v}")
|
| 566 |
+
|
| 567 |
+
lines.append("")
|
| 568 |
+
lines.append(f"Total runtime: {(time.time()-t0)/60:.1f} min")
|
| 569 |
+
|
| 570 |
+
# Strip torch tensors from results before JSON dump
|
| 571 |
+
json_out = {}
|
| 572 |
+
for variant_name, r in all_results.items():
|
| 573 |
+
json_out[variant_name] = {
|
| 574 |
+
"code_dim": r["code_dim"],
|
| 575 |
+
"within": r["within"],
|
| 576 |
+
"topsim": r.get("topsim", None),
|
| 577 |
+
"posdis": r.get("posdis", None),
|
| 578 |
+
"causal_spec_max": r.get("causal_spec_max", None),
|
| 579 |
+
"cross": r.get("cross", {}),
|
| 580 |
+
}
|
| 581 |
+
|
| 582 |
+
summary = "\n".join(lines)
|
| 583 |
+
(OUT / "exp_m_summary.txt").write_text(summary + "\n")
|
| 584 |
+
(OUT / "exp_m_summary.json").write_text(json.dumps({
|
| 585 |
+
"config": {"n_seeds": N_SEEDS, "N_list": N_LIST,
|
| 586 |
+
"variants": list(variants.keys())},
|
| 587 |
+
"results": json_out,
|
| 588 |
+
"runtime_s": time.time() - t0,
|
| 589 |
+
}, indent=2, default=str))
|
| 590 |
+
print("\n" + summary, flush=True)
|
| 591 |
+
log(f"DONE in {(time.time()-t0)/60:.1f} min")
|
| 592 |
+
|
| 593 |
+
|
| 594 |
+
if __name__ == "__main__":
|
| 595 |
+
main()
|
code/_rev_n_multiprop_continuous.py
ADDED
|
@@ -0,0 +1,544 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP N (reviewer_response): MULTI-PROPERTY CONTINUOUS BOTTLENECK.
|
| 3 |
+
|
| 4 |
+
Exp M trained a continuous bottleneck on a SINGLE property (restitution),
|
| 5 |
+
producing PosDis 0.04-0.15 because PosDis is structurally low for single-
|
| 6 |
+
attribute supervision. The discrete battery's PosDis 0.76 was on multi-
|
| 7 |
+
property training. To make the comparison fair, train continuous codes on
|
| 8 |
+
TWO properties (mass_bin + restitution_bin, both 3-class) with a 2-headed
|
| 9 |
+
receiver, then measure multi-property PosDis.
|
| 10 |
+
|
| 11 |
+
Architecture: same TemporalEncoder + multi-agent (4) ContinuousSender as
|
| 12 |
+
Exp M. Receiver gets the concatenated continuous message and decodes BOTH
|
| 13 |
+
properties via two parallel heads. Loss = sum of two CE losses. Iterated
|
| 14 |
+
learning: 3 receivers, reset every 40 epochs.
|
| 15 |
+
|
| 16 |
+
After training, compute multi-property metrics:
|
| 17 |
+
- TopSim: Spearman corr between L2 message-distance and label-vector
|
| 18 |
+
distance (concatenated [mass_bin, restitution_bin]).
|
| 19 |
+
- PosDis (multi-prop): for each code dimension, MI with each property.
|
| 20 |
+
PosDis = mean over dims of (top_MI - second_MI) / max(top_MI, eps).
|
| 21 |
+
Range: [0, 1]. High = each dim specializes for one property.
|
| 22 |
+
- CausalSpec: zero out each code dim, measure per-property accuracy
|
| 23 |
+
drop. CausalSpec = max specialization across (dim, property) pairs.
|
| 24 |
+
|
| 25 |
+
Cross-scenario eval on RESTITUTION only (mass is constant in ramp/flat).
|
| 26 |
+
"""
|
| 27 |
+
import json, time, sys, os, math
|
| 28 |
+
from pathlib import Path
|
| 29 |
+
from datetime import datetime, timezone
|
| 30 |
+
import numpy as np
|
| 31 |
+
import torch
|
| 32 |
+
import torch.nn as nn
|
| 33 |
+
import torch.nn.functional as F
|
| 34 |
+
|
| 35 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 36 |
+
from _kinematics_train import (
|
| 37 |
+
DEVICE, ClassifierReceiver,
|
| 38 |
+
HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR,
|
| 39 |
+
EARLY_STOP_PATIENCE,
|
| 40 |
+
)
|
| 41 |
+
from _killer_experiment import TemporalEncoder, ContinuousSender, ContinuousMultiSender
|
| 42 |
+
from _overnight_p1_transfer import make_splits
|
| 43 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 44 |
+
from _rev_f_cnn_control import ci95
|
| 45 |
+
from _rev_m_continuous_bottleneck import (
|
| 46 |
+
build_continuous_sender, get_continuous_messages,
|
| 47 |
+
train_recv_frozen_cont,
|
| 48 |
+
)
|
| 49 |
+
|
| 50 |
+
OUT = Path("results/reviewer_response/exp_n")
|
| 51 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 52 |
+
N_EPOCHS = 150
|
| 53 |
+
N_SEEDS = 5
|
| 54 |
+
N_LIST = [0, 16, 64, 192]
|
| 55 |
+
CODE_DIM = 3 # 3 per agent x 4 = 12 total dims; small enough for factorization pressure
|
| 56 |
+
N_PROPS = 2 # mass_bin + restitution_bin
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
def log(msg):
|
| 60 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 61 |
+
print(f"[{ts}] EXP-N: {msg}", flush=True)
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 65 |
+
# Multi-property receiver: shared body + per-property head
|
| 66 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 67 |
+
|
| 68 |
+
class MultiPropReceiver(nn.Module):
|
| 69 |
+
def __init__(self, msg_dim, hidden_dim=HIDDEN_DIM, n_classes_per_prop=(3, 3)):
|
| 70 |
+
super().__init__()
|
| 71 |
+
self.body = nn.Sequential(
|
| 72 |
+
nn.Linear(msg_dim, hidden_dim), nn.ReLU(),
|
| 73 |
+
nn.Linear(hidden_dim, hidden_dim), nn.ReLU(),
|
| 74 |
+
)
|
| 75 |
+
self.heads = nn.ModuleList([
|
| 76 |
+
nn.Linear(hidden_dim, n) for n in n_classes_per_prop
|
| 77 |
+
])
|
| 78 |
+
|
| 79 |
+
def forward(self, msg):
|
| 80 |
+
h = self.body(msg)
|
| 81 |
+
return [head(h) for head in self.heads]
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
def train_multiprop_continuous_base(feat, labels_list, seed,
|
| 85 |
+
code_dim_per_agent=CODE_DIM,
|
| 86 |
+
n_epochs=N_EPOCHS):
|
| 87 |
+
"""Train continuous sender + multi-prop receivers. labels_list = list of
|
| 88 |
+
per-scene int label arrays, one per property."""
|
| 89 |
+
N, nf, dim = feat.shape
|
| 90 |
+
fpa = 1
|
| 91 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 92 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 93 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 94 |
+
|
| 95 |
+
# Stratified split using the FIRST property
|
| 96 |
+
train_ids, holdout_ids = [], []
|
| 97 |
+
primary = labels_list[0]
|
| 98 |
+
for c in np.unique(primary):
|
| 99 |
+
ids_c = np.where(primary == c)[0]
|
| 100 |
+
rng.shuffle(ids_c)
|
| 101 |
+
split = max(1, len(ids_c) // 5)
|
| 102 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 103 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 104 |
+
n_classes_per_prop = [int(lbl.max()) + 1 for lbl in labels_list]
|
| 105 |
+
chance = 1.0 / max(n_classes_per_prop)
|
| 106 |
+
|
| 107 |
+
msg_dim = code_dim_per_agent * N_AGENTS
|
| 108 |
+
sender = build_continuous_sender(dim, code_dim_per_agent, fpa)
|
| 109 |
+
receivers = [MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE)
|
| 110 |
+
for _ in range(3)]
|
| 111 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 112 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 113 |
+
labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list]
|
| 114 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 115 |
+
best_acc = 0.0; best_ep = 0
|
| 116 |
+
best_sender_state = None; best_receiver_states = None
|
| 117 |
+
best_recv_idx = 0
|
| 118 |
+
|
| 119 |
+
for ep in range(n_epochs):
|
| 120 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break
|
| 121 |
+
if ep > 0 and ep % 40 == 0:
|
| 122 |
+
for i in range(len(receivers)):
|
| 123 |
+
receivers[i] = MultiPropReceiver(msg_dim, HIDDEN_DIM,
|
| 124 |
+
n_classes_per_prop).to(DEVICE)
|
| 125 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 126 |
+
sender.train(); [r.train() for r in receivers]
|
| 127 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 128 |
+
perm = rng_ep.permutation(train_ids)
|
| 129 |
+
for b in range(n_batches):
|
| 130 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 131 |
+
if len(batch_ids) < 4: continue
|
| 132 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 133 |
+
tgts = [ld[batch_ids] for ld in labels_dev]
|
| 134 |
+
msg, _ = sender(views)
|
| 135 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 136 |
+
for r in receivers:
|
| 137 |
+
logits_list = r(msg)
|
| 138 |
+
for logits, tgt in zip(logits_list, tgts):
|
| 139 |
+
loss = loss + F.cross_entropy(logits, tgt)
|
| 140 |
+
loss = loss / (len(receivers) * len(tgts))
|
| 141 |
+
if torch.isnan(loss):
|
| 142 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 143 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 144 |
+
loss.backward()
|
| 145 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 146 |
+
so.step(); [o.step() for o in ros]
|
| 147 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 148 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 149 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 150 |
+
with torch.no_grad():
|
| 151 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 152 |
+
msg_ho, _ = sender(v_ho)
|
| 153 |
+
tgt_ho = [ld[holdout_ids] for ld in labels_dev]
|
| 154 |
+
# "Best" combined accuracy = mean across both properties
|
| 155 |
+
best_per_recv = 0.0; best_idx = 0
|
| 156 |
+
for ri, r in enumerate(receivers):
|
| 157 |
+
logits_list = r(msg_ho)
|
| 158 |
+
accs = []
|
| 159 |
+
for logits, tgt in zip(logits_list, tgt_ho):
|
| 160 |
+
accs.append((logits.argmax(-1) == tgt).float().mean().item())
|
| 161 |
+
combined = float(np.mean(accs))
|
| 162 |
+
if combined > best_per_recv:
|
| 163 |
+
best_per_recv = combined; best_idx = ri
|
| 164 |
+
if best_per_recv > best_acc:
|
| 165 |
+
best_acc = best_per_recv; best_ep = ep
|
| 166 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 167 |
+
best_receiver_states = [
|
| 168 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 169 |
+
for r in receivers]
|
| 170 |
+
best_recv_idx = best_idx
|
| 171 |
+
return {
|
| 172 |
+
"sender_state": best_sender_state,
|
| 173 |
+
"receiver_states": best_receiver_states,
|
| 174 |
+
"best_recv_idx": best_recv_idx,
|
| 175 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 176 |
+
"task_acc": best_acc, "chance": chance,
|
| 177 |
+
"n_classes_per_prop": n_classes_per_prop,
|
| 178 |
+
"fpa": 1, "dim": dim,
|
| 179 |
+
"code_dim_per_agent": code_dim_per_agent,
|
| 180 |
+
"msg_dim": msg_dim,
|
| 181 |
+
}
|
| 182 |
+
|
| 183 |
+
|
| 184 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 185 |
+
# Multi-property metrics (standard PosDis)
|
| 186 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 187 |
+
|
| 188 |
+
def _mi_continuous(col, labels, n_bins=10):
|
| 189 |
+
"""MI between one binned continuous dim and discrete labels."""
|
| 190 |
+
if col.std() < 1e-9: return 0.0
|
| 191 |
+
edges = np.quantile(col, np.linspace(0, 1, n_bins + 1)[1:-1])
|
| 192 |
+
binned = np.digitize(col, edges)
|
| 193 |
+
n = len(labels)
|
| 194 |
+
n_lbl = int(np.max(labels)) + 1
|
| 195 |
+
p_x = np.bincount(binned, minlength=n_bins) / n
|
| 196 |
+
p_y = np.bincount(labels, minlength=n_lbl) / n
|
| 197 |
+
H_x = -np.sum([p * np.log(p) for p in p_x if p > 0])
|
| 198 |
+
H_y = -np.sum([p * np.log(p) for p in p_y if p > 0])
|
| 199 |
+
joint = np.zeros((n_bins, n_lbl))
|
| 200 |
+
for x, y in zip(binned, labels):
|
| 201 |
+
joint[int(x), int(y)] += 1
|
| 202 |
+
joint /= n
|
| 203 |
+
H_xy = 0.0
|
| 204 |
+
for v in joint.ravel():
|
| 205 |
+
if v > 0: H_xy -= v * np.log(v)
|
| 206 |
+
return max(H_x + H_y - H_xy, 0.0)
|
| 207 |
+
|
| 208 |
+
|
| 209 |
+
def topsim_multiprop(messages, labels_list, n_pairs=5000):
|
| 210 |
+
"""Spearman corr between message L2 distance and label-vector L1 distance."""
|
| 211 |
+
from scipy.stats import spearmanr
|
| 212 |
+
rng = np.random.RandomState(42)
|
| 213 |
+
msg_np = messages.numpy() if isinstance(messages, torch.Tensor) else messages
|
| 214 |
+
N = msg_np.shape[0]
|
| 215 |
+
msg_d = []; lbl_d = []
|
| 216 |
+
seen = set()
|
| 217 |
+
n_pairs = min(n_pairs, N * (N - 1) // 2)
|
| 218 |
+
for _ in range(n_pairs):
|
| 219 |
+
i, j = rng.randint(0, N), rng.randint(0, N)
|
| 220 |
+
if i == j or (i, j) in seen or (j, i) in seen: continue
|
| 221 |
+
seen.add((i, j))
|
| 222 |
+
msg_d.append(np.linalg.norm(msg_np[i] - msg_np[j]))
|
| 223 |
+
ld = sum(abs(int(lbl[i]) - int(lbl[j])) for lbl in labels_list)
|
| 224 |
+
lbl_d.append(ld)
|
| 225 |
+
if len(msg_d) < 10: return float("nan")
|
| 226 |
+
if np.std(msg_d) < 1e-9 or np.std(lbl_d) < 1e-9: return float("nan")
|
| 227 |
+
rho, _ = spearmanr(msg_d, lbl_d)
|
| 228 |
+
return float(rho) if not np.isnan(rho) else 0.0
|
| 229 |
+
|
| 230 |
+
|
| 231 |
+
def posdis_multiprop(messages, labels_list, n_bins=10):
|
| 232 |
+
"""Standard PosDis on continuous codes via per-dim MI binning.
|
| 233 |
+
For each dim d: compute MI(d, prop) for each property. Disentanglement
|
| 234 |
+
of dim d = (top - second) / max(top, eps). Mean over dims."""
|
| 235 |
+
msg_np = messages.numpy() if isinstance(messages, torch.Tensor) else messages
|
| 236 |
+
D = msg_np.shape[1]
|
| 237 |
+
P = len(labels_list)
|
| 238 |
+
mi_matrix = np.zeros((D, P))
|
| 239 |
+
for d in range(D):
|
| 240 |
+
for p in range(P):
|
| 241 |
+
mi_matrix[d, p] = _mi_continuous(msg_np[:, d], labels_list[p], n_bins)
|
| 242 |
+
if mi_matrix.sum() < 1e-9: return float("nan"), mi_matrix
|
| 243 |
+
pos_dis = 0.0
|
| 244 |
+
n_active_dims = 0
|
| 245 |
+
for d in range(D):
|
| 246 |
+
sorted_mi = np.sort(mi_matrix[d])[::-1]
|
| 247 |
+
if sorted_mi[0] > 1e-6:
|
| 248 |
+
pos_dis += (sorted_mi[0] - sorted_mi[1]) / sorted_mi[0]
|
| 249 |
+
n_active_dims += 1
|
| 250 |
+
if n_active_dims == 0: return float("nan"), mi_matrix
|
| 251 |
+
return float(pos_dis / n_active_dims), mi_matrix
|
| 252 |
+
|
| 253 |
+
|
| 254 |
+
def causal_spec_multiprop(base, feat, labels_list, holdout_ids):
|
| 255 |
+
"""Per-dim x per-property accuracy drop. Returns (D, P) matrix and overall max."""
|
| 256 |
+
sender = build_continuous_sender(feat.shape[2], base["code_dim_per_agent"], base["fpa"])
|
| 257 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 258 |
+
receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes_per_prop"]).to(DEVICE)
|
| 259 |
+
for _ in range(len(base["receiver_states"]))]
|
| 260 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 261 |
+
[r.eval() for r in receivers]
|
| 262 |
+
best_recv = receivers[base.get("best_recv_idx", 0)]
|
| 263 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 264 |
+
labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list]
|
| 265 |
+
P = len(labels_list)
|
| 266 |
+
with torch.no_grad():
|
| 267 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 268 |
+
msg_ho, _ = sender(v_ho)
|
| 269 |
+
tgt_ho = [ld[holdout_ids] for ld in labels_dev]
|
| 270 |
+
D = msg_ho.shape[1]
|
| 271 |
+
baseline_per_prop = []
|
| 272 |
+
for logits, tgt in zip(best_recv(msg_ho), tgt_ho):
|
| 273 |
+
baseline_per_prop.append((logits.argmax(-1) == tgt).float().mean().item())
|
| 274 |
+
drops = np.zeros((D, P))
|
| 275 |
+
mean_vals = msg_ho.mean(dim=0)
|
| 276 |
+
for d in range(D):
|
| 277 |
+
masked = msg_ho.clone()
|
| 278 |
+
masked[:, d] = mean_vals[d]
|
| 279 |
+
for p_idx, (logits, tgt) in enumerate(zip(best_recv(masked), tgt_ho)):
|
| 280 |
+
acc = (logits.argmax(-1) == tgt).float().mean().item()
|
| 281 |
+
drops[d, p_idx] = baseline_per_prop[p_idx] - acc
|
| 282 |
+
return baseline_per_prop, drops
|
| 283 |
+
|
| 284 |
+
|
| 285 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 286 |
+
# Main
|
| 287 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 288 |
+
|
| 289 |
+
def main():
|
| 290 |
+
t0 = time.time()
|
| 291 |
+
log("=" * 60)
|
| 292 |
+
log("EXP N: Multi-property continuous bottleneck")
|
| 293 |
+
log(f" code_dim_per_agent={CODE_DIM} (msg_dim={CODE_DIM*N_AGENTS})")
|
| 294 |
+
|
| 295 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 296 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 297 |
+
feat_f = load_feat_subsampled("flat_drop", "vjepa2")
|
| 298 |
+
lbl_c_mass = load_labels("collision", "mass")
|
| 299 |
+
lbl_c_restit = load_labels("collision", "restitution")
|
| 300 |
+
lbl_r_restit = load_labels("ramp", "restitution")
|
| 301 |
+
lbl_f_restit = load_labels("flat_drop", "restitution")
|
| 302 |
+
log(f" collision feat={tuple(feat_c.shape)} mass dist={np.bincount(lbl_c_mass).tolist()} "
|
| 303 |
+
f"restit dist={np.bincount(lbl_c_restit).tolist()}")
|
| 304 |
+
|
| 305 |
+
# Train multi-prop continuous bottleneck (5 seeds)
|
| 306 |
+
log(f"\n --- Training multi-prop continuous bottleneck (5 seeds) ---")
|
| 307 |
+
bases = []
|
| 308 |
+
within_combined = [] # mean of mass + restit accuracy on holdout
|
| 309 |
+
for seed in range(N_SEEDS):
|
| 310 |
+
t_s = time.time()
|
| 311 |
+
try:
|
| 312 |
+
base = train_multiprop_continuous_base(
|
| 313 |
+
feat_c, [lbl_c_mass, lbl_c_restit], seed,
|
| 314 |
+
code_dim_per_agent=CODE_DIM, n_epochs=N_EPOCHS)
|
| 315 |
+
bases.append(base); within_combined.append(float(base["task_acc"]))
|
| 316 |
+
log(f" seed {seed}: combined within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]")
|
| 317 |
+
except Exception as e:
|
| 318 |
+
log(f" seed {seed} FAILED: {e}")
|
| 319 |
+
bases.append(None); within_combined.append(float("nan"))
|
| 320 |
+
|
| 321 |
+
# Within-scenario metrics on best base
|
| 322 |
+
valid = [(i, a) for i, a in enumerate(within_combined) if not np.isnan(a)]
|
| 323 |
+
if not valid:
|
| 324 |
+
log("ERROR: no successful base"); return
|
| 325 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 326 |
+
best_base = bases[best_idx]
|
| 327 |
+
ho_ids = best_base["holdout_ids"]
|
| 328 |
+
log(f"\n --- Within-scenario metrics on best seed ({best_idx}, ho_n={len(ho_ids)}) ---")
|
| 329 |
+
|
| 330 |
+
# Per-property accuracies on holdout
|
| 331 |
+
sender = build_continuous_sender(feat_c.shape[2], CODE_DIM, best_base["fpa"])
|
| 332 |
+
sender.load_state_dict(best_base["sender_state"]); sender.eval().to(DEVICE)
|
| 333 |
+
receivers = [MultiPropReceiver(best_base["msg_dim"], HIDDEN_DIM,
|
| 334 |
+
best_base["n_classes_per_prop"]).to(DEVICE)
|
| 335 |
+
for _ in range(len(best_base["receiver_states"]))]
|
| 336 |
+
for r, s in zip(receivers, best_base["receiver_states"]): r.load_state_dict(s)
|
| 337 |
+
[r.eval() for r in receivers]
|
| 338 |
+
best_recv = receivers[best_base["best_recv_idx"]]
|
| 339 |
+
agent_views = [feat_c[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 340 |
+
with torch.no_grad():
|
| 341 |
+
v_ho = [v[ho_ids].to(DEVICE) for v in agent_views]
|
| 342 |
+
msg_ho, _ = sender(v_ho)
|
| 343 |
+
msgs_full = sender([v.to(DEVICE) for v in agent_views])[0].cpu().float()
|
| 344 |
+
msgs_ho_cpu = msg_ho.cpu().float()
|
| 345 |
+
tgt_mass = torch.tensor(lbl_c_mass[ho_ids], dtype=torch.long).to(DEVICE)
|
| 346 |
+
tgt_rest = torch.tensor(lbl_c_restit[ho_ids], dtype=torch.long).to(DEVICE)
|
| 347 |
+
out_mass, out_rest = best_recv(msg_ho)
|
| 348 |
+
acc_mass = (out_mass.argmax(-1) == tgt_mass).float().mean().item()
|
| 349 |
+
acc_rest = (out_rest.argmax(-1) == tgt_rest).float().mean().item()
|
| 350 |
+
log(f" holdout mass acc: {acc_mass:.3f}")
|
| 351 |
+
log(f" holdout restit acc: {acc_rest:.3f}")
|
| 352 |
+
|
| 353 |
+
# TopSim, PosDis, CausalSpec on multi-prop labels
|
| 354 |
+
try:
|
| 355 |
+
ts = topsim_multiprop(msgs_ho_cpu, [lbl_c_mass[ho_ids], lbl_c_restit[ho_ids]])
|
| 356 |
+
except Exception as e:
|
| 357 |
+
log(f" TopSim error: {e}"); ts = float("nan")
|
| 358 |
+
try:
|
| 359 |
+
pd_, mi_matrix = posdis_multiprop(msgs_ho_cpu, [lbl_c_mass[ho_ids], lbl_c_restit[ho_ids]])
|
| 360 |
+
except Exception as e:
|
| 361 |
+
log(f" PosDis error: {e}"); pd_ = float("nan"); mi_matrix = None
|
| 362 |
+
try:
|
| 363 |
+
baseline_per_prop, drops = causal_spec_multiprop(best_base, feat_c,
|
| 364 |
+
[lbl_c_mass, lbl_c_restit], ho_ids)
|
| 365 |
+
# CausalSpec: max relative drop (drop / baseline) per (dim, prop)
|
| 366 |
+
cs_max = float(np.max(drops))
|
| 367 |
+
except Exception as e:
|
| 368 |
+
log(f" causal-spec error: {e}"); cs_max = float("nan")
|
| 369 |
+
log(f" TopSim: {ts:+.3f}")
|
| 370 |
+
log(f" PosDis: {pd_:.3f}")
|
| 371 |
+
log(f" CausalSpec (max-drop): {cs_max:.3f}")
|
| 372 |
+
if mi_matrix is not None:
|
| 373 |
+
log(f" MI matrix (D x [mass, restit]):")
|
| 374 |
+
for d in range(min(mi_matrix.shape[0], 12)):
|
| 375 |
+
log(f" dim {d}: mass={mi_matrix[d,0]:.3f} restit={mi_matrix[d,1]:.3f}")
|
| 376 |
+
|
| 377 |
+
# Cross-scenario N-shot for restitution (the only common property)
|
| 378 |
+
log(f"\n --- N-shot cross-scenario on RESTITUTION (5 seeds) ---")
|
| 379 |
+
cross_results = {}
|
| 380 |
+
for direction, feat_tgt, lbl_tgt in [
|
| 381 |
+
("collision->ramp", feat_r, lbl_r_restit),
|
| 382 |
+
("collision->flat_drop", feat_f, lbl_f_restit),
|
| 383 |
+
]:
|
| 384 |
+
log(f" {direction}")
|
| 385 |
+
# Each base has 2 receivers (mass+restit). For frozen-sender 16-shot
|
| 386 |
+
# cross to a NEW property/scenario, we train a fresh single-property
|
| 387 |
+
# receiver on the bottleneck messages (matching Exp M's protocol).
|
| 388 |
+
# Use the single-property version of train_recv_frozen by reconstructing
|
| 389 |
+
# a single-task base dict.
|
| 390 |
+
curve = {n: [] for n in N_LIST}
|
| 391 |
+
for seed, base in enumerate(bases):
|
| 392 |
+
if base is None:
|
| 393 |
+
for n in N_LIST: curve[n].append(float("nan"))
|
| 394 |
+
continue
|
| 395 |
+
# Build a "single-task" base view for restitution by creating a
|
| 396 |
+
# restit-only receiver state (start fresh receiver per N-shot call)
|
| 397 |
+
single_base = dict(base)
|
| 398 |
+
single_base["n_classes"] = base["n_classes_per_prop"][1] # restit
|
| 399 |
+
single_base["receiver_states"] = [] # not used by train_recv_frozen_cont N>0
|
| 400 |
+
tr_t, ho_t = make_splits(lbl_tgt, seed)
|
| 401 |
+
for n in N_LIST:
|
| 402 |
+
try:
|
| 403 |
+
if n == 0:
|
| 404 |
+
# Zero-shot using the restitution head from training
|
| 405 |
+
sender2 = build_continuous_sender(
|
| 406 |
+
feat_tgt.shape[2], base["code_dim_per_agent"], base["fpa"])
|
| 407 |
+
sender2.load_state_dict(base["sender_state"])
|
| 408 |
+
sender2.eval().to(DEVICE)
|
| 409 |
+
receivers2 = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM,
|
| 410 |
+
base["n_classes_per_prop"]).to(DEVICE)
|
| 411 |
+
for _ in range(len(base["receiver_states"]))]
|
| 412 |
+
for r, s in zip(receivers2, base["receiver_states"]): r.load_state_dict(s)
|
| 413 |
+
[r.eval() for r in receivers2]
|
| 414 |
+
ag = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 415 |
+
labels_dev = torch.tensor(lbl_tgt, dtype=torch.long).to(DEVICE)
|
| 416 |
+
with torch.no_grad():
|
| 417 |
+
v_ho2 = [v[ho_t].to(DEVICE) for v in ag]
|
| 418 |
+
msg_ho2, _ = sender2(v_ho2)
|
| 419 |
+
tgt_ho2 = labels_dev[ho_t]
|
| 420 |
+
best = 0.0
|
| 421 |
+
for r in receivers2:
|
| 422 |
+
_, restit_logits = r(msg_ho2)
|
| 423 |
+
acc_zs = (restit_logits.argmax(-1) == tgt_ho2).float().mean().item()
|
| 424 |
+
best = max(best, acc_zs)
|
| 425 |
+
acc = best
|
| 426 |
+
else:
|
| 427 |
+
acc = train_recv_frozen_cont(
|
| 428 |
+
single_base, feat_tgt, lbl_tgt, tr_t, ho_t, seed, n)
|
| 429 |
+
except Exception as e:
|
| 430 |
+
log(f" {direction} s{seed} N={n} failed: {e}")
|
| 431 |
+
acc = float("nan")
|
| 432 |
+
curve[n].append(acc)
|
| 433 |
+
cross_results[direction] = curve
|
| 434 |
+
for n in N_LIST:
|
| 435 |
+
v = [x for x in curve[n] if not np.isnan(x)]
|
| 436 |
+
if v:
|
| 437 |
+
log(f" {direction} N={n}: {np.mean(v)*100:.1f}% +/- "
|
| 438 |
+
f"{(np.std(v, ddof=1) if len(v) > 1 else 0.0)*100:.1f}")
|
| 439 |
+
|
| 440 |
+
# ── Summary ──
|
| 441 |
+
def m(vals):
|
| 442 |
+
v = [x for x in vals if not (isinstance(x, float) and np.isnan(x))]
|
| 443 |
+
if not v: return (float("nan"), float("nan"), (float("nan"), float("nan")))
|
| 444 |
+
return float(np.mean(v)), (float(np.std(v, ddof=1)) if len(v) > 1 else 0.0), ci95(v)
|
| 445 |
+
|
| 446 |
+
lines = [
|
| 447 |
+
"EXPERIMENT N -- MULTI-PROPERTY CONTINUOUS BOTTLENECK (5 seeds)",
|
| 448 |
+
"",
|
| 449 |
+
"Architecture: same continuous sender as Exp M, but trained on TWO",
|
| 450 |
+
"properties simultaneously (mass_bin + restitution_bin, both 3-class)",
|
| 451 |
+
"via a 2-headed receiver. code_dim_per_agent = 3 (msg_dim = 12).",
|
| 452 |
+
"",
|
| 453 |
+
"WITHIN-SCENARIO (collision):",
|
| 454 |
+
f"{'Architecture':<32s} | {'Acc':<14s} | {'TopSim':<8s} | "
|
| 455 |
+
f"{'PosDis':<8s} | {'CausalSpec':<12s}",
|
| 456 |
+
"-" * 90,
|
| 457 |
+
]
|
| 458 |
+
lines.append(f"{'Discrete (battery, multi-prop)':<32s} | {'94.2%':<14s} | "
|
| 459 |
+
f"{'+0.84':<8s} | {'0.76':<8s} | {'0.99':<12s}")
|
| 460 |
+
lines.append(f"{'Continuous (Exp M, single-prop)':<32s} | {'96.0%':<14s} | "
|
| 461 |
+
f"{'+0.88':<8s} | {'0.04':<8s} | {'0.01':<12s}")
|
| 462 |
+
wm, ws, _ = m(within_combined)
|
| 463 |
+
lines.append(f"{'Continuous (Exp N, multi-prop)':<32s} | "
|
| 464 |
+
f"{wm*100:5.1f}%+/-{ws*100:.1f} | "
|
| 465 |
+
f"{ts:+.2f} | "
|
| 466 |
+
f"{pd_:.2f} | "
|
| 467 |
+
f"{cs_max:.2f}")
|
| 468 |
+
lines.append(f" (per-prop: mass={acc_mass*100:.1f}%, restit={acc_rest*100:.1f}%)")
|
| 469 |
+
|
| 470 |
+
lines.append("")
|
| 471 |
+
lines.append("CROSS-SCENARIO N-SHOT on restitution (5 seeds):")
|
| 472 |
+
lines.append(f"{'N':<5s} | {'coll->ramp':<18s} | {'coll->flat_drop':<22s} | {'Mean':<10s}")
|
| 473 |
+
lines.append("-" * 60)
|
| 474 |
+
plateau_means = []
|
| 475 |
+
for n in N_LIST:
|
| 476 |
+
vr = [x for x in cross_results["collision->ramp"][n] if not np.isnan(x)]
|
| 477 |
+
vf = [x for x in cross_results["collision->flat_drop"][n] if not np.isnan(x)]
|
| 478 |
+
rm = float(np.mean(vr)) if vr else float("nan")
|
| 479 |
+
fm = float(np.mean(vf)) if vf else float("nan")
|
| 480 |
+
rs = float(np.std(vr, ddof=1)) if len(vr) > 1 else 0.0
|
| 481 |
+
fs = float(np.std(vf, ddof=1)) if len(vf) > 1 else 0.0
|
| 482 |
+
mean = float(np.nanmean([rm, fm])) if (not np.isnan(rm) or not np.isnan(fm)) else float("nan")
|
| 483 |
+
if n == 192 and not np.isnan(mean): plateau_means.append(mean)
|
| 484 |
+
lines.append(f"{n:<5d} | {rm*100:5.1f}%+/-{rs*100:.1f} | "
|
| 485 |
+
f"{fm*100:5.1f}%+/-{fs*100:.1f} | "
|
| 486 |
+
f"{mean*100:5.1f}%")
|
| 487 |
+
|
| 488 |
+
lines.append("")
|
| 489 |
+
lines.append("REFERENCE:")
|
| 490 |
+
lines.append(" Discrete bottleneck plateau: ~46%")
|
| 491 |
+
lines.append(" Continuous single-prop (Exp M): 51.2%")
|
| 492 |
+
lines.append(" Linear probe at N=192: ~73%")
|
| 493 |
+
lines.append(" Oracle one-hot (Exp A): 100.0%")
|
| 494 |
+
|
| 495 |
+
plateau = float(np.mean(plateau_means)) if plateau_means else float("nan")
|
| 496 |
+
lines.append("")
|
| 497 |
+
lines.append("VERDICT:")
|
| 498 |
+
targets_met = []
|
| 499 |
+
if not np.isnan(pd_) and pd_ >= 0.5: targets_met.append(f"PosDis={pd_:.2f} >= 0.5 [yes]")
|
| 500 |
+
elif not np.isnan(pd_): targets_met.append(f"PosDis={pd_:.2f} < 0.5 [no]")
|
| 501 |
+
if not np.isnan(cs_max) and cs_max >= 0.5: targets_met.append(f"CausalSpec={cs_max:.2f} >= 0.5 [yes]")
|
| 502 |
+
elif not np.isnan(cs_max): targets_met.append(f"CausalSpec={cs_max:.2f} < 0.5 [no]")
|
| 503 |
+
if not np.isnan(plateau) and plateau <= 0.55: targets_met.append(f"Cross plateau={plateau*100:.1f}% <= 55% [yes]")
|
| 504 |
+
elif not np.isnan(plateau): targets_met.append(f"Cross plateau={plateau*100:.1f}% > 55% [no]")
|
| 505 |
+
for line in targets_met:
|
| 506 |
+
lines.append(f" {line}")
|
| 507 |
+
|
| 508 |
+
n_yes = sum(1 for s in targets_met if "[yes]" in s)
|
| 509 |
+
if n_yes >= 3:
|
| 510 |
+
lines.append("")
|
| 511 |
+
lines.append("ALL THREE TARGETS MET. The compositionality-without-invariance dissociation")
|
| 512 |
+
lines.append("holds across BOTH discrete and continuous codes, with high TopSim AND")
|
| 513 |
+
lines.append("high PosDis AND high CausalSpec. Abstract claim defensible.")
|
| 514 |
+
elif n_yes == 2:
|
| 515 |
+
lines.append("")
|
| 516 |
+
lines.append("PARTIAL: 2 of 3 targets met. Most of the abstract claim survives.")
|
| 517 |
+
else:
|
| 518 |
+
lines.append("")
|
| 519 |
+
lines.append("LIMITED: only 1 of 3 targets met. Continuous codes do not achieve the")
|
| 520 |
+
lines.append("same factorization metrics as discrete codes; abstract claim must be")
|
| 521 |
+
lines.append("scoped to discrete codes for PosDis/CausalSpec.")
|
| 522 |
+
|
| 523 |
+
lines.append("")
|
| 524 |
+
lines.append(f"Total runtime: {(time.time()-t0)/60:.1f} min")
|
| 525 |
+
|
| 526 |
+
summary = "\n".join(lines)
|
| 527 |
+
(OUT / "exp_n_summary.txt").write_text(summary + "\n")
|
| 528 |
+
(OUT / "exp_n_summary.json").write_text(json.dumps({
|
| 529 |
+
"config": {"code_dim_per_agent": CODE_DIM, "n_seeds": N_SEEDS,
|
| 530 |
+
"N_list": N_LIST},
|
| 531 |
+
"within": within_combined,
|
| 532 |
+
"best_seed": best_idx,
|
| 533 |
+
"metrics": {"topsim": ts, "posdis": pd_, "causal_spec_max": cs_max,
|
| 534 |
+
"acc_mass": acc_mass, "acc_restit": acc_rest,
|
| 535 |
+
"mi_matrix": mi_matrix.tolist() if mi_matrix is not None else None},
|
| 536 |
+
"cross_results": {d: {str(n): v for n, v in c.items()} for d, c in cross_results.items()},
|
| 537 |
+
"runtime_s": time.time() - t0,
|
| 538 |
+
}, indent=2, default=str))
|
| 539 |
+
print("\n" + summary, flush=True)
|
| 540 |
+
log(f"DONE in {(time.time()-t0)/60:.1f} min")
|
| 541 |
+
|
| 542 |
+
|
| 543 |
+
if __name__ == "__main__":
|
| 544 |
+
main()
|
code/_rev_phys101_bottleneck_n192.py
ADDED
|
@@ -0,0 +1,136 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP REV-P101-BN-N192: Test bottleneck on Phys101 cross-scenario at N=192.
|
| 3 |
+
|
| 4 |
+
The original Phys101 experiment (P3) reported bottleneck cross-scenario at 16-shot
|
| 5 |
+
(~45%). The new LP diagnostic shows LP at N=192 reaches 74-79% on Phys101.
|
| 6 |
+
This script trains the bottleneck at N=192 to test whether the dissociation
|
| 7 |
+
replicates at matched N (the natural comparison for the Kubric N=192 numbers).
|
| 8 |
+
|
| 9 |
+
5 seeds, both per-scenario and global tertile binning.
|
| 10 |
+
"""
|
| 11 |
+
import json, time, sys, os
|
| 12 |
+
from pathlib import Path
|
| 13 |
+
from datetime import datetime, timezone
|
| 14 |
+
import numpy as np
|
| 15 |
+
import torch
|
| 16 |
+
|
| 17 |
+
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
|
| 18 |
+
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
|
| 19 |
+
T0 = time.time()
|
| 20 |
+
|
| 21 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 22 |
+
from _overnight_p1_transfer import (
|
| 23 |
+
train_base, train_receiver_frozen_sender, make_splits, N_FRAMES_SUBSAMPLE,
|
| 24 |
+
)
|
| 25 |
+
|
| 26 |
+
OUT = Path("results/reviewer_response/exp_phys101_bn_n192")
|
| 27 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 28 |
+
N_SEEDS = 5
|
| 29 |
+
N_TARGET = 192
|
| 30 |
+
DOMAINS = ("spring", "fall", "ramp")
|
| 31 |
+
PHYS_FILES = {s: f"results/phase87_phys101_{s}_features.pt" for s in DOMAINS}
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
def log(msg):
|
| 35 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 36 |
+
print(f"[{ts}] EXP-P101BN: {msg}", flush=True)
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
def load_phys(s, mass_to_label):
|
| 40 |
+
"""Load features + apply provided mass->label function."""
|
| 41 |
+
d = torch.load(PHYS_FILES[s], weights_only=False, map_location="cpu")
|
| 42 |
+
feat = d["features"].float()
|
| 43 |
+
T = feat.shape[1]
|
| 44 |
+
if T >= N_FRAMES_SUBSAMPLE:
|
| 45 |
+
idx = np.linspace(0, T-1, N_FRAMES_SUBSAMPLE).astype(int)
|
| 46 |
+
feat = feat[:, idx, :].contiguous()
|
| 47 |
+
mass = np.asarray(d["mass_values"], dtype=np.float64)
|
| 48 |
+
labels = mass_to_label(mass).astype(np.int64)
|
| 49 |
+
return feat, labels, mass
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
def main():
|
| 53 |
+
log("=" * 60)
|
| 54 |
+
log(f"EXP P101 BN N={N_TARGET}: bottleneck on Phys101 at matched N")
|
| 55 |
+
|
| 56 |
+
# First gather all masses for global tertile
|
| 57 |
+
all_masses = []
|
| 58 |
+
for s in DOMAINS:
|
| 59 |
+
d = torch.load(PHYS_FILES[s], weights_only=False, map_location="cpu")
|
| 60 |
+
all_masses.append(np.asarray(d["mass_values"], dtype=np.float64))
|
| 61 |
+
all_mass = np.concatenate(all_masses)
|
| 62 |
+
global_edges = np.quantile(all_mass, [1/3, 2/3])
|
| 63 |
+
log(f"Global tertile edges: {global_edges.tolist()}")
|
| 64 |
+
|
| 65 |
+
pairs = [(src, tgt) for src in DOMAINS for tgt in DOMAINS if src != tgt]
|
| 66 |
+
|
| 67 |
+
out = {"per_scenario": {}, "global": {}}
|
| 68 |
+
for binning_name in ["per_scenario", "global"]:
|
| 69 |
+
log(f"\n=== {binning_name.upper()} BINNING ===")
|
| 70 |
+
# Build mass->label function for this binning
|
| 71 |
+
if binning_name == "per_scenario":
|
| 72 |
+
# Per-scenario: each scenario gets its own tertile
|
| 73 |
+
data = {}
|
| 74 |
+
for s in DOMAINS:
|
| 75 |
+
d = torch.load(PHYS_FILES[s], weights_only=False, map_location="cpu")
|
| 76 |
+
m = np.asarray(d["mass_values"], dtype=np.float64)
|
| 77 |
+
edges = np.quantile(m, [1/3, 2/3])
|
| 78 |
+
f = lambda x, e=edges: np.searchsorted(e, x)
|
| 79 |
+
data[s] = load_phys(s, f)
|
| 80 |
+
else: # global
|
| 81 |
+
f = lambda x: np.searchsorted(global_edges, x)
|
| 82 |
+
data = {s: load_phys(s, f) for s in DOMAINS}
|
| 83 |
+
|
| 84 |
+
for src in DOMAINS:
|
| 85 |
+
log(f" --- {src} as source ---")
|
| 86 |
+
for seed in range(N_SEEDS):
|
| 87 |
+
feat_s, lbl_s, _ = data[src]
|
| 88 |
+
t0 = time.time()
|
| 89 |
+
try:
|
| 90 |
+
base = train_base(feat_s, lbl_s, seed, n_epochs=150)
|
| 91 |
+
log(f" {src} s{seed}: within={base['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 92 |
+
except Exception as e:
|
| 93 |
+
log(f" {src} s{seed} train FAILED: {e}")
|
| 94 |
+
continue
|
| 95 |
+
for tgt in DOMAINS:
|
| 96 |
+
if tgt == src:
|
| 97 |
+
continue
|
| 98 |
+
feat_t, lbl_t, _ = data[tgt]
|
| 99 |
+
tr, hoids = make_splits(lbl_t, seed)
|
| 100 |
+
try:
|
| 101 |
+
acc = train_receiver_frozen_sender(
|
| 102 |
+
base, feat_t, lbl_t, tr, hoids, seed,
|
| 103 |
+
max_examples=N_TARGET, n_epochs=80)
|
| 104 |
+
except Exception as e:
|
| 105 |
+
log(f" {src}->{tgt} s{seed} FAILED: {e}")
|
| 106 |
+
acc = float("nan")
|
| 107 |
+
key = f"{src}->{tgt}"
|
| 108 |
+
out[binning_name].setdefault(key, []).append(float(acc))
|
| 109 |
+
log(f" {src}->{tgt} s{seed} N=192: {acc*100:.1f}%")
|
| 110 |
+
|
| 111 |
+
# Aggregate
|
| 112 |
+
SUMMARY = [f"Phys101 cross-scenario BOTTLENECK at N={N_TARGET} (5 seeds, mean across 6 directional pairs)",
|
| 113 |
+
""]
|
| 114 |
+
for binning_name in ["per_scenario", "global"]:
|
| 115 |
+
all_accs = [a for accs in out[binning_name].values() for a in accs if not np.isnan(a)]
|
| 116 |
+
if all_accs:
|
| 117 |
+
m = np.mean(all_accs); sd = np.std(all_accs, ddof=1)
|
| 118 |
+
SUMMARY.append(f"--- {binning_name} ---")
|
| 119 |
+
SUMMARY.append(f" Mean across pairs: {m*100:5.1f}% +/- {sd*100:.1f}%")
|
| 120 |
+
for pair, accs in out[binning_name].items():
|
| 121 |
+
v = [a for a in accs if not np.isnan(a)]
|
| 122 |
+
if v:
|
| 123 |
+
SUMMARY.append(f" {pair}: {np.mean(v)*100:5.1f}% +/- {np.std(v, ddof=1) if len(v) > 1 else 0.0:.1f}%")
|
| 124 |
+
SUMMARY.append("")
|
| 125 |
+
print("\n".join(SUMMARY), flush=True)
|
| 126 |
+
with open(OUT / "exp_phys101_bn_n192_summary.txt", "w") as fh:
|
| 127 |
+
fh.write("\n".join(SUMMARY) + "\n")
|
| 128 |
+
with open(OUT / "exp_phys101_bn_n192_summary.json", "w") as fh:
|
| 129 |
+
json.dump(out, fh, indent=2)
|
| 130 |
+
end_ts = datetime.now(timezone.utc).isoformat()
|
| 131 |
+
runtime_min = (time.time() - T0) / 60.0
|
| 132 |
+
print(f"\nEND_TIME = {end_ts}\nTotal runtime: {runtime_min:.2f} min", flush=True)
|
| 133 |
+
|
| 134 |
+
|
| 135 |
+
if __name__ == "__main__":
|
| 136 |
+
main()
|
code/_rev_phys101_global_bins.py
ADDED
|
@@ -0,0 +1,208 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP REV-P101-GLOBAL: Diagnose the Phys101 LP/bottleneck dissociation collapse.
|
| 3 |
+
|
| 4 |
+
The paper currently bins Phys101 mass per-scenario tertile (source and target use
|
| 5 |
+
DIFFERENT class boundaries). This is the standard Phys101 protocol but creates a
|
| 6 |
+
class-boundary shift on top of the feature-distribution shift, which may explain
|
| 7 |
+
why the Kubric LP/bottleneck dissociation does not replicate.
|
| 8 |
+
|
| 9 |
+
This script re-runs Phys101 cross-scenario with GLOBAL mass tertile binning
|
| 10 |
+
(unified bin edges across all 3 subsets) and compares to per-scenario binning.
|
| 11 |
+
If LP recovers above the bottleneck under global binning, per-scenario binning
|
| 12 |
+
was the confound. If LP and bottleneck still both sit at ~45%, something else
|
| 13 |
+
(real-video feature variance, smaller pool size) is the cause.
|
| 14 |
+
"""
|
| 15 |
+
import json, time, sys, os
|
| 16 |
+
from pathlib import Path
|
| 17 |
+
from datetime import datetime, timezone
|
| 18 |
+
import numpy as np
|
| 19 |
+
import torch
|
| 20 |
+
from sklearn.linear_model import LogisticRegression
|
| 21 |
+
from sklearn.preprocessing import StandardScaler
|
| 22 |
+
|
| 23 |
+
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
|
| 24 |
+
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
|
| 25 |
+
T0 = time.time()
|
| 26 |
+
|
| 27 |
+
OUT = Path("results/reviewer_response/exp_phys101_global_bins")
|
| 28 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 29 |
+
N_SEEDS = 5
|
| 30 |
+
N_LIST = [16, 64, 192]
|
| 31 |
+
DOMAINS = ("spring", "fall", "ramp")
|
| 32 |
+
PHYS_FILES = {s: f"results/phase87_phys101_{s}_features.pt" for s in DOMAINS}
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
def log(msg):
|
| 36 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 37 |
+
print(f"[{ts}] EXP-P101G: {msg}", flush=True)
|
| 38 |
+
|
| 39 |
+
|
| 40 |
+
def load_data():
|
| 41 |
+
"""Load features and mass values for all three scenarios."""
|
| 42 |
+
out = {}
|
| 43 |
+
for s in DOMAINS:
|
| 44 |
+
d = torch.load(PHYS_FILES[s], weights_only=False, map_location="cpu")
|
| 45 |
+
feat = d["features"].float()
|
| 46 |
+
# L2-pool over time (8 frames -> 1024-d feature)
|
| 47 |
+
pooled = feat.mean(dim=1).numpy()
|
| 48 |
+
mass = np.asarray(d["mass_values"], dtype=np.float64)
|
| 49 |
+
out[s] = (pooled, mass)
|
| 50 |
+
log(f" {s}: pooled feat={pooled.shape}, mass=[{mass.min():.1f}, {mass.max():.1f}], n={len(mass)}")
|
| 51 |
+
return out
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
def bin_global(mass_dict):
|
| 55 |
+
"""Global tertile binning across all three subsets pooled."""
|
| 56 |
+
all_mass = np.concatenate([m for _, m in mass_dict.values()])
|
| 57 |
+
edges = np.quantile(all_mass, [1/3, 2/3])
|
| 58 |
+
log(f" GLOBAL tertile edges (computed on union of {len(all_mass)} clips): {edges.tolist()}")
|
| 59 |
+
return {s: (np.searchsorted(edges, m).astype(np.int64), edges)
|
| 60 |
+
for s, (_, m) in mass_dict.items()}
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
def bin_per_scenario(mass_dict):
|
| 64 |
+
"""Per-scenario tertile binning (the standard Phys101 protocol)."""
|
| 65 |
+
out = {}
|
| 66 |
+
for s, (_, m) in mass_dict.items():
|
| 67 |
+
edges = np.quantile(m, [1/3, 2/3])
|
| 68 |
+
out[s] = (np.searchsorted(edges, m).astype(np.int64), edges)
|
| 69 |
+
log(f" PER-SCEN {s} edges: {edges.tolist()}, bins={np.bincount(out[s][0], minlength=3).tolist()}")
|
| 70 |
+
return out
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
def stratified_subset(rng, y, n_per_class):
|
| 74 |
+
idxs = []
|
| 75 |
+
for c in np.unique(y):
|
| 76 |
+
cand = np.where(y == c)[0]
|
| 77 |
+
if len(cand) == 0:
|
| 78 |
+
continue
|
| 79 |
+
chosen = rng.choice(cand, size=min(n_per_class, len(cand)), replace=False)
|
| 80 |
+
idxs.extend(chosen.tolist())
|
| 81 |
+
return np.array(sorted(idxs))
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
def train_lp(X_tr, y_tr, X_te, y_te):
|
| 85 |
+
sc = StandardScaler().fit(X_tr)
|
| 86 |
+
Xs_tr = sc.transform(X_tr)
|
| 87 |
+
Xs_te = sc.transform(X_te)
|
| 88 |
+
model = LogisticRegression(max_iter=2000, C=1.0, solver="lbfgs")
|
| 89 |
+
model.fit(Xs_tr, y_tr)
|
| 90 |
+
return float((model.predict(Xs_te) == y_te).mean())
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
def stats(vals):
|
| 94 |
+
v = np.array([x for x in vals if not np.isnan(x)])
|
| 95 |
+
if len(v) == 0:
|
| 96 |
+
return float("nan"), float("nan")
|
| 97 |
+
return float(v.mean()), float(v.std(ddof=1) if len(v) > 1 else 0.0)
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
def evaluate_lp(features, labels_dict, src, tgt, n_classes=3):
|
| 101 |
+
"""LP cross-scenario: train on source, train fresh classifier with N target samples,
|
| 102 |
+
eval on remaining target samples."""
|
| 103 |
+
X_src = features[src]
|
| 104 |
+
y_src = labels_dict[src][0]
|
| 105 |
+
X_tgt = features[tgt]
|
| 106 |
+
y_tgt = labels_dict[tgt][0]
|
| 107 |
+
|
| 108 |
+
smallest = min(int(np.sum(y_tgt == c)) for c in np.unique(y_tgt))
|
| 109 |
+
results = {N: [] for N in N_LIST}
|
| 110 |
+
n0_results = []
|
| 111 |
+
|
| 112 |
+
# N=0: train on source only, eval on target
|
| 113 |
+
for s in range(N_SEEDS):
|
| 114 |
+
try:
|
| 115 |
+
acc = train_lp(X_src, y_src, X_tgt, y_tgt)
|
| 116 |
+
n0_results.append(acc)
|
| 117 |
+
except Exception as e:
|
| 118 |
+
log(f" {src}->{tgt} N=0 s{s}: FAILED {e}")
|
| 119 |
+
n0_results.append(float("nan"))
|
| 120 |
+
|
| 121 |
+
# N>0: train on source + N stratified target, eval on remaining target
|
| 122 |
+
for N in N_LIST:
|
| 123 |
+
per_class = max(1, N // n_classes)
|
| 124 |
+
per_class = min(per_class, int(0.7 * smallest))
|
| 125 |
+
for s in range(N_SEEDS):
|
| 126 |
+
rng = np.random.default_rng(1234 + s)
|
| 127 |
+
tgt_idx_train = stratified_subset(rng, y_tgt, per_class)
|
| 128 |
+
mask = np.ones(len(y_tgt), bool); mask[tgt_idx_train] = False
|
| 129 |
+
X_eval = X_tgt[mask]; y_eval = y_tgt[mask]
|
| 130 |
+
if len(y_eval) == 0:
|
| 131 |
+
continue
|
| 132 |
+
X_tr = np.concatenate([X_src, X_tgt[tgt_idx_train]], axis=0)
|
| 133 |
+
y_tr = np.concatenate([y_src, y_tgt[tgt_idx_train]], axis=0)
|
| 134 |
+
try:
|
| 135 |
+
acc = train_lp(X_tr, y_tr, X_eval, y_eval)
|
| 136 |
+
results[N].append(acc)
|
| 137 |
+
except Exception as e:
|
| 138 |
+
log(f" {src}->{tgt} N={N} s{s}: FAILED {e}")
|
| 139 |
+
results[N].append(float("nan"))
|
| 140 |
+
return n0_results, results
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
def main():
|
| 144 |
+
log("=" * 60)
|
| 145 |
+
log("Phys101 cross-scenario diagnostic: global vs per-scenario binning")
|
| 146 |
+
|
| 147 |
+
raw = load_data()
|
| 148 |
+
features = {s: raw[s][0] for s in DOMAINS}
|
| 149 |
+
|
| 150 |
+
# Per-scenario binning (standard, current paper protocol)
|
| 151 |
+
log("\n=== PER-SCENARIO TERTILE BINNING (paper default) ===")
|
| 152 |
+
perscen_labels = bin_per_scenario(raw)
|
| 153 |
+
|
| 154 |
+
# Global binning (diagnostic)
|
| 155 |
+
log("\n=== GLOBAL TERTILE BINNING (diagnostic) ===")
|
| 156 |
+
global_labels = bin_global(raw)
|
| 157 |
+
|
| 158 |
+
# Evaluate all 6 cross-scenario directions under both binnings
|
| 159 |
+
pairs = [(src, tgt) for src in DOMAINS for tgt in DOMAINS if src != tgt]
|
| 160 |
+
|
| 161 |
+
out = {"per_scenario": {}, "global": {}}
|
| 162 |
+
for binning_name, label_dict in [("per_scenario", perscen_labels),
|
| 163 |
+
("global", global_labels)]:
|
| 164 |
+
log(f"\n--- {binning_name} binning ---")
|
| 165 |
+
for src, tgt in pairs:
|
| 166 |
+
log(f" LP {src}->{tgt}:")
|
| 167 |
+
n0, results = evaluate_lp(features, label_dict, src, tgt)
|
| 168 |
+
n0_m, n0_s = stats(n0)
|
| 169 |
+
log(f" N=0 (src-only): {n0_m*100:5.1f}% +/- {n0_s*100:.1f}%")
|
| 170 |
+
for N in N_LIST:
|
| 171 |
+
m, sd = stats(results[N])
|
| 172 |
+
log(f" N={N:>3d}: {m*100:5.1f}% +/- {sd*100:.1f}%")
|
| 173 |
+
out[binning_name][f"{src}->{tgt}"] = {
|
| 174 |
+
"N0": [float(x) for x in n0],
|
| 175 |
+
"curve": {N: [float(x) for x in results[N]] for N in N_LIST},
|
| 176 |
+
}
|
| 177 |
+
|
| 178 |
+
# Aggregate means
|
| 179 |
+
def mean_across_pairs(binning):
|
| 180 |
+
rows = out[binning]
|
| 181 |
+
n0_all = [x for r in rows.values() for x in r["N0"] if not np.isnan(x)]
|
| 182 |
+
per_N = {N: [x for r in rows.values() for x in r["curve"][N] if not np.isnan(x)]
|
| 183 |
+
for N in N_LIST}
|
| 184 |
+
return {"N0": stats(n0_all), **{f"N{N}": stats(per_N[N]) for N in N_LIST}}
|
| 185 |
+
|
| 186 |
+
SUMMARY = ["Phys101 cross-scenario LP, per-scenario vs global tertile mass binning",
|
| 187 |
+
"(5 seeds, mean across 6 directional pairs, +/- = std across all seeds*pairs)",
|
| 188 |
+
""]
|
| 189 |
+
for binning_name in ["per_scenario", "global"]:
|
| 190 |
+
agg = mean_across_pairs(binning_name)
|
| 191 |
+
SUMMARY.append(f"--- {binning_name} ---")
|
| 192 |
+
for k in ["N0", "N16", "N64", "N192"]:
|
| 193 |
+
m, sd = agg[k]
|
| 194 |
+
SUMMARY.append(f" {k:>5s}: {m*100:5.1f}% +/- {sd*100:.1f}%")
|
| 195 |
+
SUMMARY.append("")
|
| 196 |
+
|
| 197 |
+
print("\n" + "\n".join(SUMMARY), flush=True)
|
| 198 |
+
with open(OUT / "exp_phys101_global_bins_summary.txt", "w") as fh:
|
| 199 |
+
fh.write("\n".join(SUMMARY) + "\n")
|
| 200 |
+
with open(OUT / "exp_phys101_global_bins_summary.json", "w") as fh:
|
| 201 |
+
json.dump(out, fh, indent=2)
|
| 202 |
+
end_ts = datetime.now(timezone.utc).isoformat()
|
| 203 |
+
runtime_min = (time.time() - T0) / 60.0
|
| 204 |
+
print(f"\nEND_TIME = {end_ts}\nTotal runtime: {runtime_min:.2f} min", flush=True)
|
| 205 |
+
|
| 206 |
+
|
| 207 |
+
if __name__ == "__main__":
|
| 208 |
+
main()
|
code/_rev_q_2nddir_singleprop_3seed.py
ADDED
|
@@ -0,0 +1,152 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP 2ND-DIR-3SEED: 3-seed re-run of the 12 single-property configurations on
|
| 3 |
+
collision -> flat-drop at N=192.
|
| 4 |
+
|
| 5 |
+
R2 + R3 convergent ask: the 1-seed run in `_rev_q_2nddirection_flatdrop.py`
|
| 6 |
+
collapsed all 12 single-property configs to exactly 40.0% (degenerate-receiver
|
| 7 |
+
floor on flat-drop). This replaces those rows with proper 3-seed best-of
|
| 8 |
+
numbers.
|
| 9 |
+
|
| 10 |
+
Single-prop configs (matching the existing 24-config sweep rows 1-12):
|
| 11 |
+
7 disc: L=2..5 x V=5,10 subset (matching the sweep)
|
| 12 |
+
5 cont: D=2,3,5,10,20
|
| 13 |
+
|
| 14 |
+
Multi-prop rows (the original 12 multi-prop configs in the 2nd-direction sweep)
|
| 15 |
+
already at 45-58% with 1 seed; they are not the source of the 40% floor and
|
| 16 |
+
re-running them would not move the headline.
|
| 17 |
+
"""
|
| 18 |
+
import json, time, sys, os
|
| 19 |
+
from pathlib import Path
|
| 20 |
+
from datetime import datetime, timezone
|
| 21 |
+
import numpy as np
|
| 22 |
+
import torch
|
| 23 |
+
|
| 24 |
+
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
|
| 25 |
+
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
|
| 26 |
+
T0 = time.time()
|
| 27 |
+
|
| 28 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 29 |
+
from _overnight_p1_transfer import make_splits
|
| 30 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 31 |
+
from _rev_q_posdis_scatter import (
|
| 32 |
+
train_discrete_custom, disc_train_recv_custom,
|
| 33 |
+
train_continuous_base, train_recv_frozen_cont,
|
| 34 |
+
)
|
| 35 |
+
disc_train_recv_frozen = disc_train_recv_custom # alias
|
| 36 |
+
|
| 37 |
+
OUT = Path("results/reviewer_response/exp_2nddir_singleprop_3seed")
|
| 38 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 39 |
+
N_SEEDS = 3
|
| 40 |
+
N_TARGET = 192
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
def log(msg):
|
| 44 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 45 |
+
print(f"[{ts}] EXP-3SEED: {msg}", flush=True)
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
def main():
|
| 49 |
+
log("=" * 60)
|
| 50 |
+
log(f"3-seed re-run: 12 single-property configs on coll -> flat-drop @ N={N_TARGET}")
|
| 51 |
+
|
| 52 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 53 |
+
feat_t = load_feat_subsampled("flat_drop", "vjepa2")
|
| 54 |
+
rest_3 = np.load("results/kinematics_vs_mechanics/labels_collision.npz")["restitution_bin"]
|
| 55 |
+
lbl_t_3 = load_labels("flat_drop", "restitution")
|
| 56 |
+
|
| 57 |
+
# 7 single-prop disc + 5 single-prop cont configs (matching sweep rows 1-12)
|
| 58 |
+
disc_configs = [
|
| 59 |
+
("disc_L2_V5", 2, 5),
|
| 60 |
+
("disc_L2_V10", 2, 10),
|
| 61 |
+
("disc_L3_V5", 3, 5),
|
| 62 |
+
("disc_L3_V10", 3, 10),
|
| 63 |
+
("disc_L4_V5", 4, 5),
|
| 64 |
+
("disc_L4_V10", 4, 10),
|
| 65 |
+
("disc_L5_V5", 5, 5),
|
| 66 |
+
]
|
| 67 |
+
cont_configs = [
|
| 68 |
+
("cont_dim2", 2),
|
| 69 |
+
("cont_dim3", 3),
|
| 70 |
+
("cont_dim5", 5),
|
| 71 |
+
("cont_dim10", 10),
|
| 72 |
+
("cont_dim20", 20),
|
| 73 |
+
]
|
| 74 |
+
|
| 75 |
+
rows = []
|
| 76 |
+
|
| 77 |
+
# Discrete configs
|
| 78 |
+
for name, L, V in disc_configs:
|
| 79 |
+
log(f"\n --- {name} (L={L}, V={V}) ---")
|
| 80 |
+
within_seeds = []; cross_seeds = []
|
| 81 |
+
for seed in range(N_SEEDS):
|
| 82 |
+
t0 = time.time()
|
| 83 |
+
try:
|
| 84 |
+
base = train_discrete_custom(feat_c, rest_3, seed=seed, n_heads=L, vocab_size=V, n_epochs=150)
|
| 85 |
+
tr_t, ho_t = make_splits(lbl_t_3, seed)
|
| 86 |
+
acc = disc_train_recv_frozen(base, feat_t, lbl_t_3, tr_t, ho_t, seed=seed, n_target=N_TARGET)
|
| 87 |
+
within_seeds.append(float(base["task_acc"]))
|
| 88 |
+
cross_seeds.append(float(acc))
|
| 89 |
+
log(f" s{seed}: within={base['task_acc']*100:.1f}%, cross={acc*100:.1f}% [{time.time()-t0:.0f}s]")
|
| 90 |
+
except Exception as e:
|
| 91 |
+
import traceback
|
| 92 |
+
log(f" s{seed} FAILED: {e}\n{traceback.format_exc()[:300]}")
|
| 93 |
+
if within_seeds:
|
| 94 |
+
rows.append({"name": name, "kind": "disc", "L": L, "V": V,
|
| 95 |
+
"within_mean": float(np.mean(within_seeds)), "within_std": float(np.std(within_seeds)),
|
| 96 |
+
"within_max": float(np.max(within_seeds)),
|
| 97 |
+
"cross_n192_mean": float(np.mean(cross_seeds)), "cross_n192_std": float(np.std(cross_seeds)),
|
| 98 |
+
"cross_n192_max": float(np.max(cross_seeds))})
|
| 99 |
+
|
| 100 |
+
# Continuous configs
|
| 101 |
+
for name, D in cont_configs:
|
| 102 |
+
log(f"\n --- {name} (D={D}) ---")
|
| 103 |
+
within_seeds = []; cross_seeds = []
|
| 104 |
+
for seed in range(N_SEEDS):
|
| 105 |
+
t0 = time.time()
|
| 106 |
+
try:
|
| 107 |
+
base = train_continuous_base(feat_c, rest_3, seed=seed, code_dim_per_agent=D, n_epochs=150)
|
| 108 |
+
tr_t, ho_t = make_splits(lbl_t_3, seed)
|
| 109 |
+
acc = train_recv_frozen_cont(base, feat_t, lbl_t_3, tr_t, ho_t, seed=seed, n_target=N_TARGET)
|
| 110 |
+
within_seeds.append(float(base["task_acc"]))
|
| 111 |
+
cross_seeds.append(float(acc))
|
| 112 |
+
log(f" s{seed}: within={base['task_acc']*100:.1f}%, cross={acc*100:.1f}% [{time.time()-t0:.0f}s]")
|
| 113 |
+
except Exception as e:
|
| 114 |
+
import traceback
|
| 115 |
+
log(f" s{seed} FAILED: {e}\n{traceback.format_exc()[:300]}")
|
| 116 |
+
if within_seeds:
|
| 117 |
+
rows.append({"name": name, "kind": "cont", "D": D,
|
| 118 |
+
"within_mean": float(np.mean(within_seeds)), "within_std": float(np.std(within_seeds)),
|
| 119 |
+
"within_max": float(np.max(within_seeds)),
|
| 120 |
+
"cross_n192_mean": float(np.mean(cross_seeds)), "cross_n192_std": float(np.std(cross_seeds)),
|
| 121 |
+
"cross_n192_max": float(np.max(cross_seeds))})
|
| 122 |
+
|
| 123 |
+
if rows:
|
| 124 |
+
SUMMARY = ["EXP 3-SEED single-prop coll->flat-drop @ N=192",
|
| 125 |
+
"",
|
| 126 |
+
f"{'Config':<14s} | {'Within (mean+-std)':>20s} | {'Cross (mean+-std)':>20s} | {'Cross max':>10s}",
|
| 127 |
+
"-" * 75]
|
| 128 |
+
for r in rows:
|
| 129 |
+
SUMMARY.append(
|
| 130 |
+
f"{r['name']:<14s} | {r['within_mean']*100:>6.1f}+-{r['within_std']*100:>4.1f}% | "
|
| 131 |
+
f"{r['cross_n192_mean']*100:>6.1f}+-{r['cross_n192_std']*100:>4.1f}% | "
|
| 132 |
+
f"{r['cross_n192_max']*100:>9.1f}%"
|
| 133 |
+
)
|
| 134 |
+
cross_means = [r["cross_n192_mean"] for r in rows]
|
| 135 |
+
cross_maxes = [r["cross_n192_max"] for r in rows]
|
| 136 |
+
SUMMARY.append("")
|
| 137 |
+
SUMMARY.append(f"All-config 3-seed mean cross flat-drop: {np.mean(cross_means)*100:.1f}+-{np.std(cross_means)*100:.1f}% (range {np.min(cross_means)*100:.1f}-{np.max(cross_means)*100:.1f}%)")
|
| 138 |
+
SUMMARY.append(f"All-config best-of-3 cross flat-drop: {np.mean(cross_maxes)*100:.1f}+-{np.std(cross_maxes)*100:.1f}% (range {np.min(cross_maxes)*100:.1f}-{np.max(cross_maxes)*100:.1f}%)")
|
| 139 |
+
SUMMARY.append("")
|
| 140 |
+
SUMMARY.append("Prior 1-seed reported all 12 configs at exactly 40.0% (degenerate-receiver floor).")
|
| 141 |
+
print("\n".join(SUMMARY), flush=True)
|
| 142 |
+
with open(OUT / "summary.txt", "w") as fh:
|
| 143 |
+
fh.write("\n".join(SUMMARY) + "\n")
|
| 144 |
+
with open(OUT / "summary.json", "w") as fh:
|
| 145 |
+
json.dump(rows, fh, indent=2)
|
| 146 |
+
end_ts = datetime.now(timezone.utc).isoformat()
|
| 147 |
+
runtime_min = (time.time() - T0) / 60.0
|
| 148 |
+
print(f"\nEND_TIME = {end_ts}\nTotal runtime: {runtime_min:.2f} min", flush=True)
|
| 149 |
+
|
| 150 |
+
|
| 151 |
+
if __name__ == "__main__":
|
| 152 |
+
main()
|
code/_rev_q_addendum2_high_posdis.py
ADDED
|
@@ -0,0 +1,343 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP Q ADDENDUM 2: try to legitimately reach PosDis >= 0.7 with V-JEPA 2 multi-
|
| 3 |
+
property training, so the headline 0.76 in Table 1 has a clean provenance from
|
| 4 |
+
this paper's protocol.
|
| 5 |
+
|
| 6 |
+
The Q addendum's discrete multi-prop topped at PosDis 0.51. To push higher we:
|
| 7 |
+
- use 5-class labels (mass {1..5}, restit {0.1..0.9}) instead of 3-class bins
|
| 8 |
+
- train for 400 epochs (vs 150) with iterated-learning receiver resets every
|
| 9 |
+
30 epochs (vs 40)
|
| 10 |
+
- sweep (L, V) in {(2,5), (3,5), (4,5)}
|
| 11 |
+
|
| 12 |
+
For each successful config, evaluate cross-scenario coll->ramp at N=16, N=192
|
| 13 |
+
on restitution 3-class (matches the rest of the paper's cross protocol).
|
| 14 |
+
"""
|
| 15 |
+
import json, time, sys, os, math
|
| 16 |
+
from pathlib import Path
|
| 17 |
+
from datetime import datetime, timezone
|
| 18 |
+
import numpy as np
|
| 19 |
+
import torch
|
| 20 |
+
import torch.nn as nn
|
| 21 |
+
import torch.nn.functional as F
|
| 22 |
+
|
| 23 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 24 |
+
from _kinematics_train import (
|
| 25 |
+
DEVICE, ClassifierReceiver,
|
| 26 |
+
HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR,
|
| 27 |
+
EARLY_STOP_PATIENCE,
|
| 28 |
+
)
|
| 29 |
+
from _killer_experiment import TemporalEncoder, DiscreteSender, DiscreteMultiSender
|
| 30 |
+
from _overnight_p1_transfer import make_splits
|
| 31 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 32 |
+
from _rev_f_cnn_control import ci95
|
| 33 |
+
from _rev_q_posdis_scatter import build_discrete_sender, discrete_token_extract
|
| 34 |
+
from _rev_n_multiprop_continuous import MultiPropReceiver
|
| 35 |
+
from _rev_q_addendum_multiprop import (
|
| 36 |
+
discrete_multi_topsim, discrete_multi_posdis, discrete_multi_causal,
|
| 37 |
+
disc_multi_train_recv_frozen,
|
| 38 |
+
)
|
| 39 |
+
|
| 40 |
+
OUT = Path("results/reviewer_response/exp_q_addendum2")
|
| 41 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 42 |
+
N_SEEDS = 3
|
| 43 |
+
N_LIST = [16, 192]
|
| 44 |
+
RESET_EVERY = 30
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
def log(msg):
|
| 48 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 49 |
+
print(f"[{ts}] EXP-QADD2: {msg}", flush=True)
|
| 50 |
+
|
| 51 |
+
|
| 52 |
+
def train_disc_multi_5class(feat, labels_list, seed, n_heads, vocab_size,
|
| 53 |
+
n_epochs=400):
|
| 54 |
+
"""Discrete multi-prop with aggressive iterated learning and 5-class labels."""
|
| 55 |
+
N, nf, dim = feat.shape
|
| 56 |
+
fpa = 1
|
| 57 |
+
msg_dim = vocab_size * n_heads * N_AGENTS
|
| 58 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 59 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 60 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 61 |
+
|
| 62 |
+
primary = labels_list[0]
|
| 63 |
+
train_ids, holdout_ids = [], []
|
| 64 |
+
for c in np.unique(primary):
|
| 65 |
+
ids_c = np.where(primary == c)[0]
|
| 66 |
+
rng.shuffle(ids_c)
|
| 67 |
+
split = max(1, len(ids_c) // 5)
|
| 68 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 69 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 70 |
+
n_classes_per_prop = [int(lbl.max()) + 1 for lbl in labels_list]
|
| 71 |
+
chance = 1.0 / max(n_classes_per_prop)
|
| 72 |
+
|
| 73 |
+
sender = build_discrete_sender(dim, n_heads, vocab_size, fpa)
|
| 74 |
+
receivers = [MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE)
|
| 75 |
+
for _ in range(3)]
|
| 76 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 77 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 78 |
+
labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list]
|
| 79 |
+
me = math.log(vocab_size)
|
| 80 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 81 |
+
best_acc = 0.0; best_ep = 0
|
| 82 |
+
best_sender_state = None; best_receiver_states = None; best_recv_idx = 0
|
| 83 |
+
|
| 84 |
+
for ep in range(n_epochs):
|
| 85 |
+
if ep - best_ep > EARLY_STOP_PATIENCE * 2 and best_acc > chance + 0.05: break
|
| 86 |
+
if ep > 0 and ep % RESET_EVERY == 0:
|
| 87 |
+
for i in range(len(receivers)):
|
| 88 |
+
receivers[i] = MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE)
|
| 89 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 90 |
+
sender.train(); [r.train() for r in receivers]
|
| 91 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1)
|
| 92 |
+
hard = ep >= 30
|
| 93 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 94 |
+
perm = rng_ep.permutation(train_ids)
|
| 95 |
+
for b in range(n_batches):
|
| 96 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 97 |
+
if len(batch_ids) < 4: continue
|
| 98 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 99 |
+
tgts = [ld[batch_ids] for ld in labels_dev]
|
| 100 |
+
msg, logits_list = sender(views, tau=tau, hard=hard)
|
| 101 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 102 |
+
for r in receivers:
|
| 103 |
+
head_logits = r(msg)
|
| 104 |
+
for hl, tgt in zip(head_logits, tgts):
|
| 105 |
+
loss = loss + F.cross_entropy(hl, tgt)
|
| 106 |
+
loss = loss / (len(receivers) * len(tgts))
|
| 107 |
+
for lg in logits_list:
|
| 108 |
+
lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8)
|
| 109 |
+
ent = -(p * lp).sum(-1).mean()
|
| 110 |
+
if ent / me < 0.1: loss = loss - 0.03 * ent
|
| 111 |
+
if torch.isnan(loss):
|
| 112 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 113 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 114 |
+
loss.backward()
|
| 115 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 116 |
+
so.step(); [o.step() for o in ros]
|
| 117 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 118 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 119 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 120 |
+
with torch.no_grad():
|
| 121 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 122 |
+
msg_ho, _ = sender(v_ho)
|
| 123 |
+
tgt_ho = [ld[holdout_ids] for ld in labels_dev]
|
| 124 |
+
best_per_recv = 0.0; best_idx = 0
|
| 125 |
+
for ri, r in enumerate(receivers):
|
| 126 |
+
head_logits = r(msg_ho)
|
| 127 |
+
accs = [(hl.argmax(-1) == tgt).float().mean().item()
|
| 128 |
+
for hl, tgt in zip(head_logits, tgt_ho)]
|
| 129 |
+
combined = float(np.mean(accs))
|
| 130 |
+
if combined > best_per_recv:
|
| 131 |
+
best_per_recv = combined; best_idx = ri
|
| 132 |
+
if best_per_recv > best_acc:
|
| 133 |
+
best_acc = best_per_recv; best_ep = ep
|
| 134 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 135 |
+
best_receiver_states = [
|
| 136 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 137 |
+
for r in receivers]
|
| 138 |
+
best_recv_idx = best_idx
|
| 139 |
+
return {
|
| 140 |
+
"sender_state": best_sender_state,
|
| 141 |
+
"receiver_states": best_receiver_states,
|
| 142 |
+
"best_recv_idx": best_recv_idx,
|
| 143 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 144 |
+
"task_acc": best_acc, "chance": chance,
|
| 145 |
+
"n_classes_per_prop": n_classes_per_prop,
|
| 146 |
+
"fpa": 1, "dim": dim,
|
| 147 |
+
"n_heads": n_heads, "vocab_size": vocab_size,
|
| 148 |
+
"msg_dim": msg_dim,
|
| 149 |
+
}
|
| 150 |
+
|
| 151 |
+
|
| 152 |
+
def main():
|
| 153 |
+
t0 = time.time()
|
| 154 |
+
log("=" * 60)
|
| 155 |
+
log("EXP Q ADDENDUM 2: try to reach PosDis >= 0.7 with multi-prop V-JEPA 2")
|
| 156 |
+
|
| 157 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 158 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 159 |
+
z = np.load("results/kinematics_vs_mechanics/labels_collision.npz")
|
| 160 |
+
# 5-class labels: map mass {1..5}->{0..4} and restitution {0.1..0.9}->{0..4}
|
| 161 |
+
mass_5 = (z["mass_scalar"].astype(int) - 1).astype(np.int64)
|
| 162 |
+
rest_levels = [0.1, 0.3, 0.5, 0.7, 0.9]
|
| 163 |
+
rest_5 = np.zeros(len(z["restitution_scalar"]), dtype=np.int64)
|
| 164 |
+
for i, lvl in enumerate(rest_levels):
|
| 165 |
+
rest_5[np.isclose(z["restitution_scalar"], lvl)] = i
|
| 166 |
+
log(f" collision mass(5)={np.bincount(mass_5).tolist()} "
|
| 167 |
+
f"restit(5)={np.bincount(rest_5).tolist()}")
|
| 168 |
+
lbl_r_3 = load_labels("ramp", "restitution") # for cross-scenario (3-class)
|
| 169 |
+
|
| 170 |
+
rows = []
|
| 171 |
+
configs = [(2, 5), (3, 5), (4, 5)]
|
| 172 |
+
for H, V in configs:
|
| 173 |
+
name = f"disc_multi5_L{H}_V{V}"
|
| 174 |
+
log(f"\n --- {name} (5-class mass + 5-class restit, 400 epochs) ---")
|
| 175 |
+
within_accs = []; bases = []
|
| 176 |
+
for seed in range(N_SEEDS):
|
| 177 |
+
t_s = time.time()
|
| 178 |
+
try:
|
| 179 |
+
base = train_disc_multi_5class(feat_c, [mass_5, rest_5], seed, H, V)
|
| 180 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 181 |
+
log(f" {name} s{seed}: combined within={base['task_acc']:.3f} "
|
| 182 |
+
f"[{time.time()-t_s:.0f}s]")
|
| 183 |
+
except Exception as e:
|
| 184 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 185 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 186 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 187 |
+
if not valid: continue
|
| 188 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 189 |
+
best_base = bases[best_idx]
|
| 190 |
+
ho_ids = best_base["holdout_ids"]
|
| 191 |
+
try:
|
| 192 |
+
tokens = discrete_token_extract(best_base, feat_c)
|
| 193 |
+
tokens_ho = tokens[ho_ids]
|
| 194 |
+
ts = discrete_multi_topsim(tokens_ho, [mass_5[ho_ids], rest_5[ho_ids]])
|
| 195 |
+
pd_, _ = discrete_multi_posdis(tokens_ho, [mass_5[ho_ids], rest_5[ho_ids]])
|
| 196 |
+
base_pp, drops = discrete_multi_causal(best_base, feat_c,
|
| 197 |
+
[mass_5, rest_5], ho_ids)
|
| 198 |
+
cs = float(drops.max())
|
| 199 |
+
except Exception as e:
|
| 200 |
+
log(f" metrics FAILED: {e}")
|
| 201 |
+
ts = pd_ = cs = float("nan")
|
| 202 |
+
# Cross-scenario: target = ramp restitution (3-class)
|
| 203 |
+
# Note base trained with 5-class restit, but we transfer the SENDER and
|
| 204 |
+
# train a fresh 3-class receiver, matching the rest of the paper.
|
| 205 |
+
cross = {n: [] for n in N_LIST}
|
| 206 |
+
for seed, base in enumerate(bases):
|
| 207 |
+
if base is None:
|
| 208 |
+
for n in N_LIST: cross[n].append(float("nan"))
|
| 209 |
+
continue
|
| 210 |
+
tr_t, ho_t = make_splits(lbl_r_3, seed)
|
| 211 |
+
for n in N_LIST:
|
| 212 |
+
try:
|
| 213 |
+
acc = disc_multi_train_recv_frozen(base, feat_r, lbl_r_3,
|
| 214 |
+
tr_t, ho_t, seed, n)
|
| 215 |
+
cross[n].append(float(acc))
|
| 216 |
+
except Exception as e:
|
| 217 |
+
log(f" {name} s{seed} N={n} FAILED: {e}")
|
| 218 |
+
cross[n].append(float("nan"))
|
| 219 |
+
wm = float(np.mean([a for a in within_accs if not np.isnan(a)]))
|
| 220 |
+
cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)]))
|
| 221 |
+
if any(not np.isnan(x) for x in cross[n]) else float("nan")
|
| 222 |
+
for n in N_LIST}
|
| 223 |
+
log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 224 |
+
f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}")
|
| 225 |
+
rows.append({
|
| 226 |
+
"name": name, "type": "discrete_multi5",
|
| 227 |
+
"n_heads": H, "vocab_size": V,
|
| 228 |
+
"within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs,
|
| 229 |
+
"cross_n16": cm[16], "cross_n192": cm[192],
|
| 230 |
+
})
|
| 231 |
+
|
| 232 |
+
# ─── Build full sweep table: original 12 + Q-addendum 3 + new 3 ───
|
| 233 |
+
original_12 = [
|
| 234 |
+
("disc_L2_V5", "discrete", 0.88, 0.20, 0.02, 41.7, 43.9),
|
| 235 |
+
("disc_L2_V10", "discrete", 0.84, 0.25, 0.05, 46.1, 41.7),
|
| 236 |
+
("disc_L3_V5", "discrete", 0.84, 0.13, 0.02, 43.3, 42.8),
|
| 237 |
+
("disc_L3_V10", "discrete", 0.84, 0.12, 0.01, 43.3, 45.6),
|
| 238 |
+
("disc_L4_V5", "discrete", 0.90, 0.10, 0.01, 41.1, 42.2),
|
| 239 |
+
("disc_L4_V10", "discrete", 0.82, 0.08, 0.02, 45.0, 45.0),
|
| 240 |
+
("disc_L5_V5", "discrete", 0.89, 0.07, 0.02, 40.0, 43.9),
|
| 241 |
+
("cont_dim2", "continuous", 0.92, 0.15, 0.20, 48.9, 54.4),
|
| 242 |
+
("cont_dim3", "continuous", 0.91, 0.15, 0.02, 40.6, 41.1),
|
| 243 |
+
("cont_dim5", "continuous", 0.89, 0.06, 0.03, 47.2, 43.9),
|
| 244 |
+
("cont_dim10", "continuous", 0.88, 0.04, 0.01, 47.8, 48.3),
|
| 245 |
+
("cont_dim20", "continuous", 0.90, 0.02, 0.00, 48.9, 55.0),
|
| 246 |
+
]
|
| 247 |
+
addendum_3 = [
|
| 248 |
+
("disc_multi_L3_V5", "discrete_multi", 0.59, 0.51, 0.06, 40.0, 46.1),
|
| 249 |
+
("disc_multi_L4_V10", "discrete_multi", 0.68, 0.48, 0.01, 45.6, 50.6),
|
| 250 |
+
("cont_multi_dim3", "continuous_multi", 0.72, 0.40, 0.10, 50.6, 55.0),
|
| 251 |
+
]
|
| 252 |
+
all_rows = list(original_12) + list(addendum_3)
|
| 253 |
+
for r in rows:
|
| 254 |
+
all_rows.append((
|
| 255 |
+
r["name"], r["type"], r["topsim"], r["posdis"], r["causal_spec"],
|
| 256 |
+
r["cross_n16"] * 100 if r["cross_n16"] <= 1 else r["cross_n16"],
|
| 257 |
+
r["cross_n192"] * 100 if r["cross_n192"] <= 1 else r["cross_n192"],
|
| 258 |
+
))
|
| 259 |
+
|
| 260 |
+
from scipy.stats import spearmanr
|
| 261 |
+
def safe_corr(idx_x, idx_y):
|
| 262 |
+
x = []; y = []
|
| 263 |
+
for r in all_rows:
|
| 264 |
+
if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y])):
|
| 265 |
+
x.append(r[idx_x]); y.append(r[idx_y])
|
| 266 |
+
if len(x) < 4 or np.std(x) < 1e-9 or np.std(y) < 1e-9:
|
| 267 |
+
return float("nan"), float("nan")
|
| 268 |
+
rho, p = spearmanr(x, y)
|
| 269 |
+
return float(rho), float(p)
|
| 270 |
+
|
| 271 |
+
# Bootstrap CIs on Spearman
|
| 272 |
+
def bootstrap_ci(idx_x, idx_y, n_boot=2000):
|
| 273 |
+
x_arr = np.array([r[idx_x] for r in all_rows
|
| 274 |
+
if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y]))])
|
| 275 |
+
y_arr = np.array([r[idx_y] for r in all_rows
|
| 276 |
+
if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y]))])
|
| 277 |
+
if len(x_arr) < 4: return (float("nan"), float("nan"))
|
| 278 |
+
rng = np.random.RandomState(42)
|
| 279 |
+
rhos = []
|
| 280 |
+
for _ in range(n_boot):
|
| 281 |
+
idx = rng.randint(0, len(x_arr), len(x_arr))
|
| 282 |
+
xs = x_arr[idx]; ys = y_arr[idx]
|
| 283 |
+
if np.std(xs) < 1e-9 or np.std(ys) < 1e-9: continue
|
| 284 |
+
rho, _ = spearmanr(xs, ys)
|
| 285 |
+
if not np.isnan(rho): rhos.append(rho)
|
| 286 |
+
if len(rhos) < 100: return (float("nan"), float("nan"))
|
| 287 |
+
return float(np.percentile(rhos, 2.5)), float(np.percentile(rhos, 97.5))
|
| 288 |
+
|
| 289 |
+
corrs = {}
|
| 290 |
+
cis = {}
|
| 291 |
+
for met_name, met_idx in [("topsim", 2), ("posdis", 3), ("causal", 4)]:
|
| 292 |
+
for tgt_name, tgt_idx in [("n16", 5), ("n192", 6)]:
|
| 293 |
+
corrs[(met_name, tgt_name)] = safe_corr(met_idx, tgt_idx)
|
| 294 |
+
cis[(met_name, tgt_name)] = bootstrap_ci(met_idx, tgt_idx)
|
| 295 |
+
|
| 296 |
+
lines = [
|
| 297 |
+
"EXP Q ADDENDUM 2 -- multi-property 5-class boost to reach high PosDis",
|
| 298 |
+
"",
|
| 299 |
+
f"{'Config':<22s} | {'TopSim':<8s} | {'PosDis':<8s} | {'CausalSpec':<12s} | "
|
| 300 |
+
f"{'Cross 16':<10s} | {'Cross 192':<10s}",
|
| 301 |
+
"-" * 90,
|
| 302 |
+
]
|
| 303 |
+
for r in all_rows:
|
| 304 |
+
name, typ, ts, pd_, cs, c16, c192 = r
|
| 305 |
+
lines.append(f"{name:<22s} | {ts:+.2f} | {pd_:.2f} | {cs:.2f} "
|
| 306 |
+
f"| {c16:5.1f}% | {c192:5.1f}%")
|
| 307 |
+
|
| 308 |
+
lines.append("")
|
| 309 |
+
lines.append(f"FULL SWEEP SPEARMAN with bootstrap 95% CI (N={len(all_rows)}):")
|
| 310 |
+
for tgt in ["n16", "n192"]:
|
| 311 |
+
lines.append(f" vs cross_{tgt}:")
|
| 312 |
+
for met, label in [("topsim", "TopSim"), ("posdis", "PosDis"),
|
| 313 |
+
("causal", "CausalSpec")]:
|
| 314 |
+
rho, p = corrs[(met, tgt)]
|
| 315 |
+
ci_lo, ci_hi = cis[(met, tgt)]
|
| 316 |
+
lines.append(f" {label:<12s}: rho={rho:+.2f} p={p:.3f} "
|
| 317 |
+
f"95% CI=[{ci_lo:+.2f}, {ci_hi:+.2f}]")
|
| 318 |
+
|
| 319 |
+
abs_max_rho = 0
|
| 320 |
+
for k, (rho, p) in corrs.items():
|
| 321 |
+
if not np.isnan(rho): abs_max_rho = max(abs_max_rho, abs(rho))
|
| 322 |
+
lines.append(f"\nMax |rho| across 6 tests: {abs_max_rho:.2f}")
|
| 323 |
+
lines.append(f"Max PosDis in this paper's sweep: "
|
| 324 |
+
f"{max((r[3] for r in all_rows if not np.isnan(r[3])), default=float('nan')):.2f}")
|
| 325 |
+
lines.append(f"\nTotal runtime: {(time.time()-t0)/60:.1f} min")
|
| 326 |
+
|
| 327 |
+
summary = "\n".join(lines)
|
| 328 |
+
(OUT / "exp_q_addendum2_summary.txt").write_text(summary + "\n")
|
| 329 |
+
(OUT / "exp_q_addendum2_summary.json").write_text(json.dumps({
|
| 330 |
+
"new_rows_5class": rows,
|
| 331 |
+
"all_rows_combined": [{"name": r[0], "type": r[1], "topsim": r[2],
|
| 332 |
+
"posdis": r[3], "causal_spec": r[4],
|
| 333 |
+
"cross_n16": r[5], "cross_n192": r[6]}
|
| 334 |
+
for r in all_rows],
|
| 335 |
+
"spearman": {f"{m}__{n}": list(v) for (m, n), v in corrs.items()},
|
| 336 |
+
"bootstrap_95ci": {f"{m}__{n}": list(v) for (m, n), v in cis.items()},
|
| 337 |
+
}, indent=2, default=str))
|
| 338 |
+
print("\n" + summary, flush=True)
|
| 339 |
+
log(f"DONE in {(time.time()-t0)/60:.1f} min")
|
| 340 |
+
|
| 341 |
+
|
| 342 |
+
if __name__ == "__main__":
|
| 343 |
+
main()
|
code/_rev_q_addendum_multiprop.py
ADDED
|
@@ -0,0 +1,576 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP Q ADDENDUM: add multi-property discrete + continuous configs to the
|
| 3 |
+
scatter sweep so the headline-PosDis configs (0.76 discrete, 0.40 continuous
|
| 4 |
+
multi-prop) are represented in the metric-vs-transfer correlation.
|
| 5 |
+
|
| 6 |
+
Trains:
|
| 7 |
+
disc_multi_L3_V5 — discrete bottleneck, 2 head per agent, V=5, multi-prop
|
| 8 |
+
(mass_bin + restit_bin, 2-headed receiver)
|
| 9 |
+
disc_multi_L4_V10 — same but L=4, V=10 (broader coverage)
|
| 10 |
+
cont_multi_dim3 — continuous bottleneck, code_dim=3 per agent, multi-prop
|
| 11 |
+
|
| 12 |
+
Per config:
|
| 13 |
+
- Within: 3 seeds, mean across 2 heads (mass + restit) for combined acc
|
| 14 |
+
- Metrics: TopSim, PosDis, CausalSpec on multi-prop labels
|
| 15 |
+
- Cross to ramp at N=16 and N=192 (restitution only, since ramp lacks mass)
|
| 16 |
+
|
| 17 |
+
Re-uses EXP N's MultiPropReceiver and the metric helpers in EXP N + Q.
|
| 18 |
+
"""
|
| 19 |
+
import json, time, sys, os, math
|
| 20 |
+
from pathlib import Path
|
| 21 |
+
from datetime import datetime, timezone
|
| 22 |
+
import numpy as np
|
| 23 |
+
import torch
|
| 24 |
+
import torch.nn as nn
|
| 25 |
+
import torch.nn.functional as F
|
| 26 |
+
|
| 27 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 28 |
+
from _kinematics_train import (
|
| 29 |
+
DEVICE, ClassifierReceiver,
|
| 30 |
+
HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR,
|
| 31 |
+
EARLY_STOP_PATIENCE,
|
| 32 |
+
)
|
| 33 |
+
from _killer_experiment import TemporalEncoder, DiscreteSender, DiscreteMultiSender
|
| 34 |
+
from _overnight_p1_transfer import make_splits
|
| 35 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 36 |
+
from _rev_f_cnn_control import ci95
|
| 37 |
+
from _rev_q_posdis_scatter import build_discrete_sender, discrete_token_extract, discrete_topsim
|
| 38 |
+
from _rev_n_multiprop_continuous import (
|
| 39 |
+
MultiPropReceiver, train_multiprop_continuous_base,
|
| 40 |
+
topsim_multiprop, posdis_multiprop, causal_spec_multiprop,
|
| 41 |
+
)
|
| 42 |
+
from _rev_m_continuous_bottleneck import (
|
| 43 |
+
build_continuous_sender, get_continuous_messages,
|
| 44 |
+
train_recv_frozen_cont,
|
| 45 |
+
)
|
| 46 |
+
|
| 47 |
+
OUT = Path("results/reviewer_response/exp_q_addendum")
|
| 48 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 49 |
+
N_SEEDS = 3
|
| 50 |
+
N_LIST = [16, 192]
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
def log(msg):
|
| 54 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 55 |
+
print(f"[{ts}] EXP-QADD: {msg}", flush=True)
|
| 56 |
+
|
| 57 |
+
|
| 58 |
+
# ─── Discrete multi-prop training ───
|
| 59 |
+
def train_discrete_multi(feat, labels_list, seed, n_heads, vocab_size,
|
| 60 |
+
n_epochs=150):
|
| 61 |
+
"""Train DiscreteSender with multi-prop receiver (2 heads per receiver)."""
|
| 62 |
+
N, nf, dim = feat.shape
|
| 63 |
+
fpa = 1
|
| 64 |
+
msg_dim = vocab_size * n_heads * N_AGENTS
|
| 65 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 66 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 67 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 68 |
+
|
| 69 |
+
primary = labels_list[0]
|
| 70 |
+
train_ids, holdout_ids = [], []
|
| 71 |
+
for c in np.unique(primary):
|
| 72 |
+
ids_c = np.where(primary == c)[0]
|
| 73 |
+
rng.shuffle(ids_c)
|
| 74 |
+
split = max(1, len(ids_c) // 5)
|
| 75 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 76 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 77 |
+
n_classes_per_prop = [int(lbl.max()) + 1 for lbl in labels_list]
|
| 78 |
+
chance = 1.0 / max(n_classes_per_prop)
|
| 79 |
+
|
| 80 |
+
sender = build_discrete_sender(dim, n_heads, vocab_size, fpa)
|
| 81 |
+
receivers = [MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE)
|
| 82 |
+
for _ in range(3)]
|
| 83 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 84 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 85 |
+
labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list]
|
| 86 |
+
me = math.log(vocab_size)
|
| 87 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 88 |
+
best_acc = 0.0; best_ep = 0
|
| 89 |
+
best_sender_state = None; best_receiver_states = None; best_recv_idx = 0
|
| 90 |
+
|
| 91 |
+
for ep in range(n_epochs):
|
| 92 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break
|
| 93 |
+
if ep > 0 and ep % 40 == 0:
|
| 94 |
+
for i in range(len(receivers)):
|
| 95 |
+
receivers[i] = MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE)
|
| 96 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 97 |
+
sender.train(); [r.train() for r in receivers]
|
| 98 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1)
|
| 99 |
+
hard = ep >= 30
|
| 100 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 101 |
+
perm = rng_ep.permutation(train_ids)
|
| 102 |
+
for b in range(n_batches):
|
| 103 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 104 |
+
if len(batch_ids) < 4: continue
|
| 105 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 106 |
+
tgts = [ld[batch_ids] for ld in labels_dev]
|
| 107 |
+
msg, logits_list = sender(views, tau=tau, hard=hard)
|
| 108 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 109 |
+
for r in receivers:
|
| 110 |
+
head_logits = r(msg)
|
| 111 |
+
for hl, tgt in zip(head_logits, tgts):
|
| 112 |
+
loss = loss + F.cross_entropy(hl, tgt)
|
| 113 |
+
loss = loss / (len(receivers) * len(tgts))
|
| 114 |
+
for lg in logits_list:
|
| 115 |
+
lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8)
|
| 116 |
+
ent = -(p * lp).sum(-1).mean()
|
| 117 |
+
if ent / me < 0.1: loss = loss - 0.03 * ent
|
| 118 |
+
if torch.isnan(loss):
|
| 119 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 120 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 121 |
+
loss.backward()
|
| 122 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 123 |
+
so.step(); [o.step() for o in ros]
|
| 124 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 125 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 126 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 127 |
+
with torch.no_grad():
|
| 128 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 129 |
+
msg_ho, _ = sender(v_ho)
|
| 130 |
+
tgt_ho = [ld[holdout_ids] for ld in labels_dev]
|
| 131 |
+
best_per_recv = 0.0; best_idx = 0
|
| 132 |
+
for ri, r in enumerate(receivers):
|
| 133 |
+
head_logits = r(msg_ho)
|
| 134 |
+
accs = [(hl.argmax(-1) == tgt).float().mean().item()
|
| 135 |
+
for hl, tgt in zip(head_logits, tgt_ho)]
|
| 136 |
+
combined = float(np.mean(accs))
|
| 137 |
+
if combined > best_per_recv:
|
| 138 |
+
best_per_recv = combined; best_idx = ri
|
| 139 |
+
if best_per_recv > best_acc:
|
| 140 |
+
best_acc = best_per_recv; best_ep = ep
|
| 141 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 142 |
+
best_receiver_states = [
|
| 143 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 144 |
+
for r in receivers]
|
| 145 |
+
best_recv_idx = best_idx
|
| 146 |
+
return {
|
| 147 |
+
"sender_state": best_sender_state,
|
| 148 |
+
"receiver_states": best_receiver_states,
|
| 149 |
+
"best_recv_idx": best_recv_idx,
|
| 150 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 151 |
+
"task_acc": best_acc, "chance": chance,
|
| 152 |
+
"n_classes_per_prop": n_classes_per_prop,
|
| 153 |
+
"fpa": 1, "dim": dim,
|
| 154 |
+
"n_heads": n_heads, "vocab_size": vocab_size,
|
| 155 |
+
"msg_dim": msg_dim,
|
| 156 |
+
}
|
| 157 |
+
|
| 158 |
+
|
| 159 |
+
# ─── Discrete multi-prop metrics ───
|
| 160 |
+
def discrete_multi_topsim(tokens, labels_list, n_pairs=5000):
|
| 161 |
+
"""Spearman corr between Hamming(message tokens) and L1(label vector)."""
|
| 162 |
+
from scipy.stats import spearmanr
|
| 163 |
+
rng = np.random.RandomState(42)
|
| 164 |
+
N = tokens.shape[0]
|
| 165 |
+
n_pairs = min(n_pairs, N * (N - 1) // 2)
|
| 166 |
+
tok_d = []; lbl_d = []
|
| 167 |
+
seen = set()
|
| 168 |
+
for _ in range(n_pairs):
|
| 169 |
+
i, j = rng.randint(0, N), rng.randint(0, N)
|
| 170 |
+
if i == j or (i, j) in seen or (j, i) in seen: continue
|
| 171 |
+
seen.add((i, j))
|
| 172 |
+
tok_d.append(int((tokens[i] != tokens[j]).sum()))
|
| 173 |
+
lbl_d.append(sum(abs(int(lbl[i]) - int(lbl[j])) for lbl in labels_list))
|
| 174 |
+
if len(tok_d) < 10 or np.std(tok_d) < 1e-9 or np.std(lbl_d) < 1e-9:
|
| 175 |
+
return float("nan")
|
| 176 |
+
rho, _ = spearmanr(tok_d, lbl_d)
|
| 177 |
+
return float(rho) if not np.isnan(rho) else 0.0
|
| 178 |
+
|
| 179 |
+
|
| 180 |
+
def _mi_disc(x, y):
|
| 181 |
+
n = len(x)
|
| 182 |
+
n_x = int(np.max(x)) + 1; n_y = int(np.max(y)) + 1
|
| 183 |
+
p_x = np.bincount(x, minlength=n_x) / n
|
| 184 |
+
p_y = np.bincount(y, minlength=n_y) / n
|
| 185 |
+
H_x = -np.sum([p * np.log(p) for p in p_x if p > 0])
|
| 186 |
+
H_y = -np.sum([p * np.log(p) for p in p_y if p > 0])
|
| 187 |
+
joint = np.zeros((n_x, n_y))
|
| 188 |
+
for xv, yv in zip(x, y): joint[int(xv), int(yv)] += 1
|
| 189 |
+
joint /= n
|
| 190 |
+
H_xy = 0.0
|
| 191 |
+
for v in joint.ravel():
|
| 192 |
+
if v > 0: H_xy -= v * np.log(v)
|
| 193 |
+
return max(H_x + H_y - H_xy, 0.0)
|
| 194 |
+
|
| 195 |
+
|
| 196 |
+
def discrete_multi_posdis(tokens, labels_list):
|
| 197 |
+
"""Standard PosDis on discrete tokens: per-position, MI with each
|
| 198 |
+
property; PosDis = mean over positions of (top-second)/top."""
|
| 199 |
+
P = tokens.shape[1]
|
| 200 |
+
K = len(labels_list)
|
| 201 |
+
mi_matrix = np.zeros((P, K))
|
| 202 |
+
for p in range(P):
|
| 203 |
+
for k in range(K):
|
| 204 |
+
mi_matrix[p, k] = _mi_disc(tokens[:, p], labels_list[k])
|
| 205 |
+
if mi_matrix.sum() < 1e-9: return float("nan"), mi_matrix
|
| 206 |
+
n_active = 0; total = 0.0
|
| 207 |
+
for p in range(P):
|
| 208 |
+
sorted_mi = np.sort(mi_matrix[p])[::-1]
|
| 209 |
+
if sorted_mi[0] > 1e-6:
|
| 210 |
+
total += (sorted_mi[0] - sorted_mi[1]) / sorted_mi[0]
|
| 211 |
+
n_active += 1
|
| 212 |
+
if n_active == 0: return float("nan"), mi_matrix
|
| 213 |
+
return float(total / n_active), mi_matrix
|
| 214 |
+
|
| 215 |
+
|
| 216 |
+
def discrete_multi_causal(base, feat, labels_list, holdout_ids):
|
| 217 |
+
"""Mask each (agent x head) block; per-property accuracy drop."""
|
| 218 |
+
sender = build_discrete_sender(feat.shape[2], base["n_heads"],
|
| 219 |
+
base["vocab_size"], base["fpa"])
|
| 220 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 221 |
+
receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM,
|
| 222 |
+
base["n_classes_per_prop"]).to(DEVICE)
|
| 223 |
+
for _ in range(len(base["receiver_states"]))]
|
| 224 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 225 |
+
[r.eval() for r in receivers]
|
| 226 |
+
best_recv = receivers[base.get("best_recv_idx", 0)]
|
| 227 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 228 |
+
labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list]
|
| 229 |
+
K = len(labels_list); V = base["vocab_size"]; H = base["n_heads"]
|
| 230 |
+
n_positions = N_AGENTS * H
|
| 231 |
+
with torch.no_grad():
|
| 232 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 233 |
+
msg_ho, _ = sender(v_ho)
|
| 234 |
+
tgt_ho = [ld[holdout_ids] for ld in labels_dev]
|
| 235 |
+
baseline_per_prop = [(hl.argmax(-1) == tgt).float().mean().item()
|
| 236 |
+
for hl, tgt in zip(best_recv(msg_ho), tgt_ho)]
|
| 237 |
+
drops = np.zeros((n_positions, K))
|
| 238 |
+
for pos in range(n_positions):
|
| 239 |
+
masked = msg_ho.clone()
|
| 240 |
+
start = pos * V; end = start + V
|
| 241 |
+
mean_block = msg_ho[:, start:end].mean(dim=0)
|
| 242 |
+
masked[:, start:end] = mean_block
|
| 243 |
+
for p_idx, (hl, tgt) in enumerate(zip(best_recv(masked), tgt_ho)):
|
| 244 |
+
acc = (hl.argmax(-1) == tgt).float().mean().item()
|
| 245 |
+
drops[pos, p_idx] = baseline_per_prop[p_idx] - acc
|
| 246 |
+
return baseline_per_prop, drops
|
| 247 |
+
|
| 248 |
+
|
| 249 |
+
# ─── Cross-scenario eval (single-property restitution on ramp) ───
|
| 250 |
+
def disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, ho_ids, restit_idx=1):
|
| 251 |
+
"""Apply discrete-multi sender + best receiver's restit head to target."""
|
| 252 |
+
sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"],
|
| 253 |
+
base["vocab_size"], base["fpa"])
|
| 254 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 255 |
+
receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM,
|
| 256 |
+
base["n_classes_per_prop"]).to(DEVICE)
|
| 257 |
+
for _ in range(len(base["receiver_states"]))]
|
| 258 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 259 |
+
[r.eval() for r in receivers]
|
| 260 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 261 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 262 |
+
with torch.no_grad():
|
| 263 |
+
v_ho = [v[ho_ids].to(DEVICE) for v in agent_views]
|
| 264 |
+
msg_ho, _ = sender(v_ho)
|
| 265 |
+
tgt_ho = labels_dev[ho_ids]
|
| 266 |
+
best = 0.0
|
| 267 |
+
for r in receivers:
|
| 268 |
+
head_logits = r(msg_ho)
|
| 269 |
+
preds = head_logits[restit_idx].argmax(-1)
|
| 270 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 271 |
+
if acc > best: best = acc
|
| 272 |
+
return best
|
| 273 |
+
|
| 274 |
+
|
| 275 |
+
def disc_multi_train_recv_frozen(base, feat_tgt, labels_tgt, train_ids, holdout_ids,
|
| 276 |
+
seed, n_target, n_epochs=80):
|
| 277 |
+
"""Freeze sender; train fresh single-property receiver on n_target stratified
|
| 278 |
+
target examples; eval on holdout. Mirrors disc_train_recv_custom but uses
|
| 279 |
+
multi-prop sender."""
|
| 280 |
+
if n_target == 0:
|
| 281 |
+
return disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, holdout_ids)
|
| 282 |
+
rng = np.random.RandomState(seed * 311 + 7 + n_target)
|
| 283 |
+
n_t_classes = int(np.max(labels_tgt)) + 1
|
| 284 |
+
per_class = max(1, n_target // n_t_classes)
|
| 285 |
+
picks = []
|
| 286 |
+
for c in range(n_t_classes):
|
| 287 |
+
ids_c = np.array([i for i in train_ids if labels_tgt[i] == c])
|
| 288 |
+
if len(ids_c) == 0: continue
|
| 289 |
+
rng.shuffle(ids_c)
|
| 290 |
+
picks.extend(ids_c[:per_class])
|
| 291 |
+
picks = np.array(picks)
|
| 292 |
+
if len(picks) > n_target: picks = picks[:n_target]
|
| 293 |
+
elif len(picks) < n_target and len(train_ids) > len(picks):
|
| 294 |
+
extras = np.array([i for i in train_ids if i not in set(picks)])
|
| 295 |
+
rng.shuffle(extras)
|
| 296 |
+
picks = np.concatenate([picks, extras[:n_target - len(picks)]])
|
| 297 |
+
if len(picks) < 2: return float("nan")
|
| 298 |
+
sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"],
|
| 299 |
+
base["vocab_size"], base["fpa"])
|
| 300 |
+
sender.load_state_dict(base["sender_state"]); sender.to(DEVICE).eval()
|
| 301 |
+
for p in sender.parameters(): p.requires_grad = False
|
| 302 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, n_t_classes).to(DEVICE)
|
| 303 |
+
for _ in range(3)]
|
| 304 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 305 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 306 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 307 |
+
bs = min(BATCH_SIZE, len(picks))
|
| 308 |
+
best = 0.0
|
| 309 |
+
for ep in range(n_epochs):
|
| 310 |
+
[r.train() for r in receivers]
|
| 311 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 312 |
+
perm = rng_ep.permutation(picks)
|
| 313 |
+
for b in range(max(1, len(picks) // bs)):
|
| 314 |
+
batch = perm[b*bs:(b+1)*bs]
|
| 315 |
+
if len(batch) < 2: continue
|
| 316 |
+
views = [v[batch].to(DEVICE) for v in agent_views]
|
| 317 |
+
with torch.no_grad():
|
| 318 |
+
msg, _ = sender(views)
|
| 319 |
+
for r, o in zip(receivers, ros):
|
| 320 |
+
logits = r(msg)
|
| 321 |
+
loss = F.cross_entropy(logits, labels_dev[batch])
|
| 322 |
+
if torch.isnan(loss): continue
|
| 323 |
+
o.zero_grad(); loss.backward(); o.step()
|
| 324 |
+
if (ep + 1) % 5 == 0:
|
| 325 |
+
[r.eval() for r in receivers]
|
| 326 |
+
with torch.no_grad():
|
| 327 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 328 |
+
msg_ho, _ = sender(v_ho)
|
| 329 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 330 |
+
for r in receivers:
|
| 331 |
+
preds = r(msg_ho).argmax(-1)
|
| 332 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 333 |
+
if acc > best: best = acc
|
| 334 |
+
return best
|
| 335 |
+
|
| 336 |
+
|
| 337 |
+
# ─── Main ───
|
| 338 |
+
def main():
|
| 339 |
+
t0 = time.time()
|
| 340 |
+
log("=" * 60)
|
| 341 |
+
log("EXP Q ADDENDUM: multi-property bottleneck configs")
|
| 342 |
+
|
| 343 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 344 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 345 |
+
lbl_c_mass = load_labels("collision", "mass")
|
| 346 |
+
lbl_c_rest = load_labels("collision", "restitution")
|
| 347 |
+
lbl_r_rest = load_labels("ramp", "restitution")
|
| 348 |
+
log(f" collision: feat={tuple(feat_c.shape)} mass={np.bincount(lbl_c_mass).tolist()} "
|
| 349 |
+
f"rest={np.bincount(lbl_c_rest).tolist()}")
|
| 350 |
+
|
| 351 |
+
rows = []
|
| 352 |
+
|
| 353 |
+
# ── Discrete multi-prop configs ──
|
| 354 |
+
discrete_specs = [
|
| 355 |
+
("disc_multi_L3_V5", 3, 5),
|
| 356 |
+
("disc_multi_L4_V10", 4, 10),
|
| 357 |
+
]
|
| 358 |
+
for name, H, V in discrete_specs:
|
| 359 |
+
log(f"\n --- {name} (L={H}, V={V}, multi-prop) ---")
|
| 360 |
+
within_accs = []; bases = []
|
| 361 |
+
for seed in range(N_SEEDS):
|
| 362 |
+
t_s = time.time()
|
| 363 |
+
try:
|
| 364 |
+
base = train_discrete_multi(feat_c, [lbl_c_mass, lbl_c_rest],
|
| 365 |
+
seed, H, V)
|
| 366 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 367 |
+
log(f" {name} s{seed}: combined within={base['task_acc']:.3f} "
|
| 368 |
+
f"[{time.time()-t_s:.0f}s]")
|
| 369 |
+
except Exception as e:
|
| 370 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 371 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 372 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 373 |
+
if not valid:
|
| 374 |
+
log(f" {name}: no successful base"); continue
|
| 375 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 376 |
+
best_base = bases[best_idx]
|
| 377 |
+
ho_ids = best_base["holdout_ids"]
|
| 378 |
+
# Within metrics on best
|
| 379 |
+
try:
|
| 380 |
+
tokens = discrete_token_extract(best_base, feat_c)
|
| 381 |
+
tokens_ho = tokens[ho_ids]
|
| 382 |
+
ts = discrete_multi_topsim(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]])
|
| 383 |
+
pd_, mi = discrete_multi_posdis(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]])
|
| 384 |
+
base_pp, drops = discrete_multi_causal(best_base, feat_c,
|
| 385 |
+
[lbl_c_mass, lbl_c_rest], ho_ids)
|
| 386 |
+
cs = float(drops.max())
|
| 387 |
+
except Exception as e:
|
| 388 |
+
log(f" {name} metrics FAILED: {e}")
|
| 389 |
+
ts = pd_ = cs = float("nan")
|
| 390 |
+
# Cross-scenario coll->ramp at N=16, N=192 (restitution)
|
| 391 |
+
cross = {n: [] for n in N_LIST}
|
| 392 |
+
for seed, base in enumerate(bases):
|
| 393 |
+
if base is None:
|
| 394 |
+
for n in N_LIST: cross[n].append(float("nan"))
|
| 395 |
+
continue
|
| 396 |
+
tr_t, ho_t = make_splits(lbl_r_rest, seed)
|
| 397 |
+
for n in N_LIST:
|
| 398 |
+
try:
|
| 399 |
+
acc = disc_multi_train_recv_frozen(base, feat_r, lbl_r_rest,
|
| 400 |
+
tr_t, ho_t, seed, n)
|
| 401 |
+
cross[n].append(float(acc))
|
| 402 |
+
except Exception as e:
|
| 403 |
+
log(f" {name} s{seed} N={n} FAILED: {e}")
|
| 404 |
+
cross[n].append(float("nan"))
|
| 405 |
+
wm = float(np.mean([a for a in within_accs if not np.isnan(a)]))
|
| 406 |
+
cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)]))
|
| 407 |
+
if any(not np.isnan(x) for x in cross[n]) else float("nan")
|
| 408 |
+
for n in N_LIST}
|
| 409 |
+
log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 410 |
+
f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}")
|
| 411 |
+
rows.append({
|
| 412 |
+
"name": name, "type": "discrete_multi",
|
| 413 |
+
"n_heads": H, "vocab_size": V,
|
| 414 |
+
"msg_dim": V * H * N_AGENTS,
|
| 415 |
+
"within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs,
|
| 416 |
+
"cross_n16": cm[16], "cross_n192": cm[192],
|
| 417 |
+
})
|
| 418 |
+
|
| 419 |
+
# ── Continuous multi-prop config (matches Exp N exactly) ──
|
| 420 |
+
cont_spec = ("cont_multi_dim3", 3)
|
| 421 |
+
name, D = cont_spec
|
| 422 |
+
log(f"\n --- {name} (D={D}, multi-prop) ---")
|
| 423 |
+
within_accs = []; bases = []
|
| 424 |
+
for seed in range(N_SEEDS):
|
| 425 |
+
t_s = time.time()
|
| 426 |
+
try:
|
| 427 |
+
base = train_multiprop_continuous_base(
|
| 428 |
+
feat_c, [lbl_c_mass, lbl_c_rest], seed,
|
| 429 |
+
code_dim_per_agent=D, n_epochs=150)
|
| 430 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 431 |
+
log(f" {name} s{seed}: combined within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]")
|
| 432 |
+
except Exception as e:
|
| 433 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 434 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 435 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 436 |
+
if valid:
|
| 437 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 438 |
+
best_base = bases[best_idx]
|
| 439 |
+
ho_ids = best_base["holdout_ids"]
|
| 440 |
+
try:
|
| 441 |
+
msgs = get_continuous_messages(best_base, feat_c)
|
| 442 |
+
msgs_ho = msgs[ho_ids]
|
| 443 |
+
ts = topsim_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]])
|
| 444 |
+
pd_, _ = posdis_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]])
|
| 445 |
+
base_pp, drops = causal_spec_multiprop(best_base, feat_c,
|
| 446 |
+
[lbl_c_mass, lbl_c_rest], ho_ids)
|
| 447 |
+
cs = float(drops.max())
|
| 448 |
+
except Exception as e:
|
| 449 |
+
log(f" {name} metrics FAILED: {e}")
|
| 450 |
+
ts = pd_ = cs = float("nan")
|
| 451 |
+
# Cross to ramp on restitution: build single-task base view; train fresh receiver
|
| 452 |
+
cross = {n: [] for n in N_LIST}
|
| 453 |
+
for seed, base in enumerate(bases):
|
| 454 |
+
if base is None:
|
| 455 |
+
for n in N_LIST: cross[n].append(float("nan"))
|
| 456 |
+
continue
|
| 457 |
+
single_base = dict(base)
|
| 458 |
+
single_base["n_classes"] = base["n_classes_per_prop"][1]
|
| 459 |
+
single_base["receiver_states"] = []
|
| 460 |
+
tr_t, ho_t = make_splits(lbl_r_rest, seed)
|
| 461 |
+
for n in N_LIST:
|
| 462 |
+
try:
|
| 463 |
+
acc = train_recv_frozen_cont(single_base, feat_r, lbl_r_rest,
|
| 464 |
+
tr_t, ho_t, seed, n)
|
| 465 |
+
cross[n].append(float(acc))
|
| 466 |
+
except Exception as e:
|
| 467 |
+
log(f" {name} s{seed} N={n} FAILED: {e}")
|
| 468 |
+
cross[n].append(float("nan"))
|
| 469 |
+
wm = float(np.mean([a for a in within_accs if not np.isnan(a)]))
|
| 470 |
+
cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)]))
|
| 471 |
+
if any(not np.isnan(x) for x in cross[n]) else float("nan")
|
| 472 |
+
for n in N_LIST}
|
| 473 |
+
log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 474 |
+
f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}")
|
| 475 |
+
rows.append({
|
| 476 |
+
"name": name, "type": "continuous_multi",
|
| 477 |
+
"code_dim": D, "msg_dim": D * N_AGENTS,
|
| 478 |
+
"within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs,
|
| 479 |
+
"cross_n16": cm[16], "cross_n192": cm[192],
|
| 480 |
+
})
|
| 481 |
+
|
| 482 |
+
# ── Combine with original Exp Q rows for full correlation ──
|
| 483 |
+
original_q_rows = [
|
| 484 |
+
# (name, type, topsim, posdis, causal_spec, cross16, cross192)
|
| 485 |
+
("disc_L2_V5", "discrete", 0.88, 0.20, 0.02, 41.7, 43.9),
|
| 486 |
+
("disc_L2_V10", "discrete", 0.84, 0.25, 0.05, 46.1, 41.7),
|
| 487 |
+
("disc_L3_V5", "discrete", 0.84, 0.13, 0.02, 43.3, 42.8),
|
| 488 |
+
("disc_L3_V10", "discrete", 0.84, 0.12, 0.01, 43.3, 45.6),
|
| 489 |
+
("disc_L4_V5", "discrete", 0.90, 0.10, 0.01, 41.1, 42.2),
|
| 490 |
+
("disc_L4_V10", "discrete", 0.82, 0.08, 0.02, 45.0, 45.0),
|
| 491 |
+
("disc_L5_V5", "discrete", 0.89, 0.07, 0.02, 40.0, 43.9),
|
| 492 |
+
("cont_dim2", "continuous", 0.92, 0.15, 0.20, 48.9, 54.4),
|
| 493 |
+
("cont_dim3", "continuous", 0.91, 0.15, 0.02, 40.6, 41.1),
|
| 494 |
+
("cont_dim5", "continuous", 0.89, 0.06, 0.03, 47.2, 43.9),
|
| 495 |
+
("cont_dim10", "continuous", 0.88, 0.04, 0.01, 47.8, 48.3),
|
| 496 |
+
("cont_dim20", "continuous", 0.90, 0.02, 0.00, 48.9, 55.0),
|
| 497 |
+
]
|
| 498 |
+
all_rows = list(original_q_rows)
|
| 499 |
+
for r in rows:
|
| 500 |
+
all_rows.append((
|
| 501 |
+
r["name"], r["type"], r["topsim"], r["posdis"], r["causal_spec"],
|
| 502 |
+
r["cross_n16"] * 100 if r["cross_n16"] <= 1 else r["cross_n16"],
|
| 503 |
+
r["cross_n192"] * 100 if r["cross_n192"] <= 1 else r["cross_n192"],
|
| 504 |
+
))
|
| 505 |
+
|
| 506 |
+
# Spearman across all rows
|
| 507 |
+
from scipy.stats import spearmanr
|
| 508 |
+
def safe_corr(idx_x, idx_y):
|
| 509 |
+
x = []; y = []
|
| 510 |
+
for r in all_rows:
|
| 511 |
+
if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y])):
|
| 512 |
+
x.append(r[idx_x]); y.append(r[idx_y])
|
| 513 |
+
if len(x) < 4 or np.std(x) < 1e-9 or np.std(y) < 1e-9:
|
| 514 |
+
return float("nan"), float("nan")
|
| 515 |
+
rho, p = spearmanr(x, y)
|
| 516 |
+
return float(rho), float(p)
|
| 517 |
+
corrs = {
|
| 518 |
+
("topsim", 16): safe_corr(2, 5),
|
| 519 |
+
("topsim", 192): safe_corr(2, 6),
|
| 520 |
+
("posdis", 16): safe_corr(3, 5),
|
| 521 |
+
("posdis", 192): safe_corr(3, 6),
|
| 522 |
+
("causal", 16): safe_corr(4, 5),
|
| 523 |
+
("causal", 192): safe_corr(4, 6),
|
| 524 |
+
}
|
| 525 |
+
|
| 526 |
+
# Print summary
|
| 527 |
+
lines = [
|
| 528 |
+
"EXP Q ADDENDUM -- multi-property bottleneck configs",
|
| 529 |
+
"",
|
| 530 |
+
f"{'Config':<22s} | {'TopSim':<8s} | {'PosDis':<8s} | {'CausalSpec':<12s} | "
|
| 531 |
+
f"{'Cross 16':<10s} | {'Cross 192':<10s}",
|
| 532 |
+
"-" * 90,
|
| 533 |
+
]
|
| 534 |
+
for r in all_rows:
|
| 535 |
+
name, typ, ts, pd_, cs, c16, c192 = r
|
| 536 |
+
lines.append(f"{name:<22s} | {ts:+.2f} | {pd_:.2f} | {cs:.2f} "
|
| 537 |
+
f"| {c16:5.1f}% | {c192:5.1f}%")
|
| 538 |
+
|
| 539 |
+
lines.append("")
|
| 540 |
+
lines.append("FULL-SWEEP SPEARMAN (now including multi-prop configs):")
|
| 541 |
+
for tgt_n in [16, 192]:
|
| 542 |
+
lines.append(f" vs cross_n{tgt_n}:")
|
| 543 |
+
for met, label in [("topsim", "TopSim"), ("posdis", "PosDis"),
|
| 544 |
+
("causal", "CausalSpec")]:
|
| 545 |
+
rho, p = corrs[(met, tgt_n)]
|
| 546 |
+
lines.append(f" {label:<12s}: rho={rho:+.2f} p={p:.3f}")
|
| 547 |
+
|
| 548 |
+
abs_max_rho = 0
|
| 549 |
+
for k, (rho, p) in corrs.items():
|
| 550 |
+
if not np.isnan(rho): abs_max_rho = max(abs_max_rho, abs(rho))
|
| 551 |
+
lines.append("")
|
| 552 |
+
if abs_max_rho < 0.30:
|
| 553 |
+
verd = "Metrics still uncorrelated with transfer."
|
| 554 |
+
elif abs_max_rho < 0.55:
|
| 555 |
+
verd = f"Weak/moderate correlation (max |rho|={abs_max_rho:.2f})."
|
| 556 |
+
else:
|
| 557 |
+
verd = f"Strong correlation (max |rho|={abs_max_rho:.2f}). Reframe."
|
| 558 |
+
lines.append(f"VERDICT: {verd}")
|
| 559 |
+
lines.append(f"\nTotal runtime: {(time.time()-t0)/60:.1f} min")
|
| 560 |
+
|
| 561 |
+
summary = "\n".join(lines)
|
| 562 |
+
(OUT / "exp_q_addendum_summary.txt").write_text(summary + "\n")
|
| 563 |
+
(OUT / "exp_q_addendum_summary.json").write_text(json.dumps({
|
| 564 |
+
"new_rows": rows,
|
| 565 |
+
"all_rows_combined": [{"name": r[0], "type": r[1], "topsim": r[2],
|
| 566 |
+
"posdis": r[3], "causal_spec": r[4],
|
| 567 |
+
"cross_n16": r[5], "cross_n192": r[6]}
|
| 568 |
+
for r in all_rows],
|
| 569 |
+
"spearman": {f"{m}__n{n}": list(v) for (m, n), v in corrs.items()},
|
| 570 |
+
}, indent=2, default=str))
|
| 571 |
+
print("\n" + summary, flush=True)
|
| 572 |
+
log(f"DONE in {(time.time()-t0)/60:.1f} min")
|
| 573 |
+
|
| 574 |
+
|
| 575 |
+
if __name__ == "__main__":
|
| 576 |
+
main()
|
code/_rev_q_posdis_scatter.py
ADDED
|
@@ -0,0 +1,600 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP Q (reviewer_response): PosDis-vs-CROSS-SCENARIO SCATTER PLOT.
|
| 3 |
+
|
| 4 |
+
Sweep bottleneck hyperparameters to get a RANGE of (TopSim, PosDis,
|
| 5 |
+
CausalSpec) values, then plot each config's cross-scenario accuracy.
|
| 6 |
+
If correlations between metrics and transfer are near zero, the metrics
|
| 7 |
+
genuinely don't predict transfer (paper claim). If positive correlations
|
| 8 |
+
emerge, claim must narrow.
|
| 9 |
+
|
| 10 |
+
Configs:
|
| 11 |
+
Discrete: vary (n_heads, vocab_size) on V-JEPA collision restitution
|
| 12 |
+
L=2 V=5, L=2 V=10, L=3 V=5, L=3 V=10, L=4 V=5, L=4 V=10, L=5 V=5
|
| 13 |
+
Continuous: vary code_dim_per_agent
|
| 14 |
+
dim=2, 3, 5, 10, 20
|
| 15 |
+
|
| 16 |
+
For each config: train within collision (3 seeds, restitution 3-class),
|
| 17 |
+
measure TopSim/PosDis/CausalSpec on holdout, then run cross-scenario
|
| 18 |
+
collision->ramp at N=16 and N=192.
|
| 19 |
+
|
| 20 |
+
Result: scatter table (TopSim, PosDis, CausalSpec, CrossN16, CrossN192)
|
| 21 |
+
across configs + Spearman correlation of each metric with transfer.
|
| 22 |
+
"""
|
| 23 |
+
import json, time, sys, os, math
|
| 24 |
+
from pathlib import Path
|
| 25 |
+
from datetime import datetime, timezone
|
| 26 |
+
import numpy as np
|
| 27 |
+
import torch
|
| 28 |
+
import torch.nn as nn
|
| 29 |
+
import torch.nn.functional as F
|
| 30 |
+
|
| 31 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 32 |
+
from _kinematics_train import (
|
| 33 |
+
DEVICE, ClassifierReceiver,
|
| 34 |
+
HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR,
|
| 35 |
+
EARLY_STOP_PATIENCE,
|
| 36 |
+
)
|
| 37 |
+
from _killer_experiment import TemporalEncoder, DiscreteSender, DiscreteMultiSender
|
| 38 |
+
from _overnight_p1_transfer import (
|
| 39 |
+
train_receiver_frozen_sender as disc_train_recv,
|
| 40 |
+
eval_zero_shot as disc_eval_zs, make_splits,
|
| 41 |
+
)
|
| 42 |
+
from _overnight_p3_matrix import load_feat_subsampled, load_labels
|
| 43 |
+
from _rev_f_cnn_control import ci95
|
| 44 |
+
from _rev_m_continuous_bottleneck import (
|
| 45 |
+
train_continuous_base, train_recv_frozen_cont,
|
| 46 |
+
get_continuous_messages, topsim_continuous,
|
| 47 |
+
posdis_continuous_per_dim, causal_specificity,
|
| 48 |
+
)
|
| 49 |
+
|
| 50 |
+
OUT = Path("results/reviewer_response/exp_q")
|
| 51 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 52 |
+
N_SEEDS = 3 # fewer seeds for sweep; we want coverage
|
| 53 |
+
N_LIST = [16, 192]
|
| 54 |
+
|
| 55 |
+
DISCRETE_CONFIGS = [
|
| 56 |
+
{"n_heads": 2, "vocab_size": 5},
|
| 57 |
+
{"n_heads": 2, "vocab_size": 10},
|
| 58 |
+
{"n_heads": 3, "vocab_size": 5},
|
| 59 |
+
{"n_heads": 3, "vocab_size": 10},
|
| 60 |
+
{"n_heads": 4, "vocab_size": 5},
|
| 61 |
+
{"n_heads": 4, "vocab_size": 10},
|
| 62 |
+
{"n_heads": 5, "vocab_size": 5},
|
| 63 |
+
]
|
| 64 |
+
CONTINUOUS_CONFIGS = [
|
| 65 |
+
{"code_dim": 2},
|
| 66 |
+
{"code_dim": 3},
|
| 67 |
+
{"code_dim": 5},
|
| 68 |
+
{"code_dim": 10},
|
| 69 |
+
{"code_dim": 20},
|
| 70 |
+
]
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
def log(msg):
|
| 74 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 75 |
+
print(f"[{ts}] EXP-Q: {msg}", flush=True)
|
| 76 |
+
|
| 77 |
+
|
| 78 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 79 |
+
# Discrete sender with custom L (n_heads), V (vocab_size)
|
| 80 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 81 |
+
|
| 82 |
+
def build_discrete_sender(feat_dim, n_heads, vocab_size, fpa=1):
|
| 83 |
+
senders = [DiscreteSender(TemporalEncoder(HIDDEN_DIM, feat_dim, fpa),
|
| 84 |
+
HIDDEN_DIM, vocab_size, n_heads)
|
| 85 |
+
for _ in range(N_AGENTS)]
|
| 86 |
+
return DiscreteMultiSender(senders).to(DEVICE)
|
| 87 |
+
|
| 88 |
+
|
| 89 |
+
def train_discrete_custom(feat, labels, seed, n_heads, vocab_size, n_epochs=150):
|
| 90 |
+
"""Train DiscreteSender with custom L=n_heads, V=vocab_size."""
|
| 91 |
+
N, nf, dim = feat.shape
|
| 92 |
+
fpa = 1
|
| 93 |
+
msg_dim = vocab_size * n_heads * N_AGENTS
|
| 94 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 95 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 96 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 97 |
+
train_ids, holdout_ids = [], []
|
| 98 |
+
for c in np.unique(labels):
|
| 99 |
+
ids_c = np.where(labels == c)[0]
|
| 100 |
+
rng.shuffle(ids_c)
|
| 101 |
+
split = max(1, len(ids_c) // 5)
|
| 102 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 103 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 104 |
+
n_classes = int(labels.max()) + 1
|
| 105 |
+
chance = 1.0 / n_classes
|
| 106 |
+
|
| 107 |
+
sender = build_discrete_sender(dim, n_heads, vocab_size, fpa)
|
| 108 |
+
receivers = [ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 109 |
+
for _ in range(3)]
|
| 110 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 111 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 112 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 113 |
+
me = math.log(vocab_size)
|
| 114 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 115 |
+
best_acc = 0.0; best_ep = 0
|
| 116 |
+
best_sender_state = None; best_receiver_states = None; best_recv_idx = 0
|
| 117 |
+
|
| 118 |
+
for ep in range(n_epochs):
|
| 119 |
+
if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break
|
| 120 |
+
if ep > 0 and ep % 40 == 0:
|
| 121 |
+
for i in range(len(receivers)):
|
| 122 |
+
receivers[i] = ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 123 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 124 |
+
sender.train(); [r.train() for r in receivers]
|
| 125 |
+
tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1)
|
| 126 |
+
hard = ep >= 30
|
| 127 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 128 |
+
perm = rng_ep.permutation(train_ids)
|
| 129 |
+
for b in range(n_batches):
|
| 130 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 131 |
+
if len(batch_ids) < 4: continue
|
| 132 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 133 |
+
tgt = labels_dev[batch_ids]
|
| 134 |
+
msg, logits_list = sender(views, tau=tau, hard=hard)
|
| 135 |
+
loss = torch.tensor(0.0, device=DEVICE)
|
| 136 |
+
for r in receivers: loss = loss + F.cross_entropy(r(msg), tgt)
|
| 137 |
+
loss = loss / len(receivers)
|
| 138 |
+
for lg in logits_list:
|
| 139 |
+
lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8)
|
| 140 |
+
ent = -(p * lp).sum(-1).mean()
|
| 141 |
+
if ent / me < 0.1: loss = loss - 0.03 * ent
|
| 142 |
+
if torch.isnan(loss):
|
| 143 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 144 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 145 |
+
loss.backward()
|
| 146 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 147 |
+
so.step(); [o.step() for o in ros]
|
| 148 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 149 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 150 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 151 |
+
with torch.no_grad():
|
| 152 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 153 |
+
msg_ho, _ = sender(v_ho)
|
| 154 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 155 |
+
best_per_recv = 0.0; best_idx = 0
|
| 156 |
+
for ri, r in enumerate(receivers):
|
| 157 |
+
preds = r(msg_ho).argmax(-1)
|
| 158 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 159 |
+
if acc > best_per_recv:
|
| 160 |
+
best_per_recv = acc; best_idx = ri
|
| 161 |
+
if best_per_recv > best_acc:
|
| 162 |
+
best_acc = best_per_recv; best_ep = ep
|
| 163 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 164 |
+
best_receiver_states = [
|
| 165 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 166 |
+
for r in receivers]
|
| 167 |
+
best_recv_idx = best_idx
|
| 168 |
+
return {
|
| 169 |
+
"sender_state": best_sender_state,
|
| 170 |
+
"receiver_states": best_receiver_states,
|
| 171 |
+
"best_recv_idx": best_recv_idx,
|
| 172 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 173 |
+
"task_acc": best_acc, "chance": chance,
|
| 174 |
+
"n_classes": n_classes, "fpa": 1, "dim": dim,
|
| 175 |
+
"n_heads": n_heads, "vocab_size": vocab_size,
|
| 176 |
+
"msg_dim": msg_dim,
|
| 177 |
+
}
|
| 178 |
+
|
| 179 |
+
|
| 180 |
+
def disc_get_messages(base, feat):
|
| 181 |
+
"""Return discrete messages as one-hot concatenated (N, msg_dim)."""
|
| 182 |
+
sender = build_discrete_sender(feat.shape[2], base["n_heads"],
|
| 183 |
+
base["vocab_size"], base["fpa"])
|
| 184 |
+
sender.load_state_dict(base["sender_state"])
|
| 185 |
+
sender.eval().to(DEVICE)
|
| 186 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 187 |
+
with torch.no_grad():
|
| 188 |
+
views = [v.to(DEVICE) for v in agent_views]
|
| 189 |
+
msg, _ = sender(views)
|
| 190 |
+
return msg.cpu().float()
|
| 191 |
+
|
| 192 |
+
|
| 193 |
+
def disc_zero_shot(base, feat_tgt, labels_tgt, ho_ids):
|
| 194 |
+
sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"],
|
| 195 |
+
base["vocab_size"], base["fpa"])
|
| 196 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 197 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 198 |
+
for _ in range(len(base["receiver_states"]))]
|
| 199 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 200 |
+
[r.eval() for r in receivers]
|
| 201 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 202 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 203 |
+
with torch.no_grad():
|
| 204 |
+
v_ho = [v[ho_ids].to(DEVICE) for v in agent_views]
|
| 205 |
+
msg_ho, _ = sender(v_ho)
|
| 206 |
+
tgt_ho = labels_dev[ho_ids]
|
| 207 |
+
best = 0.0
|
| 208 |
+
for r in receivers:
|
| 209 |
+
preds = r(msg_ho).argmax(-1)
|
| 210 |
+
best = max(best, (preds == tgt_ho).float().mean().item())
|
| 211 |
+
return best
|
| 212 |
+
|
| 213 |
+
|
| 214 |
+
def disc_train_recv_custom(base, feat_tgt, labels_tgt, train_ids, holdout_ids,
|
| 215 |
+
seed, n_target, n_epochs=80):
|
| 216 |
+
"""Mimics the canonical train_receiver_frozen_sender but using our custom
|
| 217 |
+
discrete sender architecture."""
|
| 218 |
+
if n_target == 0:
|
| 219 |
+
return disc_zero_shot(base, feat_tgt, labels_tgt, holdout_ids)
|
| 220 |
+
rng = np.random.RandomState(seed * 311 + 7 + n_target)
|
| 221 |
+
n_t_classes = int(np.max(labels_tgt)) + 1
|
| 222 |
+
per_class = max(1, n_target // n_t_classes)
|
| 223 |
+
picks = []
|
| 224 |
+
for c in range(n_t_classes):
|
| 225 |
+
ids_c = np.array([i for i in train_ids if labels_tgt[i] == c])
|
| 226 |
+
if len(ids_c) == 0: continue
|
| 227 |
+
rng.shuffle(ids_c)
|
| 228 |
+
picks.extend(ids_c[:per_class])
|
| 229 |
+
picks = np.array(picks)
|
| 230 |
+
if len(picks) > n_target: picks = picks[:n_target]
|
| 231 |
+
elif len(picks) < n_target and len(train_ids) > len(picks):
|
| 232 |
+
extras = np.array([i for i in train_ids if i not in set(picks)])
|
| 233 |
+
rng.shuffle(extras)
|
| 234 |
+
picks = np.concatenate([picks, extras[:n_target - len(picks)]])
|
| 235 |
+
if len(picks) < 2: return float("nan")
|
| 236 |
+
|
| 237 |
+
sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"],
|
| 238 |
+
base["vocab_size"], base["fpa"])
|
| 239 |
+
sender.load_state_dict(base["sender_state"]); sender.to(DEVICE).eval()
|
| 240 |
+
for p in sender.parameters(): p.requires_grad = False
|
| 241 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 242 |
+
for _ in range(3)]
|
| 243 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 244 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 245 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 246 |
+
bs = min(BATCH_SIZE, len(picks))
|
| 247 |
+
best = 0.0
|
| 248 |
+
for ep in range(n_epochs):
|
| 249 |
+
[r.train() for r in receivers]
|
| 250 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 251 |
+
perm = rng_ep.permutation(picks)
|
| 252 |
+
for b in range(max(1, len(picks) // bs)):
|
| 253 |
+
batch = perm[b*bs:(b+1)*bs]
|
| 254 |
+
if len(batch) < 2: continue
|
| 255 |
+
views = [v[batch].to(DEVICE) for v in agent_views]
|
| 256 |
+
with torch.no_grad():
|
| 257 |
+
msg, _ = sender(views)
|
| 258 |
+
for r, o in zip(receivers, ros):
|
| 259 |
+
logits = r(msg)
|
| 260 |
+
loss = F.cross_entropy(logits, labels_dev[batch])
|
| 261 |
+
if torch.isnan(loss): continue
|
| 262 |
+
o.zero_grad(); loss.backward(); o.step()
|
| 263 |
+
if (ep + 1) % 5 == 0:
|
| 264 |
+
[r.eval() for r in receivers]
|
| 265 |
+
with torch.no_grad():
|
| 266 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 267 |
+
msg_ho, _ = sender(v_ho)
|
| 268 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 269 |
+
for r in receivers:
|
| 270 |
+
preds = r(msg_ho).argmax(-1)
|
| 271 |
+
acc = (preds == tgt_ho).float().mean().item()
|
| 272 |
+
if acc > best: best = acc
|
| 273 |
+
return best
|
| 274 |
+
|
| 275 |
+
|
| 276 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 277 |
+
# Discrete TopSim/PosDis (token-based)
|
| 278 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 279 |
+
|
| 280 |
+
def discrete_token_extract(base, feat):
|
| 281 |
+
"""Get argmax tokens from each head per agent. Returns (N, n_agents*n_heads) ints."""
|
| 282 |
+
sender = build_discrete_sender(feat.shape[2], base["n_heads"],
|
| 283 |
+
base["vocab_size"], base["fpa"])
|
| 284 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 285 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 286 |
+
all_tokens = []
|
| 287 |
+
with torch.no_grad():
|
| 288 |
+
views = [v.to(DEVICE) for v in agent_views]
|
| 289 |
+
for s, v in zip(sender.senders, views):
|
| 290 |
+
h = s.encoder(v)
|
| 291 |
+
for head in s.heads:
|
| 292 |
+
logits = head(h)
|
| 293 |
+
all_tokens.append(logits.argmax(-1).cpu().numpy())
|
| 294 |
+
return np.stack(all_tokens, axis=1) # (N, n_agents*n_heads)
|
| 295 |
+
|
| 296 |
+
|
| 297 |
+
def discrete_topsim(tokens, labels, n_pairs=5000):
|
| 298 |
+
from scipy.stats import spearmanr
|
| 299 |
+
rng = np.random.RandomState(42)
|
| 300 |
+
N = len(labels)
|
| 301 |
+
n_pairs = min(n_pairs, N * (N - 1) // 2)
|
| 302 |
+
tok_d = []; lbl_d = []
|
| 303 |
+
seen = set()
|
| 304 |
+
for _ in range(n_pairs):
|
| 305 |
+
i, j = rng.randint(0, N), rng.randint(0, N)
|
| 306 |
+
if i == j or (i, j) in seen or (j, i) in seen: continue
|
| 307 |
+
seen.add((i, j))
|
| 308 |
+
tok_d.append(int((tokens[i] != tokens[j]).sum()))
|
| 309 |
+
lbl_d.append(abs(int(labels[i]) - int(labels[j])))
|
| 310 |
+
if len(tok_d) < 10 or np.std(tok_d) < 1e-9 or np.std(lbl_d) < 1e-9:
|
| 311 |
+
return float("nan")
|
| 312 |
+
rho, _ = spearmanr(tok_d, lbl_d)
|
| 313 |
+
return float(rho) if not np.isnan(rho) else 0.0
|
| 314 |
+
|
| 315 |
+
|
| 316 |
+
def _mi_discrete(x, y):
|
| 317 |
+
n = len(x)
|
| 318 |
+
n_x = int(np.max(x)) + 1; n_y = int(np.max(y)) + 1
|
| 319 |
+
p_x = np.bincount(x, minlength=n_x) / n
|
| 320 |
+
p_y = np.bincount(y, minlength=n_y) / n
|
| 321 |
+
H_x = -np.sum([p * np.log(p) for p in p_x if p > 0])
|
| 322 |
+
H_y = -np.sum([p * np.log(p) for p in p_y if p > 0])
|
| 323 |
+
joint = np.zeros((n_x, n_y))
|
| 324 |
+
for xv, yv in zip(x, y): joint[int(xv), int(yv)] += 1
|
| 325 |
+
joint /= n
|
| 326 |
+
H_xy = 0.0
|
| 327 |
+
for v in joint.ravel():
|
| 328 |
+
if v > 0: H_xy -= v * np.log(v)
|
| 329 |
+
return max(H_x + H_y - H_xy, 0.0)
|
| 330 |
+
|
| 331 |
+
|
| 332 |
+
def discrete_posdis(tokens, labels):
|
| 333 |
+
"""Per-position MI with the single label, normalized to fraction of total MI
|
| 334 |
+
concentrated in the top position (single-prop variant)."""
|
| 335 |
+
P = tokens.shape[1]
|
| 336 |
+
mis = np.zeros(P)
|
| 337 |
+
for p in range(P):
|
| 338 |
+
mis[p] = _mi_discrete(tokens[:, p], labels)
|
| 339 |
+
if mis.sum() < 1e-9: return float("nan")
|
| 340 |
+
return float(mis.max() / mis.sum())
|
| 341 |
+
|
| 342 |
+
|
| 343 |
+
def discrete_causal_spec(base, feat, labels, holdout_ids):
|
| 344 |
+
"""Per-position mask -> measure receiver accuracy drop."""
|
| 345 |
+
sender = build_discrete_sender(feat.shape[2], base["n_heads"],
|
| 346 |
+
base["vocab_size"], base["fpa"])
|
| 347 |
+
sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE)
|
| 348 |
+
receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes"]).to(DEVICE)
|
| 349 |
+
for _ in range(len(base["receiver_states"]))]
|
| 350 |
+
for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s)
|
| 351 |
+
[r.eval() for r in receivers]
|
| 352 |
+
best_recv = receivers[base.get("best_recv_idx", 0)]
|
| 353 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 354 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 355 |
+
V = base["vocab_size"]; H = base["n_heads"]
|
| 356 |
+
with torch.no_grad():
|
| 357 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 358 |
+
msg_ho, _ = sender(v_ho)
|
| 359 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 360 |
+
baseline = (best_recv(msg_ho).argmax(-1) == tgt_ho).float().mean().item()
|
| 361 |
+
# Mask each (agent, head) block
|
| 362 |
+
n_positions = N_AGENTS * H
|
| 363 |
+
drops = np.zeros(n_positions)
|
| 364 |
+
for pos in range(n_positions):
|
| 365 |
+
masked = msg_ho.clone()
|
| 366 |
+
start = pos * V
|
| 367 |
+
end = start + V
|
| 368 |
+
mean_block = msg_ho[:, start:end].mean(dim=0)
|
| 369 |
+
masked[:, start:end] = mean_block
|
| 370 |
+
acc = (best_recv(masked).argmax(-1) == tgt_ho).float().mean().item()
|
| 371 |
+
drops[pos] = baseline - acc
|
| 372 |
+
return baseline, drops
|
| 373 |
+
|
| 374 |
+
|
| 375 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 376 |
+
# Main sweep
|
| 377 |
+
# ─────────────────────────────────────────────────────────────────────────────
|
| 378 |
+
|
| 379 |
+
def main():
|
| 380 |
+
t0 = time.time()
|
| 381 |
+
log("=" * 60)
|
| 382 |
+
log("EXP Q: PosDis-vs-cross-scenario sweep")
|
| 383 |
+
|
| 384 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 385 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 386 |
+
lbl_c = load_labels("collision", "restitution")
|
| 387 |
+
lbl_r = load_labels("ramp", "restitution")
|
| 388 |
+
log(f" collision: {tuple(feat_c.shape)} dist={np.bincount(lbl_c).tolist()}")
|
| 389 |
+
log(f" ramp: {tuple(feat_r.shape)} dist={np.bincount(lbl_r).tolist()}")
|
| 390 |
+
|
| 391 |
+
rows = [] # each row: dict with config, metrics, transfer
|
| 392 |
+
|
| 393 |
+
# ── Discrete sweep ──
|
| 394 |
+
for cfg in DISCRETE_CONFIGS:
|
| 395 |
+
H, V = cfg["n_heads"], cfg["vocab_size"]
|
| 396 |
+
name = f"disc_L{H}_V{V}"
|
| 397 |
+
log(f"\n --- {name} (L={H}, V={V}) ---")
|
| 398 |
+
within_accs = []; bases = []
|
| 399 |
+
for seed in range(N_SEEDS):
|
| 400 |
+
t_s = time.time()
|
| 401 |
+
try:
|
| 402 |
+
base = train_discrete_custom(feat_c, lbl_c, seed, H, V)
|
| 403 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 404 |
+
log(f" {name} s{seed}: within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]")
|
| 405 |
+
except Exception as e:
|
| 406 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 407 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 408 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 409 |
+
if not valid:
|
| 410 |
+
log(f" {name}: no successful base"); continue
|
| 411 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 412 |
+
best_base = bases[best_idx]
|
| 413 |
+
ho_ids = best_base["holdout_ids"]
|
| 414 |
+
# Metrics on best base
|
| 415 |
+
try:
|
| 416 |
+
tokens = discrete_token_extract(best_base, feat_c)
|
| 417 |
+
tokens_ho = tokens[ho_ids]
|
| 418 |
+
ts = discrete_topsim(tokens_ho, lbl_c[ho_ids])
|
| 419 |
+
pd_ = discrete_posdis(tokens_ho, lbl_c[ho_ids])
|
| 420 |
+
base_acc, drops = discrete_causal_spec(best_base, feat_c, lbl_c, ho_ids)
|
| 421 |
+
cs = float(drops.max())
|
| 422 |
+
except Exception as e:
|
| 423 |
+
log(f" {name} metrics FAILED: {e}")
|
| 424 |
+
ts = pd_ = cs = float("nan")
|
| 425 |
+
# Cross-scenario at N=16, N=192
|
| 426 |
+
cross = {n: [] for n in N_LIST}
|
| 427 |
+
for seed, base in enumerate(bases):
|
| 428 |
+
if base is None:
|
| 429 |
+
for n in N_LIST: cross[n].append(float("nan"))
|
| 430 |
+
continue
|
| 431 |
+
tr_t, ho_t = make_splits(lbl_r, seed)
|
| 432 |
+
for n in N_LIST:
|
| 433 |
+
try:
|
| 434 |
+
if n == 0:
|
| 435 |
+
acc = disc_zero_shot(base, feat_r, lbl_r, ho_t)
|
| 436 |
+
else:
|
| 437 |
+
acc = disc_train_recv_custom(base, feat_r, lbl_r, tr_t, ho_t,
|
| 438 |
+
seed, n)
|
| 439 |
+
cross[n].append(float(acc))
|
| 440 |
+
except Exception as e:
|
| 441 |
+
log(f" {name} s{seed} N={n} FAILED: {e}")
|
| 442 |
+
cross[n].append(float("nan"))
|
| 443 |
+
wm = float(np.mean([a for a in within_accs if not np.isnan(a)]))
|
| 444 |
+
cross_means = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)]))
|
| 445 |
+
if any(not np.isnan(x) for x in cross[n]) else float("nan")
|
| 446 |
+
for n in N_LIST}
|
| 447 |
+
log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 448 |
+
f"CausalSpec={cs:.3f} cross16={cross_means[16]:.3f} cross192={cross_means[192]:.3f}")
|
| 449 |
+
rows.append({
|
| 450 |
+
"name": name, "type": "discrete",
|
| 451 |
+
"n_heads": H, "vocab_size": V,
|
| 452 |
+
"msg_dim": V * H * N_AGENTS,
|
| 453 |
+
"within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs,
|
| 454 |
+
"cross_n16": cross_means[16], "cross_n192": cross_means[192],
|
| 455 |
+
})
|
| 456 |
+
|
| 457 |
+
# ── Continuous sweep ──
|
| 458 |
+
for cfg in CONTINUOUS_CONFIGS:
|
| 459 |
+
D = cfg["code_dim"]
|
| 460 |
+
name = f"cont_dim{D}"
|
| 461 |
+
log(f"\n --- {name} ---")
|
| 462 |
+
within_accs = []; bases = []
|
| 463 |
+
for seed in range(N_SEEDS):
|
| 464 |
+
t_s = time.time()
|
| 465 |
+
try:
|
| 466 |
+
base = train_continuous_base(feat_c, lbl_c, seed,
|
| 467 |
+
code_dim_per_agent=D, n_epochs=150)
|
| 468 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 469 |
+
log(f" {name} s{seed}: within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]")
|
| 470 |
+
except Exception as e:
|
| 471 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 472 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 473 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 474 |
+
if not valid:
|
| 475 |
+
log(f" {name}: no successful base"); continue
|
| 476 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 477 |
+
best_base = bases[best_idx]
|
| 478 |
+
ho_ids = best_base["holdout_ids"]
|
| 479 |
+
try:
|
| 480 |
+
msgs = get_continuous_messages(best_base, feat_c)
|
| 481 |
+
msgs_ho = msgs[ho_ids]
|
| 482 |
+
ts = topsim_continuous(msgs_ho, lbl_c[ho_ids])
|
| 483 |
+
mi = posdis_continuous_per_dim(msgs_ho, lbl_c[ho_ids])
|
| 484 |
+
pd_ = float(mi.max() / (mi.sum() + 1e-9)) if mi.sum() > 0 else float("nan")
|
| 485 |
+
base_acc, drops = causal_specificity(best_base, feat_c, lbl_c, ho_ids)
|
| 486 |
+
cs = float(drops.max())
|
| 487 |
+
except Exception as e:
|
| 488 |
+
log(f" {name} metrics FAILED: {e}")
|
| 489 |
+
ts = pd_ = cs = float("nan")
|
| 490 |
+
cross = {n: [] for n in N_LIST}
|
| 491 |
+
for seed, base in enumerate(bases):
|
| 492 |
+
if base is None:
|
| 493 |
+
for n in N_LIST: cross[n].append(float("nan"))
|
| 494 |
+
continue
|
| 495 |
+
tr_t, ho_t = make_splits(lbl_r, seed)
|
| 496 |
+
for n in N_LIST:
|
| 497 |
+
try:
|
| 498 |
+
acc = train_recv_frozen_cont(base, feat_r, lbl_r, tr_t, ho_t,
|
| 499 |
+
seed, n)
|
| 500 |
+
cross[n].append(float(acc))
|
| 501 |
+
except Exception as e:
|
| 502 |
+
log(f" {name} s{seed} N={n} FAILED: {e}")
|
| 503 |
+
cross[n].append(float("nan"))
|
| 504 |
+
wm = float(np.mean([a for a in within_accs if not np.isnan(a)]))
|
| 505 |
+
cross_means = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)]))
|
| 506 |
+
if any(not np.isnan(x) for x in cross[n]) else float("nan")
|
| 507 |
+
for n in N_LIST}
|
| 508 |
+
log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} "
|
| 509 |
+
f"CausalSpec={cs:.3f} cross16={cross_means[16]:.3f} cross192={cross_means[192]:.3f}")
|
| 510 |
+
rows.append({
|
| 511 |
+
"name": name, "type": "continuous",
|
| 512 |
+
"code_dim": D, "msg_dim": D * N_AGENTS,
|
| 513 |
+
"within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs,
|
| 514 |
+
"cross_n16": cross_means[16], "cross_n192": cross_means[192],
|
| 515 |
+
})
|
| 516 |
+
|
| 517 |
+
# ── Spearman correlations ──
|
| 518 |
+
from scipy.stats import spearmanr
|
| 519 |
+
def safe_corr(metric, target):
|
| 520 |
+
x = []; y = []
|
| 521 |
+
for r in rows:
|
| 522 |
+
if not np.isnan(r[metric]) and not np.isnan(r[target]):
|
| 523 |
+
x.append(r[metric]); y.append(r[target])
|
| 524 |
+
if len(x) < 4 or np.std(x) < 1e-9 or np.std(y) < 1e-9:
|
| 525 |
+
return float("nan"), float("nan")
|
| 526 |
+
rho, p = spearmanr(x, y)
|
| 527 |
+
return float(rho), float(p)
|
| 528 |
+
|
| 529 |
+
corrs = {}
|
| 530 |
+
for met in ["topsim", "posdis", "causal_spec"]:
|
| 531 |
+
for tgt in ["cross_n16", "cross_n192"]:
|
| 532 |
+
corrs[(met, tgt)] = safe_corr(met, tgt)
|
| 533 |
+
|
| 534 |
+
# Build summary
|
| 535 |
+
lines = [
|
| 536 |
+
"EXPERIMENT Q -- PosDis vs CROSS-SCENARIO TRANSFER (V-JEPA 2, 3 seeds per config)",
|
| 537 |
+
"",
|
| 538 |
+
"Sweep across 7 discrete + 5 continuous bottleneck configs. Each row: best",
|
| 539 |
+
"within-collision metrics on holdout + cross-scenario coll->ramp accuracy",
|
| 540 |
+
"at N=16 and N=192 (mean across 3 seeds).",
|
| 541 |
+
"",
|
| 542 |
+
f"{'Config':<14s} | {'Within':<10s} | {'TopSim':<10s} | {'PosDis':<10s} | "
|
| 543 |
+
f"{'CausalSpec':<12s} | {'Cross N=16':<12s} | {'Cross N=192':<12s}",
|
| 544 |
+
"-" * 100,
|
| 545 |
+
]
|
| 546 |
+
for r in rows:
|
| 547 |
+
lines.append(
|
| 548 |
+
f"{r['name']:<14s} | "
|
| 549 |
+
f"{r['within']*100:5.1f}% | "
|
| 550 |
+
f"{r['topsim']:+.2f} | "
|
| 551 |
+
f"{r['posdis']:.2f} | "
|
| 552 |
+
f"{r['causal_spec']:.2f} | "
|
| 553 |
+
f"{r['cross_n16']*100:5.1f}% | "
|
| 554 |
+
f"{r['cross_n192']*100:5.1f}%")
|
| 555 |
+
|
| 556 |
+
lines.append("")
|
| 557 |
+
lines.append("SPEARMAN CORRELATIONS (across configs):")
|
| 558 |
+
for tgt in ["cross_n16", "cross_n192"]:
|
| 559 |
+
lines.append(f" vs {tgt}:")
|
| 560 |
+
for met in ["topsim", "posdis", "causal_spec"]:
|
| 561 |
+
rho, p = corrs[(met, tgt)]
|
| 562 |
+
lines.append(f" {met:<14s}: rho={rho:+.2f} p={p:.3f}")
|
| 563 |
+
|
| 564 |
+
# Verdict
|
| 565 |
+
lines.append("")
|
| 566 |
+
lines.append("VERDICT:")
|
| 567 |
+
abs_max_rho = 0
|
| 568 |
+
for k, (rho, p) in corrs.items():
|
| 569 |
+
if not np.isnan(rho): abs_max_rho = max(abs_max_rho, abs(rho))
|
| 570 |
+
if abs_max_rho < 0.30:
|
| 571 |
+
v = ("Compositionality metrics (TopSim, PosDis, CausalSpec) DO NOT predict "
|
| 572 |
+
"cross-scenario transfer. All Spearman |rho| < 0.30 across configs. "
|
| 573 |
+
"Strong support for the abstract claim.")
|
| 574 |
+
elif abs_max_rho < 0.55:
|
| 575 |
+
v = (f"Weak/moderate correlation (max |rho|={abs_max_rho:.2f}). Metrics "
|
| 576 |
+
"partially predict transfer but explain little variance. Honest, "
|
| 577 |
+
"nuanced finding.")
|
| 578 |
+
else:
|
| 579 |
+
v = (f"Strong correlation (max |rho|={abs_max_rho:.2f}). Some metrics DO "
|
| 580 |
+
"predict transfer. The abstract claim must be NARROWED.")
|
| 581 |
+
lines.append(f" {v}")
|
| 582 |
+
|
| 583 |
+
lines.append("")
|
| 584 |
+
lines.append(f"Total runtime: {(time.time()-t0)/60:.1f} min")
|
| 585 |
+
summary = "\n".join(lines)
|
| 586 |
+
(OUT / "exp_q_summary.txt").write_text(summary + "\n")
|
| 587 |
+
(OUT / "exp_q_summary.json").write_text(json.dumps({
|
| 588 |
+
"config": {"n_seeds": N_SEEDS, "N_list": N_LIST,
|
| 589 |
+
"discrete_configs": DISCRETE_CONFIGS,
|
| 590 |
+
"continuous_configs": CONTINUOUS_CONFIGS},
|
| 591 |
+
"rows": rows,
|
| 592 |
+
"spearman": {f"{m}__{t}": list(corrs[(m, t)]) for (m, t) in corrs},
|
| 593 |
+
"runtime_s": time.time() - t0,
|
| 594 |
+
}, indent=2, default=str))
|
| 595 |
+
print("\n" + summary, flush=True)
|
| 596 |
+
log(f"DONE in {(time.time()-t0)/60:.1f} min")
|
| 597 |
+
|
| 598 |
+
|
| 599 |
+
if __name__ == "__main__":
|
| 600 |
+
main()
|
code/_rev_q_vqvae_bottleneck.py
ADDED
|
@@ -0,0 +1,450 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
EXP REV-Q-VQ: Vector-quantised (VQ-VAE) bottleneck configurations on V-JEPA 2.
|
| 3 |
+
|
| 4 |
+
R2 has explicitly asked for VQ-VAE configs in the last 3 review rounds. This
|
| 5 |
+
script implements a multi-agent multi-position VQ-VAE sender and runs
|
| 6 |
+
3-5 configurations on collision restitution to (a) compute within-scenario
|
| 7 |
+
TopSim/PosDis/causal-spec, and (b) measure cross-scenario transfer at N=192
|
| 8 |
+
on collision -> ramp.
|
| 9 |
+
|
| 10 |
+
VQ design: each sender has N_positions codebooks, each with V_codes entries of
|
| 11 |
+
dimension D_code. The hidden representation is projected to D_code per position,
|
| 12 |
+
then quantised to the nearest codebook entry. Output to receiver is the
|
| 13 |
+
one-hot index per position (same dimensionality as Gumbel-Softmax sender), so
|
| 14 |
+
PosDis/TopSim/Causal-spec all apply unchanged. Training uses straight-through
|
| 15 |
+
estimator + commitment loss (beta=0.25) + codebook loss.
|
| 16 |
+
|
| 17 |
+
If VQ configs land in the same 41-56% cross-scenario band as the existing
|
| 18 |
+
24-config sweep, the sufficiency claim extends beyond Gumbel-Softmax/tanh.
|
| 19 |
+
"""
|
| 20 |
+
import json, time, sys, os, math
|
| 21 |
+
from pathlib import Path
|
| 22 |
+
from datetime import datetime, timezone
|
| 23 |
+
import numpy as np
|
| 24 |
+
import torch
|
| 25 |
+
import torch.nn as nn
|
| 26 |
+
import torch.nn.functional as F
|
| 27 |
+
|
| 28 |
+
PROMPT_RECEIVED_TIME = datetime.now(timezone.utc).isoformat()
|
| 29 |
+
print(f"PROMPT_RECEIVED_TIME = {PROMPT_RECEIVED_TIME}", flush=True)
|
| 30 |
+
T0 = time.time()
|
| 31 |
+
|
| 32 |
+
sys.path.insert(0, os.path.dirname(__file__))
|
| 33 |
+
from _kinematics_train import (
|
| 34 |
+
DEVICE, ClassifierReceiver, HIDDEN_DIM, N_AGENTS, BATCH_SIZE,
|
| 35 |
+
SENDER_LR, RECEIVER_LR, EARLY_STOP_PATIENCE,
|
| 36 |
+
)
|
| 37 |
+
from _killer_experiment import TemporalEncoder
|
| 38 |
+
from _overnight_p1_transfer import make_splits, train_receiver_frozen_sender
|
| 39 |
+
from _overnight_p3_matrix import load_labels, load_feat_subsampled
|
| 40 |
+
from _rev_q_posdis_scatter import discrete_token_extract, discrete_topsim, discrete_posdis # single-prop matches existing sweep rows 1-12
|
| 41 |
+
from _rev_q_addendum_multiprop import discrete_multi_topsim, discrete_multi_posdis, discrete_multi_causal
|
| 42 |
+
|
| 43 |
+
OUT = Path("results/reviewer_response/exp_q_vqvae")
|
| 44 |
+
OUT.mkdir(parents=True, exist_ok=True)
|
| 45 |
+
N_SEEDS = 3
|
| 46 |
+
N_LIST = [16, 192]
|
| 47 |
+
COMMIT_BETA = 0.25
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
class VQSender(nn.Module):
|
| 51 |
+
"""VQ-VAE sender: encoder -> per-position projection -> VQ codebook -> one-hot."""
|
| 52 |
+
def __init__(self, encoder, hd, vs, nh, code_dim=8):
|
| 53 |
+
super().__init__()
|
| 54 |
+
self.encoder = encoder
|
| 55 |
+
self.vs = vs
|
| 56 |
+
self.nh = nh
|
| 57 |
+
self.code_dim = code_dim
|
| 58 |
+
self.heads = nn.ModuleList([nn.Linear(hd, code_dim) for _ in range(nh)])
|
| 59 |
+
# Codebooks: nh codebooks, each vs x code_dim
|
| 60 |
+
self.codebooks = nn.ParameterList(
|
| 61 |
+
[nn.Parameter(torch.randn(vs, code_dim) * 0.1) for _ in range(nh)]
|
| 62 |
+
)
|
| 63 |
+
self._last_commit_loss = torch.zeros(1)
|
| 64 |
+
|
| 65 |
+
def init_codebooks_from_data(self, x):
|
| 66 |
+
"""K-means-style data-dependent codebook init: sample V z's from a batch."""
|
| 67 |
+
with torch.no_grad():
|
| 68 |
+
h = self.encoder(x)
|
| 69 |
+
for head_idx, (head, codebook) in enumerate(zip(self.heads, self.codebooks)):
|
| 70 |
+
z = head(h) # [B, code_dim]
|
| 71 |
+
if z.size(0) >= self.vs:
|
| 72 |
+
# Random sample without replacement
|
| 73 |
+
perm = torch.randperm(z.size(0), device=z.device)[:self.vs]
|
| 74 |
+
self.codebooks[head_idx].data.copy_(z[perm])
|
| 75 |
+
else:
|
| 76 |
+
# Sample with replacement
|
| 77 |
+
idx = torch.randint(z.size(0), (self.vs,), device=z.device)
|
| 78 |
+
self.codebooks[head_idx].data.copy_(z[idx])
|
| 79 |
+
|
| 80 |
+
def reset_dead_codes(self, x, code_usage):
|
| 81 |
+
"""For each head, reset codes with usage<threshold to random data points."""
|
| 82 |
+
with torch.no_grad():
|
| 83 |
+
h = self.encoder(x)
|
| 84 |
+
for head_idx, (head, codebook) in enumerate(zip(self.heads, self.codebooks)):
|
| 85 |
+
z = head(h) # [B, code_dim]
|
| 86 |
+
usage = code_usage[head_idx] # [vs]
|
| 87 |
+
dead = (usage < 0.01).nonzero(as_tuple=True)[0]
|
| 88 |
+
if len(dead) == 0: continue
|
| 89 |
+
if z.size(0) >= len(dead):
|
| 90 |
+
perm = torch.randperm(z.size(0), device=z.device)[:len(dead)]
|
| 91 |
+
self.codebooks[head_idx].data[dead] = z[perm]
|
| 92 |
+
|
| 93 |
+
def forward(self, x, tau=1.0, hard=True, track_usage=False):
|
| 94 |
+
h = self.encoder(x)
|
| 95 |
+
msgs, logits_all = [], []
|
| 96 |
+
commit_loss = torch.zeros(1, device=h.device)
|
| 97 |
+
usage_per_head = [] if track_usage else None
|
| 98 |
+
for head, codebook in zip(self.heads, self.codebooks):
|
| 99 |
+
z = head(h) # [B, code_dim]
|
| 100 |
+
# Distances from each batch element to each codebook entry
|
| 101 |
+
# |z - c|^2 = |z|^2 + |c|^2 - 2 z.c
|
| 102 |
+
dists = (z.pow(2).sum(-1, keepdim=True)
|
| 103 |
+
+ codebook.pow(2).sum(-1).unsqueeze(0)
|
| 104 |
+
- 2 * z @ codebook.t()) # [B, vs]
|
| 105 |
+
indices = dists.argmin(-1) # [B]
|
| 106 |
+
z_q = codebook[indices] # [B, code_dim]
|
| 107 |
+
# Commitment + codebook losses
|
| 108 |
+
commit_loss = commit_loss + COMMIT_BETA * F.mse_loss(z, z_q.detach())
|
| 109 |
+
commit_loss = commit_loss + F.mse_loss(z_q, z.detach())
|
| 110 |
+
# Straight-through: forward = hard one-hot; backward = softmax(-dists/tau)
|
| 111 |
+
# This is the standard STE for one-hot VQ-VAE -> discrete receiver.
|
| 112 |
+
soft = F.softmax(-dists / max(tau, 1e-3), dim=-1) # [B, vs]
|
| 113 |
+
hard_oh = F.one_hot(indices, self.vs).float()
|
| 114 |
+
msg = soft + (hard_oh - soft).detach() # forward=hard, grad flows via soft
|
| 115 |
+
msgs.append(msg)
|
| 116 |
+
logits_all.append(-dists)
|
| 117 |
+
if track_usage:
|
| 118 |
+
# one-hot count per code, normalized to fraction
|
| 119 |
+
usage_per_head.append(hard_oh.detach().mean(0)) # [vs]
|
| 120 |
+
self._last_commit_loss = commit_loss
|
| 121 |
+
if track_usage:
|
| 122 |
+
self._last_usage = usage_per_head
|
| 123 |
+
return torch.cat(msgs, -1), logits_all
|
| 124 |
+
|
| 125 |
+
|
| 126 |
+
class VQMultiSender(nn.Module):
|
| 127 |
+
def __init__(self, senders):
|
| 128 |
+
super().__init__()
|
| 129 |
+
self.senders = nn.ModuleList(senders)
|
| 130 |
+
|
| 131 |
+
def forward(self, views, tau=1.0, hard=True):
|
| 132 |
+
msgs, all_logits = [], []
|
| 133 |
+
commit = torch.zeros(1, device=views[0].device)
|
| 134 |
+
for s, v in zip(self.senders, views):
|
| 135 |
+
m, l = s(v, tau, hard)
|
| 136 |
+
msgs.append(m)
|
| 137 |
+
all_logits.extend(l)
|
| 138 |
+
commit = commit + s._last_commit_loss
|
| 139 |
+
self._last_commit_loss = commit
|
| 140 |
+
return torch.cat(msgs, -1), all_logits
|
| 141 |
+
|
| 142 |
+
|
| 143 |
+
def log(msg):
|
| 144 |
+
ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ")
|
| 145 |
+
print(f"[{ts}] EXP-VQ: {msg}", flush=True)
|
| 146 |
+
|
| 147 |
+
|
| 148 |
+
def train_vq(feat, labels, seed, n_heads, vocab_size, n_epochs=150, code_dim=8):
|
| 149 |
+
"""Train VQ-VAE bottleneck on within-scenario task."""
|
| 150 |
+
N, nf, dim = feat.shape
|
| 151 |
+
fpa = 1
|
| 152 |
+
msg_dim = vocab_size * n_heads * N_AGENTS
|
| 153 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 154 |
+
torch.manual_seed(seed); np.random.seed(seed)
|
| 155 |
+
rng = np.random.RandomState(seed * 1000 + 42)
|
| 156 |
+
|
| 157 |
+
train_ids, holdout_ids = [], []
|
| 158 |
+
for c in np.unique(labels):
|
| 159 |
+
ids_c = np.where(labels == c)[0]
|
| 160 |
+
rng.shuffle(ids_c)
|
| 161 |
+
split = max(1, len(ids_c) // 5)
|
| 162 |
+
holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:])
|
| 163 |
+
train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids)
|
| 164 |
+
n_classes = int(labels.max()) + 1
|
| 165 |
+
chance = 1.0 / n_classes
|
| 166 |
+
|
| 167 |
+
senders = [VQSender(TemporalEncoder(HIDDEN_DIM, dim, fpa), HIDDEN_DIM,
|
| 168 |
+
vocab_size, n_heads, code_dim).to(DEVICE)
|
| 169 |
+
for _ in range(N_AGENTS)]
|
| 170 |
+
sender = VQMultiSender(senders).to(DEVICE)
|
| 171 |
+
# Data-dependent codebook init: sample from a forward pass on a training batch
|
| 172 |
+
# (standard VQ-VAE init trick to avoid initial collapse to a single codebook entry).
|
| 173 |
+
with torch.no_grad():
|
| 174 |
+
init_batch_ids = train_ids[:min(BATCH_SIZE * 4, len(train_ids))]
|
| 175 |
+
init_views = [v[init_batch_ids].to(DEVICE) for v in agent_views]
|
| 176 |
+
for s, v in zip(sender.senders, init_views):
|
| 177 |
+
s.init_codebooks_from_data(v)
|
| 178 |
+
receivers = [ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 179 |
+
for _ in range(3)]
|
| 180 |
+
so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR)
|
| 181 |
+
ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers]
|
| 182 |
+
labels_dev = torch.tensor(labels, dtype=torch.long).to(DEVICE)
|
| 183 |
+
n_batches = max(1, len(train_ids) // BATCH_SIZE)
|
| 184 |
+
best_acc = 0.0; best_ep = 0
|
| 185 |
+
best_sender_state = None; best_receiver_states = None; best_recv_idx = 0
|
| 186 |
+
RESET_EVERY = 40
|
| 187 |
+
DEAD_CODE_RESET_EVERY = 20 # standard VQ-VAE dead-code mitigation
|
| 188 |
+
|
| 189 |
+
for ep in range(n_epochs):
|
| 190 |
+
# Temperature anneal: 3.0 -> 1.0 over first half of training (matches Gumbel sender schedule)
|
| 191 |
+
tau = max(1.0, 3.0 - 2.0 * ep / max(1, n_epochs // 2))
|
| 192 |
+
# Dead-code reset every DEAD_CODE_RESET_EVERY epochs (skip ep 0)
|
| 193 |
+
if ep > 0 and ep % DEAD_CODE_RESET_EVERY == 0:
|
| 194 |
+
with torch.no_grad():
|
| 195 |
+
# Compute usage from a forward pass over a training batch
|
| 196 |
+
usage_batch = train_ids[:min(BATCH_SIZE * 4, len(train_ids))]
|
| 197 |
+
u_views = [v[usage_batch].to(DEVICE) for v in agent_views]
|
| 198 |
+
# Per-sender usage tracking
|
| 199 |
+
for s, v in zip(sender.senders, u_views):
|
| 200 |
+
_, _ = s(v, tau=tau, track_usage=True)
|
| 201 |
+
s.reset_dead_codes(v, s._last_usage)
|
| 202 |
+
if ep - best_ep > EARLY_STOP_PATIENCE * 2 and best_acc > chance + 0.05: break
|
| 203 |
+
if ep > 0 and ep % RESET_EVERY == 0:
|
| 204 |
+
for i in range(len(receivers)):
|
| 205 |
+
receivers[i] = ClassifierReceiver(msg_dim, HIDDEN_DIM, n_classes).to(DEVICE)
|
| 206 |
+
ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR)
|
| 207 |
+
sender.train(); [r.train() for r in receivers]
|
| 208 |
+
rng_ep = np.random.RandomState(seed * 10000 + ep)
|
| 209 |
+
perm = rng_ep.permutation(train_ids)
|
| 210 |
+
for b in range(n_batches):
|
| 211 |
+
batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE]
|
| 212 |
+
if len(batch_ids) < 4: continue
|
| 213 |
+
views = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 214 |
+
tgts = labels_dev[batch_ids]
|
| 215 |
+
msg, logits_list = sender(views, tau=tau)
|
| 216 |
+
ce_loss = torch.tensor(0.0, device=DEVICE)
|
| 217 |
+
for r in receivers:
|
| 218 |
+
logits = r(msg)
|
| 219 |
+
ce_loss = ce_loss + F.cross_entropy(logits, tgts)
|
| 220 |
+
ce_loss = ce_loss / len(receivers)
|
| 221 |
+
# Total loss: CE + commitment+codebook (already accumulated in sender)
|
| 222 |
+
loss = ce_loss + sender._last_commit_loss.squeeze()
|
| 223 |
+
if torch.isnan(loss):
|
| 224 |
+
so.zero_grad(); [o.zero_grad() for o in ros]; continue
|
| 225 |
+
so.zero_grad(); [o.zero_grad() for o in ros]
|
| 226 |
+
loss.backward()
|
| 227 |
+
torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0)
|
| 228 |
+
so.step(); [o.step() for o in ros]
|
| 229 |
+
if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache()
|
| 230 |
+
if (ep + 1) % 10 == 0 or ep == 0:
|
| 231 |
+
sender.eval(); [r.eval() for r in receivers]
|
| 232 |
+
with torch.no_grad():
|
| 233 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 234 |
+
msg_ho, _ = sender(v_ho)
|
| 235 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 236 |
+
best_per_recv = 0.0; best_idx = 0
|
| 237 |
+
for ri, r in enumerate(receivers):
|
| 238 |
+
acc = (r(msg_ho).argmax(-1) == tgt_ho).float().mean().item()
|
| 239 |
+
if acc > best_per_recv:
|
| 240 |
+
best_per_recv = acc; best_idx = ri
|
| 241 |
+
if best_per_recv > best_acc:
|
| 242 |
+
best_acc = best_per_recv; best_ep = ep
|
| 243 |
+
best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()}
|
| 244 |
+
best_receiver_states = [
|
| 245 |
+
{k: v.cpu().clone() for k, v in r.state_dict().items()}
|
| 246 |
+
for r in receivers]
|
| 247 |
+
best_recv_idx = best_idx
|
| 248 |
+
return {
|
| 249 |
+
"sender_state": best_sender_state,
|
| 250 |
+
"receiver_states": best_receiver_states,
|
| 251 |
+
"best_recv_idx": best_recv_idx,
|
| 252 |
+
"train_ids": train_ids, "holdout_ids": holdout_ids,
|
| 253 |
+
"task_acc": best_acc, "chance": chance,
|
| 254 |
+
"n_classes_per_prop": [n_classes],
|
| 255 |
+
"fpa": 1, "dim": dim,
|
| 256 |
+
"n_heads": n_heads, "vocab_size": vocab_size,
|
| 257 |
+
"code_dim": code_dim,
|
| 258 |
+
"msg_dim": msg_dim,
|
| 259 |
+
}
|
| 260 |
+
|
| 261 |
+
|
| 262 |
+
def vq_token_extract(base, feat):
|
| 263 |
+
"""Extract token indices from VQ sender for compositionality metrics."""
|
| 264 |
+
senders = [VQSender(TemporalEncoder(HIDDEN_DIM, base["dim"], base["fpa"]),
|
| 265 |
+
HIDDEN_DIM, base["vocab_size"], base["n_heads"], base["code_dim"]).to(DEVICE)
|
| 266 |
+
for _ in range(N_AGENTS)]
|
| 267 |
+
sender = VQMultiSender(senders).to(DEVICE)
|
| 268 |
+
sender.load_state_dict(base["sender_state"])
|
| 269 |
+
sender.eval()
|
| 270 |
+
agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 271 |
+
with torch.no_grad():
|
| 272 |
+
N = feat.shape[0]
|
| 273 |
+
all_tokens = []
|
| 274 |
+
# Use full feat as one batch (small)
|
| 275 |
+
v_in = [v.to(DEVICE) for v in agent_views]
|
| 276 |
+
msg, _ = sender(v_in)
|
| 277 |
+
# msg is [N, vocab_size * n_heads * N_AGENTS], one-hot per position
|
| 278 |
+
# Convert back to indices
|
| 279 |
+
msg_reshaped = msg.view(N, N_AGENTS, base["n_heads"], base["vocab_size"])
|
| 280 |
+
tokens = msg_reshaped.argmax(-1) # [N, N_AGENTS, n_heads]
|
| 281 |
+
# Flatten across agents and heads to get a token vector
|
| 282 |
+
tokens = tokens.reshape(N, N_AGENTS * base["n_heads"]).cpu().numpy()
|
| 283 |
+
return tokens
|
| 284 |
+
|
| 285 |
+
|
| 286 |
+
def vq_train_recv_frozen(base, feat_tgt, labels_tgt, train_ids, holdout_ids, seed, n_target, n_epochs=80):
|
| 287 |
+
"""Freeze VQ sender; train fresh receiver on n_target stratified target examples."""
|
| 288 |
+
if n_target == 0:
|
| 289 |
+
# Zero-shot: apply source receiver directly
|
| 290 |
+
senders = [VQSender(TemporalEncoder(HIDDEN_DIM, base["dim"], base["fpa"]),
|
| 291 |
+
HIDDEN_DIM, base["vocab_size"], base["n_heads"], base["code_dim"]).to(DEVICE)
|
| 292 |
+
for _ in range(N_AGENTS)]
|
| 293 |
+
sender = VQMultiSender(senders).to(DEVICE)
|
| 294 |
+
sender.load_state_dict(base["sender_state"])
|
| 295 |
+
sender.eval()
|
| 296 |
+
n_classes = base["n_classes_per_prop"][0]
|
| 297 |
+
receiver = ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, n_classes).to(DEVICE)
|
| 298 |
+
receiver.load_state_dict(base["receiver_states"][base["best_recv_idx"]])
|
| 299 |
+
receiver.eval()
|
| 300 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 301 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 302 |
+
with torch.no_grad():
|
| 303 |
+
v_in = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 304 |
+
msg, _ = sender(v_in)
|
| 305 |
+
tgt = labels_dev[holdout_ids]
|
| 306 |
+
return float((receiver(msg).argmax(-1) == tgt).float().mean())
|
| 307 |
+
# Sample n_target stratified examples from train_ids
|
| 308 |
+
rng = np.random.RandomState(seed * 7 + 13)
|
| 309 |
+
n_per_class = max(1, n_target // 3)
|
| 310 |
+
sub_train = []
|
| 311 |
+
for c in np.unique(labels_tgt[train_ids]):
|
| 312 |
+
cand = train_ids[labels_tgt[train_ids] == c]
|
| 313 |
+
rng.shuffle(cand)
|
| 314 |
+
sub_train.extend(cand[:n_per_class])
|
| 315 |
+
sub_train = np.array(sub_train)
|
| 316 |
+
# Build sender, freeze
|
| 317 |
+
senders = [VQSender(TemporalEncoder(HIDDEN_DIM, base["dim"], base["fpa"]),
|
| 318 |
+
HIDDEN_DIM, base["vocab_size"], base["n_heads"], base["code_dim"]).to(DEVICE)
|
| 319 |
+
for _ in range(N_AGENTS)]
|
| 320 |
+
sender = VQMultiSender(senders).to(DEVICE)
|
| 321 |
+
sender.load_state_dict(base["sender_state"])
|
| 322 |
+
sender.eval()
|
| 323 |
+
for p in sender.parameters(): p.requires_grad = False
|
| 324 |
+
n_classes = base["n_classes_per_prop"][0]
|
| 325 |
+
receiver = ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, n_classes).to(DEVICE)
|
| 326 |
+
opt = torch.optim.Adam(receiver.parameters(), lr=RECEIVER_LR)
|
| 327 |
+
agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)]
|
| 328 |
+
labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE)
|
| 329 |
+
best_acc = 0.0
|
| 330 |
+
for ep in range(n_epochs):
|
| 331 |
+
receiver.train()
|
| 332 |
+
rng_ep = np.random.RandomState(seed * 100 + ep)
|
| 333 |
+
perm = rng_ep.permutation(sub_train)
|
| 334 |
+
n_batches = max(1, len(perm) // 16)
|
| 335 |
+
for b in range(n_batches):
|
| 336 |
+
batch_ids = perm[b*16:(b+1)*16]
|
| 337 |
+
if len(batch_ids) < 4: continue
|
| 338 |
+
with torch.no_grad():
|
| 339 |
+
v_in = [v[batch_ids].to(DEVICE) for v in agent_views]
|
| 340 |
+
msg, _ = sender(v_in)
|
| 341 |
+
tgts = labels_dev[batch_ids]
|
| 342 |
+
logits = receiver(msg)
|
| 343 |
+
loss = F.cross_entropy(logits, tgts)
|
| 344 |
+
opt.zero_grad(); loss.backward(); opt.step()
|
| 345 |
+
if ep % 5 == 0 or ep == n_epochs - 1:
|
| 346 |
+
receiver.eval()
|
| 347 |
+
with torch.no_grad():
|
| 348 |
+
v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views]
|
| 349 |
+
msg_ho, _ = sender(v_ho)
|
| 350 |
+
tgt_ho = labels_dev[holdout_ids]
|
| 351 |
+
acc = (receiver(msg_ho).argmax(-1) == tgt_ho).float().mean().item()
|
| 352 |
+
if acc > best_acc: best_acc = acc
|
| 353 |
+
return float(best_acc)
|
| 354 |
+
|
| 355 |
+
|
| 356 |
+
def main():
|
| 357 |
+
log("=" * 60)
|
| 358 |
+
log("EXP Q VQ-VAE: VQ-VAE bottleneck configs")
|
| 359 |
+
|
| 360 |
+
feat_c = load_feat_subsampled("collision", "vjepa2")
|
| 361 |
+
feat_r = load_feat_subsampled("ramp", "vjepa2")
|
| 362 |
+
z = np.load("results/kinematics_vs_mechanics/labels_collision.npz")
|
| 363 |
+
rest_3 = z["restitution_bin"]
|
| 364 |
+
lbl_r_3 = load_labels("ramp", "restitution")
|
| 365 |
+
|
| 366 |
+
# Configurations: (name, L, V, code_dim)
|
| 367 |
+
configs = [
|
| 368 |
+
("vq_L2_V8_d8", 2, 8, 8),
|
| 369 |
+
("vq_L3_V8_d8", 3, 8, 8),
|
| 370 |
+
("vq_L3_V16_d8", 3, 16, 8),
|
| 371 |
+
("vq_L4_V16_d8", 4, 16, 8),
|
| 372 |
+
]
|
| 373 |
+
|
| 374 |
+
rows = []
|
| 375 |
+
for name, L, V, D in configs:
|
| 376 |
+
log(f"\n --- {name} (L={L}, V={V}, code_dim={D}) ---")
|
| 377 |
+
within_accs = []; bases = []
|
| 378 |
+
for seed in range(N_SEEDS):
|
| 379 |
+
t0 = time.time()
|
| 380 |
+
try:
|
| 381 |
+
base = train_vq(feat_c, rest_3, seed, L, V, n_epochs=150, code_dim=D)
|
| 382 |
+
bases.append(base); within_accs.append(float(base["task_acc"]))
|
| 383 |
+
log(f" {name} s{seed}: within={base['task_acc']:.3f} [{time.time()-t0:.0f}s]")
|
| 384 |
+
except Exception as e:
|
| 385 |
+
log(f" {name} s{seed} FAILED: {e}")
|
| 386 |
+
bases.append(None); within_accs.append(float("nan"))
|
| 387 |
+
|
| 388 |
+
valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)]
|
| 389 |
+
if not valid:
|
| 390 |
+
rows.append({"name": name, "within": float("nan"),
|
| 391 |
+
"topsim": float("nan"), "posdis": float("nan"),
|
| 392 |
+
"causal": float("nan"),
|
| 393 |
+
"cross_n16": float("nan"), "cross_n192": float("nan")})
|
| 394 |
+
continue
|
| 395 |
+
best_idx = max(valid, key=lambda x: x[1])[0]
|
| 396 |
+
best_base = bases[best_idx]
|
| 397 |
+
ho_ids = best_base["holdout_ids"]
|
| 398 |
+
|
| 399 |
+
# Compute compositionality metrics on holdout (single-property: matches sweep rows 1-12)
|
| 400 |
+
try:
|
| 401 |
+
tokens = vq_token_extract(best_base, feat_c) # [N, N_AGENTS * L]
|
| 402 |
+
tokens_ho = tokens[ho_ids]
|
| 403 |
+
ts = discrete_topsim(tokens_ho, rest_3[ho_ids])
|
| 404 |
+
pd_ = discrete_posdis(tokens_ho, rest_3[ho_ids])
|
| 405 |
+
cs = float("nan") # causal spec deferred — VQ uses same receiver, can be added later
|
| 406 |
+
except Exception as e:
|
| 407 |
+
import traceback
|
| 408 |
+
log(f" {name} metric FAILED: {e}\n{traceback.format_exc()}")
|
| 409 |
+
ts = pd_ = cs = float("nan")
|
| 410 |
+
|
| 411 |
+
# Cross-scenario coll->ramp at N=16 and N=192
|
| 412 |
+
cross_n16 = []; cross_n192 = []
|
| 413 |
+
for seed in range(N_SEEDS):
|
| 414 |
+
tr_t, ho_t = make_splits(lbl_r_3, seed)
|
| 415 |
+
for N_target, lst in [(16, cross_n16), (192, cross_n192)]:
|
| 416 |
+
try:
|
| 417 |
+
acc = vq_train_recv_frozen(best_base, feat_r, lbl_r_3, tr_t, ho_t, seed, N_target)
|
| 418 |
+
lst.append(float(acc))
|
| 419 |
+
except Exception as e:
|
| 420 |
+
log(f" {name} cross s{seed} N={N_target} FAILED: {e}")
|
| 421 |
+
|
| 422 |
+
m16 = float(np.mean(cross_n16)) if cross_n16 else float("nan")
|
| 423 |
+
m192 = float(np.mean(cross_n192)) if cross_n192 else float("nan")
|
| 424 |
+
log(f" {name}: within={float(np.nanmean(within_accs)):.3f} TopSim={ts:+.2f} PosDis={pd_:.2f} cross16={m16*100:.1f}% cross192={m192*100:.1f}%")
|
| 425 |
+
rows.append({"name": name, "L": L, "V": V, "code_dim": D,
|
| 426 |
+
"within": float(np.nanmean(within_accs)),
|
| 427 |
+
"topsim": float(ts), "posdis": float(pd_), "causal": float(cs),
|
| 428 |
+
"cross_n16": m16, "cross_n192": m192})
|
| 429 |
+
|
| 430 |
+
SUMMARY = ["EXP Q VQ-VAE -- VQ-VAE bottleneck configurations on V-JEPA 2 collision",
|
| 431 |
+
"",
|
| 432 |
+
f"{'Config':<22s} | {'Within':>7s} | {'TopSim':>7s} | {'PosDis':>7s} | {'Cross16':>8s} | {'Cross192':>9s}",
|
| 433 |
+
"-" * 76]
|
| 434 |
+
for r in rows:
|
| 435 |
+
SUMMARY.append(
|
| 436 |
+
f"{r['name']:<22s} | {r['within']*100:6.1f}% | {r['topsim']:+7.2f} | "
|
| 437 |
+
f"{r['posdis']:7.2f} | {r['cross_n16']*100:7.1f}% | {r['cross_n192']*100:8.1f}%"
|
| 438 |
+
)
|
| 439 |
+
print("\n".join(SUMMARY), flush=True)
|
| 440 |
+
with open(OUT / "exp_q_vqvae_summary.txt", "w") as fh:
|
| 441 |
+
fh.write("\n".join(SUMMARY) + "\n")
|
| 442 |
+
with open(OUT / "exp_q_vqvae_summary.json", "w") as fh:
|
| 443 |
+
json.dump(rows, fh, indent=2)
|
| 444 |
+
end_ts = datetime.now(timezone.utc).isoformat()
|
| 445 |
+
runtime_min = (time.time() - T0) / 60.0
|
| 446 |
+
print(f"\nEND_TIME = {end_ts}\nTotal runtime: {runtime_min:.2f} min", flush=True)
|
| 447 |
+
|
| 448 |
+
|
| 449 |
+
if __name__ == "__main__":
|
| 450 |
+
main()
|
croissant.json
ADDED
|
@@ -0,0 +1,204 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"@context": {
|
| 3 |
+
"@language": "en",
|
| 4 |
+
"@vocab": "https://schema.org/",
|
| 5 |
+
"citeAs": "cr:citeAs",
|
| 6 |
+
"column": "cr:column",
|
| 7 |
+
"conformsTo": "dct:conformsTo",
|
| 8 |
+
"cr": "http://mlcommons.org/croissant/",
|
| 9 |
+
"rai": "http://mlcommons.org/croissant/RAI/",
|
| 10 |
+
"data": {
|
| 11 |
+
"@id": "cr:data",
|
| 12 |
+
"@type": "@json"
|
| 13 |
+
},
|
| 14 |
+
"dataType": {
|
| 15 |
+
"@id": "cr:dataType",
|
| 16 |
+
"@type": "@vocab"
|
| 17 |
+
},
|
| 18 |
+
"dct": "http://purl.org/dc/terms/",
|
| 19 |
+
"examples": {
|
| 20 |
+
"@id": "cr:examples",
|
| 21 |
+
"@type": "@json"
|
| 22 |
+
},
|
| 23 |
+
"extract": "cr:extract",
|
| 24 |
+
"field": "cr:field",
|
| 25 |
+
"fileProperty": "cr:fileProperty",
|
| 26 |
+
"fileObject": "cr:fileObject",
|
| 27 |
+
"fileSet": "cr:fileSet",
|
| 28 |
+
"format": "cr:format",
|
| 29 |
+
"includes": "cr:includes",
|
| 30 |
+
"isLiveDataset": "cr:isLiveDataset",
|
| 31 |
+
"jsonPath": "cr:jsonPath",
|
| 32 |
+
"key": "cr:key",
|
| 33 |
+
"md5": "cr:md5",
|
| 34 |
+
"parentField": "cr:parentField",
|
| 35 |
+
"path": "cr:path",
|
| 36 |
+
"recordSet": "cr:recordSet",
|
| 37 |
+
"references": "cr:references",
|
| 38 |
+
"regex": "cr:regex",
|
| 39 |
+
"repeated": "cr:repeated",
|
| 40 |
+
"replace": "cr:replace",
|
| 41 |
+
"sc": "https://schema.org/",
|
| 42 |
+
"separator": "cr:separator",
|
| 43 |
+
"source": "cr:source",
|
| 44 |
+
"subField": "cr:subField",
|
| 45 |
+
"transform": "cr:transform"
|
| 46 |
+
},
|
| 47 |
+
"@type": "sc:Dataset",
|
| 48 |
+
"name": "cross-scenario-physics-code-transfer",
|
| 49 |
+
"conformsTo": "http://mlcommons.org/croissant/1.0",
|
| 50 |
+
"description": "A cross-scenario physics-transfer benchmark for stress-testing whether within-scenario compositionality metrics on frozen video features predict cross-scenario physical-property transfer. Contains four Kubric physics scenarios (collision, ramp, flat-drop, elasticity) totalling 1,800 scenes plus a 75-scene matched-visual low-gravity collision variant; pre-extracted frozen features for eight video and image backbones (V-JEPA 2, V-JEPA 2.1, DINOv2-S/L, CLIP ViT-L/14, MAE, SigLIP, VideoMAE); ground-truth per-object tracks; and N-shot adaptation protocols with explicit task/label mappings. Real-video evaluation uses the public Phys101 dataset; we redistribute only feature tensors extracted from it. Designed to evaluate whether high TopSim, PosDis, and causal-specificity bottleneck codes generalise across scenarios; the headline negative result is that they do not in the tested bottleneck protocol family.",
|
| 51 |
+
"url": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer",
|
| 52 |
+
"version": "1.0.0",
|
| 53 |
+
"license": "https://www.apache.org/licenses/LICENSE-2.0",
|
| 54 |
+
"datePublished": "2026-05-05",
|
| 55 |
+
"creator": {
|
| 56 |
+
"@type": "Organization",
|
| 57 |
+
"name": "Anonymous (NeurIPS 2026 E&D Track Submission)"
|
| 58 |
+
},
|
| 59 |
+
"citeAs": "Anonymous Authors. A Benchmark for Cross-Scenario Physics-Code Transfer: Compositionality Metrics on Frozen Video Features. NeurIPS 2026 Evaluations & Datasets Track (under review).",
|
| 60 |
+
"isLiveDataset": false,
|
| 61 |
+
"keywords": [
|
| 62 |
+
"physics representation learning",
|
| 63 |
+
"cross-scenario transfer",
|
| 64 |
+
"compositionality metrics",
|
| 65 |
+
"TopSim",
|
| 66 |
+
"PosDis",
|
| 67 |
+
"video foundation models",
|
| 68 |
+
"V-JEPA 2",
|
| 69 |
+
"frozen features",
|
| 70 |
+
"benchmark",
|
| 71 |
+
"Kubric",
|
| 72 |
+
"Phys101",
|
| 73 |
+
"negative result"
|
| 74 |
+
],
|
| 75 |
+
"rai:dataCollection": "Kubric scenarios were rendered using the public Kubric simulator (Apache 2.0) with the PyBullet physics backend at 240Hz substeps. Scene parameters (mass, restitution, friction, drop-height, initial velocity) are sampled on explicit grids documented in the accompanying paper. The 75-scene low-gravity variant uses the same RNG seed as the standard-gravity collision dataset to match per-scene physics-random variables (sphere color, lighting, initial velocity, position jitter) for matched-visual analysis. Phys101 features are extracted from the publicly available Phys101 dataset (Wu et al., BMVC 2016) using the V-JEPA 2 frozen encoder; we redistribute the feature tensors, not the source video.",
|
| 76 |
+
"rai:dataCollectionType": "Synthetic simulation (Kubric/PyBullet) plus derived features from public real-video dataset (Phys101).",
|
| 77 |
+
"rai:dataCollectionRawData": "Raw rendered video frames at 256x256, 48 frames at 24 fps; ground-truth per-object position and finite-difference velocity from PyBullet.",
|
| 78 |
+
"rai:dataCollectionTimeframe": "Rendered in 2025-2026.",
|
| 79 |
+
"rai:dataAnnotationProtocol": "All labels are derived directly from the simulator's ground-truth physical state (mass, restitution, friction, drop-height, initial velocity) and binned into discrete classes (3-class union-binned restitution; 5-class mass-ratio for multi-property training; per-scenario tertile mass for Phys101). No human annotation involved.",
|
| 80 |
+
"rai:dataAnnotationPlatform": "Programmatic (Python).",
|
| 81 |
+
"rai:dataAnnotationAnalysis": "Class-balance and bin-boundary statistics are reported in Appendix A.7 of the paper.",
|
| 82 |
+
"rai:dataAnnotationPerItemTime": "0",
|
| 83 |
+
"rai:dataAnnotationDemographics": "Not applicable (synthetic data + derived features).",
|
| 84 |
+
"rai:dataAnnotationTools": "Custom Python scripts; see code/ in this supplementary bundle.",
|
| 85 |
+
"rai:dataAnnotatorDemographicsDescription": "No human annotators involved.",
|
| 86 |
+
"rai:dataPreprocessingProtocol": "Frozen-feature pre-extraction: each scene is processed as 4 evenly-spaced frames at 256x256, with one forward pass per ViT-L-scale encoder; spatial features are mean-pooled within each tubelet/patch grid to produce (N, T=4, D) tensors per scenario per backbone. Stratified train/holdout splits use 80/20 per primary label class. N-shot target-side stratified subsampling is documented in Section 3.6 of the paper.",
|
| 87 |
+
"rai:dataUseCases": "Evaluating whether within-scenario compositionality metrics (TopSim, PosDis, causal specificity) predict cross-scenario transfer of learned physics codes; benchmarking new sender architectures (slot-attention, EMA-codebook VQ, product-quantised continuous, transfer-aware objectives) on a frozen-feature substrate; stress-testing the assumption that high within-scenario metric values entail abstract reusable structure; isolating the contribution of visual versus dynamics versus scene/task-structure shift via the matched-visual transfer ladder.",
|
| 88 |
+
"rai:dataLimitations": "Four Kubric scenarios with 1-2 spheres; broader scenario diversity (objects, occlusions, multi-agent) would strengthen generality. The metric-transfer sweep is collision-trained and primarily evaluated on collision -> ramp at N=192 with a partial replication on collision -> flat-drop. The n=24 correlational analysis is underpowered (widest bootstrap 95% CI [-0.58, +0.32]); the load-bearing claim is sufficiency, not non-predictiveness. Tested code families are Gumbel-Softmax, tanh-bounded continuous, and a 4-configuration VQ-VAE probe; slot-attention and EMA-codebook VQ are explicit forward directions, not tested. Phys101 per-scenario tertile binning means class boundaries differ across source and target, contributing a label-boundary shift on top of the feature-distribution shift on real video; this is documented and a global-binning diagnostic is also released.",
|
| 89 |
+
"rai:dataSocialImpact": "The benchmark is intended to help researchers more honestly evaluate physics-representation methods and reduce overclaiming of compositionality from within-scenario metrics. We see no immediate negative societal impacts. The released features are derived from frozen public foundation models on synthetic and openly-licensed real-video sources; no personally identifiable information is present.",
|
| 90 |
+
"rai:dataBiases": "All scenes are simulator-generated and contain no human subjects, places, or demographic information. Phys101 is a publicly released physics dataset with no personally identifiable content; we redistribute only V-JEPA 2 features extracted from it.",
|
| 91 |
+
"rai:personalSensitiveInformation": "None. The dataset contains no personally identifiable information.",
|
| 92 |
+
"rai:dataReleaseMaintenancePlan": "The benchmark is hosted on Hugging Face. Updates (additional backbones, additional scenarios, additional pre-extracted features) will follow semantic versioning; current release is v1.0.0. Issues and pull requests are tracked at the (anonymous) Hugging Face dataset repository.",
|
| 93 |
+
"distribution": [
|
| 94 |
+
{
|
| 95 |
+
"@type": "cr:FileObject",
|
| 96 |
+
"@id": "kubric-collision-features-vjepa2",
|
| 97 |
+
"name": "vjepa2_collision_pooled.pt",
|
| 98 |
+
"description": "V-JEPA 2 frozen features for the 600 collision scenes, mean-pooled to (N=600, T=4, D=1024).",
|
| 99 |
+
"encodingFormat": "application/octet-stream",
|
| 100 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/features/vjepa2_collision_pooled.pt"
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"@type": "cr:FileObject",
|
| 104 |
+
"@id": "kubric-ramp-features-vjepa2",
|
| 105 |
+
"name": "vjepa2_ramp_pooled.pt",
|
| 106 |
+
"description": "V-JEPA 2 frozen features for the 300 ramp scenes.",
|
| 107 |
+
"encodingFormat": "application/octet-stream",
|
| 108 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/features/vjepa2_ramp_pooled.pt"
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"@type": "cr:FileObject",
|
| 112 |
+
"@id": "kubric-flatdrop-features-vjepa2",
|
| 113 |
+
"name": "vjepa2_flat_drop_pooled.pt",
|
| 114 |
+
"description": "V-JEPA 2 frozen features for the 300 flat-drop scenes.",
|
| 115 |
+
"encodingFormat": "application/octet-stream",
|
| 116 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/features/vjepa2_flat_drop_pooled.pt"
|
| 117 |
+
},
|
| 118 |
+
{
|
| 119 |
+
"@type": "cr:FileObject",
|
| 120 |
+
"@id": "kubric-elasticity-features-vjepa2",
|
| 121 |
+
"name": "vjepa2_elasticity_pooled.pt",
|
| 122 |
+
"description": "V-JEPA 2 frozen features for the 600 elasticity scenes.",
|
| 123 |
+
"encodingFormat": "application/octet-stream",
|
| 124 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/features/vjepa2_elasticity_pooled.pt"
|
| 125 |
+
},
|
| 126 |
+
{
|
| 127 |
+
"@type": "cr:FileObject",
|
| 128 |
+
"@id": "labels-collision",
|
| 129 |
+
"name": "labels_collision.npz",
|
| 130 |
+
"description": "Ground-truth physics labels for the 600 collision scenes: per-scene mass scalars, restitution scalars, mass quantile bins (3-class and 5-class), and restitution quantile bins (3-class and 5-class).",
|
| 131 |
+
"encodingFormat": "application/zip",
|
| 132 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/labels/labels_collision.npz"
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"@type": "cr:FileObject",
|
| 136 |
+
"@id": "low-gravity-variant",
|
| 137 |
+
"name": "low_gravity_collision_75scene.tar.gz",
|
| 138 |
+
"description": "75-scene matched-visual low-gravity (g=-3.0 m/s^2) collision variant. Same RNG seed as standard collision so per-scene physics-random variables match scene-by-scene; rendered at 256x256, 48 frames at 24 fps.",
|
| 139 |
+
"encodingFormat": "application/gzip",
|
| 140 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/scenarios/low_gravity_collision_75scene.tar.gz"
|
| 141 |
+
},
|
| 142 |
+
{
|
| 143 |
+
"@type": "cr:FileObject",
|
| 144 |
+
"@id": "phys101-features-vjepa2",
|
| 145 |
+
"name": "phys101_vjepa2_features.tar.gz",
|
| 146 |
+
"description": "V-JEPA 2 frozen features for Phys101 spring (206 clips), ramp (1,801 clips), and fall (666 clips) subsets. Source video remains under its CC-BY license at the Phys101 project URL.",
|
| 147 |
+
"encodingFormat": "application/gzip",
|
| 148 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/phys101/phys101_vjepa2_features.tar.gz"
|
| 149 |
+
},
|
| 150 |
+
{
|
| 151 |
+
"@type": "cr:FileObject",
|
| 152 |
+
"@id": "code-bundle",
|
| 153 |
+
"name": "code.tar.gz",
|
| 154 |
+
"description": "Reproduction scripts for all paper experiments. Mirrors the contents of this supplementary bundle's code/ directory.",
|
| 155 |
+
"encodingFormat": "application/gzip",
|
| 156 |
+
"contentUrl": "https://huggingface.co/datasets/physics-code-transfer-bench/cross-scenario-physics-code-transfer/resolve/main/code/code.tar.gz"
|
| 157 |
+
}
|
| 158 |
+
],
|
| 159 |
+
"recordSet": [
|
| 160 |
+
{
|
| 161 |
+
"@type": "cr:RecordSet",
|
| 162 |
+
"@id": "kubric-collision-scenes",
|
| 163 |
+
"name": "kubric-collision-scenes",
|
| 164 |
+
"description": "One record per collision scene, indexing into the V-JEPA 2 (or alternative-backbone) feature tensor.",
|
| 165 |
+
"field": [
|
| 166 |
+
{
|
| 167 |
+
"@type": "cr:Field",
|
| 168 |
+
"@id": "kubric-collision-scenes/scene-id",
|
| 169 |
+
"name": "scene_id",
|
| 170 |
+
"description": "Integer index into the (N=600) feature and label tensors.",
|
| 171 |
+
"dataType": "sc:Integer"
|
| 172 |
+
},
|
| 173 |
+
{
|
| 174 |
+
"@type": "cr:Field",
|
| 175 |
+
"@id": "kubric-collision-scenes/mass-scalar",
|
| 176 |
+
"name": "mass_scalar",
|
| 177 |
+
"description": "Mass-ratio of the two colliding spheres on a 5-point grid: 1, 2, 3, 4, 5.",
|
| 178 |
+
"dataType": "sc:Integer"
|
| 179 |
+
},
|
| 180 |
+
{
|
| 181 |
+
"@type": "cr:Field",
|
| 182 |
+
"@id": "kubric-collision-scenes/restitution-scalar",
|
| 183 |
+
"name": "restitution_scalar",
|
| 184 |
+
"description": "Coefficient of restitution on a 5-point grid: 0.1, 0.3, 0.5, 0.7, 0.9.",
|
| 185 |
+
"dataType": "sc:Number"
|
| 186 |
+
},
|
| 187 |
+
{
|
| 188 |
+
"@type": "cr:Field",
|
| 189 |
+
"@id": "kubric-collision-scenes/restitution-bin-3class",
|
| 190 |
+
"name": "restitution_bin_3class",
|
| 191 |
+
"description": "Restitution binned to 3 quantile classes by union over all four Kubric scenarios so the same scalar maps to the same class regardless of source/target scenario. Chance accuracy 33.3%.",
|
| 192 |
+
"dataType": "sc:Integer"
|
| 193 |
+
},
|
| 194 |
+
{
|
| 195 |
+
"@type": "cr:Field",
|
| 196 |
+
"@id": "kubric-collision-scenes/feature",
|
| 197 |
+
"name": "feature",
|
| 198 |
+
"description": "(T=4, D=1024) V-JEPA 2 frozen feature tensor for the scene; 4 evenly-spaced frames mean-pooled spatially.",
|
| 199 |
+
"dataType": "cr:Float32"
|
| 200 |
+
}
|
| 201 |
+
]
|
| 202 |
+
}
|
| 203 |
+
]
|
| 204 |
+
}
|
features/vjepa2_collision_pooled.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2bee296d965bf6a7d8e47c196575aff422e78788136957d7bc272ea647fb5fc2
|
| 3 |
+
size 29604716
|
labels/labels_collision.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aaa5f2940f4d7bb822e84873c2674bd2a853feefad7e4988044b2e9294eac1dc
|
| 3 |
+
size 75934
|
labels/labels_elasticity.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3e328c84a275f739ce81cbb35cf3dd4ab0550085a6674e6d65fd0321f8d43086
|
| 3 |
+
size 55650
|
labels/labels_flat_drop.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1e1b6ad78d53cd4a9ecfc648e3f30af61727f2f6669aa429aec9ab2987a230b
|
| 3 |
+
size 23936
|
labels/labels_ramp.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:142cf1154948e3698f954326c0f7ef33a0b71ea997fd46a4181297e53341b4c6
|
| 3 |
+
size 39930
|
labels/labels_ramp_3prop.npz
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:954eea264e382b56dcca1fa2c8c4501fb3fb8c46d485ecf596b985b662ff3a2f
|
| 3 |
+
size 38340
|