| { | |
| "best_metric": 0.5980878186968839, | |
| "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine/checkpoint-95000", | |
| "epoch": 9.863878477017163, | |
| "eval_steps": 2500, | |
| "global_step": 100000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.049319392385085814, | |
| "grad_norm": 0.9255660772323608, | |
| "learning_rate": 3.57653791130186e-05, | |
| "loss": 2.3662, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.09863878477017163, | |
| "grad_norm": 0.6119515895843506, | |
| "learning_rate": 4.987090502485352e-05, | |
| "loss": 0.3224, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.14795817715525744, | |
| "grad_norm": 0.4047534763813019, | |
| "learning_rate": 4.9090579537757233e-05, | |
| "loss": 0.1093, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.19727756954034326, | |
| "grad_norm": 0.3516473174095154, | |
| "learning_rate": 4.762477336752997e-05, | |
| "loss": 0.0539, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.24659696192542907, | |
| "grad_norm": 0.2518511414527893, | |
| "learning_rate": 4.5515195009118114e-05, | |
| "loss": 0.0324, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.2959163543105149, | |
| "grad_norm": 0.18100735545158386, | |
| "learning_rate": 4.282187104726926e-05, | |
| "loss": 0.0224, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.3452357466956007, | |
| "grad_norm": 0.17784442007541656, | |
| "learning_rate": 3.962143814181834e-05, | |
| "loss": 0.0167, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3945551390806865, | |
| "grad_norm": 0.15665775537490845, | |
| "learning_rate": 3.600496238484107e-05, | |
| "loss": 0.0127, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.4438745314657723, | |
| "grad_norm": 0.190831258893013, | |
| "learning_rate": 3.207534807835037e-05, | |
| "loss": 0.0101, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "grad_norm": 0.17124462127685547, | |
| "learning_rate": 2.7944409663963666e-05, | |
| "loss": 0.0086, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.49319392385085814, | |
| "eval_all_ligands_equal": 0.4899079320113314, | |
| "eval_e3_equal": 0.7831975920679887, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9948654390934845, | |
| "eval_e3_heavy_atoms_difference": 0.35915368271954673, | |
| "eval_e3_heavy_atoms_difference_norm": 0.004063330129381431, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9948654390934845, | |
| "eval_has_all_attachment_points": 0.9820290368271954, | |
| "eval_has_three_substructures": 0.9988491501416431, | |
| "eval_heavy_atoms_difference": 7.435021246458923, | |
| "eval_heavy_atoms_difference_norm": 0.09742815646256632, | |
| "eval_linker_equal": 0.7091890934844193, | |
| "eval_linker_graph_edit_distance": 4.603399433427762e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.054428620142835865, | |
| "eval_linker_has_attachment_point(s)": 0.9953966005665722, | |
| "eval_linker_heavy_atoms_difference": 0.6527089235127479, | |
| "eval_linker_heavy_atoms_difference_norm": 0.02000991525067224, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9953966005665722, | |
| "eval_loss": 0.293075829744339, | |
| "eval_num_fragments": 3.0004426345609065, | |
| "eval_poi_equal": 0.7415899433427762, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9247521246458924, | |
| "eval_poi_heavy_atoms_difference": 2.4198831444759206, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07853231753644796, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9247521246458924, | |
| "eval_reassembly": 0.4971671388101983, | |
| "eval_reassembly_nostereo": 0.5238137393767706, | |
| "eval_runtime": 2417.0464, | |
| "eval_samples_per_second": 4.673, | |
| "eval_steps_per_second": 0.073, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9182011331444759, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.542513316235944, | |
| "grad_norm": 0.12277983129024506, | |
| "learning_rate": 2.372969012074575e-05, | |
| "loss": 0.007, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5918327086210298, | |
| "grad_norm": 0.11204977333545685, | |
| "learning_rate": 1.9551116361503457e-05, | |
| "loss": 0.006, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.6411521010061156, | |
| "grad_norm": 0.1196436733007431, | |
| "learning_rate": 1.55275867959039e-05, | |
| "loss": 0.0051, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.6904714933912014, | |
| "grad_norm": 0.12161414325237274, | |
| "learning_rate": 1.1773588158934163e-05, | |
| "loss": 0.0045, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "grad_norm": 0.14829416573047638, | |
| "learning_rate": 8.395937870495241e-06, | |
| "loss": 0.004, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7397908857762873, | |
| "eval_all_ligands_equal": 0.5380665722379604, | |
| "eval_e3_equal": 0.8043555240793201, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9923866855524079, | |
| "eval_e3_heavy_atoms_difference": 0.4004957507082153, | |
| "eval_e3_heavy_atoms_difference_norm": 0.00806031722408992, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9923866855524079, | |
| "eval_has_all_attachment_points": 0.9817634560906515, | |
| "eval_has_three_substructures": 0.9977868271954674, | |
| "eval_heavy_atoms_difference": 6.986809490084986, | |
| "eval_heavy_atoms_difference_norm": 0.09237955806043094, | |
| "eval_linker_equal": 0.7604461756373938, | |
| "eval_linker_graph_edit_distance": 5.93130311614731e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9940686968838527, | |
| "eval_linker_heavy_atoms_difference": 0.24982294617563738, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0010343910857730165, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9940686968838527, | |
| "eval_loss": 0.30975914001464844, | |
| "eval_num_fragments": 3.000619688385269, | |
| "eval_poi_equal": 0.7602691218130312, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9225389518413598, | |
| "eval_poi_heavy_atoms_difference": 2.202372521246459, | |
| "eval_poi_heavy_atoms_difference_norm": 0.07091974582407527, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9225389518413598, | |
| "eval_reassembly": 0.5450601983002833, | |
| "eval_reassembly_nostereo": 0.5729461756373938, | |
| "eval_runtime": 2364.3909, | |
| "eval_samples_per_second": 4.778, | |
| "eval_steps_per_second": 0.075, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9147485835694051, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.789110278161373, | |
| "grad_norm": 0.11315891146659851, | |
| "learning_rate": 5.490744620021837e-06, | |
| "loss": 0.0036, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.8384296705464589, | |
| "grad_norm": 0.10538846999406815, | |
| "learning_rate": 3.1406736605785858e-06, | |
| "loss": 0.0034, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.8877490629315447, | |
| "grad_norm": 0.09490078687667847, | |
| "learning_rate": 1.412594626586694e-06, | |
| "loss": 0.0032, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.9370684553166305, | |
| "grad_norm": 0.08204534649848938, | |
| "learning_rate": 3.5567880489165705e-07, | |
| "loss": 0.0031, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "grad_norm": 0.10336631536483765, | |
| "learning_rate": 0.0, | |
| "loss": 0.003, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9863878477017163, | |
| "eval_all_ligands_equal": 0.5477160056657224, | |
| "eval_e3_equal": 0.8034702549575071, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9942457507082153, | |
| "eval_e3_heavy_atoms_difference": 0.3627832861189802, | |
| "eval_e3_heavy_atoms_difference_norm": 0.004441809193160064, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9942457507082153, | |
| "eval_has_all_attachment_points": 0.9856586402266289, | |
| "eval_has_three_substructures": 0.9983179886685553, | |
| "eval_heavy_atoms_difference": 6.492917847025495, | |
| "eval_heavy_atoms_difference_norm": 0.08538313201004577, | |
| "eval_linker_equal": 0.7725743626062322, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9951310198300283, | |
| "eval_linker_heavy_atoms_difference": 0.32524787535410765, | |
| "eval_linker_heavy_atoms_difference_norm": 0.005000241007082242, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9951310198300283, | |
| "eval_loss": 0.312394380569458, | |
| "eval_num_fragments": 3.0007967422096318, | |
| "eval_poi_equal": 0.7672627478753541, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9293555240793201, | |
| "eval_poi_heavy_atoms_difference": 2.0848973087818696, | |
| "eval_poi_heavy_atoms_difference_norm": 0.06681110633224353, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9293555240793201, | |
| "eval_reassembly": 0.5548866855524079, | |
| "eval_reassembly_nostereo": 0.5845432011331445, | |
| "eval_runtime": 2338.9582, | |
| "eval_samples_per_second": 4.83, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9231586402266289, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.0357072400868022, | |
| "grad_norm": 0.10573304444551468, | |
| "learning_rate": 4.867483222184158e-05, | |
| "loss": 0.0134, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.085026632471888, | |
| "grad_norm": 0.07383856922388077, | |
| "learning_rate": 4.8545628694730624e-05, | |
| "loss": 0.0058, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.1343460248569737, | |
| "grad_norm": 0.08897178620100021, | |
| "learning_rate": 4.841060399978481e-05, | |
| "loss": 0.0045, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.1836654172420595, | |
| "grad_norm": 0.08299541473388672, | |
| "learning_rate": 4.826979151905655e-05, | |
| "loss": 0.0042, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.2329848096271454, | |
| "grad_norm": 0.08171670883893967, | |
| "learning_rate": 4.812322606550813e-05, | |
| "loss": 0.004, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.2823042020122313, | |
| "grad_norm": 0.08833315223455429, | |
| "learning_rate": 4.797094387440491e-05, | |
| "loss": 0.0037, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.3316235943973171, | |
| "grad_norm": 0.04986245930194855, | |
| "learning_rate": 4.78129825943569e-05, | |
| "loss": 0.0034, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.3809429867824028, | |
| "grad_norm": 0.07238644361495972, | |
| "learning_rate": 4.7649381278011e-05, | |
| "loss": 0.0038, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.4302623791674887, | |
| "grad_norm": 0.05512389540672302, | |
| "learning_rate": 4.748018037239592e-05, | |
| "loss": 0.003, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.4795817715525745, | |
| "grad_norm": 0.058981988579034805, | |
| "learning_rate": 4.73054217089226e-05, | |
| "loss": 0.0027, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.5289011639376602, | |
| "grad_norm": 0.07365080714225769, | |
| "learning_rate": 4.712514849304219e-05, | |
| "loss": 0.003, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.578220556322746, | |
| "grad_norm": 0.04787248373031616, | |
| "learning_rate": 4.693940529356444e-05, | |
| "loss": 0.0025, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.627539948707832, | |
| "grad_norm": 0.0840035229921341, | |
| "learning_rate": 4.674823803163899e-05, | |
| "loss": 0.0023, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.6768593410929178, | |
| "grad_norm": 0.05743684619665146, | |
| "learning_rate": 4.655169396940229e-05, | |
| "loss": 0.0024, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.7261787334780037, | |
| "grad_norm": 0.048048391938209534, | |
| "learning_rate": 4.6349821698293025e-05, | |
| "loss": 0.0022, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.7754981258630895, | |
| "grad_norm": 0.062112707644701004, | |
| "learning_rate": 4.6142671127038905e-05, | |
| "loss": 0.0026, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.8248175182481752, | |
| "grad_norm": 0.034723177552223206, | |
| "learning_rate": 4.593029346931777e-05, | |
| "loss": 0.0028, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.874136910633261, | |
| "grad_norm": 0.0403965599834919, | |
| "learning_rate": 4.571274123109606e-05, | |
| "loss": 0.0016, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.9234563030183467, | |
| "grad_norm": 0.022397898137569427, | |
| "learning_rate": 4.549006819764779e-05, | |
| "loss": 0.0016, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.9727756954034326, | |
| "grad_norm": 0.04490479454398155, | |
| "learning_rate": 4.52623294202573e-05, | |
| "loss": 0.0017, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.0220950877885184, | |
| "grad_norm": 0.047866348177194595, | |
| "learning_rate": 4.502958120260894e-05, | |
| "loss": 0.0017, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.0714144801736043, | |
| "grad_norm": 0.053165681660175323, | |
| "learning_rate": 4.479188108686714e-05, | |
| "loss": 0.0019, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.12073387255869, | |
| "grad_norm": 0.04063253104686737, | |
| "learning_rate": 4.4549287839450324e-05, | |
| "loss": 0.0017, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.170053264943776, | |
| "grad_norm": 0.059268295764923096, | |
| "learning_rate": 4.4301861436502156e-05, | |
| "loss": 0.0015, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.219372657328862, | |
| "grad_norm": 0.021407226100564003, | |
| "learning_rate": 4.404966304906363e-05, | |
| "loss": 0.0014, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.2686920497139473, | |
| "grad_norm": 0.027945173904299736, | |
| "learning_rate": 4.379275502794983e-05, | |
| "loss": 0.0014, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.318011442099033, | |
| "grad_norm": 0.03261112794280052, | |
| "learning_rate": 4.353120088833501e-05, | |
| "loss": 0.0014, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.367330834484119, | |
| "grad_norm": 0.05259308964014053, | |
| "learning_rate": 4.326506529404972e-05, | |
| "loss": 0.0013, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.416650226869205, | |
| "grad_norm": 0.0584435798227787, | |
| "learning_rate": 4.2994959806435226e-05, | |
| "loss": 0.0015, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.465969619254291, | |
| "grad_norm": 0.022548576816916466, | |
| "learning_rate": 4.2719868638689734e-05, | |
| "loss": 0.0012, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.5152890116393767, | |
| "grad_norm": 0.03915947675704956, | |
| "learning_rate": 4.244039660132742e-05, | |
| "loss": 0.0013, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.5646084040244626, | |
| "grad_norm": 0.02655004896223545, | |
| "learning_rate": 4.2156612788007085e-05, | |
| "loss": 0.0012, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.6139277964095484, | |
| "grad_norm": 0.030596962198615074, | |
| "learning_rate": 4.186858735838457e-05, | |
| "loss": 0.0012, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.6632471887946343, | |
| "grad_norm": 0.03012579120695591, | |
| "learning_rate": 4.15763915207673e-05, | |
| "loss": 0.0012, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.7125665811797197, | |
| "grad_norm": 0.015622408129274845, | |
| "learning_rate": 4.1280097514509486e-05, | |
| "loss": 0.001, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.7618859735648056, | |
| "grad_norm": 0.03695574030280113, | |
| "learning_rate": 4.097977859215244e-05, | |
| "loss": 0.0011, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.8112053659498915, | |
| "grad_norm": 0.03273458406329155, | |
| "learning_rate": 4.0675509001314444e-05, | |
| "loss": 0.0011, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.8605247583349773, | |
| "grad_norm": 0.023473775014281273, | |
| "learning_rate": 4.036798407361931e-05, | |
| "loss": 0.0011, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.909844150720063, | |
| "grad_norm": 0.0245220847427845, | |
| "learning_rate": 4.005604729882533e-05, | |
| "loss": 0.0009, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.959163543105149, | |
| "grad_norm": 0.025120645761489868, | |
| "learning_rate": 3.974038822893393e-05, | |
| "loss": 0.001, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.008482935490235, | |
| "grad_norm": 0.03636594116687775, | |
| "learning_rate": 3.942108490409635e-05, | |
| "loss": 0.0008, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.0578023278753204, | |
| "grad_norm": 0.02662217803299427, | |
| "learning_rate": 3.9098865507980035e-05, | |
| "loss": 0.0011, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.1071217202604062, | |
| "grad_norm": 0.014715551398694515, | |
| "learning_rate": 3.8772518268796315e-05, | |
| "loss": 0.0009, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.156441112645492, | |
| "grad_norm": 0.024017762392759323, | |
| "learning_rate": 3.8442766060418925e-05, | |
| "loss": 0.0008, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.205760505030578, | |
| "grad_norm": 0.042992159724235535, | |
| "learning_rate": 3.8109690407234974e-05, | |
| "loss": 0.0008, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.255079897415664, | |
| "grad_norm": 0.05144192650914192, | |
| "learning_rate": 3.777337365528434e-05, | |
| "loss": 0.0009, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.3043992898007497, | |
| "grad_norm": 0.02588796801865101, | |
| "learning_rate": 3.7433898951901246e-05, | |
| "loss": 0.0008, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 3.3537186821858356, | |
| "grad_norm": 0.03916551172733307, | |
| "learning_rate": 3.709135022515784e-05, | |
| "loss": 0.0008, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 3.4030380745709214, | |
| "grad_norm": 0.013717463240027428, | |
| "learning_rate": 3.674581216311475e-05, | |
| "loss": 0.0008, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 3.4523574669560073, | |
| "grad_norm": 0.07564926147460938, | |
| "learning_rate": 3.639806991778405e-05, | |
| "loss": 0.0009, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 3.501676859341093, | |
| "grad_norm": 0.023237833753228188, | |
| "learning_rate": 3.604752099655735e-05, | |
| "loss": 0.0009, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 3.5509962517261786, | |
| "grad_norm": 0.034029532223939896, | |
| "learning_rate": 3.569354109150707e-05, | |
| "loss": 0.0006, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 3.6003156441112645, | |
| "grad_norm": 0.04032694920897484, | |
| "learning_rate": 3.533691743044375e-05, | |
| "loss": 0.0007, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 3.6496350364963503, | |
| "grad_norm": 0.03434913605451584, | |
| "learning_rate": 3.497773818116328e-05, | |
| "loss": 0.0007, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 3.698954428881436, | |
| "grad_norm": 0.012562757357954979, | |
| "learning_rate": 3.461609214327767e-05, | |
| "loss": 0.0007, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 3.748273821266522, | |
| "grad_norm": 0.03795718029141426, | |
| "learning_rate": 3.4253529436582336e-05, | |
| "loss": 0.0012, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 3.797593213651608, | |
| "grad_norm": 0.036550700664520264, | |
| "learning_rate": 3.3887227607163154e-05, | |
| "loss": 0.0007, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 3.8469126060366934, | |
| "grad_norm": 0.009776749648153782, | |
| "learning_rate": 3.3518728595236816e-05, | |
| "loss": 0.0006, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 3.8962319984217793, | |
| "grad_norm": 0.010442009195685387, | |
| "learning_rate": 3.314812350453317e-05, | |
| "loss": 0.0006, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 3.945551390806865, | |
| "grad_norm": 0.023100633174180984, | |
| "learning_rate": 3.27755039594663e-05, | |
| "loss": 0.0006, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 3.994870783191951, | |
| "grad_norm": 0.028563622385263443, | |
| "learning_rate": 3.2400962082482306e-05, | |
| "loss": 0.0007, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.044190175577037, | |
| "grad_norm": 0.012048379518091679, | |
| "learning_rate": 3.2024590471283946e-05, | |
| "loss": 0.0006, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.093509567962123, | |
| "grad_norm": 0.01336103118956089, | |
| "learning_rate": 3.164648217593777e-05, | |
| "loss": 0.0006, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.142828960347209, | |
| "grad_norm": 0.04073134437203407, | |
| "learning_rate": 3.1266730675869434e-05, | |
| "loss": 0.0006, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 4.1921483527322945, | |
| "grad_norm": 0.03634357079863548, | |
| "learning_rate": 3.088542985675283e-05, | |
| "loss": 0.0006, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 4.24146774511738, | |
| "grad_norm": 0.03682945668697357, | |
| "learning_rate": 3.0502673987298837e-05, | |
| "loss": 0.0007, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 4.290787137502466, | |
| "grad_norm": 0.030003152787685394, | |
| "learning_rate": 3.0119327223166655e-05, | |
| "loss": 0.0007, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 4.340106529887552, | |
| "grad_norm": 0.025118093937635422, | |
| "learning_rate": 2.973471986909034e-05, | |
| "loss": 0.0008, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 4.389425922272638, | |
| "grad_norm": 0.01078966073691845, | |
| "learning_rate": 2.9348172431675086e-05, | |
| "loss": 0.0005, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 4.438745314657724, | |
| "grad_norm": 0.020486511290073395, | |
| "learning_rate": 2.8960549998903897e-05, | |
| "loss": 0.0005, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 4.48806470704281, | |
| "grad_norm": 0.06772090494632721, | |
| "learning_rate": 2.857194840237535e-05, | |
| "loss": 0.0005, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 4.537384099427895, | |
| "grad_norm": 0.05354917794466019, | |
| "learning_rate": 2.818246371576595e-05, | |
| "loss": 0.0005, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 4.5867034918129805, | |
| "grad_norm": 0.034519363194704056, | |
| "learning_rate": 2.7792192231077896e-05, | |
| "loss": 0.0005, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 4.636022884198066, | |
| "grad_norm": 0.02402963489294052, | |
| "learning_rate": 2.740123043483287e-05, | |
| "loss": 0.0005, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 4.685342276583152, | |
| "grad_norm": 0.03265873342752457, | |
| "learning_rate": 2.7009674984217738e-05, | |
| "loss": 0.0005, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 4.734661668968238, | |
| "grad_norm": 0.04922989383339882, | |
| "learning_rate": 2.661762268318811e-05, | |
| "loss": 0.0005, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 4.783981061353324, | |
| "grad_norm": 0.02939719334244728, | |
| "learning_rate": 2.6225170458535525e-05, | |
| "loss": 0.0005, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 4.83330045373841, | |
| "grad_norm": 0.043153587728738785, | |
| "learning_rate": 2.5832415335924354e-05, | |
| "loss": 0.0005, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 4.882619846123496, | |
| "grad_norm": 0.024349920451641083, | |
| "learning_rate": 2.5439454415904175e-05, | |
| "loss": 0.0005, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 4.931939238508582, | |
| "grad_norm": 0.04341769590973854, | |
| "learning_rate": 2.504638484990372e-05, | |
| "loss": 0.0004, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 4.9812586308936675, | |
| "grad_norm": 0.04032417759299278, | |
| "learning_rate": 2.4653303816212188e-05, | |
| "loss": 0.0005, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 5.030578023278753, | |
| "grad_norm": 0.008274455554783344, | |
| "learning_rate": 2.42603084959539e-05, | |
| "loss": 0.0005, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 5.079897415663839, | |
| "grad_norm": 0.028657300397753716, | |
| "learning_rate": 2.386749604906234e-05, | |
| "loss": 0.0004, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 5.129216808048925, | |
| "grad_norm": 0.006416209042072296, | |
| "learning_rate": 2.3474963590259334e-05, | |
| "loss": 0.0004, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 5.178536200434011, | |
| "grad_norm": 0.014230284839868546, | |
| "learning_rate": 2.3082808165045468e-05, | |
| "loss": 0.0004, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 5.227855592819097, | |
| "grad_norm": 0.014515167102217674, | |
| "learning_rate": 2.2691126725707605e-05, | |
| "loss": 0.0004, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 5.277174985204182, | |
| "grad_norm": 0.0125635527074337, | |
| "learning_rate": 2.2300797694611534e-05, | |
| "loss": 0.0004, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 5.326494377589268, | |
| "grad_norm": 0.022368459030985832, | |
| "learning_rate": 2.191035315978832e-05, | |
| "loss": 0.0004, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 5.375813769974354, | |
| "grad_norm": 0.007136535365134478, | |
| "learning_rate": 2.1520672475998373e-05, | |
| "loss": 0.0004, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 5.425133162359439, | |
| "grad_norm": 0.00662201177328825, | |
| "learning_rate": 2.1131851983700107e-05, | |
| "loss": 0.0004, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 5.474452554744525, | |
| "grad_norm": 0.004521891940385103, | |
| "learning_rate": 2.074398781068747e-05, | |
| "loss": 0.0004, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 5.523771947129611, | |
| "grad_norm": 0.010427785106003284, | |
| "learning_rate": 2.0357175848324307e-05, | |
| "loss": 0.0004, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 5.573091339514697, | |
| "grad_norm": 0.01812533475458622, | |
| "learning_rate": 1.997228184710801e-05, | |
| "loss": 0.0004, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 5.622410731899783, | |
| "grad_norm": 0.01055191457271576, | |
| "learning_rate": 1.958785833460854e-05, | |
| "loss": 0.0004, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 5.671730124284869, | |
| "grad_norm": 0.02802406996488571, | |
| "learning_rate": 1.9204772861767664e-05, | |
| "loss": 0.0004, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 5.721049516669955, | |
| "grad_norm": 0.02212759293615818, | |
| "learning_rate": 1.8823120138514845e-05, | |
| "loss": 0.0004, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 5.7703689090550405, | |
| "grad_norm": 0.020376645028591156, | |
| "learning_rate": 1.8442994520561926e-05, | |
| "loss": 0.0004, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 5.819688301440126, | |
| "grad_norm": 0.004514740779995918, | |
| "learning_rate": 1.8064489986075706e-05, | |
| "loss": 0.0004, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 5.869007693825212, | |
| "grad_norm": 0.014004266820847988, | |
| "learning_rate": 1.768845191897044e-05, | |
| "loss": 0.0004, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 5.918327086210298, | |
| "grad_norm": 0.020170176401734352, | |
| "learning_rate": 1.7313466151363385e-05, | |
| "loss": 0.0004, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 5.967646478595384, | |
| "grad_norm": 0.005268739070743322, | |
| "learning_rate": 1.6940380719660764e-05, | |
| "loss": 0.0003, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 6.01696587098047, | |
| "grad_norm": 0.006257560569792986, | |
| "learning_rate": 1.6569287861489484e-05, | |
| "loss": 0.0003, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 6.066285263365556, | |
| "grad_norm": 0.012183616869151592, | |
| "learning_rate": 1.620027932185401e-05, | |
| "loss": 0.0003, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 6.115604655750641, | |
| "grad_norm": 0.005832422524690628, | |
| "learning_rate": 1.5834177764868226e-05, | |
| "loss": 0.0003, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 6.164924048135727, | |
| "grad_norm": 0.004268107004463673, | |
| "learning_rate": 1.5469606390923247e-05, | |
| "loss": 0.0003, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 6.2142434405208125, | |
| "grad_norm": 0.004789270460605621, | |
| "learning_rate": 1.5107391208925827e-05, | |
| "loss": 0.0003, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 6.263562832905898, | |
| "grad_norm": 0.012517180293798447, | |
| "learning_rate": 1.474762176905931e-05, | |
| "loss": 0.0003, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 6.312882225290984, | |
| "grad_norm": 0.01253846287727356, | |
| "learning_rate": 1.439038701684809e-05, | |
| "loss": 0.0003, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 6.36220161767607, | |
| "grad_norm": 0.014298639260232449, | |
| "learning_rate": 1.4036481818516661e-05, | |
| "loss": 0.0003, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 6.411521010061156, | |
| "grad_norm": 0.0054347398690879345, | |
| "learning_rate": 1.3684575241351546e-05, | |
| "loss": 0.0003, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 6.460840402446242, | |
| "grad_norm": 0.01365108322352171, | |
| "learning_rate": 1.3335466168024479e-05, | |
| "loss": 0.0003, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 6.510159794831328, | |
| "grad_norm": 0.01254518423229456, | |
| "learning_rate": 1.2989240908505564e-05, | |
| "loss": 0.0003, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 6.5594791872164135, | |
| "grad_norm": 0.0054794480092823505, | |
| "learning_rate": 1.2645985059801791e-05, | |
| "loss": 0.0003, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 6.608798579601499, | |
| "grad_norm": 0.007447356358170509, | |
| "learning_rate": 1.2306460783748302e-05, | |
| "loss": 0.0003, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 6.658117971986585, | |
| "grad_norm": 0.0031654227059334517, | |
| "learning_rate": 1.196939123002433e-05, | |
| "loss": 0.0003, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 6.707437364371671, | |
| "grad_norm": 0.009027605876326561, | |
| "learning_rate": 1.1635543223774156e-05, | |
| "loss": 0.0003, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 6.756756756756757, | |
| "grad_norm": 0.008913296274840832, | |
| "learning_rate": 1.130499930198588e-05, | |
| "loss": 0.0003, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 6.806076149141843, | |
| "grad_norm": 0.005660299211740494, | |
| "learning_rate": 1.0977841184781343e-05, | |
| "loss": 0.0003, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 6.855395541526929, | |
| "grad_norm": 0.006413063500076532, | |
| "learning_rate": 1.0654793624981977e-05, | |
| "loss": 0.0003, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "grad_norm": 0.004667436238378286, | |
| "learning_rate": 1.0334641736320572e-05, | |
| "loss": 0.0003, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.904714933912015, | |
| "eval_all_ligands_equal": 0.5963172804532578, | |
| "eval_e3_equal": 0.8271954674220963, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9876062322946175, | |
| "eval_e3_heavy_atoms_difference": 0.32790368271954673, | |
| "eval_e3_heavy_atoms_difference_norm": 0.004086581188234727, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9876062322946175, | |
| "eval_has_all_attachment_points": 0.9836225212464589, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 6.53257790368272, | |
| "eval_heavy_atoms_difference_norm": 0.08691394560206064, | |
| "eval_linker_equal": 0.8436614730878187, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.2073300283286119, | |
| "eval_linker_heavy_atoms_difference_norm": -0.00028415548137912695, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.3754393458366394, | |
| "eval_num_fragments": 3.0, | |
| "eval_poi_equal": 0.7956798866855525, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.93546388101983, | |
| "eval_poi_heavy_atoms_difference": 1.9802584985835694, | |
| "eval_poi_heavy_atoms_difference_norm": 0.06374572149016507, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.93546388101983, | |
| "eval_reassembly": 0.6042847025495751, | |
| "eval_reassembly_nostereo": 0.6316395184135978, | |
| "eval_runtime": 2293.8906, | |
| "eval_samples_per_second": 4.924, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.921742209631728, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 6.9540343262971, | |
| "grad_norm": 0.006098458077758551, | |
| "learning_rate": 1.001811555300422e-05, | |
| "loss": 0.0003, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 7.0033537186821855, | |
| "grad_norm": 0.01404863316565752, | |
| "learning_rate": 9.705293329559965e-06, | |
| "loss": 0.0003, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 7.052673111067271, | |
| "grad_norm": 0.005748235620558262, | |
| "learning_rate": 9.396252404787646e-06, | |
| "loss": 0.0003, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 7.101992503452357, | |
| "grad_norm": 0.022364113479852676, | |
| "learning_rate": 9.091069182639542e-06, | |
| "loss": 0.0003, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "grad_norm": 0.0024993352126330137, | |
| "learning_rate": 8.790417638516349e-06, | |
| "loss": 0.0003, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.151311895837443, | |
| "eval_all_ligands_equal": 0.5958746458923513, | |
| "eval_e3_equal": 0.8270184135977338, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9891111898016998, | |
| "eval_e3_heavy_atoms_difference": 0.3385269121813031, | |
| "eval_e3_heavy_atoms_difference_norm": 0.005225861500269275, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9891111898016998, | |
| "eval_has_all_attachment_points": 0.9861012747875354, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 6.21264164305949, | |
| "eval_heavy_atoms_difference_norm": 0.08315062421329239, | |
| "eval_linker_equal": 0.8451664305949008, | |
| "eval_linker_graph_edit_distance": 2.8328611898017e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9971671388101983, | |
| "eval_linker_heavy_atoms_difference": 0.22910764872521247, | |
| "eval_linker_heavy_atoms_difference_norm": 0.002949945736119303, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9971671388101983, | |
| "eval_loss": 0.37781959772109985, | |
| "eval_num_fragments": 2.9996458923512748, | |
| "eval_poi_equal": 0.7947946175637394, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9374114730878187, | |
| "eval_poi_heavy_atoms_difference": 1.8699539660056657, | |
| "eval_poi_heavy_atoms_difference_norm": 0.06056721262069715, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9374114730878187, | |
| "eval_reassembly": 0.6030453257790368, | |
| "eval_reassembly_nostereo": 0.6300460339943342, | |
| "eval_runtime": 2304.732, | |
| "eval_samples_per_second": 4.901, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9257259206798867, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 7.200631288222529, | |
| "grad_norm": 0.008329563774168491, | |
| "learning_rate": 8.493167110937786e-06, | |
| "loss": 0.0002, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 7.249950680607615, | |
| "grad_norm": 0.00801061000674963, | |
| "learning_rate": 8.199997555072267e-06, | |
| "loss": 0.0002, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 7.299270072992701, | |
| "grad_norm": 0.003677870612591505, | |
| "learning_rate": 7.910981451005626e-06, | |
| "loss": 0.0002, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 7.348589465377787, | |
| "grad_norm": 0.01143123209476471, | |
| "learning_rate": 7.626190251968981e-06, | |
| "loss": 0.0002, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "grad_norm": 0.010139146819710732, | |
| "learning_rate": 7.3456943666733776e-06, | |
| "loss": 0.0002, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.397908857762872, | |
| "eval_all_ligands_equal": 0.5968484419263456, | |
| "eval_e3_equal": 0.827992209631728, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9871635977337111, | |
| "eval_e3_heavy_atoms_difference": 0.40084985835694054, | |
| "eval_e3_heavy_atoms_difference_norm": 0.009092149214275373, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9871635977337111, | |
| "eval_has_all_attachment_points": 0.9868094900849859, | |
| "eval_has_three_substructures": 0.9996458923512748, | |
| "eval_heavy_atoms_difference": 6.3768590651558075, | |
| "eval_heavy_atoms_difference_norm": 0.08539069759580305, | |
| "eval_linker_equal": 0.8485304532577904, | |
| "eval_linker_graph_edit_distance": 2.6558073654390935e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9973441926345609, | |
| "eval_linker_heavy_atoms_difference": 0.2073300283286119, | |
| "eval_linker_heavy_atoms_difference_norm": 0.002126790340110303, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9973441926345609, | |
| "eval_loss": 0.3792908787727356, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.796742209631728, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9372344192634561, | |
| "eval_poi_heavy_atoms_difference": 1.8756196883852692, | |
| "eval_poi_heavy_atoms_difference_norm": 0.060811687345212743, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9372344192634561, | |
| "eval_reassembly": 0.6037535410764873, | |
| "eval_reassembly_nostereo": 0.6298689801699717, | |
| "eval_runtime": 2326.6144, | |
| "eval_samples_per_second": 4.855, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9239553824362606, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 7.447228250147958, | |
| "grad_norm": 0.006876158062368631, | |
| "learning_rate": 7.069563141902743e-06, | |
| "loss": 0.0003, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 7.496547642533044, | |
| "grad_norm": 0.00790834054350853, | |
| "learning_rate": 6.798403773068926e-06, | |
| "loss": 0.0003, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 7.54586703491813, | |
| "grad_norm": 0.003300942713394761, | |
| "learning_rate": 6.531196510034746e-06, | |
| "loss": 0.0003, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 7.595186427303216, | |
| "grad_norm": 0.01328230183571577, | |
| "learning_rate": 6.268555275209789e-06, | |
| "loss": 0.0002, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "grad_norm": 0.005011397879570723, | |
| "learning_rate": 6.010545001186354e-06, | |
| "loss": 0.0002, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.644505819688302, | |
| "eval_all_ligands_equal": 0.59578611898017, | |
| "eval_e3_equal": 0.8279036827195467, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9863668555240793, | |
| "eval_e3_heavy_atoms_difference": 0.45910056657223797, | |
| "eval_e3_heavy_atoms_difference_norm": 0.009712045117076571, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9863668555240793, | |
| "eval_has_all_attachment_points": 0.9822060906515581, | |
| "eval_has_three_substructures": 0.9992917847025495, | |
| "eval_heavy_atoms_difference": 6.255223087818697, | |
| "eval_heavy_atoms_difference_norm": 0.08376456594434907, | |
| "eval_linker_equal": 0.8460516997167139, | |
| "eval_linker_graph_edit_distance": 2.9213881019830026e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.17103399433427763, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0013874128154392413, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.37478843331336975, | |
| "eval_num_fragments": 3.0003541076487252, | |
| "eval_poi_equal": 0.7935552407932012, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9388279036827195, | |
| "eval_poi_heavy_atoms_difference": 1.778594192634561, | |
| "eval_poi_heavy_atoms_difference_norm": 0.05832234698491925, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9388279036827195, | |
| "eval_reassembly": 0.6029567988668555, | |
| "eval_reassembly_nostereo": 0.6311968838526912, | |
| "eval_runtime": 2350.0687, | |
| "eval_samples_per_second": 4.807, | |
| "eval_steps_per_second": 0.075, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9256373937677054, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 7.693825212073387, | |
| "grad_norm": 0.002521380316466093, | |
| "learning_rate": 5.757731379518924e-06, | |
| "loss": 0.0003, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 7.7431446044584735, | |
| "grad_norm": 0.0021294711623340845, | |
| "learning_rate": 5.509163652996047e-06, | |
| "loss": 0.0002, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 7.7924639968435585, | |
| "grad_norm": 0.012009754776954651, | |
| "learning_rate": 5.265414631086604e-06, | |
| "loss": 0.0002, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 7.841783389228644, | |
| "grad_norm": 0.008709125220775604, | |
| "learning_rate": 5.026544575675502e-06, | |
| "loss": 0.0002, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "grad_norm": 0.008389169350266457, | |
| "learning_rate": 4.792612542424446e-06, | |
| "loss": 0.0002, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.89110278161373, | |
| "eval_all_ligands_equal": 0.5972025495750708, | |
| "eval_e3_equal": 0.8281692634560907, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9868094900849859, | |
| "eval_e3_heavy_atoms_difference": 0.3861543909348442, | |
| "eval_e3_heavy_atoms_difference_norm": 0.007623719724388173, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9868094900849859, | |
| "eval_has_all_attachment_points": 0.9819405099150141, | |
| "eval_has_three_substructures": 0.9990262039660056, | |
| "eval_heavy_atoms_difference": 6.23778328611898, | |
| "eval_heavy_atoms_difference_norm": 0.08363425468652734, | |
| "eval_linker_equal": 0.8481763456090652, | |
| "eval_linker_graph_edit_distance": 2.921388101983003e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9970786118980169, | |
| "eval_linker_heavy_atoms_difference": 0.162092776203966, | |
| "eval_linker_heavy_atoms_difference_norm": -0.002065998280888638, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9970786118980169, | |
| "eval_loss": 0.3756082355976105, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7962110481586402, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9375885269121813, | |
| "eval_poi_heavy_atoms_difference": 1.8251593484419264, | |
| "eval_poi_heavy_atoms_difference_norm": 0.05892174445701429, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9375885269121813, | |
| "eval_reassembly": 0.603399433427762, | |
| "eval_reassembly_nostereo": 0.6274787535410765, | |
| "eval_runtime": 2295.8147, | |
| "eval_samples_per_second": 4.92, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9251062322946175, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 7.940422173998816, | |
| "grad_norm": 0.004469925072044134, | |
| "learning_rate": 4.564129214831858e-06, | |
| "loss": 0.0002, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 7.989741566383902, | |
| "grad_norm": 0.013919726945459843, | |
| "learning_rate": 4.340235334721807e-06, | |
| "loss": 0.0002, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 8.039060958768989, | |
| "grad_norm": 0.011285272426903248, | |
| "learning_rate": 4.121449152482312e-06, | |
| "loss": 0.0002, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 8.088380351154074, | |
| "grad_norm": 0.003505601081997156, | |
| "learning_rate": 3.907824758454454e-06, | |
| "loss": 0.0002, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "grad_norm": 0.005240774247795343, | |
| "learning_rate": 3.6994149668343274e-06, | |
| "loss": 0.0002, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.137699743539159, | |
| "eval_all_ligands_equal": 0.5972910764872521, | |
| "eval_e3_equal": 0.828257790368272, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9884029745042493, | |
| "eval_e3_heavy_atoms_difference": 0.315421388101983, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0055262845315828465, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9884029745042493, | |
| "eval_has_all_attachment_points": 0.984507790368272, | |
| "eval_has_three_substructures": 0.9993803116147308, | |
| "eval_heavy_atoms_difference": 6.1099504249291785, | |
| "eval_heavy_atoms_difference_norm": 0.08167488349727248, | |
| "eval_linker_equal": 0.8495042492917847, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9973441926345609, | |
| "eval_linker_heavy_atoms_difference": 0.15492209631728046, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0014936588027761213, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9973441926345609, | |
| "eval_loss": 0.38009196519851685, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7943519830028328, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9383852691218131, | |
| "eval_poi_heavy_atoms_difference": 1.8310906515580736, | |
| "eval_poi_heavy_atoms_difference_norm": 0.06023915382430907, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9383852691218131, | |
| "eval_reassembly": 0.6043732294617564, | |
| "eval_reassembly_nostereo": 0.6323477337110481, | |
| "eval_runtime": 2301.9928, | |
| "eval_samples_per_second": 4.907, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9263456090651558, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 8.187019135924245, | |
| "grad_norm": 0.009014414623379707, | |
| "learning_rate": 3.4962713026158694e-06, | |
| "loss": 0.0002, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 8.23633852830933, | |
| "grad_norm": 0.003539947560057044, | |
| "learning_rate": 3.298443988852326e-06, | |
| "loss": 0.0002, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 8.285657920694417, | |
| "grad_norm": 0.006800816394388676, | |
| "learning_rate": 3.1059819342396962e-06, | |
| "loss": 0.0002, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 8.334977313079502, | |
| "grad_norm": 0.00440778024494648, | |
| "learning_rate": 2.9196700737611133e-06, | |
| "loss": 0.0002, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "grad_norm": 0.01367681659758091, | |
| "learning_rate": 2.7380580193039963e-06, | |
| "loss": 0.0002, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.384296705464589, | |
| "eval_all_ligands_equal": 0.597556657223796, | |
| "eval_e3_equal": 0.8276381019830028, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9881373937677054, | |
| "eval_e3_heavy_atoms_difference": 0.3526026912181303, | |
| "eval_e3_heavy_atoms_difference_norm": 0.007320876979799362, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9881373937677054, | |
| "eval_has_all_attachment_points": 0.984507790368272, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 6.028063031161473, | |
| "eval_heavy_atoms_difference_norm": 0.08074758321724956, | |
| "eval_linker_equal": 0.8491501416430595, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9974327195467422, | |
| "eval_linker_heavy_atoms_difference": 0.17422096317280453, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0005777300511874596, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9974327195467422, | |
| "eval_loss": 0.3847833275794983, | |
| "eval_num_fragments": 2.9998229461756374, | |
| "eval_poi_equal": 0.7947946175637394, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9398016997167139, | |
| "eval_poi_heavy_atoms_difference": 1.7613314447592068, | |
| "eval_poi_heavy_atoms_difference_norm": 0.05652376268722852, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9398016997167139, | |
| "eval_reassembly": 0.6047273371104815, | |
| "eval_reassembly_nostereo": 0.6298689801699717, | |
| "eval_runtime": 2301.847, | |
| "eval_samples_per_second": 4.907, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9274964589235127, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 8.433616097849674, | |
| "grad_norm": 0.01910337246954441, | |
| "learning_rate": 2.5619497677938975e-06, | |
| "loss": 0.0002, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 8.48293549023476, | |
| "grad_norm": 0.00587499001994729, | |
| "learning_rate": 2.391388858339605e-06, | |
| "loss": 0.0002, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 8.532254882619846, | |
| "grad_norm": 0.011989916674792767, | |
| "learning_rate": 2.2264174585847454e-06, | |
| "loss": 0.0002, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 8.581574275004932, | |
| "grad_norm": 0.0059767174534499645, | |
| "learning_rate": 2.0673893910399504e-06, | |
| "loss": 0.0002, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "grad_norm": 0.010840820148587227, | |
| "learning_rate": 1.913706598207185e-06, | |
| "loss": 0.0002, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.630893667390017, | |
| "eval_all_ligands_equal": 0.5987075070821529, | |
| "eval_e3_equal": 0.8281692634560907, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9884029745042493, | |
| "eval_e3_heavy_atoms_difference": 0.3751770538243626, | |
| "eval_e3_heavy_atoms_difference_norm": 0.0073208764593784646, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9884029745042493, | |
| "eval_has_all_attachment_points": 0.9848618980169972, | |
| "eval_has_three_substructures": 0.9991147308781869, | |
| "eval_heavy_atoms_difference": 6.101805949008498, | |
| "eval_heavy_atoms_difference_norm": 0.08142256101607198, | |
| "eval_linker_equal": 0.8494157223796034, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9968130311614731, | |
| "eval_linker_heavy_atoms_difference": 0.1990084985835694, | |
| "eval_linker_heavy_atoms_difference_norm": 0.0010097837219711044, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9968130311614731, | |
| "eval_loss": 0.38322362303733826, | |
| "eval_num_fragments": 3.000531161473088, | |
| "eval_poi_equal": 0.7955028328611898, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9403328611898017, | |
| "eval_poi_heavy_atoms_difference": 1.7714235127478755, | |
| "eval_poi_heavy_atoms_difference_norm": 0.05643854255359148, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9403328611898017, | |
| "eval_reassembly": 0.6055240793201133, | |
| "eval_reassembly_nostereo": 0.6305771954674221, | |
| "eval_runtime": 2315.3072, | |
| "eval_samples_per_second": 4.879, | |
| "eval_steps_per_second": 0.076, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9280276203966006, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 8.680213059775104, | |
| "grad_norm": 0.010447048582136631, | |
| "learning_rate": 1.7657314120949286e-06, | |
| "loss": 0.0002, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 8.72953245216019, | |
| "grad_norm": 0.00754762114956975, | |
| "learning_rate": 1.6235004164956435e-06, | |
| "loss": 0.0002, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 8.778851844545276, | |
| "grad_norm": 0.0072602503933012486, | |
| "learning_rate": 1.4870487750699602e-06, | |
| "loss": 0.0002, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 8.82817123693036, | |
| "grad_norm": 0.015462782233953476, | |
| "learning_rate": 1.3564102226531817e-06, | |
| "loss": 0.0002, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "grad_norm": 0.003628707956522703, | |
| "learning_rate": 1.2316170569150438e-06, | |
| "loss": 0.0002, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.877490629315448, | |
| "eval_all_ligands_equal": 0.59782223796034, | |
| "eval_e3_equal": 0.8274610481586402, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9896423512747875, | |
| "eval_e3_heavy_atoms_difference": 0.303735835694051, | |
| "eval_e3_heavy_atoms_difference_norm": 0.004488732672379424, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9896423512747875, | |
| "eval_has_all_attachment_points": 0.9854815864022662, | |
| "eval_has_three_substructures": 0.9994688385269122, | |
| "eval_heavy_atoms_difference": 5.897751416430595, | |
| "eval_heavy_atoms_difference_norm": 0.07867299424690383, | |
| "eval_linker_equal": 0.8490616147308782, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.99796388101983, | |
| "eval_linker_heavy_atoms_difference": 0.14686614730878186, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0023561778228989327, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.99796388101983, | |
| "eval_loss": 0.3838784992694855, | |
| "eval_num_fragments": 3.0, | |
| "eval_poi_equal": 0.7959454674220963, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9398902266288952, | |
| "eval_poi_heavy_atoms_difference": 1.7939978753541077, | |
| "eval_poi_heavy_atoms_difference_norm": 0.057305995399255676, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9398902266288952, | |
| "eval_reassembly": 0.605435552407932, | |
| "eval_reassembly_nostereo": 0.6316395184135978, | |
| "eval_runtime": 2309.1207, | |
| "eval_samples_per_second": 4.892, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9291784702549575, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 8.926810021700533, | |
| "grad_norm": 0.00641515152528882, | |
| "learning_rate": 1.1127001303747858e-06, | |
| "loss": 0.0002, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 8.97612941408562, | |
| "grad_norm": 0.009313106536865234, | |
| "learning_rate": 9.999089527066409e-07, | |
| "loss": 0.0002, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 9.025448806470704, | |
| "grad_norm": 0.004975590389221907, | |
| "learning_rate": 8.928193496717002e-07, | |
| "loss": 0.0002, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 9.07476819885579, | |
| "grad_norm": 0.0008887302246876061, | |
| "learning_rate": 7.916897465335582e-07, | |
| "loss": 0.0002, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "grad_norm": 0.002454654313623905, | |
| "learning_rate": 6.965451454869154e-07, | |
| "loss": 0.0002, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.124087591240876, | |
| "eval_all_ligands_equal": 0.5979992917847026, | |
| "eval_e3_equal": 0.8277266288951841, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9879603399433428, | |
| "eval_e3_heavy_atoms_difference": 0.3639341359773371, | |
| "eval_e3_heavy_atoms_difference_norm": 0.006711238872370429, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9879603399433428, | |
| "eval_has_all_attachment_points": 0.9851274787535411, | |
| "eval_has_three_substructures": 0.9992032577903682, | |
| "eval_heavy_atoms_difference": 6.029745042492918, | |
| "eval_heavy_atoms_difference_norm": 0.0802538771820065, | |
| "eval_linker_equal": 0.8500354107648725, | |
| "eval_linker_graph_edit_distance": 2.4787535410764876e+61, | |
| "eval_linker_graph_edit_distance_norm": 0.028018674227771625, | |
| "eval_linker_has_attachment_point(s)": 0.9975212464589235, | |
| "eval_linker_heavy_atoms_difference": 0.16306657223796034, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0023061965798732965, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9975212464589235, | |
| "eval_loss": 0.3835347592830658, | |
| "eval_num_fragments": 3.000265580736544, | |
| "eval_poi_equal": 0.7953257790368272, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9397131728045326, | |
| "eval_poi_heavy_atoms_difference": 1.7906338526912182, | |
| "eval_poi_heavy_atoms_difference_norm": 0.057265091162753884, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9397131728045326, | |
| "eval_reassembly": 0.6052584985835694, | |
| "eval_reassembly_nostereo": 0.6319050991501416, | |
| "eval_runtime": 2302.4169, | |
| "eval_samples_per_second": 4.906, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9276735127478754, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 9.173406983625961, | |
| "grad_norm": 0.004000749904662371, | |
| "learning_rate": 6.07581329809051e-07, | |
| "loss": 0.0002, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 9.222726376011048, | |
| "grad_norm": 0.0045508430339396, | |
| "learning_rate": 5.244637329178403e-07, | |
| "loss": 0.0002, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 9.272045768396133, | |
| "grad_norm": 0.006930803880095482, | |
| "learning_rate": 4.4739720420178545e-07, | |
| "loss": 0.0002, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 9.32136516078122, | |
| "grad_norm": 0.003500725608319044, | |
| "learning_rate": 3.7640079676004027e-07, | |
| "loss": 0.0002, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "grad_norm": 0.00568777322769165, | |
| "learning_rate": 3.114920629802981e-07, | |
| "loss": 0.0002, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.370684553166305, | |
| "eval_all_ligands_equal": 0.5980878186968839, | |
| "eval_e3_equal": 0.8281692634560907, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9871635977337111, | |
| "eval_e3_heavy_atoms_difference": 0.3919971671388102, | |
| "eval_e3_heavy_atoms_difference_norm": 0.00803532425791878, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9871635977337111, | |
| "eval_has_all_attachment_points": 0.9850389518413598, | |
| "eval_has_three_substructures": 0.9992917847025495, | |
| "eval_heavy_atoms_difference": 5.996901558073654, | |
| "eval_heavy_atoms_difference_norm": 0.08024716789835033, | |
| "eval_linker_equal": 0.8501239376770539, | |
| "eval_linker_graph_edit_distance": 2.1246458923512745e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9978753541076487, | |
| "eval_linker_heavy_atoms_difference": 0.14660056657223797, | |
| "eval_linker_heavy_atoms_difference_norm": -0.002835503777451444, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9978753541076487, | |
| "eval_loss": 0.3829340636730194, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.7951487252124646, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9405984419263456, | |
| "eval_poi_heavy_atoms_difference": 1.7691218130311614, | |
| "eval_poi_heavy_atoms_difference_norm": 0.056437930515668894, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9405984419263456, | |
| "eval_reassembly": 0.6056126062322946, | |
| "eval_reassembly_nostereo": 0.6318165722379604, | |
| "eval_runtime": 2300.3144, | |
| "eval_samples_per_second": 4.911, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9279390934844193, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 9.420003945551391, | |
| "grad_norm": 0.0018721942324191332, | |
| "learning_rate": 2.526870501993378e-07, | |
| "loss": 0.0002, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 9.469323337936476, | |
| "grad_norm": 0.0017144909361377358, | |
| "learning_rate": 2.0009955530982117e-07, | |
| "loss": 0.0002, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 9.518642730321563, | |
| "grad_norm": 0.0021374619100242853, | |
| "learning_rate": 1.535318123066204e-07, | |
| "loss": 0.0002, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 9.567962122706648, | |
| "grad_norm": 0.005559207871556282, | |
| "learning_rate": 1.1310684269457139e-07, | |
| "loss": 0.0002, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "grad_norm": 0.0035916175693273544, | |
| "learning_rate": 7.883464070810032e-08, | |
| "loss": 0.0002, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.617281515091735, | |
| "eval_all_ligands_equal": 0.5979992917847026, | |
| "eval_e3_equal": 0.8281692634560907, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9868980169971672, | |
| "eval_e3_heavy_atoms_difference": 0.40625, | |
| "eval_e3_heavy_atoms_difference_norm": 0.008543233459256977, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9868980169971672, | |
| "eval_has_all_attachment_points": 0.9848618980169972, | |
| "eval_has_three_substructures": 0.9992917847025495, | |
| "eval_heavy_atoms_difference": 6.117032577903683, | |
| "eval_heavy_atoms_difference_norm": 0.08171100798099956, | |
| "eval_linker_equal": 0.8498583569405099, | |
| "eval_linker_graph_edit_distance": 2.2131728045325776e+61, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9977868271954674, | |
| "eval_linker_heavy_atoms_difference": 0.15430240793201133, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0017717876495697477, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9977868271954674, | |
| "eval_loss": 0.38293081521987915, | |
| "eval_num_fragments": 3.0001770538243626, | |
| "eval_poi_equal": 0.7955028328611898, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9394475920679887, | |
| "eval_poi_heavy_atoms_difference": 1.805771954674221, | |
| "eval_poi_heavy_atoms_difference_norm": 0.057646246942032714, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9394475920679887, | |
| "eval_reassembly": 0.605435552407932, | |
| "eval_reassembly_nostereo": 0.6317280453257791, | |
| "eval_runtime": 2300.5871, | |
| "eval_samples_per_second": 4.91, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9265226628895185, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 9.66660090747682, | |
| "grad_norm": 0.0018907383782789111, | |
| "learning_rate": 5.0723679437578165e-08, | |
| "loss": 0.0002, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 9.715920299861907, | |
| "grad_norm": 0.00224716286174953, | |
| "learning_rate": 2.8818634561947334e-08, | |
| "loss": 0.0002, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 9.765239692246992, | |
| "grad_norm": 0.0026547384914010763, | |
| "learning_rate": 1.3037127690179729e-08, | |
| "loss": 0.0002, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 9.814559084632076, | |
| "grad_norm": 0.005669564940035343, | |
| "learning_rate": 3.4331286033528754e-09, | |
| "loss": 0.0002, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "grad_norm": 0.003081223228946328, | |
| "learning_rate": 9.011690840776687e-12, | |
| "loss": 0.0002, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 9.863878477017163, | |
| "eval_all_ligands_equal": 0.5980878186968839, | |
| "eval_e3_equal": 0.828257790368272, | |
| "eval_e3_graph_edit_distance": Infinity, | |
| "eval_e3_graph_edit_distance_norm": Infinity, | |
| "eval_e3_has_attachment_point(s)": 0.9870750708215298, | |
| "eval_e3_heavy_atoms_difference": 0.40421388101983, | |
| "eval_e3_heavy_atoms_difference_norm": 0.008500938679118707, | |
| "eval_e3_tanimoto_similarity": 0.0, | |
| "eval_e3_valid": 0.9870750708215298, | |
| "eval_has_all_attachment_points": 0.9846848441926346, | |
| "eval_has_three_substructures": 0.9992032577903682, | |
| "eval_heavy_atoms_difference": 6.106409348441926, | |
| "eval_heavy_atoms_difference_norm": 0.0815233025391124, | |
| "eval_linker_equal": 0.849592776203966, | |
| "eval_linker_graph_edit_distance": Infinity, | |
| "eval_linker_graph_edit_distance_norm": Infinity, | |
| "eval_linker_has_attachment_point(s)": 0.9977868271954674, | |
| "eval_linker_heavy_atoms_difference": 0.15465651558073654, | |
| "eval_linker_heavy_atoms_difference_norm": -0.0017519301091180534, | |
| "eval_linker_tanimoto_similarity": 0.0, | |
| "eval_linker_valid": 0.9977868271954674, | |
| "eval_loss": 0.38294461369514465, | |
| "eval_num_fragments": 3.0000885269121813, | |
| "eval_poi_equal": 0.7955028328611898, | |
| "eval_poi_graph_edit_distance": Infinity, | |
| "eval_poi_graph_edit_distance_norm": Infinity, | |
| "eval_poi_has_attachment_point(s)": 0.9393590651558074, | |
| "eval_poi_heavy_atoms_difference": 1.8055949008498584, | |
| "eval_poi_heavy_atoms_difference_norm": 0.05770569057556065, | |
| "eval_poi_tanimoto_similarity": 0.0, | |
| "eval_poi_valid": 0.9393590651558074, | |
| "eval_reassembly": 0.6055240793201133, | |
| "eval_reassembly_nostereo": 0.6318165722379604, | |
| "eval_runtime": 2300.3904, | |
| "eval_samples_per_second": 4.91, | |
| "eval_steps_per_second": 0.077, | |
| "eval_tanimoto_similarity": 0.0, | |
| "eval_valid": 0.9266997167138811, | |
| "step": 100000 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 100000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 5000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.663173632603694e+17, | |
| "train_batch_size": 128, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |