{ "name": "root", "gauges": { "Pyramids.Policy.Entropy.mean": { "value": 0.1385202556848526, "min": 0.13323648273944855, "max": 1.421204686164856, "count": 100 }, "Pyramids.Policy.Entropy.sum": { "value": 4135.66064453125, "min": 3990.698974609375, "max": 43113.6640625, "count": 100 }, "Pyramids.Step.mean": { "value": 2999941.0, "min": 29955.0, "max": 2999941.0, "count": 100 }, "Pyramids.Step.sum": { "value": 2999941.0, "min": 29955.0, "max": 2999941.0, "count": 100 }, "Pyramids.Policy.ExtrinsicValueEstimate.mean": { "value": 0.794061541557312, "min": -0.10698705911636353, "max": 0.8489585518836975, "count": 100 }, "Pyramids.Policy.ExtrinsicValueEstimate.sum": { "value": 236.63034057617188, "min": -25.78388214111328, "max": 252.9895782470703, "count": 100 }, "Pyramids.Policy.RndValueEstimate.mean": { "value": 0.0014153291704133153, "min": -0.04450678825378418, "max": 0.47370487451553345, "count": 100 }, "Pyramids.Policy.RndValueEstimate.sum": { "value": 0.42176809906959534, "min": -12.99598217010498, "max": 113.21546173095703, "count": 100 }, "Pyramids.Losses.PolicyLoss.mean": { "value": 0.06642072090263744, "min": 0.06206131367466455, "max": 0.07442717246536715, "count": 100 }, "Pyramids.Losses.PolicyLoss.sum": { "value": 0.9298900926369242, "min": 0.5954173797229372, "max": 1.1024439942314832, "count": 100 }, "Pyramids.Losses.ValueLoss.mean": { "value": 0.018062023364195955, "min": 0.00023003910917380216, "max": 0.018062023364195955, "count": 100 }, "Pyramids.Losses.ValueLoss.sum": { "value": 0.25286832709874335, "min": 0.002990508419259428, "max": 0.25286832709874335, "count": 100 }, "Pyramids.Policy.LearningRate.mean": { "value": 1.4489066599214254e-06, "min": 1.4489066599214254e-06, "max": 0.0002983308380563875, "count": 100 }, "Pyramids.Policy.LearningRate.sum": { "value": 2.0284693238899957e-05, "min": 2.0284693238899957e-05, "max": 0.003927850090716666, "count": 100 }, "Pyramids.Policy.Epsilon.mean": { "value": 0.10048293571428572, "min": 0.10048293571428572, "max": 0.19944361250000003, "count": 100 }, "Pyramids.Policy.Epsilon.sum": { "value": 1.4067611000000002, "min": 1.4067611000000002, "max": 2.7975554999999996, "count": 100 }, "Pyramids.Policy.Beta.mean": { "value": 5.824527785714276e-05, "min": 5.824527785714276e-05, "max": 0.009944416888749998, "count": 100 }, "Pyramids.Policy.Beta.sum": { "value": 0.0008154338899999987, "min": 0.0008154338899999987, "max": 0.130937405, "count": 100 }, "Pyramids.Losses.RNDLoss.mean": { "value": 0.005995760206133127, "min": 0.00539572536945343, "max": 0.5341712236404419, "count": 100 }, "Pyramids.Losses.RNDLoss.sum": { "value": 0.08394064009189606, "min": 0.07554015517234802, "max": 4.273369789123535, "count": 100 }, "Pyramids.Environment.EpisodeLength.mean": { "value": 222.8560606060606, "min": 220.37956204379563, "max": 999.0, "count": 100 }, "Pyramids.Environment.EpisodeLength.sum": { "value": 29417.0, "min": 17520.0, "max": 33061.0, "count": 100 }, "Pyramids.Environment.CumulativeReward.mean": { "value": 1.746828774667599, "min": -0.9997161795054713, "max": 1.746828774667599, "count": 100 }, "Pyramids.Environment.CumulativeReward.sum": { "value": 230.58139825612307, "min": -30.99120156466961, "max": 240.47379745543003, "count": 100 }, "Pyramids.Policy.ExtrinsicReward.mean": { "value": 1.746828774667599, "min": -0.9997161795054713, "max": 1.746828774667599, "count": 100 }, "Pyramids.Policy.ExtrinsicReward.sum": { "value": 230.58139825612307, "min": -30.99120156466961, "max": 240.47379745543003, "count": 100 }, "Pyramids.Policy.RndReward.mean": { "value": 0.013966089139406298, "min": 0.01261693318397823, "max": 10.272860352538133, "count": 100 }, "Pyramids.Policy.RndReward.sum": { "value": 1.8435237664016313, "min": 1.6173795241866173, "max": 195.18434669822454, "count": 100 }, "Pyramids.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 100 }, "Pyramids.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 100 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1676321851", "python_version": "3.10.8 (main, Feb 13 2023, 21:19:31) [GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./trained-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1676327542" }, "total": 5691.576500532916, "count": 1, "self": 0.42790146288461983, "children": { "run_training.setup": { "total": 0.008906654082238674, "count": 1, "self": 0.008906654082238674 }, "TrainerController.start_learning": { "total": 5691.139692415949, "count": 1, "self": 3.6156710099894553, "children": { "TrainerController._reset_env": { "total": 3.4092272468842566, "count": 1, "self": 3.4092272468842566 }, "TrainerController.advance": { "total": 5684.031574906083, "count": 194859, "self": 3.457825487712398, "children": { "env_step": { "total": 3736.9711535903625, "count": 194859, "self": 3463.093571543228, "children": { "SubprocessEnvManager._take_step": { "total": 271.57532442128286, "count": 194859, "self": 13.570822451496497, "children": { "TorchPolicy.evaluate": { "total": 258.00450196978636, "count": 187555, "self": 258.00450196978636 } } }, "workers": { "total": 2.30225762585178, "count": 194859, "self": 0.0, "children": { "worker_root": { "total": 5683.934014911763, "count": 194859, "is_parallel": true, "self": 2540.3886847316753, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0043431001249700785, "count": 1, "is_parallel": true, "self": 0.0013831316027790308, "children": { "_process_rank_one_or_two_observation": { "total": 0.0029599685221910477, "count": 8, "is_parallel": true, "self": 0.0029599685221910477 } } }, "UnityEnvironment.step": { "total": 0.03385546593926847, "count": 1, "is_parallel": true, "self": 0.0005877050571143627, "children": { "UnityEnvironment._generate_step_input": { "total": 0.0007269070483744144, "count": 1, "is_parallel": true, "self": 0.0007269070483744144 }, "communicator.exchange": { "total": 0.030462343944236636, "count": 1, "is_parallel": true, "self": 0.030462343944236636 }, "steps_from_proto": { "total": 0.0020785098895430565, "count": 1, "is_parallel": true, "self": 0.0004306999035179615, "children": { "_process_rank_one_or_two_observation": { "total": 0.001647809986025095, "count": 8, "is_parallel": true, "self": 0.001647809986025095 } } } } } } }, "UnityEnvironment.step": { "total": 3143.545330180088, "count": 194858, "is_parallel": true, "self": 95.26298781647347, "children": { "UnityEnvironment._generate_step_input": { "total": 77.98528197151609, "count": 194858, "is_parallel": true, "self": 77.98528197151609 }, "communicator.exchange": { "total": 2671.6618833495304, "count": 194858, "is_parallel": true, "self": 2671.6618833495304 }, "steps_from_proto": { "total": 298.6351770425681, "count": 194858, "is_parallel": true, "self": 60.68810750730336, "children": { "_process_rank_one_or_two_observation": { "total": 237.94706953526475, "count": 1558864, "is_parallel": true, "self": 237.94706953526475 } } } } } } } } } } }, "trainer_advance": { "total": 1943.602595828008, "count": 194859, "self": 7.752323070308194, "children": { "process_trajectory": { "total": 282.4697275806684, "count": 194859, "self": 281.7832751579117, "children": { "RLTrainer._checkpoint": { "total": 0.6864524227567017, "count": 6, "self": 0.6864524227567017 } } }, "_update_policy": { "total": 1653.3805451770313, "count": 1406, "self": 1031.2318110633641, "children": { "TorchPPOOptimizer.update": { "total": 622.1487341136672, "count": 68382, "self": 622.1487341136672 } } } } } } }, "trainer_threads": { "total": 6.910413503646851e-07, "count": 1, "self": 6.910413503646851e-07 }, "TrainerController._save_models": { "total": 0.08321856195107102, "count": 1, "self": 0.001867838902398944, "children": { "RLTrainer._checkpoint": { "total": 0.08135072304867208, "count": 1, "self": 0.08135072304867208 } } } } } } }