python_dispatcher_mode =(enable_python_dispatcher()ifshape_envisnotNoneelsenullcontext())
-
- deftry_record_chromium_data(**kwargs):
- # `backend_compile` only exists as an event if we are compiling with dynamo
- # In some unit tests we don't use dynamo, so we ignore those cases
- chromium_log=get_chromium_event_logger()
- if"backend_compile"inchromium_log.get_stack():
- chromium_log.add_event_data("backend_compile",**kwargs)
-
+ chromium_log=get_chromium_event_logger()# See NOTE: [Deferring tensor pack/unpack hooks until runtime]# If any saved tensor hooks are active, we **don't** want to trace them.# Instead, we'll let them run at runtime, around the custom autograd.Function
@@ -1030,8 +1023,8 @@
]
recomputable_banned_nodes=get_recomputable_banned_nodes(banned_nodes)
+ # sort first by name, to ensure determinism when multiple nodes have same size
+ recomputable_banned_nodes=sorted(recomputable_banned_nodes,key=lambdax:x.name)# default: runtime_optimized_saved_values# more aggressive: more_aggressive_saved_values
diff --git a/nightly/aot_autograd.html b/nightly/aot_autograd.html
index 863949b0b..ede1525d8 100644
--- a/nightly/aot_autograd.html
+++ b/nightly/aot_autograd.html
@@ -267,7 +267,7 @@