pytorch

Форк
0
/
conf.py 
3870 строк · 100.9 Кб
1
#
2
# PyTorch documentation build configuration file, created by
3
# sphinx-quickstart on Fri Dec 23 13:31:47 2016.
4
#
5
# This file is execfile()d with the current directory set to its
6
# containing dir.
7
#
8
# Note that not all possible configuration values are present in this
9
# autogenerated file.
10
#
11
# All configuration values have a default; values that are commented out
12
# serve to show the default.
13

14
# If extensions (or modules to document with autodoc) are in another directory,
15
# add these directories to sys.path here. If the directory is relative to the
16
# documentation root, use os.path.abspath to make it absolute, like shown here.
17
#
18
import os
19

20
# import sys
21
import pkgutil
22
import re
23
from os import path
24

25
# source code directory, relative to this file, for sphinx-autobuild
26
# sys.path.insert(0, os.path.abspath('../..'))
27

28
import torch
29

30
try:
31
    import torchvision  # noqa: F401
32
except ImportError:
33
    import warnings
34

35
    warnings.warn('unable to load "torchvision" package')
36

37
RELEASE = os.environ.get("RELEASE", False)
38

39
import pytorch_sphinx_theme
40

41
# -- General configuration ------------------------------------------------
42

43
# If your documentation needs a minimal Sphinx version, state it here.
44
#
45
needs_sphinx = "3.1.2"
46

47
# Add any Sphinx extension module names here, as strings. They can be
48
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
49
# ones.
50
extensions = [
51
    "sphinx.ext.autodoc",
52
    "sphinx.ext.autosummary",
53
    "sphinx.ext.doctest",
54
    "sphinx.ext.intersphinx",
55
    "sphinx.ext.todo",
56
    "sphinx.ext.coverage",
57
    "sphinx.ext.napoleon",
58
    "sphinx.ext.viewcode",
59
    "sphinxcontrib.katex",
60
    "sphinx.ext.autosectionlabel",
61
    "sphinx_copybutton",
62
    "sphinx_panels",
63
    "myst_parser",
64
]
65

66
# build the templated autosummary files
67
autosummary_generate = True
68
numpydoc_show_class_members = False
69

70
# Theme has bootstrap already
71
panels_add_bootstrap_css = False
72

73
# autosectionlabel throws warnings if section names are duplicated.
74
# The following tells autosectionlabel to not throw a warning for
75
# duplicated section names that are in different documents.
76
autosectionlabel_prefix_document = True
77

78
# katex options
79
#
80
#
81

82
katex_prerender = True
83

84
napoleon_use_ivar = True
85

86
# Add any paths that contain templates here, relative to this directory.
87
templates_path = ["_templates"]
88

89
# TODO: document these and remove them from here.
90

91
coverage_ignore_functions = [
92
    # torch
93
    "typename",
94
    # torch.cuda
95
    "check_error",
96
    "cudart",
97
    "is_bf16_supported",
98
    # torch.cuda._sanitizer
99
    "zip_arguments",
100
    "zip_by_key",
101
    # torch.distributed.autograd
102
    "is_available",
103
    # torch.distributed.checkpoint.state_dict
104
    "gc_context",
105
    "state_dict",
106
    # torch.distributed.elastic.events
107
    "construct_and_record_rdzv_event",
108
    "record_rdzv_event",
109
    # torch.distributed.elastic.metrics
110
    "initialize_metrics",
111
    # torch.distributed.elastic.rendezvous.registry
112
    "get_rendezvous_handler",
113
    # torch.distributed.launch
114
    "launch",
115
    "main",
116
    "parse_args",
117
    # torch.distributed.rpc
118
    "is_available",
119
    # torch.distributed.run
120
    "config_from_args",
121
    "determine_local_world_size",
122
    "get_args_parser",
123
    "get_rdzv_endpoint",
124
    "get_use_env",
125
    "main",
126
    "parse_args",
127
    "parse_min_max_nnodes",
128
    "run",
129
    "run_script_path",
130
    # torch.distributions.constraints
131
    "is_dependent",
132
    # torch.hub
133
    "import_module",
134
    # torch.jit
135
    "export_opnames",
136
    # torch.jit.unsupported_tensor_ops
137
    "execWrapper",
138
    # torch.onnx
139
    "unregister_custom_op_symbolic",
140
    # torch.ao.quantization
141
    "default_eval_fn",
142
    # torch.backends
143
    "disable_global_flags",
144
    "flags_frozen",
145
    # torch.distributed.algorithms.ddp_comm_hooks
146
    "register_ddp_comm_hook",
147
    # torch.nn
148
    "factory_kwargs",
149
    # torch.nn.parallel
150
    "DistributedDataParallelCPU",
151
    # torch.utils
152
    "set_module",
153
    # torch.utils.model_dump
154
    "burn_in_info",
155
    "get_info_and_burn_skeleton",
156
    "get_inline_skeleton",
157
    "get_model_info",
158
    "get_storage_info",
159
    "hierarchical_pickle",
160
    # torch.amp.autocast_mode
161
    "autocast_decorator",
162
    # torch.ao.nn.quantized.dynamic.modules.rnn
163
    "apply_permutation",
164
    "pack_weight_bias",
165
    # torch.ao.nn.quantized.reference.modules.rnn
166
    "get_quantized_weight",
167
    # torch.ao.ns.fx.graph_matcher
168
    "get_matching_subgraph_pairs",
169
    # torch.ao.ns.fx.graph_passes
170
    "add_loggers_to_model",
171
    "create_a_shadows_b",
172
    # torch.ao.ns.fx.mappings
173
    "add_op_to_sets_of_related_ops",
174
    "get_base_name_for_op",
175
    "get_base_name_to_sets_of_related_ops",
176
    "get_node_type_to_io_type_map",
177
    "get_unmatchable_types_map",
178
    # torch.ao.ns.fx.n_shadows_utils
179
    "create_add_loggers_graph",
180
    "create_n_transformed_and_logged_copies_of_subgraph",
181
    "create_one_transformed_and_logged_copy_of_subgraph",
182
    "create_results_comparison",
183
    "create_submodule_from_subgraph",
184
    "extract_weight_comparison",
185
    "group_results_by_subgraph",
186
    "print_n_shadows_summary",
187
    # torch.ao.ns.fx.pattern_utils
188
    "end_node_matches_reversed_fusion",
189
    "get_reversed_fusions",
190
    "get_type_a_related_to_b",
191
    # torch.ao.ns.fx.utils
192
    "get_arg_indices_of_inputs_to_log",
193
    "get_node_first_input_and_output_type",
194
    "get_node_input_qparams",
195
    "get_normalized_nth_input",
196
    "get_number_of_non_param_args",
197
    "get_target_type_str",
198
    "maybe_add_missing_fqns",
199
    "maybe_dequantize_first_two_tensor_args_and_handle_tuples",
200
    "op_type_supports_shadowing",
201
    "rekey_logger_info_on_node_name_of_model",
202
    "return_first_non_observer_node",
203
    # torch.ao.ns.fx.weight_utils
204
    "extract_weight_from_node",
205
    "get_conv_fun_weight",
206
    "get_conv_mod_weight",
207
    "get_linear_fun_weight",
208
    "get_linear_mod_weight",
209
    "get_lstm_mod_weights",
210
    "get_lstm_weight",
211
    "get_op_to_type_to_weight_extraction_fn",
212
    "get_qconv_fun_weight",
213
    "get_qlinear_fun_weight",
214
    "get_qlstm_weight",
215
    "mod_0_weight_detach",
216
    "mod_weight_bias_0",
217
    "mod_weight_detach",
218
    # torch.ao.pruning.sparsifier.utils
219
    "fqn_to_module",
220
    "get_arg_info_from_tensor_fqn",
221
    "module_contains_param",
222
    "module_to_fqn",
223
    "swap_module",
224
    # torch.ao.quantization.backend_config.executorch
225
    "get_executorch_backend_config",
226
    # torch.ao.quantization.backend_config.fbgemm
227
    "get_fbgemm_backend_config",
228
    # torch.ao.quantization.backend_config.native
229
    "get_native_backend_config",
230
    "get_native_backend_config_dict",
231
    "get_test_only_legacy_native_backend_config",
232
    "get_test_only_legacy_native_backend_config_dict",
233
    # torch.ao.quantization.backend_config.onednn
234
    "get_onednn_backend_config",
235
    # torch.ao.quantization.backend_config.qnnpack
236
    "get_qnnpack_backend_config",
237
    # torch.ao.quantization.backend_config.tensorrt
238
    "get_tensorrt_backend_config",
239
    "get_tensorrt_backend_config_dict",
240
    # torch.ao.quantization.backend_config.utils
241
    "entry_to_pretty_str",
242
    "get_fused_module_classes",
243
    "get_fuser_method_mapping",
244
    "get_fusion_pattern_to_extra_inputs_getter",
245
    "get_fusion_pattern_to_root_node_getter",
246
    "get_module_to_qat_module",
247
    "get_pattern_to_dtype_configs",
248
    "get_pattern_to_input_type_to_index",
249
    "get_qat_module_classes",
250
    "get_root_module_to_quantized_reference_module",
251
    "pattern_to_human_readable",
252
    "remove_boolean_dispatch_from_name",
253
    # torch.ao.quantization.backend_config.x86
254
    "get_x86_backend_config",
255
    # torch.ao.quantization.fuse_modules
256
    "fuse_known_modules",
257
    "fuse_modules_qat",
258
    # torch.ao.quantization.fuser_method_mappings
259
    "fuse_conv_bn",
260
    "fuse_conv_bn_relu",
261
    "fuse_convtranspose_bn",
262
    "fuse_linear_bn",
263
    "get_fuser_method",
264
    "get_fuser_method_new",
265
    # torch.ao.quantization.fx.convert
266
    "convert",
267
    "convert_custom_module",
268
    "convert_standalone_module",
269
    "convert_weighted_module",
270
    # torch.ao.quantization.fx.fuse
271
    "fuse",
272
    # torch.ao.quantization.fx.lower_to_fbgemm
273
    "lower_to_fbgemm",
274
    # torch.ao.quantization.fx.lower_to_qnnpack
275
    "lower_to_qnnpack",
276
    # torch.ao.quantization.fx.pattern_utils
277
    "get_default_fusion_patterns",
278
    "get_default_output_activation_post_process_map",
279
    "get_default_quant_patterns",
280
    # torch.ao.quantization.fx.prepare
281
    "insert_observers_for_model",
282
    "prepare",
283
    "propagate_dtypes_for_known_nodes",
284
    # torch.ao.quantization.fx.utils
285
    "all_node_args_except_first",
286
    "all_node_args_have_no_tensors",
287
    "assert_and_get_unique_device",
288
    "collect_producer_nodes",
289
    "create_getattr_from_value",
290
    "create_node_from_old_node_preserve_meta",
291
    "get_custom_module_class_keys",
292
    "get_linear_prepack_op_for_dtype",
293
    "get_new_attr_name_with_prefix",
294
    "get_non_observable_arg_indexes_and_types",
295
    "get_qconv_prepack_op",
296
    "get_skipped_module_name_and_classes",
297
    "graph_module_from_producer_nodes",
298
    "maybe_get_next_module",
299
    "node_arg_is_bias",
300
    "node_arg_is_weight",
301
    "return_arg_list",
302
    # torch.ao.quantization.pt2e.graph_utils
303
    "find_sequential_partitions",
304
    "get_equivalent_types",
305
    "update_equivalent_types_dict",
306
    # torch.ao.quantization.pt2e.prepare
307
    "prepare",
308
    # torch.ao.quantization.pt2e.representation.rewrite
309
    "reference_representation_rewrite",
310
    # torch.ao.quantization.pt2e.utils
311
    "fold_bn_weights_into_conv_node",
312
    "get_aten_graph_module",
313
    "remove_tensor_overload_for_qdq_ops",
314
    # torch.ao.quantization.qconfig
315
    "get_default_qat_qconfig",
316
    "get_default_qat_qconfig_dict",
317
    "get_default_qconfig",
318
    "get_default_qconfig_dict",
319
    "qconfig_equals",
320
    # torch.ao.quantization.quantization_mappings
321
    "get_default_compare_output_module_list",
322
    "get_default_dynamic_quant_module_mappings",
323
    "get_default_dynamic_sparse_quant_module_mappings",
324
    "get_default_float_to_quantized_operator_mappings",
325
    "get_default_qat_module_mappings",
326
    "get_default_qconfig_propagation_list",
327
    "get_default_static_quant_module_mappings",
328
    "get_default_static_quant_reference_module_mappings",
329
    "get_default_static_sparse_quant_module_mappings",
330
    "get_dynamic_quant_module_class",
331
    "get_embedding_qat_module_mappings",
332
    "get_embedding_static_quant_module_mappings",
333
    "get_quantized_operator",
334
    "get_static_quant_module_class",
335
    "no_observer_set",
336
    # torch.ao.quantization.quantize
337
    "get_default_custom_config_dict",
338
    # torch.ao.quantization.quantize_fx
339
    "attach_preserved_attrs_to_model",
340
    "convert_to_reference_fx",
341
    # torch.ao.quantization.quantize_jit
342
    "convert_dynamic_jit",
343
    "convert_jit",
344
    "fuse_conv_bn_jit",
345
    "prepare_dynamic_jit",
346
    "prepare_jit",
347
    "quantize_dynamic_jit",
348
    "quantize_jit",
349
    "script_qconfig",
350
    "script_qconfig_dict",
351
    # torch.ao.quantization.quantize_pt2e
352
    "convert_pt2e",
353
    "prepare_pt2e",
354
    "prepare_qat_pt2e",
355
    # torch.ao.quantization.quantizer.embedding_quantizer
356
    "get_embedding_operators_config",
357
    # torch.ao.quantization.quantizer.xnnpack_quantizer_utils
358
    "get_bias_qspec",
359
    "get_input_act_qspec",
360
    "get_output_act_qspec",
361
    "get_weight_qspec",
362
    "propagate_annotation",
363
    "register_annotator",
364
    # torch.ao.quantization.utils
365
    "activation_dtype",
366
    "activation_is_dynamically_quantized",
367
    "activation_is_int32_quantized",
368
    "activation_is_int8_quantized",
369
    "activation_is_statically_quantized",
370
    "calculate_qmin_qmax",
371
    "check_min_max_valid",
372
    "check_node",
373
    "determine_qparams",
374
    "get_combined_dict",
375
    "get_fqn_to_example_inputs",
376
    "get_qconfig_dtypes",
377
    "get_qparam_dict",
378
    "get_quant_type",
379
    "get_swapped_custom_module_class",
380
    "getattr_from_fqn",
381
    "has_no_children_ignoring_parametrizations",
382
    "is_per_channel",
383
    "is_per_tensor",
384
    "op_is_int8_dynamically_quantized",
385
    "to_underlying_dtype",
386
    "validate_qmin_qmax",
387
    "weight_dtype",
388
    "weight_is_quantized",
389
    "weight_is_statically_quantized",
390
    # torch.backends.cudnn.rnn
391
    "get_cudnn_mode",
392
    "init_dropout_state",
393
    # torch.backends.xeon.run_cpu
394
    "create_args",
395
    # torch.cuda.amp.autocast_mode
396
    "custom_bwd",
397
    "custom_fwd",
398
    # torch.cuda.amp.common
399
    "amp_definitely_not_available",
400
    # torch.cuda.graphs
401
    "graph_pool_handle",
402
    "is_current_stream_capturing",
403
    "make_graphed_callables",
404
    # torch.cuda.memory
405
    "caching_allocator_alloc",
406
    "caching_allocator_delete",
407
    "change_current_allocator",
408
    "empty_cache",
409
    "get_allocator_backend",
410
    "list_gpu_processes",
411
    "max_memory_allocated",
412
    "max_memory_cached",
413
    "max_memory_reserved",
414
    "mem_get_info",
415
    "memory_allocated",
416
    "memory_cached",
417
    "memory_reserved",
418
    "memory_snapshot",
419
    "memory_stats",
420
    "memory_stats_as_nested_dict",
421
    "memory_summary",
422
    "reset_accumulated_memory_stats",
423
    "reset_max_memory_allocated",
424
    "reset_max_memory_cached",
425
    "reset_peak_memory_stats",
426
    "set_per_process_memory_fraction",
427
    # torch.cuda.nccl
428
    "all_gather",
429
    "all_reduce",
430
    "broadcast",
431
    "init_rank",
432
    "reduce",
433
    "reduce_scatter",
434
    "unique_id",
435
    "version",
436
    # torch.cuda.nvtx
437
    "range",
438
    "range_end",
439
    "range_start",
440
    # torch.cuda.profiler
441
    "init",
442
    "profile",
443
    "start",
444
    "stop",
445
    # torch.cuda.random
446
    "get_rng_state",
447
    "get_rng_state_all",
448
    "initial_seed",
449
    "manual_seed",
450
    "manual_seed_all",
451
    "seed",
452
    "seed_all",
453
    "set_rng_state",
454
    "set_rng_state_all",
455
    # torch.distributed.algorithms.ddp_comm_hooks.ddp_zero_hook
456
    "hook_with_zero_step",
457
    "hook_with_zero_step_interleaved",
458
    # torch.distributed.algorithms.ddp_comm_hooks.post_localSGD_hook
459
    "post_localSGD_hook",
460
    # torch.distributed.algorithms.ddp_comm_hooks.quantization_hooks
461
    "quantization_perchannel_hook",
462
    "quantization_pertensor_hook",
463
    # torch.distributed.algorithms.model_averaging.utils
464
    "average_parameters",
465
    "average_parameters_or_parameter_groups",
466
    "get_params_to_average",
467
    # torch.distributed.checkpoint.default_planner
468
    "create_default_global_load_plan",
469
    "create_default_global_save_plan",
470
    "create_default_local_load_plan",
471
    "create_default_local_save_plan",
472
    # torch.distributed.checkpoint.optimizer
473
    "load_sharded_optimizer_state_dict",
474
    # torch.distributed.checkpoint.planner_helpers
475
    "create_read_items_for_chunk_list",
476
    # torch.distributed.checkpoint.state_dict_loader
477
    "load_state_dict",
478
    # torch.distributed.checkpoint.state_dict_saver
479
    "save_state_dict",
480
    # torch.distributed.checkpoint.utils
481
    "find_state_dict_object",
482
    "find_tensor_shard",
483
    # torch.distributed.collective_utils
484
    "all_gather",
485
    "all_gather_object_enforce_type",
486
    "broadcast",
487
    # torch.distributed.distributed_c10d
488
    "all_gather",
489
    "all_gather_coalesced",
490
    "all_gather_into_tensor",
491
    "all_gather_object",
492
    "all_reduce",
493
    "all_reduce_coalesced",
494
    "all_to_all",
495
    "all_to_all_single",
496
    "barrier",
497
    "batch_isend_irecv",
498
    "broadcast",
499
    "broadcast_object_list",
500
    "destroy_process_group",
501
    "gather",
502
    "gather_object",
503
    "get_backend",
504
    "get_backend_config",
505
    "get_global_rank",
506
    "get_group_rank",
507
    "get_process_group_ranks",
508
    "get_rank",
509
    "get_world_size",
510
    "init_process_group",
511
    "irecv",
512
    "is_backend_available",
513
    "is_gloo_available",
514
    "is_initialized",
515
    "is_mpi_available",
516
    "is_nccl_available",
517
    "is_torchelastic_launched",
518
    "is_ucc_available",
519
    "isend",
520
    "monitored_barrier",
521
    "new_group",
522
    "new_subgroups",
523
    "new_subgroups_by_enumeration",
524
    "recv",
525
    "reduce",
526
    "reduce_scatter",
527
    "reduce_scatter_tensor",
528
    "scatter",
529
    "scatter_object_list",
530
    "send",
531
    "supports_complex",
532
    # torch.distributed.elastic.events.handlers
533
    "get_logging_handler",
534
    # torch.distributed.elastic.metrics.api
535
    "configure",
536
    "getStream",
537
    "get_elapsed_time_ms",
538
    "prof",
539
    "profile",
540
    "publish_metric",
541
    "put_metric",
542
    # torch.distributed.elastic.multiprocessing.api
543
    "get_std_cm",
544
    "to_map",
545
    # torch.distributed.elastic.multiprocessing.errors.handlers
546
    "get_error_handler",
547
    # torch.distributed.elastic.multiprocessing.redirects
548
    "get_libc",
549
    "redirect",
550
    # torch.distributed.elastic.multiprocessing.tail_log
551
    "tail_logfile",
552
    # torch.distributed.elastic.rendezvous.dynamic_rendezvous
553
    "get_method_name",
554
    # torch.distributed.elastic.rendezvous.etcd_rendezvous
555
    "create_rdzv_handler",
556
    # torch.distributed.elastic.rendezvous.etcd_server
557
    "find_free_port",
558
    "stop_etcd",
559
    # torch.distributed.elastic.rendezvous.etcd_store
560
    "cas_delay",
561
    # torch.distributed.elastic.rendezvous.static_tcp_rendezvous
562
    "create_rdzv_handler",
563
    # torch.distributed.elastic.rendezvous.utils
564
    "parse_rendezvous_endpoint",
565
    # torch.distributed.elastic.timer.api
566
    "configure",
567
    "expires",
568
    # torch.distributed.elastic.utils.api
569
    "get_env_variable_or_raise",
570
    "get_socket_with_port",
571
    # torch.distributed.elastic.utils.distributed
572
    "create_c10d_store",
573
    "get_free_port",
574
    "get_socket_with_port",
575
    # torch.distributed.elastic.utils.log_level
576
    "get_log_level",
577
    # torch.distributed.elastic.utils.logging
578
    "get_logger",
579
    # torch.distributed.elastic.utils.store
580
    "barrier",
581
    "get_all",
582
    "synchronize",
583
    # torch.distributed.fsdp.wrap
584
    "always_wrap_policy",
585
    "enable_wrap",
586
    "lambda_auto_wrap_policy",
587
    "size_based_auto_wrap_policy",
588
    "transformer_auto_wrap_policy",
589
    "wrap",
590
    # torch.distributed.nn.functional
591
    "all_gather",
592
    "all_reduce",
593
    "all_to_all",
594
    "all_to_all_single",
595
    "broadcast",
596
    "gather",
597
    "reduce",
598
    "reduce_scatter",
599
    "scatter",
600
    # torch.distributed.nn.jit.instantiator
601
    "get_arg_return_types_from_interface",
602
    "instantiate_non_scriptable_remote_module_template",
603
    "instantiate_scriptable_remote_module_template",
604
    # torch.distributed.nn.jit.templates.remote_module_template
605
    "get_remote_module_template",
606
    # torch.distributed.optim.utils
607
    "as_functional_optim",
608
    "register_functional_optim",
609
    # torch.distributed.pipeline.sync.checkpoint
610
    "checkpoint",
611
    "enable_checkpointing",
612
    "enable_recomputing",
613
    "is_checkpointing",
614
    "is_recomputing",
615
    "restore_rng_states",
616
    "save_rng_states",
617
    # torch.distributed.pipeline.sync.dependency
618
    "fork",
619
    "join",
620
    # torch.distributed.pipeline.sync.microbatch
621
    "check",
622
    "gather",
623
    "scatter",
624
    # torch.distributed.pipeline.sync.phony
625
    "get_phony",
626
    # torch.distributed.pipeline.sync.skip.layout
627
    "inspect_skip_layout",
628
    # torch.distributed.pipeline.sync.skip.tracker
629
    "current_skip_tracker",
630
    "use_skip_tracker",
631
    # torch.distributed.pipeline.sync.stream
632
    "as_cuda",
633
    "current_stream",
634
    "default_stream",
635
    "get_device",
636
    "is_cuda",
637
    "new_stream",
638
    "record_stream",
639
    "use_device",
640
    "use_stream",
641
    "wait_stream",
642
    # torch.distributed.pipeline.sync.utils
643
    "partition_model",
644
    # torch.distributed.pipeline.sync.worker
645
    "create_workers",
646
    "spawn_workers",
647
    "worker",
648
    # torch.distributed.rendezvous
649
    "register_rendezvous_handler",
650
    "rendezvous",
651
    # torch.distributed.rpc.api
652
    "get_worker_info",
653
    "method_factory",
654
    "new_method",
655
    "remote",
656
    "rpc_async",
657
    "rpc_sync",
658
    "shutdown",
659
    # torch.distributed.rpc.backend_registry
660
    "backend_registered",
661
    "construct_rpc_backend_options",
662
    "init_backend",
663
    "register_backend",
664
    # torch.distributed.rpc.internal
665
    "deserialize",
666
    "serialize",
667
    # torch.distributed.tensor.parallel.api
668
    "parallelize_module",
669
    # torch.distributed.tensor.parallel.input_reshard
670
    "input_reshard",
671
    # torch.distributed.tensor.parallel.loss
672
    "loss_parallel",
673
    # torch.distributed.tensor.parallel.style
674
    "make_sharded_output_tensor",
675
    # torch.distributions.utils
676
    "broadcast_all",
677
    "clamp_probs",
678
    "logits_to_probs",
679
    "probs_to_logits",
680
    "tril_matrix_to_vec",
681
    "vec_to_tril_matrix",
682
    # torch.functional
683
    "align_tensors",
684
    "atleast_1d",
685
    "atleast_2d",
686
    "atleast_3d",
687
    "block_diag",
688
    "broadcast_shapes",
689
    "broadcast_tensors",
690
    "cartesian_prod",
691
    "cdist",
692
    "chain_matmul",
693
    "einsum",
694
    "lu",
695
    "meshgrid",
696
    "norm",
697
    "split",
698
    "stft",
699
    "tensordot",
700
    "unique",
701
    "unique_consecutive",
702
    "unravel_index",
703
    # torch.fx.annotate
704
    "annotate",
705
    # torch.fx.experimental.accelerator_partitioner
706
    "check_dependency",
707
    "combine_two_partitions",
708
    "get_bfs_level_partition",
709
    "get_device_partition_stats",
710
    "get_device_to_partitions_mapping",
711
    "get_logical_id_to_device",
712
    "get_node_to_partition_mapping",
713
    "reorganize_partitions",
714
    "reset_partition_device",
715
    "set_parents_and_children",
716
    # torch.fx.experimental.const_fold
717
    "get_unique_attr_name_in_module",
718
    "split_const_subgraphs",
719
    # torch.fx.experimental.debug
720
    "set_trace",
721
    # torch.fx.experimental.graph_gradual_typechecker
722
    "adaptiveavgpool2d_check",
723
    "adaptiveavgpool2d_inference_rule",
724
    "add_inference_rule",
725
    "all_eq",
726
    "bn2d_inference_rule",
727
    "broadcast_types",
728
    "calculate_out_dimension",
729
    "conv2d_inference_rule",
730
    "conv_refinement_rule",
731
    "conv_rule",
732
    "element_wise_eq",
733
    "expand_to_tensor_dim",
734
    "first_two_eq",
735
    "flatten_check",
736
    "flatten_inference_rule",
737
    "flatten_refinement_rule",
738
    "get_attr_inference_rule",
739
    "get_greatest_upper_bound",
740
    "get_parameter",
741
    "linear_check",
742
    "linear_inference_rule",
743
    "linear_refinement_rule",
744
    "maxpool2d_check",
745
    "maxpool2d_inference_rule",
746
    "register_algebraic_expressions_inference_rule",
747
    "register_inference_rule",
748
    "register_refinement_rule",
749
    "relu_inference_rule",
750
    "reshape_inference_rule",
751
    "transpose_inference_rule",
752
    # torch.fx.experimental.merge_matmul
753
    "are_nodes_independent",
754
    "may_depend_on",
755
    "merge_matmul",
756
    "split_result_tensors",
757
    # torch.fx.experimental.meta_tracer
758
    "embedding_override",
759
    "functional_relu_override",
760
    "gen_constructor_wrapper",
761
    "nn_layernorm_override",
762
    "proxys_to_metas",
763
    "symbolic_trace",
764
    "torch_abs_override",
765
    "torch_nn_relu_override",
766
    "torch_relu_override",
767
    "torch_where_override",
768
    # torch.fx.experimental.migrate_gradual_types.constraint
769
    "is_algebraic_expression",
770
    "is_bool_expr",
771
    "is_dim",
772
    # torch.fx.experimental.migrate_gradual_types.constraint_generator
773
    "adaptive_inference_rule",
774
    "add_layer_norm_constraints",
775
    "add_linear_constraints",
776
    "arange_inference_rule",
777
    "assert_inference_rule",
778
    "batchnorm_inference_rule",
779
    "bmm_inference_rule",
780
    "broadcasting_inference_rule",
781
    "conv2d_inference_rule",
782
    "cumsum_inference_rule",
783
    "embedding_inference_rule",
784
    "embedding_inference_rule_functional",
785
    "eq_inference_rule",
786
    "equality_inference_rule",
787
    "expand_inference_rule",
788
    "flatten_inference_rule",
789
    "full_inference_rule",
790
    "gen_broadcasting_constraints",
791
    "gen_embedding_rules",
792
    "gen_layer_norm_constraints",
793
    "generate_flatten_constraints",
794
    "get_attr_inference_rule",
795
    "getitem_inference_rule",
796
    "gt_inference_rule",
797
    "index_select_inference_rule",
798
    "layer_norm_functional",
799
    "layer_norm_inference_rule",
800
    "linear_constraints",
801
    "linear_inference_rule",
802
    "lt_inference_rule",
803
    "masked_fill_inference_rule",
804
    "maxpool_inference_rule",
805
    "neq_inference_rule",
806
    "range_check",
807
    "register_inference_rule",
808
    "relu_inference_rule",
809
    "reshape_inference_rule",
810
    "size_inference_rule",
811
    "tensor_inference_rule",
812
    "torch_dim_inference_rule",
813
    "torch_linear_inference_rule",
814
    "transpose_inference_rule",
815
    "type_inference_rule",
816
    "view_inference_rule",
817
    # torch.fx.experimental.migrate_gradual_types.constraint_transformation
818
    "apply_padding",
819
    "broadcast_dim",
820
    "calc_last_two_dims",
821
    "create_equality_constraints_for_broadcasting",
822
    "gen_all_reshape_possibilities",
823
    "gen_broadcasting_constraints",
824
    "gen_consistency_constraints",
825
    "gen_greatest_upper_bound",
826
    "gen_lists_of_dims",
827
    "generate_all_broadcasting_possibilities_no_padding",
828
    "generate_all_int_dyn_dim_possibilities",
829
    "generate_binconstraint_d",
830
    "generate_binconstraint_t",
831
    "generate_broadcasting",
832
    "generate_calc_conv",
833
    "generate_calc_maxpool",
834
    "generate_calc_product",
835
    "generate_conj",
836
    "generate_d_gub",
837
    "generate_disj",
838
    "generate_gub",
839
    "generate_reshape",
840
    "is_dim_div_by_target",
841
    "is_target_div_by_dim",
842
    "no_broadcast_dim_with_index",
843
    "register_transformation_rule",
844
    "transform_constraint",
845
    "transform_get_item",
846
    "transform_get_item_tensor",
847
    "transform_index_select",
848
    "transform_transpose",
849
    "valid_index",
850
    "valid_index_tensor",
851
    # torch.fx.experimental.migrate_gradual_types.transform_to_z3
852
    "evaluate_conditional_with_constraints",
853
    # torch.fx.experimental.migrate_gradual_types.util
854
    "gen_bvar",
855
    "gen_dvar",
856
    "gen_nat_constraints",
857
    "gen_tensor_dims",
858
    "gen_tvar",
859
    # torch.fx.experimental.optimization
860
    "extract_subgraph",
861
    "fuse",
862
    "gen_mkl_autotuner",
863
    "matches_module_pattern",
864
    "modules_to_mkldnn",
865
    "optimize_for_inference",
866
    "remove_dropout",
867
    "replace_node_module",
868
    "reset_modules",
869
    "use_mkl_length",
870
    # torch.fx.experimental.partitioner_utils
871
    "get_comm_latency_between",
872
    "get_extra_size_of",
873
    "get_latency_of_one_partition",
874
    "get_latency_of_partitioned_graph",
875
    "get_partition_to_latency_mapping",
876
    # torch.fx.experimental.proxy_tensor
877
    "decompose",
878
    "disable_autocast_cache",
879
    "disable_proxy_modes_tracing",
880
    "dispatch_trace",
881
    "extract_val",
882
    "fake_signature",
883
    "fetch_sym_proxy",
884
    "fetch_object_proxy",
885
    "get_innermost_proxy_mode",
886
    "get_isolated_graphmodule",
887
    "get_proxy_slot",
888
    "get_torch_dispatch_modes",
889
    "has_proxy_slot",
890
    "is_sym_node",
891
    "make_fx",
892
    "maybe_disable_fake_tensor_mode",
893
    "maybe_handle_decomp",
894
    "proxy_call",
895
    "set_meta",
896
    "set_original_aten_op",
897
    "set_proxy_slot",
898
    "snapshot_fake",
899
    "thunkify",
900
    "track_tensor",
901
    "track_tensor_tree",
902
    "wrap_key",
903
    "wrapper_and_args_for_make_fx",
904
    # torch.fx.experimental.recording
905
    "record_shapeenv_event",
906
    "replay_shape_env_events",
907
    "shape_env_check_state_equal",
908
    # torch.fx.experimental.sym_node
909
    "ceil_impl",
910
    "floor_ceil_helper",
911
    "floor_impl",
912
    "method_to_operator",
913
    "sympy_is_channels_last_contiguous_2d",
914
    "sympy_is_channels_last_contiguous_3d",
915
    "sympy_is_channels_last_strides_2d",
916
    "sympy_is_channels_last_strides_3d",
917
    "sympy_is_channels_last_strides_generic",
918
    "sympy_is_contiguous",
919
    "sympy_is_contiguous_generic",
920
    "to_node",
921
    "wrap_node",
922
    "sym_sqrt",
923
    "sym_ite",
924
    # torch.fx.experimental.symbolic_shapes
925
    "bind_symbols",
926
    "cast_symbool_to_symint_guardless",
927
    "create_contiguous",
928
    "error",
929
    "eval_guards",
930
    "eval_is_non_overlapping_and_dense",
931
    "expect_true",
932
    "find_symbol_binding_fx_nodes",
933
    "free_symbols",
934
    "free_unbacked_symbols",
935
    "fx_placeholder_targets",
936
    "fx_placeholder_vals",
937
    "guard_bool",
938
    "guard_float",
939
    "guard_int",
940
    "guard_scalar",
941
    "has_hint",
942
    "has_symbolic_sizes_strides",
943
    "is_channels_last_contiguous_2d",
944
    "is_channels_last_contiguous_3d",
945
    "is_channels_last_strides_2d",
946
    "is_channels_last_strides_3d",
947
    "is_contiguous",
948
    "is_non_overlapping_and_dense_indicator",
949
    "is_nested_int",
950
    "is_symbol_binding_fx_node",
951
    "is_symbolic",
952
    # torch.fx.experimental.unification.core
953
    "reify",
954
    # torch.fx.experimental.unification.match
955
    "edge",
956
    "match",
957
    "ordering",
958
    "supercedes",
959
    # torch.fx.experimental.unification.more
960
    "reify_object",
961
    "unifiable",
962
    "unify_object",
963
    # torch.fx.experimental.unification.multipledispatch.conflict
964
    "ambiguities",
965
    "ambiguous",
966
    "consistent",
967
    "edge",
968
    "ordering",
969
    "super_signature",
970
    "supercedes",
971
    # torch.fx.experimental.unification.multipledispatch.core
972
    "dispatch",
973
    "ismethod",
974
    # torch.fx.experimental.unification.multipledispatch.dispatcher
975
    "ambiguity_warn",
976
    "halt_ordering",
977
    "restart_ordering",
978
    "source",
979
    "str_signature",
980
    "variadic_signature_matches",
981
    "variadic_signature_matches_iter",
982
    "warning_text",
983
    # torch.fx.experimental.unification.multipledispatch.utils
984
    "expand_tuples",
985
    "groupby",
986
    "raises",
987
    "reverse_dict",
988
    # torch.fx.experimental.unification.multipledispatch.variadic
989
    "isvariadic",
990
    # torch.fx.experimental.unification.unification_tools
991
    "assoc",
992
    "assoc_in",
993
    "dissoc",
994
    "first",
995
    "get_in",
996
    "getter",
997
    "groupby",
998
    "itemfilter",
999
    "itemmap",
1000
    "keyfilter",
1001
    "keymap",
1002
    "merge",
1003
    "merge_with",
1004
    "update_in",
1005
    "valfilter",
1006
    "valmap",
1007
    # torch.fx.experimental.unification.utils
1008
    "freeze",
1009
    "hashable",
1010
    "raises",
1011
    "reverse_dict",
1012
    "transitive_get",
1013
    "xfail",
1014
    # torch.fx.experimental.unification.variable
1015
    "var",
1016
    "vars",
1017
    # torch.fx.experimental.unify_refinements
1018
    "check_for_type_equality",
1019
    "convert_eq",
1020
    "infer_symbolic_types",
1021
    "infer_symbolic_types_single_pass",
1022
    "substitute_all_types",
1023
    "substitute_solution_one_type",
1024
    "unify_eq",
1025
    # torch.fx.experimental.validator
1026
    "bisect",
1027
    "translation_validation_enabled",
1028
    "translation_validation_timeout",
1029
    "z3op",
1030
    "z3str",
1031
    # torch.fx.graph_module
1032
    "reduce_deploy_graph_module",
1033
    "reduce_graph_module",
1034
    "reduce_package_graph_module",
1035
    # torch.fx.node
1036
    "has_side_effect",
1037
    "map_aggregate",
1038
    "map_arg",
1039
    # torch.fx.operator_schemas
1040
    "check_for_mutable_operation",
1041
    "create_type_hint",
1042
    "get_signature_for_torch_op",
1043
    "normalize_function",
1044
    "normalize_module",
1045
    "type_matches",
1046
    # torch.fx.passes.annotate_getitem_nodes
1047
    "annotate_getitem_nodes",
1048
    # torch.fx.passes.backends.cudagraphs
1049
    "partition_cudagraphs",
1050
    # torch.fx.passes.dialect.common.cse_pass
1051
    "get_CSE_banned_ops",
1052
    # torch.fx.passes.graph_manipulation
1053
    "get_size_of_all_nodes",
1054
    "get_size_of_node",
1055
    "get_tensor_meta",
1056
    "replace_target_nodes_with",
1057
    # torch.fx.passes.infra.pass_manager
1058
    "pass_result_wrapper",
1059
    "this_before_that_pass_constraint",
1060
    # torch.fx.passes.operator_support
1061
    "any_chain",
1062
    "chain",
1063
    "create_op_support",
1064
    # torch.fx.passes.param_fetch
1065
    "default_matching",
1066
    "extract_attrs_for_lowering",
1067
    "lift_lowering_attrs_to_nodes",
1068
    # torch.fx.passes.pass_manager
1069
    "inplace_wrapper",
1070
    "log_hook",
1071
    "loop_pass",
1072
    "these_before_those_pass_constraint",
1073
    "this_before_that_pass_constraint",
1074
    # torch.fx.passes.reinplace
1075
    "reinplace",
1076
    # torch.fx.passes.split_module
1077
    "split_module",
1078
    # torch.fx.passes.split_utils
1079
    "getattr_recursive",
1080
    "setattr_recursive",
1081
    "split_by_tags",
1082
    # torch.fx.passes.splitter_base
1083
    "generate_inputs_for_submodules",
1084
    # torch.fx.passes.tools_common
1085
    "get_acc_ops_name",
1086
    "get_node_target",
1087
    "is_node_output_tensor",
1088
    "legalize_graph",
1089
    # torch.fx.passes.utils.common
1090
    "compare_graphs",
1091
    "lift_subgraph_as_module",
1092
    # torch.fx.passes.utils.fuser_utils
1093
    "erase_nodes",
1094
    "fuse_as_graphmodule",
1095
    "fuse_by_partitions",
1096
    "insert_subgm",
1097
    "topo_sort",
1098
    "validate_partition",
1099
    # torch.fx.passes.utils.source_matcher_utils
1100
    "check_subgraphs_connected",
1101
    "get_source_partitions",
1102
    # torch.fx.proxy
1103
    "assert_fn",
1104
    # torch.fx.subgraph_rewriter
1105
    "replace_pattern",
1106
    "replace_pattern_with_filters",
1107
    # torch.fx.tensor_type
1108
    "is_consistent",
1109
    "is_more_precise",
1110
    # torch.fx.traceback
1111
    "format_stack",
1112
    "get_current_meta",
1113
    "has_preserved_node_meta",
1114
    "preserve_node_meta",
1115
    "reset_grad_fn_seq_nr",
1116
    "set_current_meta",
1117
    "set_grad_fn_seq_nr",
1118
    "set_stack_trace",
1119
    # torch.jit.annotations
1120
    "ann_to_type",
1121
    "check_fn",
1122
    "get_enum_value_type",
1123
    "get_param_names",
1124
    "get_signature",
1125
    "get_type_line",
1126
    "is_function_or_method",
1127
    "is_tensor",
1128
    "is_vararg",
1129
    "parse_type_line",
1130
    "split_type_line",
1131
    "try_ann_to_type",
1132
    "try_real_annotations",
1133
    # torch.jit.frontend
1134
    "build_class_def",
1135
    "build_def",
1136
    "build_ignore_context_manager",
1137
    "build_param",
1138
    "build_param_list",
1139
    "build_stmts",
1140
    "build_withitems",
1141
    "find_before",
1142
    "get_class_assigns",
1143
    "get_class_properties",
1144
    "get_default_args",
1145
    "get_default_args_for_class",
1146
    "get_jit_class_def",
1147
    "get_jit_def",
1148
    "is_reserved_name",
1149
    "is_torch_jit_ignore_context_manager",
1150
    # torch.jit.generate_bytecode
1151
    "format_bytecode",
1152
    "generate_upgraders_bytecode",
1153
    # torch.jit.quantized
1154
    "apply_permutation",
1155
    "quantize_linear_modules",
1156
    "quantize_rnn_cell_modules",
1157
    "quantize_rnn_modules",
1158
    # torch.library
1159
    "define",
1160
    "get_ctx",
1161
    "impl",
1162
    "impl_abstract",
1163
    # torch.masked.maskedtensor.core
1164
    "is_masked_tensor",
1165
    # torch.masked.maskedtensor.creation
1166
    "as_masked_tensor",
1167
    "masked_tensor",
1168
    # torch.multiprocessing.pool
1169
    "clean_worker",
1170
    # torch.multiprocessing.reductions
1171
    "fd_id",
1172
    "init_reductions",
1173
    "rebuild_cuda_tensor",
1174
    "rebuild_event",
1175
    "rebuild_nested_tensor",
1176
    "rebuild_sparse_coo_tensor",
1177
    "rebuild_sparse_compressed_tensor",
1178
    "rebuild_storage_empty",
1179
    "rebuild_storage_fd",
1180
    "rebuild_storage_filename",
1181
    "rebuild_tensor",
1182
    "rebuild_typed_storage",
1183
    "rebuild_typed_storage_child",
1184
    "reduce_event",
1185
    "reduce_storage",
1186
    "reduce_tensor",
1187
    "reduce_typed_storage",
1188
    "reduce_typed_storage_child",
1189
    "storage_from_cache",
1190
    # torch.multiprocessing.spawn
1191
    "start_processes",
1192
    # torch.nn.functional
1193
    "adaptive_max_pool1d_with_indices",
1194
    "adaptive_max_pool2d_with_indices",
1195
    "adaptive_max_pool3d_with_indices",
1196
    "assert_int_or_pair",
1197
    "fractional_max_pool2d_with_indices",
1198
    "fractional_max_pool3d_with_indices",
1199
    "max_pool1d_with_indices",
1200
    "max_pool2d_with_indices",
1201
    "max_pool3d_with_indices",
1202
    "multi_head_attention_forward",
1203
    # torch.nn.grad
1204
    "conv1d_input",
1205
    "conv1d_weight",
1206
    "conv2d_input",
1207
    "conv2d_weight",
1208
    "conv3d_input",
1209
    "conv3d_weight",
1210
    # torch.nn.init
1211
    "constant",
1212
    "dirac",
1213
    "eye",
1214
    "kaiming_normal",
1215
    "kaiming_uniform",
1216
    "normal",
1217
    "orthogonal",
1218
    "sparse",
1219
    "uniform",
1220
    "xavier_normal",
1221
    "xavier_uniform",
1222
    # torch.nn.modules.rnn
1223
    "apply_permutation",
1224
    # torch.nn.modules.utils
1225
    "consume_prefix_in_state_dict_if_present",
1226
    # torch.nn.parallel.comm
1227
    "broadcast",
1228
    "broadcast_coalesced",
1229
    "gather",
1230
    "reduce_add",
1231
    "reduce_add_coalesced",
1232
    "scatter",
1233
    # torch.nn.parallel.data_parallel
1234
    "data_parallel",
1235
    # torch.nn.parallel.parallel_apply
1236
    "get_a_var",
1237
    "parallel_apply",
1238
    # torch.nn.parallel.replicate
1239
    "replicate",
1240
    # torch.nn.parallel.scatter_gather
1241
    "gather",
1242
    "is_namedtuple",
1243
    "scatter",
1244
    "scatter_kwargs",
1245
    # torch.nn.parameter
1246
    "is_lazy",
1247
    # torch.nn.utils.clip_grad
1248
    "clip_grad_norm",
1249
    "clip_grad_norm_",
1250
    "clip_grad_value_",
1251
    # torch.nn.utils.convert_parameters
1252
    "parameters_to_vector",
1253
    "vector_to_parameters",
1254
    # torch.nn.utils.fusion
1255
    "fuse_conv_bn_eval",
1256
    "fuse_conv_bn_weights",
1257
    "fuse_linear_bn_eval",
1258
    "fuse_linear_bn_weights",
1259
    # torch.nn.utils.init
1260
    "skip_init",
1261
    # torch.nn.utils.memory_format
1262
    "convert_conv2d_weight_memory_format",
1263
    # torch.nn.utils.parametrizations
1264
    "weight_norm",
1265
    # torch.nn.utils.parametrize
1266
    "transfer_parametrizations_and_params",
1267
    "type_before_parametrizations",
1268
    # torch.nn.utils.rnn
1269
    "bind",
1270
    "invert_permutation",
1271
    # torch.nn.utils.spectral_norm
1272
    "remove_spectral_norm",
1273
    "spectral_norm",
1274
    # torch.nn.utils.weight_norm
1275
    "remove_weight_norm",
1276
    "weight_norm",
1277
    # torch.onnx.operators
1278
    "reshape_from_tensor_shape",
1279
    "shape_as_tensor",
1280
    # torch.onnx.symbolic_caffe2
1281
    "add",
1282
    "avg_pool2d",
1283
    "cat",
1284
    "conv2d",
1285
    "conv2d_relu",
1286
    "conv_prepack",
1287
    "dequantize",
1288
    "linear",
1289
    "linear_prepack",
1290
    "max_pool2d",
1291
    "nchw2nhwc",
1292
    "nhwc2nchw",
1293
    "quantize_per_tensor",
1294
    "register_quantized_ops",
1295
    "relu",
1296
    "reshape",
1297
    "sigmoid",
1298
    "slice",
1299
    "upsample_nearest2d",
1300
    # torch.onnx.symbolic_helper
1301
    "args_have_same_dtype",
1302
    "check_training_mode",
1303
    "dequantize_helper",
1304
    "is_caffe2_aten_fallback",
1305
    "is_complex_value",
1306
    "quantize_helper",
1307
    "quantized_args",
1308
    "requantize_bias_helper",
1309
    # torch.onnx.symbolic_opset10
1310
    "dequantize",
1311
    "div",
1312
    "embedding_bag",
1313
    "fake_quantize_per_tensor_affine",
1314
    "flip",
1315
    "fmod",
1316
    "isfinite",
1317
    "isinf",
1318
    "nan_to_num",
1319
    "quantize_per_tensor",
1320
    "quantized_add",
1321
    "quantized_add_relu",
1322
    "quantized_cat",
1323
    "quantized_conv1d",
1324
    "quantized_conv1d_relu",
1325
    "quantized_conv2d",
1326
    "quantized_conv2d_relu",
1327
    "quantized_conv3d",
1328
    "quantized_conv3d_relu",
1329
    "quantized_conv_transpose1d",
1330
    "quantized_conv_transpose2d",
1331
    "quantized_conv_transpose3d",
1332
    "quantized_group_norm",
1333
    "quantized_hardswish",
1334
    "quantized_instance_norm",
1335
    "quantized_layer_norm",
1336
    "quantized_leaky_relu",
1337
    "quantized_linear",
1338
    "quantized_linear_relu",
1339
    "quantized_mul",
1340
    "quantized_sigmoid",
1341
    "slice",
1342
    "sort",
1343
    "topk",
1344
    # torch.onnx.symbolic_opset11
1345
    "Delete",
1346
    "add",
1347
    "append",
1348
    "arange",
1349
    "argsort",
1350
    "atleast_1d",
1351
    "atleast_2d",
1352
    "atleast_3d",
1353
    "cat",
1354
    "chunk",
1355
    "clamp",
1356
    "clamp_max",
1357
    "clamp_min",
1358
    "constant_pad_nd",
1359
    "cumsum",
1360
    "embedding_bag",
1361
    "embedding_renorm",
1362
    "flatten",
1363
    "gather",
1364
    "hardtanh",
1365
    "hstack",
1366
    "im2col",
1367
    "index",
1368
    "index_copy",
1369
    "index_fill",
1370
    "index_put",
1371
    "insert",
1372
    "linalg_det",
1373
    "linalg_vector_norm",
1374
    "logdet",
1375
    "masked_scatter",
1376
    "masked_select",
1377
    "mm",
1378
    "narrow",
1379
    "normal",
1380
    "pad",
1381
    "pixel_shuffle",
1382
    "pop",
1383
    "prim_constant_chunk",
1384
    "reflection_pad",
1385
    "relu6",
1386
    "remainder",
1387
    "replication_pad",
1388
    "round",
1389
    "scatter",
1390
    "select",
1391
    "size",
1392
    "sort",
1393
    "split",
1394
    "split_with_sizes",
1395
    "squeeze",
1396
    "stack",
1397
    "topk",
1398
    "unbind",
1399
    "unique_dim",
1400
    "unsqueeze",
1401
    "vstack",
1402
    # torch.onnx.symbolic_opset12
1403
    "argmax",
1404
    "argmin",
1405
    "binary_cross_entropy_with_logits",
1406
    "celu",
1407
    "cross_entropy_loss",
1408
    "dropout",
1409
    "einsum",
1410
    "ge",
1411
    "le",
1412
    "native_dropout",
1413
    "nll_loss",
1414
    "nll_loss2d",
1415
    "nll_loss_nd",
1416
    "outer",
1417
    "pow",
1418
    "tensordot",
1419
    "unfold",
1420
    # torch.onnx.symbolic_opset13
1421
    "diagonal",
1422
    "fake_quantize_per_channel_affine",
1423
    "fake_quantize_per_tensor_affine",
1424
    "frobenius_norm",
1425
    "log_softmax",
1426
    "nonzero_numpy",
1427
    "quantized_conv1d",
1428
    "quantized_conv1d_relu",
1429
    "quantized_conv2d",
1430
    "quantized_conv2d_relu",
1431
    "quantized_conv3d",
1432
    "quantized_conv3d_relu",
1433
    "quantized_conv_transpose1d",
1434
    "quantized_conv_transpose2d",
1435
    "quantized_conv_transpose3d",
1436
    "quantized_linear",
1437
    "quantized_linear_relu",
1438
    "repeat_interleave",
1439
    "softmax",
1440
    "split",
1441
    "split_with_sizes",
1442
    "tensor_split",
1443
    "tile",
1444
    "unbind",
1445
    "unflatten",
1446
    "unsafe_chunk",
1447
    "unsafe_split",
1448
    "unsafe_split_with_sizes",
1449
    "where",
1450
    # torch.onnx.symbolic_opset14
1451
    "batch_norm",
1452
    "hardswish",
1453
    "quantized_hardswish",
1454
    "reshape",
1455
    "scaled_dot_product_attention",
1456
    "tril",
1457
    "triu",
1458
    # torch.onnx.symbolic_opset15
1459
    "aten__is_",
1460
    "aten__isnot_",
1461
    "bernoulli",
1462
    "prim_unchecked_cast",
1463
    # torch.onnx.symbolic_opset16
1464
    "grid_sampler",
1465
    "scatter_add",
1466
    "scatter_reduce",
1467
    # torch.onnx.symbolic_opset17
1468
    "layer_norm",
1469
    "stft",
1470
    # torch.onnx.symbolic_opset18
1471
    "col2im",
1472
    # torch.onnx.symbolic_opset7
1473
    "max",
1474
    "min",
1475
    # torch.onnx.symbolic_opset8
1476
    "addmm",
1477
    "bmm",
1478
    "empty",
1479
    "empty_like",
1480
    "flatten",
1481
    "full",
1482
    "full_like",
1483
    "gt",
1484
    "lt",
1485
    "matmul",
1486
    "mm",
1487
    "ones",
1488
    "ones_like",
1489
    "prelu",
1490
    "repeat",
1491
    "zeros",
1492
    "zeros_like",
1493
    # torch.onnx.symbolic_opset9
1494
    "abs",
1495
    "acos",
1496
    "adaptive_avg_pool1d",
1497
    "adaptive_avg_pool2d",
1498
    "adaptive_avg_pool3d",
1499
    "adaptive_max_pool1d",
1500
    "adaptive_max_pool2d",
1501
    "adaptive_max_pool3d",
1502
    "add",
1503
    "addcmul",
1504
    "addmm",
1505
    "alias",
1506
    "amax",
1507
    "amin",
1508
    "aminmax",
1509
    "arange",
1510
    "argmax",
1511
    "argmin",
1512
    "as_strided",
1513
    "as_tensor",
1514
    "asin",
1515
    "atan",
1516
    "atan2",
1517
    "avg_pool1d",
1518
    "avg_pool2d",
1519
    "avg_pool3d",
1520
    "baddbmm",
1521
    "batch_norm",
1522
    "bernoulli",
1523
    "bitwise_not",
1524
    "bitwise_or",
1525
    "bmm",
1526
    "broadcast_tensors",
1527
    "broadcast_to",
1528
    "bucketize",
1529
    "cat",
1530
    "cdist",
1531
    "ceil",
1532
    "clamp",
1533
    "clamp_max",
1534
    "clamp_min",
1535
    "clone",
1536
    "constant_pad_nd",
1537
    "contiguous",
1538
    "conv1d",
1539
    "conv2d",
1540
    "conv3d",
1541
    "conv_tbc",
1542
    "conv_transpose1d",
1543
    "conv_transpose2d",
1544
    "conv_transpose3d",
1545
    "convert_element_type",
1546
    "convolution",
1547
    "cos",
1548
    "cosine_similarity",
1549
    "cross",
1550
    "cumsum",
1551
    "detach",
1552
    "dim",
1553
    "div",
1554
    "dot",
1555
    "dropout",
1556
    "elu",
1557
    "embedding",
1558
    "embedding_bag",
1559
    "empty",
1560
    "empty_like",
1561
    "eq",
1562
    "erf",
1563
    "exp",
1564
    "expand",
1565
    "expand_as",
1566
    "eye",
1567
    "fill",
1568
    "flatten",
1569
    "floor",
1570
    "floor_divide",
1571
    "floordiv",
1572
    "frobenius_norm",
1573
    "full",
1574
    "full_like",
1575
    "gather",
1576
    "ge",
1577
    "gelu",
1578
    "get_pool_ceil_padding",
1579
    "glu",
1580
    "group_norm",
1581
    "gru",
1582
    "gt",
1583
    "hann_window",
1584
    "hardshrink",
1585
    "hardsigmoid",
1586
    "hardswish",
1587
    "hardtanh",
1588
    "index",
1589
    "index_add",
1590
    "index_copy",
1591
    "index_fill",
1592
    "index_put",
1593
    "index_select",
1594
    "instance_norm",
1595
    "is_floating_point",
1596
    "is_pinned",
1597
    "isnan",
1598
    "item",
1599
    "kl_div",
1600
    "layer_norm",
1601
    "le",
1602
    "leaky_relu",
1603
    "lerp",
1604
    "lift",
1605
    "linalg_cross",
1606
    "linalg_matrix_norm",
1607
    "linalg_norm",
1608
    "linalg_vector_norm",
1609
    "linear",
1610
    "linspace",
1611
    "log",
1612
    "log10",
1613
    "log1p",
1614
    "log2",
1615
    "log_sigmoid",
1616
    "log_softmax",
1617
    "logical_and",
1618
    "logical_not",
1619
    "logical_or",
1620
    "logical_xor",
1621
    "logit",
1622
    "logsumexp",
1623
    "lstm",
1624
    "lstm_cell",
1625
    "lt",
1626
    "masked_fill",
1627
    "masked_fill_",
1628
    "matmul",
1629
    "max",
1630
    "max_pool1d",
1631
    "max_pool1d_with_indices",
1632
    "max_pool2d",
1633
    "max_pool2d_with_indices",
1634
    "max_pool3d",
1635
    "max_pool3d_with_indices",
1636
    "maximum",
1637
    "meshgrid",
1638
    "min",
1639
    "minimum",
1640
    "mish",
1641
    "mm",
1642
    "movedim",
1643
    "mse_loss",
1644
    "mul",
1645
    "multinomial",
1646
    "mv",
1647
    "narrow",
1648
    "native_layer_norm",
1649
    "ne",
1650
    "neg",
1651
    "new_empty",
1652
    "new_full",
1653
    "new_ones",
1654
    "new_zeros",
1655
    "nonzero",
1656
    "nonzero_numpy",
1657
    "noop_complex_operators",
1658
    "norm",
1659
    "numel",
1660
    "numpy_T",
1661
    "one_hot",
1662
    "ones",
1663
    "ones_like",
1664
    "onnx_placeholder",
1665
    "overload_by_arg_count",
1666
    "pad",
1667
    "pairwise_distance",
1668
    "permute",
1669
    "pixel_shuffle",
1670
    "pixel_unshuffle",
1671
    "pow",
1672
    "prelu",
1673
    "prim_constant",
1674
    "prim_constant_chunk",
1675
    "prim_constant_split",
1676
    "prim_data",
1677
    "prim_device",
1678
    "prim_dtype",
1679
    "prim_if",
1680
    "prim_layout",
1681
    "prim_list_construct",
1682
    "prim_list_unpack",
1683
    "prim_loop",
1684
    "prim_max",
1685
    "prim_min",
1686
    "prim_shape",
1687
    "prim_tolist",
1688
    "prim_tuple_construct",
1689
    "prim_type",
1690
    "prim_unchecked_cast",
1691
    "prim_uninitialized",
1692
    "rand",
1693
    "rand_like",
1694
    "randint",
1695
    "randint_like",
1696
    "randn",
1697
    "randn_like",
1698
    "reciprocal",
1699
    "reflection_pad",
1700
    "relu",
1701
    "relu6",
1702
    "remainder",
1703
    "repeat",
1704
    "repeat_interleave",
1705
    "replication_pad",
1706
    "reshape",
1707
    "reshape_as",
1708
    "rnn_relu",
1709
    "rnn_tanh",
1710
    "roll",
1711
    "rrelu",
1712
    "rsqrt",
1713
    "rsub",
1714
    "scalar_tensor",
1715
    "scatter",
1716
    "scatter_add",
1717
    "select",
1718
    "selu",
1719
    "sigmoid",
1720
    "sign",
1721
    "silu",
1722
    "sin",
1723
    "size",
1724
    "slice",
1725
    "softmax",
1726
    "softplus",
1727
    "softshrink",
1728
    "sort",
1729
    "split",
1730
    "split_with_sizes",
1731
    "sqrt",
1732
    "square",
1733
    "squeeze",
1734
    "stack",
1735
    "std",
1736
    "std_mean",
1737
    "sub",
1738
    "t",
1739
    "take",
1740
    "tan",
1741
    "tanh",
1742
    "tanhshrink",
1743
    "tensor",
1744
    "threshold",
1745
    "to",
1746
    "topk",
1747
    "transpose",
1748
    "true_divide",
1749
    "type_as",
1750
    "unbind",
1751
    "unfold",
1752
    "unsafe_chunk",
1753
    "unsafe_split",
1754
    "unsafe_split_with_sizes",
1755
    "unsqueeze",
1756
    "unsupported_complex_operators",
1757
    "unused",
1758
    "upsample_bilinear2d",
1759
    "upsample_linear1d",
1760
    "upsample_nearest1d",
1761
    "upsample_nearest2d",
1762
    "upsample_nearest3d",
1763
    "upsample_trilinear3d",
1764
    "var",
1765
    "var_mean",
1766
    "view",
1767
    "view_as",
1768
    "where",
1769
    "wrap_logical_op_with_cast_to",
1770
    "wrap_logical_op_with_negation",
1771
    "zero",
1772
    "zeros",
1773
    "zeros_like",
1774
    # torch.onnx.utils
1775
    "disable_apex_o2_state_dict_hook",
1776
    "export",
1777
    "export_to_pretty_string",
1778
    "exporter_context",
1779
    "is_in_onnx_export",
1780
    "model_signature",
1781
    "register_custom_op_symbolic",
1782
    "select_model_mode_for_export",
1783
    "setup_onnx_logging",
1784
    "unconvertible_ops",
1785
    "unpack_quantized_tensor",
1786
    "warn_on_static_input_change",
1787
    # torch.onnx.verification
1788
    "check_export_model_diff",
1789
    "verify",
1790
    "verify_aten_graph",
1791
    # torch.optim.adadelta
1792
    "adadelta",
1793
    # torch.optim.adagrad
1794
    "adagrad",
1795
    # torch.optim.adam
1796
    "adam",
1797
    # torch.optim.adamax
1798
    "adamax",
1799
    # torch.optim.adamw
1800
    "adamw",
1801
    # torch.optim.asgd
1802
    "asgd",
1803
    # torch.optim.nadam
1804
    "nadam",
1805
    # torch.optim.optimizer
1806
    "register_optimizer_step_post_hook",
1807
    "register_optimizer_step_pre_hook",
1808
    # torch.optim.radam
1809
    "radam",
1810
    # torch.optim.rmsprop
1811
    "rmsprop",
1812
    # torch.optim.rprop
1813
    "rprop",
1814
    # torch.optim.sgd
1815
    "sgd",
1816
    # torch.optim.swa_utils
1817
    "get_ema_avg_fn",
1818
    "get_ema_multi_avg_fn",
1819
    "get_swa_avg_fn",
1820
    "get_swa_multi_avg_fn",
1821
    "update_bn",
1822
    # torch.overrides
1823
    "enable_reentrant_dispatch",
1824
    # torch.package.analyze.find_first_use_of_broken_modules
1825
    "find_first_use_of_broken_modules",
1826
    # torch.package.analyze.is_from_package
1827
    "is_from_package",
1828
    # torch.package.analyze.trace_dependencies
1829
    "trace_dependencies",
1830
    # torch.profiler.itt
1831
    "range",
1832
    # torch.profiler.profiler
1833
    "schedule",
1834
    "supported_activities",
1835
    "tensorboard_trace_handler",
1836
    # torch.return_types
1837
    "pytree_register_structseq",
1838
    # torch.serialization
1839
    "check_module_version_greater_or_equal",
1840
    "default_restore_location",
1841
    "load",
1842
    "location_tag",
1843
    "mkdtemp",
1844
    "normalize_storage_type",
1845
    "save",
1846
    "storage_to_tensor_type",
1847
    "validate_cuda_device",
1848
    "validate_hpu_device",
1849
    # torch.signal.windows.windows
1850
    "bartlett",
1851
    "blackman",
1852
    "cosine",
1853
    "exponential",
1854
    "gaussian",
1855
    "general_cosine",
1856
    "general_hamming",
1857
    "hamming",
1858
    "hann",
1859
    "kaiser",
1860
    "nuttall",
1861
    # torch.sparse.semi_structured
1862
    "to_sparse_semi_structured",
1863
    # torch.utils.backend_registration
1864
    "generate_methods_for_privateuse1_backend",
1865
    "rename_privateuse1_backend",
1866
    # torch.utils.benchmark.examples.blas_compare_setup
1867
    "conda_run",
1868
    # torch.utils.benchmark.examples.op_benchmark
1869
    "assert_dicts_equal",
1870
    # torch.utils.benchmark.op_fuzzers.spectral
1871
    "power_range",
1872
    # torch.utils.benchmark.utils.common
1873
    "ordered_unique",
1874
    "select_unit",
1875
    "set_torch_threads",
1876
    "trim_sigfig",
1877
    "unit_to_english",
1878
    # torch.utils.benchmark.utils.compare
1879
    "optional_min",
1880
    # torch.utils.benchmark.utils.compile
1881
    "bench_all",
1882
    "bench_loop",
1883
    "benchmark_compile",
1884
    # torch.utils.benchmark.utils.cpp_jit
1885
    "compile_callgrind_template",
1886
    "compile_timeit_template",
1887
    "get_compat_bindings",
1888
    # torch.utils.benchmark.utils.fuzzer
1889
    "dtype_size",
1890
    "prod",
1891
    # torch.utils.benchmark.utils.timer
1892
    "timer",
1893
    # torch.utils.benchmark.utils.valgrind_wrapper.timer_interface
1894
    "wrapper_singleton",
1895
    # torch.utils.bundled_inputs
1896
    "augment_many_model_functions_with_bundled_inputs",
1897
    "augment_model_with_bundled_inputs",
1898
    "bundle_inputs",
1899
    "bundle_large_tensor",
1900
    "bundle_randn",
1901
    # torch.utils.checkpoint
1902
    "check_backward_validity",
1903
    "detach_variable",
1904
    "get_device_states",
1905
    "noop_context_fn",
1906
    "set_checkpoint_early_stop",
1907
    "set_device_states",
1908
    # torch.utils.collect_env
1909
    "check_release_file",
1910
    "get_cachingallocator_config",
1911
    "get_clang_version",
1912
    "get_cmake_version",
1913
    "get_conda_packages",
1914
    "get_cpu_info",
1915
    "get_cuda_module_loading_config",
1916
    "get_cudnn_version",
1917
    "get_env_info",
1918
    "get_gcc_version",
1919
    "get_gpu_info",
1920
    "get_libc_version",
1921
    "get_lsb_version",
1922
    "get_mac_version",
1923
    "get_nvidia_driver_version",
1924
    "get_nvidia_smi",
1925
    "get_os",
1926
    "get_pip_packages",
1927
    "get_platform",
1928
    "get_pretty_env_info",
1929
    "get_python_platform",
1930
    "get_running_cuda_version",
1931
    "get_windows_version",
1932
    "is_xnnpack_available",
1933
    "pretty_str",
1934
    # torch.utils.cpp_backtrace
1935
    "get_cpp_backtrace",
1936
    # torch.utils.cpp_extension
1937
    "check_compiler_is_gcc",
1938
    "check_compiler_ok_for_platform",
1939
    "get_cxx_compiler",
1940
    "get_default_build_root",
1941
    "library_paths",
1942
    "remove_extension_h_precompiler_headers",
1943
    # torch.utils.data.backward_compatibility
1944
    "worker_init_fn",
1945
    # torch.utils.data.datapipes.dataframe.dataframe_wrapper
1946
    "concat",
1947
    "create_dataframe",
1948
    "get_columns",
1949
    "get_df_wrapper",
1950
    "get_item",
1951
    "get_len",
1952
    "is_column",
1953
    "is_dataframe",
1954
    "iterate",
1955
    "set_df_wrapper",
1956
    # torch.utils.data.datapipes.dataframe.dataframes
1957
    "disable_capture",
1958
    "get_val",
1959
    # torch.utils.data.datapipes.gen_pyi
1960
    "extract_class_name",
1961
    "extract_method_name",
1962
    "find_file_paths",
1963
    "gen_from_template",
1964
    "get_method_definitions",
1965
    "materialize_lines",
1966
    "parse_datapipe_file",
1967
    "parse_datapipe_files",
1968
    "process_signature",
1969
    "split_outside_bracket",
1970
    # torch.utils.data.datapipes.map.callable
1971
    "default_fn",
1972
    # torch.utils.data.datapipes.utils.common
1973
    "get_file_binaries_from_pathnames",
1974
    "get_file_pathnames_from_root",
1975
    "match_masks",
1976
    "validate_input_col",
1977
    "validate_pathname_binary_tuple",
1978
    # torch.utils.data.datapipes.utils.decoder
1979
    "audiohandler",
1980
    "basichandlers",
1981
    "extension_extract_fn",
1982
    "handle_extension",
1983
    "imagehandler",
1984
    "mathandler",
1985
    "videohandler",
1986
    # torch.utils.data.dataset
1987
    "random_split",
1988
    # torch.utils.data.graph
1989
    "traverse",
1990
    "traverse_dps",
1991
    # torch.utils.data.graph_settings
1992
    "apply_random_seed",
1993
    "apply_sharding",
1994
    "apply_shuffle_seed",
1995
    "apply_shuffle_settings",
1996
    "get_all_graph_pipes",
1997
    # torch.utils.flop_counter
1998
    "addmm_flop",
1999
    "baddbmm_flop",
2000
    "bmm_flop",
2001
    "conv_backward_flop",
2002
    "conv_flop",
2003
    "conv_flop_count",
2004
    "convert_num_with_suffix",
2005
    "get_shape",
2006
    "get_suffix_str",
2007
    "mm_flop",
2008
    "normalize_tuple",
2009
    "register_flop_formula",
2010
    "sdpa_backward_flop",
2011
    "sdpa_backward_flop_count",
2012
    "sdpa_flop",
2013
    "sdpa_flop_count",
2014
    "shape_wrapper",
2015
    "transpose_shape",
2016
    # torch.utils.hipify.hipify_python
2017
    "add_dim3",
2018
    "compute_stats",
2019
    "extract_arguments",
2020
    "file_add_header",
2021
    "file_specific_replacement",
2022
    "find_bracket_group",
2023
    "find_closure_group",
2024
    "find_parentheses_group",
2025
    "fix_static_global_kernels",
2026
    "get_hip_file_path",
2027
    "hip_header_magic",
2028
    "hipify",
2029
    "is_caffe2_gpu_file",
2030
    "is_cusparse_file",
2031
    "is_out_of_place",
2032
    "is_pytorch_file",
2033
    "is_special_file",
2034
    "match_extensions",
2035
    "matched_files_iter",
2036
    "openf",
2037
    "preprocess_file_and_save_result",
2038
    "preprocessor",
2039
    "processKernelLaunches",
2040
    "replace_extern_shared",
2041
    "replace_math_functions",
2042
    "str2bool",
2043
    # torch.utils.hooks
2044
    "unserializable_hook",
2045
    "warn_if_has_hooks",
2046
    # torch.utils.jit.log_extract
2047
    "extract_ir",
2048
    "load_graph_and_inputs",
2049
    "make_tensor_from_type",
2050
    "no_fuser",
2051
    "time_cpu",
2052
    "time_cuda",
2053
    # torch.utils.mkldnn
2054
    "to_mkldnn",
2055
    # torch.utils.mobile_optimizer
2056
    "generate_mobile_module_lints",
2057
    # torch.utils.tensorboard.summary
2058
    "audio",
2059
    "compute_curve",
2060
    "custom_scalars",
2061
    "draw_boxes",
2062
    "half_to_int",
2063
    "histogram",
2064
    "histogram_raw",
2065
    "hparams",
2066
    "image",
2067
    "image_boxes",
2068
    "int_to_half",
2069
    "make_histogram",
2070
    "make_image",
2071
    "make_video",
2072
    "mesh",
2073
    "pr_curve",
2074
    "pr_curve_raw",
2075
    "scalar",
2076
    "tensor_proto",
2077
    "text",
2078
    "video",
2079
    # torch.utils.throughput_benchmark
2080
    "format_time",
2081
]
2082

2083
coverage_ignore_classes = [
2084
    # torch
2085
    "FatalError",
2086
    "QUInt2x4Storage",
2087
    "Size",
2088
    "Storage",
2089
    "Stream",
2090
    "Tensor",
2091
    "finfo",
2092
    "iinfo",
2093
    "qscheme",
2094
    "AggregationType",
2095
    "AliasDb",
2096
    "AnyType",
2097
    "Argument",
2098
    "ArgumentSpec",
2099
    "AwaitType",
2100
    "BenchmarkConfig",
2101
    "BenchmarkExecutionStats",
2102
    "Block",
2103
    "BoolType",
2104
    "BufferDict",
2105
    "CallStack",
2106
    "Capsule",
2107
    "ClassType",
2108
    "Code",
2109
    "CompleteArgumentSpec",
2110
    "ComplexType",
2111
    "ConcreteModuleType",
2112
    "ConcreteModuleTypeBuilder",
2113
    "DeepCopyMemoTable",
2114
    "DeserializationStorageContext",
2115
    "DeviceObjType",
2116
    "DictType",
2117
    "DispatchKey",
2118
    "DispatchKeySet",
2119
    "EnumType",
2120
    "ExcludeDispatchKeyGuard",
2121
    "ExecutionPlan",
2122
    "FileCheck",
2123
    "FloatType",
2124
    "FunctionSchema",
2125
    "Gradient",
2126
    "Graph",
2127
    "GraphExecutorState",
2128
    "IODescriptor",
2129
    "InferredType",
2130
    "IntType",
2131
    "InterfaceType",
2132
    "ListType",
2133
    "LockingLogger",
2134
    "MobileOptimizerType",
2135
    "ModuleDict",
2136
    "Node",
2137
    "NoneType",
2138
    "NoopLogger",
2139
    "NumberType",
2140
    "OperatorInfo",
2141
    "OptionalType",
2142
    "ParameterDict",
2143
    "PyObjectType",
2144
    "PyTorchFileReader",
2145
    "PyTorchFileWriter",
2146
    "RRefType",
2147
    "ScriptClass",
2148
    "ScriptClassFunction",
2149
    "ScriptDict",
2150
    "ScriptDictIterator",
2151
    "ScriptDictKeyIterator",
2152
    "ScriptList",
2153
    "ScriptListIterator",
2154
    "ScriptMethod",
2155
    "ScriptModule",
2156
    "ScriptModuleSerializer",
2157
    "ScriptObject",
2158
    "ScriptObjectProperty",
2159
    "SerializationStorageContext",
2160
    "StaticModule",
2161
    "StringType",
2162
    "SymIntType",
2163
    "SymBoolType",
2164
    "ThroughputBenchmark",
2165
    "TracingState",
2166
    "TupleType",
2167
    "Type",
2168
    "UnionType",
2169
    "Use",
2170
    "Value",
2171
    # torch.cuda
2172
    "BFloat16Storage",
2173
    "BFloat16Tensor",
2174
    "BoolStorage",
2175
    "BoolTensor",
2176
    "ByteStorage",
2177
    "ByteTensor",
2178
    "CharStorage",
2179
    "CharTensor",
2180
    "ComplexDoubleStorage",
2181
    "ComplexFloatStorage",
2182
    "CudaError",
2183
    "DeferredCudaCallError",
2184
    "DoubleStorage",
2185
    "DoubleTensor",
2186
    "FloatStorage",
2187
    "FloatTensor",
2188
    "HalfStorage",
2189
    "HalfTensor",
2190
    "IntStorage",
2191
    "IntTensor",
2192
    "LongStorage",
2193
    "LongTensor",
2194
    "ShortStorage",
2195
    "ShortTensor",
2196
    "cudaStatus",
2197
    # torch.cuda._sanitizer
2198
    "Access",
2199
    "AccessType",
2200
    "Await",
2201
    "CUDASanitizer",
2202
    "CUDASanitizerDispatchMode",
2203
    "CUDASanitizerErrors",
2204
    "EventHandler",
2205
    "SynchronizationError",
2206
    "UnsynchronizedAccessError",
2207
    # torch.distributed.elastic.multiprocessing.errors
2208
    "ChildFailedError",
2209
    "ProcessFailure",
2210
    # torch.distributions.constraints
2211
    "cat",
2212
    "greater_than",
2213
    "greater_than_eq",
2214
    "half_open_interval",
2215
    "independent",
2216
    "integer_interval",
2217
    "interval",
2218
    "less_than",
2219
    "multinomial",
2220
    "stack",
2221
    # torch.distributions.transforms
2222
    "AffineTransform",
2223
    "CatTransform",
2224
    "ComposeTransform",
2225
    "CorrCholeskyTransform",
2226
    "CumulativeDistributionTransform",
2227
    "ExpTransform",
2228
    "IndependentTransform",
2229
    "PowerTransform",
2230
    "ReshapeTransform",
2231
    "SigmoidTransform",
2232
    "SoftmaxTransform",
2233
    "SoftplusTransform",
2234
    "StackTransform",
2235
    "StickBreakingTransform",
2236
    "TanhTransform",
2237
    "Transform",
2238
    # torch.jit
2239
    "CompilationUnit",
2240
    "Error",
2241
    "Future",
2242
    "ScriptFunction",
2243
    # torch.onnx
2244
    "CheckerError",
2245
    "ExportTypes",
2246
    # torch.backends
2247
    "ContextProp",
2248
    "PropModule",
2249
    # torch.backends.cuda
2250
    "cuBLASModule",
2251
    "cuFFTPlanCache",
2252
    "cuFFTPlanCacheAttrContextProp",
2253
    "cuFFTPlanCacheManager",
2254
    # torch.distributed.algorithms.ddp_comm_hooks
2255
    "DDPCommHookType",
2256
    # torch.jit.mobile
2257
    "LiteScriptModule",
2258
    # torch.ao.nn.quantized.modules
2259
    "DeQuantize",
2260
    "Quantize",
2261
    # torch.utils.backcompat
2262
    "Warning",
2263
    # torch.ao.nn.intrinsic.modules.fused
2264
    "ConvAdd2d",
2265
    "ConvAddReLU2d",
2266
    "LinearBn1d",
2267
    "LinearLeakyReLU",
2268
    "LinearTanh",
2269
    # torch.ao.nn.intrinsic.qat.modules.conv_fused
2270
    "ConvBnReLU1d",
2271
    "ConvBnReLU2d",
2272
    "ConvBnReLU3d",
2273
    "ConvReLU1d",
2274
    "ConvReLU2d",
2275
    "ConvReLU3d",
2276
    # torch.ao.nn.intrinsic.qat.modules.linear_fused
2277
    "LinearBn1d",
2278
    # torch.ao.nn.intrinsic.qat.modules.linear_relu
2279
    "LinearReLU",
2280
    # torch.ao.nn.intrinsic.quantized.dynamic.modules.linear_relu
2281
    "LinearReLU",
2282
    # torch.ao.nn.intrinsic.quantized.modules.bn_relu
2283
    "BNReLU2d",
2284
    "BNReLU3d",
2285
    # torch.ao.nn.intrinsic.quantized.modules.conv_add
2286
    "ConvAdd2d",
2287
    "ConvAddReLU2d",
2288
    # torch.ao.nn.intrinsic.quantized.modules.conv_relu
2289
    "ConvReLU1d",
2290
    "ConvReLU2d",
2291
    "ConvReLU3d",
2292
    # torch.ao.nn.intrinsic.quantized.modules.linear_relu
2293
    "LinearLeakyReLU",
2294
    "LinearReLU",
2295
    "LinearTanh",
2296
    # torch.ao.nn.qat.modules.conv
2297
    "Conv1d",
2298
    "Conv2d",
2299
    "Conv3d",
2300
    # torch.ao.nn.qat.modules.embedding_ops
2301
    "Embedding",
2302
    "EmbeddingBag",
2303
    # torch.ao.nn.qat.modules.linear
2304
    "Linear",
2305
    # torch.ao.nn.quantizable.modules.activation
2306
    "MultiheadAttention",
2307
    # torch.ao.nn.quantizable.modules.rnn
2308
    "LSTM",
2309
    "LSTMCell",
2310
    # torch.ao.nn.quantized.dynamic.modules.conv
2311
    "Conv1d",
2312
    "Conv2d",
2313
    "Conv3d",
2314
    "ConvTranspose1d",
2315
    "ConvTranspose2d",
2316
    "ConvTranspose3d",
2317
    # torch.ao.nn.quantized.dynamic.modules.linear
2318
    "Linear",
2319
    # torch.ao.nn.quantized.dynamic.modules.rnn
2320
    "GRU",
2321
    "GRUCell",
2322
    "LSTM",
2323
    "LSTMCell",
2324
    "PackedParameter",
2325
    "RNNBase",
2326
    "RNNCell",
2327
    "RNNCellBase",
2328
    # torch.ao.nn.quantized.modules.activation
2329
    "ELU",
2330
    "Hardswish",
2331
    "LeakyReLU",
2332
    "MultiheadAttention",
2333
    "PReLU",
2334
    "ReLU6",
2335
    "Sigmoid",
2336
    "Softmax",
2337
    # torch.ao.nn.quantized.modules.batchnorm
2338
    "BatchNorm2d",
2339
    "BatchNorm3d",
2340
    # torch.ao.nn.quantized.modules.conv
2341
    "Conv1d",
2342
    "Conv2d",
2343
    "Conv3d",
2344
    "ConvTranspose1d",
2345
    "ConvTranspose2d",
2346
    "ConvTranspose3d",
2347
    # torch.ao.nn.quantized.modules.dropout
2348
    "Dropout",
2349
    # torch.ao.nn.quantized.modules.embedding_ops
2350
    "Embedding",
2351
    "EmbeddingBag",
2352
    "EmbeddingPackedParams",
2353
    # torch.ao.nn.quantized.modules.functional_modules
2354
    "FXFloatFunctional",
2355
    "FloatFunctional",
2356
    "QFunctional",
2357
    # torch.ao.nn.quantized.modules.linear
2358
    "Linear",
2359
    "LinearPackedParams",
2360
    # torch.ao.nn.quantized.modules.normalization
2361
    "GroupNorm",
2362
    "InstanceNorm1d",
2363
    "InstanceNorm2d",
2364
    "InstanceNorm3d",
2365
    "LayerNorm",
2366
    # torch.ao.nn.quantized.modules.rnn
2367
    "LSTM",
2368
    # torch.ao.nn.quantized.modules.utils
2369
    "WeightedQuantizedModule",
2370
    # torch.ao.nn.quantized.reference.modules.conv
2371
    "Conv1d",
2372
    "Conv2d",
2373
    "Conv3d",
2374
    "ConvTranspose1d",
2375
    "ConvTranspose2d",
2376
    "ConvTranspose3d",
2377
    # torch.ao.nn.quantized.reference.modules.linear
2378
    "Linear",
2379
    # torch.ao.nn.quantized.reference.modules.rnn
2380
    "GRU",
2381
    "GRUCell",
2382
    "LSTM",
2383
    "LSTMCell",
2384
    "RNNBase",
2385
    "RNNCell",
2386
    "RNNCellBase",
2387
    # torch.ao.nn.quantized.reference.modules.sparse
2388
    "Embedding",
2389
    "EmbeddingBag",
2390
    # torch.ao.nn.quantized.reference.modules.utils
2391
    "ReferenceQuantizedModule",
2392
    # torch.ao.nn.sparse.quantized.dynamic.linear
2393
    "Linear",
2394
    # torch.ao.nn.sparse.quantized.linear
2395
    "Linear",
2396
    "LinearPackedParams",
2397
    # torch.ao.nn.sparse.quantized.utils
2398
    "LinearBlockSparsePattern",
2399
    # torch.ao.ns.fx.graph_matcher
2400
    "SubgraphTypeRelationship",
2401
    # torch.ao.ns.fx.n_shadows_utils
2402
    "OutputProp",
2403
    # torch.ao.ns.fx.ns_types
2404
    "NSSingleResultValuesType",
2405
    "NSSubgraph",
2406
    # torch.ao.ns.fx.qconfig_multi_mapping
2407
    "QConfigMultiMapping",
2408
    # torch.ao.pruning.scheduler.base_scheduler
2409
    "BaseScheduler",
2410
    # torch.ao.pruning.scheduler.cubic_scheduler
2411
    "CubicSL",
2412
    # torch.ao.pruning.scheduler.lambda_scheduler
2413
    "LambdaSL",
2414
    # torch.ao.pruning.sparsifier.base_sparsifier
2415
    "BaseSparsifier",
2416
    # torch.ao.pruning.sparsifier.nearly_diagonal_sparsifier
2417
    "NearlyDiagonalSparsifier",
2418
    # torch.ao.pruning.sparsifier.utils
2419
    "FakeSparsity",
2420
    # torch.ao.pruning.sparsifier.weight_norm_sparsifier
2421
    "WeightNormSparsifier",
2422
    # torch.ao.quantization.backend_config.backend_config
2423
    "BackendConfig",
2424
    "BackendPatternConfig",
2425
    "DTypeConfig",
2426
    # torch.ao.quantization.fake_quantize
2427
    "FakeQuantize",
2428
    "FakeQuantizeBase",
2429
    "FixedQParamsFakeQuantize",
2430
    "FusedMovingAvgObsFakeQuantize",
2431
    # torch.ao.quantization.fx.fuse_handler
2432
    "DefaultFuseHandler",
2433
    "FuseHandler",
2434
    # torch.ao.quantization.fx.graph_module
2435
    "FusedGraphModule",
2436
    "ObservedGraphModule",
2437
    "ObservedStandaloneGraphModule",
2438
    # torch.ao.quantization.fx.quantize_handler
2439
    "BatchNormQuantizeHandler",
2440
    "BinaryOpQuantizeHandler",
2441
    "CatQuantizeHandler",
2442
    "ConvReluQuantizeHandler",
2443
    "CopyNodeQuantizeHandler",
2444
    "CustomModuleQuantizeHandler",
2445
    "DefaultNodeQuantizeHandler",
2446
    "EmbeddingQuantizeHandler",
2447
    "FixedQParamsOpQuantizeHandler",
2448
    "GeneralTensorShapeOpQuantizeHandler",
2449
    "LinearReLUQuantizeHandler",
2450
    "RNNDynamicQuantizeHandler",
2451
    "StandaloneModuleQuantizeHandler",
2452
    # torch.ao.quantization.fx.tracer
2453
    "QuantizationTracer",
2454
    "ScopeContextManager",
2455
    # torch.ao.quantization.fx.utils
2456
    "ObservedGraphModuleAttrs",
2457
    # torch.ao.quantization.observer
2458
    "FixedQParamsObserver",
2459
    "HistogramObserver",
2460
    "MinMaxObserver",
2461
    "MovingAverageMinMaxObserver",
2462
    "MovingAveragePerChannelMinMaxObserver",
2463
    "NoopObserver",
2464
    "ObserverBase",
2465
    "PerChannelMinMaxObserver",
2466
    "PlaceholderObserver",
2467
    "RecordingObserver",
2468
    "ReuseInputObserver",
2469
    "UniformQuantizationObserverBase",
2470
    "default_debug_observer",
2471
    "default_placeholder_observer",
2472
    "default_reuse_input_observer",
2473
    # torch.ao.quantization.pt2e.duplicate_dq_pass
2474
    "DuplicateDQPass",
2475
    # torch.ao.quantization.pt2e.port_metadata_pass
2476
    "PortNodeMetaForQDQ",
2477
    # torch.ao.quantization.qconfig
2478
    "QConfigDynamic",
2479
    # torch.ao.quantization.quant_type
2480
    "QuantType",
2481
    # torch.ao.quantization.quantizer.composable_quantizer
2482
    "ComposableQuantizer",
2483
    # torch.ao.quantization.quantizer.embedding_quantizer
2484
    "EmbeddingQuantizer",
2485
    # torch.ao.quantization.quantizer.quantizer
2486
    "DerivedQuantizationSpec",
2487
    "FixedQParamsQuantizationSpec",
2488
    "QuantizationAnnotation",
2489
    "QuantizationSpec",
2490
    "QuantizationSpecBase",
2491
    "SharedQuantizationSpec",
2492
    # torch.ao.quantization.quantizer.x86_inductor_quantizer
2493
    "X86InductorQuantizer",
2494
    # torch.ao.quantization.quantizer.xnnpack_quantizer
2495
    "XNNPACKQuantizer",
2496
    # torch.ao.quantization.quantizer.xnnpack_quantizer_utils
2497
    "OperatorConfig",
2498
    "QuantizationConfig",
2499
    # torch.ao.quantization.stubs
2500
    "DeQuantStub",
2501
    "QuantStub",
2502
    "QuantWrapper",
2503
    # torch.ao.quantization.utils
2504
    "MatchAllNode",
2505
    # torch.backends.cudnn.rnn
2506
    "Unserializable",
2507
    # torch.amp.grad_scaler
2508
    "GradScaler",
2509
    "OptState",
2510
    # torch.cuda.graphs
2511
    "CUDAGraph",
2512
    # torch.cuda.streams
2513
    "Event",
2514
    # torch.distributed.algorithms.ddp_comm_hooks.post_localSGD_hook
2515
    "PostLocalSGDState",
2516
    # torch.distributed.algorithms.ddp_comm_hooks.powerSGD_hook
2517
    "PowerSGDState",
2518
    # torch.distributed.algorithms.join
2519
    "Join",
2520
    "JoinHook",
2521
    "Joinable",
2522
    # torch.distributed.algorithms.model_averaging.averagers
2523
    "ModelAverager",
2524
    "PeriodicModelAverager",
2525
    # torch.distributed.algorithms.model_averaging.hierarchical_model_averager
2526
    "HierarchicalModelAverager",
2527
    # torch.distributed.argparse_util
2528
    "check_env",
2529
    "env",
2530
    # torch.distributed.checkpoint.api
2531
    "CheckpointException",
2532
    # torch.distributed.checkpoint.default_planner
2533
    "DefaultLoadPlanner",
2534
    "DefaultSavePlanner",
2535
    # torch.distributed.checkpoint.filesystem
2536
    "FileSystemReader",
2537
    "FileSystemWriter",
2538
    # torch.distributed.checkpoint.metadata
2539
    "BytesStorageMetadata",
2540
    "ChunkStorageMetadata",
2541
    "Metadata",
2542
    "MetadataIndex",
2543
    # torch.distributed.checkpoint.planner
2544
    "LoadItemType",
2545
    "LoadPlanner",
2546
    "SavePlanner",
2547
    "WriteItemType",
2548
    # torch.distributed.checkpoint.state_dict
2549
    "DistributedStateDictOptions",
2550
    # torch.distributed.checkpoint.storage
2551
    "WriteResult",
2552
    # torch.distributed.collective_utils
2553
    "SyncPayload",
2554
    # torch.distributed.distributed_c10d
2555
    "AllToAllOptions",
2556
    "AllreduceCoalescedOptions",
2557
    "AllreduceOptions",
2558
    "Backend",
2559
    "BackendConfig",
2560
    "BarrierOptions",
2561
    "BroadcastOptions",
2562
    "DebugLevel",
2563
    "GatherOptions",
2564
    "GroupMember",
2565
    "ProcessGroup",
2566
    "ProcessGroupGloo",
2567
    "ProcessGroupNCCL",
2568
    "ReduceOptions",
2569
    "ReduceScatterOptions",
2570
    "ScatterOptions",
2571
    "Work",
2572
    "group",
2573
    # torch.distributed.elastic.agent.server.api
2574
    "ElasticAgent",
2575
    "RunResult",
2576
    "SimpleElasticAgent",
2577
    "WorkerSpec",
2578
    # torch.distributed.elastic.events.api
2579
    "Event",
2580
    "RdzvEvent",
2581
    # torch.distributed.elastic.metrics.api
2582
    "ConsoleMetricHandler",
2583
    "MetricData",
2584
    "MetricHandler",
2585
    "MetricStream",
2586
    "MetricsConfig",
2587
    "NullMetricHandler",
2588
    # torch.distributed.elastic.multiprocessing.api
2589
    "MultiprocessContext",
2590
    "PContext",
2591
    "RunProcsResult",
2592
    "SignalException",
2593
    "Std",
2594
    "SubprocessContext",
2595
    "SubprocessHandler",
2596
    # torch.distributed.elastic.multiprocessing.tail_log
2597
    "TailLog",
2598
    # torch.distributed.elastic.rendezvous.api
2599
    "RendezvousHandler",
2600
    "RendezvousHandlerRegistry",
2601
    "RendezvousParameters",
2602
    # torch.distributed.elastic.rendezvous.dynamic_rendezvous
2603
    "DynamicRendezvousHandler",
2604
    "RendezvousSettings",
2605
    # torch.distributed.elastic.rendezvous.etcd_rendezvous
2606
    "EtcdRendezvous",
2607
    "EtcdRendezvousHandler",
2608
    "EtcdRendezvousRetryImmediately",
2609
    "EtcdRendezvousRetryableFailure",
2610
    # torch.distributed.elastic.rendezvous.etcd_server
2611
    "EtcdServer",
2612
    # torch.distributed.elastic.rendezvous.static_tcp_rendezvous
2613
    "StaticTCPRendezvous",
2614
    # torch.distributed.elastic.timer.api
2615
    "RequestQueue",
2616
    "TimerClient",
2617
    "TimerServer",
2618
    # torch.distributed.elastic.timer.file_based_local_timer
2619
    "FileTimerClient",
2620
    "FileTimerRequest",
2621
    "FileTimerServer",
2622
    # torch.distributed.elastic.timer.local_timer
2623
    "LocalTimerClient",
2624
    "LocalTimerServer",
2625
    "MultiprocessingRequestQueue",
2626
    # torch.distributed.elastic.utils.api
2627
    "macros",
2628
    # torch.distributed.elastic.utils.data.cycling_iterator
2629
    "CyclingIterator",
2630
    # torch.distributed.elastic.utils.data.elastic_distributed_sampler
2631
    "ElasticDistributedSampler",
2632
    # torch.distributed.fsdp.api
2633
    "StateDictType",
2634
    # torch.distributed.fsdp.fully_sharded_data_parallel
2635
    "FullyShardedDataParallel",
2636
    "OptimStateKeyType",
2637
    # torch.distributed.fsdp.sharded_grad_scaler
2638
    "ShardedGradScaler",
2639
    # torch.distributed.fsdp.wrap
2640
    "CustomPolicy",
2641
    "ModuleWrapPolicy",
2642
    # torch.distributed.launcher.api
2643
    "LaunchConfig",
2644
    "elastic_launch",
2645
    # torch.distributed.optim.optimizer
2646
    "DistributedOptimizer",
2647
    # torch.distributed.optim.post_localSGD_optimizer
2648
    "PostLocalSGDOptimizer",
2649
    # torch.distributed.optim.zero_redundancy_optimizer
2650
    "ZeroRedundancyOptimizer",
2651
    # torch.distributed.pipeline.sync.batchnorm
2652
    "DeferredBatchNorm",
2653
    # torch.distributed.pipeline.sync.checkpoint
2654
    "Checkpoint",
2655
    "Checkpointing",
2656
    "Context",
2657
    "Function",
2658
    "Recompute",
2659
    "ThreadLocal",
2660
    # torch.distributed.pipeline.sync.copy
2661
    "Context",
2662
    "Copy",
2663
    "Wait",
2664
    # torch.distributed.pipeline.sync.dependency
2665
    "Fork",
2666
    "Join",
2667
    # torch.distributed.pipeline.sync.microbatch
2668
    "Batch",
2669
    "NoChunk",
2670
    # torch.distributed.pipeline.sync.pipe
2671
    "BalanceError",
2672
    "Pipe",
2673
    "PipeSequential",
2674
    "WithDevice",
2675
    # torch.distributed.pipeline.sync.pipeline
2676
    "Pipeline",
2677
    # torch.distributed.pipeline.sync.skip.layout
2678
    "SkipLayout",
2679
    # torch.distributed.pipeline.sync.skip.namespace
2680
    "Namespace",
2681
    # torch.distributed.pipeline.sync.skip.portal
2682
    "Context",
2683
    "Portal",
2684
    "PortalBlue",
2685
    "PortalCopy",
2686
    "PortalOrange",
2687
    # torch.distributed.pipeline.sync.skip.skippable
2688
    "Skippable",
2689
    # torch.distributed.pipeline.sync.skip.tracker
2690
    "SkipTracker",
2691
    "SkipTrackerThroughPotals",
2692
    "ThreadLocal",
2693
    # torch.distributed.pipeline.sync.stream
2694
    "CPUStreamType",
2695
    # torch.distributed.pipeline.sync.worker
2696
    "Task",
2697
    # torch.distributed.rpc.api
2698
    "AllGatherStates",
2699
    "RRef",
2700
    # torch.distributed.rpc.backend_registry
2701
    "BackendValue",
2702
    # torch.distributed.rpc.internal
2703
    "PythonUDF",
2704
    "RPCExecMode",
2705
    "RemoteException",
2706
    # torch.distributed.rpc.rref_proxy
2707
    "RRefProxy",
2708
    # torch.distributed.tensor.parallel.fsdp
2709
    "DTensorExtensions",
2710
    # torch.distributed.tensor.parallel.style
2711
    "ParallelStyle",
2712
    # torch.distributions.logistic_normal
2713
    "LogisticNormal",
2714
    # torch.distributions.one_hot_categorical
2715
    "OneHotCategoricalStraightThrough",
2716
    # torch.distributions.relaxed_categorical
2717
    "ExpRelaxedCategorical",
2718
    # torch.distributions.utils
2719
    "lazy_property",
2720
    # torch.export.exported_program
2721
    "ConstantArgument",
2722
    "ExportedProgram",
2723
    # torch.fx.experimental.accelerator_partitioner
2724
    "DAG",
2725
    "DAGNode",
2726
    "PartitionResult",
2727
    "Partitioner",
2728
    # torch.fx.experimental.const_fold
2729
    "FoldedGraphModule",
2730
    # torch.fx.experimental.graph_gradual_typechecker
2731
    "Refine",
2732
    # torch.fx.experimental.meta_tracer
2733
    "MetaAttribute",
2734
    "MetaDeviceAttribute",
2735
    "MetaProxy",
2736
    "MetaTracer",
2737
    # torch.fx.experimental.migrate_gradual_types.constraint
2738
    "ApplyBroadcasting",
2739
    "BVar",
2740
    "BinConstraintD",
2741
    "BinConstraintT",
2742
    "BinaryConstraint",
2743
    "CalcConv",
2744
    "CalcMaxPool",
2745
    "CalcProduct",
2746
    "CanReshape",
2747
    "Conj",
2748
    "Constraint",
2749
    "DGreatestUpperBound",
2750
    "DVar",
2751
    "Disj",
2752
    "F",
2753
    "GetItem",
2754
    "GetItemTensor",
2755
    "IndexSelect",
2756
    "Prod",
2757
    "T",
2758
    "TGreatestUpperBound",
2759
    "TVar",
2760
    "Transpose",
2761
    # torch.fx.experimental.migrate_gradual_types.constraint_generator
2762
    "ConstraintGenerator",
2763
    # torch.fx.experimental.normalize
2764
    "NormalizeArgs",
2765
    "NormalizeOperators",
2766
    # torch.fx.experimental.optimization
2767
    "MklSubgraph",
2768
    "UnionFind",
2769
    # torch.fx.experimental.partitioner_utils
2770
    "Device",
2771
    "Partition",
2772
    "PartitionLatency",
2773
    "PartitionMode",
2774
    "PartitionerConfig",
2775
    # torch.fx.experimental.proxy_tensor
2776
    "DecompositionInterpreter",
2777
    "PreDispatchTorchFunctionMode",
2778
    "ProxySymDispatchMode",
2779
    "ProxyTorchDispatchMode",
2780
    "PythonKeyTracer",
2781
    # torch.fx.experimental.recording
2782
    "FakeTensorMeta",
2783
    "NotEqualError",
2784
    "ShapeEnvEvent",
2785
    # torch.fx.experimental.refinement_types
2786
    "Equality",
2787
    # torch.fx.experimental.rewriter
2788
    "AST_Rewriter",
2789
    "RewritingTracer",
2790
    # torch.fx.experimental.schema_type_annotation
2791
    "AnnotateTypesWithSchema",
2792
    # torch.fx.experimental.sym_node
2793
    "SymNode",
2794
    # torch.fx.experimental.symbolic_shapes
2795
    "Constraint",
2796
    "ConstraintViolationError",
2797
    "DynamicDimConstraintPrinter",
2798
    "GuardOnDataDependentSymNode",
2799
    "LoggingShapeGuardPrinter",
2800
    "RelaxedUnspecConstraint",
2801
    "RuntimeAssert",
2802
    "ShapeGuardPrinter",
2803
    "SymDispatchMode",
2804
    "SymbolicContext",
2805
    # torch.fx.experimental.unification.match
2806
    "Dispatcher",
2807
    "VarDispatcher",
2808
    # torch.fx.experimental.unification.multipledispatch.conflict
2809
    "AmbiguityWarning",
2810
    # torch.fx.experimental.unification.multipledispatch.dispatcher
2811
    "Dispatcher",
2812
    "MDNotImplementedError",
2813
    "MethodDispatcher",
2814
    # torch.fx.experimental.unification.multipledispatch.variadic
2815
    "Variadic",
2816
    "VariadicSignatureMeta",
2817
    "VariadicSignatureType",
2818
    # torch.fx.experimental.unification.variable
2819
    "Var",
2820
    # torch.fx.experimental.validator
2821
    "BisectValidationException",
2822
    "PopulateValidator",
2823
    "SympyToZ3",
2824
    "ValidationException",
2825
    # torch.fx.graph
2826
    "PythonCode",
2827
    # torch.fx.immutable_collections
2828
    "immutable_dict",
2829
    "immutable_list",
2830
    # torch.fx.interpreter
2831
    "Interpreter",
2832
    # torch.fx.operator_schemas
2833
    "ArgsKwargsPair",
2834
    # torch.fx.passes.backends.cudagraphs
2835
    "CudaGraphsSupport",
2836
    # torch.fx.passes.dialect.common.cse_pass
2837
    "CSEPass",
2838
    # torch.fx.passes.fake_tensor_prop
2839
    "FakeTensorProp",
2840
    # torch.fx.passes.graph_drawer
2841
    "FxGraphDrawer",
2842
    # torch.fx.passes.graph_manipulation
2843
    "size_bytes",
2844
    # torch.fx.passes.infra.partitioner
2845
    "CapabilityBasedPartitioner",
2846
    "Partition",
2847
    # torch.fx.passes.infra.pass_base
2848
    "PassBase",
2849
    "PassResult",
2850
    # torch.fx.passes.infra.pass_manager
2851
    "PassManager",
2852
    # torch.fx.passes.net_min_base
2853
    "FxNetMinimizerBadModuleError",
2854
    "FxNetMinimizerResultMismatchError",
2855
    "FxNetMinimizerRunFuncError",
2856
    # torch.fx.passes.operator_support
2857
    "OpSupports",
2858
    "OperatorSupport",
2859
    "OperatorSupportBase",
2860
    # torch.fx.passes.pass_manager
2861
    "PassManager",
2862
    # torch.fx.passes.shape_prop
2863
    "ShapeProp",
2864
    # torch.fx.passes.split_module
2865
    "Partition",
2866
    # torch.fx.passes.split_utils
2867
    "Component",
2868
    # torch.fx.passes.splitter_base
2869
    "FxNetAccNodesFinder",
2870
    "FxNetSplitterInternalError",
2871
    "SplitResult",
2872
    "Subgraph",
2873
    # torch.fx.passes.tests.test_pass_manager
2874
    "TestPassManager",
2875
    # torch.fx.passes.tools_common
2876
    "FxNetAccFusionsFinder",
2877
    # torch.fx.passes.utils.common
2878
    "HolderModule",
2879
    # torch.fx.passes.utils.matcher_utils
2880
    "InternalMatch",
2881
    "SubgraphMatcher",
2882
    # torch.fx.passes.utils.source_matcher_utils
2883
    "SourcePartition",
2884
    # torch.fx.proxy
2885
    "Attribute",
2886
    "ParameterProxy",
2887
    "Proxy",
2888
    "Scope",
2889
    "ScopeContextManager",
2890
    "TraceError",
2891
    "TracerBase",
2892
    # torch.fx.subgraph_rewriter
2893
    "Match",
2894
    "ReplacedPatterns",
2895
    # torch.jit.annotations
2896
    "EvalEnv",
2897
    "Module",
2898
    # torch.jit.frontend
2899
    "Builder",
2900
    "ExprBuilder",
2901
    "FrontendError",
2902
    "FrontendTypeError",
2903
    "NotSupportedError",
2904
    "StmtBuilder",
2905
    "UnsupportedNodeError",
2906
    "WithItemBuilder",
2907
    # torch.masked.maskedtensor.core
2908
    "MaskedTensor",
2909
    # torch.multiprocessing.pool
2910
    "Pool",
2911
    # torch.multiprocessing.queue
2912
    "ConnectionWrapper",
2913
    "Queue",
2914
    "SimpleQueue",
2915
    # torch.multiprocessing.reductions
2916
    "SharedCache",
2917
    # torch.multiprocessing.spawn
2918
    "ProcessContext",
2919
    "ProcessException",
2920
    "ProcessExitedException",
2921
    "ProcessRaisedException",
2922
    "SpawnContext",
2923
    # torch.nn.cpp
2924
    "ModuleWrapper",
2925
    "OrderedDictWrapper",
2926
    # torch.nn.modules.activation
2927
    "CELU",
2928
    "ELU",
2929
    "GELU",
2930
    "GLU",
2931
    "Hardshrink",
2932
    "Hardsigmoid",
2933
    "Hardswish",
2934
    "Hardtanh",
2935
    "LeakyReLU",
2936
    "LogSigmoid",
2937
    "LogSoftmax",
2938
    "Mish",
2939
    "MultiheadAttention",
2940
    "PReLU",
2941
    "RReLU",
2942
    "ReLU",
2943
    "ReLU6",
2944
    "SELU",
2945
    "SiLU",
2946
    "Sigmoid",
2947
    "Softmax",
2948
    "Softmax2d",
2949
    "Softmin",
2950
    "Softplus",
2951
    "Softshrink",
2952
    "Softsign",
2953
    "Tanh",
2954
    "Tanhshrink",
2955
    "Threshold",
2956
    # torch.nn.modules.adaptive
2957
    "AdaptiveLogSoftmaxWithLoss",
2958
    # torch.nn.modules.batchnorm
2959
    "SyncBatchNorm",
2960
    # torch.nn.modules.channelshuffle
2961
    "ChannelShuffle",
2962
    # torch.nn.modules.container
2963
    "Container",
2964
    "ModuleList",
2965
    "ParameterList",
2966
    "Sequential",
2967
    # torch.nn.modules.conv
2968
    "Conv1d",
2969
    "Conv2d",
2970
    "Conv3d",
2971
    "ConvTranspose1d",
2972
    "ConvTranspose2d",
2973
    "ConvTranspose3d",
2974
    # torch.nn.modules.distance
2975
    "CosineSimilarity",
2976
    "PairwiseDistance",
2977
    # torch.nn.modules.dropout
2978
    "AlphaDropout",
2979
    "Dropout",
2980
    "Dropout1d",
2981
    "Dropout2d",
2982
    "Dropout3d",
2983
    "FeatureAlphaDropout",
2984
    # torch.nn.modules.flatten
2985
    "Flatten",
2986
    "Unflatten",
2987
    # torch.nn.modules.fold
2988
    "Fold",
2989
    "Unfold",
2990
    # torch.nn.modules.linear
2991
    "Bilinear",
2992
    "Identity",
2993
    "LazyLinear",
2994
    "Linear",
2995
    "NonDynamicallyQuantizableLinear",
2996
    # torch.nn.modules.loss
2997
    "BCELoss",
2998
    "BCEWithLogitsLoss",
2999
    "CTCLoss",
3000
    "CosineEmbeddingLoss",
3001
    "CrossEntropyLoss",
3002
    "GaussianNLLLoss",
3003
    "HingeEmbeddingLoss",
3004
    "HuberLoss",
3005
    "KLDivLoss",
3006
    "L1Loss",
3007
    "MSELoss",
3008
    "MarginRankingLoss",
3009
    "MultiLabelMarginLoss",
3010
    "MultiLabelSoftMarginLoss",
3011
    "MultiMarginLoss",
3012
    "NLLLoss",
3013
    "NLLLoss2d",
3014
    "PoissonNLLLoss",
3015
    "SmoothL1Loss",
3016
    "SoftMarginLoss",
3017
    "TripletMarginLoss",
3018
    "TripletMarginWithDistanceLoss",
3019
    # torch.nn.modules.module
3020
    "Module",
3021
    # torch.nn.modules.normalization
3022
    "CrossMapLRN2d",
3023
    "GroupNorm",
3024
    "LayerNorm",
3025
    "LocalResponseNorm",
3026
    # torch.nn.modules.padding
3027
    "CircularPad1d",
3028
    "CircularPad2d",
3029
    "CircularPad3d",
3030
    "ZeroPad1d",
3031
    "ZeroPad2d",
3032
    "ZeroPad3d",
3033
    # torch.nn.modules.pixelshuffle
3034
    "PixelShuffle",
3035
    "PixelUnshuffle",
3036
    # torch.nn.modules.pooling
3037
    "AdaptiveAvgPool1d",
3038
    "AdaptiveAvgPool2d",
3039
    "AdaptiveAvgPool3d",
3040
    "AdaptiveMaxPool1d",
3041
    "AdaptiveMaxPool2d",
3042
    "AdaptiveMaxPool3d",
3043
    "AvgPool1d",
3044
    "AvgPool2d",
3045
    "AvgPool3d",
3046
    "FractionalMaxPool2d",
3047
    "FractionalMaxPool3d",
3048
    "LPPool1d",
3049
    "LPPool2d",
3050
    "LPPool3d",
3051
    "MaxPool1d",
3052
    "MaxPool2d",
3053
    "MaxPool3d",
3054
    "MaxUnpool1d",
3055
    "MaxUnpool2d",
3056
    "MaxUnpool3d",
3057
    # torch.nn.modules.rnn
3058
    "GRU",
3059
    "GRUCell",
3060
    "LSTM",
3061
    "LSTMCell",
3062
    "RNN",
3063
    "RNNBase",
3064
    "RNNCell",
3065
    "RNNCellBase",
3066
    # torch.nn.modules.sparse
3067
    "Embedding",
3068
    "EmbeddingBag",
3069
    # torch.nn.modules.upsampling
3070
    "Upsample",
3071
    # torch.nn.parallel.data_parallel
3072
    "DataParallel",
3073
    # torch.nn.parallel.distributed
3074
    "DistributedDataParallel",
3075
    # torch.nn.parameter
3076
    "UninitializedTensorMixin",
3077
    # torch.nn.utils.parametrize
3078
    "ParametrizationList",
3079
    # torch.nn.utils.prune
3080
    "CustomFromMask",
3081
    "Identity",
3082
    "L1Unstructured",
3083
    "RandomUnstructured",
3084
    # torch.nn.utils.rnn
3085
    "PackedSequence",
3086
    "PackedSequence_",
3087
    # torch.nn.utils.spectral_norm
3088
    "SpectralNorm",
3089
    "SpectralNormLoadStateDictPreHook",
3090
    "SpectralNormStateDictHook",
3091
    # torch.nn.utils.weight_norm
3092
    "WeightNorm",
3093
    # torch.onnx.errors
3094
    "OnnxExporterError",
3095
    "OnnxExporterWarning",
3096
    "SymbolicValueError",
3097
    "UnsupportedOperatorError",
3098
    # torch.onnx.verification
3099
    "OnnxBackend",
3100
    "OnnxTestCaseRepro",
3101
    # torch.optim.adadelta
3102
    "Adadelta",
3103
    # torch.optim.adagrad
3104
    "Adagrad",
3105
    # torch.optim.adam
3106
    "Adam",
3107
    # torch.optim.adamax
3108
    "Adamax",
3109
    # torch.optim.adamw
3110
    "AdamW",
3111
    # torch.optim.asgd
3112
    "ASGD",
3113
    # torch.optim.lbfgs
3114
    "LBFGS",
3115
    # torch.optim.lr_scheduler
3116
    "ChainedScheduler",
3117
    "ConstantLR",
3118
    "CosineAnnealingLR",
3119
    "CosineAnnealingWarmRestarts",
3120
    "CyclicLR",
3121
    "ExponentialLR",
3122
    "LRScheduler",
3123
    "LambdaLR",
3124
    "LinearLR",
3125
    "MultiStepLR",
3126
    "MultiplicativeLR",
3127
    "OneCycleLR",
3128
    "PolynomialLR",
3129
    "ReduceLROnPlateau",
3130
    "SequentialLR",
3131
    "StepLR",
3132
    # torch.optim.nadam
3133
    "NAdam",
3134
    # torch.optim.optimizer
3135
    "Optimizer",
3136
    # torch.optim.radam
3137
    "RAdam",
3138
    # torch.optim.rmsprop
3139
    "RMSprop",
3140
    # torch.optim.rprop
3141
    "Rprop",
3142
    # torch.optim.sgd
3143
    "SGD",
3144
    # torch.optim.sparse_adam
3145
    "SparseAdam",
3146
    # torch.optim.swa_utils
3147
    "AveragedModel",
3148
    "SWALR",
3149
    # torch.overrides
3150
    "BaseTorchFunctionMode",
3151
    "TorchFunctionMode",
3152
    # torch.package.file_structure_representation
3153
    "Directory",
3154
    # torch.package.glob_group
3155
    "GlobGroup",
3156
    # torch.package.importer
3157
    "Importer",
3158
    "ObjMismatchError",
3159
    "ObjNotFoundError",
3160
    "OrderedImporter",
3161
    # torch.package.package_exporter
3162
    "PackageExporter",
3163
    "PackagingErrorReason",
3164
    # torch.package.package_importer
3165
    "PackageImporter",
3166
    # torch.profiler.profiler
3167
    "ExecutionTraceObserver",
3168
    "profile",
3169
    # torch.return_types
3170
    "aminmax",
3171
    "aminmax_out",
3172
    "cummax",
3173
    "cummax_out",
3174
    "cummin",
3175
    "cummin_out",
3176
    "frexp",
3177
    "frexp_out",
3178
    "geqrf",
3179
    "geqrf_out",
3180
    "histogram",
3181
    "histogram_out",
3182
    "histogramdd",
3183
    "kthvalue",
3184
    "kthvalue_out",
3185
    "linalg_cholesky_ex",
3186
    "linalg_cholesky_ex_out",
3187
    "linalg_eig",
3188
    "linalg_eig_out",
3189
    "linalg_eigh",
3190
    "linalg_eigh_out",
3191
    "linalg_inv_ex",
3192
    "linalg_inv_ex_out",
3193
    "linalg_ldl_factor",
3194
    "linalg_ldl_factor_ex",
3195
    "linalg_ldl_factor_ex_out",
3196
    "linalg_ldl_factor_out",
3197
    "linalg_lstsq",
3198
    "linalg_lstsq_out",
3199
    "linalg_lu",
3200
    "linalg_lu_factor",
3201
    "linalg_lu_factor_ex",
3202
    "linalg_lu_factor_ex_out",
3203
    "linalg_lu_factor_out",
3204
    "linalg_lu_out",
3205
    "linalg_qr",
3206
    "linalg_qr_out",
3207
    "linalg_slogdet",
3208
    "linalg_slogdet_out",
3209
    "linalg_solve_ex",
3210
    "linalg_solve_ex_out",
3211
    "linalg_svd",
3212
    "linalg_svd_out",
3213
    "lu_unpack",
3214
    "lu_unpack_out",
3215
    "max",
3216
    "max_out",
3217
    "median",
3218
    "median_out",
3219
    "min",
3220
    "min_out",
3221
    "mode",
3222
    "mode_out",
3223
    "nanmedian",
3224
    "nanmedian_out",
3225
    "qr",
3226
    "qr_out",
3227
    "slogdet",
3228
    "slogdet_out",
3229
    "sort",
3230
    "sort_out",
3231
    "svd",
3232
    "svd_out",
3233
    "topk",
3234
    "topk_out",
3235
    "triangular_solve",
3236
    "triangular_solve_out",
3237
    # torch.serialization
3238
    "LoadEndianness",
3239
    "SourceChangeWarning",
3240
    # torch.sparse.semi_structured
3241
    "SparseSemiStructuredTensor",
3242
    # torch.storage
3243
    "UntypedStorage",
3244
    # torch.torch_version
3245
    "TorchVersion",
3246
    # torch.types
3247
    "SymInt",
3248
    # torch.utils.benchmark.examples.blas_compare_setup
3249
    "SubEnvSpec",
3250
    # torch.utils.benchmark.examples.compare
3251
    "FauxTorch",
3252
    # torch.utils.benchmark.examples.spectral_ops_fuzz_test
3253
    "Benchmark",
3254
    # torch.utils.benchmark.op_fuzzers.binary
3255
    "BinaryOpFuzzer",
3256
    # torch.utils.benchmark.op_fuzzers.sparse_binary
3257
    "BinaryOpSparseFuzzer",
3258
    # torch.utils.benchmark.op_fuzzers.sparse_unary
3259
    "UnaryOpSparseFuzzer",
3260
    # torch.utils.benchmark.op_fuzzers.spectral
3261
    "SpectralOpFuzzer",
3262
    # torch.utils.benchmark.op_fuzzers.unary
3263
    "UnaryOpFuzzer",
3264
    # torch.utils.benchmark.utils.common
3265
    "Measurement",
3266
    "TaskSpec",
3267
    # torch.utils.benchmark.utils.compare
3268
    "Colorize",
3269
    "Compare",
3270
    "Table",
3271
    # torch.utils.benchmark.utils.fuzzer
3272
    "FuzzedParameter",
3273
    "FuzzedTensor",
3274
    "Fuzzer",
3275
    "ParameterAlias",
3276
    # torch.utils.benchmark.utils.sparse_fuzzer
3277
    "FuzzedSparseTensor",
3278
    # torch.utils.benchmark.utils.timer
3279
    "CPPTimer",
3280
    "Language",
3281
    "Timer",
3282
    # torch.utils.benchmark.utils.valgrind_wrapper.timer_interface
3283
    "CallgrindStats",
3284
    "CopyIfCallgrind",
3285
    "FunctionCount",
3286
    "FunctionCounts",
3287
    "GlobalsBridge",
3288
    "Serialization",
3289
    # torch.utils.bundled_inputs
3290
    "InflatableArg",
3291
    # torch.utils.checkpoint
3292
    "CheckpointError",
3293
    "CheckpointFunction",
3294
    "DefaultDeviceType",
3295
    # torch.utils.collect_env
3296
    "SystemEnv",
3297
    # torch.utils.cpp_extension
3298
    "BuildExtension",
3299
    # torch.utils.data.dataloader
3300
    "DataLoader",
3301
    # torch.utils.data.datapipes.dataframe.dataframe_wrapper
3302
    "PandasWrapper",
3303
    "default_wrapper",
3304
    # torch.utils.data.datapipes.dataframe.dataframes
3305
    "Capture",
3306
    "CaptureA",
3307
    "CaptureAdd",
3308
    "CaptureCall",
3309
    "CaptureControl",
3310
    "CaptureDataFrame",
3311
    "CaptureDataFrameWithDataPipeOps",
3312
    "CaptureF",
3313
    "CaptureGetAttr",
3314
    "CaptureGetItem",
3315
    "CaptureInitial",
3316
    "CaptureLikeMock",
3317
    "CaptureMul",
3318
    "CaptureSetItem",
3319
    "CaptureSub",
3320
    "CaptureVariable",
3321
    "CaptureVariableAssign",
3322
    "DataFrameTracedOps",
3323
    "DataFrameTracer",
3324
    # torch.utils.data.datapipes.dataframe.datapipes
3325
    "ConcatDataFramesPipe",
3326
    "DataFramesAsTuplesPipe",
3327
    "ExampleAggregateAsDataFrames",
3328
    "FilterDataFramesPipe",
3329
    "PerRowDataFramesPipe",
3330
    "ShuffleDataFramesPipe",
3331
    # torch.utils.data.datapipes.dataframe.structures
3332
    "DataChunkDF",
3333
    # torch.utils.data.datapipes.datapipe
3334
    "DFIterDataPipe",
3335
    "DataChunk",
3336
    "IterDataPipe",
3337
    "MapDataPipe",
3338
    # torch.utils.data.datapipes.iter.callable
3339
    "CollatorIterDataPipe",
3340
    "MapperIterDataPipe",
3341
    # torch.utils.data.datapipes.iter.combinatorics
3342
    "SamplerIterDataPipe",
3343
    "ShufflerIterDataPipe",
3344
    # torch.utils.data.datapipes.iter.combining
3345
    "ConcaterIterDataPipe",
3346
    "DemultiplexerIterDataPipe",
3347
    "ForkerIterDataPipe",
3348
    "MultiplexerIterDataPipe",
3349
    "ZipperIterDataPipe",
3350
    # torch.utils.data.datapipes.iter.filelister
3351
    "FileListerIterDataPipe",
3352
    # torch.utils.data.datapipes.iter.fileopener
3353
    "FileOpenerIterDataPipe",
3354
    # torch.utils.data.datapipes.iter.grouping
3355
    "BatcherIterDataPipe",
3356
    "GrouperIterDataPipe",
3357
    "UnBatcherIterDataPipe",
3358
    # torch.utils.data.datapipes.iter.routeddecoder
3359
    "RoutedDecoderIterDataPipe",
3360
    # torch.utils.data.datapipes.iter.selecting
3361
    "FilterIterDataPipe",
3362
    # torch.utils.data.datapipes.iter.sharding
3363
    "SHARDING_PRIORITIES",
3364
    "ShardingFilterIterDataPipe",
3365
    # torch.utils.data.datapipes.iter.utils
3366
    "IterableWrapperIterDataPipe",
3367
    # torch.utils.data.datapipes.map.callable
3368
    "MapperMapDataPipe",
3369
    # torch.utils.data.datapipes.map.combinatorics
3370
    "ShufflerIterDataPipe",
3371
    # torch.utils.data.datapipes.map.combining
3372
    "ConcaterMapDataPipe",
3373
    "ZipperMapDataPipe",
3374
    # torch.utils.data.datapipes.map.grouping
3375
    "BatcherMapDataPipe",
3376
    # torch.utils.data.datapipes.map.utils
3377
    "SequenceWrapperMapDataPipe",
3378
    # torch.utils.data.datapipes.utils.decoder
3379
    "Decoder",
3380
    "ImageHandler",
3381
    "MatHandler",
3382
    # torch.utils.data.dataset
3383
    "ConcatDataset",
3384
    # torch.utils.data.distributed
3385
    "DistributedSampler",
3386
    # torch.utils.dlpack
3387
    "DLDeviceType",
3388
    # torch.utils.file_baton
3389
    "FileBaton",
3390
    # torch.utils.flop_counter
3391
    "FlopCounterMode",
3392
    # torch.utils.hipify.hipify_python
3393
    "CurrentState",
3394
    "GeneratedFileCleaner",
3395
    "HipifyResult",
3396
    "InputError",
3397
    "Trie",
3398
    "bcolors",
3399
    # torch.utils.hooks
3400
    "BackwardHook",
3401
    "RemovableHandle",
3402
    # torch.utils.mkldnn
3403
    "MkldnnBatchNorm",
3404
    "MkldnnConv1d",
3405
    "MkldnnConv2d",
3406
    "MkldnnConv3d",
3407
    "MkldnnLinear",
3408
    "MkldnnPrelu",
3409
    # torch.utils.mobile_optimizer
3410
    "LintCode",
3411
    # torch.utils.show_pickle
3412
    "DumpUnpickler",
3413
    "FakeClass",
3414
    "FakeObject",
3415
    # torch.utils.tensorboard.writer
3416
    "FileWriter",
3417
    "SummaryWriter",
3418
    # torch.utils.throughput_benchmark
3419
    "ExecutionStats",
3420
    # torch.utils.weak
3421
    "WeakIdKeyDictionary",
3422
    "WeakIdRef",
3423
    "WeakTensorKeyDictionary",
3424
]
3425

3426
# The suffix(es) of source filenames.
3427
# You can specify multiple suffix as a list of string:
3428
#
3429
# source_suffix = ['.rst', '.md']
3430
source_suffix = ".rst"
3431

3432
# The master toctree document.
3433
master_doc = "index"
3434

3435
# General information about the project.
3436
project = "PyTorch"
3437
copyright = "2023, PyTorch Contributors"
3438
author = "PyTorch Contributors"
3439
torch_version = str(torch.__version__)
3440

3441
# The version info for the project you're documenting, acts as replacement for
3442
# |version| and |release|, also used in various other places throughout the
3443
# built documents.
3444
#
3445
# The short X.Y version.
3446
# TODO: change to [:2] at v1.0
3447
version = "main (" + torch_version + " )"
3448
# The full version, including alpha/beta/rc tags.
3449
# TODO: verify this works as expected
3450
release = "main"
3451

3452
# Customized html_title here.
3453
# Default is " ".join(project, release, "documentation") if not set
3454
if RELEASE:
3455
    # Turn 1.11.0aHASH into 1.11
3456
    # Note: the release candidates should no longer have the aHASH suffix, but in any
3457
    # case we wish to leave only major.minor, even for rc builds.
3458
    version = ".".join(torch_version.split(".")[:2])
3459
    html_title = " ".join((project, version, "documentation"))
3460
    release = version
3461

3462
# The language for content autogenerated by Sphinx. Refer to documentation
3463
# for a list of supported languages.
3464
#
3465
# This is also used if you do content translation via gettext catalogs.
3466
# Usually you set "language" from the command line for these cases.
3467
language = "en"
3468

3469
# List of patterns, relative to source directory, that match files and
3470
# directories to ignore when looking for source files.
3471
# This patterns also effect to html_static_path and html_extra_path
3472
exclude_patterns = []
3473

3474
# The name of the Pygments (syntax highlighting) style to use.
3475
pygments_style = "sphinx"
3476

3477
# If true, `todo` and `todoList` produce output, else they produce nothing.
3478
todo_include_todos = True
3479

3480
# Disable docstring inheritance
3481
autodoc_inherit_docstrings = False
3482

3483
# Show type hints in the description
3484
autodoc_typehints = "description"
3485

3486
# Add parameter types if the parameter is documented in the docstring
3487
autodoc_typehints_description_target = "documented_params"
3488

3489
# Type aliases for common types
3490
# Sphinx type aliases only works with Postponed Evaluation of Annotations
3491
# (PEP 563) enabled (via `from __future__ import annotations`), which keeps the
3492
# type annotations in string form instead of resolving them to actual types.
3493
# However, PEP 563 does not work well with JIT, which uses the type information
3494
# to generate the code. Therefore, the following dict does not have any effect
3495
# until PEP 563 is supported by JIT and enabled in files.
3496
autodoc_type_aliases = {
3497
    "_size_1_t": "int or tuple[int]",
3498
    "_size_2_t": "int or tuple[int, int]",
3499
    "_size_3_t": "int or tuple[int, int, int]",
3500
    "_size_4_t": "int or tuple[int, int, int, int]",
3501
    "_size_5_t": "int or tuple[int, int, int, int, int]",
3502
    "_size_6_t": "int or tuple[int, int, int, int, int, int]",
3503
    "_size_any_opt_t": "int or None or tuple",
3504
    "_size_2_opt_t": "int or None or 2-tuple",
3505
    "_size_3_opt_t": "int or None or 3-tuple",
3506
    "_ratio_2_t": "float or tuple[float, float]",
3507
    "_ratio_3_t": "float or tuple[float, float, float]",
3508
    "_ratio_any_t": "float or tuple",
3509
    "_tensor_list_t": "Tensor or tuple[Tensor]",
3510
}
3511

3512
# Enable overriding of function signatures in the first line of the docstring.
3513
autodoc_docstring_signature = True
3514

3515
# -- katex javascript in header
3516
#
3517
#    def setup(app):
3518
#    app.add_javascript("https://cdn.jsdelivr.net/npm/katex@0.10.0-beta/dist/katex.min.js")
3519

3520

3521
# -- Options for HTML output ----------------------------------------------
3522
#
3523
# The theme to use for HTML and HTML Help pages.  See the documentation for
3524
# a list of builtin themes.
3525
#
3526
#
3527
#
3528

3529
html_theme = "pytorch_sphinx_theme"
3530
html_theme_path = [pytorch_sphinx_theme.get_html_theme_path()]
3531

3532
# Theme options are theme-specific and customize the look and feel of a theme
3533
# further.  For a list of options available for each theme, see the
3534
# documentation.
3535

3536
html_theme_options = {
3537
    "pytorch_project": "docs",
3538
    "canonical_url": "https://pytorch.org/docs/stable/",
3539
    "collapse_navigation": False,
3540
    "display_version": True,
3541
    "logo_only": True,
3542
    "analytics_id": "GTM-T8XT4PS",
3543
}
3544

3545
html_logo = "_static/img/pytorch-logo-dark-unstable.png"
3546
if RELEASE:
3547
    html_logo = "_static/img/pytorch-logo-dark.svg"
3548

3549

3550
# Add any paths that contain custom static files (such as style sheets) here,
3551
# relative to this directory. They are copied after the builtin static files,
3552
# so a file named "default.css" will overwrite the builtin "default.css".
3553
html_static_path = ["_static"]
3554

3555
html_css_files = [
3556
    "css/jit.css",
3557
]
3558

3559
from sphinx.ext.coverage import CoverageBuilder
3560

3561

3562
def coverage_post_process(app, exception):
3563
    if exception is not None:
3564
        return
3565

3566
    # Only run this test for the coverage build
3567
    if not isinstance(app.builder, CoverageBuilder):
3568
        return
3569

3570
    if not torch.distributed.is_available():
3571
        raise RuntimeError(
3572
            "The coverage tool cannot run with a version "
3573
            "of PyTorch that was built with USE_DISTRIBUTED=0 "
3574
            "as this module's API changes."
3575
        )
3576

3577
    # These are all the modules that have "automodule" in an rst file
3578
    # These modules are the ones for which coverage is checked
3579
    # Here, we make sure that no module is missing from that list
3580
    modules = app.env.domaindata["py"]["modules"]
3581

3582
    # We go through all the torch submodules and make sure they are
3583
    # properly tested
3584
    missing = set()
3585

3586
    def is_not_internal(modname):
3587
        split_name = modname.split(".")
3588
        for name in split_name:
3589
            if name[0] == "_":
3590
                return False
3591
        return True
3592

3593
    # The walk function does not return the top module
3594
    if "torch" not in modules:
3595
        missing.add("torch")
3596

3597
    for _, modname, ispkg in pkgutil.walk_packages(
3598
        path=torch.__path__, prefix=torch.__name__ + "."
3599
    ):
3600
        if is_not_internal(modname):
3601
            if modname not in modules:
3602
                missing.add(modname)
3603

3604
    output = []
3605

3606
    if missing:
3607
        mods = ", ".join(missing)
3608
        output.append(
3609
            f"\nYou added the following module(s) to the PyTorch namespace '{mods}' "
3610
            "but they have no corresponding entry in a doc .rst file. You should "
3611
            "either make sure that the .rst file that contains the module's documentation "
3612
            "properly contains either '.. automodule:: mod_name' (if you do not want "
3613
            "the paragraph added by the automodule, you can simply use '.. py:module:: mod_name') "
3614
            " or make the module private (by appending an '_' at the beginning of its name)."
3615
        )
3616

3617
    # The output file is hard-coded by the coverage tool
3618
    # Our CI is setup to fail if any line is added to this file
3619
    output_file = path.join(app.outdir, "python.txt")
3620

3621
    if output:
3622
        with open(output_file, "a") as f:
3623
            for o in output:
3624
                f.write(o)
3625

3626

3627
def process_docstring(app, what_, name, obj, options, lines):
3628
    """
3629
    Custom process to transform docstring lines Remove "Ignore" blocks
3630

3631
    Args:
3632
        app (sphinx.application.Sphinx): the Sphinx application object
3633

3634
        what (str):
3635
            the type of the object which the docstring belongs to (one of
3636
            "module", "class", "exception", "function", "method", "attribute")
3637

3638
        name (str): the fully qualified name of the object
3639

3640
        obj: the object itself
3641

3642
        options: the options given to the directive: an object with
3643
            attributes inherited_members, undoc_members, show_inheritance
3644
            and noindex that are true if the flag option of same name was
3645
            given to the auto directive
3646

3647
        lines (List[str]): the lines of the docstring, see above
3648

3649
    References:
3650
        https://www.sphinx-doc.org/en/1.5.1/_modules/sphinx/ext/autodoc.html
3651
        https://www.sphinx-doc.org/en/master/usage/extensions/autodoc.html
3652
    """
3653
    import re
3654

3655
    remove_directives = [
3656
        # Remove all xdoctest directives
3657
        re.compile(r"\s*>>>\s*#\s*x?doctest:\s*.*"),
3658
        re.compile(r"\s*>>>\s*#\s*x?doc:\s*.*"),
3659
    ]
3660
    filtered_lines = [
3661
        line for line in lines if not any(pat.match(line) for pat in remove_directives)
3662
    ]
3663
    # Modify the lines inplace
3664
    lines[:] = filtered_lines
3665

3666
    # make sure there is a blank line at the end
3667
    if lines and lines[-1].strip():
3668
        lines.append("")
3669

3670

3671
# Called automatically by Sphinx, making this `conf.py` an "extension".
3672
def setup(app):
3673
    # NOTE: in Sphinx 1.8+ `html_css_files` is an official configuration value
3674
    # and can be moved outside of this function (and the setup(app) function
3675
    # can be deleted).
3676
    html_css_files = [
3677
        "https://cdn.jsdelivr.net/npm/katex@0.10.0-beta/dist/katex.min.css"
3678
    ]
3679

3680
    # In Sphinx 1.8 it was renamed to `add_css_file`, 1.7 and prior it is
3681
    # `add_stylesheet` (deprecated in 1.8).
3682
    add_css = getattr(app, "add_css_file", app.add_stylesheet)
3683
    for css_file in html_css_files:
3684
        add_css(css_file)
3685

3686
    app.connect("build-finished", coverage_post_process)
3687
    app.connect("autodoc-process-docstring", process_docstring)
3688

3689

3690
# From PyTorch 1.5, we now use autogenerated files to document classes and
3691
# functions. This breaks older references since
3692
# https://pytorch.org/docs/stable/torch.html#torch.flip
3693
# moved to
3694
# https://pytorch.org/docs/stable/generated/torch.flip.html
3695
# which breaks older links from blog posts, stack overflow answers and more.
3696
# To mitigate that, we add an id="torch.flip" in an appropriated place
3697
# in torch.html by overriding the visit_reference method of html writers.
3698
# Someday this can be removed, once the old links fade away
3699

3700
from sphinx.writers import html, html5
3701

3702

3703
def replace(Klass):
3704
    old_call = Klass.visit_reference
3705

3706
    def visit_reference(self, node):
3707
        if "refuri" in node and "generated" in node.get("refuri"):
3708
            ref = node.get("refuri")
3709
            ref_anchor = ref.split("#")
3710
            if len(ref_anchor) > 1:
3711
                # Only add the id if the node href and the text match,
3712
                # i.e. the href is "torch.flip#torch.flip" and the content is
3713
                # "torch.flip" or "flip" since that is a signal the node refers
3714
                # to autogenerated content
3715
                anchor = ref_anchor[1]
3716
                txt = node.parent.astext()
3717
                if txt == anchor or txt == anchor.split(".")[-1]:
3718
                    self.body.append(f'<p id="{ref_anchor[1]}"/>')
3719
        return old_call(self, node)
3720

3721
    Klass.visit_reference = visit_reference
3722

3723

3724
replace(html.HTMLTranslator)
3725
replace(html5.HTML5Translator)
3726

3727
# -- Options for HTMLHelp output ------------------------------------------
3728

3729
# Output file base name for HTML help builder.
3730
htmlhelp_basename = "PyTorchdoc"
3731

3732

3733
# -- Options for LaTeX output ---------------------------------------------
3734

3735
latex_elements = {
3736
    # The paper size ('letterpaper' or 'a4paper').
3737
    #
3738
    # 'papersize': 'letterpaper',
3739
    # The font size ('10pt', '11pt' or '12pt').
3740
    #
3741
    # 'pointsize': '10pt',
3742
    # Additional stuff for the LaTeX preamble.
3743
    #
3744
    # 'preamble': '',
3745
    # Latex figure (float) alignment
3746
    #
3747
    # 'figure_align': 'htbp',
3748
}
3749

3750
# Grouping the document tree into LaTeX files. List of tuples
3751
# (source start file, target name, title,
3752
#  author, documentclass [howto, manual, or own class]).
3753
latex_documents = [
3754
    (
3755
        master_doc,
3756
        "pytorch.tex",
3757
        "PyTorch Documentation",
3758
        "Torch Contributors",
3759
        "manual",
3760
    ),
3761
]
3762

3763

3764
# -- Options for manual page output ---------------------------------------
3765

3766
# One entry per manual page. List of tuples
3767
# (source start file, name, description, authors, manual section).
3768
man_pages = [(master_doc, "PyTorch", "PyTorch Documentation", [author], 1)]
3769

3770

3771
# -- Options for Texinfo output -------------------------------------------
3772

3773
# Grouping the document tree into Texinfo files. List of tuples
3774
# (source start file, target name, title, author,
3775
#  dir menu entry, description, category)
3776
texinfo_documents = [
3777
    (
3778
        master_doc,
3779
        "PyTorch",
3780
        "PyTorch Documentation",
3781
        author,
3782
        "PyTorch",
3783
        "One line description of project.",
3784
        "Miscellaneous",
3785
    ),
3786
]
3787

3788

3789
# Example configuration for intersphinx: refer to the Python standard library.
3790
intersphinx_mapping = {
3791
    "python": ("https://docs.python.org/3", None),
3792
    "numpy": ("https://numpy.org/doc/stable", None),
3793
}
3794

3795
import sphinx.ext.doctest
3796

3797
# -- A patch that prevents Sphinx from cross-referencing ivar tags -------
3798
# See http://stackoverflow.com/a/41184353/3343043
3799

3800
from docutils import nodes
3801
from sphinx import addnodes
3802
from sphinx.util.docfields import TypedField
3803

3804
# Without this, doctest adds any example with a `>>>` as a test
3805
doctest_test_doctest_blocks = ""
3806
doctest_default_flags = sphinx.ext.doctest.doctest.ELLIPSIS
3807
doctest_global_setup = """
3808
import torch
3809
try:
3810
    import torchvision
3811
except ImportError:
3812
    torchvision = None
3813
"""
3814

3815

3816
def patched_make_field(self, types, domain, items, **kw):
3817
    # `kw` catches `env=None` needed for newer sphinx while maintaining
3818
    #  backwards compatibility when passed along further down!
3819

3820
    # type: (List, unicode, Tuple) -> nodes.field
3821
    def handle_item(fieldarg, content):
3822
        par = nodes.paragraph()
3823
        par += addnodes.literal_strong("", fieldarg)  # Patch: this line added
3824
        # par.extend(self.make_xrefs(self.rolename, domain, fieldarg,
3825
        #                           addnodes.literal_strong))
3826
        if fieldarg in types:
3827
            par += nodes.Text(" (")
3828
            # NOTE: using .pop() here to prevent a single type node to be
3829
            # inserted twice into the doctree, which leads to
3830
            # inconsistencies later when references are resolved
3831
            fieldtype = types.pop(fieldarg)
3832
            if len(fieldtype) == 1 and isinstance(fieldtype[0], nodes.Text):
3833
                typename = fieldtype[0].astext()
3834
                builtin_types = ["int", "long", "float", "bool", "type"]
3835
                for builtin_type in builtin_types:
3836
                    pattern = rf"(?<![\w.]){builtin_type}(?![\w.])"
3837
                    repl = f"python:{builtin_type}"
3838
                    typename = re.sub(pattern, repl, typename)
3839
                par.extend(
3840
                    self.make_xrefs(
3841
                        self.typerolename,
3842
                        domain,
3843
                        typename,
3844
                        addnodes.literal_emphasis,
3845
                        **kw,
3846
                    )
3847
                )
3848
            else:
3849
                par += fieldtype
3850
            par += nodes.Text(")")
3851
        par += nodes.Text(" -- ")
3852
        par += content
3853
        return par
3854

3855
    fieldname = nodes.field_name("", self.label)
3856
    if len(items) == 1 and self.can_collapse:
3857
        fieldarg, content = items[0]
3858
        bodynode = handle_item(fieldarg, content)
3859
    else:
3860
        bodynode = self.list_type()
3861
        for fieldarg, content in items:
3862
            bodynode += nodes.list_item("", handle_item(fieldarg, content))
3863
    fieldbody = nodes.field_body("", bodynode)
3864
    return nodes.field("", fieldname, fieldbody)
3865

3866

3867
TypedField.make_field = patched_make_field
3868

3869
copybutton_prompt_text = r">>> |\.\.\. "
3870
copybutton_prompt_is_regexp = True
3871

Использование cookies

Мы используем файлы cookie в соответствии с Политикой конфиденциальности и Политикой использования cookies.

Нажимая кнопку «Принимаю», Вы даете АО «СберТех» согласие на обработку Ваших персональных данных в целях совершенствования нашего веб-сайта и Сервиса GitVerse, а также повышения удобства их использования.

Запретить использование cookies Вы можете самостоятельно в настройках Вашего браузера.