/aosp_15_r20/external/pytorch/torch/ao/nn/quantizable/modules/ |
H A D | rnn.py | 202 bidirectional: bool = False, 209 self.bidirectional = bidirectional 213 if self.bidirectional: 226 if self.bidirectional: 247 if hasattr(self, "layer_bw") and self.bidirectional: 285 bidirectional = kwargs.get("bidirectional", other.bidirectional) 287 layer = cls(input_size, hidden_size, bias, batch_first, bidirectional) 296 if other.bidirectional: 342 bidirectional: bool = False, 354 self.bidirectional = bidirectional [all …]
|
/aosp_15_r20/external/pytorch/torch/nn/modules/ |
H A D | rnn.py | 81 bidirectional: bool 93 bidirectional: bool = False, 107 self.bidirectional = bidirectional 110 num_directions = 2 if bidirectional else 1 278 bool(self.bidirectional), 323 num_directions = 2 if self.bidirectional else 1 383 if self.bidirectional is not False: 412 num_directions = 2 if self.bidirectional else 1 606 bidirectional: bool = False, 653 num_directions = 2 if self.bidirectional else 1 [all …]
|
/aosp_15_r20/external/pytorch/torch/csrc/api/src/nn/modules/ |
H A D | rnn.cpp | 61 const int64_t num_directions = options_base.bidirectional() ? 2 : 1; in reset() 236 options_base.bidirectional()); in flatten_parameters() 316 int64_t num_directions = options_base.bidirectional() ? 2 : 1; in get_expected_hidden_size() 375 << ", bidirectional=" << options_base.bidirectional(); in pretty_print() 425 .bidirectional(options_.bidirectional())), in RNNImpl() 435 int64_t num_directions = options_base.bidirectional() ? 2 : 1; in forward_helper() 460 options_base.bidirectional(), in forward_helper() 472 options_base.bidirectional(), in forward_helper() 491 options_base.bidirectional()); in forward_helper() 503 options_base.bidirectional()); in forward_helper() [all …]
|
/aosp_15_r20/external/pytorch/benchmarks/functional_autograd_benchmark/ |
H A D | torchaudio_models.py | 186 bidirectional=False, argument 192 self.bidirectional = bidirectional 199 bidirectional=bidirectional, 202 self.num_directions = 2 if bidirectional else 1 213 if self.bidirectional: 269 bidirectional, argument 279 self.bidirectional = bidirectional 306 bidirectional=bidirectional, 315 bidirectional=bidirectional, 325 if not bidirectional [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mkldnn/ |
H A D | RNN.cpp | 41 bool bidirectional, in mkldnn_rnn_layer() argument 67 bool bidirectional, in mkldnn_rnn_layer_backward() argument 105 bool bidirectional, in RNNParams() 114 num_directions = bidirectional ? 2 : 1; in RNNParams() 229 bool bidirectional, in mkldnn_rnn_layer() argument 238 bidirectional, in mkldnn_rnn_layer() 314 bool bidirectional, in mkldnn_rnn_layer_backward() argument 333 bidirectional, in mkldnn_rnn_layer_backward() 459 bool train, bool bidirectional, IntArrayRef batch_sizes) { in mkldnn_rnn() argument 476 auto num_directions = bidirectional ? 2 : 1; in mkldnn_rnn() [all …]
|
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/dynamic/modules/ |
H A D | rnn.py | 105 bidirectional=False, argument 117 self.bidirectional = bidirectional 121 num_directions = 2 if bidirectional else 1 205 if self.bidirectional is not False: 257 num_directions = 2 if self.bidirectional else 1 316 num_directions = 2 if self.bidirectional else 1 387 mod.bidirectional, 398 mod.bidirectional, 406 num_directions = 2 if mod.bidirectional else 1 474 num_directions = 2 if self.bidirectional else 1 [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/ |
H A D | RNN.cpp | 1112 int64_t num_layers, double dropout_p, bool train, bool bidirectional) { in _rnn_impl() argument 1115 if (bidirectional) { in _rnn_impl() 1129 int64_t num_layers, double dropout_p, bool train, bool bidirectional) { in _rnn_impl_with_concat() argument 1130 …ellType, LayerT, BidirLayerT>(input, params, hiddens, num_layers, dropout_p, train, bidirectional); in _rnn_impl_with_concat() 1138 int64_t num_layers, double dropout_p, bool train, bool bidirectional) { in _lstm_impl() argument 1150 …params>, LayerT, BidirLayerT>(input, params, hiddens, num_layers, dropout_p, train, bidirectional); in _lstm_impl() 1203 bool bidirectional, \ 1218 bidirectional, \ 1235 bidirectional, \ 1250 bidirectional); \ [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/cudnn/ |
H A D | RNN.cpp | 88 bool bidirectional, in _cudnn_rnn_backward() argument 156 cudnnDirectionMode_t bidirectional; member 164 return bidirectional ? 2 : 1; in num_directions() 190 bidirectional = in set_bidirectional() 204 bool bidirectional, in set() 215 bool bidirectional, in set() 227 this->set_bidirectional(bidirectional); in set() 243 bidirectional, in set() 259 bidirectional, in set() 1248 bool bidirectional, in copy_weights_to_flat_buf_views() argument [all …]
|
/aosp_15_r20/external/autotest/server/cros/network/ |
H A D | iperf_runner.py | 74 if config.bidirectional: 235 self.bidirectional = False 238 self.bidirectional = False 241 self.bidirectional = True 244 self.bidirectional = False 247 self.bidirectional = False 250 self.bidirectional = True 330 self._bidirectional_flag = '-d' if config.bidirectional else ''
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/miopen/ |
H A D | RNN_miopen.cpp | 43 …double dropout, bool train, bool bidirectional, IntArrayRef batch_sizes, const std::optional<Tenso… in miopen_rnn_backward() argument 117 …void set(int64_t mode, int64_t hidden_size, int64_t num_layers, bool bidirectional, miopenDataType… in set() 121 this->set_bidirectional(bidirectional); in set() 763 …double dropout, bool train, bool bidirectional, IntArrayRef batch_sizes, const std::optional<Tenso… in miopen_rnn_backward() argument 781 … grad_cy, mode, hidden_size, num_layers, batch_first, dropout, train, bidirectional, batch_sizes, … in miopen_rnn_backward() 784 …, output, mode, hidden_size, num_layers, batch_first, dropout, train, bidirectional, batch_sizes, … in miopen_rnn_backward() 825 int64_t num_layers, double dropout_p, bool train, bool bidirectional) { in _miopen_impl() argument 837 dropout_p, train, bidirectional, batch_sizes, dropout_state); in _miopen_impl() 847 int64_t num_layers, double dropout_p, bool train, bool bidirectional, bool batch_first) { in _miopen_impl() argument 856 train, bidirectional, /*batch_sizes=*/{}, dropout_state); in _miopen_impl() [all …]
|
/aosp_15_r20/external/python/cpython2/Tools/pybench/ |
D | Unicode.py | 473 bidirectional = unicodedata.bidirectional 486 bidirectional(c) 495 bidirectional(c) 504 bidirectional(c) 513 bidirectional(c) 522 bidirectional(c) 535 bidirectional = unicodedata.bidirectional
|
/aosp_15_r20/external/pytorch/torch/ao/nn/quantized/reference/modules/ |
H A D | rnn.py | 403 bidirectional: bool = False, 417 bidirectional, 509 num_directions = 2 if self.bidirectional else 1 589 num_directions = 2 if self.bidirectional else 1 640 self.bidirectional, 653 self.bidirectional, 681 mod.bidirectional, 774 num_directions = 2 if self.bidirectional else 1 797 self.bidirectional, 810 self.bidirectional, [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/mps/operations/ |
H A D | RnnOps.mm | 37 bool bidirectional, 43 if (bidirectional) { 93 bool bidirectional, 108 const int64_t total_layers = num_layers * (bidirectional ? 2 : 1); 135 … std::to_string(num_layers) + "_bidirectional_" + std::to_string(bidirectional) + "_has_biases_" + 161 opDesc.bidirectional = bidirectional; 192 bidirectional, 217 if (bidirectional) { 356 bool bidirectional, 373 hidden_size * (bidirectional ? 2 : 1)}, [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/native/vulkan/ops/ |
H A D | Gru.cpp | 50 bool bidirectional, in gru_input() argument 64 !bidirectional, "Vulkan gru expects 'bidirectional' to be false."); in gru_input() 193 bool bidirectional, in GruPackedContext() argument 199 !bidirectional, "Vulkan gru expects 'bidirectional' to be false."); in GruPackedContext() 210 packed_.emplace_back(bidirectional); in GruPackedContext() 268 bool bidirectional, in create_gru_context() argument 276 bidirectional, in create_gru_context()
|
H A D | Lstm.cpp | 59 bool bidirectional, in lstm_input() argument 79 !bidirectional, "Vulkan LSTM expects 'bidirectional' to be false."); in lstm_input() 235 bool bidirectional, in LstmPackedContext() argument 241 !bidirectional, "Vulkan LSTM expects 'bidirectional' to be false."); in LstmPackedContext() 252 packed_.emplace_back(bidirectional); in LstmPackedContext() 310 bool bidirectional, in create_lstm_context() argument 318 bidirectional, in create_lstm_context()
|
/aosp_15_r20/external/skia/third_party/libgrapheme/ |
H A D | BUILD.gn | 62 compile_tool("bidirectional") { 63 sources = [ "../externals/libgrapheme/gen/bidirectional.c" ] 99 "$_src/bidirectional.c", 109 ":bidirectional", 147 "$_src/bidirectional.c",
|
/aosp_15_r20/external/python/cpython2/Lib/test/ |
D | test_unicodedata.py | 96 self.db.bidirectional(char), 149 self.assertEqual(self.db.bidirectional(u'\uFFFE'), '') 150 self.assertEqual(self.db.bidirectional(u' '), 'WS') 151 self.assertEqual(self.db.bidirectional(u'A'), 'L') 152 self.assertEqual(self.db.bidirectional(u'\U00020000'), 'L') 154 self.assertRaises(TypeError, self.db.bidirectional) 155 self.assertRaises(TypeError, self.db.bidirectional, u'xx')
|
/aosp_15_r20/external/pytorch/torch/csrc/api/include/torch/nn/options/ |
H A D | rnn.h | 44 TORCH_ARG(bool, bidirectional) = false; 87 TORCH_ARG(bool, bidirectional) = false; 120 TORCH_ARG(bool, bidirectional) = false; 155 TORCH_ARG(bool, bidirectional) = false;
|
/aosp_15_r20/external/python/cpython3/Lib/test/ |
D | test_unicodedata.py | 90 self.db.bidirectional(char), 154 self.assertEqual(self.db.bidirectional('\uFFFE'), '') 155 self.assertEqual(self.db.bidirectional(' '), 'WS') 156 self.assertEqual(self.db.bidirectional('A'), 'L') 157 self.assertEqual(self.db.bidirectional('\U00020000'), 'L') 159 self.assertRaises(TypeError, self.db.bidirectional) 160 self.assertRaises(TypeError, self.db.bidirectional, 'xx')
|
/aosp_15_r20/external/pytorch/test/quantization/core/ |
H A D | test_quantized_module.py | 1691 bidirectional=st.booleans(), 1694 def test_lstm_api(self, dtype, bidirectional): argument 1707 num_directions = 2 if bidirectional else 1 1723 h = torch.randn(num_layers * (bidirectional + 1), batch, hidden_size) 1724 c = torch.randn(num_layers * (bidirectional + 1), batch, hidden_size) 1731 bidirectional=bidirectional, 1739 bidirectional=bidirectional, 1749 bidirectional, 1781 bidirectional = False 1784 h = torch.rand(num_layers * (bidirectional + 1), batch, hidden_size) [all …]
|
/aosp_15_r20/external/pytorch/aten/src/ATen/cudnn/ |
H A D | AutocastRNN.cpp | 35 bool bidirectional, in _cudnn_rnn_cast_reflatten() argument 91 bidirectional, in _cudnn_rnn_cast_reflatten() 112 bidirectional, in _cudnn_rnn_cast_reflatten()
|
/aosp_15_r20/external/webrtc/pc/ |
H A D | sdp_serializer.cc | 221 bool bidirectional = tokens.size() == 4; // indicates both send and recv in DeserializeSimulcastDescription() local 225 (bidirectional && tokens[2] != kSendDirection && in DeserializeSimulcastDescription() 227 (bidirectional && tokens[0] == tokens[2])) { in DeserializeSimulcastDescription() 238 if (bidirectional) { in DeserializeSimulcastDescription()
|
/aosp_15_r20/external/pytorch/benchmarks/fastrnns/ |
H A D | custom_lstms.py | 42 bidirectional=False, argument 50 if bidirectional: 78 bidirectional=False, argument 88 if bidirectional: 458 rnn = script_lstm(input_size, hidden_size, num_layers, bidirectional=True) 463 lstm = nn.LSTM(input_size, hidden_size, num_layers, bidirectional=True)
|
/aosp_15_r20/external/sdv/vsomeip/third_party/boost/mpl/doc/src/refmanual/ |
D | advance.rst | 25 Moves ``Iterator`` by the distance ``N``. For |bidirectional| and 126 .. |bidirectional| replace:: `bidirectional`_ substdef in See also
|
/aosp_15_r20/external/pytorch/test/cpp/api/ |
H A D | rnn.cpp | 399 GRU gru(GRUOptions(100, 256).num_layers(2).bidirectional(true)); in TEST_F() 453 GRU bi_grus{gru_options.bidirectional(true)}; in BidirectionalGRUReverseForward() 454 GRU reverse_gru{gru_options.bidirectional(false)}; in BidirectionalGRUReverseForward() 508 LSTM bi_lstm{lstm_opt.bidirectional(true)}; in BidirectionalLSTMReverseForwardTest() 509 LSTM reverse_lstm{lstm_opt.bidirectional(false)}; in BidirectionalLSTMReverseForwardTest() 557 GRUOptions(2, 4).num_layers(3).batch_first(false).bidirectional(true); in TEST_F() 612 LSTMOptions(2, 4).num_layers(3).batch_first(false).bidirectional(true); in TEST_F() 668 .bidirectional(true) in TEST_F()
|