/frameworks/ml/nn/runtime/test/generated/models/ |
D | hashtable_lookup_float.model.cpp | 7 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {4}, 1.f, 0); in CreateModel() local
|
D | hashtable_lookup_float_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {4}, 1.f, 0); in CreateModel() local
|
D | hashtable_lookup_quant8.model.cpp | 7 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {4}, 1.f, 0); in CreateModel() local
|
D | fully_connected_quant8_weights_as_inputs.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | lsh_projection_weights_as_inputs_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {8}); in CreateModel() local
|
D | lsh_projection_weights_as_inputs.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {8}); in CreateModel() local
|
D | lsh_projection.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {8}); in CreateModel() local
|
D | lsh_projection_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {8}); in CreateModel() local
|
D | lsh_projection_2_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {4}); in CreateModel() local
|
D | lsh_projection_2.model.cpp | 7 OperandType type4(Type::TENSOR_INT32, {4}); in CreateModel() local
|
D | fully_connected_quant8_2.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_float_3.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_float_4d_simple.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_quant8.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_float_4d_simple_relaxed.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_float_2.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | fully_connected_float_2_relaxed.model.cpp | 3 OperandType type4(Type::INT32, {}); in CreateModel() local
|
D | rnn_state.model.cpp | 7 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); in CreateModel() local
|
D | conv_quant8_overflow_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 3, 3}, 1.0, 0); in CreateModel() local
|
D | rnn_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); in CreateModel() local
|
D | conv_quant8_channels_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 1, 1, 3}, 1.0, 0); in CreateModel() local
|
D | conv_quant8_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 2, 1}, 1.f, 0); in CreateModel() local
|
D | conv_quant8_large_weights_as_inputs.model.cpp | 6 OperandType type4(Type::TENSOR_QUANT8_ASYMM, {1, 2, 3, 3}, 1.0, 0); in CreateModel() local
|
D | rnn.model.cpp | 7 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); in CreateModel() local
|
D | rnn_state_relaxed.model.cpp | 7 OperandType type4(Type::TENSOR_FLOAT32, {2, 16}); in CreateModel() local
|