-
Notifications
You must be signed in to change notification settings - Fork 2.2k
/
Copy pathprediction_service_pb2.py
129 lines (111 loc) · 5.63 KB
/
prediction_service_pb2.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
# Copyright 2016 Google Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ==============================================================================
#
# To regenerate run
# python -m grpc.tools.protoc --python_out=. --grpc_python_out=. -I. tensorflow_serving/apis/prediction_service.proto
### @@AUTOGENERATED SECTION STARTS HERE@@
# Generated by the protocol buffer compiler. DO NOT EDIT!
# source: tensorflow_serving/apis/prediction_service.proto
import sys
_b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1'))
from google.protobuf import descriptor as _descriptor
from google.protobuf import descriptor_pb2
from google.protobuf import symbol_database as _symbol_database
# @@protoc_insertion_point(imports)
_sym_db = _symbol_database.Default()
from tensorflow_serving.apis import (
classification_pb2 as tensorflow__serving_dot_apis_dot_classification__pb2,
)
from tensorflow_serving.apis import (
get_model_metadata_pb2 as tensorflow__serving_dot_apis_dot_get__model__metadata__pb2,
)
from tensorflow_serving.apis import (
inference_pb2 as tensorflow__serving_dot_apis_dot_inference__pb2,
)
from tensorflow_serving.apis import (
predict_pb2 as tensorflow__serving_dot_apis_dot_predict__pb2,
)
from tensorflow_serving.apis import (
regression_pb2 as tensorflow__serving_dot_apis_dot_regression__pb2,
)
DESCRIPTOR = _descriptor.FileDescriptor(
name='tensorflow_serving/apis/prediction_service.proto',
package='tensorflow.serving',
syntax='proto3',
serialized_pb=_b('\n0tensorflow_serving/apis/prediction_service.proto\x12\x12tensorflow.serving\x1a,tensorflow_serving/apis/classification.proto\x1a\x30tensorflow_serving/apis/get_model_metadata.proto\x1a\'tensorflow_serving/apis/inference.proto\x1a%tensorflow_serving/apis/predict.proto\x1a(tensorflow_serving/apis/regression.proto2\xfc\x03\n\x11PredictionService\x12\x61\n\x08\x43lassify\x12).tensorflow.serving.ClassificationRequest\x1a*.tensorflow.serving.ClassificationResponse\x12X\n\x07Regress\x12%.tensorflow.serving.RegressionRequest\x1a&.tensorflow.serving.RegressionResponse\x12R\n\x07Predict\x12\".tensorflow.serving.PredictRequest\x1a#.tensorflow.serving.PredictResponse\x12g\n\x0eMultiInference\x12).tensorflow.serving.MultiInferenceRequest\x1a*.tensorflow.serving.MultiInferenceResponse\x12m\n\x10GetModelMetadata\x12+.tensorflow.serving.GetModelMetadataRequest\x1a,.tensorflow.serving.GetModelMetadataResponseB\x03\xf8\x01\x01\x62\x06proto3')
,
dependencies=[tensorflow__serving_dot_apis_dot_classification__pb2.DESCRIPTOR,tensorflow__serving_dot_apis_dot_get__model__metadata__pb2.DESCRIPTOR,tensorflow__serving_dot_apis_dot_inference__pb2.DESCRIPTOR,tensorflow__serving_dot_apis_dot_predict__pb2.DESCRIPTOR,tensorflow__serving_dot_apis_dot_regression__pb2.DESCRIPTOR,])
_sym_db.RegisterFileDescriptor(DESCRIPTOR)
DESCRIPTOR.has_options = True
DESCRIPTOR._options = _descriptor._ParseOptions(descriptor_pb2.FileOptions(), _b('\370\001\001'))
_PREDICTIONSERVICE = _descriptor.ServiceDescriptor(
name='PredictionService',
full_name='tensorflow.serving.PredictionService',
file=DESCRIPTOR,
index=0,
options=None,
serialized_start=291,
serialized_end=799,
methods=[
_descriptor.MethodDescriptor(
name='Classify',
full_name='tensorflow.serving.PredictionService.Classify',
index=0,
containing_service=None,
input_type=tensorflow__serving_dot_apis_dot_classification__pb2._CLASSIFICATIONREQUEST,
output_type=tensorflow__serving_dot_apis_dot_classification__pb2._CLASSIFICATIONRESPONSE,
options=None,
),
_descriptor.MethodDescriptor(
name='Regress',
full_name='tensorflow.serving.PredictionService.Regress',
index=1,
containing_service=None,
input_type=tensorflow__serving_dot_apis_dot_regression__pb2._REGRESSIONREQUEST,
output_type=tensorflow__serving_dot_apis_dot_regression__pb2._REGRESSIONRESPONSE,
options=None,
),
_descriptor.MethodDescriptor(
name='Predict',
full_name='tensorflow.serving.PredictionService.Predict',
index=2,
containing_service=None,
input_type=tensorflow__serving_dot_apis_dot_predict__pb2._PREDICTREQUEST,
output_type=tensorflow__serving_dot_apis_dot_predict__pb2._PREDICTRESPONSE,
options=None,
),
_descriptor.MethodDescriptor(
name='MultiInference',
full_name='tensorflow.serving.PredictionService.MultiInference',
index=3,
containing_service=None,
input_type=tensorflow__serving_dot_apis_dot_inference__pb2._MULTIINFERENCEREQUEST,
output_type=tensorflow__serving_dot_apis_dot_inference__pb2._MULTIINFERENCERESPONSE,
options=None,
),
_descriptor.MethodDescriptor(
name='GetModelMetadata',
full_name='tensorflow.serving.PredictionService.GetModelMetadata',
index=4,
containing_service=None,
input_type=tensorflow__serving_dot_apis_dot_get__model__metadata__pb2._GETMODELMETADATAREQUEST,
output_type=tensorflow__serving_dot_apis_dot_get__model__metadata__pb2._GETMODELMETADATARESPONSE,
options=None,
),
])
_sym_db.RegisterServiceDescriptor(_PREDICTIONSERVICE)
DESCRIPTOR.services_by_name['PredictionService'] = _PREDICTIONSERVICE
# @@protoc_insertion_point(module_scope)
### @@AUTOGENERATED SECTION ENDS HERE@@