3030)
3131from datarobot_drum .drum .root_predictors .transform_helpers import (
3232 is_sparse ,
33- make_arrow_payload ,
3433 make_csv_payload ,
3534 make_mtx_payload ,
3635)
@@ -74,7 +73,7 @@ def _fetch_data_from_request(file_key, logger=None):
7473 else :
7574 wrong_key_error_message = (
7675 "Samples should be provided as: "
77- " - a csv, mtx, or arrow file under `{}` form-data param key."
76+ " - a csv or mtx under `{}` form-data param key."
7877 " - binary data" .format (file_key )
7978 )
8079 if logger is not None :
@@ -126,8 +125,6 @@ def _check_mimetype_support(self, mimetype):
126125 )
127126 + "Make DRUM support the format or implement `read_input_data` hook to read the data. "
128127 )
129- if mimetype == PredictionServerMimetypes .APPLICATION_X_APACHE_ARROW_STREAM :
130- error_message += "pyarrow package may be missing, try to install."
131128 return {"message" : error_message }, HTTP_422_UNPROCESSABLE_ENTITY
132129 return None
133130
@@ -193,16 +190,6 @@ def _build_drum_response_json_str(predict_response):
193190 def _transform (self , logger = None ):
194191 response_status = HTTP_200_OK
195192
196- arrow_key = "arrow_version"
197- arrow_version = request .files .get (arrow_key )
198- # TODO: check implementation of how arrow_version is passed
199- # Currently it is passed as a file content,
200- # so arrow_version is of type werkzeug.datastructures.FileStorage,
201- # that's why io.BytesIO getvalue is called on it.
202- if arrow_version is not None :
203- arrow_version = arrow_version .getvalue ().decode ("utf-8" )
204- use_arrow = arrow_version is not None
205-
206193 try :
207194 feature_binary_data , feature_mimetype , feature_charset = self ._fetch_data_from_request (
208195 "X" , logger = logger
@@ -256,32 +243,13 @@ def _transform(self, logger=None):
256243
257244 # make output
258245 if is_sparse (out_data ):
259- if use_arrow :
260- target_payload = (
261- make_arrow_payload (out_target , arrow_version )
262- if out_target is not None
263- else None
264- )
265- target_out_format = "arrow"
266- else :
267- target_payload = make_csv_payload (out_target ) if out_target is not None else None
268- target_out_format = "csv"
246+ target_payload = make_csv_payload (out_target ) if out_target is not None else None
269247 feature_payload , colnames = make_mtx_payload (out_data )
270248 out_format = "sparse"
271249 else :
272- if use_arrow :
273- feature_payload = make_arrow_payload (out_data , arrow_version )
274- target_payload = (
275- make_arrow_payload (out_target , arrow_version )
276- if out_target is not None
277- else None
278- )
279- out_format = "arrow"
280- else :
281- feature_payload = make_csv_payload (out_data )
282- target_payload = make_csv_payload (out_target ) if out_target is not None else None
283- out_format = "csv"
284- target_out_format = out_format
250+ feature_payload = make_csv_payload (out_data )
251+ target_payload = make_csv_payload (out_target ) if out_target is not None else None
252+ out_format = "csv"
285253
286254 out_fields = {
287255 "X.format" : out_format ,
@@ -306,7 +274,7 @@ def _transform(self, logger=None):
306274 if target_payload is not None :
307275 out_fields .update (
308276 {
309- "y.format" : target_out_format ,
277+ "y.format" : "csv" ,
310278 Y_TRANSFORM_KEY : (
311279 Y_TRANSFORM_KEY ,
312280 target_payload ,
0 commit comments