Skip to content

Commit d6de377

Browse files
authored
optimize phi::CPUPlace in paddle/fluid/inference/ (#77208)
1 parent 39facf8 commit d6de377

28 files changed

+97
-105
lines changed

paddle/fluid/inference/analysis/passes/convert_to_mixed_precision.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ ConvertToMixedPrecisionPass::ConvertToMixedPrecisionPass(
7474
}
7575

7676
void ConvertToMixedPrecisionPass::LoadModel() {
77-
framework::Executor exe{phi::CPUPlace{}};
77+
framework::Executor exe{CPUPlace{}};
7878
// If we did not find the provided weight path,
7979
// we assume that the model to be converted only has a model file and no
8080
// params file, we believe this situation is reasonable. In this case, weight
@@ -177,7 +177,7 @@ void ConvertToMixedPrecisionPass::SaveMixedModel() {
177177
op->SetAttr("file_path", save_params_path);
178178
op->CheckAttrs();
179179

180-
framework::Executor exe(phi::CPUPlace{});
180+
framework::Executor exe(CPUPlace{});
181181
exe.Run(save_program, &scope_, 0, true, true);
182182
};
183183

paddle/fluid/inference/analysis/passes/ir_graph_build_pass.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,7 @@ void IrGraphBuildPass::RunImpl(Argument *argument) {
4242
// so that the parameters will on the same device, or they will keep copying
4343
// between difference devices.
4444
phi::Place place;
45-
place = phi::CPUPlace();
45+
place = CPUPlace();
4646

4747
if (argument->model_dir_valid()) {
4848
auto program =

paddle/fluid/inference/analysis/passes/ir_params_sync_among_devices_pass.cc

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ void IrParamsSyncAmongDevicesPass::CopyParamsToGpu(Argument *argument) {
5353
true,
5454
common::errors::PreconditionNotMet(
5555
"The gpu_device_id field should be valid"));
56-
phi::Place place = phi::GPUPlace(argument->gpu_device_id());
56+
phi::Place place = GPUPlace(argument->gpu_device_id());
5757
auto *scope = argument->scope_ptr();
5858
std::vector<std::string> all_vars = scope->LocalVarNames();
5959

@@ -130,7 +130,7 @@ void IrParamsSyncAmongDevicesPass::CopyParamsToGpu(Argument *argument) {
130130
dst_ptr = dev_ctx->Alloc(t, t->dtype());
131131
phi::memory_utils::Copy(place,
132132
dst_ptr,
133-
phi::CPUPlace(),
133+
CPUPlace(),
134134
src_ptr,
135135
t->numel() * phi::SizeOf(t->dtype()),
136136
stream);
@@ -269,7 +269,7 @@ void IrParamsSyncAmongDevicesPass::RunImpl(Argument *argument) {
269269
CopyParamsToXpu(argument);
270270
}
271271
#endif
272-
paddle::memory::Release(phi::CPUPlace());
272+
paddle::memory::Release(CPUPlace());
273273
}
274274

275275
std::string IrParamsSyncAmongDevicesPass::repr() const {

paddle/fluid/inference/analysis/passes/save_optimized_model_pass.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -91,7 +91,7 @@ void SaveOptimizedModelPass::SaveOptimizedModel(Argument* argument) {
9191
op->SetAttr("file_path", save_params_path);
9292
op->CheckAttrs();
9393

94-
framework::Executor exe(phi::CPUPlace{});
94+
framework::Executor exe(CPUPlace{});
9595
exe.Run(save_program, &scope, 0, true, true);
9696
};
9797
// TODO(shentanyue01): Setting hardware and version identification for

paddle/fluid/inference/io.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -273,7 +273,7 @@ void SaveVars(const framework::Scope& scope,
273273
op->SetAttr("file_path", dirname + "/param");
274274
op->CheckAttrs();
275275

276-
phi::CPUPlace place;
276+
CPUPlace place;
277277
framework::Executor exe(place);
278278
exe.Run(prog, const_cast<framework::Scope*>(&scope), 0, true, true);
279279
}

paddle/fluid/inference/tensorrt/convert/batch_norm_op.cc

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -70,18 +70,18 @@ class BatchNormOpConverter : public OpConverter {
7070
scale_tensor.Resize(Scale_t->dims());
7171
variance_tensor.Resize(Variance_t->dims());
7272

73-
phi::CPUPlace cpu_place;
73+
CPUPlace cpu_place;
7474
// copy data from gpu to cpu
7575
paddle::framework::TensorCopySync((*Bias_t), cpu_place, &bias_tensor);
7676
paddle::framework::TensorCopySync((*Mean_t), cpu_place, &mean_tensor);
7777
paddle::framework::TensorCopySync((*Scale_t), cpu_place, &scale_tensor);
7878
paddle::framework::TensorCopySync(
7979
(*Variance_t), cpu_place, &variance_tensor);
8080

81-
auto* bias_data = bias_tensor.mutable_data<float>(phi::CPUPlace());
82-
auto* mean_data = mean_tensor.mutable_data<float>(phi::CPUPlace());
83-
auto* scale_data = scale_tensor.mutable_data<float>(phi::CPUPlace());
84-
auto* variance_data = variance_tensor.mutable_data<float>(phi::CPUPlace());
81+
auto* bias_data = bias_tensor.mutable_data<float>(CPUPlace());
82+
auto* mean_data = mean_tensor.mutable_data<float>(CPUPlace());
83+
auto* scale_data = scale_tensor.mutable_data<float>(CPUPlace());
84+
auto* variance_data = variance_tensor.mutable_data<float>(CPUPlace());
8585

8686
std::unique_ptr<phi::DenseTensor> combine_scale_tensor(
8787
new phi::DenseTensor());
@@ -92,9 +92,9 @@ class BatchNormOpConverter : public OpConverter {
9292
combine_bias_tensor->Resize(bias_tensor.dims());
9393

9494
auto* combine_scale_data =
95-
combine_scale_tensor->mutable_data<float>(phi::CPUPlace());
95+
combine_scale_tensor->mutable_data<float>(CPUPlace());
9696
auto* combine_bias_data =
97-
combine_bias_tensor->mutable_data<float>(phi::CPUPlace());
97+
combine_bias_tensor->mutable_data<float>(CPUPlace());
9898

9999
size_t ele_num = combine_scale_tensor->memory_size() / sizeof(float);
100100

paddle/fluid/inference/tensorrt/convert/dropout_op.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -45,10 +45,10 @@ class DropoutOpConverter : public OpConverter {
4545
return;
4646
}
4747

48-
phi::CPUPlace cpu_place;
48+
CPUPlace cpu_place;
4949
std::unique_ptr<phi::DenseTensor> weight_tensor(new phi::DenseTensor());
5050
weight_tensor->Resize(common::make_ddim({1}));
51-
auto* weight_data = weight_tensor->mutable_data<float>(phi::CPUPlace());
51+
auto* weight_data = weight_tensor->mutable_data<float>(CPUPlace());
5252
weight_data[0] = 1 - dropout_prob;
5353

5454
TensorRTEngine::Weight scale_weights{

paddle/fluid/inference/tensorrt/convert/fill_constant_op.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -95,13 +95,13 @@ class FillConstantOpConverter : public OpConverter {
9595
void* trt_data = nullptr;
9696
size_t trt_num;
9797
if (dtype == 2 || dtype == 3) { // int,int64
98-
auto* tmp_ptr = out_tensor->mutable_data<int>(phi::CPUPlace());
98+
auto* tmp_ptr = out_tensor->mutable_data<int>(CPUPlace());
9999
for (int64_t i = 0; i < out_tensor->numel(); i++)
100100
tmp_ptr[i] = std::stoi(str_value);
101101
trt_dtype = nvinfer1::DataType::kINT32;
102102
trt_data = static_cast<void*>(tmp_ptr);
103103
} else if (dtype == 5) { // float
104-
auto* tmp_ptr = out_tensor->mutable_data<float>(phi::CPUPlace());
104+
auto* tmp_ptr = out_tensor->mutable_data<float>(CPUPlace());
105105
for (int64_t i = 0; i < out_tensor->numel(); i++)
106106
tmp_ptr[i] = std::stof(str_value);
107107
trt_data = static_cast<void*>(tmp_ptr);

paddle/fluid/inference/tensorrt/convert/gelu_op.cc

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ class GeluOpConverter : public OpConverter {
4545
auto create_weights = [&](float data, std::string type) -> float* {
4646
std::unique_ptr<phi::DenseTensor> tmp_tensor(new phi::DenseTensor());
4747
tmp_tensor->Resize({1});
48-
auto* tmp_data = tmp_tensor->mutable_data<float>(phi::CPUPlace());
48+
auto* tmp_data = tmp_tensor->mutable_data<float>(CPUPlace());
4949
tmp_data[0] = data;
5050
engine_->SetWeights(out_name + "_gelu_op_" + type,
5151
std::move(tmp_tensor));
@@ -146,7 +146,7 @@ class GeluOpConverter : public OpConverter {
146146
auto create_weights = [&](float data, std::string type) -> float* {
147147
std::unique_ptr<phi::DenseTensor> tmp_tensor(new phi::DenseTensor());
148148
tmp_tensor->Resize({1});
149-
auto* tmp_data = tmp_tensor->mutable_data<float>(phi::CPUPlace());
149+
auto* tmp_data = tmp_tensor->mutable_data<float>(CPUPlace());
150150
tmp_data[0] = data;
151151
engine_->SetWeights(out_name + "_gelu_op_" + type,
152152
std::move(tmp_tensor));

paddle/fluid/inference/tensorrt/convert/leaky_relu_op.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@ class LeakyReluOpConverter : public OpConverter {
4747
engine_->SetTensorDynamicRange(input, in_scale);
4848
}
4949
#else
50-
phi::CPUPlace place;
50+
CPUPlace place;
5151
std::unique_ptr<phi::DenseTensor> alpha_tensor(new phi::DenseTensor());
5252
alpha_tensor->Resize(common::make_ddim({2}));
5353
float* alpha_data = alpha_tensor->mutable_data<float>(place);

0 commit comments

Comments
 (0)