diff --git a/api_version.lock b/api_version.lock index 720b1f147..b60e1d2c3 100644 --- a/api_version.lock +++ b/api_version.lock @@ -1 +1 @@ -v0.1.457 +v0.1.460 diff --git a/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelines.java b/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelines.java index f3a9128aa..ad12621e3 100644 --- a/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelines.java +++ b/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelines.java @@ -4674,13 +4674,13 @@ public static com.google.protobuf.Parser parser() { } } - public interface UpdateDataPipelineRequestOrBuilder extends - // @@protoc_insertion_point(interface_extends:viam.app.datapipelines.v1.UpdateDataPipelineRequest) + public interface RenameDataPipelineRequestOrBuilder extends + // @@protoc_insertion_point(interface_extends:viam.app.datapipelines.v1.RenameDataPipelineRequest) com.google.protobuf.MessageLiteOrBuilder { /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4689,7 +4689,7 @@ public interface UpdateDataPipelineRequestOrBuilder extends java.lang.String getId(); /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4700,7 +4700,7 @@ public interface UpdateDataPipelineRequestOrBuilder extends /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4709,7 +4709,7 @@ public interface UpdateDataPipelineRequestOrBuilder extends java.lang.String getName(); /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4717,109 +4717,24 @@ public interface UpdateDataPipelineRequestOrBuilder extends */ com.google.protobuf.ByteString getNameBytes(); - - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return A list containing the mqlBinary. - */ - java.util.List getMqlBinaryList(); - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return The count of mqlBinary. - */ - int getMqlBinaryCount(); - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param index The index of the element to return. - * @return The mqlBinary at the given index. - */ - com.google.protobuf.ByteString getMqlBinary(int index); - - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The schedule. - */ - java.lang.String getSchedule(); - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The bytes for schedule. - */ - com.google.protobuf.ByteString - getScheduleBytes(); - - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return Whether the dataSourceType field is set. - */ - boolean hasDataSourceType(); - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The enum numeric value on the wire for dataSourceType. - */ - int getDataSourceTypeValue(); - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The dataSourceType. - */ - viam.app.data.v1.Data.TabularDataSourceType getDataSourceType(); } /** - * Protobuf type {@code viam.app.datapipelines.v1.UpdateDataPipelineRequest} + * Protobuf type {@code viam.app.datapipelines.v1.RenameDataPipelineRequest} */ - public static final class UpdateDataPipelineRequest extends + public static final class RenameDataPipelineRequest extends com.google.protobuf.GeneratedMessageLite< - UpdateDataPipelineRequest, UpdateDataPipelineRequest.Builder> implements - // @@protoc_insertion_point(message_implements:viam.app.datapipelines.v1.UpdateDataPipelineRequest) - UpdateDataPipelineRequestOrBuilder { - private UpdateDataPipelineRequest() { + RenameDataPipelineRequest, RenameDataPipelineRequest.Builder> implements + // @@protoc_insertion_point(message_implements:viam.app.datapipelines.v1.RenameDataPipelineRequest) + RenameDataPipelineRequestOrBuilder { + private RenameDataPipelineRequest() { id_ = ""; name_ = ""; - mqlBinary_ = emptyProtobufList(); - schedule_ = ""; } - private int bitField0_; public static final int ID_FIELD_NUMBER = 1; private java.lang.String id_; /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4831,7 +4746,7 @@ public java.lang.String getId() { } /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4844,7 +4759,7 @@ public java.lang.String getId() { } /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4858,7 +4773,7 @@ private void setId( } /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4869,7 +4784,7 @@ private void clearId() { } /** *
-     * The ID of the data pipeline to update.
+     * The ID of the data pipeline to rename.
      * 
* * string id = 1 [json_name = "id"]; @@ -4886,7 +4801,7 @@ private void setIdBytes( private java.lang.String name_; /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4898,7 +4813,7 @@ public java.lang.String getName() { } /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4911,7 +4826,7 @@ public java.lang.String getName() { } /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4925,7 +4840,7 @@ private void setName( } /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4936,7 +4851,7 @@ private void clearName() { } /** *
-     * A unique identifier at the org level.
+     * A unique identifier at the organization level.
      * 
* * string name = 2 [json_name = "name"]; @@ -4949,328 +4864,75 @@ private void setNameBytes( } - public static final int MQL_BINARY_FIELD_NUMBER = 3; - private com.google.protobuf.Internal.ProtobufList mqlBinary_; - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return A list containing the mqlBinary. - */ - @java.lang.Override - public java.util.List - getMqlBinaryList() { - return mqlBinary_; + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + java.nio.ByteBuffer data) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return The count of mqlBinary. - */ - @java.lang.Override - public int getMqlBinaryCount() { - return mqlBinary_.size(); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + java.nio.ByteBuffer data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data, extensionRegistry); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param index The index of the element to return. - * @return The mqlBinary at the given index. - */ - @java.lang.Override - public com.google.protobuf.ByteString getMqlBinary(int index) { - return mqlBinary_.get(index); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + com.google.protobuf.ByteString data) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data); } - private void ensureMqlBinaryIsMutable() { - com.google.protobuf.Internal.ProtobufList tmp = mqlBinary_; - if (!tmp.isModifiable()) { - mqlBinary_ = - com.google.protobuf.GeneratedMessageLite.mutableCopy(tmp); - } + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + com.google.protobuf.ByteString data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data, extensionRegistry); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param index The index to set the value at. - * @param value The mqlBinary to set. - */ - private void setMqlBinary( - int index, com.google.protobuf.ByteString value) { - java.lang.Class valueClass = value.getClass(); - ensureMqlBinaryIsMutable(); - mqlBinary_.set(index, value); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom(byte[] data) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param value The mqlBinary to add. - */ - private void addMqlBinary(com.google.protobuf.ByteString value) { - java.lang.Class valueClass = value.getClass(); - ensureMqlBinaryIsMutable(); - mqlBinary_.add(value); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + byte[] data, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws com.google.protobuf.InvalidProtocolBufferException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, data, extensionRegistry); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param values The mqlBinary to add. - */ - private void addAllMqlBinary( - java.lang.Iterable values) { - ensureMqlBinaryIsMutable(); - com.google.protobuf.AbstractMessageLite.addAll( - values, mqlBinary_); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom(java.io.InputStream input) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, input); } - /** - *
-     * A MongoDB aggregation pipeline as a list of BSON documents, where
-     * each document is one stage in the pipeline.
-     * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - */ - private void clearMqlBinary() { - mqlBinary_ = emptyProtobufList(); + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return com.google.protobuf.GeneratedMessageLite.parseFrom( + DEFAULT_INSTANCE, input, extensionRegistry); } - public static final int SCHEDULE_FIELD_NUMBER = 4; - private java.lang.String schedule_; - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The schedule. - */ - @java.lang.Override - public java.lang.String getSchedule() { - return schedule_; + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseDelimitedFrom(java.io.InputStream input) + throws java.io.IOException { + return parseDelimitedFrom(DEFAULT_INSTANCE, input); } - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The bytes for schedule. - */ - @java.lang.Override - public com.google.protobuf.ByteString - getScheduleBytes() { - return com.google.protobuf.ByteString.copyFromUtf8(schedule_); + + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseDelimitedFrom( + java.io.InputStream input, + com.google.protobuf.ExtensionRegistryLite extensionRegistry) + throws java.io.IOException { + return parseDelimitedFrom(DEFAULT_INSTANCE, input, extensionRegistry); } - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @param value The schedule to set. - */ - private void setSchedule( - java.lang.String value) { - java.lang.Class valueClass = value.getClass(); - - schedule_ = value; - } - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - */ - private void clearSchedule() { - - schedule_ = getDefaultInstance().getSchedule(); - } - /** - *
-     * A cron expression representing the expected execution schedule in UTC (note this also
-     * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-     * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @param value The bytes for schedule to set. - */ - private void setScheduleBytes( - com.google.protobuf.ByteString value) { - checkByteStringIsUtf8(value); - schedule_ = value.toStringUtf8(); - - } - - public static final int DATA_SOURCE_TYPE_FIELD_NUMBER = 5; - private int dataSourceType_; - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return Whether the dataSourceType field is set. - */ - @java.lang.Override - public boolean hasDataSourceType() { - return ((bitField0_ & 0x00000001) != 0); - } - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The enum numeric value on the wire for dataSourceType. - */ - @java.lang.Override - public int getDataSourceTypeValue() { - return dataSourceType_; - } - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The dataSourceType. - */ - @java.lang.Override - public viam.app.data.v1.Data.TabularDataSourceType getDataSourceType() { - viam.app.data.v1.Data.TabularDataSourceType result = viam.app.data.v1.Data.TabularDataSourceType.forNumber(dataSourceType_); - return result == null ? viam.app.data.v1.Data.TabularDataSourceType.UNRECOGNIZED : result; - } - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @param value The enum numeric value on the wire for dataSourceType to set. - */ - private void setDataSourceTypeValue(int value) { - bitField0_ |= 0x00000001; dataSourceType_ = value; - } - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @param value The dataSourceType to set. - */ - private void setDataSourceType(viam.app.data.v1.Data.TabularDataSourceType value) { - dataSourceType_ = value.getNumber(); - bitField0_ |= 0x00000001; - } - /** - *
-     * The type of data source for the pipeline. If not specified, default is standard data storage.
-     * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - */ - private void clearDataSourceType() { - bitField0_ = (bitField0_ & ~0x00000001); - dataSourceType_ = 0; - } - - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - java.nio.ByteBuffer data) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - java.nio.ByteBuffer data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data, extensionRegistry); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - com.google.protobuf.ByteString data) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - com.google.protobuf.ByteString data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data, extensionRegistry); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom(byte[] data) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - byte[] data, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws com.google.protobuf.InvalidProtocolBufferException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, data, extensionRegistry); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom(java.io.InputStream input) - throws java.io.IOException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, input); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return com.google.protobuf.GeneratedMessageLite.parseFrom( - DEFAULT_INSTANCE, input, extensionRegistry); - } - - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return parseDelimitedFrom(DEFAULT_INSTANCE, input); - } - - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseDelimitedFrom( - java.io.InputStream input, - com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return parseDelimitedFrom(DEFAULT_INSTANCE, input, extensionRegistry); - } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, input); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5281,19 +4943,19 @@ public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest public static Builder newBuilder() { return (Builder) DEFAULT_INSTANCE.createBuilder(); } - public static Builder newBuilder(viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest prototype) { + public static Builder newBuilder(viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest prototype) { return DEFAULT_INSTANCE.createBuilder(prototype); } /** - * Protobuf type {@code viam.app.datapipelines.v1.UpdateDataPipelineRequest} + * Protobuf type {@code viam.app.datapipelines.v1.RenameDataPipelineRequest} */ public static final class Builder extends com.google.protobuf.GeneratedMessageLite.Builder< - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest, Builder> implements - // @@protoc_insertion_point(builder_implements:viam.app.datapipelines.v1.UpdateDataPipelineRequest) - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequestOrBuilder { - // Construct using viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest.newBuilder() + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest, Builder> implements + // @@protoc_insertion_point(builder_implements:viam.app.datapipelines.v1.RenameDataPipelineRequest) + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequestOrBuilder { + // Construct using viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest.newBuilder() private Builder() { super(DEFAULT_INSTANCE); } @@ -5301,7 +4963,7 @@ private Builder() { /** *
-       * The ID of the data pipeline to update.
+       * The ID of the data pipeline to rename.
        * 
* * string id = 1 [json_name = "id"]; @@ -5313,7 +4975,7 @@ public java.lang.String getId() { } /** *
-       * The ID of the data pipeline to update.
+       * The ID of the data pipeline to rename.
        * 
* * string id = 1 [json_name = "id"]; @@ -5326,7 +4988,7 @@ public java.lang.String getId() { } /** *
-       * The ID of the data pipeline to update.
+       * The ID of the data pipeline to rename.
        * 
* * string id = 1 [json_name = "id"]; @@ -5341,7 +5003,7 @@ public Builder setId( } /** *
-       * The ID of the data pipeline to update.
+       * The ID of the data pipeline to rename.
        * 
* * string id = 1 [json_name = "id"]; @@ -5354,7 +5016,7 @@ public Builder clearId() { } /** *
-       * The ID of the data pipeline to update.
+       * The ID of the data pipeline to rename.
        * 
* * string id = 1 [json_name = "id"]; @@ -5370,7 +5032,7 @@ public Builder setIdBytes( /** *
-       * A unique identifier at the org level.
+       * A unique identifier at the organization level.
        * 
* * string name = 2 [json_name = "name"]; @@ -5382,7 +5044,7 @@ public java.lang.String getName() { } /** *
-       * A unique identifier at the org level.
+       * A unique identifier at the organization level.
        * 
* * string name = 2 [json_name = "name"]; @@ -5395,7 +5057,7 @@ public java.lang.String getName() { } /** *
-       * A unique identifier at the org level.
+       * A unique identifier at the organization level.
        * 
* * string name = 2 [json_name = "name"]; @@ -5410,7 +5072,7 @@ public Builder setName( } /** *
-       * A unique identifier at the org level.
+       * A unique identifier at the organization level.
        * 
* * string name = 2 [json_name = "name"]; @@ -5423,7 +5085,7 @@ public Builder clearName() { } /** *
-       * A unique identifier at the org level.
+       * A unique identifier at the organization level.
        * 
* * string name = 2 [json_name = "name"]; @@ -5437,263 +5099,7 @@ public Builder setNameBytes( return this; } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return A list containing the mqlBinary. - */ - @java.lang.Override - public java.util.List - getMqlBinaryList() { - return java.util.Collections.unmodifiableList( - instance.getMqlBinaryList()); - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return The count of mqlBinary. - */ - @java.lang.Override - public int getMqlBinaryCount() { - return instance.getMqlBinaryCount(); - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param index The index of the element to return. - * @return The mqlBinary at the given index. - */ - @java.lang.Override - public com.google.protobuf.ByteString getMqlBinary(int index) { - return instance.getMqlBinary(index); - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param value The mqlBinary to set. - * @return This builder for chaining. - */ - public Builder setMqlBinary( - int index, com.google.protobuf.ByteString value) { - copyOnWrite(); - instance.setMqlBinary(index, value); - return this; - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param value The mqlBinary to add. - * @return This builder for chaining. - */ - public Builder addMqlBinary(com.google.protobuf.ByteString value) { - copyOnWrite(); - instance.addMqlBinary(value); - return this; - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @param values The mqlBinary to add. - * @return This builder for chaining. - */ - public Builder addAllMqlBinary( - java.lang.Iterable values) { - copyOnWrite(); - instance.addAllMqlBinary(values); - return this; - } - /** - *
-       * A MongoDB aggregation pipeline as a list of BSON documents, where
-       * each document is one stage in the pipeline.
-       * 
- * - * repeated bytes mql_binary = 3 [json_name = "mqlBinary"]; - * @return This builder for chaining. - */ - public Builder clearMqlBinary() { - copyOnWrite(); - instance.clearMqlBinary(); - return this; - } - - /** - *
-       * A cron expression representing the expected execution schedule in UTC (note this also
-       * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-       * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The schedule. - */ - @java.lang.Override - public java.lang.String getSchedule() { - return instance.getSchedule(); - } - /** - *
-       * A cron expression representing the expected execution schedule in UTC (note this also
-       * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-       * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return The bytes for schedule. - */ - @java.lang.Override - public com.google.protobuf.ByteString - getScheduleBytes() { - return instance.getScheduleBytes(); - } - /** - *
-       * A cron expression representing the expected execution schedule in UTC (note this also
-       * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-       * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @param value The schedule to set. - * @return This builder for chaining. - */ - public Builder setSchedule( - java.lang.String value) { - copyOnWrite(); - instance.setSchedule(value); - return this; - } - /** - *
-       * A cron expression representing the expected execution schedule in UTC (note this also
-       * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-       * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @return This builder for chaining. - */ - public Builder clearSchedule() { - copyOnWrite(); - instance.clearSchedule(); - return this; - } - /** - *
-       * A cron expression representing the expected execution schedule in UTC (note this also
-       * defines the input time window; an hourly schedule would process 1 hour of data at a time).
-       * 
- * - * string schedule = 4 [json_name = "schedule"]; - * @param value The bytes for schedule to set. - * @return This builder for chaining. - */ - public Builder setScheduleBytes( - com.google.protobuf.ByteString value) { - copyOnWrite(); - instance.setScheduleBytes(value); - return this; - } - - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return Whether the dataSourceType field is set. - */ - @java.lang.Override - public boolean hasDataSourceType() { - return instance.hasDataSourceType(); - } - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The enum numeric value on the wire for dataSourceType. - */ - @java.lang.Override - public int getDataSourceTypeValue() { - return instance.getDataSourceTypeValue(); - } - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @param value The dataSourceType to set. - * @return This builder for chaining. - */ - public Builder setDataSourceTypeValue(int value) { - copyOnWrite(); - instance.setDataSourceTypeValue(value); - return this; - } - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return The dataSourceType. - */ - @java.lang.Override - public viam.app.data.v1.Data.TabularDataSourceType getDataSourceType() { - return instance.getDataSourceType(); - } - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @param value The enum numeric value on the wire for dataSourceType to set. - * @return This builder for chaining. - */ - public Builder setDataSourceType(viam.app.data.v1.Data.TabularDataSourceType value) { - copyOnWrite(); - instance.setDataSourceType(value); - return this; - } - /** - *
-       * The type of data source for the pipeline. If not specified, default is standard data storage.
-       * 
- * - * optional .viam.app.data.v1.TabularDataSourceType data_source_type = 5 [json_name = "dataSourceType"]; - * @return This builder for chaining. - */ - public Builder clearDataSourceType() { - copyOnWrite(); - instance.clearDataSourceType(); - return this; - } - - // @@protoc_insertion_point(builder_scope:viam.app.datapipelines.v1.UpdateDataPipelineRequest) + // @@protoc_insertion_point(builder_scope:viam.app.datapipelines.v1.RenameDataPipelineRequest) } @java.lang.Override @java.lang.SuppressWarnings({"unchecked", "fallthrough"}) @@ -5702,23 +5108,19 @@ protected final java.lang.Object dynamicMethod( java.lang.Object arg0, java.lang.Object arg1) { switch (method) { case NEW_MUTABLE_INSTANCE: { - return new viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest(); + return new viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest(); } case NEW_BUILDER: { return new Builder(); } case BUILD_MESSAGE_INFO: { java.lang.Object[] objects = new java.lang.Object[] { - "bitField0_", "id_", "name_", - "mqlBinary_", - "schedule_", - "dataSourceType_", }; java.lang.String info = - "\u0000\u0005\u0000\u0001\u0001\u0005\u0005\u0000\u0001\u0000\u0001\u0208\u0002\u0208" + - "\u0003\u001c\u0004\u0208\u0005\u100c\u0000"; + "\u0000\u0002\u0000\u0000\u0001\u0002\u0002\u0000\u0000\u0000\u0001\u0208\u0002\u0208" + + ""; return newMessageInfo(DEFAULT_INSTANCE, info, objects); } // fall through @@ -5726,13 +5128,13 @@ protected final java.lang.Object dynamicMethod( return DEFAULT_INSTANCE; } case GET_PARSER: { - com.google.protobuf.Parser parser = PARSER; + com.google.protobuf.Parser parser = PARSER; if (parser == null) { - synchronized (viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest.class) { + synchronized (viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest.class) { parser = PARSER; if (parser == null) { parser = - new DefaultInstanceBasedParser( + new DefaultInstanceBasedParser( DEFAULT_INSTANCE); PARSER = parser; } @@ -5751,86 +5153,86 @@ protected final java.lang.Object dynamicMethod( } - // @@protoc_insertion_point(class_scope:viam.app.datapipelines.v1.UpdateDataPipelineRequest) - private static final viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest DEFAULT_INSTANCE; + // @@protoc_insertion_point(class_scope:viam.app.datapipelines.v1.RenameDataPipelineRequest) + private static final viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest DEFAULT_INSTANCE; static { - UpdateDataPipelineRequest defaultInstance = new UpdateDataPipelineRequest(); + RenameDataPipelineRequest defaultInstance = new RenameDataPipelineRequest(); // New instances are implicitly immutable so no need to make // immutable. DEFAULT_INSTANCE = defaultInstance; com.google.protobuf.GeneratedMessageLite.registerDefaultInstance( - UpdateDataPipelineRequest.class, defaultInstance); + RenameDataPipelineRequest.class, defaultInstance); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest getDefaultInstance() { + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest getDefaultInstance() { return DEFAULT_INSTANCE; } - private static volatile com.google.protobuf.Parser PARSER; + private static volatile com.google.protobuf.Parser PARSER; - public static com.google.protobuf.Parser parser() { + public static com.google.protobuf.Parser parser() { return DEFAULT_INSTANCE.getParserForType(); } } - public interface UpdateDataPipelineResponseOrBuilder extends - // @@protoc_insertion_point(interface_extends:viam.app.datapipelines.v1.UpdateDataPipelineResponse) + public interface RenameDataPipelineResponseOrBuilder extends + // @@protoc_insertion_point(interface_extends:viam.app.datapipelines.v1.RenameDataPipelineResponse) com.google.protobuf.MessageLiteOrBuilder { } /** - * Protobuf type {@code viam.app.datapipelines.v1.UpdateDataPipelineResponse} + * Protobuf type {@code viam.app.datapipelines.v1.RenameDataPipelineResponse} */ - public static final class UpdateDataPipelineResponse extends + public static final class RenameDataPipelineResponse extends com.google.protobuf.GeneratedMessageLite< - UpdateDataPipelineResponse, UpdateDataPipelineResponse.Builder> implements - // @@protoc_insertion_point(message_implements:viam.app.datapipelines.v1.UpdateDataPipelineResponse) - UpdateDataPipelineResponseOrBuilder { - private UpdateDataPipelineResponse() { + RenameDataPipelineResponse, RenameDataPipelineResponse.Builder> implements + // @@protoc_insertion_point(message_implements:viam.app.datapipelines.v1.RenameDataPipelineResponse) + RenameDataPipelineResponseOrBuilder { + private RenameDataPipelineResponse() { } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( java.nio.ByteBuffer data) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( java.nio.ByteBuffer data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data, extensionRegistry); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data, extensionRegistry); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom(byte[] data) + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, data, extensionRegistry); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom(java.io.InputStream input) + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, input); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5838,24 +5240,24 @@ public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse DEFAULT_INSTANCE, input, extensionRegistry); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseDelimitedFrom(java.io.InputStream input) + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return parseDelimitedFrom(DEFAULT_INSTANCE, input); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseDelimitedFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return parseDelimitedFrom(DEFAULT_INSTANCE, input, extensionRegistry); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return com.google.protobuf.GeneratedMessageLite.parseFrom( DEFAULT_INSTANCE, input); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse parseFrom( + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { @@ -5866,25 +5268,25 @@ public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse public static Builder newBuilder() { return (Builder) DEFAULT_INSTANCE.createBuilder(); } - public static Builder newBuilder(viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse prototype) { + public static Builder newBuilder(viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse prototype) { return DEFAULT_INSTANCE.createBuilder(prototype); } /** - * Protobuf type {@code viam.app.datapipelines.v1.UpdateDataPipelineResponse} + * Protobuf type {@code viam.app.datapipelines.v1.RenameDataPipelineResponse} */ public static final class Builder extends com.google.protobuf.GeneratedMessageLite.Builder< - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse, Builder> implements - // @@protoc_insertion_point(builder_implements:viam.app.datapipelines.v1.UpdateDataPipelineResponse) - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponseOrBuilder { - // Construct using viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse.newBuilder() + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse, Builder> implements + // @@protoc_insertion_point(builder_implements:viam.app.datapipelines.v1.RenameDataPipelineResponse) + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponseOrBuilder { + // Construct using viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse.newBuilder() private Builder() { super(DEFAULT_INSTANCE); } - // @@protoc_insertion_point(builder_scope:viam.app.datapipelines.v1.UpdateDataPipelineResponse) + // @@protoc_insertion_point(builder_scope:viam.app.datapipelines.v1.RenameDataPipelineResponse) } @java.lang.Override @java.lang.SuppressWarnings({"unchecked", "fallthrough"}) @@ -5893,7 +5295,7 @@ protected final java.lang.Object dynamicMethod( java.lang.Object arg0, java.lang.Object arg1) { switch (method) { case NEW_MUTABLE_INSTANCE: { - return new viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse(); + return new viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse(); } case NEW_BUILDER: { return new Builder(); @@ -5908,13 +5310,13 @@ protected final java.lang.Object dynamicMethod( return DEFAULT_INSTANCE; } case GET_PARSER: { - com.google.protobuf.Parser parser = PARSER; + com.google.protobuf.Parser parser = PARSER; if (parser == null) { - synchronized (viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse.class) { + synchronized (viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse.class) { parser = PARSER; if (parser == null) { parser = - new DefaultInstanceBasedParser( + new DefaultInstanceBasedParser( DEFAULT_INSTANCE); PARSER = parser; } @@ -5933,24 +5335,24 @@ protected final java.lang.Object dynamicMethod( } - // @@protoc_insertion_point(class_scope:viam.app.datapipelines.v1.UpdateDataPipelineResponse) - private static final viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse DEFAULT_INSTANCE; + // @@protoc_insertion_point(class_scope:viam.app.datapipelines.v1.RenameDataPipelineResponse) + private static final viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse DEFAULT_INSTANCE; static { - UpdateDataPipelineResponse defaultInstance = new UpdateDataPipelineResponse(); + RenameDataPipelineResponse defaultInstance = new RenameDataPipelineResponse(); // New instances are implicitly immutable so no need to make // immutable. DEFAULT_INSTANCE = defaultInstance; com.google.protobuf.GeneratedMessageLite.registerDefaultInstance( - UpdateDataPipelineResponse.class, defaultInstance); + RenameDataPipelineResponse.class, defaultInstance); } - public static viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse getDefaultInstance() { + public static viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse getDefaultInstance() { return DEFAULT_INSTANCE; } - private static volatile com.google.protobuf.Parser PARSER; + private static volatile com.google.protobuf.Parser PARSER; - public static com.google.protobuf.Parser parser() { + public static com.google.protobuf.Parser parser() { return DEFAULT_INSTANCE.getParserForType(); } } diff --git a/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelinesServiceGrpc.java b/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelinesServiceGrpc.java index 4111582d9..5336c3250 100644 --- a/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelinesServiceGrpc.java +++ b/core/sdk/src/main/gen/viam/app/datapipelines/v1/DataPipelinesServiceGrpc.java @@ -108,34 +108,34 @@ viam.app.datapipelines.v1.DataPipelines.CreateDataPipelineResponse> getCreateDat return getCreateDataPipelineMethod; } - private static volatile io.grpc.MethodDescriptor getUpdateDataPipelineMethod; + private static volatile io.grpc.MethodDescriptor getRenameDataPipelineMethod; @io.grpc.stub.annotations.RpcMethod( - fullMethodName = SERVICE_NAME + '/' + "UpdateDataPipeline", - requestType = viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest.class, - responseType = viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse.class, + fullMethodName = SERVICE_NAME + '/' + "RenameDataPipeline", + requestType = viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest.class, + responseType = viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse.class, methodType = io.grpc.MethodDescriptor.MethodType.UNARY) - public static io.grpc.MethodDescriptor getUpdateDataPipelineMethod() { - io.grpc.MethodDescriptor getUpdateDataPipelineMethod; - if ((getUpdateDataPipelineMethod = DataPipelinesServiceGrpc.getUpdateDataPipelineMethod) == null) { + public static io.grpc.MethodDescriptor getRenameDataPipelineMethod() { + io.grpc.MethodDescriptor getRenameDataPipelineMethod; + if ((getRenameDataPipelineMethod = DataPipelinesServiceGrpc.getRenameDataPipelineMethod) == null) { synchronized (DataPipelinesServiceGrpc.class) { - if ((getUpdateDataPipelineMethod = DataPipelinesServiceGrpc.getUpdateDataPipelineMethod) == null) { - DataPipelinesServiceGrpc.getUpdateDataPipelineMethod = getUpdateDataPipelineMethod = - io.grpc.MethodDescriptor.newBuilder() + if ((getRenameDataPipelineMethod = DataPipelinesServiceGrpc.getRenameDataPipelineMethod) == null) { + DataPipelinesServiceGrpc.getRenameDataPipelineMethod = getRenameDataPipelineMethod = + io.grpc.MethodDescriptor.newBuilder() .setType(io.grpc.MethodDescriptor.MethodType.UNARY) - .setFullMethodName(generateFullMethodName(SERVICE_NAME, "UpdateDataPipeline")) + .setFullMethodName(generateFullMethodName(SERVICE_NAME, "RenameDataPipeline")) .setSampledToLocalTracing(true) .setRequestMarshaller(io.grpc.protobuf.lite.ProtoLiteUtils.marshaller( - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest.getDefaultInstance())) + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest.getDefaultInstance())) .setResponseMarshaller(io.grpc.protobuf.lite.ProtoLiteUtils.marshaller( - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse.getDefaultInstance())) + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse.getDefaultInstance())) .build(); } } } - return getUpdateDataPipelineMethod; + return getRenameDataPipelineMethod; } private static volatile io.grpc.MethodDescriptor - * UpdateDataPipeline modifies an existing data pipeline's configuration. + * RenameDataPipeline changes a data pipeline's name. * */ - default void updateDataPipeline(viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest request, - io.grpc.stub.StreamObserver responseObserver) { - io.grpc.stub.ServerCalls.asyncUnimplementedUnaryCall(getUpdateDataPipelineMethod(), responseObserver); + default void renameDataPipeline(viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest request, + io.grpc.stub.StreamObserver responseObserver) { + io.grpc.stub.ServerCalls.asyncUnimplementedUnaryCall(getRenameDataPipelineMethod(), responseObserver); } /** @@ -458,13 +458,13 @@ public void createDataPipeline(viam.app.datapipelines.v1.DataPipelines.CreateDat /** *
-     * UpdateDataPipeline modifies an existing data pipeline's configuration.
+     * RenameDataPipeline changes a data pipeline's name.
      * 
*/ - public void updateDataPipeline(viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest request, - io.grpc.stub.StreamObserver responseObserver) { + public void renameDataPipeline(viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest request, + io.grpc.stub.StreamObserver responseObserver) { io.grpc.stub.ClientCalls.asyncUnaryCall( - getChannel().newCall(getUpdateDataPipelineMethod(), getCallOptions()), request, responseObserver); + getChannel().newCall(getRenameDataPipelineMethod(), getCallOptions()), request, responseObserver); } /** @@ -563,12 +563,12 @@ public viam.app.datapipelines.v1.DataPipelines.CreateDataPipelineResponse create /** *
-     * UpdateDataPipeline modifies an existing data pipeline's configuration.
+     * RenameDataPipeline changes a data pipeline's name.
      * 
*/ - public viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse updateDataPipeline(viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest request) { + public viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse renameDataPipeline(viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest request) { return io.grpc.stub.ClientCalls.blockingUnaryCall( - getChannel(), getUpdateDataPipelineMethod(), getCallOptions(), request); + getChannel(), getRenameDataPipelineMethod(), getCallOptions(), request); } /** @@ -666,13 +666,13 @@ public com.google.common.util.concurrent.ListenableFuture - * UpdateDataPipeline modifies an existing data pipeline's configuration. + * RenameDataPipeline changes a data pipeline's name. * */ - public com.google.common.util.concurrent.ListenableFuture updateDataPipeline( - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest request) { + public com.google.common.util.concurrent.ListenableFuture renameDataPipeline( + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest request) { return io.grpc.stub.ClientCalls.futureUnaryCall( - getChannel().newCall(getUpdateDataPipelineMethod(), getCallOptions()), request); + getChannel().newCall(getRenameDataPipelineMethod(), getCallOptions()), request); } /** @@ -723,7 +723,7 @@ public com.google.common.util.concurrent.ListenableFuture responseObserv serviceImpl.createDataPipeline((viam.app.datapipelines.v1.DataPipelines.CreateDataPipelineRequest) request, (io.grpc.stub.StreamObserver) responseObserver); break; - case METHODID_UPDATE_DATA_PIPELINE: - serviceImpl.updateDataPipeline((viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest) request, - (io.grpc.stub.StreamObserver) responseObserver); + case METHODID_RENAME_DATA_PIPELINE: + serviceImpl.renameDataPipeline((viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest) request, + (io.grpc.stub.StreamObserver) responseObserver); break; case METHODID_DELETE_DATA_PIPELINE: serviceImpl.deleteDataPipeline((viam.app.datapipelines.v1.DataPipelines.DeleteDataPipelineRequest) request, @@ -818,12 +818,12 @@ public static final io.grpc.ServerServiceDefinition bindService(AsyncService ser viam.app.datapipelines.v1.DataPipelines.CreateDataPipelineResponse>( service, METHODID_CREATE_DATA_PIPELINE))) .addMethod( - getUpdateDataPipelineMethod(), + getRenameDataPipelineMethod(), io.grpc.stub.ServerCalls.asyncUnaryCall( new MethodHandlers< - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineRequest, - viam.app.datapipelines.v1.DataPipelines.UpdateDataPipelineResponse>( - service, METHODID_UPDATE_DATA_PIPELINE))) + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineRequest, + viam.app.datapipelines.v1.DataPipelines.RenameDataPipelineResponse>( + service, METHODID_RENAME_DATA_PIPELINE))) .addMethod( getDeleteDataPipelineMethod(), io.grpc.stub.ServerCalls.asyncUnaryCall( @@ -867,7 +867,7 @@ public static io.grpc.ServiceDescriptor getServiceDescriptor() { .addMethod(getGetDataPipelineMethod()) .addMethod(getListDataPipelinesMethod()) .addMethod(getCreateDataPipelineMethod()) - .addMethod(getUpdateDataPipelineMethod()) + .addMethod(getRenameDataPipelineMethod()) .addMethod(getDeleteDataPipelineMethod()) .addMethod(getEnableDataPipelineMethod()) .addMethod(getDisableDataPipelineMethod()) diff --git a/core/sdk/src/main/gen/viam/app/v1/Robot.java b/core/sdk/src/main/gen/viam/app/v1/Robot.java index c55dd7c0e..aa4465234 100644 --- a/core/sdk/src/main/gen/viam/app/v1/Robot.java +++ b/core/sdk/src/main/gen/viam/app/v1/Robot.java @@ -20206,6 +20206,26 @@ public interface RemoteConfigOrBuilder extends */ com.google.protobuf.ByteString getSecretBytes(); + + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The prefix. + */ + java.lang.String getPrefix(); + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The bytes for prefix. + */ + com.google.protobuf.ByteString + getPrefixBytes(); } /** *
@@ -20228,6 +20248,7 @@ private RemoteConfig() {
       managedBy_ = "";
       serviceConfigs_ = emptyProtobufList();
       secret_ = "";
+      prefix_ = "";
     }
     private int bitField0_;
     public static final int NAME_FIELD_NUMBER = 1;
@@ -20742,6 +20763,73 @@ private void setSecretBytes(
 
     }
 
+    public static final int PREFIX_FIELD_NUMBER = 11;
+    private java.lang.String prefix_;
+    /**
+     * 
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The prefix. + */ + @java.lang.Override + public java.lang.String getPrefix() { + return prefix_; + } + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The bytes for prefix. + */ + @java.lang.Override + public com.google.protobuf.ByteString + getPrefixBytes() { + return com.google.protobuf.ByteString.copyFromUtf8(prefix_); + } + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @param value The prefix to set. + */ + private void setPrefix( + java.lang.String value) { + java.lang.Class valueClass = value.getClass(); + + prefix_ = value; + } + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + */ + private void clearPrefix() { + + prefix_ = getDefaultInstance().getPrefix(); + } + /** + *
+     * A string with which to prefix all resource names fetched from this remote.
+     * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @param value The bytes for prefix to set. + */ + private void setPrefixBytes( + com.google.protobuf.ByteString value) { + checkByteStringIsUtf8(value); + prefix_ = value.toStringUtf8(); + + } + public static viam.app.v1.Robot.RemoteConfig parseFrom( java.nio.ByteBuffer data) throws com.google.protobuf.InvalidProtocolBufferException { @@ -21380,6 +21468,75 @@ public Builder setSecretBytes( return this; } + /** + *
+       * A string with which to prefix all resource names fetched from this remote.
+       * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The prefix. + */ + @java.lang.Override + public java.lang.String getPrefix() { + return instance.getPrefix(); + } + /** + *
+       * A string with which to prefix all resource names fetched from this remote.
+       * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return The bytes for prefix. + */ + @java.lang.Override + public com.google.protobuf.ByteString + getPrefixBytes() { + return instance.getPrefixBytes(); + } + /** + *
+       * A string with which to prefix all resource names fetched from this remote.
+       * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @param value The prefix to set. + * @return This builder for chaining. + */ + public Builder setPrefix( + java.lang.String value) { + copyOnWrite(); + instance.setPrefix(value); + return this; + } + /** + *
+       * A string with which to prefix all resource names fetched from this remote.
+       * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @return This builder for chaining. + */ + public Builder clearPrefix() { + copyOnWrite(); + instance.clearPrefix(); + return this; + } + /** + *
+       * A string with which to prefix all resource names fetched from this remote.
+       * 
+ * + * string prefix = 11 [json_name = "prefix"]; + * @param value The bytes for prefix to set. + * @return This builder for chaining. + */ + public Builder setPrefixBytes( + com.google.protobuf.ByteString value) { + copyOnWrite(); + instance.setPrefixBytes(value); + return this; + } + // @@protoc_insertion_point(builder_scope:viam.app.v1.RemoteConfig) } @java.lang.Override @@ -21408,11 +21565,12 @@ protected final java.lang.Object dynamicMethod( "serviceConfigs_", viam.app.v1.Robot.ResourceLevelServiceConfig.class, "secret_", + "prefix_", }; java.lang.String info = - "\u0000\n\u0000\u0001\u0001\n\n\u0000\u0001\u0000\u0001\u0208\u0002\u0208\u0003\u1009" + - "\u0000\u0004\u1009\u0001\u0005\u0208\u0006\u0007\u0007\u1009\u0002\b\u1009\u0003" + - "\t\u001b\n\u0208"; + "\u0000\u000b\u0000\u0001\u0001\u000b\u000b\u0000\u0001\u0000\u0001\u0208\u0002\u0208" + + "\u0003\u1009\u0000\u0004\u1009\u0001\u0005\u0208\u0006\u0007\u0007\u1009\u0002\b" + + "\u1009\u0003\t\u001b\n\u0208\u000b\u0208"; return newMessageInfo(DEFAULT_INSTANCE, info, objects); } // fall through diff --git a/core/sdk/src/main/resources/protofds/viam.api b/core/sdk/src/main/resources/protofds/viam.api index a67bbdea8..fe8e2dce4 100644 Binary files a/core/sdk/src/main/resources/protofds/viam.api and b/core/sdk/src/main/resources/protofds/viam.api differ