From dba1e516cd442626897e6d62080a2777aa12c97b Mon Sep 17 00:00:00 2001 From: kmontemayor Date: Mon, 4 May 2026 19:22:43 +0000 Subject: [PATCH] Run presubmit checks (no eol whitespace, empty line at eof) on all files --- .../actions/assert-is-collaborator/action.yml | 7 +-- .github/actions/comment-on-pr/action.yml | 4 +- .github/actions/get-pr-src-branch/action.yml | 2 +- .../gbml/postprocessed_metadata.proto | 2 +- .../gbml/subgraph_sampling_strategy.proto | 2 +- scala/.scalafix.conf | 2 +- .../dataset_metadata/DatasetMetadata.scala | 4 +- .../DatasetMetadataProto.scala | 2 +- ...NodeAnchorBasedLinkPredictionDataset.scala | 38 ++++++------ .../SupervisedLinkBasedTaskSplitDataset.scala | 8 +-- .../SupervisedNodeClassificationDataset.scala | 8 +-- .../FlattenedGraphMetadata.scala | 4 +- .../FlattenedGraphMetadataProto.scala | 2 +- .../NodeAnchorBasedLinkPredictionOutput.scala | 14 ++--- .../SupervisedLinkBasedTaskOutput.scala | 6 +- .../SupervisedNodeClassificationOutput.scala | 6 +- .../gbml/gigl_resource_config/Component.scala | 20 +++---- .../DataPreprocessorConfig.scala | 2 +- .../DistributedInferencerConfig.scala | 4 +- .../KFPTrainerConfig.scala | 12 ++-- .../LocalTrainerConfig.scala | 4 +- .../SparkResourceConfig.scala | 8 +-- .../VertexAiTrainerConfig.scala | 10 ++-- .../research/gbml/graph_schema/Edge.scala | 6 +- .../research/gbml/graph_schema/EdgeType.scala | 8 +-- .../research/gbml/graph_schema/Graph.scala | 2 +- .../gbml/graph_schema/GraphMetadata.scala | 20 +++---- .../gbml/graph_schema/GraphSchemaProto.scala | 2 +- .../research/gbml/graph_schema/Node.scala | 4 +- .../InferenceMetadata.scala | 10 ++-- .../InferenceMetadataProto.scala | 2 +- .../inference_metadata/InferenceOutput.scala | 2 +- .../PostProcessedMetadata.scala | 4 +- .../PostprocessedMetadataProto.scala | 2 +- .../PreprocessedMetadata.scala | 60 +++++++++---------- .../PreprocessedMetadataProto.scala | 2 +- .../GlobalRandomUniformStrategy.scala | 4 +- .../MessagePassingPath.scala | 4 +- .../MessagePassingPathStrategy.scala | 2 +- .../RandomUniform.scala | 6 +- .../RandomWeighted.scala | 6 +- .../SamplingDirection.scala | 8 +-- .../SamplingOp.scala | 8 +-- .../SubgraphSamplingStrategy.scala | 4 +- .../SubgraphSamplingStrategyProto.scala | 2 +- .../subgraph_sampling_strategy/TopK.scala | 6 +- .../UserDefined.scala | 14 ++--- .../TrainedModelMetadata.scala | 10 ++-- .../TrainedModelMetadataProto.scala | 2 +- .../gbml/training_samples_schema/Label.scala | 6 +- .../NodeAnchorBasedLinkPredictionSample.scala | 2 +- .../RootedNodeNeighborhood.scala | 2 +- .../SupervisedLinkBasedTaskSample.scala | 2 +- .../SupervisedNodeClassificationSample.scala | 2 +- .../TrainingSamplesSchemaProto.scala | 2 +- .../src/test/assets/resource_config.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../preprocessed_metadata.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- ...frozen_gbml_config_graphdb_dblp_local.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../preprocessed_metadata.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../lib/assigners/AbstractAssigners.scala | 4 +- .../lib/split_strategies/SplitStrategy.scala | 2 +- ...horBasedSupervisionEdgeSplitStrategy.scala | 2 +- .../scala/libs/task/TaskOutputValidator.scala | 4 +- scala_spark35/.gitignore | 2 +- scala_spark35/.scalafix.conf | 2 +- .../graphdb/nebula/NebulaGraphDBClient.scala | 4 +- .../NebulaQueryResponseTranslator.scala | 22 +++---- .../dataset_metadata/DatasetMetadata.scala | 4 +- .../DatasetMetadataProto.scala | 2 +- ...NodeAnchorBasedLinkPredictionDataset.scala | 38 ++++++------ .../SupervisedLinkBasedTaskSplitDataset.scala | 8 +-- .../SupervisedNodeClassificationDataset.scala | 8 +-- .../FlattenedGraphMetadata.scala | 4 +- .../FlattenedGraphMetadataProto.scala | 2 +- .../NodeAnchorBasedLinkPredictionOutput.scala | 14 ++--- .../SupervisedLinkBasedTaskOutput.scala | 6 +- .../SupervisedNodeClassificationOutput.scala | 6 +- .../gbml/gigl_resource_config/Component.scala | 20 +++---- .../DataPreprocessorConfig.scala | 2 +- .../DistributedInferencerConfig.scala | 4 +- .../KFPTrainerConfig.scala | 12 ++-- .../LocalTrainerConfig.scala | 4 +- .../SparkResourceConfig.scala | 8 +-- .../VertexAiTrainerConfig.scala | 10 ++-- .../research/gbml/graph_schema/Edge.scala | 6 +- .../research/gbml/graph_schema/EdgeType.scala | 8 +-- .../research/gbml/graph_schema/Graph.scala | 2 +- .../gbml/graph_schema/GraphMetadata.scala | 20 +++---- .../gbml/graph_schema/GraphSchemaProto.scala | 2 +- .../research/gbml/graph_schema/Node.scala | 4 +- .../InferenceMetadata.scala | 10 ++-- .../InferenceMetadataProto.scala | 2 +- .../inference_metadata/InferenceOutput.scala | 2 +- .../PostProcessedMetadata.scala | 4 +- .../PostprocessedMetadataProto.scala | 2 +- .../PreprocessedMetadata.scala | 60 +++++++++---------- .../PreprocessedMetadataProto.scala | 2 +- .../GlobalRandomUniformStrategy.scala | 4 +- .../MessagePassingPath.scala | 4 +- .../MessagePassingPathStrategy.scala | 2 +- .../RandomUniform.scala | 6 +- .../RandomWeighted.scala | 6 +- .../SamplingDirection.scala | 8 +-- .../SamplingOp.scala | 8 +-- .../SubgraphSamplingStrategy.scala | 4 +- .../SubgraphSamplingStrategyProto.scala | 2 +- .../subgraph_sampling_strategy/TopK.scala | 6 +- .../UserDefined.scala | 14 ++--- .../TrainedModelMetadata.scala | 10 ++-- .../TrainedModelMetadataProto.scala | 2 +- .../gbml/training_samples_schema/Label.scala | 6 +- .../NodeAnchorBasedLinkPredictionSample.scala | 2 +- .../RootedNodeNeighborhood.scala | 2 +- .../SupervisedLinkBasedTaskSample.scala | 2 +- .../SupervisedNodeClassificationSample.scala | 2 +- .../TrainingSamplesSchemaProto.scala | 2 +- .../userDefinedAggregators/RnnUDAF.scala | 12 ++-- .../src/main/scala/utils/SlottedJoiner.scala | 14 ++--- .../src/test/assets/resource_config.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../preprocessed_metadata.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../frozen_gbml_config.yaml | 2 +- .../userDefinedAggregators/RnnUDAFTest.scala | 8 +-- .../test/scala/utils/SlottedJoinerTest.scala | 10 ++-- .../split_generator/src/main/scala/Main.scala | 2 +- .../lib/assigners/AbstractAssigners.scala | 4 +- .../lib/split_strategies/SplitStrategy.scala | 2 +- ...horBasedSupervisionEdgeSplitStrategy.scala | 2 +- .../scala/libs/task/TaskOutputValidator.scala | 4 +- .../task/pureSparkV2/EgoNetGeneration.scala | 10 ++-- 135 files changed, 445 insertions(+), 446 deletions(-) diff --git a/.github/actions/assert-is-collaborator/action.yml b/.github/actions/assert-is-collaborator/action.yml index cd1e6feb8..2551f5bea 100644 --- a/.github/actions/assert-is-collaborator/action.yml +++ b/.github/actions/assert-is-collaborator/action.yml @@ -5,7 +5,7 @@ inputs: description: The GitHub username to check required: true initiating-pr-number: - description: The PR number that the check may be associated with, if provided will comment on the PR incase of failures + description: The PR number that the check may be associated with, if provided will comment on the PR incase of failures required: false runs: using: "composite" @@ -22,7 +22,7 @@ runs: script: | try { const username = "${{ inputs.username }}"; - const result = await github.rest.repos.checkCollaborator({ + const result = await github.rest.repos.checkCollaborator({ owner: context.repo.owner, repo: context.repo.repo, username: username @@ -41,7 +41,7 @@ runs: console.log(`Error checking collaborator status: ${error.message}`); } } - + - name: Comment workflow permissions if: ${{ failure() && steps.assert-is-collaborator.conclusion == 'failure' && inputs.initiating-pr-number != '' }} uses: snapchat/gigl/.github/actions/comment-on-pr@main @@ -49,4 +49,3 @@ runs: pr_number: ${{ inputs.initiating-pr-number }} message: | 🔒 User ${{ inputs.username }} does not have permissions to run this workflow - \ No newline at end of file diff --git a/.github/actions/comment-on-pr/action.yml b/.github/actions/comment-on-pr/action.yml index 1f7bdf894..9151c2c5d 100644 --- a/.github/actions/comment-on-pr/action.yml +++ b/.github/actions/comment-on-pr/action.yml @@ -16,7 +16,7 @@ outputs: comment_id: description: 'The ID of the created or updated comment' value: ${{steps.comment.outputs.result}} - + runs: using: 'composite' @@ -67,4 +67,4 @@ runs: }); return response.data.id; - } \ No newline at end of file + } diff --git a/.github/actions/get-pr-src-branch/action.yml b/.github/actions/get-pr-src-branch/action.yml index 56c83ae1b..17cf87821 100644 --- a/.github/actions/get-pr-src-branch/action.yml +++ b/.github/actions/get-pr-src-branch/action.yml @@ -30,4 +30,4 @@ runs: }); const branch_name = pr.data.head.ref; console.log("Branch name is:", branch_name); - return branch_name; \ No newline at end of file + return branch_name; diff --git a/proto/snapchat/research/gbml/postprocessed_metadata.proto b/proto/snapchat/research/gbml/postprocessed_metadata.proto index c03b7caed..a093ecbb3 100644 --- a/proto/snapchat/research/gbml/postprocessed_metadata.proto +++ b/proto/snapchat/research/gbml/postprocessed_metadata.proto @@ -5,4 +5,4 @@ package snapchat.research.gbml; message PostProcessedMetadata{ // The path to the post processor evaluation results string post_processor_log_metrics_uri = 1; -} \ No newline at end of file +} diff --git a/proto/snapchat/research/gbml/subgraph_sampling_strategy.proto b/proto/snapchat/research/gbml/subgraph_sampling_strategy.proto index 6f55e457b..19fbc5665 100644 --- a/proto/snapchat/research/gbml/subgraph_sampling_strategy.proto +++ b/proto/snapchat/research/gbml/subgraph_sampling_strategy.proto @@ -4,7 +4,7 @@ package snapchat.research.gbml; import "snapchat/research/gbml/graph_schema.proto"; -message RandomUniform { // Randomly sample nodes from the neighborhood without replacement. +message RandomUniform { // Randomly sample nodes from the neighborhood without replacement. int32 num_nodes_to_sample = 1; } diff --git a/scala/.scalafix.conf b/scala/.scalafix.conf index 195ccd431..9e45edd17 100644 --- a/scala/.scalafix.conf +++ b/scala/.scalafix.conf @@ -1,7 +1,7 @@ rules = [ ExplicitResultTypes NoValInForComprehension, - OrganizeImports, + OrganizeImports, ProcedureSyntax, RedundantSyntax, ] diff --git a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala index 0cafea1ca..afe385a0d 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala @@ -38,7 +38,7 @@ final case class DatasetMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { outputMetadata.supervisedNodeClassificationDataset.foreach { __v => @@ -165,7 +165,7 @@ object DatasetMetadata extends scalapb.GeneratedMessageCompanion[snapchat.resear override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class SupervisedNodeClassificationDataset(value: snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset) extends snapchat.research.gbml.dataset_metadata.DatasetMetadata.OutputMetadata { type ValueType = snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset diff --git a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala index 7f2bfe943..5393b2b91 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala @@ -60,4 +60,4 @@ object DatasetMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala index 63cd8933f..44fd8509f 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala @@ -21,21 +21,21 @@ final case class NodeAnchorBasedLinkPredictionDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainMainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testMainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valMainDataUri if (!__value.isEmpty) { @@ -64,7 +64,7 @@ final case class NodeAnchorBasedLinkPredictionDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -250,14 +250,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -274,7 +274,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -318,7 +318,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry]) } - + object TrainNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry = { @@ -383,7 +383,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry]) } - + @SerialVersionUID(0L) final case class ValNodeTypeToRandomNegativeDataUriEntry( key: _root_.scala.Predef.String = "", @@ -394,14 +394,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -418,7 +418,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -462,7 +462,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry]) } - + object ValNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry = { @@ -527,7 +527,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry]) } - + @SerialVersionUID(0L) final case class TestNodeTypeToRandomNegativeDataUriEntry( key: _root_.scala.Predef.String = "", @@ -538,14 +538,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -562,7 +562,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -606,7 +606,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry]) } - + object TestNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry = { @@ -671,7 +671,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry]) } - + implicit class NodeAnchorBasedLinkPredictionDatasetLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset](_l) { def trainMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.trainMainDataUri)((c_, f_) => c_.copy(trainMainDataUri = f_)) def testMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.testMainDataUri)((c_, f_) => c_.copy(testMainDataUri = f_)) diff --git a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala index 2aca6bc33..453be5b92 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala @@ -18,21 +18,21 @@ final case class SupervisedLinkBasedTaskSplitDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valDataUri if (!__value.isEmpty) { @@ -49,7 +49,7 @@ final case class SupervisedLinkBasedTaskSplitDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala index ac5f71b11..2411d36d2 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala @@ -18,21 +18,21 @@ final case class SupervisedNodeClassificationDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valDataUri if (!__value.isEmpty) { @@ -49,7 +49,7 @@ final case class SupervisedNodeClassificationDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala index 0395d2bd1..6f8ab4d22 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala @@ -38,7 +38,7 @@ final case class FlattenedGraphMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { outputMetadata.supervisedNodeClassificationOutput.foreach { __v => @@ -165,7 +165,7 @@ object FlattenedGraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class SupervisedNodeClassificationOutput(value: snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput) extends snapchat.research.gbml.flattened_graph_metadata.FlattenedGraphMetadata.OutputMetadata { type ValueType = snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput diff --git a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala index b0102ad23..1018c5787 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala @@ -50,4 +50,4 @@ object FlattenedGraphMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala index c31af112a..375313e97 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala @@ -22,7 +22,7 @@ final case class NodeAnchorBasedLinkPredictionOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { @@ -43,7 +43,7 @@ final case class NodeAnchorBasedLinkPredictionOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -153,14 +153,14 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -177,7 +177,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -221,7 +221,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa def companion: snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry.type = snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry]) } - + object NodeTypeToRandomNegativeTfrecordUriPrefixEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry = { @@ -286,7 +286,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry]) } - + implicit class NodeAnchorBasedLinkPredictionOutputLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput](_l) { def tfrecordUriPrefix: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.tfrecordUriPrefix)((c_, f_) => c_.copy(tfrecordUriPrefix = f_)) def nodeTypeToRandomNegativeTfrecordUriPrefix: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, _root_.scala.Predef.String]] = field(_.nodeTypeToRandomNegativeTfrecordUriPrefix)((c_, f_) => c_.copy(nodeTypeToRandomNegativeTfrecordUriPrefix = f_)) diff --git a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala index 8cfd94948..7af22ab07 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala @@ -20,14 +20,14 @@ final case class SupervisedLinkBasedTaskOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labeledTfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = unlabeledTfrecordUriPrefix if (!__value.isEmpty) { @@ -44,7 +44,7 @@ final case class SupervisedLinkBasedTaskOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala index cb3ed0bb3..538828f18 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala @@ -20,14 +20,14 @@ final case class SupervisedNodeClassificationOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labeledTfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = unlabeledTfrecordUriPrefix if (!__value.isEmpty) { @@ -44,7 +44,7 @@ final case class SupervisedNodeClassificationOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala index 871d66432..849972498 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala @@ -24,63 +24,63 @@ sealed abstract class Component(val value: _root_.scala.Int) extends _root_.scal object Component extends _root_.scalapb.GeneratedEnumCompanion[Component] { sealed trait Recognized extends Component implicit def enumCompanion: _root_.scalapb.GeneratedEnumCompanion[Component] = this - + @SerialVersionUID(0L) case object Component_Unknown extends Component(0) with Component.Recognized { val index = 0 val name = "Component_Unknown" override def isComponentUnknown: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Config_Validator extends Component(1) with Component.Recognized { val index = 1 val name = "Component_Config_Validator" override def isComponentConfigValidator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Config_Populator extends Component(2) with Component.Recognized { val index = 2 val name = "Component_Config_Populator" override def isComponentConfigPopulator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Data_Preprocessor extends Component(3) with Component.Recognized { val index = 3 val name = "Component_Data_Preprocessor" override def isComponentDataPreprocessor: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Subgraph_Sampler extends Component(4) with Component.Recognized { val index = 4 val name = "Component_Subgraph_Sampler" override def isComponentSubgraphSampler: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Split_Generator extends Component(5) with Component.Recognized { val index = 5 val name = "Component_Split_Generator" override def isComponentSplitGenerator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Trainer extends Component(6) with Component.Recognized { val index = 6 val name = "Component_Trainer" override def isComponentTrainer: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Inferencer extends Component(7) with Component.Recognized { val index = 7 val name = "Component_Inferencer" override def isComponentInferencer: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) final case class Unrecognized(unrecognizedValue: _root_.scala.Int) extends Component(unrecognizedValue) with _root_.scalapb.UnrecognizedEnum lazy val values = scala.collection.immutable.Seq(Component_Unknown, Component_Config_Validator, Component_Config_Populator, Component_Data_Preprocessor, Component_Subgraph_Sampler, Component_Split_Generator, Component_Trainer, Component_Inferencer) @@ -97,4 +97,4 @@ object Component extends _root_.scalapb.GeneratedEnumCompanion[Component] { } def javaDescriptor: _root_.com.google.protobuf.Descriptors.EnumDescriptor = GiglResourceConfigProto.javaDescriptor.getEnumTypes().get(0) def scalaDescriptor: _root_.scalapb.descriptors.EnumDescriptor = GiglResourceConfigProto.scalaDescriptor.enums(0) -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala index 75c2e54af..7d9951c80 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala @@ -35,7 +35,7 @@ final case class DataPreprocessorConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { edgePreprocessorConfig.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala index 8363bdb1f..2198a2eb5 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala @@ -38,7 +38,7 @@ final case class DistributedInferencerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { trainerConfig.vertexAiInferencerConfig.foreach { __v => @@ -165,7 +165,7 @@ object DistributedInferencerConfig extends scalapb.GeneratedMessageCompanion[sna override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class VertexAiInferencerConfig(value: snapchat.research.gbml.gigl_resource_config.VertexAiResourceConfig) extends snapchat.research.gbml.gigl_resource_config.DistributedInferencerConfig.TrainerConfig { type ValueType = snapchat.research.gbml.gigl_resource_config.VertexAiResourceConfig diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala index 909ec979b..1225ba210 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala @@ -32,35 +32,35 @@ final case class KFPTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = cpuRequest if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = memoryRequest if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = gpuType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = gpuLimit if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(4, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -77,7 +77,7 @@ final case class KFPTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala index ba2cc9389..86e238074 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala @@ -17,7 +17,7 @@ final case class LocalTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numWorkers if (__value != 0) { @@ -34,7 +34,7 @@ final case class LocalTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala index d32c915cb..96f354f47 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala @@ -25,21 +25,21 @@ final case class SparkResourceConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = machineType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = numLocalSsds if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(2, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -56,7 +56,7 @@ final case class SparkResourceConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala index c088bafc2..37d730799 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala @@ -29,28 +29,28 @@ final case class VertexAiTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = machineType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = gpuType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = gpuLimit if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(3, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -67,7 +67,7 @@ final case class VertexAiTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala index cd1501b1f..dc359d1e3 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala @@ -33,14 +33,14 @@ final case class Edge( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = srcNodeId if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(1, __value) } }; - + { val __value = dstNodeId if (__value != 0) { @@ -65,7 +65,7 @@ final case class Edge( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala index d2a68d8b9..439ae2229 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala @@ -21,21 +21,21 @@ final case class EdgeType( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = relation if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = srcNodeType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = dstNodeType if (!__value.isEmpty) { @@ -52,7 +52,7 @@ final case class EdgeType( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala index 0e4e3105d..7c317bb6d 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala @@ -35,7 +35,7 @@ final case class Graph( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodes.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala index 8c6307580..856a159a0 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala @@ -58,7 +58,7 @@ final case class GraphMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodeTypes.foreach { __v => @@ -205,7 +205,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -226,7 +226,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -269,7 +269,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research def companion: snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry.type = snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.GraphMetadata.CondensedEdgeTypeMapEntry]) } - + object CondensedEdgeTypeMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry = { @@ -341,7 +341,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.GraphMetadata.CondensedEdgeTypeMapEntry]) } - + @SerialVersionUID(0L) final case class CondensedNodeTypeMapEntry( key: _root_.scala.Int = 0, @@ -352,14 +352,14 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -376,7 +376,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -420,7 +420,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research def companion: snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry.type = snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.GraphMetadata.CondensedNodeTypeMapEntry]) } - + object CondensedNodeTypeMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry = { @@ -485,7 +485,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.GraphMetadata.CondensedNodeTypeMapEntry]) } - + implicit class GraphMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.graph_schema.GraphMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.graph_schema.GraphMetadata](_l) { def nodeTypes: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Seq[_root_.scala.Predef.String]] = field(_.nodeTypes)((c_, f_) => c_.copy(nodeTypes = f_)) def edgeTypes: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Seq[snapchat.research.gbml.graph_schema.EdgeType]] = field(_.edgeTypes)((c_, f_) => c_.copy(edgeTypes = f_)) diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala index 28c2eff67..0cd9e60ca 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala @@ -50,4 +50,4 @@ object GraphSchemaProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala index e8a4c6f98..7e4c5c8b9 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala @@ -27,7 +27,7 @@ final case class Node( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = nodeId if (__value != 0) { @@ -52,7 +52,7 @@ final case class Node( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala index 6668d67ce..a58d891c0 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala @@ -31,7 +31,7 @@ final case class InferenceMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodeTypeToInferencerOutputInfoMap.foreach { __v => @@ -123,7 +123,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { @@ -144,7 +144,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -187,7 +187,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese def companion: snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry.type = snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry]) } - + object NodeTypeToInferencerOutputInfoMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry = { @@ -259,7 +259,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry]) } - + implicit class InferenceMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.inference_metadata.InferenceMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.inference_metadata.InferenceMetadata](_l) { def nodeTypeToInferencerOutputInfoMap: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, snapchat.research.gbml.inference_metadata.InferenceOutput]] = field(_.nodeTypeToInferencerOutputInfoMap)((c_, f_) => c_.copy(nodeTypeToInferencerOutputInfoMap = f_)) } diff --git a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala index 7e1424d6e..9c335401b 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala @@ -35,4 +35,4 @@ object InferenceMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala index 845fbb29c..66c68b47b 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala @@ -38,7 +38,7 @@ final case class InferenceOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { embeddingsPath.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala index a0399d909..8ebed0a48 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala @@ -17,7 +17,7 @@ final case class PostProcessedMetadata( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = postProcessorLogMetricsUri if (!__value.isEmpty) { @@ -34,7 +34,7 @@ final case class PostProcessedMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala index 4b0e94597..d36401685 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala @@ -28,4 +28,4 @@ object PostprocessedMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala index 6a160cdd4..80160636b 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala @@ -38,7 +38,7 @@ final case class PreprocessedMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { condensedNodeTypeToPreprocessedMetadata.foreach { __v => @@ -181,7 +181,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = nodeIdKey if (!__value.isEmpty) { @@ -196,28 +196,28 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = __item __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(4, __value) } }; - + { val __value = schemaUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(5, __value) } }; - + { val __value = enumeratedNodeIdsBqTable if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(6, __value) } }; - + { val __value = enumeratedNodeDataBqTable if (!__value.isEmpty) { @@ -228,7 +228,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = featureDim.get __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(8, __value) }; - + { val __value = transformFnAssetsUri if (!__value.isEmpty) { @@ -245,7 +245,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -366,7 +366,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.NodeMetadataOutput]) } - + object NodeMetadataOutput extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput = { @@ -499,7 +499,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.NodeMetadataOutput]) } - + /** Houses metadata of edge features output from DataPreprocessor * * @param featureKeys @@ -540,21 +540,21 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = __item __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = schemaUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(4, __value) } }; - + { val __value = enumeratedEdgeDataBqTable if (!__value.isEmpty) { @@ -565,7 +565,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = featureDim.get __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(6, __value) }; - + { val __value = transformFnAssetsUri if (!__value.isEmpty) { @@ -582,7 +582,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { featureKeys.foreach { __v => @@ -679,7 +679,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataInfo]) } - + object EdgeMetadataInfo extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo = { @@ -792,7 +792,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataInfo]) } - + /** Houses metadata about edge TFTransform output from DataPreprocessor. * * @param srcNodeIdKey @@ -819,14 +819,14 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = srcNodeIdKey if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = dstNodeIdKey if (!__value.isEmpty) { @@ -855,7 +855,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -932,7 +932,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataOutput]) } - + object EdgeMetadataOutput extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput = { @@ -1035,7 +1035,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataOutput]) } - + @SerialVersionUID(0L) final case class CondensedNodeTypeToPreprocessedMetadataEntry( key: _root_.scala.Int = 0, @@ -1046,7 +1046,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -1067,7 +1067,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -1110,7 +1110,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry]) } - + object CondensedNodeTypeToPreprocessedMetadataEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry = { @@ -1182,7 +1182,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry]) } - + @SerialVersionUID(0L) final case class CondensedEdgeTypeToPreprocessedMetadataEntry( key: _root_.scala.Int = 0, @@ -1193,7 +1193,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -1214,7 +1214,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -1257,7 +1257,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry]) } - + object CondensedEdgeTypeToPreprocessedMetadataEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry = { @@ -1329,7 +1329,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry]) } - + implicit class PreprocessedMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata](_l) { def condensedNodeTypeToPreprocessedMetadata: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Int, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput]] = field(_.condensedNodeTypeToPreprocessedMetadata)((c_, f_) => c_.copy(condensedNodeTypeToPreprocessedMetadata = f_)) def condensedEdgeTypeToPreprocessedMetadata: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Int, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput]] = field(_.condensedEdgeTypeToPreprocessedMetadata)((c_, f_) => c_.copy(condensedEdgeTypeToPreprocessedMetadata = f_)) diff --git a/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala index b6e8d0d6d..becc2d068 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala @@ -61,4 +61,4 @@ object PreprocessedMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala index c56f47fa4..2a6227235 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala @@ -15,7 +15,7 @@ final case class GlobalRandomUniformStrategy( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numHops if (__value != 0) { @@ -36,7 +36,7 @@ final case class GlobalRandomUniformStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala index dc3069c8b..4aad9e19c 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala @@ -15,7 +15,7 @@ final case class MessagePassingPath( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = rootNodeType if (!__value.isEmpty) { @@ -36,7 +36,7 @@ final case class MessagePassingPath( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala index fd8a906af..ee3aa77a0 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala @@ -34,7 +34,7 @@ final case class MessagePassingPathStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { paths.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala index 5579eca5a..ef0fde958 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala @@ -5,7 +5,7 @@ package snapchat.research.gbml.subgraph_sampling_strategy -/** Randomly sample nodes from the neighborhood without replacement. +/** Randomly sample nodes from the neighborhood without replacement. */ @SerialVersionUID(0L) final case class RandomUniform( @@ -16,7 +16,7 @@ final case class RandomUniform( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { @@ -33,7 +33,7 @@ final case class RandomUniform( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala index 389735b80..395f3876b 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala @@ -17,14 +17,14 @@ final case class RandomWeighted( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeInt32Size(1, __value) } }; - + { val __value = edgeFeatName if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class RandomWeighted( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala index 0563e9b71..8f521cf2e 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala @@ -16,7 +16,7 @@ sealed abstract class SamplingDirection(val value: _root_.scala.Int) extends _ro object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingDirection] { sealed trait Recognized extends SamplingDirection implicit def enumCompanion: _root_.scalapb.GeneratedEnumCompanion[SamplingDirection] = this - + /** Sample incoming edges to the dst nodes (default) */ @SerialVersionUID(0L) @@ -25,7 +25,7 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD val name = "INCOMING" override def isIncoming: _root_.scala.Boolean = true } - + /** Sample outgoing edges from the src nodes */ @SerialVersionUID(0L) @@ -34,7 +34,7 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD val name = "OUTGOING" override def isOutgoing: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) final case class Unrecognized(unrecognizedValue: _root_.scala.Int) extends SamplingDirection(unrecognizedValue) with _root_.scalapb.UnrecognizedEnum lazy val values = scala.collection.immutable.Seq(INCOMING, OUTGOING) @@ -45,4 +45,4 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD } def javaDescriptor: _root_.com.google.protobuf.Descriptors.EnumDescriptor = SubgraphSamplingStrategyProto.javaDescriptor.getEnumTypes().get(0) def scalaDescriptor: _root_.scalapb.descriptors.EnumDescriptor = SubgraphSamplingStrategyProto.scalaDescriptor.enums(0) -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala index 513ceaf83..6afa2e8b6 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala @@ -28,7 +28,7 @@ final case class SamplingOp( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = opName if (!__value.isEmpty) { @@ -59,7 +59,7 @@ final case class SamplingOp( val __value = samplingMethod.userDefined.get __size += 2 + _root_.com.google.protobuf.CodedOutputStream.computeUInt32SizeNoTag(__value.serializedSize) + __value.serializedSize }; - + { val __value = samplingDirection.value if (__value != 0) { @@ -76,7 +76,7 @@ final case class SamplingOp( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -293,7 +293,7 @@ object SamplingOp extends scalapb.GeneratedMessageCompanion[snapchat.research.gb override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class RandomUniform(value: snapchat.research.gbml.subgraph_sampling_strategy.RandomUniform) extends snapchat.research.gbml.subgraph_sampling_strategy.SamplingOp.SamplingMethod { type ValueType = snapchat.research.gbml.subgraph_sampling_strategy.RandomUniform diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala index 5f8e3dee6..4abc7182b 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala @@ -32,7 +32,7 @@ final case class SubgraphSamplingStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { strategy.messagePassingPaths.foreach { __v => @@ -143,7 +143,7 @@ object SubgraphSamplingStrategy extends scalapb.GeneratedMessageCompanion[snapch override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class MessagePassingPaths(value: snapchat.research.gbml.subgraph_sampling_strategy.MessagePassingPathStrategy) extends snapchat.research.gbml.subgraph_sampling_strategy.SubgraphSamplingStrategy.Strategy { type ValueType = snapchat.research.gbml.subgraph_sampling_strategy.MessagePassingPathStrategy diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala index 709ae1159..e6ae31c4c 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala @@ -67,4 +67,4 @@ object SubgraphSamplingStrategyProto extends _root_.scalapb.GeneratedFileObject } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala index 86bda274e..39c5388e4 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala @@ -17,14 +17,14 @@ final case class TopK( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeInt32Size(1, __value) } }; - + { val __value = edgeFeatName if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class TopK( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala index a055bfbaa..cc00d0f9d 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala @@ -20,7 +20,7 @@ final case class UserDefined( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = pathToUdf if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class UserDefined( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -151,14 +151,14 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -175,7 +175,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -219,7 +219,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g def companion: snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry.type = snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.UserDefined.ParamsEntry]) } - + object ParamsEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry = { @@ -284,7 +284,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.UserDefined.ParamsEntry]) } - + implicit class UserDefinedLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.subgraph_sampling_strategy.UserDefined]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.subgraph_sampling_strategy.UserDefined](_l) { def pathToUdf: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.pathToUdf)((c_, f_) => c_.copy(pathToUdf = f_)) def params: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, _root_.scala.Predef.String]] = field(_.params)((c_, f_) => c_.copy(params = f_)) diff --git a/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala b/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala index bcf95c046..2c5a042f9 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala @@ -26,28 +26,28 @@ final case class TrainedModelMetadata( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainedModelUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = scriptedModelUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = evalMetricsUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = tensorboardLogsUri if (!__value.isEmpty) { @@ -64,7 +64,7 @@ final case class TrainedModelMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala index b06e0d55a..1262d9517 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala @@ -31,4 +31,4 @@ object TrainedModelMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala index 21884234b..3289f2d76 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala @@ -15,14 +15,14 @@ final case class Label( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labelType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = label if (__value != 0) { @@ -39,7 +39,7 @@ final case class Label( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala index a05a6ef18..a1f7def5a 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala @@ -64,7 +64,7 @@ final case class NodeAnchorBasedLinkPredictionSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala index f326f0375..cdd1d4a4c 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala @@ -41,7 +41,7 @@ final case class RootedNodeNeighborhood( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala index 4ffbeb328..9707ea68a 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala @@ -48,7 +48,7 @@ final case class SupervisedLinkBasedTaskSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootEdge.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala index 35c594ca3..00304a823 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala @@ -43,7 +43,7 @@ final case class SupervisedNodeClassificationSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala index f4187ea60..f02891e26 100644 --- a/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala +++ b/scala/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala @@ -51,4 +51,4 @@ object TrainingSamplesSchemaProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala/common/src/test/assets/resource_config.yaml b/scala/common/src/test/assets/resource_config.yaml index 5d0979b72..702b9f917 100644 --- a/scala/common/src/test/assets/resource_config.yaml +++ b/scala/common/src/test/assets/resource_config.yaml @@ -42,4 +42,4 @@ inferencer_config: num_workers: 1 max_num_workers: 256 machine_type: "c3-standard-22" - disk_size_gb: 100 \ No newline at end of file + disk_size_gb: 100 diff --git a/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml b/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml index b1a1421cd..a0a2ba944 100644 --- a/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml +++ b/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml @@ -44,4 +44,4 @@ graphMetadata: relation: engage srcNodeType: user nodeTypes: - - user \ No newline at end of file + - user diff --git a/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml b/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml index 429035cbb..f44cdbb44 100644 --- a/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml +++ b/scala/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml @@ -15,4 +15,4 @@ condensedNodeTypeToPreprocessedMetadata: - f1 nodeIdKey: node_id schemaUri: not.used.for.test - tfrecordUriPrefix: not.used.for.test \ No newline at end of file + tfrecordUriPrefix: not.used.for.test diff --git a/scala/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml b/scala/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml index 90762002c..68711a211 100644 --- a/scala/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml +++ b/scala/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml @@ -28,4 +28,4 @@ sharedConfig: supervisedNodeClassificationOutput: labeledTfrecordUriPrefix: common/src/test/assets/split_generator/supervised_node_classification/sgs_output/labeled/samples/ unlabeledTfrecordUriPrefix: common/src/test/assets/split_generator/supervised_node_classification/sgs_output/unlabeled/samples/ - preprocessedMetadataUri: common/src/test/assets/split_generator/supervised_node_classification/preprocessed_metadata.yaml \ No newline at end of file + preprocessedMetadataUri: common/src/test/assets/split_generator/supervised_node_classification/preprocessed_metadata.yaml diff --git a/scala/common/src/test/assets/subgraph_sampler/heterogeneous/node_anchor_based_link_prediction/frozen_gbml_config_graphdb_dblp_local.yaml b/scala/common/src/test/assets/subgraph_sampler/heterogeneous/node_anchor_based_link_prediction/frozen_gbml_config_graphdb_dblp_local.yaml index 02a166cf1..6fea68ba6 100755 --- a/scala/common/src/test/assets/subgraph_sampler/heterogeneous/node_anchor_based_link_prediction/frozen_gbml_config_graphdb_dblp_local.yaml +++ b/scala/common/src/test/assets/subgraph_sampler/heterogeneous/node_anchor_based_link_prediction/frozen_gbml_config_graphdb_dblp_local.yaml @@ -50,4 +50,4 @@ graphMetadata: srcNodeType: paper nodeTypes: - author - - paper \ No newline at end of file + - paper diff --git a/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/frozen_gbml_config.yaml b/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/frozen_gbml_config.yaml index 404c5b0f9..c2cde89f6 100755 --- a/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/frozen_gbml_config.yaml +++ b/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/frozen_gbml_config.yaml @@ -37,4 +37,4 @@ graphMetadata: relation: friend srcNodeType: user nodeTypes: - - user \ No newline at end of file + - user diff --git a/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/preprocessed_metadata.yaml b/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/preprocessed_metadata.yaml index a9ed15158..0102fe86c 100755 --- a/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/preprocessed_metadata.yaml +++ b/scala/common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/preprocessed_metadata.yaml @@ -31,4 +31,4 @@ condensedNodeTypeToPreprocessedMetadata: - f1 nodeIdKey: node_id schemaUri: not.used.for.test - tfrecordUriPrefix: common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/node_data \ No newline at end of file + tfrecordUriPrefix: common/src/test/assets/subgraph_sampler/node_anchor_based_link_prediction/node_data diff --git a/scala/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml b/scala/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml index cb1587a83..798350dd0 100755 --- a/scala/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml +++ b/scala/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml @@ -24,4 +24,4 @@ sharedConfig: supervisedNodeClassificationOutput: labeledTfrecordUriPrefix: common/src/test/assets/subgraph_sampler/supervised_node_classification/output/labeled/samples/ unlabeledTfrecordUriPrefix: common/src/test/assets/subgraph_sampler/supervised_node_classification/output/unlabeled/samples/ - preprocessedMetadataUri: common/src/test/assets/subgraph_sampler/supervised_node_classification/preprocessed_metadata.yaml \ No newline at end of file + preprocessedMetadataUri: common/src/test/assets/subgraph_sampler/supervised_node_classification/preprocessed_metadata.yaml diff --git a/scala/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala b/scala/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala index 90524f6a0..891198953 100644 --- a/scala/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala +++ b/scala/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala @@ -22,7 +22,7 @@ object AbstractAssigners { * e.g. could be assigning a NodePb (T) to some Enum (S). * * @param obj the object to hash - * @return + * @return */ def assign(obj: T): S } @@ -59,7 +59,7 @@ object AbstractAssigners { /** * Relative width of each bucket in the hash space. e.g. [0.2, 0.4, 0.4] would indicate 3 buckets, where - * the second and third bucket are twice as prominent as the first bucket. + * the second and third bucket are twice as prominent as the first bucket. */ lazy val weights: Seq[Float] = bucketWeights.values.toList diff --git a/scala/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala b/scala/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala index ba15966a3..b0d7e1146 100644 --- a/scala/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala +++ b/scala/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala @@ -32,7 +32,7 @@ abstract class SplitStrategy[A](splitStrategyArgs: Map[String, String]) extends val graphMetadataPbWrapper: GraphMetadataPbWrapper /** - * Takes in a single "un-split" training sample instance output by SubgraphSampler, + * Takes in a single "un-split" training sample instance output by SubgraphSampler, * and a DatasetSplit(TRAIN, TEST, VAL) and outputs the the "split" samples for that dataset split * * @param sample : Input Sample from SGS diff --git a/scala/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala b/scala/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala index 61fefef6f..12e378ebe 100644 --- a/scala/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala +++ b/scala/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala @@ -37,7 +37,7 @@ class UDLAnchorBasedSupervisionEdgeSplitStrategy( * (a) All pos_edges and hard_neg_edges belonging to the split. * (b) message passing structure which should be pb.neighborhood and therefore the same across all splits * (i.e. no masking). - * (c) The message passing structure may be filtered down to only include edges that are not in the pos_edges + * (c) The message passing structure may be filtered down to only include edges that are not in the pos_edges * and hard_neg_edges. * An output train-split sample needs to have >0 pos_edges in this setting for loss computation. * Output val/test-split samples may have 0 pos_edges (and even 0 hard_neg_edges), since these diff --git a/scala/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala b/scala/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala index f50bb5d0f..8327695b3 100644 --- a/scala/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala +++ b/scala/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala @@ -17,7 +17,7 @@ object TaskOutputValidator { * is present in the neighborhood nodes. * This method does a dataset.map() on the final output produced by SGS and returns the same dataset * if there is no validation failure. Raises and excpetion if there is some error - * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any + * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any * duplication of computation due to this validation code. * * @param mainSampleDS @@ -50,7 +50,7 @@ object TaskOutputValidator { * is present in the neighborhood nodes. * This method does a dataset.map() on the final output produced by SGS and returns the same dataset * if there is no validation failure. Raises and excpetion if there is some error - * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any + * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any * duplication of computation due to this validation code. * * @param mainSampleDS diff --git a/scala_spark35/.gitignore b/scala_spark35/.gitignore index f7d538fde..3c3cd13b5 100644 --- a/scala_spark35/.gitignore +++ b/scala_spark35/.gitignore @@ -267,4 +267,4 @@ spark-warehouse/ .metals/ .bloop/ .ammonite/ -metals.sbt \ No newline at end of file +metals.sbt diff --git a/scala_spark35/.scalafix.conf b/scala_spark35/.scalafix.conf index 195ccd431..9e45edd17 100644 --- a/scala_spark35/.scalafix.conf +++ b/scala_spark35/.scalafix.conf @@ -1,7 +1,7 @@ rules = [ ExplicitResultTypes NoValInForComprehension, - OrganizeImports, + OrganizeImports, ProcedureSyntax, RedundantSyntax, ] diff --git a/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaGraphDBClient.scala b/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaGraphDBClient.scala index 79cdf9889..449ddd653 100644 --- a/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaGraphDBClient.scala +++ b/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaGraphDBClient.scala @@ -27,8 +27,8 @@ import scala.collection.JavaConversions._ To use SessionPool, you must config the graph space to connect for SessionPool. The SessionPool is thread-safe, and support retry(release old session and get available session from SessionPool) for both connection error, session error and execution error(caused by bad storaged server), and the retry mechanism needs users to config retryTimes and intervalTime between retrys. - - + + This class needs to be serializable if defined outside of mapPartitions, nebula client however is has underlying classes that are not serializable ConnectionPool + getSession - java.io.NotSerializableException: com.vesoft.nebula.client.graph.net.RoundRobinLoadBalancer SessionPool - Task not serializable: java.io.NotSerializableException: java.util.concurrent.ScheduledThreadPoolExecutor diff --git a/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaQueryResponseTranslator.scala b/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaQueryResponseTranslator.scala index 61731262b..769f432a5 100644 --- a/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaQueryResponseTranslator.scala +++ b/scala_spark35/common/src/main/scala/graphdb/nebula/NebulaQueryResponseTranslator.scala @@ -60,7 +60,7 @@ class NebulaQueryResponseTranslator( case SamplingOp.SamplingMethod.RandomUniform(value) => { val numNodesToSample = value.numNodesToSample s"""GO 1 STEP - FROM ${nebulaVID} + FROM ${nebulaVID} OVER ${nebulaEdgeType} ${outgoingEdgesModifier} YIELD src(edge) as ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, @@ -74,33 +74,33 @@ class NebulaQueryResponseTranslator( case SamplingOp.SamplingMethod.RandomWeighted(value) => { val numNodesToSample = value.numNodesToSample val edgeFeatName = value.edgeFeatName - s"""GO 1 STEP - FROM ${nebulaVID} + s"""GO 1 STEP + FROM ${nebulaVID} OVER ${nebulaEdgeType} ${outgoingEdgesModifier} - YIELD + YIELD src(edge) as ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, dst(edge) as ${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME}, ${nebulaEdgeType}.${edgeFeatName} * rand() as ${edgeFeatName} | ORDER BY $$-.${edgeFeatName} DESC | LIMIT ${numNodesToSample} | - YIELD - $$-.${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, + YIELD + $$-.${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, $$-.${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME}""" } case SamplingOp.SamplingMethod.TopK(value) => { val numNodesToSample = value.numNodesToSample val edgeFeatName = value.edgeFeatName - s"""GO 1 STEP - FROM ${nebulaVID} + s"""GO 1 STEP + FROM ${nebulaVID} OVER ${nebulaEdgeType} ${outgoingEdgesModifier} - YIELD + YIELD src(edge) as ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, dst(edge) as ${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME}, ${nebulaEdgeType}.${edgeFeatName} as ${edgeFeatName} | ORDER BY $$-.${edgeFeatName} DESC | LIMIT ${numNodesToSample} | - YIELD - $$-.${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, + YIELD + $$-.${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_SRC_NODE_ID_COL_NAME}, $$-.${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME} AS ${NebulaQueryResponseTranslator.RESULT_DST_NODE_ID_COL_NAME}""" } case SamplingOp.SamplingMethod.UserDefined(value) => { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala index 0cafea1ca..afe385a0d 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadata.scala @@ -38,7 +38,7 @@ final case class DatasetMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { outputMetadata.supervisedNodeClassificationDataset.foreach { __v => @@ -165,7 +165,7 @@ object DatasetMetadata extends scalapb.GeneratedMessageCompanion[snapchat.resear override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class SupervisedNodeClassificationDataset(value: snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset) extends snapchat.research.gbml.dataset_metadata.DatasetMetadata.OutputMetadata { type ValueType = snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala index 7f2bfe943..5393b2b91 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/DatasetMetadataProto.scala @@ -60,4 +60,4 @@ object DatasetMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala index 63cd8933f..44fd8509f 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/NodeAnchorBasedLinkPredictionDataset.scala @@ -21,21 +21,21 @@ final case class NodeAnchorBasedLinkPredictionDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainMainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testMainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valMainDataUri if (!__value.isEmpty) { @@ -64,7 +64,7 @@ final case class NodeAnchorBasedLinkPredictionDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -250,14 +250,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -274,7 +274,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -318,7 +318,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry]) } - + object TrainNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry = { @@ -383,7 +383,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry]) } - + @SerialVersionUID(0L) final case class ValNodeTypeToRandomNegativeDataUriEntry( key: _root_.scala.Predef.String = "", @@ -394,14 +394,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -418,7 +418,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -462,7 +462,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry]) } - + object ValNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry = { @@ -527,7 +527,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry]) } - + @SerialVersionUID(0L) final case class TestNodeTypeToRandomNegativeDataUriEntry( key: _root_.scala.Predef.String = "", @@ -538,14 +538,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -562,7 +562,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -606,7 +606,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry]) } - + object TestNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry = { @@ -671,7 +671,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry]) } - + implicit class NodeAnchorBasedLinkPredictionDatasetLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset](_l) { def trainMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.trainMainDataUri)((c_, f_) => c_.copy(trainMainDataUri = f_)) def testMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.testMainDataUri)((c_, f_) => c_.copy(testMainDataUri = f_)) diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala index 2aca6bc33..453be5b92 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedLinkBasedTaskSplitDataset.scala @@ -18,21 +18,21 @@ final case class SupervisedLinkBasedTaskSplitDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valDataUri if (!__value.isEmpty) { @@ -49,7 +49,7 @@ final case class SupervisedLinkBasedTaskSplitDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala index ac5f71b11..2411d36d2 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/dataset_metadata/SupervisedNodeClassificationDataset.scala @@ -18,21 +18,21 @@ final case class SupervisedNodeClassificationDataset( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = testDataUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = valDataUri if (!__value.isEmpty) { @@ -49,7 +49,7 @@ final case class SupervisedNodeClassificationDataset( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala index 0395d2bd1..6f8ab4d22 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadata.scala @@ -38,7 +38,7 @@ final case class FlattenedGraphMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { outputMetadata.supervisedNodeClassificationOutput.foreach { __v => @@ -165,7 +165,7 @@ object FlattenedGraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class SupervisedNodeClassificationOutput(value: snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput) extends snapchat.research.gbml.flattened_graph_metadata.FlattenedGraphMetadata.OutputMetadata { type ValueType = snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala index b0102ad23..1018c5787 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/FlattenedGraphMetadataProto.scala @@ -50,4 +50,4 @@ object FlattenedGraphMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala index c31af112a..375313e97 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/NodeAnchorBasedLinkPredictionOutput.scala @@ -22,7 +22,7 @@ final case class NodeAnchorBasedLinkPredictionOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { @@ -43,7 +43,7 @@ final case class NodeAnchorBasedLinkPredictionOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -153,14 +153,14 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -177,7 +177,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -221,7 +221,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa def companion: snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry.type = snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry]) } - + object NodeTypeToRandomNegativeTfrecordUriPrefixEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry = { @@ -286,7 +286,7 @@ object NodeAnchorBasedLinkPredictionOutput extends scalapb.GeneratedMessageCompa ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionOutput.NodeTypeToRandomNegativeTfrecordUriPrefixEntry]) } - + implicit class NodeAnchorBasedLinkPredictionOutputLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.flattened_graph_metadata.NodeAnchorBasedLinkPredictionOutput](_l) { def tfrecordUriPrefix: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.tfrecordUriPrefix)((c_, f_) => c_.copy(tfrecordUriPrefix = f_)) def nodeTypeToRandomNegativeTfrecordUriPrefix: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, _root_.scala.Predef.String]] = field(_.nodeTypeToRandomNegativeTfrecordUriPrefix)((c_, f_) => c_.copy(nodeTypeToRandomNegativeTfrecordUriPrefix = f_)) diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala index 8cfd94948..7af22ab07 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedLinkBasedTaskOutput.scala @@ -20,14 +20,14 @@ final case class SupervisedLinkBasedTaskOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labeledTfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = unlabeledTfrecordUriPrefix if (!__value.isEmpty) { @@ -44,7 +44,7 @@ final case class SupervisedLinkBasedTaskOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala index cb3ed0bb3..538828f18 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/flattened_graph_metadata/SupervisedNodeClassificationOutput.scala @@ -20,14 +20,14 @@ final case class SupervisedNodeClassificationOutput( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labeledTfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = unlabeledTfrecordUriPrefix if (!__value.isEmpty) { @@ -44,7 +44,7 @@ final case class SupervisedNodeClassificationOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala index d61aa9295..3169265e7 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/Component.scala @@ -25,63 +25,63 @@ sealed abstract class Component(val value: _root_.scala.Int) extends _root_.scal object Component extends _root_.scalapb.GeneratedEnumCompanion[Component] { sealed trait Recognized extends Component implicit def enumCompanion: _root_.scalapb.GeneratedEnumCompanion[Component] = this - + @SerialVersionUID(0L) case object Component_Unknown extends Component(0) with Component.Recognized { val index = 0 val name = "Component_Unknown" override def isComponentUnknown: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Config_Validator extends Component(1) with Component.Recognized { val index = 1 val name = "Component_Config_Validator" override def isComponentConfigValidator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Config_Populator extends Component(2) with Component.Recognized { val index = 2 val name = "Component_Config_Populator" override def isComponentConfigPopulator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Data_Preprocessor extends Component(3) with Component.Recognized { val index = 3 val name = "Component_Data_Preprocessor" override def isComponentDataPreprocessor: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Subgraph_Sampler extends Component(4) with Component.Recognized { val index = 4 val name = "Component_Subgraph_Sampler" override def isComponentSubgraphSampler: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Split_Generator extends Component(5) with Component.Recognized { val index = 5 val name = "Component_Split_Generator" override def isComponentSplitGenerator: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Trainer extends Component(6) with Component.Recognized { val index = 6 val name = "Component_Trainer" override def isComponentTrainer: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) case object Component_Inferencer extends Component(7) with Component.Recognized { val index = 7 val name = "Component_Inferencer" override def isComponentInferencer: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) final case class Unrecognized(unrecognizedValue: _root_.scala.Int) extends Component(unrecognizedValue) with _root_.scalapb.UnrecognizedEnum lazy val values: scala.collection.immutable.Seq[ValueType] = scala.collection.immutable.Seq(Component_Unknown, Component_Config_Validator, Component_Config_Populator, Component_Data_Preprocessor, Component_Subgraph_Sampler, Component_Split_Generator, Component_Trainer, Component_Inferencer) @@ -98,4 +98,4 @@ object Component extends _root_.scalapb.GeneratedEnumCompanion[Component] { } def javaDescriptor: _root_.com.google.protobuf.Descriptors.EnumDescriptor = GiglResourceConfigProto.javaDescriptor.getEnumTypes().get(0) def scalaDescriptor: _root_.scalapb.descriptors.EnumDescriptor = GiglResourceConfigProto.scalaDescriptor.enums(0) -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala index 75c2e54af..7d9951c80 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DataPreprocessorConfig.scala @@ -35,7 +35,7 @@ final case class DataPreprocessorConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { edgePreprocessorConfig.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala index 8363bdb1f..2198a2eb5 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/DistributedInferencerConfig.scala @@ -38,7 +38,7 @@ final case class DistributedInferencerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { trainerConfig.vertexAiInferencerConfig.foreach { __v => @@ -165,7 +165,7 @@ object DistributedInferencerConfig extends scalapb.GeneratedMessageCompanion[sna override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class VertexAiInferencerConfig(value: snapchat.research.gbml.gigl_resource_config.VertexAiResourceConfig) extends snapchat.research.gbml.gigl_resource_config.DistributedInferencerConfig.TrainerConfig { type ValueType = snapchat.research.gbml.gigl_resource_config.VertexAiResourceConfig diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala index 909ec979b..1225ba210 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/KFPTrainerConfig.scala @@ -32,35 +32,35 @@ final case class KFPTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = cpuRequest if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = memoryRequest if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = gpuType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = gpuLimit if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(4, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -77,7 +77,7 @@ final case class KFPTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala index ba2cc9389..86e238074 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/LocalTrainerConfig.scala @@ -17,7 +17,7 @@ final case class LocalTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numWorkers if (__value != 0) { @@ -34,7 +34,7 @@ final case class LocalTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala index d32c915cb..96f354f47 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/SparkResourceConfig.scala @@ -25,21 +25,21 @@ final case class SparkResourceConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = machineType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = numLocalSsds if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(2, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -56,7 +56,7 @@ final case class SparkResourceConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala index c088bafc2..37d730799 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/gigl_resource_config/VertexAiTrainerConfig.scala @@ -29,28 +29,28 @@ final case class VertexAiTrainerConfig( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = machineType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = gpuType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = gpuLimit if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(3, __value) } }; - + { val __value = numReplicas if (__value != 0) { @@ -67,7 +67,7 @@ final case class VertexAiTrainerConfig( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala index cd1501b1f..dc359d1e3 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Edge.scala @@ -33,14 +33,14 @@ final case class Edge( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = srcNodeId if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(1, __value) } }; - + { val __value = dstNodeId if (__value != 0) { @@ -65,7 +65,7 @@ final case class Edge( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala index d2a68d8b9..439ae2229 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/EdgeType.scala @@ -21,21 +21,21 @@ final case class EdgeType( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = relation if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = srcNodeType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = dstNodeType if (!__value.isEmpty) { @@ -52,7 +52,7 @@ final case class EdgeType( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala index 0e4e3105d..7c317bb6d 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Graph.scala @@ -35,7 +35,7 @@ final case class Graph( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodes.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala index 8c6307580..856a159a0 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphMetadata.scala @@ -58,7 +58,7 @@ final case class GraphMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodeTypes.foreach { __v => @@ -205,7 +205,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -226,7 +226,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -269,7 +269,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research def companion: snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry.type = snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.GraphMetadata.CondensedEdgeTypeMapEntry]) } - + object CondensedEdgeTypeMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.graph_schema.GraphMetadata.CondensedEdgeTypeMapEntry = { @@ -341,7 +341,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.GraphMetadata.CondensedEdgeTypeMapEntry]) } - + @SerialVersionUID(0L) final case class CondensedNodeTypeMapEntry( key: _root_.scala.Int = 0, @@ -352,14 +352,14 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -376,7 +376,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -420,7 +420,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research def companion: snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry.type = snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.GraphMetadata.CondensedNodeTypeMapEntry]) } - + object CondensedNodeTypeMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.graph_schema.GraphMetadata.CondensedNodeTypeMapEntry = { @@ -485,7 +485,7 @@ object GraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat.research ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.GraphMetadata.CondensedNodeTypeMapEntry]) } - + implicit class GraphMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.graph_schema.GraphMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.graph_schema.GraphMetadata](_l) { def nodeTypes: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Seq[_root_.scala.Predef.String]] = field(_.nodeTypes)((c_, f_) => c_.copy(nodeTypes = f_)) def edgeTypes: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Seq[snapchat.research.gbml.graph_schema.EdgeType]] = field(_.edgeTypes)((c_, f_) => c_.copy(edgeTypes = f_)) diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala index 28c2eff67..0cd9e60ca 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/GraphSchemaProto.scala @@ -50,4 +50,4 @@ object GraphSchemaProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala index e8a4c6f98..7e4c5c8b9 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/graph_schema/Node.scala @@ -27,7 +27,7 @@ final case class Node( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = nodeId if (__value != 0) { @@ -52,7 +52,7 @@ final case class Node( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala index 6668d67ce..a58d891c0 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadata.scala @@ -31,7 +31,7 @@ final case class InferenceMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { nodeTypeToInferencerOutputInfoMap.foreach { __v => @@ -123,7 +123,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { @@ -144,7 +144,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -187,7 +187,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese def companion: snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry.type = snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry]) } - + object NodeTypeToInferencerOutputInfoMapEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.inference_metadata.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry = { @@ -259,7 +259,7 @@ object InferenceMetadata extends scalapb.GeneratedMessageCompanion[snapchat.rese ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.InferenceMetadata.NodeTypeToInferencerOutputInfoMapEntry]) } - + implicit class InferenceMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.inference_metadata.InferenceMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.inference_metadata.InferenceMetadata](_l) { def nodeTypeToInferencerOutputInfoMap: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, snapchat.research.gbml.inference_metadata.InferenceOutput]] = field(_.nodeTypeToInferencerOutputInfoMap)((c_, f_) => c_.copy(nodeTypeToInferencerOutputInfoMap = f_)) } diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala index 7e1424d6e..9c335401b 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceMetadataProto.scala @@ -35,4 +35,4 @@ object InferenceMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala index 845fbb29c..66c68b47b 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/inference_metadata/InferenceOutput.scala @@ -38,7 +38,7 @@ final case class InferenceOutput( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { embeddingsPath.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala index a0399d909..8ebed0a48 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostProcessedMetadata.scala @@ -17,7 +17,7 @@ final case class PostProcessedMetadata( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = postProcessorLogMetricsUri if (!__value.isEmpty) { @@ -34,7 +34,7 @@ final case class PostProcessedMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala index 4b0e94597..d36401685 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/postprocessed_metadata/PostprocessedMetadataProto.scala @@ -28,4 +28,4 @@ object PostprocessedMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala index 6a160cdd4..80160636b 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadata.scala @@ -38,7 +38,7 @@ final case class PreprocessedMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { condensedNodeTypeToPreprocessedMetadata.foreach { __v => @@ -181,7 +181,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = nodeIdKey if (!__value.isEmpty) { @@ -196,28 +196,28 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = __item __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(4, __value) } }; - + { val __value = schemaUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(5, __value) } }; - + { val __value = enumeratedNodeIdsBqTable if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(6, __value) } }; - + { val __value = enumeratedNodeDataBqTable if (!__value.isEmpty) { @@ -228,7 +228,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = featureDim.get __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(8, __value) }; - + { val __value = transformFnAssetsUri if (!__value.isEmpty) { @@ -245,7 +245,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -366,7 +366,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.NodeMetadataOutput]) } - + object NodeMetadataOutput extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput = { @@ -499,7 +499,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.NodeMetadataOutput]) } - + /** Houses metadata of edge features output from DataPreprocessor * * @param featureKeys @@ -540,21 +540,21 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = __item __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } - + { val __value = tfrecordUriPrefix if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = schemaUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(4, __value) } }; - + { val __value = enumeratedEdgeDataBqTable if (!__value.isEmpty) { @@ -565,7 +565,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r val __value = featureDim.get __size += _root_.com.google.protobuf.CodedOutputStream.computeUInt32Size(6, __value) }; - + { val __value = transformFnAssetsUri if (!__value.isEmpty) { @@ -582,7 +582,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { featureKeys.foreach { __v => @@ -679,7 +679,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataInfo]) } - + object EdgeMetadataInfo extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataInfo = { @@ -792,7 +792,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataInfo]) } - + /** Houses metadata about edge TFTransform output from DataPreprocessor. * * @param srcNodeIdKey @@ -819,14 +819,14 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = srcNodeIdKey if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = dstNodeIdKey if (!__value.isEmpty) { @@ -855,7 +855,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -932,7 +932,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataOutput]) } - + object EdgeMetadataOutput extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput = { @@ -1035,7 +1035,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.EdgeMetadataOutput]) } - + @SerialVersionUID(0L) final case class CondensedNodeTypeToPreprocessedMetadataEntry( key: _root_.scala.Int = 0, @@ -1046,7 +1046,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -1067,7 +1067,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -1110,7 +1110,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry]) } - + object CondensedNodeTypeToPreprocessedMetadataEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry = { @@ -1182,7 +1182,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.CondensedNodeTypeToPreprocessedMetadataEntry]) } - + @SerialVersionUID(0L) final case class CondensedEdgeTypeToPreprocessedMetadataEntry( key: _root_.scala.Int = 0, @@ -1193,7 +1193,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (__value != 0) { @@ -1214,7 +1214,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -1257,7 +1257,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r def companion: snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry.type = snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry]) } - + object CondensedEdgeTypeToPreprocessedMetadataEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry = { @@ -1329,7 +1329,7 @@ object PreprocessedMetadata extends scalapb.GeneratedMessageCompanion[snapchat.r ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.PreprocessedMetadata.CondensedEdgeTypeToPreprocessedMetadataEntry]) } - + implicit class PreprocessedMetadataLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata](_l) { def condensedNodeTypeToPreprocessedMetadata: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Int, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.NodeMetadataOutput]] = field(_.condensedNodeTypeToPreprocessedMetadata)((c_, f_) => c_.copy(condensedNodeTypeToPreprocessedMetadata = f_)) def condensedEdgeTypeToPreprocessedMetadata: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Int, snapchat.research.gbml.preprocessed_metadata.PreprocessedMetadata.EdgeMetadataOutput]] = field(_.condensedEdgeTypeToPreprocessedMetadata)((c_, f_) => c_.copy(condensedEdgeTypeToPreprocessedMetadata = f_)) diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala index b6e8d0d6d..becc2d068 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/preprocessed_metadata/PreprocessedMetadataProto.scala @@ -61,4 +61,4 @@ object PreprocessedMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala index c56f47fa4..2a6227235 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/GlobalRandomUniformStrategy.scala @@ -15,7 +15,7 @@ final case class GlobalRandomUniformStrategy( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numHops if (__value != 0) { @@ -36,7 +36,7 @@ final case class GlobalRandomUniformStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala index dc3069c8b..4aad9e19c 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPath.scala @@ -15,7 +15,7 @@ final case class MessagePassingPath( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = rootNodeType if (!__value.isEmpty) { @@ -36,7 +36,7 @@ final case class MessagePassingPath( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala index fd8a906af..ee3aa77a0 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/MessagePassingPathStrategy.scala @@ -34,7 +34,7 @@ final case class MessagePassingPathStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { paths.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala index 5579eca5a..ef0fde958 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomUniform.scala @@ -5,7 +5,7 @@ package snapchat.research.gbml.subgraph_sampling_strategy -/** Randomly sample nodes from the neighborhood without replacement. +/** Randomly sample nodes from the neighborhood without replacement. */ @SerialVersionUID(0L) final case class RandomUniform( @@ -16,7 +16,7 @@ final case class RandomUniform( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { @@ -33,7 +33,7 @@ final case class RandomUniform( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala index 389735b80..395f3876b 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/RandomWeighted.scala @@ -17,14 +17,14 @@ final case class RandomWeighted( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeInt32Size(1, __value) } }; - + { val __value = edgeFeatName if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class RandomWeighted( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala index b35d80157..6b6b756d5 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingDirection.scala @@ -17,7 +17,7 @@ sealed abstract class SamplingDirection(val value: _root_.scala.Int) extends _ro object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingDirection] { sealed trait Recognized extends SamplingDirection implicit def enumCompanion: _root_.scalapb.GeneratedEnumCompanion[SamplingDirection] = this - + /** Sample incoming edges to the dst nodes (default) */ @SerialVersionUID(0L) @@ -26,7 +26,7 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD val name = "INCOMING" override def isIncoming: _root_.scala.Boolean = true } - + /** Sample outgoing edges from the src nodes */ @SerialVersionUID(0L) @@ -35,7 +35,7 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD val name = "OUTGOING" override def isOutgoing: _root_.scala.Boolean = true } - + @SerialVersionUID(0L) final case class Unrecognized(unrecognizedValue: _root_.scala.Int) extends SamplingDirection(unrecognizedValue) with _root_.scalapb.UnrecognizedEnum lazy val values: scala.collection.immutable.Seq[ValueType] = scala.collection.immutable.Seq(INCOMING, OUTGOING) @@ -46,4 +46,4 @@ object SamplingDirection extends _root_.scalapb.GeneratedEnumCompanion[SamplingD } def javaDescriptor: _root_.com.google.protobuf.Descriptors.EnumDescriptor = SubgraphSamplingStrategyProto.javaDescriptor.getEnumTypes().get(0) def scalaDescriptor: _root_.scalapb.descriptors.EnumDescriptor = SubgraphSamplingStrategyProto.scalaDescriptor.enums(0) -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala index 513ceaf83..6afa2e8b6 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SamplingOp.scala @@ -28,7 +28,7 @@ final case class SamplingOp( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = opName if (!__value.isEmpty) { @@ -59,7 +59,7 @@ final case class SamplingOp( val __value = samplingMethod.userDefined.get __size += 2 + _root_.com.google.protobuf.CodedOutputStream.computeUInt32SizeNoTag(__value.serializedSize) + __value.serializedSize }; - + { val __value = samplingDirection.value if (__value != 0) { @@ -76,7 +76,7 @@ final case class SamplingOp( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -293,7 +293,7 @@ object SamplingOp extends scalapb.GeneratedMessageCompanion[snapchat.research.gb override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class RandomUniform(value: snapchat.research.gbml.subgraph_sampling_strategy.RandomUniform) extends snapchat.research.gbml.subgraph_sampling_strategy.SamplingOp.SamplingMethod { type ValueType = snapchat.research.gbml.subgraph_sampling_strategy.RandomUniform diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala index 5f8e3dee6..4abc7182b 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategy.scala @@ -32,7 +32,7 @@ final case class SubgraphSamplingStrategy( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { strategy.messagePassingPaths.foreach { __v => @@ -143,7 +143,7 @@ object SubgraphSamplingStrategy extends scalapb.GeneratedMessageCompanion[snapch override def number: _root_.scala.Int = 0 override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value") } - + @SerialVersionUID(0L) final case class MessagePassingPaths(value: snapchat.research.gbml.subgraph_sampling_strategy.MessagePassingPathStrategy) extends snapchat.research.gbml.subgraph_sampling_strategy.SubgraphSamplingStrategy.Strategy { type ValueType = snapchat.research.gbml.subgraph_sampling_strategy.MessagePassingPathStrategy diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala index 709ae1159..e6ae31c4c 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/SubgraphSamplingStrategyProto.scala @@ -67,4 +67,4 @@ object SubgraphSamplingStrategyProto extends _root_.scalapb.GeneratedFileObject } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala index 86bda274e..39c5388e4 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/TopK.scala @@ -17,14 +17,14 @@ final case class TopK( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = numNodesToSample if (__value != 0) { __size += _root_.com.google.protobuf.CodedOutputStream.computeInt32Size(1, __value) } }; - + { val __value = edgeFeatName if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class TopK( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala index a055bfbaa..cc00d0f9d 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/subgraph_sampling_strategy/UserDefined.scala @@ -20,7 +20,7 @@ final case class UserDefined( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = pathToUdf if (!__value.isEmpty) { @@ -41,7 +41,7 @@ final case class UserDefined( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -151,14 +151,14 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = key if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = value if (!__value.isEmpty) { @@ -175,7 +175,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { @@ -219,7 +219,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g def companion: snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry.type = snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry // @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.UserDefined.ParamsEntry]) } - + object ParamsEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry] { implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry] = this def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.subgraph_sampling_strategy.UserDefined.ParamsEntry = { @@ -284,7 +284,7 @@ object UserDefined extends scalapb.GeneratedMessageCompanion[snapchat.research.g ) // @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.UserDefined.ParamsEntry]) } - + implicit class UserDefinedLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.subgraph_sampling_strategy.UserDefined]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.subgraph_sampling_strategy.UserDefined](_l) { def pathToUdf: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.pathToUdf)((c_, f_) => c_.copy(pathToUdf = f_)) def params: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.collection.immutable.Map[_root_.scala.Predef.String, _root_.scala.Predef.String]] = field(_.params)((c_, f_) => c_.copy(params = f_)) diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala index bcf95c046..2c5a042f9 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadata.scala @@ -26,28 +26,28 @@ final case class TrainedModelMetadata( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = trainedModelUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = scriptedModelUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value) } }; - + { val __value = evalMetricsUri if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(3, __value) } }; - + { val __value = tensorboardLogsUri if (!__value.isEmpty) { @@ -64,7 +64,7 @@ final case class TrainedModelMetadata( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala index b06e0d55a..1262d9517 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/trained_model_metadata/TrainedModelMetadataProto.scala @@ -31,4 +31,4 @@ object TrainedModelMetadataProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala index 21884234b..3289f2d76 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/Label.scala @@ -15,14 +15,14 @@ final case class Label( private[this] var __serializedSizeMemoized: _root_.scala.Int = 0 private[this] def __computeSerializedSize(): _root_.scala.Int = { var __size = 0 - + { val __value = labelType if (!__value.isEmpty) { __size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value) } }; - + { val __value = label if (__value != 0) { @@ -39,7 +39,7 @@ final case class Label( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { { diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala index a05a6ef18..a1f7def5a 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/NodeAnchorBasedLinkPredictionSample.scala @@ -64,7 +64,7 @@ final case class NodeAnchorBasedLinkPredictionSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala index f326f0375..cdd1d4a4c 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/RootedNodeNeighborhood.scala @@ -41,7 +41,7 @@ final case class RootedNodeNeighborhood( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala index 4ffbeb328..9707ea68a 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedLinkBasedTaskSample.scala @@ -48,7 +48,7 @@ final case class SupervisedLinkBasedTaskSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootEdge.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala index 35c594ca3..00304a823 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/SupervisedNodeClassificationSample.scala @@ -43,7 +43,7 @@ final case class SupervisedNodeClassificationSample( __serializedSizeMemoized = __size } __size - 1 - + } def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = { rootNode.foreach { __v => diff --git a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala index f4187ea60..f02891e26 100644 --- a/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala +++ b/scala_spark35/common/src/main/scala/snapchat/research/gbml/training_samples_schema/TrainingSamplesSchemaProto.scala @@ -51,4 +51,4 @@ object TrainingSamplesSchemaProto extends _root_.scalapb.GeneratedFileObject { } @deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47") def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor -} \ No newline at end of file +} diff --git a/scala_spark35/common/src/main/scala/userDefinedAggregators/RnnUDAF.scala b/scala_spark35/common/src/main/scala/userDefinedAggregators/RnnUDAF.scala index 1900f8e07..856e64e47 100644 --- a/scala_spark35/common/src/main/scala/userDefinedAggregators/RnnUDAF.scala +++ b/scala_spark35/common/src/main/scala/userDefinedAggregators/RnnUDAF.scala @@ -70,16 +70,16 @@ class RnnUDAF( extends Aggregator[RnnUDAF.InTwoHopData, RnnUDAF.BufferRNN, Array[Byte]] { /** - * Introduces a custom user defined aggregation function that + * Introduces a custom user defined aggregation function that * allows for more efficient "GROUP BY" on "root_node_id" when formulating a 2 hop subgraph, * as compared to using default Spark aggregate functions like array_append, array_union, array_agg, et al. * These functions are quite expensive and not suitable for aggregating all types of columns. - * + * * The UDAF is used to aggregate the 2-hop subgraph information into a single RootedNodeNeighborhood * protobuf message (byte array). - * + * * sampleN: Option[Int] - The number of edges to sample from the 1-hop and 2-hop neighbors of the root node. - * + * * Example usage: * spark.udf.register("rnnUDAF", F.udaf(new RnnUDAF(sampleN = Some(VAL)))) * ... @@ -99,9 +99,9 @@ class RnnUDAF( * _2_hop_node_features, * _2_hop_edge_features, * _2_hop_edge_type - * ) as result + * ) as result * FROM - * ... + * ... * GROUP BY * _root_node_id, _root_node_type */ diff --git a/scala_spark35/common/src/main/scala/utils/SlottedJoiner.scala b/scala_spark35/common/src/main/scala/utils/SlottedJoiner.scala index d9635d857..2006b079d 100644 --- a/scala_spark35/common/src/main/scala/utils/SlottedJoiner.scala +++ b/scala_spark35/common/src/main/scala/utils/SlottedJoiner.scala @@ -9,7 +9,7 @@ object SlottedJoiner { /** * This class helps handle OOM and disk space issues in Spark jobs during large table joins. - * Instead of one big join, it partitions the left table into smaller tables and joins + * Instead of one big join, it partitions the left table into smaller tables and joins * them iteratively with the right table, ensuring better scalability with commodity hardware. * * Usage: @@ -17,11 +17,11 @@ object SlottedJoiner { * vla rightDf = ... * val numSlots = 10 * val slottedLeftDf = SlottedJoiner.computeSlotsOnDataframe( - * df=leftDf, - * columnToComputeSlotOn="joinKey", + * df=leftDf, + * columnToComputeSlotOn="joinKey", * numSlots=numSlots * ) - * + * * // Caching helps us avoid recomputing the tables * cacher.createDiskPartitionedTable( * df = leftSlottedDF, @@ -119,7 +119,7 @@ object SlottedJoiner { val queryWithSlottedTables = f""" WITH ${leftTableName} as ( - SELECT * + SELECT * FROM ${leftSlottedTableName} WHERE ${SLOT_NUM_COLUMN_NAME} = ${slotNum} ) ${currSlotQuery} @@ -171,12 +171,12 @@ object SlottedJoiner { .replace(rightSlottedTableName, rightTableName) val queryWithSlottedTables = f""" WITH ${leftTableName} as ( - SELECT * + SELECT * FROM ${leftSlottedTableName} WHERE ${SLOT_NUM_COLUMN_NAME} = ${slotNum} ), ${rightTableName} as ( - SELECT * + SELECT * FROM ${rightSlottedTableName} WHERE ${SLOT_NUM_COLUMN_NAME} = ${slotNum} ) ${currSlotQuery} diff --git a/scala_spark35/common/src/test/assets/resource_config.yaml b/scala_spark35/common/src/test/assets/resource_config.yaml index 1a795ef14..8eff18103 100644 --- a/scala_spark35/common/src/test/assets/resource_config.yaml +++ b/scala_spark35/common/src/test/assets/resource_config.yaml @@ -41,4 +41,4 @@ inferencer_config: num_workers: 1 max_num_workers: 256 machine_type: "c3-standard-22" - disk_size_gb: 100 \ No newline at end of file + disk_size_gb: 100 diff --git a/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml b/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml index b1a1421cd..a0a2ba944 100644 --- a/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml +++ b/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/frozen_gbml_config.yaml @@ -44,4 +44,4 @@ graphMetadata: relation: engage srcNodeType: user nodeTypes: - - user \ No newline at end of file + - user diff --git a/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml b/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml index 429035cbb..f44cdbb44 100644 --- a/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml +++ b/scala_spark35/common/src/test/assets/split_generator/node_anchor_based_link_prediction/preprocessed_metadata.yaml @@ -15,4 +15,4 @@ condensedNodeTypeToPreprocessedMetadata: - f1 nodeIdKey: node_id schemaUri: not.used.for.test - tfrecordUriPrefix: not.used.for.test \ No newline at end of file + tfrecordUriPrefix: not.used.for.test diff --git a/scala_spark35/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml b/scala_spark35/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml index 90762002c..68711a211 100644 --- a/scala_spark35/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml +++ b/scala_spark35/common/src/test/assets/split_generator/supervised_node_classification/frozen_gbml_config.yaml @@ -28,4 +28,4 @@ sharedConfig: supervisedNodeClassificationOutput: labeledTfrecordUriPrefix: common/src/test/assets/split_generator/supervised_node_classification/sgs_output/labeled/samples/ unlabeledTfrecordUriPrefix: common/src/test/assets/split_generator/supervised_node_classification/sgs_output/unlabeled/samples/ - preprocessedMetadataUri: common/src/test/assets/split_generator/supervised_node_classification/preprocessed_metadata.yaml \ No newline at end of file + preprocessedMetadataUri: common/src/test/assets/split_generator/supervised_node_classification/preprocessed_metadata.yaml diff --git a/scala_spark35/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml b/scala_spark35/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml index cb1587a83..798350dd0 100644 --- a/scala_spark35/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml +++ b/scala_spark35/common/src/test/assets/subgraph_sampler/supervised_node_classification/frozen_gbml_config.yaml @@ -24,4 +24,4 @@ sharedConfig: supervisedNodeClassificationOutput: labeledTfrecordUriPrefix: common/src/test/assets/subgraph_sampler/supervised_node_classification/output/labeled/samples/ unlabeledTfrecordUriPrefix: common/src/test/assets/subgraph_sampler/supervised_node_classification/output/unlabeled/samples/ - preprocessedMetadataUri: common/src/test/assets/subgraph_sampler/supervised_node_classification/preprocessed_metadata.yaml \ No newline at end of file + preprocessedMetadataUri: common/src/test/assets/subgraph_sampler/supervised_node_classification/preprocessed_metadata.yaml diff --git a/scala_spark35/common/src/test/scala/userDefinedAggregators/RnnUDAFTest.scala b/scala_spark35/common/src/test/scala/userDefinedAggregators/RnnUDAFTest.scala index 699ae40f5..ca3a1b984 100644 --- a/scala_spark35/common/src/test/scala/userDefinedAggregators/RnnUDAFTest.scala +++ b/scala_spark35/common/src/test/scala/userDefinedAggregators/RnnUDAFTest.scala @@ -263,10 +263,10 @@ class RnnUDAFTest extends AnyFunSuite with BeforeAndAfterAll with SharedSparkSes _2_hop_node_features, _2_hop_edge_features, _2_hop_edge_type - ) as result - FROM - test_view - GROUP BY + ) as result + FROM + test_view + GROUP BY _root_node_id, _root_node_type """) diff --git a/scala_spark35/common/src/test/scala/utils/SlottedJoinerTest.scala b/scala_spark35/common/src/test/scala/utils/SlottedJoinerTest.scala index 49a77ce58..f6fa5a1d8 100644 --- a/scala_spark35/common/src/test/scala/utils/SlottedJoinerTest.scala +++ b/scala_spark35/common/src/test/scala/utils/SlottedJoinerTest.scala @@ -89,11 +89,11 @@ class SlottedJoinerTest extends AnyFunSuite with BeforeAndAfterAll with SharedSp slottedOnSrc.dst_node as root_node, slottedOnDst.dst_node as 1_hop_node, slottedOnDst.src_node as 2_hop_node - FROM - slottedOnSrc - JOIN - slottedOnDst - ON + FROM + slottedOnSrc + JOIN + slottedOnDst + ON slottedOnSrc.src_node = slottedOnDst.dst_node """, numSlots = numSlots, diff --git a/scala_spark35/split_generator/src/main/scala/Main.scala b/scala_spark35/split_generator/src/main/scala/Main.scala index b54f019e4..8c8fa5b2e 100644 --- a/scala_spark35/split_generator/src/main/scala/Main.scala +++ b/scala_spark35/split_generator/src/main/scala/Main.scala @@ -17,7 +17,7 @@ object Main { val resourceConfigYamlGcsUri = args(2) println(f""" - Starting Split Generator with the following arguments: + Starting Split Generator with the following arguments: sparkAppName=${sparkAppName}, frozenGbmlConfigYamlGcsUri=${frozenGbmlConfigYamlGcsUri}, resourceConfigYamlGcsUri=${resourceConfigYamlGcsUri} diff --git a/scala_spark35/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala b/scala_spark35/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala index 90524f6a0..891198953 100644 --- a/scala_spark35/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala +++ b/scala_spark35/split_generator/src/main/scala/lib/assigners/AbstractAssigners.scala @@ -22,7 +22,7 @@ object AbstractAssigners { * e.g. could be assigning a NodePb (T) to some Enum (S). * * @param obj the object to hash - * @return + * @return */ def assign(obj: T): S } @@ -59,7 +59,7 @@ object AbstractAssigners { /** * Relative width of each bucket in the hash space. e.g. [0.2, 0.4, 0.4] would indicate 3 buckets, where - * the second and third bucket are twice as prominent as the first bucket. + * the second and third bucket are twice as prominent as the first bucket. */ lazy val weights: Seq[Float] = bucketWeights.values.toList diff --git a/scala_spark35/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala b/scala_spark35/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala index ba15966a3..b0d7e1146 100644 --- a/scala_spark35/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala +++ b/scala_spark35/split_generator/src/main/scala/lib/split_strategies/SplitStrategy.scala @@ -32,7 +32,7 @@ abstract class SplitStrategy[A](splitStrategyArgs: Map[String, String]) extends val graphMetadataPbWrapper: GraphMetadataPbWrapper /** - * Takes in a single "un-split" training sample instance output by SubgraphSampler, + * Takes in a single "un-split" training sample instance output by SubgraphSampler, * and a DatasetSplit(TRAIN, TEST, VAL) and outputs the the "split" samples for that dataset split * * @param sample : Input Sample from SGS diff --git a/scala_spark35/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala b/scala_spark35/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala index 61fefef6f..12e378ebe 100644 --- a/scala_spark35/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala +++ b/scala_spark35/split_generator/src/main/scala/lib/split_strategies/UDLAnchorBasedSupervisionEdgeSplitStrategy.scala @@ -37,7 +37,7 @@ class UDLAnchorBasedSupervisionEdgeSplitStrategy( * (a) All pos_edges and hard_neg_edges belonging to the split. * (b) message passing structure which should be pb.neighborhood and therefore the same across all splits * (i.e. no masking). - * (c) The message passing structure may be filtered down to only include edges that are not in the pos_edges + * (c) The message passing structure may be filtered down to only include edges that are not in the pos_edges * and hard_neg_edges. * An output train-split sample needs to have >0 pos_edges in this setting for loss computation. * Output val/test-split samples may have 0 pos_edges (and even 0 hard_neg_edges), since these diff --git a/scala_spark35/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala b/scala_spark35/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala index a1c5b9081..dd63cc113 100644 --- a/scala_spark35/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala +++ b/scala_spark35/subgraph_sampler/src/main/scala/libs/task/TaskOutputValidator.scala @@ -17,7 +17,7 @@ object TaskOutputValidator { * is present in the neighborhood nodes. * This method does a dataset.map() on the final output produced by SGS and returns the same dataset * if there is no validation failure. Raises and excpetion if there is some error - * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any + * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any * duplication of computation due to this validation code. * * @param mainSampleDS @@ -50,7 +50,7 @@ object TaskOutputValidator { * is present in the neighborhood nodes. * This method does a dataset.map() on the final output produced by SGS and returns the same dataset * if there is no validation failure. Raises and excpetion if there is some error - * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any + * @spark: dataset.map() is not an action (unlike foreach) and does not lead to any * duplication of computation due to this validation code. * * @param mainSampleDS diff --git a/scala_spark35/subgraph_sampler/src/main/scala/libs/task/pureSparkV2/EgoNetGeneration.scala b/scala_spark35/subgraph_sampler/src/main/scala/libs/task/pureSparkV2/EgoNetGeneration.scala index b66300d9e..68ad26f3d 100644 --- a/scala_spark35/subgraph_sampler/src/main/scala/libs/task/pureSparkV2/EgoNetGeneration.scala +++ b/scala_spark35/subgraph_sampler/src/main/scala/libs/task/pureSparkV2/EgoNetGeneration.scala @@ -46,7 +46,7 @@ object EgoNetGeneration { spark.sql( s""" SELECT DISTINCT * FROM ( - SELECT + SELECT dst_node_id as src_node_id, src_node_id as dst_node_id, ${DEFAULT_EDGE_TYPE} as edge_type, @@ -454,10 +454,10 @@ class EgoNetGeneration( val toNodeIdColumn = flags.to_node_id_column println(s""" - Running EgoNetGeneration Job w/ - nodeTableName: ${nodeTableName}, - edgeTableName: ${edgeTableName}, - fromNodeIdColumn: ${fromNodeIdColumn}, + Running EgoNetGeneration Job w/ + nodeTableName: ${nodeTableName}, + edgeTableName: ${edgeTableName}, + fromNodeIdColumn: ${fromNodeIdColumn}, toNodeIdColumn: ${toNodeIdColumn} """)