Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 3 additions & 4 deletions .github/actions/assert-is-collaborator/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ inputs:
description: The GitHub username to check
required: true
initiating-pr-number:
description: The PR number that the check may be associated with, if provided will comment on the PR incase of failures
description: The PR number that the check may be associated with, if provided will comment on the PR incase of failures
required: false
runs:
using: "composite"
Expand All @@ -22,7 +22,7 @@ runs:
script: |
try {
const username = "${{ inputs.username }}";
const result = await github.rest.repos.checkCollaborator({
const result = await github.rest.repos.checkCollaborator({
owner: context.repo.owner,
repo: context.repo.repo,
username: username
Expand All @@ -41,12 +41,11 @@ runs:
console.log(`Error checking collaborator status: ${error.message}`);
}
}

Comment thread
kmontemayor2-sc marked this conversation as resolved.
- name: Comment workflow permissions
if: ${{ failure() && steps.assert-is-collaborator.conclusion == 'failure' && inputs.initiating-pr-number != '' }}
uses: snapchat/gigl/.github/actions/comment-on-pr@main
with:
pr_number: ${{ inputs.initiating-pr-number }}
message: |
🔒 User ${{ inputs.username }} does not have permissions to run this workflow

4 changes: 2 additions & 2 deletions .github/actions/comment-on-pr/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@ outputs:
comment_id:
description: 'The ID of the created or updated comment'
value: ${{steps.comment.outputs.result}}


runs:
using: 'composite'
Expand Down Expand Up @@ -67,4 +67,4 @@ runs:
});

return response.data.id;
}
}
2 changes: 1 addition & 1 deletion .github/actions/get-pr-src-branch/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -30,4 +30,4 @@ runs:
});
const branch_name = pr.data.head.ref;
console.log("Branch name is:", branch_name);
return branch_name;
return branch_name;
Original file line number Diff line number Diff line change
Expand Up @@ -5,4 +5,4 @@ package snapchat.research.gbml;
message PostProcessedMetadata{
// The path to the post processor evaluation results
string post_processor_log_metrics_uri = 1;
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ package snapchat.research.gbml;

import "snapchat/research/gbml/graph_schema.proto";

message RandomUniform { // Randomly sample nodes from the neighborhood without replacement.
message RandomUniform { // Randomly sample nodes from the neighborhood without replacement.
int32 num_nodes_to_sample = 1;
}

Expand Down
2 changes: 1 addition & 1 deletion scala/.scalafix.conf
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
rules = [
ExplicitResultTypes
NoValInForComprehension,
OrganizeImports,
OrganizeImports,
ProcedureSyntax,
RedundantSyntax,
]
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ final case class DatasetMetadata(
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
outputMetadata.supervisedNodeClassificationDataset.foreach { __v =>
Expand Down Expand Up @@ -165,7 +165,7 @@ object DatasetMetadata extends scalapb.GeneratedMessageCompanion[snapchat.resear
override def number: _root_.scala.Int = 0
override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value")
}

@SerialVersionUID(0L)
final case class SupervisedNodeClassificationDataset(value: snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset) extends snapchat.research.gbml.dataset_metadata.DatasetMetadata.OutputMetadata {
type ValueType = snapchat.research.gbml.dataset_metadata.SupervisedNodeClassificationDataset
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -60,4 +60,4 @@ object DatasetMetadataProto extends _root_.scalapb.GeneratedFileObject {
}
@deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47")
def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -21,21 +21,21 @@ final case class NodeAnchorBasedLinkPredictionDataset(
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = trainMainDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = testMainDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value)
}
};

{
val __value = valMainDataUri
if (!__value.isEmpty) {
Expand Down Expand Up @@ -64,7 +64,7 @@ final case class NodeAnchorBasedLinkPredictionDataset(
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down Expand Up @@ -250,14 +250,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = key
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = value
if (!__value.isEmpty) {
Expand All @@ -274,7 +274,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down Expand Up @@ -318,7 +318,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry
// @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry])
}

object TrainNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] {
implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry] = this
def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry = {
Expand Down Expand Up @@ -383,7 +383,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
)
// @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TrainNodeTypeToRandomNegativeDataUriEntry])
}

@SerialVersionUID(0L)
final case class ValNodeTypeToRandomNegativeDataUriEntry(
key: _root_.scala.Predef.String = "",
Expand All @@ -394,14 +394,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = key
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = value
if (!__value.isEmpty) {
Expand All @@ -418,7 +418,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down Expand Up @@ -462,7 +462,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry
// @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry])
}

object ValNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] {
implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry] = this
def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry = {
Expand Down Expand Up @@ -527,7 +527,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
)
// @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.ValNodeTypeToRandomNegativeDataUriEntry])
}

@SerialVersionUID(0L)
final case class TestNodeTypeToRandomNegativeDataUriEntry(
key: _root_.scala.Predef.String = "",
Expand All @@ -538,14 +538,14 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = key
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = value
if (!__value.isEmpty) {
Expand All @@ -562,7 +562,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down Expand Up @@ -606,7 +606,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
def companion: snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry.type = snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry
// @@protoc_insertion_point(GeneratedMessage[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry])
}

object TestNodeTypeToRandomNegativeDataUriEntry extends scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] {
implicit def messageCompanion: scalapb.GeneratedMessageCompanion[snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry] = this
def parseFrom(`_input__`: _root_.com.google.protobuf.CodedInputStream): snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry = {
Expand Down Expand Up @@ -671,7 +671,7 @@ object NodeAnchorBasedLinkPredictionDataset extends scalapb.GeneratedMessageComp
)
// @@protoc_insertion_point(GeneratedMessageCompanion[snapchat.research.gbml.NodeAnchorBasedLinkPredictionDataset.TestNodeTypeToRandomNegativeDataUriEntry])
}

implicit class NodeAnchorBasedLinkPredictionDatasetLens[UpperPB](_l: _root_.scalapb.lenses.Lens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset]) extends _root_.scalapb.lenses.ObjectLens[UpperPB, snapchat.research.gbml.dataset_metadata.NodeAnchorBasedLinkPredictionDataset](_l) {
def trainMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.trainMainDataUri)((c_, f_) => c_.copy(trainMainDataUri = f_))
def testMainDataUri: _root_.scalapb.lenses.Lens[UpperPB, _root_.scala.Predef.String] = field(_.testMainDataUri)((c_, f_) => c_.copy(testMainDataUri = f_))
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,21 +18,21 @@ final case class SupervisedLinkBasedTaskSplitDataset(
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = trainDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = testDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value)
}
};

{
val __value = valDataUri
if (!__value.isEmpty) {
Expand All @@ -49,7 +49,7 @@ final case class SupervisedLinkBasedTaskSplitDataset(
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,21 +18,21 @@ final case class SupervisedNodeClassificationDataset(
private[this] var __serializedSizeMemoized: _root_.scala.Int = 0
private[this] def __computeSerializedSize(): _root_.scala.Int = {
var __size = 0

{
val __value = trainDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(1, __value)
}
};

{
val __value = testDataUri
if (!__value.isEmpty) {
__size += _root_.com.google.protobuf.CodedOutputStream.computeStringSize(2, __value)
}
};

{
val __value = valDataUri
if (!__value.isEmpty) {
Expand All @@ -49,7 +49,7 @@ final case class SupervisedNodeClassificationDataset(
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
{
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ final case class FlattenedGraphMetadata(
__serializedSizeMemoized = __size
}
__size - 1

}
def writeTo(`_output__`: _root_.com.google.protobuf.CodedOutputStream): _root_.scala.Unit = {
outputMetadata.supervisedNodeClassificationOutput.foreach { __v =>
Expand Down Expand Up @@ -165,7 +165,7 @@ object FlattenedGraphMetadata extends scalapb.GeneratedMessageCompanion[snapchat
override def number: _root_.scala.Int = 0
override def value: _root_.scala.Nothing = throw new java.util.NoSuchElementException("Empty.value")
}

@SerialVersionUID(0L)
final case class SupervisedNodeClassificationOutput(value: snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput) extends snapchat.research.gbml.flattened_graph_metadata.FlattenedGraphMetadata.OutputMetadata {
type ValueType = snapchat.research.gbml.flattened_graph_metadata.SupervisedNodeClassificationOutput
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -50,4 +50,4 @@ object FlattenedGraphMetadataProto extends _root_.scalapb.GeneratedFileObject {
}
@deprecated("Use javaDescriptor instead. In a future version this will refer to scalaDescriptor.", "ScalaPB 0.5.47")
def descriptor: com.google.protobuf.Descriptors.FileDescriptor = javaDescriptor
}
}
Loading