graphlearn_torch/python/sampler/negative_sampler.py (18 lines of code) (raw):

# Copyright 2022 Alibaba Group Holding Limited. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== import torch from .. import py_graphlearn_torch as pywrap class RandomNegativeSampler(object): r""" Random negative Sampler. Args: graph: A ``graphlearn_torch.data.Graph`` object. mode: Execution mode of sampling, 'CUDA' means sampling on GPU, 'CPU' means sampling on CPU. edge_dir: The direction of edges to be sampled, determines the order of rows and columns returned. """ def __init__(self, graph, mode='CUDA', edge_dir='out'): self._mode = mode self.edge_dir = edge_dir if mode == 'CUDA': self._sampler = pywrap.CUDARandomNegativeSampler(graph.graph_handler) else: self._sampler = pywrap.CPURandomNegativeSampler(graph.graph_handler) def sample(self, req_num, trials_num=5, padding=False): r""" Negative sampling. Args: req_num: The number of request(max) negative samples. trials_num: The number of trials for negative sampling. padding: Whether to patch the negative sampling results to req_num. If True, after trying trials_num times, if the number of true negative samples is still less than req_num, just random sample edges(non-strict negative) as negative samples. Returns: negative edge_index(non-strict when padding is True). """ if self.edge_dir == 'out': rows, cols = self._sampler.sample(req_num, trials_num, padding) elif self.edge_dir == 'in': cols, rows = self._sampler.sample(req_num, trials_num, padding) return torch.stack([rows, cols], dim=0)