public static void main()

in spark/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/example/hbasecontext/JavaHBaseMapGetPutExample.java [51:96]


  public static void main(String[] args) {
    if (args.length < 1) {
      System.out.println("JavaHBaseBulkGetExample {tableName}");
      return;
    }

    final String tableName = args[0];

    SparkConf sparkConf = new SparkConf().setAppName("JavaHBaseBulkGetExample " + tableName);
    JavaSparkContext jsc = new JavaSparkContext(sparkConf);

    try {
      List<byte[]> list = new ArrayList<>(5);
      list.add(Bytes.toBytes("1"));
      list.add(Bytes.toBytes("2"));
      list.add(Bytes.toBytes("3"));
      list.add(Bytes.toBytes("4"));
      list.add(Bytes.toBytes("5"));

      JavaRDD<byte[]> rdd = jsc.parallelize(list);
      Configuration conf = HBaseConfiguration.create();

      JavaHBaseContext hbaseContext = new JavaHBaseContext(jsc, conf);

      hbaseContext.foreachPartition(rdd, new VoidFunction<Tuple2<Iterator<byte[]>, Connection>>() {
        public void call(Tuple2<Iterator<byte[]>, Connection> t) throws Exception {
          Table table = t._2().getTable(TableName.valueOf(tableName));
          BufferedMutator mutator = t._2().getBufferedMutator(TableName.valueOf(tableName));

          while (t._1().hasNext()) {
            byte[] b = t._1().next();
            Result r = table.get(new Get(b));
            if (r.getExists()) {
              mutator.mutate(new Put(b));
            }
          }

          mutator.flush();
          mutator.close();
          table.close();
        }
      });
    } finally {
      jsc.stop();
    }
  }