Package co.cask.cdap.api.data.batch

Examples of co.cask.cdap.api.data.batch.RecordScannable


      throw new IOException(
        String.format("Dataset %s does not implement RecordScannable, and hence cannot be queried in Hive.",
                      conf.get(Constants.Explore.DATASET_NAME)));
    }

    RecordScannable recordScannable = (RecordScannable) dataset;

    if (recordScannable instanceof TransactionAware) {
      startTransaction(conf, (TransactionAware) recordScannable);
    }
View Full Code Here


  private static final Gson GSON = new Gson();

  @Override
  public InputSplit[] getSplits(JobConf jobConf, int numSplits) throws IOException {

    RecordScannable recordScannable = DatasetAccessor.getRecordScannable(jobConf);

    try {
      Job job = new Job(jobConf);
      JobContext jobContext = ShimLoader.getHadoopShims().newJobContext(job);
      Path[] tablePaths = FileInputFormat.getInputPaths(jobContext);

      List<Split> dsSplits = recordScannable.getSplits();

      InputSplit[] inputSplits = new InputSplit[dsSplits.size()];
      for (int i = 0; i < dsSplits.size(); i++) {
        inputSplits[i] = new DatasetInputSplit(dsSplits.get(i), tablePaths[0]);
      }
      return inputSplits;
    } finally {
      recordScannable.close();
    }
  }
View Full Code Here

  @Override
  public RecordReader<Void, ObjectWritable> getRecordReader(final InputSplit split, JobConf jobConf, Reporter reporter)
    throws IOException {

    final RecordScannable recordScannable = DatasetAccessor.getRecordScannable(jobConf);

    if (!(split instanceof DatasetInputSplit)) {
      throw new IOException("Invalid type for InputSplit: " + split.getClass().getName());
    }
    final DatasetInputSplit datasetInputSplit = (DatasetInputSplit) split;

    final RecordScanner recordScanner = recordScannable.createSplitRecordScanner(
        new Split() {
          @Override
          public long getLength() {
            try {
              return split.getLength();
            } catch (IOException e) {
              throw new RuntimeException(e);
            }
          }
        }
    );

    return new RecordReader<Void, ObjectWritable>() {
      private final AtomicBoolean initialized = new AtomicBoolean(false);

      private void initialize() throws IOException {
        try {
          recordScanner.initialize(datasetInputSplit.getDataSetSplit());
        } catch (InterruptedException ie) {
          Thread.currentThread().interrupt();
          throw new IOException("Interrupted while initializing reader", ie);
        }
        initialized.set(true);
      }

      @Override
      public boolean next(Void key, ObjectWritable value) throws IOException {
        if (!initialized.get()) {
          initialize();
        }

        try {
          boolean retVal = recordScanner.nextRecord();
          if (retVal) {
            value.set(recordScanner.getCurrentRecord());
          }
          return retVal;
        } catch (InterruptedException e) {
          Thread.currentThread().interrupt();
          throw new IOException(e);
        }
      }

      @Override
      public Void createKey() {
        return null;
      }

      @Override
      public ObjectWritable createValue() {
        return new ObjectWritable();
      }

      @Override
      public long getPos() throws IOException {
        // Not required.
        return 0;
      }

      @Override
      public void close() throws IOException {
        try {
          recordScanner.close();
        } finally {
          recordScannable.close();
        }
      }

      @Override
      public float getProgress() throws IOException {
View Full Code Here

TOP

Related Classes of co.cask.cdap.api.data.batch.RecordScannable

Copyright © 2018 www.massapicom. All rights reserved.
All source code are property of their respective owners. Java is a trademark of Sun Microsystems, Inc and owned by ORACLE Inc. Contact coftware#gmail.com.