All files / src/utils importer.js

Press n or j to go to the next uncovered block, b, p or k for the previous block.

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 1281x 1x 1x 1x 1x 1x 1x   1x   40016x 20005x 20005x 20005x 1x 1x 2x 1x 1x 1x   20005x 40015x                       40015x   20005x     9x 9x 9x 9x 9x   9x       58x 9x 9x 9x 9x 9x 9x   9x 9x 9x   9x 9x 9x 9x 20005x   20005x 20005x 20004x 20004x 2000x 2000x     1x     20005x 20005x 200x 200x   200x 200x             20005x 20x     9x       9x 9x 9x 9x 1x 8x 2x   9x   9x 9x 9x 9x                           1x  
const Promise = require('bluebird');
const _ = require('lodash');
const util = require('util');
const fs = require('fs');
const path = require('path');
const jsonStream = require('JSONStream');
const debug = require('debug')('express-cassandra');
 
const importer = {
  buildTableQueryForDataRow(keyspace, tableInfo, row) {
    row = _.omitBy(row, (item) => (item === null));
    let query = util.format('INSERT INTO "%s"."%s" ("%s") VALUES (?%s)', keyspace, tableInfo.name, _.keys(row).join('","'), _.repeat(',?', _.keys(row).length - 1));
    let params = _.values(row);
    if (tableInfo.isCounterTable) {
      const primaryKeyFields = _.pick(row, tableInfo.primaryKeys);
      const otherKeyFields = _.omit(row, tableInfo.primaryKeys);
      const setQueries = _.map(_.keys(otherKeyFields), (key) => util.format('"%s"="%s" + ?', key, key));
      const whereQueries = _.map(_.keys(primaryKeyFields), (key) => util.format('"%s"=?', key));
      query = util.format('UPDATE "%s"."%s" SET %s WHERE %s', keyspace, tableInfo.name, setQueries.join(', '), whereQueries.join(' AND '));
      params = _.values(otherKeyFields).concat(_.values(primaryKeyFields));
    }
    params = _.map(params, (param) => {
      Iif (_.isPlainObject(param)) {
        if (param.type === 'Buffer') {
          return Buffer.from(param);
        }
        const omittedParams = _.omitBy(param, (item) => (item === null));
        Object.keys(omittedParams).forEach((key) => {
          if (_.isObject(omittedParams[key]) && omittedParams[key].type === 'Buffer') {
            omittedParams[key] = Buffer.from(omittedParams[key]);
          }
        });
        return omittedParams;
      }
      return param;
    });
    return { query, params };
  },
 
  processTableImport(systemClient, fixtureDirectory, keyspace, table, batchSize) {
    return new Promise((resolve, reject) => {
      debug('==================================================');
      debug(`Reading metadata for table: ${table}`);
      systemClient.metadata.getTable(keyspace, table)
        .then((tableInfo) => {
          Iif (!tableInfo) {
            resolve();
            return;
          }
          const isCounterTable = _.some(tableInfo.columns, (column) => (column.type.code === 5));
          if (isCounterTable) batchSize = 1;
          let primaryKeys = [];
          primaryKeys = primaryKeys.concat(_.map(tableInfo.partitionKeys, (item) => item.name));
          primaryKeys = primaryKeys.concat(_.map(tableInfo.clusteringKeys, (item) => item.name));
          tableInfo.isCounterTable = isCounterTable;
          tableInfo.primaryKeys = primaryKeys;
 
          let queryPromises = [];
          let queries = [];
          let processed = 0;
 
          debug(`Creating read stream from: ${table}.json`);
          const jsonfile = fs.createReadStream(path.join(fixtureDirectory, `${table}.json`), { encoding: 'utf8' });
          const readStream = jsonfile.pipe(jsonStream.parse('*'));
          readStream.on('data', (row) => {
            processed++;
 
            const query = this.buildTableQueryForDataRow(keyspace, tableInfo, row);
            if (batchSize > 1) {
              queries.push(query);
              if (queries.length >= batchSize) {
                queryPromises.push(systemClient.batch(queries, { prepare: true }));
                queries = [];
              }
            } else {
              queryPromises.push(systemClient.execute(query.query, query.params, { prepare: true }));
            }
 
            const processPauseSize = (batchSize >= 10) ? batchSize * 10 : 100;
            if (processed % processPauseSize === 0) {
              readStream.pause();
              Promise.all(queryPromises)
                .then(() => {
                  queryPromises = [];
                  readStream.resume();
                })
                .catch((err) => {
                  reject(err);
                });
            }
 
            if (processed % 1000 === 0) {
              debug(`Streaming ${processed} rows to table: ${table}`);
            }
          });
          readStream.on('error', (err) => {
            reject(err);
          });
 
          const startTime = Date.now();
          readStream.on('end', () => {
            debug(`Streaming ${processed} rows to table: ${table}`);
            if (queries.length > 1) {
              queryPromises.push(systemClient.batch(queries, { prepare: true }));
            } else if (queries.length === 1) {
              queryPromises.push(systemClient.execute(queries[0].query, queries[0].params, { prepare: true }));
            }
            Promise.all(queryPromises)
              .then(() => {
                const timeTaken = (Date.now() - startTime) / 1000;
                const throughput = timeTaken ? processed / timeTaken : 0.00;
                debug(`Done with table, throughput: ${throughput.toFixed(1)} rows/s`);
                resolve();
              })
              .catch((err) => {
                reject(err);
              });
          });
        })
        .catch((err) => {
          reject(err);
        });
    });
  },
};
 
module.exports = importer;