Important: This documentation covers Yarn 1 (Classic).
For Yarn 2+ docs and migration guide, see yarnpkg.com.

Package detail

csv-streamify

klaemo190.2kBSD-2-Clause4.0.0

Streaming CSV Parser. Made entirely out of streams.

csv, parser, stream, cli

readme

csv-streamify Build Status Greenkeeper badge

NPM

Parses csv files. Accepts options. No coffee script, no weird APIs. Just streams. Tested against csv-spectrum and used in production. It is also "fast enough" (around 60,000 rows per second, but that varies with data obviously).

Works in node 4, 6, 8 and 9. Might work in earlier versions, but is not tested in it.

Installation

npm install csv-streamify

Usage

This module implements a simple node stream.Transform stream. You can write to it, read from it and use .pipe as you would expect.

const csv = require('csv-streamify')
const fs = require('fs')

const parser = csv()

// emits each line as a buffer or as a string representing an array of fields
parser.on('data', function (line) {
  console.log(line)
})

// now pipe some data into it
fs.createReadStream('/path/to/file.csv').pipe(parser)

with options and callback

The first argument can either be an options object (see below) or a callback function.

Note: If you pass a callback to csv-streamify it will buffer the parsed data for you and pass it to the callback when it's done. This behaviour can obviously lead to out of memory errors with very large csv files.

const csv = require('csv-streamify')
const fs = require('fs')

const parser = csv({ objectMode: true }, function (err, result) {
  if (err) throw err
  // our csv has been parsed succesfully
  result.forEach(function (line) { console.log(line) })
})

// now pipe some data into it
fs.createReadStream('/path/to/file.csv').pipe(parser)

Options

You can pass some options to the parser. All of them are optional.

The options are also passed to the underlying transform stream, so you can pass in any standard node core stream options.

{
  delimiter: ',', // comma, semicolon, whatever
  newline: '\n', // newline character (use \r\n for CRLF files)
  quote: '"', // what's considered a quote
  empty: '', // empty fields are replaced by this,

  // if true, emit arrays instead of stringified arrays or buffers
  objectMode: false,

  // if set to true, uses first row as keys -> [ { column1: value1, column2: value2 }, ...]
  columns: false
}

Also, take a look at iconv-lite (npm install iconv-lite --save), it provides pure javascript streaming character encoding conversion.

CLI

To use on the command line install it globally:

$ npm install csv-streamify -g

This should add the csv-streamify command to your $PATH.

Then, you either pipe data into it or give it a filename:

# pipe data in
$ cat some_data.csv | csv-streamify
# pass a filename
$ csv-streamify some_data.csv > output.json
# tell csv-streamify to read from + wait on stdin
$ csv-streamify -

Wishlist

  • browser support
  • better CLI

If you would like to contribute either of those just open an issue so we can discuss it further. :)

Contributors

Nicolas Hery (objectMode)

changelog

4.0.0 / 2018-03-25

  • BREAKING: make objectMode: true the default. The parser now emits objects/arrays per line instead of buffers.
  • drop support for old, unsupported node versions (0.12, 5) and test in node 8 and 9.
  • update dependencies

3.0.4 / 2016-05-24

  • ensure compatibility with newer node version

3.0.3 / 2015-11-26

  • tweak readme

3.0.2 / 2015-11-25

  • fixes longstanding issue #14

3.0.1 / 2015-11-25

  • fixes #17

3.0.0 / 2015-11-01

  • drop node 0.10 support, add node 4 support
  • use through2 internally
  • deprecate parser.lineNo and parser.body

2.0.0 / 2015-02-15

  • iconv-lite should be used seperatly.
  • iojs and 0.12 compat.

1.0.0 / 2014-03-30

  • No changes, I just want to start doing semver properly.

0.9.1 / 2014-03-25

  • add simple CLI tool

0.9.0 / 2013-12-27

  • fix #9
  • fix #8

0.8.1 / 2013-11-21

  • fix: correctly parse empty quoted cells (#6)
  • update csv-spectrum devDep

0.8.0 / 2013-11-16

  • change: use csv-spectrum as a node module and comply to its changed tests

0.7.0 / 2013-11-14

  • change: use iconv-lite for encoding conversion (please test und submit issues)

0.6.1 / 2013-11-13

  • fix: properly handle CRLF in quoted sequences

0.6.0 / 2013-11-12

  • feature: add support for columns
  • fix: some small changes regarding quotes to align output with csv-spectrum
  • internal: implement _flush

0.5.1 / 2013-11-10

  • fix: last line wasn't parsed (damn! please update!)

0.5.0 / 2013-07-19

  • change: rename 'encoding' to 'inputEncoding'
  • fix: parsing of CRLF files (thanks ktk for reporting)