fix: 修复中介导入成功条数计算错误
问题: - 导入成功条数显示为负数 - 原因:成功数量计算使用 validRecords.size() - failures.size() - 但没有使用实际的数据库操作返回值 修复: - saveBatchWithUpsert 和 saveBatch 方法现在返回 int - 累加实际的数据库影响行数 - 使用 actualSuccessCount 变量跟踪真实成功数量 影响范围: - CcdiIntermediaryPersonImportServiceImpl - CcdiIntermediaryEntityImportServiceImpl
This commit is contained in:
1
doc/test-data/purchase_transaction/node_modules/binary/.npmignore
generated
vendored
Normal file
1
doc/test-data/purchase_transaction/node_modules/binary/.npmignore
generated
vendored
Normal file
@@ -0,0 +1 @@
|
||||
node_modules
|
||||
4
doc/test-data/purchase_transaction/node_modules/binary/.travis.yml
generated
vendored
Normal file
4
doc/test-data/purchase_transaction/node_modules/binary/.travis.yml
generated
vendored
Normal file
@@ -0,0 +1,4 @@
|
||||
language: node_js
|
||||
node_js:
|
||||
- 0.4
|
||||
- 0.6
|
||||
177
doc/test-data/purchase_transaction/node_modules/binary/README.markdown
generated
vendored
Normal file
177
doc/test-data/purchase_transaction/node_modules/binary/README.markdown
generated
vendored
Normal file
@@ -0,0 +1,177 @@
|
||||
binary
|
||||
======
|
||||
|
||||
Unpack multibyte binary values from buffers and streams.
|
||||
You can specify the endianness and signedness of the fields to be unpacked too.
|
||||
|
||||
This module is a cleaner and more complete version of
|
||||
[bufferlist](https://github.com/substack/node-bufferlist)'s binary module that
|
||||
runs on pre-allocated buffers instead of a linked list.
|
||||
|
||||
[](http://travis-ci.org/substack/node-binary)
|
||||
|
||||
examples
|
||||
========
|
||||
|
||||
stream.js
|
||||
---------
|
||||
|
||||
``` js
|
||||
var binary = require('binary');
|
||||
|
||||
var ws = binary()
|
||||
.word32lu('x')
|
||||
.word16bs('y')
|
||||
.word16bu('z')
|
||||
.tap(function (vars) {
|
||||
console.dir(vars);
|
||||
})
|
||||
;
|
||||
process.stdin.pipe(ws);
|
||||
process.stdin.resume();
|
||||
```
|
||||
|
||||
output:
|
||||
|
||||
```
|
||||
$ node examples/stream.js
|
||||
abcdefgh
|
||||
{ x: 1684234849, y: 25958, z: 26472 }
|
||||
^D
|
||||
```
|
||||
|
||||
parse.js
|
||||
--------
|
||||
|
||||
``` js
|
||||
var buf = new Buffer([ 97, 98, 99, 100, 101, 102, 0 ]);
|
||||
|
||||
var binary = require('binary');
|
||||
var vars = binary.parse(buf)
|
||||
.word16ls('ab')
|
||||
.word32bu('cf')
|
||||
.word8('x')
|
||||
.vars
|
||||
;
|
||||
console.dir(vars);
|
||||
```
|
||||
|
||||
output:
|
||||
|
||||
```
|
||||
{ ab: 25185, cf: 1667523942, x: 0 }
|
||||
```
|
||||
|
||||
methods
|
||||
=======
|
||||
|
||||
`var binary = require('binary')`
|
||||
|
||||
var b = binary()
|
||||
----------------
|
||||
|
||||
Return a new writable stream `b` that has the chainable methods documented below
|
||||
for buffering binary input.
|
||||
|
||||
binary.parse(buf)
|
||||
-----------------
|
||||
|
||||
Parse a static buffer in one pass. Returns a chainable interface with the
|
||||
methods below plus a `vars` field to get at the variable stash as the last item
|
||||
in a chain.
|
||||
|
||||
In parse mode, methods will set their keys to `null` if the buffer isn't big
|
||||
enough except `buffer()` and `scan()` which read up up to the end of the buffer
|
||||
and stop.
|
||||
|
||||
b.word{8,16,32,64}{l,b}{e,u,s}(key)
|
||||
-----------------------------------
|
||||
|
||||
Parse bytes in the buffer or stream given:
|
||||
|
||||
* number of bits
|
||||
* endianness ( l : little, b : big ),
|
||||
* signedness ( u and e : unsigned, s : signed )
|
||||
|
||||
These functions won't start parsing until all previous parser functions have run
|
||||
and the data is available.
|
||||
|
||||
The result of the parse goes into the variable stash at `key`.
|
||||
If `key` has dots (`.`s), it refers to a nested address. If parent container
|
||||
values don't exist they will be created automatically, so for instance you can
|
||||
assign into `dst.addr` and `dst.port` and the `dst` key in the variable stash
|
||||
will be `{ addr : x, port : y }` afterwards.
|
||||
|
||||
b.buffer(key, size)
|
||||
-------------------
|
||||
|
||||
Take `size` bytes directly off the buffer stream, putting the resulting buffer
|
||||
slice in the variable stash at `key`. If `size` is a string, use the value at
|
||||
`vars[size]`. The key follows the same dotted address rules as the word
|
||||
functions.
|
||||
|
||||
b.scan(key, buffer)
|
||||
-------------------
|
||||
|
||||
Search for `buffer` in the stream and store all the intervening data in the
|
||||
stash at at `key`, excluding the search buffer. If `buffer` passed as a string,
|
||||
it will be converted into a Buffer internally.
|
||||
|
||||
For example, to read in a line you can just do:
|
||||
|
||||
``` js
|
||||
var b = binary()
|
||||
.scan('line', new Buffer('\r\n'))
|
||||
.tap(function (vars) {
|
||||
console.log(vars.line)
|
||||
})
|
||||
;
|
||||
stream.pipe(b);
|
||||
```
|
||||
|
||||
b.tap(cb)
|
||||
---------
|
||||
|
||||
The callback `cb` is provided with the variable stash from all the previous
|
||||
actions once they've all finished.
|
||||
|
||||
You can nest additional actions onto `this` inside the callback.
|
||||
|
||||
b.into(key, cb)
|
||||
---------------
|
||||
|
||||
Like `.tap()`, except all nested actions will assign into a `key` in the `vars`
|
||||
stash.
|
||||
|
||||
b.loop(cb)
|
||||
----------
|
||||
|
||||
Loop, each time calling `cb(end, vars)` for function `end` and the variable
|
||||
stash with `this` set to a new chain for nested parsing. The loop terminates
|
||||
once `end` is called.
|
||||
|
||||
b.flush()
|
||||
---------
|
||||
|
||||
Clear the variable stash entirely.
|
||||
|
||||
installation
|
||||
============
|
||||
|
||||
To install with [npm](http://github.com/isaacs/npm):
|
||||
|
||||
```
|
||||
npm install binary
|
||||
```
|
||||
|
||||
notes
|
||||
=====
|
||||
|
||||
The word64 functions will only return approximations since javascript uses ieee
|
||||
floating point for all number types. Mind the loss of precision.
|
||||
|
||||
license
|
||||
=======
|
||||
|
||||
MIT
|
||||
|
||||
11
doc/test-data/purchase_transaction/node_modules/binary/example/buf.js
generated
vendored
Normal file
11
doc/test-data/purchase_transaction/node_modules/binary/example/buf.js
generated
vendored
Normal file
@@ -0,0 +1,11 @@
|
||||
var buf = new Buffer([ 97, 98, 99, 100, 101, 102, 0 ]);
|
||||
|
||||
var binary = require('binary');
|
||||
binary(buf)
|
||||
.word16ls('ab')
|
||||
.word32bu('cf')
|
||||
.word8('x')
|
||||
.tap(function (vars) {
|
||||
console.dir(vars);
|
||||
})
|
||||
;
|
||||
10
doc/test-data/purchase_transaction/node_modules/binary/example/parse.js
generated
vendored
Normal file
10
doc/test-data/purchase_transaction/node_modules/binary/example/parse.js
generated
vendored
Normal file
@@ -0,0 +1,10 @@
|
||||
var buf = new Buffer([ 97, 98, 99, 100, 101, 102, 0 ]);
|
||||
|
||||
var binary = require('binary');
|
||||
var vars = binary.parse(buf)
|
||||
.word16ls('ab')
|
||||
.word32bu('cf')
|
||||
.word8('x')
|
||||
.vars
|
||||
;
|
||||
console.dir(vars);
|
||||
12
doc/test-data/purchase_transaction/node_modules/binary/example/stream.js
generated
vendored
Normal file
12
doc/test-data/purchase_transaction/node_modules/binary/example/stream.js
generated
vendored
Normal file
@@ -0,0 +1,12 @@
|
||||
var binary = require('binary');
|
||||
|
||||
var ws = binary()
|
||||
.word32lu('x')
|
||||
.word16bs('y')
|
||||
.word16bu('z')
|
||||
.tap(function (vars) {
|
||||
console.dir(vars);
|
||||
})
|
||||
;
|
||||
process.stdin.pipe(ws);
|
||||
process.stdin.resume();
|
||||
397
doc/test-data/purchase_transaction/node_modules/binary/index.js
generated
vendored
Normal file
397
doc/test-data/purchase_transaction/node_modules/binary/index.js
generated
vendored
Normal file
@@ -0,0 +1,397 @@
|
||||
var Chainsaw = require('chainsaw');
|
||||
var EventEmitter = require('events').EventEmitter;
|
||||
var Buffers = require('buffers');
|
||||
var Vars = require('./lib/vars.js');
|
||||
var Stream = require('stream').Stream;
|
||||
|
||||
exports = module.exports = function (bufOrEm, eventName) {
|
||||
if (Buffer.isBuffer(bufOrEm)) {
|
||||
return exports.parse(bufOrEm);
|
||||
}
|
||||
|
||||
var s = exports.stream();
|
||||
if (bufOrEm && bufOrEm.pipe) {
|
||||
bufOrEm.pipe(s);
|
||||
}
|
||||
else if (bufOrEm) {
|
||||
bufOrEm.on(eventName || 'data', function (buf) {
|
||||
s.write(buf);
|
||||
});
|
||||
|
||||
bufOrEm.on('end', function () {
|
||||
s.end();
|
||||
});
|
||||
}
|
||||
return s;
|
||||
};
|
||||
|
||||
exports.stream = function (input) {
|
||||
if (input) return exports.apply(null, arguments);
|
||||
|
||||
var pending = null;
|
||||
function getBytes (bytes, cb, skip) {
|
||||
pending = {
|
||||
bytes : bytes,
|
||||
skip : skip,
|
||||
cb : function (buf) {
|
||||
pending = null;
|
||||
cb(buf);
|
||||
},
|
||||
};
|
||||
dispatch();
|
||||
}
|
||||
|
||||
var offset = null;
|
||||
function dispatch () {
|
||||
if (!pending) {
|
||||
if (caughtEnd) done = true;
|
||||
return;
|
||||
}
|
||||
if (typeof pending === 'function') {
|
||||
pending();
|
||||
}
|
||||
else {
|
||||
var bytes = offset + pending.bytes;
|
||||
|
||||
if (buffers.length >= bytes) {
|
||||
var buf;
|
||||
if (offset == null) {
|
||||
buf = buffers.splice(0, bytes);
|
||||
if (!pending.skip) {
|
||||
buf = buf.slice();
|
||||
}
|
||||
}
|
||||
else {
|
||||
if (!pending.skip) {
|
||||
buf = buffers.slice(offset, bytes);
|
||||
}
|
||||
offset = bytes;
|
||||
}
|
||||
|
||||
if (pending.skip) {
|
||||
pending.cb();
|
||||
}
|
||||
else {
|
||||
pending.cb(buf);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function builder (saw) {
|
||||
function next () { if (!done) saw.next() }
|
||||
|
||||
var self = words(function (bytes, cb) {
|
||||
return function (name) {
|
||||
getBytes(bytes, function (buf) {
|
||||
vars.set(name, cb(buf));
|
||||
next();
|
||||
});
|
||||
};
|
||||
});
|
||||
|
||||
self.tap = function (cb) {
|
||||
saw.nest(cb, vars.store);
|
||||
};
|
||||
|
||||
self.into = function (key, cb) {
|
||||
if (!vars.get(key)) vars.set(key, {});
|
||||
var parent = vars;
|
||||
vars = Vars(parent.get(key));
|
||||
|
||||
saw.nest(function () {
|
||||
cb.apply(this, arguments);
|
||||
this.tap(function () {
|
||||
vars = parent;
|
||||
});
|
||||
}, vars.store);
|
||||
};
|
||||
|
||||
self.flush = function () {
|
||||
vars.store = {};
|
||||
next();
|
||||
};
|
||||
|
||||
self.loop = function (cb) {
|
||||
var end = false;
|
||||
|
||||
saw.nest(false, function loop () {
|
||||
this.vars = vars.store;
|
||||
cb.call(this, function () {
|
||||
end = true;
|
||||
next();
|
||||
}, vars.store);
|
||||
this.tap(function () {
|
||||
if (end) saw.next()
|
||||
else loop.call(this)
|
||||
}.bind(this));
|
||||
}, vars.store);
|
||||
};
|
||||
|
||||
self.buffer = function (name, bytes) {
|
||||
if (typeof bytes === 'string') {
|
||||
bytes = vars.get(bytes);
|
||||
}
|
||||
|
||||
getBytes(bytes, function (buf) {
|
||||
vars.set(name, buf);
|
||||
next();
|
||||
});
|
||||
};
|
||||
|
||||
self.skip = function (bytes) {
|
||||
if (typeof bytes === 'string') {
|
||||
bytes = vars.get(bytes);
|
||||
}
|
||||
|
||||
getBytes(bytes, function () {
|
||||
next();
|
||||
});
|
||||
};
|
||||
|
||||
self.scan = function find (name, search) {
|
||||
if (typeof search === 'string') {
|
||||
search = new Buffer(search);
|
||||
}
|
||||
else if (!Buffer.isBuffer(search)) {
|
||||
throw new Error('search must be a Buffer or a string');
|
||||
}
|
||||
|
||||
var taken = 0;
|
||||
pending = function () {
|
||||
var pos = buffers.indexOf(search, offset + taken);
|
||||
var i = pos-offset-taken;
|
||||
if (pos !== -1) {
|
||||
pending = null;
|
||||
if (offset != null) {
|
||||
vars.set(
|
||||
name,
|
||||
buffers.slice(offset, offset + taken + i)
|
||||
);
|
||||
offset += taken + i + search.length;
|
||||
}
|
||||
else {
|
||||
vars.set(
|
||||
name,
|
||||
buffers.slice(0, taken + i)
|
||||
);
|
||||
buffers.splice(0, taken + i + search.length);
|
||||
}
|
||||
next();
|
||||
dispatch();
|
||||
} else {
|
||||
i = Math.max(buffers.length - search.length - offset - taken, 0);
|
||||
}
|
||||
taken += i;
|
||||
};
|
||||
dispatch();
|
||||
};
|
||||
|
||||
self.peek = function (cb) {
|
||||
offset = 0;
|
||||
saw.nest(function () {
|
||||
cb.call(this, vars.store);
|
||||
this.tap(function () {
|
||||
offset = null;
|
||||
});
|
||||
});
|
||||
};
|
||||
|
||||
return self;
|
||||
};
|
||||
|
||||
var stream = Chainsaw.light(builder);
|
||||
stream.writable = true;
|
||||
|
||||
var buffers = Buffers();
|
||||
|
||||
stream.write = function (buf) {
|
||||
buffers.push(buf);
|
||||
dispatch();
|
||||
};
|
||||
|
||||
var vars = Vars();
|
||||
|
||||
var done = false, caughtEnd = false;
|
||||
stream.end = function () {
|
||||
caughtEnd = true;
|
||||
};
|
||||
|
||||
stream.pipe = Stream.prototype.pipe;
|
||||
Object.getOwnPropertyNames(EventEmitter.prototype).forEach(function (name) {
|
||||
stream[name] = EventEmitter.prototype[name];
|
||||
});
|
||||
|
||||
return stream;
|
||||
};
|
||||
|
||||
exports.parse = function parse (buffer) {
|
||||
var self = words(function (bytes, cb) {
|
||||
return function (name) {
|
||||
if (offset + bytes <= buffer.length) {
|
||||
var buf = buffer.slice(offset, offset + bytes);
|
||||
offset += bytes;
|
||||
vars.set(name, cb(buf));
|
||||
}
|
||||
else {
|
||||
vars.set(name, null);
|
||||
}
|
||||
return self;
|
||||
};
|
||||
});
|
||||
|
||||
var offset = 0;
|
||||
var vars = Vars();
|
||||
self.vars = vars.store;
|
||||
|
||||
self.tap = function (cb) {
|
||||
cb.call(self, vars.store);
|
||||
return self;
|
||||
};
|
||||
|
||||
self.into = function (key, cb) {
|
||||
if (!vars.get(key)) {
|
||||
vars.set(key, {});
|
||||
}
|
||||
var parent = vars;
|
||||
vars = Vars(parent.get(key));
|
||||
cb.call(self, vars.store);
|
||||
vars = parent;
|
||||
return self;
|
||||
};
|
||||
|
||||
self.loop = function (cb) {
|
||||
var end = false;
|
||||
var ender = function () { end = true };
|
||||
while (end === false) {
|
||||
cb.call(self, ender, vars.store);
|
||||
}
|
||||
return self;
|
||||
};
|
||||
|
||||
self.buffer = function (name, size) {
|
||||
if (typeof size === 'string') {
|
||||
size = vars.get(size);
|
||||
}
|
||||
var buf = buffer.slice(offset, Math.min(buffer.length, offset + size));
|
||||
offset += size;
|
||||
vars.set(name, buf);
|
||||
|
||||
return self;
|
||||
};
|
||||
|
||||
self.skip = function (bytes) {
|
||||
if (typeof bytes === 'string') {
|
||||
bytes = vars.get(bytes);
|
||||
}
|
||||
offset += bytes;
|
||||
|
||||
return self;
|
||||
};
|
||||
|
||||
self.scan = function (name, search) {
|
||||
if (typeof search === 'string') {
|
||||
search = new Buffer(search);
|
||||
}
|
||||
else if (!Buffer.isBuffer(search)) {
|
||||
throw new Error('search must be a Buffer or a string');
|
||||
}
|
||||
vars.set(name, null);
|
||||
|
||||
// simple but slow string search
|
||||
for (var i = 0; i + offset <= buffer.length - search.length + 1; i++) {
|
||||
for (
|
||||
var j = 0;
|
||||
j < search.length && buffer[offset+i+j] === search[j];
|
||||
j++
|
||||
);
|
||||
if (j === search.length) break;
|
||||
}
|
||||
|
||||
vars.set(name, buffer.slice(offset, offset + i));
|
||||
offset += i + search.length;
|
||||
return self;
|
||||
};
|
||||
|
||||
self.peek = function (cb) {
|
||||
var was = offset;
|
||||
cb.call(self, vars.store);
|
||||
offset = was;
|
||||
return self;
|
||||
};
|
||||
|
||||
self.flush = function () {
|
||||
vars.store = {};
|
||||
return self;
|
||||
};
|
||||
|
||||
self.eof = function () {
|
||||
return offset >= buffer.length;
|
||||
};
|
||||
|
||||
return self;
|
||||
};
|
||||
|
||||
// convert byte strings to unsigned little endian numbers
|
||||
function decodeLEu (bytes) {
|
||||
var acc = 0;
|
||||
for (var i = 0; i < bytes.length; i++) {
|
||||
acc += Math.pow(256,i) * bytes[i];
|
||||
}
|
||||
return acc;
|
||||
}
|
||||
|
||||
// convert byte strings to unsigned big endian numbers
|
||||
function decodeBEu (bytes) {
|
||||
var acc = 0;
|
||||
for (var i = 0; i < bytes.length; i++) {
|
||||
acc += Math.pow(256, bytes.length - i - 1) * bytes[i];
|
||||
}
|
||||
return acc;
|
||||
}
|
||||
|
||||
// convert byte strings to signed big endian numbers
|
||||
function decodeBEs (bytes) {
|
||||
var val = decodeBEu(bytes);
|
||||
if ((bytes[0] & 0x80) == 0x80) {
|
||||
val -= Math.pow(256, bytes.length);
|
||||
}
|
||||
return val;
|
||||
}
|
||||
|
||||
// convert byte strings to signed little endian numbers
|
||||
function decodeLEs (bytes) {
|
||||
var val = decodeLEu(bytes);
|
||||
if ((bytes[bytes.length - 1] & 0x80) == 0x80) {
|
||||
val -= Math.pow(256, bytes.length);
|
||||
}
|
||||
return val;
|
||||
}
|
||||
|
||||
function words (decode) {
|
||||
var self = {};
|
||||
|
||||
[ 1, 2, 4, 8 ].forEach(function (bytes) {
|
||||
var bits = bytes * 8;
|
||||
|
||||
self['word' + bits + 'le']
|
||||
= self['word' + bits + 'lu']
|
||||
= decode(bytes, decodeLEu);
|
||||
|
||||
self['word' + bits + 'ls']
|
||||
= decode(bytes, decodeLEs);
|
||||
|
||||
self['word' + bits + 'be']
|
||||
= self['word' + bits + 'bu']
|
||||
= decode(bytes, decodeBEu);
|
||||
|
||||
self['word' + bits + 'bs']
|
||||
= decode(bytes, decodeBEs);
|
||||
});
|
||||
|
||||
// word8be(n) == word8le(n) for all n
|
||||
self.word8 = self.word8u = self.word8be;
|
||||
self.word8s = self.word8bs;
|
||||
|
||||
return self;
|
||||
}
|
||||
28
doc/test-data/purchase_transaction/node_modules/binary/lib/vars.js
generated
vendored
Normal file
28
doc/test-data/purchase_transaction/node_modules/binary/lib/vars.js
generated
vendored
Normal file
@@ -0,0 +1,28 @@
|
||||
module.exports = function (store) {
|
||||
function getset (name, value) {
|
||||
var node = vars.store;
|
||||
var keys = name.split('.');
|
||||
keys.slice(0,-1).forEach(function (k) {
|
||||
if (node[k] === undefined) node[k] = {};
|
||||
node = node[k]
|
||||
});
|
||||
var key = keys[keys.length - 1];
|
||||
if (arguments.length == 1) {
|
||||
return node[key];
|
||||
}
|
||||
else {
|
||||
return node[key] = value;
|
||||
}
|
||||
}
|
||||
|
||||
var vars = {
|
||||
get : function (name) {
|
||||
return getset(name);
|
||||
},
|
||||
set : function (name, value) {
|
||||
return getset(name, value);
|
||||
},
|
||||
store : store || {},
|
||||
};
|
||||
return vars;
|
||||
};
|
||||
38
doc/test-data/purchase_transaction/node_modules/binary/package.json
generated
vendored
Normal file
38
doc/test-data/purchase_transaction/node_modules/binary/package.json
generated
vendored
Normal file
@@ -0,0 +1,38 @@
|
||||
{
|
||||
"name" : "binary",
|
||||
"version" : "0.3.0",
|
||||
"description" : "Unpack multibyte binary values from buffers",
|
||||
"main" : "./index.js",
|
||||
"repository" : {
|
||||
"type" : "git",
|
||||
"url" : "http://github.com/substack/node-binary.git"
|
||||
},
|
||||
"keywords": [
|
||||
"binary",
|
||||
"decode",
|
||||
"endian",
|
||||
"unpack",
|
||||
"signed",
|
||||
"unsigned"
|
||||
],
|
||||
"author" : {
|
||||
"name" : "James Halliday",
|
||||
"email" : "mail@substack.net",
|
||||
"url" : "http://substack.net"
|
||||
},
|
||||
"dependencies" : {
|
||||
"chainsaw" : "~0.1.0",
|
||||
"buffers" : "~0.1.1"
|
||||
},
|
||||
"devDependencies" : {
|
||||
"seq" : "~0.2.5",
|
||||
"tap" : "~0.2.4"
|
||||
},
|
||||
"scripts" : {
|
||||
"test" : "tap test/*.js"
|
||||
},
|
||||
"license" : "MIT",
|
||||
"engine" : {
|
||||
"node" : ">=0.4.0"
|
||||
}
|
||||
}
|
||||
92
doc/test-data/purchase_transaction/node_modules/binary/perf/loop.js
generated
vendored
Normal file
92
doc/test-data/purchase_transaction/node_modules/binary/perf/loop.js
generated
vendored
Normal file
@@ -0,0 +1,92 @@
|
||||
var Seq = require('seq');
|
||||
var Hash = require('hashish');
|
||||
var EventEmitter = require('events').EventEmitter;
|
||||
|
||||
var Bin = require('binary');
|
||||
var Buf = require('bufferlist/binary');
|
||||
var BufferList = require('bufferlist');
|
||||
|
||||
console.log('loop');
|
||||
function emitter () {
|
||||
var em = new EventEmitter;
|
||||
|
||||
var i = 0;
|
||||
var iv = setInterval(function () {
|
||||
var buf = new Buffer(10000);
|
||||
buf[0] = 0xff;
|
||||
|
||||
if (++ i >= 2000) {
|
||||
buf[0] = 0;
|
||||
clearInterval(iv);
|
||||
}
|
||||
em.emit('data', buf);
|
||||
}, 1);
|
||||
|
||||
return em;
|
||||
}
|
||||
|
||||
Seq()
|
||||
.seq(function () {
|
||||
var next = this.bind({}, null);
|
||||
bufferlist(next);
|
||||
})
|
||||
.seq(function () {
|
||||
var next = this.bind({}, null);
|
||||
binary(next);
|
||||
})
|
||||
;
|
||||
|
||||
function binary (next) {
|
||||
var em = emitter();
|
||||
var t0 = Date.now();
|
||||
|
||||
Bin(em)
|
||||
.loop(function (end) {
|
||||
this
|
||||
.word8('x')
|
||||
.word8('y')
|
||||
.word32be('z')
|
||||
.word32le('w')
|
||||
.buffer('buf', 10000 - 10)
|
||||
.tap(function (vars) {
|
||||
if (vars.x === 0) {
|
||||
var tf = Date.now();
|
||||
console.log(' binary: ' + (tf - t0) + ' ms');
|
||||
end();
|
||||
setTimeout(next, 20);
|
||||
}
|
||||
})
|
||||
})
|
||||
;
|
||||
}
|
||||
|
||||
function bufferlist (next) {
|
||||
var em = emitter();
|
||||
var t0 = Date.now();
|
||||
|
||||
var blist = new BufferList;
|
||||
em.on('data', function (buf) {
|
||||
blist.push(buf);
|
||||
});
|
||||
|
||||
Buf(blist)
|
||||
.forever(function () {
|
||||
var top = this;
|
||||
this
|
||||
.getWord8('x')
|
||||
.getWord8('y')
|
||||
.getWord32be('z')
|
||||
.getWord32le('w')
|
||||
.getBuffer('buf', 10000 - 10)
|
||||
.tap(function (vars) {
|
||||
if (vars.x === 0) {
|
||||
var tf = Date.now();
|
||||
console.log(' bufferlist: ' + (tf - t0) + ' ms');
|
||||
top.exit();
|
||||
setTimeout(next, 20);
|
||||
}
|
||||
})
|
||||
})
|
||||
.end()
|
||||
;
|
||||
}
|
||||
80
doc/test-data/purchase_transaction/node_modules/binary/perf/small.js
generated
vendored
Normal file
80
doc/test-data/purchase_transaction/node_modules/binary/perf/small.js
generated
vendored
Normal file
@@ -0,0 +1,80 @@
|
||||
var Seq = require('seq');
|
||||
var Hash = require('hashish');
|
||||
|
||||
var Bin = require('binary');
|
||||
var Buf = require('bufferlist/binary');
|
||||
var BufferList = require('bufferlist');
|
||||
var EventEmitter = require('events').EventEmitter;
|
||||
|
||||
function binary (buf, cb) {
|
||||
Bin(buf)
|
||||
.word32le('x')
|
||||
.word16be('y')
|
||||
.word16be('z')
|
||||
.word32le('w')
|
||||
.tap(cb)
|
||||
;
|
||||
};
|
||||
|
||||
function stream (buf, cb) {
|
||||
var em = new EventEmitter;
|
||||
Bin(em)
|
||||
.word32le('x')
|
||||
.word16be('y')
|
||||
.word16be('z')
|
||||
.word32le('w')
|
||||
.tap(cb)
|
||||
;
|
||||
em.emit('data', buf);
|
||||
};
|
||||
|
||||
function parse (buf, cb) {
|
||||
cb(Bin.parse(buf)
|
||||
.word32le('x')
|
||||
.word16be('y')
|
||||
.word16be('z')
|
||||
.word32le('w')
|
||||
.vars
|
||||
);
|
||||
};
|
||||
|
||||
function bufferlist (buf, cb) {
|
||||
var blist = new BufferList;
|
||||
blist.push(buf);
|
||||
Buf(blist)
|
||||
.getWord32le('x')
|
||||
.getWord16be('y')
|
||||
.getWord16be('z')
|
||||
.getWord32le('w')
|
||||
.tap(cb)
|
||||
.end()
|
||||
;
|
||||
};
|
||||
|
||||
|
||||
var buffers = [];
|
||||
for (var i = 0; i < 200; i++) {
|
||||
buffers.push(new Buffer(12));
|
||||
}
|
||||
|
||||
console.log('small');
|
||||
Seq(binary, stream, parse, bufferlist)
|
||||
.seqEach(function (f) {
|
||||
var t = this;
|
||||
var t0 = Date.now();
|
||||
Seq()
|
||||
.extend(buffers)
|
||||
.seqEach(function (buf) {
|
||||
f(buf, this.bind(this, null));
|
||||
})
|
||||
.seq(function () {
|
||||
var tf = Date.now();
|
||||
console.log(' ' + f.name + ': ' + (tf - t0));
|
||||
t(null);
|
||||
})
|
||||
;
|
||||
})
|
||||
.seq(function () {
|
||||
this(null);
|
||||
})
|
||||
;
|
||||
Reference in New Issue
Block a user