| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583 |
- // Experimental
- var numpy = numpy || {};
- var python = python || require('./python');
- numpy.ModelFactory = class {
- match(context) {
- const stream = context.stream;
- const signature = [ 0x93, 0x4E, 0x55, 0x4D, 0x50, 0x59 ];
- if (stream && signature.length <= stream.length && stream.peek(signature.length).every((value, index) => value === signature[index])) {
- return { name: 'npy' };
- }
- const entries = context.entries('zip');
- if (entries.size > 0 && Array.from(entries.keys()).every((name) => name.endsWith('.npy'))) {
- return { name: 'npz', value: entries };
- }
- const obj = context.open('pkl');
- if (obj) {
- if (numpy.Utility.isTensor(obj)) {
- return { name: 'numpy.ndarray', value: obj };
- }
- if (Array.isArray(obj) && obj.every((obj) => obj && obj.__class__ && obj.__class__.__name__ === 'Network' && (obj.__class__.__module__ === 'dnnlib.tflib.network' || obj.__class__.__module__ === 'tfutil'))) {
- return { name: 'dnnlib.tflib.network', value: obj };
- }
- const weights = numpy.Utility.weights(obj);
- if (weights) {
- return { name: 'pickle', value: weights };
- }
- }
- return undefined;
- }
- open(context, match) {
- let format = '';
- const graphs = [];
- switch (match.name) {
- case 'npy': {
- format = 'NumPy Array';
- const execution = new python.Execution(null);
- const stream = context.stream;
- const buffer = stream.peek();
- const bytes = execution.invoke('io.BytesIO', [ buffer ]);
- const array = execution.invoke('numpy.load', [ bytes ]);
- const layer = { type: 'numpy.ndarray', parameters: [ { name: 'value', tensor: { name: '', array: array } } ] };
- graphs.push({ layers: [ layer ] });
- break;
- }
- case 'npz': {
- format = 'NumPy Zip';
- const layers = new Map();
- const execution = new python.Execution(null);
- for (const entry of match.value) {
- if (!entry[0].endsWith('.npy')) {
- throw new numpy.Error("Invalid file name '" + entry.name + "'.");
- }
- const name = entry[0].replace(/\.npy$/, '');
- const parts = name.split('/');
- const parameterName = parts.pop();
- const groupName = parts.join('/');
- if (!layers.has(groupName)) {
- layers.set(groupName, { name: groupName, parameters: [] });
- }
- const layer = layers.get(groupName);
- const stream = entry[1];
- const buffer = stream.peek();
- const bytes = execution.invoke('io.BytesIO', [ buffer ]);
- let array = execution.invoke('numpy.load', [ bytes ]);
- if (array.dtype.byteorder === '|' && array.dtype.itemsize !== 1) {
- if (array.dtype.kind !== 'O') {
- throw new numpy.Error("Invalid data type '" + array.dataType + "'.");
- }
- const unpickler = python.Unpickler.open(array.data, execution);
- array = unpickler.load();
- }
- layer.parameters.push({
- name: parameterName,
- tensor: { name: name, array: array }
- });
- }
- graphs.push({ layers: Array.from(layers.values()) });
- break;
- }
- case 'pickle': {
- format = 'NumPy Weights';
- const layers = new Map();
- const weights = match.value;
- let separator = '.';
- if (Array.from(weights.keys()).filter((key) => key.indexOf('_') !== -1) &&
- Array.from(weights.keys()).every((key) => key.indexOf('_') > key.indexOf('.'))) {
- separator = '_';
- }
- for (const pair of weights) {
- const name = pair[0];
- const array = pair[1];
- const parts = name.split(separator);
- const parameterName = parts.length > 1 ? parts.pop() : '?';
- const layerName = parts.join(separator);
- if (!layers.has(layerName)) {
- layers.set(layerName, { name: layerName, parameters: [] });
- }
- const layer = layers.get(layerName);
- layer.parameters.push({
- name: parameterName,
- tensor: { name: name, array: array }
- });
- }
- graphs.push({ layers: Array.from(layers.values()) });
- break;
- }
- case 'numpy.ndarray': {
- format = 'NumPy NDArray';
- const layer = {
- type: 'numpy.ndarray',
- parameters: [ { name: 'value', tensor: { name: '', array: match.value } } ]
- };
- graphs.push({ layers: [ layer ] });
- break;
- }
- case 'dnnlib.tflib.network': {
- format = 'dnnlib';
- for (const obj of match.value) {
- const layers = new Map();
- for (const entry of obj.variables) {
- const name = entry[0];
- const value = entry[1];
- if (numpy.Utility.isTensor(value)) {
- const parts = name.split('/');
- const parameterName = parts.length > 1 ? parts.pop() : '?';
- const layerName = parts.join('/');
- if (!layers.has(layerName)) {
- layers.set(layerName, { name: layerName, parameters: [] });
- }
- const layer = layers.get(layerName);
- layer.parameters.push({
- name: parameterName,
- tensor: { name: name, array: value }
- });
- }
- }
- graphs.push({ name: obj.name, layers: Array.from(layers.values()) });
- }
- break;
- }
- default: {
- throw new numpy.Error("Unsupported NumPy format '" + match.name + "'.");
- }
- }
- const model = new numpy.Model(format, graphs);
- return Promise.resolve(model);
- }
- };
- numpy.Model = class {
- constructor(format, graphs) {
- this._format = format;
- this._graphs = graphs.map((graph) => new numpy.Graph(graph));
- }
- get format() {
- return this._format;
- }
- get graphs() {
- return this._graphs;
- }
- };
- numpy.Graph = class {
- constructor(graph) {
- this._name = graph.name || '';
- this._nodes = graph.layers.map((layer) => new numpy.Node(layer));
- }
- get name() {
- return this._name;
- }
- get inputs() {
- return [];
- }
- get outputs() {
- return [];
- }
- get nodes() {
- return this._nodes;
- }
- };
- numpy.Parameter = class {
- constructor(name, args) {
- this._name = name;
- this._arguments = args;
- }
- get name() {
- return this._name;
- }
- get visible() {
- return true;
- }
- get arguments() {
- return this._arguments;
- }
- };
- numpy.Argument = class {
- constructor(name, initializer) {
- if (typeof name !== 'string') {
- throw new numpy.Error("Invalid argument identifier '" + JSON.stringify(name) + "'.");
- }
- this._name = name;
- this._initializer = initializer || null;
- }
- get name() {
- return this._name;
- }
- get type() {
- return this._initializer.type;
- }
- get initializer() {
- return this._initializer;
- }
- };
- numpy.Node = class {
- constructor(layer) {
- this._name = layer.name || '';
- this._type = { name: layer.type || '{}' };
- this._inputs = [];
- for (const parameter of layer.parameters) {
- const initializer = new numpy.Tensor(parameter.tensor.array);
- this._inputs.push(new numpy.Parameter(parameter.name, [
- new numpy.Argument(parameter.tensor.name || '', initializer)
- ]));
- }
- }
- get type() {
- return this._type;
- }
- get name() {
- return this._name;
- }
- get inputs() {
- return this._inputs;
- }
- get outputs() {
- return [];
- }
- get attributes() {
- return [];
- }
- };
- numpy.Tensor = class {
- constructor(array) {
- this._type = new numpy.TensorType(array.dtype.__name__, new numpy.TensorShape(array.shape));
- this._data = array.tobytes();
- this._byteorder = array.dtype.byteorder;
- this._itemsize = array.dtype.itemsize;
- }
- get type(){
- return this._type;
- }
- get state() {
- return this._context().state;
- }
- get value() {
- const context = this._context();
- if (context.state) {
- return null;
- }
- context.limit = Number.MAX_SAFE_INTEGER;
- return this._decode(context, 0);
- }
- toString() {
- const context = this._context();
- if (context.state) {
- return '';
- }
- context.limit = 10000;
- const value = this._decode(context, 0);
- return numpy.Tensor._stringify(value, '', ' ');
- }
- _context() {
- const context = {};
- context.index = 0;
- context.count = 0;
- context.state = null;
- if (this._byteorder !== '<' && this._byteorder !== '>' && this._type.dataType !== 'uint8' && this._type.dataType !== 'int8') {
- context.state = 'Tensor byte order is not supported.';
- return context;
- }
- if (!this._data || this._data.length == 0) {
- context.state = 'Tensor data is empty.';
- return context;
- }
- context.itemSize = this._itemsize;
- context.dimensions = this._type.shape.dimensions;
- context.dataType = this._type.dataType;
- context.littleEndian = this._byteorder == '<';
- context.data = this._data;
- context.rawData = new DataView(this._data.buffer, this._data.byteOffset, this._data.byteLength);
- return context;
- }
- _decode(context, dimension) {
- const littleEndian = context.littleEndian;
- const shape = context.dimensions.length == 0 ? [ 1 ] : context.dimensions;
- const results = [];
- const size = shape[dimension];
- if (dimension == shape.length - 1) {
- for (let i = 0; i < size; i++) {
- if (context.count > context.limit) {
- results.push('...');
- return results;
- }
- if (context.rawData) {
- switch (context.dataType) {
- case 'float16':
- results.push(context.rawData.getFloat16(context.index, littleEndian));
- break;
- case 'float32':
- results.push(context.rawData.getFloat32(context.index, littleEndian));
- break;
- case 'float64':
- results.push(context.rawData.getFloat64(context.index, littleEndian));
- break;
- case 'int8':
- results.push(context.rawData.getInt8(context.index, littleEndian));
- break;
- case 'int16':
- results.push(context.rawData.getInt16(context.index, littleEndian));
- break;
- case 'int32':
- results.push(context.rawData.getInt32(context.index, littleEndian));
- break;
- case 'int64':
- results.push(context.rawData.getInt64(context.index, littleEndian));
- break;
- case 'uint8':
- results.push(context.rawData.getUint8(context.index, littleEndian));
- break;
- case 'uint16':
- results.push(context.rawData.getUint16(context.index, littleEndian));
- break;
- case 'uint32':
- results.push(context.rawData.getUint32(context.index, littleEndian));
- break;
- default:
- throw new numpy.Error("Unsupported tensor data type '" + context.dataType + "'.");
- }
- context.index += context.itemSize;
- context.count++;
- }
- }
- }
- else {
- for (let j = 0; j < size; j++) {
- if (context.count > context.limit) {
- results.push('...');
- return results;
- }
- results.push(this._decode(context, dimension + 1));
- }
- }
- if (context.dimensions.length == 0) {
- return results[0];
- }
- return results;
- }
- static _stringify(value, indentation, indent) {
- if (Array.isArray(value)) {
- const result = [];
- result.push(indentation + '[');
- const items = value.map((item) => numpy.Tensor._stringify(item, indentation + indent, indent));
- if (items.length > 0) {
- result.push(items.join(',\n'));
- }
- result.push(indentation + ']');
- return result.join('\n');
- }
- if (typeof value == 'string') {
- return indentation + value;
- }
- if (value == Infinity) {
- return indentation + 'Infinity';
- }
- if (value == -Infinity) {
- return indentation + '-Infinity';
- }
- if (isNaN(value)) {
- return indentation + 'NaN';
- }
- return indentation + value.toString();
- }
- };
- numpy.TensorType = class {
- constructor(dataType, shape) {
- this._dataType = dataType;
- this._shape = shape;
- }
- get dataType() {
- return this._dataType || '?';
- }
- get shape() {
- return this._shape;
- }
- toString() {
- return this.dataType + this._shape.toString();
- }
- };
- numpy.TensorShape = class {
- constructor(dimensions) {
- this._dimensions = dimensions;
- }
- get dimensions() {
- return this._dimensions;
- }
- toString() {
- if (!this._dimensions || this._dimensions.length == 0) {
- return '';
- }
- return '[' + this._dimensions.join(',') + ']';
- }
- };
- numpy.Utility = class {
- static isTensor(obj) {
- return obj && obj.__class__ &&
- ((obj.__class__.__module__ === 'numpy' && obj.__class__.__name__ === 'ndarray') ||
- (obj.__class__.__module__ === 'numpy.core.memmap' && obj.__class__.__name__ === 'memmap'));
- }
- static weights(obj) {
- const dict = (obj, key) => {
- const dict = key === '' ? obj : obj[key];
- if (dict) {
- const weights = new Map();
- if (dict instanceof Map) {
- for (const pair of dict) {
- const key = pair[0];
- const obj = pair[1];
- if (numpy.Utility.isTensor(obj)) {
- weights.set(key, obj);
- continue;
- }
- else if (obj instanceof Map && Array.from(obj).every((pair) => numpy.Utility.isTensor(pair[1]))) {
- for (const pair of obj) {
- weights.set(key + '.' + pair[0], pair[1]);
- }
- continue;
- }
- else if (key === '_metadata') {
- continue;
- }
- return null;
- }
- return weights;
- }
- else if (!Array.isArray(dict)) {
- const set = new Set([ 'weight_order', 'lr', 'model_iter', '__class__' ]);
- for (const entry of Object.entries(dict)) {
- const key = entry[0];
- const value = entry[1];
- if (key) {
- if (numpy.Utility.isTensor(value)) {
- weights.set(key, value);
- continue;
- }
- if (set.has(key)) {
- continue;
- }
- if (value && !Array.isArray(value) && Object.entries(value).every((entry) => numpy.Utility.isTensor(entry[1]))) {
- const name = key;
- for (const entry of Object.entries(value)) {
- weights.set(name + '.' + entry[0], entry[1]);
- }
- continue;
- }
- }
- return null;
- }
- return weights;
- }
- }
- return null;
- };
- const list = (obj, key) => {
- const list = key === '' ? obj : obj[key];
- if (list && Array.isArray(list) && list.every((obj) => Object.entries(obj).every((entry) => numpy.Utility.isTensor(entry[1])))) {
- const weights = new Map();
- for (let i = 0; i < list.length; i++) {
- const obj = list[i];
- for (const entry of Object.entries(obj)) {
- weights.set(i.toString() + '.' + entry[0], entry[1]);
- }
- }
- return weights;
- }
- if (list && Array.isArray(list)) {
- const weights = new Map();
- for (let i = 0; i < list.length; i++) {
- const obj = list[i];
- if (numpy.Utility.isTensor(obj)) {
- weights.set(i.toString(), obj);
- continue;
- }
- else if (obj instanceof Map && Array.from(obj).every((pair) => numpy.Utility.isTensor(pair[1]))) {
- for (const pair of obj) {
- weights.set(i.toString() + '.' + pair[0], pair[1]);
- }
- continue;
- }
- return null;
- }
- return weights;
- }
- return null;
- };
- const keys = [ '', 'blobs', 'model' ];
- for (const key of keys) {
- const weights = dict(obj, key);
- if (weights && weights.size > 0) {
- return weights;
- }
- }
- for (const key of keys) {
- const weights = list(obj, key);
- if (weights) {
- return weights;
- }
- }
- return null;
- }
- };
- numpy.Error = class extends Error {
- constructor(message) {
- super(message);
- this.name = 'Error loading Chainer model.';
- }
- };
- if (typeof module !== 'undefined' && typeof module.exports === 'object') {
- module.exports.ModelFactory = numpy.ModelFactory;
- }
|