Add Fluence transport implementation

This commit is contained in:
Igor Krivchun 2019-09-20 18:08:12 +07:00
parent 3fa9f3ac5f
commit bf61103c18
5 changed files with 323 additions and 4 deletions

312
TransportFluence.js Normal file
View File

@ -0,0 +1,312 @@
const Url = require('url');
const Transport = require('./Transport'); // Base class for TransportXyz
const Transports = require('./Transports'); // Manage all Transports that are loaded
const errors = require('./Errors'); // Standard Dweb Errors
const canonicaljson = require('@stratumn/canonicaljson');
const fluence = require('fluence');
const debug = require('debug')('dweb-transports:fluence');
const defaultOptions = {
nodeUrl: 'https://ia-redis.fluence.one',
nodePort: 443,
appId: '4' // Redis
};
class TransportFluence extends Transport {
constructor(options) {
super(options);
this.options = options; // Dictionary of options
this.session = undefined;
this.name = 'FLUENCE'; // For console log etc
this.supportURLs = ['fluence'];
this.supportFunctions = [
// General data functions
'fetch', // p_rawfetch(url, {timeoutMS, start, end, relay}) Fetch some bytes based on a url
// Lists functions
'list', // p_rawlist(url) Fetch all the objects in a list .. identified by the url of the .. 'signedby' parameter of the p_rawadd call
'add', // p_rawadd(url, sig) Store a new list item, it should be stored so that it can be retrieved either by "signedby" (using p_rawlist) or by "url" (with p_rawreverse).
'newlisturls', // p_newlisturls(cl) Obtain a pair of URLs for a new list
// KeyValueTable functions
'newdatabase', // p_newdatabase(pubkey) Create a new database based on some existing object
'newtable', // p_newtable(pubkey, table) Create a new table
'get', // p_get(url, keys) Get one or more keys from a table
'set', // p_set(url, keyvalues, value) Set one or more keys in a table.
'getall',// p_getall(url) Return a dictionary representing the table
'keys', // p_keys(url) Return a list of keys in a table (suitable for iterating through)
];
this.supportFeatures = [];
this.status = Transport.STATUS_LOADED;
}
static setup0(options) {
const combinedOptions = Transport.mergeoptions(defaultOptions, options.fluence);
console.assert(combinedOptions.nodeUrl, 'Fluence Node url should be specified');
console.assert(combinedOptions.nodePort, 'Fluence Node port should be specified');
console.assert(combinedOptions.appId, 'Fluence AppId should be specified');
let t = new TransportFluence(combinedOptions);
Transports.addtransport(t);
return t;
}
async p_setup1(cb) {
try {
this.status = Transport.STATUS_STARTING;
debug('connecting...');
if (cb) cb(this);
const rndString = Math.random().toString(36).substring(2, 15) + Math.random().toString(36).substring(2, 15);
this.session = fluence.directConnect(this.options.nodeUrl, this.options.nodePort, this.options.appId, rndString);
debug('connected.');
this.status = Transport.STATUS_CONNECTED;
} catch (err) {
console.error(this.name, 'failed to start', err);
this.status = Transport.STATUS_FAILED;
}
if (cb) cb(this);
return this;
}
async p_status() {
if (this.session !== null && this.session !== undefined) {
this.status = Transport.STATUS_CONNECTED;
}
return super.p_status();
}
parseUrl(url) {
const parsedUrl = Url.parse(url);
if (parsedUrl.protocol !== 'fluence:') {
throw new errors.TransportError(`TransportFLUENCE Error encountered retrieving val: url (${parsedUrl.href}) is not a valid FLUENCE url | protocol = ${parsedUrl.protocol}`);
}
debug('retrieve url', parsedUrl.href);
return parsedUrl;
}
newKey(pubkey) {
if (pubkey.hasOwnProperty("keypair")) {
pubkey = pubkey.keypair.signingexport();
}
// By this point pubkey should be an export of a public key of form xyz:abc where xyz
// specifies the type of public key (NACL VERIFY being the only kind we expect currently)
return `fluence:/fluence/${encodeURIComponent(pubkey)}`;
}
parseRedisResponse(result) {
if (result.startsWith('$-1')) {
return undefined;
}
function parseResponsePart(result) {
const [, type, countStr] = /^([\+\-\:\$\*]{1})([0-9]+)/.exec(result);
const count = Number(countStr);
switch (type) {
case '+': { // Simple string
const offset = 1;
const [, data] = /([^\n\r]+)/.exec(result.substring(offset));
return {
data: data,
offset: offset + data.length + 2
};
}
case ':': { // Integer
const offset = 1;
const [, data] = /([0-9]+)/.exec(result.substring(offset));
return {
data: Number(data),
offset: offset + data.length + 2
};
}
case '$': { // Bulk string
const offset = 1 + String(count).length + 2;
return {
data: result.substring(offset, offset + count),
offset: offset + count + 2
};
}
case '*': { // Array
let offset = 1 + String(count).length + 2;
const list = [];
for(let i = 0;i < count;i++) {
const parsedListItem = parseResponsePart(result.substring(offset));
list.push(parsedListItem.data);
offset += parsedListItem.offset;
}
return {
data: list,
offset: null
};
}
default: {
throw new errors.TransportError(`TransportFLUENCE Error unsupprted Redis response type: ${type}, response: ${result}`);
}
}
}
return parseResponsePart(result).data;
}
// General data functions (uses Redis basic GET\SET)
async p_rawfetch(url) {
const parsedUrl = this.parseUrl(url);
const key = parsedUrl.path;
const result = await this.session.request(`GET ${key}`);
const data = this.parseRedisResponse(result.asString());
if (!data) {
throw new errors.TransportError(`TransportFLUENCE unable to retrieve: ${url.href}`);
}
return typeof data === 'string' ? JSON.parse(data) : data;
}
// List functions (uses Redis list)
async p_rawlist(url) {
const parsedUrl = this.parseUrl(url);
const key = parsedUrl.path;
const result = await this.session.request(`LRANGE ${key} 0 -1`);
const data = this.parseRedisResponse(result.asString());
if (!data) {
throw new errors.TransportError(`TransportFLUENCE unable to retrieve list: ${url.href}`);
}
return data.map(listItem => typeof listItem === 'string' ? JSON.parse(listItem) : listItem);
}
async p_rawadd(url, sig) {
const parsedUrl = this.parseUrl(url);
const key = parsedUrl.path;
const data = canonicaljson.stringify( sig.preflight( Object.assign({}, sig)));
await this.session.request(`RPUSH ${key} ${data}`);
}
async p_newlisturls(cl) {
const key = this.newKey(cl);
return [key, key];
}
// KeyValueTable functions (uses Redis hashes)
async p_newdatabase(pubkey) {
/*
Request a new database
returns: { publicurl: "fluence:/fluence/<publickey>", privateurl: "fluence:/fluence/<publickey>"> }
*/
let key = await this.newKey(pubkey);
return {
publicurl: key,
privateurl: key
};
}
async p_newtable(pubkey, table) {
/*
Request a new table
returns: {publicurl: "fluence:/fluence/<publickey>/<table>", privateurl: "fluence:/fluence/<publickey>/<table>">
*/
if (!pubkey) {
throw new errors.CodingError("p_newtable currently requires a pubkey");
}
const { publicurl, privateurl } = await this.p_newdatabase(pubkey);
return {
privateurl: `${privateurl}/${table}`,
publicurl: `${publicurl}/${table}`
};
}
async p_set(url, keyvalues, value) { // url = fluence:/fluence/<publickey>/<table>
/*
Set key values
keyvalues: string (key) in which case value should be set there OR
object in which case value is ignored
*/
if (typeof keyvalues === 'string') {
await this.session.request(`HSET ${url} ${keyvalues} ${canonicaljson.stringify(value)}`);
} else {
// Store all key-value pairs without destroying any other key/value pairs previously set
console.assert(!Array.isArray(keyvalues), 'TransportFLUENCE - shouldnt pass an array as the keyvalues');
await Promise.all(
Object.keys(keyvalues).map(hKey => this.session.request(`HSET ${url} ${hKey} ${canonicaljson.stringify(keyvalues[hKey])}`))
);
}
}
async p_get(url, keys) {
if (Array.isArray(keys)) {
const result = await this.session.request(`HMGET ${url} ${keys.join(' ')}`);
const data = this.parseRedisResponse(result.asString());
return keys.reduce((store, key, index) => {
const keyValue = data[index];
store[key] = typeof keyValue === "string" ? JSON.parse(keyValue) : keyValue;
return store;
}, {});
} else {
const result = await this.session.request(`HGET ${url} ${keys}`);
let data = this.parseRedisResponse(result.asString());
return typeof data === 'string' ? JSON.parse(data) : data;
}
}
async p_keys(url) {
const result = await this.session.request(`HKEYS ${url}`);
return this.parseRedisResponse(result.asString());
}
async p_getall(url) {
const result = await this.session.request(`HGETALL ${url}`);
const dataArray = this.parseRedisResponse(result.asString());
return dataArray.reduce((store, key, index, dataArray) => {
if (index % 2 !== 0) {
return store;
}
const keyValue = dataArray[index + 1];
store[key] = typeof keyValue === "string" ? JSON.parse(keyValue) : keyValue;
return store;
}, {});
}
async p_delete(url, keys) {
if (typeof keys === "string") {
await this.session.request(`HDEL ${url} ${keys}`);
} else {
await this.session.request(`HDEL ${url} ${keys.join(' ')}`);
}
}
}
Transports._transportclasses['FLUENCE'] = TransportFluence;
exports = module.exports = TransportFluence;

View File

@ -116,6 +116,11 @@ class Transports {
return Transports._connected().find((t) => t.name === "GUN")
}
static fluence() {
// Find a FLUENCE transport if it exists
return Transports._connected().find((t) => t.name === "FLUENCE")
}
static async p_resolveNames(urls) {
/* Resolve urls that might be names, returning a modified array.
*/
@ -720,7 +725,7 @@ class Transports {
// WOLK is currently working (I think)
// GUN is turned off by default because it fills up localstorage on browser and stops working, https://github.com/internetarchive/dweb-archive/issues/106
//if (! tabbrevs.length) { tabbrevs = ["HTTP", "YJS", "IPFS", "WEBTORRENT", "GUN", "WOLK"]; } // SEE-OTHER-ADDTRANSPORT
if (! tabbrevs.length) { tabbrevs = ["HTTP", "IPFS", "WEBTORRENT", "WOLK"]; } // SEE-OTHER-ADDTRANSPORT
if (! tabbrevs.length) { tabbrevs = ["HTTP", "IPFS", "WEBTORRENT", "WOLK", "FLUENCE"]; } // SEE-OTHER-ADDTRANSPORT
tabbrevs = tabbrevs.map(n => n.toUpperCase());
return tabbrevs;
}
@ -822,7 +827,7 @@ class Transports {
// In patterns below http or https; and :/ or :// are treated the same
const gateways = ["dweb.me", "ipfs.io"]; // Known gateways, may dynamically load this at some point
// SEE-OTHER-ADDTRANSPORT
const protocols = ["ipfs","gun","magnet","yjs","wolk","arc", "contenthash", "http", "https"];
const protocols = ["ipfs","gun","magnet","yjs","wolk","arc", "contenthash", "http", "https", "fluence"];
const protocolsWantingDomains = ["arc", "http", "https"];
const gatewaypatts = [ // Must be before patts because gateway names often start with a valid proto
/^http[s]?:[/]+([^/]+)[/](\w+)[/](.*)/i, // https://(gateway)/proto/(internal) + gateway in list (IPFS gateways. dweb.me)

View File

@ -113,7 +113,7 @@ Please follow conventions i.e.
### Options for integration: Hijack, Push, Hybrid
The actual choices to be made will depend on some of the differences between transports, specifically.
* Is data immutable, and refered to by a content address or hash (IPFS, WEBTORRENT), or is it mutable and refered to by a name. (GUN, YJS)
* Is data immutable, and refered to by a content address or hash (IPFS, WEBTORRENT), or is it mutable and refered to by a name. (GUN, YJS, FLUENCE)
* Will it be easier to
1. 'hijack' specific addresses and use the peer to initiate retrieval from our servers (GUN)
2. Have the server Push data into the platform and share the hash generated by the platform in the metadata (IPFS) and/or pass a URL to the platform which it can pull and return its hash.
@ -266,4 +266,4 @@ Typically your code for integrating into Docker would then look something like t
ENV XYZABC="some environment info you need"
# Expose any defaults you need
EXPOSE 1234
```
```

View File

@ -8,5 +8,6 @@ require("./TransportYJS.js");
require("./TransportWEBTORRENT.js");
require("./TransportWOLK.js");
require("./TransportGUN.js");
require("./TransportFluence.js");
if (typeof window !== "undefined") { window.DwebTransports = DwebTransports; }
exports = module.exports = DwebTransports;

View File

@ -13,6 +13,7 @@
"dependencies": {
"@stratumn/canonicaljson": "^1.0.3",
"debug": "^4.1.1",
"fluence": "^0.3.13",
"ipfs-http-client": "^33.1.1",
"node-fetch": "^2.3.0",
"readable-stream": "^3.3.0",