fix(basic.gblib): Allow MERGE keyword in storage #380. @othonlima
This commit is contained in:
parent
a7308ef1e8
commit
f5f64fd09b
3 changed files with 115 additions and 94 deletions
|
@ -41,7 +41,6 @@ import Path from 'path';
|
|||
*/
|
||||
export class KeywordsExpressions {
|
||||
private static getParams = (text: string, names) => {
|
||||
let ret = {};
|
||||
const splitParamsButIgnoreCommasInDoublequotes = (str: string) => {
|
||||
return str.split(',').reduce(
|
||||
(accum, curr) => {
|
||||
|
@ -428,16 +427,16 @@ export class KeywordsExpressions {
|
|||
|
||||
keywords[i++] = [
|
||||
/^\s*(DELETE)(\s*)(.*)/gim,
|
||||
($0, $1, $2, $3, $4) => {
|
||||
const params = this.getParams($4, ['file']);
|
||||
($0, $1, $2, $3) => {
|
||||
const params = this.getParams($3, ['file']);
|
||||
return `await sys.deleteFile ({pid: pid, ${params}})`;
|
||||
}
|
||||
];
|
||||
|
||||
keywords[i++] = [
|
||||
/^\s*(.*)\=\s*(UPLOAD)(\s*)(.*)/gim,
|
||||
($0, $1, $2, $3, $4) => {
|
||||
const params = this.getParams($4, ['file']);
|
||||
($0, $1, $2, $3) => {
|
||||
const params = this.getParams($3, ['file']);
|
||||
return `await sys.uploadFile ({pid: pid, ${params}})`;
|
||||
}
|
||||
];
|
||||
|
|
|
@ -2139,7 +2139,37 @@ export class SystemKeywords {
|
|||
if (!this.cachedMerge[pid][file]) {
|
||||
await retry(
|
||||
async (bail) => {
|
||||
rows = await t.findAll({});
|
||||
let rows = [];
|
||||
|
||||
const paginate = (query, { page, pageSize }) => {
|
||||
const offset = page * pageSize;
|
||||
const limit = pageSize;
|
||||
|
||||
return {
|
||||
...query,
|
||||
offset,
|
||||
limit,
|
||||
};
|
||||
};
|
||||
|
||||
let page = 0, pageSize = 1000;
|
||||
let count = 0;
|
||||
|
||||
do {
|
||||
|
||||
rows = [
|
||||
t.findAll(
|
||||
paginate(
|
||||
{
|
||||
where: {},
|
||||
},
|
||||
{ page, pageSize },
|
||||
),
|
||||
), ...rows];
|
||||
|
||||
count = rows.length;
|
||||
|
||||
} while (count !== 1000)
|
||||
},
|
||||
{
|
||||
retries: 5,
|
||||
|
@ -2543,16 +2573,14 @@ export class SystemKeywords {
|
|||
/**
|
||||
* Loads all para from tabular file Config.xlsx.
|
||||
*/
|
||||
public async dirFolder(
|
||||
min: GBMinInstance,
|
||||
remotePath: string,
|
||||
baseUrl: string = null,
|
||||
array = null
|
||||
): Promise<any> {
|
||||
public async dirFolder({ pid, remotePath, baseUrl = null, array = null }) {
|
||||
|
||||
const { min } = await DialogKeywords.getProcessInfo(pid);
|
||||
GBLogEx.info(min, `dirFolder: remotePath=${remotePath}, baseUrl=${baseUrl}`);
|
||||
|
||||
if (!baseUrl) {
|
||||
let { baseUrl, client } = await GBDeployer.internalGetDriveClient(min);
|
||||
let client;
|
||||
[baseUrl, client] = await GBDeployer.internalGetDriveClient(min);
|
||||
|
||||
remotePath = remotePath.replace(/\\/gi, '/');
|
||||
|
||||
|
@ -2574,8 +2602,8 @@ export class SystemKeywords {
|
|||
await CollectionUtil.asyncForEach(documents, async item => {
|
||||
|
||||
if (item.folder) {
|
||||
const nextFolder = urlJoin(remotePath, item.name);
|
||||
array = [array, ... await this.dirFolder(min, null, nextFolder, array)];
|
||||
remotePath = urlJoin(remotePath, item.name);
|
||||
array = [array, ... await this.dirFolder({ pid, remotePath, baseUrl, array })];
|
||||
} else {
|
||||
|
||||
// TODO: https://raw.githubusercontent.com/ishanarora04/quickxorhash/master/quickxorhash.js
|
||||
|
|
|
@ -106,17 +106,11 @@ export class GBDeployer implements IGBDeployer {
|
|||
public static async internalGetDriveClient(min: GBMinInstance) {
|
||||
let token;
|
||||
|
||||
// TODO: Add expiration logic.
|
||||
|
||||
if (min['cacheToken'] && null) {
|
||||
return min['cacheToken'];
|
||||
} else {
|
||||
|
||||
// Get token as root only if the bot does not have
|
||||
// an custom tenant for retrieving packages.
|
||||
|
||||
token = await (min.adminService as any)['acquireElevatedToken']
|
||||
(min.instance.instanceId, min.instance.authenticatorTenant?false:true);
|
||||
(min.instance.instanceId, min.instance.authenticatorTenant ? false : true);
|
||||
|
||||
const siteId = process.env.STORAGE_SITE_ID;
|
||||
const libraryId = process.env.STORAGE_LIBRARY;
|
||||
|
@ -130,7 +124,7 @@ export class GBDeployer implements IGBDeployer {
|
|||
min['cacheToken'] = { baseUrl, client };
|
||||
|
||||
return min['cacheToken'];
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -624,12 +618,12 @@ export class GBDeployer implements IGBDeployer {
|
|||
const connectionName = t.replace(strFind, '');
|
||||
let con = {};
|
||||
con['name'] = connectionName;
|
||||
con['storageServer']= min.core.getParam<string>(min.instance, `${connectionName} Server`, null),
|
||||
con['storageName']= min.core.getParam<string>(min.instance, `${connectionName} Name`, null),
|
||||
con['storageUsername']= min.core.getParam<string>(min.instance, `${connectionName} Username`, null),
|
||||
con['storagePort']= min.core.getParam<string>(min.instance, `${connectionName} Port`, null),
|
||||
con['storagePassword']= min.core.getParam<string>(min.instance, `${connectionName} Password`, null),
|
||||
con['storageDriver']= min.core.getParam<string>(min.instance, `${connectionName} Driver`, null)
|
||||
con['storageServer'] = min.core.getParam<string>(min.instance, `${connectionName} Server`, null),
|
||||
con['storageName'] = min.core.getParam<string>(min.instance, `${connectionName} Name`, null),
|
||||
con['storageUsername'] = min.core.getParam<string>(min.instance, `${connectionName} Username`, null),
|
||||
con['storagePort'] = min.core.getParam<string>(min.instance, `${connectionName} Port`, null),
|
||||
con['storagePassword'] = min.core.getParam<string>(min.instance, `${connectionName} Password`, null),
|
||||
con['storageDriver'] = min.core.getParam<string>(min.instance, `${connectionName} Driver`, null)
|
||||
connections.push(con);
|
||||
});
|
||||
|
||||
|
@ -685,14 +679,14 @@ export class GBDeployer implements IGBDeployer {
|
|||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Removes the package local files from cache.
|
||||
*/
|
||||
public async cleanupPackage(instance: IGBInstance, packageName: string) {
|
||||
const path = DialogKeywords.getGBAIPath(instance.botId, null, packageName);
|
||||
const localFolder = Path.join('work', path);
|
||||
rimraf.sync(localFolder);
|
||||
}
|
||||
/**
|
||||
* Removes the package local files from cache.
|
||||
*/
|
||||
public async cleanupPackage(instance: IGBInstance, packageName: string) {
|
||||
const path = DialogKeywords.getGBAIPath(instance.botId, null, packageName);
|
||||
const localFolder = Path.join('work', path);
|
||||
rimraf.sync(localFolder);
|
||||
}
|
||||
|
||||
/**
|
||||
* Removes the package from the storage and local work folders.
|
||||
|
@ -731,7 +725,7 @@ export class GBDeployer implements IGBDeployer {
|
|||
const service = new KBService(this.core.sequelize);
|
||||
rimraf.sync(localPath);
|
||||
|
||||
if (p){
|
||||
if (p) {
|
||||
await service.undeployKbFromStorage(instance, this, p.packageId);
|
||||
}
|
||||
|
||||
|
@ -769,68 +763,68 @@ export class GBDeployer implements IGBDeployer {
|
|||
|
||||
// TODO: Semaphore logic.
|
||||
//try {
|
||||
GBLogEx.info(instance.instanceId, `rebuildIndex running...`);
|
||||
// release = await GBServer.globals.indexSemaphore.acquire();
|
||||
// GBLogEx.info(instance.instanceId, `Acquire rebuildIndex done.`);
|
||||
GBLogEx.info(instance.instanceId, `rebuildIndex running...`);
|
||||
// release = await GBServer.globals.indexSemaphore.acquire();
|
||||
// GBLogEx.info(instance.instanceId, `Acquire rebuildIndex done.`);
|
||||
|
||||
const key = instance.searchKey ? instance.searchKey : GBServer.globals.minBoot.instance.searchKey;
|
||||
const searchIndex = instance.searchIndex ? instance.searchIndex : GBServer.globals.minBoot.instance.searchIndex;
|
||||
const searchIndexer = instance.searchIndexer
|
||||
? instance.searchIndexer
|
||||
: GBServer.globals.minBoot.instance.searchIndexer;
|
||||
const host = instance.searchHost ? instance.searchHost : GBServer.globals.minBoot.instance.searchHost;
|
||||
const key = instance.searchKey ? instance.searchKey : GBServer.globals.minBoot.instance.searchKey;
|
||||
const searchIndex = instance.searchIndex ? instance.searchIndex : GBServer.globals.minBoot.instance.searchIndex;
|
||||
const searchIndexer = instance.searchIndexer
|
||||
? instance.searchIndexer
|
||||
: GBServer.globals.minBoot.instance.searchIndexer;
|
||||
const host = instance.searchHost ? instance.searchHost : GBServer.globals.minBoot.instance.searchHost;
|
||||
|
||||
// Prepares search.
|
||||
// Prepares search.
|
||||
|
||||
const search = new AzureSearch(
|
||||
key,
|
||||
host,
|
||||
searchIndex,
|
||||
searchIndexer
|
||||
);
|
||||
const connectionString = GBDeployer.getConnectionStringFromInstance(GBServer.globals.minBoot.instance);
|
||||
const dsName = 'gb';
|
||||
const search = new AzureSearch(
|
||||
key,
|
||||
host,
|
||||
searchIndex,
|
||||
searchIndexer
|
||||
);
|
||||
const connectionString = GBDeployer.getConnectionStringFromInstance(GBServer.globals.minBoot.instance);
|
||||
const dsName = 'gb';
|
||||
|
||||
// Removes any previous index.
|
||||
// Removes any previous index.
|
||||
|
||||
try {
|
||||
await search.deleteDataSource(dsName);
|
||||
} catch (err) {
|
||||
// If it is a 404 there is nothing to delete as it is the first creation.
|
||||
try {
|
||||
await search.deleteDataSource(dsName);
|
||||
} catch (err) {
|
||||
// If it is a 404 there is nothing to delete as it is the first creation.
|
||||
|
||||
if (err.code !== 404) {
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
// Removes the index.
|
||||
|
||||
try {
|
||||
await search.deleteIndex();
|
||||
} catch (err) {
|
||||
// If it is a 404 there is nothing to delete as it is the first creation.
|
||||
|
||||
if (err.code !== 404 && err.code !== 'OperationNotAllowed') {
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
// Creates the data source and index on the cloud.
|
||||
|
||||
try {
|
||||
await search.createDataSource(dsName, dsName, 'GuaribasQuestion', 'azuresql', connectionString);
|
||||
} catch (err) {
|
||||
GBLog.error(err);
|
||||
if (err.code !== 404) {
|
||||
throw err;
|
||||
}
|
||||
await search.createIndex(searchSchema, dsName);
|
||||
}
|
||||
|
||||
// release();
|
||||
GBLogEx.info(instance.instanceId, `Released rebuildIndex mutex.`);
|
||||
// Removes the index.
|
||||
|
||||
try {
|
||||
await search.deleteIndex();
|
||||
} catch (err) {
|
||||
// If it is a 404 there is nothing to delete as it is the first creation.
|
||||
|
||||
if (err.code !== 404 && err.code !== 'OperationNotAllowed') {
|
||||
throw err;
|
||||
}
|
||||
}
|
||||
|
||||
// Creates the data source and index on the cloud.
|
||||
|
||||
try {
|
||||
await search.createDataSource(dsName, dsName, 'GuaribasQuestion', 'azuresql', connectionString);
|
||||
} catch (err) {
|
||||
GBLog.error(err);
|
||||
throw err;
|
||||
}
|
||||
await search.createIndex(searchSchema, dsName);
|
||||
|
||||
// release();
|
||||
GBLogEx.info(instance.instanceId, `Released rebuildIndex mutex.`);
|
||||
//} catch {
|
||||
// if (release) {
|
||||
// release();
|
||||
// }
|
||||
// if (release) {
|
||||
// release();
|
||||
// }
|
||||
//}
|
||||
}
|
||||
|
||||
|
|
Loading…
Add table
Reference in a new issue