mirror of
https://github.com/wavetermdev/waveterm.git
synced 2024-12-21 16:38:23 +01:00
Merge branch 'main' into cole/file-backend-changes
This commit is contained in:
commit
25259428ab
2
.github/workflows/regression.yml
vendored
2
.github/workflows/regression.yml
vendored
@ -22,7 +22,7 @@ jobs:
|
||||
- uses: dashcamio/testdriver@main
|
||||
id: testdriver
|
||||
with:
|
||||
version: v2.10.2
|
||||
version: v2.12.5
|
||||
prerun: |
|
||||
rm ~/Desktop/WITH-LOVE-FROM-AMERICA.txt
|
||||
cd ~/actions-runner/_work/testdriver/testdriver/
|
||||
|
@ -34,9 +34,6 @@ let wasActive = true;
|
||||
let wasInFg = true;
|
||||
let currentGlobalShortcut: string | null = null;
|
||||
let initialClientData: ClientDataType = null;
|
||||
let windows: Windows = {};
|
||||
|
||||
interface Windows extends Record<string, Electron.BrowserWindow> {}
|
||||
|
||||
checkPromptMigrate();
|
||||
ensureDir(waveHome);
|
||||
@ -325,10 +322,7 @@ function shFrameNavHandler(event: Electron.Event<Electron.WebContentsWillFrameNa
|
||||
console.log("frame navigation canceled");
|
||||
}
|
||||
|
||||
function createWindow(id: string, clientData: ClientDataType | null): Electron.BrowserWindow {
|
||||
if (windows[id]) {
|
||||
console.error(`createWindow called for existing window ${id}`);
|
||||
}
|
||||
function createWindow(clientData: ClientDataType | null): Electron.BrowserWindow {
|
||||
const bounds = calcBounds(clientData);
|
||||
setKeyUtilPlatform(platform());
|
||||
const win = new electron.BrowserWindow({
|
||||
@ -376,18 +370,9 @@ function createWindow(id: string, clientData: ClientDataType | null): Electron.B
|
||||
wasInFg = true;
|
||||
wasActive = true;
|
||||
});
|
||||
win.on("close", () => {
|
||||
delete windows[id];
|
||||
});
|
||||
win.webContents.on("zoom-changed", (e) => {
|
||||
win.webContents.send("zoom-changed");
|
||||
});
|
||||
windows[id] = win;
|
||||
return win;
|
||||
}
|
||||
|
||||
function createMainWindow(clientData: ClientDataType | null) {
|
||||
const win = createWindow("main", clientData);
|
||||
win.webContents.setWindowOpenHandler(({ url, frameName }) => {
|
||||
if (url.startsWith("https://docs.waveterm.dev/")) {
|
||||
console.log("openExternal docs", url);
|
||||
@ -408,6 +393,7 @@ function createMainWindow(clientData: ClientDataType | null) {
|
||||
console.log("window-open denied", url);
|
||||
return { action: "deny" };
|
||||
});
|
||||
return win;
|
||||
}
|
||||
|
||||
function mainResizeHandler(_: any, win: Electron.BrowserWindow) {
|
||||
@ -673,13 +659,13 @@ async function getClientData(willRetry: boolean, retryNum: number): Promise<Clie
|
||||
}
|
||||
|
||||
function sendWSSC() {
|
||||
if (windows["main"] != null) {
|
||||
electron.BrowserWindow.getAllWindows().forEach((win) => {
|
||||
if (waveSrvProc == null) {
|
||||
windows["main"].webContents.send("wavesrv-status-change", false);
|
||||
return;
|
||||
win.webContents.send("wavesrv-status-change", false);
|
||||
} else {
|
||||
win.webContents.send("wavesrv-status-change", true, waveSrvProc.pid);
|
||||
}
|
||||
windows["main"].webContents.send("wavesrv-status-change", true, waveSrvProc.pid);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
function runWaveSrv() {
|
||||
@ -747,7 +733,7 @@ electron.ipcMain.on("context-editmenu", (_, { x, y }, opts) => {
|
||||
menu.popup({ x, y });
|
||||
});
|
||||
|
||||
async function createMainWindowWrap() {
|
||||
async function createWindowWrap() {
|
||||
let clientData: ClientDataType | null = null;
|
||||
try {
|
||||
clientData = await getClientDataPoll(1);
|
||||
@ -755,9 +741,9 @@ async function createMainWindowWrap() {
|
||||
} catch (e) {
|
||||
console.log("error getting wavesrv clientdata", e.toString());
|
||||
}
|
||||
createMainWindow(clientData);
|
||||
const win = createWindow(clientData);
|
||||
if (clientData?.winsize.fullscreen) {
|
||||
windows["main"].setFullScreen(true);
|
||||
win.setFullScreen(true);
|
||||
}
|
||||
configureAutoUpdaterStartup(clientData);
|
||||
}
|
||||
@ -776,7 +762,7 @@ function logActiveState() {
|
||||
console.log("error logging active state", err);
|
||||
});
|
||||
// for next iteration
|
||||
wasInFg = windows["main"]?.isFocused();
|
||||
wasInFg = electron.BrowserWindow.getFocusedWindow()?.isFocused() ?? false;
|
||||
wasActive = false;
|
||||
}
|
||||
|
||||
@ -802,9 +788,13 @@ function reregisterGlobalShortcut(shortcut: string) {
|
||||
currentGlobalShortcut = null;
|
||||
return;
|
||||
}
|
||||
const ok = electron.globalShortcut.register(shortcut, () => {
|
||||
const ok = electron.globalShortcut.register(shortcut, async () => {
|
||||
console.log("global shortcut triggered, showing window");
|
||||
windows["main"]?.show();
|
||||
if (electron.BrowserWindow.getAllWindows().length == 0) {
|
||||
await createWindowWrap();
|
||||
}
|
||||
const winToShow = electron.BrowserWindow.getFocusedWindow() ?? electron.BrowserWindow.getAllWindows()[0];
|
||||
winToShow?.show();
|
||||
});
|
||||
console.log("registered global shortcut", shortcut, ok ? "ok" : "failed");
|
||||
if (!ok) {
|
||||
@ -829,9 +819,9 @@ let lastUpdateCheck: Date = null;
|
||||
*/
|
||||
function setAppUpdateStatus(status: string) {
|
||||
appUpdateStatus = status;
|
||||
if (windows["main"] != null) {
|
||||
windows["main"].webContents.send("app-update-status", appUpdateStatus);
|
||||
}
|
||||
electron.BrowserWindow.getAllWindows().forEach((window) => {
|
||||
window.webContents.send("app-update-status", appUpdateStatus);
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
@ -915,9 +905,14 @@ async function installAppUpdate() {
|
||||
detail: "A new version has been downloaded. Restart the application to apply the updates.",
|
||||
};
|
||||
|
||||
await electron.dialog.showMessageBox(windows["main"], dialogOpts).then(({ response }) => {
|
||||
if (response === 0) autoUpdater.quitAndInstall();
|
||||
});
|
||||
const allWindows = electron.BrowserWindow.getAllWindows();
|
||||
if (allWindows.length > 0) {
|
||||
await electron.dialog
|
||||
.showMessageBox(electron.BrowserWindow.getFocusedWindow() ?? allWindows[0], dialogOpts)
|
||||
.then(({ response }) => {
|
||||
if (response === 0) autoUpdater.quitAndInstall();
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
electron.ipcMain.on("install-app-update", () => fireAndForget(() => installAppUpdate()));
|
||||
@ -989,10 +984,11 @@ function configureAutoUpdater(enabled: boolean) {
|
||||
}
|
||||
setTimeout(runActiveTimer, 5000); // start active timer, wait 5s just to be safe
|
||||
await app.whenReady();
|
||||
await createMainWindowWrap();
|
||||
await createWindowWrap();
|
||||
|
||||
app.on("activate", () => {
|
||||
if (electron.BrowserWindow.getAllWindows().length === 0) {
|
||||
createMainWindowWrap().then();
|
||||
createWindowWrap().then();
|
||||
}
|
||||
checkForUpdates();
|
||||
});
|
||||
|
@ -1128,6 +1128,11 @@ func main() {
|
||||
log.Printf("[error] migrate up: %v\n", err)
|
||||
return
|
||||
}
|
||||
// err = blockstore.MigrateBlockstore()
|
||||
// if err != nil {
|
||||
// log.Printf("[error] migrate blockstore: %v\n", err)
|
||||
// return
|
||||
// }
|
||||
clientData, err := sstore.EnsureClientData(context.Background())
|
||||
if err != nil {
|
||||
log.Printf("[error] ensuring client data: %v\n", err)
|
||||
|
1
wavesrv/db/blockstore-migrations/000001_init.down.sql
Normal file
1
wavesrv/db/blockstore-migrations/000001_init.down.sql
Normal file
@ -0,0 +1 @@
|
||||
-- nothing
|
19
wavesrv/db/blockstore-migrations/000001_init.up.sql
Normal file
19
wavesrv/db/blockstore-migrations/000001_init.up.sql
Normal file
@ -0,0 +1,19 @@
|
||||
CREATE TABLE block_file (
|
||||
blockid varchar(36) NOT NULL,
|
||||
name varchar(200) NOT NULL,
|
||||
maxsize bigint NOT NULL,
|
||||
circular boolean NOT NULL,
|
||||
size bigint NOT NULL,
|
||||
createdts bigint NOT NULL,
|
||||
modts bigint NOT NULL,
|
||||
meta json NOT NULL,
|
||||
PRIMARY KEY (blockid, name)
|
||||
);
|
||||
|
||||
CREATE TABLE block_data (
|
||||
blockid varchar(36) NOT NULL,
|
||||
name varchar(200) NOT NULL,
|
||||
partidx int NOT NULL,
|
||||
data blob NOT NULL,
|
||||
PRIMARY KEY(blockid, name, partidx)
|
||||
);
|
@ -10,3 +10,6 @@ import "embed"
|
||||
|
||||
//go:embed migrations/*.sql
|
||||
var MigrationFS embed.FS
|
||||
|
||||
//go:embed blockstore-migrations/*.sql
|
||||
var BlockstoreMigrationFS embed.FS
|
||||
|
@ -5,7 +5,6 @@ go 1.22
|
||||
toolchain go1.22.0
|
||||
|
||||
require (
|
||||
github.com/alecthomas/units v0.0.0-20231202071711-9a357b53e9c9
|
||||
github.com/alessio/shellescape v1.4.1
|
||||
github.com/armon/circbuf v0.0.0-20190214190532-5111143e8da2
|
||||
github.com/creack/pty v1.1.18
|
||||
|
@ -1,5 +1,3 @@
|
||||
github.com/alecthomas/units v0.0.0-20231202071711-9a357b53e9c9 h1:ez/4by2iGztzR4L0zgAOR8lTQK9VlyBVVd7G4omaOQs=
|
||||
github.com/alecthomas/units v0.0.0-20231202071711-9a357b53e9c9/go.mod h1:OMCwj8VM1Kc9e19TLln2VL61YJF0x1XFtfdL4JdbSyE=
|
||||
github.com/alessio/shellescape v1.4.1 h1:V7yhSDDn8LP4lc4jS8pFkt0zCnzVJlG5JXy9BVKJUX0=
|
||||
github.com/alessio/shellescape v1.4.1/go.mod h1:PZAiSCk0LJaZkiCSkPv8qIobYglO3FPpyFjDCtHLS30=
|
||||
github.com/armon/circbuf v0.0.0-20190214190532-5111143e8da2 h1:7Ip0wMmLHLRJdrloDxZfhMm0xrLXZS8+COSu2bXmEQs=
|
||||
@ -57,7 +55,6 @@ github.com/sawka/txwrap v0.1.2 h1:v8xS0Z1LE7/6vMZA81PYihI+0TSR6Zm1MalzzBIuXKc=
|
||||
github.com/sawka/txwrap v0.1.2/go.mod h1:T3nlw2gVpuolo6/XEetvBbk1oMXnY978YmBFy1UyHvw=
|
||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||
github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI=
|
||||
github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4=
|
||||
github.com/stretchr/testify v1.8.1 h1:w7B6lhMri9wdJUVmEZPGGhZzrYTPvgJArz7wNPgYKsk=
|
||||
github.com/stretchr/testify v1.8.1/go.mod h1:w2LPCIKwWwSfY2zedu0+kehJoqGctiVI29o6fzry7u4=
|
||||
github.com/wavetermdev/ssh_config v0.0.0-20240306041034-17e2087ebde2 h1:onqZrJVap1sm15AiIGTfWzdr6cEF0KdtddeuuOVhzyY=
|
||||
@ -74,8 +71,6 @@ golang.org/x/sys v0.15.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA=
|
||||
golang.org/x/term v0.15.0 h1:y/Oo/a/q3IXu26lQgl04j/gjuBDOBlx7X6Om1j2CPW4=
|
||||
golang.org/x/term v0.15.0/go.mod h1:BDl952bC7+uMoWR75FIrCDx79TPU9oHkTZ9yRbYOrX0=
|
||||
golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||
gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI=
|
||||
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
||||
gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||
mvdan.cc/sh/v3 v3.7.0 h1:lSTjdP/1xsddtaKfGg7Myu7DnlHItd3/M2tomOcNNBg=
|
||||
|
@ -10,8 +10,6 @@ import (
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/alecthomas/units"
|
||||
)
|
||||
|
||||
type FileOptsType struct {
|
||||
@ -32,7 +30,11 @@ type FileInfo struct {
|
||||
Meta FileMeta
|
||||
}
|
||||
|
||||
const MaxBlockSize = int64(128 * units.Kilobyte)
|
||||
const UnitsKB = 1024 * 1024
|
||||
const UnitsMB = 1024 * UnitsKB
|
||||
const UnitsGB = 1024 * UnitsMB
|
||||
|
||||
const MaxBlockSize = int64(128 * UnitsKB)
|
||||
const DefaultFlushTimeout = 1 * time.Second
|
||||
|
||||
type CacheEntry struct {
|
||||
@ -79,16 +81,23 @@ type BlockStore interface {
|
||||
GetAllBlockIds(ctx context.Context) []string
|
||||
}
|
||||
|
||||
var cache map[string]*CacheEntry = make(map[string]*CacheEntry)
|
||||
var blockstoreCache map[string]*CacheEntry = make(map[string]*CacheEntry)
|
||||
var globalLock *sync.Mutex = &sync.Mutex{}
|
||||
var appendLock *sync.Mutex = &sync.Mutex{}
|
||||
var flushTimeout = DefaultFlushTimeout
|
||||
var lastWriteTime time.Time
|
||||
|
||||
// for testing
|
||||
func clearCache() {
|
||||
globalLock.Lock()
|
||||
defer globalLock.Unlock()
|
||||
blockstoreCache = make(map[string]*CacheEntry)
|
||||
}
|
||||
|
||||
func InsertFileIntoDB(ctx context.Context, fileInfo FileInfo) error {
|
||||
metaJson, err := json.Marshal(fileInfo.Meta)
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error writing file %s to db: %v", fileInfo.Name, err)
|
||||
return fmt.Errorf("error writing file %s to db: %v", fileInfo.Name, err)
|
||||
}
|
||||
txErr := WithTx(ctx, func(tx *TxWrap) error {
|
||||
query := `INSERT INTO block_file VALUES (?, ?, ?, ?, ?, ?, ?, ?)`
|
||||
@ -96,7 +105,7 @@ func InsertFileIntoDB(ctx context.Context, fileInfo FileInfo) error {
|
||||
return nil
|
||||
})
|
||||
if txErr != nil {
|
||||
return fmt.Errorf("Error writing file %s to db: %v", fileInfo.Name, txErr)
|
||||
return fmt.Errorf("error writing file %s to db: %v", fileInfo.Name, txErr)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@ -104,7 +113,7 @@ func InsertFileIntoDB(ctx context.Context, fileInfo FileInfo) error {
|
||||
func WriteFileToDB(ctx context.Context, fileInfo FileInfo) error {
|
||||
metaJson, err := json.Marshal(fileInfo.Meta)
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error writing file %s to db: %v", fileInfo.Name, err)
|
||||
return fmt.Errorf("error writing file %s to db: %v", fileInfo.Name, err)
|
||||
}
|
||||
txErr := WithTx(ctx, func(tx *TxWrap) error {
|
||||
query := `UPDATE block_file SET blockid = ?, name = ?, maxsize = ?, circular = ?, size = ?, createdts = ?, modts = ?, meta = ? where blockid = ? and name = ?`
|
||||
@ -112,7 +121,7 @@ func WriteFileToDB(ctx context.Context, fileInfo FileInfo) error {
|
||||
return nil
|
||||
})
|
||||
if txErr != nil {
|
||||
return fmt.Errorf("Error writing file %s to db: %v", fileInfo.Name, txErr)
|
||||
return fmt.Errorf("error writing file %s to db: %v", fileInfo.Name, txErr)
|
||||
}
|
||||
return nil
|
||||
|
||||
@ -125,7 +134,7 @@ func WriteDataBlockToDB(ctx context.Context, blockId string, name string, index
|
||||
return nil
|
||||
})
|
||||
if txErr != nil {
|
||||
return fmt.Errorf("Error writing data block to db: %v", txErr)
|
||||
return fmt.Errorf("error writing data block to db: %v", txErr)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@ -152,7 +161,7 @@ func WriteToCacheBlockNum(ctx context.Context, blockId string, name string, p []
|
||||
defer cacheEntry.Lock.Unlock()
|
||||
block, err := GetCacheBlock(ctx, blockId, name, cacheNum, pullFromDB)
|
||||
if err != nil {
|
||||
return 0, 0, fmt.Errorf("Error getting cache block: %v", err)
|
||||
return 0, 0, fmt.Errorf("error getting cache block: %v", err)
|
||||
}
|
||||
var bytesWritten = 0
|
||||
blockLen := len(block.data)
|
||||
@ -192,7 +201,7 @@ func ReadFromCacheBlock(ctx context.Context, blockId string, name string, block
|
||||
}
|
||||
}()
|
||||
if pos > len(block.data) {
|
||||
return 0, fmt.Errorf("Reading past end of cache block, should never happen")
|
||||
return 0, fmt.Errorf("reading past end of cache block, should never happen")
|
||||
}
|
||||
bytesWritten := 0
|
||||
index := pos
|
||||
@ -216,7 +225,7 @@ func ReadFromCacheBlock(ctx context.Context, blockId string, name string, block
|
||||
return bytesWritten, nil
|
||||
}
|
||||
|
||||
const MaxSizeError = "Hit Max Size"
|
||||
const MaxSizeError = "MaxSizeError"
|
||||
|
||||
func WriteToCacheBuf(buf *[]byte, p []byte, pos int, length int, maxWrite int64) (int, error) {
|
||||
bytesToWrite := length
|
||||
@ -260,7 +269,7 @@ func GetValuesFromCacheId(cacheId string) (blockId string, name string) {
|
||||
func GetCacheEntry(ctx context.Context, blockId string, name string) (*CacheEntry, bool) {
|
||||
globalLock.Lock()
|
||||
defer globalLock.Unlock()
|
||||
if curCacheEntry, found := cache[GetCacheId(blockId, name)]; found {
|
||||
if curCacheEntry, found := blockstoreCache[GetCacheId(blockId, name)]; found {
|
||||
return curCacheEntry, true
|
||||
} else {
|
||||
return nil, false
|
||||
@ -279,7 +288,7 @@ func GetCacheEntryOrPopulate(ctx context.Context, blockId string, name string) (
|
||||
if cacheEntry, found := GetCacheEntry(ctx, blockId, name); found {
|
||||
return cacheEntry, nil
|
||||
} else {
|
||||
return nil, fmt.Errorf("Error getting cache entry %v %v", blockId, name)
|
||||
return nil, fmt.Errorf("error getting cache entry %v %v", blockId, name)
|
||||
}
|
||||
}
|
||||
|
||||
@ -288,16 +297,16 @@ func GetCacheEntryOrPopulate(ctx context.Context, blockId string, name string) (
|
||||
func SetCacheEntry(ctx context.Context, cacheId string, cacheEntry *CacheEntry) {
|
||||
globalLock.Lock()
|
||||
defer globalLock.Unlock()
|
||||
if _, found := cache[cacheId]; found {
|
||||
if _, found := blockstoreCache[cacheId]; found {
|
||||
return
|
||||
}
|
||||
cache[cacheId] = cacheEntry
|
||||
blockstoreCache[cacheId] = cacheEntry
|
||||
}
|
||||
|
||||
func DeleteCacheEntry(ctx context.Context, blockId string, name string) {
|
||||
globalLock.Lock()
|
||||
defer globalLock.Unlock()
|
||||
delete(cache, GetCacheId(blockId, name))
|
||||
delete(blockstoreCache, GetCacheId(blockId, name))
|
||||
}
|
||||
|
||||
func GetCacheBlock(ctx context.Context, blockId string, name string, cacheNum int, pullFromDB bool) (*CacheBlock, error) {
|
||||
@ -392,7 +401,7 @@ func WriteAtHelper(ctx context.Context, blockId string, name string, p []byte, o
|
||||
}
|
||||
fInfo, err := Stat(ctx, blockId, name)
|
||||
if err != nil {
|
||||
return 0, fmt.Errorf("Write At err: %v", err)
|
||||
return 0, fmt.Errorf("WriteAt err: %v", err)
|
||||
}
|
||||
if off > fInfo.Opts.MaxSize && fInfo.Opts.Circular {
|
||||
numOver := off / fInfo.Opts.MaxSize
|
||||
@ -416,12 +425,12 @@ func WriteAtHelper(ctx context.Context, blockId string, name string, p []byte, o
|
||||
b, err := WriteAtHelper(ctx, blockId, name, p, 0, false)
|
||||
bytesWritten += b
|
||||
if err != nil {
|
||||
return bytesWritten, fmt.Errorf("Write to cache error: %v", err)
|
||||
return bytesWritten, fmt.Errorf("write to cache error: %v", err)
|
||||
}
|
||||
break
|
||||
}
|
||||
} else {
|
||||
return bytesWritten, fmt.Errorf("Write to cache error: %v", err)
|
||||
return bytesWritten, fmt.Errorf("write to cache error: %v", err)
|
||||
}
|
||||
}
|
||||
if len(p) == b {
|
||||
@ -452,7 +461,7 @@ func GetAllBlockSizes(dataBlocks []*CacheBlock) (int, int) {
|
||||
}
|
||||
|
||||
func FlushCache(ctx context.Context) error {
|
||||
for _, cacheEntry := range cache {
|
||||
for _, cacheEntry := range blockstoreCache {
|
||||
err := WriteFileToDB(ctx, *cacheEntry.Info)
|
||||
if err != nil {
|
||||
return err
|
||||
@ -485,14 +494,14 @@ func ReadAt(ctx context.Context, blockId string, name string, p *[]byte, off int
|
||||
bytesRead := 0
|
||||
fInfo, err := Stat(ctx, blockId, name)
|
||||
if err != nil {
|
||||
return 0, fmt.Errorf("Read At err: %v", err)
|
||||
return 0, fmt.Errorf("ReadAt err: %v", err)
|
||||
}
|
||||
if off > fInfo.Opts.MaxSize && fInfo.Opts.Circular {
|
||||
numOver := off / fInfo.Opts.MaxSize
|
||||
off = off - (numOver * fInfo.Opts.MaxSize)
|
||||
}
|
||||
if off > fInfo.Size {
|
||||
return 0, fmt.Errorf("Read At error: tried to read past the end of the file")
|
||||
return 0, fmt.Errorf("ReadAt error: tried to read past the end of the file")
|
||||
}
|
||||
endReadPos := math.Min(float64(int64(len(*p))+off), float64(fInfo.Size))
|
||||
bytesToRead := int64(endReadPos) - off
|
||||
@ -505,7 +514,7 @@ func ReadAt(ctx context.Context, blockId string, name string, p *[]byte, off int
|
||||
for index := curCacheNum; index < curCacheNum+numCaches; index++ {
|
||||
curCacheBlock, err := GetCacheBlock(ctx, blockId, name, index, true)
|
||||
if err != nil {
|
||||
return bytesRead, fmt.Errorf("Error getting cache block: %v", err)
|
||||
return bytesRead, fmt.Errorf("error getting cache block: %v", err)
|
||||
}
|
||||
cacheOffset := off - (int64(index) * MaxBlockSize)
|
||||
if cacheOffset < 0 {
|
||||
@ -540,7 +549,7 @@ func ReadAt(ctx context.Context, blockId string, name string, p *[]byte, off int
|
||||
break
|
||||
}
|
||||
} else {
|
||||
return bytesRead, fmt.Errorf("Read from cache error: %v", err)
|
||||
return bytesRead, fmt.Errorf("read from cache error: %v", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -552,7 +561,7 @@ func AppendData(ctx context.Context, blockId string, name string, p []byte) (int
|
||||
defer appendLock.Unlock()
|
||||
fInfo, err := Stat(ctx, blockId, name)
|
||||
if err != nil {
|
||||
return 0, fmt.Errorf("Append stat error: %v", err)
|
||||
return 0, fmt.Errorf("append stat error: %v", err)
|
||||
}
|
||||
return WriteAt(ctx, blockId, name, p, fInfo.Size)
|
||||
}
|
||||
@ -564,12 +573,12 @@ func DeleteFile(ctx context.Context, blockId string, name string) error {
|
||||
}
|
||||
|
||||
func DeleteBlock(ctx context.Context, blockId string) error {
|
||||
for cacheId, _ := range cache {
|
||||
for cacheId := range blockstoreCache {
|
||||
curBlockId, name := GetValuesFromCacheId(cacheId)
|
||||
if curBlockId == blockId {
|
||||
err := DeleteFile(ctx, blockId, name)
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error deleting %v %v: %v", blockId, name, err)
|
||||
return fmt.Errorf("error deleting %v %v: %v", blockId, name, err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -8,11 +8,16 @@ import (
|
||||
"path"
|
||||
"sync"
|
||||
|
||||
"github.com/golang-migrate/migrate/v4"
|
||||
_ "github.com/golang-migrate/migrate/v4/database/sqlite3"
|
||||
"github.com/golang-migrate/migrate/v4/source/iofs"
|
||||
"github.com/jmoiron/sqlx"
|
||||
_ "github.com/mattn/go-sqlite3"
|
||||
"github.com/sawka/txwrap"
|
||||
"github.com/wavetermdev/waveterm/wavesrv/pkg/dbutil"
|
||||
"github.com/wavetermdev/waveterm/wavesrv/pkg/scbase"
|
||||
|
||||
dbfs "github.com/wavetermdev/waveterm/wavesrv/db"
|
||||
)
|
||||
|
||||
const DBFileName = "blockstore.db"
|
||||
@ -21,12 +26,64 @@ type SingleConnDBGetter struct {
|
||||
SingleConnLock *sync.Mutex
|
||||
}
|
||||
|
||||
var dbWrap *SingleConnDBGetter
|
||||
var dbWrap *SingleConnDBGetter = &SingleConnDBGetter{SingleConnLock: &sync.Mutex{}}
|
||||
|
||||
type TxWrap = txwrap.TxWrap
|
||||
|
||||
func InitDBState() {
|
||||
dbWrap = &SingleConnDBGetter{SingleConnLock: &sync.Mutex{}}
|
||||
func MakeBlockstoreMigrate() (*migrate.Migrate, error) {
|
||||
fsVar, err := iofs.New(dbfs.BlockstoreMigrationFS, "blockstore-migrations")
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("opening iofs: %w", err)
|
||||
}
|
||||
dbUrl := fmt.Sprintf("sqlite3://%s", GetDBName())
|
||||
m, err := migrate.NewWithSourceInstance("iofs", fsVar, dbUrl)
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("making blockstore migration db[%s]: %w", GetDBName(), err)
|
||||
}
|
||||
return m, nil
|
||||
}
|
||||
|
||||
func MigrateBlockstore() error {
|
||||
log.Printf("migrate blockstore\n")
|
||||
m, err := MakeBlockstoreMigrate()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
curVersion, dirty, err := GetMigrateVersion(m)
|
||||
if dirty {
|
||||
return fmt.Errorf("cannot migrate up, database is dirty")
|
||||
}
|
||||
if err != nil {
|
||||
return fmt.Errorf("cannot get current migration version: %v", err)
|
||||
}
|
||||
defer m.Close()
|
||||
err = m.Up()
|
||||
if err != nil && err != migrate.ErrNoChange {
|
||||
return fmt.Errorf("migrating blockstore: %w", err)
|
||||
}
|
||||
newVersion, _, err := GetMigrateVersion(m)
|
||||
if err != nil {
|
||||
return fmt.Errorf("cannot get new migration version: %v", err)
|
||||
}
|
||||
if newVersion != curVersion {
|
||||
log.Printf("[db] blockstore migration done, version %d -> %d\n", curVersion, newVersion)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func GetMigrateVersion(m *migrate.Migrate) (uint, bool, error) {
|
||||
if m == nil {
|
||||
var err error
|
||||
m, err = MakeBlockstoreMigrate()
|
||||
if err != nil {
|
||||
return 0, false, err
|
||||
}
|
||||
}
|
||||
curVersion, dirty, err := m.Version()
|
||||
if err == migrate.ErrNilVersion {
|
||||
return 0, false, nil
|
||||
}
|
||||
return curVersion, dirty, err
|
||||
}
|
||||
|
||||
func (dbg *SingleConnDBGetter) GetDB(ctx context.Context) (*sqlx.DB, error) {
|
||||
@ -62,8 +119,12 @@ func WithTxRtn[RT any](ctx context.Context, fn func(tx *TxWrap) (RT, error)) (RT
|
||||
var globalDBLock = &sync.Mutex{}
|
||||
var globalDB *sqlx.DB
|
||||
var globalDBErr error
|
||||
var overrideDBName string
|
||||
|
||||
func GetDBName() string {
|
||||
if overrideDBName != "" {
|
||||
return overrideDBName
|
||||
}
|
||||
scHome := scbase.GetWaveHomeDir()
|
||||
return path.Join(scHome, DBFileName)
|
||||
}
|
||||
|
@ -6,15 +6,17 @@ import (
|
||||
"crypto/md5"
|
||||
"crypto/rand"
|
||||
"log"
|
||||
"os"
|
||||
"sync"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/alecthomas/units"
|
||||
|
||||
"github.com/wavetermdev/waveterm/wavesrv/pkg/dbutil"
|
||||
)
|
||||
|
||||
const testOverrideDBName = "test-blockstore.db"
|
||||
const bigFileSize = 10 * UnitsMB
|
||||
|
||||
type TestBlockType struct {
|
||||
BlockId string
|
||||
Name string
|
||||
@ -22,6 +24,22 @@ type TestBlockType struct {
|
||||
Data []byte
|
||||
}
|
||||
|
||||
func initTestDb(t *testing.T) {
|
||||
log.Printf("initTestDb: %v", t.Name())
|
||||
os.Remove(testOverrideDBName)
|
||||
overrideDBName = testOverrideDBName
|
||||
err := MigrateBlockstore()
|
||||
if err != nil {
|
||||
t.Fatalf("MigrateBlockstore error: %v", err)
|
||||
}
|
||||
}
|
||||
|
||||
func cleanupTestDB(t *testing.T) {
|
||||
clearCache()
|
||||
CloseDB()
|
||||
os.Remove(testOverrideDBName)
|
||||
}
|
||||
|
||||
func (b *TestBlockType) ToMap() map[string]interface{} {
|
||||
rtn := make(map[string]interface{})
|
||||
return rtn
|
||||
@ -35,22 +53,17 @@ func (b *TestBlockType) FromMap(m map[string]interface{}) bool {
|
||||
return true
|
||||
}
|
||||
|
||||
func Cleanup(t *testing.T, ctx context.Context) {
|
||||
DeleteBlock(ctx, "test-block-id")
|
||||
}
|
||||
|
||||
func CleanupName(t *testing.T, ctx context.Context, blockId string) {
|
||||
DeleteBlock(ctx, blockId)
|
||||
}
|
||||
|
||||
func TestGetDB(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
GetDBTimeout := 10 * time.Second
|
||||
ctx, _ := context.WithTimeout(context.Background(), GetDBTimeout)
|
||||
ctx, cancelFn := context.WithTimeout(context.Background(), GetDBTimeout)
|
||||
defer cancelFn()
|
||||
_, err := GetDB(ctx)
|
||||
if err != nil {
|
||||
t.Errorf("TestInitDB error: %v", err)
|
||||
}
|
||||
CloseDB()
|
||||
}
|
||||
|
||||
func SimpleAssert(t *testing.T, condition bool, description string) {
|
||||
@ -82,9 +95,11 @@ func InsertIntoBlockData(t *testing.T, ctx context.Context, blockId string, name
|
||||
}
|
||||
|
||||
func TestTx(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
SetFlushTimeout(2 * time.Minute)
|
||||
InitDBState()
|
||||
txErr := WithTx(ctx, func(tx *TxWrap) error {
|
||||
query := `INSERT into block_data values ('test-block-id', 'test-file-name', 0, 256)`
|
||||
tx.Exec(query)
|
||||
@ -127,11 +142,13 @@ func TestTx(t *testing.T) {
|
||||
if txErr != nil {
|
||||
t.Errorf("TestTx error deleting test entries: %v", txErr)
|
||||
}
|
||||
CloseDB()
|
||||
}
|
||||
|
||||
func TestMultipleChunks(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
InitDBState()
|
||||
InsertIntoBlockData(t, ctx, "test-block-id", "file-1", 0, make([]byte, 5))
|
||||
InsertIntoBlockData(t, ctx, "test-block-id", "file-1", 1, make([]byte, 5))
|
||||
InsertIntoBlockData(t, ctx, "test-block-id", "file-1", 2, make([]byte, 5))
|
||||
@ -178,7 +195,9 @@ func TestMultipleChunks(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestMakeFile(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -205,7 +224,7 @@ func TestMakeFile(t *testing.T) {
|
||||
log.Printf("cur file info: %v", curFileInfo)
|
||||
SimpleAssert(t, curFileInfo.Name == "file-1", "correct file name")
|
||||
SimpleAssert(t, curFileInfo.Meta["test-descriptor"] == true, "meta correct")
|
||||
curCacheEntry := cache[GetCacheId("test-block-id", "file-1")]
|
||||
curCacheEntry := blockstoreCache[GetCacheId("test-block-id", "file-1")]
|
||||
curFileInfo = curCacheEntry.Info
|
||||
log.Printf("cache entry: %v", curCacheEntry)
|
||||
SimpleAssert(t, curFileInfo.Name == "file-1", "cache correct file name")
|
||||
@ -218,15 +237,16 @@ func TestMakeFile(t *testing.T) {
|
||||
if txErr != nil {
|
||||
t.Errorf("TestTx error deleting test entries: %v", txErr)
|
||||
}
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteAt(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -244,7 +264,10 @@ func TestWriteAt(t *testing.T) {
|
||||
} else {
|
||||
log.Printf("Write at no errors: %v", bytesWritten)
|
||||
}
|
||||
SimpleAssert(t, bytesWritten == len(testBytesToWrite), "Correct num bytes written")
|
||||
if bytesWritten != len(testBytesToWrite) {
|
||||
t.Errorf("WriteAt error: towrite:%d written:%d err:%v\n", len(testBytesToWrite), bytesWritten, err)
|
||||
return
|
||||
}
|
||||
cacheData, err = GetCacheBlock(ctx, "test-block-id", "file-1", 0, false)
|
||||
if err != nil {
|
||||
t.Errorf("Error getting cache: %v", err)
|
||||
@ -313,15 +336,16 @@ func TestWriteAt(t *testing.T) {
|
||||
}
|
||||
log.Printf("Got stat: %v", fInfo)
|
||||
SimpleAssert(t, int64(len(cacheData.data)) == fInfo.Size, "Correct fInfo size")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteAtLeftPad(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -349,14 +373,16 @@ func TestWriteAtLeftPad(t *testing.T) {
|
||||
}
|
||||
log.Printf("Got stat: %v %v %v", fInfo, fInfo.Size, len(cacheData.data))
|
||||
SimpleAssert(t, int64(len(cacheData.data)) == fInfo.Size, "Correct fInfo size")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestReadAt(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -399,14 +425,16 @@ func TestReadAt(t *testing.T) {
|
||||
}
|
||||
SimpleAssert(t, bytesRead == (11-4), "Correct num bytes read")
|
||||
log.Printf("bytes read: %v string: %s", read, string(read))
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestFlushCache(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -461,17 +489,16 @@ func TestFlushCache(t *testing.T) {
|
||||
t.Errorf("get data from db error: %v", txErr)
|
||||
}
|
||||
log.Printf("DB Data: %v", dbData)
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
var largeDataFlushFullWriteSize int64 = int64(1024 * units.Megabyte)
|
||||
var largeDataFlushFullWriteSize int64 = 64 * UnitsKB
|
||||
|
||||
func WriteLargeDataFlush(t *testing.T, ctx context.Context) {
|
||||
writeSize := int64(64 - 16)
|
||||
fullWriteSize := largeDataFlushFullWriteSize
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -524,6 +551,9 @@ func WriteLargeDataFlush(t *testing.T, ctx context.Context) {
|
||||
SimpleAssert(t, bytes.Equal(readHashBuf, hashBuf), "hashes are equal")
|
||||
}
|
||||
func TestWriteAtMaxSize(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -544,11 +574,12 @@ func TestWriteAtMaxSize(t *testing.T) {
|
||||
log.Printf("readbuf: %v\n", readBuf)
|
||||
SimpleAssert(t, bytesRead == 4, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf[:4], readTest), "Correct bytes read")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteAtMaxSizeMultipleBlocks(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -569,11 +600,12 @@ func TestWriteAtMaxSizeMultipleBlocks(t *testing.T) {
|
||||
log.Printf("readbuf multiple: %v %v %v\n", readBuf, bytesRead, bytesWritten)
|
||||
SimpleAssert(t, bytesRead == 4, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf[:4], readTest), "Correct bytes read")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteAtCircular(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -603,11 +635,12 @@ func TestWriteAtCircular(t *testing.T) {
|
||||
SimpleAssert(t, bytesRead == 7, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf[:7], readTest), "Correct bytes read")
|
||||
log.Printf("readbuf circular %v %v, %v", readBuf, string(readBuf), bytesRead)
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteAtCircularWierdOffset(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -646,11 +679,12 @@ func TestWriteAtCircularWierdOffset(t *testing.T) {
|
||||
SimpleAssert(t, bytesRead == 7, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf[:7], readTest), "Correct bytes read")
|
||||
log.Printf("readbuf circular %v %v, %v", readBuf, string(readBuf), bytesRead)
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestAppend(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
@ -691,7 +725,6 @@ func TestAppend(t *testing.T) {
|
||||
}
|
||||
SimpleAssert(t, bytesRead == bytesWritten+4, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf, readTestBytes), "Correct bytes read")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func AppendSyncWorker(t *testing.T, ctx context.Context, wg *sync.WaitGroup) {
|
||||
@ -705,13 +738,15 @@ func AppendSyncWorker(t *testing.T, ctx context.Context, wg *sync.WaitGroup) {
|
||||
SimpleAssert(t, bytesWritten == 1, "Correct bytes written")
|
||||
}
|
||||
func TestAppendSync(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
var wg sync.WaitGroup
|
||||
numWorkers := 10
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id-sync", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -729,15 +764,6 @@ func TestAppendSync(t *testing.T) {
|
||||
}
|
||||
log.Printf("read buf : %v", readBuf)
|
||||
SimpleAssert(t, bytesRead == numWorkers, "Correct bytes read")
|
||||
CleanupName(t, ctx, "test-block-id-sync")
|
||||
}
|
||||
|
||||
func TestAppendSyncMultiple(t *testing.T) {
|
||||
numTests := 100
|
||||
for index := 0; index < numTests; index++ {
|
||||
TestAppendSync(t)
|
||||
log.Printf("finished test: %v", index)
|
||||
}
|
||||
}
|
||||
|
||||
func WriteAtSyncWorker(t *testing.T, ctx context.Context, wg *sync.WaitGroup, index int64) {
|
||||
@ -753,13 +779,15 @@ func WriteAtSyncWorker(t *testing.T, ctx context.Context, wg *sync.WaitGroup, in
|
||||
}
|
||||
|
||||
func TestWriteAtSync(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
var wg sync.WaitGroup
|
||||
numWorkers := 10
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id-sync", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -777,22 +805,16 @@ func TestWriteAtSync(t *testing.T) {
|
||||
}
|
||||
log.Printf("read buf : %v", readBuf)
|
||||
SimpleAssert(t, bytesRead == numWorkers, "Correct num bytes read")
|
||||
CleanupName(t, ctx, "test-block-id-sync")
|
||||
}
|
||||
|
||||
func TestWriteAtSyncMultiple(t *testing.T) {
|
||||
numTests := 100
|
||||
for index := 0; index < numTests; index++ {
|
||||
TestWriteAtSync(t)
|
||||
}
|
||||
}
|
||||
|
||||
func TestWriteFile(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
testBytesToWrite := []byte{'T', 'E', 'S', 'T', 'M', 'E', 'S', 'S', 'A', 'G', 'E'}
|
||||
bytesWritten, err := WriteFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts, testBytesToWrite)
|
||||
if err != nil {
|
||||
@ -807,15 +829,16 @@ func TestWriteFile(t *testing.T) {
|
||||
SimpleAssert(t, bytesRead == bytesWritten, "Correct num bytes read")
|
||||
log.Printf("bytes read: %v string: %s", read, string(read))
|
||||
SimpleAssert(t, bytes.Equal(read, testBytesToWrite), "Correct bytes read")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteMeta(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -843,15 +866,16 @@ func TestWriteMeta(t *testing.T) {
|
||||
}
|
||||
log.Printf("meta: %v", fInfo.Meta)
|
||||
SimpleAssert(t, fInfo.Meta["second-test-descriptor"] == "test1", "Retrieved second meta correctly")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestGetAllBlockIds(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
err = MakeFile(ctx, "test-block-id-2", "file-1", fileMeta, fileOpts)
|
||||
err = MakeFile(ctx, "test-block-id-2", "file-2", fileMeta, fileOpts)
|
||||
@ -864,16 +888,17 @@ func TestGetAllBlockIds(t *testing.T) {
|
||||
testBlockIdArr := []string{"test-block-id", "test-block-id-2", "test-block-id-3"}
|
||||
for idx, val := range blockIds {
|
||||
SimpleAssert(t, testBlockIdArr[idx] == val, "Correct blockid value")
|
||||
CleanupName(t, ctx, val)
|
||||
}
|
||||
}
|
||||
|
||||
func TestListFiles(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
err = MakeFile(ctx, "test-block-id-2", "file-1", fileMeta, fileOpts)
|
||||
err = MakeFile(ctx, "test-block-id-2", "file-2", fileMeta, fileOpts)
|
||||
@ -893,19 +918,18 @@ func TestListFiles(t *testing.T) {
|
||||
for idx, val := range files {
|
||||
SimpleAssert(t, val.Name == blockid_1_files[idx], "Correct file name")
|
||||
}
|
||||
CleanupName(t, ctx, "test-block-id")
|
||||
CleanupName(t, ctx, "test-block-id-2")
|
||||
CleanupName(t, ctx, "test-block-id-3")
|
||||
}
|
||||
|
||||
func TestFlushTimer(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
testFlushTimeout := 10 * time.Second
|
||||
SetFlushTimeout(testFlushTimeout)
|
||||
InitDBState()
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -958,22 +982,12 @@ func TestFlushTimer(t *testing.T) {
|
||||
t.Errorf("get data from db error: %v", txErr)
|
||||
}
|
||||
log.Printf("DB Data: %v", dbData)
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestFlushTimerMultiple(t *testing.T) {
|
||||
testFlushTimeout := 1 * time.Second
|
||||
SetFlushTimeout(testFlushTimeout)
|
||||
numTests := 10
|
||||
for index := 0; index < numTests; index++ {
|
||||
TestWriteAt(t)
|
||||
time.Sleep(500 * time.Millisecond)
|
||||
}
|
||||
}
|
||||
|
||||
// time consuming test
|
||||
|
||||
func TestWriteAtMiddle(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
WriteLargeDataFlush(t, ctx)
|
||||
testBytesToWrite := []byte{'T', 'E', 'S', 'T', 'M', 'E', 'S', 'S', 'A', 'G', 'E'}
|
||||
@ -989,23 +1003,26 @@ func TestWriteAtMiddle(t *testing.T) {
|
||||
log.Printf("readBuf: %v %v", readBuf, string(readBuf))
|
||||
SimpleAssert(t, bytesRead == bytesWritten, "Correct num bytes read")
|
||||
SimpleAssert(t, bytes.Equal(readBuf, testBytesToWrite), "read correct bytes")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteLargeDataFlush(t *testing.T) {
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
ctx := context.Background()
|
||||
WriteLargeDataFlush(t, ctx)
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
func TestWriteLargeDataNoFlush(t *testing.T) {
|
||||
InitDBState()
|
||||
initTestDb(t)
|
||||
defer cleanupTestDB(t)
|
||||
|
||||
writeSize := int64(64 - 16)
|
||||
fullWriteSize := int64(1024 * units.Megabyte)
|
||||
fullWriteSize := int64(64 * UnitsKB)
|
||||
ctx := context.Background()
|
||||
fileMeta := make(FileMeta)
|
||||
fileMeta["test-descriptor"] = true
|
||||
fileOpts := FileOptsType{MaxSize: int64(5 * units.Gigabyte), Circular: false, IJson: false}
|
||||
fileOpts := FileOptsType{MaxSize: bigFileSize, Circular: false, IJson: false}
|
||||
err := MakeFile(ctx, "test-block-id", "file-1", fileMeta, fileOpts)
|
||||
if err != nil {
|
||||
t.Fatalf("MakeFile error: %v", err)
|
||||
@ -1028,11 +1045,13 @@ func TestWriteLargeDataNoFlush(t *testing.T) {
|
||||
copy(hashBuf, hash.Sum(nil))
|
||||
bytesWritten, err := WriteAt(ctx, "test-block-id", "file-1", writeBuf, writeIndex)
|
||||
if int64(bytesWritten) != writeSize {
|
||||
log.Printf("write issue: %v %v \n", bytesWritten, writeSize)
|
||||
t.Errorf("write issue: %v %v %v err:%v\n", bytesWritten, writeSize, writeIndex, err)
|
||||
return
|
||||
}
|
||||
if err != nil {
|
||||
log.Printf("error: %v", err)
|
||||
t.Errorf("Write At error: %v\n", err)
|
||||
return
|
||||
}
|
||||
writeIndex += int64(bytesWritten)
|
||||
}
|
||||
@ -1060,7 +1079,6 @@ func TestWriteLargeDataNoFlush(t *testing.T) {
|
||||
}
|
||||
log.Printf("final hash: %v readBuf: %v, bytesRead: %v", readHashBuf, readBuf, readIndex)
|
||||
SimpleAssert(t, bytes.Equal(readHashBuf, hashBuf), "hashes are equal")
|
||||
Cleanup(t, ctx)
|
||||
}
|
||||
|
||||
// saving this code for later
|
||||
|
@ -17,7 +17,7 @@ import (
|
||||
_ "github.com/golang-migrate/migrate/v4/source/file"
|
||||
"github.com/golang-migrate/migrate/v4/source/iofs"
|
||||
_ "github.com/mattn/go-sqlite3"
|
||||
sh2db "github.com/wavetermdev/waveterm/wavesrv/db"
|
||||
dbfs "github.com/wavetermdev/waveterm/wavesrv/db"
|
||||
|
||||
"github.com/golang-migrate/migrate/v4"
|
||||
)
|
||||
@ -29,7 +29,7 @@ const CmdLineSpecialMigration = 20
|
||||
const RISpecialMigration = 30
|
||||
|
||||
func MakeMigrate() (*migrate.Migrate, error) {
|
||||
fsVar, err := iofs.New(sh2db.MigrationFS, "migrations")
|
||||
fsVar, err := iofs.New(dbfs.MigrationFS, "migrations")
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("opening iofs: %w", err)
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user