1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495 |
- package mount
- import (
- "fmt"
- "io"
- "sync"
- "github.com/seaweedfs/seaweedfs/weed/glog"
- "github.com/seaweedfs/seaweedfs/weed/mount/page_writer"
- "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
- )
- type ChunkedDirtyPages struct {
- fh *FileHandle
- writeWaitGroup sync.WaitGroup
- lastErr error
- collection string
- replication string
- uploadPipeline *page_writer.UploadPipeline
- hasWrites bool
- }
- var (
- _ = page_writer.DirtyPages(&ChunkedDirtyPages{})
- )
- func newMemoryChunkPages(fh *FileHandle, chunkSize int64) *ChunkedDirtyPages {
- dirtyPages := &ChunkedDirtyPages{
- fh: fh,
- }
- swapFileDir := fh.wfs.option.getUniqueCacheDirForWrite()
- dirtyPages.uploadPipeline = page_writer.NewUploadPipeline(fh.wfs.concurrentWriters, chunkSize,
- dirtyPages.saveChunkedFileIntervalToStorage, fh.wfs.option.ConcurrentWriters, swapFileDir)
- return dirtyPages
- }
- func (pages *ChunkedDirtyPages) AddPage(offset int64, data []byte, isSequential bool, tsNs int64) {
- pages.hasWrites = true
- glog.V(4).Infof("%v memory AddPage [%d, %d)", pages.fh.fh, offset, offset+int64(len(data)))
- pages.uploadPipeline.SaveDataAt(data, offset, isSequential, tsNs)
- return
- }
- func (pages *ChunkedDirtyPages) FlushData() error {
- if !pages.hasWrites {
- return nil
- }
- pages.uploadPipeline.FlushAll()
- if pages.lastErr != nil {
- return fmt.Errorf("flush data: %v", pages.lastErr)
- }
- return nil
- }
- func (pages *ChunkedDirtyPages) ReadDirtyDataAt(data []byte, startOffset int64, tsNs int64) (maxStop int64) {
- if !pages.hasWrites {
- return
- }
- return pages.uploadPipeline.MaybeReadDataAt(data, startOffset, tsNs)
- }
- func (pages *ChunkedDirtyPages) saveChunkedFileIntervalToStorage(reader io.Reader, offset int64, size int64, modifiedTsNs int64, cleanupFn func()) {
- defer cleanupFn()
- fileFullPath := pages.fh.FullPath()
- fileName := fileFullPath.Name()
- chunk, err := pages.fh.wfs.saveDataAsChunk(fileFullPath)(reader, fileName, offset, modifiedTsNs)
- if err != nil {
- glog.V(0).Infof("%v saveToStorage [%d,%d): %v", fileFullPath, offset, offset+size, err)
- pages.lastErr = err
- return
- }
- pages.fh.AddChunks([]*filer_pb.FileChunk{chunk})
- pages.fh.entryChunkGroup.AddChunk(chunk)
- glog.V(3).Infof("%v saveToStorage %s [%d,%d)", fileFullPath, chunk.FileId, offset, offset+size)
- }
- func (pages *ChunkedDirtyPages) Destroy() {
- pages.uploadPipeline.Shutdown()
- }
- func (pages *ChunkedDirtyPages) LockForRead(startOffset, stopOffset int64) {
- pages.uploadPipeline.LockForRead(startOffset, stopOffset)
- }
- func (pages *ChunkedDirtyPages) UnlockForRead(startOffset, stopOffset int64) {
- pages.uploadPipeline.UnlockForRead(startOffset, stopOffset)
- }
|