123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147 |
- package filer2
- import (
- "context"
- "fmt"
- "io"
- "sync"
- "github.com/chrislusf/seaweedfs/weed/glog"
- "github.com/chrislusf/seaweedfs/weed/pb/filer_pb"
- "github.com/chrislusf/seaweedfs/weed/util/chunk_cache"
- "github.com/chrislusf/seaweedfs/weed/wdclient"
- )
- type ChunkReadAt struct {
- masterClient *wdclient.MasterClient
- chunkViews []*ChunkView
- buffer []byte
- bufferOffset int64
- lookupFileId func(fileId string) (targetUrl string, err error)
- readerLock sync.Mutex
- chunkCache *chunk_cache.ChunkCache
- }
- // var _ = io.ReaderAt(&ChunkReadAt{})
- type LookupFileIdFunctionType func(fileId string) (targetUrl string, err error)
- func LookupFn(filerClient filer_pb.FilerClient) LookupFileIdFunctionType {
- return func(fileId string) (targetUrl string, err error) {
- err = filerClient.WithFilerClient(func(client filer_pb.SeaweedFilerClient) error {
- vid := VolumeId(fileId)
- resp, err := client.LookupVolume(context.Background(), &filer_pb.LookupVolumeRequest{
- VolumeIds: []string{vid},
- })
- if err != nil {
- return err
- }
- locations := resp.LocationsMap[vid]
- if locations == nil || len(locations.Locations) == 0 {
- glog.V(0).Infof("failed to locate %s", fileId)
- return fmt.Errorf("failed to locate %s", fileId)
- }
- volumeServerAddress := filerClient.AdjustedUrl(locations.Locations[0].Url)
- targetUrl = fmt.Sprintf("http://%s/%s", volumeServerAddress, fileId)
- return nil
- })
- return
- }
- }
- func NewChunkReaderAtFromClient(filerClient filer_pb.FilerClient, chunkViews []*ChunkView, chunkCache *chunk_cache.ChunkCache) *ChunkReadAt {
- return &ChunkReadAt{
- chunkViews: chunkViews,
- lookupFileId: LookupFn(filerClient),
- bufferOffset: -1,
- chunkCache: chunkCache,
- }
- }
- func (c *ChunkReadAt) ReadAt(p []byte, offset int64) (n int, err error) {
- c.readerLock.Lock()
- defer c.readerLock.Unlock()
- for n < len(p) && err == nil {
- readCount, readErr := c.doReadAt(p[n:], offset+int64(n))
- n += readCount
- err = readErr
- if readCount == 0 {
- return n, io.EOF
- }
- }
- return
- }
- func (c *ChunkReadAt) doReadAt(p []byte, offset int64) (n int, err error) {
- var found bool
- for _, chunk := range c.chunkViews {
- if chunk.LogicOffset <= offset && offset < chunk.LogicOffset+int64(chunk.Size) {
- found = true
- if c.bufferOffset != chunk.LogicOffset {
- c.buffer, err = c.fetchChunkData(chunk)
- if err != nil {
- glog.Errorf("fetching chunk %+v: %v\n", chunk, err)
- }
- c.bufferOffset = chunk.LogicOffset
- }
- break
- }
- }
- if !found {
- return 0, io.EOF
- }
- if err == nil {
- n = copy(p, c.buffer[offset-c.bufferOffset:])
- }
- // fmt.Printf("> doReadAt [%d,%d), buffer:[%d,%d)\n", offset, offset+int64(n), c.bufferOffset, c.bufferOffset+int64(len(c.buffer)))
- return
- }
- func (c *ChunkReadAt) fetchChunkData(chunkView *ChunkView) (data []byte, err error) {
- glog.V(4).Infof("fetchChunkData %s [%d,%d)\n", chunkView.FileId, chunkView.LogicOffset, chunkView.LogicOffset+int64(chunkView.Size))
- hasDataInCache := false
- chunkData := c.chunkCache.GetChunk(chunkView.FileId, chunkView.ChunkSize)
- if chunkData != nil {
- glog.V(4).Infof("cache hit %s [%d,%d)", chunkView.FileId, chunkView.LogicOffset, chunkView.LogicOffset+int64(chunkView.Size))
- hasDataInCache = true
- } else {
- chunkData, err = c.doFetchFullChunkData(chunkView.FileId, chunkView.CipherKey, chunkView.IsGzipped)
- if err != nil {
- return nil, err
- }
- }
- if int64(len(chunkData)) < chunkView.Offset+int64(chunkView.Size) {
- glog.Errorf("unexpected larger cached:%v chunk %s [%d,%d) than %d", hasDataInCache, chunkView.FileId, chunkView.Offset, chunkView.Offset+int64(chunkView.Size), len(chunkData))
- return nil, fmt.Errorf("unexpected larger cached:%v chunk %s [%d,%d) than %d", hasDataInCache, chunkView.FileId, chunkView.Offset, chunkView.Offset+int64(chunkView.Size), len(chunkData))
- }
- data = chunkData[chunkView.Offset : chunkView.Offset+int64(chunkView.Size)]
- if !hasDataInCache {
- c.chunkCache.SetChunk(chunkView.FileId, chunkData)
- }
- return data, nil
- }
- func (c *ChunkReadAt) doFetchFullChunkData(fileId string, cipherKey []byte, isGzipped bool) ([]byte, error) {
- return fetchChunk(c.lookupFileId, fileId, cipherKey, isGzipped)
- }
|