filehandle_read.go 2.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100
  1. package mount
  2. import (
  3. "context"
  4. "fmt"
  5. "io"
  6. "github.com/seaweedfs/seaweedfs/weed/filer"
  7. "github.com/seaweedfs/seaweedfs/weed/glog"
  8. "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb"
  9. )
  10. func (fh *FileHandle) lockForRead(startOffset int64, size int) {
  11. fh.dirtyPages.LockForRead(startOffset, startOffset+int64(size))
  12. }
  13. func (fh *FileHandle) unlockForRead(startOffset int64, size int) {
  14. fh.dirtyPages.UnlockForRead(startOffset, startOffset+int64(size))
  15. }
  16. func (fh *FileHandle) readFromDirtyPages(buff []byte, startOffset int64, tsNs int64) (maxStop int64) {
  17. maxStop = fh.dirtyPages.ReadDirtyDataAt(buff, startOffset, tsNs)
  18. return
  19. }
  20. func (fh *FileHandle) readFromChunks(buff []byte, offset int64) (int64, int64, error) {
  21. fh.entryLock.RLock()
  22. defer fh.entryLock.RUnlock()
  23. fileFullPath := fh.FullPath()
  24. entry := fh.GetEntry()
  25. if entry.IsInRemoteOnly() {
  26. glog.V(4).Infof("download remote entry %s", fileFullPath)
  27. err := fh.downloadRemoteEntry(entry)
  28. if err != nil {
  29. glog.V(1).Infof("download remote entry %s: %v", fileFullPath, err)
  30. return 0, 0, err
  31. }
  32. }
  33. fileSize := int64(entry.Attributes.FileSize)
  34. if fileSize == 0 {
  35. fileSize = int64(filer.FileSize(entry.GetEntry()))
  36. }
  37. if fileSize == 0 {
  38. glog.V(1).Infof("empty fh %v", fileFullPath)
  39. return 0, 0, io.EOF
  40. } else if offset == fileSize {
  41. return 0, 0, io.EOF
  42. } else if offset >= fileSize {
  43. glog.V(1).Infof("invalid read, fileSize %d, offset %d for %s", fileSize, offset, fileFullPath)
  44. return 0, 0, io.EOF
  45. }
  46. if offset < int64(len(entry.Content)) {
  47. totalRead := copy(buff, entry.Content[offset:])
  48. glog.V(4).Infof("file handle read cached %s [%d,%d] %d", fileFullPath, offset, offset+int64(totalRead), totalRead)
  49. return int64(totalRead), 0, nil
  50. }
  51. totalRead, ts, err := fh.entryChunkGroup.ReadDataAt(fileSize, buff, offset)
  52. if err != nil && err != io.EOF {
  53. glog.Errorf("file handle read %s: %v", fileFullPath, err)
  54. }
  55. // glog.V(4).Infof("file handle read %s [%d,%d] %d : %v", fileFullPath, offset, offset+int64(totalRead), totalRead, err)
  56. return int64(totalRead), ts, err
  57. }
  58. func (fh *FileHandle) downloadRemoteEntry(entry *LockedEntry) error {
  59. fileFullPath := fh.FullPath()
  60. dir, _ := fileFullPath.DirAndName()
  61. err := fh.wfs.WithFilerClient(false, func(client filer_pb.SeaweedFilerClient) error {
  62. request := &filer_pb.CacheRemoteObjectToLocalClusterRequest{
  63. Directory: string(dir),
  64. Name: entry.Name,
  65. }
  66. glog.V(4).Infof("download entry: %v", request)
  67. resp, err := client.CacheRemoteObjectToLocalCluster(context.Background(), request)
  68. if err != nil {
  69. return fmt.Errorf("CacheRemoteObjectToLocalCluster file %s: %v", fileFullPath, err)
  70. }
  71. entry.SetEntry(resp.Entry)
  72. fh.wfs.metaCache.InsertEntry(context.Background(), filer.FromPbEntry(request.Directory, resp.Entry))
  73. return nil
  74. })
  75. return err
  76. }