123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445 |
- package storage
- import (
- "fmt"
- "os"
- "path/filepath"
- "runtime"
- "strconv"
- "strings"
- "sync"
- "time"
- "github.com/google/uuid"
- "github.com/seaweedfs/seaweedfs/weed/glog"
- "github.com/seaweedfs/seaweedfs/weed/stats"
- "github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
- "github.com/seaweedfs/seaweedfs/weed/storage/needle"
- "github.com/seaweedfs/seaweedfs/weed/storage/types"
- "github.com/seaweedfs/seaweedfs/weed/util"
- )
- type DiskLocation struct {
- Directory string
- DirectoryUuid string
- IdxDirectory string
- DiskType types.DiskType
- MaxVolumeCount int32
- OriginalMaxVolumeCount int32
- MinFreeSpace util.MinFreeSpace
- volumes map[needle.VolumeId]*Volume
- volumesLock sync.RWMutex
- // erasure coding
- ecVolumes map[needle.VolumeId]*erasure_coding.EcVolume
- ecVolumesLock sync.RWMutex
- isDiskSpaceLow bool
- }
- func GenerateDirUuid(dir string) (dirUuidString string, err error) {
- glog.V(1).Infof("Getting uuid of volume directory:%s", dir)
- dirUuidString = ""
- fileName := dir + "/vol_dir.uuid"
- if !util.FileExists(fileName) {
- dirUuid, _ := uuid.NewRandom()
- dirUuidString = dirUuid.String()
- writeErr := util.WriteFile(fileName, []byte(dirUuidString), 0644)
- if writeErr != nil {
- return "", fmt.Errorf("failed to write uuid to %s : %v", fileName, writeErr)
- }
- } else {
- uuidData, readErr := os.ReadFile(fileName)
- if readErr != nil {
- return "", fmt.Errorf("failed to read uuid from %s : %v", fileName, readErr)
- }
- dirUuidString = string(uuidData)
- }
- return dirUuidString, nil
- }
- func NewDiskLocation(dir string, maxVolumeCount int32, minFreeSpace util.MinFreeSpace, idxDir string, diskType types.DiskType) *DiskLocation {
- dir = util.ResolvePath(dir)
- if idxDir == "" {
- idxDir = dir
- } else {
- idxDir = util.ResolvePath(idxDir)
- }
- dirUuid, err := GenerateDirUuid(dir)
- if err != nil {
- glog.Fatalf("cannot generate uuid of dir %s: %v", dir, err)
- }
- location := &DiskLocation{
- Directory: dir,
- DirectoryUuid: dirUuid,
- IdxDirectory: idxDir,
- DiskType: diskType,
- MaxVolumeCount: maxVolumeCount,
- OriginalMaxVolumeCount: maxVolumeCount,
- MinFreeSpace: minFreeSpace,
- }
- location.volumes = make(map[needle.VolumeId]*Volume)
- location.ecVolumes = make(map[needle.VolumeId]*erasure_coding.EcVolume)
- go location.CheckDiskSpace()
- return location
- }
- func volumeIdFromFileName(filename string) (needle.VolumeId, string, error) {
- if isValidVolume(filename) {
- base := filename[:len(filename)-4]
- collection, volumeId, err := parseCollectionVolumeId(base)
- return volumeId, collection, err
- }
- return 0, "", fmt.Errorf("file is not a volume: %s", filename)
- }
- func parseCollectionVolumeId(base string) (collection string, vid needle.VolumeId, err error) {
- i := strings.LastIndex(base, "_")
- if i > 0 {
- collection, base = base[0:i], base[i+1:]
- }
- vol, err := needle.NewVolumeId(base)
- return collection, vol, err
- }
- func isValidVolume(basename string) bool {
- return strings.HasSuffix(basename, ".idx") || strings.HasSuffix(basename, ".vif")
- }
- func getValidVolumeName(basename string) string {
- if isValidVolume(basename) {
- return basename[:len(basename)-4]
- }
- return ""
- }
- func (l *DiskLocation) loadExistingVolume(dirEntry os.DirEntry, needleMapKind NeedleMapKind, skipIfEcVolumesExists bool, ldbTimeout int64) bool {
- basename := dirEntry.Name()
- if dirEntry.IsDir() {
- return false
- }
- volumeName := getValidVolumeName(basename)
- if volumeName == "" {
- return false
- }
- // skip if ec volumes exists
- if skipIfEcVolumesExists {
- if util.FileExists(l.Directory + "/" + volumeName + ".ecx") {
- return false
- }
- }
- // check for incomplete volume
- noteFile := l.Directory + "/" + volumeName + ".note"
- if util.FileExists(noteFile) {
- note, _ := os.ReadFile(noteFile)
- glog.Warningf("volume %s was not completed: %s", volumeName, string(note))
- removeVolumeFiles(l.Directory + "/" + volumeName)
- removeVolumeFiles(l.IdxDirectory + "/" + volumeName)
- return false
- }
- // parse out collection, volume id
- vid, collection, err := volumeIdFromFileName(basename)
- if err != nil {
- glog.Warningf("get volume id failed, %s, err : %s", volumeName, err)
- return false
- }
- // avoid loading one volume more than once
- l.volumesLock.RLock()
- _, found := l.volumes[vid]
- l.volumesLock.RUnlock()
- if found {
- glog.V(1).Infof("loaded volume, %v", vid)
- return true
- }
- // load the volume
- v, e := NewVolume(l.Directory, l.IdxDirectory, collection, vid, needleMapKind, nil, nil, 0, 0, ldbTimeout)
- if e != nil {
- glog.V(0).Infof("new volume %s error %s", volumeName, e)
- return false
- }
- l.SetVolume(vid, v)
- size, _, _ := v.FileStat()
- glog.V(0).Infof("data file %s, replication=%s v=%d size=%d ttl=%s",
- l.Directory+"/"+volumeName+".dat", v.ReplicaPlacement, v.Version(), size, v.Ttl.String())
- return true
- }
- func (l *DiskLocation) concurrentLoadingVolumes(needleMapKind NeedleMapKind, concurrency int, ldbTimeout int64) {
- task_queue := make(chan os.DirEntry, 10*concurrency)
- go func() {
- foundVolumeNames := make(map[string]bool)
- if dirEntries, err := os.ReadDir(l.Directory); err == nil {
- for _, entry := range dirEntries {
- volumeName := getValidVolumeName(entry.Name())
- if volumeName == "" {
- continue
- }
- if _, found := foundVolumeNames[volumeName]; !found {
- foundVolumeNames[volumeName] = true
- task_queue <- entry
- }
- }
- }
- close(task_queue)
- }()
- var wg sync.WaitGroup
- for workerNum := 0; workerNum < concurrency; workerNum++ {
- wg.Add(1)
- go func() {
- defer wg.Done()
- for fi := range task_queue {
- _ = l.loadExistingVolume(fi, needleMapKind, true, ldbTimeout)
- }
- }()
- }
- wg.Wait()
- }
- func (l *DiskLocation) loadExistingVolumes(needleMapKind NeedleMapKind, ldbTimeout int64) {
- workerNum := runtime.NumCPU()
- val, ok := os.LookupEnv("GOMAXPROCS")
- if ok {
- num, err := strconv.Atoi(val)
- if err != nil || num < 1 {
- num = 10
- glog.Warningf("failed to set worker number from GOMAXPROCS , set to default:10")
- }
- workerNum = num
- } else {
- if workerNum <= 10 {
- workerNum = 10
- }
- }
- l.concurrentLoadingVolumes(needleMapKind, workerNum, ldbTimeout)
- glog.V(0).Infof("Store started on dir: %s with %d volumes max %d", l.Directory, len(l.volumes), l.MaxVolumeCount)
- l.loadAllEcShards()
- glog.V(0).Infof("Store started on dir: %s with %d ec shards", l.Directory, len(l.ecVolumes))
- }
- func (l *DiskLocation) DeleteCollectionFromDiskLocation(collection string) (e error) {
- l.volumesLock.Lock()
- delVolsMap := l.unmountVolumeByCollection(collection)
- l.volumesLock.Unlock()
- l.ecVolumesLock.Lock()
- delEcVolsMap := l.unmountEcVolumeByCollection(collection)
- l.ecVolumesLock.Unlock()
- errChain := make(chan error, 2)
- var wg sync.WaitGroup
- wg.Add(2)
- go func() {
- for _, v := range delVolsMap {
- if err := v.Destroy(); err != nil {
- errChain <- err
- }
- }
- wg.Done()
- }()
- go func() {
- for _, v := range delEcVolsMap {
- v.Destroy()
- }
- wg.Done()
- }()
- go func() {
- wg.Wait()
- close(errChain)
- }()
- errBuilder := strings.Builder{}
- for err := range errChain {
- errBuilder.WriteString(err.Error())
- errBuilder.WriteString("; ")
- }
- if errBuilder.Len() > 0 {
- e = fmt.Errorf(errBuilder.String())
- }
- return
- }
- func (l *DiskLocation) deleteVolumeById(vid needle.VolumeId) (found bool, e error) {
- v, ok := l.volumes[vid]
- if !ok {
- return
- }
- e = v.Destroy()
- if e != nil {
- return
- }
- found = true
- delete(l.volumes, vid)
- return
- }
- func (l *DiskLocation) LoadVolume(vid needle.VolumeId, needleMapKind NeedleMapKind) bool {
- if fileInfo, found := l.LocateVolume(vid); found {
- return l.loadExistingVolume(fileInfo, needleMapKind, false, 0)
- }
- return false
- }
- var ErrVolumeNotFound = fmt.Errorf("volume not found")
- func (l *DiskLocation) DeleteVolume(vid needle.VolumeId) error {
- l.volumesLock.Lock()
- defer l.volumesLock.Unlock()
- _, ok := l.volumes[vid]
- if !ok {
- return ErrVolumeNotFound
- }
- _, err := l.deleteVolumeById(vid)
- return err
- }
- func (l *DiskLocation) UnloadVolume(vid needle.VolumeId) error {
- l.volumesLock.Lock()
- defer l.volumesLock.Unlock()
- v, ok := l.volumes[vid]
- if !ok {
- return ErrVolumeNotFound
- }
- v.Close()
- delete(l.volumes, vid)
- return nil
- }
- func (l *DiskLocation) unmountVolumeByCollection(collectionName string) map[needle.VolumeId]*Volume {
- deltaVols := make(map[needle.VolumeId]*Volume, 0)
- for k, v := range l.volumes {
- if v.Collection == collectionName && !v.isCompacting && !v.isCommitCompacting {
- deltaVols[k] = v
- }
- }
- for k := range deltaVols {
- delete(l.volumes, k)
- }
- return deltaVols
- }
- func (l *DiskLocation) SetVolume(vid needle.VolumeId, volume *Volume) {
- l.volumesLock.Lock()
- defer l.volumesLock.Unlock()
- l.volumes[vid] = volume
- volume.location = l
- }
- func (l *DiskLocation) FindVolume(vid needle.VolumeId) (*Volume, bool) {
- l.volumesLock.RLock()
- defer l.volumesLock.RUnlock()
- v, ok := l.volumes[vid]
- return v, ok
- }
- func (l *DiskLocation) VolumesLen() int {
- l.volumesLock.RLock()
- defer l.volumesLock.RUnlock()
- return len(l.volumes)
- }
- func (l *DiskLocation) SetStopping() {
- l.volumesLock.Lock()
- for _, v := range l.volumes {
- v.SyncToDisk()
- }
- l.volumesLock.Unlock()
- return
- }
- func (l *DiskLocation) Close() {
- l.volumesLock.Lock()
- for _, v := range l.volumes {
- v.Close()
- }
- l.volumesLock.Unlock()
- l.ecVolumesLock.Lock()
- for _, ecVolume := range l.ecVolumes {
- ecVolume.Close()
- }
- l.ecVolumesLock.Unlock()
- return
- }
- func (l *DiskLocation) LocateVolume(vid needle.VolumeId) (os.DirEntry, bool) {
- // println("LocateVolume", vid, "on", l.Directory)
- if dirEntries, err := os.ReadDir(l.Directory); err == nil {
- for _, entry := range dirEntries {
- // println("checking", entry.Name(), "...")
- volId, _, err := volumeIdFromFileName(entry.Name())
- // println("volId", volId, "err", err)
- if vid == volId && err == nil {
- return entry, true
- }
- }
- }
- return nil, false
- }
- func (l *DiskLocation) UnUsedSpace(volumeSizeLimit uint64) (unUsedSpace uint64) {
- l.volumesLock.RLock()
- defer l.volumesLock.RUnlock()
- for _, vol := range l.volumes {
- if vol.IsReadOnly() {
- continue
- }
- datSize, idxSize, _ := vol.FileStat()
- unUsedSpace += volumeSizeLimit - (datSize + idxSize)
- }
- return
- }
- func (l *DiskLocation) CheckDiskSpace() {
- for {
- if dir, e := filepath.Abs(l.Directory); e == nil {
- s := stats.NewDiskStatus(dir)
- stats.VolumeServerResourceGauge.WithLabelValues(l.Directory, "all").Set(float64(s.All))
- stats.VolumeServerResourceGauge.WithLabelValues(l.Directory, "used").Set(float64(s.Used))
- stats.VolumeServerResourceGauge.WithLabelValues(l.Directory, "free").Set(float64(s.Free))
- isLow, desc := l.MinFreeSpace.IsLow(s.Free, s.PercentFree)
- if isLow != l.isDiskSpaceLow {
- l.isDiskSpaceLow = !l.isDiskSpaceLow
- }
- logLevel := glog.Level(4)
- if l.isDiskSpaceLow {
- logLevel = glog.Level(0)
- }
- glog.V(logLevel).Infof("dir %s %s", dir, desc)
- }
- time.Sleep(time.Minute)
- }
- }
|