123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271 |
- package topology
- import (
- "fmt"
- "github.com/seaweedfs/seaweedfs/weed/storage/types"
- "github.com/seaweedfs/seaweedfs/weed/util"
- "sync"
- "sync/atomic"
- "github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
- "github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding"
- "github.com/seaweedfs/seaweedfs/weed/storage/needle"
- "github.com/seaweedfs/seaweedfs/weed/storage"
- )
- type Disk struct {
- NodeImpl
- volumes map[needle.VolumeId]storage.VolumeInfo
- ecShards map[needle.VolumeId]*erasure_coding.EcVolumeInfo
- ecShardsLock sync.RWMutex
- }
- func NewDisk(diskType string) *Disk {
- s := &Disk{}
- s.id = NodeId(diskType)
- s.nodeType = "Disk"
- s.diskUsages = newDiskUsages()
- s.volumes = make(map[needle.VolumeId]storage.VolumeInfo, 2)
- s.ecShards = make(map[needle.VolumeId]*erasure_coding.EcVolumeInfo, 2)
- s.NodeImpl.value = s
- return s
- }
- type DiskUsages struct {
- sync.RWMutex
- usages map[types.DiskType]*DiskUsageCounts
- }
- func newDiskUsages() *DiskUsages {
- return &DiskUsages{
- usages: make(map[types.DiskType]*DiskUsageCounts),
- }
- }
- func (d *DiskUsages) negative() *DiskUsages {
- d.RLock()
- defer d.RUnlock()
- t := newDiskUsages()
- for diskType, b := range d.usages {
- a := t.getOrCreateDisk(diskType)
- a.volumeCount = -b.volumeCount
- a.remoteVolumeCount = -b.remoteVolumeCount
- a.activeVolumeCount = -b.activeVolumeCount
- a.ecShardCount = -b.ecShardCount
- a.maxVolumeCount = -b.maxVolumeCount
- }
- return t
- }
- func (d *DiskUsages) ToDiskInfo() map[string]*master_pb.DiskInfo {
- ret := make(map[string]*master_pb.DiskInfo)
- for diskType, diskUsageCounts := range d.usages {
- m := &master_pb.DiskInfo{
- VolumeCount: diskUsageCounts.volumeCount,
- MaxVolumeCount: diskUsageCounts.maxVolumeCount,
- FreeVolumeCount: diskUsageCounts.maxVolumeCount - diskUsageCounts.volumeCount,
- ActiveVolumeCount: diskUsageCounts.activeVolumeCount,
- RemoteVolumeCount: diskUsageCounts.remoteVolumeCount,
- }
- ret[string(diskType)] = m
- }
- return ret
- }
- func (d *DiskUsages) FreeSpace() (freeSpace int64) {
- d.RLock()
- defer d.RUnlock()
- for _, diskUsage := range d.usages {
- freeSpace += diskUsage.FreeSpace()
- }
- return
- }
- func (d *DiskUsages) GetMaxVolumeCount() (maxVolumeCount int64) {
- d.RLock()
- defer d.RUnlock()
- for _, diskUsage := range d.usages {
- maxVolumeCount += diskUsage.maxVolumeCount
- }
- return
- }
- type DiskUsageCounts struct {
- volumeCount int64
- remoteVolumeCount int64
- activeVolumeCount int64
- ecShardCount int64
- maxVolumeCount int64
- }
- func (a *DiskUsageCounts) addDiskUsageCounts(b *DiskUsageCounts) {
- atomic.AddInt64(&a.volumeCount, b.volumeCount)
- atomic.AddInt64(&a.remoteVolumeCount, b.remoteVolumeCount)
- atomic.AddInt64(&a.activeVolumeCount, b.activeVolumeCount)
- atomic.AddInt64(&a.ecShardCount, b.ecShardCount)
- atomic.AddInt64(&a.maxVolumeCount, b.maxVolumeCount)
- }
- func (a *DiskUsageCounts) FreeSpace() int64 {
- freeVolumeSlotCount := a.maxVolumeCount + a.remoteVolumeCount - a.volumeCount
- if a.ecShardCount > 0 {
- freeVolumeSlotCount = freeVolumeSlotCount - a.ecShardCount/erasure_coding.DataShardsCount - 1
- }
- return freeVolumeSlotCount
- }
- func (a *DiskUsageCounts) minus(b *DiskUsageCounts) *DiskUsageCounts {
- return &DiskUsageCounts{
- volumeCount: a.volumeCount - b.volumeCount,
- remoteVolumeCount: a.remoteVolumeCount - b.remoteVolumeCount,
- activeVolumeCount: a.activeVolumeCount - b.activeVolumeCount,
- ecShardCount: a.ecShardCount - b.ecShardCount,
- maxVolumeCount: a.maxVolumeCount - b.maxVolumeCount,
- }
- }
- func (du *DiskUsages) getOrCreateDisk(diskType types.DiskType) *DiskUsageCounts {
- du.Lock()
- defer du.Unlock()
- t, found := du.usages[diskType]
- if found {
- return t
- }
- t = &DiskUsageCounts{}
- du.usages[diskType] = t
- return t
- }
- func (d *Disk) String() string {
- d.RLock()
- defer d.RUnlock()
- return fmt.Sprintf("Disk:%s, volumes:%v, ecShards:%v", d.NodeImpl.String(), d.volumes, d.ecShards)
- }
- func (d *Disk) AddOrUpdateVolume(v storage.VolumeInfo) (isNew, isChanged bool) {
- d.Lock()
- defer d.Unlock()
- return d.doAddOrUpdateVolume(v)
- }
- func (d *Disk) doAddOrUpdateVolume(v storage.VolumeInfo) (isNew, isChanged bool) {
- deltaDiskUsages := newDiskUsages()
- deltaDiskUsage := deltaDiskUsages.getOrCreateDisk(types.ToDiskType(v.DiskType))
- if oldV, ok := d.volumes[v.Id]; !ok {
- d.volumes[v.Id] = v
- deltaDiskUsage.volumeCount = 1
- if v.IsRemote() {
- deltaDiskUsage.remoteVolumeCount = 1
- }
- if !v.ReadOnly {
- deltaDiskUsage.activeVolumeCount = 1
- }
- d.UpAdjustMaxVolumeId(v.Id)
- d.UpAdjustDiskUsageDelta(deltaDiskUsages)
- isNew = true
- } else {
- if oldV.IsRemote() != v.IsRemote() {
- if v.IsRemote() {
- deltaDiskUsage.remoteVolumeCount = 1
- }
- if oldV.IsRemote() {
- deltaDiskUsage.remoteVolumeCount = -1
- }
- d.UpAdjustDiskUsageDelta(deltaDiskUsages)
- }
- isChanged = d.volumes[v.Id].ReadOnly != v.ReadOnly
- d.volumes[v.Id] = v
- }
- return
- }
- func (d *Disk) GetVolumes() (ret []storage.VolumeInfo) {
- d.RLock()
- for _, v := range d.volumes {
- ret = append(ret, v)
- }
- d.RUnlock()
- return ret
- }
- func (d *Disk) GetVolumesById(id needle.VolumeId) (storage.VolumeInfo, error) {
- d.RLock()
- defer d.RUnlock()
- vInfo, ok := d.volumes[id]
- if ok {
- return vInfo, nil
- } else {
- return storage.VolumeInfo{}, fmt.Errorf("volumeInfo not found")
- }
- }
- func (d *Disk) GetDataCenter() *DataCenter {
- dn := d.Parent()
- rack := dn.Parent()
- dcNode := rack.Parent()
- dcValue := dcNode.GetValue()
- return dcValue.(*DataCenter)
- }
- func (d *Disk) GetRack() *Rack {
- return d.Parent().Parent().(*NodeImpl).value.(*Rack)
- }
- func (d *Disk) GetTopology() *Topology {
- p := d.Parent()
- for p.Parent() != nil {
- p = p.Parent()
- }
- t := p.(*Topology)
- return t
- }
- func (d *Disk) ToMap() interface{} {
- ret := make(map[string]interface{})
- diskUsage := d.diskUsages.getOrCreateDisk(types.ToDiskType(string(d.Id())))
- ret["Volumes"] = diskUsage.volumeCount
- ret["VolumeIds"] = d.GetVolumeIds()
- ret["EcShards"] = diskUsage.ecShardCount
- ret["Max"] = diskUsage.maxVolumeCount
- ret["Free"] = d.FreeSpace()
- return ret
- }
- func (d *Disk) FreeSpace() int64 {
- t := d.diskUsages.getOrCreateDisk(types.ToDiskType(string(d.Id())))
- return t.FreeSpace()
- }
- func (d *Disk) ToDiskInfo() *master_pb.DiskInfo {
- diskUsage := d.diskUsages.getOrCreateDisk(types.ToDiskType(string(d.Id())))
- m := &master_pb.DiskInfo{
- Type: string(d.Id()),
- VolumeCount: diskUsage.volumeCount,
- MaxVolumeCount: diskUsage.maxVolumeCount,
- FreeVolumeCount: diskUsage.maxVolumeCount - diskUsage.volumeCount,
- ActiveVolumeCount: diskUsage.activeVolumeCount,
- RemoteVolumeCount: diskUsage.remoteVolumeCount,
- }
- for _, v := range d.GetVolumes() {
- m.VolumeInfos = append(m.VolumeInfos, v.ToVolumeInformationMessage())
- }
- for _, ecv := range d.GetEcShards() {
- m.EcShardInfos = append(m.EcShardInfos, ecv.ToVolumeEcShardInformationMessage())
- }
- return m
- }
- // GetVolumeIds returns the human readable volume ids limited to count of max 100.
- func (d *Disk) GetVolumeIds() string {
- d.RLock()
- defer d.RUnlock()
- ids := make([]int, 0, len(d.volumes))
- for k := range d.volumes {
- ids = append(ids, int(k))
- }
- return util.HumanReadableIntsMax(100, ids...)
- }
|