b00d328107
This is not currently working correctly — when setting the limit to 0 all existing files are destroyed. When setting it back to a non-zero value the old files are returned from the allocated disk. This at least starts getting the code pathways in place to handle it and introduces some additional lock safety on the VHD management code.
288 lines
8.4 KiB
Go
288 lines
8.4 KiB
Go
package server
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"os"
|
|
"runtime"
|
|
"sync"
|
|
"time"
|
|
|
|
"emperror.dev/errors"
|
|
"github.com/apex/log"
|
|
"github.com/gammazero/workerpool"
|
|
"github.com/goccy/go-json"
|
|
|
|
"github.com/pterodactyl/wings/config"
|
|
"github.com/pterodactyl/wings/environment"
|
|
"github.com/pterodactyl/wings/environment/docker"
|
|
"github.com/pterodactyl/wings/remote"
|
|
"github.com/pterodactyl/wings/server/filesystem"
|
|
)
|
|
|
|
type Manager struct {
|
|
mu sync.RWMutex
|
|
client remote.Client
|
|
skipVhdInitialization bool
|
|
servers []*Server
|
|
}
|
|
|
|
// NewManager returns a new server manager instance. This will boot up all the
|
|
// servers that are currently present on the filesystem and set them into the
|
|
// manager.
|
|
func NewManager(ctx context.Context, client remote.Client, skipVhdInit bool) (*Manager, error) {
|
|
m := NewEmptyManager(client)
|
|
m.skipVhdInitialization = skipVhdInit
|
|
if err := m.init(ctx); err != nil {
|
|
return nil, err
|
|
}
|
|
return m, nil
|
|
}
|
|
|
|
// NewEmptyManager returns a new empty manager collection without actually
|
|
// loading any of the servers from the disk. This allows the caller to set their
|
|
// own servers into the collection as needed.
|
|
func NewEmptyManager(client remote.Client) *Manager {
|
|
return &Manager{client: client}
|
|
}
|
|
|
|
// Client returns the HTTP client interface that allows interaction with the
|
|
// Panel API.
|
|
func (m *Manager) Client() remote.Client {
|
|
return m.client
|
|
}
|
|
|
|
// Len returns the count of servers stored in the manager instance.
|
|
func (m *Manager) Len() int {
|
|
m.mu.RLock()
|
|
defer m.mu.RUnlock()
|
|
return len(m.servers)
|
|
}
|
|
|
|
// Keys returns all of the server UUIDs stored in the manager set.
|
|
func (m *Manager) Keys() []string {
|
|
m.mu.RLock()
|
|
defer m.mu.RUnlock()
|
|
keys := make([]string, len(m.servers))
|
|
for i, s := range m.servers {
|
|
keys[i] = s.ID()
|
|
}
|
|
return keys
|
|
}
|
|
|
|
// Put replaces all the current values in the collection with the value that
|
|
// is passed through.
|
|
func (m *Manager) Put(s []*Server) {
|
|
m.mu.Lock()
|
|
m.servers = s
|
|
m.mu.Unlock()
|
|
}
|
|
|
|
// All returns all the items in the collection.
|
|
func (m *Manager) All() []*Server {
|
|
m.mu.RLock()
|
|
defer m.mu.RUnlock()
|
|
return m.servers
|
|
}
|
|
|
|
// Add adds an item to the collection store.
|
|
func (m *Manager) Add(s *Server) {
|
|
m.mu.Lock()
|
|
m.servers = append(m.servers, s)
|
|
m.mu.Unlock()
|
|
}
|
|
|
|
// Get returns a single server instance and a boolean value indicating if it was
|
|
// found in the global collection or not.
|
|
func (m *Manager) Get(uuid string) (*Server, bool) {
|
|
match := m.Find(func(server *Server) bool {
|
|
return server.ID() == uuid
|
|
})
|
|
return match, match != nil
|
|
}
|
|
|
|
// Filter returns only those items matching the filter criteria.
|
|
func (m *Manager) Filter(filter func(match *Server) bool) []*Server {
|
|
m.mu.RLock()
|
|
defer m.mu.RUnlock()
|
|
r := make([]*Server, 0)
|
|
for _, v := range m.servers {
|
|
if filter(v) {
|
|
r = append(r, v)
|
|
}
|
|
}
|
|
return r
|
|
}
|
|
|
|
// Find returns a single element from the collection matching the filter. If
|
|
// nothing is found a nil result is returned.
|
|
func (m *Manager) Find(filter func(match *Server) bool) *Server {
|
|
m.mu.RLock()
|
|
defer m.mu.RUnlock()
|
|
for _, v := range m.servers {
|
|
if filter(v) {
|
|
return v
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Remove removes all items from the collection that match the filter function.
|
|
func (m *Manager) Remove(filter func(match *Server) bool) {
|
|
m.mu.Lock()
|
|
defer m.mu.Unlock()
|
|
r := make([]*Server, 0)
|
|
for _, v := range m.servers {
|
|
if !filter(v) {
|
|
r = append(r, v)
|
|
}
|
|
}
|
|
m.servers = r
|
|
}
|
|
|
|
// PersistStates writes the current environment states to the disk for each
|
|
// server. This is generally called at a specific interval defined in the root
|
|
// runner command to avoid hammering disk I/O when tons of server switch states
|
|
// at once. It is fine if this file falls slightly out of sync, it is just here
|
|
// to make recovering from an unexpected system reboot a little easier.
|
|
func (m *Manager) PersistStates() error {
|
|
states := map[string]string{}
|
|
for _, s := range m.All() {
|
|
states[s.ID()] = s.Environment.State()
|
|
}
|
|
data, err := json.Marshal(states)
|
|
if err != nil {
|
|
return errors.WithStack(err)
|
|
}
|
|
if err := os.WriteFile(config.Get().System.GetStatesPath(), data, 0o644); err != nil {
|
|
return errors.WithStack(err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// ReadStates returns the state of the servers.
|
|
func (m *Manager) ReadStates() (map[string]string, error) {
|
|
f, err := os.OpenFile(config.Get().System.GetStatesPath(), os.O_RDONLY|os.O_CREATE, 0o644)
|
|
if err != nil {
|
|
return nil, errors.WithStack(err)
|
|
}
|
|
defer f.Close()
|
|
var states map[string]string
|
|
if err := json.NewDecoder(f).Decode(&states); err != nil && err != io.EOF {
|
|
return nil, errors.WithStack(err)
|
|
}
|
|
out := make(map[string]string, 0)
|
|
// Only return states for servers that we're currently tracking in the system.
|
|
for id, state := range states {
|
|
if _, ok := m.Get(id); ok {
|
|
out[id] = state
|
|
}
|
|
}
|
|
return out, nil
|
|
}
|
|
|
|
// InitServer initializes a server using a data byte array. This will be
|
|
// marshaled into the given struct using a YAML marshaler. This will also
|
|
// configure the given environment for a server.
|
|
func (m *Manager) InitServer(ctx context.Context, data remote.ServerConfigurationResponse) (*Server, error) {
|
|
s, err := New(m.client)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Setup the base server configuration data which will be used for all of the
|
|
// remaining functionality in this call.
|
|
if err := s.SyncWithConfiguration(data); err != nil {
|
|
return nil, errors.WithStackIf(err)
|
|
}
|
|
|
|
s.fs = filesystem.New(s.ID(), s.DiskSpace(), s.Config().Egg.FileDenylist)
|
|
// If this is a virtual filesystem we need to go ahead and mount the disk
|
|
// so that everything is accessible.
|
|
if s.fs.IsVirtual() && !m.skipVhdInitialization {
|
|
log.WithField("server", s.ID()).Info("mounting virtual disk for server")
|
|
if err := s.fs.MountDisk(ctx); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
// Right now we only support a Docker based environment, so I'm going to hard code
|
|
// this logic in. When we're ready to support other environment we'll need to make
|
|
// some modifications here, obviously.
|
|
settings := environment.Settings{
|
|
Mounts: s.Mounts(),
|
|
Allocations: s.cfg.Allocations,
|
|
Limits: s.cfg.Build,
|
|
}
|
|
|
|
envCfg := environment.NewConfiguration(settings, s.GetEnvironmentVariables())
|
|
meta := docker.Metadata{
|
|
Image: s.Config().Container.Image,
|
|
}
|
|
|
|
if env, err := docker.New(s.ID(), &meta, envCfg); err != nil {
|
|
return nil, err
|
|
} else {
|
|
s.Environment = env
|
|
s.StartEventListeners()
|
|
}
|
|
|
|
// If the server's data directory exists, force disk usage calculation.
|
|
if _, err := os.Stat(s.Filesystem().Path()); err == nil {
|
|
s.Filesystem().HasSpaceAvailable(true)
|
|
}
|
|
|
|
return s, nil
|
|
}
|
|
|
|
// initializeFromRemoteSource iterates over a given directory and loads all
|
|
// the servers listed before returning them to the calling function.
|
|
func (m *Manager) init(ctx context.Context) error {
|
|
log.Info("fetching list of servers from API")
|
|
servers, err := m.client.GetServers(ctx, config.Get().RemoteQuery.BootServersPerPage)
|
|
if err != nil {
|
|
if !remote.IsRequestError(err) {
|
|
return errors.WithStackIf(err)
|
|
}
|
|
return errors.WrapIf(err, "manager: failed to retrieve server configurations")
|
|
}
|
|
|
|
start := time.Now()
|
|
log.WithField("total_configs", len(servers)).Info("processing servers returned by the API")
|
|
|
|
pool := workerpool.New(runtime.NumCPU())
|
|
log.Debugf("using %d workerpools to instantiate server instances", runtime.NumCPU())
|
|
for _, data := range servers {
|
|
data := data
|
|
pool.Submit(func() {
|
|
// Parse the json.RawMessage into an expected struct value. We do this here so that a single broken
|
|
// server does not cause the entire boot process to hang, and allows us to show more useful error
|
|
// messaging in the output.
|
|
d := remote.ServerConfigurationResponse{
|
|
Settings: data.Settings,
|
|
}
|
|
log.WithField("server", data.Uuid).Info("creating new server object from API response")
|
|
if err := json.Unmarshal(data.ProcessConfiguration, &d.ProcessConfiguration); err != nil {
|
|
log.WithField("server", data.Uuid).WithField("error", err).Error("failed to parse server configuration from API response, skipping...")
|
|
return
|
|
}
|
|
s, err := m.InitServer(ctx, d)
|
|
if err != nil {
|
|
log.WithField("server", data.Uuid).WithField("error", err).Error("failed to load server, skipping...")
|
|
return
|
|
}
|
|
m.Add(s)
|
|
})
|
|
}
|
|
|
|
// Wait until we've processed all the configuration files in the directory
|
|
// before continuing.
|
|
pool.StopWait()
|
|
|
|
diff := time.Now().Sub(start)
|
|
log.WithField("duration", fmt.Sprintf("%s", diff)).Info("finished processing server configurations")
|
|
|
|
return nil
|
|
}
|