package router import ( "bufio" "context" "github.com/pterodactyl/wings/internal/models" "io" "mime/multipart" "net/http" "net/url" "os" "path" "path/filepath" "strconv" "strings" "github.com/pterodactyl/wings/config" "emperror.dev/errors" "github.com/apex/log" "github.com/gin-gonic/gin" "golang.org/x/sync/errgroup" "github.com/pterodactyl/wings/router/downloader" "github.com/pterodactyl/wings/router/middleware" "github.com/pterodactyl/wings/router/tokens" "github.com/pterodactyl/wings/server" "github.com/pterodactyl/wings/server/filesystem" ) // getServerFileContents returns the contents of a file on the server. func getServerFileContents(c *gin.Context) { s := middleware.ExtractServer(c) p := "/" + strings.TrimLeft(c.Query("file"), "/") f, st, err := s.Filesystem().File(p) if err != nil { middleware.CaptureAndAbort(c, err) return } defer f.Close() // Don't allow a named pipe to be opened. // // @see https://github.com/pterodactyl/panel/issues/4059 if st.Mode()&os.ModeNamedPipe != 0 { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Cannot open files of this type.", }) return } c.Header("X-Mime-Type", st.Mimetype) c.Header("Content-Length", strconv.Itoa(int(st.Size()))) // If a download parameter is included in the URL go ahead and attach the necessary headers // so that the file can be downloaded. if c.Query("download") != "" { c.Header("Content-Disposition", "attachment; filename="+strconv.Quote(st.Name())) c.Header("Content-Type", "application/octet-stream") } defer c.Writer.Flush() // If you don't do a limited reader here you will trigger a panic on write when // a different server process writes content to the file after you've already // determined the file size. This could lead to some weird content output but // it would technically be accurate based on the content at the time of the request. // // "http: wrote more than the declared Content-Length" // // @see https://github.com/pterodactyl/panel/issues/3131 r := io.LimitReader(f, st.Size()) if _, err = bufio.NewReader(r).WriteTo(c.Writer); err != nil { // Pretty sure this will unleash chaos on the response, but its a risk we can // take since a panic will at least be recovered and this should be incredibly // rare? middleware.CaptureAndAbort(c, err) return } } // Returns the contents of a directory for a server. func getServerListDirectory(c *gin.Context) { s := ExtractServer(c) dir := c.Query("directory") if stats, err := s.Filesystem().ListDirectory(dir); err != nil { WithError(c, err) } else { c.JSON(http.StatusOK, stats) } } type renameFile struct { To string `json:"to"` From string `json:"from"` } // Renames (or moves) files for a server. func putServerRenameFiles(c *gin.Context) { s := ExtractServer(c) var data struct { Root string `json:"root"` Files []renameFile `json:"files"` } // BindJSON sends 400 if the request fails, all we need to do is return if err := c.BindJSON(&data); err != nil { return } if len(data.Files) == 0 { c.AbortWithStatusJSON(http.StatusUnprocessableEntity, gin.H{ "error": "No files to move or rename were provided.", }) return } g, ctx := errgroup.WithContext(c.Request.Context()) // Loop over the array of files passed in and perform the move or rename action against each. for _, p := range data.Files { pf := path.Join(data.Root, p.From) pt := path.Join(data.Root, p.To) g.Go(func() error { select { case <-ctx.Done(): return ctx.Err() default: fs := s.Filesystem() // Ignore renames on a file that is on the denylist (both as the rename from or // the rename to value). if err := fs.IsIgnored(pf, pt); err != nil { return err } if err := fs.Rename(pf, pt); err != nil { // Return nil if the error is an is not exists. // NOTE: os.IsNotExist() does not work if the error is wrapped. if errors.Is(err, os.ErrNotExist) { s.Log().WithField("error", err). WithField("from_path", pf). WithField("to_path", pt). Warn("failed to rename: source or target does not exist") return nil } return err } return nil } }) } if err := g.Wait(); err != nil { if errors.Is(err, os.ErrExist) { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Cannot move or rename file, destination already exists.", }) return } NewServerError(err, s).AbortFilesystemError(c) return } c.Status(http.StatusNoContent) } // Copies a server file. func postServerCopyFile(c *gin.Context) { s := ExtractServer(c) var data struct { Location string `json:"location"` } // BindJSON sends 400 if the request fails, all we need to do is return if err := c.BindJSON(&data); err != nil { return } if err := s.Filesystem().IsIgnored(data.Location); err != nil { NewServerError(err, s).Abort(c) return } if err := s.Filesystem().Copy(data.Location); err != nil { NewServerError(err, s).AbortFilesystemError(c) return } c.Status(http.StatusNoContent) } // Deletes files from a server. func postServerDeleteFiles(c *gin.Context) { s := ExtractServer(c) var data struct { Root string `json:"root"` Files []string `json:"files"` } if err := c.BindJSON(&data); err != nil { return } if len(data.Files) == 0 { c.AbortWithStatusJSON(http.StatusUnprocessableEntity, gin.H{ "error": "No files were specified for deletion.", }) return } g, ctx := errgroup.WithContext(context.Background()) // Loop over the array of files passed in and delete them. If any of the file deletions // fail just abort the process entirely. for _, p := range data.Files { pi := path.Join(data.Root, p) g.Go(func() error { select { case <-ctx.Done(): return ctx.Err() default: return s.Filesystem().Delete(pi) } }) } if err := g.Wait(); err != nil { NewServerError(err, s).Abort(c) return } c.Status(http.StatusNoContent) } // Writes the contents of the request to a file on a server. func postServerWriteFile(c *gin.Context) { s := ExtractServer(c) f := c.Query("file") f = "/" + strings.TrimLeft(f, "/") if err := s.Filesystem().IsIgnored(f); err != nil { NewServerError(err, s).Abort(c) return } if err := s.Filesystem().Writefile(f, c.Request.Body); err != nil { if filesystem.IsErrorCode(err, filesystem.ErrCodeIsDirectory) { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Cannot write file, name conflicts with an existing directory by the same name.", }) return } NewServerError(err, s).AbortFilesystemError(c) return } c.Status(http.StatusNoContent) } // Returns all of the currently in-progress file downloads and their current download // progress. The progress is also pushed out via a websocket event allowing you to just // call this once to get current downloads, and then listen to targeted websocket events // with the current progress for everything. func getServerPullingFiles(c *gin.Context) { s := ExtractServer(c) c.JSON(http.StatusOK, gin.H{ "downloads": downloader.ByServer(s.ID()), }) } // Writes the contents of the remote URL to a file on a server. func postServerPullRemoteFile(c *gin.Context) { s := ExtractServer(c) var data struct { // Deprecated Directory string `binding:"required_without=RootPath,omitempty" json:"directory"` RootPath string `binding:"required_without=Directory,omitempty" json:"root"` URL string `binding:"required" json:"url"` FileName string `json:"file_name"` UseHeader bool `json:"use_header"` Foreground bool `json:"foreground"` } if err := c.BindJSON(&data); err != nil { return } // Handle the deprecated Directory field in the struct until it is removed. if data.Directory != "" && data.RootPath == "" { data.RootPath = data.Directory } u, err := url.Parse(data.URL) if err != nil { if e, ok := err.(*url.Error); ok { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "An error occurred while parsing that URL: " + e.Err.Error(), }) return } WithError(c, err) return } if err := s.Filesystem().HasSpaceErr(true); err != nil { WithError(c, err) return } // Do not allow more than three simultaneous remote file downloads at one time. if len(downloader.ByServer(s.ID())) >= 3 { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "This server has reached its limit of 3 simultaneous remote file downloads at once. Please wait for one to complete before trying again.", }) return } dl := downloader.New(s, downloader.DownloadRequest{ Directory: data.RootPath, URL: u, FileName: data.FileName, UseHeader: data.UseHeader, }) download := func() error { s.Log().WithField("download_id", dl.Identifier).WithField("url", u.String()).Info("starting pull of remote file to disk") if err := dl.Execute(); err != nil { s.Log().WithField("download_id", dl.Identifier).WithField("error", err).Error("failed to pull remote file") return err } else { s.Log().WithField("download_id", dl.Identifier).Info("completed pull of remote file") } return nil } if !data.Foreground { go func() { _ = download() }() c.JSON(http.StatusAccepted, gin.H{ "identifier": dl.Identifier, }) return } if err := download(); err != nil { NewServerError(err, s).Abort(c) return } st, err := s.Filesystem().Stat(dl.Path()) if err != nil { NewServerError(err, s).AbortFilesystemError(c) return } c.JSON(http.StatusOK, &st) } // Stops a remote file download if it exists and belongs to this server. func deleteServerPullRemoteFile(c *gin.Context) { s := ExtractServer(c) if dl := downloader.ByID(c.Param("download")); dl != nil && dl.BelongsTo(s) { dl.Cancel() } c.Status(http.StatusNoContent) } // Create a directory on a server. func postServerCreateDirectory(c *gin.Context) { s := ExtractServer(c) var data struct { Name string `json:"name"` Path string `json:"path"` } // BindJSON sends 400 if the request fails, all we need to do is return if err := c.BindJSON(&data); err != nil { return } if err := s.Filesystem().CreateDirectory(data.Name, data.Path); err != nil { if err.Error() == "not a directory" { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Part of the path being created is not a directory (ENOTDIR).", }) return } NewServerError(err, s).Abort(c) return } c.Status(http.StatusNoContent) } func postServerCompressFiles(c *gin.Context) { s := ExtractServer(c) var data struct { RootPath string `json:"root"` Files []string `json:"files"` } if err := c.BindJSON(&data); err != nil { return } if len(data.Files) == 0 { c.AbortWithStatusJSON(http.StatusUnprocessableEntity, gin.H{ "error": "No files were passed through to be compressed.", }) return } if !s.Filesystem().HasSpaceAvailable(true) { c.AbortWithStatusJSON(http.StatusConflict, gin.H{ "error": "This server does not have enough available disk space to generate a compressed archive.", }) return } f, err := s.Filesystem().CompressFiles(data.RootPath, data.Files) if err != nil { NewServerError(err, s).AbortFilesystemError(c) return } c.JSON(http.StatusOK, &filesystem.Stat{ FileInfo: f, Mimetype: "application/tar+gzip", }) } // postServerDecompressFiles receives the HTTP request and starts the process // of unpacking an archive that exists on the server into the provided RootPath // for the server. func postServerDecompressFiles(c *gin.Context) { var data struct { RootPath string `json:"root"` File string `json:"file"` } if err := c.BindJSON(&data); err != nil { return } s := middleware.ExtractServer(c) lg := middleware.ExtractLogger(c).WithFields(log.Fields{"root_path": data.RootPath, "file": data.File}) lg.Debug("checking if space is available for file decompression") err := s.Filesystem().SpaceAvailableForDecompression(data.RootPath, data.File) if err != nil { if filesystem.IsErrorCode(err, filesystem.ErrCodeUnknownArchive) { lg.WithField("error", err).Warn("failed to decompress file: unknown archive format") c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{"error": "The archive provided is in a format Wings does not understand."}) return } middleware.CaptureAndAbort(c, err) return } lg.Info("starting file decompression") if err := s.Filesystem().DecompressFile(data.RootPath, data.File); err != nil { // If the file is busy for some reason just return a nicer error to the user since there is not // much we specifically can do. They'll need to stop the running server process in order to overwrite // a file like this. if strings.Contains(err.Error(), "text file busy") { lg.WithField("error", errors.WithStackIf(err)).Warn("failed to decompress file: text file busy") c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "One or more files this archive is attempting to overwrite are currently in use by another process. Please try again.", }) return } middleware.CaptureAndAbort(c, err) return } c.Status(http.StatusNoContent) } type chmodFile struct { File string `json:"file"` Mode string `json:"mode"` } var errInvalidFileMode = errors.New("invalid file mode") func postServerChmodFile(c *gin.Context) { s := ExtractServer(c) var data struct { Root string `json:"root"` Files []chmodFile `json:"files"` } if err := c.BindJSON(&data); err != nil { log.Debug(err.Error()) return } if len(data.Files) == 0 { c.AbortWithStatusJSON(http.StatusUnprocessableEntity, gin.H{ "error": "No files to chmod were provided.", }) return } g, ctx := errgroup.WithContext(context.Background()) // Loop over the array of files passed in and perform the move or rename action against each. for _, p := range data.Files { g.Go(func() error { select { case <-ctx.Done(): return ctx.Err() default: mode, err := strconv.ParseUint(p.Mode, 8, 32) if err != nil { return errInvalidFileMode } if err := s.Filesystem().Chmod(path.Join(data.Root, p.File), os.FileMode(mode)); err != nil { // Return nil if the error is an is not exists. // NOTE: os.IsNotExist() does not work if the error is wrapped. if errors.Is(err, os.ErrNotExist) { return nil } return err } return nil } }) } if err := g.Wait(); err != nil { if errors.Is(err, errInvalidFileMode) { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Invalid file mode.", }) return } NewServerError(err, s).AbortFilesystemError(c) return } c.Status(http.StatusNoContent) } func postServerUploadFiles(c *gin.Context) { manager := middleware.ExtractManager(c) token := tokens.UploadPayload{} if err := tokens.ParseToken([]byte(c.Query("token")), &token); err != nil { NewTrackedError(err).Abort(c) return } s, ok := manager.Get(token.ServerUuid) if !ok || !token.IsUniqueRequest() { c.AbortWithStatusJSON(http.StatusNotFound, gin.H{ "error": "The requested resource was not found on this server.", }) return } form, err := c.MultipartForm() if err != nil { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "Failed to get multipart form data from request.", }) return } headers, ok := form.File["files"] if !ok { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "No files were found on the request body.", }) return } directory := c.Query("directory") maxFileSize := config.Get().Api.UploadLimit maxFileSizeBytes := maxFileSize * 1024 * 1024 var totalSize int64 for _, header := range headers { if header.Size > maxFileSizeBytes { c.AbortWithStatusJSON(http.StatusBadRequest, gin.H{ "error": "File " + header.Filename + " is larger than the maximum file upload size of " + strconv.FormatInt(maxFileSize, 10) + " MB.", }) return } totalSize += header.Size } for _, header := range headers { p, err := s.Filesystem().SafePath(filepath.Join(directory, header.Filename)) if err != nil { NewServerError(err, s).Abort(c) return } // We run this in a different method so I can use defer without any of // the consequences caused by calling it in a loop. if err := handleFileUpload(p, s, header); err != nil { NewServerError(err, s).Abort(c) return } else { s.SaveActivity(s.NewRequestActivity(token.UserUuid, c.ClientIP()), server.ActivityFileUploaded, models.ActivityMeta{ "file": header.Filename, "directory": filepath.Clean(directory), }) } } } func handleFileUpload(p string, s *server.Server, header *multipart.FileHeader) error { file, err := header.Open() if err != nil { return err } defer file.Close() if err := s.Filesystem().IsIgnored(p); err != nil { return err } if err := s.Filesystem().Writefile(p, file); err != nil { return err } return nil }