mirror of
https://github.com/gohugoio/hugo.git
synced 2024-12-01 18:33:08 -05:00
29ccb36069
In `v0.103.0` we added support for `resources.PostProcess` for all file types, not just HTML. We had benchmarks that said we were fine in that department, but those did not consider the static file syncing. This fixes that by: * Making sure that the /static syncer always gets its own file system without any checks for the post process token. * For dynamic files (e.g. rendered HTML files) we add an additional check to make sure that we skip binary files (e.g. images) Fixes #10328
204 lines
4.5 KiB
Go
204 lines
4.5 KiB
Go
// Copyright 2019 The Hugo Authors. All rights reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package hugolib
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"path/filepath"
|
|
|
|
"github.com/gohugoio/hugo/config"
|
|
"github.com/gohugoio/hugo/source"
|
|
|
|
"github.com/gohugoio/hugo/hugofs/files"
|
|
"golang.org/x/sync/errgroup"
|
|
|
|
"github.com/gohugoio/hugo/common/herrors"
|
|
"github.com/gohugoio/hugo/hugofs"
|
|
)
|
|
|
|
func newPagesProcessor(h *HugoSites, sp *source.SourceSpec) *pagesProcessor {
|
|
procs := make(map[string]pagesCollectorProcessorProvider)
|
|
for _, s := range h.Sites {
|
|
procs[s.Lang()] = &sitePagesProcessor{
|
|
m: s.pageMap,
|
|
errorSender: s.h,
|
|
itemChan: make(chan interface{}, config.GetNumWorkerMultiplier()*2),
|
|
}
|
|
}
|
|
return &pagesProcessor{
|
|
procs: procs,
|
|
}
|
|
}
|
|
|
|
type pagesCollectorProcessorProvider interface {
|
|
Process(item any) error
|
|
Start(ctx context.Context) context.Context
|
|
Wait() error
|
|
}
|
|
|
|
type pagesProcessor struct {
|
|
// Per language/Site
|
|
procs map[string]pagesCollectorProcessorProvider
|
|
}
|
|
|
|
func (proc *pagesProcessor) Process(item any) error {
|
|
switch v := item.(type) {
|
|
// Page bundles mapped to their language.
|
|
case pageBundles:
|
|
for _, vv := range v {
|
|
proc.getProcFromFi(vv.header).Process(vv)
|
|
}
|
|
case hugofs.FileMetaInfo:
|
|
proc.getProcFromFi(v).Process(v)
|
|
default:
|
|
panic(fmt.Sprintf("unrecognized item type in Process: %T", item))
|
|
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (proc *pagesProcessor) Start(ctx context.Context) context.Context {
|
|
for _, p := range proc.procs {
|
|
ctx = p.Start(ctx)
|
|
}
|
|
return ctx
|
|
}
|
|
|
|
func (proc *pagesProcessor) Wait() error {
|
|
var err error
|
|
for _, p := range proc.procs {
|
|
if e := p.Wait(); e != nil {
|
|
err = e
|
|
}
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (proc *pagesProcessor) getProcFromFi(fi hugofs.FileMetaInfo) pagesCollectorProcessorProvider {
|
|
if p, found := proc.procs[fi.Meta().Lang]; found {
|
|
return p
|
|
}
|
|
return defaultPageProcessor
|
|
}
|
|
|
|
type nopPageProcessor int
|
|
|
|
func (nopPageProcessor) Process(item any) error {
|
|
return nil
|
|
}
|
|
|
|
func (nopPageProcessor) Start(ctx context.Context) context.Context {
|
|
return context.Background()
|
|
}
|
|
|
|
func (nopPageProcessor) Wait() error {
|
|
return nil
|
|
}
|
|
|
|
var defaultPageProcessor = new(nopPageProcessor)
|
|
|
|
type sitePagesProcessor struct {
|
|
m *pageMap
|
|
errorSender herrors.ErrorSender
|
|
|
|
ctx context.Context
|
|
itemChan chan any
|
|
itemGroup *errgroup.Group
|
|
}
|
|
|
|
func (p *sitePagesProcessor) Process(item any) error {
|
|
select {
|
|
case <-p.ctx.Done():
|
|
return nil
|
|
default:
|
|
p.itemChan <- item
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (p *sitePagesProcessor) Start(ctx context.Context) context.Context {
|
|
p.itemGroup, ctx = errgroup.WithContext(ctx)
|
|
p.ctx = ctx
|
|
p.itemGroup.Go(func() error {
|
|
for item := range p.itemChan {
|
|
if err := p.doProcess(item); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
})
|
|
return ctx
|
|
}
|
|
|
|
func (p *sitePagesProcessor) Wait() error {
|
|
close(p.itemChan)
|
|
return p.itemGroup.Wait()
|
|
}
|
|
|
|
func (p *sitePagesProcessor) copyFile(fim hugofs.FileMetaInfo) error {
|
|
meta := fim.Meta()
|
|
f, err := meta.Open()
|
|
if err != nil {
|
|
return fmt.Errorf("copyFile: failed to open: %w", err)
|
|
}
|
|
|
|
s := p.m.s
|
|
|
|
target := filepath.Join(s.PathSpec.GetTargetLanguageBasePath(), meta.Path)
|
|
|
|
defer f.Close()
|
|
|
|
fs := s.PublishFsStatic
|
|
|
|
return s.publish(&s.PathSpec.ProcessingStats.Files, target, f, fs)
|
|
}
|
|
|
|
func (p *sitePagesProcessor) doProcess(item any) error {
|
|
m := p.m
|
|
switch v := item.(type) {
|
|
case *fileinfoBundle:
|
|
if err := m.AddFilesBundle(v.header, v.resources...); err != nil {
|
|
return err
|
|
}
|
|
case hugofs.FileMetaInfo:
|
|
if p.shouldSkip(v) {
|
|
return nil
|
|
}
|
|
meta := v.Meta()
|
|
|
|
classifier := meta.Classifier
|
|
switch classifier {
|
|
case files.ContentClassContent:
|
|
if err := m.AddFilesBundle(v); err != nil {
|
|
return err
|
|
}
|
|
case files.ContentClassFile:
|
|
if err := p.copyFile(v); err != nil {
|
|
return err
|
|
}
|
|
default:
|
|
panic(fmt.Sprintf("invalid classifier: %q", classifier))
|
|
}
|
|
default:
|
|
panic(fmt.Sprintf("unrecognized item type in Process: %T", item))
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (p *sitePagesProcessor) shouldSkip(fim hugofs.FileMetaInfo) bool {
|
|
// TODO(ep) unify
|
|
return p.m.s.SourceSpec.DisabledLanguages[fim.Meta().Lang]
|
|
}
|