mirror of
https://github.com/gohugoio/hugo.git
synced 2024-11-29 11:22:08 -05:00
dea71670c0
Before this commit, you would have to use page bundles to do image processing etc. in Hugo. This commit adds * A new `/assets` top-level project or theme dir (configurable via `assetDir`) * A new template func, `resources.Get` which can be used to "get a resource" that can be further processed. This means that you can now do this in your templates (or shortcodes): ```bash {{ $sunset := (resources.Get "images/sunset.jpg").Fill "300x200" }} ``` This also adds a new `extended` build tag that enables powerful SCSS/SASS support with source maps. To compile this from source, you will also need a C compiler installed: ``` HUGO_BUILD_TAGS=extended mage install ``` Note that you can use output of the SCSS processing later in a non-SCSSS-enabled Hugo. The `SCSS` processor is a _Resource transformation step_ and it can be chained with the many others in a pipeline: ```bash {{ $css := resources.Get "styles.scss" | resources.ToCSS | resources.PostCSS | resources.Minify | resources.Fingerprint }} <link rel="stylesheet" href="{{ $styles.RelPermalink }}" integrity="{{ $styles.Data.Digest }}" media="screen"> ``` The transformation funcs above have aliases, so it can be shortened to: ```bash {{ $css := resources.Get "styles.scss" | toCSS | postCSS | minify | fingerprint }} <link rel="stylesheet" href="{{ $styles.RelPermalink }}" integrity="{{ $styles.Data.Digest }}" media="screen"> ``` A quick tip would be to avoid the fingerprinting part, and possibly also the not-superfast `postCSS` when you're doing development, as it allows Hugo to be smarter about the rebuilding. Documentation will follow, but have a look at the demo repo in https://github.com/bep/hugo-sass-test New functions to create `Resource` objects: * `resources.Get` (see above) * `resources.FromString`: Create a Resource from a string. New `Resource` transformation funcs: * `resources.ToCSS`: Compile `SCSS` or `SASS` into `CSS`. * `resources.PostCSS`: Process your CSS with PostCSS. Config file support (project or theme or passed as an option). * `resources.Minify`: Currently supports `css`, `js`, `json`, `html`, `svg`, `xml`. * `resources.Fingerprint`: Creates a fingerprinted version of the given Resource with Subresource Integrity.. * `resources.Concat`: Concatenates a list of Resource objects. Think of this as a poor man's bundler. * `resources.ExecuteAsTemplate`: Parses and executes the given Resource and data context (e.g. .Site) as a Go template. Fixes #4381 Fixes #4903 Fixes #4858
225 lines
6 KiB
Go
225 lines
6 KiB
Go
// Copyright 2016 The Hugo Authors. All rights reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package hugolib
|
|
|
|
import (
|
|
"path"
|
|
"path/filepath"
|
|
"strings"
|
|
|
|
"github.com/gohugoio/hugo/cache"
|
|
"github.com/gohugoio/hugo/helpers"
|
|
)
|
|
|
|
// PageCollections contains the page collections for a site.
|
|
type PageCollections struct {
|
|
// Includes only pages of all types, and only pages in the current language.
|
|
Pages Pages
|
|
|
|
// Includes all pages in all languages, including the current one.
|
|
// Includes pages of all types.
|
|
AllPages Pages
|
|
|
|
// A convenience cache for the traditional index types, taxonomies, home page etc.
|
|
// This is for the current language only.
|
|
indexPages Pages
|
|
|
|
// A convenience cache for the regular pages.
|
|
// This is for the current language only.
|
|
RegularPages Pages
|
|
|
|
// A convenience cache for the all the regular pages.
|
|
AllRegularPages Pages
|
|
|
|
// Includes absolute all pages (of all types), including drafts etc.
|
|
rawAllPages Pages
|
|
|
|
// Includes headless bundles, i.e. bundles that produce no output for its content page.
|
|
headlessPages Pages
|
|
|
|
pageCache *cache.PartitionedLazyCache
|
|
}
|
|
|
|
func (c *PageCollections) refreshPageCaches() {
|
|
c.indexPages = c.findPagesByKindNotIn(KindPage, c.Pages)
|
|
c.RegularPages = c.findPagesByKindIn(KindPage, c.Pages)
|
|
c.AllRegularPages = c.findPagesByKindIn(KindPage, c.AllPages)
|
|
|
|
var s *Site
|
|
|
|
if len(c.Pages) > 0 {
|
|
s = c.Pages[0].s
|
|
}
|
|
|
|
cacheLoader := func(kind string) func() (map[string]interface{}, error) {
|
|
return func() (map[string]interface{}, error) {
|
|
cache := make(map[string]interface{})
|
|
switch kind {
|
|
case KindPage:
|
|
// Note that we deliberately use the pages from all sites
|
|
// in this cache, as we intend to use this in the ref and relref
|
|
// shortcodes. If the user says "sect/doc1.en.md", he/she knows
|
|
// what he/she is looking for.
|
|
for _, pageCollection := range []Pages{c.AllRegularPages, c.headlessPages} {
|
|
for _, p := range pageCollection {
|
|
cache[filepath.ToSlash(p.Source.Path())] = p
|
|
|
|
if s != nil && p.s == s {
|
|
// Ref/Relref supports this potentially ambiguous lookup.
|
|
cache[p.Source.LogicalName()] = p
|
|
|
|
translasionBaseName := p.Source.TranslationBaseName()
|
|
dir := filepath.ToSlash(strings.TrimSuffix(p.Dir(), helpers.FilePathSeparator))
|
|
|
|
if translasionBaseName == "index" {
|
|
_, name := path.Split(dir)
|
|
cache[name] = p
|
|
cache[dir] = p
|
|
} else {
|
|
// Again, ambigous
|
|
cache[translasionBaseName] = p
|
|
}
|
|
|
|
// We need a way to get to the current language version.
|
|
pathWithNoExtensions := path.Join(dir, translasionBaseName)
|
|
cache[pathWithNoExtensions] = p
|
|
}
|
|
}
|
|
|
|
}
|
|
default:
|
|
for _, p := range c.indexPages {
|
|
key := path.Join(p.sections...)
|
|
cache[key] = p
|
|
}
|
|
}
|
|
|
|
return cache, nil
|
|
}
|
|
}
|
|
|
|
partitions := make([]cache.Partition, len(allKindsInPages))
|
|
|
|
for i, kind := range allKindsInPages {
|
|
partitions[i] = cache.Partition{Key: kind, Load: cacheLoader(kind)}
|
|
}
|
|
|
|
c.pageCache = cache.NewPartitionedLazyCache(partitions...)
|
|
}
|
|
|
|
func newPageCollections() *PageCollections {
|
|
return &PageCollections{}
|
|
}
|
|
|
|
func newPageCollectionsFromPages(pages Pages) *PageCollections {
|
|
return &PageCollections{rawAllPages: pages}
|
|
}
|
|
|
|
func (c *PageCollections) getPage(typ string, sections ...string) *Page {
|
|
var key string
|
|
if len(sections) == 1 {
|
|
key = filepath.ToSlash(sections[0])
|
|
} else {
|
|
key = path.Join(sections...)
|
|
}
|
|
|
|
p, _ := c.pageCache.Get(typ, key)
|
|
if p == nil {
|
|
return nil
|
|
}
|
|
return p.(*Page)
|
|
|
|
}
|
|
|
|
func (*PageCollections) findPagesByKindIn(kind string, inPages Pages) Pages {
|
|
var pages Pages
|
|
for _, p := range inPages {
|
|
if p.Kind == kind {
|
|
pages = append(pages, p)
|
|
}
|
|
}
|
|
return pages
|
|
}
|
|
|
|
func (*PageCollections) findFirstPageByKindIn(kind string, inPages Pages) *Page {
|
|
for _, p := range inPages {
|
|
if p.Kind == kind {
|
|
return p
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (*PageCollections) findPagesByKindNotIn(kind string, inPages Pages) Pages {
|
|
var pages Pages
|
|
for _, p := range inPages {
|
|
if p.Kind != kind {
|
|
pages = append(pages, p)
|
|
}
|
|
}
|
|
return pages
|
|
}
|
|
|
|
func (c *PageCollections) findPagesByKind(kind string) Pages {
|
|
return c.findPagesByKindIn(kind, c.Pages)
|
|
}
|
|
|
|
func (c *PageCollections) addPage(page *Page) {
|
|
c.rawAllPages = append(c.rawAllPages, page)
|
|
}
|
|
|
|
func (c *PageCollections) removePageFilename(filename string) {
|
|
if i := c.rawAllPages.findPagePosByFilename(filename); i >= 0 {
|
|
c.clearResourceCacheForPage(c.rawAllPages[i])
|
|
c.rawAllPages = append(c.rawAllPages[:i], c.rawAllPages[i+1:]...)
|
|
}
|
|
|
|
}
|
|
|
|
func (c *PageCollections) removePage(page *Page) {
|
|
if i := c.rawAllPages.findPagePos(page); i >= 0 {
|
|
c.clearResourceCacheForPage(c.rawAllPages[i])
|
|
c.rawAllPages = append(c.rawAllPages[:i], c.rawAllPages[i+1:]...)
|
|
}
|
|
|
|
}
|
|
|
|
func (c *PageCollections) findPagesByShortcode(shortcode string) Pages {
|
|
var pages Pages
|
|
|
|
for _, p := range c.rawAllPages {
|
|
if p.shortcodeState != nil {
|
|
if _, ok := p.shortcodeState.nameSet[shortcode]; ok {
|
|
pages = append(pages, p)
|
|
}
|
|
}
|
|
}
|
|
return pages
|
|
}
|
|
|
|
func (c *PageCollections) replacePage(page *Page) {
|
|
// will find existing page that matches filepath and remove it
|
|
c.removePage(page)
|
|
c.addPage(page)
|
|
}
|
|
|
|
func (c *PageCollections) clearResourceCacheForPage(page *Page) {
|
|
if len(page.Resources) > 0 {
|
|
first := page.Resources[0]
|
|
dir := path.Dir(first.RelPermalink())
|
|
dir = strings.TrimPrefix(dir, page.LanguagePrefix())
|
|
// This is done to keep the memory usage in check when doing live reloads.
|
|
page.s.ResourceSpec.DeleteCacheByPrefix(dir)
|
|
}
|
|
}
|