mirror of
https://github.com/gohugoio/hugo.git
synced 2024-11-29 23:22:11 -05:00
c6b599a06d
If a content file contains shortcode(s), we have logic in place to re-render it per output format. We also have logic in place that avoids making a copy of the content used for this process if we don't need it. This was before this commit limited to server mode and if the page should be output to multiple formats. But there is a third case: If a site (language) borrows and renders `.Content` from another language. This would, before this commit, behave oddly for content with shortcodes. Fixes #4986
2307 lines
54 KiB
Go
2307 lines
54 KiB
Go
// Copyright 2018 The Hugo Authors. All rights reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package hugolib
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"reflect"
|
|
"unicode"
|
|
|
|
"github.com/gohugoio/hugo/media"
|
|
|
|
"github.com/gohugoio/hugo/common/maps"
|
|
|
|
"github.com/gohugoio/hugo/langs"
|
|
|
|
"github.com/gohugoio/hugo/related"
|
|
|
|
"github.com/bep/gitmap"
|
|
|
|
"github.com/gohugoio/hugo/helpers"
|
|
"github.com/gohugoio/hugo/hugolib/pagemeta"
|
|
"github.com/gohugoio/hugo/resource"
|
|
|
|
"github.com/gohugoio/hugo/output"
|
|
"github.com/gohugoio/hugo/parser"
|
|
"github.com/mitchellh/mapstructure"
|
|
|
|
"html/template"
|
|
"io"
|
|
"path"
|
|
"path/filepath"
|
|
"regexp"
|
|
"runtime"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
"unicode/utf8"
|
|
|
|
bp "github.com/gohugoio/hugo/bufferpool"
|
|
"github.com/gohugoio/hugo/compare"
|
|
"github.com/gohugoio/hugo/source"
|
|
"github.com/spf13/cast"
|
|
)
|
|
|
|
var (
|
|
cjk = regexp.MustCompile(`\p{Han}|\p{Hangul}|\p{Hiragana}|\p{Katakana}`)
|
|
|
|
// This is all the kinds we can expect to find in .Site.Pages.
|
|
allKindsInPages = []string{KindPage, KindHome, KindSection, KindTaxonomy, KindTaxonomyTerm}
|
|
|
|
allKinds = append(allKindsInPages, []string{kindRSS, kindSitemap, kindRobotsTXT, kind404}...)
|
|
|
|
// Assert that it implements the Eqer interface.
|
|
_ compare.Eqer = (*Page)(nil)
|
|
_ compare.Eqer = (*PageOutput)(nil)
|
|
|
|
// Assert that it implements the interface needed for related searches.
|
|
_ related.Document = (*Page)(nil)
|
|
)
|
|
|
|
const (
|
|
KindPage = "page"
|
|
|
|
// The rest are node types; home page, sections etc.
|
|
|
|
KindHome = "home"
|
|
KindSection = "section"
|
|
KindTaxonomy = "taxonomy"
|
|
KindTaxonomyTerm = "taxonomyTerm"
|
|
|
|
// Temporary state.
|
|
kindUnknown = "unknown"
|
|
|
|
// The following are (currently) temporary nodes,
|
|
// i.e. nodes we create just to render in isolation.
|
|
kindRSS = "RSS"
|
|
kindSitemap = "sitemap"
|
|
kindRobotsTXT = "robotsTXT"
|
|
kind404 = "404"
|
|
|
|
pageResourceType = "page"
|
|
)
|
|
|
|
type Page struct {
|
|
*pageInit
|
|
*pageContentInit
|
|
|
|
// Kind is the discriminator that identifies the different page types
|
|
// in the different page collections. This can, as an example, be used
|
|
// to to filter regular pages, find sections etc.
|
|
// Kind will, for the pages available to the templates, be one of:
|
|
// page, home, section, taxonomy and taxonomyTerm.
|
|
// It is of string type to make it easy to reason about in
|
|
// the templates.
|
|
Kind string
|
|
|
|
// Since Hugo 0.18 we got rid of the Node type. So now all pages are ...
|
|
// pages (regular pages, home page, sections etc.).
|
|
// Sections etc. will have child pages. These were earlier placed in .Data.Pages,
|
|
// but can now be more intuitively also be fetched directly from .Pages.
|
|
// This collection will be nil for regular pages.
|
|
Pages Pages
|
|
|
|
// Since Hugo 0.32, a Page can have resources such as images and CSS associated
|
|
// with itself. The resource will typically be placed relative to the Page,
|
|
// but templates should use the links (Permalink and RelPermalink)
|
|
// provided by the Resource object.
|
|
Resources resource.Resources
|
|
|
|
// This is the raw front matter metadata that is going to be assigned to
|
|
// the Resources above.
|
|
resourcesMetadata []map[string]interface{}
|
|
|
|
// translations will contain references to this page in other language
|
|
// if available.
|
|
translations Pages
|
|
|
|
// A key that maps to translation(s) of this page. This value is fetched
|
|
// from the page front matter.
|
|
translationKey string
|
|
|
|
// Params contains configuration defined in the params section of page frontmatter.
|
|
params map[string]interface{}
|
|
|
|
// Content sections
|
|
contentv template.HTML
|
|
summary template.HTML
|
|
TableOfContents template.HTML
|
|
// Passed to the shortcodes
|
|
pageWithoutContent *PageWithoutContent
|
|
|
|
Aliases []string
|
|
|
|
Images []Image
|
|
Videos []Video
|
|
|
|
truncated bool
|
|
Draft bool
|
|
Status string
|
|
|
|
// PageMeta contains page stats such as word count etc.
|
|
PageMeta
|
|
|
|
// Markup contains the markup type for the content.
|
|
Markup string
|
|
|
|
extension string
|
|
contentType string
|
|
renderable bool
|
|
|
|
Layout string
|
|
|
|
// For npn-renderable pages (see IsRenderable), the content itself
|
|
// is used as template and the template name is stored here.
|
|
selfLayout string
|
|
|
|
linkTitle string
|
|
|
|
frontmatter []byte
|
|
|
|
// rawContent is the raw content read from the content file.
|
|
rawContent []byte
|
|
|
|
// workContent is a copy of rawContent that may be mutated during site build.
|
|
workContent []byte
|
|
|
|
// whether the content is in a CJK language.
|
|
isCJKLanguage bool
|
|
|
|
shortcodeState *shortcodeHandler
|
|
|
|
// the content stripped for HTML
|
|
plain string // TODO should be []byte
|
|
plainWords []string
|
|
|
|
// rendering configuration
|
|
renderingConfig *helpers.BlackFriday
|
|
|
|
// menus
|
|
pageMenus PageMenus
|
|
|
|
Source
|
|
|
|
Position `json:"-"`
|
|
|
|
GitInfo *gitmap.GitInfo
|
|
|
|
// This was added as part of getting the Nodes (taxonomies etc.) to work as
|
|
// Pages in Hugo 0.18.
|
|
// It is deliberately named similar to Section, but not exported (for now).
|
|
// We currently have only one level of section in Hugo, but the page can live
|
|
// any number of levels down the file path.
|
|
// To support taxonomies like /categories/hugo etc. we will need to keep track
|
|
// of that information in a general way.
|
|
// So, sections represents the path to the content, i.e. a content file or a
|
|
// virtual content file in the situations where a taxonomy or a section etc.
|
|
// isn't accomanied by one.
|
|
sections []string
|
|
|
|
// Will only be set for sections and regular pages.
|
|
parent *Page
|
|
|
|
// When we create paginator pages, we create a copy of the original,
|
|
// but keep track of it here.
|
|
origOnCopy *Page
|
|
|
|
// Will only be set for section pages and the home page.
|
|
subSections Pages
|
|
|
|
s *Site
|
|
|
|
// Pulled over from old Node. TODO(bep) reorg and group (embed)
|
|
|
|
Site *SiteInfo `json:"-"`
|
|
|
|
title string
|
|
Description string
|
|
Keywords []string
|
|
data map[string]interface{}
|
|
|
|
pagemeta.PageDates
|
|
|
|
Sitemap Sitemap
|
|
pagemeta.URLPath
|
|
frontMatterURL string
|
|
|
|
permalink string
|
|
relPermalink string
|
|
|
|
// relative target path without extension and any base path element
|
|
// from the baseURL or the language code.
|
|
// This is used to construct paths in the page resources.
|
|
relTargetPathBase string
|
|
// Is set to a forward slashed path if this is a Page resources living in a folder below its owner.
|
|
resourcePath string
|
|
|
|
// This is enabled if it is a leaf bundle (the "index.md" type) and it is marked as headless in front matter.
|
|
// Being headless means that
|
|
// 1. The page itself is not rendered to disk
|
|
// 2. It is not available in .Site.Pages etc.
|
|
// 3. But you can get it via .Site.GetPage
|
|
headless bool
|
|
|
|
layoutDescriptor output.LayoutDescriptor
|
|
|
|
scratch *maps.Scratch
|
|
|
|
// It would be tempting to use the language set on the Site, but in they way we do
|
|
// multi-site processing, these values may differ during the initial page processing.
|
|
language *langs.Language
|
|
|
|
lang string
|
|
|
|
// The output formats this page will be rendered to.
|
|
outputFormats output.Formats
|
|
|
|
// This is the PageOutput that represents the first item in outputFormats.
|
|
// Use with care, as there are potential for inifinite loops.
|
|
mainPageOutput *PageOutput
|
|
|
|
targetPathDescriptorPrototype *targetPathDescriptor
|
|
}
|
|
|
|
func stackTrace(length int) string {
|
|
trace := make([]byte, length)
|
|
runtime.Stack(trace, true)
|
|
return string(trace)
|
|
}
|
|
|
|
func (p *Page) Data() interface{} {
|
|
return p.data
|
|
}
|
|
|
|
func (p *Page) initContent() {
|
|
|
|
p.contentInit.Do(func() {
|
|
// This careful dance is here to protect against circular loops in shortcode/content
|
|
// constructs.
|
|
// TODO(bep) context vs the remote shortcodes
|
|
ctx, cancel := context.WithTimeout(context.Background(), p.s.Timeout)
|
|
defer cancel()
|
|
c := make(chan error, 1)
|
|
|
|
go func() {
|
|
var err error
|
|
p.contentInitMu.Lock()
|
|
defer p.contentInitMu.Unlock()
|
|
|
|
err = p.prepareForRender()
|
|
if err != nil {
|
|
p.s.Log.ERROR.Printf("Failed to prepare page %q for render: %s", p.Path(), err)
|
|
return
|
|
}
|
|
|
|
if len(p.summary) == 0 {
|
|
if err = p.setAutoSummary(); err != nil {
|
|
err = fmt.Errorf("Failed to set user auto summary for page %q: %s", p.pathOrTitle(), err)
|
|
}
|
|
}
|
|
c <- err
|
|
}()
|
|
|
|
select {
|
|
case <-ctx.Done():
|
|
p.s.Log.WARN.Printf("WARNING: Timed out creating content for page %q (.Content will be empty). This is most likely a circular shortcode content loop that should be fixed. If this is just a shortcode calling a slow remote service, try to set \"timeout=20000\" (or higher, value is in milliseconds) in config.toml.\n", p.pathOrTitle())
|
|
case err := <-c:
|
|
if err != nil {
|
|
p.s.Log.ERROR.Println(err)
|
|
}
|
|
}
|
|
})
|
|
|
|
}
|
|
|
|
// This is sent to the shortcodes for this page. Not doing that will create an infinite regress. So,
|
|
// shortcodes can access .Page.TableOfContents, but not .Page.Content etc.
|
|
func (p *Page) withoutContent() *PageWithoutContent {
|
|
p.pageInit.withoutContentInit.Do(func() {
|
|
p.pageWithoutContent = &PageWithoutContent{Page: p}
|
|
})
|
|
return p.pageWithoutContent
|
|
}
|
|
|
|
func (p *Page) Content() (interface{}, error) {
|
|
return p.content(), nil
|
|
}
|
|
|
|
func (p *Page) Truncated() bool {
|
|
p.initContent()
|
|
return p.truncated
|
|
}
|
|
|
|
func (p *Page) content() template.HTML {
|
|
p.initContent()
|
|
return p.contentv
|
|
}
|
|
|
|
func (p *Page) Summary() template.HTML {
|
|
p.initContent()
|
|
return p.summary
|
|
}
|
|
|
|
// Sites is a convenience method to get all the Hugo sites/languages configured.
|
|
func (p *Page) Sites() SiteInfos {
|
|
infos := make(SiteInfos, len(p.s.owner.Sites))
|
|
for i, site := range p.s.owner.Sites {
|
|
infos[i] = &site.Info
|
|
}
|
|
|
|
return infos
|
|
}
|
|
|
|
// SearchKeywords implements the related.Document interface needed for fast page searches.
|
|
func (p *Page) SearchKeywords(cfg related.IndexConfig) ([]related.Keyword, error) {
|
|
|
|
v, err := p.Param(cfg.Name)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return cfg.ToKeywords(v)
|
|
}
|
|
|
|
// PubDate is when this page was or will be published.
|
|
// NOTE: This is currently used for search only and is not meant to be used
|
|
// directly in templates. We need to consolidate the dates in this struct.
|
|
// TODO(bep) see https://github.com/gohugoio/hugo/issues/3854
|
|
func (p *Page) PubDate() time.Time {
|
|
if !p.PublishDate.IsZero() {
|
|
return p.PublishDate
|
|
}
|
|
return p.Date
|
|
}
|
|
|
|
func (*Page) ResourceType() string {
|
|
return pageResourceType
|
|
}
|
|
|
|
func (p *Page) RSSLink() template.URL {
|
|
f, found := p.outputFormats.GetByName(output.RSSFormat.Name)
|
|
if !found {
|
|
return ""
|
|
}
|
|
return template.URL(newOutputFormat(p, f).Permalink())
|
|
}
|
|
|
|
func (p *Page) createLayoutDescriptor() output.LayoutDescriptor {
|
|
var section string
|
|
|
|
switch p.Kind {
|
|
case KindSection:
|
|
// In Hugo 0.22 we introduce nested sections, but we still only
|
|
// use the first level to pick the correct template. This may change in
|
|
// the future.
|
|
section = p.sections[0]
|
|
case KindTaxonomy, KindTaxonomyTerm:
|
|
section = p.s.taxonomiesPluralSingular[p.sections[0]]
|
|
default:
|
|
}
|
|
|
|
return output.LayoutDescriptor{
|
|
Kind: p.Kind,
|
|
Type: p.Type(),
|
|
Lang: p.Lang(),
|
|
Layout: p.Layout,
|
|
Section: section,
|
|
}
|
|
}
|
|
|
|
// pageInit lazy initializes different parts of the page. It is extracted
|
|
// into its own type so we can easily create a copy of a given page.
|
|
type pageInit struct {
|
|
languageInit sync.Once
|
|
pageMenusInit sync.Once
|
|
pageMetaInit sync.Once
|
|
renderingConfigInit sync.Once
|
|
withoutContentInit sync.Once
|
|
}
|
|
|
|
type pageContentInit struct {
|
|
contentInitMu sync.Mutex
|
|
contentInit sync.Once
|
|
plainInit sync.Once
|
|
plainWordsInit sync.Once
|
|
}
|
|
|
|
func (p *Page) resetContent() {
|
|
p.pageContentInit = &pageContentInit{}
|
|
}
|
|
|
|
// IsNode returns whether this is an item of one of the list types in Hugo,
|
|
// i.e. not a regular content page.
|
|
func (p *Page) IsNode() bool {
|
|
return p.Kind != KindPage
|
|
}
|
|
|
|
// IsHome returns whether this is the home page.
|
|
func (p *Page) IsHome() bool {
|
|
return p.Kind == KindHome
|
|
}
|
|
|
|
// IsSection returns whether this is a section page.
|
|
func (p *Page) IsSection() bool {
|
|
return p.Kind == KindSection
|
|
}
|
|
|
|
// IsPage returns whether this is a regular content page.
|
|
func (p *Page) IsPage() bool {
|
|
return p.Kind == KindPage
|
|
}
|
|
|
|
// BundleType returns the bundle type: "leaf", "branch" or an empty string if it is none.
|
|
// See https://gohugo.io/content-management/page-bundles/
|
|
func (p *Page) BundleType() string {
|
|
if p.IsNode() {
|
|
return "branch"
|
|
}
|
|
|
|
var source interface{} = p.Source.File
|
|
if fi, ok := source.(*fileInfo); ok {
|
|
switch fi.bundleTp {
|
|
case bundleBranch:
|
|
return "branch"
|
|
case bundleLeaf:
|
|
return "leaf"
|
|
}
|
|
}
|
|
|
|
return ""
|
|
}
|
|
|
|
func (p *Page) MediaType() media.Type {
|
|
return media.OctetType
|
|
}
|
|
|
|
type Source struct {
|
|
Frontmatter []byte
|
|
Content []byte
|
|
source.File
|
|
}
|
|
type PageMeta struct {
|
|
wordCount int
|
|
fuzzyWordCount int
|
|
readingTime int
|
|
Weight int
|
|
}
|
|
|
|
type Position struct {
|
|
Prev *Page
|
|
Next *Page
|
|
PrevInSection *Page
|
|
NextInSection *Page
|
|
}
|
|
|
|
type Pages []*Page
|
|
|
|
func (ps Pages) String() string {
|
|
return fmt.Sprintf("Pages(%d)", len(ps))
|
|
}
|
|
|
|
func (ps Pages) findPagePosByFilename(filename string) int {
|
|
for i, x := range ps {
|
|
if x.Source.Filename() == filename {
|
|
return i
|
|
}
|
|
}
|
|
return -1
|
|
}
|
|
|
|
func (ps Pages) removeFirstIfFound(p *Page) Pages {
|
|
ii := -1
|
|
for i, pp := range ps {
|
|
if pp == p {
|
|
ii = i
|
|
break
|
|
}
|
|
}
|
|
|
|
if ii != -1 {
|
|
ps = append(ps[:ii], ps[ii+1:]...)
|
|
}
|
|
return ps
|
|
}
|
|
|
|
func (ps Pages) findPagePosByFilnamePrefix(prefix string) int {
|
|
if prefix == "" {
|
|
return -1
|
|
}
|
|
|
|
lenDiff := -1
|
|
currPos := -1
|
|
prefixLen := len(prefix)
|
|
|
|
// Find the closest match
|
|
for i, x := range ps {
|
|
if strings.HasPrefix(x.Source.Filename(), prefix) {
|
|
diff := len(x.Source.Filename()) - prefixLen
|
|
if lenDiff == -1 || diff < lenDiff {
|
|
lenDiff = diff
|
|
currPos = i
|
|
}
|
|
}
|
|
}
|
|
return currPos
|
|
}
|
|
|
|
// findPagePos Given a page, it will find the position in Pages
|
|
// will return -1 if not found
|
|
func (ps Pages) findPagePos(page *Page) int {
|
|
for i, x := range ps {
|
|
if x.Source.Filename() == page.Source.Filename() {
|
|
return i
|
|
}
|
|
}
|
|
return -1
|
|
}
|
|
|
|
func (p *Page) createWorkContentCopy() {
|
|
p.workContent = make([]byte, len(p.rawContent))
|
|
copy(p.workContent, p.rawContent)
|
|
}
|
|
|
|
func (p *Page) Plain() string {
|
|
p.initContent()
|
|
p.initPlain(true)
|
|
return p.plain
|
|
}
|
|
|
|
func (p *Page) initPlain(lock bool) {
|
|
p.plainInit.Do(func() {
|
|
if lock {
|
|
p.contentInitMu.Lock()
|
|
defer p.contentInitMu.Unlock()
|
|
}
|
|
p.plain = helpers.StripHTML(string(p.contentv))
|
|
})
|
|
}
|
|
|
|
func (p *Page) PlainWords() []string {
|
|
p.initContent()
|
|
p.initPlainWords(true)
|
|
return p.plainWords
|
|
}
|
|
|
|
func (p *Page) initPlainWords(lock bool) {
|
|
p.plainWordsInit.Do(func() {
|
|
if lock {
|
|
p.contentInitMu.Lock()
|
|
defer p.contentInitMu.Unlock()
|
|
}
|
|
p.plainWords = strings.Fields(p.plain)
|
|
})
|
|
}
|
|
|
|
// Param is a convenience method to do lookups in Page's and Site's Params map,
|
|
// in that order.
|
|
//
|
|
// This method is also implemented on Node and SiteInfo.
|
|
func (p *Page) Param(key interface{}) (interface{}, error) {
|
|
keyStr, err := cast.ToStringE(key)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
keyStr = strings.ToLower(keyStr)
|
|
result, _ := p.traverseDirect(keyStr)
|
|
if result != nil {
|
|
return result, nil
|
|
}
|
|
|
|
keySegments := strings.Split(keyStr, ".")
|
|
if len(keySegments) == 1 {
|
|
return nil, nil
|
|
}
|
|
|
|
return p.traverseNested(keySegments)
|
|
}
|
|
|
|
func (p *Page) traverseDirect(key string) (interface{}, error) {
|
|
keyStr := strings.ToLower(key)
|
|
if val, ok := p.params[keyStr]; ok {
|
|
return val, nil
|
|
}
|
|
|
|
return p.Site.Params[keyStr], nil
|
|
}
|
|
|
|
func (p *Page) traverseNested(keySegments []string) (interface{}, error) {
|
|
result := traverse(keySegments, p.params)
|
|
if result != nil {
|
|
return result, nil
|
|
}
|
|
|
|
result = traverse(keySegments, p.Site.Params)
|
|
if result != nil {
|
|
return result, nil
|
|
}
|
|
|
|
// Didn't find anything, but also no problems.
|
|
return nil, nil
|
|
}
|
|
|
|
func traverse(keys []string, m map[string]interface{}) interface{} {
|
|
// Shift first element off.
|
|
firstKey, rest := keys[0], keys[1:]
|
|
result := m[firstKey]
|
|
|
|
// No point in continuing here.
|
|
if result == nil {
|
|
return result
|
|
}
|
|
|
|
if len(rest) == 0 {
|
|
// That was the last key.
|
|
return result
|
|
}
|
|
|
|
// That was not the last key.
|
|
return traverse(rest, cast.ToStringMap(result))
|
|
}
|
|
|
|
func (p *Page) Author() Author {
|
|
authors := p.Authors()
|
|
|
|
for _, author := range authors {
|
|
return author
|
|
}
|
|
return Author{}
|
|
}
|
|
|
|
func (p *Page) Authors() AuthorList {
|
|
authorKeys, ok := p.params["authors"]
|
|
if !ok {
|
|
return AuthorList{}
|
|
}
|
|
authors := authorKeys.([]string)
|
|
if len(authors) < 1 || len(p.Site.Authors) < 1 {
|
|
return AuthorList{}
|
|
}
|
|
|
|
al := make(AuthorList)
|
|
for _, author := range authors {
|
|
a, ok := p.Site.Authors[author]
|
|
if ok {
|
|
al[author] = a
|
|
}
|
|
}
|
|
return al
|
|
}
|
|
|
|
func (p *Page) UniqueID() string {
|
|
return p.Source.UniqueID()
|
|
}
|
|
|
|
// for logging
|
|
func (p *Page) lineNumRawContentStart() int {
|
|
return bytes.Count(p.frontmatter, []byte("\n")) + 1
|
|
}
|
|
|
|
var (
|
|
internalSummaryDivider = []byte("HUGOMORE42")
|
|
)
|
|
|
|
// replaceDivider replaces the <!--more--> with an internal value and returns
|
|
// whether the contentis truncated or not.
|
|
// Note: The content slice will be modified if needed.
|
|
func replaceDivider(content, from, to []byte) ([]byte, bool) {
|
|
dividerIdx := bytes.Index(content, from)
|
|
if dividerIdx == -1 {
|
|
return content, false
|
|
}
|
|
|
|
afterSummary := content[dividerIdx+len(from):]
|
|
|
|
// If the raw content has nothing but whitespace after the summary
|
|
// marker then the page shouldn't be marked as truncated. This check
|
|
// is simplest against the raw content because different markup engines
|
|
// (rst and asciidoc in particular) add div and p elements after the
|
|
// summary marker.
|
|
truncated := bytes.IndexFunc(afterSummary, func(r rune) bool { return !unicode.IsSpace(r) }) != -1
|
|
|
|
content = append(content[:dividerIdx], append(to, afterSummary...)...)
|
|
|
|
return content, truncated
|
|
|
|
}
|
|
|
|
// We have to replace the <!--more--> with something that survives all the
|
|
// rendering engines.
|
|
func (p *Page) replaceDivider(content []byte) []byte {
|
|
summaryDivider := helpers.SummaryDivider
|
|
if p.Markup == "org" {
|
|
summaryDivider = []byte("# more")
|
|
}
|
|
|
|
replaced, truncated := replaceDivider(content, summaryDivider, internalSummaryDivider)
|
|
|
|
p.truncated = truncated
|
|
|
|
return replaced
|
|
}
|
|
|
|
// Returns the page as summary and main if a user defined split is provided.
|
|
func (p *Page) setUserDefinedSummaryIfProvided(rawContentCopy []byte) (*summaryContent, error) {
|
|
|
|
sc, err := splitUserDefinedSummaryAndContent(p.Markup, rawContentCopy)
|
|
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if sc == nil {
|
|
// No divider found
|
|
return nil, nil
|
|
}
|
|
|
|
p.summary = helpers.BytesToHTML(sc.summary)
|
|
|
|
return sc, nil
|
|
}
|
|
|
|
// Make this explicit so there is no doubt about what is what.
|
|
type summaryContent struct {
|
|
summary []byte
|
|
content []byte
|
|
}
|
|
|
|
func splitUserDefinedSummaryAndContent(markup string, c []byte) (sc *summaryContent, err error) {
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
err = fmt.Errorf("summary split failed: %s", r)
|
|
}
|
|
}()
|
|
|
|
c = bytes.TrimSpace(c)
|
|
startDivider := bytes.Index(c, internalSummaryDivider)
|
|
|
|
if startDivider == -1 {
|
|
return
|
|
}
|
|
|
|
endDivider := startDivider + len(internalSummaryDivider)
|
|
endSummary := startDivider
|
|
|
|
var (
|
|
startMarkup []byte
|
|
endMarkup []byte
|
|
addDiv bool
|
|
)
|
|
|
|
switch markup {
|
|
default:
|
|
startMarkup = []byte("<p>")
|
|
endMarkup = []byte("</p>")
|
|
case "asciidoc":
|
|
startMarkup = []byte("<div class=\"paragraph\">")
|
|
endMarkup = []byte("</div>")
|
|
case "rst":
|
|
startMarkup = []byte("<p>")
|
|
endMarkup = []byte("</p>")
|
|
addDiv = true
|
|
}
|
|
|
|
// Find the closest end/start markup string to the divider
|
|
fromStart := -1
|
|
fromIdx := bytes.LastIndex(c[:startDivider], startMarkup)
|
|
if fromIdx != -1 {
|
|
fromStart = startDivider - fromIdx - len(startMarkup)
|
|
}
|
|
fromEnd := bytes.Index(c[endDivider:], endMarkup)
|
|
|
|
if fromEnd != -1 && fromEnd <= fromStart {
|
|
endSummary = startDivider + fromEnd + len(endMarkup)
|
|
} else if fromStart != -1 && fromEnd != -1 {
|
|
endSummary = startDivider - fromStart - len(startMarkup)
|
|
}
|
|
|
|
withoutDivider := bytes.TrimSpace(append(c[:startDivider], c[endDivider:]...))
|
|
var (
|
|
summary []byte
|
|
)
|
|
|
|
if len(withoutDivider) > 0 {
|
|
summary = bytes.TrimSpace(withoutDivider[:endSummary])
|
|
}
|
|
|
|
if addDiv {
|
|
// For the rst
|
|
summary = append(append([]byte(nil), summary...), []byte("</div>")...)
|
|
}
|
|
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
sc = &summaryContent{
|
|
summary: summary,
|
|
content: withoutDivider,
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func (p *Page) setAutoSummary() error {
|
|
var summary string
|
|
var truncated bool
|
|
// This careful init dance could probably be refined, but it is purely for performance
|
|
// reasons. These "plain" methods are expensive if the plain content is never actually
|
|
// used.
|
|
p.initPlain(false)
|
|
if p.isCJKLanguage {
|
|
p.initPlainWords(false)
|
|
summary, truncated = p.s.ContentSpec.TruncateWordsByRune(p.plainWords)
|
|
} else {
|
|
summary, truncated = p.s.ContentSpec.TruncateWordsToWholeSentence(p.plain)
|
|
}
|
|
p.summary = template.HTML(summary)
|
|
p.truncated = truncated
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
func (p *Page) renderContent(content []byte) []byte {
|
|
return p.s.ContentSpec.RenderBytes(&helpers.RenderingContext{
|
|
Content: content, RenderTOC: true, PageFmt: p.Markup,
|
|
Cfg: p.Language(),
|
|
DocumentID: p.UniqueID(), DocumentName: p.Path(),
|
|
Config: p.getRenderingConfig()})
|
|
}
|
|
|
|
func (p *Page) getRenderingConfig() *helpers.BlackFriday {
|
|
p.renderingConfigInit.Do(func() {
|
|
bfParam := p.getParamToLower("blackfriday")
|
|
if bfParam == nil {
|
|
p.renderingConfig = p.s.ContentSpec.BlackFriday
|
|
return
|
|
}
|
|
// Create a copy so we can modify it.
|
|
bf := *p.s.ContentSpec.BlackFriday
|
|
p.renderingConfig = &bf
|
|
|
|
if p.Language() == nil {
|
|
panic(fmt.Sprintf("nil language for %s with source lang %s", p.BaseFileName(), p.lang))
|
|
}
|
|
|
|
pageParam := cast.ToStringMap(bfParam)
|
|
if err := mapstructure.Decode(pageParam, &p.renderingConfig); err != nil {
|
|
p.s.Log.FATAL.Printf("Failed to get rendering config for %s:\n%s", p.BaseFileName(), err.Error())
|
|
}
|
|
|
|
})
|
|
|
|
return p.renderingConfig
|
|
}
|
|
|
|
func (s *Site) newPage(filename string) *Page {
|
|
fi := newFileInfo(
|
|
s.SourceSpec,
|
|
s.absContentDir(),
|
|
filename,
|
|
nil,
|
|
bundleNot,
|
|
)
|
|
return s.newPageFromFile(fi)
|
|
}
|
|
|
|
func (s *Site) newPageFromFile(fi *fileInfo) *Page {
|
|
return &Page{
|
|
pageInit: &pageInit{},
|
|
pageContentInit: &pageContentInit{},
|
|
Kind: kindFromFileInfo(fi),
|
|
contentType: "",
|
|
Source: Source{File: fi},
|
|
Keywords: []string{}, Sitemap: Sitemap{Priority: -1},
|
|
params: make(map[string]interface{}),
|
|
translations: make(Pages, 0),
|
|
sections: sectionsFromFile(fi),
|
|
Site: &s.Info,
|
|
s: s,
|
|
}
|
|
}
|
|
|
|
func (p *Page) IsRenderable() bool {
|
|
return p.renderable
|
|
}
|
|
|
|
func (p *Page) Type() string {
|
|
if p.contentType != "" {
|
|
return p.contentType
|
|
}
|
|
|
|
if x := p.Section(); x != "" {
|
|
return x
|
|
}
|
|
|
|
return "page"
|
|
}
|
|
|
|
// Section returns the first path element below the content root. Note that
|
|
// since Hugo 0.22 we support nested sections, but this will always be the first
|
|
// element of any nested path.
|
|
func (p *Page) Section() string {
|
|
if p.Kind == KindSection || p.Kind == KindTaxonomy || p.Kind == KindTaxonomyTerm {
|
|
return p.sections[0]
|
|
}
|
|
return p.Source.Section()
|
|
}
|
|
|
|
func (s *Site) NewPageFrom(buf io.Reader, name string) (*Page, error) {
|
|
p, err := s.NewPage(name)
|
|
if err != nil {
|
|
return p, err
|
|
}
|
|
_, err = p.ReadFrom(buf)
|
|
|
|
return p, err
|
|
}
|
|
|
|
func (s *Site) NewPage(name string) (*Page, error) {
|
|
if len(name) == 0 {
|
|
return nil, errors.New("Zero length page name")
|
|
}
|
|
|
|
// Create new page
|
|
p := s.newPage(name)
|
|
p.s = s
|
|
p.Site = &s.Info
|
|
|
|
return p, nil
|
|
}
|
|
|
|
func (p *Page) ReadFrom(buf io.Reader) (int64, error) {
|
|
// Parse for metadata & body
|
|
if err := p.parse(buf); err != nil {
|
|
p.s.Log.ERROR.Printf("%s for %s", err, p.File.Path())
|
|
return 0, err
|
|
}
|
|
|
|
return int64(len(p.rawContent)), nil
|
|
}
|
|
|
|
func (p *Page) WordCount() int {
|
|
p.initContentPlainAndMeta()
|
|
return p.wordCount
|
|
}
|
|
|
|
func (p *Page) ReadingTime() int {
|
|
p.initContentPlainAndMeta()
|
|
return p.readingTime
|
|
}
|
|
|
|
func (p *Page) FuzzyWordCount() int {
|
|
p.initContentPlainAndMeta()
|
|
return p.fuzzyWordCount
|
|
}
|
|
|
|
func (p *Page) initContentPlainAndMeta() {
|
|
p.initContent()
|
|
p.initPlain(true)
|
|
p.initPlainWords(true)
|
|
p.initMeta()
|
|
}
|
|
|
|
func (p *Page) initContentAndMeta() {
|
|
p.initContent()
|
|
p.initMeta()
|
|
}
|
|
|
|
func (p *Page) initMeta() {
|
|
p.pageMetaInit.Do(func() {
|
|
if p.isCJKLanguage {
|
|
p.wordCount = 0
|
|
for _, word := range p.plainWords {
|
|
runeCount := utf8.RuneCountInString(word)
|
|
if len(word) == runeCount {
|
|
p.wordCount++
|
|
} else {
|
|
p.wordCount += runeCount
|
|
}
|
|
}
|
|
} else {
|
|
p.wordCount = helpers.TotalWords(p.plain)
|
|
}
|
|
|
|
// TODO(bep) is set in a test. Fix that.
|
|
if p.fuzzyWordCount == 0 {
|
|
p.fuzzyWordCount = (p.wordCount + 100) / 100 * 100
|
|
}
|
|
|
|
if p.isCJKLanguage {
|
|
p.readingTime = (p.wordCount + 500) / 501
|
|
} else {
|
|
p.readingTime = (p.wordCount + 212) / 213
|
|
}
|
|
})
|
|
}
|
|
|
|
// HasShortcode return whether the page has a shortcode with the given name.
|
|
// This method is mainly motivated with the Hugo Docs site's need for a list
|
|
// of pages with the `todo` shortcode in it.
|
|
func (p *Page) HasShortcode(name string) bool {
|
|
if p.shortcodeState == nil {
|
|
return false
|
|
}
|
|
|
|
return p.shortcodeState.nameSet[name]
|
|
}
|
|
|
|
// AllTranslations returns all translations, including the current Page.
|
|
func (p *Page) AllTranslations() Pages {
|
|
return p.translations
|
|
}
|
|
|
|
// IsTranslated returns whether this content file is translated to
|
|
// other language(s).
|
|
func (p *Page) IsTranslated() bool {
|
|
return len(p.translations) > 1
|
|
}
|
|
|
|
// Translations returns the translations excluding the current Page.
|
|
func (p *Page) Translations() Pages {
|
|
translations := make(Pages, 0)
|
|
for _, t := range p.translations {
|
|
if t.Lang() != p.Lang() {
|
|
translations = append(translations, t)
|
|
}
|
|
}
|
|
return translations
|
|
}
|
|
|
|
// TranslationKey returns the key used to map language translations of this page.
|
|
// It will use the translationKey set in front matter if set, or the content path and
|
|
// filename (excluding any language code and extension), e.g. "about/index".
|
|
// The Page Kind is always prepended.
|
|
func (p *Page) TranslationKey() string {
|
|
if p.translationKey != "" {
|
|
return p.Kind + "/" + p.translationKey
|
|
}
|
|
|
|
if p.IsNode() {
|
|
return path.Join(p.Kind, path.Join(p.sections...), p.TranslationBaseName())
|
|
}
|
|
|
|
return path.Join(p.Kind, filepath.ToSlash(p.Dir()), p.TranslationBaseName())
|
|
}
|
|
|
|
func (p *Page) LinkTitle() string {
|
|
if len(p.linkTitle) > 0 {
|
|
return p.linkTitle
|
|
}
|
|
return p.title
|
|
}
|
|
|
|
func (p *Page) shouldBuild() bool {
|
|
return shouldBuild(p.s.BuildFuture, p.s.BuildExpired,
|
|
p.s.BuildDrafts, p.Draft, p.PublishDate, p.ExpiryDate)
|
|
}
|
|
|
|
func shouldBuild(buildFuture bool, buildExpired bool, buildDrafts bool, Draft bool,
|
|
publishDate time.Time, expiryDate time.Time) bool {
|
|
if !(buildDrafts || !Draft) {
|
|
return false
|
|
}
|
|
if !buildFuture && !publishDate.IsZero() && publishDate.After(time.Now()) {
|
|
return false
|
|
}
|
|
if !buildExpired && !expiryDate.IsZero() && expiryDate.Before(time.Now()) {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
func (p *Page) IsDraft() bool {
|
|
return p.Draft
|
|
}
|
|
|
|
func (p *Page) IsFuture() bool {
|
|
if p.PublishDate.IsZero() {
|
|
return false
|
|
}
|
|
return p.PublishDate.After(time.Now())
|
|
}
|
|
|
|
func (p *Page) IsExpired() bool {
|
|
if p.ExpiryDate.IsZero() {
|
|
return false
|
|
}
|
|
return p.ExpiryDate.Before(time.Now())
|
|
}
|
|
|
|
func (p *Page) URL() string {
|
|
|
|
if p.IsPage() && p.URLPath.URL != "" {
|
|
// This is the url set in front matter
|
|
return p.URLPath.URL
|
|
}
|
|
// Fall back to the relative permalink.
|
|
u := p.RelPermalink()
|
|
return u
|
|
}
|
|
|
|
// Permalink returns the absolute URL to this Page.
|
|
func (p *Page) Permalink() string {
|
|
if p.headless {
|
|
return ""
|
|
}
|
|
return p.permalink
|
|
}
|
|
|
|
// RelPermalink gets a URL to the resource relative to the host.
|
|
func (p *Page) RelPermalink() string {
|
|
if p.headless {
|
|
return ""
|
|
}
|
|
return p.relPermalink
|
|
}
|
|
|
|
// See resource.Resource
|
|
// This value is used, by default, in Resources.ByPrefix etc.
|
|
func (p *Page) Name() string {
|
|
if p.resourcePath != "" {
|
|
return p.resourcePath
|
|
}
|
|
return p.title
|
|
}
|
|
|
|
func (p *Page) Title() string {
|
|
return p.title
|
|
}
|
|
|
|
func (p *Page) Params() map[string]interface{} {
|
|
return p.params
|
|
}
|
|
|
|
func (p *Page) subResourceTargetPathFactory(base string) string {
|
|
return path.Join(p.relTargetPathBase, base)
|
|
}
|
|
|
|
func (p *Page) initMainOutputFormat() error {
|
|
if p.mainPageOutput != nil {
|
|
return nil
|
|
}
|
|
|
|
outFormat := p.outputFormats[0]
|
|
pageOutput, err := newPageOutput(p, false, false, outFormat)
|
|
|
|
if err != nil {
|
|
return fmt.Errorf("Failed to create output page for type %q for page %q: %s", outFormat.Name, p.pathOrTitle(), err)
|
|
}
|
|
|
|
p.mainPageOutput = pageOutput
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
func (p *Page) setContentInit(start bool) error {
|
|
|
|
if start {
|
|
// This is a new language.
|
|
p.shortcodeState.clearDelta()
|
|
}
|
|
updated := true
|
|
if p.shortcodeState != nil {
|
|
updated = p.shortcodeState.updateDelta()
|
|
}
|
|
|
|
if updated {
|
|
p.resetContent()
|
|
}
|
|
|
|
for _, r := range p.Resources.ByType(pageResourceType) {
|
|
p.s.PathSpec.ProcessingStats.Incr(&p.s.PathSpec.ProcessingStats.Pages)
|
|
bp := r.(*Page)
|
|
if start {
|
|
bp.shortcodeState.clearDelta()
|
|
}
|
|
if bp.shortcodeState != nil {
|
|
updated = bp.shortcodeState.updateDelta()
|
|
}
|
|
if updated {
|
|
bp.resetContent()
|
|
}
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
func (p *Page) prepareForRender() error {
|
|
s := p.s
|
|
|
|
// If we got this far it means that this is either a new Page pointer
|
|
// or a template or similar has changed so wee need to do a rerendering
|
|
// of the shortcodes etc.
|
|
|
|
// If in watch mode or if we have multiple sites or output formats,
|
|
// we need to keep the original so we can
|
|
// potentially repeat this process on rebuild.
|
|
needsACopy := s.running() || len(s.owner.Sites) > 1 || len(p.outputFormats) > 1
|
|
var workContentCopy []byte
|
|
if needsACopy {
|
|
workContentCopy = make([]byte, len(p.workContent))
|
|
copy(workContentCopy, p.workContent)
|
|
} else {
|
|
// Just reuse the same slice.
|
|
workContentCopy = p.workContent
|
|
}
|
|
|
|
var err error
|
|
// Note: The shortcodes in a page cannot access the page content it lives in,
|
|
// hence the withoutContent().
|
|
if workContentCopy, err = handleShortcodes(p.withoutContent(), workContentCopy); err != nil {
|
|
s.Log.ERROR.Printf("Failed to handle shortcodes for page %s: %s", p.BaseFileName(), err)
|
|
}
|
|
|
|
if p.Markup != "html" {
|
|
|
|
// Now we know enough to create a summary of the page and count some words
|
|
summaryContent, err := p.setUserDefinedSummaryIfProvided(workContentCopy)
|
|
|
|
if err != nil {
|
|
s.Log.ERROR.Printf("Failed to set user defined summary for page %q: %s", p.Path(), err)
|
|
} else if summaryContent != nil {
|
|
workContentCopy = summaryContent.content
|
|
}
|
|
|
|
p.contentv = helpers.BytesToHTML(workContentCopy)
|
|
|
|
} else {
|
|
p.contentv = helpers.BytesToHTML(workContentCopy)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
var ErrHasDraftAndPublished = errors.New("both draft and published parameters were found in page's frontmatter")
|
|
|
|
func (p *Page) update(frontmatter map[string]interface{}) error {
|
|
if frontmatter == nil {
|
|
return errors.New("missing frontmatter data")
|
|
}
|
|
// Needed for case insensitive fetching of params values
|
|
maps.ToLower(frontmatter)
|
|
|
|
var mtime time.Time
|
|
if p.Source.FileInfo() != nil {
|
|
mtime = p.Source.FileInfo().ModTime()
|
|
}
|
|
|
|
var gitAuthorDate time.Time
|
|
if p.GitInfo != nil {
|
|
gitAuthorDate = p.GitInfo.AuthorDate
|
|
}
|
|
|
|
descriptor := &pagemeta.FrontMatterDescriptor{
|
|
Frontmatter: frontmatter,
|
|
Params: p.params,
|
|
Dates: &p.PageDates,
|
|
PageURLs: &p.URLPath,
|
|
BaseFilename: p.BaseFileName(),
|
|
ModTime: mtime,
|
|
GitAuthorDate: gitAuthorDate,
|
|
}
|
|
|
|
// Handle the date separately
|
|
// TODO(bep) we need to "do more" in this area so this can be split up and
|
|
// more easily tested without the Page, but the coupling is strong.
|
|
err := p.s.frontmatterHandler.HandleDates(descriptor)
|
|
if err != nil {
|
|
p.s.Log.ERROR.Printf("Failed to handle dates for page %q: %s", p.Path(), err)
|
|
}
|
|
|
|
var draft, published, isCJKLanguage *bool
|
|
for k, v := range frontmatter {
|
|
loki := strings.ToLower(k)
|
|
|
|
if loki == "published" { // Intentionally undocumented
|
|
vv, err := cast.ToBoolE(v)
|
|
if err == nil {
|
|
published = &vv
|
|
}
|
|
// published may also be a date
|
|
continue
|
|
}
|
|
|
|
if p.s.frontmatterHandler.IsDateKey(loki) {
|
|
continue
|
|
}
|
|
|
|
switch loki {
|
|
case "title":
|
|
p.title = cast.ToString(v)
|
|
p.params[loki] = p.title
|
|
case "linktitle":
|
|
p.linkTitle = cast.ToString(v)
|
|
p.params[loki] = p.linkTitle
|
|
case "description":
|
|
p.Description = cast.ToString(v)
|
|
p.params[loki] = p.Description
|
|
case "slug":
|
|
p.Slug = cast.ToString(v)
|
|
p.params[loki] = p.Slug
|
|
case "url":
|
|
if url := cast.ToString(v); strings.HasPrefix(url, "http://") || strings.HasPrefix(url, "https://") {
|
|
return fmt.Errorf("Only relative URLs are supported, %v provided", url)
|
|
}
|
|
p.URLPath.URL = cast.ToString(v)
|
|
p.frontMatterURL = p.URLPath.URL
|
|
p.params[loki] = p.URLPath.URL
|
|
case "type":
|
|
p.contentType = cast.ToString(v)
|
|
p.params[loki] = p.contentType
|
|
case "extension", "ext":
|
|
p.extension = cast.ToString(v)
|
|
p.params[loki] = p.extension
|
|
case "keywords":
|
|
p.Keywords = cast.ToStringSlice(v)
|
|
p.params[loki] = p.Keywords
|
|
case "headless":
|
|
// For now, only the leaf bundles ("index.md") can be headless (i.e. produce no output).
|
|
// We may expand on this in the future, but that gets more complex pretty fast.
|
|
if p.TranslationBaseName() == "index" {
|
|
p.headless = cast.ToBool(v)
|
|
}
|
|
p.params[loki] = p.headless
|
|
case "outputs":
|
|
o := cast.ToStringSlice(v)
|
|
if len(o) > 0 {
|
|
// Output formats are exlicitly set in front matter, use those.
|
|
outFormats, err := p.s.outputFormatsConfig.GetByNames(o...)
|
|
|
|
if err != nil {
|
|
p.s.Log.ERROR.Printf("Failed to resolve output formats: %s", err)
|
|
} else {
|
|
p.outputFormats = outFormats
|
|
p.params[loki] = outFormats
|
|
}
|
|
|
|
}
|
|
case "draft":
|
|
draft = new(bool)
|
|
*draft = cast.ToBool(v)
|
|
case "layout":
|
|
p.Layout = cast.ToString(v)
|
|
p.params[loki] = p.Layout
|
|
case "markup":
|
|
p.Markup = cast.ToString(v)
|
|
p.params[loki] = p.Markup
|
|
case "weight":
|
|
p.Weight = cast.ToInt(v)
|
|
p.params[loki] = p.Weight
|
|
case "aliases":
|
|
p.Aliases = cast.ToStringSlice(v)
|
|
for _, alias := range p.Aliases {
|
|
if strings.HasPrefix(alias, "http://") || strings.HasPrefix(alias, "https://") {
|
|
return fmt.Errorf("Only relative aliases are supported, %v provided", alias)
|
|
}
|
|
}
|
|
p.params[loki] = p.Aliases
|
|
case "status":
|
|
p.Status = cast.ToString(v)
|
|
p.params[loki] = p.Status
|
|
case "sitemap":
|
|
p.Sitemap = parseSitemap(cast.ToStringMap(v))
|
|
p.params[loki] = p.Sitemap
|
|
case "iscjklanguage":
|
|
isCJKLanguage = new(bool)
|
|
*isCJKLanguage = cast.ToBool(v)
|
|
case "translationkey":
|
|
p.translationKey = cast.ToString(v)
|
|
p.params[loki] = p.translationKey
|
|
case "resources":
|
|
var resources []map[string]interface{}
|
|
handled := true
|
|
|
|
switch vv := v.(type) {
|
|
case []map[interface{}]interface{}:
|
|
for _, vvv := range vv {
|
|
resources = append(resources, cast.ToStringMap(vvv))
|
|
}
|
|
case []map[string]interface{}:
|
|
for _, vvv := range vv {
|
|
resources = append(resources, vvv)
|
|
}
|
|
case []interface{}:
|
|
for _, vvv := range vv {
|
|
switch vvvv := vvv.(type) {
|
|
case map[interface{}]interface{}:
|
|
resources = append(resources, cast.ToStringMap(vvvv))
|
|
case map[string]interface{}:
|
|
resources = append(resources, vvvv)
|
|
}
|
|
}
|
|
default:
|
|
handled = false
|
|
}
|
|
|
|
if handled {
|
|
p.params[loki] = resources
|
|
p.resourcesMetadata = resources
|
|
break
|
|
}
|
|
fallthrough
|
|
|
|
default:
|
|
// If not one of the explicit values, store in Params
|
|
switch vv := v.(type) {
|
|
case bool:
|
|
p.params[loki] = vv
|
|
case string:
|
|
p.params[loki] = vv
|
|
case int64, int32, int16, int8, int:
|
|
p.params[loki] = vv
|
|
case float64, float32:
|
|
p.params[loki] = vv
|
|
case time.Time:
|
|
p.params[loki] = vv
|
|
default: // handle array of strings as well
|
|
switch vvv := vv.(type) {
|
|
case []interface{}:
|
|
if len(vvv) > 0 {
|
|
switch vvv[0].(type) {
|
|
case map[interface{}]interface{}: // Proper parsing structured array from YAML based FrontMatter
|
|
p.params[loki] = vvv
|
|
case map[string]interface{}: // Proper parsing structured array from JSON based FrontMatter
|
|
p.params[loki] = vvv
|
|
case []interface{}:
|
|
p.params[loki] = vvv
|
|
default:
|
|
a := make([]string, len(vvv))
|
|
for i, u := range vvv {
|
|
a[i] = cast.ToString(u)
|
|
}
|
|
|
|
p.params[loki] = a
|
|
}
|
|
} else {
|
|
p.params[loki] = []string{}
|
|
}
|
|
default:
|
|
p.params[loki] = vv
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Try markup explicitly set in the frontmatter
|
|
p.Markup = helpers.GuessType(p.Markup)
|
|
if p.Markup == "unknown" {
|
|
// Fall back to file extension (might also return "unknown")
|
|
p.Markup = helpers.GuessType(p.Source.Ext())
|
|
}
|
|
|
|
if draft != nil && published != nil {
|
|
p.Draft = *draft
|
|
p.s.Log.ERROR.Printf("page %s has both draft and published settings in its frontmatter. Using draft.", p.File.Path())
|
|
return ErrHasDraftAndPublished
|
|
} else if draft != nil {
|
|
p.Draft = *draft
|
|
} else if published != nil {
|
|
p.Draft = !*published
|
|
}
|
|
p.params["draft"] = p.Draft
|
|
|
|
if isCJKLanguage != nil {
|
|
p.isCJKLanguage = *isCJKLanguage
|
|
} else if p.s.Cfg.GetBool("hasCJKLanguage") {
|
|
if cjk.Match(p.rawContent) {
|
|
p.isCJKLanguage = true
|
|
} else {
|
|
p.isCJKLanguage = false
|
|
}
|
|
}
|
|
p.params["iscjklanguage"] = p.isCJKLanguage
|
|
|
|
return nil
|
|
}
|
|
|
|
func (p *Page) GetParam(key string) interface{} {
|
|
return p.getParam(key, false)
|
|
}
|
|
|
|
func (p *Page) getParamToLower(key string) interface{} {
|
|
return p.getParam(key, true)
|
|
}
|
|
|
|
func (p *Page) getParam(key string, stringToLower bool) interface{} {
|
|
v := p.params[strings.ToLower(key)]
|
|
|
|
if v == nil {
|
|
return nil
|
|
}
|
|
|
|
switch val := v.(type) {
|
|
case bool:
|
|
return val
|
|
case string:
|
|
if stringToLower {
|
|
return strings.ToLower(val)
|
|
}
|
|
return val
|
|
case int64, int32, int16, int8, int:
|
|
return cast.ToInt(v)
|
|
case float64, float32:
|
|
return cast.ToFloat64(v)
|
|
case time.Time:
|
|
return val
|
|
case []string:
|
|
if stringToLower {
|
|
return helpers.SliceToLower(val)
|
|
}
|
|
return v
|
|
case map[string]interface{}: // JSON and TOML
|
|
return v
|
|
case map[interface{}]interface{}: // YAML
|
|
return v
|
|
}
|
|
|
|
p.s.Log.ERROR.Printf("GetParam(\"%s\"): Unknown type %s\n", key, reflect.TypeOf(v))
|
|
return nil
|
|
}
|
|
|
|
func (p *Page) HasMenuCurrent(menuID string, me *MenuEntry) bool {
|
|
|
|
sectionPagesMenu := p.Site.sectionPagesMenu
|
|
|
|
// page is labeled as "shadow-member" of the menu with the same identifier as the section
|
|
if sectionPagesMenu != "" {
|
|
section := p.Section()
|
|
|
|
if section != "" && sectionPagesMenu == menuID && section == me.Identifier {
|
|
return true
|
|
}
|
|
}
|
|
|
|
if !me.HasChildren() {
|
|
return false
|
|
}
|
|
|
|
menus := p.Menus()
|
|
|
|
if m, ok := menus[menuID]; ok {
|
|
|
|
for _, child := range me.Children {
|
|
if child.IsEqual(m) {
|
|
return true
|
|
}
|
|
if p.HasMenuCurrent(menuID, child) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
}
|
|
|
|
if p.IsPage() {
|
|
return false
|
|
}
|
|
|
|
// The following logic is kept from back when Hugo had both Page and Node types.
|
|
// TODO(bep) consolidate / clean
|
|
nme := MenuEntry{Page: p, Name: p.title, URL: p.URL()}
|
|
|
|
for _, child := range me.Children {
|
|
if nme.IsSameResource(child) {
|
|
return true
|
|
}
|
|
if p.HasMenuCurrent(menuID, child) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
func (p *Page) IsMenuCurrent(menuID string, inme *MenuEntry) bool {
|
|
|
|
menus := p.Menus()
|
|
|
|
if me, ok := menus[menuID]; ok {
|
|
if me.IsEqual(inme) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
if p.IsPage() {
|
|
return false
|
|
}
|
|
|
|
// The following logic is kept from back when Hugo had both Page and Node types.
|
|
// TODO(bep) consolidate / clean
|
|
me := MenuEntry{Page: p, Name: p.title, URL: p.URL()}
|
|
|
|
if !me.IsSameResource(inme) {
|
|
return false
|
|
}
|
|
|
|
// this resource may be included in several menus
|
|
// search for it to make sure that it is in the menu with the given menuId
|
|
if menu, ok := (*p.Site.Menus)[menuID]; ok {
|
|
for _, menuEntry := range *menu {
|
|
if menuEntry.IsSameResource(inme) {
|
|
return true
|
|
}
|
|
|
|
descendantFound := p.isSameAsDescendantMenu(inme, menuEntry)
|
|
if descendantFound {
|
|
return descendantFound
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
func (p *Page) isSameAsDescendantMenu(inme *MenuEntry, parent *MenuEntry) bool {
|
|
if parent.HasChildren() {
|
|
for _, child := range parent.Children {
|
|
if child.IsSameResource(inme) {
|
|
return true
|
|
}
|
|
descendantFound := p.isSameAsDescendantMenu(inme, child)
|
|
if descendantFound {
|
|
return descendantFound
|
|
}
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func (p *Page) Menus() PageMenus {
|
|
p.pageMenusInit.Do(func() {
|
|
p.pageMenus = PageMenus{}
|
|
|
|
if ms, ok := p.params["menu"]; ok {
|
|
link := p.RelPermalink()
|
|
|
|
me := MenuEntry{Page: p, Name: p.LinkTitle(), Weight: p.Weight, URL: link}
|
|
|
|
// Could be the name of the menu to attach it to
|
|
mname, err := cast.ToStringE(ms)
|
|
|
|
if err == nil {
|
|
me.Menu = mname
|
|
p.pageMenus[mname] = &me
|
|
return
|
|
}
|
|
|
|
// Could be a slice of strings
|
|
mnames, err := cast.ToStringSliceE(ms)
|
|
|
|
if err == nil {
|
|
for _, mname := range mnames {
|
|
me.Menu = mname
|
|
p.pageMenus[mname] = &me
|
|
}
|
|
return
|
|
}
|
|
|
|
// Could be a structured menu entry
|
|
menus, err := cast.ToStringMapE(ms)
|
|
|
|
if err != nil {
|
|
p.s.Log.ERROR.Printf("unable to process menus for %q\n", p.title)
|
|
}
|
|
|
|
for name, menu := range menus {
|
|
menuEntry := MenuEntry{Page: p, Name: p.LinkTitle(), URL: link, Weight: p.Weight, Menu: name}
|
|
if menu != nil {
|
|
p.s.Log.DEBUG.Printf("found menu: %q, in %q\n", name, p.title)
|
|
ime, err := cast.ToStringMapE(menu)
|
|
if err != nil {
|
|
p.s.Log.ERROR.Printf("unable to process menus for %q: %s", p.title, err)
|
|
}
|
|
|
|
menuEntry.marshallMap(ime)
|
|
}
|
|
p.pageMenus[name] = &menuEntry
|
|
|
|
}
|
|
}
|
|
})
|
|
|
|
return p.pageMenus
|
|
}
|
|
|
|
func (p *Page) shouldRenderTo(f output.Format) bool {
|
|
_, found := p.outputFormats.GetByName(f.Name)
|
|
return found
|
|
}
|
|
|
|
func (p *Page) parse(reader io.Reader) error {
|
|
psr, err := parser.ReadFrom(reader)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
p.renderable = psr.IsRenderable()
|
|
p.frontmatter = psr.FrontMatter()
|
|
p.rawContent = psr.Content()
|
|
p.lang = p.Source.File.Lang()
|
|
|
|
meta, err := psr.Metadata()
|
|
if err != nil {
|
|
return fmt.Errorf("failed to parse page metadata for %q: %s", p.File.Path(), err)
|
|
}
|
|
if meta == nil {
|
|
// missing frontmatter equivalent to empty frontmatter
|
|
meta = map[string]interface{}{}
|
|
}
|
|
|
|
if p.s != nil && p.s.owner != nil {
|
|
gi, enabled := p.s.owner.gitInfo.forPage(p)
|
|
if gi != nil {
|
|
p.GitInfo = gi
|
|
} else if enabled {
|
|
p.s.Log.WARN.Printf("Failed to find GitInfo for page %q", p.Path())
|
|
}
|
|
}
|
|
|
|
return p.update(meta)
|
|
}
|
|
|
|
func (p *Page) RawContent() string {
|
|
return string(p.rawContent)
|
|
}
|
|
|
|
func (p *Page) SetSourceContent(content []byte) {
|
|
p.Source.Content = content
|
|
}
|
|
|
|
func (p *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
|
|
// See https://github.com/gohugoio/hugo/issues/2458
|
|
defer func() {
|
|
if r := recover(); r != nil {
|
|
var ok bool
|
|
err, ok = r.(error)
|
|
if !ok {
|
|
err = fmt.Errorf("error from marshal: %v", r)
|
|
}
|
|
}
|
|
}()
|
|
|
|
buf := bp.GetBuffer()
|
|
defer bp.PutBuffer(buf)
|
|
|
|
err = parser.InterfaceToFrontMatter(in, mark, buf)
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
_, err = buf.WriteRune('\n')
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
p.Source.Frontmatter = buf.Bytes()
|
|
|
|
return
|
|
}
|
|
|
|
func (p *Page) SafeSaveSourceAs(path string) error {
|
|
return p.saveSourceAs(path, true)
|
|
}
|
|
|
|
func (p *Page) SaveSourceAs(path string) error {
|
|
return p.saveSourceAs(path, false)
|
|
}
|
|
|
|
func (p *Page) saveSourceAs(path string, safe bool) error {
|
|
b := bp.GetBuffer()
|
|
defer bp.PutBuffer(b)
|
|
|
|
b.Write(p.Source.Frontmatter)
|
|
b.Write(p.Source.Content)
|
|
|
|
bc := make([]byte, b.Len(), b.Len())
|
|
copy(bc, b.Bytes())
|
|
|
|
return p.saveSource(bc, path, safe)
|
|
}
|
|
|
|
func (p *Page) saveSource(by []byte, inpath string, safe bool) (err error) {
|
|
if !filepath.IsAbs(inpath) {
|
|
inpath = p.s.PathSpec.AbsPathify(inpath)
|
|
}
|
|
p.s.Log.INFO.Println("creating", inpath)
|
|
if safe {
|
|
err = helpers.SafeWriteToDisk(inpath, bytes.NewReader(by), p.s.Fs.Source)
|
|
} else {
|
|
err = helpers.WriteToDisk(inpath, bytes.NewReader(by), p.s.Fs.Source)
|
|
}
|
|
if err != nil {
|
|
return
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (p *Page) SaveSource() error {
|
|
return p.SaveSourceAs(p.FullFilePath())
|
|
}
|
|
|
|
// TODO(bep) lazy consolidate
|
|
func (p *Page) processShortcodes() error {
|
|
p.shortcodeState = newShortcodeHandler(p)
|
|
tmpContent, err := p.shortcodeState.extractShortcodes(string(p.workContent), p.withoutContent())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
p.workContent = []byte(tmpContent)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
func (p *Page) FullFilePath() string {
|
|
return filepath.Join(p.Dir(), p.LogicalName())
|
|
}
|
|
|
|
// Returns the canonical, absolute fully-qualifed logical reference used by
|
|
// methods such as GetPage and ref/relref shortcodes to refer to
|
|
// this page. It is prefixed with a "/".
|
|
//
|
|
// For pages that have a source file, it is returns the path to this file as an
|
|
// absolute path rooted in this site's content dir.
|
|
// For pages that do not (sections witout content page etc.), it returns the
|
|
// virtual path, consistent with where you would add a source file.
|
|
func (p *Page) absoluteSourceRef() string {
|
|
if p.Source.File != nil {
|
|
sourcePath := p.Source.Path()
|
|
if sourcePath != "" {
|
|
return "/" + filepath.ToSlash(sourcePath)
|
|
}
|
|
}
|
|
|
|
if len(p.sections) > 0 {
|
|
// no backing file, return the virtual source path
|
|
return "/" + path.Join(p.sections...)
|
|
}
|
|
|
|
return ""
|
|
}
|
|
|
|
// Pre render prepare steps
|
|
|
|
func (p *Page) prepareLayouts() error {
|
|
// TODO(bep): Check the IsRenderable logic.
|
|
if p.Kind == KindPage {
|
|
if !p.IsRenderable() {
|
|
self := "__" + p.UniqueID()
|
|
err := p.s.TemplateHandler().AddLateTemplate(self, string(p.content()))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
p.selfLayout = self
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (p *Page) prepareData(s *Site) error {
|
|
if p.Kind != KindSection {
|
|
var pages Pages
|
|
p.data = make(map[string]interface{})
|
|
|
|
switch p.Kind {
|
|
case KindPage:
|
|
case KindHome:
|
|
pages = s.RegularPages
|
|
case KindTaxonomy:
|
|
plural := p.sections[0]
|
|
term := p.sections[1]
|
|
|
|
if s.Info.preserveTaxonomyNames {
|
|
if v, ok := s.taxonomiesOrigKey[fmt.Sprintf("%s-%s", plural, term)]; ok {
|
|
term = v
|
|
}
|
|
}
|
|
|
|
singular := s.taxonomiesPluralSingular[plural]
|
|
taxonomy := s.Taxonomies[plural].Get(term)
|
|
|
|
p.data[singular] = taxonomy
|
|
p.data["Singular"] = singular
|
|
p.data["Plural"] = plural
|
|
p.data["Term"] = term
|
|
pages = taxonomy.Pages()
|
|
case KindTaxonomyTerm:
|
|
plural := p.sections[0]
|
|
singular := s.taxonomiesPluralSingular[plural]
|
|
|
|
p.data["Singular"] = singular
|
|
p.data["Plural"] = plural
|
|
p.data["Terms"] = s.Taxonomies[plural]
|
|
// keep the following just for legacy reasons
|
|
p.data["OrderedIndex"] = p.data["Terms"]
|
|
p.data["Index"] = p.data["Terms"]
|
|
|
|
// A list of all KindTaxonomy pages with matching plural
|
|
for _, p := range s.findPagesByKind(KindTaxonomy) {
|
|
if p.sections[0] == plural {
|
|
pages = append(pages, p)
|
|
}
|
|
}
|
|
}
|
|
|
|
p.data["Pages"] = pages
|
|
p.Pages = pages
|
|
}
|
|
|
|
// Now we know enough to set missing dates on home page etc.
|
|
p.updatePageDates()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (p *Page) updatePageDates() {
|
|
// TODO(bep) there is a potential issue with page sorting for home pages
|
|
// etc. without front matter dates set, but let us wrap the head around
|
|
// that in another time.
|
|
if !p.IsNode() {
|
|
return
|
|
}
|
|
|
|
if !p.Date.IsZero() {
|
|
if p.Lastmod.IsZero() {
|
|
p.Lastmod = p.Date
|
|
}
|
|
return
|
|
} else if !p.Lastmod.IsZero() {
|
|
if p.Date.IsZero() {
|
|
p.Date = p.Lastmod
|
|
}
|
|
return
|
|
}
|
|
|
|
// Set it to the first non Zero date in children
|
|
var foundDate, foundLastMod bool
|
|
|
|
for _, child := range p.Pages {
|
|
if !child.Date.IsZero() {
|
|
p.Date = child.Date
|
|
foundDate = true
|
|
}
|
|
if !child.Lastmod.IsZero() {
|
|
p.Lastmod = child.Lastmod
|
|
foundLastMod = true
|
|
}
|
|
|
|
if foundDate && foundLastMod {
|
|
break
|
|
}
|
|
}
|
|
}
|
|
|
|
// copy creates a copy of this page with the lazy sync.Once vars reset
|
|
// so they will be evaluated again, for word count calculations etc.
|
|
func (p *Page) copy(initContent bool) *Page {
|
|
p.contentInitMu.Lock()
|
|
c := *p
|
|
p.contentInitMu.Unlock()
|
|
c.pageInit = &pageInit{}
|
|
if initContent {
|
|
if len(p.outputFormats) < 2 {
|
|
panic(fmt.Sprintf("programming error: page %q should not need to rebuild content as it has only %d outputs", p.Path(), len(p.outputFormats)))
|
|
}
|
|
c.pageContentInit = &pageContentInit{}
|
|
}
|
|
return &c
|
|
}
|
|
|
|
func (p *Page) Hugo() *HugoInfo {
|
|
return hugoInfo
|
|
}
|
|
|
|
// GetPage looks up a page for the given ref.
|
|
// {{ with .GetPage "blog" }}{{ .Title }}{{ end }}
|
|
//
|
|
// This will return nil when no page could be found, and will return
|
|
// an error if the ref is ambiguous.
|
|
func (p *Page) GetPage(ref string) (*Page, error) {
|
|
return p.s.getPageNew(p, ref)
|
|
}
|
|
|
|
type refArgs struct {
|
|
Path string
|
|
Lang string
|
|
OutputFormat string
|
|
}
|
|
|
|
func (p *Page) decodeRefArgs(args map[string]interface{}) (refArgs, *SiteInfo, error) {
|
|
var ra refArgs
|
|
err := mapstructure.WeakDecode(args, &ra)
|
|
if err != nil {
|
|
return ra, nil, nil
|
|
}
|
|
s := p.Site
|
|
|
|
if ra.Lang != "" && ra.Lang != p.Lang() {
|
|
// Find correct site
|
|
found := false
|
|
for _, ss := range p.s.owner.Sites {
|
|
if ss.Lang() == ra.Lang {
|
|
found = true
|
|
s = &ss.Info
|
|
}
|
|
}
|
|
|
|
if !found {
|
|
p.s.siteRefLinker.logNotFound(ra.Path, fmt.Sprintf("no site found with lang %q", ra.Lang), p)
|
|
return ra, nil, nil
|
|
}
|
|
}
|
|
|
|
return ra, s, nil
|
|
}
|
|
|
|
func (p *Page) Ref(argsm map[string]interface{}) (string, error) {
|
|
args, s, err := p.decodeRefArgs(argsm)
|
|
if err != nil {
|
|
return "", fmt.Errorf("invalid arguments to Ref: %s", err)
|
|
}
|
|
|
|
if s == nil {
|
|
return p.s.siteRefLinker.notFoundURL, nil
|
|
}
|
|
|
|
if args.Path == "" {
|
|
return "", nil
|
|
}
|
|
|
|
if args.OutputFormat != "" {
|
|
return s.Ref(args.Path, p, args.OutputFormat)
|
|
}
|
|
return s.Ref(args.Path, p)
|
|
}
|
|
|
|
func (p *Page) RelRef(argsm map[string]interface{}) (string, error) {
|
|
args, s, err := p.decodeRefArgs(argsm)
|
|
if err != nil {
|
|
return "", fmt.Errorf("invalid arguments to Ref: %s", err)
|
|
}
|
|
|
|
if s == nil {
|
|
return p.s.siteRefLinker.notFoundURL, nil
|
|
}
|
|
|
|
if args.Path == "" {
|
|
return "", nil
|
|
}
|
|
|
|
if args.OutputFormat != "" {
|
|
return s.RelRef(args.Path, p, args.OutputFormat)
|
|
}
|
|
return s.RelRef(args.Path, p)
|
|
}
|
|
|
|
func (p *Page) String() string {
|
|
if sourceRef := p.absoluteSourceRef(); sourceRef != "" {
|
|
return fmt.Sprintf("Page(%s)", sourceRef)
|
|
}
|
|
return fmt.Sprintf("Page(%q)", p.title)
|
|
}
|
|
|
|
// Scratch returns the writable context associated with this Page.
|
|
func (p *Page) Scratch() *maps.Scratch {
|
|
if p.scratch == nil {
|
|
p.scratch = maps.NewScratch()
|
|
}
|
|
return p.scratch
|
|
}
|
|
|
|
func (p *Page) Language() *langs.Language {
|
|
p.initLanguage()
|
|
return p.language
|
|
}
|
|
|
|
func (p *Page) Lang() string {
|
|
// When set, Language can be different from lang in the case where there is a
|
|
// content file (doc.sv.md) with language indicator, but there is no language
|
|
// config for that language. Then the language will fall back on the site default.
|
|
if p.Language() != nil {
|
|
return p.Language().Lang
|
|
}
|
|
return p.lang
|
|
}
|
|
|
|
func (p *Page) isNewTranslation(candidate *Page) bool {
|
|
|
|
if p.Kind != candidate.Kind {
|
|
return false
|
|
}
|
|
|
|
if p.Kind == KindPage || p.Kind == kindUnknown {
|
|
panic("Node type not currently supported for this op")
|
|
}
|
|
|
|
// At this point, we know that this is a traditional Node (home page, section, taxonomy)
|
|
// It represents the same node, but different language, if the sections is the same.
|
|
if len(p.sections) != len(candidate.sections) {
|
|
return false
|
|
}
|
|
|
|
for i := 0; i < len(p.sections); i++ {
|
|
if p.sections[i] != candidate.sections[i] {
|
|
return false
|
|
}
|
|
}
|
|
|
|
// Finally check that it is not already added.
|
|
for _, translation := range p.translations {
|
|
if candidate == translation {
|
|
return false
|
|
}
|
|
}
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
func (p *Page) shouldAddLanguagePrefix() bool {
|
|
if !p.Site.IsMultiLingual() {
|
|
return false
|
|
}
|
|
|
|
if p.s.owner.IsMultihost() {
|
|
return true
|
|
}
|
|
|
|
if p.Lang() == "" {
|
|
return false
|
|
}
|
|
|
|
if !p.Site.defaultContentLanguageInSubdir && p.Lang() == p.s.multilingual().DefaultLang.Lang {
|
|
return false
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
func (p *Page) initLanguage() {
|
|
p.languageInit.Do(func() {
|
|
if p.language != nil {
|
|
return
|
|
}
|
|
|
|
ml := p.s.multilingual()
|
|
if ml == nil {
|
|
panic("Multilanguage not set")
|
|
}
|
|
if p.lang == "" {
|
|
p.lang = ml.DefaultLang.Lang
|
|
p.language = ml.DefaultLang
|
|
return
|
|
}
|
|
|
|
language := ml.Language(p.lang)
|
|
|
|
if language == nil {
|
|
// It can be a file named stefano.chiodino.md.
|
|
p.s.Log.WARN.Printf("Page language (if it is that) not found in multilang setup: %s.", p.lang)
|
|
language = ml.DefaultLang
|
|
}
|
|
|
|
p.language = language
|
|
|
|
})
|
|
}
|
|
|
|
func (p *Page) LanguagePrefix() string {
|
|
return p.Site.LanguagePrefix
|
|
}
|
|
|
|
func (p *Page) addLangPathPrefixIfFlagSet(outfile string, should bool) string {
|
|
if helpers.IsAbsURL(outfile) {
|
|
return outfile
|
|
}
|
|
|
|
if !should {
|
|
return outfile
|
|
}
|
|
|
|
hadSlashSuffix := strings.HasSuffix(outfile, "/")
|
|
|
|
outfile = "/" + path.Join(p.Lang(), outfile)
|
|
if hadSlashSuffix {
|
|
outfile += "/"
|
|
}
|
|
return outfile
|
|
}
|
|
|
|
func sectionsFromFile(fi *fileInfo) []string {
|
|
dirname := fi.Dir()
|
|
dirname = strings.Trim(dirname, helpers.FilePathSeparator)
|
|
if dirname == "" {
|
|
return nil
|
|
}
|
|
parts := strings.Split(dirname, helpers.FilePathSeparator)
|
|
|
|
if fi.bundleTp == bundleLeaf && len(parts) > 0 {
|
|
// my-section/mybundle/index.md => my-section
|
|
return parts[:len(parts)-1]
|
|
}
|
|
|
|
return parts
|
|
}
|
|
|
|
func kindFromFileInfo(fi *fileInfo) string {
|
|
if fi.TranslationBaseName() == "_index" {
|
|
if fi.Dir() == "" {
|
|
return KindHome
|
|
}
|
|
// Could be index for section, taxonomy, taxonomy term
|
|
// We don't know enough yet to determine which
|
|
return kindUnknown
|
|
}
|
|
return KindPage
|
|
}
|
|
|
|
func (p *Page) setValuesForKind(s *Site) {
|
|
if p.Kind == kindUnknown {
|
|
// This is either a taxonomy list, taxonomy term or a section
|
|
nodeType := s.kindFromSections(p.sections)
|
|
|
|
if nodeType == kindUnknown {
|
|
panic(fmt.Sprintf("Unable to determine page kind from %q", p.sections))
|
|
}
|
|
|
|
p.Kind = nodeType
|
|
}
|
|
|
|
switch p.Kind {
|
|
case KindHome:
|
|
p.URLPath.URL = "/"
|
|
case KindPage:
|
|
default:
|
|
if p.URLPath.URL == "" {
|
|
p.URLPath.URL = "/" + path.Join(p.sections...) + "/"
|
|
}
|
|
}
|
|
}
|
|
|
|
// Used in error logs.
|
|
func (p *Page) pathOrTitle() string {
|
|
if p.Path() != "" {
|
|
return p.Path()
|
|
}
|
|
return p.title
|
|
}
|