mirror of
https://github.com/gohugoio/hugo.git
synced 2025-04-27 22:21:07 +03:00

In a multi-language setup, before this commit the Node's Translations() method would return some "dummy nodes" that would point to the correct page (Permalink), but would not be the same as the node it points to -- it would not have the translated title etc. The node creation is, however, so mingled with rendering, whihc is too early to have any global state, so the nodes has to be split in a prepare and a render phase. This commits does that with as small a change as possible. This implementation is a temp solution until we fix #2297. Updates #2309
2456 lines
61 KiB
Go
2456 lines
61 KiB
Go
// Copyright 2016 The Hugo Authors. All rights reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package hugolib
|
|
|
|
import (
|
|
"errors"
|
|
"fmt"
|
|
"html/template"
|
|
"io"
|
|
"net/url"
|
|
"os"
|
|
"path"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"sync/atomic"
|
|
|
|
"github.com/bep/inflect"
|
|
"github.com/fsnotify/fsnotify"
|
|
"github.com/spf13/afero"
|
|
"github.com/spf13/cast"
|
|
bp "github.com/spf13/hugo/bufferpool"
|
|
"github.com/spf13/hugo/helpers"
|
|
"github.com/spf13/hugo/hugofs"
|
|
"github.com/spf13/hugo/parser"
|
|
"github.com/spf13/hugo/source"
|
|
"github.com/spf13/hugo/target"
|
|
"github.com/spf13/hugo/tpl"
|
|
"github.com/spf13/hugo/transform"
|
|
jww "github.com/spf13/jwalterweatherman"
|
|
"github.com/spf13/nitro"
|
|
"github.com/spf13/viper"
|
|
)
|
|
|
|
var _ = transform.AbsURL
|
|
|
|
// used to indicate if run as a test.
|
|
var testMode bool
|
|
|
|
var defaultTimer *nitro.B
|
|
|
|
var (
|
|
distinctErrorLogger = helpers.NewDistinctErrorLogger()
|
|
distinctFeedbackLogger = helpers.NewDistinctFeedbackLogger()
|
|
)
|
|
|
|
type nodeCache struct {
|
|
m map[string]*Node
|
|
sync.RWMutex
|
|
}
|
|
|
|
func (c *nodeCache) reset() {
|
|
c.m = make(map[string]*Node)
|
|
}
|
|
|
|
// Site contains all the information relevant for constructing a static
|
|
// site. The basic flow of information is as follows:
|
|
//
|
|
// 1. A list of Files is parsed and then converted into Pages.
|
|
//
|
|
// 2. Pages contain sections (based on the file they were generated from),
|
|
// aliases and slugs (included in a pages frontmatter) which are the
|
|
// various targets that will get generated. There will be canonical
|
|
// listing. The canonical path can be overruled based on a pattern.
|
|
//
|
|
// 3. Taxonomies are created via configuration and will present some aspect of
|
|
// the final page and typically a perm url.
|
|
//
|
|
// 4. All Pages are passed through a template based on their desired
|
|
// layout based on numerous different elements.
|
|
//
|
|
// 5. The entire collection of files is written to disk.
|
|
type Site struct {
|
|
owner *HugoSites
|
|
|
|
// Used internally to discover duplicates.
|
|
nodeCache *nodeCache
|
|
nodeCacheInit sync.Once
|
|
|
|
Pages Pages
|
|
AllPages Pages
|
|
rawAllPages Pages
|
|
Files []*source.File
|
|
Taxonomies TaxonomyList
|
|
Source source.Input
|
|
Sections Taxonomy
|
|
Info SiteInfo
|
|
Menus Menus
|
|
timer *nitro.B
|
|
targets targetList
|
|
targetListInit sync.Once
|
|
draftCount int
|
|
futureCount int
|
|
expiredCount int
|
|
Data map[string]interface{}
|
|
Language *helpers.Language
|
|
}
|
|
|
|
// reset returns a new Site prepared for rebuild.
|
|
func (s *Site) reset() *Site {
|
|
return &Site{Language: s.Language}
|
|
}
|
|
|
|
// newSite creates a new site in the given language.
|
|
func newSite(lang *helpers.Language) *Site {
|
|
return &Site{Language: lang, Info: SiteInfo{multilingual: newMultiLingualForLanguage(lang)}}
|
|
}
|
|
|
|
// newSite creates a new site in the default language.
|
|
func newSiteDefaultLang() *Site {
|
|
return newSite(helpers.NewDefaultLanguage())
|
|
}
|
|
|
|
// Convenience func used in tests.
|
|
func newSiteFromSources(pathContentPairs ...string) *Site {
|
|
if len(pathContentPairs)%2 != 0 {
|
|
panic("pathContentPairs must come in pairs")
|
|
}
|
|
|
|
sources := make([]source.ByteSource, 0)
|
|
|
|
for i := 0; i < len(pathContentPairs); i += 2 {
|
|
path := pathContentPairs[i]
|
|
content := pathContentPairs[i+1]
|
|
sources = append(sources, source.ByteSource{Name: filepath.FromSlash(path), Content: []byte(content)})
|
|
}
|
|
|
|
return &Site{
|
|
Source: &source.InMemorySource{ByteSource: sources},
|
|
Language: helpers.NewDefaultLanguage(),
|
|
}
|
|
}
|
|
|
|
type targetList struct {
|
|
page target.Output
|
|
pageUgly target.Output
|
|
file target.Output
|
|
alias target.AliasPublisher
|
|
languageAlias target.AliasPublisher
|
|
}
|
|
|
|
type SiteInfo struct {
|
|
BaseURL template.URL
|
|
Taxonomies TaxonomyList
|
|
Authors AuthorList
|
|
Social SiteSocial
|
|
Sections Taxonomy
|
|
Pages *Pages // Includes only pages in this language
|
|
AllPages *Pages // Includes other translated pages, excluding those in this language.
|
|
rawAllPages *Pages // Includes absolute all pages, including drafts etc.
|
|
Files *[]*source.File
|
|
Menus *Menus
|
|
Hugo *HugoInfo
|
|
Title string
|
|
RSSLink string
|
|
Author map[string]interface{}
|
|
LanguageCode string
|
|
DisqusShortname string
|
|
GoogleAnalytics string
|
|
Copyright string
|
|
LastChange time.Time
|
|
Permalinks PermalinkOverrides
|
|
Params map[string]interface{}
|
|
BuildDrafts bool
|
|
canonifyURLs bool
|
|
preserveTaxonomyNames bool
|
|
paginationPageCount uint64
|
|
Data *map[string]interface{}
|
|
|
|
owner *HugoSites
|
|
multilingual *Multilingual
|
|
Language *helpers.Language
|
|
LanguagePrefix string
|
|
Languages helpers.Languages
|
|
defaultContentLanguageInSubdir bool
|
|
}
|
|
|
|
// Used in tests.
|
|
func newSiteInfoDefaultLanguage(baseURL string, pages ...*Page) *SiteInfo {
|
|
ps := Pages(pages)
|
|
|
|
return &SiteInfo{
|
|
BaseURL: template.URL(baseURL),
|
|
rawAllPages: &ps,
|
|
multilingual: newMultiLingualDefaultLanguage(),
|
|
}
|
|
}
|
|
|
|
// SiteSocial is a place to put social details on a site level. These are the
|
|
// standard keys that themes will expect to have available, but can be
|
|
// expanded to any others on a per site basis
|
|
// github
|
|
// facebook
|
|
// facebook_admin
|
|
// twitter
|
|
// twitter_domain
|
|
// googleplus
|
|
// pinterest
|
|
// instagram
|
|
// youtube
|
|
// linkedin
|
|
type SiteSocial map[string]string
|
|
|
|
// GetParam gets a site parameter value if found, nil if not.
|
|
func (s *SiteInfo) GetParam(key string) interface{} {
|
|
v := s.Params[strings.ToLower(key)]
|
|
|
|
if v == nil {
|
|
return nil
|
|
}
|
|
|
|
switch val := v.(type) {
|
|
case bool:
|
|
return val
|
|
case string:
|
|
return val
|
|
case int64, int32, int16, int8, int:
|
|
return cast.ToInt(v)
|
|
case float64, float32:
|
|
return cast.ToFloat64(v)
|
|
case time.Time:
|
|
return val
|
|
case []string:
|
|
return v
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *SiteInfo) IsMultiLingual() bool {
|
|
return len(s.Languages) > 1
|
|
}
|
|
|
|
func (s *SiteInfo) refLink(ref string, page *Page, relative bool) (string, error) {
|
|
var refURL *url.URL
|
|
var err error
|
|
|
|
refURL, err = url.Parse(ref)
|
|
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
var target *Page
|
|
var link string
|
|
|
|
if refURL.Path != "" {
|
|
for _, page := range []*Page(*s.AllPages) {
|
|
refPath := filepath.FromSlash(refURL.Path)
|
|
if page.Source.Path() == refPath || page.Source.LogicalName() == refPath {
|
|
target = page
|
|
break
|
|
}
|
|
}
|
|
|
|
if target == nil {
|
|
return "", fmt.Errorf("No page found with path or logical name \"%s\".\n", refURL.Path)
|
|
}
|
|
|
|
if relative {
|
|
link, err = target.RelPermalink()
|
|
} else {
|
|
link, err = target.Permalink()
|
|
}
|
|
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
}
|
|
|
|
if refURL.Fragment != "" {
|
|
link = link + "#" + refURL.Fragment
|
|
|
|
if refURL.Path != "" && target != nil && !target.getRenderingConfig().PlainIDAnchors {
|
|
link = link + ":" + target.UniqueID()
|
|
} else if page != nil && !page.getRenderingConfig().PlainIDAnchors {
|
|
link = link + ":" + page.UniqueID()
|
|
}
|
|
}
|
|
|
|
return link, nil
|
|
}
|
|
|
|
// Ref will give an absolute URL to ref in the given Page.
|
|
func (s *SiteInfo) Ref(ref string, page *Page) (string, error) {
|
|
return s.refLink(ref, page, false)
|
|
}
|
|
|
|
// RelRef will give an relative URL to ref in the given Page.
|
|
func (s *SiteInfo) RelRef(ref string, page *Page) (string, error) {
|
|
return s.refLink(ref, page, true)
|
|
}
|
|
|
|
// SourceRelativeLink attempts to convert any source page relative links (like [../another.md]) into absolute links
|
|
func (s *SiteInfo) SourceRelativeLink(ref string, currentPage *Page) (string, error) {
|
|
var refURL *url.URL
|
|
var err error
|
|
|
|
refURL, err = url.Parse(strings.TrimPrefix(ref, currentPage.getRenderingConfig().SourceRelativeLinksProjectFolder))
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
if refURL.Scheme != "" {
|
|
// Not a relative source level path
|
|
return ref, nil
|
|
}
|
|
|
|
var target *Page
|
|
var link string
|
|
|
|
if refURL.Path != "" {
|
|
refPath := filepath.Clean(filepath.FromSlash(refURL.Path))
|
|
|
|
if strings.IndexRune(refPath, os.PathSeparator) == 0 { // filepath.IsAbs fails to me.
|
|
refPath = refPath[1:]
|
|
} else {
|
|
if currentPage != nil {
|
|
refPath = filepath.Join(currentPage.Source.Dir(), refURL.Path)
|
|
}
|
|
}
|
|
|
|
for _, page := range []*Page(*s.AllPages) {
|
|
if page.Source.Path() == refPath {
|
|
target = page
|
|
break
|
|
}
|
|
}
|
|
// need to exhaust the test, then try with the others :/
|
|
// if the refPath doesn't end in a filename with extension `.md`, then try with `.md` , and then `/index.md`
|
|
mdPath := strings.TrimSuffix(refPath, string(os.PathSeparator)) + ".md"
|
|
for _, page := range []*Page(*s.AllPages) {
|
|
if page.Source.Path() == mdPath {
|
|
target = page
|
|
break
|
|
}
|
|
}
|
|
indexPath := filepath.Join(refPath, "index.md")
|
|
for _, page := range []*Page(*s.AllPages) {
|
|
if page.Source.Path() == indexPath {
|
|
target = page
|
|
break
|
|
}
|
|
}
|
|
|
|
if target == nil {
|
|
return "", fmt.Errorf("No page found for \"%s\" on page \"%s\".\n", ref, currentPage.Source.Path())
|
|
}
|
|
|
|
link, err = target.RelPermalink()
|
|
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
}
|
|
|
|
if refURL.Fragment != "" {
|
|
link = link + "#" + refURL.Fragment
|
|
|
|
if refURL.Path != "" && target != nil && !target.getRenderingConfig().PlainIDAnchors {
|
|
link = link + ":" + target.UniqueID()
|
|
} else if currentPage != nil && !currentPage.getRenderingConfig().PlainIDAnchors {
|
|
link = link + ":" + currentPage.UniqueID()
|
|
}
|
|
}
|
|
|
|
return link, nil
|
|
}
|
|
|
|
// SourceRelativeLinkFile attempts to convert any non-md source relative links (like [../another.gif]) into absolute links
|
|
func (s *SiteInfo) SourceRelativeLinkFile(ref string, currentPage *Page) (string, error) {
|
|
var refURL *url.URL
|
|
var err error
|
|
|
|
refURL, err = url.Parse(strings.TrimPrefix(ref, currentPage.getRenderingConfig().SourceRelativeLinksProjectFolder))
|
|
if err != nil {
|
|
return "", err
|
|
}
|
|
|
|
if refURL.Scheme != "" {
|
|
// Not a relative source level path
|
|
return ref, nil
|
|
}
|
|
|
|
var target *source.File
|
|
var link string
|
|
|
|
if refURL.Path != "" {
|
|
refPath := filepath.Clean(filepath.FromSlash(refURL.Path))
|
|
|
|
if strings.IndexRune(refPath, os.PathSeparator) == 0 { // filepath.IsAbs fails to me.
|
|
refPath = refPath[1:]
|
|
} else {
|
|
if currentPage != nil {
|
|
refPath = filepath.Join(currentPage.Source.Dir(), refURL.Path)
|
|
}
|
|
}
|
|
|
|
for _, file := range *s.Files {
|
|
if file.Path() == refPath {
|
|
target = file
|
|
break
|
|
}
|
|
}
|
|
|
|
if target == nil {
|
|
return "", fmt.Errorf("No file found for \"%s\" on page \"%s\".\n", ref, currentPage.Source.Path())
|
|
}
|
|
|
|
link = target.Path()
|
|
return "/" + filepath.ToSlash(link), nil
|
|
}
|
|
|
|
return "", fmt.Errorf("failed to find a file to match \"%s\" on page \"%s\"", ref, currentPage.Source.Path())
|
|
}
|
|
|
|
func (s *SiteInfo) addToPaginationPageCount(cnt uint64) {
|
|
atomic.AddUint64(&s.paginationPageCount, cnt)
|
|
}
|
|
|
|
type runmode struct {
|
|
Watching bool
|
|
}
|
|
|
|
func (s *Site) running() bool {
|
|
return s.owner.runMode.Watching
|
|
}
|
|
|
|
func init() {
|
|
defaultTimer = nitro.Initalize()
|
|
}
|
|
|
|
func (s *Site) timerStep(step string) {
|
|
if s.timer == nil {
|
|
s.timer = defaultTimer
|
|
}
|
|
s.timer.Step(step)
|
|
}
|
|
|
|
// reBuild partially rebuilds a site given the filesystem events.
|
|
// It returns whetever the content source was changed.
|
|
func (s *Site) reBuild(events []fsnotify.Event) (bool, error) {
|
|
|
|
jww.DEBUG.Printf("Rebuild for events %q", events)
|
|
|
|
s.timerStep("initialize rebuild")
|
|
|
|
// First we need to determine what changed
|
|
|
|
sourceChanged := []fsnotify.Event{}
|
|
sourceReallyChanged := []fsnotify.Event{}
|
|
tmplChanged := []fsnotify.Event{}
|
|
dataChanged := []fsnotify.Event{}
|
|
i18nChanged := []fsnotify.Event{}
|
|
|
|
// prevent spamming the log on changes
|
|
logger := helpers.NewDistinctFeedbackLogger()
|
|
|
|
for _, ev := range events {
|
|
// Need to re-read source
|
|
if strings.HasPrefix(ev.Name, s.absContentDir()) {
|
|
logger.Println("Source changed", ev.Name)
|
|
sourceChanged = append(sourceChanged, ev)
|
|
}
|
|
if strings.HasPrefix(ev.Name, s.absLayoutDir()) || strings.HasPrefix(ev.Name, s.absThemeDir()) {
|
|
logger.Println("Template changed", ev.Name)
|
|
tmplChanged = append(tmplChanged, ev)
|
|
}
|
|
if strings.HasPrefix(ev.Name, s.absDataDir()) {
|
|
logger.Println("Data changed", ev.Name)
|
|
dataChanged = append(dataChanged, ev)
|
|
}
|
|
if strings.HasPrefix(ev.Name, s.absI18nDir()) {
|
|
logger.Println("i18n changed", ev.Name)
|
|
i18nChanged = append(dataChanged, ev)
|
|
}
|
|
}
|
|
|
|
if len(tmplChanged) > 0 {
|
|
s.prepTemplates(nil)
|
|
s.owner.tmpl.PrintErrors()
|
|
s.timerStep("template prep")
|
|
}
|
|
|
|
if len(dataChanged) > 0 {
|
|
s.readDataFromSourceFS()
|
|
}
|
|
|
|
if len(i18nChanged) > 0 {
|
|
// TODO(bep ml
|
|
s.readI18nSources()
|
|
}
|
|
|
|
// If a content file changes, we need to reload only it and re-render the entire site.
|
|
|
|
// First step is to read the changed files and (re)place them in site.AllPages
|
|
// This includes processing any meta-data for that content
|
|
|
|
// The second step is to convert the content into HTML
|
|
// This includes processing any shortcodes that may be present.
|
|
|
|
// We do this in parallel... even though it's likely only one file at a time.
|
|
// We need to process the reading prior to the conversion for each file, but
|
|
// we can convert one file while another one is still reading.
|
|
errs := make(chan error)
|
|
readResults := make(chan HandledResult)
|
|
filechan := make(chan *source.File)
|
|
convertResults := make(chan HandledResult)
|
|
pageChan := make(chan *Page)
|
|
fileConvChan := make(chan *source.File)
|
|
coordinator := make(chan bool)
|
|
|
|
wg := &sync.WaitGroup{}
|
|
wg.Add(2)
|
|
for i := 0; i < 2; i++ {
|
|
go sourceReader(s, filechan, readResults, wg)
|
|
}
|
|
|
|
wg2 := &sync.WaitGroup{}
|
|
wg2.Add(4)
|
|
for i := 0; i < 2; i++ {
|
|
go fileConverter(s, fileConvChan, convertResults, wg2)
|
|
go pageConverter(s, pageChan, convertResults, wg2)
|
|
}
|
|
|
|
for _, ev := range sourceChanged {
|
|
// The incrementalReadCollator below will also make changes to the site's pages,
|
|
// so we do this first to prevent races.
|
|
if ev.Op&fsnotify.Remove == fsnotify.Remove {
|
|
//remove the file & a create will follow
|
|
path, _ := helpers.GetRelativePath(ev.Name, s.absContentDir())
|
|
s.removePageByPath(path)
|
|
continue
|
|
}
|
|
|
|
// Some editors (Vim) sometimes issue only a Rename operation when writing an existing file
|
|
// Sometimes a rename operation means that file has been renamed other times it means
|
|
// it's been updated
|
|
if ev.Op&fsnotify.Rename == fsnotify.Rename {
|
|
// If the file is still on disk, it's only been updated, if it's not, it's been moved
|
|
if ex, err := afero.Exists(hugofs.Source(), ev.Name); !ex || err != nil {
|
|
path, _ := helpers.GetRelativePath(ev.Name, s.absContentDir())
|
|
s.removePageByPath(path)
|
|
continue
|
|
}
|
|
}
|
|
|
|
sourceReallyChanged = append(sourceReallyChanged, ev)
|
|
}
|
|
|
|
go incrementalReadCollator(s, readResults, pageChan, fileConvChan, coordinator, errs)
|
|
go converterCollator(s, convertResults, errs)
|
|
|
|
if len(tmplChanged) > 0 || len(dataChanged) > 0 {
|
|
// Do not need to read the files again, but they need conversion
|
|
// for shortocde re-rendering.
|
|
for _, p := range s.rawAllPages {
|
|
if p.shouldBuild() {
|
|
pageChan <- p
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, ev := range sourceReallyChanged {
|
|
|
|
file, err := s.reReadFile(ev.Name)
|
|
|
|
if err != nil {
|
|
jww.ERROR.Println("Error reading file", ev.Name, ";", err)
|
|
}
|
|
|
|
if file != nil {
|
|
filechan <- file
|
|
}
|
|
|
|
}
|
|
|
|
// we close the filechan as we have sent everything we want to send to it.
|
|
// this will tell the sourceReaders to stop iterating on that channel
|
|
close(filechan)
|
|
|
|
// waiting for the sourceReaders to all finish
|
|
wg.Wait()
|
|
// Now closing readResults as this will tell the incrementalReadCollator to
|
|
// stop iterating over that.
|
|
close(readResults)
|
|
|
|
// once readResults is finished it will close coordinator and move along
|
|
<-coordinator
|
|
// allow that routine to finish, then close page & fileconvchan as we've sent
|
|
// everything to them we need to.
|
|
close(pageChan)
|
|
close(fileConvChan)
|
|
|
|
wg2.Wait()
|
|
close(convertResults)
|
|
|
|
s.timerStep("read & convert pages from source")
|
|
|
|
return len(sourceChanged) > 0, nil
|
|
|
|
}
|
|
|
|
func (s *Site) loadTemplates() {
|
|
s.owner.tmpl = tpl.InitializeT()
|
|
s.owner.tmpl.LoadTemplates(s.absLayoutDir())
|
|
if s.hasTheme() {
|
|
s.owner.tmpl.LoadTemplatesWithPrefix(s.absThemeDir()+"/layouts", "theme")
|
|
}
|
|
}
|
|
|
|
func (s *Site) prepTemplates(withTemplate func(templ tpl.Template) error) error {
|
|
s.loadTemplates()
|
|
|
|
if withTemplate != nil {
|
|
if err := withTemplate(s.owner.tmpl); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
s.owner.tmpl.MarkReady()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) loadData(sources []source.Input) (err error) {
|
|
jww.DEBUG.Printf("Load Data from %q", sources)
|
|
s.Data = make(map[string]interface{})
|
|
var current map[string]interface{}
|
|
for _, currentSource := range sources {
|
|
for _, r := range currentSource.Files() {
|
|
// Crawl in data tree to insert data
|
|
current = s.Data
|
|
for _, key := range strings.Split(r.Dir(), helpers.FilePathSeparator) {
|
|
if key != "" {
|
|
if _, ok := current[key]; !ok {
|
|
current[key] = make(map[string]interface{})
|
|
}
|
|
current = current[key].(map[string]interface{})
|
|
}
|
|
}
|
|
|
|
data, err := readData(r)
|
|
if err != nil {
|
|
return fmt.Errorf("Failed to read data from %s: %s", filepath.Join(r.Path(), r.LogicalName()), err)
|
|
}
|
|
|
|
if data == nil {
|
|
continue
|
|
}
|
|
|
|
// Copy content from current to data when needed
|
|
if _, ok := current[r.BaseFileName()]; ok {
|
|
data := data.(map[string]interface{})
|
|
|
|
for key, value := range current[r.BaseFileName()].(map[string]interface{}) {
|
|
if _, override := data[key]; override {
|
|
// filepath.Walk walks the files in lexical order, '/' comes before '.'
|
|
// this warning could happen if
|
|
// 1. A theme uses the same key; the main data folder wins
|
|
// 2. A sub folder uses the same key: the sub folder wins
|
|
jww.WARN.Printf("Data for key '%s' in path '%s' is overridden in subfolder", key, r.Path())
|
|
}
|
|
data[key] = value
|
|
}
|
|
}
|
|
|
|
// Insert data
|
|
current[r.BaseFileName()] = data
|
|
}
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func readData(f *source.File) (interface{}, error) {
|
|
switch f.Extension() {
|
|
case "yaml", "yml":
|
|
return parser.HandleYAMLMetaData(f.Bytes())
|
|
case "json":
|
|
return parser.HandleJSONMetaData(f.Bytes())
|
|
case "toml":
|
|
return parser.HandleTOMLMetaData(f.Bytes())
|
|
default:
|
|
jww.WARN.Printf("Data not supported for extension '%s'", f.Extension())
|
|
return nil, nil
|
|
}
|
|
}
|
|
|
|
func (s *Site) readI18nSources() error {
|
|
|
|
i18nSources := []source.Input{&source.Filesystem{Base: s.absI18nDir()}}
|
|
|
|
themeI18nDir, err := helpers.GetThemeI18nDirPath()
|
|
if err == nil {
|
|
i18nSources = []source.Input{&source.Filesystem{Base: themeI18nDir}, i18nSources[0]}
|
|
}
|
|
|
|
if err = loadI18n(i18nSources); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) readDataFromSourceFS() error {
|
|
dataSources := make([]source.Input, 0, 2)
|
|
dataSources = append(dataSources, &source.Filesystem{Base: s.absDataDir()})
|
|
|
|
// have to be last - duplicate keys in earlier entries will win
|
|
themeDataDir, err := helpers.GetThemeDataDirPath()
|
|
if err == nil {
|
|
dataSources = append(dataSources, &source.Filesystem{Base: themeDataDir})
|
|
}
|
|
|
|
err = s.loadData(dataSources)
|
|
s.timerStep("load data")
|
|
return err
|
|
}
|
|
|
|
func (s *Site) preProcess(config BuildCfg) (err error) {
|
|
s.timerStep("Go initialization")
|
|
if err = s.initialize(); err != nil {
|
|
return
|
|
}
|
|
s.prepTemplates(config.withTemplate)
|
|
s.owner.tmpl.PrintErrors()
|
|
s.timerStep("initialize & template prep")
|
|
|
|
if err = s.readDataFromSourceFS(); err != nil {
|
|
return
|
|
}
|
|
|
|
if err = s.readI18nSources(); err != nil {
|
|
return
|
|
}
|
|
|
|
s.timerStep("load i18n")
|
|
return s.createPages()
|
|
|
|
}
|
|
|
|
func (s *Site) postProcess() (err error) {
|
|
|
|
s.setupPrevNext()
|
|
|
|
if err = s.buildSiteMeta(); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("build taxonomies")
|
|
return
|
|
}
|
|
|
|
func (s *Site) setupPrevNext() {
|
|
for i, page := range s.Pages {
|
|
if i < len(s.Pages)-1 {
|
|
page.Next = s.Pages[i+1]
|
|
}
|
|
|
|
if i > 0 {
|
|
page.Prev = s.Pages[i-1]
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Site) render() (err error) {
|
|
// There are sadly some global template funcs etc. that needs the language information.
|
|
viper.Set("Multilingual", s.multilingualEnabled())
|
|
viper.Set("CurrentContentLanguage", s.Language)
|
|
if err = tpl.SetTranslateLang(s.Language.Lang); err != nil {
|
|
return
|
|
}
|
|
|
|
if err = s.renderAliases(); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write aliases")
|
|
if err = s.renderTaxonomiesLists(false); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write taxonomies")
|
|
if err = s.renderListsOfTaxonomyTerms(false); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render & write taxonomy lists")
|
|
if err = s.renderSectionLists(false); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write lists")
|
|
if err = s.renderPages(); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write pages")
|
|
if err = s.renderHomePage(false); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write homepage")
|
|
if err = s.renderSitemap(); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write Sitemap")
|
|
|
|
if err = s.renderRobotsTXT(); err != nil {
|
|
return
|
|
}
|
|
s.timerStep("render and write robots.txt")
|
|
|
|
return
|
|
}
|
|
|
|
func (s *Site) Initialise() (err error) {
|
|
return s.initialize()
|
|
}
|
|
|
|
func (s *Site) initialize() (err error) {
|
|
defer s.initializeSiteInfo()
|
|
s.Menus = Menus{}
|
|
|
|
// May be supplied in tests.
|
|
if s.Source != nil && len(s.Source.Files()) > 0 {
|
|
jww.DEBUG.Println("initialize: Source is already set")
|
|
return
|
|
}
|
|
|
|
if err = s.checkDirectories(); err != nil {
|
|
return err
|
|
}
|
|
|
|
staticDir := helpers.AbsPathify(viper.GetString("StaticDir") + "/")
|
|
|
|
s.Source = &source.Filesystem{
|
|
AvoidPaths: []string{staticDir},
|
|
Base: s.absContentDir(),
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
// HomeAbsURL is a convenience method giving the absolute URL to the home page.
|
|
func (s *SiteInfo) HomeAbsURL() string {
|
|
base := ""
|
|
if s.IsMultiLingual() {
|
|
base = s.Language.Lang
|
|
}
|
|
return helpers.AbsURL(base, false)
|
|
}
|
|
|
|
// SitemapAbsURL is a convenience method giving the absolute URL to the sitemap.
|
|
func (s *SiteInfo) SitemapAbsURL() string {
|
|
sitemapDefault := parseSitemap(viper.GetStringMap("Sitemap"))
|
|
return path.Join(s.HomeAbsURL(), sitemapDefault.Filename)
|
|
}
|
|
|
|
func (s *Site) initializeSiteInfo() {
|
|
|
|
var (
|
|
lang *helpers.Language = s.Language
|
|
languages helpers.Languages
|
|
)
|
|
|
|
if s.owner != nil && s.owner.multilingual != nil {
|
|
languages = s.owner.multilingual.Languages
|
|
}
|
|
|
|
params := lang.Params()
|
|
|
|
permalinks := make(PermalinkOverrides)
|
|
for k, v := range viper.GetStringMapString("Permalinks") {
|
|
permalinks[k] = pathPattern(v)
|
|
}
|
|
|
|
defaultContentInSubDir := viper.GetBool("DefaultContentLanguageInSubdir")
|
|
defaultContentLanguage := viper.GetString("DefaultContentLanguage")
|
|
|
|
languagePrefix := ""
|
|
if s.multilingualEnabled() && (defaultContentInSubDir || lang.Lang != defaultContentLanguage) {
|
|
languagePrefix = "/" + lang.Lang
|
|
}
|
|
|
|
var multilingual *Multilingual
|
|
if s.owner != nil {
|
|
multilingual = s.owner.multilingual
|
|
}
|
|
|
|
s.Info = SiteInfo{
|
|
BaseURL: template.URL(helpers.SanitizeURLKeepTrailingSlash(viper.GetString("BaseURL"))),
|
|
Title: lang.GetString("Title"),
|
|
Author: lang.GetStringMap("author"),
|
|
Social: lang.GetStringMapString("social"),
|
|
LanguageCode: lang.GetString("languagecode"),
|
|
Copyright: lang.GetString("copyright"),
|
|
DisqusShortname: lang.GetString("DisqusShortname"),
|
|
multilingual: multilingual,
|
|
Language: lang,
|
|
LanguagePrefix: languagePrefix,
|
|
Languages: languages,
|
|
defaultContentLanguageInSubdir: defaultContentInSubDir,
|
|
GoogleAnalytics: lang.GetString("GoogleAnalytics"),
|
|
RSSLink: permalinkStr(viper.GetString("RSSUri")),
|
|
BuildDrafts: viper.GetBool("BuildDrafts"),
|
|
canonifyURLs: viper.GetBool("CanonifyURLs"),
|
|
preserveTaxonomyNames: viper.GetBool("PreserveTaxonomyNames"),
|
|
AllPages: &s.AllPages,
|
|
Pages: &s.Pages,
|
|
rawAllPages: &s.rawAllPages,
|
|
Files: &s.Files,
|
|
Menus: &s.Menus,
|
|
Params: params,
|
|
Permalinks: permalinks,
|
|
Data: &s.Data,
|
|
owner: s.owner,
|
|
}
|
|
}
|
|
|
|
func (s *Site) hasTheme() bool {
|
|
return viper.GetString("theme") != ""
|
|
}
|
|
|
|
func (s *Site) absDataDir() string {
|
|
return helpers.AbsPathify(viper.GetString("DataDir"))
|
|
}
|
|
|
|
func (s *Site) absI18nDir() string {
|
|
return helpers.AbsPathify(viper.GetString("I18nDir"))
|
|
}
|
|
|
|
func (s *Site) absThemeDir() string {
|
|
return helpers.AbsPathify(viper.GetString("themesDir") + "/" + viper.GetString("theme"))
|
|
}
|
|
|
|
func (s *Site) absLayoutDir() string {
|
|
return helpers.AbsPathify(viper.GetString("LayoutDir"))
|
|
}
|
|
|
|
func (s *Site) absContentDir() string {
|
|
return helpers.AbsPathify(viper.GetString("ContentDir"))
|
|
}
|
|
|
|
func (s *Site) absPublishDir() string {
|
|
return helpers.AbsPathify(viper.GetString("PublishDir"))
|
|
}
|
|
|
|
func (s *Site) checkDirectories() (err error) {
|
|
if b, _ := helpers.DirExists(s.absContentDir(), hugofs.Source()); !b {
|
|
return fmt.Errorf("No source directory found, expecting to find it at " + s.absContentDir())
|
|
}
|
|
return
|
|
}
|
|
|
|
// reReadFile resets file to be read from disk again
|
|
func (s *Site) reReadFile(absFilePath string) (*source.File, error) {
|
|
jww.INFO.Println("rereading", absFilePath)
|
|
var file *source.File
|
|
|
|
reader, err := source.NewLazyFileReader(hugofs.Source(), absFilePath)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
file, err = source.NewFileFromAbs(s.absContentDir(), absFilePath, reader)
|
|
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return file, nil
|
|
}
|
|
|
|
func (s *Site) readPagesFromSource() chan error {
|
|
if s.Source == nil {
|
|
panic(fmt.Sprintf("s.Source not set %s", s.absContentDir()))
|
|
}
|
|
|
|
jww.DEBUG.Printf("Read %d pages from source", len(s.Source.Files()))
|
|
|
|
errs := make(chan error)
|
|
if len(s.Source.Files()) < 1 {
|
|
close(errs)
|
|
return errs
|
|
}
|
|
|
|
files := s.Source.Files()
|
|
results := make(chan HandledResult)
|
|
filechan := make(chan *source.File)
|
|
procs := getGoMaxProcs()
|
|
wg := &sync.WaitGroup{}
|
|
|
|
wg.Add(procs * 4)
|
|
for i := 0; i < procs*4; i++ {
|
|
go sourceReader(s, filechan, results, wg)
|
|
}
|
|
|
|
// we can only have exactly one result collator, since it makes changes that
|
|
// must be synchronized.
|
|
go readCollator(s, results, errs)
|
|
|
|
for _, file := range files {
|
|
filechan <- file
|
|
}
|
|
|
|
close(filechan)
|
|
wg.Wait()
|
|
close(results)
|
|
|
|
return errs
|
|
}
|
|
|
|
func (s *Site) convertSource() chan error {
|
|
errs := make(chan error)
|
|
results := make(chan HandledResult)
|
|
pageChan := make(chan *Page)
|
|
fileConvChan := make(chan *source.File)
|
|
procs := getGoMaxProcs()
|
|
wg := &sync.WaitGroup{}
|
|
|
|
wg.Add(2 * procs * 4)
|
|
for i := 0; i < procs*4; i++ {
|
|
go fileConverter(s, fileConvChan, results, wg)
|
|
go pageConverter(s, pageChan, results, wg)
|
|
}
|
|
|
|
go converterCollator(s, results, errs)
|
|
|
|
for _, p := range s.rawAllPages {
|
|
if p.shouldBuild() {
|
|
pageChan <- p
|
|
}
|
|
}
|
|
|
|
for _, f := range s.Files {
|
|
fileConvChan <- f
|
|
}
|
|
|
|
close(pageChan)
|
|
close(fileConvChan)
|
|
wg.Wait()
|
|
close(results)
|
|
|
|
return errs
|
|
}
|
|
|
|
func (s *Site) createPages() error {
|
|
readErrs := <-s.readPagesFromSource()
|
|
s.timerStep("read pages from source")
|
|
|
|
renderErrs := <-s.convertSource()
|
|
s.timerStep("convert source")
|
|
|
|
if renderErrs == nil && readErrs == nil {
|
|
return nil
|
|
}
|
|
if renderErrs == nil {
|
|
return readErrs
|
|
}
|
|
if readErrs == nil {
|
|
return renderErrs
|
|
}
|
|
|
|
return fmt.Errorf("%s\n%s", readErrs, renderErrs)
|
|
}
|
|
|
|
func sourceReader(s *Site, files <-chan *source.File, results chan<- HandledResult, wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
for file := range files {
|
|
readSourceFile(s, file, results)
|
|
}
|
|
}
|
|
|
|
func readSourceFile(s *Site, file *source.File, results chan<- HandledResult) {
|
|
h := NewMetaHandler(file.Extension())
|
|
if h != nil {
|
|
h.Read(file, s, results)
|
|
} else {
|
|
jww.ERROR.Println("Unsupported File Type", file.Path())
|
|
}
|
|
}
|
|
|
|
func pageConverter(s *Site, pages <-chan *Page, results HandleResults, wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
for page := range pages {
|
|
var h *MetaHandle
|
|
if page.Markup != "" {
|
|
h = NewMetaHandler(page.Markup)
|
|
} else {
|
|
h = NewMetaHandler(page.File.Extension())
|
|
}
|
|
if h != nil {
|
|
h.Convert(page, s, results)
|
|
}
|
|
}
|
|
}
|
|
|
|
func fileConverter(s *Site, files <-chan *source.File, results HandleResults, wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
for file := range files {
|
|
h := NewMetaHandler(file.Extension())
|
|
if h != nil {
|
|
h.Convert(file, s, results)
|
|
}
|
|
}
|
|
}
|
|
|
|
func converterCollator(s *Site, results <-chan HandledResult, errs chan<- error) {
|
|
errMsgs := []string{}
|
|
for r := range results {
|
|
if r.err != nil {
|
|
errMsgs = append(errMsgs, r.err.Error())
|
|
continue
|
|
}
|
|
}
|
|
if len(errMsgs) == 0 {
|
|
errs <- nil
|
|
return
|
|
}
|
|
errs <- fmt.Errorf("Errors rendering pages: %s", strings.Join(errMsgs, "\n"))
|
|
}
|
|
|
|
func (s *Site) addPage(page *Page) {
|
|
s.rawAllPages = append(s.rawAllPages, page)
|
|
}
|
|
|
|
func (s *Site) removePageByPath(path string) {
|
|
if i := s.rawAllPages.FindPagePosByFilePath(path); i >= 0 {
|
|
s.rawAllPages = append(s.rawAllPages[:i], s.rawAllPages[i+1:]...)
|
|
}
|
|
}
|
|
|
|
func (s *Site) removePage(page *Page) {
|
|
if i := s.rawAllPages.FindPagePos(page); i >= 0 {
|
|
s.rawAllPages = append(s.rawAllPages[:i], s.rawAllPages[i+1:]...)
|
|
}
|
|
}
|
|
|
|
func (s *Site) replacePage(page *Page) {
|
|
// will find existing page that matches filepath and remove it
|
|
s.removePage(page)
|
|
s.addPage(page)
|
|
}
|
|
|
|
func (s *Site) replaceFile(sf *source.File) {
|
|
for i, f := range s.Files {
|
|
if f.Path() == sf.Path() {
|
|
s.Files[i] = sf
|
|
return
|
|
}
|
|
}
|
|
|
|
// If a match isn't found, then append it
|
|
s.Files = append(s.Files, sf)
|
|
}
|
|
|
|
func incrementalReadCollator(s *Site, results <-chan HandledResult, pageChan chan *Page, fileConvChan chan *source.File, coordinator chan bool, errs chan<- error) {
|
|
errMsgs := []string{}
|
|
for r := range results {
|
|
if r.err != nil {
|
|
errMsgs = append(errMsgs, r.Error())
|
|
continue
|
|
}
|
|
|
|
if r.page == nil {
|
|
s.replaceFile(r.file)
|
|
fileConvChan <- r.file
|
|
} else {
|
|
s.replacePage(r.page)
|
|
pageChan <- r.page
|
|
}
|
|
}
|
|
|
|
s.rawAllPages.Sort()
|
|
close(coordinator)
|
|
|
|
if len(errMsgs) == 0 {
|
|
errs <- nil
|
|
return
|
|
}
|
|
errs <- fmt.Errorf("Errors reading pages: %s", strings.Join(errMsgs, "\n"))
|
|
}
|
|
|
|
func readCollator(s *Site, results <-chan HandledResult, errs chan<- error) {
|
|
errMsgs := []string{}
|
|
for r := range results {
|
|
if r.err != nil {
|
|
errMsgs = append(errMsgs, r.Error())
|
|
continue
|
|
}
|
|
|
|
// !page == file
|
|
if r.page == nil {
|
|
s.Files = append(s.Files, r.file)
|
|
} else {
|
|
s.addPage(r.page)
|
|
}
|
|
}
|
|
|
|
s.rawAllPages.Sort()
|
|
if len(errMsgs) == 0 {
|
|
errs <- nil
|
|
return
|
|
}
|
|
errs <- fmt.Errorf("Errors reading pages: %s", strings.Join(errMsgs, "\n"))
|
|
}
|
|
|
|
func (s *Site) buildSiteMeta() (err error) {
|
|
s.assembleMenus()
|
|
|
|
if len(s.Pages) == 0 {
|
|
return
|
|
}
|
|
|
|
s.assembleTaxonomies()
|
|
s.assembleSections()
|
|
s.Info.LastChange = s.Pages[0].Lastmod
|
|
|
|
return
|
|
}
|
|
|
|
func (s *Site) getMenusFromConfig() Menus {
|
|
|
|
ret := Menus{}
|
|
|
|
if menus := viper.GetStringMap("menu"); menus != nil {
|
|
for name, menu := range menus {
|
|
m, err := cast.ToSliceE(menu)
|
|
if err != nil {
|
|
jww.ERROR.Printf("unable to process menus in site config\n")
|
|
jww.ERROR.Println(err)
|
|
} else {
|
|
for _, entry := range m {
|
|
jww.DEBUG.Printf("found menu: %q, in site config\n", name)
|
|
|
|
menuEntry := MenuEntry{Menu: name}
|
|
ime, err := cast.ToStringMapE(entry)
|
|
if err != nil {
|
|
jww.ERROR.Printf("unable to process menus in site config\n")
|
|
jww.ERROR.Println(err)
|
|
}
|
|
|
|
menuEntry.marshallMap(ime)
|
|
menuEntry.URL = s.Info.createNodeMenuEntryURL(menuEntry.URL)
|
|
|
|
if ret[name] == nil {
|
|
ret[name] = &Menu{}
|
|
}
|
|
*ret[name] = ret[name].add(&menuEntry)
|
|
}
|
|
}
|
|
}
|
|
return ret
|
|
}
|
|
return ret
|
|
}
|
|
|
|
func (s *SiteInfo) createNodeMenuEntryURL(in string) string {
|
|
|
|
if !strings.HasPrefix(in, "/") {
|
|
return in
|
|
}
|
|
// make it match the nodes
|
|
menuEntryURL := in
|
|
menuEntryURL = helpers.SanitizeURLKeepTrailingSlash(helpers.URLize(menuEntryURL))
|
|
if !s.canonifyURLs {
|
|
menuEntryURL = helpers.AddContextRoot(string(s.BaseURL), menuEntryURL)
|
|
}
|
|
return menuEntryURL
|
|
}
|
|
|
|
func (s *Site) assembleMenus() {
|
|
s.Menus = Menus{}
|
|
|
|
type twoD struct {
|
|
MenuName, EntryName string
|
|
}
|
|
flat := map[twoD]*MenuEntry{}
|
|
children := map[twoD]Menu{}
|
|
|
|
menuConfig := s.getMenusFromConfig()
|
|
for name, menu := range menuConfig {
|
|
for _, me := range *menu {
|
|
flat[twoD{name, me.KeyName()}] = me
|
|
}
|
|
}
|
|
|
|
sectionPagesMenu := viper.GetString("SectionPagesMenu")
|
|
sectionPagesMenus := make(map[string]interface{})
|
|
//creating flat hash
|
|
for _, p := range s.Pages {
|
|
|
|
if sectionPagesMenu != "" {
|
|
if _, ok := sectionPagesMenus[p.Section()]; !ok {
|
|
if p.Section() != "" {
|
|
me := MenuEntry{Identifier: p.Section(),
|
|
Name: helpers.MakeTitle(helpers.FirstUpper(p.Section())),
|
|
URL: s.Info.createNodeMenuEntryURL(p.addLangPathPrefix("/"+p.Section()) + "/")}
|
|
if _, ok := flat[twoD{sectionPagesMenu, me.KeyName()}]; ok {
|
|
// menu with same id defined in config, let that one win
|
|
continue
|
|
}
|
|
flat[twoD{sectionPagesMenu, me.KeyName()}] = &me
|
|
sectionPagesMenus[p.Section()] = true
|
|
}
|
|
}
|
|
}
|
|
|
|
for name, me := range p.Menus() {
|
|
if _, ok := flat[twoD{name, me.KeyName()}]; ok {
|
|
jww.ERROR.Printf("Two or more menu items have the same name/identifier in Menu %q: %q.\nRename or set an unique identifier.\n", name, me.KeyName())
|
|
continue
|
|
}
|
|
flat[twoD{name, me.KeyName()}] = me
|
|
}
|
|
}
|
|
|
|
// Create Children Menus First
|
|
for _, e := range flat {
|
|
if e.Parent != "" {
|
|
children[twoD{e.Menu, e.Parent}] = children[twoD{e.Menu, e.Parent}].add(e)
|
|
}
|
|
}
|
|
|
|
// Placing Children in Parents (in flat)
|
|
for p, childmenu := range children {
|
|
_, ok := flat[twoD{p.MenuName, p.EntryName}]
|
|
if !ok {
|
|
// if parent does not exist, create one without a URL
|
|
flat[twoD{p.MenuName, p.EntryName}] = &MenuEntry{Name: p.EntryName, URL: ""}
|
|
}
|
|
flat[twoD{p.MenuName, p.EntryName}].Children = childmenu
|
|
}
|
|
|
|
// Assembling Top Level of Tree
|
|
for menu, e := range flat {
|
|
if e.Parent == "" {
|
|
_, ok := s.Menus[menu.MenuName]
|
|
if !ok {
|
|
s.Menus[menu.MenuName] = &Menu{}
|
|
}
|
|
*s.Menus[menu.MenuName] = s.Menus[menu.MenuName].add(e)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Site) assembleTaxonomies() {
|
|
s.Taxonomies = make(TaxonomyList)
|
|
|
|
taxonomies := s.Language.GetStringMapString("Taxonomies")
|
|
jww.INFO.Printf("found taxonomies: %#v\n", taxonomies)
|
|
|
|
for _, plural := range taxonomies {
|
|
s.Taxonomies[plural] = make(Taxonomy)
|
|
for _, p := range s.Pages {
|
|
vals := p.getParam(plural, !s.Info.preserveTaxonomyNames)
|
|
weight := p.GetParam(plural + "_weight")
|
|
if weight == nil {
|
|
weight = 0
|
|
}
|
|
if vals != nil {
|
|
if v, ok := vals.([]string); ok {
|
|
for _, idx := range v {
|
|
x := WeightedPage{weight.(int), p}
|
|
s.Taxonomies[plural].add(idx, x, s.Info.preserveTaxonomyNames)
|
|
}
|
|
} else if v, ok := vals.(string); ok {
|
|
x := WeightedPage{weight.(int), p}
|
|
s.Taxonomies[plural].add(v, x, s.Info.preserveTaxonomyNames)
|
|
} else {
|
|
jww.ERROR.Printf("Invalid %s in %s\n", plural, p.File.Path())
|
|
}
|
|
}
|
|
}
|
|
for k := range s.Taxonomies[plural] {
|
|
s.Taxonomies[plural][k].Sort()
|
|
}
|
|
}
|
|
|
|
s.Info.Taxonomies = s.Taxonomies
|
|
}
|
|
|
|
// Prepare site for a new full build.
|
|
func (s *Site) resetBuildState() {
|
|
|
|
s.nodeCache.reset()
|
|
|
|
s.Pages = make(Pages, 0)
|
|
s.AllPages = make(Pages, 0)
|
|
|
|
s.Info.paginationPageCount = 0
|
|
s.draftCount = 0
|
|
s.futureCount = 0
|
|
s.expiredCount = 0
|
|
|
|
for _, p := range s.rawAllPages {
|
|
p.scratch = newScratch()
|
|
}
|
|
}
|
|
|
|
func (s *Site) assembleSections() {
|
|
s.Sections = make(Taxonomy)
|
|
s.Info.Sections = s.Sections
|
|
|
|
for i, p := range s.Pages {
|
|
s.Sections.add(p.Section(), WeightedPage{s.Pages[i].Weight, s.Pages[i]}, s.Info.preserveTaxonomyNames)
|
|
}
|
|
|
|
for k := range s.Sections {
|
|
s.Sections[k].Sort()
|
|
|
|
for i, wp := range s.Sections[k] {
|
|
if i > 0 {
|
|
wp.Page.NextInSection = s.Sections[k][i-1].Page
|
|
}
|
|
if i < len(s.Sections[k])-1 {
|
|
wp.Page.PrevInSection = s.Sections[k][i+1].Page
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// renderAliases renders shell pages that simply have a redirect in the header.
|
|
func (s *Site) renderAliases() error {
|
|
for _, p := range s.Pages {
|
|
if len(p.Aliases) == 0 {
|
|
continue
|
|
}
|
|
|
|
plink, err := p.Permalink()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
for _, a := range p.Aliases {
|
|
if err := s.writeDestAlias(a, plink); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
if s.owner.multilingual.enabled() && s.Info.defaultContentLanguageInSubdir {
|
|
mainLang := s.owner.multilingual.DefaultLang.Lang
|
|
mainLangURL := helpers.AbsURL(mainLang, false)
|
|
jww.DEBUG.Printf("Write redirect to main language %s: %s", mainLang, mainLangURL)
|
|
if err := s.publishDestAlias(s.languageAliasTarget(), "/", mainLangURL); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// renderPages renders pages each corresponding to a markdown file.
|
|
func (s *Site) renderPages() error {
|
|
|
|
results := make(chan error)
|
|
pages := make(chan *Page)
|
|
errs := make(chan error)
|
|
|
|
go errorCollator(results, errs)
|
|
|
|
procs := getGoMaxProcs()
|
|
|
|
// this cannot be fanned out to multiple Go routines
|
|
// See issue #1601
|
|
// TODO(bep): Check the IsRenderable logic.
|
|
for _, p := range s.Pages {
|
|
var layouts []string
|
|
if !p.IsRenderable() {
|
|
self := "__" + p.TargetPath()
|
|
_, err := s.owner.tmpl.GetClone().New(self).Parse(string(p.Content))
|
|
if err != nil {
|
|
results <- err
|
|
continue
|
|
}
|
|
layouts = append(layouts, self)
|
|
} else {
|
|
layouts = append(layouts, p.layouts()...)
|
|
layouts = append(layouts, "_default/single.html")
|
|
}
|
|
p.layoutsCalculated = layouts
|
|
}
|
|
|
|
wg := &sync.WaitGroup{}
|
|
|
|
for i := 0; i < procs*4; i++ {
|
|
wg.Add(1)
|
|
go pageRenderer(s, pages, results, wg)
|
|
}
|
|
|
|
for _, page := range s.Pages {
|
|
pages <- page
|
|
}
|
|
|
|
close(pages)
|
|
|
|
wg.Wait()
|
|
|
|
close(results)
|
|
|
|
err := <-errs
|
|
if err != nil {
|
|
return fmt.Errorf("Error(s) rendering pages: %s", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func pageRenderer(s *Site, pages <-chan *Page, results chan<- error, wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
for p := range pages {
|
|
err := s.renderAndWritePage("page "+p.FullFilePath(), p.TargetPath(), p, s.appendThemeTemplates(p.layouts())...)
|
|
if err != nil {
|
|
results <- err
|
|
}
|
|
}
|
|
}
|
|
|
|
func errorCollator(results <-chan error, errs chan<- error) {
|
|
errMsgs := []string{}
|
|
for err := range results {
|
|
if err != nil {
|
|
errMsgs = append(errMsgs, err.Error())
|
|
}
|
|
}
|
|
if len(errMsgs) == 0 {
|
|
errs <- nil
|
|
} else {
|
|
errs <- errors.New(strings.Join(errMsgs, "\n"))
|
|
}
|
|
close(errs)
|
|
}
|
|
|
|
func (s *Site) appendThemeTemplates(in []string) []string {
|
|
if !s.hasTheme() {
|
|
return in
|
|
}
|
|
|
|
out := []string{}
|
|
// First place all non internal templates
|
|
for _, t := range in {
|
|
if !strings.HasPrefix(t, "_internal/") {
|
|
out = append(out, t)
|
|
}
|
|
}
|
|
|
|
// Then place theme templates with the same names
|
|
for _, t := range in {
|
|
if !strings.HasPrefix(t, "_internal/") {
|
|
out = append(out, "theme/"+t)
|
|
}
|
|
}
|
|
|
|
// Lastly place internal templates
|
|
for _, t := range in {
|
|
if strings.HasPrefix(t, "_internal/") {
|
|
out = append(out, t)
|
|
}
|
|
}
|
|
return out
|
|
|
|
}
|
|
|
|
type taxRenderInfo struct {
|
|
key string
|
|
pages WeightedPages
|
|
singular string
|
|
plural string
|
|
}
|
|
|
|
// renderTaxonomiesLists renders the listing pages based on the meta data
|
|
// each unique term within a taxonomy will have a page created
|
|
func (s *Site) renderTaxonomiesLists(prepare bool) error {
|
|
wg := &sync.WaitGroup{}
|
|
|
|
taxes := make(chan taxRenderInfo)
|
|
results := make(chan error)
|
|
|
|
procs := getGoMaxProcs()
|
|
|
|
for i := 0; i < procs*4; i++ {
|
|
wg.Add(1)
|
|
go taxonomyRenderer(prepare, s, taxes, results, wg)
|
|
}
|
|
|
|
errs := make(chan error)
|
|
|
|
go errorCollator(results, errs)
|
|
|
|
taxonomies := s.Language.GetStringMapString("Taxonomies")
|
|
for singular, plural := range taxonomies {
|
|
for key, pages := range s.Taxonomies[plural] {
|
|
taxes <- taxRenderInfo{key, pages, singular, plural}
|
|
}
|
|
}
|
|
close(taxes)
|
|
|
|
wg.Wait()
|
|
|
|
close(results)
|
|
|
|
err := <-errs
|
|
if err != nil {
|
|
return fmt.Errorf("Error(s) rendering taxonomies: %s", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) newTaxonomyNode(prepare bool, t taxRenderInfo, counter int) (*Node, string) {
|
|
key := t.key
|
|
n := s.nodeLookup(fmt.Sprintf("tax-%s-%s", t.plural, key), counter, prepare)
|
|
|
|
if s.Info.preserveTaxonomyNames {
|
|
key = helpers.MakePathSanitized(key)
|
|
}
|
|
base := t.plural + "/" + key
|
|
|
|
if !prepare {
|
|
return n, base
|
|
}
|
|
|
|
if s.Info.preserveTaxonomyNames {
|
|
key = helpers.MakePathSanitized(key)
|
|
// keep as is in the title
|
|
n.Title = t.key
|
|
} else {
|
|
n.Title = strings.Replace(strings.Title(t.key), "-", " ", -1)
|
|
}
|
|
s.setURLs(n, base)
|
|
if len(t.pages) > 0 {
|
|
n.Date = t.pages[0].Page.Date
|
|
n.Lastmod = t.pages[0].Page.Lastmod
|
|
}
|
|
n.Data[t.singular] = t.pages
|
|
n.Data["Singular"] = t.singular
|
|
n.Data["Plural"] = t.plural
|
|
n.Data["Pages"] = t.pages.Pages()
|
|
return n, base
|
|
}
|
|
|
|
func taxonomyRenderer(prepare bool, s *Site, taxes <-chan taxRenderInfo, results chan<- error, wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
|
|
var n *Node
|
|
|
|
for t := range taxes {
|
|
|
|
var (
|
|
base string
|
|
baseWithLanguagePrefix string
|
|
paginatePath string
|
|
layouts []string
|
|
)
|
|
|
|
n, base = s.newTaxonomyNode(prepare, t, 0)
|
|
|
|
if prepare {
|
|
continue
|
|
}
|
|
|
|
baseWithLanguagePrefix = n.addLangPathPrefix(base)
|
|
|
|
layouts = s.appendThemeTemplates(
|
|
[]string{"taxonomy/" + t.singular + ".html", "indexes/" + t.singular + ".html", "_default/taxonomy.html", "_default/list.html"})
|
|
|
|
dest := base
|
|
if viper.GetBool("UglyURLs") {
|
|
dest = helpers.Uglify(baseWithLanguagePrefix + ".html")
|
|
} else {
|
|
dest = helpers.PrettifyPath(baseWithLanguagePrefix + "/index.html")
|
|
}
|
|
|
|
if err := s.renderAndWritePage("taxonomy "+t.singular, dest, n, layouts...); err != nil {
|
|
results <- err
|
|
continue
|
|
}
|
|
|
|
if n.paginator != nil {
|
|
|
|
paginatePath = viper.GetString("paginatePath")
|
|
|
|
// write alias for page 1
|
|
s.writeDestAlias(helpers.PaginateAliasPath(baseWithLanguagePrefix, 1), n.Permalink())
|
|
|
|
pagers := n.paginator.Pagers()
|
|
|
|
for i, pager := range pagers {
|
|
if i == 0 {
|
|
// already created
|
|
continue
|
|
}
|
|
|
|
taxonomyPagerNode, _ := s.newTaxonomyNode(true, t, i)
|
|
|
|
taxonomyPagerNode.paginator = pager
|
|
if pager.TotalPages() > 0 {
|
|
first, _ := pager.page(0)
|
|
taxonomyPagerNode.Date = first.Date
|
|
taxonomyPagerNode.Lastmod = first.Lastmod
|
|
}
|
|
|
|
pageNumber := i + 1
|
|
htmlBase := fmt.Sprintf("/%s/%s/%d", baseWithLanguagePrefix, paginatePath, pageNumber)
|
|
if err := s.renderAndWritePage(fmt.Sprintf("taxonomy %s", t.singular), htmlBase, taxonomyPagerNode, layouts...); err != nil {
|
|
results <- err
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
|
|
if prepare {
|
|
continue
|
|
}
|
|
|
|
if !viper.GetBool("DisableRSS") {
|
|
// XML Feed
|
|
c := *n
|
|
rssNode := &c
|
|
rssNode.nodeID = ""
|
|
rssuri := viper.GetString("RSSUri")
|
|
s.setURLs(rssNode, base+"/"+rssuri)
|
|
|
|
rssLayouts := []string{"taxonomy/" + t.singular + ".rss.xml", "_default/rss.xml", "rss.xml", "_internal/_default/rss.xml"}
|
|
|
|
if err := s.renderAndWriteXML("taxonomy "+t.singular+" rss", baseWithLanguagePrefix+"/"+rssuri, rssNode, s.appendThemeTemplates(rssLayouts)...); err != nil {
|
|
results <- err
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// renderListsOfTaxonomyTerms renders a page per taxonomy that lists the terms for that taxonomy
|
|
func (s *Site) renderListsOfTaxonomyTerms(prepare bool) (err error) {
|
|
taxonomies := s.Language.GetStringMapString("Taxonomies")
|
|
for singular, plural := range taxonomies {
|
|
n := s.nodeLookup(fmt.Sprintf("taxlist-%s", plural), 0, prepare)
|
|
|
|
if prepare {
|
|
n.Title = strings.Title(plural)
|
|
s.setURLs(n, plural)
|
|
n.Data["Singular"] = singular
|
|
n.Data["Plural"] = plural
|
|
n.Data["Terms"] = s.Taxonomies[plural]
|
|
// keep the following just for legacy reasons
|
|
n.Data["OrderedIndex"] = n.Data["Terms"]
|
|
n.Data["Index"] = n.Data["Terms"]
|
|
|
|
continue
|
|
}
|
|
|
|
layouts := []string{"taxonomy/" + singular + ".terms.html", "_default/terms.html", "indexes/indexes.html"}
|
|
layouts = s.appendThemeTemplates(layouts)
|
|
if s.layoutExists(layouts...) {
|
|
if err := s.renderAndWritePage("taxonomy terms for "+singular, n.addLangPathPrefix(plural+"/index.html"), n, layouts...); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
return
|
|
}
|
|
|
|
func (s *Site) newSectionListNode(prepare bool, sectionName, section string, data WeightedPages, counter int) *Node {
|
|
n := s.nodeLookup(fmt.Sprintf("sect-%s", sectionName), counter, prepare)
|
|
|
|
if !prepare {
|
|
return n
|
|
}
|
|
|
|
sectionName = helpers.FirstUpper(sectionName)
|
|
if viper.GetBool("PluralizeListTitles") {
|
|
n.Title = inflect.Pluralize(sectionName)
|
|
} else {
|
|
n.Title = sectionName
|
|
}
|
|
s.setURLs(n, section)
|
|
n.Date = data[0].Page.Date
|
|
n.Lastmod = data[0].Page.Lastmod
|
|
n.Data["Pages"] = data.Pages()
|
|
|
|
return n
|
|
}
|
|
|
|
// renderSectionLists renders a page for each section
|
|
func (s *Site) renderSectionLists(prepare bool) error {
|
|
for section, data := range s.Sections {
|
|
// section keys can be lower case (depending on site.pathifyTaxonomyKeys)
|
|
// extract the original casing from the first page to get sensible titles.
|
|
sectionName := section
|
|
if !s.Info.preserveTaxonomyNames && len(data) > 0 {
|
|
sectionName = data[0].Page.Section()
|
|
}
|
|
|
|
n := s.newSectionListNode(prepare, sectionName, section, data, 0)
|
|
|
|
if prepare {
|
|
continue
|
|
}
|
|
|
|
layouts := s.appendThemeTemplates(
|
|
[]string{"section/" + section + ".html", "_default/section.html", "_default/list.html", "indexes/" + section + ".html", "_default/indexes.html"})
|
|
|
|
if s.Info.preserveTaxonomyNames {
|
|
section = helpers.MakePathSanitized(section)
|
|
}
|
|
|
|
base := n.addLangPathPrefix(section)
|
|
|
|
if err := s.renderAndWritePage(fmt.Sprintf("section %s", section), base, n, s.appendThemeTemplates(layouts)...); err != nil {
|
|
return err
|
|
}
|
|
|
|
if n.paginator != nil {
|
|
|
|
paginatePath := viper.GetString("paginatePath")
|
|
|
|
// write alias for page 1
|
|
s.writeDestAlias(helpers.PaginateAliasPath(base, 1), permalink(base))
|
|
|
|
pagers := n.paginator.Pagers()
|
|
|
|
for i, pager := range pagers {
|
|
if i == 0 {
|
|
// already created
|
|
continue
|
|
}
|
|
|
|
sectionPagerNode := s.newSectionListNode(true, sectionName, section, data, i)
|
|
sectionPagerNode.paginator = pager
|
|
if pager.TotalPages() > 0 {
|
|
first, _ := pager.page(0)
|
|
sectionPagerNode.Date = first.Date
|
|
sectionPagerNode.Lastmod = first.Lastmod
|
|
}
|
|
pageNumber := i + 1
|
|
htmlBase := fmt.Sprintf("/%s/%s/%d", base, paginatePath, pageNumber)
|
|
if err := s.renderAndWritePage(fmt.Sprintf("section %s", section), filepath.FromSlash(htmlBase), sectionPagerNode, layouts...); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
if prepare {
|
|
return nil
|
|
}
|
|
|
|
if !viper.GetBool("DisableRSS") && section != "" {
|
|
// XML Feed
|
|
rssuri := viper.GetString("RSSUri")
|
|
s.setURLs(n, section+"/"+rssuri)
|
|
rssLayouts := []string{"section/" + section + ".rss.xml", "_default/rss.xml", "rss.xml", "_internal/_default/rss.xml"}
|
|
if err := s.renderAndWriteXML("section "+section+" rss", n.addLangPathPrefix(section+"/"+rssuri), n, s.appendThemeTemplates(rssLayouts)...); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) renderHomePage(prepare bool) error {
|
|
|
|
n := s.newHomeNode(prepare, 0)
|
|
if prepare {
|
|
return nil
|
|
}
|
|
|
|
layouts := s.appendThemeTemplates([]string{"index.html", "_default/list.html"})
|
|
base := n.addLangFilepathPrefix("")
|
|
if err := s.renderAndWritePage("homepage", base, n, layouts...); err != nil {
|
|
return err
|
|
}
|
|
|
|
if n.paginator != nil {
|
|
paginatePath := viper.GetString("paginatePath")
|
|
|
|
{
|
|
// write alias for page 1
|
|
// TODO(bep) ml all of these n.addLang ... fix.
|
|
s.writeDestAlias(n.addLangPathPrefix(helpers.PaginateAliasPath("", 1)), n.Permalink())
|
|
}
|
|
|
|
pagers := n.paginator.Pagers()
|
|
|
|
for i, pager := range pagers {
|
|
if i == 0 {
|
|
// already created
|
|
continue
|
|
}
|
|
|
|
homePagerNode := s.newHomeNode(true, i)
|
|
|
|
homePagerNode.paginator = pager
|
|
if pager.TotalPages() > 0 {
|
|
first, _ := pager.page(0)
|
|
homePagerNode.Date = first.Date
|
|
homePagerNode.Lastmod = first.Lastmod
|
|
}
|
|
|
|
pageNumber := i + 1
|
|
htmlBase := fmt.Sprintf("/%s/%d", paginatePath, pageNumber)
|
|
htmlBase = n.addLangPathPrefix(htmlBase)
|
|
if err := s.renderAndWritePage(fmt.Sprintf("homepage"),
|
|
filepath.FromSlash(htmlBase), homePagerNode, layouts...); err != nil {
|
|
return err
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
if !viper.GetBool("DisableRSS") {
|
|
// XML Feed
|
|
rssNode := s.newNode("rss-home")
|
|
s.setURLs(rssNode, viper.GetString("RSSUri"))
|
|
rssNode.Title = ""
|
|
high := 50
|
|
if len(s.Pages) < high {
|
|
high = len(s.Pages)
|
|
}
|
|
rssNode.Data["Pages"] = s.Pages[:high]
|
|
if len(s.Pages) > 0 {
|
|
rssNode.Date = s.Pages[0].Date
|
|
rssNode.Lastmod = s.Pages[0].Lastmod
|
|
}
|
|
|
|
rssLayouts := []string{"rss.xml", "_default/rss.xml", "_internal/_default/rss.xml"}
|
|
|
|
if err := s.renderAndWriteXML("homepage rss", rssNode.addLangPathPrefix(viper.GetString("RSSUri")), rssNode, s.appendThemeTemplates(rssLayouts)...); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
if viper.GetBool("Disable404") {
|
|
return nil
|
|
}
|
|
|
|
node404 := s.newNode("404")
|
|
node404.Title = "404 Page not found"
|
|
s.setURLs(node404, "404.html")
|
|
|
|
nfLayouts := []string{"404.html"}
|
|
if nfErr := s.renderAndWritePage("404 page", "404.html", node404, s.appendThemeTemplates(nfLayouts)...); nfErr != nil {
|
|
return nfErr
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) newHomeNode(prepare bool, counter int) *Node {
|
|
n := s.nodeLookup("home", counter, prepare)
|
|
n.Title = n.Site.Title
|
|
n.IsHome = true
|
|
s.setURLs(n, "/")
|
|
n.Data["Pages"] = s.Pages
|
|
if len(s.Pages) != 0 {
|
|
n.Date = s.Pages[0].Date
|
|
n.Lastmod = s.Pages[0].Lastmod
|
|
}
|
|
return n
|
|
}
|
|
|
|
func (s *Site) renderSitemap() error {
|
|
if viper.GetBool("DisableSitemap") {
|
|
return nil
|
|
}
|
|
|
|
sitemapDefault := parseSitemap(viper.GetStringMap("Sitemap"))
|
|
|
|
n := s.newNode("sitemap")
|
|
|
|
// Prepend homepage to the list of pages
|
|
pages := make(Pages, 0)
|
|
|
|
page := &Page{}
|
|
page.language = s.Language
|
|
page.Date = s.Info.LastChange
|
|
page.Lastmod = s.Info.LastChange
|
|
page.Site = &s.Info
|
|
page.URLPath.URL = ""
|
|
page.Sitemap.ChangeFreq = sitemapDefault.ChangeFreq
|
|
page.Sitemap.Priority = sitemapDefault.Priority
|
|
|
|
pages = append(pages, page)
|
|
pages = append(pages, s.Pages...)
|
|
|
|
n.Data["Pages"] = pages
|
|
|
|
for _, page := range pages {
|
|
if page.Sitemap.ChangeFreq == "" {
|
|
page.Sitemap.ChangeFreq = sitemapDefault.ChangeFreq
|
|
}
|
|
|
|
if page.Sitemap.Priority == -1 {
|
|
page.Sitemap.Priority = sitemapDefault.Priority
|
|
}
|
|
|
|
if page.Sitemap.Filename == "" {
|
|
page.Sitemap.Filename = sitemapDefault.Filename
|
|
}
|
|
}
|
|
|
|
smLayouts := []string{"sitemap.xml", "_default/sitemap.xml", "_internal/_default/sitemap.xml"}
|
|
addLanguagePrefix := n.Site.IsMultiLingual()
|
|
if err := s.renderAndWriteXML("sitemap", n.addLangPathPrefixIfFlagSet(page.Sitemap.Filename, addLanguagePrefix), n, s.appendThemeTemplates(smLayouts)...); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) renderRobotsTXT() error {
|
|
if !viper.GetBool("EnableRobotsTXT") {
|
|
return nil
|
|
}
|
|
|
|
n := s.newNode("robots")
|
|
n.Data["Pages"] = s.Pages
|
|
|
|
rLayouts := []string{"robots.txt", "_default/robots.txt", "_internal/_default/robots.txt"}
|
|
outBuffer := bp.GetBuffer()
|
|
defer bp.PutBuffer(outBuffer)
|
|
err := s.renderForLayouts("robots", n, outBuffer, s.appendThemeTemplates(rLayouts)...)
|
|
|
|
if err == nil {
|
|
err = s.writeDestFile("robots.txt", outBuffer)
|
|
}
|
|
|
|
return err
|
|
}
|
|
|
|
// Stats prints Hugo builds stats to the console.
|
|
// This is what you see after a successful hugo build.
|
|
func (s *Site) Stats() {
|
|
jww.FEEDBACK.Printf("Built site for language %s:\n", s.Language.Lang)
|
|
jww.FEEDBACK.Println(s.draftStats())
|
|
jww.FEEDBACK.Println(s.futureStats())
|
|
jww.FEEDBACK.Println(s.expiredStats())
|
|
jww.FEEDBACK.Printf("%d pages created\n", len(s.Pages))
|
|
jww.FEEDBACK.Printf("%d non-page files copied\n", len(s.Files))
|
|
jww.FEEDBACK.Printf("%d paginator pages created\n", s.Info.paginationPageCount)
|
|
taxonomies := s.Language.GetStringMapString("Taxonomies")
|
|
|
|
for _, pl := range taxonomies {
|
|
jww.FEEDBACK.Printf("%d %s created\n", len(s.Taxonomies[pl]), pl)
|
|
}
|
|
|
|
}
|
|
|
|
func (s *Site) setURLs(n *Node, in string) {
|
|
n.URLPath.URL = helpers.URLizeAndPrep(in)
|
|
n.URLPath.Permalink = permalink(n.URLPath.URL)
|
|
n.RSSLink = template.HTML(permalink(in + ".xml"))
|
|
}
|
|
|
|
func permalink(plink string) string {
|
|
return permalinkStr(plink)
|
|
}
|
|
|
|
func permalinkStr(plink string) string {
|
|
return helpers.MakePermalink(viper.GetString("BaseURL"), helpers.URLizeAndPrep(plink)).String()
|
|
}
|
|
func (s *Site) newNode(nodeID string) *Node {
|
|
return s.nodeLookup(nodeID, 0, true)
|
|
}
|
|
|
|
func (s *Site) getNode(nodeID string) *Node {
|
|
return s.getOrAddNode(nodeID, false)
|
|
}
|
|
|
|
func (s *Site) getOrAddNode(nodeID string, add bool) *Node {
|
|
s.nodeCacheInit.Do(func() {
|
|
s.nodeCache = &nodeCache{m: make(map[string]*Node)}
|
|
})
|
|
|
|
s.nodeCache.RLock()
|
|
if n, ok := s.nodeCache.m[nodeID]; ok {
|
|
s.nodeCache.RUnlock()
|
|
if !add {
|
|
return n
|
|
}
|
|
panic(fmt.Sprintf("Node with ID %q in use", nodeID))
|
|
}
|
|
|
|
s.nodeCache.RUnlock()
|
|
s.nodeCache.Lock()
|
|
|
|
if !add {
|
|
// this is a test type error, print the keys
|
|
for k, _ := range s.nodeCache.m {
|
|
fmt.Println("Node:", k)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Double check
|
|
if _, ok := s.nodeCache.m[nodeID]; ok {
|
|
s.nodeCache.Unlock()
|
|
panic(fmt.Sprintf("Node with ID %q in use", nodeID))
|
|
}
|
|
|
|
n := &Node{
|
|
nodeID: nodeID,
|
|
Data: make(map[string]interface{}),
|
|
Site: &s.Info,
|
|
language: s.Language,
|
|
}
|
|
|
|
s.nodeCache.m[nodeID] = n
|
|
s.nodeCache.Unlock()
|
|
return n
|
|
}
|
|
|
|
func (s *Site) nodeLookup(nodeIDPrefix string, counter int, add bool) *Node {
|
|
|
|
nodeID := fmt.Sprintf("%s-%d", nodeIDPrefix, counter)
|
|
|
|
n := s.getOrAddNode(nodeID, add)
|
|
|
|
// Paginator nodes (counter > 0) gets created during rendering and cannot take part in any
|
|
// global translations mapping
|
|
if add && s.owner != nil && counter == 0 {
|
|
s.owner.addNode(nodeID, n)
|
|
}
|
|
|
|
return n
|
|
}
|
|
|
|
func (s *Site) layoutExists(layouts ...string) bool {
|
|
_, found := s.findFirstLayout(layouts...)
|
|
|
|
return found
|
|
}
|
|
|
|
func (s *Site) renderAndWriteXML(name string, dest string, d interface{}, layouts ...string) error {
|
|
renderBuffer := bp.GetBuffer()
|
|
defer bp.PutBuffer(renderBuffer)
|
|
renderBuffer.WriteString("<?xml version=\"1.0\" encoding=\"utf-8\" standalone=\"yes\" ?>\n")
|
|
|
|
err := s.renderForLayouts(name, d, renderBuffer, layouts...)
|
|
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
outBuffer := bp.GetBuffer()
|
|
defer bp.PutBuffer(outBuffer)
|
|
|
|
var path []byte
|
|
if viper.GetBool("RelativeURLs") {
|
|
path = []byte(helpers.GetDottedRelativePath(dest))
|
|
} else {
|
|
s := viper.GetString("BaseURL")
|
|
if !strings.HasSuffix(s, "/") {
|
|
s += "/"
|
|
}
|
|
path = []byte(s)
|
|
}
|
|
transformer := transform.NewChain(transform.AbsURLInXML)
|
|
transformer.Apply(outBuffer, renderBuffer, path)
|
|
|
|
return s.writeDestFile(dest, outBuffer)
|
|
|
|
}
|
|
|
|
func (s *Site) renderAndWritePage(name string, dest string, d interface{}, layouts ...string) error {
|
|
renderBuffer := bp.GetBuffer()
|
|
defer bp.PutBuffer(renderBuffer)
|
|
|
|
err := s.renderForLayouts(name, d, renderBuffer, layouts...)
|
|
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
outBuffer := bp.GetBuffer()
|
|
defer bp.PutBuffer(outBuffer)
|
|
|
|
var pageTarget target.Output
|
|
|
|
if p, ok := d.(*Page); ok && path.Ext(p.URLPath.URL) != "" {
|
|
// user has explicitly set a URL with extension for this page
|
|
// make sure it sticks even if "ugly URLs" are turned off.
|
|
pageTarget = s.pageUglyTarget()
|
|
} else {
|
|
pageTarget = s.pageTarget()
|
|
}
|
|
|
|
transformLinks := transform.NewEmptyTransforms()
|
|
|
|
if viper.GetBool("RelativeURLs") || viper.GetBool("CanonifyURLs") {
|
|
transformLinks = append(transformLinks, transform.AbsURL)
|
|
}
|
|
|
|
if s.running() && viper.GetBool("watch") && !viper.GetBool("DisableLiveReload") {
|
|
transformLinks = append(transformLinks, transform.LiveReloadInject)
|
|
}
|
|
|
|
// For performance reasons we only inject the Hugo generator tag on the home page.
|
|
if n, ok := d.(*Node); ok && n.IsHome {
|
|
if !viper.GetBool("DisableHugoGeneratorInject") {
|
|
transformLinks = append(transformLinks, transform.HugoGeneratorInject)
|
|
}
|
|
}
|
|
|
|
var path []byte
|
|
|
|
if viper.GetBool("RelativeURLs") {
|
|
translated, err := pageTarget.(target.OptionalTranslator).TranslateRelative(dest)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
path = []byte(helpers.GetDottedRelativePath(translated))
|
|
} else if viper.GetBool("CanonifyURLs") {
|
|
s := viper.GetString("BaseURL")
|
|
if !strings.HasSuffix(s, "/") {
|
|
s += "/"
|
|
}
|
|
path = []byte(s)
|
|
}
|
|
|
|
transformer := transform.NewChain(transformLinks...)
|
|
transformer.Apply(outBuffer, renderBuffer, path)
|
|
|
|
if outBuffer.Len() == 0 {
|
|
|
|
jww.WARN.Printf("%q is rendered empty\n", dest)
|
|
if dest == "/" {
|
|
debugAddend := ""
|
|
if !viper.GetBool("Verbose") {
|
|
debugAddend = "* For more debugging information, run \"hugo -v\""
|
|
}
|
|
distinctFeedbackLogger.Printf(`=============================================================
|
|
Your rendered home page is blank: /index.html is zero-length
|
|
* Did you specify a theme on the command-line or in your
|
|
%q file? (Current theme: %q)
|
|
%s
|
|
=============================================================`,
|
|
filepath.Base(viper.ConfigFileUsed()),
|
|
viper.GetString("Theme"),
|
|
debugAddend)
|
|
}
|
|
|
|
}
|
|
|
|
if err == nil {
|
|
|
|
if err = s.writeDestPage(dest, pageTarget, outBuffer); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return err
|
|
}
|
|
|
|
func (s *Site) renderForLayouts(name string, d interface{}, w io.Writer, layouts ...string) error {
|
|
layout, found := s.findFirstLayout(layouts...)
|
|
if found == false {
|
|
jww.WARN.Printf("Unable to locate layout for %s: %s\n", name, layouts)
|
|
return nil
|
|
}
|
|
|
|
if err := s.renderThing(d, layout, w); err != nil {
|
|
|
|
// Behavior here should be dependent on if running in server or watch mode.
|
|
distinctErrorLogger.Printf("Error while rendering %s: %v", name, err)
|
|
if !s.running() && !testMode {
|
|
// TODO(bep) check if this can be propagated
|
|
os.Exit(-1)
|
|
} else if testMode {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (s *Site) findFirstLayout(layouts ...string) (string, bool) {
|
|
for _, layout := range layouts {
|
|
if s.owner.tmpl.Lookup(layout) != nil {
|
|
return layout, true
|
|
}
|
|
}
|
|
return "", false
|
|
}
|
|
|
|
func (s *Site) renderThing(d interface{}, layout string, w io.Writer) error {
|
|
|
|
// If the template doesn't exist, then return, but leave the Writer open
|
|
if templ := s.owner.tmpl.Lookup(layout); templ != nil {
|
|
return templ.Execute(w, d)
|
|
}
|
|
return fmt.Errorf("Layout not found: %s", layout)
|
|
|
|
}
|
|
|
|
func (s *Site) pageTarget() target.Output {
|
|
s.initTargetList()
|
|
return s.targets.page
|
|
}
|
|
|
|
func (s *Site) pageUglyTarget() target.Output {
|
|
s.initTargetList()
|
|
return s.targets.pageUgly
|
|
}
|
|
|
|
func (s *Site) fileTarget() target.Output {
|
|
s.initTargetList()
|
|
return s.targets.file
|
|
}
|
|
|
|
func (s *Site) aliasTarget() target.AliasPublisher {
|
|
s.initTargetList()
|
|
return s.targets.alias
|
|
}
|
|
|
|
func (s *Site) languageAliasTarget() target.AliasPublisher {
|
|
s.initTargetList()
|
|
return s.targets.languageAlias
|
|
}
|
|
|
|
func (s *Site) initTargetList() {
|
|
s.targetListInit.Do(func() {
|
|
if s.targets.page == nil {
|
|
s.targets.page = &target.PagePub{
|
|
PublishDir: s.absPublishDir(),
|
|
UglyURLs: viper.GetBool("UglyURLs"),
|
|
}
|
|
}
|
|
if s.targets.pageUgly == nil {
|
|
s.targets.pageUgly = &target.PagePub{
|
|
PublishDir: s.absPublishDir(),
|
|
UglyURLs: true,
|
|
}
|
|
}
|
|
if s.targets.file == nil {
|
|
s.targets.file = &target.Filesystem{
|
|
PublishDir: s.absPublishDir(),
|
|
}
|
|
}
|
|
if s.targets.alias == nil {
|
|
s.targets.alias = &target.HTMLRedirectAlias{
|
|
PublishDir: s.absPublishDir(),
|
|
}
|
|
}
|
|
if s.targets.languageAlias == nil {
|
|
s.targets.languageAlias = &target.HTMLRedirectAlias{
|
|
PublishDir: s.absPublishDir(),
|
|
AllowRoot: true,
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
func (s *Site) writeDestFile(path string, reader io.Reader) (err error) {
|
|
jww.DEBUG.Println("creating file:", path)
|
|
return s.fileTarget().Publish(path, reader)
|
|
}
|
|
|
|
func (s *Site) writeDestPage(path string, publisher target.Publisher, reader io.Reader) (err error) {
|
|
jww.DEBUG.Println("creating page:", path)
|
|
return publisher.Publish(path, reader)
|
|
}
|
|
|
|
// AliasPublisher
|
|
func (s *Site) writeDestAlias(path string, permalink string) (err error) {
|
|
return s.publishDestAlias(s.aliasTarget(), path, permalink)
|
|
}
|
|
|
|
func (s *Site) publishDestAlias(aliasPublisher target.AliasPublisher, path string, permalink string) (err error) {
|
|
if viper.GetBool("RelativeURLs") {
|
|
// convert `permalink` into URI relative to location of `path`
|
|
baseURL := helpers.SanitizeURLKeepTrailingSlash(viper.GetString("BaseURL"))
|
|
if strings.HasPrefix(permalink, baseURL) {
|
|
permalink = "/" + strings.TrimPrefix(permalink, baseURL)
|
|
}
|
|
permalink, err = helpers.GetRelativePath(permalink, path)
|
|
if err != nil {
|
|
jww.ERROR.Println("Failed to make a RelativeURL alias:", path, "redirecting to", permalink)
|
|
}
|
|
permalink = filepath.ToSlash(permalink)
|
|
}
|
|
jww.DEBUG.Println("creating alias:", path, "redirecting to", permalink)
|
|
return aliasPublisher.Publish(path, permalink)
|
|
}
|
|
|
|
func (s *Site) draftStats() string {
|
|
var msg string
|
|
|
|
switch s.draftCount {
|
|
case 0:
|
|
return "0 draft content"
|
|
case 1:
|
|
msg = "1 draft rendered"
|
|
default:
|
|
msg = fmt.Sprintf("%d drafts rendered", s.draftCount)
|
|
}
|
|
|
|
if viper.GetBool("BuildDrafts") {
|
|
return fmt.Sprintf("%d of ", s.draftCount) + msg
|
|
}
|
|
|
|
return "0 of " + msg
|
|
}
|
|
|
|
func (s *Site) futureStats() string {
|
|
var msg string
|
|
|
|
switch s.futureCount {
|
|
case 0:
|
|
return "0 future content"
|
|
case 1:
|
|
msg = "1 future rendered"
|
|
default:
|
|
msg = fmt.Sprintf("%d futures rendered", s.futureCount)
|
|
}
|
|
|
|
if viper.GetBool("BuildFuture") {
|
|
return fmt.Sprintf("%d of ", s.futureCount) + msg
|
|
}
|
|
|
|
return "0 of " + msg
|
|
}
|
|
|
|
func (s *Site) expiredStats() string {
|
|
var msg string
|
|
|
|
switch s.expiredCount {
|
|
case 0:
|
|
return "0 expired content"
|
|
case 1:
|
|
msg = "1 expired rendered"
|
|
default:
|
|
msg = fmt.Sprintf("%d expired rendered", s.expiredCount)
|
|
}
|
|
|
|
if viper.GetBool("BuildExpired") {
|
|
return fmt.Sprintf("%d of ", s.expiredCount) + msg
|
|
}
|
|
|
|
return "0 of " + msg
|
|
}
|
|
|
|
func getGoMaxProcs() int {
|
|
if gmp := os.Getenv("GOMAXPROCS"); gmp != "" {
|
|
if p, err := strconv.Atoi(gmp); err != nil {
|
|
return p
|
|
}
|
|
}
|
|
return 1
|
|
}
|