mirror of
https://github.com/gohugoio/hugo.git
synced 2025-05-06 10:46:34 +03:00
Big refactor of how source files are used. Also added default destination extension option.
This commit is contained in:
parent
24bbfe7d32
commit
5dfc1dedb8
24 changed files with 646 additions and 465 deletions
|
@ -99,20 +99,20 @@ func convertContents(mark rune) (err error) {
|
||||||
|
|
||||||
jww.FEEDBACK.Println("processing", len(site.Source.Files()), "content files")
|
jww.FEEDBACK.Println("processing", len(site.Source.Files()), "content files")
|
||||||
for _, file := range site.Source.Files() {
|
for _, file := range site.Source.Files() {
|
||||||
jww.INFO.Println("Attempting to convert", file.LogicalName)
|
jww.INFO.Println("Attempting to convert", file.LogicalName())
|
||||||
page, err := hugolib.NewPage(file.LogicalName)
|
page, err := hugolib.NewPage(file.LogicalName())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
psr, err := parser.ReadFrom(file.Contents)
|
psr, err := parser.ReadFrom(file.Contents)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
jww.ERROR.Println("Error processing file:", path.Join(file.Dir, file.LogicalName))
|
jww.ERROR.Println("Error processing file:", file.Path())
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
metadata, err := psr.Metadata()
|
metadata, err := psr.Metadata()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
jww.ERROR.Println("Error processing file:", path.Join(file.Dir, file.LogicalName))
|
jww.ERROR.Println("Error processing file:", file.Path())
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -128,7 +128,7 @@ func convertContents(mark rune) (err error) {
|
||||||
metadata = newmetadata
|
metadata = newmetadata
|
||||||
}
|
}
|
||||||
|
|
||||||
page.Dir = file.Dir
|
//page.Dir = file.Dir
|
||||||
page.SetSourceContent(psr.Content())
|
page.SetSourceContent(psr.Content())
|
||||||
page.SetSourceMetaData(metadata, mark)
|
page.SetSourceMetaData(metadata, mark)
|
||||||
|
|
||||||
|
|
|
@ -118,6 +118,7 @@ func InitializeConfig() {
|
||||||
viper.SetDefault("Permalinks", make(hugolib.PermalinkOverrides, 0))
|
viper.SetDefault("Permalinks", make(hugolib.PermalinkOverrides, 0))
|
||||||
viper.SetDefault("Sitemap", hugolib.Sitemap{Priority: -1})
|
viper.SetDefault("Sitemap", hugolib.Sitemap{Priority: -1})
|
||||||
viper.SetDefault("PygmentsStyle", "monokai")
|
viper.SetDefault("PygmentsStyle", "monokai")
|
||||||
|
viper.SetDefault("DefaultExtension", "html")
|
||||||
viper.SetDefault("PygmentsUseClasses", false)
|
viper.SetDefault("PygmentsUseClasses", false)
|
||||||
viper.SetDefault("DisableLiveReload", false)
|
viper.SetDefault("DisableLiveReload", false)
|
||||||
viper.SetDefault("PluralizeListTitles", true)
|
viper.SetDefault("PluralizeListTitles", true)
|
||||||
|
|
|
@ -94,7 +94,7 @@ func NewContent(kind, name string) (err error) {
|
||||||
newmetadata["date"] = time.Now().Format(time.RFC3339)
|
newmetadata["date"] = time.Now().Format(time.RFC3339)
|
||||||
}
|
}
|
||||||
|
|
||||||
page.Dir = viper.GetString("sourceDir")
|
//page.Dir = viper.GetString("sourceDir")
|
||||||
page.SetSourceMetaData(newmetadata, parser.FormatToLeadRune(viper.GetString("MetaDataFormat")))
|
page.SetSourceMetaData(newmetadata, parser.FormatToLeadRune(viper.GetString("MetaDataFormat")))
|
||||||
page.SetSourceContent(psr.Content())
|
page.SetSourceContent(psr.Content())
|
||||||
if err = page.SafeSaveSourceAs(path.Join(viper.GetString("contentDir"), name)); err != nil {
|
if err = page.SafeSaveSourceAs(path.Join(viper.GetString("contentDir"), name)); err != nil {
|
||||||
|
|
|
@ -31,7 +31,7 @@
|
||||||
{{end}}
|
{{end}}
|
||||||
<li> <a href="https://github.com/spf13/hugo/issues" target="blank"><i class='fa fa-life-ring'></i>Issues & Help</a> </li>
|
<li> <a href="https://github.com/spf13/hugo/issues" target="blank"><i class='fa fa-life-ring'></i>Issues & Help</a> </li>
|
||||||
{{ if .IsPage }}
|
{{ if .IsPage }}
|
||||||
{{ $File := .File }} {{with $File.FileName }}<li><a href="https://github.com/spf13/hugo/edit/master/docs/content/{{ $File.Dir }}{{ $File.FileName }}" target="blank"><i class='fa fa-edit'></i> Refine this Page</a> </li>{{end}}
|
{{ $File := .File }} {{with $File.Path }}<li><a href="https://github.com/spf13/hugo/edit/master/docs/content/{{ $File.Dir }}{{ $File.LogicalName }}" target="blank"><i class='fa fa-edit'></i> Refine this Page</a> </li>{{end}}
|
||||||
{{ end }}
|
{{ end }}
|
||||||
</ul>
|
</ul>
|
||||||
<!-- sidebar menu end-->
|
<!-- sidebar menu end-->
|
||||||
|
|
231
helpers/content.go
Normal file
231
helpers/content.go
Normal file
|
@ -0,0 +1,231 @@
|
||||||
|
// Copyright © 2014 Steve Francia <spf@spf13.com>.
|
||||||
|
//
|
||||||
|
// Licensed under the Simple Public License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
// http://opensource.org/licenses/Simple-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package helpers
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"html/template"
|
||||||
|
"os/exec"
|
||||||
|
|
||||||
|
"github.com/russross/blackfriday"
|
||||||
|
"github.com/spf13/viper"
|
||||||
|
|
||||||
|
jww "github.com/spf13/jwalterweatherman"
|
||||||
|
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
var SummaryLength = 70
|
||||||
|
var SummaryDivider = []byte("<!--more-->")
|
||||||
|
|
||||||
|
func StripHTML(s string) string {
|
||||||
|
output := ""
|
||||||
|
|
||||||
|
// Shortcut strings with no tags in them
|
||||||
|
if !strings.ContainsAny(s, "<>") {
|
||||||
|
output = s
|
||||||
|
} else {
|
||||||
|
s = strings.Replace(s, "\n", " ", -1)
|
||||||
|
s = strings.Replace(s, "</p>", " \n", -1)
|
||||||
|
s = strings.Replace(s, "<br>", " \n", -1)
|
||||||
|
s = strings.Replace(s, "</br>", " \n", -1)
|
||||||
|
|
||||||
|
// Walk through the string removing all tags
|
||||||
|
b := new(bytes.Buffer)
|
||||||
|
inTag := false
|
||||||
|
for _, r := range s {
|
||||||
|
switch r {
|
||||||
|
case '<':
|
||||||
|
inTag = true
|
||||||
|
case '>':
|
||||||
|
inTag = false
|
||||||
|
default:
|
||||||
|
if !inTag {
|
||||||
|
b.WriteRune(r)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
output = b.String()
|
||||||
|
}
|
||||||
|
return output
|
||||||
|
}
|
||||||
|
|
||||||
|
func StripEmptyNav(in []byte) []byte {
|
||||||
|
return bytes.Replace(in, []byte("<nav>\n</nav>\n\n"), []byte(``), -1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func BytesToHTML(b []byte) template.HTML {
|
||||||
|
return template.HTML(string(b))
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetHtmlRenderer(defaultFlags int, footnoteref string) blackfriday.Renderer {
|
||||||
|
renderParameters := blackfriday.HtmlRendererParameters{
|
||||||
|
FootnoteAnchorPrefix: viper.GetString("FootnoteAnchorPrefix"),
|
||||||
|
FootnoteReturnLinkContents: viper.GetString("FootnoteReturnLinkContents"),
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(footnoteref) != 0 {
|
||||||
|
renderParameters.FootnoteAnchorPrefix = footnoteref + ":" +
|
||||||
|
renderParameters.FootnoteAnchorPrefix
|
||||||
|
}
|
||||||
|
|
||||||
|
htmlFlags := defaultFlags
|
||||||
|
htmlFlags |= blackfriday.HTML_USE_XHTML
|
||||||
|
htmlFlags |= blackfriday.HTML_USE_SMARTYPANTS
|
||||||
|
htmlFlags |= blackfriday.HTML_SMARTYPANTS_FRACTIONS
|
||||||
|
htmlFlags |= blackfriday.HTML_SMARTYPANTS_LATEX_DASHES
|
||||||
|
htmlFlags |= blackfriday.HTML_FOOTNOTE_RETURN_LINKS
|
||||||
|
|
||||||
|
return blackfriday.HtmlRendererWithParameters(htmlFlags, "", "", renderParameters)
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetMarkdownExtensions() int {
|
||||||
|
return 0 | blackfriday.EXTENSION_NO_INTRA_EMPHASIS |
|
||||||
|
blackfriday.EXTENSION_TABLES | blackfriday.EXTENSION_FENCED_CODE |
|
||||||
|
blackfriday.EXTENSION_AUTOLINK | blackfriday.EXTENSION_STRIKETHROUGH |
|
||||||
|
blackfriday.EXTENSION_SPACE_HEADERS | blackfriday.EXTENSION_FOOTNOTES |
|
||||||
|
blackfriday.EXTENSION_HEADER_IDS
|
||||||
|
}
|
||||||
|
|
||||||
|
func MarkdownRender(content []byte, footnoteref string) []byte {
|
||||||
|
return blackfriday.Markdown(content, GetHtmlRenderer(0, footnoteref),
|
||||||
|
GetMarkdownExtensions())
|
||||||
|
}
|
||||||
|
|
||||||
|
func MarkdownRenderWithTOC(content []byte, footnoteref string) []byte {
|
||||||
|
return blackfriday.Markdown(content,
|
||||||
|
GetHtmlRenderer(blackfriday.HTML_TOC, footnoteref),
|
||||||
|
GetMarkdownExtensions())
|
||||||
|
}
|
||||||
|
|
||||||
|
func ExtractTOC(content []byte) (newcontent []byte, toc []byte) {
|
||||||
|
origContent := make([]byte, len(content))
|
||||||
|
copy(origContent, content)
|
||||||
|
first := []byte(`<nav>
|
||||||
|
<ul>`)
|
||||||
|
|
||||||
|
last := []byte(`</ul>
|
||||||
|
</nav>`)
|
||||||
|
|
||||||
|
replacement := []byte(`<nav id="TableOfContents">
|
||||||
|
<ul>`)
|
||||||
|
|
||||||
|
startOfTOC := bytes.Index(content, first)
|
||||||
|
|
||||||
|
peekEnd := len(content)
|
||||||
|
if peekEnd > 70+startOfTOC {
|
||||||
|
peekEnd = 70 + startOfTOC
|
||||||
|
}
|
||||||
|
|
||||||
|
if startOfTOC < 0 {
|
||||||
|
return StripEmptyNav(content), toc
|
||||||
|
}
|
||||||
|
// Need to peek ahead to see if this nav element is actually the right one.
|
||||||
|
correctNav := bytes.Index(content[startOfTOC:peekEnd], []byte(`#toc_0`))
|
||||||
|
if correctNav < 0 { // no match found
|
||||||
|
return content, toc
|
||||||
|
}
|
||||||
|
lengthOfTOC := bytes.Index(content[startOfTOC:], last) + len(last)
|
||||||
|
endOfTOC := startOfTOC + lengthOfTOC
|
||||||
|
|
||||||
|
newcontent = append(content[:startOfTOC], content[endOfTOC:]...)
|
||||||
|
toc = append(replacement, origContent[startOfTOC+len(first):endOfTOC]...)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
func RenderBytesWithTOC(content []byte, pagefmt string, footnoteref string) []byte {
|
||||||
|
switch pagefmt {
|
||||||
|
default:
|
||||||
|
return MarkdownRenderWithTOC(content, footnoteref)
|
||||||
|
case "markdown":
|
||||||
|
return MarkdownRenderWithTOC(content, footnoteref)
|
||||||
|
case "rst":
|
||||||
|
return []byte(GetRstContent(content))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func RenderBytes(content []byte, pagefmt string, footnoteref string) []byte {
|
||||||
|
switch pagefmt {
|
||||||
|
default:
|
||||||
|
return MarkdownRender(content, footnoteref)
|
||||||
|
case "markdown":
|
||||||
|
return MarkdownRender(content, footnoteref)
|
||||||
|
case "rst":
|
||||||
|
return []byte(GetRstContent(content))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TotalWords(s string) int {
|
||||||
|
return len(strings.Fields(s))
|
||||||
|
}
|
||||||
|
|
||||||
|
func WordCount(s string) map[string]int {
|
||||||
|
m := make(map[string]int)
|
||||||
|
for _, f := range strings.Fields(s) {
|
||||||
|
m[f] += 1
|
||||||
|
}
|
||||||
|
|
||||||
|
return m
|
||||||
|
}
|
||||||
|
|
||||||
|
func RemoveSummaryDivider(content []byte) []byte {
|
||||||
|
return bytes.Replace(content, SummaryDivider, []byte(""), -1)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TruncateWords(s string, max int) string {
|
||||||
|
words := strings.Fields(s)
|
||||||
|
if max > len(words) {
|
||||||
|
return strings.Join(words, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(words[:max], " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
func TruncateWordsToWholeSentence(s string, max int) string {
|
||||||
|
words := strings.Fields(s)
|
||||||
|
if max > len(words) {
|
||||||
|
return strings.Join(words, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
for counter, word := range words[max:] {
|
||||||
|
if strings.HasSuffix(word, ".") ||
|
||||||
|
strings.HasSuffix(word, "?") ||
|
||||||
|
strings.HasSuffix(word, ".\"") ||
|
||||||
|
strings.HasSuffix(word, "!") {
|
||||||
|
return strings.Join(words[:max+counter+1], " ")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return strings.Join(words[:max], " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetRstContent(content []byte) string {
|
||||||
|
cleanContent := bytes.Replace(content, SummaryDivider, []byte(""), 1)
|
||||||
|
|
||||||
|
cmd := exec.Command("rst2html.py", "--leave-comments")
|
||||||
|
cmd.Stdin = bytes.NewReader(cleanContent)
|
||||||
|
var out bytes.Buffer
|
||||||
|
cmd.Stdout = &out
|
||||||
|
if err := cmd.Run(); err != nil {
|
||||||
|
jww.ERROR.Println(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
rstLines := strings.Split(out.String(), "\n")
|
||||||
|
for i, line := range rstLines {
|
||||||
|
if strings.HasPrefix(line, "<body>") {
|
||||||
|
rstLines = (rstLines[i+1 : len(rstLines)-3])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return strings.Join(rstLines, "\n")
|
||||||
|
}
|
|
@ -15,43 +15,14 @@ package helpers
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
|
"crypto/md5"
|
||||||
|
"encoding/hex"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"net"
|
"net"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
func StripHTML(s string) string {
|
|
||||||
output := ""
|
|
||||||
|
|
||||||
// Shortcut strings with no tags in them
|
|
||||||
if !strings.ContainsAny(s, "<>") {
|
|
||||||
output = s
|
|
||||||
} else {
|
|
||||||
s = strings.Replace(s, "\n", " ", -1)
|
|
||||||
s = strings.Replace(s, "</p>", " \n", -1)
|
|
||||||
s = strings.Replace(s, "<br>", " \n", -1)
|
|
||||||
s = strings.Replace(s, "</br>", " \n", -1)
|
|
||||||
|
|
||||||
// Walk through the string removing all tags
|
|
||||||
b := new(bytes.Buffer)
|
|
||||||
inTag := false
|
|
||||||
for _, r := range s {
|
|
||||||
switch r {
|
|
||||||
case '<':
|
|
||||||
inTag = true
|
|
||||||
case '>':
|
|
||||||
inTag = false
|
|
||||||
default:
|
|
||||||
if !inTag {
|
|
||||||
b.WriteRune(r)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
output = b.String()
|
|
||||||
}
|
|
||||||
return output
|
|
||||||
}
|
|
||||||
|
|
||||||
func FindAvailablePort() (*net.TCPAddr, error) {
|
func FindAvailablePort() (*net.TCPAddr, error) {
|
||||||
l, err := net.Listen("tcp", ":0")
|
l, err := net.Listen("tcp", ":0")
|
||||||
if err == nil {
|
if err == nil {
|
||||||
|
@ -64,3 +35,42 @@ func FindAvailablePort() (*net.TCPAddr, error) {
|
||||||
}
|
}
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func GuessType(in string) string {
|
||||||
|
switch strings.ToLower(in) {
|
||||||
|
case "md", "markdown", "mdown":
|
||||||
|
return "markdown"
|
||||||
|
case "rst":
|
||||||
|
return "rst"
|
||||||
|
case "html", "htm":
|
||||||
|
return "html"
|
||||||
|
}
|
||||||
|
|
||||||
|
return "unknown"
|
||||||
|
}
|
||||||
|
|
||||||
|
func ReaderToBytes(lines io.Reader) []byte {
|
||||||
|
b := new(bytes.Buffer)
|
||||||
|
b.ReadFrom(lines)
|
||||||
|
return b.Bytes()
|
||||||
|
}
|
||||||
|
|
||||||
|
// sliceToLower goes through the source slice and lowers all values.
|
||||||
|
func SliceToLower(s []string) []string {
|
||||||
|
if s == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
l := make([]string, len(s))
|
||||||
|
for i, v := range s {
|
||||||
|
l[i] = strings.ToLower(v)
|
||||||
|
}
|
||||||
|
|
||||||
|
return l
|
||||||
|
}
|
||||||
|
|
||||||
|
func Md5String(f string) string {
|
||||||
|
h := md5.New()
|
||||||
|
h.Write([]byte(f))
|
||||||
|
return hex.EncodeToString(h.Sum([]byte{}))
|
||||||
|
}
|
||||||
|
|
|
@ -166,16 +166,52 @@ func FileAndExt(in string) (name string, ext string) {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func GetRelativePath(path, base string) (final string, err error) {
|
||||||
|
if filepath.IsAbs(path) && base == "" {
|
||||||
|
return "", errors.New("source: missing base directory")
|
||||||
|
}
|
||||||
|
name := filepath.Clean(path)
|
||||||
|
base = filepath.Clean(base)
|
||||||
|
|
||||||
|
name, err = filepath.Rel(base, name)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
name = filepath.ToSlash(name)
|
||||||
|
return name, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Given a source path, determine the section
|
||||||
func GuessSection(in string) string {
|
func GuessSection(in string) string {
|
||||||
x := strings.Split(in, "/")
|
parts := strings.Split(in, "/")
|
||||||
x = x[:len(x)-1]
|
|
||||||
if len(x) == 0 {
|
if len(parts) == 0 {
|
||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
if x[0] == "content" {
|
|
||||||
x = x[1:]
|
// trim filename
|
||||||
|
if !strings.HasSuffix(in, "/") {
|
||||||
|
parts = parts[:len(parts)-1]
|
||||||
}
|
}
|
||||||
return path.Join(x...)
|
|
||||||
|
if len(parts) == 0 {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
// if first directory is "content", return second directory
|
||||||
|
section := ""
|
||||||
|
|
||||||
|
if parts[0] == "content" && len(parts) > 1 {
|
||||||
|
section = parts[1]
|
||||||
|
} else {
|
||||||
|
section = parts[0]
|
||||||
|
}
|
||||||
|
|
||||||
|
if section == "." {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
return section
|
||||||
}
|
}
|
||||||
|
|
||||||
func PathPrep(ugly bool, in string) string {
|
func PathPrep(ugly bool, in string) string {
|
||||||
|
|
284
hugolib/page.go
284
hugolib/page.go
|
@ -15,56 +15,63 @@ package hugolib
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"crypto/md5"
|
|
||||||
"encoding/hex"
|
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"html/template"
|
"html/template"
|
||||||
"io"
|
"io"
|
||||||
"net/url"
|
"net/url"
|
||||||
"path"
|
"path"
|
||||||
"path/filepath"
|
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/russross/blackfriday"
|
|
||||||
"github.com/spf13/cast"
|
"github.com/spf13/cast"
|
||||||
"github.com/spf13/hugo/helpers"
|
"github.com/spf13/hugo/helpers"
|
||||||
"github.com/spf13/hugo/parser"
|
"github.com/spf13/hugo/parser"
|
||||||
|
"github.com/spf13/hugo/source"
|
||||||
jww "github.com/spf13/jwalterweatherman"
|
jww "github.com/spf13/jwalterweatherman"
|
||||||
"github.com/spf13/viper"
|
"github.com/spf13/viper"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Page struct {
|
type Page struct {
|
||||||
Status string
|
Params map[string]interface{}
|
||||||
Images []string
|
|
||||||
rawContent []byte
|
|
||||||
Content template.HTML
|
Content template.HTML
|
||||||
Summary template.HTML
|
Summary template.HTML
|
||||||
|
Aliases []string
|
||||||
|
Status string
|
||||||
|
Images []string
|
||||||
TableOfContents template.HTML
|
TableOfContents template.HTML
|
||||||
Truncated bool
|
Truncated bool
|
||||||
plain string // TODO should be []byte
|
|
||||||
Params map[string]interface{}
|
|
||||||
contentType string
|
|
||||||
Draft bool
|
Draft bool
|
||||||
PublishDate time.Time
|
PublishDate time.Time
|
||||||
Aliases []string
|
|
||||||
Tmpl Template
|
Tmpl Template
|
||||||
Markup string
|
Markup string
|
||||||
|
|
||||||
|
extension string
|
||||||
|
contentType string
|
||||||
renderable bool
|
renderable bool
|
||||||
layout string
|
layout string
|
||||||
linkTitle string
|
linkTitle string
|
||||||
frontmatter []byte
|
frontmatter []byte
|
||||||
sourceFrontmatter []byte
|
rawContent []byte
|
||||||
sourceContent []byte
|
plain string // TODO should be []byte
|
||||||
|
//sourceFrontmatter []byte
|
||||||
|
//sourceContent []byte
|
||||||
PageMeta
|
PageMeta
|
||||||
File
|
//SourceFile source.File
|
||||||
|
Source
|
||||||
Position
|
Position
|
||||||
Node
|
Node
|
||||||
|
//Destination source.File
|
||||||
}
|
}
|
||||||
|
|
||||||
type File struct {
|
//type File struct {
|
||||||
Name, FileName, Extension, Dir, UniqueId string
|
//Name, FileName, Extension, Dir, UniqueId string
|
||||||
|
//}
|
||||||
|
|
||||||
|
type Source struct {
|
||||||
|
Frontmatter []byte
|
||||||
|
Content []byte
|
||||||
|
source.File
|
||||||
}
|
}
|
||||||
|
|
||||||
type PageMeta struct {
|
type PageMeta struct {
|
||||||
|
@ -97,75 +104,40 @@ func (p *Page) IsPage() bool {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) UniqueId() string {
|
func (p *Page) UniqueId() string {
|
||||||
return p.File.UniqueId
|
return p.Source.UniqueId()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) setSummary() {
|
func (p *Page) setSummary() {
|
||||||
if bytes.Contains(p.rawContent, summaryDivider) {
|
if bytes.Contains(p.rawContent, helpers.SummaryDivider) {
|
||||||
// If user defines split:
|
// If user defines split:
|
||||||
// Split then render
|
// Split then render
|
||||||
p.Truncated = true // by definition
|
p.Truncated = true // by definition
|
||||||
header := bytes.Split(p.rawContent, summaryDivider)[0]
|
header := bytes.Split(p.rawContent, helpers.SummaryDivider)[0]
|
||||||
p.Summary = bytesToHTML(p.renderBytes(header))
|
p.Summary = helpers.BytesToHTML(p.renderBytes(header))
|
||||||
} else {
|
} else {
|
||||||
// If hugo defines split:
|
// If hugo defines split:
|
||||||
// render, strip html, then split
|
// render, strip html, then split
|
||||||
plain := strings.TrimSpace(p.Plain())
|
plain := strings.TrimSpace(p.Plain())
|
||||||
p.Summary = bytesToHTML([]byte(TruncateWordsToWholeSentence(plain, summaryLength)))
|
p.Summary = helpers.BytesToHTML([]byte(helpers.TruncateWordsToWholeSentence(plain, helpers.SummaryLength)))
|
||||||
p.Truncated = len(p.Summary) != len(plain)
|
p.Truncated = len(p.Summary) != len(plain)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func stripEmptyNav(in []byte) []byte {
|
|
||||||
return bytes.Replace(in, []byte("<nav>\n</nav>\n\n"), []byte(``), -1)
|
|
||||||
}
|
|
||||||
|
|
||||||
func bytesToHTML(b []byte) template.HTML {
|
|
||||||
return template.HTML(string(b))
|
|
||||||
}
|
|
||||||
|
|
||||||
func (p *Page) renderBytes(content []byte) []byte {
|
func (p *Page) renderBytes(content []byte) []byte {
|
||||||
return renderBytes(content, p.guessMarkupType(), p.UniqueId())
|
return helpers.RenderBytes(content, p.guessMarkupType(), p.UniqueId())
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) renderContent(content []byte) []byte {
|
func (p *Page) renderContent(content []byte) []byte {
|
||||||
return renderBytesWithTOC(content, p.guessMarkupType(), p.UniqueId())
|
return helpers.RenderBytesWithTOC(content, p.guessMarkupType(), p.UniqueId())
|
||||||
}
|
|
||||||
|
|
||||||
func renderBytesWithTOC(content []byte, pagefmt string, footnoteref string) []byte {
|
|
||||||
switch pagefmt {
|
|
||||||
default:
|
|
||||||
return markdownRenderWithTOC(content, footnoteref)
|
|
||||||
case "markdown":
|
|
||||||
return markdownRenderWithTOC(content, footnoteref)
|
|
||||||
case "rst":
|
|
||||||
return []byte(getRstContent(content))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func renderBytes(content []byte, pagefmt string, footnoteref string) []byte {
|
|
||||||
switch pagefmt {
|
|
||||||
default:
|
|
||||||
return markdownRender(content, footnoteref)
|
|
||||||
case "markdown":
|
|
||||||
return markdownRender(content, footnoteref)
|
|
||||||
case "rst":
|
|
||||||
return []byte(getRstContent(content))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func newPage(filename string) *Page {
|
func newPage(filename string) *Page {
|
||||||
name := filepath.Base(filename)
|
|
||||||
// strip off the extension
|
|
||||||
name = name[:len(name)-len(filepath.Ext(name))]
|
|
||||||
|
|
||||||
page := Page{contentType: "",
|
page := Page{contentType: "",
|
||||||
File: File{Name: name, FileName: filename, Extension: "html", UniqueId: md5ForFilename(filename)},
|
Source: Source{File: *source.NewFile(filename)},
|
||||||
Node: Node{Keywords: []string{}, Sitemap: Sitemap{Priority: -1}},
|
Node: Node{Keywords: []string{}, Sitemap: Sitemap{Priority: -1}},
|
||||||
Params: make(map[string]interface{})}
|
Params: make(map[string]interface{})}
|
||||||
|
|
||||||
jww.DEBUG.Println("Reading from", page.File.FileName)
|
jww.DEBUG.Println("Reading from", page.File.Path())
|
||||||
page.guessSection()
|
|
||||||
return &page
|
return &page
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -173,24 +145,22 @@ func (p *Page) IsRenderable() bool {
|
||||||
return p.renderable
|
return p.renderable
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) guessSection() {
|
|
||||||
if p.Section == "" {
|
|
||||||
p.Section = helpers.GuessSection(p.FileName)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (page *Page) Type() string {
|
func (page *Page) Type() string {
|
||||||
if page.contentType != "" {
|
if page.contentType != "" {
|
||||||
return page.contentType
|
return page.contentType
|
||||||
}
|
}
|
||||||
page.guessSection()
|
|
||||||
if x := page.Section; x != "" {
|
if x := page.Section(); x != "" {
|
||||||
return x
|
return x
|
||||||
}
|
}
|
||||||
|
|
||||||
return "page"
|
return "page"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (page *Page) Section() string {
|
||||||
|
return page.Source.Section()
|
||||||
|
}
|
||||||
|
|
||||||
func (page *Page) Layout(l ...string) []string {
|
func (page *Page) Layout(l ...string) []string {
|
||||||
if page.layout != "" {
|
if page.layout != "" {
|
||||||
return layouts(page.Type(), page.layout)
|
return layouts(page.Type(), page.layout)
|
||||||
|
@ -261,14 +231,14 @@ func (p *Page) ReadFrom(buf io.Reader) (err error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) analyzePage() {
|
func (p *Page) analyzePage() {
|
||||||
p.WordCount = TotalWords(p.Plain())
|
p.WordCount = helpers.TotalWords(p.Plain())
|
||||||
p.FuzzyWordCount = int((p.WordCount+100)/100) * 100
|
p.FuzzyWordCount = int((p.WordCount+100)/100) * 100
|
||||||
p.ReadingTime = int((p.WordCount + 212) / 213)
|
p.ReadingTime = int((p.WordCount + 212) / 213)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) permalink() (*url.URL, error) {
|
func (p *Page) permalink() (*url.URL, error) {
|
||||||
baseUrl := string(p.Site.BaseUrl)
|
baseUrl := string(p.Site.BaseUrl)
|
||||||
dir := strings.TrimSpace(p.Dir)
|
dir := strings.TrimSpace(p.Source.Dir())
|
||||||
pSlug := strings.TrimSpace(p.Slug)
|
pSlug := strings.TrimSpace(p.Slug)
|
||||||
pUrl := strings.TrimSpace(p.Url)
|
pUrl := strings.TrimSpace(p.Url)
|
||||||
var permalink string
|
var permalink string
|
||||||
|
@ -278,7 +248,7 @@ func (p *Page) permalink() (*url.URL, error) {
|
||||||
return helpers.MakePermalink(baseUrl, pUrl), nil
|
return helpers.MakePermalink(baseUrl, pUrl), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
if override, ok := p.Site.Permalinks[p.Section]; ok {
|
if override, ok := p.Site.Permalinks[p.Section()]; ok {
|
||||||
permalink, err = override.Expand(p)
|
permalink, err = override.Expand(p)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -287,16 +257,24 @@ func (p *Page) permalink() (*url.URL, error) {
|
||||||
// fmt.Printf("have a section override for %q in section %s → %s\n", p.Title, p.Section, permalink)
|
// fmt.Printf("have a section override for %q in section %s → %s\n", p.Title, p.Section, permalink)
|
||||||
} else {
|
} else {
|
||||||
if len(pSlug) > 0 {
|
if len(pSlug) > 0 {
|
||||||
permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, p.Slug+"."+p.Extension))
|
permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, p.Slug+"."+p.Extension()))
|
||||||
} else {
|
} else {
|
||||||
_, t := path.Split(p.FileName)
|
_, t := path.Split(p.Source.LogicalName())
|
||||||
permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension)))
|
permalink = helpers.UrlPrep(viper.GetBool("UglyUrls"), path.Join(dir, helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension())))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return helpers.MakePermalink(baseUrl, permalink), nil
|
return helpers.MakePermalink(baseUrl, permalink), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p *Page) Extension() string {
|
||||||
|
if p.extension != "" {
|
||||||
|
return p.extension
|
||||||
|
} else {
|
||||||
|
return viper.GetString("DefaultExtension")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (p *Page) LinkTitle() string {
|
func (p *Page) LinkTitle() string {
|
||||||
if len(p.linkTitle) > 0 {
|
if len(p.linkTitle) > 0 {
|
||||||
return p.linkTitle
|
return p.linkTitle
|
||||||
|
@ -370,6 +348,8 @@ func (page *Page) update(f interface{}) error {
|
||||||
page.Url = helpers.Urlize(cast.ToString(v))
|
page.Url = helpers.Urlize(cast.ToString(v))
|
||||||
case "type":
|
case "type":
|
||||||
page.contentType = cast.ToString(v)
|
page.contentType = cast.ToString(v)
|
||||||
|
case "extension", "ext":
|
||||||
|
page.extension = cast.ToString(v)
|
||||||
case "keywords":
|
case "keywords":
|
||||||
page.Keywords = cast.ToStringSlice(v)
|
page.Keywords = cast.ToStringSlice(v)
|
||||||
case "date":
|
case "date":
|
||||||
|
@ -445,7 +425,7 @@ func (page *Page) GetParam(key string) interface{} {
|
||||||
case time.Time:
|
case time.Time:
|
||||||
return cast.ToTime(v)
|
return cast.ToTime(v)
|
||||||
case []string:
|
case []string:
|
||||||
return sliceToLower(v.([]string))
|
return helpers.SliceToLower(v.([]string))
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -543,31 +523,13 @@ func (p *Page) Render(layout ...string) template.HTML {
|
||||||
func (page *Page) guessMarkupType() string {
|
func (page *Page) guessMarkupType() string {
|
||||||
// First try the explicitly set markup from the frontmatter
|
// First try the explicitly set markup from the frontmatter
|
||||||
if page.Markup != "" {
|
if page.Markup != "" {
|
||||||
format := guessType(page.Markup)
|
format := helpers.GuessType(page.Markup)
|
||||||
if format != "unknown" {
|
if format != "unknown" {
|
||||||
return format
|
return format
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Then try to guess from the extension
|
return helpers.GuessType(page.Source.Ext())
|
||||||
ext := strings.ToLower(path.Ext(page.FileName))
|
|
||||||
if strings.HasPrefix(ext, ".") {
|
|
||||||
return guessType(ext[1:])
|
|
||||||
}
|
|
||||||
|
|
||||||
return "unknown"
|
|
||||||
}
|
|
||||||
|
|
||||||
func guessType(in string) string {
|
|
||||||
switch strings.ToLower(in) {
|
|
||||||
case "md", "markdown", "mdown":
|
|
||||||
return "markdown"
|
|
||||||
case "rst":
|
|
||||||
return "rst"
|
|
||||||
case "html", "htm":
|
|
||||||
return "html"
|
|
||||||
}
|
|
||||||
return "unknown"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (page *Page) detectFrontMatter() (f *parser.FrontmatterType) {
|
func (page *Page) detectFrontMatter() (f *parser.FrontmatterType) {
|
||||||
|
@ -585,7 +547,7 @@ func (page *Page) parse(reader io.Reader) error {
|
||||||
meta, err := psr.Metadata()
|
meta, err := psr.Metadata()
|
||||||
if meta != nil {
|
if meta != nil {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
jww.ERROR.Printf("Error parsing page meta data for %s", page.FileName)
|
jww.ERROR.Printf("Error parsing page meta data for %s", page.File.Path())
|
||||||
jww.ERROR.Println(err)
|
jww.ERROR.Println(err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -601,7 +563,7 @@ func (page *Page) parse(reader io.Reader) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (page *Page) SetSourceContent(content []byte) {
|
func (page *Page) SetSourceContent(content []byte) {
|
||||||
page.sourceContent = content
|
page.Source.Content = content
|
||||||
}
|
}
|
||||||
|
|
||||||
func (page *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
|
func (page *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
|
||||||
|
@ -611,7 +573,7 @@ func (page *Page) SetSourceMetaData(in interface{}, mark rune) (err error) {
|
||||||
}
|
}
|
||||||
by = append(by, '\n')
|
by = append(by, '\n')
|
||||||
|
|
||||||
page.sourceFrontmatter = by
|
page.Source.Frontmatter = by
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -626,8 +588,8 @@ func (page *Page) SaveSourceAs(path string) error {
|
||||||
|
|
||||||
func (page *Page) saveSourceAs(path string, safe bool) error {
|
func (page *Page) saveSourceAs(path string, safe bool) error {
|
||||||
b := new(bytes.Buffer)
|
b := new(bytes.Buffer)
|
||||||
b.Write(page.sourceFrontmatter)
|
b.Write(page.Source.Frontmatter)
|
||||||
b.Write(page.sourceContent)
|
b.Write(page.Source.Content)
|
||||||
|
|
||||||
err := page.saveSource(b.Bytes(), path, safe)
|
err := page.saveSource(b.Bytes(), path, safe)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -666,100 +628,19 @@ func (page *Page) Convert() error {
|
||||||
markupType := page.guessMarkupType()
|
markupType := page.guessMarkupType()
|
||||||
switch markupType {
|
switch markupType {
|
||||||
case "markdown", "rst":
|
case "markdown", "rst":
|
||||||
tmpContent, tmpTableOfContents := extractTOC(page.renderContent(RemoveSummaryDivider(page.rawContent)))
|
tmpContent, tmpTableOfContents := helpers.ExtractTOC(page.renderContent(helpers.RemoveSummaryDivider(page.rawContent)))
|
||||||
page.Content = bytesToHTML(tmpContent)
|
page.Content = helpers.BytesToHTML(tmpContent)
|
||||||
page.TableOfContents = bytesToHTML(tmpTableOfContents)
|
page.TableOfContents = helpers.BytesToHTML(tmpTableOfContents)
|
||||||
case "html":
|
case "html":
|
||||||
page.Content = bytesToHTML(page.rawContent)
|
page.Content = helpers.BytesToHTML(page.rawContent)
|
||||||
default:
|
default:
|
||||||
return fmt.Errorf("Error converting unsupported file type '%s' for page '%s'", markupType, page.FileName)
|
return fmt.Errorf("Error converting unsupported file type '%s' for page '%s'", markupType, page.Source.Path())
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func getHtmlRenderer(defaultFlags int, footnoteref string) blackfriday.Renderer {
|
|
||||||
renderParameters := blackfriday.HtmlRendererParameters{
|
|
||||||
FootnoteAnchorPrefix: viper.GetString("FootnoteAnchorPrefix"),
|
|
||||||
FootnoteReturnLinkContents: viper.GetString("FootnoteReturnLinkContents"),
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(footnoteref) != 0 {
|
|
||||||
renderParameters.FootnoteAnchorPrefix = footnoteref + ":" +
|
|
||||||
renderParameters.FootnoteAnchorPrefix
|
|
||||||
}
|
|
||||||
|
|
||||||
htmlFlags := defaultFlags
|
|
||||||
htmlFlags |= blackfriday.HTML_USE_XHTML
|
|
||||||
htmlFlags |= blackfriday.HTML_USE_SMARTYPANTS
|
|
||||||
htmlFlags |= blackfriday.HTML_SMARTYPANTS_FRACTIONS
|
|
||||||
htmlFlags |= blackfriday.HTML_SMARTYPANTS_LATEX_DASHES
|
|
||||||
htmlFlags |= blackfriday.HTML_FOOTNOTE_RETURN_LINKS
|
|
||||||
|
|
||||||
return blackfriday.HtmlRendererWithParameters(htmlFlags, "", "", renderParameters)
|
|
||||||
}
|
|
||||||
|
|
||||||
func getMarkdownExtensions() int {
|
|
||||||
return 0 | blackfriday.EXTENSION_NO_INTRA_EMPHASIS |
|
|
||||||
blackfriday.EXTENSION_TABLES | blackfriday.EXTENSION_FENCED_CODE |
|
|
||||||
blackfriday.EXTENSION_AUTOLINK | blackfriday.EXTENSION_STRIKETHROUGH |
|
|
||||||
blackfriday.EXTENSION_SPACE_HEADERS | blackfriday.EXTENSION_FOOTNOTES |
|
|
||||||
blackfriday.EXTENSION_HEADER_IDS
|
|
||||||
}
|
|
||||||
|
|
||||||
func markdownRender(content []byte, footnoteref string) []byte {
|
|
||||||
return blackfriday.Markdown(content, getHtmlRenderer(0, footnoteref),
|
|
||||||
getMarkdownExtensions())
|
|
||||||
}
|
|
||||||
|
|
||||||
func markdownRenderWithTOC(content []byte, footnoteref string) []byte {
|
|
||||||
return blackfriday.Markdown(content,
|
|
||||||
getHtmlRenderer(blackfriday.HTML_TOC, footnoteref),
|
|
||||||
getMarkdownExtensions())
|
|
||||||
}
|
|
||||||
|
|
||||||
func extractTOC(content []byte) (newcontent []byte, toc []byte) {
|
|
||||||
origContent := make([]byte, len(content))
|
|
||||||
copy(origContent, content)
|
|
||||||
first := []byte(`<nav>
|
|
||||||
<ul>`)
|
|
||||||
|
|
||||||
last := []byte(`</ul>
|
|
||||||
</nav>`)
|
|
||||||
|
|
||||||
replacement := []byte(`<nav id="TableOfContents">
|
|
||||||
<ul>`)
|
|
||||||
|
|
||||||
startOfTOC := bytes.Index(content, first)
|
|
||||||
|
|
||||||
peekEnd := len(content)
|
|
||||||
if peekEnd > 70+startOfTOC {
|
|
||||||
peekEnd = 70 + startOfTOC
|
|
||||||
}
|
|
||||||
|
|
||||||
if startOfTOC < 0 {
|
|
||||||
return stripEmptyNav(content), toc
|
|
||||||
}
|
|
||||||
// Need to peek ahead to see if this nav element is actually the right one.
|
|
||||||
correctNav := bytes.Index(content[startOfTOC:peekEnd], []byte(`#toc_0`))
|
|
||||||
if correctNav < 0 { // no match found
|
|
||||||
return content, toc
|
|
||||||
}
|
|
||||||
lengthOfTOC := bytes.Index(content[startOfTOC:], last) + len(last)
|
|
||||||
endOfTOC := startOfTOC + lengthOfTOC
|
|
||||||
|
|
||||||
newcontent = append(content[:startOfTOC], content[endOfTOC:]...)
|
|
||||||
toc = append(replacement, origContent[startOfTOC+len(first):endOfTOC]...)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func ReaderToBytes(lines io.Reader) []byte {
|
|
||||||
b := new(bytes.Buffer)
|
|
||||||
b.ReadFrom(lines)
|
|
||||||
return b.Bytes()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (p *Page) FullFilePath() string {
|
func (p *Page) FullFilePath() string {
|
||||||
return path.Join(p.Dir, p.FileName)
|
return path.Join(p.Source.Dir(), p.Source.Path())
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *Page) TargetPath() (outfile string) {
|
func (p *Page) TargetPath() (outfile string) {
|
||||||
|
@ -775,7 +656,7 @@ func (p *Page) TargetPath() (outfile string) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// If there's a Permalink specification, we use that
|
// If there's a Permalink specification, we use that
|
||||||
if override, ok := p.Site.Permalinks[p.Section]; ok {
|
if override, ok := p.Site.Permalinks[p.Section()]; ok {
|
||||||
var err error
|
var err error
|
||||||
outfile, err = override.Expand(p)
|
outfile, err = override.Expand(p)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
|
@ -787,32 +668,11 @@ func (p *Page) TargetPath() (outfile string) {
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(strings.TrimSpace(p.Slug)) > 0 {
|
if len(strings.TrimSpace(p.Slug)) > 0 {
|
||||||
outfile = strings.TrimSpace(p.Slug) + "." + p.Extension
|
outfile = strings.TrimSpace(p.Slug) + "." + p.Extension()
|
||||||
} else {
|
} else {
|
||||||
// Fall back to filename
|
// Fall back to filename
|
||||||
_, t := path.Split(p.FileName)
|
outfile = helpers.ReplaceExtension(p.Source.LogicalName(), p.Extension())
|
||||||
outfile = helpers.ReplaceExtension(strings.TrimSpace(t), p.Extension)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return path.Join(p.Dir, strings.TrimSpace(outfile))
|
return path.Join(p.Source.Dir(), strings.TrimSpace(outfile))
|
||||||
}
|
|
||||||
|
|
||||||
// sliceToLower goes through the source slice and lowers all values.
|
|
||||||
func sliceToLower(s []string) []string {
|
|
||||||
if s == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
l := make([]string, len(s))
|
|
||||||
for i, v := range s {
|
|
||||||
l[i] = strings.ToLower(v)
|
|
||||||
}
|
|
||||||
|
|
||||||
return l
|
|
||||||
}
|
|
||||||
|
|
||||||
func md5ForFilename(f string) string {
|
|
||||||
h := md5.New()
|
|
||||||
h.Write([]byte(f))
|
|
||||||
return hex.EncodeToString(h.Sum([]byte{}))
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -83,21 +83,26 @@ func (p Pages) GroupBy(key string, order ...string) (PagesGroup, error) {
|
||||||
direction = "desc"
|
direction = "desc"
|
||||||
}
|
}
|
||||||
|
|
||||||
ppt := reflect.TypeOf(&Page{})
|
ppt := reflect.TypeOf(&Page{}) // *hugolib.Page
|
||||||
|
|
||||||
ft, ok := ppt.Elem().FieldByName(key)
|
ft, ok := ppt.Elem().FieldByName(key)
|
||||||
|
|
||||||
if !ok {
|
if !ok {
|
||||||
return nil, errors.New("No such field in Page struct")
|
return nil, errors.New("No such field in Page struct")
|
||||||
}
|
}
|
||||||
|
|
||||||
tmp := reflect.MakeMap(reflect.MapOf(ft.Type, reflect.SliceOf(ppt)))
|
tmp := reflect.MakeMap(reflect.MapOf(ft.Type, reflect.SliceOf(ppt)))
|
||||||
|
|
||||||
for _, e := range p {
|
for _, e := range p {
|
||||||
ppv := reflect.ValueOf(e)
|
ppv := reflect.ValueOf(e)
|
||||||
fv := ppv.Elem().FieldByName(key)
|
fv := ppv.Elem().FieldByName(key)
|
||||||
|
if !fv.IsNil() {
|
||||||
if !tmp.MapIndex(fv).IsValid() {
|
if !tmp.MapIndex(fv).IsValid() {
|
||||||
tmp.SetMapIndex(fv, reflect.MakeSlice(reflect.SliceOf(ppt), 0, 0))
|
tmp.SetMapIndex(fv, reflect.MakeSlice(reflect.SliceOf(ppt), 0, 0))
|
||||||
}
|
}
|
||||||
tmp.SetMapIndex(fv, reflect.Append(tmp.MapIndex(fv), ppv))
|
tmp.SetMapIndex(fv, reflect.Append(tmp.MapIndex(fv), ppv))
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
var r []PageGroup
|
var r []PageGroup
|
||||||
for _, k := range sortKeys(tmp.MapKeys(), direction) {
|
for _, k := range sortKeys(tmp.MapKeys(), direction) {
|
||||||
|
|
|
@ -4,6 +4,7 @@ import (
|
||||||
"html/template"
|
"html/template"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
"github.com/spf13/hugo/source"
|
||||||
"github.com/spf13/viper"
|
"github.com/spf13/viper"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -33,6 +34,8 @@ func TestPermalink(t *testing.T) {
|
||||||
{"x/y/z/boofar.md", "x/y/z", "", "", "/z/y/q/", false, "/z/y/q/", "/z/y/q/"},
|
{"x/y/z/boofar.md", "x/y/z", "", "", "/z/y/q/", false, "/z/y/q/", "/z/y/q/"},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
viper.Set("DefaultExtension", "html")
|
||||||
|
|
||||||
for i, test := range tests {
|
for i, test := range tests {
|
||||||
viper.Set("uglyurls", test.uglyurls)
|
viper.Set("uglyurls", test.uglyurls)
|
||||||
p := &Page{
|
p := &Page{
|
||||||
|
@ -45,7 +48,7 @@ func TestPermalink(t *testing.T) {
|
||||||
BaseUrl: test.base,
|
BaseUrl: test.base,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
File: File{FileName: test.file, Dir: test.dir, Extension: "html"},
|
Source: Source{File: *source.NewFile(test.file)},
|
||||||
}
|
}
|
||||||
|
|
||||||
if test.slug != "" {
|
if test.slug != "" {
|
||||||
|
|
|
@ -6,6 +6,8 @@ import (
|
||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/spf13/hugo/helpers"
|
||||||
)
|
)
|
||||||
|
|
||||||
var EMPTY_PAGE = ""
|
var EMPTY_PAGE = ""
|
||||||
|
@ -507,7 +509,7 @@ func TestDegenerateInvalidFrontMatterLeadingWhitespace(t *testing.T) {
|
||||||
func TestSectionEvaluation(t *testing.T) {
|
func TestSectionEvaluation(t *testing.T) {
|
||||||
page, _ := NewPage("blue/file1.md")
|
page, _ := NewPage("blue/file1.md")
|
||||||
page.ReadFrom(strings.NewReader(SIMPLE_PAGE))
|
page.ReadFrom(strings.NewReader(SIMPLE_PAGE))
|
||||||
if page.Section != "blue" {
|
if page.Section() != "blue" {
|
||||||
t.Errorf("Section should be %s, got: %s", "blue", page.Section)
|
t.Errorf("Section should be %s, got: %s", "blue", page.Section)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -529,12 +531,12 @@ func TestLayoutOverride(t *testing.T) {
|
||||||
path string
|
path string
|
||||||
expectedLayout []string
|
expectedLayout []string
|
||||||
}{
|
}{
|
||||||
{SIMPLE_PAGE_NOLAYOUT, path_content_two_dir, L("dub/sub/single.html", "dub/single.html", "_default/single.html")},
|
{SIMPLE_PAGE_NOLAYOUT, path_content_two_dir, L("dub/single.html", "_default/single.html")},
|
||||||
{SIMPLE_PAGE_NOLAYOUT, path_content_one_dir, L("gub/single.html", "_default/single.html")},
|
{SIMPLE_PAGE_NOLAYOUT, path_content_one_dir, L("gub/single.html", "_default/single.html")},
|
||||||
{SIMPLE_PAGE_NOLAYOUT, path_content_no_dir, L("page/single.html", "_default/single.html")},
|
{SIMPLE_PAGE_NOLAYOUT, path_content_no_dir, L("page/single.html", "_default/single.html")},
|
||||||
{SIMPLE_PAGE_NOLAYOUT, path_one_directory, L("fub/single.html", "_default/single.html")},
|
{SIMPLE_PAGE_NOLAYOUT, path_one_directory, L("fub/single.html", "_default/single.html")},
|
||||||
{SIMPLE_PAGE_NOLAYOUT, path_no_directory, L("page/single.html", "_default/single.html")},
|
{SIMPLE_PAGE_NOLAYOUT, path_no_directory, L("page/single.html", "_default/single.html")},
|
||||||
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_two_dir, L("dub/sub/foobar.html", "dub/foobar.html", "_default/foobar.html")},
|
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_two_dir, L("dub/foobar.html", "_default/foobar.html")},
|
||||||
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_one_dir, L("gub/foobar.html", "_default/foobar.html")},
|
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_content_one_dir, L("gub/foobar.html", "_default/foobar.html")},
|
||||||
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_one_directory, L("fub/foobar.html", "_default/foobar.html")},
|
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_one_directory, L("fub/foobar.html", "_default/foobar.html")},
|
||||||
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_no_directory, L("page/foobar.html", "_default/foobar.html")},
|
{SIMPLE_PAGE_LAYOUT_FOOBAR, path_no_directory, L("page/foobar.html", "_default/foobar.html")},
|
||||||
|
@ -576,7 +578,7 @@ func TestSliceToLower(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, test := range tests {
|
for _, test := range tests {
|
||||||
res := sliceToLower(test.value)
|
res := helpers.SliceToLower(test.value)
|
||||||
for i, val := range res {
|
for i, val := range res {
|
||||||
if val != test.expected[i] {
|
if val != test.expected[i] {
|
||||||
t.Errorf("Case mismatch. Expected %s, got %s", test.expected[i], res[i])
|
t.Errorf("Case mismatch. Expected %s, got %s", test.expected[i], res[i])
|
||||||
|
|
|
@ -17,7 +17,7 @@ func TestDegenerateMissingFolderInPageFilename(t *testing.T) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Error in NewPageFrom")
|
t.Fatalf("Error in NewPageFrom")
|
||||||
}
|
}
|
||||||
if p.Section != "" {
|
if p.Section() != "" {
|
||||||
t.Fatalf("No section should be set for a file path: foobar")
|
t.Fatalf("No section should be set for a file path: foobar")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -31,17 +31,16 @@ func TestNewPageWithFilePath(t *testing.T) {
|
||||||
{path.Join("sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
|
{path.Join("sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
|
||||||
{path.Join("content", "foobar.html"), "", L("page/single.html", "_default/single.html")},
|
{path.Join("content", "foobar.html"), "", L("page/single.html", "_default/single.html")},
|
||||||
{path.Join("content", "sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
|
{path.Join("content", "sub", "foobar.html"), "sub", L("sub/single.html", "_default/single.html")},
|
||||||
{path.Join("content", "dub", "sub", "foobar.html"), "dub/sub", L("dub/sub/single.html", "dub/single.html", "_default/single.html")},
|
{path.Join("content", "dub", "sub", "foobar.html"), "dub", L("dub/single.html", "_default/single.html")},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, el := range toCheck {
|
for _, el := range toCheck {
|
||||||
p, err := NewPageFrom(strings.NewReader(SIMPLE_PAGE_YAML), el.input)
|
p, err := NewPageFrom(strings.NewReader(SIMPLE_PAGE_YAML), el.input)
|
||||||
p.guessSection()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Errorf("Reading from SIMPLE_PAGE_YAML resulted in an error: %s", err)
|
t.Errorf("Reading from SIMPLE_PAGE_YAML resulted in an error: %s", err)
|
||||||
}
|
}
|
||||||
if p.Section != el.section {
|
if p.Section() != el.section {
|
||||||
t.Errorf("Section not set to %s for page %s. Got: %s", el.section, el.input, p.Section)
|
t.Errorf("Section not set to %s for page %s. Got: %s", el.section, el.input, p.Section())
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, y := range el.layout {
|
for _, y := range el.layout {
|
||||||
|
|
|
@ -3,7 +3,6 @@ package hugolib
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"path/filepath"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
@ -120,9 +119,9 @@ func pageToPermalinkTitle(p *Page, _ string) (string, error) {
|
||||||
|
|
||||||
// pageToPermalinkFilename returns the URL-safe form of the filename
|
// pageToPermalinkFilename returns the URL-safe form of the filename
|
||||||
func pageToPermalinkFilename(p *Page, _ string) (string, error) {
|
func pageToPermalinkFilename(p *Page, _ string) (string, error) {
|
||||||
var extension = filepath.Ext(p.FileName)
|
//var extension = p.Source.Ext
|
||||||
var name = p.FileName[0 : len(p.FileName)-len(extension)]
|
//var name = p.Source.Path()[0 : len(p.Source.Path())-len(extension)]
|
||||||
return helpers.Urlize(name), nil
|
return helpers.Urlize(p.Source.BaseFileName()), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// if the page has a slug, return the slug, else return the title
|
// if the page has a slug, return the slug, else return the title
|
||||||
|
@ -143,7 +142,7 @@ func pageToPermalinkSlugElseTitle(p *Page, a string) (string, error) {
|
||||||
|
|
||||||
func pageToPermalinkSection(p *Page, _ string) (string, error) {
|
func pageToPermalinkSection(p *Page, _ string) (string, error) {
|
||||||
// Page contains Node contains UrlPath which has Section
|
// Page contains Node contains UrlPath which has Section
|
||||||
return p.Section, nil
|
return p.Section(), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
|
|
@ -11,7 +11,7 @@ func (s *Site) ShowPlan(out io.Writer) (err error) {
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, p := range s.Pages {
|
for _, p := range s.Pages {
|
||||||
fmt.Fprintf(out, "%s", p.FileName)
|
fmt.Fprintf(out, "%s", p.Source.Path())
|
||||||
if p.IsRenderable() {
|
if p.IsRenderable() {
|
||||||
fmt.Fprintf(out, " (renderer: markdown)")
|
fmt.Fprintf(out, " (renderer: markdown)")
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -20,6 +20,7 @@ import (
|
||||||
"strings"
|
"strings"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
|
||||||
|
"github.com/spf13/hugo/helpers"
|
||||||
jww "github.com/spf13/jwalterweatherman"
|
jww "github.com/spf13/jwalterweatherman"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -93,7 +94,7 @@ func ShortcodesHandle(stringToParse string, p *Page, t Template) string {
|
||||||
var data = &ShortcodeWithPage{Params: params, Page: p}
|
var data = &ShortcodeWithPage{Params: params, Page: p}
|
||||||
if endStart > 0 {
|
if endStart > 0 {
|
||||||
s := stringToParse[leadEnd+3 : leadEnd+endStart]
|
s := stringToParse[leadEnd+3 : leadEnd+endStart]
|
||||||
data.Inner = template.HTML(renderBytes([]byte(CleanP(ShortcodesHandle(s, p, t))), p.guessMarkupType(), p.UniqueId()))
|
data.Inner = template.HTML(helpers.RenderBytes([]byte(CleanP(ShortcodesHandle(s, p, t))), p.guessMarkupType(), p.UniqueId()))
|
||||||
remainder := CleanP(stringToParse[leadEnd+endEnd:])
|
remainder := CleanP(stringToParse[leadEnd+endEnd:])
|
||||||
|
|
||||||
return CleanP(stringToParse[:leadStart]) +
|
return CleanP(stringToParse[:leadStart]) +
|
||||||
|
|
|
@ -236,7 +236,7 @@ func (s *Site) Render() (err error) {
|
||||||
func (s *Site) checkDescriptions() {
|
func (s *Site) checkDescriptions() {
|
||||||
for _, p := range s.Pages {
|
for _, p := range s.Pages {
|
||||||
if len(p.Description) < 60 {
|
if len(p.Description) < 60 {
|
||||||
jww.FEEDBACK.Println(p.FileName + " ")
|
jww.FEEDBACK.Println(p.Source.Path() + " ")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -340,7 +340,7 @@ func (s *Site) CreatePages() error {
|
||||||
|
|
||||||
for i := 0; i < procs*4; i++ {
|
for i := 0; i < procs*4; i++ {
|
||||||
wg.Add(1)
|
wg.Add(1)
|
||||||
go pageReader(s, filechan, results, wg)
|
go sourceReader(s, filechan, results, wg)
|
||||||
}
|
}
|
||||||
|
|
||||||
errs := make(chan error)
|
errs := make(chan error)
|
||||||
|
@ -397,18 +397,16 @@ func (s *Site) CreatePages() error {
|
||||||
return fmt.Errorf("%s\n%s", readErrs, renderErrs)
|
return fmt.Errorf("%s\n%s", readErrs, renderErrs)
|
||||||
}
|
}
|
||||||
|
|
||||||
func pageReader(s *Site, files <-chan *source.File, results chan<- pageResult, wg *sync.WaitGroup) {
|
func sourceReader(s *Site, files <-chan *source.File, results chan<- pageResult, wg *sync.WaitGroup) {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
for file := range files {
|
for file := range files {
|
||||||
page, err := NewPage(file.LogicalName)
|
page, err := NewPage(file.Path())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
results <- pageResult{nil, err}
|
results <- pageResult{nil, err}
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
page.Site = &s.Info
|
page.Site = &s.Info
|
||||||
page.Tmpl = s.Tmpl
|
page.Tmpl = s.Tmpl
|
||||||
page.Section = file.Section
|
|
||||||
page.Dir = file.Dir
|
|
||||||
if err := page.ReadFrom(file.Contents); err != nil {
|
if err := page.ReadFrom(file.Contents); err != nil {
|
||||||
results <- pageResult{nil, err}
|
results <- pageResult{nil, err}
|
||||||
continue
|
continue
|
||||||
|
@ -604,7 +602,7 @@ func (s *Site) assembleTaxonomies() {
|
||||||
x := WeightedPage{weight.(int), p}
|
x := WeightedPage{weight.(int), p}
|
||||||
s.Taxonomies[plural].Add(v, x)
|
s.Taxonomies[plural].Add(v, x)
|
||||||
} else {
|
} else {
|
||||||
jww.ERROR.Printf("Invalid %s in %s\n", plural, p.File.FileName)
|
jww.ERROR.Printf("Invalid %s in %s\n", plural, p.File.Path())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -620,7 +618,7 @@ func (s *Site) assembleTaxonomies() {
|
||||||
|
|
||||||
func (s *Site) assembleSections() {
|
func (s *Site) assembleSections() {
|
||||||
for i, p := range s.Pages {
|
for i, p := range s.Pages {
|
||||||
s.Sections.Add(p.Section, WeightedPage{s.Pages[i].Weight, s.Pages[i]})
|
s.Sections.Add(p.Section(), WeightedPage{s.Pages[i].Weight, s.Pages[i]})
|
||||||
}
|
}
|
||||||
|
|
||||||
for k := range s.Sections {
|
for k := range s.Sections {
|
||||||
|
|
|
@ -47,13 +47,13 @@ func checkShowPlanExpected(t *testing.T, s *Site, expected string) {
|
||||||
|
|
||||||
for _, x := range gotList {
|
for _, x := range gotList {
|
||||||
if !stringInSlice(x, expectedList) {
|
if !stringInSlice(x, expectedList) {
|
||||||
t.Errorf("ShowPlan expected:\n%q\ngot\n%q", expected, got)
|
t.Errorf("\nShowPlan expected:\n%q\ngot:\n%q", expected, got)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, x := range expectedList {
|
for _, x := range expectedList {
|
||||||
if !stringInSlice(x, gotList) {
|
if !stringInSlice(x, gotList) {
|
||||||
t.Errorf("ShowPlan expected:\n%q\ngot\n%q", expected, got)
|
t.Errorf("\nShowPlan expected:\n%q\ngot:\n%q", expected, got)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -213,7 +213,7 @@ func TestTargetPath(t *testing.T) {
|
||||||
t.Errorf("%s => OutFile expected: '%s', got: '%s'", test.doc, expected, p.TargetPath())
|
t.Errorf("%s => OutFile expected: '%s', got: '%s'", test.doc, expected, p.TargetPath())
|
||||||
}
|
}
|
||||||
|
|
||||||
if p.Section != test.expectedSection {
|
if p.Section() != test.expectedSection {
|
||||||
t.Errorf("%s => p.Section expected: %s, got: %s", test.doc, test.expectedSection, p.Section)
|
t.Errorf("%s => p.Section expected: %s, got: %s", test.doc, test.expectedSection, p.Section)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -223,10 +223,10 @@ func TestDraftAndFutureRender(t *testing.T) {
|
||||||
files := make(map[string][]byte)
|
files := make(map[string][]byte)
|
||||||
target := &target.InMemoryTarget{Files: files}
|
target := &target.InMemoryTarget{Files: files}
|
||||||
sources := []source.ByteSource{
|
sources := []source.ByteSource{
|
||||||
{"sect/doc1.md", []byte("---\ntitle: doc1\ndraft: true\npublishdate: \"2414-05-29\"\n---\n# doc1\n*some content*"), "sect"},
|
{"sect/doc1.md", []byte("---\ntitle: doc1\ndraft: true\npublishdate: \"2414-05-29\"\n---\n# doc1\n*some content*")},
|
||||||
{"sect/doc2.md", []byte("---\ntitle: doc2\ndraft: true\npublishdate: \"2012-05-29\"\n---\n# doc2\n*some content*"), "sect"},
|
{"sect/doc2.md", []byte("---\ntitle: doc2\ndraft: true\npublishdate: \"2012-05-29\"\n---\n# doc2\n*some content*")},
|
||||||
{"sect/doc3.md", []byte("---\ntitle: doc3\ndraft: false\npublishdate: \"2414-05-29\"\n---\n# doc3\n*some content*"), "sect"},
|
{"sect/doc3.md", []byte("---\ntitle: doc3\ndraft: false\npublishdate: \"2414-05-29\"\n---\n# doc3\n*some content*")},
|
||||||
{"sect/doc4.md", []byte("---\ntitle: doc4\ndraft: false\npublishdate: \"2012-05-29\"\n---\n# doc4\n*some content*"), "sect"},
|
{"sect/doc4.md", []byte("---\ntitle: doc4\ndraft: false\npublishdate: \"2012-05-29\"\n---\n# doc4\n*some content*")},
|
||||||
}
|
}
|
||||||
|
|
||||||
siteSetup := func() *Site {
|
siteSetup := func() *Site {
|
||||||
|
@ -283,14 +283,14 @@ func TestSkipRender(t *testing.T) {
|
||||||
files := make(map[string][]byte)
|
files := make(map[string][]byte)
|
||||||
target := &target.InMemoryTarget{Files: files}
|
target := &target.InMemoryTarget{Files: files}
|
||||||
sources := []source.ByteSource{
|
sources := []source.ByteSource{
|
||||||
{"sect/doc1.html", []byte("---\nmarkup: markdown\n---\n# title\nsome *content*"), "sect"},
|
{"sect/doc1.html", []byte("---\nmarkup: markdown\n---\n# title\nsome *content*")},
|
||||||
{"sect/doc2.html", []byte("<!doctype html><html><body>more content</body></html>"), "sect"},
|
{"sect/doc2.html", []byte("<!doctype html><html><body>more content</body></html>")},
|
||||||
{"sect/doc3.md", []byte("# doc3\n*some* content"), "sect"},
|
{"sect/doc3.md", []byte("# doc3\n*some* content")},
|
||||||
{"sect/doc4.md", []byte("---\ntitle: doc4\n---\n# doc4\n*some content*"), "sect"},
|
{"sect/doc4.md", []byte("---\ntitle: doc4\n---\n# doc4\n*some content*")},
|
||||||
{"sect/doc5.html", []byte("<!doctype html><html>{{ template \"head\" }}<body>body5</body></html>"), "sect"},
|
{"sect/doc5.html", []byte("<!doctype html><html>{{ template \"head\" }}<body>body5</body></html>")},
|
||||||
{"sect/doc6.html", []byte("<!doctype html><html>{{ template \"head_abs\" }}<body>body5</body></html>"), "sect"},
|
{"sect/doc6.html", []byte("<!doctype html><html>{{ template \"head_abs\" }}<body>body5</body></html>")},
|
||||||
{"doc7.html", []byte("<html><body>doc7 content</body></html>"), ""},
|
{"doc7.html", []byte("<html><body>doc7 content</body></html>")},
|
||||||
{"sect/doc8.html", []byte("---\nmarkup: md\n---\n# title\nsome *content*"), "sect"},
|
{"sect/doc8.html", []byte("---\nmarkup: md\n---\n# title\nsome *content*")},
|
||||||
}
|
}
|
||||||
|
|
||||||
viper.Set("verbose", true)
|
viper.Set("verbose", true)
|
||||||
|
@ -350,8 +350,8 @@ func TestAbsUrlify(t *testing.T) {
|
||||||
files := make(map[string][]byte)
|
files := make(map[string][]byte)
|
||||||
target := &target.InMemoryTarget{Files: files}
|
target := &target.InMemoryTarget{Files: files}
|
||||||
sources := []source.ByteSource{
|
sources := []source.ByteSource{
|
||||||
{"sect/doc1.html", []byte("<!doctype html><html><head></head><body><a href=\"#frag1\">link</a></body></html>"), "sect"},
|
{"sect/doc1.html", []byte("<!doctype html><html><head></head><body><a href=\"#frag1\">link</a></body></html>")},
|
||||||
{"content/blue/doc2.html", []byte("---\nf: t\n---\n<!doctype html><html><body>more content</body></html>"), "blue"},
|
{"content/blue/doc2.html", []byte("---\nf: t\n---\n<!doctype html><html><body>more content</body></html>")},
|
||||||
}
|
}
|
||||||
for _, canonify := range []bool{true, false} {
|
for _, canonify := range []bool{true, false} {
|
||||||
viper.Set("CanonifyUrls", canonify)
|
viper.Set("CanonifyUrls", canonify)
|
||||||
|
@ -428,10 +428,10 @@ date = "2012-01-01"
|
||||||
Front Matter with Ordered Pages 4. This is longer content`)
|
Front Matter with Ordered Pages 4. This is longer content`)
|
||||||
|
|
||||||
var WEIGHTED_SOURCES = []source.ByteSource{
|
var WEIGHTED_SOURCES = []source.ByteSource{
|
||||||
{"sect/doc1.md", WEIGHTED_PAGE_1, "sect"},
|
{"sect/doc1.md", WEIGHTED_PAGE_1},
|
||||||
{"sect/doc2.md", WEIGHTED_PAGE_2, "sect"},
|
{"sect/doc2.md", WEIGHTED_PAGE_2},
|
||||||
{"sect/doc3.md", WEIGHTED_PAGE_3, "sect"},
|
{"sect/doc3.md", WEIGHTED_PAGE_3},
|
||||||
{"sect/doc4.md", WEIGHTED_PAGE_4, "sect"},
|
{"sect/doc4.md", WEIGHTED_PAGE_4},
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestOrderedPages(t *testing.T) {
|
func TestOrderedPages(t *testing.T) {
|
||||||
|
@ -484,13 +484,18 @@ func TestOrderedPages(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
var GROUPED_SOURCES = []source.ByteSource{
|
var GROUPED_SOURCES = []source.ByteSource{
|
||||||
{"sect1/doc1.md", WEIGHTED_PAGE_1, "sect1"},
|
{"sect1/doc1.md", WEIGHTED_PAGE_1},
|
||||||
{"sect1/doc2.md", WEIGHTED_PAGE_2, "sect1"},
|
{"sect1/doc2.md", WEIGHTED_PAGE_2},
|
||||||
{"sect2/doc3.md", WEIGHTED_PAGE_3, "sect2"},
|
{"sect2/doc3.md", WEIGHTED_PAGE_3},
|
||||||
{"sect3/doc4.md", WEIGHTED_PAGE_4, "sect3"},
|
{"sect3/doc4.md", WEIGHTED_PAGE_4},
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestGroupedPages(t *testing.T) {
|
func TestGroupedPages(t *testing.T) {
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r != nil {
|
||||||
|
fmt.Println("Recovered in f", r)
|
||||||
|
}
|
||||||
|
}()
|
||||||
files := make(map[string][]byte)
|
files := make(map[string][]byte)
|
||||||
target := &target.InMemoryTarget{Files: files}
|
target := &target.InMemoryTarget{Files: files}
|
||||||
|
|
||||||
|
@ -583,9 +588,9 @@ func TestWeightedTaxonomies(t *testing.T) {
|
||||||
files := make(map[string][]byte)
|
files := make(map[string][]byte)
|
||||||
target := &target.InMemoryTarget{Files: files}
|
target := &target.InMemoryTarget{Files: files}
|
||||||
sources := []source.ByteSource{
|
sources := []source.ByteSource{
|
||||||
{"sect/doc1.md", PAGE_WITH_WEIGHTED_TAXONOMIES_1, "sect"},
|
{"sect/doc1.md", PAGE_WITH_WEIGHTED_TAXONOMIES_1},
|
||||||
{"sect/doc2.md", PAGE_WITH_WEIGHTED_TAXONOMIES_2, "sect"},
|
{"sect/doc2.md", PAGE_WITH_WEIGHTED_TAXONOMIES_2},
|
||||||
{"sect/doc3.md", PAGE_WITH_WEIGHTED_TAXONOMIES_3, "sect"},
|
{"sect/doc3.md", PAGE_WITH_WEIGHTED_TAXONOMIES_3},
|
||||||
}
|
}
|
||||||
taxonomies := make(map[string]string)
|
taxonomies := make(map[string]string)
|
||||||
|
|
||||||
|
|
|
@ -45,8 +45,8 @@ func (t *InMemoryAliasTarget) Publish(label string, permalink template.HTML) (er
|
||||||
}
|
}
|
||||||
|
|
||||||
var urlFakeSource = []source.ByteSource{
|
var urlFakeSource = []source.ByteSource{
|
||||||
{"content/blue/doc1.md", []byte(SLUG_DOC_1), "blue"},
|
{"content/blue/doc1.md", []byte(SLUG_DOC_1)},
|
||||||
{"content/blue/doc2.md", []byte(SLUG_DOC_2), "blue"},
|
{"content/blue/doc2.md", []byte(SLUG_DOC_2)},
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestPageCount(t *testing.T) {
|
func TestPageCount(t *testing.T) {
|
||||||
|
|
|
@ -1,76 +1 @@
|
||||||
package hugolib
|
package hugolib
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"os/exec"
|
|
||||||
"strings"
|
|
||||||
|
|
||||||
jww "github.com/spf13/jwalterweatherman"
|
|
||||||
)
|
|
||||||
|
|
||||||
var summaryLength = 70
|
|
||||||
var summaryDivider = []byte("<!--more-->")
|
|
||||||
|
|
||||||
func TotalWords(s string) int {
|
|
||||||
return len(strings.Fields(s))
|
|
||||||
}
|
|
||||||
|
|
||||||
func WordCount(s string) map[string]int {
|
|
||||||
m := make(map[string]int)
|
|
||||||
for _, f := range strings.Fields(s) {
|
|
||||||
m[f] += 1
|
|
||||||
}
|
|
||||||
|
|
||||||
return m
|
|
||||||
}
|
|
||||||
|
|
||||||
func RemoveSummaryDivider(content []byte) []byte {
|
|
||||||
return bytes.Replace(content, summaryDivider, []byte(""), -1)
|
|
||||||
}
|
|
||||||
|
|
||||||
func TruncateWords(s string, max int) string {
|
|
||||||
words := strings.Fields(s)
|
|
||||||
if max > len(words) {
|
|
||||||
return strings.Join(words, " ")
|
|
||||||
}
|
|
||||||
|
|
||||||
return strings.Join(words[:max], " ")
|
|
||||||
}
|
|
||||||
|
|
||||||
func TruncateWordsToWholeSentence(s string, max int) string {
|
|
||||||
words := strings.Fields(s)
|
|
||||||
if max > len(words) {
|
|
||||||
return strings.Join(words, " ")
|
|
||||||
}
|
|
||||||
|
|
||||||
for counter, word := range words[max:] {
|
|
||||||
if strings.HasSuffix(word, ".") ||
|
|
||||||
strings.HasSuffix(word, "?") ||
|
|
||||||
strings.HasSuffix(word, ".\"") ||
|
|
||||||
strings.HasSuffix(word, "!") {
|
|
||||||
return strings.Join(words[:max+counter+1], " ")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return strings.Join(words[:max], " ")
|
|
||||||
}
|
|
||||||
|
|
||||||
func getRstContent(content []byte) string {
|
|
||||||
cleanContent := bytes.Replace(content, summaryDivider, []byte(""), 1)
|
|
||||||
|
|
||||||
cmd := exec.Command("rst2html.py", "--leave-comments")
|
|
||||||
cmd.Stdin = bytes.NewReader(cleanContent)
|
|
||||||
var out bytes.Buffer
|
|
||||||
cmd.Stdout = &out
|
|
||||||
if err := cmd.Run(); err != nil {
|
|
||||||
jww.ERROR.Println(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
rstLines := strings.Split(out.String(), "\n")
|
|
||||||
for i, line := range rstLines {
|
|
||||||
if strings.HasPrefix(line, "<body>") {
|
|
||||||
rstLines = (rstLines[i+1 : len(rstLines)-3])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return strings.Join(rstLines, "\n")
|
|
||||||
}
|
|
||||||
|
|
114
source/file.go
Normal file
114
source/file.go
Normal file
|
@ -0,0 +1,114 @@
|
||||||
|
// Copyright © 2014 Steve Francia <spf@spf13.com>.
|
||||||
|
//
|
||||||
|
// Licensed under the Simple Public License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
// http://opensource.org/licenses/Simple-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package source
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io"
|
||||||
|
"path"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/spf13/hugo/helpers"
|
||||||
|
)
|
||||||
|
|
||||||
|
type File struct {
|
||||||
|
relpath string // Original Full Path eg. /Users/Home/Hugo/foo.txt
|
||||||
|
logicalName string // foo.txt
|
||||||
|
Contents io.Reader
|
||||||
|
section string // The first directory
|
||||||
|
dir string // The full directory Path (minus file name)
|
||||||
|
ext string // Just the ext (eg txt)
|
||||||
|
uniqueId string // MD5 of the filename
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) UniqueId() string {
|
||||||
|
if f.uniqueId == "" {
|
||||||
|
f.uniqueId = helpers.Md5String(f.LogicalName())
|
||||||
|
}
|
||||||
|
return f.uniqueId
|
||||||
|
}
|
||||||
|
|
||||||
|
// Filename without extension
|
||||||
|
func (f *File) BaseFileName() string {
|
||||||
|
return helpers.Filename(f.LogicalName())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) Section() string {
|
||||||
|
if f.section != "" {
|
||||||
|
return f.section
|
||||||
|
} else {
|
||||||
|
f.section = helpers.GuessSection(f.Dir())
|
||||||
|
return f.section
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) LogicalName() string {
|
||||||
|
if f.logicalName != "" {
|
||||||
|
return f.logicalName
|
||||||
|
} else {
|
||||||
|
_, f.logicalName = path.Split(f.relpath)
|
||||||
|
return f.logicalName
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//func (f *File) SetDir(dir string) {
|
||||||
|
//f.dir = dir
|
||||||
|
//}
|
||||||
|
|
||||||
|
func (f *File) Dir() string {
|
||||||
|
if f.dir != "" {
|
||||||
|
return f.dir
|
||||||
|
} else {
|
||||||
|
f.dir, _ = path.Split(f.relpath)
|
||||||
|
return f.dir
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) Extension() string {
|
||||||
|
if f.ext != "" {
|
||||||
|
return f.ext
|
||||||
|
} else {
|
||||||
|
f.ext = strings.TrimPrefix(filepath.Ext(f.LogicalName()), ".")
|
||||||
|
return f.ext
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) Ext() string {
|
||||||
|
return f.Extension()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *File) Path() string {
|
||||||
|
return f.relpath
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewFileWithContents(relpath string, content io.Reader) *File {
|
||||||
|
file := NewFile(relpath)
|
||||||
|
file.Contents = content
|
||||||
|
return file
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewFile(relpath string) *File {
|
||||||
|
return &File{
|
||||||
|
relpath: relpath,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewFileFromAbs(base, fullpath string, content io.Reader) (f *File, err error) {
|
||||||
|
var name string
|
||||||
|
if name, err = helpers.GetRelativePath(fullpath, base); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return NewFileWithContents(name, content), nil
|
||||||
|
}
|
|
@ -1,34 +1,56 @@
|
||||||
|
// Copyright © 2014 Steve Francia <spf@spf13.com>.
|
||||||
|
//
|
||||||
|
// Licensed under the Simple Public License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
// http://opensource.org/licenses/Simple-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
package source
|
package source
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"errors"
|
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"os"
|
"os"
|
||||||
"path"
|
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
|
"github.com/spf13/hugo/helpers"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Input interface {
|
type Input interface {
|
||||||
Files() []*File
|
Files() []*File
|
||||||
}
|
}
|
||||||
|
|
||||||
type File struct {
|
|
||||||
name string
|
|
||||||
LogicalName string
|
|
||||||
Contents io.Reader
|
|
||||||
Section string
|
|
||||||
Dir string
|
|
||||||
}
|
|
||||||
|
|
||||||
type Filesystem struct {
|
type Filesystem struct {
|
||||||
files []*File
|
files []*File
|
||||||
Base string
|
Base string
|
||||||
AvoidPaths []string
|
AvoidPaths []string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (f *Filesystem) FilesByExts(exts ...string) []*File {
|
||||||
|
var newFiles []*File
|
||||||
|
|
||||||
|
if len(exts) == 0 {
|
||||||
|
return f.Files()
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, x := range f.Files() {
|
||||||
|
for _, e := range exts {
|
||||||
|
if x.Ext() == strings.TrimPrefix(e, ".") {
|
||||||
|
newFiles = append(newFiles, x)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return newFiles
|
||||||
|
}
|
||||||
|
|
||||||
func (f *Filesystem) Files() []*File {
|
func (f *Filesystem) Files() []*File {
|
||||||
if len(f.files) < 1 {
|
if len(f.files) < 1 {
|
||||||
f.captureFiles()
|
f.captureFiles()
|
||||||
|
@ -36,47 +58,23 @@ func (f *Filesystem) Files() []*File {
|
||||||
return f.files
|
return f.files
|
||||||
}
|
}
|
||||||
|
|
||||||
var errMissingBaseDir = errors.New("source: missing base directory")
|
|
||||||
|
|
||||||
func (f *Filesystem) add(name string, reader io.Reader) (err error) {
|
func (f *Filesystem) add(name string, reader io.Reader) (err error) {
|
||||||
|
var file *File
|
||||||
|
|
||||||
if name, err = f.getRelativePath(name); err != nil {
|
//if f.Base == "" {
|
||||||
|
//file = NewFileWithContents(name, reader)
|
||||||
|
//} else {
|
||||||
|
file, err = NewFileFromAbs(f.Base, name, reader)
|
||||||
|
//}
|
||||||
|
|
||||||
|
if err == nil {
|
||||||
|
f.files = append(f.files, file)
|
||||||
|
}
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// section should be the first part of the path
|
|
||||||
dir, logical := path.Split(name)
|
|
||||||
parts := strings.Split(dir, "/")
|
|
||||||
section := parts[0]
|
|
||||||
|
|
||||||
if section == "." {
|
|
||||||
section = ""
|
|
||||||
}
|
|
||||||
|
|
||||||
f.files = append(f.files, &File{
|
|
||||||
name: name,
|
|
||||||
LogicalName: logical,
|
|
||||||
Contents: reader,
|
|
||||||
Section: section,
|
|
||||||
Dir: dir,
|
|
||||||
})
|
|
||||||
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (f *Filesystem) getRelativePath(name string) (final string, err error) {
|
func (f *Filesystem) getRelativePath(name string) (final string, err error) {
|
||||||
if filepath.IsAbs(name) && f.Base == "" {
|
return helpers.GetRelativePath(name, f.Base)
|
||||||
return "", errMissingBaseDir
|
|
||||||
}
|
|
||||||
name = filepath.Clean(name)
|
|
||||||
base := filepath.Clean(f.Base)
|
|
||||||
|
|
||||||
name, err = filepath.Rel(base, name)
|
|
||||||
if err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
name = filepath.ToSlash(name)
|
|
||||||
return name, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f *Filesystem) captureFiles() {
|
func (f *Filesystem) captureFiles() {
|
||||||
|
|
|
@ -32,16 +32,17 @@ func TestAddFile(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, src := range []*Filesystem{srcDefault, srcWithBase} {
|
for _, src := range []*Filesystem{srcDefault, srcWithBase} {
|
||||||
|
|
||||||
p := test.filename
|
p := test.filename
|
||||||
if !filepath.IsAbs(test.filename) {
|
if !filepath.IsAbs(test.filename) {
|
||||||
p = path.Join(src.Base, test.filename)
|
p = path.Join(src.Base, test.filename)
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := src.add(p, bytes.NewReader([]byte(test.content))); err != nil {
|
if err := src.add(p, bytes.NewReader([]byte(test.content))); err != nil {
|
||||||
if err == errMissingBaseDir {
|
if err.Error() == "source: missing base directory" {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
t.Fatalf("%s add returned and error: %s", p, err)
|
t.Fatalf("%s add returned an error: %s", p, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(src.Files()) != 1 {
|
if len(src.Files()) != 1 {
|
||||||
|
@ -49,8 +50,8 @@ func TestAddFile(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
f := src.Files()[0]
|
f := src.Files()[0]
|
||||||
if f.LogicalName != test.logical {
|
if f.LogicalName() != test.logical {
|
||||||
t.Errorf("Filename (Base: %q) expected: %q, got: %q", src.Base, test.logical, f.LogicalName)
|
t.Errorf("Filename (Base: %q) expected: %q, got: %q", src.Base, test.logical, f.LogicalName())
|
||||||
}
|
}
|
||||||
|
|
||||||
b := new(bytes.Buffer)
|
b := new(bytes.Buffer)
|
||||||
|
@ -59,12 +60,12 @@ func TestAddFile(t *testing.T) {
|
||||||
t.Errorf("File (Base: %q) contents should be %q, got: %q", src.Base, test.content, b.String())
|
t.Errorf("File (Base: %q) contents should be %q, got: %q", src.Base, test.content, b.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
if f.Section != test.section {
|
if f.Section() != test.section {
|
||||||
t.Errorf("File section (Base: %q) expected: %q, got: %q", src.Base, test.section, f.Section)
|
t.Errorf("File section (Base: %q) expected: %q, got: %q", src.Base, test.section, f.Section())
|
||||||
}
|
}
|
||||||
|
|
||||||
if f.Dir != test.dir {
|
if f.Dir() != test.dir {
|
||||||
t.Errorf("Dir path (Base: %q) expected: %q, got: %q", src.Base, test.dir, f.Dir)
|
t.Errorf("Dir path (Base: %q) expected: %q, got: %q", src.Base, test.dir, f.Dir())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -3,17 +3,15 @@ package source
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
"path"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type ByteSource struct {
|
type ByteSource struct {
|
||||||
Name string
|
Name string
|
||||||
Content []byte
|
Content []byte
|
||||||
Section string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *ByteSource) String() string {
|
func (b *ByteSource) String() string {
|
||||||
return fmt.Sprintf("%s %s %s", b.Name, b.Section, string(b.Content))
|
return fmt.Sprintf("%s %s", b.Name, string(b.Content))
|
||||||
}
|
}
|
||||||
|
|
||||||
type InMemorySource struct {
|
type InMemorySource struct {
|
||||||
|
@ -23,12 +21,7 @@ type InMemorySource struct {
|
||||||
func (i *InMemorySource) Files() (files []*File) {
|
func (i *InMemorySource) Files() (files []*File) {
|
||||||
files = make([]*File, len(i.ByteSource))
|
files = make([]*File, len(i.ByteSource))
|
||||||
for i, fake := range i.ByteSource {
|
for i, fake := range i.ByteSource {
|
||||||
files[i] = &File{
|
files[i] = NewFileWithContents(fake.Name, bytes.NewReader(fake.Content))
|
||||||
LogicalName: fake.Name,
|
|
||||||
Contents: bytes.NewReader(fake.Content),
|
|
||||||
Section: fake.Section,
|
|
||||||
Dir: path.Dir(fake.Name),
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue