1 Star 0 Fork 80

skykainls/faygo

forked from andeyalee/faygo 
加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
克隆/下载
fs.go 28.95 KB
一键复制 编辑 原始数据 按行查看 历史
henrylee2cn 提交于 2017-07-15 18:33 . fix template render
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035
// Copyright 2016 HenryLee. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
// HTTP file system with cache request handler
package faygo
import (
"bytes"
"errors"
"fmt"
"io"
"io/ioutil"
"mime"
"mime/multipart"
"net/http"
"net/textproto"
"net/url"
"os"
"path"
"path/filepath"
"sort"
"strconv"
"strings"
"sync"
"time"
"github.com/henrylee2cn/faygo/acceptencoder"
"github.com/henrylee2cn/faygo/freecache"
"github.com/henrylee2cn/faygo/markdown"
)
const indexPage = "/index.html"
// FileServerManager is file cache system manager
type FileServerManager struct {
files map[string]CacheFile
cache *freecache.Cache
fileExpireSeconds int
maxSizeOfSingle int64
enableCache bool
enableCompress bool
errorFunc ErrorFunc
filesLock sync.RWMutex
}
// The cache size will be set to 512KB at minimum.
// If the size is set relatively large, you should call
// `debug.SetGCPercent()`, set it to a much smaller value
// to limit the memory consumption and GC pause time.
// expireSeconds <= 0 means no expire.
func newFileServerManager(cacheSize int64, fileExpireSeconds int, enableCache bool, enableCompress bool) *FileServerManager {
manager := &FileServerManager{
enableCache: enableCache,
enableCompress: enableCompress,
}
if enableCache {
manager.fileExpireSeconds = fileExpireSeconds
manager.cache = freecache.NewCache(int(cacheSize))
manager.files = map[string]CacheFile{}
manager.maxSizeOfSingle = cacheSize / 1024
if manager.maxSizeOfSingle < 512 {
manager.maxSizeOfSingle = 512
}
}
return manager
}
// Open gets or stores the file with compression and caching options.
// If the name is larger than 65535 or body is larger than 1/1024 of the cache size,
// the entry will not be written to the cache.
func (c *FileServerManager) Open(name string, encoding string, nocache bool) (http.File, error) {
var f http.File
var err error
var compressible = encoding != "" && c.enableCompress
var cacheable = !nocache && c.enableCache
if cacheable {
f, err = c.Get(name)
if err == nil {
return f, nil
}
}
f, err = os.Open(name)
if err != nil {
return nil, err
}
fileInfo, err := f.Stat()
if err != nil || fileInfo.IsDir() {
return f, err
}
var content []byte
if compressible {
content, encoding, err = fileCompress2(f, encoding)
f.Close()
if err != nil {
return nil, err
}
if !cacheable || int64(len(content)) > c.maxSizeOfSingle {
return &CacheFile{
fileInfo: fileInfo,
encoding: encoding,
Reader: bytes.NewReader(content),
}, nil
}
} else {
if !cacheable || fileInfo.Size() > c.maxSizeOfSingle {
return f, nil
}
content, err = ioutil.ReadAll(f)
f.Close()
if err != nil {
return nil, err
}
}
return c.Set(name, content, fileInfo, encoding)
}
// OpenFS gets or stores the cache file.
// If the name is larger than 65535 or body is larger than 1/1024 of the cache size,
// the entry will not be written to the cache.
func (c *FileServerManager) OpenFS(ctx *Context, name string, fs FileSystem) (http.File, error) {
var f http.File
var err error
var compressible = !fs.Nocompress() && c.enableCompress
var cacheable = !fs.Nocache() && c.enableCache
if cacheable {
f, err = c.Get(name)
if err == nil {
if encoding := f.(*CacheFile).encoding; encoding != "" {
ctx.W.Header().Set("Content-Encoding", encoding)
}
return f, nil
}
}
f, err = fs.Open(name)
if err != nil {
return nil, err
}
fileInfo, err := f.Stat()
if err != nil || fileInfo.IsDir() {
return f, err
}
var content []byte
var encoding string
if compressible {
content, encoding, err = fileCompress(f, ctx)
f.Close()
if err != nil {
return nil, err
}
if !cacheable || int64(len(content)) > c.maxSizeOfSingle {
return &CacheFile{
fileInfo: fileInfo,
encoding: encoding,
Reader: bytes.NewReader(content),
}, nil
}
} else {
if !cacheable || fileInfo.Size() > c.maxSizeOfSingle {
return f, nil
}
content, err = ioutil.ReadAll(f)
f.Close()
if err != nil {
return nil, err
}
}
return c.Set(name, content, fileInfo, encoding)
}
// Get gets file from cache.
func (c *FileServerManager) Get(name string) (http.File, error) {
b, err := c.cache.Get([]byte(name))
if err != nil {
c.filesLock.Lock()
delete(c.files, name)
c.filesLock.Unlock()
return nil, err
}
c.filesLock.RLock()
f := c.files[name]
c.filesLock.RUnlock()
f.Reader = bytes.NewReader(b)
return &f, nil
}
// Set sets file to cache.
func (c *FileServerManager) Set(name string, body []byte, fileInfo os.FileInfo, encoding string) (http.File, error) {
err := c.cache.Set([]byte(name), body, c.fileExpireSeconds)
if err != nil {
return nil, err
}
f := CacheFile{
fileInfo: fileInfo,
encoding: encoding,
}
c.filesLock.Lock()
c.files[name] = f
c.filesLock.Unlock()
f.Reader = bytes.NewReader(body)
return &f, nil
}
type (
// FileSystem is a file system with compression and caching options
FileSystem interface {
http.FileSystem
Nocompress() bool // not allowed compress
Nocache() bool // not allowed cache
}
fileSystem struct {
http.FileSystem
nocompress bool
nocache bool
}
)
func (fs *fileSystem) Nocompress() bool {
return fs.nocompress
}
func (fs *fileSystem) Nocache() bool {
return fs.nocache
}
// FS creates a file system with compression and caching options
func FS(fs http.FileSystem, nocompressAndNocache ...bool) FileSystem {
var nocompress, nocache bool
var count = len(nocompressAndNocache)
if count == 1 {
nocompress = nocompressAndNocache[0]
} else if count >= 2 {
nocompress = nocompressAndNocache[0]
nocache = nocompressAndNocache[1]
}
return &fileSystem{
FileSystem: fs,
nocompress: nocompress,
nocache: nocache,
}
}
// DirFS creates a file system with compression and caching options, similar to http.Dir
func DirFS(root string, nocompressAndNocache ...bool) FileSystem {
return FS(http.Dir(root), nocompressAndNocache...)
}
// RenderFS creates a file system with auto-rendering.
// param `suffix` is used to specify the extension to be rendered, `*` for all extensions.
func RenderFS(root string, suffix string, tplVar Map) FileSystem {
mime.AddExtensionType(path.Ext(suffix), "text/html")
return FS(&renderFS{
dir: root,
suffix: suffix,
tplVar: tplVar,
render: GetRender(),
}, false, true)
}
type renderFS struct {
dir string
suffix string
tplVar Map
render *Render
}
func (fs *renderFS) Open(name string) (http.File, error) {
if filepath.Separator != '/' && strings.ContainsRune(name, filepath.Separator) ||
strings.Contains(name, "\x00") {
return nil, errors.New("RenderFS: invalid character in file path")
}
dir := fs.dir
if dir == "" {
dir = "."
}
fname := filepath.Join(dir, filepath.FromSlash(path.Clean("/"+name)))
if fs.suffix != "*" && !strings.HasSuffix(fname, fs.suffix) {
f, err := global.fsManager.Open(fname, "", false)
if err != nil {
// Error("RenderFS:", fname, err)
return nil, err
}
return f, nil
}
b, fileInfo, err := fs.render.renderForFS(fname, fs.tplVar)
if err != nil {
if strings.Contains(err.Error(), "not find") {
return nil, os.ErrNotExist
}
// Error("RenderFS:", fname, err)
return NewFile(b, fileInfo), err
}
return NewFile(b, fileInfo), nil
}
// MarkdownFS creates a markdown file system.
func MarkdownFS(root string, nocompressAndNocache ...bool) FileSystem {
return FS(&markdownFS{
dir: root,
}, nocompressAndNocache...)
}
type markdownFS struct {
dir string
}
func (fs *markdownFS) Open(name string) (http.File, error) {
if filepath.Separator != '/' && strings.ContainsRune(name, filepath.Separator) ||
strings.Contains(name, "\x00") {
return nil, errors.New("MarkdownFS: invalid character in file path")
}
dir := fs.dir
if dir == "" {
dir = "."
}
fname := filepath.Join(dir, filepath.FromSlash(path.Clean("/"+name)))
f, err := global.fsManager.Open(fname, "", false)
if err != nil {
// Error("MarkdownFS:", fname, err)
return nil, err
}
if !strings.HasSuffix(fname, ".md") {
return f, nil
}
fileInfo, err := f.Stat()
if err != nil {
f.Close()
// Error("MarkdownFS:", fname, err)
return nil, err
}
b, err := ioutil.ReadAll(f)
f.Close()
if err != nil {
// Error("MarkdownFS:", fname, err)
return nil, err
}
b, err = markdown.GithubMarkdown(b, false)
if err != nil {
// Error("MarkdownFS:", fname, err)
return nil, err
}
return NewFile(b, fileInfo), nil
}
// CacheFile implements os.File
type CacheFile struct {
fileInfo os.FileInfo
encoding string
*bytes.Reader
}
var _ http.File = new(CacheFile)
// NewFile creates a cacheFile
func NewFile(b []byte, fileInfo os.FileInfo) *CacheFile {
return &CacheFile{
Reader: bytes.NewReader(b),
fileInfo: fileInfo,
}
}
// Stat returns file info
func (c *CacheFile) Stat() (os.FileInfo, error) {
if c.fileInfo == nil {
c.fileInfo = &FileInfo{
size: int64(c.Len()),
modTime: time.Now(),
}
}
return c.fileInfo, nil
}
// Close closes file
func (c *CacheFile) Close() error {
c.Reader = nil
return nil
}
// Readdir gets path info
func (c *CacheFile) Readdir(count int) ([]os.FileInfo, error) {
return []os.FileInfo{}, errors.New("Readdir " + c.fileInfo.Name() + ": The system cannot find the path specified.")
}
// FileInfo implements os.FileInfo
type FileInfo struct {
name string
size int64
mode os.FileMode
modTime time.Time
isDir bool
sys interface{}
}
// Name returns base name of the file
func (info *FileInfo) Name() string {
return info.name
}
// Size returns the size in bytes for regular files; system-dependent for others
func (info *FileInfo) Size() int64 {
return info.size
}
// Mode returns file mode bits
func (info *FileInfo) Mode() os.FileMode {
return info.mode
}
// ModTime returns modification time
func (info *FileInfo) ModTime() time.Time {
return info.modTime
}
// IsDir is the abbreviation for Mode().IsDir()
func (info *FileInfo) IsDir() bool {
return info.isDir
}
// Sys returns underlying data source (can return nil)
func (info *FileInfo) Sys() interface{} {
return info.sys
}
func (c *FileServerManager) dirList(ctx *Context, f http.File) {
dirs, err := f.Readdir(-1)
if err != nil {
// TODO: log err.Error() to the Server.ErrorLog, once it's possible
// for a handler to get at its Server via the *Context. See
// Issue 12438.
global.errorFunc(ctx, "Error reading directory", http.StatusInternalServerError)
return
}
sort.Sort(byName(dirs))
ctx.W.Header().Set("Content-Type", "text/html; charset=utf-8")
fmt.Fprintf(ctx.W, "<pre>\n")
for _, d := range dirs {
name := d.Name()
if d.IsDir() {
name += "/"
}
// name may contain '?' or '#', which must be escaped to remain
// part of the URL path, and not indicate the start of a query
// string or fragment.
url := url.URL{Path: name}
fmt.Fprintf(ctx.W, "<a href=\"%s\">%s</a>\n", url.String(), htmlReplacer.Replace(name))
}
fmt.Fprintf(ctx.W, "</pre>\n")
}
var htmlReplacer = strings.NewReplacer(
"&", "&amp;",
"<", "&lt;",
">", "&gt;",
// "&#34;" is shorter than "&quot;".
`"`, "&#34;",
// "&#39;" is shorter than "&apos;" and apos was not in HTML until HTML5.
"'", "&#39;",
)
// ServeContent replies to the request using the content in the
// provided ReadSeeker. The main benefit of ServeContent over io.Copy
// is that it handles Range requests properly, sets the MIME type, and
// handles If-Modified-Since requests.
//
// If the response's Content-Type header is not set, ServeContent
// first tries to deduce the type from name's file extension and,
// if that fails, falls back to reading the first block of the content
// and passing it to DetectContentType.
// The name is otherwise unused; in particular it can be empty and is
// never sent in the response.
//
// If modtime is not the zero time or Unix epoch, ServeContent
// includes it in a Last-Modified header in the response. If the
// request includes an If-Modified-Since header, ServeContent uses
// modtime to decide whether the content needs to be sent at all.
//
// The content's Seek method must work: ServeContent uses
// a seek to the end of the content to determine its size.
//
// If the caller has set ctx's ETag header, ServeContent uses it to
// handle requests using If-Range and If-None-Match.
//
// Note that *os.File implements the io.ReadSeeker interface.
func (c *FileServerManager) ServeContent(ctx *Context, name string, modtime time.Time, content io.ReadSeeker) {
sizeFunc := func() (int64, error) {
size, err := content.Seek(0, io.SeekEnd)
if err != nil {
return 0, errSeeker
}
_, err = content.Seek(0, io.SeekStart)
if err != nil {
return 0, errSeeker
}
return size, nil
}
c.serveContent(ctx, name, modtime, sizeFunc, content)
}
// errSeeker is returned by ServeContent's sizeFunc when the content
// doesn't seek properly. The underlying Seeker's error text isn't
// included in the sizeFunc reply so it's not sent over HTTP to end
// users.
var errSeeker = errors.New("seeker can't seek")
// The algorithm uses at most sniffLen bytes to make its decision.
const sniffLen = 512
// if name is empty, filename is unknown. (used for mime type, before sniffing)
// if modtime.IsZero(), modtime is unknown.
// content must be seeked to the beginning of the file.
// The sizeFunc is called at most once. Its error, if any, is sent in the HTTP response.
func (c *FileServerManager) serveContent(ctx *Context, name string, modtime time.Time, sizeFunc func() (int64, error), content io.ReadSeeker) {
if checkLastModified(ctx, modtime) {
return
}
rangeReq, done := checkETag(ctx, modtime)
if done {
return
}
code := http.StatusOK
// If Content-Type isn't set, use the file's extension to find it, but
// if the Content-Type is unset explicitly, do not sniff the type.
ctypes, haveType := ctx.W.Header()["Content-Type"]
var ctype string
if !haveType {
ctype = mime.TypeByExtension(filepath.Ext(name))
// Warning("ctypes:", haveType, name, filepath.Ext(name), ctype)
if ctype == "" {
// read a chunk to decide between utf-8 text and binary
var buf [sniffLen]byte
n, _ := io.ReadFull(content, buf[:])
ctype = http.DetectContentType(buf[:n])
_, err := content.Seek(0, io.SeekStart) // rewind to output whole file
if err != nil {
global.errorFunc(ctx, "seeker can't seek", http.StatusInternalServerError)
return
}
}
ctx.W.Header().Set("Content-Type", ctype)
} else if len(ctypes) > 0 {
ctype = ctypes[0]
}
size, err := sizeFunc()
if err != nil {
global.errorFunc(ctx, err.Error(), http.StatusInternalServerError)
return
}
// handle Content-Range header.
sendSize := size
var sendContent io.Reader = content
if size >= 0 {
ranges, err := parseRange(rangeReq, size)
if err != nil {
global.errorFunc(ctx, err.Error(), http.StatusRequestedRangeNotSatisfiable)
return
}
if sumRangesSize(ranges) > size {
// The total number of bytes in all the ranges
// is larger than the size of the file by
// itself, so this is probably an attack, or a
// dumb client. Ignore the range request.
ranges = nil
}
switch {
case len(ranges) == 1:
// RFC 2616, Section 14.16:
// "When an HTTP message includes the content of a single
// range (for example, a response to a request for a
// single range, or to a request for a set of ranges
// that overlap without any holes), this content is
// transmitted with a Content-Range header, and a
// Content-Length header showing the number of bytes
// actually transferred.
// ...
// A response to a request for a single range MUST NOT
// be sent using the multipart/byteranges media type."
ra := ranges[0]
if _, err := content.Seek(ra.start, io.SeekStart); err != nil {
global.errorFunc(ctx, err.Error(), http.StatusRequestedRangeNotSatisfiable)
return
}
sendSize = ra.length
code = http.StatusPartialContent
ctx.W.Header().Set("Content-Range", ra.contentRange(size))
case len(ranges) > 1:
sendSize = rangesMIMESize(ranges, ctype, size)
code = http.StatusPartialContent
pr, pw := io.Pipe()
mw := multipart.NewWriter(pw)
ctx.W.Header().Set("Content-Type", "multipart/byteranges; boundary="+mw.Boundary())
sendContent = pr
defer pr.Close() // cause writing goroutine to fail and exit if CopyN doesn't finish.
go func() {
for _, ra := range ranges {
part, err := mw.CreatePart(ra.mimeHeader(ctype, size))
if err != nil {
pw.CloseWithError(err)
return
}
if _, err := content.Seek(ra.start, io.SeekStart); err != nil {
pw.CloseWithError(err)
return
}
if _, err := io.CopyN(part, content, ra.length); err != nil {
pw.CloseWithError(err)
return
}
}
mw.Close()
pw.Close()
}()
}
ctx.W.Header().Set("Accept-Ranges", "bytes")
if ctx.W.Header().Get("Content-Encoding") == "" {
ctx.W.Header().Set("Content-Length", strconv.FormatInt(sendSize, 10))
}
}
ctx.W.WriteHeader(code)
if ctx.R.Method != "HEAD" {
io.CopyN(ctx.W, sendContent, sendSize)
}
}
var unixEpochTime = time.Unix(0, 0)
// modtime is the modification time of the resource to be served, or IsZero().
// return value is whether this request is now complete.
func checkLastModified(ctx *Context, modtime time.Time) bool {
if modtime.IsZero() || modtime.Equal(unixEpochTime) {
// If the file doesn't have a modtime (IsZero), or the modtime
// is obviously garbage (Unix time == 0), then ignore modtimes
// and don't process the If-Modified-Since header.
return false
}
// The Date-Modified header truncates sub-second precision, so
// use mtime < t+1s instead of mtime <= t to check for unmodified.
if t, err := time.Parse(http.TimeFormat, ctx.R.Header.Get("If-Modified-Since")); err == nil && modtime.Before(t.Add(1*time.Second)) {
h := ctx.W.Header()
delete(h, "Content-Type")
delete(h, "Content-Length")
ctx.W.WriteHeader(http.StatusNotModified)
return true
}
ctx.W.Header().Set("Last-Modified", modtime.UTC().Format(http.TimeFormat))
return false
}
// checkETag implements If-None-Match and If-Range checks.
//
// The ETag or modtime must have been previously set in the
// *Context's headers. The modtime is only compared at second
// granularity and may be the zero value to mean unknown.
//
// The return value is the effective request "Range" header to use and
// whether this request is now considered done.
func checkETag(ctx *Context, modtime time.Time) (rangeReq string, done bool) {
etag := ctx.W.Header().Get("Etag")
rangeReq = ctx.R.Header.Get("Range")
// Invalidate the range request if the entity doesn't match the one
// the client was expecting.
// "If-Range: version" means "ignore the Range: header unless version matches the
// current file."
// We only support ETag versions.
// The caller must have set the ETag on the response already.
if ir := ctx.R.Header.Get("If-Range"); ir != "" && ir != etag {
// The If-Range value is typically the ETag value, but it may also be
// the modtime date. See golang.org/issue/8367.
timeMatches := false
if !modtime.IsZero() {
if t, err := http.ParseTime(ir); err == nil && t.Unix() == modtime.Unix() {
timeMatches = true
}
}
if !timeMatches {
rangeReq = ""
}
}
if inm := ctx.R.Header.Get("If-None-Match"); inm != "" {
// Must know ETag.
if etag == "" {
return rangeReq, false
}
// TODO(bradfitz): non-GET/HEAD requests require more work:
// sending a different status code on matches, and
// also can't use weak cache validators (those with a "W/
// prefix). But most users of ServeContent will be using
// it on GET or HEAD, so only support those for now.
if ctx.R.Method != "GET" && ctx.R.Method != "HEAD" {
return rangeReq, false
}
// TODO(bradfitz): deal with comma-separated or multiple-valued
// list of If-None-match values. For now just handle the common
// case of a single item.
if inm == etag || inm == "*" {
h := ctx.W.Header()
delete(h, "Content-Type")
delete(h, "Content-Length")
ctx.W.WriteHeader(http.StatusNotModified)
return "", true
}
}
return rangeReq, false
}
// name is '/'-separated, not filepath.Separator.
func (c *FileServerManager) serveFile(ctx *Context, fs FileSystem, name string, redirect bool) {
// redirect .../index.html to .../
// can't use Redirect() because that would make the path absolute,
// which would be a problem running under StripPrefix
//
// if strings.HasSuffix(ctx.R.URL.Path, indexPage) {
// localRedirect(ctx, "./")
// return
// }
f, err := c.OpenFS(ctx, name, fs)
if err != nil {
msg, code := toHTTPError(err)
global.errorFunc(ctx, msg, code)
return
}
defer f.Close()
d, err := f.Stat()
if err != nil {
msg, code := toHTTPError(err)
global.errorFunc(ctx, msg, code)
return
}
// if redirect {
// // redirect to canonical path: / at end of directory url
// // ctx.R.URL.Path always begins with /
// url := ctx.R.URL.Path
// if d.IsDir() {
// if url[len(url)-1] != '/' {
// localRedirect(ctx, path.Base(url)+"/")
// return
// }
// } else {
// if url[len(url)-1] == '/' {
// localRedirect(ctx, "../"+path.Base(url))
// return
// }
// }
// }
// redirect if the directory name doesn't end in a slash
if d.IsDir() {
url := ctx.R.URL.Path
if url[len(url)-1] != '/' {
localRedirect(ctx, path.Base(url)+"/")
return
}
}
// use contents of index.html for directory, if present
if d.IsDir() {
index := strings.TrimSuffix(name, "/") + indexPage
ff, err := c.OpenFS(ctx, index, fs)
if err == nil {
defer ff.Close()
dd, err := ff.Stat()
if err == nil {
// name = index
d = dd
f = ff
}
}
}
// Still a directory? (we didn't find an index.html file)
if d.IsDir() {
if checkLastModified(ctx, d.ModTime()) {
return
}
global.errorFunc(ctx, http.StatusText(http.StatusNotFound), http.StatusNotFound)
// c.dirList(ctx, f)
return
}
// serveContent will check modification time
sizeFunc := func() (int64, error) { return d.Size(), nil }
c.serveContent(ctx, d.Name(), d.ModTime(), sizeFunc, f)
}
func fileCompress(file http.File, ctx *Context) ([]byte, string, error) {
var buf = &bytes.Buffer{}
var encoding string
if b, n, _ := acceptencoder.WriteFile(acceptencoder.ParseEncoding(ctx.R), buf, file); b {
ctx.W.Header().Set("Content-Encoding", n)
encoding = n
}
return buf.Bytes(), encoding, nil
}
func fileCompress2(f http.File, encoding string) ([]byte, string, error) {
var buf = &bytes.Buffer{}
if b, n, _ := acceptencoder.WriteFile(encoding, buf, f); b {
encoding = n
}
f.Close()
return buf.Bytes(), encoding, nil
}
// toHTTPError returns a non-specific HTTP error message and status code
// for a given non-nil error value. It's important that toHTTPError does not
// actually return err.Error(), since msg and httpStatus are returned to users,
// and historically Go's ServeContent always returned just "404 Not Found" for
// all errors. We don't want to start leaking information in error messages.
func toHTTPError(err error) (msg string, httpStatus int) {
if os.IsNotExist(err) {
return "404 page not found", http.StatusNotFound
}
if os.IsPermission(err) {
return "403 Forbidden", http.StatusForbidden
}
// Default:
return err.Error(), http.StatusInternalServerError
// return "500 Internal Server Error", http.StatusInternalServerError
}
// localRedirect gives a Moved Permanently response.
// It does not convert relative paths to absolute paths like Redirect does.
func localRedirect(ctx *Context, newPath string) {
if q := ctx.R.URL.RawQuery; q != "" {
newPath += "?" + q
}
ctx.W.Header().Set("Location", newPath)
ctx.W.WriteHeader(http.StatusMovedPermanently)
}
// ServeFile replies to the request with the contents of the named
// file or directory.
//
// If the provided file or directory name is a relative path, it is
// interpreted relative to the current directory and may ascend to parent
// directories. If the provided name is constructed from user input, it
// should be sanitized before calling ServeFile. As a precaution, ServeFile
// will reject requests where r.URL.Path contains a ".." path element.
//
// As a special case, ServeFile redirects any request where r.URL.Path
// ends in "/index.html" to the same path, without the final
// "index.html". To avoid such redirects either modify the path or
// use ServeContent.
func (c *FileServerManager) ServeFile(ctx *Context, name string, nocompressAndNocache ...bool) {
if containsDotDot(ctx.R.URL.Path) {
// Too many programs use ctx.R.URL.Path to construct the argument to
// serveFile. Reject the request under the assumption that happened
// here and ".." may not be wanted.
// Note that name might not contain "..", for example if code (still
// incorrectly) used filepath.Join(myDir, ctx.R.URL.Path).
global.errorFunc(ctx, "invalid URL path", http.StatusBadRequest)
return
}
dir, file := filepath.Split(name)
c.serveFile(ctx, DirFS(dir, nocompressAndNocache...), file, false)
}
func containsDotDot(v string) bool {
if !strings.Contains(v, "..") {
return false
}
for _, ent := range strings.FieldsFunc(v, isSlashRune) {
if ent == ".." {
return true
}
}
return false
}
func isSlashRune(r rune) bool { return r == '/' || r == '\\' }
type fileHandler struct {
root FileSystem
fileServerManager *FileServerManager
}
// FileServer returns a handler that serves HTTP requests
// with the contents of the file system rooted at fs.
//
// To use the operating system's file system implementation,
// use http.Dir:
//
// http.Handle("/", http.FileServer(http.Dir("/tmp")))
//
// As a special case, the returned file server redirects any request
// ending in "/index.html" to the same path, without the final
// "index.html".
func (c *FileServerManager) FileServer(fs FileSystem) Handler {
return &fileHandler{fs, c}
}
func (f *fileHandler) Serve(ctx *Context) error {
r := ctx.R
upath := r.URL.Path
if !strings.HasPrefix(upath, "/") {
upath = "/" + upath
r.URL.Path = upath
}
f.fileServerManager.serveFile(ctx, f.root, path.Clean(upath), true)
return nil
}
// httpRange specifies the byte range to be sent to the client.
type httpRange struct {
start, length int64
}
func (r httpRange) contentRange(size int64) string {
return fmt.Sprintf("bytes %d-%d/%d", r.start, r.start+r.length-1, size)
}
func (r httpRange) mimeHeader(contentType string, size int64) textproto.MIMEHeader {
return textproto.MIMEHeader{
"Content-Range": {r.contentRange(size)},
"Content-Type": {contentType},
}
}
// parseRange parses a Range header string as per RFC 2616.
func parseRange(s string, size int64) ([]httpRange, error) {
if s == "" {
return nil, nil // header not present
}
const b = "bytes="
if !strings.HasPrefix(s, b) {
return nil, errors.New("invalid range")
}
var ranges []httpRange
for _, ra := range strings.Split(s[len(b):], ",") {
ra = strings.TrimSpace(ra)
if ra == "" {
continue
}
i := strings.Index(ra, "-")
if i < 0 {
return nil, errors.New("invalid range")
}
start, end := strings.TrimSpace(ra[:i]), strings.TrimSpace(ra[i+1:])
var r httpRange
if start == "" {
// If no start is specified, end specifies the
// range start relative to the end of the file.
i, err := strconv.ParseInt(end, 10, 64)
if err != nil {
return nil, errors.New("invalid range")
}
if i > size {
i = size
}
r.start = size - i
r.length = size - r.start
} else {
i, err := strconv.ParseInt(start, 10, 64)
if err != nil || i >= size || i < 0 {
return nil, errors.New("invalid range")
}
r.start = i
if end == "" {
// If no end is specified, range extends to end of the file.
r.length = size - r.start
} else {
i, err := strconv.ParseInt(end, 10, 64)
if err != nil || r.start > i {
return nil, errors.New("invalid range")
}
if i >= size {
i = size - 1
}
r.length = i - r.start + 1
}
}
ranges = append(ranges, r)
}
return ranges, nil
}
// countingWriter counts how many bytes have been written to it.
type countingWriter int64
func (w *countingWriter) Write(p []byte) (n int, err error) {
*w += countingWriter(len(p))
return len(p), nil
}
// rangesMIMESize returns the number of bytes it takes to encode the
// provided ranges as a multipart response.
func rangesMIMESize(ranges []httpRange, contentType string, contentSize int64) (encSize int64) {
var w countingWriter
mw := multipart.NewWriter(&w)
for _, ra := range ranges {
mw.CreatePart(ra.mimeHeader(contentType, contentSize))
encSize += ra.length
}
mw.Close()
encSize += int64(w)
return
}
func sumRangesSize(ranges []httpRange) (size int64) {
for _, ra := range ranges {
size += ra.length
}
return
}
type byName []os.FileInfo
func (s byName) Len() int { return len(s) }
func (s byName) Less(i, j int) bool { return s[i].Name() < s[j].Name() }
func (s byName) Swap(i, j int) { s[i], s[j] = s[j], s[i] }
马建仓 AI 助手
尝试更多
代码解读
代码找茬
代码优化
Go
1
https://gitee.com/skykain/faygo.git
git@gitee.com:skykain/faygo.git
skykain
faygo
faygo
master

搜索帮助