forked from mirror/oddmu
Compare commits
8 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
26033de177 | ||
|
|
e1ba007f97 | ||
|
|
e90ff9e7dd | ||
|
|
70356e850a | ||
|
|
81a59fd6ac | ||
|
|
52d6f26eed | ||
|
|
171910ff4f | ||
|
|
5fb0f57b5c |
6
.gitignore
vendored
6
.gitignore
vendored
@@ -1,3 +1,7 @@
|
||||
/oddmu
|
||||
test.md
|
||||
/testdata/
|
||||
/oddmu-darwin-*
|
||||
/oddmu-linux-*
|
||||
/oddmu-windows-*
|
||||
/oddmu.exe
|
||||
/oddmu
|
||||
|
||||
2
RELEASE
2
RELEASE
@@ -4,6 +4,8 @@ When preparing a new release
|
||||
1. Run tests
|
||||
|
||||
2. Update man/oddmu-releases.7.txt
|
||||
- add missing items
|
||||
- change "(unreleased)"
|
||||
|
||||
3. make docs
|
||||
|
||||
|
||||
15
diff_test.go
15
diff_test.go
@@ -70,6 +70,7 @@ I hate the machine!`
|
||||
I shiver at home
|
||||
the monitor glares and moans
|
||||
my grey heart grows cold`
|
||||
// create s and overwrite it with r
|
||||
p := &Page{Name: "testdata/backup/cold", Body: []byte(s)}
|
||||
p.save()
|
||||
p = &Page{Name: "testdata/backup/cold", Body: []byte(r)}
|
||||
@@ -78,19 +79,29 @@ my grey heart grows cold`
|
||||
// diff from s to r:
|
||||
assert.Contains(t, body, `<del>fear or cold, who knows?</del>`)
|
||||
assert.Contains(t, body, `<ins>I hate the machine!</ins>`)
|
||||
// save u
|
||||
p = &Page{Name: "testdata/backup/cold", Body: []byte(u)}
|
||||
p.save()
|
||||
body = string(p.Diff())
|
||||
// diff from s to u since r was not 60 min or older
|
||||
// diff from s to u since r was not 60 min or older and so the backup is kept
|
||||
assert.Contains(t, body, `<del>fear or cold, who knows?</del>`)
|
||||
assert.Contains(t, body, `<ins>my grey heart grows cold</ins>`)
|
||||
// set timestamp 2h in the past
|
||||
ts := time.Now().Add(-2 * time.Hour)
|
||||
assert.NoError(t, os.Chtimes("testdata/backup/cold.md~", ts, ts))
|
||||
assert.NoError(t, os.Chtimes("testdata/backup/cold.md", ts, ts))
|
||||
// save r
|
||||
p = &Page{Name: "testdata/backup/cold", Body: []byte(r)}
|
||||
p.save()
|
||||
body = string(p.Diff())
|
||||
// diff from u to r:
|
||||
// diff from u to r since enough time has passed and the old backup is discarded
|
||||
assert.Contains(t, body, `<del>my grey heart grows cold</del>`)
|
||||
assert.Contains(t, body, `<ins>I hate the machine!</ins>`)
|
||||
// save s
|
||||
p = &Page{Name: "testdata/backup/cold", Body: []byte(s)}
|
||||
p.save()
|
||||
body = string(p.Diff())
|
||||
// diff from u to s since this is still "the same" editing window
|
||||
assert.Contains(t, body, `<del>my grey heart grows cold</del>`)
|
||||
assert.Contains(t, body, `<ins>fear or cold, who knows?</ins>`)
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@
|
||||
<meta charset="utf-8">
|
||||
<meta name="format-detection" content="telephone=no">
|
||||
<meta name="viewport" content="width=device-width">
|
||||
<base href="/view/{{.Dir}}">
|
||||
<title>Editing {{.Title}}</title>
|
||||
<style>
|
||||
html { max-width: 70ch; padding: 2ch; margin: auto; color: #111; background-color: #ffe; }
|
||||
|
||||
4
go.mod
4
go.mod
@@ -9,7 +9,8 @@ require (
|
||||
github.com/edwvee/exiffix v0.0.0-20210922235313-0f6cbda5e58f
|
||||
github.com/fsnotify/fsnotify v1.7.0
|
||||
github.com/gabriel-vasile/mimetype v1.4.3
|
||||
github.com/gomarkdown/markdown v0.0.0-20240930133403-7e0a027d98c5
|
||||
github.com/gen2brain/heic v0.3.1
|
||||
github.com/gomarkdown/markdown v0.0.0-20250207164621-7a1f277a159e
|
||||
github.com/google/subcommands v1.2.0
|
||||
github.com/hexops/gotextdiff v1.0.3
|
||||
github.com/microcosm-cc/bluemonday v1.0.26
|
||||
@@ -24,7 +25,6 @@ require (
|
||||
github.com/aymerick/douceur v0.2.0 // indirect
|
||||
github.com/davecgh/go-spew v1.1.1 // indirect
|
||||
github.com/ebitengine/purego v0.7.1 // indirect
|
||||
github.com/gen2brain/heic v0.3.1 // indirect
|
||||
github.com/gorilla/css v1.0.1 // indirect
|
||||
github.com/mattn/go-runewidth v0.0.15 // indirect
|
||||
github.com/pmezard/go-difflib v1.0.0 // indirect
|
||||
|
||||
8
go.sum
8
go.sum
@@ -13,14 +13,12 @@ github.com/fsnotify/fsnotify v1.7.0 h1:8JEhPFa5W2WU7YfeZzPNqzMP6Lwt7L2715Ggo0nos
|
||||
github.com/fsnotify/fsnotify v1.7.0/go.mod h1:40Bi/Hjc2AVfZrqy+aj+yEI+/bRxZnMJyTJwOpGvigM=
|
||||
github.com/gabriel-vasile/mimetype v1.4.3 h1:in2uUcidCuFcDKtdcBxlR0rJ1+fsokWf+uqxgUFjbI0=
|
||||
github.com/gabriel-vasile/mimetype v1.4.3/go.mod h1:d8uq/6HKRL6CGdk+aubisF/M5GcPfT7nKyLpA0lbSSk=
|
||||
github.com/gen2brain/heic v0.0.0-20230113233934-ca402e77a786 h1:zvgtcRb2B5gynWjm+Fc9oJZPHXwmcgyH0xCcNm6Rmo4=
|
||||
github.com/gen2brain/heic v0.0.0-20230113233934-ca402e77a786/go.mod h1:aKVJoQ0cc9K5Xb058XSnnAxXLliR97qbSqWBlm5ca1E=
|
||||
github.com/gen2brain/heic v0.3.1 h1:ClY5YTdXdIanw7pe9ZVUM9XcsqH6CCCa5CZBlm58qOs=
|
||||
github.com/gen2brain/heic v0.3.1/go.mod h1:m2sVIf02O7wfO8mJm+PvE91lnq4QYJy2hseUon7So10=
|
||||
github.com/gomarkdown/markdown v0.0.0-20240730141124-034f12af3bf6 h1:ZPy+2XJ8u0bB3sNFi+I72gMEMS7MTg7aZCCXPOjV8iw=
|
||||
github.com/gomarkdown/markdown v0.0.0-20240730141124-034f12af3bf6/go.mod h1:JDGcbDT52eL4fju3sZ4TeHGsQwhG9nbDV21aMyhwPoA=
|
||||
github.com/gomarkdown/markdown v0.0.0-20240930133403-7e0a027d98c5 h1:qIhG9h8tUzKsVHn0iHtWUohq7Ve7btgA8rGp7TvrIHw=
|
||||
github.com/gomarkdown/markdown v0.0.0-20240930133403-7e0a027d98c5/go.mod h1:JDGcbDT52eL4fju3sZ4TeHGsQwhG9nbDV21aMyhwPoA=
|
||||
github.com/gomarkdown/markdown v0.0.0-20250207164621-7a1f277a159e h1:ESHlT0RVZphh4JGBz49I5R6nTdC8Qyc08vU25GQHzzQ=
|
||||
github.com/gomarkdown/markdown v0.0.0-20250207164621-7a1f277a159e/go.mod h1:JDGcbDT52eL4fju3sZ4TeHGsQwhG9nbDV21aMyhwPoA=
|
||||
github.com/google/go-cmp v0.5.8 h1:e6P7q2lk1O+qJJb4BtCQXlK8vWEO8V1ZeuEdJNOqZyg=
|
||||
github.com/google/go-cmp v0.5.8/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY=
|
||||
github.com/google/subcommands v1.2.0 h1:vWQspBTo2nEqTUFita5/KeEWlUL8kQObDFbub/EN9oE=
|
||||
@@ -68,8 +66,6 @@ golang.org/x/image v0.15.0 h1:kOELfmgrmJlw4Cdb7g/QGuB3CvDrXbqEIww/pNtNBm8=
|
||||
golang.org/x/image v0.15.0/go.mod h1:HUYqC05R2ZcZ3ejNQsIHQDQiwWM4JBqmm6MKANTp4LE=
|
||||
golang.org/x/net v0.20.0 h1:aCL9BSgETF1k+blQaYUBx9hJ9LOGP3gAVemcZlf1Kpo=
|
||||
golang.org/x/net v0.20.0/go.mod h1:z8BVo6PvndSri0LbOE3hAn0apkU+1YvI6E70E9jsnvY=
|
||||
golang.org/x/sys v0.16.0 h1:xWw16ngr6ZMtmxDyKyIgsE93KNKz5HKmMa3b8ALHidU=
|
||||
golang.org/x/sys v0.16.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA=
|
||||
golang.org/x/sys v0.21.0 h1:rF+pYz3DAGSQAxAu1CbC7catZg4ebC4UIeIhKxBZvws=
|
||||
golang.org/x/sys v0.21.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA=
|
||||
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
||||
|
||||
@@ -8,9 +8,18 @@ import (
|
||||
)
|
||||
|
||||
func TestHashtagsCmd(t *testing.T) {
|
||||
cleanup(t, "testdata/hashtag")
|
||||
p := &Page{Name: "testdata/hashtag/hash", Body: []byte(`# Hash
|
||||
|
||||
I hope for a time
|
||||
not like today, relentless,
|
||||
just crocus blooming
|
||||
|
||||
#Crocus`)}
|
||||
p.save()
|
||||
b := new(bytes.Buffer)
|
||||
s := hashtagsCli(b)
|
||||
assert.Equal(t, subcommands.ExitSuccess, s)
|
||||
x := b.String()
|
||||
assert.Contains(t, x, "#like_this\t")
|
||||
assert.Contains(t, x, "crocus\t")
|
||||
}
|
||||
|
||||
7
index.go
7
index.go
@@ -62,10 +62,12 @@ func (idx *indexStore) reset() {
|
||||
}
|
||||
|
||||
// addDocument adds the text as a new document. This assumes that the index is locked!
|
||||
// The hashtags (only!) are used as tokens. They are stored in lower case.
|
||||
func (idx *indexStore) addDocument(text []byte) docid {
|
||||
id := idx.next_id
|
||||
idx.next_id++
|
||||
for _, token := range hashtags(text) {
|
||||
token = strings.ToLower(token)
|
||||
ids := idx.token[token]
|
||||
// Don't add same ID more than once. Checking the last
|
||||
// position of the []docid works because the id is
|
||||
@@ -193,8 +195,8 @@ func (idx *indexStore) update(p *Page) {
|
||||
idx.add(p)
|
||||
}
|
||||
|
||||
// search searches the index for a query string and returns page
|
||||
// names.
|
||||
// search searches the index. The query string is parsed for tokens. Each token is turned to lower cased and looked up
|
||||
// in the index. Each page in the result must contain all the tokens. Returns page names.
|
||||
func (idx *indexStore) search(q string) []string {
|
||||
idx.RLock()
|
||||
defer idx.RUnlock()
|
||||
@@ -203,6 +205,7 @@ func (idx *indexStore) search(q string) []string {
|
||||
if len(hashtags) > 0 {
|
||||
var r []docid
|
||||
for _, token := range hashtags {
|
||||
token = strings.ToLower(token)
|
||||
if ids, ok := idx.token[token]; ok {
|
||||
if r == nil {
|
||||
r = ids
|
||||
|
||||
@@ -11,8 +11,8 @@ func TestIndexAdd(t *testing.T) {
|
||||
idx.reset()
|
||||
idx.Lock()
|
||||
defer idx.Unlock()
|
||||
tag := "#hello"
|
||||
id := idx.addDocument([]byte("oh hi " + tag))
|
||||
tag := "hello"
|
||||
id := idx.addDocument([]byte("oh hi #" + tag))
|
||||
assert.Contains(t, idx.token, tag)
|
||||
idx.deleteDocument(id)
|
||||
assert.NotContains(t, idx.token, tag)
|
||||
@@ -31,10 +31,19 @@ func TestIndex(t *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
// Lower case hashtag!
|
||||
func TestSearchHashtag(t *testing.T) {
|
||||
cleanup(t, "testdata/search-hashtag")
|
||||
p := &Page{Name: "testdata/search-hashtag/search", Body: []byte(`# Search
|
||||
|
||||
I'm back in this room
|
||||
Shelf, table, chair, and shelf again
|
||||
Where are my glasses?
|
||||
|
||||
#Searching`)}
|
||||
p.save()
|
||||
index.load()
|
||||
q := "#like_this"
|
||||
pages, _ := search(q, "", "", 1, false)
|
||||
pages, _ := search("#searching", "", "", 1, false)
|
||||
assert.NotZero(t, len(pages))
|
||||
}
|
||||
|
||||
|
||||
@@ -5,7 +5,7 @@
|
||||
.nh
|
||||
.ad l
|
||||
.\" Begin generated content:
|
||||
.TH "ODDMU-RELEASES" "7" "2024-11-09"
|
||||
.TH "ODDMU-RELEASES" "7" "2025-02-09"
|
||||
.PP
|
||||
.SH NAME
|
||||
.PP
|
||||
@@ -15,7 +15,18 @@ oddmu-releases - what'\&s new?\&
|
||||
.PP
|
||||
This page lists user-visible features and template changes to consider.\&
|
||||
.PP
|
||||
.SS 1.14 (unreleased)
|
||||
.SS 1.15 (2025)
|
||||
.PP
|
||||
Fix the hashtag detection.\& This was necessary to cut down on the many false
|
||||
positives.\& They were most obvious with the \fIhashtags\fR subcommand.\& Now the
|
||||
Markdown parser is used at startup to index the pages, making startup slower
|
||||
(about twice as long with my blog).\& The Markdown parser is also used to parse
|
||||
search terms (where it makes little difference).\&
|
||||
.PP
|
||||
Fix the timestamp for backup files.\& This was necessary because the diff didn'\&t
|
||||
work as intended.\&
|
||||
.PP
|
||||
.SS 1.14 (2024)
|
||||
.PP
|
||||
Add \fIlist\fR, \fIdelete\fR and \fIrename\fR actions.\&
|
||||
.PP
|
||||
@@ -34,7 +45,7 @@ existing templates.\& For example, from upload.\&html:
|
||||
.fi
|
||||
.RE
|
||||
.PP
|
||||
The following line was added to the "preview.\&html" template:
|
||||
The following line was added to the "preview.\&html" and "edit.\&html" template:
|
||||
.PP
|
||||
.nf
|
||||
.RS 4
|
||||
@@ -44,7 +55,8 @@ The following line was added to the "preview.\&html" template:
|
||||
.PP
|
||||
You might want to do that as well, if you have your own.\& Without this, links in
|
||||
the preview cannot be followed as they all point to \fB/preview\fR instead of
|
||||
\fB/view\fR.\&
|
||||
\fB/view\fR and the link to the list of changes cannot be followed from the edit
|
||||
page: it leads to editing the list of changes.\&
|
||||
.PP
|
||||
.SS 1.13 (2024)
|
||||
.PP
|
||||
|
||||
@@ -8,7 +8,18 @@ oddmu-releases - what's new?
|
||||
|
||||
This page lists user-visible features and template changes to consider.
|
||||
|
||||
## 1.14 (unreleased)
|
||||
## 1.15 (2025)
|
||||
|
||||
Fix the hashtag detection. This was necessary to cut down on the many false
|
||||
positives. They were most obvious with the _hashtags_ subcommand. Now the
|
||||
Markdown parser is used at startup to index the pages, making startup slower
|
||||
(about twice as long with my blog). The Markdown parser is also used to parse
|
||||
search terms (where it makes little difference).
|
||||
|
||||
Fix the timestamp for backup files. This was necessary because the diff didn't
|
||||
work as intended.
|
||||
|
||||
## 1.14 (2024)
|
||||
|
||||
Add _list_, _delete_ and _rename_ actions.
|
||||
|
||||
@@ -25,7 +36,7 @@ existing templates. For example, from upload.html:
|
||||
<p>You can rename and delete files <a href="/list/{{.Dir}}">from the file list</a>.
|
||||
```
|
||||
|
||||
The following line was added to the "preview.html" template:
|
||||
The following line was added to the "preview.html" and "edit.html" template:
|
||||
|
||||
```
|
||||
<base href="/view/{{.Dir}}">
|
||||
@@ -33,7 +44,8 @@ The following line was added to the "preview.html" template:
|
||||
|
||||
You might want to do that as well, if you have your own. Without this, links in
|
||||
the preview cannot be followed as they all point to */preview* instead of
|
||||
*/view*.
|
||||
*/view* and the link to the list of changes cannot be followed from the edit
|
||||
page: it leads to editing the list of changes.
|
||||
|
||||
## 1.13 (2024)
|
||||
|
||||
|
||||
@@ -5,7 +5,7 @@
|
||||
.nh
|
||||
.ad l
|
||||
.\" Begin generated content:
|
||||
.TH "ODDMU" "1" "2024-09-25"
|
||||
.TH "ODDMU" "1" "2024-11-16"
|
||||
.PP
|
||||
.SH NAME
|
||||
.PP
|
||||
@@ -288,9 +288,10 @@ it, it can'\&t be a page name.\& Filenames can contain slashes and Oddmu creates
|
||||
subdirectories as necessary.\&
|
||||
.PP
|
||||
Files may not end with a tilde ('\&~'\&) – these are backup files.\& When saving pages
|
||||
and file uploads, the old file renamed to the backup file unless the backup file
|
||||
is less than an hour old, thus collapsing all edits made in an hour into a
|
||||
single diff when comparing backup and current version.\&
|
||||
and file uploads, the old file is renamed to the backup file unless the backup
|
||||
file is less than an hour old, thus collapsing all edits made in an hour into a
|
||||
single diff when comparing backup and current version.\& The backup also gets an
|
||||
updated timestamp so that subsequent edits don'\&t immediately overwrite it.\&
|
||||
.PP
|
||||
The \fBindex\fR page is the default page.\& People visiting the "root" of the site are
|
||||
redirected to "/view/index".\&
|
||||
|
||||
@@ -231,9 +231,10 @@ it, it can't be a page name. Filenames can contain slashes and Oddmu creates
|
||||
subdirectories as necessary.
|
||||
|
||||
Files may not end with a tilde ('~') – these are backup files. When saving pages
|
||||
and file uploads, the old file renamed to the backup file unless the backup file
|
||||
is less than an hour old, thus collapsing all edits made in an hour into a
|
||||
single diff when comparing backup and current version.
|
||||
and file uploads, the old file is renamed to the backup file unless the backup
|
||||
file is less than an hour old, thus collapsing all edits made in an hour into a
|
||||
single diff when comparing backup and current version. The backup also gets an
|
||||
updated timestamp so that subsequent edits don't immediately overwrite it.
|
||||
|
||||
The *index* page is the default page. People visiting the "root" of the site are
|
||||
redirected to "/view/index".
|
||||
|
||||
@@ -16,9 +16,9 @@ while (<>) {
|
||||
# italic
|
||||
s/\b_([^_]+)_\b/*$1*/g;
|
||||
# move all H1 headers to H2
|
||||
s/^# /## /;
|
||||
s/^# (.*)/"## ".ucfirst(lc($1))/e;
|
||||
# the new H1 title
|
||||
s/^([A-Z.-]*\([1-9]\))( ".*")?$/# $1/;
|
||||
s/^([A-Z.-]*\([1-9]\))( ".*")?$/"# ".lc($1)/e;
|
||||
# quoted URLs
|
||||
s/"(http.*?)"/`$1`/g;
|
||||
# quoted wiki links
|
||||
|
||||
14
page.go
14
page.go
@@ -88,9 +88,10 @@ func (p *Page) save() error {
|
||||
return os.WriteFile(fp, s, 0644)
|
||||
}
|
||||
|
||||
// backup a file by renaming (!) it unless the existing backup is less than an hour old. A backup gets a tilde appended
|
||||
// to it ("~"). This is true even if the file refers to a binary file like "image.png" and most applications don't know
|
||||
// what to do with a file called "image.png~". This expects a file path. Use filepath.FromSlash(path) if necessary.
|
||||
// backup a file by renaming it unless the existing backup is less than an hour old. A backup gets a tilde appended to
|
||||
// it ("~"). This is true even if the file refers to a binary file like "image.png" and most applications don't know
|
||||
// what to do with a file called "image.png~". This expects a file path. Use filepath.FromSlash(path) if necessary. The
|
||||
// backup file gets its modification time set to now so that subsequent edits don't immediately overwrite it again.
|
||||
func backup(fp string) error {
|
||||
_, err := os.Stat(fp)
|
||||
if err != nil {
|
||||
@@ -99,7 +100,12 @@ func backup(fp string) error {
|
||||
bp := fp + "~"
|
||||
fi, err := os.Stat(bp)
|
||||
if err != nil || time.Since(fi.ModTime()).Minutes() >= 60 {
|
||||
return os.Rename(fp, bp)
|
||||
err = os.Rename(fp, bp)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
ts := time.Now()
|
||||
return os.Chtimes(bp, ts, ts)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
11
parser.go
11
parser.go
@@ -37,9 +37,13 @@ func wikiLink(fn func(p *parser.Parser, data []byte, offset int) (int, ast.Node)
|
||||
|
||||
// hashtag returns an inline parser function. This indirection is
|
||||
// required because we want to receive an array of hashtags found.
|
||||
// The hashtags in the array keep their case.
|
||||
func hashtag() (func(p *parser.Parser, data []byte, offset int) (int, ast.Node), *[]string) {
|
||||
hashtags := make([]string, 0)
|
||||
return func(p *parser.Parser, data []byte, offset int) (int, ast.Node) {
|
||||
if p.InsideLink {
|
||||
return 0, nil
|
||||
}
|
||||
data = data[offset:]
|
||||
i := 0
|
||||
n := len(data)
|
||||
@@ -148,6 +152,13 @@ func (p *Page) images() []ImageData {
|
||||
return images
|
||||
}
|
||||
|
||||
// hashtags returns an array of hashtags
|
||||
func hashtags(s []byte) []string {
|
||||
parser, hashtags := wikiParser()
|
||||
markdown.Parse(s, parser)
|
||||
return *hashtags
|
||||
}
|
||||
|
||||
// toString for a node returns the text nodes' literals, concatenated. There is no whitespace added so the expectation
|
||||
// is that there is only one child node. Otherwise, there may be a space missing between the literals, depending on the
|
||||
// exact child nodes they belong to.
|
||||
|
||||
@@ -51,11 +51,13 @@ I am cold, alone</p>
|
||||
func TestPageHtmlHashtagCornerCases(t *testing.T) {
|
||||
p := &Page{Body: []byte(`#
|
||||
|
||||
ok # #o #ok`)}
|
||||
ok # #o #ok
|
||||
[oh #ok \#nok](ok)`)}
|
||||
p.renderHtml()
|
||||
r := `<p>#</p>
|
||||
|
||||
<p>ok # <a class="tag" href="/search/?q=%23o">#o</a> <a class="tag" href="/search/?q=%23ok">#ok</a></p>
|
||||
<p>ok # <a class="tag" href="/search/?q=%23o">#o</a> <a class="tag" href="/search/?q=%23ok">#ok</a>
|
||||
<a href="ok">oh #ok #nok</a></p>
|
||||
`
|
||||
assert.Equal(t, r, string(p.Html))
|
||||
}
|
||||
|
||||
17
preview_test.go
Normal file
17
preview_test.go
Normal file
@@ -0,0 +1,17 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
"github.com/stretchr/testify/assert"
|
||||
"net/url"
|
||||
"testing"
|
||||
)
|
||||
|
||||
func TestPreview(t *testing.T) {
|
||||
cleanup(t, "testdata/preview")
|
||||
|
||||
data := url.Values{}
|
||||
data.Set("body", "**Hallo**!")
|
||||
|
||||
r := assert.HTTPBody(makeHandler(previewHandler, false), "POST", "/view/testdata/preview/alex", data)
|
||||
assert.Contains(t, r, "<strong>Hallo</strong>!")
|
||||
}
|
||||
@@ -9,6 +9,7 @@ import (
|
||||
"io"
|
||||
"net/url"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"regexp"
|
||||
"strings"
|
||||
)
|
||||
@@ -18,6 +19,8 @@ type searchCmd struct {
|
||||
page int
|
||||
all bool
|
||||
extract bool
|
||||
files bool
|
||||
quiet bool
|
||||
}
|
||||
|
||||
func (cmd *searchCmd) SetFlags(f *flag.FlagSet) {
|
||||
@@ -25,12 +28,14 @@ func (cmd *searchCmd) SetFlags(f *flag.FlagSet) {
|
||||
f.IntVar(&cmd.page, "page", 1, "the page in the search result set, default 1")
|
||||
f.BoolVar(&cmd.all, "all", false, "show all the pages and ignore -page")
|
||||
f.BoolVar(&cmd.extract, "extract", false, "print page extract instead of link list")
|
||||
f.BoolVar(&cmd.files, "files", false, "show just the filenames")
|
||||
f.BoolVar(&cmd.quiet, "quiet", false, "suppress summary line at the top")
|
||||
}
|
||||
|
||||
func (*searchCmd) Name() string { return "search" }
|
||||
func (*searchCmd) Synopsis() string { return "search pages and print a list of links" }
|
||||
func (*searchCmd) Usage() string {
|
||||
return `search [-dir string] [-page <n>|-all] [-extract] <terms>:
|
||||
return `search [-dir string] [-page <n>|-all] [-extract|-files] [-quiet] <terms>:
|
||||
Search for pages matching terms and print the result set as a
|
||||
Markdown list. Before searching, all the pages are indexed. Thus,
|
||||
startup is slow. The benefit is that the page order is exactly as
|
||||
@@ -39,24 +44,24 @@ func (*searchCmd) Usage() string {
|
||||
}
|
||||
|
||||
func (cmd *searchCmd) Execute(_ context.Context, f *flag.FlagSet, _ ...interface{}) subcommands.ExitStatus {
|
||||
return searchCli(os.Stdout, cmd.dir, cmd.page, cmd.all, cmd.extract, false, f.Args())
|
||||
return searchCli(os.Stdout, cmd, f.Args())
|
||||
}
|
||||
|
||||
// searchCli runs the search command on the command line. It is used
|
||||
// here with an io.Writer for easy testing.
|
||||
func searchCli(w io.Writer, dir string, n int, all, extract bool, quiet bool, args []string) subcommands.ExitStatus {
|
||||
dir, err := checkDir(dir)
|
||||
func searchCli(w io.Writer, cmd *searchCmd, args []string) subcommands.ExitStatus {
|
||||
dir, err := checkDir(cmd.dir)
|
||||
if err != nil {
|
||||
return subcommands.ExitFailure
|
||||
}
|
||||
index.reset()
|
||||
index.load()
|
||||
q := strings.Join(args, " ")
|
||||
items, more := search(q, dir, "", n, true)
|
||||
if !quiet {
|
||||
items, more := search(q, dir, "", cmd.page, true)
|
||||
if !cmd.quiet {
|
||||
fmt.Fprint(os.Stderr, "Search for ", q)
|
||||
if !all {
|
||||
fmt.Fprint(os.Stderr, ", page ", n)
|
||||
if !cmd.all {
|
||||
fmt.Fprint(os.Stderr, ", page ", cmd.page)
|
||||
}
|
||||
fmt.Fprint(os.Stderr, ": ", len(items))
|
||||
if len(items) == 1 {
|
||||
@@ -65,8 +70,13 @@ func searchCli(w io.Writer, dir string, n int, all, extract bool, quiet bool, ar
|
||||
fmt.Fprint(os.Stderr, " results\n")
|
||||
}
|
||||
}
|
||||
if extract {
|
||||
if cmd.extract {
|
||||
searchExtract(w, items)
|
||||
} else if cmd.files {
|
||||
for _, p := range items {
|
||||
name := filepath.FromSlash(p.Name) + ".md\n"
|
||||
fmt.Fprintf(w, name)
|
||||
}
|
||||
} else {
|
||||
for _, p := range items {
|
||||
name := p.Name
|
||||
|
||||
@@ -9,7 +9,7 @@ import (
|
||||
|
||||
func TestSearchCmd(t *testing.T) {
|
||||
b := new(bytes.Buffer)
|
||||
s := searchCli(b, "", 1, false, false, true, []string{"oddµ"})
|
||||
s := searchCli(b, &searchCmd{quiet: true}, []string{"oddµ"})
|
||||
assert.Equal(t, subcommands.ExitSuccess, s)
|
||||
r := `* [Oddµ: A minimal wiki](README)
|
||||
* [Themes](themes/index)
|
||||
@@ -26,7 +26,7 @@ that before we type and speak
|
||||
we hear that moment`)}
|
||||
p.save()
|
||||
b := new(bytes.Buffer)
|
||||
s := searchCli(b, "testdata/search", 1, false, false, true, []string{"speak"})
|
||||
s := searchCli(b, &searchCmd{dir: "testdata/search", quiet: true}, []string{"speak"})
|
||||
assert.Equal(t, subcommands.ExitSuccess, s)
|
||||
r := `* [Wait](wait)
|
||||
`
|
||||
|
||||
@@ -221,7 +221,7 @@ func staticPage(source, target string) (*Page, error) {
|
||||
func staticFeed(source, target string, p *Page, ti time.Time) error {
|
||||
// render feed, maybe
|
||||
base := filepath.Base(source)
|
||||
_, ok := index.token["#"+strings.ToLower(base)]
|
||||
_, ok := index.token[strings.ToLower(base)]
|
||||
if base == "index" || ok {
|
||||
f := feed(p, ti)
|
||||
if len(f.Items) > 0 {
|
||||
|
||||
32
tokenizer.go
32
tokenizer.go
@@ -1,10 +1,8 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"strings"
|
||||
"unicode"
|
||||
"unicode/utf8"
|
||||
)
|
||||
|
||||
// lowercaseFilter returns a slice of lower case tokens.
|
||||
@@ -133,33 +131,3 @@ func highlightTokens(q string) []string {
|
||||
tokens = lowercaseFilter(tokens)
|
||||
return noPredicateFilter(tokens)
|
||||
}
|
||||
|
||||
// hashtags returns a slice of hashtags. Use this to extract hashtags
|
||||
// from a page body. This ignores Markdown completely.
|
||||
func hashtags(s []byte) []string {
|
||||
hashtags := make([]string, 0)
|
||||
for {
|
||||
i := bytes.IndexRune(s, '#')
|
||||
if i == -1 {
|
||||
return hashtags
|
||||
}
|
||||
if i > 0 && s[i-1] == '\\' {
|
||||
s = s[i+1:]
|
||||
continue
|
||||
}
|
||||
from := i
|
||||
i++
|
||||
for {
|
||||
r, n := utf8.DecodeRune(s[i:])
|
||||
if n > 0 && (unicode.IsLetter(r) || unicode.IsNumber(r) || r == '_') {
|
||||
i += n
|
||||
} else {
|
||||
break
|
||||
}
|
||||
}
|
||||
if i > from+1 { // not just "#"
|
||||
hashtags = append(hashtags, string(bytes.ToLower(s[from:i])))
|
||||
}
|
||||
s = s[i:]
|
||||
}
|
||||
}
|
||||
|
||||
@@ -7,7 +7,7 @@ import (
|
||||
)
|
||||
|
||||
func TestHashtags(t *testing.T) {
|
||||
assert.EqualValues(t, []string{"#truth"}, hashtags([]byte("This is boring. #Truth")), "hashtags")
|
||||
assert.EqualValues(t, []string{"Truth"}, hashtags([]byte("This is boring. #Truth")), "hashtags")
|
||||
}
|
||||
|
||||
func TestEscapedHashtags(t *testing.T) {
|
||||
|
||||
Reference in New Issue
Block a user