Fix URL handling in the whole markdown module, improve test coverage (#1027)
Amended with string to bool change in API SDK. Signed-off-by: Andrew Boyarshin <andrew.boyarshin@gmail.com>
This commit is contained in:
@ -150,7 +150,7 @@ func composeTplData(subject, body, link string) map[string]interface{} {
|
|||||||
|
|
||||||
func composeIssueMessage(issue *Issue, doer *User, tplName base.TplName, tos []string, info string) *mailer.Message {
|
func composeIssueMessage(issue *Issue, doer *User, tplName base.TplName, tos []string, info string) *mailer.Message {
|
||||||
subject := issue.mailSubject()
|
subject := issue.mailSubject()
|
||||||
body := string(markdown.RenderSpecialLink([]byte(issue.Content), issue.Repo.HTMLURL(), issue.Repo.ComposeMetas()))
|
body := string(markdown.RenderString(issue.Content, issue.Repo.HTMLURL(), issue.Repo.ComposeMetas()))
|
||||||
data := composeTplData(subject, body, issue.HTMLURL())
|
data := composeTplData(subject, body, issue.HTMLURL())
|
||||||
data["Doer"] = doer
|
data["Doer"] = doer
|
||||||
|
|
||||||
|
@ -92,10 +92,10 @@ var (
|
|||||||
ShortLinkPattern = regexp.MustCompile(`(\[\[.*\]\]\w*)`)
|
ShortLinkPattern = regexp.MustCompile(`(\[\[.*\]\]\w*)`)
|
||||||
|
|
||||||
// AnySHA1Pattern allows to split url containing SHA into parts
|
// AnySHA1Pattern allows to split url containing SHA into parts
|
||||||
AnySHA1Pattern = regexp.MustCompile(`http\S+//(\S+)/(\S+)/(\S+)/(\S+)/([0-9a-f]{40})(?:/?([^#\s]+)?(?:#(\S+))?)?`)
|
AnySHA1Pattern = regexp.MustCompile(`(http\S*)://(\S+)/(\S+)/(\S+)/(\S+)/([0-9a-f]{40})(?:/?([^#\s]+)?(?:#(\S+))?)?`)
|
||||||
|
|
||||||
// IssueFullPattern allows to split issue (and pull) URLs into parts
|
// IssueFullPattern allows to split issue (and pull) URLs into parts
|
||||||
IssueFullPattern = regexp.MustCompile(`(?:^|\s|\()http\S+//((?:[^\s/]+/)+)((?:\w{1,10}-)?[1-9][0-9]*)([\?|#]\S+.(\S+)?)?\b`)
|
IssueFullPattern = regexp.MustCompile(`(?:^|\s|\()(http\S*)://((?:[^\s/]+/)+)((?:\w{1,10}-)?[1-9][0-9]*)([\?|#]\S+.(\S+)?)?\b`)
|
||||||
|
|
||||||
validLinksPattern = regexp.MustCompile(`^[a-z][\w-]+://`)
|
validLinksPattern = regexp.MustCompile(`^[a-z][\w-]+://`)
|
||||||
)
|
)
|
||||||
@ -126,10 +126,11 @@ type Renderer struct {
|
|||||||
func (r *Renderer) Link(out *bytes.Buffer, link []byte, title []byte, content []byte) {
|
func (r *Renderer) Link(out *bytes.Buffer, link []byte, title []byte, content []byte) {
|
||||||
if len(link) > 0 && !isLink(link) {
|
if len(link) > 0 && !isLink(link) {
|
||||||
if link[0] != '#' {
|
if link[0] != '#' {
|
||||||
mLink := URLJoin(r.urlPrefix, string(link))
|
lnk := string(link)
|
||||||
if r.isWikiMarkdown {
|
if r.isWikiMarkdown {
|
||||||
mLink = URLJoin(r.urlPrefix, "wiki", string(link))
|
lnk = URLJoin("wiki", lnk)
|
||||||
}
|
}
|
||||||
|
mLink := URLJoin(r.urlPrefix, lnk)
|
||||||
link = []byte(mLink)
|
link = []byte(mLink)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -206,12 +207,10 @@ func (r *Renderer) Image(out *bytes.Buffer, link []byte, title []byte, alt []byt
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if link[0] != '/' {
|
lnk := string(link)
|
||||||
if !strings.HasSuffix(prefix, "/") {
|
lnk = URLJoin(prefix, lnk)
|
||||||
prefix += "/"
|
lnk = strings.Replace(lnk, " ", "+", -1)
|
||||||
}
|
link = []byte(lnk)
|
||||||
}
|
|
||||||
link = []byte(url.QueryEscape(prefix + string(link)))
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -246,10 +245,30 @@ func URLJoin(elem ...string) string {
|
|||||||
last := len(elem) - 1
|
last := len(elem) - 1
|
||||||
for i, item := range elem {
|
for i, item := range elem {
|
||||||
res += item
|
res += item
|
||||||
if !strings.HasSuffix(res, "/") && i != last {
|
if i != last && !strings.HasSuffix(res, "/") {
|
||||||
res += "/"
|
res += "/"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
cwdIndex := strings.Index(res, "/./")
|
||||||
|
for cwdIndex != -1 {
|
||||||
|
res = strings.Replace(res, "/./", "/", 1)
|
||||||
|
cwdIndex = strings.Index(res, "/./")
|
||||||
|
}
|
||||||
|
upIndex := strings.Index(res, "/..")
|
||||||
|
for upIndex != -1 {
|
||||||
|
res = strings.Replace(res, "/..", "", 1)
|
||||||
|
prevStart := -1
|
||||||
|
for i := upIndex - 1; i >= 0; i-- {
|
||||||
|
if res[i] == '/' {
|
||||||
|
prevStart = i
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if prevStart != -1 {
|
||||||
|
res = res[:prevStart] + res[upIndex:]
|
||||||
|
}
|
||||||
|
upIndex = strings.Index(res, "/..")
|
||||||
|
}
|
||||||
return res
|
return res
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -286,6 +305,9 @@ func RenderIssueIndexPattern(rawBytes []byte, urlPrefix string, metas map[string
|
|||||||
|
|
||||||
// IsSameDomain checks if given url string has the same hostname as current Gitea instance
|
// IsSameDomain checks if given url string has the same hostname as current Gitea instance
|
||||||
func IsSameDomain(s string) bool {
|
func IsSameDomain(s string) bool {
|
||||||
|
if strings.HasPrefix(s, "/") {
|
||||||
|
return true
|
||||||
|
}
|
||||||
if uapp, err := url.Parse(setting.AppURL); err == nil {
|
if uapp, err := url.Parse(setting.AppURL); err == nil {
|
||||||
if u, err := url.Parse(s); err == nil {
|
if u, err := url.Parse(s); err == nil {
|
||||||
return u.Host == uapp.Host
|
return u.Host == uapp.Host
|
||||||
@ -300,26 +322,27 @@ func renderFullSha1Pattern(rawBytes []byte, urlPrefix string) []byte {
|
|||||||
ms := AnySHA1Pattern.FindAllSubmatch(rawBytes, -1)
|
ms := AnySHA1Pattern.FindAllSubmatch(rawBytes, -1)
|
||||||
for _, m := range ms {
|
for _, m := range ms {
|
||||||
all := m[0]
|
all := m[0]
|
||||||
paths := string(m[1])
|
protocol := string(m[1])
|
||||||
var path = "//" + paths
|
paths := string(m[2])
|
||||||
author := string(m[2])
|
path := protocol + "://" + paths
|
||||||
repoName := string(m[3])
|
author := string(m[3])
|
||||||
|
repoName := string(m[4])
|
||||||
path = URLJoin(path, author, repoName)
|
path = URLJoin(path, author, repoName)
|
||||||
ltype := "src"
|
ltype := "src"
|
||||||
itemType := m[4]
|
itemType := m[5]
|
||||||
if IsSameDomain(paths) {
|
if IsSameDomain(paths) {
|
||||||
ltype = string(itemType)
|
ltype = string(itemType)
|
||||||
} else if string(itemType) == "commit" {
|
} else if string(itemType) == "commit" {
|
||||||
ltype = "commit"
|
ltype = "commit"
|
||||||
}
|
}
|
||||||
sha := m[5]
|
sha := m[6]
|
||||||
var subtree string
|
var subtree string
|
||||||
if len(m) > 6 && len(m[6]) > 0 {
|
if len(m) > 7 && len(m[7]) > 0 {
|
||||||
subtree = string(m[6])
|
subtree = string(m[7])
|
||||||
}
|
}
|
||||||
var line []byte
|
var line []byte
|
||||||
if len(m) > 7 && len(m[7]) > 0 {
|
if len(m) > 8 && len(m[8]) > 0 {
|
||||||
line = m[7]
|
line = m[8]
|
||||||
}
|
}
|
||||||
urlSuffix := ""
|
urlSuffix := ""
|
||||||
text := base.ShortSha(string(sha))
|
text := base.ShortSha(string(sha))
|
||||||
@ -346,23 +369,18 @@ func renderFullIssuePattern(rawBytes []byte, urlPrefix string) []byte {
|
|||||||
ms := IssueFullPattern.FindAllSubmatch(rawBytes, -1)
|
ms := IssueFullPattern.FindAllSubmatch(rawBytes, -1)
|
||||||
for _, m := range ms {
|
for _, m := range ms {
|
||||||
all := m[0]
|
all := m[0]
|
||||||
paths := bytes.Split(m[1], []byte("/"))
|
protocol := string(m[1])
|
||||||
|
paths := bytes.Split(m[2], []byte("/"))
|
||||||
paths = paths[:len(paths)-1]
|
paths = paths[:len(paths)-1]
|
||||||
if bytes.HasPrefix(paths[0], []byte("gist.")) {
|
if bytes.HasPrefix(paths[0], []byte("gist.")) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
var path string
|
path := protocol + "://" + string(m[2])
|
||||||
if len(paths) > 3 {
|
id := string(m[3])
|
||||||
// Internal one
|
|
||||||
path = URLJoin(urlPrefix, "issues")
|
|
||||||
} else {
|
|
||||||
path = "//" + string(m[1])
|
|
||||||
}
|
|
||||||
id := string(m[2])
|
|
||||||
path = URLJoin(path, id)
|
path = URLJoin(path, id)
|
||||||
var comment []byte
|
var comment []byte
|
||||||
if len(m) > 3 {
|
if len(m) > 3 {
|
||||||
comment = m[3]
|
comment = m[4]
|
||||||
}
|
}
|
||||||
urlSuffix := ""
|
urlSuffix := ""
|
||||||
text := "#" + id
|
text := "#" + id
|
||||||
@ -394,8 +412,13 @@ func lastIndexOfByte(sl []byte, target byte) int {
|
|||||||
return -1
|
return -1
|
||||||
}
|
}
|
||||||
|
|
||||||
// renderShortLinks processes [[syntax]]
|
// RenderShortLinks processes [[syntax]]
|
||||||
func renderShortLinks(rawBytes []byte, urlPrefix string, noLink bool) []byte {
|
//
|
||||||
|
// noLink flag disables making link tags when set to true
|
||||||
|
// so this function just replaces the whole [[...]] with the content text
|
||||||
|
//
|
||||||
|
// isWikiMarkdown is a flag to choose linking url prefix
|
||||||
|
func RenderShortLinks(rawBytes []byte, urlPrefix string, noLink bool, isWikiMarkdown bool) []byte {
|
||||||
ms := ShortLinkPattern.FindAll(rawBytes, -1)
|
ms := ShortLinkPattern.FindAll(rawBytes, -1)
|
||||||
for _, m := range ms {
|
for _, m := range ms {
|
||||||
orig := bytes.TrimSpace(m)
|
orig := bytes.TrimSpace(m)
|
||||||
@ -482,11 +505,17 @@ func renderShortLinks(rawBytes []byte, urlPrefix string, noLink bool) []byte {
|
|||||||
}
|
}
|
||||||
absoluteLink := isLink([]byte(link))
|
absoluteLink := isLink([]byte(link))
|
||||||
if !absoluteLink {
|
if !absoluteLink {
|
||||||
link = url.QueryEscape(link)
|
link = strings.Replace(link, " ", "+", -1)
|
||||||
}
|
}
|
||||||
if image {
|
if image {
|
||||||
if !absoluteLink {
|
if !absoluteLink {
|
||||||
link = URLJoin(urlPrefix, "wiki", "raw", link)
|
if IsSameDomain(urlPrefix) {
|
||||||
|
urlPrefix = strings.Replace(urlPrefix, "/src/", "/raw/", 1)
|
||||||
|
}
|
||||||
|
if isWikiMarkdown {
|
||||||
|
link = URLJoin("wiki", "raw", link)
|
||||||
|
}
|
||||||
|
link = URLJoin(urlPrefix, link)
|
||||||
}
|
}
|
||||||
title := props["title"]
|
title := props["title"]
|
||||||
if title == "" {
|
if title == "" {
|
||||||
@ -504,7 +533,10 @@ func renderShortLinks(rawBytes []byte, urlPrefix string, noLink bool) []byte {
|
|||||||
}
|
}
|
||||||
name = fmt.Sprintf(`<img src="%s" %s title="%s" />`, link, alt, title)
|
name = fmt.Sprintf(`<img src="%s" %s title="%s" />`, link, alt, title)
|
||||||
} else if !absoluteLink {
|
} else if !absoluteLink {
|
||||||
link = URLJoin(urlPrefix, "wiki", link)
|
if isWikiMarkdown {
|
||||||
|
link = URLJoin("wiki", link)
|
||||||
|
}
|
||||||
|
link = URLJoin(urlPrefix, link)
|
||||||
}
|
}
|
||||||
if noLink {
|
if noLink {
|
||||||
rawBytes = bytes.Replace(rawBytes, orig, []byte(name), -1)
|
rawBytes = bytes.Replace(rawBytes, orig, []byte(name), -1)
|
||||||
@ -527,7 +559,7 @@ func RenderCrossReferenceIssueIndexPattern(rawBytes []byte, urlPrefix string, me
|
|||||||
repo := string(bytes.Split(m, []byte("#"))[0])
|
repo := string(bytes.Split(m, []byte("#"))[0])
|
||||||
issue := string(bytes.Split(m, []byte("#"))[1])
|
issue := string(bytes.Split(m, []byte("#"))[1])
|
||||||
|
|
||||||
link := fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(urlPrefix, repo, "issues", issue), m)
|
link := fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(setting.AppURL, repo, "issues", issue), m)
|
||||||
rawBytes = bytes.Replace(rawBytes, m, []byte(link), 1)
|
rawBytes = bytes.Replace(rawBytes, m, []byte(link), 1)
|
||||||
}
|
}
|
||||||
return rawBytes
|
return rawBytes
|
||||||
@ -548,7 +580,7 @@ func renderSha1CurrentPattern(rawBytes []byte, urlPrefix string) []byte {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// RenderSpecialLink renders mentions, indexes and SHA1 strings to corresponding links.
|
// RenderSpecialLink renders mentions, indexes and SHA1 strings to corresponding links.
|
||||||
func RenderSpecialLink(rawBytes []byte, urlPrefix string, metas map[string]string) []byte {
|
func RenderSpecialLink(rawBytes []byte, urlPrefix string, metas map[string]string, isWikiMarkdown bool) []byte {
|
||||||
ms := MentionPattern.FindAll(rawBytes, -1)
|
ms := MentionPattern.FindAll(rawBytes, -1)
|
||||||
for _, m := range ms {
|
for _, m := range ms {
|
||||||
m = m[bytes.Index(m, []byte("@")):]
|
m = m[bytes.Index(m, []byte("@")):]
|
||||||
@ -556,7 +588,7 @@ func RenderSpecialLink(rawBytes []byte, urlPrefix string, metas map[string]strin
|
|||||||
[]byte(fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(setting.AppURL, string(m[1:])), m)), -1)
|
[]byte(fmt.Sprintf(`<a href="%s">%s</a>`, URLJoin(setting.AppURL, string(m[1:])), m)), -1)
|
||||||
}
|
}
|
||||||
|
|
||||||
rawBytes = renderShortLinks(rawBytes, urlPrefix, false)
|
rawBytes = RenderShortLinks(rawBytes, urlPrefix, false, isWikiMarkdown)
|
||||||
rawBytes = RenderIssueIndexPattern(rawBytes, urlPrefix, metas)
|
rawBytes = RenderIssueIndexPattern(rawBytes, urlPrefix, metas)
|
||||||
rawBytes = RenderCrossReferenceIssueIndexPattern(rawBytes, urlPrefix, metas)
|
rawBytes = RenderCrossReferenceIssueIndexPattern(rawBytes, urlPrefix, metas)
|
||||||
rawBytes = renderFullSha1Pattern(rawBytes, urlPrefix)
|
rawBytes = renderFullSha1Pattern(rawBytes, urlPrefix)
|
||||||
@ -601,7 +633,7 @@ var noEndTags = []string{"img", "input", "br", "hr"}
|
|||||||
|
|
||||||
// PostProcess treats different types of HTML differently,
|
// PostProcess treats different types of HTML differently,
|
||||||
// and only renders special links for plain text blocks.
|
// and only renders special links for plain text blocks.
|
||||||
func PostProcess(rawHTML []byte, urlPrefix string, metas map[string]string) []byte {
|
func PostProcess(rawHTML []byte, urlPrefix string, metas map[string]string, isWikiMarkdown bool) []byte {
|
||||||
startTags := make([]string, 0, 5)
|
startTags := make([]string, 0, 5)
|
||||||
var buf bytes.Buffer
|
var buf bytes.Buffer
|
||||||
tokenizer := html.NewTokenizer(bytes.NewReader(rawHTML))
|
tokenizer := html.NewTokenizer(bytes.NewReader(rawHTML))
|
||||||
@ -611,7 +643,7 @@ OUTER_LOOP:
|
|||||||
token := tokenizer.Token()
|
token := tokenizer.Token()
|
||||||
switch token.Type {
|
switch token.Type {
|
||||||
case html.TextToken:
|
case html.TextToken:
|
||||||
buf.Write(RenderSpecialLink([]byte(token.String()), urlPrefix, metas))
|
buf.Write(RenderSpecialLink([]byte(token.String()), urlPrefix, metas, isWikiMarkdown))
|
||||||
|
|
||||||
case html.StartTagToken:
|
case html.StartTagToken:
|
||||||
buf.WriteString(token.String())
|
buf.WriteString(token.String())
|
||||||
@ -623,7 +655,7 @@ OUTER_LOOP:
|
|||||||
token = tokenizer.Token()
|
token = tokenizer.Token()
|
||||||
|
|
||||||
// Copy the token to the output verbatim
|
// Copy the token to the output verbatim
|
||||||
buf.Write(renderShortLinks([]byte(token.String()), urlPrefix, true))
|
buf.Write(RenderShortLinks([]byte(token.String()), urlPrefix, true, isWikiMarkdown))
|
||||||
|
|
||||||
if token.Type == html.StartTagToken {
|
if token.Type == html.StartTagToken {
|
||||||
if !com.IsSliceContainsStr(noEndTags, token.Data) {
|
if !com.IsSliceContainsStr(noEndTags, token.Data) {
|
||||||
@ -673,9 +705,9 @@ OUTER_LOOP:
|
|||||||
|
|
||||||
// Render renders Markdown to HTML with all specific handling stuff.
|
// Render renders Markdown to HTML with all specific handling stuff.
|
||||||
func render(rawBytes []byte, urlPrefix string, metas map[string]string, isWikiMarkdown bool) []byte {
|
func render(rawBytes []byte, urlPrefix string, metas map[string]string, isWikiMarkdown bool) []byte {
|
||||||
urlPrefix = strings.Replace(urlPrefix, " ", "%20", -1)
|
urlPrefix = strings.Replace(urlPrefix, " ", "+", -1)
|
||||||
result := RenderRaw(rawBytes, urlPrefix, isWikiMarkdown)
|
result := RenderRaw(rawBytes, urlPrefix, isWikiMarkdown)
|
||||||
result = PostProcess(result, urlPrefix, metas)
|
result = PostProcess(result, urlPrefix, metas, isWikiMarkdown)
|
||||||
result = Sanitizer.SanitizeBytes(result)
|
result = Sanitizer.SanitizeBytes(result)
|
||||||
return result
|
return result
|
||||||
}
|
}
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -27,7 +27,13 @@ func Markdown(ctx *context.APIContext, form api.MarkdownOption) {
|
|||||||
|
|
||||||
switch form.Mode {
|
switch form.Mode {
|
||||||
case "gfm":
|
case "gfm":
|
||||||
ctx.Write(markdown.Render([]byte(form.Text), markdown.URLJoin(setting.AppURL, form.Context), nil))
|
md := []byte(form.Text)
|
||||||
|
context := markdown.URLJoin(setting.AppURL, form.Context)
|
||||||
|
if form.Wiki {
|
||||||
|
ctx.Write([]byte(markdown.RenderWiki(md, context, nil)))
|
||||||
|
} else {
|
||||||
|
ctx.Write(markdown.Render(md, context, nil))
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
ctx.Write(markdown.RenderRaw([]byte(form.Text), "", false))
|
ctx.Write(markdown.RenderRaw([]byte(form.Text), "", false))
|
||||||
}
|
}
|
||||||
|
@ -53,6 +53,7 @@ func TestAPI_RenderGFM(t *testing.T) {
|
|||||||
Mode: "gfm",
|
Mode: "gfm",
|
||||||
Text: "",
|
Text: "",
|
||||||
Context: Repo,
|
Context: Repo,
|
||||||
|
Wiki: true,
|
||||||
}
|
}
|
||||||
requrl, _ := url.Parse(markdown.URLJoin(AppURL, "api", "v1", "markdown"))
|
requrl, _ := url.Parse(markdown.URLJoin(AppURL, "api", "v1", "markdown"))
|
||||||
req := &http.Request{
|
req := &http.Request{
|
||||||
@ -74,7 +75,7 @@ func TestAPI_RenderGFM(t *testing.T) {
|
|||||||
<ul>
|
<ul>
|
||||||
<li><a href="` + AppSubURL + `wiki/Links" rel="nofollow">Links, Language bindings, Engine bindings</a></li>
|
<li><a href="` + AppSubURL + `wiki/Links" rel="nofollow">Links, Language bindings, Engine bindings</a></li>
|
||||||
<li><a href="` + AppSubURL + `wiki/Tips" rel="nofollow">Tips</a></li>
|
<li><a href="` + AppSubURL + `wiki/Tips" rel="nofollow">Tips</a></li>
|
||||||
<li>Bezier widget (by <a href="` + AppURL + `r-lyeh" rel="nofollow">@r-lyeh</a>)<a href="` + AppSubURL + `issues/786" rel="nofollow">#786</a></li>
|
<li>Bezier widget (by <a href="` + AppURL + `r-lyeh" rel="nofollow">@r-lyeh</a>)<a href="https://github.com/ocornut/imgui/issues/786" rel="nofollow">#786</a></li>
|
||||||
</ul>
|
</ul>
|
||||||
`,
|
`,
|
||||||
// wine-staging wiki home extract: special wiki syntax, images
|
// wine-staging wiki home extract: special wiki syntax, images
|
||||||
@ -97,7 +98,7 @@ Here are some links to the most important topics. You can find the full list of
|
|||||||
<p>Here are some links to the most important topics. You can find the full list of pages at the sidebar.</p>
|
<p>Here are some links to the most important topics. You can find the full list of pages at the sidebar.</p>
|
||||||
|
|
||||||
<p><a href="` + AppSubURL + `wiki/Configuration" rel="nofollow">Configuration</a>
|
<p><a href="` + AppSubURL + `wiki/Configuration" rel="nofollow">Configuration</a>
|
||||||
<a href="` + AppSubURL + `wiki/raw/images%2Ficon-bug.png" rel="nofollow"><img src="` + AppSubURL + `wiki/raw/images%2Ficon-bug.png" alt="images/icon-bug.png" title="icon-bug.png"/></a></p>
|
<a href="` + AppSubURL + `wiki/raw/images/icon-bug.png" rel="nofollow"><img src="` + AppSubURL + `wiki/raw/images/icon-bug.png" alt="images/icon-bug.png" title="icon-bug.png"/></a></p>
|
||||||
`,
|
`,
|
||||||
// Guard wiki sidebar: special syntax
|
// Guard wiki sidebar: special syntax
|
||||||
`[[Guardfile-DSL / Configuring-Guard|Guardfile-DSL---Configuring-Guard]]`,
|
`[[Guardfile-DSL / Configuring-Guard|Guardfile-DSL---Configuring-Guard]]`,
|
||||||
|
1
vendor/code.gitea.io/sdk/gitea/miscellaneous.go
generated
vendored
1
vendor/code.gitea.io/sdk/gitea/miscellaneous.go
generated
vendored
@ -9,4 +9,5 @@ type MarkdownOption struct {
|
|||||||
Text string
|
Text string
|
||||||
Mode string
|
Mode string
|
||||||
Context string
|
Context string
|
||||||
|
Wiki bool
|
||||||
}
|
}
|
||||||
|
Reference in New Issue
Block a user