mirror of
https://github.com/mjl-/mox.git
synced 2025-07-12 17:44:35 +03:00
mox!
This commit is contained in:
463
imapserver/search.go
Normal file
463
imapserver/search.go
Normal file
@ -0,0 +1,463 @@
|
||||
package imapserver
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"io"
|
||||
"net/textproto"
|
||||
"strings"
|
||||
|
||||
"github.com/mjl-/bstore"
|
||||
|
||||
"github.com/mjl-/mox/message"
|
||||
"github.com/mjl-/mox/mlog"
|
||||
"github.com/mjl-/mox/store"
|
||||
)
|
||||
|
||||
// Search returns messages matching criteria specified in parameters.
|
||||
//
|
||||
// State: Selected
|
||||
func (c *conn) cmdxSearch(isUID bool, tag, cmd string, p *parser) {
|
||||
// Command: ../rfc/9051:3716 ../rfc/4731:31 ../rfc/4466:354 ../rfc/3501:2723
|
||||
// Examples: ../rfc/9051:3986 ../rfc/4731:153 ../rfc/3501:2975
|
||||
// Syntax: ../rfc/9051:6918 ../rfc/4466:611 ../rfc/3501:4954
|
||||
|
||||
// We will respond with ESEARCH instead of SEARCH if "RETURN" is present or for IMAP4rev2.
|
||||
var eargs map[string]bool // Options except SAVE. Nil means old-style SEARCH response.
|
||||
var save bool // For SAVE option. Kept separately for easier handling of MIN/MAX later.
|
||||
|
||||
// IMAP4rev2 always returns ESEARCH, even with absent RETURN.
|
||||
if c.enabled[capIMAP4rev2] {
|
||||
eargs = map[string]bool{}
|
||||
}
|
||||
// ../rfc/9051:6967
|
||||
if p.take(" RETURN (") {
|
||||
eargs = map[string]bool{}
|
||||
|
||||
for !p.take(")") {
|
||||
if len(eargs) > 0 || save {
|
||||
p.xspace()
|
||||
}
|
||||
if w, ok := p.takelist("MIN", "MAX", "ALL", "COUNT", "SAVE"); ok {
|
||||
if w == "SAVE" {
|
||||
save = true
|
||||
} else {
|
||||
eargs[w] = true
|
||||
}
|
||||
} else {
|
||||
// ../rfc/4466:378 ../rfc/9051:3745
|
||||
xsyntaxErrorf("ESEARCH result option %q not supported", w)
|
||||
}
|
||||
}
|
||||
}
|
||||
// ../rfc/4731:149 ../rfc/9051:3737
|
||||
if eargs != nil && len(eargs) == 0 && !save {
|
||||
eargs["ALL"] = true
|
||||
}
|
||||
|
||||
// If UTF8=ACCEPT is enabled, we should not accept any charset. We are a bit more
|
||||
// relaxed (reasonable?) and still allow US-ASCII and UTF-8. ../rfc/6855:198
|
||||
if p.take(" CHARSET ") {
|
||||
charset := strings.ToUpper(p.xastring())
|
||||
if charset != "US-ASCII" && charset != "UTF-8" {
|
||||
// ../rfc/3501:2771 ../rfc/9051:3836
|
||||
xusercodeErrorf("BADCHARSET", "only US-ASCII and UTF-8 supported")
|
||||
}
|
||||
}
|
||||
p.xspace()
|
||||
sk := &searchKey{
|
||||
searchKeys: []searchKey{*p.xsearchKey()},
|
||||
}
|
||||
for !p.empty() {
|
||||
p.xspace()
|
||||
sk.searchKeys = append(sk.searchKeys, *p.xsearchKey())
|
||||
}
|
||||
|
||||
// Even in case of error, we ensure search result is changed.
|
||||
if save {
|
||||
c.searchResult = []store.UID{}
|
||||
}
|
||||
|
||||
// Note: we only hold the account rlock for verifying the mailbox at the start.
|
||||
c.account.RLock()
|
||||
runlock := c.account.RUnlock
|
||||
// Note: in a defer because we replace it below.
|
||||
defer func() {
|
||||
runlock()
|
||||
}()
|
||||
|
||||
// If we only have a MIN and/or MAX, we can stop processing as soon as we
|
||||
// have those matches.
|
||||
var min, max int
|
||||
if eargs["MIN"] {
|
||||
min = 1
|
||||
}
|
||||
if eargs["MAX"] {
|
||||
max = 1
|
||||
}
|
||||
|
||||
var expungeIssued bool
|
||||
|
||||
var uids []store.UID
|
||||
c.xdbread(func(tx *bstore.Tx) {
|
||||
c.xmailboxID(tx, c.mailboxID) // Validate.
|
||||
runlock()
|
||||
runlock = func() {}
|
||||
|
||||
// Normal forward search when we don't have MAX only.
|
||||
var lastIndex = -1
|
||||
if eargs == nil || max == 0 || len(eargs) != 1 {
|
||||
for i, uid := range c.uids {
|
||||
lastIndex = i
|
||||
if c.searchMatch(tx, msgseq(i+1), uid, *sk, &expungeIssued) {
|
||||
uids = append(uids, uid)
|
||||
if min == 1 && min+max == len(eargs) {
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
// And reverse search for MAX if we have only MAX or MAX combined with MIN.
|
||||
if max == 1 && (len(eargs) == 1 || min+max == len(eargs)) {
|
||||
for i := len(c.uids) - 1; i > lastIndex; i-- {
|
||||
if c.searchMatch(tx, msgseq(i+1), c.uids[i], *sk, &expungeIssued) {
|
||||
uids = append(uids, c.uids[i])
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
})
|
||||
|
||||
if eargs == nil {
|
||||
// Old-style SEARCH response. We must spell out each number. So we may be splitting
|
||||
// into multiple responses. ../rfc/9051:6809 ../rfc/3501:4833
|
||||
for len(uids) > 0 {
|
||||
n := len(uids)
|
||||
if n > 100 {
|
||||
n = 100
|
||||
}
|
||||
s := ""
|
||||
for _, v := range uids[:n] {
|
||||
if !isUID {
|
||||
v = store.UID(c.xsequence(v))
|
||||
}
|
||||
s += " " + fmt.Sprintf("%d", v)
|
||||
}
|
||||
uids = uids[n:]
|
||||
c.bwritelinef("* SEARCH%s", s)
|
||||
}
|
||||
} else {
|
||||
// New-style ESEARCH response. ../rfc/9051:6546 ../rfc/4466:522
|
||||
|
||||
if save {
|
||||
// ../rfc/9051:3784 ../rfc/5182:13
|
||||
c.searchResult = uids
|
||||
if sanityChecks {
|
||||
checkUIDs(c.searchResult)
|
||||
}
|
||||
}
|
||||
|
||||
// No untagged ESEARCH response if nothing was requested. ../rfc/9051:4160
|
||||
if len(eargs) > 0 {
|
||||
resp := fmt.Sprintf("* ESEARCH (TAG %s)", tag)
|
||||
if isUID {
|
||||
resp += " UID"
|
||||
}
|
||||
|
||||
// NOTE: we are converting UIDs to msgseq in the uids slice (if needed) while
|
||||
// keeping the "uids" name!
|
||||
if !isUID {
|
||||
// If searchResult is hanging on to the slice, we need to work on a copy.
|
||||
if save {
|
||||
nuids := make([]store.UID, len(uids))
|
||||
copy(nuids, uids)
|
||||
uids = nuids
|
||||
}
|
||||
for i, uid := range uids {
|
||||
uids[i] = store.UID(c.xsequence(uid))
|
||||
}
|
||||
}
|
||||
|
||||
// If no matches, then no MIN/MAX response. ../rfc/4731:98 ../rfc/9051:3758
|
||||
if eargs["MIN"] && len(uids) > 0 {
|
||||
resp += fmt.Sprintf(" MIN %d", uids[0])
|
||||
}
|
||||
if eargs["MAX"] && len(uids) > 0 {
|
||||
resp += fmt.Sprintf(" MAX %d", uids[len(uids)-1])
|
||||
}
|
||||
if eargs["COUNT"] {
|
||||
resp += fmt.Sprintf(" COUNT %d", len(uids))
|
||||
}
|
||||
if eargs["ALL"] && len(uids) > 0 {
|
||||
resp += fmt.Sprintf(" ALL %s", compactUIDSet(uids).String())
|
||||
}
|
||||
c.bwritelinef("%s", resp)
|
||||
}
|
||||
}
|
||||
if expungeIssued {
|
||||
// ../rfc/9051:5102
|
||||
c.writeresultf("%s OK [EXPUNGEISSUED] done", tag)
|
||||
} else {
|
||||
c.ok(tag, cmd)
|
||||
}
|
||||
}
|
||||
|
||||
type search struct {
|
||||
c *conn
|
||||
tx *bstore.Tx
|
||||
seq msgseq
|
||||
uid store.UID
|
||||
mr *store.MsgReader
|
||||
m store.Message
|
||||
p *message.Part
|
||||
expungeIssued *bool
|
||||
}
|
||||
|
||||
func (c *conn) searchMatch(tx *bstore.Tx, seq msgseq, uid store.UID, sk searchKey, expungeIssued *bool) bool {
|
||||
s := search{c: c, tx: tx, seq: seq, uid: uid, expungeIssued: expungeIssued}
|
||||
defer func() {
|
||||
if s.mr != nil {
|
||||
err := s.mr.Close()
|
||||
c.xsanity(err, "closing messagereader")
|
||||
s.mr = nil
|
||||
}
|
||||
}()
|
||||
return s.match(sk)
|
||||
}
|
||||
|
||||
func (s *search) match(sk searchKey) bool {
|
||||
c := s.c
|
||||
|
||||
if sk.searchKeys != nil {
|
||||
for _, ssk := range sk.searchKeys {
|
||||
if !s.match(ssk) {
|
||||
return false
|
||||
}
|
||||
}
|
||||
return true
|
||||
} else if sk.seqSet != nil {
|
||||
return sk.seqSet.containsSeq(s.seq, c.uids, c.searchResult)
|
||||
}
|
||||
|
||||
filterHeader := func(field, value string) bool {
|
||||
lower := strings.ToLower(value)
|
||||
h, err := s.p.Header()
|
||||
if err != nil {
|
||||
c.log.Debugx("parsing message header", err, mlog.Field("uid", s.uid))
|
||||
return false
|
||||
}
|
||||
for _, v := range h.Values(field) {
|
||||
if strings.Contains(strings.ToLower(v), lower) {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
// We handle ops by groups that need increasing details about the message.
|
||||
|
||||
switch sk.op {
|
||||
case "ALL":
|
||||
return true
|
||||
case "NEW":
|
||||
// We do not implement the RECENT flag, so messages cannot be NEW.
|
||||
return false
|
||||
case "OLD":
|
||||
// We treat all messages as non-recent, so this means all messages.
|
||||
return true
|
||||
case "RECENT":
|
||||
// We do not implement the RECENT flag. All messages are not recent.
|
||||
return false
|
||||
case "NOT":
|
||||
return !s.match(*sk.searchKey)
|
||||
case "OR":
|
||||
return s.match(*sk.searchKey) || s.match(*sk.searchKey2)
|
||||
case "UID":
|
||||
return sk.uidSet.containsUID(s.uid, c.uids, c.searchResult)
|
||||
}
|
||||
|
||||
// Parsed message.
|
||||
if s.mr == nil {
|
||||
q := bstore.QueryTx[store.Message](s.tx)
|
||||
q.FilterNonzero(store.Message{MailboxID: c.mailboxID, UID: s.uid})
|
||||
m, err := q.Get()
|
||||
if err == bstore.ErrAbsent {
|
||||
// ../rfc/2180:607
|
||||
*s.expungeIssued = true
|
||||
return false
|
||||
}
|
||||
xcheckf(err, "get message")
|
||||
s.m = m
|
||||
|
||||
// Closed by searchMatch after all (recursive) search.match calls are finished.
|
||||
s.mr = c.account.MessageReader(m)
|
||||
|
||||
if m.ParsedBuf == nil {
|
||||
c.log.Error("missing parsed message")
|
||||
} else {
|
||||
p, err := m.LoadPart(s.mr)
|
||||
xcheckf(err, "load parsed message")
|
||||
s.p = &p
|
||||
}
|
||||
}
|
||||
|
||||
// Parsed message, basic info.
|
||||
switch sk.op {
|
||||
case "ANSWERED":
|
||||
return s.m.Answered
|
||||
case "DELETED":
|
||||
return s.m.Deleted
|
||||
case "FLAGGED":
|
||||
return s.m.Flagged
|
||||
case "KEYWORD":
|
||||
switch sk.atom {
|
||||
case "$Forwarded":
|
||||
return s.m.Forwarded
|
||||
case "$Junk":
|
||||
return s.m.Junk
|
||||
case "$NotJunk":
|
||||
return s.m.Notjunk
|
||||
case "$Phishing":
|
||||
return s.m.Phishing
|
||||
case "$MDNSent":
|
||||
return s.m.MDNSent
|
||||
default:
|
||||
c.log.Info("search with unknown keyword", mlog.Field("keyword", sk.atom))
|
||||
return false
|
||||
}
|
||||
case "SEEN":
|
||||
return s.m.Seen
|
||||
case "UNANSWERED":
|
||||
return !s.m.Answered
|
||||
case "UNDELETED":
|
||||
return !s.m.Deleted
|
||||
case "UNFLAGGED":
|
||||
return !s.m.Flagged
|
||||
case "UNKEYWORD":
|
||||
switch sk.atom {
|
||||
case "$Forwarded":
|
||||
return !s.m.Forwarded
|
||||
case "$Junk":
|
||||
return !s.m.Junk
|
||||
case "$NotJunk":
|
||||
return !s.m.Notjunk
|
||||
case "$Phishing":
|
||||
return !s.m.Phishing
|
||||
case "$MDNSent":
|
||||
return !s.m.MDNSent
|
||||
default:
|
||||
c.log.Info("search with unknown keyword", mlog.Field("keyword", sk.atom))
|
||||
return false
|
||||
}
|
||||
case "UNSEEN":
|
||||
return !s.m.Seen
|
||||
case "DRAFT":
|
||||
return s.m.Draft
|
||||
case "UNDRAFT":
|
||||
return !s.m.Draft
|
||||
case "BEFORE", "ON", "SINCE":
|
||||
skdt := sk.date.Format("2006-01-02")
|
||||
rdt := s.m.Received.Format("2006-01-02")
|
||||
switch sk.op {
|
||||
case "BEFORE":
|
||||
return rdt < skdt
|
||||
case "ON":
|
||||
return rdt == skdt
|
||||
case "SINCE":
|
||||
return rdt >= skdt
|
||||
}
|
||||
panic("missing case")
|
||||
case "LARGER":
|
||||
return s.m.Size > sk.number
|
||||
case "SMALLER":
|
||||
return s.m.Size < sk.number
|
||||
}
|
||||
|
||||
if s.p == nil {
|
||||
c.log.Info("missing parsed message, not matching", mlog.Field("uid", s.uid))
|
||||
return false
|
||||
}
|
||||
|
||||
// Parsed message, more info.
|
||||
switch sk.op {
|
||||
case "BCC":
|
||||
return filterHeader("Bcc", sk.astring)
|
||||
case "BODY", "TEXT":
|
||||
headerToo := sk.op == "TEXT"
|
||||
lower := strings.ToLower(sk.astring)
|
||||
return mailContains(c, s.uid, s.p, lower, headerToo)
|
||||
case "CC":
|
||||
return filterHeader("Cc", sk.astring)
|
||||
case "FROM":
|
||||
return filterHeader("From", sk.astring)
|
||||
case "SUBJECT":
|
||||
return filterHeader("Subject", sk.astring)
|
||||
case "TO":
|
||||
return filterHeader("To", sk.astring)
|
||||
case "HEADER":
|
||||
// ../rfc/9051:3895
|
||||
lower := strings.ToLower(sk.astring)
|
||||
h, err := s.p.Header()
|
||||
if err != nil {
|
||||
c.log.Errorx("parsing header for search", err, mlog.Field("uid", s.uid))
|
||||
return false
|
||||
}
|
||||
k := textproto.CanonicalMIMEHeaderKey(sk.headerField)
|
||||
for _, v := range h.Values(k) {
|
||||
if lower == "" || strings.Contains(strings.ToLower(v), lower) {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
case "SENTBEFORE", "SENTON", "SENTSINCE":
|
||||
if s.p.Envelope == nil || s.p.Envelope.Date.IsZero() {
|
||||
return false
|
||||
}
|
||||
dt := s.p.Envelope.Date.Format("2006-01-02")
|
||||
skdt := sk.date.Format("2006-01-02")
|
||||
switch sk.op {
|
||||
case "SENTBEFORE":
|
||||
return dt < skdt
|
||||
case "SENTON":
|
||||
return dt == skdt
|
||||
case "SENTSINCE":
|
||||
return dt > skdt
|
||||
}
|
||||
panic("missing case")
|
||||
}
|
||||
panic(serverError{fmt.Errorf("missing case for search key op %q", sk.op)})
|
||||
}
|
||||
|
||||
// mailContains returns whether the mail message or part represented by p contains (case-insensitive) string lower.
|
||||
// The (decoded) text bodies are tested for a match.
|
||||
// If headerToo is set, the header part of the message is checked as well.
|
||||
func mailContains(c *conn, uid store.UID, p *message.Part, lower string, headerToo bool) bool {
|
||||
if headerToo && mailContainsReader(c, uid, p.HeaderReader(), lower) {
|
||||
return true
|
||||
}
|
||||
|
||||
if len(p.Parts) == 0 {
|
||||
if p.MediaType != "TEXT" {
|
||||
// todo: for types we could try to find a library for parsing and search in there too
|
||||
return false
|
||||
}
|
||||
// todo: for html and perhaps other types, we could try to parse as text and filter on the text.
|
||||
return mailContainsReader(c, uid, p.Reader(), lower)
|
||||
}
|
||||
for _, pp := range p.Parts {
|
||||
headerToo = pp.MediaType == "MESSAGE" && (pp.MediaSubType == "RFC822" || pp.MediaSubType == "GLOBAL")
|
||||
if mailContains(c, uid, &pp, lower, headerToo) {
|
||||
return true
|
||||
}
|
||||
}
|
||||
return false
|
||||
}
|
||||
|
||||
func mailContainsReader(c *conn, uid store.UID, r io.Reader, lower string) bool {
|
||||
// todo: match as we read
|
||||
buf, err := io.ReadAll(r)
|
||||
if err != nil {
|
||||
c.log.Errorx("reading for search text match", err, mlog.Field("uid", uid))
|
||||
return false
|
||||
}
|
||||
return strings.Contains(strings.ToLower(string(buf)), lower)
|
||||
}
|
Reference in New Issue
Block a user