autobrr/internal/announce/announce.go
ze0s 25a165b764
feat(indexers): add support for optional baseurl override (#551)
* feat(indexers): optional baseurl override

* feat(indexers): update baseUrl parsing

* refactor(indexers): BREAKING move parse to IRC struct

* Move Parse as part of IRC struct from Indexer
* Updated definitions
* Build torrentUrl in stages
* Use new url.JoinPath to build torrentUrl
* Update tests

* refactor(indexers): select option obj

* refactor(indexers): make backwards compatible
2022-12-03 15:40:45 +01:00

324 lines
7.3 KiB
Go

package announce
import (
"bytes"
"net/url"
"regexp"
"strings"
"text/template"
"github.com/autobrr/autobrr/internal/domain"
"github.com/autobrr/autobrr/internal/release"
"github.com/autobrr/autobrr/pkg/errors"
"github.com/rs/zerolog"
)
type Processor interface {
AddLineToQueue(channel string, line string) error
}
type announceProcessor struct {
log zerolog.Logger
indexer *domain.IndexerDefinition
releaseSvc release.Service
queues map[string]chan string
}
func NewAnnounceProcessor(log zerolog.Logger, releaseSvc release.Service, indexer *domain.IndexerDefinition) Processor {
ap := &announceProcessor{
log: log.With().Str("module", "announce_processor").Logger(),
releaseSvc: releaseSvc,
indexer: indexer,
}
// setup queues and consumers
ap.setupQueues()
ap.setupQueueConsumers()
return ap
}
func (a *announceProcessor) setupQueues() {
queues := make(map[string]chan string)
for _, channel := range a.indexer.IRC.Channels {
channel = strings.ToLower(channel)
queues[channel] = make(chan string, 128)
a.log.Trace().Msgf("announce: setup queue: %v", channel)
}
a.queues = queues
}
func (a *announceProcessor) setupQueueConsumers() {
for queueName, queue := range a.queues {
go func(name string, q chan string) {
a.log.Trace().Msgf("announce: setup queue consumer: %v", name)
a.processQueue(q)
a.log.Trace().Msgf("announce: queue consumer stopped: %v", name)
}(queueName, queue)
}
}
func (a *announceProcessor) processQueue(queue chan string) {
for {
tmpVars := map[string]string{}
parseFailed := false
//patternParsed := false
for _, parseLine := range a.indexer.IRC.Parse.Lines {
line, err := a.getNextLine(queue)
if err != nil {
a.log.Error().Err(err).Msg("could not get line from queue")
return
}
a.log.Trace().Msgf("announce: process line: %v", line)
// check should ignore
match, err := a.parseLine(parseLine.Pattern, parseLine.Vars, tmpVars, line)
if err != nil {
a.log.Error().Err(err).Msgf("error parsing extract for line: %v", line)
parseFailed = true
break
}
if !match {
a.log.Debug().Msgf("line not matching expected regex pattern: %v", line)
parseFailed = true
break
}
}
if parseFailed {
continue
}
rls := domain.NewRelease(a.indexer.Identifier)
// on lines matched
if err := a.onLinesMatched(a.indexer, tmpVars, rls); err != nil {
a.log.Error().Err(err).Msg("error match line")
continue
}
// process release in a new go routine
go a.releaseSvc.Process(rls)
}
}
func (a *announceProcessor) getNextLine(queue chan string) (string, error) {
for {
line, ok := <-queue
if !ok {
return "", errors.New("could not queue line")
}
return line, nil
}
}
func (a *announceProcessor) AddLineToQueue(channel string, line string) error {
channel = strings.ToLower(channel)
queue, ok := a.queues[channel]
if !ok {
return errors.New("no queue for channel (%v) found", channel)
}
queue <- line
a.log.Trace().Msgf("announce: queued line: %v", line)
return nil
}
func (a *announceProcessor) parseLine(pattern string, vars []string, tmpVars map[string]string, line string) (bool, error) {
if len(vars) > 0 {
return a.parseExtract(pattern, vars, tmpVars, line)
}
return a.parseMatchRegexp(pattern, tmpVars, line)
}
func (a *announceProcessor) parseExtract(pattern string, vars []string, tmpVars map[string]string, line string) (bool, error) {
rxp, err := regExMatch(pattern, line)
if err != nil {
a.log.Debug().Msgf("did not match expected line: %v", line)
}
if rxp == nil {
return false, nil
}
// extract matched
for i, v := range vars {
value := ""
if rxp[i] != "" {
value = rxp[i]
// tmpVars[v] = rxp[i]
}
tmpVars[v] = value
}
return true, nil
}
func (a *announceProcessor) parseMatchRegexp(pattern string, tmpVars map[string]string, line string) (bool, error) {
var re = regexp.MustCompile(`(?mi)` + pattern)
groupNames := re.SubexpNames()
for _, match := range re.FindAllStringSubmatch(line, -1) {
for groupIdx, group := range match {
name := groupNames[groupIdx]
if name == "" {
name = "raw"
}
tmpVars[name] = group
}
}
return true, nil
}
// onLinesMatched process vars into release
func (a *announceProcessor) onLinesMatched(def *domain.IndexerDefinition, vars map[string]string, rls *domain.Release) error {
// map variables from regex capture onto release struct
if err := rls.MapVars(def, vars); err != nil {
a.log.Error().Err(err).Msg("announce: could not map vars for release")
return err
}
// parse fields
// run before ParseMatch to not potentially use a reconstructed TorrentName
rls.ParseString(rls.TorrentName)
// set baseUrl to default domain
baseUrl := def.URLS[0]
// override baseUrl
if def.BaseURL != "" {
baseUrl = def.BaseURL
}
// merge vars from regex captures on announce and vars from settings
mergedVars := mergeVars(vars, def.SettingsMap)
// parse torrentUrl
matched, err := def.IRC.Parse.ParseMatch(baseUrl, mergedVars)
if err != nil {
a.log.Error().Err(err).Msgf("announce: %v", err)
return err
}
if matched != nil {
rls.TorrentURL = matched.TorrentURL
// only used by few indexers
if matched.TorrentName != "" {
rls.TorrentName = matched.TorrentName
}
}
// handle optional cookies
if v, ok := def.SettingsMap["cookie"]; ok {
rls.RawCookie = v
}
return nil
}
func (a *announceProcessor) processTorrentUrl(match string, vars map[string]string, extraVars map[string]string, encode []string) (string, error) {
tmpVars := map[string]string{}
// copy vars to new tmp map
for k, v := range vars {
tmpVars[k] = v
}
// merge extra vars with vars
for k, v := range extraVars {
tmpVars[k] = v
}
// handle url encode of values
for _, e := range encode {
if v, ok := tmpVars[e]; ok {
// url encode value
t := url.QueryEscape(v)
tmpVars[e] = t
}
}
// setup text template to inject variables into
tmpl, err := template.New("torrenturl").Parse(match)
if err != nil {
a.log.Error().Err(err).Msg("could not create torrent url template")
return "", err
}
var b bytes.Buffer
if err := tmpl.Execute(&b, &tmpVars); err != nil {
a.log.Error().Err(err).Msg("could not write torrent url template output")
return "", err
}
a.log.Trace().Msg("torrenturl processed")
return b.String(), nil
}
// mergeVars merge maps
func mergeVars(data ...map[string]string) map[string]string {
tmpVars := map[string]string{}
for _, vars := range data {
// copy vars to new tmp map
for k, v := range vars {
tmpVars[k] = v
}
}
return tmpVars
}
func removeElement(s []string, i int) ([]string, error) {
// s is [1,2,3,4,5,6], i is 2
// perform bounds checking first to prevent a panic!
if i >= len(s) || i < 0 {
return nil, errors.New("Index is out of range. Index is %d with slice length %d", i, len(s))
}
// This creates a new slice by creating 2 slices from the original:
// s[:i] -> [1, 2]
// s[i+1:] -> [4, 5, 6]
// and joining them together using `append`
return append(s[:i], s[i+1:]...), nil
}
func regExMatch(pattern string, value string) ([]string, error) {
rxp, err := regexp.Compile(pattern)
if err != nil {
return nil, err
}
matches := rxp.FindStringSubmatch(value)
if matches == nil {
return nil, nil
}
res := make([]string, 0)
if matches != nil {
res, err = removeElement(matches, 0)
if err != nil {
return nil, err
}
}
return res, nil
}