feat(filters): skip duplicates (#1711)

* feat(filters): skip duplicates

* fix: add interface instead of any

* fix(filters): tonullint

* feat(filters): skip dupes check month day

* chore: cleanup

* feat(db): set autoincrement id

* feat(filters): add repack and proper to dupe profile

* feat(filters): add default dupe profiles

* feat(duplicates): check audio and website

* feat(duplicates): update tests

* feat(duplicates): add toggles on addform

* feat(duplicates): fix sqlite upgrade path and initialize duplicate profiles

* feat(duplicates): simplify sqlite upgrade

avoiding temp table and unwieldy select.  Besides, FK constraints
are turned off anyway in #229.

* feat(duplicates): change CheckIsDuplicateRelease treatment of PROPER and REPACK

"Proper" and "Repack" are not parallel to the other conditions like "Title",
so they do not belong as dedup conditions.  "PROPER" means there was an issue in
the previous release, and so a PROPER is never a duplicate, even if it replaces
another PROPER.  Similarly, "REPACK" means there was an issue in the previous
release by that group, and so it is a duplicate only if we previously took a
release from a DIFFERENT group.

I have not removed Proper and Repack from the UI or the schema yet.

* feat(duplicates): update postgres schema to match sqlite

* feat(duplicates): fix web build errors

* feat(duplicates): fix postgres errors

* feat(filters): do leftjoin for duplicate profile

* fix(filters): partial update dupe profile

* go fmt `internal/domain/filter.go`

* feat(duplicates): restore straightforward logic for proper/repack

* feat(duplicates): remove mostly duplicate TV duplicate profiles

Having one profile seems the cleanest.  If somebody wants multiple
resolutions then they can add Resolution to the duplicate profile.
Tested this profile with both weekly episodic releases and daily
show releases.

* feat(release): add db indexes and sub_title

* feat(release): add IsDuplicate tests

* feat(release): update action handler

* feat(release): add more tests for skip duplicates

* feat(duplicates): check audio

* feat(duplicates): add more tests

* feat(duplicates): match edition cut and more

* fix(duplicates): tests

* fix(duplicates): missing imports

* fix(duplicates): tests

* feat(duplicates): handle sub_title edition and language in ui

* fix(duplicates): tests

* feat(duplicates): check name against normalized hash

* fix(duplicates): tests

* chore: update .gitignore to ignore .pnpm-store

* fix: tests

* fix(filters): tests

* fix: bad conflict merge

* fix: update release type in test

* fix: use vendored hot-toast

* fix: release_test.go

* fix: rss_test.go

* feat(duplicates): improve title hashing for unique check

* feat(duplicates): further improve title hashing for unique check with lang

* feat(duplicates): fix tests

* feat(duplicates): add macros IsDuplicate and DuplicateProfile ID and name

* feat(duplicates): add normalized hash match option

* fix: headlessui-state prop warning

* fix(duplicates): add missing year in daily ep normalize

* fix(duplicates): check rejections len

---------

Co-authored-by: ze0s <ze0s@riseup.net>
This commit is contained in:
kenstir 2024-12-25 16:33:46 -05:00 committed by GitHub
parent d153ac44b8
commit 4009554d10
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
49 changed files with 3792 additions and 743 deletions

1
.gitignore vendored
View file

@ -29,6 +29,7 @@ node_modules/
!.yarn/versions
# Ditto for yarn, except we're using pnpm
yarn.lock
.pnpm-store
# If needed, package-lock.json shall be added manually using an explicit git add command
package-lock.json

View file

@ -21,7 +21,7 @@ import (
)
type Service interface {
Store(ctx context.Context, action domain.Action) (*domain.Action, error)
Store(ctx context.Context, action *domain.Action) error
StoreFilterActions(ctx context.Context, filterID int64, actions []*domain.Action) ([]*domain.Action, error)
List(ctx context.Context) ([]domain.Action, error)
Get(ctx context.Context, req *domain.GetActionRequest) (*domain.Action, error)
@ -63,7 +63,7 @@ func NewService(log logger.Logger, repo domain.ActionRepo, clientSvc download_cl
return s
}
func (s *service) Store(ctx context.Context, action domain.Action) (*domain.Action, error) {
func (s *service) Store(ctx context.Context, action *domain.Action) error {
return s.repo.Store(ctx, action)
}

View file

@ -706,7 +706,7 @@ func (r *ActionRepo) DeleteByFilterID(ctx context.Context, filterID int) error {
return nil
}
func (r *ActionRepo) Store(ctx context.Context, action domain.Action) (*domain.Action, error) {
func (r *ActionRepo) Store(ctx context.Context, action *domain.Action) error {
queryBuilder := r.db.squirrel.
Insert("action").
Columns(
@ -783,14 +783,14 @@ func (r *ActionRepo) Store(ctx context.Context, action domain.Action) (*domain.A
var retID int64
if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil {
return nil, errors.Wrap(err, "error executing query")
return errors.Wrap(err, "error executing query")
}
action.ID = int(retID)
r.log.Debug().Msgf("action.store: added new %d", retID)
return &action, nil
return nil
}
func (r *ActionRepo) Update(ctx context.Context, action domain.Action) (*domain.Action, error) {

View file

@ -16,8 +16,8 @@ import (
"github.com/stretchr/testify/assert"
)
func getMockAction() domain.Action {
return domain.Action{
func getMockAction() *domain.Action {
return &domain.Action{
Name: "randomAction",
Type: domain.ActionTypeTest,
Enabled: true,
@ -78,29 +78,29 @@ func TestActionRepo_Store(t *testing.T) {
mockData.FilterID = createdFilters[0].ID
// Actual test for Store
createdAction, err := repo.Store(context.Background(), mockData)
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
assert.NotNil(t, createdAction)
assert.NotNil(t, mockData)
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
t.Run(fmt.Sprintf("Store_Succeeds_With_Missing_or_empty_fields [%s]", dbType), func(t *testing.T) {
mockData := domain.Action{}
createdAction, err := repo.Store(context.Background(), mockData)
mockData := &domain.Action{}
err := repo.Store(context.Background(), mockData)
assert.NoError(t, err)
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockData.ID})
})
t.Run(fmt.Sprintf("Store_Fails_With_Invalid_ClientID [%s]", dbType), func(t *testing.T) {
mockData := getMockAction()
mockData.ClientID = 9999
_, err := repo.Store(context.Background(), mockData)
err := repo.Store(context.Background(), mockData)
assert.Error(t, err)
})
@ -110,7 +110,7 @@ func TestActionRepo_Store(t *testing.T) {
ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond)
defer cancel()
_, err := repo.Store(ctx, mockData)
err := repo.Store(ctx, mockData)
assert.Error(t, err)
})
}
@ -142,7 +142,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) {
mockData.FilterID = createdFilters[0].ID
// Actual test for StoreFilterActions
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
assert.NotNil(t, createdActions)
@ -154,7 +154,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) {
})
t.Run(fmt.Sprintf("StoreFilterActions_Fails_Invalid_FilterID [%s]", dbType), func(t *testing.T) {
_, err := repo.StoreFilterActions(context.Background(), 9999, []*domain.Action{&mockData})
_, err := repo.StoreFilterActions(context.Background(), 9999, []*domain.Action{mockData})
assert.NoError(t, err)
})
@ -186,7 +186,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) {
assert.NoError(t, err)
assert.NotNil(t, createdFilters)
_, err = repo.StoreFilterActions(ctx, int64(createdFilters[0].ID), []*domain.Action{&mockData})
_, err = repo.StoreFilterActions(ctx, int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.Error(t, err)
// Cleanup
@ -219,7 +219,7 @@ func TestActionRepo_FindByFilterID(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
// Actual test for FindByFilterID
@ -294,7 +294,7 @@ func TestActionRepo_List(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
// Actual test for List
@ -344,7 +344,7 @@ func TestActionRepo_Get(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
// Actual test for Get
@ -401,7 +401,7 @@ func TestActionRepo_Delete(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
// Actual test for Delete
@ -455,7 +455,7 @@ func TestActionRepo_DeleteByFilterID(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
err = repo.DeleteByFilterID(context.Background(), mockData.FilterID)
@ -508,7 +508,7 @@ func TestActionRepo_ToggleEnabled(t *testing.T) {
mockData.ClientID = mock.ID
mockData.FilterID = createdFilters[0].ID
mockData.Enabled = false
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData})
createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData})
assert.NoError(t, err)
// Actual test for ToggleEnabled

View file

@ -129,13 +129,9 @@ type Tx struct {
handler *DB
}
type ILikeDynamic interface {
ToSql() (sql string, args []interface{}, err error)
}
// ILike is a wrapper for sq.Like and sq.ILike
// SQLite does not support ILike but postgres does so this checks what database is being used
func (db *DB) ILike(col string, val string) ILikeDynamic {
func (db *DB) ILike(col string, val string) sq.Sqlizer {
//if databaseDriver == "sqlite" {
if db.Driver == "sqlite" {
return sq.Like{col: val}

View file

@ -240,6 +240,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter
"f.max_seeders",
"f.min_leechers",
"f.max_leechers",
"f.release_profile_duplicate_id",
"f.created_at",
"f.updated_at",
).
@ -266,6 +267,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter
var minSize, maxSize, maxDownloadsUnit, matchReleases, exceptReleases, matchReleaseGroups, exceptReleaseGroups, matchReleaseTags, exceptReleaseTags, matchDescription, exceptDescription, freeleechPercent, shows, seasons, episodes, years, months, days, artists, albums, matchCategories, exceptCategories, matchUploaders, exceptUploaders, matchRecordLabels, exceptRecordLabels, tags, exceptTags, tagsMatchLogic, exceptTagsMatchLogic sql.NullString
var useRegex, scene, freeleech, hasLog, hasCue, perfectFlac sql.NullBool
var delay, maxDownloads, logScore sql.NullInt32
var releaseProfileDuplicateId sql.NullInt64
err = row.Scan(
&f.ID,
@ -335,6 +337,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter
&f.MaxSeeders,
&f.MinLeechers,
&f.MaxLeechers,
&releaseProfileDuplicateId,
&f.CreatedAt,
&f.UpdatedAt,
)
@ -385,6 +388,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter
f.UseRegex = useRegex.Bool
f.Scene = scene.Bool
f.Freeleech = freeleech.Bool
f.ReleaseProfileDuplicateID = releaseProfileDuplicateId.Int64
return &f, nil
}
@ -466,10 +470,35 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string
"f.max_leechers",
"f.created_at",
"f.updated_at",
"f.release_profile_duplicate_id",
"rdp.id",
"rdp.name",
"rdp.release_name",
"rdp.hash",
"rdp.title",
"rdp.sub_title",
"rdp.year",
"rdp.month",
"rdp.day",
"rdp.source",
"rdp.resolution",
"rdp.codec",
"rdp.container",
"rdp.dynamic_range",
"rdp.audio",
"rdp.release_group",
"rdp.season",
"rdp.episode",
"rdp.website",
"rdp.proper",
"rdp.repack",
"rdp.edition",
"rdp.language",
).
From("filter f").
Join("filter_indexer fi ON f.id = fi.filter_id").
Join("indexer i ON i.id = fi.indexer_id").
LeftJoin("release_profile_duplicate rdp ON rdp.id = f.release_profile_duplicate_id").
Where(sq.Eq{"i.identifier": indexer}).
Where(sq.Eq{"i.enabled": true}).
Where(sq.Eq{"f.enabled": true}).
@ -495,6 +524,10 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string
var minSize, maxSize, maxDownloadsUnit, matchReleases, exceptReleases, matchReleaseGroups, exceptReleaseGroups, matchReleaseTags, exceptReleaseTags, matchDescription, exceptDescription, freeleechPercent, shows, seasons, episodes, years, months, days, artists, albums, matchCategories, exceptCategories, matchUploaders, exceptUploaders, matchRecordLabels, exceptRecordLabels, tags, exceptTags, tagsMatchLogic, exceptTagsMatchLogic sql.NullString
var useRegex, scene, freeleech, hasLog, hasCue, perfectFlac sql.NullBool
var delay, maxDownloads, logScore sql.NullInt32
var releaseProfileDuplicateID, rdpId sql.NullInt64
var rdpName sql.NullString
var rdpRelName, rdpHash, rdpTitle, rdpSubTitle, rdpYear, rdpMonth, rdpDay, rdpSource, rdpResolution, rdpCodec, rdpContainer, rdpDynRange, rdpAudio, rdpGroup, rdpSeason, rdpEpisode, rdpWebsite, rdpProper, rdpRepack, rdpEdition, rdpLanguage sql.NullBool
err := rows.Scan(
&f.ID,
@ -566,6 +599,30 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string
&f.MaxLeechers,
&f.CreatedAt,
&f.UpdatedAt,
&releaseProfileDuplicateID,
&rdpId,
&rdpName,
&rdpRelName,
&rdpHash,
&rdpTitle,
&rdpSubTitle,
&rdpYear,
&rdpMonth,
&rdpDay,
&rdpSource,
&rdpResolution,
&rdpCodec,
&rdpContainer,
&rdpDynRange,
&rdpAudio,
&rdpGroup,
&rdpSeason,
&rdpEpisode,
&rdpWebsite,
&rdpProper,
&rdpRepack,
&rdpEdition,
&rdpLanguage,
)
if err != nil {
return nil, errors.Wrap(err, "error scanning row")
@ -610,9 +667,40 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string
f.UseRegex = useRegex.Bool
f.Scene = scene.Bool
f.Freeleech = freeleech.Bool
f.ReleaseProfileDuplicateID = releaseProfileDuplicateID.Int64
f.Rejections = []string{}
if releaseProfileDuplicateID.Valid {
profile := domain.DuplicateReleaseProfile{
ID: rdpId.Int64,
//Protocol: rdpName.String,
Name: rdpName.String,
ReleaseName: rdpRelName.Bool,
Hash: rdpHash.Bool,
Title: rdpTitle.Bool,
SubTitle: rdpSubTitle.Bool,
Year: rdpYear.Bool,
Month: rdpMonth.Bool,
Day: rdpDay.Bool,
Source: rdpSource.Bool,
Resolution: rdpResolution.Bool,
Codec: rdpCodec.Bool,
Container: rdpContainer.Bool,
DynamicRange: rdpDynRange.Bool,
Audio: rdpAudio.Bool,
Group: rdpGroup.Bool,
Season: rdpSeason.Bool,
Episode: rdpEpisode.Bool,
Website: rdpWebsite.Bool,
Proper: rdpProper.Bool,
Repack: rdpRepack.Bool,
Edition: rdpEdition.Bool,
Language: rdpLanguage.Bool,
}
f.DuplicateHandling = &profile
}
filters = append(filters, &f)
}
@ -774,6 +862,7 @@ func (r *FilterRepo) Store(ctx context.Context, filter *domain.Filter) error {
"max_seeders",
"min_leechers",
"max_leechers",
"release_profile_duplicate_id",
).
Values(
filter.Name,
@ -842,6 +931,7 @@ func (r *FilterRepo) Store(ctx context.Context, filter *domain.Filter) error {
filter.MaxSeeders,
filter.MinLeechers,
filter.MaxLeechers,
toNullInt64(filter.ReleaseProfileDuplicateID),
).
Suffix("RETURNING id").RunWith(r.db.handler)
@ -928,6 +1018,7 @@ func (r *FilterRepo) Update(ctx context.Context, filter *domain.Filter) error {
Set("max_seeders", filter.MaxSeeders).
Set("min_leechers", filter.MinLeechers).
Set("max_leechers", filter.MaxLeechers).
Set("release_profile_duplicate_id", toNullInt64(filter.ReleaseProfileDuplicateID)).
Set("updated_at", time.Now().Format(time.RFC3339)).
Where(sq.Eq{"id": filter.ID})
@ -1153,6 +1244,9 @@ func (r *FilterRepo) UpdatePartial(ctx context.Context, filter domain.FilterUpda
if filter.MaxLeechers != nil {
q = q.Set("max_leechers", filter.MaxLeechers)
}
if filter.ReleaseProfileDuplicateID != nil {
q = q.Set("release_profile_duplicate_id", filter.ReleaseProfileDuplicateID)
}
q = q.Where(sq.Eq{"id": filter.ID})

View file

@ -800,7 +800,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
mockAction.FilterID = mockData.ID
mockAction.ClientID = mockClient.ID
action, err := actionRepo.Store(context.Background(), mockAction)
err = actionRepo.Store(context.Background(), mockAction)
mockReleaseActionStatus.FilterID = int64(mockData.ID)
mockRelease.FilterID = mockData.ID
@ -808,7 +808,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
err = releaseRepo.Store(context.Background(), mockRelease)
assert.NoError(t, err)
mockReleaseActionStatus.ActionID = int64(action.ID)
mockReleaseActionStatus.ActionID = int64(mockAction.ID)
mockReleaseActionStatus.ReleaseID = mockRelease.ID
err = releaseRepo.StoreReleaseActionStatus(context.Background(), mockReleaseActionStatus)
@ -827,7 +827,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
})
// Cleanup
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction.ID})
_ = repo.Delete(context.Background(), mockData.ID)
_ = downloadClientRepo.Delete(context.Background(), mockClient.ID)
_ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
@ -861,13 +861,15 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
mockAction1.FilterID = mockData.ID
mockAction1.ClientID = mockClient.ID
action1, err := actionRepo.Store(context.Background(), mockAction1)
actionErr := actionRepo.Store(context.Background(), mockAction1)
assert.NoError(t, actionErr)
mockAction2 := getMockAction()
mockAction2.FilterID = mockData.ID
mockAction2.ClientID = mockClient.ID
action2, err := actionRepo.Store(context.Background(), mockAction2)
action2Err := actionRepo.Store(context.Background(), mockAction2)
assert.NoError(t, action2Err)
mockRelease.FilterID = mockData.ID
@ -875,7 +877,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
assert.NoError(t, err)
mockReleaseActionStatus1 := getMockReleaseActionStatus()
mockReleaseActionStatus1.ActionID = int64(action1.ID)
mockReleaseActionStatus1.ActionID = int64(mockAction1.ID)
mockReleaseActionStatus1.FilterID = int64(mockData.ID)
mockReleaseActionStatus1.ReleaseID = mockRelease.ID
@ -883,7 +885,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
assert.NoError(t, err)
mockReleaseActionStatus2 := getMockReleaseActionStatus()
mockReleaseActionStatus2.ActionID = int64(action2.ID)
mockReleaseActionStatus2.ActionID = int64(mockAction2.ID)
mockReleaseActionStatus2.FilterID = int64(mockData.ID)
mockReleaseActionStatus2.ReleaseID = mockRelease.ID
@ -903,8 +905,8 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
})
// Cleanup
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action1.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action2.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction1.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction2.ID})
_ = repo.Delete(context.Background(), mockData.ID)
_ = downloadClientRepo.Delete(context.Background(), mockClient.ID)
_ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
@ -924,13 +926,15 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
mockAction.FilterID = mockData.ID
mockAction.ClientID = mockClient.ID
action, err := actionRepo.Store(context.Background(), mockAction)
err = actionRepo.Store(context.Background(), mockAction)
assert.NoError(t, err)
mockAction2 := getMockAction()
mockAction2.FilterID = mockData.ID
mockAction2.ClientID = mockClient.ID
action2, err := actionRepo.Store(context.Background(), mockAction2)
err = actionRepo.Store(context.Background(), mockAction2)
assert.NoError(t, err)
mockRelease.FilterID = mockData.ID
@ -938,7 +942,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
assert.NoError(t, err)
mockReleaseActionStatus = getMockReleaseActionStatus()
mockReleaseActionStatus.ActionID = int64(action.ID)
mockReleaseActionStatus.ActionID = int64(mockAction.ID)
mockReleaseActionStatus.FilterID = int64(mockData.ID)
mockReleaseActionStatus.ReleaseID = mockRelease.ID
mockReleaseActionStatus.Timestamp = mockReleaseActionStatus.Timestamp.AddDate(0, -1, 0)
@ -947,7 +951,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
assert.NoError(t, err)
mockReleaseActionStatus2 := getMockReleaseActionStatus()
mockReleaseActionStatus2.ActionID = int64(action2.ID)
mockReleaseActionStatus2.ActionID = int64(mockAction2.ID)
mockReleaseActionStatus2.FilterID = int64(mockData.ID)
mockReleaseActionStatus2.ReleaseID = mockRelease.ID
mockReleaseActionStatus2.Timestamp = mockReleaseActionStatus2.Timestamp.AddDate(0, -1, 0)
@ -968,7 +972,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) {
})
// Cleanup
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction.ID})
_ = repo.Delete(context.Background(), mockData.ID)
_ = downloadClientRepo.Delete(context.Background(), mockClient.ID)
_ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})

View file

@ -88,6 +88,39 @@ CREATE TABLE irc_channel
UNIQUE (network_id, name)
);
CREATE TABLE release_profile_duplicate
(
id SERIAL PRIMARY KEY,
name TEXT NOT NULL,
protocol BOOLEAN DEFAULT FALSE,
release_name BOOLEAN DEFAULT FALSE,
hash BOOLEAN DEFAULT FALSE,
title BOOLEAN DEFAULT FALSE,
sub_title BOOLEAN DEFAULT FALSE,
year BOOLEAN DEFAULT FALSE,
month BOOLEAN DEFAULT FALSE,
day BOOLEAN DEFAULT FALSE,
source BOOLEAN DEFAULT FALSE,
resolution BOOLEAN DEFAULT FALSE,
codec BOOLEAN DEFAULT FALSE,
container BOOLEAN DEFAULT FALSE,
dynamic_range BOOLEAN DEFAULT FALSE,
audio BOOLEAN DEFAULT FALSE,
release_group BOOLEAN DEFAULT FALSE,
season BOOLEAN DEFAULT FALSE,
episode BOOLEAN DEFAULT FALSE,
website BOOLEAN DEFAULT FALSE,
proper BOOLEAN DEFAULT FALSE,
repack BOOLEAN DEFAULT FALSE,
edition BOOLEAN DEFAULT FALSE,
language BOOLEAN DEFAULT FALSE
);
INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language)
VALUES (1, 'Exact release', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'),
(2, 'Movie', 'f', 'f', 'f', 't', 'f', 't', 'f', 'f', 'f', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'),
(3, 'TV', 'f', 'f', 'f', 't', 'f', 't', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f');
CREATE TABLE filter
(
id SERIAL PRIMARY KEY,
@ -159,7 +192,9 @@ CREATE TABLE filter
min_seeders INTEGER DEFAULT 0,
max_seeders INTEGER DEFAULT 0,
min_leechers INTEGER DEFAULT 0,
max_leechers INTEGER DEFAULT 0
max_leechers INTEGER DEFAULT 0,
release_profile_duplicate_id INTEGER,
FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate(id) ON DELETE SET NULL
);
CREATE INDEX filter_enabled_index
@ -270,9 +305,11 @@ CREATE TABLE "release"
group_id TEXT,
torrent_id TEXT,
torrent_name TEXT,
normalized_hash TEXT,
size BIGINT,
raw TEXT,
title TEXT,
sub_title TEXT,
category TEXT,
season INTEGER,
episode INTEGER,
@ -285,15 +322,18 @@ CREATE TABLE "release"
container TEXT,
hdr TEXT,
audio TEXT,
audio_channels TEXT,
release_group TEXT,
region TEXT,
language TEXT,
edition TEXT,
cut TEXT,
unrated BOOLEAN,
hybrid BOOLEAN,
proper BOOLEAN,
repack BOOLEAN,
website TEXT,
media_processing TEXT,
artists TEXT [] DEFAULT '{}' NOT NULL,
type TEXT,
format TEXT,
@ -308,6 +348,7 @@ CREATE TABLE "release"
freeleech_percent INTEGER,
uploader TEXT,
pre_time TEXT,
other TEXT [] DEFAULT '{}' NOT NULL,
filter_id INTEGER
CONSTRAINT release_filter_id_fk
REFERENCES filter
@ -326,6 +367,81 @@ CREATE INDEX release_timestamp_index
CREATE INDEX release_torrent_name_index
ON "release" (torrent_name);
CREATE INDEX release_normalized_hash_index
ON "release" (normalized_hash);
CREATE INDEX release_title_index
ON "release" (title);
CREATE INDEX release_sub_title_index
ON "release" (sub_title);
CREATE INDEX release_season_index
ON "release" (season);
CREATE INDEX release_episode_index
ON "release" (episode);
CREATE INDEX release_year_index
ON "release" (year);
CREATE INDEX release_month_index
ON "release" (month);
CREATE INDEX release_day_index
ON "release" (day);
CREATE INDEX release_resolution_index
ON "release" (resolution);
CREATE INDEX release_source_index
ON "release" (source);
CREATE INDEX release_codec_index
ON "release" (codec);
CREATE INDEX release_container_index
ON "release" (container);
CREATE INDEX release_hdr_index
ON "release" (hdr);
CREATE INDEX release_audio_index
ON "release" (audio);
CREATE INDEX release_audio_channels_index
ON "release" (audio_channels);
CREATE INDEX release_release_group_index
ON "release" (release_group);
CREATE INDEX release_language_index
ON "release" (language);
CREATE INDEX release_proper_index
ON "release" (proper);
CREATE INDEX release_repack_index
ON "release" (repack);
CREATE INDEX release_website_index
ON "release" (website);
CREATE INDEX release_media_processing_index
ON "release" (media_processing);
CREATE INDEX release_region_index
ON "release" (region);
CREATE INDEX release_edition_index
ON "release" (edition);
CREATE INDEX release_cut_index
ON "release" (cut);
CREATE INDEX release_hybrid_index
ON "release" (hybrid);
CREATE TABLE release_action_status
(
id SERIAL PRIMARY KEY,
@ -1074,5 +1190,154 @@ CREATE TABLE list_filter
ALTER TABLE filter
ADD COLUMN except_record_labels TEXT DEFAULT '';
`,
`CREATE TABLE release_profile_duplicate
(
id SERIAL PRIMARY KEY,
name TEXT NOT NULL,
protocol BOOLEAN DEFAULT FALSE,
release_name BOOLEAN DEFAULT FALSE,
hash BOOLEAN DEFAULT FALSE,
title BOOLEAN DEFAULT FALSE,
sub_title BOOLEAN DEFAULT FALSE,
year BOOLEAN DEFAULT FALSE,
month BOOLEAN DEFAULT FALSE,
day BOOLEAN DEFAULT FALSE,
source BOOLEAN DEFAULT FALSE,
resolution BOOLEAN DEFAULT FALSE,
codec BOOLEAN DEFAULT FALSE,
container BOOLEAN DEFAULT FALSE,
dynamic_range BOOLEAN DEFAULT FALSE,
audio BOOLEAN DEFAULT FALSE,
release_group BOOLEAN DEFAULT FALSE,
season BOOLEAN DEFAULT FALSE,
episode BOOLEAN DEFAULT FALSE,
website BOOLEAN DEFAULT FALSE,
proper BOOLEAN DEFAULT FALSE,
repack BOOLEAN DEFAULT FALSE,
edition BOOLEAN DEFAULT FALSE,
language BOOLEAN DEFAULT FALSE
);
INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language)
VALUES (1, 'Exact release', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'),
(2, 'Movie', 'f', 'f', 'f', 't', 'f', 't', 'f', 'f', 'f', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'),
(3, 'TV', 'f', 'f', 'f', 't', 'f', 't', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f');
ALTER TABLE filter
ADD release_profile_duplicate_id INTEGER;
ALTER TABLE filter
ADD CONSTRAINT filter_release_profile_duplicate_id_fk
FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate (id)
ON DELETE SET NULL;
ALTER TABLE "release"
ADD normalized_hash TEXT;
ALTER TABLE "release"
ADD sub_title TEXT;
ALTER TABLE "release"
ADD COLUMN IF NOT EXISTS audio TEXT;
ALTER TABLE "release"
ADD audio_channels TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS language TEXT;
ALTER TABLE "release"
ADD media_processing TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS edition TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS cut TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS hybrid TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS region TEXT;
ALTER TABLE "release"
ADD IF NOT EXISTS other TEXT [] DEFAULT '{}' NOT NULL;
CREATE INDEX release_normalized_hash_index
ON "release" (normalized_hash);
CREATE INDEX release_title_index
ON "release" (title);
CREATE INDEX release_sub_title_index
ON "release" (sub_title);
CREATE INDEX release_season_index
ON "release" (season);
CREATE INDEX release_episode_index
ON "release" (episode);
CREATE INDEX release_year_index
ON "release" (year);
CREATE INDEX release_month_index
ON "release" (month);
CREATE INDEX release_day_index
ON "release" (day);
CREATE INDEX release_resolution_index
ON "release" (resolution);
CREATE INDEX release_source_index
ON "release" (source);
CREATE INDEX release_codec_index
ON "release" (codec);
CREATE INDEX release_container_index
ON "release" (container);
CREATE INDEX release_hdr_index
ON "release" (hdr);
CREATE INDEX release_audio_index
ON "release" (audio);
CREATE INDEX release_audio_channels_index
ON "release" (audio_channels);
CREATE INDEX release_release_group_index
ON "release" (release_group);
CREATE INDEX release_proper_index
ON "release" (proper);
CREATE INDEX release_repack_index
ON "release" (repack);
CREATE INDEX release_website_index
ON "release" (website);
CREATE INDEX release_media_processing_index
ON "release" (media_processing);
CREATE INDEX release_language_index
ON "release" (language);
CREATE INDEX release_region_index
ON "release" (region);
CREATE INDEX release_edition_index
ON "release" (edition);
CREATE INDEX release_cut_index
ON "release" (cut);
CREATE INDEX release_hybrid_index
ON "release" (hybrid);
`,
}

View file

@ -33,23 +33,31 @@ func NewReleaseRepo(log logger.Logger, db *DB) domain.ReleaseRepo {
}
func (repo *ReleaseRepo) Store(ctx context.Context, r *domain.Release) error {
codecStr := strings.Join(r.Codec, ",")
hdrStr := strings.Join(r.HDR, ",")
var (
codecStr = strings.Join(r.Codec, ",")
hdrStr = strings.Join(r.HDR, ",")
audioStr = strings.Join(r.Audio, ",")
editionStr = strings.Join(r.Edition, ",")
cutStr = strings.Join(r.Cut, ",")
languageStr = strings.Join(r.Language, ",")
)
queryBuilder := repo.db.squirrel.
Insert("release").
Columns("filter_status", "rejections", "indexer", "filter", "protocol", "implementation", "timestamp", "announce_type", "group_id", "torrent_id", "info_url", "download_url", "torrent_name", "size", "title", "category", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "hdr", "release_group", "proper", "repack", "website", "type", "origin", "tags", "uploader", "pre_time", "filter_id").
Values(r.FilterStatus, pq.Array(r.Rejections), r.Indexer.Identifier, r.FilterName, r.Protocol, r.Implementation, r.Timestamp.Format(time.RFC3339), r.AnnounceType, r.GroupID, r.TorrentID, r.InfoURL, r.DownloadURL, r.TorrentName, r.Size, r.Title, r.Category, r.Season, r.Episode, r.Year, r.Month, r.Day, r.Resolution, r.Source, codecStr, r.Container, hdrStr, r.Group, r.Proper, r.Repack, r.Website, r.Type, r.Origin, pq.Array(r.Tags), r.Uploader, r.PreTime, r.FilterID).
Columns("filter_status", "rejections", "indexer", "filter", "protocol", "implementation", "timestamp", "announce_type", "group_id", "torrent_id", "info_url", "download_url", "torrent_name", "normalized_hash", "size", "title", "sub_title", "category", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "hdr", "audio", "audio_channels", "release_group", "proper", "repack", "region", "language", "cut", "edition", "hybrid", "media_processing", "website", "type", "origin", "tags", "uploader", "pre_time", "other", "filter_id").
Values(r.FilterStatus, pq.Array(r.Rejections), r.Indexer.Identifier, r.FilterName, r.Protocol, r.Implementation, r.Timestamp.Format(time.RFC3339), r.AnnounceType, r.GroupID, r.TorrentID, r.InfoURL, r.DownloadURL, r.TorrentName, r.NormalizedHash, r.Size, r.Title, r.SubTitle, r.Category, r.Season, r.Episode, r.Year, r.Month, r.Day, r.Resolution, r.Source, codecStr, r.Container, hdrStr, audioStr, r.AudioChannels, r.Group, r.Proper, r.Repack, r.Region, languageStr, cutStr, editionStr, r.Hybrid, r.MediaProcessing, r.Website, r.Type.String(), r.Origin, pq.Array(r.Tags), r.Uploader, r.PreTime, pq.Array(r.Other), r.FilterID).
Suffix("RETURNING id").RunWith(repo.db.handler)
// return values
var retID int64
if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil {
return errors.Wrap(err, "error executing query")
q, args, err := queryBuilder.ToSql()
if err != nil {
return errors.Wrap(err, "error building query")
}
r.ID = retID
repo.log.Debug().Msgf("release.store: %s %v", q, args)
if err := queryBuilder.QueryRowContext(ctx).Scan(&r.ID); err != nil {
return errors.Wrap(err, "error executing query")
}
repo.log.Debug().Msgf("release.store: %+v", r)
@ -102,14 +110,9 @@ func (repo *ReleaseRepo) StoreReleaseActionStatus(ctx context.Context, status *d
Values(status.Status, status.Action, status.ActionID, status.Type, status.Client, status.Filter, status.FilterID, pq.Array(status.Rejections), status.Timestamp.Format(time.RFC3339), status.ReleaseID).
Suffix("RETURNING id").RunWith(repo.db.handler)
// return values
var retID int64
if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil {
if err := queryBuilder.QueryRowContext(ctx).Scan(&status.ID); err != nil {
return errors.Wrap(err, "error executing query")
}
status.ID = retID
}
repo.log.Trace().Msgf("release.store_release_action_status: %+v", status)
@ -117,6 +120,62 @@ func (repo *ReleaseRepo) StoreReleaseActionStatus(ctx context.Context, status *d
return nil
}
func (repo *ReleaseRepo) StoreDuplicateProfile(ctx context.Context, profile *domain.DuplicateReleaseProfile) error {
if profile.ID == 0 {
queryBuilder := repo.db.squirrel.
Insert("release_profile_duplicate").
Columns("name", "protocol", "release_name", "hash", "title", "sub_title", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "dynamic_range", "audio", "release_group", "website", "proper", "repack").
Values(profile.Name, profile.Protocol, profile.ReleaseName, profile.Hash, profile.Title, profile.SubTitle, profile.Season, profile.Episode, profile.Year, profile.Month, profile.Day, profile.Resolution, profile.Source, profile.Codec, profile.Container, profile.DynamicRange, profile.Audio, profile.Group, profile.Website, profile.Proper, profile.Repack).
Suffix("RETURNING id").
RunWith(repo.db.handler)
// return values
var retID int64
err := queryBuilder.QueryRowContext(ctx).Scan(&retID)
if err != nil {
return errors.Wrap(err, "error executing query")
}
profile.ID = retID
} else {
queryBuilder := repo.db.squirrel.
Update("release_profile_duplicate").
Set("name", profile.Name).
Set("protocol", profile.Protocol).
Set("release_name", profile.ReleaseName).
Set("hash", profile.Hash).
Set("title", profile.Title).
Set("sub_title", profile.SubTitle).
Set("season", profile.Season).
Set("episode", profile.Episode).
Set("year", profile.Year).
Set("month", profile.Month).
Set("day", profile.Day).
Set("resolution", profile.Resolution).
Set("source", profile.Source).
Set("codec", profile.Codec).
Set("container", profile.Container).
Set("dynamic_range", profile.DynamicRange).
Set("audio", profile.Audio).
Set("release_group", profile.Group).
Set("website", profile.Website).
Set("proper", profile.Proper).
Set("repack", profile.Repack).
Where(sq.Eq{"id": profile.ID}).
RunWith(repo.db.handler)
_, err := queryBuilder.ExecContext(ctx)
if err != nil {
return errors.Wrap(err, "error executing query")
}
}
repo.log.Debug().Msgf("release.StoreDuplicateProfile: %+v", profile)
return nil
}
func (repo *ReleaseRepo) Find(ctx context.Context, params domain.ReleaseQueryParams) (*domain.FindReleasesResponse, error) {
tx, err := repo.db.BeginTx(ctx, &sql.TxOptions{Isolation: sql.LevelReadCommitted})
if err != nil {
@ -192,7 +251,7 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
whereQuery, _, err := whereQueryBuilder.ToSql()
if err != nil {
return nil, errors.Wrap(err, "error building wherequery")
return nil, errors.Wrap(err, "error building where query")
}
subQueryBuilder := repo.db.squirrel.
@ -230,8 +289,49 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
}
queryBuilder := repo.db.squirrel.
Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "i.id", "i.name", "i.identifier_external", "r.filter", "r.protocol", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.torrent_name", "r.size", "r.category", "r.season", "r.episode", "r.year", "r.resolution", "r.source", "r.codec", "r.container", "r.release_group", "r.timestamp",
"ras.id", "ras.status", "ras.action", "ras.action_id", "ras.type", "ras.client", "ras.filter", "ras.filter_id", "ras.release_id", "ras.rejections", "ras.timestamp").
Select(
"r.id",
"r.filter_status",
"r.rejections",
"r.indexer",
"i.id",
"i.name",
"i.identifier_external",
"r.filter",
"r.protocol",
"r.announce_type",
"r.info_url",
"r.download_url",
"r.title",
"r.sub_title",
"r.torrent_name",
"r.normalized_hash",
"r.size",
"r.category",
"r.season",
"r.episode",
"r.year",
"r.resolution",
"r.source",
"r.codec",
"r.container",
"r.hdr",
"r.audio",
"r.audio_channels",
"r.release_group",
"r.region",
"r.language",
"r.edition",
"r.cut",
"r.hybrid",
"r.proper",
"r.repack",
"r.website",
"r.media_processing",
"r.type",
"r.timestamp",
"ras.id", "ras.status", "ras.action", "ras.action_id", "ras.type", "ras.client", "ras.filter", "ras.filter_id", "ras.release_id", "ras.rejections", "ras.timestamp",
).
Column(sq.Alias(countQuery, "page_total")).
From("release r").
OrderBy("r.id DESC").
@ -267,7 +367,7 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
var rls domain.Release
var ras domain.ReleaseActionStatus
var rlsIndexer, rlsIndexerName, rlsIndexerExternalName, rlsFilter, rlsAnnounceType, infoUrl, downloadUrl, codec sql.NullString
var rlsIndexer, rlsIndexerName, rlsIndexerExternalName, rlsFilter, rlsAnnounceType, infoUrl, downloadUrl, subTitle, normalizedHash, codec, hdr, rlsType, audioStr, languageStr, editionStr, cutStr, website sql.NullString
var rlsIndexerID sql.NullInt64
var rasId, rasFilterId, rasReleaseId, rasActionId sql.NullInt64
@ -275,7 +375,49 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
var rasRejections []sql.NullString
var rasTimestamp sql.NullTime
if err := rows.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &rlsIndexer, &rlsIndexerID, &rlsIndexerName, &rlsIndexerExternalName, &rlsFilter, &rls.Protocol, &rlsAnnounceType, &infoUrl, &downloadUrl, &rls.Title, &rls.TorrentName, &rls.Size, &rls.Category, &rls.Season, &rls.Episode, &rls.Year, &rls.Resolution, &rls.Source, &codec, &rls.Container, &rls.Group, &rls.Timestamp, &rasId, &rasStatus, &rasAction, &rasActionId, &rasType, &rasClient, &rasFilter, &rasFilterId, &rasReleaseId, pq.Array(&rasRejections), &rasTimestamp, &resp.TotalCount); err != nil {
if err := rows.Scan(
&rls.ID,
&rls.FilterStatus,
pq.Array(&rls.Rejections),
&rlsIndexer,
&rlsIndexerID,
&rlsIndexerName,
&rlsIndexerExternalName,
&rlsFilter,
&rls.Protocol,
&rlsAnnounceType,
&infoUrl,
&downloadUrl,
&rls.Title,
&subTitle,
&rls.TorrentName,
&normalizedHash,
&rls.Size,
&rls.Category,
&rls.Season,
&rls.Episode,
&rls.Year,
&rls.Resolution,
&rls.Source,
&codec,
&rls.Container,
&hdr,
&audioStr,
&rls.AudioChannels,
&rls.Group,
&rls.Region,
&languageStr,
&editionStr,
&cutStr,
&rls.Hybrid,
&rls.Proper,
&rls.Repack,
&website,
&rls.MediaProcessing,
&rlsType,
&rls.Timestamp,
&rasId, &rasStatus, &rasAction, &rasActionId, &rasType, &rasClient, &rasFilter, &rasFilterId, &rasReleaseId, pq.Array(&rasRejections), &rasTimestamp, &resp.TotalCount,
); err != nil {
return resp, errors.Wrap(err, "error scanning row")
}
@ -324,7 +466,19 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
rls.ActionStatus = make([]domain.ReleaseActionStatus, 0)
rls.InfoURL = infoUrl.String
rls.DownloadURL = downloadUrl.String
rls.SubTitle = subTitle.String
rls.NormalizedHash = normalizedHash.String
rls.Codec = strings.Split(codec.String, ",")
rls.HDR = strings.Split(hdr.String, ",")
rls.Audio = strings.Split(audioStr.String, ",")
rls.Language = strings.Split(languageStr.String, ",")
rls.Edition = strings.Split(editionStr.String, ",")
rls.Cut = strings.Split(cutStr.String, ",")
rls.Website = website.String
//rls.Type = rlsType.String
if rlsType.Valid {
rls.ParseType(rlsType.String)
}
// only add ActionStatus if it's not empty
if ras.ID > 0 {
@ -342,6 +496,66 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain
return resp, nil
}
func (repo *ReleaseRepo) FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) {
queryBuilder := repo.db.squirrel.
Select(
"id",
"name",
"protocol",
"release_name",
"hash",
"title",
"sub_title",
"year",
"month",
"day",
"source",
"resolution",
"codec",
"container",
"dynamic_range",
"audio",
"release_group",
"season",
"episode",
"website",
"proper",
"repack",
).
From("release_profile_duplicate")
query, args, err := queryBuilder.ToSql()
if err != nil {
return nil, errors.Wrap(err, "error building query")
}
rows, err := repo.db.handler.QueryContext(ctx, query, args...)
if err != nil {
return nil, errors.Wrap(err, "error executing query")
}
defer rows.Close()
if err := rows.Err(); err != nil {
return nil, errors.Wrap(err, "error rows FindDuplicateReleaseProfiles")
}
res := make([]*domain.DuplicateReleaseProfile, 0)
for rows.Next() {
var p domain.DuplicateReleaseProfile
err := rows.Scan(&p.ID, &p.Name, &p.Protocol, &p.ReleaseName, &p.Hash, &p.Title, &p.SubTitle, &p.Year, &p.Month, &p.Day, &p.Source, &p.Resolution, &p.Codec, &p.Container, &p.DynamicRange, &p.Audio, &p.Group, &p.Season, &p.Episode, &p.Website, &p.Proper, &p.Repack)
if err != nil {
return nil, errors.Wrap(err, "error scanning row")
}
res = append(res, &p)
}
return res, nil
}
func (repo *ReleaseRepo) GetIndexerOptions(ctx context.Context) ([]string, error) {
query := `SELECT DISTINCT indexer FROM "release" UNION SELECT DISTINCT identifier indexer FROM indexer;`
@ -420,7 +634,7 @@ func (repo *ReleaseRepo) GetActionStatusByReleaseID(ctx context.Context, release
func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest) (*domain.Release, error) {
queryBuilder := repo.db.squirrel.
Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "r.filter", "r.filter_id", "r.protocol", "r.implementation", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.torrent_name", "r.category", "r.size", "r.group_id", "r.torrent_id", "r.uploader", "r.timestamp").
Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "r.filter", "r.filter_id", "r.protocol", "r.implementation", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.sub_title", "r.torrent_name", "r.category", "r.size", "r.group_id", "r.torrent_id", "r.uploader", "r.timestamp").
From("release r").
OrderBy("r.id DESC").
Where(sq.Eq{"r.id": req.Id})
@ -439,10 +653,10 @@ func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest)
var rls domain.Release
var indexerName, filterName, announceType, infoUrl, downloadUrl, groupId, torrentId, category, uploader sql.NullString
var indexerName, filterName, announceType, infoUrl, downloadUrl, subTitle, groupId, torrentId, category, uploader sql.NullString
var filterId sql.NullInt64
if err := row.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &indexerName, &filterName, &filterId, &rls.Protocol, &rls.Implementation, &announceType, &infoUrl, &downloadUrl, &rls.Title, &rls.TorrentName, &category, &rls.Size, &groupId, &torrentId, &uploader, &rls.Timestamp); err != nil {
if err := row.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &indexerName, &filterName, &filterId, &rls.Protocol, &rls.Implementation, &announceType, &infoUrl, &downloadUrl, &rls.Title, &subTitle, &rls.TorrentName, &category, &rls.Size, &groupId, &torrentId, &uploader, &rls.Timestamp); err != nil {
if errors.Is(err, sql.ErrNoRows) {
return nil, domain.ErrRecordNotFound
}
@ -457,6 +671,7 @@ func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest)
rls.AnnounceType = domain.AnnounceType(announceType.String)
rls.InfoURL = infoUrl.String
rls.DownloadURL = downloadUrl.String
rls.SubTitle = subTitle.String
rls.Category = category.String
rls.GroupID = groupId.String
rls.TorrentID = torrentId.String
@ -670,6 +885,31 @@ func (repo *ReleaseRepo) Delete(ctx context.Context, req *domain.DeleteReleaseRe
return nil
}
func (repo *ReleaseRepo) DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error {
qb := repo.db.squirrel.Delete("release_profile_duplicate").Where(sq.Eq{"id": id})
query, args, err := qb.ToSql()
if err != nil {
return errors.Wrap(err, "error building SQL query")
}
_, err = repo.db.handler.ExecContext(ctx, query, args...)
if err != nil {
return errors.Wrap(err, "error executing delete query")
}
//deletedRows, err := result.RowsAffected()
//if err != nil {
// return errors.Wrap(err, "error fetching rows affected")
//}
//
//repo.log.Debug().Msgf("deleted %d rows from release table", deletedRows)
repo.log.Debug().Msgf("deleted duplicate release profile: %d", id)
return nil
}
func (repo *ReleaseRepo) CheckSmartEpisodeCanDownload(ctx context.Context, p *domain.SmartEpisodeParams) (bool, error) {
queryBuilder := repo.db.squirrel.
Select("COUNT(*)").
@ -793,3 +1033,200 @@ func (repo *ReleaseRepo) UpdateBaseURL(ctx context.Context, indexer string, oldB
return nil
}
func (repo *ReleaseRepo) CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error) {
queryBuilder := repo.db.squirrel.
Select("r.id, r.torrent_name, r.normalized_hash, r.title, ras.action, ras.status").
From("release r").
LeftJoin("release_action_status ras ON r.id = ras.release_id").
Where("ras.status = 'PUSH_APPROVED'")
if profile.ReleaseName && profile.Hash {
//queryBuilder = queryBuilder.Where(repo.db.ILike("r.torrent_name", release.TorrentName))
queryBuilder = queryBuilder.Where(sq.Eq{"r.normalized_hash": release.NormalizedHash})
} else {
if profile.Title {
queryBuilder = queryBuilder.Where(repo.db.ILike("r.title", release.Title))
}
if profile.SubTitle {
queryBuilder = queryBuilder.Where(repo.db.ILike("r.sub_title", release.SubTitle))
}
if profile.ReleaseName && profile.Hash {
//queryBuilder = queryBuilder.Where(repo.db.ILike("r.torrent_name", release.TorrentName))
queryBuilder = queryBuilder.Where(sq.Eq{"r.normalized_hash": release.NormalizedHash})
}
if profile.Year {
queryBuilder = queryBuilder.Where(sq.Eq{"r.year": release.Year})
}
if profile.Month {
queryBuilder = queryBuilder.Where(sq.Eq{"r.month": release.Month})
}
if profile.Day {
queryBuilder = queryBuilder.Where(sq.Eq{"r.day": release.Day})
}
if profile.Source {
queryBuilder = queryBuilder.Where(sq.Eq{"r.source": release.Source})
}
if profile.Container {
queryBuilder = queryBuilder.Where(sq.Eq{"r.container": release.Container})
}
if profile.Edition {
//queryBuilder = queryBuilder.Where(sq.Eq{"r.cut": release.Cut})
if len(release.Cut) > 1 {
var and sq.And
for _, cut := range release.Cut {
//and = append(and, sq.Eq{"r.cut": "%" + cut + "%"})
and = append(and, repo.db.ILike("r.cut", "%"+cut+"%"))
}
queryBuilder = queryBuilder.Where(and)
} else if len(release.Cut) == 1 {
queryBuilder = queryBuilder.Where(repo.db.ILike("r.cut", "%"+release.Cut[0]+"%"))
}
//queryBuilder = queryBuilder.Where(sq.Eq{"r.edition": release.Edition})
if len(release.Edition) > 1 {
var and sq.And
for _, edition := range release.Edition {
and = append(and, repo.db.ILike("r.edition", "%"+edition+"%"))
}
queryBuilder = queryBuilder.Where(and)
} else if len(release.Edition) == 1 {
queryBuilder = queryBuilder.Where(repo.db.ILike("r.edition", "%"+release.Edition[0]+"%"))
}
}
// video features (hybrid, remux)
if release.IsTypeVideo() {
queryBuilder = queryBuilder.Where(sq.Eq{"r.hybrid": release.Hybrid})
queryBuilder = queryBuilder.Where(sq.Eq{"r.media_processing": release.MediaProcessing})
}
if profile.Language {
queryBuilder = queryBuilder.Where(sq.Eq{"r.region": release.Region})
if len(release.Language) > 0 {
var and sq.And
for _, lang := range release.Language {
and = append(and, repo.db.ILike("r.language", "%"+lang+"%"))
}
queryBuilder = queryBuilder.Where(and)
} else {
queryBuilder = queryBuilder.Where(sq.Eq{"r.language": ""})
}
}
if profile.Codec {
if len(release.Codec) > 1 {
var and sq.And
for _, codec := range release.Codec {
and = append(and, repo.db.ILike("r.codec", "%"+codec+"%"))
}
queryBuilder = queryBuilder.Where(and)
} else {
// FIXME this does an IN (arg)
queryBuilder = queryBuilder.Where(sq.Eq{"r.codec": release.Codec})
}
}
if profile.Resolution {
queryBuilder = queryBuilder.Where(sq.Eq{"r.resolution": release.Resolution})
}
if profile.DynamicRange {
//if len(release.HDR) > 1 {
// var and sq.And
// for _, hdr := range release.HDR {
// and = append(and, repo.db.ILike("r.hdr", "%"+hdr+"%"))
// }
// queryBuilder = queryBuilder.Where(and)
//} else {
// queryBuilder = queryBuilder.Where(sq.Eq{"r.hdr": release.HDR})
//}
queryBuilder = queryBuilder.Where(sq.Eq{"r.hdr": strings.Join(release.HDR, ",")})
}
if profile.Audio {
queryBuilder = queryBuilder.Where(sq.Eq{"r.audio": strings.Join(release.Audio, ",")})
queryBuilder = queryBuilder.Where(sq.Eq{"r.audio_channels": release.AudioChannels})
}
if profile.Group {
queryBuilder = queryBuilder.Where(repo.db.ILike("r.release_group", release.Group))
}
if profile.Season {
queryBuilder = queryBuilder.Where(sq.Eq{"r.season": release.Season})
}
if profile.Episode {
queryBuilder = queryBuilder.Where(sq.Eq{"r.episode": release.Episode})
}
if profile.Website {
queryBuilder = queryBuilder.Where(sq.Eq{"r.website": release.Website})
}
if profile.Proper {
queryBuilder = queryBuilder.Where(sq.Eq{"r.proper": release.Proper})
}
if profile.Repack {
queryBuilder = queryBuilder.Where(sq.And{
sq.Eq{"r.repack": release.Repack},
repo.db.ILike("r.release_group", release.Group),
})
}
}
query, args, err := queryBuilder.ToSql()
if err != nil {
return false, errors.Wrap(err, "error building query")
}
repo.log.Trace().Str("database", "release.FindDuplicateReleases").Msgf("query: %q, args: %q", query, args)
rows, err := repo.db.handler.QueryContext(ctx, query, args...)
if err != nil {
return false, err
}
if err := rows.Err(); err != nil {
return false, errors.Wrap(err, "error rows CheckIsDuplicateRelease")
}
type result struct {
id int
release string
hash string
title string
action string
status string
}
var res []result
for rows.Next() {
r := result{}
if err := rows.Scan(&r.id, &r.release, &r.hash, &r.title, &r.action, &r.status); err != nil {
return false, errors.Wrap(err, "error scan CheckIsDuplicateRelease")
}
res = append(res, r)
}
repo.log.Trace().Str("database", "release.FindDuplicateReleases").Msgf("found duplicate releases: %+v", res)
if len(res) == 0 {
return false, nil
}
return true, nil
}

View file

@ -13,6 +13,7 @@ import (
"github.com/autobrr/autobrr/internal/domain"
"github.com/moistari/rls"
"github.com/stretchr/testify/assert"
)
@ -49,12 +50,13 @@ func getMockRelease() *domain.Release {
Proper: true,
Repack: false,
Website: "https://example.com",
Type: "Movie",
Type: rls.Movie,
Origin: "P2P",
Tags: []string{"Action", "Adventure"},
Uploader: "john_doe",
PreTime: "10m",
FilterID: 1,
Other: []string{},
}
}
@ -108,11 +110,11 @@ func TestReleaseRepo_Store(t *testing.T) {
// Execute
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -123,7 +125,7 @@ func TestReleaseRepo_Store(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -164,11 +166,11 @@ func TestReleaseRepo_StoreReleaseActionStatus(t *testing.T) {
// Execute
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -179,7 +181,7 @@ func TestReleaseRepo_StoreReleaseActionStatus(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -328,11 +330,11 @@ func TestReleaseRepo_GetIndexerOptions(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -347,7 +349,7 @@ func TestReleaseRepo_GetIndexerOptions(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -387,11 +389,11 @@ func TestReleaseRepo_GetActionStatusByReleaseID(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -407,7 +409,7 @@ func TestReleaseRepo_GetActionStatusByReleaseID(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -447,11 +449,11 @@ func TestReleaseRepo_Get(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -467,7 +469,7 @@ func TestReleaseRepo_Get(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -507,11 +509,11 @@ func TestReleaseRepo_Stats(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -527,7 +529,7 @@ func TestReleaseRepo_Stats(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -567,11 +569,11 @@ func TestReleaseRepo_Delete(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -584,7 +586,7 @@ func TestReleaseRepo_Delete(t *testing.T) {
assert.NoError(t, err)
// Cleanup
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
@ -624,11 +626,11 @@ func TestReleaseRepo_CheckSmartEpisodeCanDownloadShow(t *testing.T) {
err = repo.Store(context.Background(), mockData)
assert.NoError(t, err)
createdAction, err := actionRepo.Store(context.Background(), actionMockData)
err = actionRepo.Store(context.Background(), actionMockData)
assert.NoError(t, err)
releaseActionMockData.ReleaseID = mockData.ID
releaseActionMockData.ActionID = int64(createdAction.ID)
releaseActionMockData.ActionID = int64(actionMockData.ID)
releaseActionMockData.FilterID = int64(createdFilters[0].ID)
err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData)
@ -652,9 +654,724 @@ func TestReleaseRepo_CheckSmartEpisodeCanDownloadShow(t *testing.T) {
// Cleanup
_ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
_ = downloadClientRepo.Delete(context.Background(), mock.ID)
})
}
}
func getMockDuplicateReleaseProfileTV() *domain.DuplicateReleaseProfile {
return &domain.DuplicateReleaseProfile{
ID: 0,
Name: "TV",
Protocol: false,
ReleaseName: false,
Hash: false,
Title: true,
SubTitle: false,
Year: false,
Month: false,
Day: false,
Source: false,
Resolution: false,
Codec: false,
Container: false,
DynamicRange: false,
Audio: false,
Group: false,
Season: true,
Episode: true,
Website: false,
Proper: false,
Repack: false,
Edition: false,
Language: false,
}
}
func getMockDuplicateReleaseProfileTVDaily() *domain.DuplicateReleaseProfile {
return &domain.DuplicateReleaseProfile{
ID: 0,
Name: "TV",
Protocol: false,
ReleaseName: false,
Hash: false,
Title: true,
SubTitle: false,
Year: true,
Month: true,
Day: true,
Source: false,
Resolution: false,
Codec: false,
Container: false,
DynamicRange: false,
Audio: false,
Group: false,
Season: false,
Episode: false,
Website: false,
Proper: false,
Repack: false,
Edition: false,
Language: false,
}
}
func getMockFilterDuplicates() *domain.Filter {
return &domain.Filter{
Name: "New Filter",
Enabled: true,
CreatedAt: time.Now(),
UpdatedAt: time.Now(),
MinSize: "10mb",
MaxSize: "20mb",
Delay: 60,
Priority: 1,
MaxDownloads: 100,
MaxDownloadsUnit: domain.FilterMaxDownloadsHour,
MatchReleases: "BRRip",
ExceptReleases: "BRRip",
UseRegex: false,
MatchReleaseGroups: "AMIABLE",
ExceptReleaseGroups: "NTb",
Scene: false,
Origins: nil,
ExceptOrigins: nil,
Bonus: nil,
Freeleech: false,
FreeleechPercent: "100%",
SmartEpisode: false,
Shows: "Is It Wrong to Try to Pick Up Girls in a Dungeon?",
Seasons: "4",
Episodes: "500",
Resolutions: []string{"1080p"},
Codecs: []string{"x264"},
Sources: []string{"BluRay"},
Containers: []string{"mkv"},
MatchHDR: []string{"HDR10"},
ExceptHDR: []string{"HDR10"},
MatchOther: []string{"Atmos"},
ExceptOther: []string{"Atmos"},
Years: "2023",
Months: "",
Days: "",
Artists: "",
Albums: "",
MatchReleaseTypes: []string{"Remux"},
ExceptReleaseTypes: "Remux",
Formats: []string{"FLAC"},
Quality: []string{"Lossless"},
Media: []string{"CD"},
PerfectFlac: true,
Cue: true,
Log: true,
LogScore: 100,
MatchCategories: "Anime",
ExceptCategories: "Anime",
MatchUploaders: "SubsPlease",
ExceptUploaders: "SubsPlease",
MatchLanguage: []string{"English", "Japanese"},
ExceptLanguage: []string{"English", "Japanese"},
Tags: "Anime, x264",
ExceptTags: "Anime, x264",
TagsAny: "Anime, x264",
ExceptTagsAny: "Anime, x264",
TagsMatchLogic: "AND",
ExceptTagsMatchLogic: "AND",
MatchReleaseTags: "Anime, x264",
ExceptReleaseTags: "Anime, x264",
UseRegexReleaseTags: true,
MatchDescription: "Anime, x264",
ExceptDescription: "Anime, x264",
UseRegexDescription: true,
}
}
func TestReleaseRepo_CheckIsDuplicateRelease(t *testing.T) {
for dbType, db := range testDBs {
log := setupLoggerForTest()
downloadClientRepo := NewDownloadClientRepo(log, db)
filterRepo := NewFilterRepo(log, db)
actionRepo := NewActionRepo(log, db, downloadClientRepo)
releaseRepo := NewReleaseRepo(log, db)
// reset
//db.handler.Exec("DELETE FROM release")
//db.handler.Exec("DELETE FROM action")
//db.handler.Exec("DELETE FROM release_action_status")
mockIndexer := domain.IndexerMinimal{ID: 0, Name: "Mock", Identifier: "mock", IdentifierExternal: "Mock"}
actionMock := &domain.Action{Name: "Test", Type: domain.ActionTypeTest, Enabled: true}
filterMock := getMockFilterDuplicates()
// Setup
err := filterRepo.Store(context.Background(), filterMock)
assert.NoError(t, err)
createdFilters, err := filterRepo.ListFilters(context.Background())
assert.NoError(t, err)
assert.NotNil(t, createdFilters)
actionMock.FilterID = filterMock.ID
err = actionRepo.Store(context.Background(), actionMock)
assert.NoError(t, err)
type fields struct {
releaseTitles []string
releaseTitle string
profile *domain.DuplicateReleaseProfile
}
tests := []struct {
name string
fields fields
isDuplicate bool
}{
{
name: "1",
fields: fields{
releaseTitles: []string{
"Inkheart 2008 BluRay 1080p DD5.1 x264-BADGROUP",
},
releaseTitle: "Inkheart 2008 BluRay 1080p DD5.1 x264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Group: true},
},
isDuplicate: false,
},
{
name: "2",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Source: true, Resolution: true},
},
isDuplicate: true,
},
{
name: "3",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true},
},
isDuplicate: true,
},
{
name: "4",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true},
},
isDuplicate: false,
},
{
name: "5",
fields: fields{
releaseTitles: []string{
"That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true},
},
isDuplicate: true,
},
{
name: "6",
fields: fields{
releaseTitles: []string{
"That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Tv.Show.2023.S01E02.BluRay.2160p.x265.DTS-HD-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true},
},
isDuplicate: false,
},
{
name: "7",
fields: fields{
releaseTitles: []string{
"That.Tv.Show.2023.S01.BluRay.2160p.x265.DTS-HD-GROUP",
},
releaseTitle: "That.Tv.Show.2023.S01.BluRay.2160p.x265.DTS-HD-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true},
},
isDuplicate: true,
},
{
name: "8",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, Group: true},
},
isDuplicate: true,
},
{
name: "9",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.HULU.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, Group: true},
},
isDuplicate: false,
},
{
name: "10",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true},
},
isDuplicate: true,
},
{
name: "11",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
"The Best Show 2020 S04E10 1080p amzn web-dl ddp 5.1 hdr dv h.264-group",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true},
},
isDuplicate: true,
},
{
name: "12",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true},
},
isDuplicate: false,
},
{
name: "13",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true},
},
isDuplicate: false,
},
{
name: "14",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true},
},
isDuplicate: true,
},
{
name: "15",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true},
},
isDuplicate: true,
},
{
name: "16",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E11 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true},
},
isDuplicate: false,
},
{
name: "17",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 Episode Title REPACK 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true},
},
isDuplicate: true,
},
{
name: "18",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The.Best.Show.2020.S04E10.Episode.Title.REPACK.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 Episode Title REPACK 1080p AMZN WEB-DL DDP 5.1 DV H.264-OTHERGROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Repack: true},
},
isDuplicate: false, // not a match because REPACK checks for the same group
},
{
name: "19",
fields: fields{
releaseTitles: []string{
"The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1",
},
releaseTitle: "The Daily Show 2024-09-21.Other.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true},
},
isDuplicate: true,
},
{
name: "20",
fields: fields{
releaseTitles: []string{
"The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1",
},
releaseTitle: "The Daily Show 2024-09-21 Other Guest 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true, SubTitle: true},
},
isDuplicate: false,
},
{
name: "21",
fields: fields{
releaseTitles: []string{
"The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP",
"The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP",
"The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1",
},
releaseTitle: "The Daily Show 2024-09-22 Other Guest 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP1",
profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true, SubTitle: true},
},
isDuplicate: false,
},
{
name: "22",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP",
},
releaseTitle: "That.Movie.2023.2160p.BluRay.DD.2.0.x265-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "23",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP",
},
releaseTitle: "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: true,
},
{
name: "24",
fields: fields{
releaseTitles: []string{
"That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP",
"That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP",
"That.Movie.2023.2160p.BluRay.DD.5.1.x265-GROUP",
},
releaseTitle: "That.Movie.2023.2160p.BluRay.AC3.5.1.x265-GROUP",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: true,
},
{
name: "25",
fields: fields{
releaseTitles: []string{
//"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "26",
fields: fields{
releaseTitles: []string{
//"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "27",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Edition: false, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: true,
},
{
name: "28",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR",
"Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Edition: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true},
},
isDuplicate: true,
},
{
name: "29",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR DV HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "30",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR DV HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "31",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "32",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true},
},
isDuplicate: false,
},
{
name: "33",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true},
},
isDuplicate: false,
},
{
name: "34",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true},
},
isDuplicate: true,
},
{
name: "35",
fields: fields{
releaseTitles: []string{
"Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
"Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
},
releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true},
},
isDuplicate: false,
},
{
name: "36",
fields: fields{
releaseTitles: []string{
"Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT",
},
releaseTitle: "Road House 1989 1080p Blu-ray AVC LPCM 2.0-MONUMENT",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true},
},
isDuplicate: false,
},
{
name: "37",
fields: fields{
releaseTitles: []string{
"Road House 1989 1080p ITA Blu-ray AVC LPCM 2.0-MONUMENT",
"Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT",
},
releaseTitle: "Road House 1989 1080p NOR Blu-ray AVC LPCM 2.0-MONUMENT",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true},
},
isDuplicate: false,
},
{
name: "38",
fields: fields{
releaseTitles: []string{
"Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT",
},
releaseTitle: "Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT",
profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true},
},
isDuplicate: true,
},
{
name: "39",
fields: fields{
releaseTitles: []string{
"The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP",
"The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP",
},
releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP",
profile: &domain.DuplicateReleaseProfile{ReleaseName: true},
},
isDuplicate: true,
},
}
for _, tt := range tests {
t.Run(fmt.Sprintf("Check_Is_Duplicate_Release %s [%s]", tt.name, dbType), func(t *testing.T) {
ctx := context.Background()
// Setup
for _, rel := range tt.fields.releaseTitles {
mockRel := domain.NewRelease(mockIndexer)
mockRel.ParseString(rel)
mockRel.FilterID = filterMock.ID
err = releaseRepo.Store(ctx, mockRel)
assert.NoError(t, err)
ras := &domain.ReleaseActionStatus{
ID: 0,
Status: domain.ReleasePushStatusApproved,
Action: "test",
ActionID: int64(actionMock.ID),
Type: domain.ActionTypeTest,
Client: "",
Filter: "Test filter",
FilterID: int64(filterMock.ID),
Rejections: []string{},
ReleaseID: mockRel.ID,
Timestamp: time.Now(),
}
err = releaseRepo.StoreReleaseActionStatus(ctx, ras)
assert.NoError(t, err)
}
releases, err := releaseRepo.Find(ctx, domain.ReleaseQueryParams{})
assert.NoError(t, err)
assert.Len(t, releases.Data, len(tt.fields.releaseTitles))
compareRel := domain.NewRelease(mockIndexer)
compareRel.ParseString(tt.fields.releaseTitle)
// Execute
isDuplicate, err := releaseRepo.CheckIsDuplicateRelease(ctx, tt.fields.profile, compareRel)
// Verify
assert.NoError(t, err)
assert.Equal(t, tt.isDuplicate, isDuplicate)
// Cleanup
_ = releaseRepo.Delete(ctx, &domain.DeleteReleaseRequest{OlderThan: 0})
})
}
// Cleanup
//_ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0})
_ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMock.ID})
_ = filterRepo.Delete(context.Background(), createdFilters[0].ID)
}
}

View file

@ -88,9 +88,42 @@ CREATE TABLE irc_channel
UNIQUE (network_id, name)
);
CREATE TABLE release_profile_duplicate
(
id INTEGER PRIMARY KEY AUTOINCREMENT,
name TEXT NOT NULL,
protocol BOOLEAN DEFAULT FALSE,
release_name BOOLEAN DEFAULT FALSE,
hash BOOLEAN DEFAULT FALSE,
title BOOLEAN DEFAULT FALSE,
sub_title BOOLEAN DEFAULT FALSE,
year BOOLEAN DEFAULT FALSE,
month BOOLEAN DEFAULT FALSE,
day BOOLEAN DEFAULT FALSE,
source BOOLEAN DEFAULT FALSE,
resolution BOOLEAN DEFAULT FALSE,
codec BOOLEAN DEFAULT FALSE,
container BOOLEAN DEFAULT FALSE,
dynamic_range BOOLEAN DEFAULT FALSE,
audio BOOLEAN DEFAULT FALSE,
release_group BOOLEAN DEFAULT FALSE,
season BOOLEAN DEFAULT FALSE,
episode BOOLEAN DEFAULT FALSE,
website BOOLEAN DEFAULT FALSE,
proper BOOLEAN DEFAULT FALSE,
repack BOOLEAN DEFAULT FALSE,
edition BOOLEAN DEFAULT FALSE,
language BOOLEAN DEFAULT FALSE
);
INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language)
VALUES (1, 'Exact release', 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0),
(2, 'Movie', 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0),
(3, 'TV', 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0);
CREATE TABLE filter
(
id INTEGER PRIMARY KEY,
id INTEGER PRIMARY KEY AUTOINCREMENT,
enabled BOOLEAN,
name TEXT NOT NULL,
min_size TEXT,
@ -159,7 +192,9 @@ CREATE TABLE filter
min_seeders INTEGER DEFAULT 0,
max_seeders INTEGER DEFAULT 0,
min_leechers INTEGER DEFAULT 0,
max_leechers INTEGER DEFAULT 0
max_leechers INTEGER DEFAULT 0,
release_profile_duplicate_id INTEGER,
FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate(id) ON DELETE SET NULL
);
CREATE INDEX filter_enabled_index
@ -273,8 +308,10 @@ CREATE TABLE "release"
group_id TEXT,
torrent_id TEXT,
torrent_name TEXT,
normalized_hash TEXT,
size INTEGER,
title TEXT,
sub_title TEXT,
category TEXT,
season INTEGER,
episode INTEGER,
@ -286,15 +323,24 @@ CREATE TABLE "release"
codec TEXT,
container TEXT,
hdr TEXT,
audio TEXT,
audio_channels TEXT,
release_group TEXT,
region TEXT,
language TEXT,
edition TEXT,
cut TEXT,
hybrid BOOLEAN,
proper BOOLEAN,
repack BOOLEAN,
website TEXT,
media_processing TEXT,
type TEXT,
origin TEXT,
tags TEXT [] DEFAULT '{}' NOT NULL,
uploader TEXT,
pre_time TEXT,
other TEXT [] DEFAULT '{}' NOT NULL,
filter_id INTEGER
REFERENCES filter
ON DELETE SET NULL
@ -312,6 +358,81 @@ CREATE INDEX release_timestamp_index
CREATE INDEX release_torrent_name_index
ON "release" (torrent_name);
CREATE INDEX release_normalized_hash_index
ON "release" (normalized_hash);
CREATE INDEX release_title_index
ON "release" (title);
CREATE INDEX release_sub_title_index
ON "release" (sub_title);
CREATE INDEX release_season_index
ON "release" (season);
CREATE INDEX release_episode_index
ON "release" (episode);
CREATE INDEX release_year_index
ON "release" (year);
CREATE INDEX release_month_index
ON "release" (month);
CREATE INDEX release_day_index
ON "release" (day);
CREATE INDEX release_resolution_index
ON "release" (resolution);
CREATE INDEX release_source_index
ON "release" (source);
CREATE INDEX release_codec_index
ON "release" (codec);
CREATE INDEX release_container_index
ON "release" (container);
CREATE INDEX release_hdr_index
ON "release" (hdr);
CREATE INDEX release_audio_index
ON "release" (audio);
CREATE INDEX release_audio_channels_index
ON "release" (audio_channels);
CREATE INDEX release_release_group_index
ON "release" (release_group);
CREATE INDEX release_language_index
ON "release" (language);
CREATE INDEX release_proper_index
ON "release" (proper);
CREATE INDEX release_repack_index
ON "release" (repack);
CREATE INDEX release_website_index
ON "release" (website);
CREATE INDEX release_media_processing_index
ON "release" (media_processing);
CREATE INDEX release_region_index
ON "release" (region);
CREATE INDEX release_edition_index
ON "release" (edition);
CREATE INDEX release_cut_index
ON "release" (cut);
CREATE INDEX release_hybrid_index
ON "release" (hybrid);
CREATE TABLE release_action_status
(
id INTEGER PRIMARY KEY,
@ -1716,5 +1837,152 @@ CREATE TABLE list_filter
ALTER TABLE filter
ADD COLUMN except_record_labels TEXT DEFAULT '';
`,
`CREATE TABLE release_profile_duplicate
(
id INTEGER PRIMARY KEY AUTOINCREMENT,
name TEXT NOT NULL,
protocol BOOLEAN DEFAULT FALSE,
release_name BOOLEAN DEFAULT FALSE,
hash BOOLEAN DEFAULT FALSE,
title BOOLEAN DEFAULT FALSE,
sub_title BOOLEAN DEFAULT FALSE,
year BOOLEAN DEFAULT FALSE,
month BOOLEAN DEFAULT FALSE,
day BOOLEAN DEFAULT FALSE,
source BOOLEAN DEFAULT FALSE,
resolution BOOLEAN DEFAULT FALSE,
codec BOOLEAN DEFAULT FALSE,
container BOOLEAN DEFAULT FALSE,
dynamic_range BOOLEAN DEFAULT FALSE,
audio BOOLEAN DEFAULT FALSE,
release_group BOOLEAN DEFAULT FALSE,
season BOOLEAN DEFAULT FALSE,
episode BOOLEAN DEFAULT FALSE,
website BOOLEAN DEFAULT FALSE,
proper BOOLEAN DEFAULT FALSE,
repack BOOLEAN DEFAULT FALSE,
edition BOOLEAN DEFAULT FALSE,
language BOOLEAN DEFAULT FALSE
);
INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language)
VALUES (1, 'Exact release', 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0),
(2, 'Movie', 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0),
(3, 'TV', 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0);
ALTER TABLE filter
ADD COLUMN release_profile_duplicate_id INTEGER
CONSTRAINT filter_release_profile_duplicate_id_fk
REFERENCES release_profile_duplicate (id)
ON DELETE SET NULL;
ALTER TABLE "release"
ADD normalized_hash TEXT;
ALTER TABLE "release"
ADD sub_title TEXT;
ALTER TABLE "release"
ADD audio TEXT;
ALTER TABLE "release"
ADD audio_channels TEXT;
ALTER TABLE "release"
ADD language TEXT;
ALTER TABLE "release"
ADD media_processing TEXT;
ALTER TABLE "release"
ADD edition TEXT;
ALTER TABLE "release"
ADD cut TEXT;
ALTER TABLE "release"
ADD hybrid TEXT;
ALTER TABLE "release"
ADD region TEXT;
ALTER TABLE "release"
ADD other TEXT [] DEFAULT '{}' NOT NULL;
CREATE INDEX release_normalized_hash_index
ON "release" (normalized_hash);
CREATE INDEX release_title_index
ON "release" (title);
CREATE INDEX release_sub_title_index
ON "release" (sub_title);
CREATE INDEX release_season_index
ON "release" (season);
CREATE INDEX release_episode_index
ON "release" (episode);
CREATE INDEX release_year_index
ON "release" (year);
CREATE INDEX release_month_index
ON "release" (month);
CREATE INDEX release_day_index
ON "release" (day);
CREATE INDEX release_resolution_index
ON "release" (resolution);
CREATE INDEX release_source_index
ON "release" (source);
CREATE INDEX release_codec_index
ON "release" (codec);
CREATE INDEX release_container_index
ON "release" (container);
CREATE INDEX release_hdr_index
ON "release" (hdr);
CREATE INDEX release_audio_index
ON "release" (audio);
CREATE INDEX release_audio_channels_index
ON "release" (audio_channels);
CREATE INDEX release_release_group_index
ON "release" (release_group);
CREATE INDEX release_proper_index
ON "release" (proper);
CREATE INDEX release_repack_index
ON "release" (repack);
CREATE INDEX release_website_index
ON "release" (website);
CREATE INDEX release_media_processing_index
ON "release" (media_processing);
CREATE INDEX release_language_index
ON "release" (language);
CREATE INDEX release_region_index
ON "release" (region);
CREATE INDEX release_edition_index
ON "release" (edition);
CREATE INDEX release_cut_index
ON "release" (cut);
CREATE INDEX release_hybrid_index
ON "release" (hybrid);
`,
}

View file

@ -29,6 +29,7 @@ func toNullInt32(s int32) sql.NullInt32 {
Valid: s != 0,
}
}
func toNullInt64(s int64) sql.NullInt64 {
return sql.NullInt64{
Int64: s,

View file

@ -11,7 +11,7 @@ import (
)
type ActionRepo interface {
Store(ctx context.Context, action Action) (*Action, error)
Store(ctx context.Context, action *Action) error
StoreFilterActions(ctx context.Context, filterID int64, actions []*Action) ([]*Action, error)
FindByFilterID(ctx context.Context, filterID int, active *bool, withClient bool) ([]*Action, error)
List(ctx context.Context) ([]Action, error)

View file

@ -91,88 +91,90 @@ type FilterQueryParams struct {
}
type Filter struct {
ID int `json:"id"`
Name string `json:"name"`
Enabled bool `json:"enabled"`
CreatedAt time.Time `json:"created_at"`
UpdatedAt time.Time `json:"updated_at"`
MinSize string `json:"min_size,omitempty"`
MaxSize string `json:"max_size,omitempty"`
Delay int `json:"delay,omitempty"`
Priority int32 `json:"priority"`
MaxDownloads int `json:"max_downloads,omitempty"`
MaxDownloadsUnit FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"`
MatchReleases string `json:"match_releases,omitempty"`
ExceptReleases string `json:"except_releases,omitempty"`
UseRegex bool `json:"use_regex,omitempty"`
MatchReleaseGroups string `json:"match_release_groups,omitempty"`
ExceptReleaseGroups string `json:"except_release_groups,omitempty"`
AnnounceTypes []string `json:"announce_types,omitempty"`
Scene bool `json:"scene,omitempty"`
Origins []string `json:"origins,omitempty"`
ExceptOrigins []string `json:"except_origins,omitempty"`
Bonus []string `json:"bonus,omitempty"`
Freeleech bool `json:"freeleech,omitempty"`
FreeleechPercent string `json:"freeleech_percent,omitempty"`
SmartEpisode bool `json:"smart_episode"`
Shows string `json:"shows,omitempty"`
Seasons string `json:"seasons,omitempty"`
Episodes string `json:"episodes,omitempty"`
Resolutions []string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p.
Codecs []string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux).
Sources []string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC
Containers []string `json:"containers,omitempty"`
MatchHDR []string `json:"match_hdr,omitempty"`
ExceptHDR []string `json:"except_hdr,omitempty"`
MatchOther []string `json:"match_other,omitempty"`
ExceptOther []string `json:"except_other,omitempty"`
Years string `json:"years,omitempty"`
Months string `json:"months,omitempty"`
Days string `json:"days,omitempty"`
Artists string `json:"artists,omitempty"`
Albums string `json:"albums,omitempty"`
MatchReleaseTypes []string `json:"match_release_types,omitempty"` // Album,Single,EP
ExceptReleaseTypes string `json:"except_release_types,omitempty"`
Formats []string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS
Quality []string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other
Media []string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other
PerfectFlac bool `json:"perfect_flac,omitempty"`
Cue bool `json:"cue,omitempty"`
Log bool `json:"log,omitempty"`
LogScore int `json:"log_score,omitempty"`
MatchCategories string `json:"match_categories,omitempty"`
ExceptCategories string `json:"except_categories,omitempty"`
MatchUploaders string `json:"match_uploaders,omitempty"`
ExceptUploaders string `json:"except_uploaders,omitempty"`
MatchRecordLabels string `json:"match_record_labels,omitempty"`
ExceptRecordLabels string `json:"except_record_labels,omitempty"`
MatchLanguage []string `json:"match_language,omitempty"`
ExceptLanguage []string `json:"except_language,omitempty"`
Tags string `json:"tags,omitempty"`
ExceptTags string `json:"except_tags,omitempty"`
TagsAny string `json:"tags_any,omitempty"`
ExceptTagsAny string `json:"except_tags_any,omitempty"`
TagsMatchLogic string `json:"tags_match_logic,omitempty"`
ExceptTagsMatchLogic string `json:"except_tags_match_logic,omitempty"`
MatchReleaseTags string `json:"match_release_tags,omitempty"`
ExceptReleaseTags string `json:"except_release_tags,omitempty"`
UseRegexReleaseTags bool `json:"use_regex_release_tags,omitempty"`
MatchDescription string `json:"match_description,omitempty"`
ExceptDescription string `json:"except_description,omitempty"`
UseRegexDescription bool `json:"use_regex_description,omitempty"`
MinSeeders int `json:"min_seeders,omitempty"`
MaxSeeders int `json:"max_seeders,omitempty"`
MinLeechers int `json:"min_leechers,omitempty"`
MaxLeechers int `json:"max_leechers,omitempty"`
ActionsCount int `json:"actions_count"`
ActionsEnabledCount int `json:"actions_enabled_count"`
IsAutoUpdated bool `json:"is_auto_updated"`
Actions []*Action `json:"actions,omitempty"`
External []FilterExternal `json:"external,omitempty"`
Indexers []Indexer `json:"indexers"`
Downloads *FilterDownloads `json:"-"`
Rejections []string `json:"-"`
RejectReasons *RejectionReasons `json:"-"`
ID int `json:"id"`
Name string `json:"name"`
Enabled bool `json:"enabled"`
CreatedAt time.Time `json:"created_at"`
UpdatedAt time.Time `json:"updated_at"`
MinSize string `json:"min_size,omitempty"`
MaxSize string `json:"max_size,omitempty"`
Delay int `json:"delay,omitempty"`
Priority int32 `json:"priority"`
MaxDownloads int `json:"max_downloads,omitempty"`
MaxDownloadsUnit FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"`
MatchReleases string `json:"match_releases,omitempty"`
ExceptReleases string `json:"except_releases,omitempty"`
UseRegex bool `json:"use_regex,omitempty"`
MatchReleaseGroups string `json:"match_release_groups,omitempty"`
ExceptReleaseGroups string `json:"except_release_groups,omitempty"`
AnnounceTypes []string `json:"announce_types,omitempty"`
Scene bool `json:"scene,omitempty"`
Origins []string `json:"origins,omitempty"`
ExceptOrigins []string `json:"except_origins,omitempty"`
Bonus []string `json:"bonus,omitempty"`
Freeleech bool `json:"freeleech,omitempty"`
FreeleechPercent string `json:"freeleech_percent,omitempty"`
SmartEpisode bool `json:"smart_episode"`
Shows string `json:"shows,omitempty"`
Seasons string `json:"seasons,omitempty"`
Episodes string `json:"episodes,omitempty"`
Resolutions []string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p.
Codecs []string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux).
Sources []string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC
Containers []string `json:"containers,omitempty"`
MatchHDR []string `json:"match_hdr,omitempty"`
ExceptHDR []string `json:"except_hdr,omitempty"`
MatchOther []string `json:"match_other,omitempty"`
ExceptOther []string `json:"except_other,omitempty"`
Years string `json:"years,omitempty"`
Months string `json:"months,omitempty"`
Days string `json:"days,omitempty"`
Artists string `json:"artists,omitempty"`
Albums string `json:"albums,omitempty"`
MatchReleaseTypes []string `json:"match_release_types,omitempty"` // Album,Single,EP
ExceptReleaseTypes string `json:"except_release_types,omitempty"`
Formats []string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS
Quality []string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other
Media []string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other
PerfectFlac bool `json:"perfect_flac,omitempty"`
Cue bool `json:"cue,omitempty"`
Log bool `json:"log,omitempty"`
LogScore int `json:"log_score,omitempty"`
MatchCategories string `json:"match_categories,omitempty"`
ExceptCategories string `json:"except_categories,omitempty"`
MatchUploaders string `json:"match_uploaders,omitempty"`
ExceptUploaders string `json:"except_uploaders,omitempty"`
MatchRecordLabels string `json:"match_record_labels,omitempty"`
ExceptRecordLabels string `json:"except_record_labels,omitempty"`
MatchLanguage []string `json:"match_language,omitempty"`
ExceptLanguage []string `json:"except_language,omitempty"`
Tags string `json:"tags,omitempty"`
ExceptTags string `json:"except_tags,omitempty"`
TagsAny string `json:"tags_any,omitempty"`
ExceptTagsAny string `json:"except_tags_any,omitempty"`
TagsMatchLogic string `json:"tags_match_logic,omitempty"`
ExceptTagsMatchLogic string `json:"except_tags_match_logic,omitempty"`
MatchReleaseTags string `json:"match_release_tags,omitempty"`
ExceptReleaseTags string `json:"except_release_tags,omitempty"`
UseRegexReleaseTags bool `json:"use_regex_release_tags,omitempty"`
MatchDescription string `json:"match_description,omitempty"`
ExceptDescription string `json:"except_description,omitempty"`
UseRegexDescription bool `json:"use_regex_description,omitempty"`
MinSeeders int `json:"min_seeders,omitempty"`
MaxSeeders int `json:"max_seeders,omitempty"`
MinLeechers int `json:"min_leechers,omitempty"`
MaxLeechers int `json:"max_leechers,omitempty"`
ActionsCount int `json:"actions_count"`
ActionsEnabledCount int `json:"actions_enabled_count"`
IsAutoUpdated bool `json:"is_auto_updated"`
Actions []*Action `json:"actions,omitempty"`
External []FilterExternal `json:"external,omitempty"`
Indexers []Indexer `json:"indexers"`
ReleaseProfileDuplicateID int64 `json:"release_profile_duplicate_id,omitempty"`
DuplicateHandling *DuplicateReleaseProfile `json:"release_profile_duplicate"`
Downloads *FilterDownloads `json:"-"`
Rejections []string `json:"-"`
RejectReasons *RejectionReasons `json:"-"`
}
type FilterExternal struct {
@ -219,80 +221,81 @@ const (
)
type FilterUpdate struct {
ID int `json:"id"`
Name *string `json:"name,omitempty"`
Enabled *bool `json:"enabled,omitempty"`
MinSize *string `json:"min_size,omitempty"`
MaxSize *string `json:"max_size,omitempty"`
Delay *int `json:"delay,omitempty"`
Priority *int32 `json:"priority,omitempty"`
AnnounceTypes *[]string `json:"announce_types,omitempty"`
MaxDownloads *int `json:"max_downloads,omitempty"`
MaxDownloadsUnit *FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"`
MatchReleases *string `json:"match_releases,omitempty"`
ExceptReleases *string `json:"except_releases,omitempty"`
UseRegex *bool `json:"use_regex,omitempty"`
MatchReleaseGroups *string `json:"match_release_groups,omitempty"`
ExceptReleaseGroups *string `json:"except_release_groups,omitempty"`
MatchReleaseTags *string `json:"match_release_tags,omitempty"`
ExceptReleaseTags *string `json:"except_release_tags,omitempty"`
UseRegexReleaseTags *bool `json:"use_regex_release_tags,omitempty"`
MatchDescription *string `json:"match_description,omitempty"`
ExceptDescription *string `json:"except_description,omitempty"`
UseRegexDescription *bool `json:"use_regex_description,omitempty"`
Scene *bool `json:"scene,omitempty"`
Origins *[]string `json:"origins,omitempty"`
ExceptOrigins *[]string `json:"except_origins,omitempty"`
Bonus *[]string `json:"bonus,omitempty"`
Freeleech *bool `json:"freeleech,omitempty"`
FreeleechPercent *string `json:"freeleech_percent,omitempty"`
SmartEpisode *bool `json:"smart_episode,omitempty"`
Shows *string `json:"shows,omitempty"`
Seasons *string `json:"seasons,omitempty"`
Episodes *string `json:"episodes,omitempty"`
Resolutions *[]string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p.
Codecs *[]string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux).
Sources *[]string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC
Containers *[]string `json:"containers,omitempty"`
MatchHDR *[]string `json:"match_hdr,omitempty"`
ExceptHDR *[]string `json:"except_hdr,omitempty"`
MatchOther *[]string `json:"match_other,omitempty"`
ExceptOther *[]string `json:"except_other,omitempty"`
Years *string `json:"years,omitempty"`
Months *string `json:"months,omitempty"`
Days *string `json:"days,omitempty"`
Artists *string `json:"artists,omitempty"`
Albums *string `json:"albums,omitempty"`
MatchReleaseTypes *[]string `json:"match_release_types,omitempty"` // Album,Single,EP
ExceptReleaseTypes *string `json:"except_release_types,omitempty"`
Formats *[]string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS
Quality *[]string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other
Media *[]string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other
PerfectFlac *bool `json:"perfect_flac,omitempty"`
Cue *bool `json:"cue,omitempty"`
Log *bool `json:"log,omitempty"`
LogScore *int `json:"log_score,omitempty"`
MatchCategories *string `json:"match_categories,omitempty"`
ExceptCategories *string `json:"except_categories,omitempty"`
MatchUploaders *string `json:"match_uploaders,omitempty"`
ExceptUploaders *string `json:"except_uploaders,omitempty"`
MatchRecordLabels *string `json:"match_record_labels,omitempty"`
ExceptRecordLabels *string `json:"except_record_labels,omitempty"`
MatchLanguage *[]string `json:"match_language,omitempty"`
ExceptLanguage *[]string `json:"except_language,omitempty"`
Tags *string `json:"tags,omitempty"`
ExceptTags *string `json:"except_tags,omitempty"`
TagsAny *string `json:"tags_any,omitempty"`
ExceptTagsAny *string `json:"except_tags_any,omitempty"`
TagsMatchLogic *string `json:"tags_match_logic,omitempty"`
ExceptTagsMatchLogic *string `json:"except_tags_match_logic,omitempty"`
MinSeeders *int `json:"min_seeders,omitempty"`
MaxSeeders *int `json:"max_seeders,omitempty"`
MinLeechers *int `json:"min_leechers,omitempty"`
MaxLeechers *int `json:"max_leechers,omitempty"`
Actions []*Action `json:"actions,omitempty"`
External []FilterExternal `json:"external,omitempty"`
Indexers []Indexer `json:"indexers,omitempty"`
ID int `json:"id"`
Name *string `json:"name,omitempty"`
Enabled *bool `json:"enabled,omitempty"`
MinSize *string `json:"min_size,omitempty"`
MaxSize *string `json:"max_size,omitempty"`
Delay *int `json:"delay,omitempty"`
Priority *int32 `json:"priority,omitempty"`
AnnounceTypes *[]string `json:"announce_types,omitempty"`
MaxDownloads *int `json:"max_downloads,omitempty"`
MaxDownloadsUnit *FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"`
MatchReleases *string `json:"match_releases,omitempty"`
ExceptReleases *string `json:"except_releases,omitempty"`
UseRegex *bool `json:"use_regex,omitempty"`
MatchReleaseGroups *string `json:"match_release_groups,omitempty"`
ExceptReleaseGroups *string `json:"except_release_groups,omitempty"`
MatchReleaseTags *string `json:"match_release_tags,omitempty"`
ExceptReleaseTags *string `json:"except_release_tags,omitempty"`
UseRegexReleaseTags *bool `json:"use_regex_release_tags,omitempty"`
MatchDescription *string `json:"match_description,omitempty"`
ExceptDescription *string `json:"except_description,omitempty"`
UseRegexDescription *bool `json:"use_regex_description,omitempty"`
Scene *bool `json:"scene,omitempty"`
Origins *[]string `json:"origins,omitempty"`
ExceptOrigins *[]string `json:"except_origins,omitempty"`
Bonus *[]string `json:"bonus,omitempty"`
Freeleech *bool `json:"freeleech,omitempty"`
FreeleechPercent *string `json:"freeleech_percent,omitempty"`
SmartEpisode *bool `json:"smart_episode,omitempty"`
Shows *string `json:"shows,omitempty"`
Seasons *string `json:"seasons,omitempty"`
Episodes *string `json:"episodes,omitempty"`
Resolutions *[]string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p.
Codecs *[]string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux).
Sources *[]string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC
Containers *[]string `json:"containers,omitempty"`
MatchHDR *[]string `json:"match_hdr,omitempty"`
ExceptHDR *[]string `json:"except_hdr,omitempty"`
MatchOther *[]string `json:"match_other,omitempty"`
ExceptOther *[]string `json:"except_other,omitempty"`
Years *string `json:"years,omitempty"`
Months *string `json:"months,omitempty"`
Days *string `json:"days,omitempty"`
Artists *string `json:"artists,omitempty"`
Albums *string `json:"albums,omitempty"`
MatchReleaseTypes *[]string `json:"match_release_types,omitempty"` // Album,Single,EP
ExceptReleaseTypes *string `json:"except_release_types,omitempty"`
Formats *[]string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS
Quality *[]string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other
Media *[]string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other
PerfectFlac *bool `json:"perfect_flac,omitempty"`
Cue *bool `json:"cue,omitempty"`
Log *bool `json:"log,omitempty"`
LogScore *int `json:"log_score,omitempty"`
MatchCategories *string `json:"match_categories,omitempty"`
ExceptCategories *string `json:"except_categories,omitempty"`
MatchUploaders *string `json:"match_uploaders,omitempty"`
ExceptUploaders *string `json:"except_uploaders,omitempty"`
MatchRecordLabels *string `json:"match_record_labels,omitempty"`
ExceptRecordLabels *string `json:"except_record_labels,omitempty"`
MatchLanguage *[]string `json:"match_language,omitempty"`
ExceptLanguage *[]string `json:"except_language,omitempty"`
Tags *string `json:"tags,omitempty"`
ExceptTags *string `json:"except_tags,omitempty"`
TagsAny *string `json:"tags_any,omitempty"`
ExceptTagsAny *string `json:"except_tags_any,omitempty"`
TagsMatchLogic *string `json:"tags_match_logic,omitempty"`
ExceptTagsMatchLogic *string `json:"except_tags_match_logic,omitempty"`
MinSeeders *int `json:"min_seeders,omitempty"`
MaxSeeders *int `json:"max_seeders,omitempty"`
MinLeechers *int `json:"min_leechers,omitempty"`
MaxLeechers *int `json:"max_leechers,omitempty"`
ReleaseProfileDuplicateID *int64 `json:"release_profile_duplicate_id,omitempty"`
Actions []*Action `json:"actions,omitempty"`
External []FilterExternal `json:"external,omitempty"`
Indexers []Indexer `json:"indexers,omitempty"`
}
func (f *Filter) Validate() error {

View file

@ -50,6 +50,7 @@ type Macro struct {
IndexerIdentifierExternal string
IndexerName string
InfoUrl string
IsDuplicate bool
Language []string
Leechers int
LogScore int
@ -66,6 +67,8 @@ type Macro struct {
Seeders int
Size uint64
SizeString string
SkipDuplicateProfileID int64
SkipDuplicateProfileName string
Source string
Tags string
Title string
@ -123,6 +126,7 @@ func NewMacro(release Release) Macro {
IndexerIdentifierExternal: release.Indexer.IdentifierExternal,
IndexerName: release.Indexer.Name,
InfoUrl: release.InfoURL,
IsDuplicate: release.IsDuplicate,
Language: release.Language,
Leechers: release.Leechers,
LogScore: release.LogScore,
@ -140,6 +144,8 @@ func NewMacro(release Release) Macro {
Size: release.Size,
SizeString: humanize.Bytes(release.Size),
Source: release.Source,
SkipDuplicateProfileID: release.SkipDuplicateProfileID,
SkipDuplicateProfileName: release.SkipDuplicateProfileName,
Tags: strings.Join(release.Tags, ", "),
Title: release.Title,
TorrentDataRawBytes: release.TorrentDataRawBytes,
@ -149,7 +155,7 @@ func NewMacro(release Release) Macro {
TorrentPathName: release.TorrentTmpFile,
TorrentUrl: release.DownloadURL,
TorrentTmpFile: release.TorrentTmpFile,
Type: release.Type,
Type: release.Type.String(),
Uploader: release.Uploader,
RecordLabel: release.RecordLabel,
Website: release.Website,

View file

@ -5,6 +5,7 @@ package domain
import (
"fmt"
"github.com/moistari/rls"
"testing"
"time"
@ -231,7 +232,7 @@ func TestMacros_Parse(t *testing.T) {
{
name: "test_type",
release: Release{
Type: "episode",
Type: rls.Episode,
},
args: args{text: "Type: {{ .Type }}"},
want: "Type: episode",

View file

@ -7,7 +7,11 @@ import (
"bufio"
"bytes"
"context"
"crypto/md5"
"encoding/hex"
"fmt"
"golang.org/x/text/transform"
"golang.org/x/text/unicode/norm"
"html"
"io"
"math"
@ -18,6 +22,7 @@ import (
"strconv"
"strings"
"time"
"unicode"
"github.com/autobrr/autobrr/pkg/errors"
"github.com/autobrr/autobrr/pkg/sharedhttp"
@ -43,6 +48,11 @@ type ReleaseRepo interface {
GetActionStatus(ctx context.Context, req *GetReleaseActionStatusRequest) (*ReleaseActionStatus, error)
StoreReleaseActionStatus(ctx context.Context, status *ReleaseActionStatus) error
StoreDuplicateProfile(ctx context.Context, profile *DuplicateReleaseProfile) error
FindDuplicateReleaseProfiles(ctx context.Context) ([]*DuplicateReleaseProfile, error)
DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error
CheckIsDuplicateRelease(ctx context.Context, profile *DuplicateReleaseProfile, release *Release) (bool, error)
}
type Release struct {
@ -55,6 +65,7 @@ type Release struct {
Implementation ReleaseImplementation `json:"implementation"` // irc, rss, api
Timestamp time.Time `json:"timestamp"`
AnnounceType AnnounceType `json:"announce_type"`
Type rls.Type `json:"type"` // rls.Type
InfoURL string `json:"info_url"`
DownloadURL string `json:"download_url"`
MagnetURI string `json:"-"`
@ -63,9 +74,11 @@ type Release struct {
TorrentTmpFile string `json:"-"`
TorrentDataRawBytes []byte `json:"-"`
TorrentHash string `json:"-"`
TorrentName string `json:"name"` // full release name
TorrentName string `json:"name"` // full release name
NormalizedHash string `json:"normalized_hash"` // normalized torrent name and md5 hashed
Size uint64 `json:"size"`
Title string `json:"title"` // Parsed title
Title string `json:"title"` // Parsed title
SubTitle string `json:"sub_title"` // Parsed secondary title for shows e.g. episode name
Description string `json:"-"`
Category string `json:"category"`
Categories []string `json:"categories,omitempty"`
@ -89,8 +102,11 @@ type Release struct {
Proper bool `json:"proper"`
Repack bool `json:"repack"`
Website string `json:"website"`
Hybrid bool `json:"hybrid"`
Edition []string `json:"edition"`
Cut []string `json:"cut"`
MediaProcessing string `json:"media_processing"` // Remux, Encode, Untouched
Artists string `json:"-"`
Type string `json:"type"` // Album,Single,EP
LogScore int `json:"-"`
HasCue bool `json:"-"`
HasLog bool `json:"-"`
@ -110,15 +126,183 @@ type Release struct {
AdditionalSizeCheckRequired bool `json:"-"`
AdditionalUploaderCheckRequired bool `json:"-"`
AdditionalRecordLabelCheckRequired bool `json:"-"`
IsDuplicate bool `json:"-"`
SkipDuplicateProfileID int64 `json:"-"`
SkipDuplicateProfileName string `json:"-"`
FilterID int `json:"-"`
Filter *Filter `json:"-"`
ActionStatus []ReleaseActionStatus `json:"action_status"`
}
// Hash return md5 hashed normalized release name
func (r *Release) Hash() string {
formatted := r.TorrentName
// for tv and movies we create the formatted title to have the best chance of matching
if r.IsTypeVideo() {
formatted = r.NormalizedTitle()
}
normalized := MustNormalize(formatted)
h := md5.Sum([]byte(normalized))
str := hex.EncodeToString(h[:])
return str
}
// MustNormalize applies the Normalize transform to s, returning a lower cased,
// clean form of s useful for matching titles.
func MustNormalize(s string) string {
s, _, err := transform.String(NewNormalizer(), s)
if err != nil {
panic(err)
}
return s
}
// NewNormalizer is a custom rls.Normalizer that keeps plus sign + for HDR10+ fx
// It creates a new a text transformer chain (similiar to
// NewCleaner) that normalizes text to lower case clean form useful for
// matching titles.
//
// See: https://go.dev/blog/normalization
func NewNormalizer() transform.Transformer {
return transform.Chain(
norm.NFD,
rls.NewCollapser(
true, true,
"`"+`':;~!@#%^*=()[]{}<>/?|\",`, " \t\r\n\f._",
func(r, prev, next rune) rune {
switch {
case r == '-' && unicode.IsSpace(prev):
return -1
case r == '$' && (unicode.IsLetter(prev) || unicode.IsLetter(next)):
return 'S'
case r == '£' && (unicode.IsLetter(prev) || unicode.IsLetter(next)):
return 'L'
case r == '$', r == '£':
return -1
}
return r
},
),
norm.NFC,
)
}
func (r *Release) NormalizedTitle() string {
var v []string
v = append(v, r.Title)
if r.Year > 0 && r.Month > 0 && r.Day > 0 {
v = append(v, fmt.Sprintf("%d %d %d", r.Year, r.Month, r.Day))
} else if r.Year > 0 {
v = append(v, fmt.Sprintf("%d", r.Year))
}
if len(r.Language) > 0 {
v = append(v, strings.Join(r.Language, " "))
}
if len(r.Cut) > 0 {
v = append(v, strings.Join(r.Cut, " "))
}
if len(r.Edition) > 0 {
v = append(v, strings.Join(r.Edition, " "))
}
if r.Season > 0 && r.Episode > 0 {
v = append(v, fmt.Sprintf("S%dE%d", r.Season, r.Episode))
} else if r.Season > 0 && r.Episode == 0 {
v = append(v, fmt.Sprintf("S%d", r.Season))
}
if r.Proper {
v = append(v, "PROPER")
}
if r.Repack {
v = append(v, r.RepackStr())
}
if r.Hybrid {
v = append(v, "HYBRiD")
}
if r.SubTitle != "" {
v = append(v, r.SubTitle)
}
if r.Resolution != "" {
v = append(v, r.Resolution)
}
if r.Website != "" {
v = append(v, r.Website)
}
if r.Region != "" {
v = append(v, r.Region)
}
if r.Source != "" {
v = append(v, r.Source)
}
// remux
if r.MediaProcessing == "REMUX" {
v = append(v, "REMUX")
}
if len(r.Codec) > 0 {
v = append(v, strings.Join(r.Codec, " "))
}
if len(r.HDR) > 0 {
v = append(v, strings.Join(r.HDR, " "))
}
if len(r.Audio) > 0 {
v = append(v, r.AudioString())
}
str := strings.Join(v, " ")
if r.Group != "" {
str = fmt.Sprintf("%s-%s", str, r.Group)
}
return str
}
func (r *Release) RepackStr() string {
if r.Other != nil {
if slices.Contains(r.Other, "REPACK") {
return "REPACK"
} else if slices.Contains(r.Other, "REREPACK") {
return "REREPACK"
} else if slices.Contains(r.Other, "REPACK2") {
return "REPACK2"
} else if slices.Contains(r.Other, "REPACK3") {
return "REPACK3"
}
}
return ""
}
func (r *Release) Raw(s string) rls.Release {
return rls.ParseString(s)
}
func (r *Release) ParseType(s string) {
r.Type = rls.ParseType(s)
}
func (r *Release) IsTypeVideo() bool {
return r.Type.Is(rls.Movie, rls.Series, rls.Episode)
}
type AnnounceType string
const (
@ -361,6 +545,10 @@ func NewRelease(indexer IndexerMinimal) *Release {
Implementation: ReleaseImplementationIRC,
Timestamp: time.Now(),
Tags: []string{},
Language: []string{},
Edition: []string{},
Cut: []string{},
Other: []string{},
Size: 0,
AnnounceType: AnnounceTypeNew,
}
@ -371,28 +559,42 @@ func NewRelease(indexer IndexerMinimal) *Release {
func (r *Release) ParseString(title string) {
rel := rls.ParseString(title)
r.Type = rel.Type.String()
r.Type = rel.Type
r.TorrentName = title
r.Source = rel.Source
r.Resolution = rel.Resolution
r.Region = rel.Region
if rel.Language != nil {
r.Language = rel.Language
}
r.Audio = rel.Audio
r.AudioChannels = rel.Channels
r.Codec = rel.Codec
r.Container = rel.Container
r.HDR = rel.HDR
r.Artists = rel.Artist
r.Language = rel.Language
r.Other = rel.Other
if rel.Other != nil {
r.Other = rel.Other
}
r.Proper = slices.Contains(r.Other, "PROPER")
r.Repack = slices.Contains(r.Other, "REPACK")
r.Repack = slices.Contains(r.Other, "REPACK") || slices.Contains(r.Other, "REREPACK")
r.Hybrid = slices.Contains(r.Other, "HYBRiD")
// TODO default to Encode and set Untouched for discs
if slices.Contains(r.Other, "REMUX") {
r.MediaProcessing = "REMUX"
}
if r.Title == "" {
r.Title = rel.Title
}
r.SubTitle = rel.Subtitle
if r.Season == 0 {
r.Season = rel.Series
@ -415,8 +617,22 @@ func (r *Release) ParseString(title string) {
r.Group = rel.Group
}
if r.Website == "" {
r.Website = rel.Collection
}
if rel.Cut != nil {
r.Cut = rel.Cut
}
if rel.Edition != nil {
r.Edition = rel.Edition
}
r.ParseReleaseTagsString(r.ReleaseTags)
r.extraParseSource(rel)
r.NormalizedHash = r.Hash()
}
func (r *Release) extraParseSource(rel rls.Release) {
@ -451,7 +667,7 @@ func (r *Release) extraParseSource(rel rls.Release) {
}
// check res to be 1080p or 2160p and codec to be AVC, HEVC or if other contains Remux, then set source to BluRay if it differs
if !basicContainsSlice(r.Source, []string{"WEB-DL", "BluRay", "UHD.BluRay"}) && basicContainsSlice(r.Resolution, []string{"1080p", "2160p"}) && basicContainsMatch(r.Codec, []string{"AVC", "HEVC"}) && basicContainsMatch(r.Other, []string{"REMUX"}) {
if !basicContainsSlice(r.Source, []string{"WEB-DL", "BluRay", "UHD.BluRay"}) && basicContainsSlice(r.Resolution, []string{"1080p", "2160p"}) && basicContainsMatch(r.Codec, []string{"AVC", "H.264", "H.265", "HEVC"}) && basicContainsMatch(r.Other, []string{"REMUX"}) {
// handle missing or unexpected source for some bluray releases
if r.Resolution == "1080p" {
r.Source = "BluRay"
@ -463,6 +679,10 @@ func (r *Release) extraParseSource(rel rls.Release) {
}
func (r *Release) ParseReleaseTagsString(tags string) {
if tags == "" {
return
}
cleanTags := CleanReleaseTags(tags)
t := ParseReleaseTagString(cleanTags)
@ -543,6 +763,20 @@ func (r *Release) OpenTorrentFile() error {
return nil
}
// AudioString takes r.Audio and r.AudioChannels and returns a string like "DDP Atmos 5.1"
func (r *Release) AudioString() string {
var audio []string
audio = append(audio, r.Audio...)
audio = append(audio, r.AudioChannels)
if len(audio) > 0 {
return strings.Join(audio, " ")
}
return ""
}
func (r *Release) DownloadTorrentFileCtx(ctx context.Context) error {
return r.downloadTorrentFile(ctx)
}
@ -992,3 +1226,30 @@ func getUniqueTags(target []string, source []string) []string {
return target
}
type DuplicateReleaseProfile struct {
ID int64 `json:"id"`
Name string `json:"name"`
Protocol bool `json:"protocol"`
ReleaseName bool `json:"release_name"`
Hash bool `json:"hash"`
Title bool `json:"title"`
SubTitle bool `json:"sub_title"`
Year bool `json:"year"`
Month bool `json:"month"`
Day bool `json:"day"`
Source bool `json:"source"`
Resolution bool `json:"resolution"`
Codec bool `json:"codec"`
Container bool `json:"container"`
DynamicRange bool `json:"dynamic_range"`
Audio bool `json:"audio"`
Group bool `json:"group"`
Season bool `json:"season"`
Episode bool `json:"episode"`
Website bool `json:"website"`
Proper bool `json:"proper"`
Repack bool `json:"repack"`
Edition bool `json:"edition"`
Language bool `json:"language"`
}

View file

@ -15,6 +15,7 @@ import (
"testing"
"time"
"github.com/moistari/rls"
"github.com/rs/zerolog"
)
@ -127,7 +128,7 @@ func TestRelease_DownloadTorrentFile(t *testing.T) {
Repack bool
Website string
Artists string
Type string
Type rls.Type
LogScore int
Origin string
Tags []string

View file

@ -6,6 +6,7 @@ package domain
import (
"testing"
"github.com/moistari/rls"
"github.com/stretchr/testify/assert"
)
@ -21,19 +22,20 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX",
},
want: Release{
TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
//Website: "ATVP",
Type: "series",
TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -42,18 +44,20 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
},
want: Release{
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Type: "series",
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -63,20 +67,22 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MKV / 2160p / WEB-DL",
},
want: Release{
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
ReleaseTags: "MKV / 2160p / WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mkv",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Type: "series",
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
ReleaseTags: "MKV / 2160p / WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mkv",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -86,20 +92,22 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MKV | 2160p | WEB-DL",
},
want: Release{
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
ReleaseTags: "MKV | 2160p | WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mkv",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Type: "series",
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
ReleaseTags: "MKV | 2160p | WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mkv",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -109,20 +117,22 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MP4 | 2160p | WEB-DL",
},
want: Release{
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
ReleaseTags: "MP4 | 2160p | WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mp4",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Type: "series",
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
ReleaseTags: "MP4 | 2160p | WEB-DL",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mp4",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -132,22 +142,24 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!",
},
want: Release{
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mp4",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Freeleech: true,
Bonus: []string{"Freeleech"},
Type: "series",
TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX",
NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0",
ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!",
Title: "Servant",
Season: 1,
Episode: 0,
Resolution: "2160p",
Source: "WEB-DL",
Container: "mp4",
Codec: []string{"HEVC"},
Audio: []string{"DDP", "Atmos"},
AudioChannels: "5.1",
HDR: []string{"DV"},
Group: "FLUX",
Freeleech: true,
Bonus: []string{"Freeleech"},
Website: "ATVP",
Type: rls.Series,
},
},
{
@ -156,15 +168,16 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP",
},
want: Release{
TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP",
Title: "Rippers Revenge",
Year: 2023,
Language: []string{"GERMAN", "DL"},
Resolution: "1080p",
Source: "BluRay",
Codec: []string{"MPEG-2"},
Group: "GROUP",
Type: "movie",
TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP",
NormalizedHash: "cb28c69ce117723d01ea6562ef3bae67",
Title: "Rippers Revenge",
Year: 2023,
Language: []string{"GERMAN", "DL"},
Resolution: "1080p",
Source: "BluRay",
Codec: []string{"MPEG-2"},
Group: "GROUP",
Type: rls.Movie,
},
},
{
@ -173,13 +186,14 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF",
},
want: Release{
TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF",
Title: "Analogue",
Resolution: "1080p", // rls does not differentiate between 1080i and 1080p which results in all 1080 releases being parsed as 1080p
Source: "AHDTV",
Codec: []string{"H.264"},
Group: "ABCDEF",
Type: "movie",
TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF",
NormalizedHash: "1a1eb25b15269d3189138326658aafe0",
Title: "Analogue",
Resolution: "1080p", // rls does not differentiate between 1080i and 1080p which results in all 1080 releases being parsed as 1080p
Source: "AHDTV",
Codec: []string{"H.264"},
Group: "ABCDEF",
Type: rls.Movie,
},
},
{
@ -189,17 +203,18 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
},
want: Release{
TorrentName: "Artist - Albumname",
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
TorrentName: "Artist - Albumname",
NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac",
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
},
},
{
@ -210,14 +225,15 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MP3 / 320 / Cassette",
},
want: Release{
TorrentName: "Various Artists - Music '21",
Tags: []string{"house, techno, tech.house, electro.house, future.house, bass.house, melodic.house"},
ReleaseTags: "MP3 / 320 / Cassette",
Title: "Various Artists - Music '21",
Source: "Cassette",
Audio: []string{"320", "MP3"},
AudioFormat: "MP3",
Bitrate: "320",
TorrentName: "Various Artists - Music '21",
NormalizedHash: "fbf7713532f4d2d03b62897fc990e37e",
Tags: []string{"house, techno, tech.house, electro.house, future.house, bass.house, melodic.house"},
ReleaseTags: "MP3 / 320 / Cassette",
Title: "Various Artists - Music '21",
Source: "Cassette",
Audio: []string{"320", "MP3"},
AudioFormat: "MP3",
Bitrate: "320",
},
},
{
@ -227,14 +243,15 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "MP3 / V0 (VBR) / CD",
},
want: Release{
TorrentName: "The artist (ザ・フリーダムユニティ) - Long album name",
ReleaseTags: "MP3 / V0 (VBR) / CD",
Title: "The artist",
Group: "name",
Source: "CD",
Audio: []string{"MP3", "VBR", "V0 (VBR)"},
AudioFormat: "MP3",
Bitrate: "V0 (VBR)",
TorrentName: "The artist (ザ・フリーダムユニティ) - Long album name",
NormalizedHash: "6b4051de7524f65d3c25be535d2c95ed",
ReleaseTags: "MP3 / V0 (VBR) / CD",
Title: "The artist",
Group: "name",
Source: "CD",
Audio: []string{"MP3", "VBR", "V0 (VBR)"},
AudioFormat: "MP3",
Bitrate: "V0 (VBR)",
},
},
{
@ -244,17 +261,18 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
},
want: Release{
TorrentName: "Artist - Albumname",
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
TorrentName: "Artist - Albumname",
NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac",
ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
},
},
{
@ -264,17 +282,18 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD",
},
want: Release{
TorrentName: "Artist - Albumname",
ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "24BIT Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
TorrentName: "Artist - Albumname",
NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac",
ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log100", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "24BIT Lossless",
HasLog: true,
LogScore: 100,
HasCue: true,
},
},
{
@ -284,17 +303,18 @@ func TestRelease_Parse(t *testing.T) {
ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD",
},
want: Release{
TorrentName: "Artist - Albumname",
ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log78", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "24BIT Lossless",
HasLog: true,
LogScore: 78,
HasCue: true,
TorrentName: "Artist - Albumname",
NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac",
ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD",
Title: "Artist",
Group: "Albumname",
Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log78", "Log"},
AudioFormat: "FLAC",
Source: "CD",
Bitrate: "24BIT Lossless",
HasLog: true,
LogScore: 78,
HasCue: true,
},
},
{
@ -303,18 +323,22 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1",
},
want: Release{
TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1",
Title: "I Am Movie",
Resolution: "2160p",
Source: "UHD.BluRay",
Codec: []string{"HEVC"},
HDR: []string{"DV"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Year: 2007,
Group: "GROUP1",
Other: []string{"HYBRiD", "REMUX"},
Type: "movie",
TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1",
NormalizedHash: "f70840f42149d7d2b539c75f5e853493",
Title: "I Am Movie",
Resolution: "2160p",
Source: "UHD.BluRay",
Codec: []string{"HEVC"},
HDR: []string{"DV"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Year: 2007,
Group: "GROUP1",
Hybrid: true,
Cut: []string{"Theatrical.Cut"},
MediaProcessing: "REMUX",
Other: []string{"HYBRiD", "REMUX"},
Type: rls.Movie,
},
},
{
@ -323,19 +347,21 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)",
},
want: Release{
TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)",
Title: "The Peripheral",
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"H.265"},
HDR: []string{"HDR10+"},
Audio: []string{"DDP"},
AudioChannels: "5.1",
Year: 2022,
Group: "GROUP1",
Season: 1,
Language: []string{"ENGLiSH"},
Type: "series",
TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)",
NormalizedHash: "33fe914c00ea5a2945883cffb4183923",
Title: "The Peripheral",
Resolution: "2160p",
Source: "WEB-DL",
Codec: []string{"H.265"},
HDR: []string{"HDR10+"},
Audio: []string{"DDP"},
AudioChannels: "5.1",
Year: 2022,
Group: "GROUP1",
Season: 1,
Language: []string{"ENGLiSH"},
Website: "AMZN",
Type: rls.Series,
},
},
{
@ -344,18 +370,20 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES",
},
want: Release{
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES",
Title: "Old Movie",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "CiNEPHiLES",
Type: "movie",
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES",
Title: "Old Movie",
NormalizedHash: "9990678babd4f24f268dfe7856b29f0e",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "CiNEPHiLES",
Type: rls.Movie,
MediaProcessing: "REMUX",
},
},
{
@ -364,17 +392,19 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD",
},
want: Release{
TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD",
Title: "Death Hunt",
Year: 1981,
Source: "BluRay",
Resolution: "1080p",
Other: []string{"REMUX"},
Codec: []string{"AVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "2.0",
Group: "playBD",
Type: "movie",
TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD",
Title: "Death Hunt",
NormalizedHash: "e6b1b708b8ea39470de6d327da2920d6",
Year: 1981,
Source: "BluRay",
Resolution: "1080p",
Other: []string{"REMUX"},
Codec: []string{"AVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "2.0",
Group: "playBD",
Type: rls.Movie,
MediaProcessing: "REMUX",
},
},
{
@ -383,18 +413,20 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS",
},
want: Release{
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS",
Title: "Old Movie",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "VHS",
Type: "movie",
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS",
Title: "Old Movie",
NormalizedHash: "1ea814c8c21979e7692efb1b24298e95",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "VHS",
Type: rls.Movie,
MediaProcessing: "REMUX",
},
},
{
@ -403,18 +435,20 @@ func TestRelease_Parse(t *testing.T) {
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS",
},
want: Release{
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS",
Title: "Old Movie",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "VHS",
Type: "movie",
TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS",
Title: "Old Movie",
NormalizedHash: "1ea814c8c21979e7692efb1b24298e95",
Year: 1954,
Source: "UHD.BluRay",
Resolution: "2160p",
Other: []string{"REMUX"},
HDR: []string{"DV", "HDR10"},
Codec: []string{"HEVC"},
Audio: []string{"DTS-HD.MA"},
AudioChannels: "5.1",
Group: "VHS",
Type: rls.Movie,
MediaProcessing: "REMUX",
},
},
}
@ -863,3 +897,220 @@ func Test_getUniqueTags(t *testing.T) {
})
}
}
func TestRelease_Hash(t *testing.T) {
type fields struct {
TorrentName string
}
tests := []struct {
name string
fields fields
want string
}{
{
name: "1",
fields: fields{
TorrentName: "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP",
},
want: "833e6c4c78e960ccbce4ef68f6564082",
},
{
name: "2",
fields: fields{
TorrentName: "That Movie 2023 2160p BluRay DTS-HD 5.1 x265-GROUP",
},
want: "833e6c4c78e960ccbce4ef68f6564082",
},
{
name: "3",
fields: fields{
TorrentName: "That Movie 2023 2160p BluRay DTS-HD 5 1 x265-GROUP",
},
want: "833e6c4c78e960ccbce4ef68f6564082",
},
{
name: "4",
fields: fields{
TorrentName: "That Movie 2023 1080p Remux DTS-HD 5 1 AVC-GROUP",
},
want: "ee15fe3c926a5b40370aec970f6b4c43",
},
{
name: "5",
fields: fields{
TorrentName: "That Movie 2023 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP",
},
want: "ee15fe3c926a5b40370aec970f6b4c43",
},
{
name: "6",
fields: fields{
TorrentName: "That Movie 2023 Theatrical Cut Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP",
},
want: "db7e55fb3471e76ba27bf8a2e1e59236",
},
{
name: "7",
fields: fields{
TorrentName: "That Movie 2023 Directors Cut Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP",
},
want: "bccfe6762b34cbc3a54784351b5424fd",
},
{
name: "8",
fields: fields{
TorrentName: "That Show 2023 S01 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP",
},
want: "d1c74c33e1c36f3545d06aa8857610b1",
},
{
name: "9",
fields: fields{
TorrentName: "That Show 2023 S01 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2",
},
want: "29e228f705f070041e72ebaa7b29239e",
},
{
name: "10",
fields: fields{
TorrentName: "That Show 2023 S1 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2",
},
want: "29e228f705f070041e72ebaa7b29239e",
},
{
name: "11",
fields: fields{
TorrentName: "That Show 2023 S2 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2",
},
want: "cbc67f4eec93428361eddf040654e077",
},
{
name: "12",
fields: fields{
TorrentName: "That Show 2023 S01E01 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "806240c76cc32413efd5bcf033abf512",
},
{
name: "13",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "33c866b898a59d255a09c10bd23a5e3f",
},
{
name: "14",
fields: fields{
TorrentName: "That Show 2023 REPACK S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "7be5ebeee948820f53f24091054fc7cd",
},
{
name: "15",
fields: fields{
TorrentName: "That Show 2023 REPACK2 S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "84e3e70f6c983f96f10cba50213a338d",
},
{
name: "16",
fields: fields{
TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "52d5d1a28a340bc5a836c0244c206c3b",
},
{
name: "17",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DV DTS-HD 5.1 HEVC-GROUP",
},
want: "5e196039d426649ab5221c3e87ed108d",
},
{
name: "18",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DoVi DTS-HD 5.1 HEVC-GROUP",
},
want: "5e196039d426649ab5221c3e87ed108d",
},
{
name: "19",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DoVi HDR10 DTS-HD 5.1 HEVC-GROUP",
},
want: "f52a9a2f26e94c90bdaa8eaca2136cc3",
},
{
name: "20",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP",
},
want: "f52a9a2f26e94c90bdaa8eaca2136cc3",
},
{
name: "21",
fields: fields{
TorrentName: "that show 2023 S01E01 this is the start 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP",
},
want: "f52a9a2f26e94c90bdaa8eaca2136cc3",
},
{
name: "22",
fields: fields{
TorrentName: "That Show 2023 S01E01 This is the start 2160p HYBRID WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP",
},
want: "414f792a345148ed4c8e070ef77340ff",
},
{
name: "23",
fields: fields{
TorrentName: "That Movie 2008 1080p GBR Blu-ray AVC TrueHD 5.1",
},
want: "70bb23ff5da4209419b6d83628310a43",
},
{
name: "24",
fields: fields{
TorrentName: "That Movie 2008 1080p GER Blu-ray AVC TrueHD 5.1",
},
want: "369ba9f92fb7819e0d07791fddf3c85f",
},
{
name: "25",
fields: fields{
TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p AMZN WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "28e215fee790fbad3957383162f7c5a6",
},
{
name: "26",
fields: fields{
TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p DSNP WEB-DL DTS-HD 5.1 HEVC-GROUP",
},
want: "6a26cd652170d0efc4b0e0320fbd38c1",
},
{
name: "27",
fields: fields{
TorrentName: "That Other Show 2023 S01E01 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP",
},
want: "544c2fc4024ece5664c0f623a5feaeab",
},
{
name: "28",
fields: fields{
TorrentName: "That Other Show 2023 S01E01 2160p WEB-DL DV HDR10+ DTS-HD 5.1 HEVC-GROUP",
},
want: "63b5d87abe5fb49131785de426708d31",
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
r := &Release{
TorrentName: tt.fields.TorrentName,
}
r.ParseString(tt.fields.TorrentName)
assert.Equalf(t, tt.want, r.NormalizedHash, "Hash()")
})
}
}

View file

@ -425,9 +425,14 @@ func ParseReleaseTags(tags []string) ReleaseTags {
return releaseTags
}
func ParseReleaseTagString(tags string) ReleaseTags {
releaseTags := ReleaseTags{}
if tags == "" {
return releaseTags
}
for tagType, tagInfos := range types {
for _, info := range tagInfos {

View file

@ -12,6 +12,7 @@ import (
"github.com/dustin/go-humanize"
"github.com/mmcdole/gofeed"
"github.com/moistari/rls"
"github.com/rs/zerolog"
"github.com/stretchr/testify/assert"
)
@ -73,7 +74,67 @@ func TestRSSJob_processItem(t *testing.T) {
Link: "/details.php?id=00000&hit=1",
GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
}},
want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)},
want: &domain.Release{
ID: 0,
FilterStatus: "PENDING",
Rejections: []string{},
Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"},
FilterName: "",
Protocol: "torrent",
Implementation: "RSS",
AnnounceType: domain.AnnounceTypeNew,
Timestamp: now,
GroupID: "",
TorrentID: "",
DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1",
TorrentTmpFile: "",
TorrentDataRawBytes: []uint8(nil),
TorrentHash: "",
TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
NormalizedHash: "edfbe552ccde335f34b801e15930bc35",
Size: 1490000000,
Title: "Some Release Title",
Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n",
Category: "",
Season: 0,
Episode: 0,
Year: 2022,
Month: 9,
Day: 22,
Resolution: "720p",
Source: "WEB",
Codec: []string{"H.264"},
Container: "",
HDR: []string(nil),
Audio: []string(nil),
AudioChannels: "",
Group: "GROUP",
Region: "",
Language: []string{},
Proper: false,
Repack: false,
Edition: []string{},
Cut: []string{},
Website: "",
Artists: "",
Type: rls.Episode,
LogScore: 0,
Origin: "",
Tags: []string{},
ReleaseTags: "",
Freeleech: false,
FreeleechPercent: 0,
Bonus: []string(nil),
Uploader: "",
PreTime: "",
Other: []string{},
RawCookie: "",
AdditionalSizeCheckRequired: false,
AdditionalUploaderCheckRequired: false,
FilterID: 0,
Filter: (*domain.Filter)(nil),
ActionStatus: []domain.ReleaseActionStatus(nil),
},
},
{
name: "with_baseurl",
@ -107,7 +168,67 @@ func TestRSSJob_processItem(t *testing.T) {
Link: "https://fake-feed.com/details.php?id=00000&hit=1",
GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
}},
want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)},
want: &domain.Release{
ID: 0,
FilterStatus: "PENDING",
Rejections: []string{},
Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"},
FilterName: "",
Protocol: "torrent",
Implementation: "RSS",
AnnounceType: domain.AnnounceTypeNew,
Timestamp: now,
GroupID: "",
TorrentID: "",
DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1",
TorrentTmpFile: "",
TorrentDataRawBytes: []uint8(nil),
TorrentHash: "",
TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
NormalizedHash: "edfbe552ccde335f34b801e15930bc35",
Size: 1490000000,
Title: "Some Release Title",
Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n",
Category: "",
Season: 0,
Episode: 0,
Year: 2022,
Month: 9,
Day: 22,
Resolution: "720p",
Source: "WEB",
Codec: []string{"H.264"},
Container: "",
HDR: []string(nil),
Audio: []string(nil),
AudioChannels: "",
Group: "GROUP",
Region: "",
Language: []string{},
Proper: false,
Repack: false,
Edition: []string{},
Cut: []string{},
Website: "",
Artists: "",
Type: rls.Episode,
LogScore: 0,
Origin: "",
Tags: []string{},
ReleaseTags: "",
Freeleech: false,
FreeleechPercent: 0,
Bonus: []string(nil),
Uploader: "",
PreTime: "",
Other: []string{},
RawCookie: "",
AdditionalSizeCheckRequired: false,
AdditionalUploaderCheckRequired: false,
FilterID: 0,
Filter: (*domain.Filter)(nil),
ActionStatus: []domain.ReleaseActionStatus(nil),
},
},
{
name: "time_parse",
@ -142,7 +263,67 @@ func TestRSSJob_processItem(t *testing.T) {
GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
//PublishedParsed: &nowMinusTime,
}},
want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)},
want: &domain.Release{
ID: 0,
FilterStatus: "PENDING",
Rejections: []string{},
Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"},
FilterName: "",
Protocol: "torrent",
Implementation: "RSS",
AnnounceType: domain.AnnounceTypeNew,
Timestamp: now,
GroupID: "",
TorrentID: "",
DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1",
TorrentTmpFile: "",
TorrentDataRawBytes: []uint8(nil),
TorrentHash: "",
TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
NormalizedHash: "edfbe552ccde335f34b801e15930bc35",
Size: 1490000000,
Title: "Some Release Title",
Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n",
Category: "",
Season: 0,
Episode: 0,
Year: 2022,
Month: 9,
Day: 22,
Resolution: "720p",
Source: "WEB",
Codec: []string{"H.264"},
Container: "",
HDR: []string(nil),
Audio: []string(nil),
AudioChannels: "",
Group: "GROUP",
Region: "",
Language: []string{},
Proper: false,
Repack: false,
Edition: []string{},
Cut: []string{},
Website: "",
Artists: "",
Type: rls.Episode,
LogScore: 0,
Origin: "",
Tags: []string{},
ReleaseTags: "",
Freeleech: false,
FreeleechPercent: 0,
Bonus: []string(nil),
Uploader: "",
PreTime: "",
Other: []string{},
RawCookie: "",
AdditionalSizeCheckRequired: false,
AdditionalUploaderCheckRequired: false,
FilterID: 0,
Filter: (*domain.Filter)(nil),
ActionStatus: []domain.ReleaseActionStatus(nil),
},
},
{
name: "time_parse",
@ -208,7 +389,68 @@ func TestRSSJob_processItem(t *testing.T) {
},
},
}},
want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, MagnetURI: "magnet:?xt=this-not-a-valid-magnet", GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 0, Title: "Some Release Title", Description: "Category: Example", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)},
want: &domain.Release{
ID: 0,
FilterStatus: "PENDING",
Rejections: []string{},
Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"},
FilterName: "",
Protocol: "torrent",
Implementation: "RSS",
AnnounceType: domain.AnnounceTypeNew,
Timestamp: now,
GroupID: "",
TorrentID: "",
DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1",
MagnetURI: "magnet:?xt=this-not-a-valid-magnet",
TorrentTmpFile: "",
TorrentDataRawBytes: []uint8(nil),
TorrentHash: "",
TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP",
NormalizedHash: "edfbe552ccde335f34b801e15930bc35",
Size: 0,
Title: "Some Release Title",
Description: "Category: Example",
Category: "",
Season: 0,
Episode: 0,
Year: 2022,
Month: 9,
Day: 22,
Resolution: "720p",
Source: "WEB",
Codec: []string{"H.264"},
Container: "",
HDR: []string(nil),
Audio: []string(nil),
AudioChannels: "",
Group: "GROUP",
Region: "",
Language: []string{},
Proper: false,
Repack: false,
Edition: []string{},
Cut: []string{},
Website: "",
Artists: "",
Type: rls.Episode,
LogScore: 0,
Origin: "",
Tags: []string{},
ReleaseTags: "",
Freeleech: false,
FreeleechPercent: 0,
Bonus: []string(nil),
Uploader: "",
PreTime: "",
Other: []string{},
RawCookie: "",
AdditionalSizeCheckRequired: false,
AdditionalUploaderCheckRequired: false,
FilterID: 0,
Filter: (*domain.Filter)(nil),
ActionStatus: []domain.ReleaseActionStatus(nil),
},
},
}
for _, tt := range tests {

View file

@ -46,6 +46,7 @@ type Service interface {
AdditionalRecordLabelCheck(ctx context.Context, f *domain.Filter, release *domain.Release) (bool, error)
CheckSmartEpisodeCanDownload(ctx context.Context, params *domain.SmartEpisodeParams) (bool, error)
GetDownloadsByFilterId(ctx context.Context, filterID int) (*domain.FilterDownloads, error)
CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error)
}
type service struct {
@ -374,6 +375,8 @@ func (s *service) Delete(ctx context.Context, filterID int) error {
func (s *service) CheckFilter(ctx context.Context, f *domain.Filter, release *domain.Release) (bool, error) {
l := s.log.With().Str("method", "CheckFilter").Logger()
l.Debug().Msgf("checking filter: %s with release %s", f.Name, release.TorrentName)
l.Trace().Msgf("checking filter: %s %+v", f.Name, f)
l.Trace().Msgf("checking filter: %s for release: %+v", f.Name, release)
@ -393,110 +396,130 @@ func (s *service) CheckFilter(ctx context.Context, f *domain.Filter, release *do
return false, nil
}
if matchedFilter {
// smartEpisode check
if f.SmartEpisode {
params := &domain.SmartEpisodeParams{
Title: release.Title,
Season: release.Season,
Episode: release.Episode,
Year: release.Year,
Month: release.Month,
Day: release.Day,
Repack: release.Repack,
Proper: release.Proper,
Group: release.Group,
}
canDownloadShow, err := s.CheckSmartEpisodeCanDownload(ctx, params)
if err != nil {
l.Trace().Msgf("failed smart episode check: %s", f.Name)
return false, nil
}
if !canDownloadShow {
l.Trace().Msgf("failed smart episode check: %s", f.Name)
if params.IsDailyEpisode() {
f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day), fmt.Sprintf("expected newer than (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day))
} else {
f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) season: %d ep: %d", release.Title, release.Season, release.Episode), fmt.Sprintf("expected newer than (%s) season: %d ep: %d", release.Title, release.Season, release.Episode))
}
return false, nil
}
}
// if matched, do additional size check if needed, attach actions and return the filter
l.Debug().Msgf("found and matched filter: %s", f.Name)
// If size constraints are set in a filter and the indexer did not
// announce the size, we need to do an additional out of band size check.
if release.AdditionalSizeCheckRequired {
l.Debug().Msgf("(%s) additional size check required", f.Name)
ok, err := s.AdditionalSizeCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional size check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional size check not matching what filter wanted", f.Name)
return false, nil
}
}
// check uploader if the indexer supports check via api
if release.AdditionalUploaderCheckRequired {
l.Debug().Msgf("(%s) additional uploader check required", f.Name)
ok, err := s.AdditionalUploaderCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional uploader check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional uploader check not matching what filter wanted", f.Name)
return false, nil
}
}
if release.AdditionalRecordLabelCheckRequired {
l.Debug().Msgf("(%s) additional record label check required", f.Name)
ok, err := s.AdditionalRecordLabelCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional record label check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional record label check not matching what filter wanted", f.Name)
return false, nil
}
}
// run external filters
if f.External != nil {
externalOk, err := s.RunExternalFilters(ctx, f, f.External, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) external filter check error", f.Name)
return false, err
}
if !externalOk {
l.Debug().Msgf("(%s) external filter check not matching what filter wanted", f.Name)
return false, nil
}
}
return true, nil
if !matchedFilter {
// if no match, return nil
return false, nil
}
// if no match, return nil
return false, nil
// smartEpisode check
if f.SmartEpisode {
params := &domain.SmartEpisodeParams{
Title: release.Title,
Season: release.Season,
Episode: release.Episode,
Year: release.Year,
Month: release.Month,
Day: release.Day,
Repack: release.Repack,
Proper: release.Proper,
Group: release.Group,
}
canDownloadShow, err := s.CheckSmartEpisodeCanDownload(ctx, params)
if err != nil {
l.Trace().Msgf("failed smart episode check: %s", f.Name)
return false, nil
}
if !canDownloadShow {
l.Trace().Msgf("failed smart episode check: %s", f.Name)
if params.IsDailyEpisode() {
f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day), fmt.Sprintf("expected newer than (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day))
} else {
f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) season: %d ep: %d", release.Title, release.Season, release.Episode), fmt.Sprintf("expected newer than (%s) season: %d ep: %d", release.Title, release.Season, release.Episode))
}
return false, nil
}
}
// check duplicates
if f.DuplicateHandling != nil {
l.Debug().Msgf("(%s) check is duplicate with profile %s", f.Name, f.DuplicateHandling.Name)
release.SkipDuplicateProfileID = f.DuplicateHandling.ID
release.SkipDuplicateProfileName = f.DuplicateHandling.Name
isDuplicate, err := s.CheckIsDuplicateRelease(ctx, f.DuplicateHandling, release)
if err != nil {
return false, errors.Wrap(err, "error finding duplicate handle")
}
if isDuplicate {
l.Debug().Msgf("filter %s rejected release %q as duplicate with profile %q", f.Name, release.TorrentName, f.DuplicateHandling.Name)
f.RejectReasons.Add("duplicate", "duplicate", "not duplicate")
// let it continue so external filters can trigger checks
//return false, nil
release.IsDuplicate = true
}
}
// if matched, do additional size check if needed, attach actions and return the filter
l.Debug().Msgf("found and matched filter: %s", f.Name)
// If size constraints are set in a filter and the indexer did not
// announce the size, we need to do an additional out of band size check.
if release.AdditionalSizeCheckRequired {
l.Debug().Msgf("(%s) additional size check required", f.Name)
ok, err := s.AdditionalSizeCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional size check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional size check not matching what filter wanted", f.Name)
return false, nil
}
}
// check uploader if the indexer supports check via api
if release.AdditionalUploaderCheckRequired {
l.Debug().Msgf("(%s) additional uploader check required", f.Name)
ok, err := s.AdditionalUploaderCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional uploader check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional uploader check not matching what filter wanted", f.Name)
return false, nil
}
}
if release.AdditionalRecordLabelCheckRequired {
l.Debug().Msgf("(%s) additional record label check required", f.Name)
ok, err := s.AdditionalRecordLabelCheck(ctx, f, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) additional record label check error", f.Name)
return false, err
}
if !ok {
l.Trace().Msgf("(%s) additional record label check not matching what filter wanted", f.Name)
return false, nil
}
}
// run external filters
if f.External != nil {
externalOk, err := s.RunExternalFilters(ctx, f, f.External, release)
if err != nil {
l.Error().Err(err).Msgf("(%s) external filter check error", f.Name)
return false, err
}
if !externalOk {
l.Debug().Msgf("(%s) external filter check not matching what filter wanted", f.Name)
return false, nil
}
}
return true, nil
}
// AdditionalSizeCheck performs additional out-of-band checks to determine the
@ -732,10 +755,18 @@ func (s *service) CheckSmartEpisodeCanDownload(ctx context.Context, params *doma
return s.releaseRepo.CheckSmartEpisodeCanDownload(ctx, params)
}
func (s *service) CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error) {
return s.releaseRepo.CheckIsDuplicateRelease(ctx, profile, release)
}
func (s *service) RunExternalFilters(ctx context.Context, f *domain.Filter, externalFilters []domain.FilterExternal, release *domain.Release) (ok bool, err error) {
defer func() {
// try recover panic if anything went wrong with the external filter checks
errors.RecoverPanic(recover(), &err)
if err != nil {
s.log.Error().Err(err).Msgf("filter %s external filter check panic", f.Name)
ok = false
}
}()
// sort filters by index

View file

@ -17,7 +17,7 @@ import (
type actionService interface {
List(ctx context.Context) ([]domain.Action, error)
Store(ctx context.Context, action domain.Action) (*domain.Action, error)
Store(ctx context.Context, action *domain.Action) error
Delete(ctx context.Context, req *domain.DeleteActionRequest) error
ToggleEnabled(actionID int) error
}
@ -56,35 +56,35 @@ func (h actionHandler) getActions(w http.ResponseWriter, r *http.Request) {
}
func (h actionHandler) storeAction(w http.ResponseWriter, r *http.Request) {
var data domain.Action
var data *domain.Action
if err := json.NewDecoder(r.Body).Decode(&data); err != nil {
h.encoder.Error(w, err)
return
}
action, err := h.service.Store(r.Context(), data)
err := h.service.Store(r.Context(), data)
if err != nil {
h.encoder.Error(w, err)
return
}
h.encoder.StatusResponse(w, http.StatusCreated, action)
h.encoder.StatusResponse(w, http.StatusCreated, data)
}
func (h actionHandler) updateAction(w http.ResponseWriter, r *http.Request) {
var data domain.Action
var data *domain.Action
if err := json.NewDecoder(r.Body).Decode(&data); err != nil {
h.encoder.Error(w, err)
return
}
action, err := h.service.Store(r.Context(), data)
err := h.service.Store(r.Context(), data)
if err != nil {
h.encoder.Error(w, err)
return
}
h.encoder.StatusResponse(w, http.StatusCreated, action)
h.encoder.StatusResponse(w, http.StatusCreated, data)
}
func (h actionHandler) deleteAction(w http.ResponseWriter, r *http.Request) {

View file

@ -25,6 +25,10 @@ type releaseService interface {
Delete(ctx context.Context, req *domain.DeleteReleaseRequest) error
Retry(ctx context.Context, req *domain.ReleaseActionRetryReq) error
ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error
StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error
FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error)
DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error
}
type releaseHandler struct {
@ -52,6 +56,13 @@ func (h releaseHandler) Routes(r chi.Router) {
r.Get("/", h.getReleaseByID)
r.Post("/actions/{actionStatusID}/retry", h.retryAction)
})
r.Route("/profiles/duplicate", func(r chi.Router) {
r.Get("/", h.findReleaseProfileDuplicate)
r.Post("/", h.storeReleaseProfileDuplicate)
r.Delete("/{profileId}", h.deleteReleaseProfileDuplicate)
})
}
func (h releaseHandler) findReleases(w http.ResponseWriter, r *http.Request) {
@ -309,3 +320,55 @@ func (h releaseHandler) retryAction(w http.ResponseWriter, r *http.Request) {
h.encoder.NoContent(w)
}
func (h releaseHandler) storeReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) {
var data *domain.DuplicateReleaseProfile
if err := json.NewDecoder(r.Body).Decode(&data); err != nil {
h.encoder.Error(w, err)
return
}
if err := h.service.StoreReleaseProfileDuplicate(r.Context(), data); err != nil {
h.encoder.Error(w, err)
return
}
h.encoder.StatusCreatedData(w, data)
}
func (h releaseHandler) findReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) {
profiles, err := h.service.FindDuplicateReleaseProfiles(r.Context())
if err != nil {
h.encoder.StatusResponse(w, http.StatusInternalServerError, map[string]interface{}{
"code": "INTERNAL_SERVER_ERROR",
"message": err.Error(),
})
return
}
//ret := struct {
// Data []*domain.DuplicateReleaseProfile `json:"data"`
//}{
// Data: profiles,
//}
h.encoder.StatusResponse(w, http.StatusOK, profiles)
}
func (h releaseHandler) deleteReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) {
//profileIdParam := chi.URLParam(r, "releaseId")
profileId, err := strconv.Atoi(chi.URLParam(r, "profileId"))
if err != nil {
h.encoder.StatusError(w, http.StatusBadRequest, err)
return
}
if err := h.service.DeleteReleaseProfileDuplicate(r.Context(), int64(profileId)); err != nil {
h.encoder.Error(w, err)
return
}
h.encoder.NoContent(w)
}

View file

@ -32,6 +32,10 @@ type Service interface {
ProcessMultiple(releases []*domain.Release)
ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error
Retry(ctx context.Context, req *domain.ReleaseActionRetryReq) error
StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error
FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error)
DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error
}
type actionClientTypeKey struct {
@ -94,6 +98,18 @@ func (s *service) Delete(ctx context.Context, req *domain.DeleteReleaseRequest)
return s.repo.Delete(ctx, req)
}
func (s *service) FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) {
return s.repo.FindDuplicateReleaseProfiles(ctx)
}
func (s *service) StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error {
return s.repo.StoreDuplicateProfile(ctx, profile)
}
func (s *service) DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error {
return s.repo.DeleteReleaseProfileDuplicate(ctx, id)
}
func (s *service) ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error {
// get indexer definition with data
def, err := s.indexerSvc.GetMappedDefinitionByName(req.IndexerIdentifier)
@ -183,8 +199,6 @@ func (s *service) Process(release *domain.Release) {
s.log.Error().Err(err).Msgf("release.Process: error processing filters for indexer: %s", release.Indexer.Name)
return
}
return
}
func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, release *domain.Release) error {
@ -201,6 +215,11 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter,
release.FilterName = f.Name
release.FilterID = f.ID
// reset IsDuplicate
release.IsDuplicate = false
release.SkipDuplicateProfileID = 0
release.SkipDuplicateProfileName = ""
// test filter
match, err := s.filterSvc.CheckFilter(ctx, f, release)
if err != nil {
@ -208,10 +227,10 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter,
return err
}
if !match {
if !match || f.RejectReasons.Len() > 0 {
l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s, no match. rejections: %s", release.Indexer.Name, release.FilterName, release.TorrentName, f.RejectReasons.String())
l.Debug().Msgf("filter %s rejected release: %s", f.Name, release.TorrentName)
l.Debug().Msgf("filter %s rejected release: %s with reasons: %s", f.Name, release.TorrentName, f.RejectReasons.String())
continue
}
@ -231,13 +250,6 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter,
continue
}
// sleep for the delay period specified in the filter before running actions
delay := release.Filter.Delay
if delay > 0 {
l.Debug().Msgf("release.Process: delaying processing of '%s' (%s) for %s by %d seconds as specified in the filter", release.TorrentName, release.FilterName, release.Indexer.Name, delay)
time.Sleep(time.Duration(delay) * time.Second)
}
// save release here to only save those with rejections from actions instead of all releases
if release.ID == 0 {
release.FilterStatus = domain.ReleaseStatusFilterApproved
@ -251,24 +263,40 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter,
var rejections []string
// run actions (watchFolder, test, exec, qBittorrent, Deluge, arr etc.)
for _, act := range actions {
for idx, act := range actions {
// only run enabled actions
if !act.Enabled {
l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s action '%s' not enabled, skip", release.Indexer.Name, release.FilterName, release.TorrentName, act.Name)
continue
}
// add action status as pending
actionStatus := domain.NewReleaseActionStatus(act, release)
if err := s.StoreReleaseActionStatus(ctx, actionStatus); err != nil {
s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName)
}
if idx == 0 {
// sleep for the delay period specified in the filter before running actions
delay := release.Filter.Delay
if delay > 0 {
l.Debug().Msgf("release.Process: delaying processing of '%s' (%s) for %s by %d seconds as specified in the filter", release.TorrentName, release.FilterName, release.Indexer.Name, delay)
time.Sleep(time.Duration(delay) * time.Second)
}
}
l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s , run action: %s", release.Indexer.Name, release.FilterName, release.TorrentName, act.Name)
// keep track of action clients to avoid sending the same thing all over again
_, tried := triedActionClients[actionClientTypeKey{Type: act.Type, ClientID: act.ClientID}]
if tried {
l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s action client already tried, skip", release.Indexer.Name, release.FilterName, release.TorrentName)
l.Debug().Msgf("release.Process: indexer: %s, filter: %s release: %s action client already tried, skip", release.Indexer.Name, release.FilterName, release.TorrentName)
continue
}
// run action
status, err := s.runAction(ctx, act, release)
status, err := s.runAction(ctx, act, release, actionStatus)
if err != nil {
l.Error().Err(err).Msgf("release.Process: error running actions for filter: %s", release.FilterName)
//continue
@ -320,13 +348,13 @@ func (s *service) ProcessMultiple(releases []*domain.Release) {
}
}
func (s *service) runAction(ctx context.Context, action *domain.Action, release *domain.Release) (*domain.ReleaseActionStatus, error) {
func (s *service) runAction(ctx context.Context, action *domain.Action, release *domain.Release, status *domain.ReleaseActionStatus) (*domain.ReleaseActionStatus, error) {
// add action status as pending
status := domain.NewReleaseActionStatus(action, release)
if err := s.StoreReleaseActionStatus(ctx, status); err != nil {
s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName)
}
//status := domain.NewReleaseActionStatus(action, release)
//
//if err := s.StoreReleaseActionStatus(ctx, status); err != nil {
// s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName)
//}
rejections, err := s.actionSvc.RunAction(ctx, action, release)
if err != nil {
@ -351,7 +379,14 @@ func (s *service) runAction(ctx context.Context, action *domain.Action, release
}
func (s *service) retryAction(ctx context.Context, action *domain.Action, release *domain.Release) error {
actionStatus, err := s.runAction(ctx, action, release)
// add action status as pending
status := domain.NewReleaseActionStatus(action, release)
if err := s.StoreReleaseActionStatus(ctx, status); err != nil {
s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName)
}
actionStatus, err := s.runAction(ctx, action, release, status)
if err != nil {
s.log.Error().Err(err).Msgf("release.retryAction: error running actions for filter: %s", release.FilterName)

View file

@ -498,7 +498,16 @@ export const APIClient = {
},
replayAction: (releaseId: number, actionId: number) => appClient.Post(
`api/release/${releaseId}/actions/${actionId}/retry`
)
),
profiles: {
duplicates: {
list: () => appClient.Get<ReleaseProfileDuplicate[]>(`api/release/profiles/duplicate`),
delete: (id: number) => appClient.Delete(`api/release/profiles/duplicate/${id}`),
store: (profile: ReleaseProfileDuplicate) => appClient.Post(`api/release/profiles/duplicate`, {
body: profile
}),
}
}
},
updates: {
check: () => appClient.Get("api/updates/check"),

View file

@ -12,7 +12,7 @@ import {
FilterKeys,
IndexerKeys,
IrcKeys, ListKeys, NotificationKeys, ProxyKeys,
ReleaseKeys,
ReleaseKeys, ReleaseProfileDuplicateKeys,
SettingsKeys
} from "@api/query_keys";
import { ColumnFilter } from "@tanstack/react-table";
@ -165,6 +165,14 @@ export const ReleasesIndexersQueryOptions = () =>
staleTime: Infinity
});
export const ReleaseProfileDuplicateList = () =>
queryOptions({
queryKey: ReleaseProfileDuplicateKeys.lists(),
queryFn: () => APIClient.release.profiles.duplicates.list(),
staleTime: 5000,
refetchOnWindowFocus: true,
});
export const ProxiesQueryOptions = () =>
queryOptions({
queryKey: ProxyKeys.lists(),

View file

@ -35,6 +35,13 @@ export const ReleaseKeys = {
latestActivity: () => [...ReleaseKeys.all, "latest-activity"] as const,
};
export const ReleaseProfileDuplicateKeys = {
all: ["releaseProfileDuplicate"] as const,
lists: () => [...ReleaseProfileDuplicateKeys.all, "list"] as const,
details: () => [...ReleaseProfileDuplicateKeys.all, "detail"] as const,
detail: (id: number) => [...ReleaseProfileDuplicateKeys.details(), id] as const,
};
export const ApiKeys = {
all: ["api_keys"] as const,
lists: () => [...ApiKeys.all, "list"] as const,

View file

@ -251,7 +251,7 @@ export function DownloadClientSelect({
export interface SelectFieldOption {
label: string;
value: string;
value: string | number | null;
}
export interface SelectFieldProps {
@ -293,7 +293,7 @@ export const Select = ({
onChange={(value) => setFieldValue(field.name, value)}
>
{({ open }) => (
<>
<div>
<Label className="flex text-xs font-bold text-gray-800 dark:text-gray-100 uppercase tracking-wide">
{tooltip ? (
<DocsTooltip label={label}>{tooltip}</DocsTooltip>
@ -364,7 +364,7 @@ export const Select = ({
</ListboxOptions>
</Transition>
</div>
</>
</div>
)}
</Listbox>
)}

15
web/src/forms/_shared.ts Normal file
View file

@ -0,0 +1,15 @@
/*
* Copyright (c) 2021 - 2024, Ludvig Lundgren and the autobrr contributors.
* SPDX-License-Identifier: GPL-2.0-or-later
*/
export interface AddFormProps {
isOpen: boolean;
toggle: () => void;
}
export interface UpdateFormProps<T> {
isOpen: boolean;
toggle: () => void;
data: T;
}

View file

@ -16,14 +16,9 @@ import { FilterKeys } from "@api/query_keys";
import { DEBUG } from "@components/debug";
import { toast } from "@components/hot-toast";
import Toast from "@components/notifications/Toast";
import { AddFormProps } from "@forms/_shared";
interface filterAddFormProps {
isOpen: boolean;
toggle: () => void;
}
export function FilterAddForm({ isOpen, toggle }: filterAddFormProps) {
export function FilterAddForm({ isOpen, toggle }: AddFormProps) {
const inputRef = useRef(null)
const queryClient = useQueryClient();
const navigate = useNavigate();

View file

@ -15,13 +15,9 @@ import { ApiKeys } from "@api/query_keys";
import { DEBUG } from "@components/debug";
import { toast } from "@components/hot-toast";
import Toast from "@components/notifications/Toast";
import { AddFormProps } from "@forms/_shared";
interface apiKeyAddFormProps {
isOpen: boolean;
toggle: () => void;
}
export function APIKeyAddForm({ isOpen, toggle }: apiKeyAddFormProps) {
export function APIKeyAddForm({ isOpen, toggle }: AddFormProps) {
const queryClient = useQueryClient();
const mutation = useMutation({

View file

@ -27,6 +27,7 @@ import {
} from "@components/inputs";
import { DocsLink, ExternalLink } from "@components/ExternalLink";
import { SelectFieldBasic } from "@components/inputs/select_wide";
import { AddFormProps, UpdateFormProps } from "@forms/_shared";
interface InitialValuesSettings {
basic?: {
@ -691,12 +692,7 @@ function DownloadClientFormButtons({
);
}
interface formProps {
isOpen: boolean;
toggle: () => void;
}
export function DownloadClientAddForm({ isOpen, toggle }: formProps) {
export function DownloadClientAddForm({ isOpen, toggle }: AddFormProps) {
const [isTesting, setIsTesting] = useState(false);
const [isSuccessfulTest, setIsSuccessfulTest] = useState(false);
const [isErrorTest, setIsErrorTest] = useState(false);
@ -856,13 +852,7 @@ export function DownloadClientAddForm({ isOpen, toggle }: formProps) {
);
}
interface updateFormProps {
isOpen: boolean;
toggle: () => void;
client: DownloadClient;
}
export function DownloadClientUpdateForm({ client, isOpen, toggle }: updateFormProps) {
export function DownloadClientUpdateForm({ isOpen, toggle, data: client}: UpdateFormProps<DownloadClient>) {
const [isTesting, setIsTesting] = useState(false);
const [isSuccessfulTest, setIsSuccessfulTest] = useState(false);
const [isErrorTest, setIsErrorTest] = useState(false);

View file

@ -18,13 +18,7 @@ import { componentMapType } from "./DownloadClientForms";
import { sleep } from "@utils";
import { ImplementationBadges } from "@screens/settings/Indexer";
import { FeedDownloadTypeOptions } from "@domain/constants";
interface UpdateProps {
isOpen: boolean;
toggle: () => void;
feed: Feed;
}
import { UpdateFormProps } from "@forms/_shared";
interface InitialValues {
id: number;
@ -41,7 +35,8 @@ interface InitialValues {
settings: FeedSettings;
}
export function FeedUpdateForm({ isOpen, toggle, feed }: UpdateProps) {
export function FeedUpdateForm({ isOpen, toggle, data}: UpdateFormProps<Feed>) {
const feed = data;
const [isTesting, setIsTesting] = useState(false);
const [isTestSuccessful, setIsSuccessfulTest] = useState(false);
const [isTestError, setIsErrorTest] = useState(false);

View file

@ -25,6 +25,7 @@ import { FeedDownloadTypeOptions } from "@domain/constants";
import { DocsLink } from "@components/ExternalLink";
import * as common from "@components/inputs/common";
import { SelectField } from "@forms/settings/IrcForms";
import { AddFormProps, UpdateFormProps } from "@forms/_shared";
// const isRequired = (message: string) => (value?: string | undefined) => (!!value ? undefined : message);
@ -255,12 +256,7 @@ type SelectValue = {
value: string;
};
export interface AddProps {
isOpen: boolean;
toggle: () => void;
}
export function IndexerAddForm({ isOpen, toggle }: AddProps) {
export function IndexerAddForm({ isOpen, toggle }: AddFormProps) {
const [indexer, setIndexer] = useState<IndexerDefinition>({} as IndexerDefinition);
const queryClient = useQueryClient();
@ -729,13 +725,7 @@ interface IndexerUpdateInitialValues {
}
}
interface UpdateProps {
isOpen: boolean;
toggle: () => void;
indexer: IndexerDefinition;
}
export function IndexerUpdateForm({ isOpen, toggle, indexer }: UpdateProps) {
export function IndexerUpdateForm({ isOpen, toggle, data: indexer }: UpdateFormProps<IndexerDefinition>) {
const queryClient = useQueryClient();
const proxies = useQuery(ProxiesQueryOptions());

View file

@ -22,6 +22,7 @@ import Toast from "@components/notifications/Toast";
import * as common from "@components/inputs/common";
import { classNames } from "@utils";
import { ProxiesQueryOptions } from "@api/queries";
import { AddFormProps, UpdateFormProps } from "@forms/_shared";
interface ChannelsFieldArrayProps {
channels: IrcChannel[];
@ -122,11 +123,6 @@ interface IrcNetworkAddFormValues {
channels: IrcChannel[];
}
interface AddFormProps {
isOpen: boolean;
toggle: () => void;
}
export function IrcNetworkAddForm({ isOpen, toggle }: AddFormProps) {
const queryClient = useQueryClient();
@ -275,17 +271,11 @@ interface IrcNetworkUpdateFormValues {
proxy_id: number;
}
interface IrcNetworkUpdateFormProps {
isOpen: boolean;
toggle: () => void;
network: IrcNetwork;
}
export function IrcNetworkUpdateForm({
isOpen,
toggle,
network
}: IrcNetworkUpdateFormProps) {
data: network
}: UpdateFormProps<IrcNetwork>) {
const queryClient = useQueryClient();
const proxies = useQuery(ProxiesQueryOptions());

View file

@ -23,6 +23,7 @@ import * as common from "@components/inputs/common";
import { NumberFieldWide, PasswordFieldWide, SwitchGroupWide, TextFieldWide } from "@components/inputs";
import { componentMapType } from "./DownloadClientForms";
import { AddFormProps, UpdateFormProps } from "@forms/_shared";
function FormFieldsDiscord() {
return (
@ -311,12 +312,7 @@ interface NotificationAddFormValues {
enabled: boolean;
}
interface AddProps {
isOpen: boolean;
toggle: () => void;
}
export function NotificationAddForm({ isOpen, toggle }: AddProps) {
export function NotificationAddForm({ isOpen, toggle }: AddFormProps) {
const queryClient = useQueryClient();
const createMutation = useMutation({
@ -565,12 +561,6 @@ const EventCheckBoxes = () => (
</fieldset>
);
interface UpdateProps {
isOpen: boolean;
toggle: () => void;
notification: ServiceNotification;
}
interface InitialValues {
id: number;
enabled: boolean;
@ -587,7 +577,7 @@ interface InitialValues {
username?: string
}
export function NotificationUpdateForm({ isOpen, toggle, notification }: UpdateProps) {
export function NotificationUpdateForm({ isOpen, toggle, data: notification }: UpdateFormProps<ServiceNotification>) {
const queryClient = useQueryClient();
const mutation = useMutation({

View file

@ -9,7 +9,7 @@ import { Dialog, DialogPanel, DialogTitle, Transition, TransitionChild } from "@
import { XMarkIcon } from "@heroicons/react/24/solid";
import { useMutation, useQueryClient } from "@tanstack/react-query";
import { AddProps } from "@forms/settings/IndexerForms";
import { AddFormProps } from "@forms/_shared";
import { DEBUG } from "@components/debug.tsx";
import { PasswordFieldWide, SwitchGroupWide, TextFieldWide } from "@components/inputs";
import { SelectFieldBasic } from "@components/inputs/select_wide";
@ -20,7 +20,7 @@ import { toast } from "@components/hot-toast";
import Toast from "@components/notifications/Toast";
import { SlideOver } from "@components/panels";
export function ProxyAddForm({ isOpen, toggle }: AddProps) {
export function ProxyAddForm({ isOpen, toggle }: AddFormProps) {
const queryClient = useQueryClient();
const createMutation = useMutation({

View file

@ -0,0 +1,198 @@
/*
* Copyright (c) 2021 - 2024, Ludvig Lundgren and the autobrr contributors.
* SPDX-License-Identifier: GPL-2.0-or-later
*/
import { useMutation, useQueryClient } from "@tanstack/react-query";
import { APIClient } from "@api/APIClient.ts";
import { ReleaseProfileDuplicateKeys } from "@api/query_keys.ts";
import { toast } from "@components/hot-toast";
import Toast from "@components/notifications/Toast.tsx";
import { SwitchGroupWide, TextFieldWide } from "@components/inputs";
import { SlideOver } from "@components/panels";
import { AddFormProps, UpdateFormProps } from "@forms/_shared";
export function ReleaseProfileDuplicateAddForm({ isOpen, toggle }: AddFormProps) {
const queryClient = useQueryClient();
const addMutation = useMutation({
mutationFn: (profile: ReleaseProfileDuplicate) => APIClient.release.profiles.duplicates.store(profile),
onSuccess: () => {
queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() });
toast.custom((t) => <Toast type="success" body="Profile was added" t={t} />);
toggle();
},
onError: () => {
toast.custom((t) => <Toast type="error" body="Profile could not be added" t={t} />);
}
});
const onSubmit = (data: unknown) => addMutation.mutate(data as ReleaseProfileDuplicate);
const initialValues: ReleaseProfileDuplicate = {
id: 0,
name: "",
protocol: false,
release_name: false,
hash: false,
title: false,
sub_title: false,
year: false,
month: false,
day: false,
source: false,
resolution: false,
codec: false,
container: false,
dynamic_range: false,
audio: false,
group: false,
season: false,
episode: false,
website: false,
proper: false,
repack: false,
edition: false,
language: false,
};
return (
<SlideOver
type="CREATE"
title="Duplicate Profile"
isOpen={isOpen}
toggle={toggle}
onSubmit={onSubmit}
initialValues={initialValues}
>
{() => (
<div className="py-2 space-y-6 sm:py-0 sm:space-y-0 divide-y divide-gray-200 dark:divide-gray-700">
<TextFieldWide required name="name" label="Name"/>
<SwitchGroupWide name="release_name" label="Release name" description="Full release name" />
<SwitchGroupWide name="hash" label="Hash" description="Normalized hash of the release name. Use with Releae name for exact match" />
<SwitchGroupWide name="title" label="Title" description="Parsed title" />
<SwitchGroupWide name="sub_title" label="Sub Title" description="Parsed Sub Title like Episode Name" />
<SwitchGroupWide name="year" label="Year" />
<SwitchGroupWide name="month" label="Month" description="For daily releases" />
<SwitchGroupWide name="day" label="Day" description="For daily releases" />
<SwitchGroupWide name="source" label="Source" />
<SwitchGroupWide name="resolution" label="Resolution" />
<SwitchGroupWide name="codec" label="Codec" />
<SwitchGroupWide name="container" label="Container" />
<SwitchGroupWide name="dynamic_range" label="Dynamic Range" />
<SwitchGroupWide name="audio" label="Audio" />
<SwitchGroupWide name="group" label="Group" description="Release group" />
<SwitchGroupWide name="season" label="Season" />
<SwitchGroupWide name="episode" label="Episode" />
<SwitchGroupWide name="website" label="Website/Service" description="Services such as AMZN/HULU/NF" />
<SwitchGroupWide name="proper" label="Proper" />
<SwitchGroupWide name="repack" label="Repack" />
<SwitchGroupWide name="edition" label="Edition" />
<SwitchGroupWide name="language" label="Language" />
</div>
)}
</SlideOver>
);
}
export function ReleaseProfileDuplicateUpdateForm({ isOpen, toggle, data: profile }: UpdateFormProps<ReleaseProfileDuplicate>) {
const queryClient = useQueryClient();
const storeMutation = useMutation({
mutationFn: (profile: ReleaseProfileDuplicate) => APIClient.release.profiles.duplicates.store(profile),
onSuccess: () => {
queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() });
toast.custom((t) => <Toast type="success" body="Profile was added" t={t} />);
toggle();
},
onError: () => {
toast.custom((t) => <Toast type="error" body="Profile could not be added" t={t} />);
}
});
const onSubmit = (data: unknown) => storeMutation.mutate(data as ReleaseProfileDuplicate);
const deleteMutation = useMutation({
mutationFn: (profileId: number) => APIClient.release.profiles.duplicates.delete(profileId),
onSuccess: () => {
queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() });
queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.detail(profile.id) });
toast.custom((t) => <Toast type="success" body={`Profile ${profile.name} was deleted!`} t={t} />);
toggle();
},
});
const onDelete = () => deleteMutation.mutate(profile.id);
const initialValues: ReleaseProfileDuplicate = {
id: profile.id,
name: profile.name,
protocol: profile.protocol,
release_name: profile.release_name,
hash: profile.hash,
title: profile.title,
sub_title: profile.sub_title,
year: profile.year,
month: profile.month,
day: profile.day,
source: profile.source,
resolution: profile.resolution,
codec: profile.codec,
container: profile.container,
dynamic_range: profile.dynamic_range,
audio: profile.audio,
group: profile.group,
season: profile.season,
episode: profile.episode,
website: profile.website,
proper: profile.proper,
repack: profile.repack,
edition: profile.edition,
language: profile.language,
};
return (
<SlideOver
type="UPDATE"
title="Duplicate Profile"
isOpen={isOpen}
toggle={toggle}
deleteAction={onDelete}
onSubmit={onSubmit}
initialValues={initialValues}
>
{() => (
<div className="py-2 space-y-6 sm:py-0 sm:space-y-0 divide-y divide-gray-200 dark:divide-gray-700">
<TextFieldWide required name="name" label="Name"/>
<SwitchGroupWide name="release_name" label="Release name" description="Full release name" />
<SwitchGroupWide name="hash" label="Hash" description="Normalized hash of the release name. Use with Releae name for exact match" />
<SwitchGroupWide name="title" label="Title" description="Parsed title" />
<SwitchGroupWide name="sub_title" label="Sub Title" description="Parsed Sub Title like Episode Name" />
<SwitchGroupWide name="year" label="Year" />
<SwitchGroupWide name="month" label="Month" description="For daily releases" />
<SwitchGroupWide name="day" label="Day" description="For daily releases" />
<SwitchGroupWide name="source" label="Source" />
<SwitchGroupWide name="resolution" label="Resolution" />
<SwitchGroupWide name="codec" label="Codec" />
<SwitchGroupWide name="container" label="Container" />
<SwitchGroupWide name="dynamic_range" label="Dynamic Range (HDR,DV etc)" />
<SwitchGroupWide name="audio" label="Audio" />
<SwitchGroupWide name="group" label="Group" description="Release group" />
<SwitchGroupWide name="season" label="Season" />
<SwitchGroupWide name="episode" label="Episode" />
<SwitchGroupWide name="website" label="Website/Service" description="Services such as AMZN/HULU/NF" />
<SwitchGroupWide name="repack" label="Repack" />
<SwitchGroupWide name="proper" label="Proper" />
<SwitchGroupWide name="edition" label="Edition and Cut" />
<SwitchGroupWide name="language" label="Language and Region" />
</div>
)}
</SlideOver>
);
}

View file

@ -455,7 +455,8 @@ export const FilterDetails = () => {
max_leechers: filter.max_leechers,
indexers: filter.indexers || [],
actions: filter.actions || [],
external: filter.external || []
external: filter.external || [],
release_profile_duplicate_id: filter.release_profile_duplicate_id,
} as Filter}
onSubmit={handleSubmit}
enableReinitialize={true}

View file

@ -6,7 +6,7 @@
import { useSuspenseQuery } from "@tanstack/react-query";
import { downloadsPerUnitOptions } from "@domain/constants";
import { IndexersOptionsQueryOptions } from "@api/queries";
import { IndexersOptionsQueryOptions, ReleaseProfileDuplicateList } from "@api/queries";
import { DocsLink } from "@components/ExternalLink";
import { FilterLayout, FilterPage, FilterSection } from "./_components";
@ -16,20 +16,27 @@ import {
MultiSelectOption,
NumberField,
Select,
SelectFieldOption,
SwitchGroup,
TextField
} from "@components/inputs";
import * as CONSTS from "@domain/constants.ts";
const MapIndexer = (indexer: Indexer) => (
{ label: indexer.name, value: indexer.id } as MultiSelectOption
);
const MapReleaseProfile = (profile: ReleaseProfileDuplicate) => (
{ label: profile.name, value: profile.id } as SelectFieldOption
);
export const General = () => {
const indexersQuery = useSuspenseQuery(IndexersOptionsQueryOptions())
const indexerOptions = indexersQuery.data && indexersQuery.data.map(MapIndexer)
const duplicateProfilesQuery = useSuspenseQuery(ReleaseProfileDuplicateList())
const duplicateProfilesOptions = duplicateProfilesQuery.data && duplicateProfilesQuery.data.map(MapReleaseProfile)
// const indexerOptions = data?.map(MapIndexer) ?? [];
return (
@ -129,6 +136,13 @@ export const General = () => {
</div>
}
/>
<Select
name={`release_profile_duplicate_id`}
label="Skip Duplicates profile"
optionDefaultText="Select profile"
options={[{label: "Select profile", value: null}, ...duplicateProfilesOptions]}
tooltip={<div><p>Select the skip duplicate profile.</p></div>}
/>
</FilterLayout>
<FilterLayout>

View file

@ -106,9 +106,9 @@ function ListItem({ client }: DLSettingsItemProps) {
<li>
<div className="grid grid-cols-12 items-center py-2">
<DownloadClientUpdateForm
client={client}
isOpen={updateClientIsOpen}
toggle={toggleUpdateClient}
data={client}
/>
<div className="col-span-2 sm:col-span-1 pl-1 sm:pl-6 flex items-center">
<Checkbox

View file

@ -167,7 +167,7 @@ function ListItem({ feed }: ListItemProps) {
return (
<li key={feed.id}>
<FeedUpdateForm isOpen={updateFormIsOpen} toggle={toggleUpdateForm} feed={feed} />
<FeedUpdateForm isOpen={updateFormIsOpen} toggle={toggleUpdateForm} data={feed} />
<div className="grid grid-cols-12 items-center text-sm font-medium text-gray-900 dark:text-gray-500">
<div className="col-span-2 sm:col-span-1 pl-6 flex items-center">

View file

@ -136,7 +136,7 @@ const ListItem = ({ indexer }: ListItemProps) => {
<IndexerUpdateForm
isOpen={updateIsOpen}
toggle={toggleUpdate}
indexer={indexer}
data={indexer}
/>
<div className="col-span-2 sm:col-span-1 flex pl-1 sm:pl-5 items-center">
<Checkbox value={indexer.enabled ?? false} setValue={onToggleMutation} />

View file

@ -237,7 +237,7 @@ const ListItem = ({ network, expanded }: ListItemProps) => {
<IrcNetworkUpdateForm
isOpen={updateIsOpen}
toggle={toggleUpdate}
network={network}
data={network}
/>
<div className="col-span-2 md:col-span-1 flex pl-1 sm:pl-2.5 text-gray-500 dark:text-gray-400">
<Checkbox

View file

@ -105,7 +105,7 @@ function ListItem({ notification }: ListItemProps) {
return (
<li key={notification.id} className="text-gray-500 dark:text-gray-400">
<NotificationUpdateForm isOpen={updateFormIsOpen} toggle={toggleUpdateForm} notification={notification} />
<NotificationUpdateForm isOpen={updateFormIsOpen} toggle={toggleUpdateForm} data={notification} />
<div className="grid grid-cols-12 items-center py-2">
<div className="col-span-2 sm:col-span-1 pl-1 py-0.5 sm:pl-6 flex items-center">

View file

@ -4,7 +4,7 @@
*/
import { useRef, useState } from "react";
import { useMutation, useQueryClient, useQuery } from "@tanstack/react-query";
import { useMutation, useQueryClient, useQuery, useSuspenseQuery } from "@tanstack/react-query";
import { MultiSelect as RMSC } from "react-multi-select-component";
import { AgeSelect } from "@components/inputs"
@ -15,21 +15,150 @@ import Toast from "@components/notifications/Toast";
import { useToggle } from "@hooks/hooks";
import { DeleteModal } from "@components/modals";
import { Section } from "./_components";
import { ReleaseProfileDuplicateList } from "@api/queries.ts";
import { EmptySimple } from "@components/emptystates";
import { PlusIcon } from "@heroicons/react/24/solid";
import { ReleaseProfileDuplicateAddForm, ReleaseProfileDuplicateUpdateForm } from "@forms/settings/ReleaseForms.tsx";
import { classNames } from "@utils";
const ReleaseSettings = () => (
<Section
title="Releases"
description="Manage release history."
>
<div className="border border-red-500 rounded">
<div className="py-6 px-4 sm:p-6">
<DeleteReleases />
<div className="lg:col-span-9">
<ReleaseProfileDuplicates/>
<div className="py-6 px-4 sm:p-6">
<div className="border border-red-500 rounded">
<div className="py-6 px-4 sm:p-6">
<DeleteReleases/>
</div>
</div>
</div>
</Section>
</div>
);
interface ReleaseProfileProps {
profile: ReleaseProfileDuplicate;
}
function ReleaseProfileListItem({ profile }: ReleaseProfileProps) {
const [updatePanelIsOpen, toggleUpdatePanel] = useToggle(false);
return (
<li>
<div className="grid grid-cols-12 items-center py-2">
<ReleaseProfileDuplicateUpdateForm isOpen={updatePanelIsOpen} toggle={toggleUpdatePanel} data={profile}/>
<div
className="col-span-2 sm:col-span-2 lg:col-span-2 pl-4 sm:pl-4 pr-6 py-3 block flex-col text-sm font-medium text-gray-900 dark:text-white truncate"
title={profile.name}>
{profile.name}
</div>
<div className="col-span-9 sm:col-span-9 lg:col-span-9 pl-4 sm:pl-4 pr-6 py-3 flex gap-x-0.5 flex-row text-sm font-medium text-gray-900 dark:text-white truncate">
{profile.release_name && <EnabledPill value={profile.release_name} label="RLS" title="Release name" />}
{profile.hash && <EnabledPill value={profile.hash} label="Hash" title="Normalized hash of the release name. Use with Releae name for exact match" />}
{profile.title && <EnabledPill value={profile.title} label="Title" title="Parsed titel" />}
{profile.sub_title && <EnabledPill value={profile.sub_title} label="Sub Title" title="Parsed sub titel like Episode name" />}
{profile.group && <EnabledPill value={profile.group} label="Group" title="Releae group" />}
{profile.year && <EnabledPill value={profile.year} label="Year" title="Year" />}
{profile.month && <EnabledPill value={profile.month} label="Month" title="Month" />}
{profile.day && <EnabledPill value={profile.day} label="Day" title="Day" />}
{profile.source && <EnabledPill value={profile.source} label="Source" title="Source" />}
{profile.resolution && <EnabledPill value={profile.resolution} label="Resolution" title="Resolution" />}
{profile.codec && <EnabledPill value={profile.codec} label="Codec" title="Codec" />}
{profile.container && <EnabledPill value={profile.container} label="Container" title="Container" />}
{profile.dynamic_range && <EnabledPill value={profile.dynamic_range} label="Dynamic Range" title="Dynamic Range (HDR,DV)" />}
{profile.audio && <EnabledPill value={profile.audio} label="Audio" title="Audio formats" />}
{profile.season && <EnabledPill value={profile.season} label="Season" title="Season number" />}
{profile.episode && <EnabledPill value={profile.episode} label="Episode" title="Episode number" />}
{profile.website && <EnabledPill value={profile.website} label="Website" title="Website/Service" />}
{profile.proper && <EnabledPill value={profile.proper} label="Proper" title="Scene proper" />}
{profile.repack && <EnabledPill value={profile.repack} label="Repack" title="Scene repack" />}
{profile.edition && <EnabledPill value={profile.edition} label="Edition" title="Edition (eg. Collectors Edition) and Cut (eg. Directors Cut)" />}
{profile.language && <EnabledPill value={profile.language} label="Language" title="Language and Region" />}
</div>
<div className="col-span-1 pl-0.5 whitespace-nowrap text-center text-sm font-medium">
<span className="text-blue-600 dark:text-gray-300 hover:text-blue-900 cursor-pointer"
onClick={toggleUpdatePanel}
>
Edit
</span>
</div>
</div>
</li>
)
}
interface PillProps {
value: boolean;
label: string;
title: string;
}
const EnabledPill = ({ value, label, title }: PillProps) => (
<span title={title} className={classNames("inline-flex items-center rounded-md px-1.5 py-0.5 text-xs font-medium ring-1 ring-inset", value ? "bg-blue-100 dark:bg-blue-400/10 text-blue-700 dark:text-blue-400 ring-blue-700/10 dark:ring-blue-400/30" : "bg-gray-100 dark:bg-gray-400/10 text-gray-600 dark:text-gray-400 ring-gray-500/10 dark:ring-gray-400/30")}>
{label}
</span>
);
function ReleaseProfileDuplicates() {
const [addPanelIsOpen, toggleAdd] = useToggle(false);
const releaseProfileQuery = useSuspenseQuery(ReleaseProfileDuplicateList())
return (
<Section
title="Release Duplicate Profiles"
description="Manage duplicate profiles."
rightSide={
<button
type="button"
className="relative inline-flex items-center px-4 py-2 border border-transparent shadow-sm text-sm font-medium rounded-md text-white bg-blue-600 dark:bg-blue-600 hover:bg-blue-700 dark:hover:bg-blue-700 focus:outline-none focus:ring-2 focus:ring-offset-2 focus:ring-blue-500 dark:focus:ring-blue-500"
onClick={toggleAdd}
>
<PlusIcon className="h-5 w-5 mr-1"/>
Add new
</button>
}
>
<ReleaseProfileDuplicateAddForm isOpen={addPanelIsOpen} toggle={toggleAdd}/>
<div className="flex flex-col">
{releaseProfileQuery.data.length > 0 ? (
<ul className="min-w-full relative">
<li className="grid grid-cols-12 border-b border-gray-200 dark:border-gray-700">
<div
className="col-span-2 sm:col-span-1 pl-1 sm:pl-4 py-3 text-left text-xs font-medium text-gray-500 dark:text-gray-400 uppercase tracking-wider">Name
</div>
{/*<div*/}
{/* className="col-span-6 sm:col-span-4 lg:col-span-4 pl-10 sm:pl-12 py-3 text-left text-xs font-medium text-gray-500 dark:text-gray-400 uppercase tracking-wider cursor-pointer"*/}
{/* // onClick={() => sortedClients.requestSort("name")}*/}
{/*>*/}
{/* Name*/}
{/*</div>*/}
{/*<div*/}
{/* className="hidden sm:flex col-span-4 py-3 text-left text-xs font-medium text-gray-500 dark:text-gray-400 uppercase tracking-wider cursor-pointer"*/}
{/* onClick={() => sortedClients.requestSort("host")}*/}
{/*>*/}
{/* Host <span className="sort-indicator">{sortedClients.getSortIndicator("host")}</span>*/}
{/*</div>*/}
{/*<div className="hidden sm:flex col-span-3 py-3 text-left text-xs font-medium text-gray-500 dark:text-gray-400 uppercase tracking-wider cursor-pointer"*/}
{/* onClick={() => sortedClients.requestSort("type")}*/}
{/*>*/}
{/* Type <span className="sort-indicator">{sortedClients.getSortIndicator("type")}</span>*/}
{/*</div>*/}
</li>
{releaseProfileQuery.data.map((profile) => (
<ReleaseProfileListItem key={profile.id} profile={profile}/>
))}
</ul>
) : (
<EmptySimple title="No duplicate rlease profiles" subtitle="" buttonText="Add new profile"
buttonAction={toggleAdd}/>
)}
</div>
</Section>
)
}
const getDurationLabel = (durationValue: number): string => {
const durationOptions: Record<number, string> = {
@ -87,11 +216,12 @@ function DeleteReleases() {
onSuccess: () => {
if (parsedDuration === 0) {
toast.custom((t) => (
<Toast type="success" body={"All releases based on criteria were deleted."} t={t} />
<Toast type="success" body={"All releases based on criteria were deleted."} t={t}/>
));
} else {
toast.custom((t) => (
<Toast type="success" body={`Releases older than ${getDurationLabel(parsedDuration ?? 0)} were deleted.`} t={t} />
<Toast type="success" body={`Releases older than ${getDurationLabel(parsedDuration ?? 0)} were deleted.`}
t={t}/>
));
}
@ -101,11 +231,15 @@ function DeleteReleases() {
const deleteOlderReleases = () => {
if (parsedDuration === undefined || isNaN(parsedDuration) || parsedDuration < 0) {
toast.custom((t) => <Toast type="error" body={"Please select a valid age."} t={t} />);
toast.custom((t) => <Toast type="error" body={"Please select a valid age."} t={t}/>);
return;
}
deleteOlderMutation.mutate({ olderThan: parsedDuration, indexers: indexers.map(i => i.value), releaseStatuses: releaseStatuses.map(rs => rs.value) });
deleteOlderMutation.mutate({
olderThan: parsedDuration,
indexers: indexers.map(i => i.value),
releaseStatuses: releaseStatuses.map(rs => rs.value)
});
};
return (
@ -122,19 +256,22 @@ function DeleteReleases() {
<div className="flex flex-col gap-2 w-full">
<div>
<h2 className="text-lg leading-4 font-bold text-gray-900 dark:text-white">Delete release history</h2>
<p className="text-sm mt-1 text-gray-500 dark:text-gray-400">
Select the criteria below to permanently delete release history records that are older than the chosen age and optionally match the selected indexers and release statuses:
<ul className="list-disc pl-5 mt-2">
<p className="text-sm mt-2 text-gray-500 dark:text-gray-400">
Select the criteria below to permanently delete release history records that are older than the chosen age
and optionally match the selected indexers and release statuses:
</p>
<ul className="list-disc pl-5 my-4 text-sm text-gray-500 dark:text-gray-400">
<li>
Older than (e.g., 6 months - all records older than 6 months will be deleted) - <strong className="text-gray-600 dark:text-gray-300">Required</strong>
Older than (e.g., 6 months - all records older than 6 months will be deleted) - <strong
className="text-gray-600 dark:text-gray-300">Required</strong>
</li>
<li>Indexers - Optional (if none selected, applies to all indexers)</li>
<li>Release statuses - Optional (if none selected, applies to all release statuses)</li>
</ul>
<p className="mt-2 text-red-600 dark:text-red-500">
<strong>Warning:</strong> If no indexers or release statuses are selected, all release history records older than the selected age will be permanently deleted, regardless of indexer or status.
</p>
</p>
<span className="pt-2 text-red-600 dark:text-red-500">
<strong>Warning:</strong> If no indexers or release statuses are selected, all release history records
older than the selected age will be permanently deleted, regardless of indexer or status.
</span>
</div>
<div className="flex flex-col sm:flex-row gap-2 pt-4 items-center text-sm">
@ -146,19 +283,23 @@ function DeleteReleases() {
<span className="text-red-600 dark:text-red-500"> *</span>
</>
),
content: <AgeSelect duration={duration} setDuration={setDuration} setParsedDuration={setParsedDuration} />
content: <AgeSelect duration={duration} setDuration={setDuration} setParsedDuration={setParsedDuration}/>
},
{
label: 'Indexers:',
content: <RMSC options={indexerOptions?.map(option => ({ value: option.identifier, label: option.name })) || []} value={indexers} onChange={setIndexers} labelledBy="Select indexers" />
content: <RMSC
options={indexerOptions?.map(option => ({ value: option.identifier, label: option.name })) || []}
value={indexers} onChange={setIndexers} labelledBy="Select indexers"/>
},
{
label: 'Release statuses:',
content: <RMSC options={releaseStatusOptions} value={releaseStatuses} onChange={setReleaseStatuses} labelledBy="Select release statuses" />
content: <RMSC options={releaseStatusOptions} value={releaseStatuses} onChange={setReleaseStatuses}
labelledBy="Select release statuses"/>
}
].map((item, index) => (
<div key={index} className="flex flex-col w-full">
<p className="text-xs font-bold text-gray-800 dark:text-gray-100 uppercase p-1 cursor-default">{item.label}</p>
<p
className="text-xs font-bold text-gray-800 dark:text-gray-100 uppercase p-1 cursor-default">{item.label}</p>
{item.content}
</div>
))}

View file

@ -82,6 +82,7 @@ interface Filter {
actions: Action[];
indexers: Indexer[];
external: ExternalFilter[];
release_profile_duplicate_id?: number;
}
interface Action {

View file

@ -75,3 +75,30 @@ interface DeleteParams {
indexers?: string[];
releaseStatuses?: string[];
}
interface ReleaseProfileDuplicate {
id: number;
name: string;
protocol: boolean;
release_name: boolean;
hash: boolean;
title: boolean;
sub_title: boolean;
year: boolean;
month: boolean;
day: boolean;
source: boolean;
resolution: boolean;
codec: boolean;
container: boolean;
dynamic_range: boolean;
audio: boolean;
group: boolean;
season: boolean;
episode: boolean;
website: boolean;
proper: boolean;
repack: boolean;
edition: boolean;
language: boolean;
}