From 4009554d10165b10ca7d705fff1bf576e9bf2a30 Mon Sep 17 00:00:00 2001 From: kenstir Date: Wed, 25 Dec 2024 16:33:46 -0500 Subject: [PATCH] feat(filters): skip duplicates (#1711) * feat(filters): skip duplicates * fix: add interface instead of any * fix(filters): tonullint * feat(filters): skip dupes check month day * chore: cleanup * feat(db): set autoincrement id * feat(filters): add repack and proper to dupe profile * feat(filters): add default dupe profiles * feat(duplicates): check audio and website * feat(duplicates): update tests * feat(duplicates): add toggles on addform * feat(duplicates): fix sqlite upgrade path and initialize duplicate profiles * feat(duplicates): simplify sqlite upgrade avoiding temp table and unwieldy select. Besides, FK constraints are turned off anyway in #229. * feat(duplicates): change CheckIsDuplicateRelease treatment of PROPER and REPACK "Proper" and "Repack" are not parallel to the other conditions like "Title", so they do not belong as dedup conditions. "PROPER" means there was an issue in the previous release, and so a PROPER is never a duplicate, even if it replaces another PROPER. Similarly, "REPACK" means there was an issue in the previous release by that group, and so it is a duplicate only if we previously took a release from a DIFFERENT group. I have not removed Proper and Repack from the UI or the schema yet. * feat(duplicates): update postgres schema to match sqlite * feat(duplicates): fix web build errors * feat(duplicates): fix postgres errors * feat(filters): do leftjoin for duplicate profile * fix(filters): partial update dupe profile * go fmt `internal/domain/filter.go` * feat(duplicates): restore straightforward logic for proper/repack * feat(duplicates): remove mostly duplicate TV duplicate profiles Having one profile seems the cleanest. If somebody wants multiple resolutions then they can add Resolution to the duplicate profile. Tested this profile with both weekly episodic releases and daily show releases. * feat(release): add db indexes and sub_title * feat(release): add IsDuplicate tests * feat(release): update action handler * feat(release): add more tests for skip duplicates * feat(duplicates): check audio * feat(duplicates): add more tests * feat(duplicates): match edition cut and more * fix(duplicates): tests * fix(duplicates): missing imports * fix(duplicates): tests * feat(duplicates): handle sub_title edition and language in ui * fix(duplicates): tests * feat(duplicates): check name against normalized hash * fix(duplicates): tests * chore: update .gitignore to ignore .pnpm-store * fix: tests * fix(filters): tests * fix: bad conflict merge * fix: update release type in test * fix: use vendored hot-toast * fix: release_test.go * fix: rss_test.go * feat(duplicates): improve title hashing for unique check * feat(duplicates): further improve title hashing for unique check with lang * feat(duplicates): fix tests * feat(duplicates): add macros IsDuplicate and DuplicateProfile ID and name * feat(duplicates): add normalized hash match option * fix: headlessui-state prop warning * fix(duplicates): add missing year in daily ep normalize * fix(duplicates): check rejections len --------- Co-authored-by: ze0s --- .gitignore | 1 + internal/action/service.go | 4 +- internal/database/action.go | 6 +- internal/database/action_test.go | 38 +- internal/database/database.go | 6 +- internal/database/filter.go | 94 +++ internal/database/filter_test.go | 32 +- internal/database/postgres_migrate.go | 267 +++++- internal/database/release.go | 485 ++++++++++- internal/database/release_test.go | 767 +++++++++++++++++- internal/database/sqlite_migrate.go | 272 ++++++- internal/database/utils.go | 1 + internal/domain/action.go | 2 +- internal/domain/filter.go | 315 +++---- internal/domain/macros.go | 8 +- internal/domain/macros_test.go | 3 +- internal/domain/release.go | 277 ++++++- internal/domain/release_download_test.go | 3 +- internal/domain/release_test.go | 713 ++++++++++------ internal/domain/releasetags.go | 5 + internal/feed/rss_test.go | 250 +++++- internal/filter/service.go | 235 +++--- internal/http/action.go | 14 +- internal/http/release.go | 63 ++ internal/release/service.go | 77 +- web/src/api/APIClient.ts | 11 +- web/src/api/queries.ts | 10 +- web/src/api/query_keys.ts | 7 + web/src/components/inputs/select.tsx | 6 +- web/src/forms/_shared.ts | 15 + web/src/forms/filters/FilterAddForm.tsx | 9 +- web/src/forms/settings/APIKeyAddForm.tsx | 8 +- .../forms/settings/DownloadClientForms.tsx | 16 +- web/src/forms/settings/FeedForms.tsx | 11 +- web/src/forms/settings/IndexerForms.tsx | 16 +- web/src/forms/settings/IrcForms.tsx | 16 +- web/src/forms/settings/NotificationForms.tsx | 16 +- web/src/forms/settings/ProxyForms.tsx | 4 +- web/src/forms/settings/ReleaseForms.tsx | 198 +++++ web/src/screens/filters/Details.tsx | 3 +- web/src/screens/filters/sections/General.tsx | 18 +- web/src/screens/settings/DownloadClient.tsx | 2 +- web/src/screens/settings/Feed.tsx | 2 +- web/src/screens/settings/Indexer.tsx | 2 +- web/src/screens/settings/Irc.tsx | 2 +- web/src/screens/settings/Notifications.tsx | 2 +- web/src/screens/settings/Releases.tsx | 193 ++++- web/src/types/Filter.d.ts | 1 + web/src/types/Release.d.ts | 29 +- 49 files changed, 3792 insertions(+), 743 deletions(-) create mode 100644 web/src/forms/_shared.ts create mode 100644 web/src/forms/settings/ReleaseForms.tsx diff --git a/.gitignore b/.gitignore index 00bda49..bf1f425 100644 --- a/.gitignore +++ b/.gitignore @@ -29,6 +29,7 @@ node_modules/ !.yarn/versions # Ditto for yarn, except we're using pnpm yarn.lock +.pnpm-store # If needed, package-lock.json shall be added manually using an explicit git add command package-lock.json diff --git a/internal/action/service.go b/internal/action/service.go index daa6155..806b006 100644 --- a/internal/action/service.go +++ b/internal/action/service.go @@ -21,7 +21,7 @@ import ( ) type Service interface { - Store(ctx context.Context, action domain.Action) (*domain.Action, error) + Store(ctx context.Context, action *domain.Action) error StoreFilterActions(ctx context.Context, filterID int64, actions []*domain.Action) ([]*domain.Action, error) List(ctx context.Context) ([]domain.Action, error) Get(ctx context.Context, req *domain.GetActionRequest) (*domain.Action, error) @@ -63,7 +63,7 @@ func NewService(log logger.Logger, repo domain.ActionRepo, clientSvc download_cl return s } -func (s *service) Store(ctx context.Context, action domain.Action) (*domain.Action, error) { +func (s *service) Store(ctx context.Context, action *domain.Action) error { return s.repo.Store(ctx, action) } diff --git a/internal/database/action.go b/internal/database/action.go index 821affa..cdf4290 100644 --- a/internal/database/action.go +++ b/internal/database/action.go @@ -706,7 +706,7 @@ func (r *ActionRepo) DeleteByFilterID(ctx context.Context, filterID int) error { return nil } -func (r *ActionRepo) Store(ctx context.Context, action domain.Action) (*domain.Action, error) { +func (r *ActionRepo) Store(ctx context.Context, action *domain.Action) error { queryBuilder := r.db.squirrel. Insert("action"). Columns( @@ -783,14 +783,14 @@ func (r *ActionRepo) Store(ctx context.Context, action domain.Action) (*domain.A var retID int64 if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil { - return nil, errors.Wrap(err, "error executing query") + return errors.Wrap(err, "error executing query") } action.ID = int(retID) r.log.Debug().Msgf("action.store: added new %d", retID) - return &action, nil + return nil } func (r *ActionRepo) Update(ctx context.Context, action domain.Action) (*domain.Action, error) { diff --git a/internal/database/action_test.go b/internal/database/action_test.go index 471f809..0d59102 100644 --- a/internal/database/action_test.go +++ b/internal/database/action_test.go @@ -16,8 +16,8 @@ import ( "github.com/stretchr/testify/assert" ) -func getMockAction() domain.Action { - return domain.Action{ +func getMockAction() *domain.Action { + return &domain.Action{ Name: "randomAction", Type: domain.ActionTypeTest, Enabled: true, @@ -78,29 +78,29 @@ func TestActionRepo_Store(t *testing.T) { mockData.FilterID = createdFilters[0].ID // Actual test for Store - createdAction, err := repo.Store(context.Background(), mockData) + err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - assert.NotNil(t, createdAction) + assert.NotNil(t, mockData) // Cleanup - _ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) t.Run(fmt.Sprintf("Store_Succeeds_With_Missing_or_empty_fields [%s]", dbType), func(t *testing.T) { - mockData := domain.Action{} - createdAction, err := repo.Store(context.Background(), mockData) + mockData := &domain.Action{} + err := repo.Store(context.Background(), mockData) assert.NoError(t, err) // Cleanup - _ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = repo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockData.ID}) }) t.Run(fmt.Sprintf("Store_Fails_With_Invalid_ClientID [%s]", dbType), func(t *testing.T) { mockData := getMockAction() mockData.ClientID = 9999 - _, err := repo.Store(context.Background(), mockData) + err := repo.Store(context.Background(), mockData) assert.Error(t, err) }) @@ -110,7 +110,7 @@ func TestActionRepo_Store(t *testing.T) { ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond) defer cancel() - _, err := repo.Store(ctx, mockData) + err := repo.Store(ctx, mockData) assert.Error(t, err) }) } @@ -142,7 +142,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) { mockData.FilterID = createdFilters[0].ID // Actual test for StoreFilterActions - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) assert.NotNil(t, createdActions) @@ -154,7 +154,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) { }) t.Run(fmt.Sprintf("StoreFilterActions_Fails_Invalid_FilterID [%s]", dbType), func(t *testing.T) { - _, err := repo.StoreFilterActions(context.Background(), 9999, []*domain.Action{&mockData}) + _, err := repo.StoreFilterActions(context.Background(), 9999, []*domain.Action{mockData}) assert.NoError(t, err) }) @@ -186,7 +186,7 @@ func TestActionRepo_StoreFilterActions(t *testing.T) { assert.NoError(t, err) assert.NotNil(t, createdFilters) - _, err = repo.StoreFilterActions(ctx, int64(createdFilters[0].ID), []*domain.Action{&mockData}) + _, err = repo.StoreFilterActions(ctx, int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.Error(t, err) // Cleanup @@ -219,7 +219,7 @@ func TestActionRepo_FindByFilterID(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) // Actual test for FindByFilterID @@ -294,7 +294,7 @@ func TestActionRepo_List(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) // Actual test for List @@ -344,7 +344,7 @@ func TestActionRepo_Get(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) // Actual test for Get @@ -401,7 +401,7 @@ func TestActionRepo_Delete(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) // Actual test for Delete @@ -455,7 +455,7 @@ func TestActionRepo_DeleteByFilterID(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) err = repo.DeleteByFilterID(context.Background(), mockData.FilterID) @@ -508,7 +508,7 @@ func TestActionRepo_ToggleEnabled(t *testing.T) { mockData.ClientID = mock.ID mockData.FilterID = createdFilters[0].ID mockData.Enabled = false - createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{&mockData}) + createdActions, err := repo.StoreFilterActions(context.Background(), int64(createdFilters[0].ID), []*domain.Action{mockData}) assert.NoError(t, err) // Actual test for ToggleEnabled diff --git a/internal/database/database.go b/internal/database/database.go index edd56d8..033cb34 100644 --- a/internal/database/database.go +++ b/internal/database/database.go @@ -129,13 +129,9 @@ type Tx struct { handler *DB } -type ILikeDynamic interface { - ToSql() (sql string, args []interface{}, err error) -} - // ILike is a wrapper for sq.Like and sq.ILike // SQLite does not support ILike but postgres does so this checks what database is being used -func (db *DB) ILike(col string, val string) ILikeDynamic { +func (db *DB) ILike(col string, val string) sq.Sqlizer { //if databaseDriver == "sqlite" { if db.Driver == "sqlite" { return sq.Like{col: val} diff --git a/internal/database/filter.go b/internal/database/filter.go index 4bcd05d..39078a0 100644 --- a/internal/database/filter.go +++ b/internal/database/filter.go @@ -240,6 +240,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter "f.max_seeders", "f.min_leechers", "f.max_leechers", + "f.release_profile_duplicate_id", "f.created_at", "f.updated_at", ). @@ -266,6 +267,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter var minSize, maxSize, maxDownloadsUnit, matchReleases, exceptReleases, matchReleaseGroups, exceptReleaseGroups, matchReleaseTags, exceptReleaseTags, matchDescription, exceptDescription, freeleechPercent, shows, seasons, episodes, years, months, days, artists, albums, matchCategories, exceptCategories, matchUploaders, exceptUploaders, matchRecordLabels, exceptRecordLabels, tags, exceptTags, tagsMatchLogic, exceptTagsMatchLogic sql.NullString var useRegex, scene, freeleech, hasLog, hasCue, perfectFlac sql.NullBool var delay, maxDownloads, logScore sql.NullInt32 + var releaseProfileDuplicateId sql.NullInt64 err = row.Scan( &f.ID, @@ -335,6 +337,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter &f.MaxSeeders, &f.MinLeechers, &f.MaxLeechers, + &releaseProfileDuplicateId, &f.CreatedAt, &f.UpdatedAt, ) @@ -385,6 +388,7 @@ func (r *FilterRepo) FindByID(ctx context.Context, filterID int) (*domain.Filter f.UseRegex = useRegex.Bool f.Scene = scene.Bool f.Freeleech = freeleech.Bool + f.ReleaseProfileDuplicateID = releaseProfileDuplicateId.Int64 return &f, nil } @@ -466,10 +470,35 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string "f.max_leechers", "f.created_at", "f.updated_at", + "f.release_profile_duplicate_id", + "rdp.id", + "rdp.name", + "rdp.release_name", + "rdp.hash", + "rdp.title", + "rdp.sub_title", + "rdp.year", + "rdp.month", + "rdp.day", + "rdp.source", + "rdp.resolution", + "rdp.codec", + "rdp.container", + "rdp.dynamic_range", + "rdp.audio", + "rdp.release_group", + "rdp.season", + "rdp.episode", + "rdp.website", + "rdp.proper", + "rdp.repack", + "rdp.edition", + "rdp.language", ). From("filter f"). Join("filter_indexer fi ON f.id = fi.filter_id"). Join("indexer i ON i.id = fi.indexer_id"). + LeftJoin("release_profile_duplicate rdp ON rdp.id = f.release_profile_duplicate_id"). Where(sq.Eq{"i.identifier": indexer}). Where(sq.Eq{"i.enabled": true}). Where(sq.Eq{"f.enabled": true}). @@ -495,6 +524,10 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string var minSize, maxSize, maxDownloadsUnit, matchReleases, exceptReleases, matchReleaseGroups, exceptReleaseGroups, matchReleaseTags, exceptReleaseTags, matchDescription, exceptDescription, freeleechPercent, shows, seasons, episodes, years, months, days, artists, albums, matchCategories, exceptCategories, matchUploaders, exceptUploaders, matchRecordLabels, exceptRecordLabels, tags, exceptTags, tagsMatchLogic, exceptTagsMatchLogic sql.NullString var useRegex, scene, freeleech, hasLog, hasCue, perfectFlac sql.NullBool var delay, maxDownloads, logScore sql.NullInt32 + var releaseProfileDuplicateID, rdpId sql.NullInt64 + + var rdpName sql.NullString + var rdpRelName, rdpHash, rdpTitle, rdpSubTitle, rdpYear, rdpMonth, rdpDay, rdpSource, rdpResolution, rdpCodec, rdpContainer, rdpDynRange, rdpAudio, rdpGroup, rdpSeason, rdpEpisode, rdpWebsite, rdpProper, rdpRepack, rdpEdition, rdpLanguage sql.NullBool err := rows.Scan( &f.ID, @@ -566,6 +599,30 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string &f.MaxLeechers, &f.CreatedAt, &f.UpdatedAt, + &releaseProfileDuplicateID, + &rdpId, + &rdpName, + &rdpRelName, + &rdpHash, + &rdpTitle, + &rdpSubTitle, + &rdpYear, + &rdpMonth, + &rdpDay, + &rdpSource, + &rdpResolution, + &rdpCodec, + &rdpContainer, + &rdpDynRange, + &rdpAudio, + &rdpGroup, + &rdpSeason, + &rdpEpisode, + &rdpWebsite, + &rdpProper, + &rdpRepack, + &rdpEdition, + &rdpLanguage, ) if err != nil { return nil, errors.Wrap(err, "error scanning row") @@ -610,9 +667,40 @@ func (r *FilterRepo) findByIndexerIdentifier(ctx context.Context, indexer string f.UseRegex = useRegex.Bool f.Scene = scene.Bool f.Freeleech = freeleech.Bool + f.ReleaseProfileDuplicateID = releaseProfileDuplicateID.Int64 f.Rejections = []string{} + if releaseProfileDuplicateID.Valid { + profile := domain.DuplicateReleaseProfile{ + ID: rdpId.Int64, + //Protocol: rdpName.String, + Name: rdpName.String, + ReleaseName: rdpRelName.Bool, + Hash: rdpHash.Bool, + Title: rdpTitle.Bool, + SubTitle: rdpSubTitle.Bool, + Year: rdpYear.Bool, + Month: rdpMonth.Bool, + Day: rdpDay.Bool, + Source: rdpSource.Bool, + Resolution: rdpResolution.Bool, + Codec: rdpCodec.Bool, + Container: rdpContainer.Bool, + DynamicRange: rdpDynRange.Bool, + Audio: rdpAudio.Bool, + Group: rdpGroup.Bool, + Season: rdpSeason.Bool, + Episode: rdpEpisode.Bool, + Website: rdpWebsite.Bool, + Proper: rdpProper.Bool, + Repack: rdpRepack.Bool, + Edition: rdpEdition.Bool, + Language: rdpLanguage.Bool, + } + f.DuplicateHandling = &profile + } + filters = append(filters, &f) } @@ -774,6 +862,7 @@ func (r *FilterRepo) Store(ctx context.Context, filter *domain.Filter) error { "max_seeders", "min_leechers", "max_leechers", + "release_profile_duplicate_id", ). Values( filter.Name, @@ -842,6 +931,7 @@ func (r *FilterRepo) Store(ctx context.Context, filter *domain.Filter) error { filter.MaxSeeders, filter.MinLeechers, filter.MaxLeechers, + toNullInt64(filter.ReleaseProfileDuplicateID), ). Suffix("RETURNING id").RunWith(r.db.handler) @@ -928,6 +1018,7 @@ func (r *FilterRepo) Update(ctx context.Context, filter *domain.Filter) error { Set("max_seeders", filter.MaxSeeders). Set("min_leechers", filter.MinLeechers). Set("max_leechers", filter.MaxLeechers). + Set("release_profile_duplicate_id", toNullInt64(filter.ReleaseProfileDuplicateID)). Set("updated_at", time.Now().Format(time.RFC3339)). Where(sq.Eq{"id": filter.ID}) @@ -1153,6 +1244,9 @@ func (r *FilterRepo) UpdatePartial(ctx context.Context, filter domain.FilterUpda if filter.MaxLeechers != nil { q = q.Set("max_leechers", filter.MaxLeechers) } + if filter.ReleaseProfileDuplicateID != nil { + q = q.Set("release_profile_duplicate_id", filter.ReleaseProfileDuplicateID) + } q = q.Where(sq.Eq{"id": filter.ID}) diff --git a/internal/database/filter_test.go b/internal/database/filter_test.go index 887afe2..133d248 100644 --- a/internal/database/filter_test.go +++ b/internal/database/filter_test.go @@ -800,7 +800,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { mockAction.FilterID = mockData.ID mockAction.ClientID = mockClient.ID - action, err := actionRepo.Store(context.Background(), mockAction) + err = actionRepo.Store(context.Background(), mockAction) mockReleaseActionStatus.FilterID = int64(mockData.ID) mockRelease.FilterID = mockData.ID @@ -808,7 +808,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { err = releaseRepo.Store(context.Background(), mockRelease) assert.NoError(t, err) - mockReleaseActionStatus.ActionID = int64(action.ID) + mockReleaseActionStatus.ActionID = int64(mockAction.ID) mockReleaseActionStatus.ReleaseID = mockRelease.ID err = releaseRepo.StoreReleaseActionStatus(context.Background(), mockReleaseActionStatus) @@ -827,7 +827,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { }) // Cleanup - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction.ID}) _ = repo.Delete(context.Background(), mockData.ID) _ = downloadClientRepo.Delete(context.Background(), mockClient.ID) _ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) @@ -861,13 +861,15 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { mockAction1.FilterID = mockData.ID mockAction1.ClientID = mockClient.ID - action1, err := actionRepo.Store(context.Background(), mockAction1) + actionErr := actionRepo.Store(context.Background(), mockAction1) + assert.NoError(t, actionErr) mockAction2 := getMockAction() mockAction2.FilterID = mockData.ID mockAction2.ClientID = mockClient.ID - action2, err := actionRepo.Store(context.Background(), mockAction2) + action2Err := actionRepo.Store(context.Background(), mockAction2) + assert.NoError(t, action2Err) mockRelease.FilterID = mockData.ID @@ -875,7 +877,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { assert.NoError(t, err) mockReleaseActionStatus1 := getMockReleaseActionStatus() - mockReleaseActionStatus1.ActionID = int64(action1.ID) + mockReleaseActionStatus1.ActionID = int64(mockAction1.ID) mockReleaseActionStatus1.FilterID = int64(mockData.ID) mockReleaseActionStatus1.ReleaseID = mockRelease.ID @@ -883,7 +885,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { assert.NoError(t, err) mockReleaseActionStatus2 := getMockReleaseActionStatus() - mockReleaseActionStatus2.ActionID = int64(action2.ID) + mockReleaseActionStatus2.ActionID = int64(mockAction2.ID) mockReleaseActionStatus2.FilterID = int64(mockData.ID) mockReleaseActionStatus2.ReleaseID = mockRelease.ID @@ -903,8 +905,8 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { }) // Cleanup - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action1.ID}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action2.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction1.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction2.ID}) _ = repo.Delete(context.Background(), mockData.ID) _ = downloadClientRepo.Delete(context.Background(), mockClient.ID) _ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) @@ -924,13 +926,15 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { mockAction.FilterID = mockData.ID mockAction.ClientID = mockClient.ID - action, err := actionRepo.Store(context.Background(), mockAction) + err = actionRepo.Store(context.Background(), mockAction) + assert.NoError(t, err) mockAction2 := getMockAction() mockAction2.FilterID = mockData.ID mockAction2.ClientID = mockClient.ID - action2, err := actionRepo.Store(context.Background(), mockAction2) + err = actionRepo.Store(context.Background(), mockAction2) + assert.NoError(t, err) mockRelease.FilterID = mockData.ID @@ -938,7 +942,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { assert.NoError(t, err) mockReleaseActionStatus = getMockReleaseActionStatus() - mockReleaseActionStatus.ActionID = int64(action.ID) + mockReleaseActionStatus.ActionID = int64(mockAction.ID) mockReleaseActionStatus.FilterID = int64(mockData.ID) mockReleaseActionStatus.ReleaseID = mockRelease.ID mockReleaseActionStatus.Timestamp = mockReleaseActionStatus.Timestamp.AddDate(0, -1, 0) @@ -947,7 +951,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { assert.NoError(t, err) mockReleaseActionStatus2 := getMockReleaseActionStatus() - mockReleaseActionStatus2.ActionID = int64(action2.ID) + mockReleaseActionStatus2.ActionID = int64(mockAction2.ID) mockReleaseActionStatus2.FilterID = int64(mockData.ID) mockReleaseActionStatus2.ReleaseID = mockRelease.ID mockReleaseActionStatus2.Timestamp = mockReleaseActionStatus2.Timestamp.AddDate(0, -1, 0) @@ -968,7 +972,7 @@ func TestFilterRepo_GetDownloadsByFilterId(t *testing.T) { }) // Cleanup - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: action.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: mockAction.ID}) _ = repo.Delete(context.Background(), mockData.ID) _ = downloadClientRepo.Delete(context.Background(), mockClient.ID) _ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) diff --git a/internal/database/postgres_migrate.go b/internal/database/postgres_migrate.go index 358e4b7..6b005f7 100644 --- a/internal/database/postgres_migrate.go +++ b/internal/database/postgres_migrate.go @@ -88,6 +88,39 @@ CREATE TABLE irc_channel UNIQUE (network_id, name) ); +CREATE TABLE release_profile_duplicate +( + id SERIAL PRIMARY KEY, + name TEXT NOT NULL, + protocol BOOLEAN DEFAULT FALSE, + release_name BOOLEAN DEFAULT FALSE, + hash BOOLEAN DEFAULT FALSE, + title BOOLEAN DEFAULT FALSE, + sub_title BOOLEAN DEFAULT FALSE, + year BOOLEAN DEFAULT FALSE, + month BOOLEAN DEFAULT FALSE, + day BOOLEAN DEFAULT FALSE, + source BOOLEAN DEFAULT FALSE, + resolution BOOLEAN DEFAULT FALSE, + codec BOOLEAN DEFAULT FALSE, + container BOOLEAN DEFAULT FALSE, + dynamic_range BOOLEAN DEFAULT FALSE, + audio BOOLEAN DEFAULT FALSE, + release_group BOOLEAN DEFAULT FALSE, + season BOOLEAN DEFAULT FALSE, + episode BOOLEAN DEFAULT FALSE, + website BOOLEAN DEFAULT FALSE, + proper BOOLEAN DEFAULT FALSE, + repack BOOLEAN DEFAULT FALSE, + edition BOOLEAN DEFAULT FALSE, + language BOOLEAN DEFAULT FALSE +); + +INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language) +VALUES (1, 'Exact release', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'), + (2, 'Movie', 'f', 'f', 'f', 't', 'f', 't', 'f', 'f', 'f', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'), + (3, 'TV', 'f', 'f', 'f', 't', 'f', 't', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f'); + CREATE TABLE filter ( id SERIAL PRIMARY KEY, @@ -159,7 +192,9 @@ CREATE TABLE filter min_seeders INTEGER DEFAULT 0, max_seeders INTEGER DEFAULT 0, min_leechers INTEGER DEFAULT 0, - max_leechers INTEGER DEFAULT 0 + max_leechers INTEGER DEFAULT 0, + release_profile_duplicate_id INTEGER, + FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate(id) ON DELETE SET NULL ); CREATE INDEX filter_enabled_index @@ -270,9 +305,11 @@ CREATE TABLE "release" group_id TEXT, torrent_id TEXT, torrent_name TEXT, + normalized_hash TEXT, size BIGINT, raw TEXT, title TEXT, + sub_title TEXT, category TEXT, season INTEGER, episode INTEGER, @@ -285,15 +322,18 @@ CREATE TABLE "release" container TEXT, hdr TEXT, audio TEXT, + audio_channels TEXT, release_group TEXT, region TEXT, language TEXT, edition TEXT, + cut TEXT, unrated BOOLEAN, hybrid BOOLEAN, proper BOOLEAN, repack BOOLEAN, website TEXT, + media_processing TEXT, artists TEXT [] DEFAULT '{}' NOT NULL, type TEXT, format TEXT, @@ -308,6 +348,7 @@ CREATE TABLE "release" freeleech_percent INTEGER, uploader TEXT, pre_time TEXT, + other TEXT [] DEFAULT '{}' NOT NULL, filter_id INTEGER CONSTRAINT release_filter_id_fk REFERENCES filter @@ -326,6 +367,81 @@ CREATE INDEX release_timestamp_index CREATE INDEX release_torrent_name_index ON "release" (torrent_name); +CREATE INDEX release_normalized_hash_index + ON "release" (normalized_hash); + +CREATE INDEX release_title_index + ON "release" (title); + +CREATE INDEX release_sub_title_index + ON "release" (sub_title); + +CREATE INDEX release_season_index + ON "release" (season); + +CREATE INDEX release_episode_index + ON "release" (episode); + +CREATE INDEX release_year_index + ON "release" (year); + +CREATE INDEX release_month_index + ON "release" (month); + +CREATE INDEX release_day_index + ON "release" (day); + +CREATE INDEX release_resolution_index + ON "release" (resolution); + +CREATE INDEX release_source_index + ON "release" (source); + +CREATE INDEX release_codec_index + ON "release" (codec); + +CREATE INDEX release_container_index + ON "release" (container); + +CREATE INDEX release_hdr_index + ON "release" (hdr); + +CREATE INDEX release_audio_index + ON "release" (audio); + +CREATE INDEX release_audio_channels_index + ON "release" (audio_channels); + +CREATE INDEX release_release_group_index + ON "release" (release_group); + +CREATE INDEX release_language_index + ON "release" (language); + +CREATE INDEX release_proper_index + ON "release" (proper); + +CREATE INDEX release_repack_index + ON "release" (repack); + +CREATE INDEX release_website_index + ON "release" (website); + +CREATE INDEX release_media_processing_index + ON "release" (media_processing); + +CREATE INDEX release_region_index + ON "release" (region); + +CREATE INDEX release_edition_index + ON "release" (edition); + +CREATE INDEX release_cut_index + ON "release" (cut); + +CREATE INDEX release_hybrid_index + ON "release" (hybrid); + CREATE TABLE release_action_status ( id SERIAL PRIMARY KEY, @@ -1074,5 +1190,154 @@ CREATE TABLE list_filter ALTER TABLE filter ADD COLUMN except_record_labels TEXT DEFAULT ''; +`, + `CREATE TABLE release_profile_duplicate +( + id SERIAL PRIMARY KEY, + name TEXT NOT NULL, + protocol BOOLEAN DEFAULT FALSE, + release_name BOOLEAN DEFAULT FALSE, + hash BOOLEAN DEFAULT FALSE, + title BOOLEAN DEFAULT FALSE, + sub_title BOOLEAN DEFAULT FALSE, + year BOOLEAN DEFAULT FALSE, + month BOOLEAN DEFAULT FALSE, + day BOOLEAN DEFAULT FALSE, + source BOOLEAN DEFAULT FALSE, + resolution BOOLEAN DEFAULT FALSE, + codec BOOLEAN DEFAULT FALSE, + container BOOLEAN DEFAULT FALSE, + dynamic_range BOOLEAN DEFAULT FALSE, + audio BOOLEAN DEFAULT FALSE, + release_group BOOLEAN DEFAULT FALSE, + season BOOLEAN DEFAULT FALSE, + episode BOOLEAN DEFAULT FALSE, + website BOOLEAN DEFAULT FALSE, + proper BOOLEAN DEFAULT FALSE, + repack BOOLEAN DEFAULT FALSE, + edition BOOLEAN DEFAULT FALSE, + language BOOLEAN DEFAULT FALSE +); + +INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language) +VALUES (1, 'Exact release', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'), + (2, 'Movie', 'f', 'f', 'f', 't', 'f', 't', 'f', 'f', 'f', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 'f'), + (3, 'TV', 'f', 'f', 'f', 't', 'f', 't', 't', 't', 'f', 'f', 'f', 'f', 'f', 'f', 'f', 't', 't', 'f', 'f', 'f', 'f', 'f'); + +ALTER TABLE filter + ADD release_profile_duplicate_id INTEGER; + +ALTER TABLE filter + ADD CONSTRAINT filter_release_profile_duplicate_id_fk + FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate (id) + ON DELETE SET NULL; + +ALTER TABLE "release" + ADD normalized_hash TEXT; + +ALTER TABLE "release" + ADD sub_title TEXT; + +ALTER TABLE "release" + ADD COLUMN IF NOT EXISTS audio TEXT; + +ALTER TABLE "release" + ADD audio_channels TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS language TEXT; + +ALTER TABLE "release" + ADD media_processing TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS edition TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS cut TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS hybrid TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS region TEXT; + +ALTER TABLE "release" + ADD IF NOT EXISTS other TEXT [] DEFAULT '{}' NOT NULL; + +CREATE INDEX release_normalized_hash_index + ON "release" (normalized_hash); + +CREATE INDEX release_title_index + ON "release" (title); + +CREATE INDEX release_sub_title_index + ON "release" (sub_title); + +CREATE INDEX release_season_index + ON "release" (season); + +CREATE INDEX release_episode_index + ON "release" (episode); + +CREATE INDEX release_year_index + ON "release" (year); + +CREATE INDEX release_month_index + ON "release" (month); + +CREATE INDEX release_day_index + ON "release" (day); + +CREATE INDEX release_resolution_index + ON "release" (resolution); + +CREATE INDEX release_source_index + ON "release" (source); + +CREATE INDEX release_codec_index + ON "release" (codec); + +CREATE INDEX release_container_index + ON "release" (container); + +CREATE INDEX release_hdr_index + ON "release" (hdr); + +CREATE INDEX release_audio_index + ON "release" (audio); + +CREATE INDEX release_audio_channels_index + ON "release" (audio_channels); + +CREATE INDEX release_release_group_index + ON "release" (release_group); + +CREATE INDEX release_proper_index + ON "release" (proper); + +CREATE INDEX release_repack_index + ON "release" (repack); + +CREATE INDEX release_website_index + ON "release" (website); + +CREATE INDEX release_media_processing_index + ON "release" (media_processing); + +CREATE INDEX release_language_index + ON "release" (language); + +CREATE INDEX release_region_index + ON "release" (region); + +CREATE INDEX release_edition_index + ON "release" (edition); + +CREATE INDEX release_cut_index + ON "release" (cut); + +CREATE INDEX release_hybrid_index + ON "release" (hybrid); `, } diff --git a/internal/database/release.go b/internal/database/release.go index 25a9202..36aeff6 100644 --- a/internal/database/release.go +++ b/internal/database/release.go @@ -33,23 +33,31 @@ func NewReleaseRepo(log logger.Logger, db *DB) domain.ReleaseRepo { } func (repo *ReleaseRepo) Store(ctx context.Context, r *domain.Release) error { - codecStr := strings.Join(r.Codec, ",") - hdrStr := strings.Join(r.HDR, ",") + var ( + codecStr = strings.Join(r.Codec, ",") + hdrStr = strings.Join(r.HDR, ",") + audioStr = strings.Join(r.Audio, ",") + editionStr = strings.Join(r.Edition, ",") + cutStr = strings.Join(r.Cut, ",") + languageStr = strings.Join(r.Language, ",") + ) queryBuilder := repo.db.squirrel. Insert("release"). - Columns("filter_status", "rejections", "indexer", "filter", "protocol", "implementation", "timestamp", "announce_type", "group_id", "torrent_id", "info_url", "download_url", "torrent_name", "size", "title", "category", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "hdr", "release_group", "proper", "repack", "website", "type", "origin", "tags", "uploader", "pre_time", "filter_id"). - Values(r.FilterStatus, pq.Array(r.Rejections), r.Indexer.Identifier, r.FilterName, r.Protocol, r.Implementation, r.Timestamp.Format(time.RFC3339), r.AnnounceType, r.GroupID, r.TorrentID, r.InfoURL, r.DownloadURL, r.TorrentName, r.Size, r.Title, r.Category, r.Season, r.Episode, r.Year, r.Month, r.Day, r.Resolution, r.Source, codecStr, r.Container, hdrStr, r.Group, r.Proper, r.Repack, r.Website, r.Type, r.Origin, pq.Array(r.Tags), r.Uploader, r.PreTime, r.FilterID). + Columns("filter_status", "rejections", "indexer", "filter", "protocol", "implementation", "timestamp", "announce_type", "group_id", "torrent_id", "info_url", "download_url", "torrent_name", "normalized_hash", "size", "title", "sub_title", "category", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "hdr", "audio", "audio_channels", "release_group", "proper", "repack", "region", "language", "cut", "edition", "hybrid", "media_processing", "website", "type", "origin", "tags", "uploader", "pre_time", "other", "filter_id"). + Values(r.FilterStatus, pq.Array(r.Rejections), r.Indexer.Identifier, r.FilterName, r.Protocol, r.Implementation, r.Timestamp.Format(time.RFC3339), r.AnnounceType, r.GroupID, r.TorrentID, r.InfoURL, r.DownloadURL, r.TorrentName, r.NormalizedHash, r.Size, r.Title, r.SubTitle, r.Category, r.Season, r.Episode, r.Year, r.Month, r.Day, r.Resolution, r.Source, codecStr, r.Container, hdrStr, audioStr, r.AudioChannels, r.Group, r.Proper, r.Repack, r.Region, languageStr, cutStr, editionStr, r.Hybrid, r.MediaProcessing, r.Website, r.Type.String(), r.Origin, pq.Array(r.Tags), r.Uploader, r.PreTime, pq.Array(r.Other), r.FilterID). Suffix("RETURNING id").RunWith(repo.db.handler) - // return values - var retID int64 - - if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil { - return errors.Wrap(err, "error executing query") + q, args, err := queryBuilder.ToSql() + if err != nil { + return errors.Wrap(err, "error building query") } - r.ID = retID + repo.log.Debug().Msgf("release.store: %s %v", q, args) + + if err := queryBuilder.QueryRowContext(ctx).Scan(&r.ID); err != nil { + return errors.Wrap(err, "error executing query") + } repo.log.Debug().Msgf("release.store: %+v", r) @@ -102,14 +110,9 @@ func (repo *ReleaseRepo) StoreReleaseActionStatus(ctx context.Context, status *d Values(status.Status, status.Action, status.ActionID, status.Type, status.Client, status.Filter, status.FilterID, pq.Array(status.Rejections), status.Timestamp.Format(time.RFC3339), status.ReleaseID). Suffix("RETURNING id").RunWith(repo.db.handler) - // return values - var retID int64 - - if err := queryBuilder.QueryRowContext(ctx).Scan(&retID); err != nil { + if err := queryBuilder.QueryRowContext(ctx).Scan(&status.ID); err != nil { return errors.Wrap(err, "error executing query") } - - status.ID = retID } repo.log.Trace().Msgf("release.store_release_action_status: %+v", status) @@ -117,6 +120,62 @@ func (repo *ReleaseRepo) StoreReleaseActionStatus(ctx context.Context, status *d return nil } +func (repo *ReleaseRepo) StoreDuplicateProfile(ctx context.Context, profile *domain.DuplicateReleaseProfile) error { + if profile.ID == 0 { + queryBuilder := repo.db.squirrel. + Insert("release_profile_duplicate"). + Columns("name", "protocol", "release_name", "hash", "title", "sub_title", "season", "episode", "year", "month", "day", "resolution", "source", "codec", "container", "dynamic_range", "audio", "release_group", "website", "proper", "repack"). + Values(profile.Name, profile.Protocol, profile.ReleaseName, profile.Hash, profile.Title, profile.SubTitle, profile.Season, profile.Episode, profile.Year, profile.Month, profile.Day, profile.Resolution, profile.Source, profile.Codec, profile.Container, profile.DynamicRange, profile.Audio, profile.Group, profile.Website, profile.Proper, profile.Repack). + Suffix("RETURNING id"). + RunWith(repo.db.handler) + + // return values + var retID int64 + + err := queryBuilder.QueryRowContext(ctx).Scan(&retID) + if err != nil { + return errors.Wrap(err, "error executing query") + } + + profile.ID = retID + } else { + queryBuilder := repo.db.squirrel. + Update("release_profile_duplicate"). + Set("name", profile.Name). + Set("protocol", profile.Protocol). + Set("release_name", profile.ReleaseName). + Set("hash", profile.Hash). + Set("title", profile.Title). + Set("sub_title", profile.SubTitle). + Set("season", profile.Season). + Set("episode", profile.Episode). + Set("year", profile.Year). + Set("month", profile.Month). + Set("day", profile.Day). + Set("resolution", profile.Resolution). + Set("source", profile.Source). + Set("codec", profile.Codec). + Set("container", profile.Container). + Set("dynamic_range", profile.DynamicRange). + Set("audio", profile.Audio). + Set("release_group", profile.Group). + Set("website", profile.Website). + Set("proper", profile.Proper). + Set("repack", profile.Repack). + Where(sq.Eq{"id": profile.ID}). + RunWith(repo.db.handler) + + _, err := queryBuilder.ExecContext(ctx) + if err != nil { + return errors.Wrap(err, "error executing query") + } + } + + repo.log.Debug().Msgf("release.StoreDuplicateProfile: %+v", profile) + + return nil +} + func (repo *ReleaseRepo) Find(ctx context.Context, params domain.ReleaseQueryParams) (*domain.FindReleasesResponse, error) { tx, err := repo.db.BeginTx(ctx, &sql.TxOptions{Isolation: sql.LevelReadCommitted}) if err != nil { @@ -192,7 +251,7 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain whereQuery, _, err := whereQueryBuilder.ToSql() if err != nil { - return nil, errors.Wrap(err, "error building wherequery") + return nil, errors.Wrap(err, "error building where query") } subQueryBuilder := repo.db.squirrel. @@ -230,8 +289,49 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain } queryBuilder := repo.db.squirrel. - Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "i.id", "i.name", "i.identifier_external", "r.filter", "r.protocol", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.torrent_name", "r.size", "r.category", "r.season", "r.episode", "r.year", "r.resolution", "r.source", "r.codec", "r.container", "r.release_group", "r.timestamp", - "ras.id", "ras.status", "ras.action", "ras.action_id", "ras.type", "ras.client", "ras.filter", "ras.filter_id", "ras.release_id", "ras.rejections", "ras.timestamp"). + Select( + "r.id", + "r.filter_status", + "r.rejections", + "r.indexer", + "i.id", + "i.name", + "i.identifier_external", + "r.filter", + "r.protocol", + "r.announce_type", + "r.info_url", + "r.download_url", + "r.title", + "r.sub_title", + "r.torrent_name", + "r.normalized_hash", + "r.size", + "r.category", + "r.season", + "r.episode", + "r.year", + "r.resolution", + "r.source", + "r.codec", + "r.container", + "r.hdr", + "r.audio", + "r.audio_channels", + "r.release_group", + "r.region", + "r.language", + "r.edition", + "r.cut", + "r.hybrid", + "r.proper", + "r.repack", + "r.website", + "r.media_processing", + "r.type", + "r.timestamp", + "ras.id", "ras.status", "ras.action", "ras.action_id", "ras.type", "ras.client", "ras.filter", "ras.filter_id", "ras.release_id", "ras.rejections", "ras.timestamp", + ). Column(sq.Alias(countQuery, "page_total")). From("release r"). OrderBy("r.id DESC"). @@ -267,7 +367,7 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain var rls domain.Release var ras domain.ReleaseActionStatus - var rlsIndexer, rlsIndexerName, rlsIndexerExternalName, rlsFilter, rlsAnnounceType, infoUrl, downloadUrl, codec sql.NullString + var rlsIndexer, rlsIndexerName, rlsIndexerExternalName, rlsFilter, rlsAnnounceType, infoUrl, downloadUrl, subTitle, normalizedHash, codec, hdr, rlsType, audioStr, languageStr, editionStr, cutStr, website sql.NullString var rlsIndexerID sql.NullInt64 var rasId, rasFilterId, rasReleaseId, rasActionId sql.NullInt64 @@ -275,7 +375,49 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain var rasRejections []sql.NullString var rasTimestamp sql.NullTime - if err := rows.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &rlsIndexer, &rlsIndexerID, &rlsIndexerName, &rlsIndexerExternalName, &rlsFilter, &rls.Protocol, &rlsAnnounceType, &infoUrl, &downloadUrl, &rls.Title, &rls.TorrentName, &rls.Size, &rls.Category, &rls.Season, &rls.Episode, &rls.Year, &rls.Resolution, &rls.Source, &codec, &rls.Container, &rls.Group, &rls.Timestamp, &rasId, &rasStatus, &rasAction, &rasActionId, &rasType, &rasClient, &rasFilter, &rasFilterId, &rasReleaseId, pq.Array(&rasRejections), &rasTimestamp, &resp.TotalCount); err != nil { + if err := rows.Scan( + &rls.ID, + &rls.FilterStatus, + pq.Array(&rls.Rejections), + &rlsIndexer, + &rlsIndexerID, + &rlsIndexerName, + &rlsIndexerExternalName, + &rlsFilter, + &rls.Protocol, + &rlsAnnounceType, + &infoUrl, + &downloadUrl, + &rls.Title, + &subTitle, + &rls.TorrentName, + &normalizedHash, + &rls.Size, + &rls.Category, + &rls.Season, + &rls.Episode, + &rls.Year, + &rls.Resolution, + &rls.Source, + &codec, + &rls.Container, + &hdr, + &audioStr, + &rls.AudioChannels, + &rls.Group, + &rls.Region, + &languageStr, + &editionStr, + &cutStr, + &rls.Hybrid, + &rls.Proper, + &rls.Repack, + &website, + &rls.MediaProcessing, + &rlsType, + &rls.Timestamp, + &rasId, &rasStatus, &rasAction, &rasActionId, &rasType, &rasClient, &rasFilter, &rasFilterId, &rasReleaseId, pq.Array(&rasRejections), &rasTimestamp, &resp.TotalCount, + ); err != nil { return resp, errors.Wrap(err, "error scanning row") } @@ -324,7 +466,19 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain rls.ActionStatus = make([]domain.ReleaseActionStatus, 0) rls.InfoURL = infoUrl.String rls.DownloadURL = downloadUrl.String + rls.SubTitle = subTitle.String + rls.NormalizedHash = normalizedHash.String rls.Codec = strings.Split(codec.String, ",") + rls.HDR = strings.Split(hdr.String, ",") + rls.Audio = strings.Split(audioStr.String, ",") + rls.Language = strings.Split(languageStr.String, ",") + rls.Edition = strings.Split(editionStr.String, ",") + rls.Cut = strings.Split(cutStr.String, ",") + rls.Website = website.String + //rls.Type = rlsType.String + if rlsType.Valid { + rls.ParseType(rlsType.String) + } // only add ActionStatus if it's not empty if ras.ID > 0 { @@ -342,6 +496,66 @@ func (repo *ReleaseRepo) findReleases(ctx context.Context, tx *Tx, params domain return resp, nil } +func (repo *ReleaseRepo) FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) { + queryBuilder := repo.db.squirrel. + Select( + "id", + "name", + "protocol", + "release_name", + "hash", + "title", + "sub_title", + "year", + "month", + "day", + "source", + "resolution", + "codec", + "container", + "dynamic_range", + "audio", + "release_group", + "season", + "episode", + "website", + "proper", + "repack", + ). + From("release_profile_duplicate") + + query, args, err := queryBuilder.ToSql() + if err != nil { + return nil, errors.Wrap(err, "error building query") + } + + rows, err := repo.db.handler.QueryContext(ctx, query, args...) + if err != nil { + return nil, errors.Wrap(err, "error executing query") + } + + defer rows.Close() + + if err := rows.Err(); err != nil { + return nil, errors.Wrap(err, "error rows FindDuplicateReleaseProfiles") + } + + res := make([]*domain.DuplicateReleaseProfile, 0) + + for rows.Next() { + var p domain.DuplicateReleaseProfile + + err := rows.Scan(&p.ID, &p.Name, &p.Protocol, &p.ReleaseName, &p.Hash, &p.Title, &p.SubTitle, &p.Year, &p.Month, &p.Day, &p.Source, &p.Resolution, &p.Codec, &p.Container, &p.DynamicRange, &p.Audio, &p.Group, &p.Season, &p.Episode, &p.Website, &p.Proper, &p.Repack) + if err != nil { + return nil, errors.Wrap(err, "error scanning row") + } + + res = append(res, &p) + } + + return res, nil +} + func (repo *ReleaseRepo) GetIndexerOptions(ctx context.Context) ([]string, error) { query := `SELECT DISTINCT indexer FROM "release" UNION SELECT DISTINCT identifier indexer FROM indexer;` @@ -420,7 +634,7 @@ func (repo *ReleaseRepo) GetActionStatusByReleaseID(ctx context.Context, release func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest) (*domain.Release, error) { queryBuilder := repo.db.squirrel. - Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "r.filter", "r.filter_id", "r.protocol", "r.implementation", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.torrent_name", "r.category", "r.size", "r.group_id", "r.torrent_id", "r.uploader", "r.timestamp"). + Select("r.id", "r.filter_status", "r.rejections", "r.indexer", "r.filter", "r.filter_id", "r.protocol", "r.implementation", "r.announce_type", "r.info_url", "r.download_url", "r.title", "r.sub_title", "r.torrent_name", "r.category", "r.size", "r.group_id", "r.torrent_id", "r.uploader", "r.timestamp"). From("release r"). OrderBy("r.id DESC"). Where(sq.Eq{"r.id": req.Id}) @@ -439,10 +653,10 @@ func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest) var rls domain.Release - var indexerName, filterName, announceType, infoUrl, downloadUrl, groupId, torrentId, category, uploader sql.NullString + var indexerName, filterName, announceType, infoUrl, downloadUrl, subTitle, groupId, torrentId, category, uploader sql.NullString var filterId sql.NullInt64 - if err := row.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &indexerName, &filterName, &filterId, &rls.Protocol, &rls.Implementation, &announceType, &infoUrl, &downloadUrl, &rls.Title, &rls.TorrentName, &category, &rls.Size, &groupId, &torrentId, &uploader, &rls.Timestamp); err != nil { + if err := row.Scan(&rls.ID, &rls.FilterStatus, pq.Array(&rls.Rejections), &indexerName, &filterName, &filterId, &rls.Protocol, &rls.Implementation, &announceType, &infoUrl, &downloadUrl, &rls.Title, &subTitle, &rls.TorrentName, &category, &rls.Size, &groupId, &torrentId, &uploader, &rls.Timestamp); err != nil { if errors.Is(err, sql.ErrNoRows) { return nil, domain.ErrRecordNotFound } @@ -457,6 +671,7 @@ func (repo *ReleaseRepo) Get(ctx context.Context, req *domain.GetReleaseRequest) rls.AnnounceType = domain.AnnounceType(announceType.String) rls.InfoURL = infoUrl.String rls.DownloadURL = downloadUrl.String + rls.SubTitle = subTitle.String rls.Category = category.String rls.GroupID = groupId.String rls.TorrentID = torrentId.String @@ -670,6 +885,31 @@ func (repo *ReleaseRepo) Delete(ctx context.Context, req *domain.DeleteReleaseRe return nil } +func (repo *ReleaseRepo) DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error { + qb := repo.db.squirrel.Delete("release_profile_duplicate").Where(sq.Eq{"id": id}) + + query, args, err := qb.ToSql() + if err != nil { + return errors.Wrap(err, "error building SQL query") + } + + _, err = repo.db.handler.ExecContext(ctx, query, args...) + if err != nil { + return errors.Wrap(err, "error executing delete query") + } + + //deletedRows, err := result.RowsAffected() + //if err != nil { + // return errors.Wrap(err, "error fetching rows affected") + //} + // + //repo.log.Debug().Msgf("deleted %d rows from release table", deletedRows) + + repo.log.Debug().Msgf("deleted duplicate release profile: %d", id) + + return nil +} + func (repo *ReleaseRepo) CheckSmartEpisodeCanDownload(ctx context.Context, p *domain.SmartEpisodeParams) (bool, error) { queryBuilder := repo.db.squirrel. Select("COUNT(*)"). @@ -793,3 +1033,200 @@ func (repo *ReleaseRepo) UpdateBaseURL(ctx context.Context, indexer string, oldB return nil } + +func (repo *ReleaseRepo) CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error) { + queryBuilder := repo.db.squirrel. + Select("r.id, r.torrent_name, r.normalized_hash, r.title, ras.action, ras.status"). + From("release r"). + LeftJoin("release_action_status ras ON r.id = ras.release_id"). + Where("ras.status = 'PUSH_APPROVED'") + + if profile.ReleaseName && profile.Hash { + //queryBuilder = queryBuilder.Where(repo.db.ILike("r.torrent_name", release.TorrentName)) + queryBuilder = queryBuilder.Where(sq.Eq{"r.normalized_hash": release.NormalizedHash}) + } else { + if profile.Title { + queryBuilder = queryBuilder.Where(repo.db.ILike("r.title", release.Title)) + } + + if profile.SubTitle { + queryBuilder = queryBuilder.Where(repo.db.ILike("r.sub_title", release.SubTitle)) + } + + if profile.ReleaseName && profile.Hash { + //queryBuilder = queryBuilder.Where(repo.db.ILike("r.torrent_name", release.TorrentName)) + queryBuilder = queryBuilder.Where(sq.Eq{"r.normalized_hash": release.NormalizedHash}) + } + + if profile.Year { + queryBuilder = queryBuilder.Where(sq.Eq{"r.year": release.Year}) + } + + if profile.Month { + queryBuilder = queryBuilder.Where(sq.Eq{"r.month": release.Month}) + } + + if profile.Day { + queryBuilder = queryBuilder.Where(sq.Eq{"r.day": release.Day}) + } + + if profile.Source { + queryBuilder = queryBuilder.Where(sq.Eq{"r.source": release.Source}) + } + + if profile.Container { + queryBuilder = queryBuilder.Where(sq.Eq{"r.container": release.Container}) + } + + if profile.Edition { + //queryBuilder = queryBuilder.Where(sq.Eq{"r.cut": release.Cut}) + if len(release.Cut) > 1 { + var and sq.And + for _, cut := range release.Cut { + //and = append(and, sq.Eq{"r.cut": "%" + cut + "%"}) + and = append(and, repo.db.ILike("r.cut", "%"+cut+"%")) + } + queryBuilder = queryBuilder.Where(and) + } else if len(release.Cut) == 1 { + queryBuilder = queryBuilder.Where(repo.db.ILike("r.cut", "%"+release.Cut[0]+"%")) + } + + //queryBuilder = queryBuilder.Where(sq.Eq{"r.edition": release.Edition}) + if len(release.Edition) > 1 { + var and sq.And + for _, edition := range release.Edition { + and = append(and, repo.db.ILike("r.edition", "%"+edition+"%")) + } + queryBuilder = queryBuilder.Where(and) + } else if len(release.Edition) == 1 { + queryBuilder = queryBuilder.Where(repo.db.ILike("r.edition", "%"+release.Edition[0]+"%")) + } + } + + // video features (hybrid, remux) + if release.IsTypeVideo() { + queryBuilder = queryBuilder.Where(sq.Eq{"r.hybrid": release.Hybrid}) + queryBuilder = queryBuilder.Where(sq.Eq{"r.media_processing": release.MediaProcessing}) + } + + if profile.Language { + queryBuilder = queryBuilder.Where(sq.Eq{"r.region": release.Region}) + + if len(release.Language) > 0 { + var and sq.And + for _, lang := range release.Language { + and = append(and, repo.db.ILike("r.language", "%"+lang+"%")) + } + + queryBuilder = queryBuilder.Where(and) + } else { + queryBuilder = queryBuilder.Where(sq.Eq{"r.language": ""}) + } + } + + if profile.Codec { + if len(release.Codec) > 1 { + var and sq.And + for _, codec := range release.Codec { + and = append(and, repo.db.ILike("r.codec", "%"+codec+"%")) + } + queryBuilder = queryBuilder.Where(and) + } else { + // FIXME this does an IN (arg) + queryBuilder = queryBuilder.Where(sq.Eq{"r.codec": release.Codec}) + } + } + + if profile.Resolution { + queryBuilder = queryBuilder.Where(sq.Eq{"r.resolution": release.Resolution}) + } + + if profile.DynamicRange { + //if len(release.HDR) > 1 { + // var and sq.And + // for _, hdr := range release.HDR { + // and = append(and, repo.db.ILike("r.hdr", "%"+hdr+"%")) + // } + // queryBuilder = queryBuilder.Where(and) + //} else { + // queryBuilder = queryBuilder.Where(sq.Eq{"r.hdr": release.HDR}) + //} + queryBuilder = queryBuilder.Where(sq.Eq{"r.hdr": strings.Join(release.HDR, ",")}) + } + + if profile.Audio { + queryBuilder = queryBuilder.Where(sq.Eq{"r.audio": strings.Join(release.Audio, ",")}) + queryBuilder = queryBuilder.Where(sq.Eq{"r.audio_channels": release.AudioChannels}) + } + + if profile.Group { + queryBuilder = queryBuilder.Where(repo.db.ILike("r.release_group", release.Group)) + } + + if profile.Season { + queryBuilder = queryBuilder.Where(sq.Eq{"r.season": release.Season}) + } + + if profile.Episode { + queryBuilder = queryBuilder.Where(sq.Eq{"r.episode": release.Episode}) + } + + if profile.Website { + queryBuilder = queryBuilder.Where(sq.Eq{"r.website": release.Website}) + } + + if profile.Proper { + queryBuilder = queryBuilder.Where(sq.Eq{"r.proper": release.Proper}) + } + + if profile.Repack { + queryBuilder = queryBuilder.Where(sq.And{ + sq.Eq{"r.repack": release.Repack}, + repo.db.ILike("r.release_group", release.Group), + }) + } + } + + query, args, err := queryBuilder.ToSql() + if err != nil { + return false, errors.Wrap(err, "error building query") + } + + repo.log.Trace().Str("database", "release.FindDuplicateReleases").Msgf("query: %q, args: %q", query, args) + + rows, err := repo.db.handler.QueryContext(ctx, query, args...) + if err != nil { + return false, err + } + + if err := rows.Err(); err != nil { + return false, errors.Wrap(err, "error rows CheckIsDuplicateRelease") + } + + type result struct { + id int + release string + hash string + title string + action string + status string + } + + var res []result + + for rows.Next() { + r := result{} + if err := rows.Scan(&r.id, &r.release, &r.hash, &r.title, &r.action, &r.status); err != nil { + return false, errors.Wrap(err, "error scan CheckIsDuplicateRelease") + } + res = append(res, r) + } + + repo.log.Trace().Str("database", "release.FindDuplicateReleases").Msgf("found duplicate releases: %+v", res) + + if len(res) == 0 { + return false, nil + } + + return true, nil +} diff --git a/internal/database/release_test.go b/internal/database/release_test.go index 779b912..a8792bb 100644 --- a/internal/database/release_test.go +++ b/internal/database/release_test.go @@ -13,6 +13,7 @@ import ( "github.com/autobrr/autobrr/internal/domain" + "github.com/moistari/rls" "github.com/stretchr/testify/assert" ) @@ -49,12 +50,13 @@ func getMockRelease() *domain.Release { Proper: true, Repack: false, Website: "https://example.com", - Type: "Movie", + Type: rls.Movie, Origin: "P2P", Tags: []string{"Action", "Adventure"}, Uploader: "john_doe", PreTime: "10m", FilterID: 1, + Other: []string{}, } } @@ -108,11 +110,11 @@ func TestReleaseRepo_Store(t *testing.T) { // Execute err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -123,7 +125,7 @@ func TestReleaseRepo_Store(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -164,11 +166,11 @@ func TestReleaseRepo_StoreReleaseActionStatus(t *testing.T) { // Execute err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -179,7 +181,7 @@ func TestReleaseRepo_StoreReleaseActionStatus(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -328,11 +330,11 @@ func TestReleaseRepo_GetIndexerOptions(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -347,7 +349,7 @@ func TestReleaseRepo_GetIndexerOptions(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -387,11 +389,11 @@ func TestReleaseRepo_GetActionStatusByReleaseID(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -407,7 +409,7 @@ func TestReleaseRepo_GetActionStatusByReleaseID(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -447,11 +449,11 @@ func TestReleaseRepo_Get(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -467,7 +469,7 @@ func TestReleaseRepo_Get(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -507,11 +509,11 @@ func TestReleaseRepo_Stats(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -527,7 +529,7 @@ func TestReleaseRepo_Stats(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -567,11 +569,11 @@ func TestReleaseRepo_Delete(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -584,7 +586,7 @@ func TestReleaseRepo_Delete(t *testing.T) { assert.NoError(t, err) // Cleanup - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) @@ -624,11 +626,11 @@ func TestReleaseRepo_CheckSmartEpisodeCanDownloadShow(t *testing.T) { err = repo.Store(context.Background(), mockData) assert.NoError(t, err) - createdAction, err := actionRepo.Store(context.Background(), actionMockData) + err = actionRepo.Store(context.Background(), actionMockData) assert.NoError(t, err) releaseActionMockData.ReleaseID = mockData.ID - releaseActionMockData.ActionID = int64(createdAction.ID) + releaseActionMockData.ActionID = int64(actionMockData.ID) releaseActionMockData.FilterID = int64(createdFilters[0].ID) err = repo.StoreReleaseActionStatus(context.Background(), releaseActionMockData) @@ -652,9 +654,724 @@ func TestReleaseRepo_CheckSmartEpisodeCanDownloadShow(t *testing.T) { // Cleanup _ = repo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) - _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: createdAction.ID}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMockData.ID}) _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) _ = downloadClientRepo.Delete(context.Background(), mock.ID) }) } } + +func getMockDuplicateReleaseProfileTV() *domain.DuplicateReleaseProfile { + return &domain.DuplicateReleaseProfile{ + ID: 0, + Name: "TV", + Protocol: false, + ReleaseName: false, + Hash: false, + Title: true, + SubTitle: false, + Year: false, + Month: false, + Day: false, + Source: false, + Resolution: false, + Codec: false, + Container: false, + DynamicRange: false, + Audio: false, + Group: false, + Season: true, + Episode: true, + Website: false, + Proper: false, + Repack: false, + Edition: false, + Language: false, + } +} + +func getMockDuplicateReleaseProfileTVDaily() *domain.DuplicateReleaseProfile { + return &domain.DuplicateReleaseProfile{ + ID: 0, + Name: "TV", + Protocol: false, + ReleaseName: false, + Hash: false, + Title: true, + SubTitle: false, + Year: true, + Month: true, + Day: true, + Source: false, + Resolution: false, + Codec: false, + Container: false, + DynamicRange: false, + Audio: false, + Group: false, + Season: false, + Episode: false, + Website: false, + Proper: false, + Repack: false, + Edition: false, + Language: false, + } +} + +func getMockFilterDuplicates() *domain.Filter { + return &domain.Filter{ + Name: "New Filter", + Enabled: true, + CreatedAt: time.Now(), + UpdatedAt: time.Now(), + MinSize: "10mb", + MaxSize: "20mb", + Delay: 60, + Priority: 1, + MaxDownloads: 100, + MaxDownloadsUnit: domain.FilterMaxDownloadsHour, + MatchReleases: "BRRip", + ExceptReleases: "BRRip", + UseRegex: false, + MatchReleaseGroups: "AMIABLE", + ExceptReleaseGroups: "NTb", + Scene: false, + Origins: nil, + ExceptOrigins: nil, + Bonus: nil, + Freeleech: false, + FreeleechPercent: "100%", + SmartEpisode: false, + Shows: "Is It Wrong to Try to Pick Up Girls in a Dungeon?", + Seasons: "4", + Episodes: "500", + Resolutions: []string{"1080p"}, + Codecs: []string{"x264"}, + Sources: []string{"BluRay"}, + Containers: []string{"mkv"}, + MatchHDR: []string{"HDR10"}, + ExceptHDR: []string{"HDR10"}, + MatchOther: []string{"Atmos"}, + ExceptOther: []string{"Atmos"}, + Years: "2023", + Months: "", + Days: "", + Artists: "", + Albums: "", + MatchReleaseTypes: []string{"Remux"}, + ExceptReleaseTypes: "Remux", + Formats: []string{"FLAC"}, + Quality: []string{"Lossless"}, + Media: []string{"CD"}, + PerfectFlac: true, + Cue: true, + Log: true, + LogScore: 100, + MatchCategories: "Anime", + ExceptCategories: "Anime", + MatchUploaders: "SubsPlease", + ExceptUploaders: "SubsPlease", + MatchLanguage: []string{"English", "Japanese"}, + ExceptLanguage: []string{"English", "Japanese"}, + Tags: "Anime, x264", + ExceptTags: "Anime, x264", + TagsAny: "Anime, x264", + ExceptTagsAny: "Anime, x264", + TagsMatchLogic: "AND", + ExceptTagsMatchLogic: "AND", + MatchReleaseTags: "Anime, x264", + ExceptReleaseTags: "Anime, x264", + UseRegexReleaseTags: true, + MatchDescription: "Anime, x264", + ExceptDescription: "Anime, x264", + UseRegexDescription: true, + } +} + +func TestReleaseRepo_CheckIsDuplicateRelease(t *testing.T) { + for dbType, db := range testDBs { + log := setupLoggerForTest() + + downloadClientRepo := NewDownloadClientRepo(log, db) + filterRepo := NewFilterRepo(log, db) + actionRepo := NewActionRepo(log, db, downloadClientRepo) + releaseRepo := NewReleaseRepo(log, db) + + // reset + //db.handler.Exec("DELETE FROM release") + //db.handler.Exec("DELETE FROM action") + //db.handler.Exec("DELETE FROM release_action_status") + + mockIndexer := domain.IndexerMinimal{ID: 0, Name: "Mock", Identifier: "mock", IdentifierExternal: "Mock"} + actionMock := &domain.Action{Name: "Test", Type: domain.ActionTypeTest, Enabled: true} + filterMock := getMockFilterDuplicates() + + // Setup + err := filterRepo.Store(context.Background(), filterMock) + assert.NoError(t, err) + + createdFilters, err := filterRepo.ListFilters(context.Background()) + assert.NoError(t, err) + assert.NotNil(t, createdFilters) + + actionMock.FilterID = filterMock.ID + + err = actionRepo.Store(context.Background(), actionMock) + assert.NoError(t, err) + + type fields struct { + releaseTitles []string + releaseTitle string + profile *domain.DuplicateReleaseProfile + } + + tests := []struct { + name string + fields fields + isDuplicate bool + }{ + { + name: "1", + fields: fields{ + releaseTitles: []string{ + "Inkheart 2008 BluRay 1080p DD5.1 x264-BADGROUP", + }, + releaseTitle: "Inkheart 2008 BluRay 1080p DD5.1 x264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "2", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Source: true, Resolution: true}, + }, + isDuplicate: true, + }, + { + name: "3", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true}, + }, + isDuplicate: true, + }, + { + name: "4", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.WEB.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "5", + fields: fields{ + releaseTitles: []string{ + "That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "6", + fields: fields{ + releaseTitles: []string{ + "That.Tv.Show.2023.S01E01.BluRay.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Tv.Show.2023.S01E02.BluRay.2160p.x265.DTS-HD-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "7", + fields: fields{ + releaseTitles: []string{ + "That.Tv.Show.2023.S01.BluRay.2160p.x265.DTS-HD-GROUP", + }, + releaseTitle: "That.Tv.Show.2023.S01.BluRay.2160p.x265.DTS-HD-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "8", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "9", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.HULU.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 SDR H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "10", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "11", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + "The Best Show 2020 S04E10 1080p amzn web-dl ddp 5.1 hdr dv h.264-group", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true}, + }, + isDuplicate: true, + }, + { + name: "12", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "13", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "14", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Year: true, Season: true, Episode: true, Source: true, Codec: true, Resolution: true, Website: true, DynamicRange: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "15", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true}, + }, + isDuplicate: true, + }, + { + name: "16", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E11 Episode Title 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true}, + }, + isDuplicate: false, + }, + { + name: "17", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.Episode.Title.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 Episode Title REPACK 1080p AMZN WEB-DL DDP 5.1 HDR DV H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, SubTitle: true, Season: true, Episode: true, DynamicRange: true}, + }, + isDuplicate: true, + }, + { + name: "18", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The.Best.Show.2020.S04E10.Episode.Title.REPACK.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 Episode Title REPACK 1080p AMZN WEB-DL DDP 5.1 DV H.264-OTHERGROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Repack: true}, + }, + isDuplicate: false, // not a match because REPACK checks for the same group + }, + { + name: "19", + fields: fields{ + releaseTitles: []string{ + "The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1", + }, + releaseTitle: "The Daily Show 2024-09-21.Other.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true}, + }, + isDuplicate: true, + }, + { + name: "20", + fields: fields{ + releaseTitles: []string{ + "The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1", + }, + releaseTitle: "The Daily Show 2024-09-21 Other Guest 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true, SubTitle: true}, + }, + isDuplicate: false, + }, + { + name: "21", + fields: fields{ + releaseTitles: []string{ + "The Daily Show 2024-09-21 1080p HULU WEB-DL DDP 5.1 SDR H.264-GROUP", + "The Daily Show 2024-09-21.1080p.AMZN.WEB-DL.DDP.5.1.SDR.H.264-GROUP", + "The Daily Show 2024-09-21.Guest.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP1", + }, + releaseTitle: "The Daily Show 2024-09-22 Other Guest 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP1", + profile: &domain.DuplicateReleaseProfile{Title: true, Season: true, Episode: true, Year: true, Month: true, Day: true, SubTitle: true}, + }, + isDuplicate: false, + }, + { + name: "22", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP", + }, + releaseTitle: "That.Movie.2023.2160p.BluRay.DD.2.0.x265-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "23", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP", + }, + releaseTitle: "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "24", + fields: fields{ + releaseTitles: []string{ + "That.Movie.2023.BluRay.2160p.x265.DTS-HD-GROUP", + "That.Movie.2023.BluRay.720p.x265.DTS-HD-GROUP", + "That.Movie.2023.2160p.BluRay.DD.5.1.x265-GROUP", + }, + releaseTitle: "That.Movie.2023.2160p.BluRay.AC3.5.1.x265-GROUP", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "25", + fields: fields{ + releaseTitles: []string{ + //"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "26", + fields: fields{ + releaseTitles: []string{ + //"Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "27", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Edition: false, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "28", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX-FraMeSToR", + "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 Collectors Edition UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Edition: true, Source: true, Codec: true, Resolution: true, Audio: true, Group: true}, + }, + isDuplicate: true, + }, + { + name: "29", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR DV HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "30", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR DV HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "31", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 DV HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "32", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR10 HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HDR HEVC REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true}, + }, + isDuplicate: false, + }, + { + name: "33", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true}, + }, + isDuplicate: false, + }, + { + name: "34", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true}, + }, + isDuplicate: true, + }, + { + name: "35", + fields: fields{ + releaseTitles: []string{ + "Despicable Me 4 2024 FRENCH UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + "Despicable Me 4 2024 GERMAN UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + }, + releaseTitle: "Despicable Me 4 2024 UHD BluRay 2160p TrueHD Atmos 7.1 HEVC DV REMUX Hybrid-FraMeSToR", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, DynamicRange: true, Audio: true, Group: true, Language: true}, + }, + isDuplicate: false, + }, + { + name: "36", + fields: fields{ + releaseTitles: []string{ + "Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT", + }, + releaseTitle: "Road House 1989 1080p Blu-ray AVC LPCM 2.0-MONUMENT", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true}, + }, + isDuplicate: false, + }, + { + name: "37", + fields: fields{ + releaseTitles: []string{ + "Road House 1989 1080p ITA Blu-ray AVC LPCM 2.0-MONUMENT", + "Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT", + }, + releaseTitle: "Road House 1989 1080p NOR Blu-ray AVC LPCM 2.0-MONUMENT", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true}, + }, + isDuplicate: false, + }, + { + name: "38", + fields: fields{ + releaseTitles: []string{ + "Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT", + }, + releaseTitle: "Road House 1989 1080p GER Blu-ray AVC LPCM 2.0-MONUMENT", + profile: &domain.DuplicateReleaseProfile{Title: true, Year: true, Source: true, Codec: true, Resolution: true, Group: true, Language: true}, + }, + isDuplicate: true, + }, + { + name: "39", + fields: fields{ + releaseTitles: []string{ + "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.H.264-GROUP", + "The.Best.Show.2020.S04E10.1080p.AMZN.WEB-DL.DDP.5.1.HDR.DV.H.264-GROUP", + }, + releaseTitle: "The Best Show 2020 S04E10 1080p AMZN WEB-DL DDP 5.1 H.264-GROUP", + profile: &domain.DuplicateReleaseProfile{ReleaseName: true}, + }, + isDuplicate: true, + }, + } + + for _, tt := range tests { + t.Run(fmt.Sprintf("Check_Is_Duplicate_Release %s [%s]", tt.name, dbType), func(t *testing.T) { + ctx := context.Background() + + // Setup + for _, rel := range tt.fields.releaseTitles { + mockRel := domain.NewRelease(mockIndexer) + mockRel.ParseString(rel) + + mockRel.FilterID = filterMock.ID + + err = releaseRepo.Store(ctx, mockRel) + assert.NoError(t, err) + + ras := &domain.ReleaseActionStatus{ + ID: 0, + Status: domain.ReleasePushStatusApproved, + Action: "test", + ActionID: int64(actionMock.ID), + Type: domain.ActionTypeTest, + Client: "", + Filter: "Test filter", + FilterID: int64(filterMock.ID), + Rejections: []string{}, + ReleaseID: mockRel.ID, + Timestamp: time.Now(), + } + + err = releaseRepo.StoreReleaseActionStatus(ctx, ras) + assert.NoError(t, err) + } + + releases, err := releaseRepo.Find(ctx, domain.ReleaseQueryParams{}) + assert.NoError(t, err) + assert.Len(t, releases.Data, len(tt.fields.releaseTitles)) + + compareRel := domain.NewRelease(mockIndexer) + compareRel.ParseString(tt.fields.releaseTitle) + + // Execute + isDuplicate, err := releaseRepo.CheckIsDuplicateRelease(ctx, tt.fields.profile, compareRel) + + // Verify + assert.NoError(t, err) + assert.Equal(t, tt.isDuplicate, isDuplicate) + + // Cleanup + _ = releaseRepo.Delete(ctx, &domain.DeleteReleaseRequest{OlderThan: 0}) + }) + } + + // Cleanup + //_ = releaseRepo.Delete(context.Background(), &domain.DeleteReleaseRequest{OlderThan: 0}) + _ = actionRepo.Delete(context.Background(), &domain.DeleteActionRequest{ActionId: actionMock.ID}) + _ = filterRepo.Delete(context.Background(), createdFilters[0].ID) + } +} diff --git a/internal/database/sqlite_migrate.go b/internal/database/sqlite_migrate.go index 5ab998f..cb9b6c1 100644 --- a/internal/database/sqlite_migrate.go +++ b/internal/database/sqlite_migrate.go @@ -88,9 +88,42 @@ CREATE TABLE irc_channel UNIQUE (network_id, name) ); +CREATE TABLE release_profile_duplicate +( + id INTEGER PRIMARY KEY AUTOINCREMENT, + name TEXT NOT NULL, + protocol BOOLEAN DEFAULT FALSE, + release_name BOOLEAN DEFAULT FALSE, + hash BOOLEAN DEFAULT FALSE, + title BOOLEAN DEFAULT FALSE, + sub_title BOOLEAN DEFAULT FALSE, + year BOOLEAN DEFAULT FALSE, + month BOOLEAN DEFAULT FALSE, + day BOOLEAN DEFAULT FALSE, + source BOOLEAN DEFAULT FALSE, + resolution BOOLEAN DEFAULT FALSE, + codec BOOLEAN DEFAULT FALSE, + container BOOLEAN DEFAULT FALSE, + dynamic_range BOOLEAN DEFAULT FALSE, + audio BOOLEAN DEFAULT FALSE, + release_group BOOLEAN DEFAULT FALSE, + season BOOLEAN DEFAULT FALSE, + episode BOOLEAN DEFAULT FALSE, + website BOOLEAN DEFAULT FALSE, + proper BOOLEAN DEFAULT FALSE, + repack BOOLEAN DEFAULT FALSE, + edition BOOLEAN DEFAULT FALSE, + language BOOLEAN DEFAULT FALSE +); + +INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language) +VALUES (1, 'Exact release', 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0), + (2, 'Movie', 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0), + (3, 'TV', 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0); + CREATE TABLE filter ( - id INTEGER PRIMARY KEY, + id INTEGER PRIMARY KEY AUTOINCREMENT, enabled BOOLEAN, name TEXT NOT NULL, min_size TEXT, @@ -159,7 +192,9 @@ CREATE TABLE filter min_seeders INTEGER DEFAULT 0, max_seeders INTEGER DEFAULT 0, min_leechers INTEGER DEFAULT 0, - max_leechers INTEGER DEFAULT 0 + max_leechers INTEGER DEFAULT 0, + release_profile_duplicate_id INTEGER, + FOREIGN KEY (release_profile_duplicate_id) REFERENCES release_profile_duplicate(id) ON DELETE SET NULL ); CREATE INDEX filter_enabled_index @@ -273,8 +308,10 @@ CREATE TABLE "release" group_id TEXT, torrent_id TEXT, torrent_name TEXT, + normalized_hash TEXT, size INTEGER, title TEXT, + sub_title TEXT, category TEXT, season INTEGER, episode INTEGER, @@ -286,15 +323,24 @@ CREATE TABLE "release" codec TEXT, container TEXT, hdr TEXT, + audio TEXT, + audio_channels TEXT, release_group TEXT, + region TEXT, + language TEXT, + edition TEXT, + cut TEXT, + hybrid BOOLEAN, proper BOOLEAN, repack BOOLEAN, website TEXT, + media_processing TEXT, type TEXT, origin TEXT, tags TEXT [] DEFAULT '{}' NOT NULL, uploader TEXT, pre_time TEXT, + other TEXT [] DEFAULT '{}' NOT NULL, filter_id INTEGER REFERENCES filter ON DELETE SET NULL @@ -312,6 +358,81 @@ CREATE INDEX release_timestamp_index CREATE INDEX release_torrent_name_index ON "release" (torrent_name); +CREATE INDEX release_normalized_hash_index + ON "release" (normalized_hash); + +CREATE INDEX release_title_index + ON "release" (title); + +CREATE INDEX release_sub_title_index + ON "release" (sub_title); + +CREATE INDEX release_season_index + ON "release" (season); + +CREATE INDEX release_episode_index + ON "release" (episode); + +CREATE INDEX release_year_index + ON "release" (year); + +CREATE INDEX release_month_index + ON "release" (month); + +CREATE INDEX release_day_index + ON "release" (day); + +CREATE INDEX release_resolution_index + ON "release" (resolution); + +CREATE INDEX release_source_index + ON "release" (source); + +CREATE INDEX release_codec_index + ON "release" (codec); + +CREATE INDEX release_container_index + ON "release" (container); + +CREATE INDEX release_hdr_index + ON "release" (hdr); + +CREATE INDEX release_audio_index + ON "release" (audio); + +CREATE INDEX release_audio_channels_index + ON "release" (audio_channels); + +CREATE INDEX release_release_group_index + ON "release" (release_group); + +CREATE INDEX release_language_index + ON "release" (language); + +CREATE INDEX release_proper_index + ON "release" (proper); + +CREATE INDEX release_repack_index + ON "release" (repack); + +CREATE INDEX release_website_index + ON "release" (website); + +CREATE INDEX release_media_processing_index + ON "release" (media_processing); + +CREATE INDEX release_region_index + ON "release" (region); + +CREATE INDEX release_edition_index + ON "release" (edition); + +CREATE INDEX release_cut_index + ON "release" (cut); + +CREATE INDEX release_hybrid_index + ON "release" (hybrid); + CREATE TABLE release_action_status ( id INTEGER PRIMARY KEY, @@ -1716,5 +1837,152 @@ CREATE TABLE list_filter ALTER TABLE filter ADD COLUMN except_record_labels TEXT DEFAULT ''; +`, + `CREATE TABLE release_profile_duplicate +( + id INTEGER PRIMARY KEY AUTOINCREMENT, + name TEXT NOT NULL, + protocol BOOLEAN DEFAULT FALSE, + release_name BOOLEAN DEFAULT FALSE, + hash BOOLEAN DEFAULT FALSE, + title BOOLEAN DEFAULT FALSE, + sub_title BOOLEAN DEFAULT FALSE, + year BOOLEAN DEFAULT FALSE, + month BOOLEAN DEFAULT FALSE, + day BOOLEAN DEFAULT FALSE, + source BOOLEAN DEFAULT FALSE, + resolution BOOLEAN DEFAULT FALSE, + codec BOOLEAN DEFAULT FALSE, + container BOOLEAN DEFAULT FALSE, + dynamic_range BOOLEAN DEFAULT FALSE, + audio BOOLEAN DEFAULT FALSE, + release_group BOOLEAN DEFAULT FALSE, + season BOOLEAN DEFAULT FALSE, + episode BOOLEAN DEFAULT FALSE, + website BOOLEAN DEFAULT FALSE, + proper BOOLEAN DEFAULT FALSE, + repack BOOLEAN DEFAULT FALSE, + edition BOOLEAN DEFAULT FALSE, + language BOOLEAN DEFAULT FALSE +); + +INSERT INTO release_profile_duplicate (id, name, protocol, release_name, hash, title, sub_title, year, month, day, source, resolution, codec, container, dynamic_range, audio, release_group, season, episode, website, proper, repack, edition, language) +VALUES (1, 'Exact release', 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0), + (2, 'Movie', 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0), + (3, 'TV', 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0); + +ALTER TABLE filter + ADD COLUMN release_profile_duplicate_id INTEGER + CONSTRAINT filter_release_profile_duplicate_id_fk + REFERENCES release_profile_duplicate (id) + ON DELETE SET NULL; + +ALTER TABLE "release" + ADD normalized_hash TEXT; + +ALTER TABLE "release" + ADD sub_title TEXT; + +ALTER TABLE "release" + ADD audio TEXT; + +ALTER TABLE "release" + ADD audio_channels TEXT; + +ALTER TABLE "release" + ADD language TEXT; + +ALTER TABLE "release" + ADD media_processing TEXT; + +ALTER TABLE "release" + ADD edition TEXT; + +ALTER TABLE "release" + ADD cut TEXT; + +ALTER TABLE "release" + ADD hybrid TEXT; + +ALTER TABLE "release" + ADD region TEXT; + +ALTER TABLE "release" + ADD other TEXT [] DEFAULT '{}' NOT NULL; + +CREATE INDEX release_normalized_hash_index + ON "release" (normalized_hash); + +CREATE INDEX release_title_index + ON "release" (title); + +CREATE INDEX release_sub_title_index + ON "release" (sub_title); + +CREATE INDEX release_season_index + ON "release" (season); + +CREATE INDEX release_episode_index + ON "release" (episode); + +CREATE INDEX release_year_index + ON "release" (year); + +CREATE INDEX release_month_index + ON "release" (month); + +CREATE INDEX release_day_index + ON "release" (day); + +CREATE INDEX release_resolution_index + ON "release" (resolution); + +CREATE INDEX release_source_index + ON "release" (source); + +CREATE INDEX release_codec_index + ON "release" (codec); + +CREATE INDEX release_container_index + ON "release" (container); + +CREATE INDEX release_hdr_index + ON "release" (hdr); + +CREATE INDEX release_audio_index + ON "release" (audio); + +CREATE INDEX release_audio_channels_index + ON "release" (audio_channels); + +CREATE INDEX release_release_group_index + ON "release" (release_group); + +CREATE INDEX release_proper_index + ON "release" (proper); + +CREATE INDEX release_repack_index + ON "release" (repack); + +CREATE INDEX release_website_index + ON "release" (website); + +CREATE INDEX release_media_processing_index + ON "release" (media_processing); + +CREATE INDEX release_language_index + ON "release" (language); + +CREATE INDEX release_region_index + ON "release" (region); + +CREATE INDEX release_edition_index + ON "release" (edition); + +CREATE INDEX release_cut_index + ON "release" (cut); + +CREATE INDEX release_hybrid_index + ON "release" (hybrid); `, } diff --git a/internal/database/utils.go b/internal/database/utils.go index c5d6b2e..9851835 100644 --- a/internal/database/utils.go +++ b/internal/database/utils.go @@ -29,6 +29,7 @@ func toNullInt32(s int32) sql.NullInt32 { Valid: s != 0, } } + func toNullInt64(s int64) sql.NullInt64 { return sql.NullInt64{ Int64: s, diff --git a/internal/domain/action.go b/internal/domain/action.go index e85f826..6117cbe 100644 --- a/internal/domain/action.go +++ b/internal/domain/action.go @@ -11,7 +11,7 @@ import ( ) type ActionRepo interface { - Store(ctx context.Context, action Action) (*Action, error) + Store(ctx context.Context, action *Action) error StoreFilterActions(ctx context.Context, filterID int64, actions []*Action) ([]*Action, error) FindByFilterID(ctx context.Context, filterID int, active *bool, withClient bool) ([]*Action, error) List(ctx context.Context) ([]Action, error) diff --git a/internal/domain/filter.go b/internal/domain/filter.go index 8eaf605..cd6fb4b 100644 --- a/internal/domain/filter.go +++ b/internal/domain/filter.go @@ -91,88 +91,90 @@ type FilterQueryParams struct { } type Filter struct { - ID int `json:"id"` - Name string `json:"name"` - Enabled bool `json:"enabled"` - CreatedAt time.Time `json:"created_at"` - UpdatedAt time.Time `json:"updated_at"` - MinSize string `json:"min_size,omitempty"` - MaxSize string `json:"max_size,omitempty"` - Delay int `json:"delay,omitempty"` - Priority int32 `json:"priority"` - MaxDownloads int `json:"max_downloads,omitempty"` - MaxDownloadsUnit FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"` - MatchReleases string `json:"match_releases,omitempty"` - ExceptReleases string `json:"except_releases,omitempty"` - UseRegex bool `json:"use_regex,omitempty"` - MatchReleaseGroups string `json:"match_release_groups,omitempty"` - ExceptReleaseGroups string `json:"except_release_groups,omitempty"` - AnnounceTypes []string `json:"announce_types,omitempty"` - Scene bool `json:"scene,omitempty"` - Origins []string `json:"origins,omitempty"` - ExceptOrigins []string `json:"except_origins,omitempty"` - Bonus []string `json:"bonus,omitempty"` - Freeleech bool `json:"freeleech,omitempty"` - FreeleechPercent string `json:"freeleech_percent,omitempty"` - SmartEpisode bool `json:"smart_episode"` - Shows string `json:"shows,omitempty"` - Seasons string `json:"seasons,omitempty"` - Episodes string `json:"episodes,omitempty"` - Resolutions []string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p. - Codecs []string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux). - Sources []string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC - Containers []string `json:"containers,omitempty"` - MatchHDR []string `json:"match_hdr,omitempty"` - ExceptHDR []string `json:"except_hdr,omitempty"` - MatchOther []string `json:"match_other,omitempty"` - ExceptOther []string `json:"except_other,omitempty"` - Years string `json:"years,omitempty"` - Months string `json:"months,omitempty"` - Days string `json:"days,omitempty"` - Artists string `json:"artists,omitempty"` - Albums string `json:"albums,omitempty"` - MatchReleaseTypes []string `json:"match_release_types,omitempty"` // Album,Single,EP - ExceptReleaseTypes string `json:"except_release_types,omitempty"` - Formats []string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS - Quality []string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other - Media []string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other - PerfectFlac bool `json:"perfect_flac,omitempty"` - Cue bool `json:"cue,omitempty"` - Log bool `json:"log,omitempty"` - LogScore int `json:"log_score,omitempty"` - MatchCategories string `json:"match_categories,omitempty"` - ExceptCategories string `json:"except_categories,omitempty"` - MatchUploaders string `json:"match_uploaders,omitempty"` - ExceptUploaders string `json:"except_uploaders,omitempty"` - MatchRecordLabels string `json:"match_record_labels,omitempty"` - ExceptRecordLabels string `json:"except_record_labels,omitempty"` - MatchLanguage []string `json:"match_language,omitempty"` - ExceptLanguage []string `json:"except_language,omitempty"` - Tags string `json:"tags,omitempty"` - ExceptTags string `json:"except_tags,omitempty"` - TagsAny string `json:"tags_any,omitempty"` - ExceptTagsAny string `json:"except_tags_any,omitempty"` - TagsMatchLogic string `json:"tags_match_logic,omitempty"` - ExceptTagsMatchLogic string `json:"except_tags_match_logic,omitempty"` - MatchReleaseTags string `json:"match_release_tags,omitempty"` - ExceptReleaseTags string `json:"except_release_tags,omitempty"` - UseRegexReleaseTags bool `json:"use_regex_release_tags,omitempty"` - MatchDescription string `json:"match_description,omitempty"` - ExceptDescription string `json:"except_description,omitempty"` - UseRegexDescription bool `json:"use_regex_description,omitempty"` - MinSeeders int `json:"min_seeders,omitempty"` - MaxSeeders int `json:"max_seeders,omitempty"` - MinLeechers int `json:"min_leechers,omitempty"` - MaxLeechers int `json:"max_leechers,omitempty"` - ActionsCount int `json:"actions_count"` - ActionsEnabledCount int `json:"actions_enabled_count"` - IsAutoUpdated bool `json:"is_auto_updated"` - Actions []*Action `json:"actions,omitempty"` - External []FilterExternal `json:"external,omitempty"` - Indexers []Indexer `json:"indexers"` - Downloads *FilterDownloads `json:"-"` - Rejections []string `json:"-"` - RejectReasons *RejectionReasons `json:"-"` + ID int `json:"id"` + Name string `json:"name"` + Enabled bool `json:"enabled"` + CreatedAt time.Time `json:"created_at"` + UpdatedAt time.Time `json:"updated_at"` + MinSize string `json:"min_size,omitempty"` + MaxSize string `json:"max_size,omitempty"` + Delay int `json:"delay,omitempty"` + Priority int32 `json:"priority"` + MaxDownloads int `json:"max_downloads,omitempty"` + MaxDownloadsUnit FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"` + MatchReleases string `json:"match_releases,omitempty"` + ExceptReleases string `json:"except_releases,omitempty"` + UseRegex bool `json:"use_regex,omitempty"` + MatchReleaseGroups string `json:"match_release_groups,omitempty"` + ExceptReleaseGroups string `json:"except_release_groups,omitempty"` + AnnounceTypes []string `json:"announce_types,omitempty"` + Scene bool `json:"scene,omitempty"` + Origins []string `json:"origins,omitempty"` + ExceptOrigins []string `json:"except_origins,omitempty"` + Bonus []string `json:"bonus,omitempty"` + Freeleech bool `json:"freeleech,omitempty"` + FreeleechPercent string `json:"freeleech_percent,omitempty"` + SmartEpisode bool `json:"smart_episode"` + Shows string `json:"shows,omitempty"` + Seasons string `json:"seasons,omitempty"` + Episodes string `json:"episodes,omitempty"` + Resolutions []string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p. + Codecs []string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux). + Sources []string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC + Containers []string `json:"containers,omitempty"` + MatchHDR []string `json:"match_hdr,omitempty"` + ExceptHDR []string `json:"except_hdr,omitempty"` + MatchOther []string `json:"match_other,omitempty"` + ExceptOther []string `json:"except_other,omitempty"` + Years string `json:"years,omitempty"` + Months string `json:"months,omitempty"` + Days string `json:"days,omitempty"` + Artists string `json:"artists,omitempty"` + Albums string `json:"albums,omitempty"` + MatchReleaseTypes []string `json:"match_release_types,omitempty"` // Album,Single,EP + ExceptReleaseTypes string `json:"except_release_types,omitempty"` + Formats []string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS + Quality []string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other + Media []string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other + PerfectFlac bool `json:"perfect_flac,omitempty"` + Cue bool `json:"cue,omitempty"` + Log bool `json:"log,omitempty"` + LogScore int `json:"log_score,omitempty"` + MatchCategories string `json:"match_categories,omitempty"` + ExceptCategories string `json:"except_categories,omitempty"` + MatchUploaders string `json:"match_uploaders,omitempty"` + ExceptUploaders string `json:"except_uploaders,omitempty"` + MatchRecordLabels string `json:"match_record_labels,omitempty"` + ExceptRecordLabels string `json:"except_record_labels,omitempty"` + MatchLanguage []string `json:"match_language,omitempty"` + ExceptLanguage []string `json:"except_language,omitempty"` + Tags string `json:"tags,omitempty"` + ExceptTags string `json:"except_tags,omitempty"` + TagsAny string `json:"tags_any,omitempty"` + ExceptTagsAny string `json:"except_tags_any,omitempty"` + TagsMatchLogic string `json:"tags_match_logic,omitempty"` + ExceptTagsMatchLogic string `json:"except_tags_match_logic,omitempty"` + MatchReleaseTags string `json:"match_release_tags,omitempty"` + ExceptReleaseTags string `json:"except_release_tags,omitempty"` + UseRegexReleaseTags bool `json:"use_regex_release_tags,omitempty"` + MatchDescription string `json:"match_description,omitempty"` + ExceptDescription string `json:"except_description,omitempty"` + UseRegexDescription bool `json:"use_regex_description,omitempty"` + MinSeeders int `json:"min_seeders,omitempty"` + MaxSeeders int `json:"max_seeders,omitempty"` + MinLeechers int `json:"min_leechers,omitempty"` + MaxLeechers int `json:"max_leechers,omitempty"` + ActionsCount int `json:"actions_count"` + ActionsEnabledCount int `json:"actions_enabled_count"` + IsAutoUpdated bool `json:"is_auto_updated"` + Actions []*Action `json:"actions,omitempty"` + External []FilterExternal `json:"external,omitempty"` + Indexers []Indexer `json:"indexers"` + ReleaseProfileDuplicateID int64 `json:"release_profile_duplicate_id,omitempty"` + DuplicateHandling *DuplicateReleaseProfile `json:"release_profile_duplicate"` + Downloads *FilterDownloads `json:"-"` + Rejections []string `json:"-"` + RejectReasons *RejectionReasons `json:"-"` } type FilterExternal struct { @@ -219,80 +221,81 @@ const ( ) type FilterUpdate struct { - ID int `json:"id"` - Name *string `json:"name,omitempty"` - Enabled *bool `json:"enabled,omitempty"` - MinSize *string `json:"min_size,omitempty"` - MaxSize *string `json:"max_size,omitempty"` - Delay *int `json:"delay,omitempty"` - Priority *int32 `json:"priority,omitempty"` - AnnounceTypes *[]string `json:"announce_types,omitempty"` - MaxDownloads *int `json:"max_downloads,omitempty"` - MaxDownloadsUnit *FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"` - MatchReleases *string `json:"match_releases,omitempty"` - ExceptReleases *string `json:"except_releases,omitempty"` - UseRegex *bool `json:"use_regex,omitempty"` - MatchReleaseGroups *string `json:"match_release_groups,omitempty"` - ExceptReleaseGroups *string `json:"except_release_groups,omitempty"` - MatchReleaseTags *string `json:"match_release_tags,omitempty"` - ExceptReleaseTags *string `json:"except_release_tags,omitempty"` - UseRegexReleaseTags *bool `json:"use_regex_release_tags,omitempty"` - MatchDescription *string `json:"match_description,omitempty"` - ExceptDescription *string `json:"except_description,omitempty"` - UseRegexDescription *bool `json:"use_regex_description,omitempty"` - Scene *bool `json:"scene,omitempty"` - Origins *[]string `json:"origins,omitempty"` - ExceptOrigins *[]string `json:"except_origins,omitempty"` - Bonus *[]string `json:"bonus,omitempty"` - Freeleech *bool `json:"freeleech,omitempty"` - FreeleechPercent *string `json:"freeleech_percent,omitempty"` - SmartEpisode *bool `json:"smart_episode,omitempty"` - Shows *string `json:"shows,omitempty"` - Seasons *string `json:"seasons,omitempty"` - Episodes *string `json:"episodes,omitempty"` - Resolutions *[]string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p. - Codecs *[]string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux). - Sources *[]string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC - Containers *[]string `json:"containers,omitempty"` - MatchHDR *[]string `json:"match_hdr,omitempty"` - ExceptHDR *[]string `json:"except_hdr,omitempty"` - MatchOther *[]string `json:"match_other,omitempty"` - ExceptOther *[]string `json:"except_other,omitempty"` - Years *string `json:"years,omitempty"` - Months *string `json:"months,omitempty"` - Days *string `json:"days,omitempty"` - Artists *string `json:"artists,omitempty"` - Albums *string `json:"albums,omitempty"` - MatchReleaseTypes *[]string `json:"match_release_types,omitempty"` // Album,Single,EP - ExceptReleaseTypes *string `json:"except_release_types,omitempty"` - Formats *[]string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS - Quality *[]string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other - Media *[]string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other - PerfectFlac *bool `json:"perfect_flac,omitempty"` - Cue *bool `json:"cue,omitempty"` - Log *bool `json:"log,omitempty"` - LogScore *int `json:"log_score,omitempty"` - MatchCategories *string `json:"match_categories,omitempty"` - ExceptCategories *string `json:"except_categories,omitempty"` - MatchUploaders *string `json:"match_uploaders,omitempty"` - ExceptUploaders *string `json:"except_uploaders,omitempty"` - MatchRecordLabels *string `json:"match_record_labels,omitempty"` - ExceptRecordLabels *string `json:"except_record_labels,omitempty"` - MatchLanguage *[]string `json:"match_language,omitempty"` - ExceptLanguage *[]string `json:"except_language,omitempty"` - Tags *string `json:"tags,omitempty"` - ExceptTags *string `json:"except_tags,omitempty"` - TagsAny *string `json:"tags_any,omitempty"` - ExceptTagsAny *string `json:"except_tags_any,omitempty"` - TagsMatchLogic *string `json:"tags_match_logic,omitempty"` - ExceptTagsMatchLogic *string `json:"except_tags_match_logic,omitempty"` - MinSeeders *int `json:"min_seeders,omitempty"` - MaxSeeders *int `json:"max_seeders,omitempty"` - MinLeechers *int `json:"min_leechers,omitempty"` - MaxLeechers *int `json:"max_leechers,omitempty"` - Actions []*Action `json:"actions,omitempty"` - External []FilterExternal `json:"external,omitempty"` - Indexers []Indexer `json:"indexers,omitempty"` + ID int `json:"id"` + Name *string `json:"name,omitempty"` + Enabled *bool `json:"enabled,omitempty"` + MinSize *string `json:"min_size,omitempty"` + MaxSize *string `json:"max_size,omitempty"` + Delay *int `json:"delay,omitempty"` + Priority *int32 `json:"priority,omitempty"` + AnnounceTypes *[]string `json:"announce_types,omitempty"` + MaxDownloads *int `json:"max_downloads,omitempty"` + MaxDownloadsUnit *FilterMaxDownloadsUnit `json:"max_downloads_unit,omitempty"` + MatchReleases *string `json:"match_releases,omitempty"` + ExceptReleases *string `json:"except_releases,omitempty"` + UseRegex *bool `json:"use_regex,omitempty"` + MatchReleaseGroups *string `json:"match_release_groups,omitempty"` + ExceptReleaseGroups *string `json:"except_release_groups,omitempty"` + MatchReleaseTags *string `json:"match_release_tags,omitempty"` + ExceptReleaseTags *string `json:"except_release_tags,omitempty"` + UseRegexReleaseTags *bool `json:"use_regex_release_tags,omitempty"` + MatchDescription *string `json:"match_description,omitempty"` + ExceptDescription *string `json:"except_description,omitempty"` + UseRegexDescription *bool `json:"use_regex_description,omitempty"` + Scene *bool `json:"scene,omitempty"` + Origins *[]string `json:"origins,omitempty"` + ExceptOrigins *[]string `json:"except_origins,omitempty"` + Bonus *[]string `json:"bonus,omitempty"` + Freeleech *bool `json:"freeleech,omitempty"` + FreeleechPercent *string `json:"freeleech_percent,omitempty"` + SmartEpisode *bool `json:"smart_episode,omitempty"` + Shows *string `json:"shows,omitempty"` + Seasons *string `json:"seasons,omitempty"` + Episodes *string `json:"episodes,omitempty"` + Resolutions *[]string `json:"resolutions,omitempty"` // SD, 480i, 480p, 576p, 720p, 810p, 1080i, 1080p. + Codecs *[]string `json:"codecs,omitempty"` // XviD, DivX, x264, h.264 (or h264), mpeg2 (or mpeg-2), VC-1 (or VC1), WMV, Remux, h.264 Remux (or h264 Remux), VC-1 Remux (or VC1 Remux). + Sources *[]string `json:"sources,omitempty"` // DSR, PDTV, HDTV, HR.PDTV, HR.HDTV, DVDRip, DVDScr, BDr, BD5, BD9, BDRip, BRRip, DVDR, MDVDR, HDDVD, HDDVDRip, BluRay, WEB-DL, TVRip, CAM, R5, TELESYNC, TS, TELECINE, TC. TELESYNC and TS are synonyms (you don't need both). Same for TELECINE and TC + Containers *[]string `json:"containers,omitempty"` + MatchHDR *[]string `json:"match_hdr,omitempty"` + ExceptHDR *[]string `json:"except_hdr,omitempty"` + MatchOther *[]string `json:"match_other,omitempty"` + ExceptOther *[]string `json:"except_other,omitempty"` + Years *string `json:"years,omitempty"` + Months *string `json:"months,omitempty"` + Days *string `json:"days,omitempty"` + Artists *string `json:"artists,omitempty"` + Albums *string `json:"albums,omitempty"` + MatchReleaseTypes *[]string `json:"match_release_types,omitempty"` // Album,Single,EP + ExceptReleaseTypes *string `json:"except_release_types,omitempty"` + Formats *[]string `json:"formats,omitempty"` // MP3, FLAC, Ogg, AAC, AC3, DTS + Quality *[]string `json:"quality,omitempty"` // 192, 320, APS (VBR), V2 (VBR), V1 (VBR), APX (VBR), V0 (VBR), q8.x (VBR), Lossless, 24bit Lossless, Other + Media *[]string `json:"media,omitempty"` // CD, DVD, Vinyl, Soundboard, SACD, DAT, Cassette, WEB, Other + PerfectFlac *bool `json:"perfect_flac,omitempty"` + Cue *bool `json:"cue,omitempty"` + Log *bool `json:"log,omitempty"` + LogScore *int `json:"log_score,omitempty"` + MatchCategories *string `json:"match_categories,omitempty"` + ExceptCategories *string `json:"except_categories,omitempty"` + MatchUploaders *string `json:"match_uploaders,omitempty"` + ExceptUploaders *string `json:"except_uploaders,omitempty"` + MatchRecordLabels *string `json:"match_record_labels,omitempty"` + ExceptRecordLabels *string `json:"except_record_labels,omitempty"` + MatchLanguage *[]string `json:"match_language,omitempty"` + ExceptLanguage *[]string `json:"except_language,omitempty"` + Tags *string `json:"tags,omitempty"` + ExceptTags *string `json:"except_tags,omitempty"` + TagsAny *string `json:"tags_any,omitempty"` + ExceptTagsAny *string `json:"except_tags_any,omitempty"` + TagsMatchLogic *string `json:"tags_match_logic,omitempty"` + ExceptTagsMatchLogic *string `json:"except_tags_match_logic,omitempty"` + MinSeeders *int `json:"min_seeders,omitempty"` + MaxSeeders *int `json:"max_seeders,omitempty"` + MinLeechers *int `json:"min_leechers,omitempty"` + MaxLeechers *int `json:"max_leechers,omitempty"` + ReleaseProfileDuplicateID *int64 `json:"release_profile_duplicate_id,omitempty"` + Actions []*Action `json:"actions,omitempty"` + External []FilterExternal `json:"external,omitempty"` + Indexers []Indexer `json:"indexers,omitempty"` } func (f *Filter) Validate() error { diff --git a/internal/domain/macros.go b/internal/domain/macros.go index 883a2c6..ac7659b 100644 --- a/internal/domain/macros.go +++ b/internal/domain/macros.go @@ -50,6 +50,7 @@ type Macro struct { IndexerIdentifierExternal string IndexerName string InfoUrl string + IsDuplicate bool Language []string Leechers int LogScore int @@ -66,6 +67,8 @@ type Macro struct { Seeders int Size uint64 SizeString string + SkipDuplicateProfileID int64 + SkipDuplicateProfileName string Source string Tags string Title string @@ -123,6 +126,7 @@ func NewMacro(release Release) Macro { IndexerIdentifierExternal: release.Indexer.IdentifierExternal, IndexerName: release.Indexer.Name, InfoUrl: release.InfoURL, + IsDuplicate: release.IsDuplicate, Language: release.Language, Leechers: release.Leechers, LogScore: release.LogScore, @@ -140,6 +144,8 @@ func NewMacro(release Release) Macro { Size: release.Size, SizeString: humanize.Bytes(release.Size), Source: release.Source, + SkipDuplicateProfileID: release.SkipDuplicateProfileID, + SkipDuplicateProfileName: release.SkipDuplicateProfileName, Tags: strings.Join(release.Tags, ", "), Title: release.Title, TorrentDataRawBytes: release.TorrentDataRawBytes, @@ -149,7 +155,7 @@ func NewMacro(release Release) Macro { TorrentPathName: release.TorrentTmpFile, TorrentUrl: release.DownloadURL, TorrentTmpFile: release.TorrentTmpFile, - Type: release.Type, + Type: release.Type.String(), Uploader: release.Uploader, RecordLabel: release.RecordLabel, Website: release.Website, diff --git a/internal/domain/macros_test.go b/internal/domain/macros_test.go index 08d0f27..7fb63c5 100644 --- a/internal/domain/macros_test.go +++ b/internal/domain/macros_test.go @@ -5,6 +5,7 @@ package domain import ( "fmt" + "github.com/moistari/rls" "testing" "time" @@ -231,7 +232,7 @@ func TestMacros_Parse(t *testing.T) { { name: "test_type", release: Release{ - Type: "episode", + Type: rls.Episode, }, args: args{text: "Type: {{ .Type }}"}, want: "Type: episode", diff --git a/internal/domain/release.go b/internal/domain/release.go index c544dbf..764763f 100644 --- a/internal/domain/release.go +++ b/internal/domain/release.go @@ -7,7 +7,11 @@ import ( "bufio" "bytes" "context" + "crypto/md5" + "encoding/hex" "fmt" + "golang.org/x/text/transform" + "golang.org/x/text/unicode/norm" "html" "io" "math" @@ -18,6 +22,7 @@ import ( "strconv" "strings" "time" + "unicode" "github.com/autobrr/autobrr/pkg/errors" "github.com/autobrr/autobrr/pkg/sharedhttp" @@ -43,6 +48,11 @@ type ReleaseRepo interface { GetActionStatus(ctx context.Context, req *GetReleaseActionStatusRequest) (*ReleaseActionStatus, error) StoreReleaseActionStatus(ctx context.Context, status *ReleaseActionStatus) error + + StoreDuplicateProfile(ctx context.Context, profile *DuplicateReleaseProfile) error + FindDuplicateReleaseProfiles(ctx context.Context) ([]*DuplicateReleaseProfile, error) + DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error + CheckIsDuplicateRelease(ctx context.Context, profile *DuplicateReleaseProfile, release *Release) (bool, error) } type Release struct { @@ -55,6 +65,7 @@ type Release struct { Implementation ReleaseImplementation `json:"implementation"` // irc, rss, api Timestamp time.Time `json:"timestamp"` AnnounceType AnnounceType `json:"announce_type"` + Type rls.Type `json:"type"` // rls.Type InfoURL string `json:"info_url"` DownloadURL string `json:"download_url"` MagnetURI string `json:"-"` @@ -63,9 +74,11 @@ type Release struct { TorrentTmpFile string `json:"-"` TorrentDataRawBytes []byte `json:"-"` TorrentHash string `json:"-"` - TorrentName string `json:"name"` // full release name + TorrentName string `json:"name"` // full release name + NormalizedHash string `json:"normalized_hash"` // normalized torrent name and md5 hashed Size uint64 `json:"size"` - Title string `json:"title"` // Parsed title + Title string `json:"title"` // Parsed title + SubTitle string `json:"sub_title"` // Parsed secondary title for shows e.g. episode name Description string `json:"-"` Category string `json:"category"` Categories []string `json:"categories,omitempty"` @@ -89,8 +102,11 @@ type Release struct { Proper bool `json:"proper"` Repack bool `json:"repack"` Website string `json:"website"` + Hybrid bool `json:"hybrid"` + Edition []string `json:"edition"` + Cut []string `json:"cut"` + MediaProcessing string `json:"media_processing"` // Remux, Encode, Untouched Artists string `json:"-"` - Type string `json:"type"` // Album,Single,EP LogScore int `json:"-"` HasCue bool `json:"-"` HasLog bool `json:"-"` @@ -110,15 +126,183 @@ type Release struct { AdditionalSizeCheckRequired bool `json:"-"` AdditionalUploaderCheckRequired bool `json:"-"` AdditionalRecordLabelCheckRequired bool `json:"-"` + IsDuplicate bool `json:"-"` + SkipDuplicateProfileID int64 `json:"-"` + SkipDuplicateProfileName string `json:"-"` FilterID int `json:"-"` Filter *Filter `json:"-"` ActionStatus []ReleaseActionStatus `json:"action_status"` } +// Hash return md5 hashed normalized release name +func (r *Release) Hash() string { + formatted := r.TorrentName + + // for tv and movies we create the formatted title to have the best chance of matching + if r.IsTypeVideo() { + formatted = r.NormalizedTitle() + } + + normalized := MustNormalize(formatted) + h := md5.Sum([]byte(normalized)) + str := hex.EncodeToString(h[:]) + return str +} + +// MustNormalize applies the Normalize transform to s, returning a lower cased, +// clean form of s useful for matching titles. +func MustNormalize(s string) string { + s, _, err := transform.String(NewNormalizer(), s) + if err != nil { + panic(err) + } + return s +} + +// NewNormalizer is a custom rls.Normalizer that keeps plus sign + for HDR10+ fx +// It creates a new a text transformer chain (similiar to +// NewCleaner) that normalizes text to lower case clean form useful for +// matching titles. +// +// See: https://go.dev/blog/normalization +func NewNormalizer() transform.Transformer { + return transform.Chain( + norm.NFD, + rls.NewCollapser( + true, true, + "`"+`':;~!@#%^*=()[]{}<>/?|\",`, " \t\r\n\f._", + func(r, prev, next rune) rune { + switch { + case r == '-' && unicode.IsSpace(prev): + return -1 + case r == '$' && (unicode.IsLetter(prev) || unicode.IsLetter(next)): + return 'S' + case r == '£' && (unicode.IsLetter(prev) || unicode.IsLetter(next)): + return 'L' + case r == '$', r == '£': + return -1 + } + return r + }, + ), + norm.NFC, + ) +} + +func (r *Release) NormalizedTitle() string { + var v []string + + v = append(v, r.Title) + + if r.Year > 0 && r.Month > 0 && r.Day > 0 { + v = append(v, fmt.Sprintf("%d %d %d", r.Year, r.Month, r.Day)) + } else if r.Year > 0 { + v = append(v, fmt.Sprintf("%d", r.Year)) + } + + if len(r.Language) > 0 { + v = append(v, strings.Join(r.Language, " ")) + } + + if len(r.Cut) > 0 { + v = append(v, strings.Join(r.Cut, " ")) + } + + if len(r.Edition) > 0 { + v = append(v, strings.Join(r.Edition, " ")) + } + + if r.Season > 0 && r.Episode > 0 { + v = append(v, fmt.Sprintf("S%dE%d", r.Season, r.Episode)) + } else if r.Season > 0 && r.Episode == 0 { + v = append(v, fmt.Sprintf("S%d", r.Season)) + } + + if r.Proper { + v = append(v, "PROPER") + } + + if r.Repack { + v = append(v, r.RepackStr()) + } + + if r.Hybrid { + v = append(v, "HYBRiD") + } + + if r.SubTitle != "" { + v = append(v, r.SubTitle) + } + + if r.Resolution != "" { + v = append(v, r.Resolution) + } + + if r.Website != "" { + v = append(v, r.Website) + } + + if r.Region != "" { + v = append(v, r.Region) + } + + if r.Source != "" { + v = append(v, r.Source) + } + + // remux + if r.MediaProcessing == "REMUX" { + v = append(v, "REMUX") + } + + if len(r.Codec) > 0 { + v = append(v, strings.Join(r.Codec, " ")) + } + + if len(r.HDR) > 0 { + v = append(v, strings.Join(r.HDR, " ")) + } + + if len(r.Audio) > 0 { + v = append(v, r.AudioString()) + } + + str := strings.Join(v, " ") + + if r.Group != "" { + str = fmt.Sprintf("%s-%s", str, r.Group) + } + + return str +} + +func (r *Release) RepackStr() string { + if r.Other != nil { + if slices.Contains(r.Other, "REPACK") { + return "REPACK" + } else if slices.Contains(r.Other, "REREPACK") { + return "REREPACK" + } else if slices.Contains(r.Other, "REPACK2") { + return "REPACK2" + } else if slices.Contains(r.Other, "REPACK3") { + return "REPACK3" + } + } + return "" +} + func (r *Release) Raw(s string) rls.Release { return rls.ParseString(s) } +func (r *Release) ParseType(s string) { + r.Type = rls.ParseType(s) +} + +func (r *Release) IsTypeVideo() bool { + return r.Type.Is(rls.Movie, rls.Series, rls.Episode) +} + type AnnounceType string const ( @@ -361,6 +545,10 @@ func NewRelease(indexer IndexerMinimal) *Release { Implementation: ReleaseImplementationIRC, Timestamp: time.Now(), Tags: []string{}, + Language: []string{}, + Edition: []string{}, + Cut: []string{}, + Other: []string{}, Size: 0, AnnounceType: AnnounceTypeNew, } @@ -371,28 +559,42 @@ func NewRelease(indexer IndexerMinimal) *Release { func (r *Release) ParseString(title string) { rel := rls.ParseString(title) - r.Type = rel.Type.String() + r.Type = rel.Type r.TorrentName = title + r.Source = rel.Source r.Resolution = rel.Resolution r.Region = rel.Region + + if rel.Language != nil { + r.Language = rel.Language + } + r.Audio = rel.Audio r.AudioChannels = rel.Channels r.Codec = rel.Codec r.Container = rel.Container r.HDR = rel.HDR r.Artists = rel.Artist - r.Language = rel.Language - r.Other = rel.Other + if rel.Other != nil { + r.Other = rel.Other + } r.Proper = slices.Contains(r.Other, "PROPER") - r.Repack = slices.Contains(r.Other, "REPACK") + r.Repack = slices.Contains(r.Other, "REPACK") || slices.Contains(r.Other, "REREPACK") + r.Hybrid = slices.Contains(r.Other, "HYBRiD") + + // TODO default to Encode and set Untouched for discs + if slices.Contains(r.Other, "REMUX") { + r.MediaProcessing = "REMUX" + } if r.Title == "" { r.Title = rel.Title } + r.SubTitle = rel.Subtitle if r.Season == 0 { r.Season = rel.Series @@ -415,8 +617,22 @@ func (r *Release) ParseString(title string) { r.Group = rel.Group } + if r.Website == "" { + r.Website = rel.Collection + } + + if rel.Cut != nil { + r.Cut = rel.Cut + } + + if rel.Edition != nil { + r.Edition = rel.Edition + } + r.ParseReleaseTagsString(r.ReleaseTags) r.extraParseSource(rel) + + r.NormalizedHash = r.Hash() } func (r *Release) extraParseSource(rel rls.Release) { @@ -451,7 +667,7 @@ func (r *Release) extraParseSource(rel rls.Release) { } // check res to be 1080p or 2160p and codec to be AVC, HEVC or if other contains Remux, then set source to BluRay if it differs - if !basicContainsSlice(r.Source, []string{"WEB-DL", "BluRay", "UHD.BluRay"}) && basicContainsSlice(r.Resolution, []string{"1080p", "2160p"}) && basicContainsMatch(r.Codec, []string{"AVC", "HEVC"}) && basicContainsMatch(r.Other, []string{"REMUX"}) { + if !basicContainsSlice(r.Source, []string{"WEB-DL", "BluRay", "UHD.BluRay"}) && basicContainsSlice(r.Resolution, []string{"1080p", "2160p"}) && basicContainsMatch(r.Codec, []string{"AVC", "H.264", "H.265", "HEVC"}) && basicContainsMatch(r.Other, []string{"REMUX"}) { // handle missing or unexpected source for some bluray releases if r.Resolution == "1080p" { r.Source = "BluRay" @@ -463,6 +679,10 @@ func (r *Release) extraParseSource(rel rls.Release) { } func (r *Release) ParseReleaseTagsString(tags string) { + if tags == "" { + return + } + cleanTags := CleanReleaseTags(tags) t := ParseReleaseTagString(cleanTags) @@ -543,6 +763,20 @@ func (r *Release) OpenTorrentFile() error { return nil } +// AudioString takes r.Audio and r.AudioChannels and returns a string like "DDP Atmos 5.1" +func (r *Release) AudioString() string { + var audio []string + + audio = append(audio, r.Audio...) + audio = append(audio, r.AudioChannels) + + if len(audio) > 0 { + return strings.Join(audio, " ") + } + + return "" +} + func (r *Release) DownloadTorrentFileCtx(ctx context.Context) error { return r.downloadTorrentFile(ctx) } @@ -992,3 +1226,30 @@ func getUniqueTags(target []string, source []string) []string { return target } + +type DuplicateReleaseProfile struct { + ID int64 `json:"id"` + Name string `json:"name"` + Protocol bool `json:"protocol"` + ReleaseName bool `json:"release_name"` + Hash bool `json:"hash"` + Title bool `json:"title"` + SubTitle bool `json:"sub_title"` + Year bool `json:"year"` + Month bool `json:"month"` + Day bool `json:"day"` + Source bool `json:"source"` + Resolution bool `json:"resolution"` + Codec bool `json:"codec"` + Container bool `json:"container"` + DynamicRange bool `json:"dynamic_range"` + Audio bool `json:"audio"` + Group bool `json:"group"` + Season bool `json:"season"` + Episode bool `json:"episode"` + Website bool `json:"website"` + Proper bool `json:"proper"` + Repack bool `json:"repack"` + Edition bool `json:"edition"` + Language bool `json:"language"` +} diff --git a/internal/domain/release_download_test.go b/internal/domain/release_download_test.go index 234593a..ed8859c 100644 --- a/internal/domain/release_download_test.go +++ b/internal/domain/release_download_test.go @@ -15,6 +15,7 @@ import ( "testing" "time" + "github.com/moistari/rls" "github.com/rs/zerolog" ) @@ -127,7 +128,7 @@ func TestRelease_DownloadTorrentFile(t *testing.T) { Repack bool Website string Artists string - Type string + Type rls.Type LogScore int Origin string Tags []string diff --git a/internal/domain/release_test.go b/internal/domain/release_test.go index 22e1281..f89cd95 100644 --- a/internal/domain/release_test.go +++ b/internal/domain/release_test.go @@ -6,6 +6,7 @@ package domain import ( "testing" + "github.com/moistari/rls" "github.com/stretchr/testify/assert" ) @@ -21,19 +22,20 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX", }, want: Release{ - TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - //Website: "ATVP", - Type: "series", + TorrentName: "Servant S01 2160p ATVP WEB-DL DDP 5.1 Atmos DV HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Website: "ATVP", + Type: rls.Series, }, }, { @@ -42,18 +44,20 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", }, want: Release{ - TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - Type: "series", + TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Website: "ATVP", + Type: rls.Series, }, }, { @@ -63,20 +67,22 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MKV / 2160p / WEB-DL", }, want: Release{ - TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", - ReleaseTags: "MKV / 2160p / WEB-DL", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Container: "mkv", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - Type: "series", + TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + ReleaseTags: "MKV / 2160p / WEB-DL", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Container: "mkv", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Website: "ATVP", + Type: rls.Series, }, }, { @@ -86,20 +92,22 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MKV | 2160p | WEB-DL", }, want: Release{ - TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", - ReleaseTags: "MKV | 2160p | WEB-DL", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Container: "mkv", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - Type: "series", + TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + ReleaseTags: "MKV | 2160p | WEB-DL", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Container: "mkv", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Website: "ATVP", + Type: rls.Series, }, }, { @@ -109,20 +117,22 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MP4 | 2160p | WEB-DL", }, want: Release{ - TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", - ReleaseTags: "MP4 | 2160p | WEB-DL", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Container: "mp4", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - Type: "series", + TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + ReleaseTags: "MP4 | 2160p | WEB-DL", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Container: "mp4", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Website: "ATVP", + Type: rls.Series, }, }, { @@ -132,22 +142,24 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!", }, want: Release{ - TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", - ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!", - Title: "Servant", - Season: 1, - Episode: 0, - Resolution: "2160p", - Source: "WEB-DL", - Container: "mp4", - Codec: []string{"HEVC"}, - Audio: []string{"DDP", "Atmos"}, - AudioChannels: "5.1", - HDR: []string{"DV"}, - Group: "FLUX", - Freeleech: true, - Bonus: []string{"Freeleech"}, - Type: "series", + TorrentName: "Servant.S01.2160p.ATVP.WEB-DL.DDP.5.1.Atmos.DV.HEVC-FLUX", + NormalizedHash: "9b73e77d51fb0b69367ea96c761577b0", + ReleaseTags: "MP4 | 2160p | WEB-DL | Freeleech!", + Title: "Servant", + Season: 1, + Episode: 0, + Resolution: "2160p", + Source: "WEB-DL", + Container: "mp4", + Codec: []string{"HEVC"}, + Audio: []string{"DDP", "Atmos"}, + AudioChannels: "5.1", + HDR: []string{"DV"}, + Group: "FLUX", + Freeleech: true, + Bonus: []string{"Freeleech"}, + Website: "ATVP", + Type: rls.Series, }, }, { @@ -156,15 +168,16 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP", }, want: Release{ - TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP", - Title: "Rippers Revenge", - Year: 2023, - Language: []string{"GERMAN", "DL"}, - Resolution: "1080p", - Source: "BluRay", - Codec: []string{"MPEG-2"}, - Group: "GROUP", - Type: "movie", + TorrentName: "Rippers.Revenge.2023.German.DL.1080p.BluRay.MPEG2-GROUP", + NormalizedHash: "cb28c69ce117723d01ea6562ef3bae67", + Title: "Rippers Revenge", + Year: 2023, + Language: []string{"GERMAN", "DL"}, + Resolution: "1080p", + Source: "BluRay", + Codec: []string{"MPEG-2"}, + Group: "GROUP", + Type: rls.Movie, }, }, { @@ -173,13 +186,14 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF", }, want: Release{ - TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF", - Title: "Analogue", - Resolution: "1080p", // rls does not differentiate between 1080i and 1080p which results in all 1080 releases being parsed as 1080p - Source: "AHDTV", - Codec: []string{"H.264"}, - Group: "ABCDEF", - Type: "movie", + TorrentName: "Analogue.1080i.AHDTV.H264-ABCDEF", + NormalizedHash: "1a1eb25b15269d3189138326658aafe0", + Title: "Analogue", + Resolution: "1080p", // rls does not differentiate between 1080i and 1080p which results in all 1080 releases being parsed as 1080p + Source: "AHDTV", + Codec: []string{"H.264"}, + Group: "ABCDEF", + Type: rls.Movie, }, }, { @@ -189,17 +203,18 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", }, want: Release{ - TorrentName: "Artist - Albumname", - ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", - Title: "Artist", - Group: "Albumname", - Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"}, - AudioFormat: "FLAC", - Source: "CD", - Bitrate: "Lossless", - HasLog: true, - LogScore: 100, - HasCue: true, + TorrentName: "Artist - Albumname", + NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac", + ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", + Title: "Artist", + Group: "Albumname", + Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"}, + AudioFormat: "FLAC", + Source: "CD", + Bitrate: "Lossless", + HasLog: true, + LogScore: 100, + HasCue: true, }, }, { @@ -210,14 +225,15 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MP3 / 320 / Cassette", }, want: Release{ - TorrentName: "Various Artists - Music '21", - Tags: []string{"house, techno, tech.house, electro.house, future.house, bass.house, melodic.house"}, - ReleaseTags: "MP3 / 320 / Cassette", - Title: "Various Artists - Music '21", - Source: "Cassette", - Audio: []string{"320", "MP3"}, - AudioFormat: "MP3", - Bitrate: "320", + TorrentName: "Various Artists - Music '21", + NormalizedHash: "fbf7713532f4d2d03b62897fc990e37e", + Tags: []string{"house, techno, tech.house, electro.house, future.house, bass.house, melodic.house"}, + ReleaseTags: "MP3 / 320 / Cassette", + Title: "Various Artists - Music '21", + Source: "Cassette", + Audio: []string{"320", "MP3"}, + AudioFormat: "MP3", + Bitrate: "320", }, }, { @@ -227,14 +243,15 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "MP3 / V0 (VBR) / CD", }, want: Release{ - TorrentName: "The artist (ザ・フリーダムユニティ) - Long album name", - ReleaseTags: "MP3 / V0 (VBR) / CD", - Title: "The artist", - Group: "name", - Source: "CD", - Audio: []string{"MP3", "VBR", "V0 (VBR)"}, - AudioFormat: "MP3", - Bitrate: "V0 (VBR)", + TorrentName: "The artist (ザ・フリーダムユニティ) - Long album name", + NormalizedHash: "6b4051de7524f65d3c25be535d2c95ed", + ReleaseTags: "MP3 / V0 (VBR) / CD", + Title: "The artist", + Group: "name", + Source: "CD", + Audio: []string{"MP3", "VBR", "V0 (VBR)"}, + AudioFormat: "MP3", + Bitrate: "V0 (VBR)", }, }, { @@ -244,17 +261,18 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", }, want: Release{ - TorrentName: "Artist - Albumname", - ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", - Title: "Artist", - Group: "Albumname", - Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"}, - AudioFormat: "FLAC", - Source: "CD", - Bitrate: "Lossless", - HasLog: true, - LogScore: 100, - HasCue: true, + TorrentName: "Artist - Albumname", + NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac", + ReleaseTags: "FLAC / Lossless / Log / 100% / Cue / CD", + Title: "Artist", + Group: "Albumname", + Audio: []string{"Cue", "FLAC", "Lossless", "Log100", "Log"}, + AudioFormat: "FLAC", + Source: "CD", + Bitrate: "Lossless", + HasLog: true, + LogScore: 100, + HasCue: true, }, }, { @@ -264,17 +282,18 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD", }, want: Release{ - TorrentName: "Artist - Albumname", - ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD", - Title: "Artist", - Group: "Albumname", - Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log100", "Log"}, - AudioFormat: "FLAC", - Source: "CD", - Bitrate: "24BIT Lossless", - HasLog: true, - LogScore: 100, - HasCue: true, + TorrentName: "Artist - Albumname", + NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac", + ReleaseTags: "FLAC / 24bit Lossless / Log / 100% / Cue / CD", + Title: "Artist", + Group: "Albumname", + Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log100", "Log"}, + AudioFormat: "FLAC", + Source: "CD", + Bitrate: "24BIT Lossless", + HasLog: true, + LogScore: 100, + HasCue: true, }, }, { @@ -284,17 +303,18 @@ func TestRelease_Parse(t *testing.T) { ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD", }, want: Release{ - TorrentName: "Artist - Albumname", - ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD", - Title: "Artist", - Group: "Albumname", - Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log78", "Log"}, - AudioFormat: "FLAC", - Source: "CD", - Bitrate: "24BIT Lossless", - HasLog: true, - LogScore: 78, - HasCue: true, + TorrentName: "Artist - Albumname", + NormalizedHash: "5de820bfae23f2bdc3a56c827a5230ac", + ReleaseTags: "FLAC / 24bit Lossless / Log / 78% / Cue / CD", + Title: "Artist", + Group: "Albumname", + Audio: []string{"24BIT Lossless", "Cue", "FLAC", "Log78", "Log"}, + AudioFormat: "FLAC", + Source: "CD", + Bitrate: "24BIT Lossless", + HasLog: true, + LogScore: 78, + HasCue: true, }, }, { @@ -303,18 +323,22 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1", }, want: Release{ - TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1", - Title: "I Am Movie", - Resolution: "2160p", - Source: "UHD.BluRay", - Codec: []string{"HEVC"}, - HDR: []string{"DV"}, - Audio: []string{"DTS-HD.MA"}, - AudioChannels: "5.1", - Year: 2007, - Group: "GROUP1", - Other: []string{"HYBRiD", "REMUX"}, - Type: "movie", + TorrentName: "I Am Movie 2007 Theatrical UHD BluRay 2160p DTS-HD MA 5.1 DV HEVC HYBRID REMUX-GROUP1", + NormalizedHash: "f70840f42149d7d2b539c75f5e853493", + Title: "I Am Movie", + Resolution: "2160p", + Source: "UHD.BluRay", + Codec: []string{"HEVC"}, + HDR: []string{"DV"}, + Audio: []string{"DTS-HD.MA"}, + AudioChannels: "5.1", + Year: 2007, + Group: "GROUP1", + Hybrid: true, + Cut: []string{"Theatrical.Cut"}, + MediaProcessing: "REMUX", + Other: []string{"HYBRiD", "REMUX"}, + Type: rls.Movie, }, }, { @@ -323,19 +347,21 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)", }, want: Release{ - TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)", - Title: "The Peripheral", - Resolution: "2160p", - Source: "WEB-DL", - Codec: []string{"H.265"}, - HDR: []string{"HDR10+"}, - Audio: []string{"DDP"}, - AudioChannels: "5.1", - Year: 2022, - Group: "GROUP1", - Season: 1, - Language: []string{"ENGLiSH"}, - Type: "series", + TorrentName: "The Peripheral (2022) S01 (2160p AMZN WEB-DL H265 HDR10+ DDP 5.1 English - GROUP1)", + NormalizedHash: "33fe914c00ea5a2945883cffb4183923", + Title: "The Peripheral", + Resolution: "2160p", + Source: "WEB-DL", + Codec: []string{"H.265"}, + HDR: []string{"HDR10+"}, + Audio: []string{"DDP"}, + AudioChannels: "5.1", + Year: 2022, + Group: "GROUP1", + Season: 1, + Language: []string{"ENGLiSH"}, + Website: "AMZN", + Type: rls.Series, }, }, { @@ -344,18 +370,20 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES", }, want: Release{ - TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES", - Title: "Old Movie", - Year: 1954, - Source: "UHD.BluRay", - Resolution: "2160p", - Other: []string{"REMUX"}, - HDR: []string{"DV", "HDR10"}, - Codec: []string{"HEVC"}, - Audio: []string{"DTS-HD.MA"}, - AudioChannels: "5.1", - Group: "CiNEPHiLES", - Type: "movie", + TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-CiNEPHiLES", + Title: "Old Movie", + NormalizedHash: "9990678babd4f24f268dfe7856b29f0e", + Year: 1954, + Source: "UHD.BluRay", + Resolution: "2160p", + Other: []string{"REMUX"}, + HDR: []string{"DV", "HDR10"}, + Codec: []string{"HEVC"}, + Audio: []string{"DTS-HD.MA"}, + AudioChannels: "5.1", + Group: "CiNEPHiLES", + Type: rls.Movie, + MediaProcessing: "REMUX", }, }, { @@ -364,17 +392,19 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD", }, want: Release{ - TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD", - Title: "Death Hunt", - Year: 1981, - Source: "BluRay", - Resolution: "1080p", - Other: []string{"REMUX"}, - Codec: []string{"AVC"}, - Audio: []string{"DTS-HD.MA"}, - AudioChannels: "2.0", - Group: "playBD", - Type: "movie", + TorrentName: "Death Hunt 1981 1080p Remux AVC DTS-HD MA 2.0-playBD", + Title: "Death Hunt", + NormalizedHash: "e6b1b708b8ea39470de6d327da2920d6", + Year: 1981, + Source: "BluRay", + Resolution: "1080p", + Other: []string{"REMUX"}, + Codec: []string{"AVC"}, + Audio: []string{"DTS-HD.MA"}, + AudioChannels: "2.0", + Group: "playBD", + Type: rls.Movie, + MediaProcessing: "REMUX", }, }, { @@ -383,18 +413,20 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS", }, want: Release{ - TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS", - Title: "Old Movie", - Year: 1954, - Source: "UHD.BluRay", - Resolution: "2160p", - Other: []string{"REMUX"}, - HDR: []string{"DV", "HDR10"}, - Codec: []string{"HEVC"}, - Audio: []string{"DTS-HD.MA"}, - AudioChannels: "5.1", - Group: "VHS", - Type: "movie", + TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1-VHS", + Title: "Old Movie", + NormalizedHash: "1ea814c8c21979e7692efb1b24298e95", + Year: 1954, + Source: "UHD.BluRay", + Resolution: "2160p", + Other: []string{"REMUX"}, + HDR: []string{"DV", "HDR10"}, + Codec: []string{"HEVC"}, + Audio: []string{"DTS-HD.MA"}, + AudioChannels: "5.1", + Group: "VHS", + Type: rls.Movie, + MediaProcessing: "REMUX", }, }, { @@ -403,18 +435,20 @@ func TestRelease_Parse(t *testing.T) { TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS", }, want: Release{ - TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS", - Title: "Old Movie", - Year: 1954, - Source: "UHD.BluRay", - Resolution: "2160p", - Other: []string{"REMUX"}, - HDR: []string{"DV", "HDR10"}, - Codec: []string{"HEVC"}, - Audio: []string{"DTS-HD.MA"}, - AudioChannels: "5.1", - Group: "VHS", - Type: "movie", + TorrentName: "Old Movie 1954 2160p Remux DoVi HDR10 HEVC DTS-HD MA 5.1 VHS", + Title: "Old Movie", + NormalizedHash: "1ea814c8c21979e7692efb1b24298e95", + Year: 1954, + Source: "UHD.BluRay", + Resolution: "2160p", + Other: []string{"REMUX"}, + HDR: []string{"DV", "HDR10"}, + Codec: []string{"HEVC"}, + Audio: []string{"DTS-HD.MA"}, + AudioChannels: "5.1", + Group: "VHS", + Type: rls.Movie, + MediaProcessing: "REMUX", }, }, } @@ -863,3 +897,220 @@ func Test_getUniqueTags(t *testing.T) { }) } } + +func TestRelease_Hash(t *testing.T) { + type fields struct { + TorrentName string + } + tests := []struct { + name string + fields fields + want string + }{ + { + name: "1", + fields: fields{ + TorrentName: "That.Movie.2023.2160p.BluRay.DTS-HD.5.1.x265-GROUP", + }, + want: "833e6c4c78e960ccbce4ef68f6564082", + }, + { + name: "2", + fields: fields{ + TorrentName: "That Movie 2023 2160p BluRay DTS-HD 5.1 x265-GROUP", + }, + want: "833e6c4c78e960ccbce4ef68f6564082", + }, + { + name: "3", + fields: fields{ + TorrentName: "That Movie 2023 2160p BluRay DTS-HD 5 1 x265-GROUP", + }, + want: "833e6c4c78e960ccbce4ef68f6564082", + }, + { + name: "4", + fields: fields{ + TorrentName: "That Movie 2023 1080p Remux DTS-HD 5 1 AVC-GROUP", + }, + want: "ee15fe3c926a5b40370aec970f6b4c43", + }, + { + name: "5", + fields: fields{ + TorrentName: "That Movie 2023 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP", + }, + want: "ee15fe3c926a5b40370aec970f6b4c43", + }, + { + name: "6", + fields: fields{ + TorrentName: "That Movie 2023 Theatrical Cut Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP", + }, + want: "db7e55fb3471e76ba27bf8a2e1e59236", + }, + { + name: "7", + fields: fields{ + TorrentName: "That Movie 2023 Directors Cut Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP", + }, + want: "bccfe6762b34cbc3a54784351b5424fd", + }, + { + name: "8", + fields: fields{ + TorrentName: "That Show 2023 S01 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP", + }, + want: "d1c74c33e1c36f3545d06aa8857610b1", + }, + { + name: "9", + fields: fields{ + TorrentName: "That Show 2023 S01 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2", + }, + want: "29e228f705f070041e72ebaa7b29239e", + }, + { + name: "10", + fields: fields{ + TorrentName: "That Show 2023 S1 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2", + }, + want: "29e228f705f070041e72ebaa7b29239e", + }, + { + name: "11", + fields: fields{ + TorrentName: "That Show 2023 S2 Bluray 1080p Remux DTS-HD 5.1 AVC-GROUP2", + }, + want: "cbc67f4eec93428361eddf040654e077", + }, + { + name: "12", + fields: fields{ + TorrentName: "That Show 2023 S01E01 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "806240c76cc32413efd5bcf033abf512", + }, + { + name: "13", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "33c866b898a59d255a09c10bd23a5e3f", + }, + { + name: "14", + fields: fields{ + TorrentName: "That Show 2023 REPACK S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "7be5ebeee948820f53f24091054fc7cd", + }, + { + name: "15", + fields: fields{ + TorrentName: "That Show 2023 REPACK2 S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "84e3e70f6c983f96f10cba50213a338d", + }, + { + name: "16", + fields: fields{ + TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "52d5d1a28a340bc5a836c0244c206c3b", + }, + { + name: "17", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DV DTS-HD 5.1 HEVC-GROUP", + }, + want: "5e196039d426649ab5221c3e87ed108d", + }, + { + name: "18", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DoVi DTS-HD 5.1 HEVC-GROUP", + }, + want: "5e196039d426649ab5221c3e87ed108d", + }, + { + name: "19", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DoVi HDR10 DTS-HD 5.1 HEVC-GROUP", + }, + want: "f52a9a2f26e94c90bdaa8eaca2136cc3", + }, + { + name: "20", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP", + }, + want: "f52a9a2f26e94c90bdaa8eaca2136cc3", + }, + { + name: "21", + fields: fields{ + TorrentName: "that show 2023 S01E01 this is the start 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP", + }, + want: "f52a9a2f26e94c90bdaa8eaca2136cc3", + }, + { + name: "22", + fields: fields{ + TorrentName: "That Show 2023 S01E01 This is the start 2160p HYBRID WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP", + }, + want: "414f792a345148ed4c8e070ef77340ff", + }, + { + name: "23", + fields: fields{ + TorrentName: "That Movie 2008 1080p GBR Blu-ray AVC TrueHD 5.1", + }, + want: "70bb23ff5da4209419b6d83628310a43", + }, + { + name: "24", + fields: fields{ + TorrentName: "That Movie 2008 1080p GER Blu-ray AVC TrueHD 5.1", + }, + want: "369ba9f92fb7819e0d07791fddf3c85f", + }, + { + name: "25", + fields: fields{ + TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p AMZN WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "28e215fee790fbad3957383162f7c5a6", + }, + { + name: "26", + fields: fields{ + TorrentName: "That Show 2023 PROPER S01E01 This is the start 2160p DSNP WEB-DL DTS-HD 5.1 HEVC-GROUP", + }, + want: "6a26cd652170d0efc4b0e0320fbd38c1", + }, + { + name: "27", + fields: fields{ + TorrentName: "That Other Show 2023 S01E01 2160p WEB-DL DV HDR10 DTS-HD 5.1 HEVC-GROUP", + }, + want: "544c2fc4024ece5664c0f623a5feaeab", + }, + { + name: "28", + fields: fields{ + TorrentName: "That Other Show 2023 S01E01 2160p WEB-DL DV HDR10+ DTS-HD 5.1 HEVC-GROUP", + }, + want: "63b5d87abe5fb49131785de426708d31", + }, + } + for _, tt := range tests { + t.Run(tt.name, func(t *testing.T) { + r := &Release{ + TorrentName: tt.fields.TorrentName, + } + r.ParseString(tt.fields.TorrentName) + assert.Equalf(t, tt.want, r.NormalizedHash, "Hash()") + }) + } +} diff --git a/internal/domain/releasetags.go b/internal/domain/releasetags.go index 8977da5..c415cce 100644 --- a/internal/domain/releasetags.go +++ b/internal/domain/releasetags.go @@ -425,9 +425,14 @@ func ParseReleaseTags(tags []string) ReleaseTags { return releaseTags } + func ParseReleaseTagString(tags string) ReleaseTags { releaseTags := ReleaseTags{} + if tags == "" { + return releaseTags + } + for tagType, tagInfos := range types { for _, info := range tagInfos { diff --git a/internal/feed/rss_test.go b/internal/feed/rss_test.go index 0c45257..f0611f6 100644 --- a/internal/feed/rss_test.go +++ b/internal/feed/rss_test.go @@ -12,6 +12,7 @@ import ( "github.com/dustin/go-humanize" "github.com/mmcdole/gofeed" + "github.com/moistari/rls" "github.com/rs/zerolog" "github.com/stretchr/testify/assert" ) @@ -73,7 +74,67 @@ func TestRSSJob_processItem(t *testing.T) { Link: "/details.php?id=00000&hit=1", GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", }}, - want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)}, + want: &domain.Release{ + ID: 0, + FilterStatus: "PENDING", + Rejections: []string{}, + Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, + FilterName: "", + Protocol: "torrent", + Implementation: "RSS", + AnnounceType: domain.AnnounceTypeNew, + Timestamp: now, + GroupID: "", + TorrentID: "", + DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", + TorrentTmpFile: "", + TorrentDataRawBytes: []uint8(nil), + TorrentHash: "", + TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", + NormalizedHash: "edfbe552ccde335f34b801e15930bc35", + Size: 1490000000, + Title: "Some Release Title", + Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", + Category: "", + Season: 0, + Episode: 0, + Year: 2022, + Month: 9, + Day: 22, + Resolution: "720p", + Source: "WEB", + Codec: []string{"H.264"}, + Container: "", + HDR: []string(nil), + Audio: []string(nil), + AudioChannels: "", + Group: "GROUP", + Region: "", + Language: []string{}, + Proper: false, + Repack: false, + Edition: []string{}, + Cut: []string{}, + Website: "", + Artists: "", + Type: rls.Episode, + LogScore: 0, + Origin: "", + Tags: []string{}, + ReleaseTags: "", + Freeleech: false, + FreeleechPercent: 0, + Bonus: []string(nil), + Uploader: "", + PreTime: "", + Other: []string{}, + RawCookie: "", + AdditionalSizeCheckRequired: false, + AdditionalUploaderCheckRequired: false, + FilterID: 0, + Filter: (*domain.Filter)(nil), + ActionStatus: []domain.ReleaseActionStatus(nil), + }, }, { name: "with_baseurl", @@ -107,7 +168,67 @@ func TestRSSJob_processItem(t *testing.T) { Link: "https://fake-feed.com/details.php?id=00000&hit=1", GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", }}, - want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)}, + want: &domain.Release{ + ID: 0, + FilterStatus: "PENDING", + Rejections: []string{}, + Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, + FilterName: "", + Protocol: "torrent", + Implementation: "RSS", + AnnounceType: domain.AnnounceTypeNew, + Timestamp: now, + GroupID: "", + TorrentID: "", + DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", + TorrentTmpFile: "", + TorrentDataRawBytes: []uint8(nil), + TorrentHash: "", + TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", + NormalizedHash: "edfbe552ccde335f34b801e15930bc35", + Size: 1490000000, + Title: "Some Release Title", + Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", + Category: "", + Season: 0, + Episode: 0, + Year: 2022, + Month: 9, + Day: 22, + Resolution: "720p", + Source: "WEB", + Codec: []string{"H.264"}, + Container: "", + HDR: []string(nil), + Audio: []string(nil), + AudioChannels: "", + Group: "GROUP", + Region: "", + Language: []string{}, + Proper: false, + Repack: false, + Edition: []string{}, + Cut: []string{}, + Website: "", + Artists: "", + Type: rls.Episode, + LogScore: 0, + Origin: "", + Tags: []string{}, + ReleaseTags: "", + Freeleech: false, + FreeleechPercent: 0, + Bonus: []string(nil), + Uploader: "", + PreTime: "", + Other: []string{}, + RawCookie: "", + AdditionalSizeCheckRequired: false, + AdditionalUploaderCheckRequired: false, + FilterID: 0, + Filter: (*domain.Filter)(nil), + ActionStatus: []domain.ReleaseActionStatus(nil), + }, }, { name: "time_parse", @@ -142,7 +263,67 @@ func TestRSSJob_processItem(t *testing.T) { GUID: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", //PublishedParsed: &nowMinusTime, }}, - want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 1490000000, Title: "Some Release Title", Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, AdditionalUploaderCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)}, + want: &domain.Release{ + ID: 0, + FilterStatus: "PENDING", + Rejections: []string{}, + Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, + FilterName: "", + Protocol: "torrent", + Implementation: "RSS", + AnnounceType: domain.AnnounceTypeNew, + Timestamp: now, + GroupID: "", + TorrentID: "", + DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", + TorrentTmpFile: "", + TorrentDataRawBytes: []uint8(nil), + TorrentHash: "", + TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", + NormalizedHash: "edfbe552ccde335f34b801e15930bc35", + Size: 1490000000, + Title: "Some Release Title", + Description: "Category: Example\n Size: 1.49 GB\n Status: 27 seeders and 1 leechers\n Speed: 772.16 kB/s\n Added: 2022-09-29 16:06:08\n", + Category: "", + Season: 0, + Episode: 0, + Year: 2022, + Month: 9, + Day: 22, + Resolution: "720p", + Source: "WEB", + Codec: []string{"H.264"}, + Container: "", + HDR: []string(nil), + Audio: []string(nil), + AudioChannels: "", + Group: "GROUP", + Region: "", + Language: []string{}, + Proper: false, + Repack: false, + Edition: []string{}, + Cut: []string{}, + Website: "", + Artists: "", + Type: rls.Episode, + LogScore: 0, + Origin: "", + Tags: []string{}, + ReleaseTags: "", + Freeleech: false, + FreeleechPercent: 0, + Bonus: []string(nil), + Uploader: "", + PreTime: "", + Other: []string{}, + RawCookie: "", + AdditionalSizeCheckRequired: false, + AdditionalUploaderCheckRequired: false, + FilterID: 0, + Filter: (*domain.Filter)(nil), + ActionStatus: []domain.ReleaseActionStatus(nil), + }, }, { name: "time_parse", @@ -208,7 +389,68 @@ func TestRSSJob_processItem(t *testing.T) { }, }, }}, - want: &domain.Release{ID: 0, FilterStatus: "PENDING", Rejections: []string{}, Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, FilterName: "", Protocol: "torrent", Implementation: "RSS", AnnounceType: domain.AnnounceTypeNew, Timestamp: now, MagnetURI: "magnet:?xt=this-not-a-valid-magnet", GroupID: "", TorrentID: "", DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", TorrentTmpFile: "", TorrentDataRawBytes: []uint8(nil), TorrentHash: "", TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", Size: 0, Title: "Some Release Title", Description: "Category: Example", Category: "", Season: 0, Episode: 0, Year: 2022, Month: 9, Day: 22, Resolution: "720p", Source: "WEB", Codec: []string{"H.264"}, Container: "", HDR: []string(nil), Audio: []string(nil), AudioChannels: "", Group: "GROUP", Region: "", Language: nil, Proper: false, Repack: false, Website: "", Artists: "", Type: "episode", LogScore: 0, Origin: "", Tags: []string{}, ReleaseTags: "", Freeleech: false, FreeleechPercent: 0, Bonus: []string(nil), Uploader: "", PreTime: "", Other: []string(nil), RawCookie: "", AdditionalSizeCheckRequired: false, FilterID: 0, Filter: (*domain.Filter)(nil), ActionStatus: []domain.ReleaseActionStatus(nil)}, + want: &domain.Release{ + ID: 0, + FilterStatus: "PENDING", + Rejections: []string{}, + Indexer: domain.IndexerMinimal{0, "Mock Feed", "mock-feed", "Mock Indexer"}, + FilterName: "", + Protocol: "torrent", + Implementation: "RSS", + AnnounceType: domain.AnnounceTypeNew, + Timestamp: now, + GroupID: "", + TorrentID: "", + DownloadURL: "https://fake-feed.com/details.php?id=00000&hit=1", + MagnetURI: "magnet:?xt=this-not-a-valid-magnet", + TorrentTmpFile: "", + TorrentDataRawBytes: []uint8(nil), + TorrentHash: "", + TorrentName: "Some.Release.Title.2022.09.22.720p.WEB.h264-GROUP", + NormalizedHash: "edfbe552ccde335f34b801e15930bc35", + Size: 0, + Title: "Some Release Title", + Description: "Category: Example", + Category: "", + Season: 0, + Episode: 0, + Year: 2022, + Month: 9, + Day: 22, + Resolution: "720p", + Source: "WEB", + Codec: []string{"H.264"}, + Container: "", + HDR: []string(nil), + Audio: []string(nil), + AudioChannels: "", + Group: "GROUP", + Region: "", + Language: []string{}, + Proper: false, + Repack: false, + Edition: []string{}, + Cut: []string{}, + Website: "", + Artists: "", + Type: rls.Episode, + LogScore: 0, + Origin: "", + Tags: []string{}, + ReleaseTags: "", + Freeleech: false, + FreeleechPercent: 0, + Bonus: []string(nil), + Uploader: "", + PreTime: "", + Other: []string{}, + RawCookie: "", + AdditionalSizeCheckRequired: false, + AdditionalUploaderCheckRequired: false, + FilterID: 0, + Filter: (*domain.Filter)(nil), + ActionStatus: []domain.ReleaseActionStatus(nil), + }, }, } for _, tt := range tests { diff --git a/internal/filter/service.go b/internal/filter/service.go index 95740ea..47e11a9 100644 --- a/internal/filter/service.go +++ b/internal/filter/service.go @@ -46,6 +46,7 @@ type Service interface { AdditionalRecordLabelCheck(ctx context.Context, f *domain.Filter, release *domain.Release) (bool, error) CheckSmartEpisodeCanDownload(ctx context.Context, params *domain.SmartEpisodeParams) (bool, error) GetDownloadsByFilterId(ctx context.Context, filterID int) (*domain.FilterDownloads, error) + CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error) } type service struct { @@ -374,6 +375,8 @@ func (s *service) Delete(ctx context.Context, filterID int) error { func (s *service) CheckFilter(ctx context.Context, f *domain.Filter, release *domain.Release) (bool, error) { l := s.log.With().Str("method", "CheckFilter").Logger() + l.Debug().Msgf("checking filter: %s with release %s", f.Name, release.TorrentName) + l.Trace().Msgf("checking filter: %s %+v", f.Name, f) l.Trace().Msgf("checking filter: %s for release: %+v", f.Name, release) @@ -393,110 +396,130 @@ func (s *service) CheckFilter(ctx context.Context, f *domain.Filter, release *do return false, nil } - if matchedFilter { - // smartEpisode check - if f.SmartEpisode { - params := &domain.SmartEpisodeParams{ - Title: release.Title, - Season: release.Season, - Episode: release.Episode, - Year: release.Year, - Month: release.Month, - Day: release.Day, - Repack: release.Repack, - Proper: release.Proper, - Group: release.Group, - } - canDownloadShow, err := s.CheckSmartEpisodeCanDownload(ctx, params) - if err != nil { - l.Trace().Msgf("failed smart episode check: %s", f.Name) - return false, nil - } - - if !canDownloadShow { - l.Trace().Msgf("failed smart episode check: %s", f.Name) - - if params.IsDailyEpisode() { - f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day), fmt.Sprintf("expected newer than (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day)) - } else { - f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) season: %d ep: %d", release.Title, release.Season, release.Episode), fmt.Sprintf("expected newer than (%s) season: %d ep: %d", release.Title, release.Season, release.Episode)) - } - - return false, nil - } - } - - // if matched, do additional size check if needed, attach actions and return the filter - - l.Debug().Msgf("found and matched filter: %s", f.Name) - - // If size constraints are set in a filter and the indexer did not - // announce the size, we need to do an additional out of band size check. - if release.AdditionalSizeCheckRequired { - l.Debug().Msgf("(%s) additional size check required", f.Name) - - ok, err := s.AdditionalSizeCheck(ctx, f, release) - if err != nil { - l.Error().Err(err).Msgf("(%s) additional size check error", f.Name) - return false, err - } - - if !ok { - l.Trace().Msgf("(%s) additional size check not matching what filter wanted", f.Name) - return false, nil - } - } - - // check uploader if the indexer supports check via api - if release.AdditionalUploaderCheckRequired { - l.Debug().Msgf("(%s) additional uploader check required", f.Name) - - ok, err := s.AdditionalUploaderCheck(ctx, f, release) - if err != nil { - l.Error().Err(err).Msgf("(%s) additional uploader check error", f.Name) - return false, err - } - - if !ok { - l.Trace().Msgf("(%s) additional uploader check not matching what filter wanted", f.Name) - return false, nil - } - } - - if release.AdditionalRecordLabelCheckRequired { - l.Debug().Msgf("(%s) additional record label check required", f.Name) - - ok, err := s.AdditionalRecordLabelCheck(ctx, f, release) - if err != nil { - l.Error().Err(err).Msgf("(%s) additional record label check error", f.Name) - return false, err - } - - if !ok { - l.Trace().Msgf("(%s) additional record label check not matching what filter wanted", f.Name) - return false, nil - } - } - - // run external filters - if f.External != nil { - externalOk, err := s.RunExternalFilters(ctx, f, f.External, release) - if err != nil { - l.Error().Err(err).Msgf("(%s) external filter check error", f.Name) - return false, err - } - - if !externalOk { - l.Debug().Msgf("(%s) external filter check not matching what filter wanted", f.Name) - return false, nil - } - } - - return true, nil + if !matchedFilter { + // if no match, return nil + return false, nil } - // if no match, return nil - return false, nil + // smartEpisode check + if f.SmartEpisode { + params := &domain.SmartEpisodeParams{ + Title: release.Title, + Season: release.Season, + Episode: release.Episode, + Year: release.Year, + Month: release.Month, + Day: release.Day, + Repack: release.Repack, + Proper: release.Proper, + Group: release.Group, + } + canDownloadShow, err := s.CheckSmartEpisodeCanDownload(ctx, params) + if err != nil { + l.Trace().Msgf("failed smart episode check: %s", f.Name) + return false, nil + } + + if !canDownloadShow { + l.Trace().Msgf("failed smart episode check: %s", f.Name) + if params.IsDailyEpisode() { + f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day), fmt.Sprintf("expected newer than (%s) daily: %d-%d-%d", release.Title, release.Year, release.Month, release.Day)) + } else { + f.RejectReasons.Add("smart episode", fmt.Sprintf("not new (%s) season: %d ep: %d", release.Title, release.Season, release.Episode), fmt.Sprintf("expected newer than (%s) season: %d ep: %d", release.Title, release.Season, release.Episode)) + } + return false, nil + } + } + + // check duplicates + if f.DuplicateHandling != nil { + l.Debug().Msgf("(%s) check is duplicate with profile %s", f.Name, f.DuplicateHandling.Name) + + release.SkipDuplicateProfileID = f.DuplicateHandling.ID + release.SkipDuplicateProfileName = f.DuplicateHandling.Name + + isDuplicate, err := s.CheckIsDuplicateRelease(ctx, f.DuplicateHandling, release) + if err != nil { + return false, errors.Wrap(err, "error finding duplicate handle") + } + + if isDuplicate { + l.Debug().Msgf("filter %s rejected release %q as duplicate with profile %q", f.Name, release.TorrentName, f.DuplicateHandling.Name) + f.RejectReasons.Add("duplicate", "duplicate", "not duplicate") + + // let it continue so external filters can trigger checks + //return false, nil + release.IsDuplicate = true + } + } + + // if matched, do additional size check if needed, attach actions and return the filter + + l.Debug().Msgf("found and matched filter: %s", f.Name) + + // If size constraints are set in a filter and the indexer did not + // announce the size, we need to do an additional out of band size check. + if release.AdditionalSizeCheckRequired { + l.Debug().Msgf("(%s) additional size check required", f.Name) + + ok, err := s.AdditionalSizeCheck(ctx, f, release) + if err != nil { + l.Error().Err(err).Msgf("(%s) additional size check error", f.Name) + return false, err + } + + if !ok { + l.Trace().Msgf("(%s) additional size check not matching what filter wanted", f.Name) + return false, nil + } + } + + // check uploader if the indexer supports check via api + if release.AdditionalUploaderCheckRequired { + l.Debug().Msgf("(%s) additional uploader check required", f.Name) + + ok, err := s.AdditionalUploaderCheck(ctx, f, release) + if err != nil { + l.Error().Err(err).Msgf("(%s) additional uploader check error", f.Name) + return false, err + } + + if !ok { + l.Trace().Msgf("(%s) additional uploader check not matching what filter wanted", f.Name) + return false, nil + } + } + + if release.AdditionalRecordLabelCheckRequired { + l.Debug().Msgf("(%s) additional record label check required", f.Name) + + ok, err := s.AdditionalRecordLabelCheck(ctx, f, release) + if err != nil { + l.Error().Err(err).Msgf("(%s) additional record label check error", f.Name) + return false, err + } + + if !ok { + l.Trace().Msgf("(%s) additional record label check not matching what filter wanted", f.Name) + return false, nil + } + } + + // run external filters + if f.External != nil { + externalOk, err := s.RunExternalFilters(ctx, f, f.External, release) + if err != nil { + l.Error().Err(err).Msgf("(%s) external filter check error", f.Name) + return false, err + } + + if !externalOk { + l.Debug().Msgf("(%s) external filter check not matching what filter wanted", f.Name) + return false, nil + } + } + + return true, nil } // AdditionalSizeCheck performs additional out-of-band checks to determine the @@ -732,10 +755,18 @@ func (s *service) CheckSmartEpisodeCanDownload(ctx context.Context, params *doma return s.releaseRepo.CheckSmartEpisodeCanDownload(ctx, params) } +func (s *service) CheckIsDuplicateRelease(ctx context.Context, profile *domain.DuplicateReleaseProfile, release *domain.Release) (bool, error) { + return s.releaseRepo.CheckIsDuplicateRelease(ctx, profile, release) +} + func (s *service) RunExternalFilters(ctx context.Context, f *domain.Filter, externalFilters []domain.FilterExternal, release *domain.Release) (ok bool, err error) { defer func() { // try recover panic if anything went wrong with the external filter checks errors.RecoverPanic(recover(), &err) + if err != nil { + s.log.Error().Err(err).Msgf("filter %s external filter check panic", f.Name) + ok = false + } }() // sort filters by index diff --git a/internal/http/action.go b/internal/http/action.go index 20ffaf5..e747ec4 100644 --- a/internal/http/action.go +++ b/internal/http/action.go @@ -17,7 +17,7 @@ import ( type actionService interface { List(ctx context.Context) ([]domain.Action, error) - Store(ctx context.Context, action domain.Action) (*domain.Action, error) + Store(ctx context.Context, action *domain.Action) error Delete(ctx context.Context, req *domain.DeleteActionRequest) error ToggleEnabled(actionID int) error } @@ -56,35 +56,35 @@ func (h actionHandler) getActions(w http.ResponseWriter, r *http.Request) { } func (h actionHandler) storeAction(w http.ResponseWriter, r *http.Request) { - var data domain.Action + var data *domain.Action if err := json.NewDecoder(r.Body).Decode(&data); err != nil { h.encoder.Error(w, err) return } - action, err := h.service.Store(r.Context(), data) + err := h.service.Store(r.Context(), data) if err != nil { h.encoder.Error(w, err) return } - h.encoder.StatusResponse(w, http.StatusCreated, action) + h.encoder.StatusResponse(w, http.StatusCreated, data) } func (h actionHandler) updateAction(w http.ResponseWriter, r *http.Request) { - var data domain.Action + var data *domain.Action if err := json.NewDecoder(r.Body).Decode(&data); err != nil { h.encoder.Error(w, err) return } - action, err := h.service.Store(r.Context(), data) + err := h.service.Store(r.Context(), data) if err != nil { h.encoder.Error(w, err) return } - h.encoder.StatusResponse(w, http.StatusCreated, action) + h.encoder.StatusResponse(w, http.StatusCreated, data) } func (h actionHandler) deleteAction(w http.ResponseWriter, r *http.Request) { diff --git a/internal/http/release.go b/internal/http/release.go index 2d36f75..be30c81 100644 --- a/internal/http/release.go +++ b/internal/http/release.go @@ -25,6 +25,10 @@ type releaseService interface { Delete(ctx context.Context, req *domain.DeleteReleaseRequest) error Retry(ctx context.Context, req *domain.ReleaseActionRetryReq) error ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error + + StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error + FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) + DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error } type releaseHandler struct { @@ -52,6 +56,13 @@ func (h releaseHandler) Routes(r chi.Router) { r.Get("/", h.getReleaseByID) r.Post("/actions/{actionStatusID}/retry", h.retryAction) }) + + r.Route("/profiles/duplicate", func(r chi.Router) { + r.Get("/", h.findReleaseProfileDuplicate) + r.Post("/", h.storeReleaseProfileDuplicate) + + r.Delete("/{profileId}", h.deleteReleaseProfileDuplicate) + }) } func (h releaseHandler) findReleases(w http.ResponseWriter, r *http.Request) { @@ -309,3 +320,55 @@ func (h releaseHandler) retryAction(w http.ResponseWriter, r *http.Request) { h.encoder.NoContent(w) } + +func (h releaseHandler) storeReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) { + var data *domain.DuplicateReleaseProfile + + if err := json.NewDecoder(r.Body).Decode(&data); err != nil { + h.encoder.Error(w, err) + return + } + + if err := h.service.StoreReleaseProfileDuplicate(r.Context(), data); err != nil { + h.encoder.Error(w, err) + return + } + + h.encoder.StatusCreatedData(w, data) +} + +func (h releaseHandler) findReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) { + profiles, err := h.service.FindDuplicateReleaseProfiles(r.Context()) + if err != nil { + h.encoder.StatusResponse(w, http.StatusInternalServerError, map[string]interface{}{ + "code": "INTERNAL_SERVER_ERROR", + "message": err.Error(), + }) + return + } + + //ret := struct { + // Data []*domain.DuplicateReleaseProfile `json:"data"` + //}{ + // Data: profiles, + //} + + h.encoder.StatusResponse(w, http.StatusOK, profiles) +} + +func (h releaseHandler) deleteReleaseProfileDuplicate(w http.ResponseWriter, r *http.Request) { + //profileIdParam := chi.URLParam(r, "releaseId") + + profileId, err := strconv.Atoi(chi.URLParam(r, "profileId")) + if err != nil { + h.encoder.StatusError(w, http.StatusBadRequest, err) + return + } + + if err := h.service.DeleteReleaseProfileDuplicate(r.Context(), int64(profileId)); err != nil { + h.encoder.Error(w, err) + return + } + + h.encoder.NoContent(w) +} diff --git a/internal/release/service.go b/internal/release/service.go index 3f84684..f1eee9a 100644 --- a/internal/release/service.go +++ b/internal/release/service.go @@ -32,6 +32,10 @@ type Service interface { ProcessMultiple(releases []*domain.Release) ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error Retry(ctx context.Context, req *domain.ReleaseActionRetryReq) error + + StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error + FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) + DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error } type actionClientTypeKey struct { @@ -94,6 +98,18 @@ func (s *service) Delete(ctx context.Context, req *domain.DeleteReleaseRequest) return s.repo.Delete(ctx, req) } +func (s *service) FindDuplicateReleaseProfiles(ctx context.Context) ([]*domain.DuplicateReleaseProfile, error) { + return s.repo.FindDuplicateReleaseProfiles(ctx) +} + +func (s *service) StoreReleaseProfileDuplicate(ctx context.Context, profile *domain.DuplicateReleaseProfile) error { + return s.repo.StoreDuplicateProfile(ctx, profile) +} + +func (s *service) DeleteReleaseProfileDuplicate(ctx context.Context, id int64) error { + return s.repo.DeleteReleaseProfileDuplicate(ctx, id) +} + func (s *service) ProcessManual(ctx context.Context, req *domain.ReleaseProcessReq) error { // get indexer definition with data def, err := s.indexerSvc.GetMappedDefinitionByName(req.IndexerIdentifier) @@ -183,8 +199,6 @@ func (s *service) Process(release *domain.Release) { s.log.Error().Err(err).Msgf("release.Process: error processing filters for indexer: %s", release.Indexer.Name) return } - - return } func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, release *domain.Release) error { @@ -201,6 +215,11 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, release.FilterName = f.Name release.FilterID = f.ID + // reset IsDuplicate + release.IsDuplicate = false + release.SkipDuplicateProfileID = 0 + release.SkipDuplicateProfileName = "" + // test filter match, err := s.filterSvc.CheckFilter(ctx, f, release) if err != nil { @@ -208,10 +227,10 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, return err } - if !match { + if !match || f.RejectReasons.Len() > 0 { l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s, no match. rejections: %s", release.Indexer.Name, release.FilterName, release.TorrentName, f.RejectReasons.String()) - l.Debug().Msgf("filter %s rejected release: %s", f.Name, release.TorrentName) + l.Debug().Msgf("filter %s rejected release: %s with reasons: %s", f.Name, release.TorrentName, f.RejectReasons.String()) continue } @@ -231,13 +250,6 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, continue } - // sleep for the delay period specified in the filter before running actions - delay := release.Filter.Delay - if delay > 0 { - l.Debug().Msgf("release.Process: delaying processing of '%s' (%s) for %s by %d seconds as specified in the filter", release.TorrentName, release.FilterName, release.Indexer.Name, delay) - time.Sleep(time.Duration(delay) * time.Second) - } - // save release here to only save those with rejections from actions instead of all releases if release.ID == 0 { release.FilterStatus = domain.ReleaseStatusFilterApproved @@ -251,24 +263,40 @@ func (s *service) processFilters(ctx context.Context, filters []*domain.Filter, var rejections []string // run actions (watchFolder, test, exec, qBittorrent, Deluge, arr etc.) - for _, act := range actions { + for idx, act := range actions { // only run enabled actions if !act.Enabled { l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s action '%s' not enabled, skip", release.Indexer.Name, release.FilterName, release.TorrentName, act.Name) continue } + // add action status as pending + actionStatus := domain.NewReleaseActionStatus(act, release) + + if err := s.StoreReleaseActionStatus(ctx, actionStatus); err != nil { + s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName) + } + + if idx == 0 { + // sleep for the delay period specified in the filter before running actions + delay := release.Filter.Delay + if delay > 0 { + l.Debug().Msgf("release.Process: delaying processing of '%s' (%s) for %s by %d seconds as specified in the filter", release.TorrentName, release.FilterName, release.Indexer.Name, delay) + time.Sleep(time.Duration(delay) * time.Second) + } + } + l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s , run action: %s", release.Indexer.Name, release.FilterName, release.TorrentName, act.Name) // keep track of action clients to avoid sending the same thing all over again _, tried := triedActionClients[actionClientTypeKey{Type: act.Type, ClientID: act.ClientID}] if tried { - l.Trace().Msgf("release.Process: indexer: %s, filter: %s release: %s action client already tried, skip", release.Indexer.Name, release.FilterName, release.TorrentName) + l.Debug().Msgf("release.Process: indexer: %s, filter: %s release: %s action client already tried, skip", release.Indexer.Name, release.FilterName, release.TorrentName) continue } // run action - status, err := s.runAction(ctx, act, release) + status, err := s.runAction(ctx, act, release, actionStatus) if err != nil { l.Error().Err(err).Msgf("release.Process: error running actions for filter: %s", release.FilterName) //continue @@ -320,13 +348,13 @@ func (s *service) ProcessMultiple(releases []*domain.Release) { } } -func (s *service) runAction(ctx context.Context, action *domain.Action, release *domain.Release) (*domain.ReleaseActionStatus, error) { +func (s *service) runAction(ctx context.Context, action *domain.Action, release *domain.Release, status *domain.ReleaseActionStatus) (*domain.ReleaseActionStatus, error) { // add action status as pending - status := domain.NewReleaseActionStatus(action, release) - - if err := s.StoreReleaseActionStatus(ctx, status); err != nil { - s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName) - } + //status := domain.NewReleaseActionStatus(action, release) + // + //if err := s.StoreReleaseActionStatus(ctx, status); err != nil { + // s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName) + //} rejections, err := s.actionSvc.RunAction(ctx, action, release) if err != nil { @@ -351,7 +379,14 @@ func (s *service) runAction(ctx context.Context, action *domain.Action, release } func (s *service) retryAction(ctx context.Context, action *domain.Action, release *domain.Release) error { - actionStatus, err := s.runAction(ctx, action, release) + // add action status as pending + status := domain.NewReleaseActionStatus(action, release) + + if err := s.StoreReleaseActionStatus(ctx, status); err != nil { + s.log.Error().Err(err).Msgf("release.runAction: error storing action for filter: %s", release.FilterName) + } + + actionStatus, err := s.runAction(ctx, action, release, status) if err != nil { s.log.Error().Err(err).Msgf("release.retryAction: error running actions for filter: %s", release.FilterName) diff --git a/web/src/api/APIClient.ts b/web/src/api/APIClient.ts index 5139b17..4f6a390 100644 --- a/web/src/api/APIClient.ts +++ b/web/src/api/APIClient.ts @@ -498,7 +498,16 @@ export const APIClient = { }, replayAction: (releaseId: number, actionId: number) => appClient.Post( `api/release/${releaseId}/actions/${actionId}/retry` - ) + ), + profiles: { + duplicates: { + list: () => appClient.Get(`api/release/profiles/duplicate`), + delete: (id: number) => appClient.Delete(`api/release/profiles/duplicate/${id}`), + store: (profile: ReleaseProfileDuplicate) => appClient.Post(`api/release/profiles/duplicate`, { + body: profile + }), + } + } }, updates: { check: () => appClient.Get("api/updates/check"), diff --git a/web/src/api/queries.ts b/web/src/api/queries.ts index 37d8fc7..39924ed 100644 --- a/web/src/api/queries.ts +++ b/web/src/api/queries.ts @@ -12,7 +12,7 @@ import { FilterKeys, IndexerKeys, IrcKeys, ListKeys, NotificationKeys, ProxyKeys, - ReleaseKeys, + ReleaseKeys, ReleaseProfileDuplicateKeys, SettingsKeys } from "@api/query_keys"; import { ColumnFilter } from "@tanstack/react-table"; @@ -165,6 +165,14 @@ export const ReleasesIndexersQueryOptions = () => staleTime: Infinity }); +export const ReleaseProfileDuplicateList = () => + queryOptions({ + queryKey: ReleaseProfileDuplicateKeys.lists(), + queryFn: () => APIClient.release.profiles.duplicates.list(), + staleTime: 5000, + refetchOnWindowFocus: true, + }); + export const ProxiesQueryOptions = () => queryOptions({ queryKey: ProxyKeys.lists(), diff --git a/web/src/api/query_keys.ts b/web/src/api/query_keys.ts index 5614e1f..95a9ac0 100644 --- a/web/src/api/query_keys.ts +++ b/web/src/api/query_keys.ts @@ -35,6 +35,13 @@ export const ReleaseKeys = { latestActivity: () => [...ReleaseKeys.all, "latest-activity"] as const, }; +export const ReleaseProfileDuplicateKeys = { + all: ["releaseProfileDuplicate"] as const, + lists: () => [...ReleaseProfileDuplicateKeys.all, "list"] as const, + details: () => [...ReleaseProfileDuplicateKeys.all, "detail"] as const, + detail: (id: number) => [...ReleaseProfileDuplicateKeys.details(), id] as const, +}; + export const ApiKeys = { all: ["api_keys"] as const, lists: () => [...ApiKeys.all, "list"] as const, diff --git a/web/src/components/inputs/select.tsx b/web/src/components/inputs/select.tsx index 42bbcba..f8698ab 100644 --- a/web/src/components/inputs/select.tsx +++ b/web/src/components/inputs/select.tsx @@ -251,7 +251,7 @@ export function DownloadClientSelect({ export interface SelectFieldOption { label: string; - value: string; + value: string | number | null; } export interface SelectFieldProps { @@ -293,7 +293,7 @@ export const Select = ({ onChange={(value) => setFieldValue(field.name, value)} > {({ open }) => ( - <> +
- + )} )} diff --git a/web/src/forms/_shared.ts b/web/src/forms/_shared.ts new file mode 100644 index 0000000..d307c09 --- /dev/null +++ b/web/src/forms/_shared.ts @@ -0,0 +1,15 @@ +/* + * Copyright (c) 2021 - 2024, Ludvig Lundgren and the autobrr contributors. + * SPDX-License-Identifier: GPL-2.0-or-later + */ + +export interface AddFormProps { + isOpen: boolean; + toggle: () => void; +} + +export interface UpdateFormProps { + isOpen: boolean; + toggle: () => void; + data: T; +} diff --git a/web/src/forms/filters/FilterAddForm.tsx b/web/src/forms/filters/FilterAddForm.tsx index 9f24e0d..3345d1d 100644 --- a/web/src/forms/filters/FilterAddForm.tsx +++ b/web/src/forms/filters/FilterAddForm.tsx @@ -16,14 +16,9 @@ import { FilterKeys } from "@api/query_keys"; import { DEBUG } from "@components/debug"; import { toast } from "@components/hot-toast"; import Toast from "@components/notifications/Toast"; +import { AddFormProps } from "@forms/_shared"; - -interface filterAddFormProps { - isOpen: boolean; - toggle: () => void; -} - -export function FilterAddForm({ isOpen, toggle }: filterAddFormProps) { +export function FilterAddForm({ isOpen, toggle }: AddFormProps) { const inputRef = useRef(null) const queryClient = useQueryClient(); const navigate = useNavigate(); diff --git a/web/src/forms/settings/APIKeyAddForm.tsx b/web/src/forms/settings/APIKeyAddForm.tsx index 7c2a975..ae4350a 100644 --- a/web/src/forms/settings/APIKeyAddForm.tsx +++ b/web/src/forms/settings/APIKeyAddForm.tsx @@ -15,13 +15,9 @@ import { ApiKeys } from "@api/query_keys"; import { DEBUG } from "@components/debug"; import { toast } from "@components/hot-toast"; import Toast from "@components/notifications/Toast"; +import { AddFormProps } from "@forms/_shared"; -interface apiKeyAddFormProps { - isOpen: boolean; - toggle: () => void; -} - -export function APIKeyAddForm({ isOpen, toggle }: apiKeyAddFormProps) { +export function APIKeyAddForm({ isOpen, toggle }: AddFormProps) { const queryClient = useQueryClient(); const mutation = useMutation({ diff --git a/web/src/forms/settings/DownloadClientForms.tsx b/web/src/forms/settings/DownloadClientForms.tsx index 87cfdb5..3c39c1a 100644 --- a/web/src/forms/settings/DownloadClientForms.tsx +++ b/web/src/forms/settings/DownloadClientForms.tsx @@ -27,6 +27,7 @@ import { } from "@components/inputs"; import { DocsLink, ExternalLink } from "@components/ExternalLink"; import { SelectFieldBasic } from "@components/inputs/select_wide"; +import { AddFormProps, UpdateFormProps } from "@forms/_shared"; interface InitialValuesSettings { basic?: { @@ -691,12 +692,7 @@ function DownloadClientFormButtons({ ); } -interface formProps { - isOpen: boolean; - toggle: () => void; -} - -export function DownloadClientAddForm({ isOpen, toggle }: formProps) { +export function DownloadClientAddForm({ isOpen, toggle }: AddFormProps) { const [isTesting, setIsTesting] = useState(false); const [isSuccessfulTest, setIsSuccessfulTest] = useState(false); const [isErrorTest, setIsErrorTest] = useState(false); @@ -856,13 +852,7 @@ export function DownloadClientAddForm({ isOpen, toggle }: formProps) { ); } -interface updateFormProps { - isOpen: boolean; - toggle: () => void; - client: DownloadClient; -} - -export function DownloadClientUpdateForm({ client, isOpen, toggle }: updateFormProps) { +export function DownloadClientUpdateForm({ isOpen, toggle, data: client}: UpdateFormProps) { const [isTesting, setIsTesting] = useState(false); const [isSuccessfulTest, setIsSuccessfulTest] = useState(false); const [isErrorTest, setIsErrorTest] = useState(false); diff --git a/web/src/forms/settings/FeedForms.tsx b/web/src/forms/settings/FeedForms.tsx index bed1f0e..172c561 100644 --- a/web/src/forms/settings/FeedForms.tsx +++ b/web/src/forms/settings/FeedForms.tsx @@ -18,13 +18,7 @@ import { componentMapType } from "./DownloadClientForms"; import { sleep } from "@utils"; import { ImplementationBadges } from "@screens/settings/Indexer"; import { FeedDownloadTypeOptions } from "@domain/constants"; - - -interface UpdateProps { - isOpen: boolean; - toggle: () => void; - feed: Feed; -} +import { UpdateFormProps } from "@forms/_shared"; interface InitialValues { id: number; @@ -41,7 +35,8 @@ interface InitialValues { settings: FeedSettings; } -export function FeedUpdateForm({ isOpen, toggle, feed }: UpdateProps) { +export function FeedUpdateForm({ isOpen, toggle, data}: UpdateFormProps) { + const feed = data; const [isTesting, setIsTesting] = useState(false); const [isTestSuccessful, setIsSuccessfulTest] = useState(false); const [isTestError, setIsErrorTest] = useState(false); diff --git a/web/src/forms/settings/IndexerForms.tsx b/web/src/forms/settings/IndexerForms.tsx index bc894de..c4eaa5c 100644 --- a/web/src/forms/settings/IndexerForms.tsx +++ b/web/src/forms/settings/IndexerForms.tsx @@ -25,6 +25,7 @@ import { FeedDownloadTypeOptions } from "@domain/constants"; import { DocsLink } from "@components/ExternalLink"; import * as common from "@components/inputs/common"; import { SelectField } from "@forms/settings/IrcForms"; +import { AddFormProps, UpdateFormProps } from "@forms/_shared"; // const isRequired = (message: string) => (value?: string | undefined) => (!!value ? undefined : message); @@ -255,12 +256,7 @@ type SelectValue = { value: string; }; -export interface AddProps { - isOpen: boolean; - toggle: () => void; -} - -export function IndexerAddForm({ isOpen, toggle }: AddProps) { +export function IndexerAddForm({ isOpen, toggle }: AddFormProps) { const [indexer, setIndexer] = useState({} as IndexerDefinition); const queryClient = useQueryClient(); @@ -729,13 +725,7 @@ interface IndexerUpdateInitialValues { } } -interface UpdateProps { - isOpen: boolean; - toggle: () => void; - indexer: IndexerDefinition; -} - -export function IndexerUpdateForm({ isOpen, toggle, indexer }: UpdateProps) { +export function IndexerUpdateForm({ isOpen, toggle, data: indexer }: UpdateFormProps) { const queryClient = useQueryClient(); const proxies = useQuery(ProxiesQueryOptions()); diff --git a/web/src/forms/settings/IrcForms.tsx b/web/src/forms/settings/IrcForms.tsx index d00a2bf..135e00d 100644 --- a/web/src/forms/settings/IrcForms.tsx +++ b/web/src/forms/settings/IrcForms.tsx @@ -22,6 +22,7 @@ import Toast from "@components/notifications/Toast"; import * as common from "@components/inputs/common"; import { classNames } from "@utils"; import { ProxiesQueryOptions } from "@api/queries"; +import { AddFormProps, UpdateFormProps } from "@forms/_shared"; interface ChannelsFieldArrayProps { channels: IrcChannel[]; @@ -122,11 +123,6 @@ interface IrcNetworkAddFormValues { channels: IrcChannel[]; } -interface AddFormProps { - isOpen: boolean; - toggle: () => void; -} - export function IrcNetworkAddForm({ isOpen, toggle }: AddFormProps) { const queryClient = useQueryClient(); @@ -275,17 +271,11 @@ interface IrcNetworkUpdateFormValues { proxy_id: number; } -interface IrcNetworkUpdateFormProps { - isOpen: boolean; - toggle: () => void; - network: IrcNetwork; -} - export function IrcNetworkUpdateForm({ isOpen, toggle, - network -}: IrcNetworkUpdateFormProps) { + data: network +}: UpdateFormProps) { const queryClient = useQueryClient(); const proxies = useQuery(ProxiesQueryOptions()); diff --git a/web/src/forms/settings/NotificationForms.tsx b/web/src/forms/settings/NotificationForms.tsx index 9eaca02..0d57cb5 100644 --- a/web/src/forms/settings/NotificationForms.tsx +++ b/web/src/forms/settings/NotificationForms.tsx @@ -23,6 +23,7 @@ import * as common from "@components/inputs/common"; import { NumberFieldWide, PasswordFieldWide, SwitchGroupWide, TextFieldWide } from "@components/inputs"; import { componentMapType } from "./DownloadClientForms"; +import { AddFormProps, UpdateFormProps } from "@forms/_shared"; function FormFieldsDiscord() { return ( @@ -311,12 +312,7 @@ interface NotificationAddFormValues { enabled: boolean; } -interface AddProps { - isOpen: boolean; - toggle: () => void; -} - -export function NotificationAddForm({ isOpen, toggle }: AddProps) { +export function NotificationAddForm({ isOpen, toggle }: AddFormProps) { const queryClient = useQueryClient(); const createMutation = useMutation({ @@ -565,12 +561,6 @@ const EventCheckBoxes = () => ( ); -interface UpdateProps { - isOpen: boolean; - toggle: () => void; - notification: ServiceNotification; -} - interface InitialValues { id: number; enabled: boolean; @@ -587,7 +577,7 @@ interface InitialValues { username?: string } -export function NotificationUpdateForm({ isOpen, toggle, notification }: UpdateProps) { +export function NotificationUpdateForm({ isOpen, toggle, data: notification }: UpdateFormProps) { const queryClient = useQueryClient(); const mutation = useMutation({ diff --git a/web/src/forms/settings/ProxyForms.tsx b/web/src/forms/settings/ProxyForms.tsx index 06882d7..a011f71 100644 --- a/web/src/forms/settings/ProxyForms.tsx +++ b/web/src/forms/settings/ProxyForms.tsx @@ -9,7 +9,7 @@ import { Dialog, DialogPanel, DialogTitle, Transition, TransitionChild } from "@ import { XMarkIcon } from "@heroicons/react/24/solid"; import { useMutation, useQueryClient } from "@tanstack/react-query"; -import { AddProps } from "@forms/settings/IndexerForms"; +import { AddFormProps } from "@forms/_shared"; import { DEBUG } from "@components/debug.tsx"; import { PasswordFieldWide, SwitchGroupWide, TextFieldWide } from "@components/inputs"; import { SelectFieldBasic } from "@components/inputs/select_wide"; @@ -20,7 +20,7 @@ import { toast } from "@components/hot-toast"; import Toast from "@components/notifications/Toast"; import { SlideOver } from "@components/panels"; -export function ProxyAddForm({ isOpen, toggle }: AddProps) { +export function ProxyAddForm({ isOpen, toggle }: AddFormProps) { const queryClient = useQueryClient(); const createMutation = useMutation({ diff --git a/web/src/forms/settings/ReleaseForms.tsx b/web/src/forms/settings/ReleaseForms.tsx new file mode 100644 index 0000000..ac28b11 --- /dev/null +++ b/web/src/forms/settings/ReleaseForms.tsx @@ -0,0 +1,198 @@ +/* + * Copyright (c) 2021 - 2024, Ludvig Lundgren and the autobrr contributors. + * SPDX-License-Identifier: GPL-2.0-or-later + */ + +import { useMutation, useQueryClient } from "@tanstack/react-query"; +import { APIClient } from "@api/APIClient.ts"; +import { ReleaseProfileDuplicateKeys } from "@api/query_keys.ts"; +import { toast } from "@components/hot-toast"; +import Toast from "@components/notifications/Toast.tsx"; +import { SwitchGroupWide, TextFieldWide } from "@components/inputs"; +import { SlideOver } from "@components/panels"; +import { AddFormProps, UpdateFormProps } from "@forms/_shared"; + +export function ReleaseProfileDuplicateAddForm({ isOpen, toggle }: AddFormProps) { + const queryClient = useQueryClient(); + + const addMutation = useMutation({ + mutationFn: (profile: ReleaseProfileDuplicate) => APIClient.release.profiles.duplicates.store(profile), + onSuccess: () => { + queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() }); + toast.custom((t) => ); + + toggle(); + }, + onError: () => { + toast.custom((t) => ); + } + }); + + const onSubmit = (data: unknown) => addMutation.mutate(data as ReleaseProfileDuplicate); + + const initialValues: ReleaseProfileDuplicate = { + id: 0, + name: "", + protocol: false, + release_name: false, + hash: false, + title: false, + sub_title: false, + year: false, + month: false, + day: false, + source: false, + resolution: false, + codec: false, + container: false, + dynamic_range: false, + audio: false, + group: false, + season: false, + episode: false, + website: false, + proper: false, + repack: false, + edition: false, + language: false, + }; + + return ( + + {() => ( +
+ + + + + + + + + + + + + + + + + + + + + + + +
+ )} +
+ ); +} + +export function ReleaseProfileDuplicateUpdateForm({ isOpen, toggle, data: profile }: UpdateFormProps) { + const queryClient = useQueryClient(); + + const storeMutation = useMutation({ + mutationFn: (profile: ReleaseProfileDuplicate) => APIClient.release.profiles.duplicates.store(profile), + onSuccess: () => { + queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() }); + toast.custom((t) => ); + + toggle(); + }, + onError: () => { + toast.custom((t) => ); + } + }); + + const onSubmit = (data: unknown) => storeMutation.mutate(data as ReleaseProfileDuplicate); + + const deleteMutation = useMutation({ + mutationFn: (profileId: number) => APIClient.release.profiles.duplicates.delete(profileId), + onSuccess: () => { + queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.lists() }); + queryClient.invalidateQueries({ queryKey: ReleaseProfileDuplicateKeys.detail(profile.id) }); + + toast.custom((t) => ); + + toggle(); + }, + }); + + const onDelete = () => deleteMutation.mutate(profile.id); + + const initialValues: ReleaseProfileDuplicate = { + id: profile.id, + name: profile.name, + protocol: profile.protocol, + release_name: profile.release_name, + hash: profile.hash, + title: profile.title, + sub_title: profile.sub_title, + year: profile.year, + month: profile.month, + day: profile.day, + source: profile.source, + resolution: profile.resolution, + codec: profile.codec, + container: profile.container, + dynamic_range: profile.dynamic_range, + audio: profile.audio, + group: profile.group, + season: profile.season, + episode: profile.episode, + website: profile.website, + proper: profile.proper, + repack: profile.repack, + edition: profile.edition, + language: profile.language, + }; + + return ( + + {() => ( +
+ + + + + + + + + + + + + + + + + + + + + + + +
+ )} +
+ ); +} diff --git a/web/src/screens/filters/Details.tsx b/web/src/screens/filters/Details.tsx index cb04f6c..e2f79c4 100644 --- a/web/src/screens/filters/Details.tsx +++ b/web/src/screens/filters/Details.tsx @@ -455,7 +455,8 @@ export const FilterDetails = () => { max_leechers: filter.max_leechers, indexers: filter.indexers || [], actions: filter.actions || [], - external: filter.external || [] + external: filter.external || [], + release_profile_duplicate_id: filter.release_profile_duplicate_id, } as Filter} onSubmit={handleSubmit} enableReinitialize={true} diff --git a/web/src/screens/filters/sections/General.tsx b/web/src/screens/filters/sections/General.tsx index e42c1de..1236520 100644 --- a/web/src/screens/filters/sections/General.tsx +++ b/web/src/screens/filters/sections/General.tsx @@ -6,7 +6,7 @@ import { useSuspenseQuery } from "@tanstack/react-query"; import { downloadsPerUnitOptions } from "@domain/constants"; -import { IndexersOptionsQueryOptions } from "@api/queries"; +import { IndexersOptionsQueryOptions, ReleaseProfileDuplicateList } from "@api/queries"; import { DocsLink } from "@components/ExternalLink"; import { FilterLayout, FilterPage, FilterSection } from "./_components"; @@ -16,20 +16,27 @@ import { MultiSelectOption, NumberField, Select, + SelectFieldOption, SwitchGroup, TextField } from "@components/inputs"; import * as CONSTS from "@domain/constants.ts"; - const MapIndexer = (indexer: Indexer) => ( { label: indexer.name, value: indexer.id } as MultiSelectOption ); +const MapReleaseProfile = (profile: ReleaseProfileDuplicate) => ( + { label: profile.name, value: profile.id } as SelectFieldOption +); + export const General = () => { const indexersQuery = useSuspenseQuery(IndexersOptionsQueryOptions()) const indexerOptions = indexersQuery.data && indexersQuery.data.map(MapIndexer) + const duplicateProfilesQuery = useSuspenseQuery(ReleaseProfileDuplicateList()) + const duplicateProfilesOptions = duplicateProfilesQuery.data && duplicateProfilesQuery.data.map(MapReleaseProfile) + // const indexerOptions = data?.map(MapIndexer) ?? []; return ( @@ -129,6 +136,13 @@ export const General = () => { } /> +