[ / / / / / / / / / ] [ dir / cute / egy / fur / kind / kpop / miku / waifuist / wooo ]

/hydrus/ - Hydrus Network

Bug reports, feature requests, and other discussion for the hydrus network.

Catalog

Name
Email
Subject
Comment *
File
* = required field[▶ Show post options & limits]
Confused? See the FAQ.
Embed
(replaces files and can be used instead)
Options
Password (For file and post deletion.)

Allowed file types:jpg, jpeg, gif, png, webm, mp4, swf, pdf
Max filesize is 12 MB.
Max image dimensions are 10000 x 10000.
You may upload 5 per post.


New user? Start here ---> http://hydrusnetwork.github.io/hydrus/

Current to-do list has: 714 items

Current big job: finishing off duplicate search/filtering workflow


YouTube embed. Click thumbnail to play.

5e0d36 No.4781

windows

zip: https://github.com/hydrusnetwork/hydrus/releases/download/v237/Hydrus.Network.237.-.Windows.-.Extract.only.zip

exe: https://github.com/hydrusnetwork/hydrus/releases/download/v237/Hydrus.Network.237.-.Windows.-.Installer.exe

os x

app: https://github.com/hydrusnetwork/hydrus/releases/download/v237/Hydrus.Network.237.-.OS.X.-.App.dmg

tar.gz: https://github.com/hydrusnetwork/hydrus/releases/download/v237/Hydrus.Network.237.-.OS.X.-.Extract.only.tar.gz

linux

tar.gz: https://github.com/hydrusnetwork/hydrus/releases/download/v237/Hydrus.Network.237.-.Linux.-.Executable.tar.gz

source

tar.gz: https://github.com/hydrusnetwork/hydrus/archive/v237.tar.gz

I had a good week. I fixed some important bugs and overhauled how local files are managed in the db.

If you have a lot of files and tags, updating may take a couple of minutes.

After you update, you'll get a new message about regenerating the autocomplete cache. Please do this when convenient, but bear in mind it can take up to twenty minutes, during which time you can't use the client.

all local files

I have added a new file service, 'all local files', which unifies 'local files' and 'trash'. You still have the old services, but if you want to search them both at once, you can use this new one. It also has an entry in review services. This is helpful on my end for a bunch of code-cleaning reasons, but it will also be useful in the ongoing dupe search stuff and is a big step forward in ultimately supporting multiple 'local files' domains.

As the original 'local files' domain becomes less and less exclusive, and as 'all local files' is confusing alongside it, I've changed its name this week to 'my files'. If you don't like this, you can now change it in manage services! The idea is to make local file domains increasingly editable while keeping 'trash' and 'all local files' static.

One nice part of this change is that the trash timestamp is easier for me to access. Trashed files will now report how long they have been in the trash when you right-click on them, and if the file domain is 'trash', then sorting by newest/oldest will sort by the trash timestamp.

full list

- added 'all local files' service that spans all local file domains

- improved trash service code

- trashed files now report their trashed timestamp in right-click menus

- trash views will sort oldest/newest by the trash timestamp

- renamed 'local files' to 'my files' to reduce initial confusion

- if you don't like 'my files', you can now rename the local files service under manage services!

- improved how some local file service metadata is stored

- cleaned and possibly fixed up some delete code

- cleaned up a bunch of misc service and file service code

- the client is more intelligent about what local files are – you can now 'open externally' trashed files, for instance

- on multiple monitor systems, the new sizing system now bounds itself by the appropriate monitor's dimensions (previously, it was always consulting the primary, I think)

- if an expansion event causes a frame to grow off screen, the new sizing system will now attempt to move it up and left so it is completely visible

- fixed an important bug in the specific service autocomplete cache that was leading to several kinds of miscount–please regen your autocomplete cache at your convenience

- the client can now post additional messages on boot (you'll see one!)

- improved how errors with unusual characters are applied to failed import file status objects

- left double-click on the main gui greyspace now opens the new page chooser as well

- restored session pages now recover more gracefully from missing services

- shortened the 'stop after this many files' phrase, which was mis-sizing downloader panels

- changed several bits of old jank in the 'import options - files' collapsible. it is also now thinner

- misc gui improvements

- updated more of the menu system code

- the selection tags box now sizes its height more reasonably

- fixed 'check now' and 'reset cache' buttons on edit sub dialog

- subscriptions report better file import status on their popup

- misc cleanup

next week

I am not sure. Christmas has snuck up to me, so my plans are a bit out of whack. I would like to do a release on the 28th, but I expect I will only work on some simple little features and cleanup. If I have time, I'll improve the dupe search tree generation code.

Post last edited at

8dabc3 No.4782

>>4781

Hey boss,

Does this release fix the performance degradation that occurs as download import list expands?

Or is that unfixable due to how sqlite databases work.


fb15f4 No.4784

Hey admin, I exchanged some emails with you in the past, topic "Tag Repository not processing" some time ago. 15 days ago I wrote you again with another set of doubs and you didnt reply. I bumped it 6 days ago and nothing.


e1059f No.4787

File: 99e59697eb25d30⋯.webm (1.91 MB, 492x462, 82:77, cockli christmas.opus.webm)

Have a very, merry, Chrismasu!

(I posted the wrong one before)


f3adfc No.4790

Hi dev!

I currently find myself struggling with a subscription, that hits a lot of deleted files on e621. Every 5th encountered error the sync stops and I have to edit subs & force check on dialog ok.

Since these are deleted files, they're not going to come back and from my point of view it doesn't make a lot of sense to count them towards the error threshold.

Ideally I'd want the downloader to ignore these files towards the error count (some kind of configurable "deleted files" handling per booru) and just march on (and, ideally sure, tell me how many files had to be skipped because they were deleted).

Given that'd likely take some work on your end, is there a way to, as a workaround, increase the error count limit from currently 5 to something much higher on my end? I'm prepared to manually sift through the error list to figure out which ones are "real" errors to retry them later.

Thanks for all your work!


bc1c22 No.4791

>>4781

Tried regenerating the autocomplete cache, gave me this

OperationalError
table specific_current_mappings_cache_1_6 already exists
Traceback (most recent call last):
File "/opt/hydrus/include/HydrusDB.py", line 469, in _ProcessJob
elif job_type in ( 'write' ): result = self._Write( action, *args, **kwargs )
File "/opt/hydrus/include/ClientDB.py", line 9380, in _Write
elif action == 'regenerate_ac_cache': result = self._RegenerateACCache( *args, **kwargs )
File "/opt/hydrus/include/ClientDB.py", line 7026, in _RegenerateACCache
self._CacheSpecificMappingsGenerate( file_service_id, tag_service_id )
File "/opt/hydrus/include/ClientDB.py", line 2273, in _CacheSpecificMappingsGenerate
self._c.execute( 'CREATE TABLE ' + cache_current_mappings_table_name + ' ( hash_id INTEGER, namespace_id INTEGER, tag_id INTEGER, PRIMARY KEY( hash_id, namespace_id, tag_id ) ) WITHOUT ROWID;' )
OperationalError: table specific_current_mappings_cache_1_6 already exists

Running on Linux, if that's relevant


5e0d36 No.4796

File: 3e65e21fc06d37a⋯.jpg (68.74 KB, 460x450, 46:45, 3e65e21fc06d37a413097f8942….jpg)

>>4782

I don't remember working specifically on anything like that, so probably not! Can you describe a specific scenario that leads to slowdown? Are you importing from hard drive, or on a gallery, or through a subscription? And how many files are we talking–100 or 10,000?

I can nearly always speed up slow stuff–even if the underlying architecture is having a problem, you just have to interface with it in a different way.

>>4784

I apologise, I think it got lost in the mix and then deleted. Can you resend me the email?

>>4787

𝕸𝖊𝖗𝖗𝖞 𝕮𝖍𝖗𝖎𝖘𝖙𝖒𝖆𝖘!

I did, thanks. Spent nice time with family. I'm now feeling super optimistic about 2017.

>>4790

Thank you for this report. This is deleted files on e621's end, right? Like the page on their site is 'this file was deleted for x reason' and my parser throws an error because it couldn't parse a file url?

I might be able to add a quick fix for this, or it might have to wait for the new downloader system which will have veto conditions to more gracefully catch this stuff.

Or yeah, I can just increase the error threshold.

Can you give me an example search that produces a bunch of these deleted results?

>>4791

Thank you for this report. I'm not sure what's going on here. The direct previous line of that code is supposed to drop that table, which shouldn't be a controversial thing. I have updated that function's error reporting for the next version (which I hope will be tomorrow if I can get enough done today, but otherwise will be put off until next week) to catch any problems with the drop call. Please try to regen your a/c cache again in that and let me know what new error message you get, if any.


1405fc No.4800

>>4796

Exactly, deleted on e621's end.

To be honest I'm not even sure these files show up anymore in a search, probably not. I think for me specifically the problem is, that I created (and synced) a subscription a while ago, when those files still did exist. And now that the downloader is finally getting around to them, they are no longer there.

You can look at a takedown notice for a sample of deleted files:

https://e621.net/takedown/show/4937

As always, thanks for all your work.


8dabc3 No.4801

>>4796

Welcome back devbro, hope you enjoyed your holiday.

I'm not sure whether the degradation is due to algorithm or because of software aging as I do have a tendency to run hydrus for very long periods of time without termination.

Anyway, heres my situation: I'm in the process of scraping all the major boorus using the in-built downloader. I like to import in batches of 100000 as theyre easier to keep track of in the long run.

I've used hydrus for 6 months now. And for a while now I've noticed that hydrus has sort of a "sweet spot" regarding interaction with the file import status menu where everything responds quickly and smoothly. This sweet spot is around n=~20000 objects. Past this number and performs noticeably degrades when interacting with the import status list as the system hangs.

This sweet spot kind of shoots down alternate theories as to why this happens, such as running out of system resources, as something like that would happen gradually and would be felt at all levels.

I'm also definitely not running out of RAM (8GB here).

Now, this is purely conjecture. But it feels to me like hydrus must refresh every other item in the list before acting on the selected object, leading me to believe some sort of O(n^2) type operation is going on. I dont have any hard numbers supporting this, but last week I had to redownload roughly 60000 files that timed out (my own fault) out of a 98000 object list. This operation took roughly 40 hours to complete. This is obviously a blue moon type of scenario, but even normal attrition rates after downloading a batch is usually in the 1k-2k range due to server refusing requests and timeouts, and even these takes roughly an hour or so to clear.


8dabc3 No.4802

>>4801

Sorry, replace "redownload" with "refresh", as in "try again".


8f7911 No.4803

The linux tar.gz just links to the windows .exe, just so you know.


5e0d36 No.4804

File: 627de2f3eb38957⋯.jpg (157.9 KB, 707x1450, 707:1450, 627de2f3eb389570608aa08234….jpg)

>>4800

Great. I have added a different error catch for your case that'll not increment the error count. It'll be in this afternoon's release. Let me know if it doesn't work for you.

>>4801

Ah yeah, I should think it is just down to raw number. I've never tested the list with more than, say, four thousand or so entries.

I will make a note to check how I tell it to refresh its rows and so on–I think I might be telling it to refresh everything on occasion, which is probably what is blatting you.

My feeling is that most of the problem here is probably on the gui side, so I recommend you not leave the import status window open if it has more than 10k or whatever files in it for now. I think the operations on the underlying data (which is just a list) should be snappy up to very high numbers.

>>4803

Thanks–I messed up my post template!


bd819e No.4809

>>4804

I'm near-speechless you included it this quickly. I gave it a whirl over night and it's working fantastically.

Thank you so much! :D




[Return][Go to top][Catalog][Post a Reply]
Delete Post [ ]
[]
[ / / / / / / / / / ] [ dir / cute / egy / fur / kind / kpop / miku / waifuist / wooo ]