[ / / / / / / / / / / / / / ] [ dir / random / cloveros / cuteboys / erp / fast / hydrus / in / strek / tftg ]

/hydrus/ - Hydrus Network

Bug reports, feature requests, and other discussion for the hydrus network.
Name
Email
Subject
Comment *
File
Password (Randomized for file and post deletion; you may also set your own.)
Archive
* = required field[▶ Show post options & limits]
Confused? See the FAQ.
Embed
(replaces files and can be used instead)
Voice recorder Show voice recorder

(the Stop button will be clickable 5 seconds after you press Record)
Options

Allowed file types:jpg, jpeg, gif, png, webm, mp4, swf, pdf
Max filesize is 16 MB.
Max image dimensions are 15000 x 15000.
You may upload 5 per post.


New user? Start here ---> http://hydrusnetwork.github.io/hydrus/

Experienced user with a bit of cash who wants to help out? ---> Patreon

Current to-do list has: 2,017 items

Current big job: Catching up on Qt, MPV, tag work, and small jobs. New poll once things have calmed down.


File: f7061c9177b1b35⋯.gif (144.02 KB, 300x300, 1:1, f7061c9177b1b35d11ed034276….gif)

c5a1fe  No.9115

As planned, I took this week easy for E3. While I did get a little neat work done, there is not enough to justify a release, so there will be no release tomorrow.

I'll do a bit of regular work tomorrow, and then I am back to normal schedule, meaning v311 will be on the 20th.

____________________________
Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

d48979  No.9122

>>9115

What was your favorite least hated thing about this year's E3?

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

6dee4a  No.9129

>>9122

>tomato GOTY

>xbox only

>PC version still has denuvo

Fucking nothing.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

768a36  No.9142

File: 786afc8260327b6⋯.png (206.78 KB, 940x990, 94:99, neural_net_tagging.png)

Hi hydrus dev!

I did a little bit of work and was able to get some basic neural network tagging to work (with the Illustration2Vec model).

It's very unstable at the moment and freezes up everything with WriteSynchronous, but it is able to assign and commit local tags that are somewhat accurate.

You can see my changes at https://github.com/antonpaquin/hydrus

I don't know if this is even on your radar at the moment so no worries if you don't want to get to it right away, but if you're interested I'm happy to clean up / rework / polish what I wrote.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

c5a1fe  No.9145

File: 372ee9b49b8a3d1⋯.jpg (158.33 KB, 1280x1034, 640:517, 372ee9b49b8a3d1e85e00045d5….jpg)

>>9122

The line I am going with is: I am very thankful to the industry for giving me a year to catch up on my backlog. :^)

It was a weaker than normal year for my gaming interests. Pretty much the only space babes were in Smash or otherwise on switch, and I am neither smashfag nor switchfag (single-player offline pc master race). Most of the characters were too hamfisted or ugly for me to be interested. I still had a great time shitposting with other Anons about it, so I am glad I watched.

Metal Wolf Chaos is great news.

Cyberpunk is obviously potentially breddy gud and a possible exception, and the clear winner of the show, although we haven't seen anything real yet. I don't trust them not to compromise and downgrade for actual release, particularly if MS and Sony push for it to work on current gen (or even next gen) consoles. Still, that trailer showed what their world was about in a clear and exciting and vivacious way that the more mainstream West is unwilling to go for.

Cuphead expansion and Nioh 2 look good, but I still haven't got around to the originals, so that is all more backlog-fodder.

The only other one that really piqued my interest was Satisfactory. I am crossing my fingers it is at least somewhat approaching Factorio-level complexity and hasn't got any always-online shit.

I enjoy business news and politics, so that side was interesting. MS seem to be solving some of their problems, Sony seem to have a little hubris, Nintendo shot their wad.

>>9142

Thanks m8, this looks really interesting. I am impressed you were able to get this up and actually integrated into my shit code. I had a shallow look at your code, but yeah I am not ready to get into this properly until the current downloader work is finished. I never dove too deep into the nuts and bolts of Illustration2Vec–does it use Keras, or did you have to alter the I2V db a bit to make it work on that? Is the tag.py in your .tar.gz your file? How is Keras to work with?

I am a complete fucking sperg about integrating code from other people, but if you were able to figure out a really simple module where I could do simple stuff like:

import nntag

nntag.learnimage( path, learning_data )

tag_info = nntag.predictags( path )

sort of stuff, I'd absolutely add it to hydrus. I don't understand the requirements of the learning loop of ML yet, so I'll need to research that first when I get around to this stuff. I am particularly interested in expanding on the I2V example to create our own, personalised tag dbs, built with our own CPU cycles. If you are read up on how they built their db, I'd be interested in a high-level explanation of what is needed to do this stuff and what data goes back and forth. If you are simply pulling from what they provide, I'd still be interested in the brief version of what happens here–what format and how reasonable are the confidence levels you get with the tags? I see you send a 224x224 image at the model–is that based on what they trained the db with, or is there a magic number in that size?

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

b00a8b  No.9147

File: 4b0ba06143e1338⋯.jpg (61.29 KB, 600x921, 200:307, 37614bfe27fe646d8350f52f58….jpg)

>>9122

>Be me

>I always like smash

>I'm looking forward to seeing more about it

<40 MINUTES OF SMASH

Just why.

Felt like one of the weakest Nintendo e3 directs in ages. Why wasnt the details about smash pushed to a separate smash direct? Why not release those details through the treehouse with Sakurai sitting in, which they did anyways? Smash's everyone is here trailer did more than enough for hype, a simple couple notes after words that returning mechanics and to play with characters on e3 to notice tweaks would have been an easy hook to get people playing.

I didn't need to see 40 fucking minutes of it going into every little change for every character and no Metroid Prime 4, no retrostudios title, or anything. I would have even taken more indies showcasing something. It really felt like they had practically nothing and banked it all on Smash. I dont even think the e3 BotW direct was this empty.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

768a36  No.9148

>>9145

I think I can actually do all this via a parsing script, which would require no code changes on your end. I'll update here if I can get that to work.

Illustration2vec uses caffe, but after going through hell to get that installed I decided to use mmdnn to translate illustration2vec into keras. Either of pytorch or tensorflow is probably the best way to distribute, and keras uses a tensorflow backend. I personally like the keras saved model format – it's just one '.h5' file that can load pretty consistently.

Keras is also pretty easy to train with, so if you want to train on your own DBs and you don't want to spend too much time learning machine learning that's what I'd go with.

The "tag.py" and all its dependencies are in the .tar.gz, that's the bit that loads i2v.

I2V was trained on a 224x224 image, so that's what it can tag with, but 224 isn't special on its own. The output is a list of 1539 numbers from 0 to 1, corresponding to the probabilities of the tags they trained on. Most probabilities are near 0 for most images – I've been getting 2-10 above a 0.3 threshold.

For something like i2v where you have a lot of possible tags, you generally need a lot of data. i2v trained on 1.2M examples, but I guess this is within reach for a lot of hydrus users. More complex tags and tags with few examples need a larger training set.

Training from scratch involves first picking the tags you want, then getting numpy images and tag labels for around 100 images at a time. Training on your own cycles is definitely possible, but we'd probably have to do some experimentation or research to figure out what is the minimum number of images per tag to be useful, and what model / parameters to use.

I'll set up the parsing script first, then I can build that module and show you the API.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

768a36  No.9151

File: 37855e4fa864da5⋯.png (49.77 KB, 686x973, 98:139, illust2vec_aya.png)

File: ee8a0435436fba4⋯.jpg (460.48 KB, 1300x1300, 1:1, 1522129239438.jpg)

>>9148

Results:

https://github.com/antonpaquin/Hydrus-Autotagging/tree/master/illust2vec-flask

Classifies with a flask server that drives a parsing script.

Advantage: can work without code changes in current hydrus

Disadvantage: this will only tag one image at a time and takes a few button presses per image

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

b99a25  No.9152

>>9145

Hey hdev, just a note for auto tagging.would you be able to set auto tagging tags to their own category?

It honestly sees to be able to get many of the basic tags for images,

It seems like it has great potential to get basics for tagging done

the example they give being

1girl, twintails, solo, aqua hair, long hair, very long hair, detached sleeves, skirt, necktie, aqua eyes

and it was able to see the image was safe

but we have all seen the meme of the safe for work ai, where it gives things that are completely unsafe nsfw status, and things that are safe nsfw

Having the tags be there and functional but also be a seperate tagging system would allow you to be able to only see what a human has tagged, and potentially also flag to go through what was automatically tagged.

Its fully possible that the ai in time will be able to tag things perfectly, and at that point a human confirmation to commit tags could be done though a ui much like duplicates are, however its 1 tag that you are confirming each time so you are able to rapidly go though them and commit the tags.

also having robo tags add a tag to an image called 'robo tagged' or something makes it so you double check it that it is a fully tagged image rather than just the things ai tagging would notice.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

c5a1fe  No.9164

File: e7f7b0e9a9b4469⋯.jpg (473.41 KB, 1400x979, 1400:979, e7f7b0e9a9b4469d90cfe49a91….jpg)

>>9151

>>9148

Thanks for this info. This looks great stuff–if it is all correct with you, I'll point people at your repos here in the v311 release post if they would like to experiment. I am interested in knowing further how these different systems work well and badly.

For the one-image-at-a-time issue, it has long been a thought to convert the file-lookup system to an automated 'figure out any tags for these 10,000 files in idle time please' maintenance routine, particularly once I had introduced the new bandwidth system, which will help us not DDoS the services we are pulling from. Your use case here is another vote in doing something like this, and generally to generalise the way(s) hydrus can pull tag recommendations from other services and scripts.

I'll reiterate that I can't put any time into this atm, but I am interested in having an ongoing conversation about how I can make these workflows better for you. I am keen to get hydrus and The Imageboard Community into the ML game in the coming years.

>>9152

Yes. In making hydrus, I have come to make a point about putting 'human eyes' in front of certain decisions, especially for large automated systems. Figuring out a good workflow is often as difficult as getting the technical side working. It is easy to fuck up a script, and if that script touches 100,000 files, it can be a huge pain to fix. Tag siblings is a good example of this–it seemed simple when I got in, but it turns out for multiple reasons–human preference and complexity of language and translation issues and simple human mistake and technical complexity at the data and gui levels and low CPU availability at certain critical moments in the sibling processing pipeline–to be much more complicated.

If we gain the ability to harvest millions of new tag mappings based on ML systems or otherwise, I would imagine putting them in a separate low-trust cache on the side of the 'current' mappings–probably just in their own service in the hydrus context–until you approved them. The decision to apply these tags to files or not would start being approved by human feedback, with human eyes seeing every decision, until some critical threshold of success were reached, at which point we could trust it.

I am keen to train our own models, so I can imagine us starting with just a handful of new tags, like say 'character:pepe'. The model could inherit its opinion of what is and isn't pepe based on the existing mappings in your db, and then it could ask you about cases in which it was not sure. After refining its model, it can start suggesting pepe for new files. Again, you say yes, yes, no, yes, no, until you are saying yes enough times in a row for it to be 99.8% or whatever confident about pepe, at which point you can say, 'ok, you are ready to pepe anything'. You can then move onto the next tag to train on.

I'll reiterate that I am not read up on this stuff yet, but if it is possible to separate or even quantize these decisions and how they affect the model, I could even write a ML repository where this info could be shared, or even make it exportable to pngs or something, so Anons could share recognition models for different tags and thus share the workload. "I spent ten hours teaching my client to recognise the top ten 2hus so you don't have to." "Here is an official model that determines the difference between 'medium breasts' and 'large breasts'." "Here is a 'feminine penis' detector." And so on, depending on how all this data shakes out, which kinds of tags it works well and badly for, and how people want to use and share it. The public tag repository has been a great success in a bunch of ways, but I think the future has a lot more sharing metadata about what tags are, and less sharing 'this file has this tag'. I am certain these systems will be complicated and require multiple iterations of work to be useful in broad ways.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

768a36  No.9171

>>9164

>if it is all correct with you

Sure! I think the code in that repo is currently workable enough for some user testing. I'll check back if you do so and try to debug any issues that come up.

>I can't put any time into this atm

Don't worry, stick to your schedule. I'm happy to help whenever you're ready.

Though I don't check 8ch much, so if you're trying to find me try pinging my github or this email.

>"Here is a 'feminine penis' detector."

I want to build this just to freak out the rest of the world

>>9152

>would you be able to set auto tagging tags to their own category?

One solution would be to assign the auto-tags to their own namespace, which I think is possible from the parsing script menus.

>it gives things that are completely unsafe nsfw status, and things that are safe nsfw

Yep, I cherry picked an image where it did decently for the example. "shameimaru aya" is an i2v tag but it missed it in >>9151, and there were a few where it incorrectly guessed safe/questionable/explicit (about 5%).

I'll admit that my ML experience is primarily academic / personal, so it's pretty interesting to see what some real user-facing requirements look like. I'm pretty sure everything you mention is possible, but I'd have to do some reading to tweak the standard process to make it fit (except export to png, unless you like 200MB png's).

I would like to build that "nntag" module, and I would like to make it so that tagging models can be shared, hot-swapped, and trained on a distributed network. Whenever you're ready to think about this stuff, hit me at either of the contacts above and we can talk about how to make it work.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

d91cf0  No.9172

yo hydev, i think the URL downloader broke. im copypasting links from 8chan itself and they give me a "could not find parser for 8chan file URL class" error

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

c5a1fe  No.9174

>>9171

👍

>>9172

Thank you for this report. 8chan direct file urls have a 'file' url class, and the url downloader is lumping them in with parsable url classes rather than just trying to download. I will make sure to have this fixed for v311.

For now, pasting one of the alternate 'file_dl' links, like the (h) or (u) ones for 8ch, should work ok.

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.

d91cf0  No.9179

>>9174

cool, have a nice week hydev

Disclaimer: this post and the subject matter and contents thereof - text, media, or otherwise - do not necessarily reflect the views of the 8kun administration.



[Return][Go to top][Catalog][Nerve Center][Random][Post a Reply]
Delete Post [ ]
[]
[ / / / / / / / / / / / / / ] [ dir / random / cloveros / cuteboys / erp / fast / hydrus / in / strek / tftg ]