>>24219
I recommend you only check this for sites like pixiv, where a 'Post URL' really can produce multiple URLs. If hydrus knows that a Post URL can only produce one file, like for most boorus, it can perform efficient 'already in db'/'previously deleted' file import results, without having to hit the URL, when it has seen that URL before. If you set this checkmark too liberally, hydrus will perform less efficiently when it visits sites more than once, and for no benefit since the URL does not produce multiple files.
>>24220
>>24223
Sorry for the shitty UI around here. Some of the string processing is half multi-line capable but half still on 'one string at a time', so a few panels around here are bad about providing 'example strings' and such to test dialogs. I do intend to improve this. If you 
>>24220 are talking about the normal 'test parse' button of a content parser though, yeah it should deliver everything the formula grabs. If the formula is set to grab the 1st item, that's what it will see, but if you have something like (JSON):
- get the "posts" item
- get the "urls" item
- then get all the items
It should get everything it sees. However, if Ebay is complicated, with dynamically named fields that you have to look up with a more specific 'type' fetch, then I'm sorry but it might be too complicated for hydrus. I have some support for this sort of thing, but nothing nice like a 'temporary variable' system yet, not really. You can sometimes scrape by by saying 'get the nth item' or getting all items and then using a string processing to filter the outputs to just stuff that fits a regex for 'https...' etc...
If you like, post an example of the JSON and we can look at it.
>>24224
Thanks. Yeah, I generally agree. It would be lovely to have the client more duplicate-aware in general at the UI level. One really big overhaul that is still missing here, that I keep meaning to do, is to get thumbnails (i.e. their UI-level media objects) aware of duplicate relationships, which is both a storage fetch problem (I need to load all that shit and store it in a nicer manager for each file on search load), and a content update pipeline problem (every time I make a db level update to a duplicate relationship, which can get a little complicated sometimes, I need to update that UI-level manager pretty promptly.
Once I have thumbnails able to talk about their duplicates without a db hit, I could potentially do a bunch of this stuff since I could fetch an importer's deleted file and go 'ok, here is the king, provide a menu option to show the user the king' or whatever without having to write a new db access routine for every thing we wanted to do. I could also do fun stuff like sort-grouping files by duplicate/alternate group.
Ultimately, I think this is natural. We've got some tech to apply duplicates in powerful new ways, and now the gaps in the UI are showing. I'll keep pushing.
>>24225
Sorry that paste is private! I was too late. Did it look like nice JSON, or did it have some weird characters at the start/end?
I think you can attach a .json file to an 8chan post? I'm not sure, I forget, but I think you can probably just post it here as a json/zip without too much trouble.
>>24233
Ah, thanks for letting me know. Many such cases!
>>24238
Yeah for now I think the simplest solution is to figure out a script or a notepad somewhere where you prepend the IPFS multihashes you want to download with your local gateway address, and then spam those URLs into the hydrus url downloader. You might be able to do that automatically with hydrus with some hacked up gallery downloader, but best to be KISS I think.
Yeah if you make a parser, try and grab/generate-with-a-prefix a valid URL that will simply produce the file, even if it is localhost. Hydrus can handle an URL no problem, although timeouts can be an issue with IPFS. I would like to grab multihashes directly from the IPFS daemon again, but I don't know when I will get it done.
>>24245
Sorry for the trouble. The older a client gets, the more cruft that builds up in the downloader system, and it can cause some crossed wires. Please feel free to delete anything that you don't need any more. I will investigate how the updated parser could have 'jumped' link--if there is an existing link, it isn't supposed to change on update, but I know I made some mistakes, particularly with 'www' domains, in the URL Domain Mask work.