added cache folder to pull from if an image exists (to remove needless refetches in the future)
This commit is contained in:
parent
0d5bbfa465
commit
d1d5153874
|
@ -1,6 +1,7 @@
|
||||||
{
|
{
|
||||||
"booru": "e621",
|
"booru": "e621",
|
||||||
"output": "./images/downloaded/",
|
"output": "./images/downloaded/",
|
||||||
|
"images": "./images/cache/",
|
||||||
"limit": 320,
|
"limit": 320,
|
||||||
"concurrency": 4,
|
"concurrency": 4,
|
||||||
"userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.88 Safari/537.36",
|
"userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.88 Safari/537.36",
|
||||||
|
|
0
images/cache/.gitkeep
vendored
Executable file
0
images/cache/.gitkeep
vendored
Executable file
11
src/fetch.js
11
src/fetch.js
|
@ -45,6 +45,7 @@ let config = {
|
||||||
|
|
||||||
output: `./images/downloaded/`, // directory to save your files
|
output: `./images/downloaded/`, // directory to save your files
|
||||||
cache: `./data/cache.json`, // JSON file of cached tags, will speed up processing when used for the renamer script
|
cache: `./data/cache.json`, // JSON file of cached tags, will speed up processing when used for the renamer script
|
||||||
|
images: `./images/cache/`, // total cache of images, will copy if file exists here
|
||||||
|
|
||||||
limit: 320, // how many posts to pull in one go
|
limit: 320, // how many posts to pull in one go
|
||||||
concurrency: 4, // how many file requests to keep in flight at the same time
|
concurrency: 4, // how many file requests to keep in flight at the same time
|
||||||
|
@ -139,6 +140,11 @@ let parse = async () => {
|
||||||
console.log(`Skipping existing file: ${booru.urls.posts}${post.id}`)
|
console.log(`Skipping existing file: ${booru.urls.posts}${post.id}`)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
if ( FS.existsSync(`${config.images}${post.filename}`) ) {
|
||||||
|
console.log(`Copying cached file: ${booru.urls.posts}${post.id}`)
|
||||||
|
FS.copyFileSync(`${config.images}${post.filename}`, `${config.output}${post.filename}`)
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
if ( config.filter ) {
|
if ( config.filter ) {
|
||||||
let filtered = null;
|
let filtered = null;
|
||||||
|
@ -173,6 +179,11 @@ let parse = async () => {
|
||||||
res.body.pipe(dest);
|
res.body.pipe(dest);
|
||||||
dest.on('close', () => {
|
dest.on('close', () => {
|
||||||
console.log(`Downloaded: ${booru.urls.posts}${post.id}`)
|
console.log(`Downloaded: ${booru.urls.posts}${post.id}`)
|
||||||
|
|
||||||
|
if ( FS.existsSync(`${config.images}`) ) {
|
||||||
|
FS.copyFileSync(`${config.output}${post.filename}`, `${config.images}${post.filename}`)
|
||||||
|
}
|
||||||
|
|
||||||
resolve()
|
resolve()
|
||||||
});
|
});
|
||||||
dest.on('error', reject);
|
dest.on('error', reject);
|
||||||
|
|
|
@ -50,6 +50,7 @@ config = {
|
||||||
|
|
||||||
'output': './images/downloaded/', # directory to save your files
|
'output': './images/downloaded/', # directory to save your files
|
||||||
'cache': './data/cache.json', # JSON file of cached tags, will speed up processing when used for the renamer script
|
'cache': './data/cache.json', # JSON file of cached tags, will speed up processing when used for the renamer script
|
||||||
|
'images': './images/cache/', # total cache of images, will copy if file exists here
|
||||||
|
|
||||||
'limit': 320, # how many posts to pull in one go
|
'limit': 320, # how many posts to pull in one go
|
||||||
|
|
||||||
|
@ -127,6 +128,11 @@ def parse():
|
||||||
if os.path.exists(f"{config['output']}{post['filename']}"):
|
if os.path.exists(f"{config['output']}{post['filename']}"):
|
||||||
print(f"Skipping existing file: {booru['urls']['posts']}{post['id']}")
|
print(f"Skipping existing file: {booru['urls']['posts']}{post['id']}")
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
if os.path.exists(f"{config['images']}{post['filename']}"):
|
||||||
|
print(f"Copying cached file: {booru['urls']['posts']}{post['id']}")
|
||||||
|
shutil.copy(os.path.join(config['images'], post['filename']), os.path.join(config['output'], post['filename']))
|
||||||
|
continue
|
||||||
|
|
||||||
if post['url'] is None:
|
if post['url'] is None:
|
||||||
print(f"Skipping file that requires logging in: {booru['urls']['posts']}{post['id']}")
|
print(f"Skipping file that requires logging in: {booru['urls']['posts']}{post['id']}")
|
||||||
|
@ -151,6 +157,9 @@ def parse():
|
||||||
|
|
||||||
if not config['skipSave']:
|
if not config['skipSave']:
|
||||||
urllib.request.urlretrieve(post['url'], f"{config['output']}{post['filename']}")
|
urllib.request.urlretrieve(post['url'], f"{config['output']}{post['filename']}")
|
||||||
|
if os.path.exists(f"{config['images']}"):
|
||||||
|
shutil.copy(os.path.join(config['output'], post['filename']), os.path.join(config['images'], post['filename']))
|
||||||
|
|
||||||
print(f"Downloaded : {booru['urls']['posts']}{post['id']}")
|
print(f"Downloaded : {booru['urls']['posts']}{post['id']}")
|
||||||
|
|
||||||
if config['rateLimit']:
|
if config['rateLimit']:
|
||||||
|
|
Loading…
Reference in New Issue
Block a user