Merge branch 'main' into javalsai-changes

This commit is contained in:
javalsai 2024-10-20 21:31:03 +02:00
commit 235e13230b
Signed by: javalsai
SSH Key Fingerprint: SHA256:3G83yKhBUWVABVX/vPWH88xnK4+ptMtHkZGCRXD4Mk8
7 changed files with 98 additions and 122 deletions

1
.gitignore vendored
View File

@ -1 +1,2 @@
/target /target
/downloads

59
Cargo.lock generated
View File

@ -833,16 +833,6 @@ version = "0.4.14"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "78b3ae25bc7c8c38cec158d1f2757ee79e9b3740fbc7ccf0e59e4b08d793fa89" checksum = "78b3ae25bc7c8c38cec158d1f2757ee79e9b3740fbc7ccf0e59e4b08d793fa89"
[[package]]
name = "lock_api"
version = "0.4.12"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "07af8b9cdd281b7915f413fa73f29ebd5d55d0d3f0155584dade1ff18cea1b17"
dependencies = [
"autocfg",
"scopeguard",
]
[[package]] [[package]]
name = "log" name = "log"
version = "0.4.22" version = "0.4.22"
@ -973,29 +963,6 @@ version = "2.2.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f38d5652c16fde515bb1ecef450ab0f6a219d619a7274976324d5e377f7dceba" checksum = "f38d5652c16fde515bb1ecef450ab0f6a219d619a7274976324d5e377f7dceba"
[[package]]
name = "parking_lot"
version = "0.12.3"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f1bf18183cf54e8d6059647fc3063646a1801cf30896933ec2311622cc4b9a27"
dependencies = [
"lock_api",
"parking_lot_core",
]
[[package]]
name = "parking_lot_core"
version = "0.9.10"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "1e401f977ab385c9e4e3ab30627d6f26d00e2c73eef317493c4ec6d468726cf8"
dependencies = [
"cfg-if",
"libc",
"redox_syscall",
"smallvec",
"windows-targets",
]
[[package]] [[package]]
name = "percent-encoding" name = "percent-encoding"
version = "2.3.1" version = "2.3.1"
@ -1083,15 +1050,6 @@ dependencies = [
"tokio", "tokio",
] ]
[[package]]
name = "redox_syscall"
version = "0.5.7"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "9b6dfecf2c74bce2466cabf93f6664d6998a69eb21e39f4207930065b27b771f"
dependencies = [
"bitflags",
]
[[package]] [[package]]
name = "regex" name = "regex"
version = "1.11.0" version = "1.11.0"
@ -1253,12 +1211,6 @@ dependencies = [
"windows-sys 0.59.0", "windows-sys 0.59.0",
] ]
[[package]]
name = "scopeguard"
version = "1.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "94143f37725109f92c262ed2cf5e59bce7498c01bcc1502d7b9afe439a4e9f49"
[[package]] [[package]]
name = "security-framework" name = "security-framework"
version = "2.11.1" version = "2.11.1"
@ -1332,15 +1284,6 @@ version = "1.3.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "0fda2ff0d084019ba4d7c6f371c95d8fd75ce3524c3cb8fb653a3023f6323e64" checksum = "0fda2ff0d084019ba4d7c6f371c95d8fd75ce3524c3cb8fb653a3023f6323e64"
[[package]]
name = "signal-hook-registry"
version = "1.4.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a9e9e0b4211b72e7b8b6e85c807d36c212bdb33ea8587f7569562a84df5465b1"
dependencies = [
"libc",
]
[[package]] [[package]]
name = "slab" name = "slab"
version = "0.4.9" version = "0.4.9"
@ -1463,9 +1406,7 @@ dependencies = [
"bytes", "bytes",
"libc", "libc",
"mio", "mio",
"parking_lot",
"pin-project-lite", "pin-project-lite",
"signal-hook-registry",
"socket2", "socket2",
"tokio-macros", "tokio-macros",
"windows-sys 0.52.0", "windows-sys 0.52.0",

View File

@ -1,13 +1,13 @@
[package] [package]
name = "r34-scraper" name = "r34-scraper"
version = "0.1.0" version = "1.0.0"
edition = "2021" edition = "2021"
[dependencies] [dependencies]
async-std = "1.13.0" async-std = "1"
clap = { version = "4.5.20", features = ["derive"] } clap = { version = "4", features = ["derive"] }
futures = "0.3.31" futures = "0"
indicatif = "0.17.8" indicatif = "0"
regex = "1.11.0" regex = "1"
reqwest = { version = "0.12.8", features = ["blocking"] } reqwest = { version = "0", features = ["blocking"] }
tokio = { version = "1", features = ["full"] } tokio = { version = "1", features = ["macros", "rt-multi-thread"] }

View File

@ -4,7 +4,7 @@ a scraper that well scrapes r34
## note ## note
this thing is still not completed, it only gathers links, it doesnt download things yet this program is pretty much complete, although i am planning to add a few extra features.
## example usage image ## example usage image
![example image](./image.png) ![example image](./image.png)

BIN
image.png

Binary file not shown.

Before

Width:  |  Height:  |  Size: 79 KiB

After

Width:  |  Height:  |  Size: 122 KiB

View File

@ -11,11 +11,15 @@ pub struct Args {
pub user_agent: String, pub user_agent: String,
/// Tags to search for /// Tags to search for
#[arg(short, long)] #[arg(short, long, required = true)]
pub tags: Option<Vec<String>>, pub tags: Vec<String>,
/// Page to start scraping from
#[arg(short, long, default_value_t = 1)]
pub page: usize,
/// Async jobs to use for fetching /// Async jobs to use for fetching
#[arg(short, long, default_value = "4")] #[arg(short, long, default_value_t = 4)]
pub jobs: usize, pub jobs: usize,
/// Delay for rate-limits (ms) /// Delay for rate-limits (ms)

View File

@ -1,35 +1,31 @@
#![feature(async_closure, iter_intersperse)] #![feature(async_closure)]
pub mod args; pub mod args;
use clap::Parser; use clap::Parser;
use futures::{stream, StreamExt}; use futures::{stream, StreamExt};
use indicatif::ProgressBar;
use regex::Regex; use regex::Regex;
use reqwest::Client; use reqwest::Client;
use tokio::time::{sleep, Duration}; use tokio::time::{sleep, Duration};
use std::io::Write;
use std::process::ExitCode; use std::process::ExitCode;
const BAR_LENGTH: u64 = 8;
#[tokio::main] #[tokio::main]
async fn main() -> ExitCode { async fn main() -> ExitCode {
let args = args::Args::parse(); let args = args::Args::parse();
let tags = args.tags.unwrap_or_else(|| { let uri_tags = &args.tags.join("+");
println!("which tags do you want to scrape? ex: 1girls 1boys yomama"); let _ = std::fs::create_dir(uri_tags);
let tags_binding = std::io::stdin().lines().next().unwrap().unwrap();
tags_binding
.split(' ')
.filter(|item| !item.is_empty())
.map(|item| item.to_owned())
.collect()
});
let uri_tags: String = tags.into_iter().intersperse(String::from("+")).collect();
let client = Client::builder() let client = Client::builder()
.user_agent(&args.user_agent) .user_agent(&args.user_agent)
.build() .build()
.unwrap(); .unwrap();
for page in 0.. { for page in args.page - 1.. {
println!("now scraping page {}", page + 1); println!("now scraping page {}", page + 1);
println!( println!(
"https://rule34.xxx/index.php?page=post&s=list&tags={uri_tags}&pid={}", "https://rule34.xxx/index.php?page=post&s=list&tags={uri_tags}&pid={}",
@ -53,75 +49,65 @@ async fn main() -> ExitCode {
}; };
let mut urls = post_html(&client).await; let mut urls = post_html(&client).await;
let mut wait_time = 5000;
if urls.is_empty() { if urls.is_empty() {
for reconnection_attempts in 0..4 { let mut reconnection_attempts = 0;
println!("no urls found, retrying in {} seconds...", wait_time / 1000); loop {
sleep(Duration::from_millis(wait_time)).await; println!("no urls found, retrying in 5 seconds...");
sleep(Duration::from_millis(5000)).await;
urls = post_html(&client).await; urls = post_html(&client).await;
if !urls.is_empty() { if !urls.is_empty() {
println!("urls found! continuing..."); println!("urls found! continuing...");
break; break;
} }
if reconnection_attempts == 3 { reconnection_attempts += 1;
println!("no urls found in 4 attempts, exiting..."); if reconnection_attempts == 12 {
println!("no urls found in 1 minute, exiting...");
return ExitCode::FAILURE; return ExitCode::FAILURE;
} }
wait_time += 5000;
} }
} }
let multi_prog = indicatif::MultiProgress::new(); let multi_prog = indicatif::MultiProgress::new();
let urls_ammount = urls.len(); let urls_amount = urls.len();
let responses = stream::iter(urls.into_iter().enumerate()) let responses = stream::iter(urls.into_iter().enumerate())
.map(|(i, url)| { .map(|(i, url)| {
let i = i + 1; let i = i + 1;
let client = &client; let client = &client;
let this_bar = indicatif::ProgressBar::new_spinner(); let this_bar = indicatif::ProgressBar::new(BAR_LENGTH);
this_bar.enable_steady_tick(Duration::from_millis(50)); this_bar.set_style(indicatif::ProgressStyle::with_template("[{bar}] {msg}").unwrap().progress_chars("=> "));
let this_prog = multi_prog.insert(i, this_bar); let this_bar = multi_prog.insert(i, this_bar);
async move { async move {
// "thread" // "thread"
loop { loop {
this_prog.set_message(format!("\x1b[30m[{i: >4}/{urls_ammount}] \x1b[36mscraping {url:?}\x1b[0m")); this_bar.set_message(format!("\x1b[37m[{i: >4}/{urls_amount}] \x1b[36mscraping {url:?}\x1b[0m"));
let resp = client.get(&url).send().await.unwrap(); let resp = client.get(&url).send().await.unwrap();
match extract_img_url(&resp.text().await.unwrap()) { if let Ok(img_url) = extract_img_url(&resp.text().await.unwrap()) {
Ok(img_url) => {
if img_url.is_empty() { if img_url.is_empty() {
this_prog.abandon_with_message(format!( this_bar.abandon_with_message(format!(
"\x1b[30m[{i: >4}/{urls_ammount}] \x1b[1;31mimage url not found\x1b[0m" "\x1b[37m[{i: >4}/{urls_amount}] \x1b[1;31mimage url not found\x1b[0m"
)); ));
} else { } else {
this_prog.finish_with_message(format!( download_file(&img_url, this_bar, i, urls_amount, uri_tags).await;
"\x1b[30m[{i: >4}/{urls_ammount}] \x1b[32mfound image url: {img_url}\x1b[0m"
));
} }
break img_url; break;
} }
Err(_) => {
this_prog this_bar
.set_message(format!( .set_message(format!(
"\x1b[30m[{i: >4}/{urls_ammount}] \x1b[31mratelimited, retrying after {}ms\x1b[0m", "\x1b[37m[{i: >4}/{urls_amount}] \x1b[31mratelimited, retrying after {}ms\x1b[0m",
args.delay.as_millis()) args.delay.as_millis())
); );
tokio::time::sleep(args.delay).await; tokio::time::sleep(args.delay).await;
continue;
}
}
} }
} }
}) })
.buffered(args.jobs); .buffered(args.jobs);
let _ = responses.for_each(|_| async {}).await; let _ = responses.for_each(|()| async {}).await;
} }
return ExitCode::SUCCESS; ExitCode::SUCCESS
} }
fn extract_urls(html: &str) -> Vec<String> { fn extract_urls(html: &str) -> Vec<String> {
@ -145,3 +131,47 @@ fn extract_img_url(html: &str) -> Result<String, &'static str> {
Ok(String::new()) Ok(String::new())
} }
} }
async fn download_file(
img_url: &str,
this_bar: ProgressBar,
i: usize,
urls_amount: usize,
uri_tags: &str,
) {
let args = args::Args::parse();
let file_name = Regex::new(r"[^/]+$")
.unwrap()
.find(img_url)
.map(|m| m.as_str())
.unwrap();
let file_path = uri_tags.to_owned() + "/" + file_name;
let mut file = if std::fs::File::open(&file_path).is_ok() {
this_bar.finish_with_message(format!(
"\x1b[37m[{i: >4}/{urls_amount}] \x1b[33m{file_name} exists, skipping...\x1b[0m"
));
return;
} else {
std::fs::File::create(file_path).unwrap()
};
let mut res = Client::new()
.get(img_url)
.header("User-Agent", &args.user_agent)
.send()
.await
.unwrap();
let file_length = res.content_length().unwrap();
let mut written = 0;
while let Some(img_chunk) = res.chunk().await.unwrap() {
file.write_all(&img_chunk).unwrap();
written += img_chunk.len();
this_bar.set_position((written as f64 / file_length as f64 * BAR_LENGTH as f64) as u64);
}
this_bar.finish_with_message(format!(
"\x1b[37m[{i: >4}/{urls_amount}] \x1b[32mdownloaded {img_url}\x1b[0m"
));
}