rust写个简单爬虫

[dependencies]
scraper = "0.12.0"
reqwest = { version = "0.11.10", features = ["blocking", "json"] }
surf = "2.3.2"
tokio = { version = "1.17.0", features = ["full"] }
futures = "0.3.21"
use futures::future::join_all;
use std::env;
use scraper::{Html, Selector};
 
async fn fetch_path(path:String) -> surf::Result<String>{
    let mut back_string = String::new();
    match surf::get(&path).await {
        Ok(mut response) => {
            match response.body_string().await{
                Ok(text) =>{
                    back_string = format!("{}",text)
                }
                Err(_) => {
                    println!("Read response text Error!")
                }
            };
        }
        Err(_) => {
            println!("reqwest get Error!")
        }
    }
    Ok(back_string)
}
 
 
#[tokio::main]
async fn main() -> surf::Result<()>{
    let stdin = env::args().nth(1).take().unwrap();
    let paths = vec![stdin.to_string(),];
    let result_list = join_all(paths.into_iter().map(|path|{
        fetch_path(path)
    })).await;
 
    let mut list_string:Vec<String> = vec![];
    for ele in result_list.into_iter(){
        if ele.is_ok(){
            list_string.push(ele.unwrap())
        }else {
            return Err(ele.unwrap_err())
        }
    }
    let v = list_string.get(0).take().unwrap();
   //  println!("{}",v);
    let fragment = Html::parse_fragment(v);
    let ul_selector = Selector::parse("script").unwrap();
 
    for element in fragment.select(&ul_selector) {
        println!("{}",element.inner_html());
    }   
    // println!("请求输出:{:?}",list_string);
    Ok(())
}
cargo run -- https://www.baidu.com/

您还可以看看它们