@@ -47,14 +47,14 @@ use url::Url;
4747#[ cfg( feature = "http" ) ]
4848use reqwest:: Client ;
4949#[ cfg( feature = "http" ) ]
50- use reqwest:: header:: UserAgent ;
50+ use reqwest:: header:: USER_AGENT ;
5151#[ cfg( feature = "http" ) ]
5252use reqwest:: StatusCode ;
5353#[ cfg( feature = "http" ) ]
5454use reqwest:: Response ;
5555
5656#[ cfg( feature = "http" ) ]
57- const USER_AGENT : & str = "robotparser-rs (https://crates.io/crates/robotparser)" ;
57+ const RP_USER_AGENT : & str = "robotparser-rs (https://crates.io/crates/robotparser)" ;
5858
5959/// A rule line is a single "Allow:" (allowance==True) or "Disallow:"
6060/// (allowance==False) followed by a path."""
@@ -255,8 +255,8 @@ impl<'a> RobotFileParser<'a> {
255255 /// Reads the robots.txt URL and feeds it to the parser.
256256 pub fn read ( & self ) {
257257 let client = Client :: new ( ) ;
258- let mut request = client. get ( self . url . clone ( ) ) ;
259- let request = request. header ( UserAgent :: new ( USER_AGENT . to_owned ( ) ) ) ;
258+ let request = client. get ( self . url . clone ( ) ) ;
259+ let request = request. header ( USER_AGENT , RP_USER_AGENT . to_owned ( ) ) ;
260260 let mut res = match request. send ( ) {
261261 Ok ( res) => res,
262262 Err ( _) => {
@@ -265,13 +265,13 @@ impl<'a> RobotFileParser<'a> {
265265 } ;
266266 let status = res. status ( ) ;
267267 match status {
268- StatusCode :: Unauthorized | StatusCode :: Forbidden => {
268+ StatusCode :: UNAUTHORIZED | StatusCode :: FORBIDDEN => {
269269 self . disallow_all . set ( true ) ;
270270 }
271- status if status >= StatusCode :: BadRequest && status < StatusCode :: InternalServerError => {
271+ status if status >= StatusCode :: BAD_REQUEST && status < StatusCode :: INTERNAL_SERVER_ERROR => {
272272 self . allow_all . set ( true ) ;
273273 }
274- StatusCode :: Ok => self . from_response ( & mut res) ,
274+ StatusCode :: OK => self . from_response ( & mut res) ,
275275 _ => { }
276276 }
277277 }
0 commit comments