uawdijnntqw1x1x1
IP : 18.226.4.132
Hostname : ns1.eurodns.top
Kernel : Linux ns1.eurodns.top 4.18.0-553.5.1.lve.1.el7h.x86_64 #1 SMP Fri Jun 14 14:24:52 UTC 2024 x86_64
Disable Function : mail,sendmail,exec,passthru,shell_exec,system,popen,curl_multi_exec,show_source,eval,open_base
OS : Linux
PATH:
/
home
/
sudancam
/
.
/
..
/
.
/
..
/
.
/
lib64
/
python2.7
/
.
/
robotparser.pyo
/
/
� �mec@s}dZddlZddlZdgZddd��YZddd��YZdd d��YZd ejfd ��YZdS(s< robotparser.py Copyright (C) 2000 Bastian Kleineidam You can choose between two licenses when using this package: 1) GNU GPLv2 2) PSF license for Python 2.2 The robots.txt Exclusion Protocol is implemented as specified in http://info.webcrawler.com/mak/projects/robots/norobots-rfc.html i����NtRobotFileParsercBsbeZdZdd�Zd�Zd�Zd�Zd�Zd�Zd�Z d �Z d �ZRS(ss This class provides a set of methods to read, parse and answer questions about a single robots.txt file. tcCs>g|_d|_t|_t|_|j|�d|_dS(Ni(tentriestNonet default_entrytFalsetdisallow_allt allow_alltset_urltlast_checked(tselfturl((s#/usr/lib64/python2.7/robotparser.pyt__init__s cCs|jS(s�Returns the time the robots.txt file was last fetched. This is useful for long-running web spiders that need to check for new robots.txt files periodically. (R (R ((s#/usr/lib64/python2.7/robotparser.pytmtime scCsddl}|j�|_dS(sYSets the time the robots.txt file was last fetched to the current time. i����N(ttimeR (R R((s#/usr/lib64/python2.7/robotparser.pytmodified)scCs/||_tj|�dd!\|_|_dS(s,Sets the URL referring to a robots.txt file.iiN(Rturlparsethosttpath(R R((s#/usr/lib64/python2.7/robotparser.pyR1s cCs�t�}|j|j�}g|D]}|j�^q"}|j�|j|_|jdkrkt|_n@|jdkr�t|_n%|jdkr�|r�|j |�ndS(s4Reads the robots.txt URL and feeds it to the parser.i�i�i�i�N(i�i�( t URLopenertopenRtstriptcloseterrcodetTrueRRtparse(R topenertftlinetlines((s#/usr/lib64/python2.7/robotparser.pytread6s cCsAd|jkr-|jdkr=||_q=n|jj|�dS(Nt*(t useragentsRRRtappend(R tentry((s#/usr/lib64/python2.7/robotparser.pyt _add_entryDscCsd}d}t�}x�|D]�}|d7}|s~|dkrPt�}d}q~|dkr~|j|�t�}d}q~n|jd�}|dkr�|| }n|j�}|s�qn|jdd�}t|�dkr|dj�j�|d<tj|dj��|d<|ddkrk|dkrN|j|�t�}n|j j |d�d}q�|ddkr�|dkr�|jj t|dt ��d}q�q�|ddkr�|dkr�|jj t|dt��d}q�q�qqW|dkr|j|�nd S( s�parse the input lines from a robots.txt file. We allow that a user-agent: line is not preceded by one or more blank lines.iiit#t:s user-agenttdisallowtallowN(tEntryR#tfindRtsplittlentlowerturllibtunquoteR R!t rulelinestRuleLineRR(R Rtstatet linenumberR"Rti((s#/usr/lib64/python2.7/robotparser.pyRMsN cCs�|jr tS|jrtStjtj|��}tjdd|j|j |j |jf�}tj|�}|s}d}nx-|j D]"}|j|�r�|j|�Sq�W|jr�|jj|�StS(s=using the parsed robots.txt decide if useragent can fetch urlRt/(RRRRRR-R.t urlunparseRtparamstquerytfragmenttquoteRt applies_tot allowanceR(R t useragentRt parsed_urlR"((s#/usr/lib64/python2.7/robotparser.pyt can_fetch�s cCs-djg|jD]}t|�d^q�S(NRs (tjoinRtstr(R R"((s#/usr/lib64/python2.7/robotparser.pyt__str__�s(t__name__t __module__t__doc__RR RRRR#RR>RA(((s#/usr/lib64/python2.7/robotparser.pyRs 3 R0cBs)eZdZd�Zd�Zd�ZRS(soA rule line is a single "Allow:" (allowance==True) or "Disallow:" (allowance==False) followed by a path.cCs;|dkr|rt}ntj|�|_||_dS(NR(RR-R9RR;(R RR;((s#/usr/lib64/python2.7/robotparser.pyR�s cCs|jdkp|j|j�S(NR(Rt startswith(R tfilename((s#/usr/lib64/python2.7/robotparser.pyR:�scCs|jrdpdd|jS(NtAllowtDisallows: (R;R(R ((s#/usr/lib64/python2.7/robotparser.pyRA�s(RBRCRDRR:RA(((s#/usr/lib64/python2.7/robotparser.pyR0�s R(cBs2eZdZd�Zd�Zd�Zd�ZRS(s?An entry has one or more user-agents and zero or more rulelinescCsg|_g|_dS(N(R R/(R ((s#/usr/lib64/python2.7/robotparser.pyR�s cCsjg}x'|jD]}|jd|dg�qWx*|jD]}|jt|�dg�q:Wdj|�S(NsUser-agent: s R(R textendR/R@R?(R trettagentR((s#/usr/lib64/python2.7/robotparser.pyRA�scCs]|jd�dj�}x=|jD]2}|dkr9tS|j�}||kr#tSq#WtS(s2check if this entry applies to the specified agentR4iR(R*R,R RR(R R<RK((s#/usr/lib64/python2.7/robotparser.pyR:�scCs.x'|jD]}|j|�r |jSq WtS(sZPreconditions: - our agent applies to this entry - filename is URL decoded(R/R:R;R(R RFR((s#/usr/lib64/python2.7/robotparser.pyR;�s(RBRCRDRRAR:R;(((s#/usr/lib64/python2.7/robotparser.pyR(�s RcBs#eZd�Zd�Zd�ZRS(cGs tjj||�d|_dS(Ni�(R-tFancyURLopenerRR(R targs((s#/usr/lib64/python2.7/robotparser.pyR�scCsdS(N(NN(R(R Rtrealm((s#/usr/lib64/python2.7/robotparser.pytprompt_user_passwd�scCs(||_tjj||||||�S(N(RR-RLthttp_error_default(R RtfpRterrmsgtheaders((s#/usr/lib64/python2.7/robotparser.pyRP�s (RBRCRRORP(((s#/usr/lib64/python2.7/robotparser.pyR�s (((( RDRR-t__all__RR0R(RLR(((s#/usr/lib64/python2.7/robotparser.pyt<module>s �$
/home/sudancam/./.././.././lib64/python2.7/./robotparser.pyo